var/home/core/zuul-output/0000755000175000017500000000000015155014264014530 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015155040722015472 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log.gz0000644000175000017500000444503415155040602020262 0ustar corecoreAikubelet.log_o[;r)Br'o -n(!9t%Cs7}g/غIs,r.k9Gfͅ|YʋI_翪|mvſFެxۻf+ovpZj!Kޒ/h3_.gSeq5v(×_~^ǿq]n>߮}+ԏbś E^"Y^-Vۋz7wH׋0g"ŒGǯguz|ny;#)a "b BLc?^^4[ftlR%KF^j 8DΆgS^Kz۞_W#|`zIlp_@oEy5 fs&2x*g+W4m ɭiE߳Kfn!#Šgv cXk?`;'`&R7߿YKS'owHF6":=3Ȑ 3xҝd){Ts}cZ%BdARO#-o"D"ޮrFg4" 0ʡPBU[fi;dYu' IAgfPF:c0Ys66q tH6#.`$vlLH}ޭA㑝V0>|J\Pg\W#NqɌDSd1d9nT#Abn q1J# !8,$RNI? j!bE"o j/o\E`r"hA ós yi\_.!=A(%Ud,QwC}F][UVYE NQGn0Ƞɻ>.ww}(o./WY<͉#5O H 'wo6C9yg|O~ €' S[q?,!yq%a:y<\tunL h%$Ǥ].v y[W_` \r/Ɛ%aޗ' B.-^ mQYd'xP2ewEڊL|^ͣrZg7n͐AG%ʷr<>; 2W>h?y|(G>ClsXT(VIx$(J:&~CQpkۗgVKx*lJ3o|s`<՛=JPBUGߩnX#;4ٻO2{Fݫr~AreFj?wQC9yO|$UvވkZoIfzC|]|[>ӸUKҳt17ä$ ֈm maUNvS_$qrMY QOΨN!㞊;4U^Z/ QB?q3En.اeI"X#gZ+Xk?povR]8~깮$b@n3xh!|t{: CºC{ 8Ѿm[ ~z/9آs;DPsif39HoN λC?; H^-¸oZ( +"@@%'0MtW#:7erԮoQ#% H!PK)~U,jxQV^pΣ@Klb5)%L%7׷v] gv6دϾDD}c6  %T%St{kJ_O{*Z8Y CEO+'HqZY PTUJ2dic3w ?YQgpa` Z_0΁?kMPc_Ԝ*΄Bs`kmJ?t 53@հ1hr}=5t;at 9:I_|AאM'NO;uD,z҄R K&Nh c{A`?2ZҘ[a-0V&2D[d#L6l\Jk}8gf) afs'oIf'mf\>UxR ks J)'u4iLaNIc2qdNA&aLQVD R0*06V۽棬mpھ*V I{a 0Ҟҝ>Ϗ ,ȓw`Ȅ/2Zjǽ}W4D)3N*[kPF =trSE *b9ē7$ M_8.Ç"q ChCMAgSdL0#W+CUu"k"圀̲F9,,&h'ZJz4U\d +( 7EqڏuC+]CEF 8'9@OVvnNbm: X„RDXfיa }fqG*YƩ{P0K=( $hC=h2@M+ `@P4Re]1he}k|]eO,v^ȹ [=zX[tꆯI7c<ۃ'BdIc*Qqk&60XdGY!D ' @{!b4ִ s Exb 5dKߤKߒ'&YILұ4q6y{&G`%$8Tt ȥ#5vGVO2Қ;m#NS8}d0Q?zLV3\LuOx:,|$;rVauNjk-ؘPꐤ`FD'JɻXC&{>.}y7Z,).Y톯h7n%PAUË/_xY~7.w47mnjGgG{9_e552s4IG^ۃn󨔖I@[ tWv Fyw9J֥WmN^<.eܢMρ'JÖŢո%gQ=p2YaI"&ư%# yCùXz!bm5uAߙXC90뼯nNNXYt\oP@gOV ]cӰJ:^q';E=-dZB4']a.QO:#'6RE'E3 */HAYk|z|ءPQgOJӚ:ƞŵ׉5'{#ޢ1c qw zǽ0 2mK:ȔsGdurWMF*֢v|EC#{usSMiI S/jﴍ8wPVC P2EU:F4!ʢlQHZ9E CBU)Y(S8)c yO[E}Lc&ld\{ELO3芷AgX*;RgXGdCgX JgX2*Ъ3:O7ǭ3ږA :}d,ZByXϯ&Ksg3["66hŢFD&iQCFd4%h= z{tKmdߟ9i {A.:Mw~^`X\u6|6rcIF3b9O:j 2IN…D% YCUI}~;XI썋Fqil><UKkZ{iqi :íy˧FR1u)X9 f΁U ~5batx|ELU:T'Tស[G*ݧ ؽZK̡O6rLmȰ (T$ n#b@hpj:˾ojs)M/8`$:) X+ҧSaۥzw}^P1J%+P:Dsƫ%z; +g 0հc0E) 3jƯ?e|miȄ{g6R/wD_tՄ.F+HP'AE; J jK=! I,($F{ձ7*Oy 6EK( EF #31J8mN .TTF9㕴/5~RxCe,&v3,JE- ZF5%Da,Gܠ*qI@qlG6s푻jÝ$ >8ȕ$eZ1j[h0SH,qf<"${/ksBK}xnwDb%M6:K<~̓9*u᛹Q{FЖt~6S#G1(zr6<ߜ!?U\(0EmG4 4c~J~]ps/9܎ms4gZY-07`-Id,9õ԰t+-b[uemNi_󈛥^g+!SKq<>78NBx;c4<ニ)H .Pd^cR^p_G+E--ۥ_F]a|v@|3p%kzh|k*BBRib\J3Yn|뇱[FfP%M:<`pz?]6laz5`ZQs{>3ư_o%oU׆]YL`d[9ɃO>z+ϟMn?!ԫ5H&=JkܓhkB\LQ"<LxeLo4l_m24^3.{oɼʪ~75/nQοs d|pxu\uw?=QR -Mݞίk@Pc n1æ*m$=4Dbs+J \EƄզ}@۶(ߐ/ۼ𹫘qݎt7!2~.볿j HmE]j `7ruuŨԀ![Z !iHlf[7Ua6BEZEkuS#q * CƂ lu" yo6"3껝I~flQ~NCBX`]ڦÞhkXO _-Qy2$?T3ͤEZ긊mۘ$XD.bͮW`AީClСw5/lbl[N*t*@56."D/< {Dۥ sLxZn$N(lYiV =?_e^0)?]{ @| 6+#gPX>Bk2_@L `CZ?z3~ }[ tŪ)۲-9ֆP}b&x Uhm._O 4m6^^osVЦ+*@5Fˢg'!>$]0 5_glg}릅h:@61Xv` 5DFnx ˭jCtu,R|ۯG8`&ו:ݓ3<:~iXN9`2ŦzhѤ^ MW`c?&d.'[\]}7A[?~R6*.9t,綨 3 6DFe^u; +֡X< paan}7ftJ^%0\?mg5k][ip4@]p6Uu|܀|Kx6خQU2KTǺ.ȕPQVzWuk{n#NWj8+\[ ?yiI~fs[:.۽ '5nGppH? 8>X+m7_Z`V j[ s3nϏT=1:T ?<= pDCm3-b _F(/f<8sl, 0۬Z"X.~b٦G3TE.֣eմi<~ik[m9뀥!cNIl8y$~\T B "2j*ҕ;ێIs ɛqQQKY`\ +\0(FęRQ hN œ@埝n|Vo|6 8~J[,o%l%!%tyNL`'žeVVޖ~;BLv[n|viPjbMeO?!hEfޮ])4 ?KN1o<]0Bg9lldXuT ʑ!Iu2ʌnB5*<^I^~G;Ja߄b&VY+yn~F8I !6c> "NqK$2$ Ri ?2,ᙌEK@-V3ʱd:/4Kwm2$'dW<qIE2Ľ)5kJҼMЌ DR3csf6rRSr[I߽ogCc;S5ׂdKZ=M3դ#F;SYƘK`K<<ƛ G׌MU.APf\M*t*vw]xo{:l[n=`smFQµtxx7/W%g!&^=SzDNew(æ*m3D Bo.hI"!A6:uQզ}@j=Mo<}nYUw1Xw:]e/sm lˣaVۤkĨdԖ)RtS2 "E I"{;ōCb{yex&Td >@).p$`XKxnX~E膂Og\IGֻq<-uˮ◶>waPcPw3``m- } vS¢=j=1 W=&;JW(7b ?+]m`F3 W((!5F-9]dDqL&RΖd}})7 k11 K ;%v'_3 dG8d t#MTU']h7^)O>?~?_ȿM4ə#a&Xi`O}6a-xm`8@;of,![0-7 4f kUy:M֖Esa./zʕy[/ݩqz2¼&'QxJE{cZ7C:?pM z*"#窾+ HsOt۩%͟A498SwWv|jNQ=-[ӓI+mj(^>c/"ɭex^k$# $V :]PGszyH(^jJ=䄸-m!AdEږG)շj#v;#y/hbv BO Iߒ {I7!UՆGIl HƗbd#HAF:iI }+2kK:Sov3b:1)'A6@\2X#Ih9N ̢t-mfeF;gUаQ/ .D%ES*;OLRX[vDb:7a}YF30H #iSpʳ]'_'ĕ -׉6tfЮ$zͪO_sYq+q艻*vzh5~Yy;,DiYTP;o./~^.6+zZFD& m@WXe{sa 2tc^XS?irG#^ŲDI'H_Ȯ;RJ&GT.Kwj;of¬zHmmS2ҒN'=zAΈ\b*K ڤUy""&D@iS=3&N+ǵtX^7ǩX"CA⥎å+4@{D/-:u5I꾧fY iʱ= %lHsd6+H~ Δ,&颒$tSL{yєYa$ H>t~q؈xRmkscXQG~gD20zQ*%iQI$!h/Vo^:y1(t˥C"*FFDEMAƚh $ /ɓzwG1Ƙl"oN:*xmS}V<"dH,^)?CpҒ7UΊ,*n.֙J߾?Ϲhӷƀc"@9Fў-Zm1_tH[A$lVE%BDI yȒv $FO[axr Y#%b Hw)j4&hCU_8xS] _N_Z6KhwefӞ@蹃DROo X"%q7<# '9l%w:9^1ee-EKQ'<1=iUNiAp(-I*#iq&CpB.$lٴާt!jU_L~Tb_,֪r>8P_䅱lw1ù=LAЦz38ckʖYz ~kQRL Q rGQ/ȆMC)vg1Xa!&'0Dp\~^=7jv "8O AfI; P|ޓܜ 8qܦzl5tw@,Mڴg$%82h7էoaz32h>`XT>%)pQ}Tgĸ6Coɲ=8f`KݜȆqDDbZ:B#O^?tNGw\Q.pPO @:Cg9dTcxRk&%])ў}VLN]Nbjgg`d]LGϸ.yҵUCL(us6*>B 2K^ sBciۨvtl:J;quӋkKϮ듃ԁ6Y.0O۾'8V%1M@)uIw].5km~Ҷ綝R(mtV3rșjmjJItHڒz>6nOj5~IJ|~!yKڮ2 h 3x}~ے4WYr9Ts] AA$ұ}21;qbUwRK #}u'tLi'^Y&,mCM)eu㠥Ѻ\a}1:V1zMzT}R,IA e<%!vĉq|?mtB|A ?dXuWLGml?*uTC̶V`FVY>ECmDnG+UaKtȃbeb筃kݴO~f^⊈ 8MK?:mM;ߵoz+O~e3݌ƺ(ܸf)*gCQE*pp^~x܃`U'A~E90t~8-2S󹞙nk56s&"mgVKA: X>7QQ-CDC'| #]Y1E-$nP4N0#C'dvܸȯ.vIH"ŐR ;@~y>Kv{) 9AG ćͩ$.!б~N8i"1KФ\L7/,U@.ڮO?mُa ې!rGHw@56DǑq LA!&mYJ*ixz2*{_;IYJXFfQ* 0kA".mݡ"3`Rd1_u6d逖`7xGMf}k/⨼0Κ_pLq7k!dTm|_>9|dUA"{!$jKx E$K3hN(tÊ-#v#O N, 9g80Ǭ&VdӞ5W1!1KYd`,-*&>F~⯰&jb.~cNk BL_OG]Bv.A|'qT(Ol.' 4IE|@Iі)<-p JkQm1 `qacܗVc?)cl*&<}P媠E{-sVU>߇GUt\+n3X]Byoz)li$2cPs6D>TE-n# rve{椱I |p)U݋7yJw&PzDgi xs  xh\L r Ѥo Zt(I >|$>tnMdэo'$0&*m.)HzzBvU0h} -_.7^nya+Cs 6K!x^' ^7 l 2Jj.S֔(*CjaS:vp/N6I*x8"EȿQa[qVM/)fpOj4r!:V_IG^nILVG#A7jF};qPU嗈M9VS;a+Ӧ8E8zmMs*7NM~@6 ' 8jp*:'SOANa0rӍ?DT%l)gvN}JT(Ȋqm|dc+lQai,|Dߟ|, d#EjZܴv]pEO7}&gbXԈedKX :+Z|p8"81,w:$TiVD7ֶ]cga@>\X=4OZSܿ* %xccDa.E h :R.qɱMu$ơI8>^V Y. ,BLq~z&0o- ,BLqfx9y:9244ANb n\"X>Y`bb*h%)(*_Gra^ sh6"BzƾH( ."e)B QlKlXt҈t9՚$ضz]'.!-r"1MCĦʸ"66pE{ =CNc\ESD[T4azry !5yY~ :3;Y[Iȧ q:i Ǟ/"8Wxç,vܰtX-LE7 |-D`JLw9|fb>4Nu ߏ3ap5k_JA+A.A~ C~`[KaQ-Ģn9ѧf q:cT >to^ X]j?-ȇlCf0hM`~ ó}0W@o  K[{d+`ze"l |d;L2k%x90ݙ^Oe ]nHfS+.4<#/5߁ݛǪ0q,7FeV/!; 瓠 Li% z}ɯww"O-]J`sdN$@"J`Y13K/9`VTElsX|D^c%֯T][$m;ԝ!,Z5f`XFzȁ=nrSA8; P=uY}r/27OUa%~0;үM3Tu ȩ*'3IC~LG,?.?C3tBYpm_g.~>3ʄ55[c&-Wgy_jVo,?s w*n\7[cpMY<~/"˘oV܉T6nn \_ߋV_}Z=k-nn sn.*upw pX\_ U-C_wS!|q?E-S_w$-#9?mh{R 4ѭm_9p -h2 dֲ 1"j {]]Nk"䁖%5'32hDz O\!f3KX0kIKq"H~%.b@:Oec6^:V8FDza5H`:&Q5 ^hI8nʁu EA~V O8Z-mYO!tO֠υ9G`6qmJc,Qh: ݢKNw2taC0Z' O > f-`:F_Ѫ2)sCj1THɩhS-^p b~?.>, `0!E%ҏ:H =VՑӄ| Ć.lL t1]}r^nʂI-|i*'yW='W6M$oeB,޳X$I6c>EK# 15ۑO2Jh)8Vgl0v/eNEU"Ik dRu˜6Uǖ xs%P ع omWl҈sApX!^ Ɩgv{Xn|$̇d`>1Ljn떚F+B9l"UP۾u2Ja>0c0Vvގj$]p^M+f~@9{bOe@7ȱ^%u~-B竟} |23 Z.`oqD>t@N _7c$h3`lg\)[h+pHBr^J |r\8czEnv@qZbRT1e8V Scc6:$[|a.fpU`ZR֩bKgTlѩynۢ, "1LӰW&jDkM~# (C>ϭQ3{ߤ%EN;?P%ٱm -{2k 8Vbv"wŏݙmn&O1^'}plM)0\n ή ?Cֲa9H] lX9^vCο -vd+OUgRy2Я\ B0!% #>bJPUck\Ul'F瘏Y4Ew`[x٘p,>9V"R1I>bJ` UL'5m1Ԥ:t6I >jz(:W֪Ƹ)!fꠗe[XLE4atGS1px#S]MF˦NJPYDX%ܠꡗhl}i9f?q>b-E'V"mNf""ŦK9kǍ-vU #`uVi<s)/=r=nlӗЩsdLyVIUI':4^6& t,O669Ȁ,EʿkڍfC58$5?DX 4q]ll9W@/zNaZf% >Ę_"+BLu>'Ɩ=xɮ[⠋X((6I#z)2S zp&m?e8 "(O+:Y EaSD]<^(]В|Ǚ8"oRs?]\McZ0ϕ!1hKS`h0O{!L-w]ln2&0Ǚ'0=.T4G7! H/ͺ|@lX)+{{^s1V63 ۗI"*al NJ`Q8B\pup6_3XqCXznL9:{o qcuו8`n{ave=}OR9~yL Z1=W8>É R$|L ]OfJl˪VVg:lDԒ͢Zu[kWۗw{{7st08`J0ꨴU1|z:9dX)z2!S:'q9| 76"Q;D*04Zٚ ?V¼r8/G:T6Fw/ɚ~h?lUc3MکEen壹n\殸,˛_uu.Jssu/*47U0)l?R_^Uon̝f-nnZTeuu nn/*0׷տ·sHH?Et _I`[>>0ւcz/Adh.$@bѨLtT=cKGX nݔ͆!`c|Lu_~ǴT?crO e9d ljB?K_z>p%'3JQK-͗R>KkΤOq,*I|0]Sj%|-Ԟ = Ʃ%>H&t;9`>$& nIdE Ͻq*nŘʰNҁV_+mI/K(l 7kSY,RGKmȾ*LުA ]] >mCst 45q]8N! #iʋt{uൺޤB s景Bv'㱩΅n^.trƖnx6kͣ,:zX{1ƀI\oUym\%B O=˵uwX}+i'dn똿DrO$T讼eXbC|Y~V3Rd2^k"+KXoR~W9ovKV⨞ =3&=26wjŒfg*$V ^vWyx;B0y\zXj(Q,f#]<ˑfYF`~$e;Lpt#7=kL:J0h%kOguu:H$q:ugq 蟧s.tr덜a^w}z^q6 ) :oxx)3&9 oo6:>u4{Yg(>7<<ӟ3h &6h@|l?}KƐ}f{-Ty"\qbg= Iʼn\ 'aEDȲ kػh_9x>T~~sSm̧GCopJyŢ=}x0) D@B7eg7FE h[ɖۖ iLghmsCI~FW'e]4zM1Ҩ|ɮ|2 sXGd=^CEʧ\I)tA31^j_AC+(lSA ,l),Q5LDI2MWMsY%[0m&j]ʳme"F<h;mʩm6@Ve (iMbTB$%M5[+2Hz /KE w/p kCGna!2H<憔T>]ֽk+QT9iwfXo}"hlK R JkjLB@ QpT~gʳ8U5eBbmmN cEU$ ̱Դqx{ caDRLm6ZJƁf {[nm-_h(TyysfZ3vTM5҈ ] NݤSW8Cm)1ͱaA\mU%\(DZ]_=r ~W8,gaN!SoUm_d+IŘmN)kIn1U0S }HPb0uczݣ6խiIcrՌ{mYL&  v"_M#7# u=; ,+t].eb*O]zlBX78f[f%9`>[O}U7-ɰ^$Sޖܘb*k^ mbLa萖g*:Dd?!k \p&EtLJnHwe\{w0|}t.sDz-"ȟ 5:JNu&l e>ѽC?誊`ëK]Eb]nwWiqY8Hֵ,%2>醱+KeT+xizEY0bX "wU,c+&`hdĝ(Ȣ][opAWu),j4yٌÃ|#xOv{}0Oqc,O3y,7&b)sư1g]EykN}7B_"b.R@uEnO{,Qha_/j;,E2Kwn>&]Id&ji8ve>0x;08Y9H%;+[S"EMK;员U=@[QgqDA'ywˣ')0mj@i1M ^ֲMM0Ȧ8gЦCo 6u;Lmj`{n aG~88LCL{z}xi0e II5nXyN 6I$E# M@H53P |t$|=<4bH &L1ޒYdTbmTKVA*?K'Ags?@{ ٘4KXFkf]WSVTh]gTE/e nJAقR6) ^O3j UOhpjZMLtG@E6CёpizCO Ų(nG d[uˋzC( <,e]Ѩ.vp;$>^T \G=z`}DaS{E嘡LdG!4X'0̮9$9N|VX*sUt ar7_sV.OVcX4&xFl(vyB")l/= Xbix2pKstʈ,HC[nje|1fPllqOw+0 iqҶτV.gf$:|:48D҂U=Vy Ib4h꥚?az]=|CIbF=&7#Or Iz=N"QmGi{W%( J<8pTސ!BGM2{In6]wD+O+Ix7\EHG^L)|r=|2d{o`5'\sUpWu>׹؞X7:Xu>օ 0֍2jf2Cԫ<Œ 4w1gi|3Fm F^{y/ !!iacc3&MI^ ?`NȻ팰.W%n^:;Opx-)uthea6@aw%]nn*ORo>wOGyxቭlvۡx~4A}r + ՟)SV w'zyqs9u0 GQ1 g|n۽vQy"^"mbG2xCG[=߉{^l 4(Q6mTN~\zs%߁@ &~ xZ稬.^SJ/)>+lXnSW݃Blb7KcϖQϡS=PH@ihz6l}.Tqq$w?- G#S} |F[ߗh&6V!4l(]d/f 2\9n$#VmzClMBmp%u}LqJi жAN g/[]~#([ (]P4A#eVl:eTZeVSWN5}j>rw]*-AXiGJw^AP{wA j?RP{AŪbA *)APgUPgAu&HAuWuw]PiAPoUPoA&HAWw]PiA`U`A &hHA-.咛9BI,L)48HVz|;NKx1/p6H Tl&2/`uk&_9/s<>=)ͣ##^ˣd4wEI"%xp^&-6n#*$wѸӞ^ߡ}O W`8yM}Y?9W(ʔͦr+[]P(Qr͠h0͕HitY xŤV1"&],X)0<F bXd0+xJRЕZ$\?2MrX&h-v\WA} MԭX1x^Q]a FYxW1dʔ\Ăe!|?xN(C"TZTwpoʷPTV/Yb~`?)o+߿?9DZ {A-kW(疴QXrJ!¥ )exk'i`T?JGb u\3Q,eQzn}G۞j:ʋ8z8~8=k'2ͷbp׋NtvN"5n=Db P{3NU>${ SfjR(}凫h W )?`mdPs1FC< ~SDq D{I6=KqM4rG(gOIK"5F@B*>YAG=E4E@wKWP'40 X;+9!/٫G/ 3Pr\Az ݛꘕ4Kʷ$Mm &/7\8|}.}60nƳjeaz` pl_$ ELbF'kS[ǔs2Aj`F8<^r,_o>ʞBЛ6êEX$ e\[wRXo]a:E7I r.ܢ(4CRt6@zc6iP%kT~m4*.F%ך$pO]xyyժsXh̪Hi,אN#GK`˝AB.H4XAFHc4dSe0 VhԊA +RNԋr1+TkM1p)}&B0L74Z5+t3ͳY*XmقKrT*e絖#{؇N˵1mwz_\Ey)Ix~B~]Y&r3pk8k~Y&HGc9) 5L*Rqů.S`QF2BۭҐS1?Y% ɛtNv4#}ڀ?¶| >r&<>gpK?OU]ԀdG2ǿ_7sUV^"^x왭"f[GyZNI4hYók붭^m])A[Gӂ3k ;Xt/Eʿ  TZa_-Ԫ,Tgߞ^mrjGҷmUǪ'#מz 'c`ƾ$+Xt![TU^{rmmQIQi&^ c Ld/x49c3Tw1QlţroD33HZ$8+M -#<0kq$2j (~1*rJ(`+MrYV:%!VmqHoHSzUZ䟳rQJTQ) TW ĵuWjqBD!oѾ ;&8 مjrG_ʕxU:uRHxn4"`ǚEAȂ u$ط~K)C Њbw\أ4JaF]PB)e=Ӝs߷5u(Qņ,Ѕl*k{k;?5:#{ֹh~}Q kVe0V5޹%JQ,JoA1|ЍBd͍ݥ>yۨL=}h'kl<- d;AbD^;Kq1VI^MV10D6_`&]3wq?Wg{t-M|Ùä)L tkK޵.qWI~9U-THJDIeЩȖg>5Gq_g+9ٷ1=h?6lm;YK]fulQ!A$8.eUȜkD}%-8>5qeȴ1H1X5#$uz ["XC}4%(oE3~aUy orqzDro |Q9|vG%☹Z +-kx,Xlڻ,BF#Go SHc\U+BHLH Gp|ms5/#3B{ZE&T!"6t  VJC_^e.NYL9X&=\\:RL_͈VO<=kxbYy?#Arp)hEGIMC&Ne ň9ږÚZlZn0vykTw[~ݜ \;kHN;aࡑ#]쪲\f6!iƅ;P N)M, oK+#ÊT'[4gu&;,WâkTY0c$]q˿RB(ZG-8Q*%*֝JqvALZWlx<;B2dcr#;%#voW'vzH0\hXYi`*5.8ct1ش]h~M^5kp:RD` ͽ=Sq?kV2LY`6SKG< $][$c]hcdv,xg-Ka{/E9uF !^rڮ+]P,g ;fHpc&Z[aC뤍b=T&ZYҨ`$]r0Lcӭ,gGػt <x"Đ_v0$U{4xgB1.ׇg N<뙃/xp'RScmG1*q 3N 6B̄er]va.'!Ե22h`dN-Qz,RvX.?BeHr#'b5rA,u>(=Z?qi0:vG\IuF`8Id?Z,O.  S &9Vk' !d@2i0R;V8O)k{9wݷ#Ip4g)}.ӬU G*4"iS` .L"1n6ILC5Zc!L UYjcuNk9؄zaU"QEb*zɞkI̳cdOm\LB#ժw .`$]N{{=8 `J8m}hVE9dgFD8o6 ,2&Oik>Tv1PKDo.p=R2-4&R2d21 j99xX#Sbg6qJB |ۊ 1Ay3gtPO g)jȎEK*o$ ԲZ"N£!\Tf>eu`O*EHŢ\^tF W1+aFV|0ʡ\f(g]/$Ƞ ޱ{XDo se<>cg#$e?~JX( B S਌^TY<ѓ9\-4g#ivQg>E2g /,_V&Q nQ35LȈY^s^pl~mYQ>`B2pJ7}Pg:uhGGJٖM&I=E+(8]絝dzk|Jgλ^nYK­W cRr̘g10Z$#=7q!FM#H`NM=yYUXkPI*Lo8z˧՘.'Z_)>x2&Yh}OŚ31%"O!b$Sۓ6r *)*p,Y+)x--YqH4<bzk+'.!I:1Xu !yv5{ZM64Go,Rh2tQ}ʹB&Vd#r }z8-t֖`jF1:93أvwI7S“k OUaM?։@xFo]>)_f 8mO׺س d nE4휍{L"áJ$CEs NEh.뒪k>ȖEz?U59۝l.d;N)oʼayIW)%-ƼD}Zv@E3ul{C̅i)T5y1]j^^rZ:{˱:}KgJ88b^SG%OkԢv6X> z 팂 @+Y%c!r5c?t3 y Q<DGFM-i: !Wټ^]ژE<7!Qv0E,I#:v|QǺ+C̞EΊʅM|nFJVfY-,~(K1I|lx8#rS~kuofzsQIchKm8oIpfR]Lї"=]жt0'RŸШN ̞ ]db\M<(*G`Ե'0ф3yvvQ TwjL;U[ex(86Us{u78) .S\eM^>PfM}yE]'スMAw |ൎK_yɹPl{UڅƙR}+z ~!V!qDH1^Hpܦ:c 6xq A[u U>n( rlf+r!}Sp(}S/KHM,P 9- 58qI\oDqnTglX;8:NYm0 = Sflz;^OkW[(|vlPsr# x<0 ?j0 1al-2t3-!1EwŌw:]I\FZWOwlclo?v{YQs/}{3jA+e}|;X[SgŒ?ƈޓ7\fpcχ, f+!S* bN엛yQL441;0b51C@/@/?:G&rxbbhC8"NppڒTˑL`%ɨ̳22 L84e-`u|Ҟ 2 ƒ }E p $He/sQֈc؅o#q=-BƘ뼯ES,1S?XU^(]D1by}PN /hsD]9k,ꂚ61PM]y ) }ˤR2W1˔CHϢMcPgt9#rJ xa}}U9#b7_g3qnșJut uyW4Ҹp֦4L{CDcE$EI%T^y[bhov 3IoR7np;IDeҸzް :el@aO1?M7`Pۑg^͓֠N t7= Rƒcܻ ![,^oH:֡D7x(5t1̒]i5B kۗbv<Հ~Po$fi#WaQ#nrI->+$IbzvGLQ )xod!s˪4>TvFc]K఻q1w4lnm l;c܈,yLE-vmJGj^mzէ"BrwơϢZ;8 JmGt2n&Z7Jwzd<6On䐍sZ^-&ʭ1,жkj9:HXLyHʂp`ܸUYWif×e/^Xm-g$ХEPvPfH.!ȭ9b]4yTZBUGY(FEy]BnĞy8*b2@96jy|B2u͌u['Wǃߓ8o7>GEŇW0<{?&h GCT\df:5 RD1;vΆ䢠gUfZ^GG1$=-X0FQæOv|ݰ('dz@ g2ˌCȡUWG/?'#ԳM8rb\-G" Lѹ\>I ԞJ4:?GB4"/Wa.1nSUxoxICS⮢50hPXb H@J&DM&10_T쯇L}T bVfߛP굏3iu捯ԯ"qvOq3SSQq n33}ڻt.U\U|S\\}hl@s1H-ܐSQ7;@7oI H"hgnކ l2` S k@8usVk+1c~aLOGhXGmg; ; GKaRq @l|YL;Yn& G{Q8ETjD lM*1NrL(~F@>^\ˆ"?Qqm IӮ"QiH5ka0"_duzڰ<`| ݜazsP*t2̓In0rb:mފWc4Лgf0XRWG)+O )G4֗uT]nZu{T/Nwr<`Tv>ɀ!tѡEsx޳̎/.ʍj RY,OpFWݻfTEqh 7|]YGf~cxg~Ă 05"e((^ TxuT[7@ZupYñˠx~)`DiZJapάHݣuDZWkZ/g$g!g)E<4~S_v% i: oO ک:ZrKȵ *dp,|ךRdYdqze~? | nX3ηM1R=/%waTα@;gtDa[AirNTL&zXz^&.uI,nTed iY&BzpzfjKӻSMӭ'39EEC?]#' m1ؚA^N_oxzM@Di[Ku f U7@Zs>$m >Tow| |iVZWTcdQ4O^oWNcV\JV͠+aAVT 9Diu._W >56bVp3)}>%ONA1|\x]!"?_{yJ@dqv=@=0$O =ҙVPTlHOo}5s34gg>9sdSYohz>o[N D{eO@u'ϔs?V:XnHOҧBWhziD iZqt?R!ئS,iG;j./f"_""? njyl-~#|*St[@u?KKT|yZ0%kzn vu(~1B??inۈo' ?X_8y pDvywE Y$ݷ[/`8=R(pPm'X=>呬xt|“ٶnOkO77o]]cUqXw{&eu}oIeB&xK]Nݰ('v0]sUrf"2Unބl-p i`aøw %0NnqaZ]Ѥ!R8v&c|o ?t}-$qSa } j+MW4$ߺiۄ~ֽQ2 g(Oaƨ(>)V0—;۱%"V|41J4 "Y{m [\^z{h8o/DT͝xcz>B;6*awJFLK[ Ѯ/n'VξglɫSL|ګ+<(eddO3gIi`jgkgqSαf.1օ<XsgV63YY|Y/nJ)dC3o+oFbsMYVr~|__OV]`Fz~fN@%#~d^B:N+D"8F$1ERmq"r#HS+{H;W {,"b`EX"ޛ`1KT_EJ6 vụ |ckC\iϑ&!`T*it1eXZf,"^ jM|ǰ{\R#RxMY*zZ4;]KHz|c+\10&O /!gRzh{3Ds9^5IʱLj\R8B+ԅ0a=5@ɃFVy㈢y8F'8W:p0¹cD(yr@H7 D!95^Ac8Aɐ %$ vOE1ӻqcy`re FI.Tyϔ9[dt AFn!ҫ<Y9~.)mD)!&HJJJ2MpI3esIUVjas?'W;Πv?Q^џR;a+JbY"/e!_dc s.5Th/wh{iRcCA H,w wN\+*OʃC. ihn sϔ26ZRtJNcKz$ۃ$S7߇yH.$miƒG4mhC^OUA@\L]<#~×S)ǜ ʩ40&&<'JBHỶ#|D> nSNmY_XY[ˬ2ZEC#IbDAvulI6(0L=h !%]R=cb I GwWTJgy E* ORΔgLNNH!yB5M2k6.56Սa~|spaAL'aL(g098yV:)M-'Wk.5|.eGXe"``:|aj٩ThC:\&:~J,jڝpjc]-Z,;h=\Eh0{yS\d9ygTZEIP>֪A4gyDETnT63ŸxIoeʘT"utej;xCdc`rFdtXȞ/|}A.:~.rE!;:zr el[C@"8ˆ絧\~Owakae`f@} `kE%tHˢ;Rjth501ũTH8c;*yFؽج nt[A,JD1ypRË,M}bLGU,y5G*0S%I StH:B̥) ]$&6˙sVs5^LڄL.HZI"w ,$u3U+i?aš e2˴CfK$aSE5u{Ub]&Iq$6ɓ$7IZTzgs.̝3r'i4U~8,x\I*zyPp%ꑶBtóGiTϕw`ZMc2"Ik V%9Wnv34Z嶵5}bk}@ HKY:cLIp}hg,*G\:T_)j<:V yכ6! bpi&e0R vMI{9If3T:?ZKi?]o|zfn0ضdx)xRf`(:w>aB;C|X,[/_]昔#(!^GܳRuVm _4Ppɏ߄?iw^<ݷpO7K7klRZC-]}3{:]xije%,+'ÿ>/P+[h%Yw|Wvjt^B΀Υkgル;iۼ|N,G@y":wOvx MF<7>E{'`Mb;y?YQdrK+Z}khnт\qS^zGޏ-^i a`_ցN}9k4~Abxx!G0'ۇxI v6&LڈU~)G=6(ҭnoڻ͝ږTε3.g]\,[\K|bZgV.[sMW|NA-O_ V'Kg#ϊ:x$v\ \()Yu ?$3r3= ?tNR ]vx1)ʞ=V)Ӷ 64HZ@Rea Ke``vSbzy>~0Ir7xhQqj^?KZ'PU[' 'آ 2Oa%3yv0l2\cQ-LvYZ媉fzϪaG<ᘒ6K<H̚Hi[ƩRϻg߀"XQjž[Td8$8( o4\(J>N |E>Tj`;3+wpz  h4=sw1 aqna]ˋw/G'!< Y= ^!gYe8|j$S-vo,GЛWPQ<9Uf9cz:<ƝMbK/ރ]D_fa~'ߏU(J/\' a~7/sq6tŇW_;y PZOc6x71`;U[ \/^-ۃޢy| [x\nzX8p>x]>e\NJǗW`|.OVG|tv`=_q@GN1-umΏBFˉdyyyW 0ydrP?yVb`l ;3:rWD\DLXU3&<7oy֛7iD VaO1u._cM3^ʛ,c*qcfZiʥܥ2N#."o4 )4%6|(; ח^J^2YeퟔAǎC/i,6 LKU|V`&cj8&2-'LvjP+)Pj)jfny+ԋuj-Ycc7C+-:b-#SV.Y=ceeyX-T-LChk{VJ53:3oj5)EɤXȇRH2A#|wz^S^RƦ8;%]rb쒃XU NT땶(x6;J}֎(V RKr Ι|]G(ꪽFɶ:81A:b 4Kbn=k$ߠ>CgvWtƎ{ ~}}ίJ4 JN9k*} c٬ȃʃak"S<8N"PG<ľavO r f E$AHuvp'+(嵆}= bX7.@]&!@Ie*6 \B)Fyf"r[Ckk&cfY8%jCksgY7*|)Cg c-Ĺ-~ OgEu|r!7n-Wdt {Vu Z]3ȖkR rk r 6x؞z{Y[*1E{vz%YgɆГ dG6Øxv#Zrkv@JŠZ/j+IuNB'$j'=U }ݲ55T"{ޅ&PO7No#3Y{Uke6=;T.iX2LA$ˋLqP4*6DD(,2DBaO;Q  a}`I# 8G1h-9 l^6|K7*=˝_V!oT2)J1bFmlhZ4L""KAX-ԈK `R"(K- Ev1h)g-q'*ԊQDhռDh=Zm_7't*$j"}V*H+ ɦ?. d%Ƒ#ihd#.0ŗ`u[ZIg}W2Kɒ+K.{豬ʯ7dN ~m"*/kߝGwg=;;kӏ ԏ?]}YevCY0wQv̞RzJ[%xy`˾_>->]C?,?})Qqs:~r/Ծ~_{yl{?bbӲ߉qt>N__q:ot'u ʔXl;m`aU};.wv` eoy蚲F8 ] L^ĸXxa/nnh =ټB{~ͳwLyӵzu|M/ǧK>\LÍ/+?YfԃڛgACa+Zf%~cPuesK}]=B6.B~׻ۘڤ\D!\Dz/݊XH;zvbjHn=htk}_LL,=X+)gMwsuZx*⦐ؓp̔\3FQq5;JvޭIeM*_~IZ"T㗪1*; d$SG8%S`WLS"]X);$h$Rd4%RG ƫV{ фE1u/U`%cSӳ) FPmoGh{5'Ej۴ z=K^XޒK cn`J^hA*`MuܤHFG\Hiu`dBNcR*뭮8]o|L%Oi2Uj{z$qSu@ Ҩ)eMs&\/#T!֥JJ}LH5rRR[ԚcR){Pz5?b)1T*c|Yu$k%{8ߥ^ДH`9"BS+Ht$>&~,U1،L 7(mqdjǾ;k˻p{ygq߄.}Uw} /yr'o?5m_}Cz mjl*-6`k\W]&=OO{ o8l&M.FY(ٚŘ2"m:wGѕx3>l- HCAQꍈ> s$ 5}QQx߰m9 ɷz5= B-L )CTUpy*H7mӝ_\{K,ZҖoz>m95PM W=@}0V%x {5ꞁ囫ǦƍR-e9 ,G%!i[fkVmXOsNnE*wL]4(6<84]~^<#p)J<=TA1:f 20dj6H9zeo?ղdެaϻ>գ575E !g)FB\kcg5ggt'e}nݙA! ;}ci ]fCHMrZ=cAl6H1#ui blہBlY͏n +CllOY5clVbX⍢,aI6jf|"uX(O&!%Aa XL,{ 1OMiVc?:(vGV<|fC;E0MN;-7CSiwpڽ=ko1ij﵉ MFRxО*7[ʛCYܡxb0]*OT36\>ӥ:XKs#i%AG9a)-e)F`ٙ]vc*כ[ss[wCCĜ%1Bςlr 9׮0]aP]avsAO_X]И-veS @ dAF[^Wn_X6!fm|m3m  $Ή/ea9R ,2i%+Dkrem?;",M+R?ώ m_b"6o.Dg4ְUccfgPLgZc\뱩dLec,kǚYlQlS6:>3>4/ HⵘP"Kf#st.ZB}W˔2=Hz#r1v(BخZFeY>.©˷>c+ ŀ+A @|[\_}gهTsC,!968H10@79:$Gi2Mil5Ȩ,2Ly W(R;'=jW;3=̲Nv)Sv 08op@-k y+~6Z蔛&54&CϢ͈oe5ҋt  2E1hlbP @V֡y/rEiY2RsT>sYPUQʮ3֣ex A1R ,= mZ PS,ZdΎ΂|br&LZ]\ŮƒKIXۤ5@xtPⱯx nKi]lA0q [.>i aFج ŀB, 4ל%#gy**,g4sD>3x<] RE%^sh۪B*>GATEzBr"^ \#3QHd d))^vaF]T*S& _8F/Ա%u(YAJrR|9{Ʀ'22 *E`r<! xsBE竐WצZ )ʘSlb|@qu(@KHZo f K.CAʑZ:]O,2P%~>HR-) KS%|m7~̄fJeVGpe6F1XqA`6SXن)=\BЈ;16al3I+wJ& H9#',iTI?N{kݼcƥ0"BR{3 j:;AB` 4уG,)+lԆ`WlJ|ez vF=#3iR#NXdg;_)4>+Pܝ1, 2#i_wmH_i!X|@63yl[[ꑬo0٭duKVڏbbWdb7,a onݝh144*t)cٻ*b)+X P i p5.8,e<4,0 %'6Y{K.^7Y7ӻ+XG: QP*؏JytP(h7]Jd-P>PnҼuF4Py"Aa8TC<]~ @lx{=6iK-& l/W~{\а sP38ЏT)E2aJm[C,i=!^CÈx*HVR_R3I>rѩpdE.HIvYOĴpl4M00>"x,ӄ٢^t; Q۟a%_}a2gCN؁c{s)yAYjπWn{ak:(?^al ;+|wd #5 ϊ‹%_ B;X?W1˲Kagm2Jy2\6^^CkhPT O* ZAe_)',٦z?hck^Oz0)E_S'/F^N"892kahq$!Z޻b)/q&XZ}_ eO~o-I952I&P*.kz&]~rvX~B)L#tQ,Ep LKat1t\] +pC:RBc#TY-W<fY.ګNYߡ]GLlJe nhir Ff PHb)e#xJ)LXf!׭EDZ3#q$DŽu*AY dHך@Ria icN'=Ҕ %l26PIC>BOBQi.sDYZ;-ttg)>K5X(@("6"H"96m|2E=|sZٿMnZ@ En P6~FRhg:ۙvlxRZc! ϲL؋!ۉv|)%jV8T<68<ᤁ҇ Aڅ B ADIT$Θ*䩊("}-R@(_lMG>Sphtd\fD8/< G?A/q{eŋہ<"Mإ,`d^q6߯i/cbyjP+Zb!'I$3 ^٦Lj&Llv Kcd!=brGE^]xAB'7$aoiT7Ccf?8qۣj zm4[4ӑK<0D1`RȺRn%&-YOl۶s=L%k(BpuEiԘKb5f*j 92u ꢂW ObYrwн1!냼P:bN oWXp iy$8h cZNe~0,CXAc.Z3lJ $lW6 [URИLGJ3arsx97汻y'3XU!('wd߾8k(3Npmr:OV"뵑gy{ 6tcC G(X8]3-UsAyM1mXx3bOs|3_φyP2-`X|t"rͽ>O6& n#"zo}_ Ma7 q@W#8C1a]y')' Oon. OĎ?ȇů|8v۝v7 (1Fp)+@HI鰝ak6)!R5t Rƙn- aeY[O aЈjNtXQAK5&iF$ȊrDUXVdi1]QcqZjTԘ /xA1;C6`%Hmw_6rCud j|fMi<]WWuєҪVx?ИeЎҘeCo1awVηF5֒V/dTf%a^Y!Z^nh47|}^ݸk㖉}z=Pؚ[IWw4`LXCR v?U''SnvwƓV;Qfy.4Bpøja,q8ƨK&?P'ب8<npX id)%0IŰ`N2Q(n4͆7gemϓEcŖJ,%#i.>Ce  SӸ@7j?C?/uhxste35#!::(-Z8 ^.éN8#t\p]6֪Rq .QۣX@ahhZ1%q^O` 6̡:H^ '#D d@wsBÈx4˰G݆,Ef#=#L̖x_ų%"<Aqx2Ӓu$-l/{[ Zcr\5]ax Ń_\\F~;ܻkFV77Oˊ ܛJǫͧAg3 o*C ]T"_Є܄ +7gA% MF9Bx[fݎ _֌.a-[pyۜuM^pKMSsN$3⬥7_/cK:c^?7#BCr.YP?|pf;4Si̙4zѿ.cdM-sj8"b{33;\B[u3: 5sV)pH!e2)r- ,xKk\7ݤ0%Ei}k讋ߕgX l?p_3q4b pT\X2``F Ik*CSrОgRBc:}T(fќx6;vw)D8 ە}i9s\84bͩBcQ-6').:[ ;FˇVyoG Ta+e8@ΨX5JRSZTf0^אtj-K)*a[qGsQd,E&%Fx>ܭliN tflƜR h1sIaiFܥ8gkgD*dv{,Иᴾ֋vvTBf&@@a%†ð 8ehMԡuj)iMCTyJjfJ0\<3\ID<2X! ?o蓟^̯Vhm4 kgISDIZ!urID+4qs*E|sIͬϕ8Gh.g"%Ltv;XG2ߓ ^C&i4gPǹ4< oͨ Q'X:+J`$`XU;I`ya&c, !HYhvan`ɐ@Mv 0'xFRCra1BSTOŹּA*\RY3I:@&mHP HP +㕂~BieU&<erXVP]YxQJ$mʑhU|d- m{`1 0ҲB3^d\XY"DF9tƇJFXګdh5!=A,P|Lh∞ 閁̯#pHPn1R.:8rϢ.5@C,'iMJCH%[r( g˯)Y+1 .gom9z0(ؠ1}-FŨS~j?Fsr5y:jԩFjԩFjԩFjԩFjԩFjԩFjԩFjԩFjԩFjԩFjԩFjԩFjԩFjԩFjԩFjԩFjԩFjԩFjԩFjԩFjԩFjԩFjԩFWkfHFNQp`SJ:@yu5:XAQuQuQuQuQuQuQuQuQuQuQuQuQuQuQuQuQuQuQuQuQuQuQu^QGQQp QG]P=uj3u4xu}N5TN5TN5TN5TN5TN5TN5TN5TN5TN5TN5TN5TN5TN5TN5TN5TN5TN5TN5TN5TN5TN5TN5΃eٝ<:qZv{lw~oү!Xm0y_cQ]Zivtd\@{@j$\G{R&O `Xl1)~};V1a6Li[.]8HIb@Ʋ2`ePXvo5t` >*X#XuJWX-s ,Āv `ɩJJ#J*#JF+kbb(`)' V ЁRMk5زÆҲQ@J˸,dͫo-Z\KNL5=⻏0mC dlw֜+K,Yb `JHk"WV[+PBq[֞| `),#L1e?ґdv[Nz `%B l1ۙ 0MԱg펊` cSO4qLv􉃽Kf7s?]xm_Y#AfO}O>2CfJq2$x$1+#1ˠ-NNڋlßkc.NEjK#n`90hځ2eQBx`+$ˮ\6+%R;t )3D\h= ~-{5TĬJwt7P-lI@c7Xgf `CU+|(`˙d(aGBXCFK)om0n]d|)H'wFw6-//x0Q>B/|]X&:n|S[^ XZC X~܍{L# q7vVr@*f-}`k<ootfw3jn0ckI{5V9I:[ yR H/DB]|!^^&ݻl F#˱j,WUHͯѲW6r͇ٝg錋MҼ)&SywHumm[擟sA-^'m06n6 ţȃ+"s<җ09Ϗ4=XJ~tCioh|1S-9x{M / ufZ}Yb$6VKcɖg?wན XYZI"N_m´Pm& 7YXsSr L}PJn>S^x):ks~^@R[`C2'QB$ZN 6^z"R"d _;;PoZi_|ynf1k뎡3-RHo? Ychan*˃&O&~ ͛Ex; &|.mge/=eZAޚti%wJ-E㵗/&e WץOizoWO'>(qܜ㻋mJHK̷s0Tx1wJ,lW~t)nypƎדR~o.9ʷWh6oJkΆ/V&ip>='ϝ@VHPAoϗ61ndx,~6zb"Y_mt޼QzOCyr<-!z^N__ WNy {oU =ܾŻ?tËmэ{YfO|+?ۍwny7?Z?[eё%4%ޅpsSfɝVv:Y5r~}.zHw.|3^ζ1x݀j_Zt:jm+ewNeg'='xʣ7iW/\f&5][8*z_ʕv*U|F!Mm`IAv( ZA^Ⴒ{$N \_'u¿:_'(Ф*f(yzT0;_RLfִuZ:j1;Wxa2 N q.]֥Hl`YgZq_qi;bcγ1uLeI]e"둕UJVzB1=6d||AUR{\SD'Lw6N MpMz!9'ڧʙQo(JKwv9f{!{z{z{z{z{z{zN_@q%0PkԍAw{{8ȋp¿#H[#y)2o}rw-m}(?jZLlV=t9 _ˎ(ܯ9&Xd w`%P7Ґeklt_vcE_|o/ Hmԯh}f%p+ eNa:|{Z븳&ض.$* .[PZ(Tsx!}#{RFtK#[K`p/meXݵpox⠙b#PgDuMiUQ{Pl@#j)f.Uo\Q.: *%F/c!+IR)zJOUobǛp(G 8lmLYw & +D>UG}cV"m 9#DQŏDq^)g \ iF>UNTٓ|m%*@zZN-F"A;6Ҟ}WLBQ!7bir\Q&﫚D>U|]k71ZzJe v6rd(B `5a|VȧZE\m%' !&GgC .%`#`5|n҅qZh ? fM ,8SM{h5+ݷ3acdJ Td0 r&<W8EO4 l>G[ &̗{>~چi{s4/pm + 8t6xJ*|Ve |xkQ-"'Kwy,6GiA A5FCAqk{SxTN#r¡xI뒵q :1\z٨2I'|>cͭ!vMJ,\,P\!\RIZOxQLw8@T `s{ a9ؖc[VmPY)6 T*9. !WUՆ ^.y;0??mKia׻o"Xrc^o?skד;T7_?^_?j!|2d,di:owVF"Vl}#3-ܧAsޙ"3x|?%svZysyWۨΠ;(\;IO i_L% ffg= z$!׷6/<қý)O"$pA))V9B\oYRxƽ F$iNWջ' XpRWBVSGqnʣz vʱ(XB <*'OU/a5ElwY n* !G:$y.g:F>Uǖ-&≵T6 B ? sRHH{g,%}G_+X!_r_)}kfUV/&x2% 6'.DM<繤A}'*Uz3 UDG)BU%2p 1\Is2F~B#|ʭd#^ gH CB<s}'*^3DR'F){$&<:,EYD<k\L3/xЉemD!*#()#92:'9ȣ^^F*Ii- MH s $KIZ( 3@"raj|.ϸ/1b4]8X̃q8T2 b26%Sc *Fx+F'|΂ ,nr1PbV7X٥D?NjjZG\%#/bl ) JJKȧZDj[bHL P*'MkS pN^=ȧzH/#Nc=˫l+fZfZ2_L'/13իC0ѳ~f{NY=hf肪T#Ӓ= Y@ZKMwBEFg;N> q rƖ x3!xkrYVqHsxJ!~ǡVU<间ī ʣ/*mD$fV)g<ɇߖ:9Wb.7GS_>D:9:9ϻShWnEa_:)EPC!~ᐤ|+!s >U2ߙD>q(+oJujQ=诨Oi, Ŀ@W$fއf#WFhz6j.p@Ƃ__H-5LrҊdXk0,+YULQzU/uWR|,v42H(m 'mGx2TDYPX}@r;} 붰3t?h5z3Lj「%G m9CQ);B"^ft?c*7'!C jk@pg9;]f~f5. #cQb"g=eTyKZnvSu vpF2G| xc~qs}+m^#zEHPAmQϸծf|ˁCG3^p6d# ^1X?㧻ϣY.)1ԾuiHd5m~zi~P|"l>KXΏdd!4Ǽy 2D~2k6 cT"I\D>RU|xq6=`Ls30jQ5 LK/%ikwZOG7 u#B~(F!^捈2'm􂾤Kp("~|~EΝK>t7O6eD#JǨ]#~#TF$:|;#V;s- N ipʼk@o@"d՝4/q+]Z&){~Gzx@LuqG&'pQ)58}Y 8Pkȧ_1w߯cnon&̷•|Ji|] Ǜ͘k/ EH>3f˲@o0D>QWY\ivs ih X/yvf! <788:AÐ)LzOոJsqD7WWq1dt4Elw>[i;L?IU.x=W]GD(=dtRCtD>U^]ۏOOczK>% #6n"K%A5:ekltƜlӔ!:$4z# ddg|z,? ]&c]htmhXr X֢a Z\x[\ȧ| km A!$%2lqsȚ2_D>Q|wC\FE/d.۟\"8KNXDCeڡQM`mmD>QbU.ԧ蛔5 cXƘ@wh0:cr^si8cTʒs;97 \e{%ڍ6r?[}2D>Qnư]%%XL}Ԁ.T,x8v@¸~D>U- tS|݄q}M"Z|Xc?<<<)7 j\*}lr&߾i# .gQu{R# (2|]oŌ%Y] V52M*.D=Q VxV^Z=H;)5Ud*`_dt~D>QR Gvn4+ 5ޕj 6#2 Z,M pN"+P^6j-ܙ^d go]e}{󈪆趙,pDǒ*'2NN$*b̰Rp [vT Cd,84+Գ}-u2 ?Ʃ kIU:AY"(#mWxL3g>W^k%l72[Wкքgy Ck3 2[=c/ Q |mM5#:hN&{$\UȨ >6?8t}QgP D_HSgj0ǖӾu(ԓT l lZhJ3hvwR ˔kdH{X Vݧp] R&U:%5S% '!go\5_ ? eǘ%4(-@09IzmVbՀdoJBO+cJ"sPGoa0g98zq`S- g^I,LvJʄɨ#9`6AG.f&eY^G9Z}!gK熌&DF/vCb)eGF$bsR#AMR#G%@%ƤPTba*MI^Qj@jfEGkY+T:y`v ~)EC>{X-zu@;Z}]Bhbۅ,2Ef<%HiML0VG6 䣶~!QYCsAS:r!ߵu Kʒ `¼Oe9ҡfB;jEhm\EL}Q&$Yx( pV`} yU^,V.ÑYX0 :0HUҌ#ܢsAȫզ )c댧~wFIF8YPY0sű hb@P@={p'9|u7Ҿp{p,f0ycm.!9nL}ێa:6C[{|;3Ϧ:]kV/;LmFp70c4rǮc~YáT w[ gJT'ucpchd;kIn#D HKQДzp/Q~Wh| #DƳਮVOP@zg(oG`]T[# uB[xarleދ[G|w-FA|Ȩ7nT^tEw*w\ER4`2Q1JR=?9/6JP \B2rqxva),e)ui `\:"q1,u>0yJI]z͖F 3Dզt9\ݪ䴨7i;L2.x\trE5+_~^.*U?յ{s1.4DY@0^wz&5GYxb*<<Ya7\9V >'t;cx<>;ClQ}cWI|v_#+u:esξ_[x̩9 Õ KK)yB{`\FbA{zwu˔8eCi晲+X ]r1 : ]RwuthȩAIK=8"gZ*5$Q0 Gddy]&fGKGƩjCt/GG%uMao,>$ G{ SdZQyu9*Հsj X8=8"G^NIӠɃ#SB]Ȝ3WHy<g119'":T ;c J A\Qg9Պ`0MWgEF%A &i5{՞]pQ "N>1Il3{s'^JDDV#.h8ġu@MPoӂRpQȭ[F9Sr,QHy#.q(ɮ|:@bao1ȭ-]-n[%۩ a"(ȴlZfÀ PU2^-2aNojcJ"sr,}8"GJ"Nu`/*%6 HrHboa 1*!؄; .[zqދl2$(%-:\<29/yt~횑|^ X%bP8@jS*_ѤRJkʥɁBZKhn$nqD&'B] Gd* s@4 ɯS1dLuN)Xs։BI޶5YUN=2G ǥEQr)s92.V)8Iӡf0 JE1[.LW-Pc,k+FMULgz/ ѼMkoYԬg]bW%#rL \ZL+#gg9S!~##5pm fA*$rd^g%+5eyld*HQ=/DUdj˟ v1"k7LPώ#ޫg _(#NЛfhc4YJ|[ue7fT2h|Z-|6 >\<.9rvMmwӑ6qrL.;@ }i*75W%z~\drHd˓gK߼2ygL~ir~7 C'~1/zdo*1-͝Q؟w=:?0ϋ(S\/Hc]T _VEO{n4|ϓ?,ۨ?OaW_'>s5G6O?'ZElUp혶b\mǷ?mlwpo׳y{U$&R+WO9ʄt+=eoY.7PؓTnlp@O6?^|ՕW3&u޾S}*Ryu[3˟<[M?{$3YɿdžVxZ֞ɿ񇾣ݹ@\~H[!3arjG7ZyQ9-Ͽa~jTTE(~Fz~u7-&]1{ c,'QTGXYt}'I5ڸWmBo# #&QҪ$P[!;[*v^nmcmr_lȔC]]f\oKp/GwTX((pH'.zQGAP6۸GLSeU*ސYE~^M-f7&8Ig0Y34o ֍?# |PCA{]m\Tt;mQ# ^fxD9n{AAk -x~^nvo c&H~5Ʃ$QA lmY}8p/ؓ3d TI܄ 3ߤ2m@p]ٵՆ,#w/A`Ŀ@y9㵭\:zgd@}\ۣU_|m9I:ŭɿM>:~?\S>k1MAl/d%[8CE;p[ԏw].[ Jg\`s üL'C^]r`k~{j5۸5ru:>@'WQq?4ڋ8H@ (W~CnnnetTjCZT/E/ \- TNKf,1DJD&t8kk9-! 5uz0 |g>ol*jNȮ졞p_>O%Xi_m3Gr9@o9 (f7( U(3FzQD%a88S0G7oo@mqJC,OO+0GX>Vrm8t hǃz57y2fzYvKӞ}ž\tL)GF3/t[1HbҙL H/2`x CR+\ʽ5SYw}K: 꾁߿5۸KU Y{(5J"m.8L$/,IV-[MKmZy![L>/9onQA."m.b[TD"Ow1KB7L T i.s>Òaa\B/&fQ?ҧ^y16^L pά1TԨ#_0 ݔǡpXۧ1!gn 55 frz# *ys臈ֵHGsL ΐ v3ard'=ws}Ct>8e5h*:v!u'eᔩBR 'cpԋ0 ^9.CvuBq|[[1 GukϨJ.z |lA6Ѵ ٝ2z1|$X$mƜI:..:䐬.{8ѥ>zkCMiע4m ixNd־!&2aSTTtAS8a^co)DYXrٻ{rX"V8,1 {rG%fxV><e&RvaḊ҇bbN۷1T'RQc|9g#X_C%}z=Y7v{g&yu2ᴀP}[`ː>W {1RЖַOW~6~1Sm 3) ?Fo'_۷70 Jՙ%1k6.f 9k.VJEg3fV{^nf$lh5^!Oz!;־!:蟣1Z.4\ c}e-NJMm6IOs_اIt]mMD¡Qo;>E5tr׾A_3.əEܚ9a jQFnD\׶v5=FrJ'hTU6qee4GkbZ7&e[|?QA'φi ,᭛0 :Mb@yuڐ`]N_KzUoSzpT`!Q?*H7#m)ŏ; Sf#k$M=M鿔[_.8;]܍gՅW ލf.?y{oFY9;d)ۜѩԇs}`FQu o^;\ K' pATh Á֋Qۨq3*x73 L9v0.F \Ɉb*fhC0eJ }$3cG?՟t?,8Fi_}^;Lǟ|Cwv@=9H C) m(&X~Ʌ8vL6Mf[J_-^}kS}!8!Coxo8"Ty#v#%qWƹ[~E/>` kބU߷o Dp:d |ZN󱽝8[7Joyy<6_vDU]淛Fا}AٵM7(4sFMM/*Ei˧+A<#T|oj7'>8 =f(ܱ_^+ᘾ(}ߊX4zu>~/_Ms˜bOԙSCW69O?!9Ł>M]^@2hV9"G)4*rg4 ];w~-`+UFMF pt jC\JR; KDwKu'R~Q`|X櫜ޘM-:]G;F0X]1uid2Ac=*OW uv ݥRL<%WW2hU9^ >SӧAM߼nj_ ~v]S:o_xsjWunjðl S\v,A쳴cQ֥R]yrR]S IfΧ1BbDBL13ʹB6qDGOl*2 [~l낦[n-1@R}\{ +A2PA=(QC7[J0bJ0JL$X8풱l!"}B"pcIgBSpm4 )8/mCrtGOU z-׽Ur8jt@U&k* }*m4Oy+TLz,z1X>譐=7W6K$Pe2byh06\L Ffg\RPqx '7L(tu/5?ߦ?8:ͳ缁fyA U6vYG.NfG%?M ?G|2w }$N_%R5NkPQZ`9-Shpjps6M<#úL~)RKzԬݍᤆ?jN>u{.hK޹ UUZ,$Vࠔ=SaclDhն4U61NrqG!M@ ,"ˤC9«A?^u߅}"^`HH*jжsrߋhN> }Cx4 ˄L΅BY-ϑ#)qp Ok-b!.ZKiKx%1!B%F0*aTXƤ7B 52ɰMb: 寓^IܘTMڢ7ߎOoDz7'w7 qG VӄWзR :}f4-׀_` <᫫e>a D"9X@d &'C Kf[33 8b BT4P&2rsHƜ!+g91+6x|Τ ya !v$T~굵H.^pK+ֲE8w;;) iZ1 SN&^fMЊԃp6By2CٿSR#kxm>-6*I/&I/IcCO1'bϱ'heX*NGAbkM=a+[ 9L\C1$(Rxײ{{Ƿ-CU`,ksp/2ʝSn qZYguS_E փh >>8 SW0۪'D/p }g52ʄx1iI;aӬiTCx*3*l3]6ʰ\ |Z<֥a@US(w i,Ҹ}d8c0|2:'Ur/8n(>Ʃj!borUAS=@6:%%{1aw 4*pʽG,qd$T"w/mTcSQI<xK̰v{LBG ށ#g]ΈŎ(k;>`gJAaV+D(K*C{pL#{ $87e8_N+8 ;8ar%؞0 "k L‡EN;:>aL{HvS9˱_t5׈.e{ m=zigz ֦EǺuwI"y79'"jў?_0'~r >MJፋqfK32̼B$Ҋf*#&?/OϏ(PΈ 9JIas8v-]sq1ثHg>եT _nSy'(ײ~UuL߬Y}9]Ht ,=#Ζ{&f=Z !Q)eCc B˲I}z'Cx܌މCMę¨\=8z&3!}8xG̤;,%@qᆛHUp!q;g: q }Ǐ%]OM=EvZ\,6@p8pH2jX[vƚK=@|1|wRh(\{a%z5C]GӮMIo[:YgP5\5v  =i( IcS* S@-+b7mh/GejNd5|WQ]צVqįŬ#V>D-XTǗ-Hozo }_kcibS$v%% `Lo PUX Pe;[&翄9OzȷMTQ`53u9@5v2 PrSU 8We;w$-evT2z׶Xi4xV?MtY˳.Om"Զ1݀:\!RY0Z&N{;蔆.~e*II*h^kGL %ŠY;`1k9gʷ`Z=e4&S~#iARh WPPl7(i擄<04+kj ,K2V:4P4e :T3jZ|9:n^La*dB)NJF֋`L@%VrZk.A oխ"[u Ygi=eL0Z&YeJr`z?Éeq[S]r˙p10Z&N| /~Dؔ9B SdcIĩk Fq|5tJ5h8 lרzJi0[j^,}ΰVU#FqKӫV%[@%J,kSxX W%-VQ/]i.xc*[ױ duJ*uOzVB0[VWnLqdW2 Eu2IՀ;ga$ZD5V2h8 x˳ʐne*O- \%%a#DnR(f̥#13- q(*"I%'TV*M'LIE*5lh8zc e5"SD6L0t]EXm#00&N*mq_w!ISIfy]c}DJ.(o Xz>DJJ"e%G-Dg OipJ~As€RU Rv,M;@eKJk&| NkUJ/2qH(<,IrZr©T[2rTJZߖ;+@ӚC$.V/(%wc$'`hQ9\i%)bIP+=bUk#qƳ/y2U `L&I,2*mx邜+][sPjke޴NT0/\Y(0s%;p9pe4"y\T娽om Dóܔ_wTr>e4ۨٻ&]SSMNzeӳRVٝF KqJ|r߀ΦJl'QÐJ miP  g0Z&/Bu._]5Wn1j({|5fujetZTi ƞ~`al'[K/]&~e"Ig=D-e72JI+Ә;B^Vݙ_ǍXt&9w4;(r~(GIz ¾5G^]$5xo5l2[(Y`sy$I805yt`P'o_u˱t;%N3 2 w94#05v7XaJg,ȷd( !OXF:POywux{[iq49 7us|ƙ!4I=75lRl2<֝pGm)6 F21 |6;Km[)zâw*h,XA}310\Nb '%`/zl\X[fɍ3ĠQ$4C~3L`R?g9IMOOj2@-A-3cf,x} sî WɍI#mWLl>FfNzu *iwXR7NKpU1QylK\l|Pt&y“;=.?n\^< ߏ4v;N?'4.޴ɦ]^IQ7x;= ||@Pr1V۶L- IN(Ғ D dَ,&tzg1tF1ҩX*&LHFQ!đ6-'=p0p*D,XR0l;LpX`c,m,!_َ6U yU;?ьs?AI V/9u%t8piaA<_8,Xn'Q&_ ).=Bpr(/g{mW8;|,q\jޠYcAf> ,(Oq,']oy0<)^݀X300Rʌ,=پ# @b}6q$@w7BYO9#=¸ص%wD{0@PBOiiw?6"Et my!H4߃60X))@bzI`\N6n`X0l`VA.ay.E auޝ q)mѳ(>‚;mpg?y}yݞ^`?wnßg{G;e7J)t=~wrrZ̭ACSJb`=f{bKW"R feqQ`QTEtzwONvz&8V1{H[ aڋr _ظv`d 8%ֻzOO&(Wmoqo3=;tfu3}Av ڬnQLr8f ޕKZq%n>%Ϡ߄p\Fǥ @2 Hˑm&!\!||;WOo ē"<,4.Ě+@Dp havs45TabOH*=jy[1-0$U,"N,'?`8c'M(#1C鴍Dռv?T!NuG2t3Xy Kmo§{HMIk.޼uo~ Iʑ ݽ0˯K'2Ǔque[\W 6lkdIht1*C9Pa=Z뇯7v#RpoAd˫Ř'̯^~]vt"7 smLX>{$Mp-.t'0ZSywIḅ§?Ȕ[$i+h=(#bB~Z]7zc ^ghՏ6JKWC42"R"%C~wόk|<$ ew_%J0+5/lpLb+k6CFb$gY?.ܗ+w^^Co'7' r;7^LE8{iha\D%.PBedBM0]L\;6k즒R͜3ʌs:-C~27q#3hnh †gtf$p. OeNϿg2ݾa!'> 멦,iúi6˘+mw0 9wڣ8Ge&@Q잯YXֿMHn_pHޒ#qE0\!>Kgʑ۠۷X-_VBf8wFSyz K"zY&G"X3"tLze;q:9a<mIzV.ȢmQd `F)ž Jчc/[gI{#[㋢r`~wKϓc;Y2K3,et˫+az}u}ua&@{)gRQE sӭ"`tK_ߎ iqDceD a?ly ss7Ə7ؙiM-[yA26~}éZ}|l-r@a/#4zv0;mX贽zZ|mt^,)0H v0+zGR=W@o݊$>5V&Z}Oo#ʖIEQ)3+oQQ>KGzCg-avۃuݴ'(nPyг ްwrzƒo?y,jŵͶ8G]wR.mOd8>T.vLbyD_I8`wnC`1dw\W'Q$d߯Ȗm[v-:lDjUb"CQ$Tߢ!,KrBzK`_zh pQ]=YX.5\ EF-GvU^`8'~T>P@T>pGe"7MSaD }|GM[ܶwNvc vvµ:Dwu ={?j ߧs/>yA*Į&B{Y{JaWr˖0GW]MAirrj/1+^{Ӗu[ y0_yoW(^xW(^xw_@aV^c) > ReT2Cck<5RoT(/"DI‚ ˱)z7@g7#؞1QlG7SJlmXFt@|0>ŅzuTى"% 1^xk2nE`k"c=[cl=[cl=[c5״^7.iз^kT5kFZzQ֨^kT5kFZzQ֨^kT5kFZzQ֨^kO!kFDŽ VkTZӵt5]kMZӵ&䶚VDrӦ`x/suMt<|kX'N?9X^{}t%||}4>6Gĭ7䅖']s/}V%՗m-893; lx 1Cpͳr_oN\wܞ[nhU>[rf?"8 tNS0CYWY2˾YhT{vr-V= ߝIuzL^_?~ ?X}jzp~zN׫˶~`I{N9$LRI|j%x8 IDb6/j ZzE+2N V,Q^#h,P)Qo?z_+ rr?cjLxl%J7ϧH~~O%׽XH|ZYc|Zo>7֛OͧӝtBv(}>5XNZ/yRL<猪Sx>g0`mHF? Ć+uc8 =BaÃC5qg52r8jvv`Ӑ闇{~>\"9Տ(Z3e_nF$ u@? CLTLyiM{`_M1_nN01nd\|(m.K1?O4;V8N&tr:׽saE'ɧ'1|9RMVs?"@ojgu 8g0gަɮ,vu_M&V\t=q}j5xi ;C1tv H%N7=L VMiA`w*+P,ȻŠgR*{0PLpzq#Dob!!{=z<" ^êըC \FQ1+DL %'JL-MB + +4,PByjL^q$acGO ȩbHJ(xq&dhHڇ F;",+uʏPBTWhE{m"4=Bj""Ka`ΏPByA/z|iHJcpL)ύʏPDyK@ Âώ+qmdH(|:WZy([4-t1]KR~+V9 3b;Ƨ`c$P>mV^2B@l.]Aa )c$Q\\y5EDct`-ET'[Cj6H(4w^hE&"oSqk@]1 (0UW?\ZPՉ8i@@ E7\C`Bbg$+T xzՍPByrxb s0ZOs`,*@ %x2l-6@K m#YEX%]]]OA$O$XHwd^"),E$j9]]}/r/IZ]awB1/<.F,M͖5;L1x'mVQTLJØAÓZ`!] ~wbl&,B\JhԊoD=ڶ >^I)KE$8b]aGE?Gf6jJ9^ݣ1:,lj>UlxꑮMW]X^QB ej/DW+>cЌ =Pj% ] :] |HWΐt>G] ]9cwڞNW@ֺ/͘]pa6tF\ hJkt*;toWș{ۼm^/?`"׼i9"&ˋVK2MMIϾ>] ~~f|X~OM`hNm!d3~E؂ Ek`w;OX|53472~~#Fd"~#jص~#_do55[zNO IvN~$__E]"7busl6;%zutZ =4%v{w󶝜^ߜ\6$7}Gs7QઇG?)6?)ɘߗ(RUJ19mt+,R쟶 {n_fvθvAT`$1M V3M}˶ѾM.|pnFx!>nnyw+eyϛ~λ{FoJkyˉ_~Jrzkq3W۶8+skq@eMν^k7[Ou)е>⯹񲞨#"cfrlq|+kuךNQ5Uo0dIJb퍿c ϟïbv[Pt+'J kX =o/ZH[UYf2SfS~^«.<_n /6?U]NxB?`\0l.(< Jk0+ p@ (;ѕZ?Ka~rk}eJ:G"^ Mn//]8Iy;y00^5;%'칯`~Gq;!gnx^7O|zy~O<2Nk=ȧrX8fyY{>a1t/Vmw ztQ;:w hZGhJ>:hPǬȌ*Nn6t5@rtixӕqc2j?^p_\]퇖_$FW~Gz{gDWՀ76<]:++7#@4pͅZNWGztXfDWppWg=t(HvN 78s*p̅@NW@yp>7+3+&Ź@+|t5PQ]Fv+t/i?v6fpw5P>+I8*͆ZC2k+^uڹ@j\{^]EaV[gpF] j\_n|<] ɲ:0"ouV8f<Oo-ϕ.S9v3ѩbp&%zt_/nړX)+>IT4S˿]ͫ?#{/u/wKb]ٮP>/#²af5G~y|rd~Ic8}sh9{ާ㉺r-jE\|7fㅿ~slɲv-97u5y6LnPEIR ̷OǛ;Hcwշit/4_\-Nk\Nb{bL.5rӞps09&l ꄳdOɌX>ZT}T/vmu~9DSfL9BH8gR5m[RՖlw8`#la~\jPʡuS&bmXC@F'2yn`̉-GU)цRkPFRlb\0j \zNѥ賄le#ZV]M9$զ\Ru7D&p7@&2oF1=FICvȥmx,fLcV }ظ\l5C1qb^rJFB 2^{-MHflz9n2Հv1[c(W 6|.v!lEv,1U`ː@KDcUyjz=$1/o>C*DA l^!H~n.Ri_hT0^:fdyKƜh49wP8&w͍hBYlM%'i(%%Ql.r- l:־uJ>,|GtS;RsE_&qu~kk]„b&`VCI :З 3kv7J"FSy#X'rJqʦTCi&h ]MɁ1sAEFR@ F ubQ =56TnG0Lc&ˌBфi&Y")!< ݫ  1ZX m'mV-G¤iP8^JLA2Cȳ±&6V7.Jɛ+)d Bhyv`k`+k.Mx5- WxlFƇqȤ9+x*PPz**F@pɖ ,5L #N-`U_θtTb0Vjp2QlH`) ndőLqJ +*AqR %.CiZ|W2TBug d#Xr nZEZPB]yځLQ!ՠ໴\Q?<1l aXoa iJB‚("3} ؙsf_)QUK5g@xb͂Gd0a \BC)q88)$ԙg*!%@ `=J `N!\dV!7W-g@Qā.QQ6A*4l;"KQ` ;ljA!O9Ð]IF).6RwYqO J HOTIy=94G3lcHHė3' Pl `d"Qy`PmB@ lkkAE2&m2*-iPU3TǎYcL±Sr"'D!` rg7``E9{?zEeR\z]/K&A ڦkka%a ƛKy*}r@ֱ+E$]:\e ,#<*(v6' " {jN\`2Q*ڂUkfT,Ѵ<0{̋&0A) |&HVBݺ2[ q;*@@8Ƶk2,ՏoM?/j[DbyҶMRFNEok#6ٽ}_]w<=NQ'K3ͅ2Xy<lwd,Z챨wPR~ȋI}L! % |+Ĭ a{vwH H1 SX4iĐ-ChgW1C;V DGWH!-v7!4I5O(3{:ByZLqk-)Y cwa3XTԙ$ fjb2Ō,%&H?<(BFXpZz,Iejn}HjKMʌؓtaQC^(bC` \8 7 .4gSM8pU.]oƖWalCyM@m{q (u#$E=ze☧[f9͙ D``.:Mj)<;pL.bC/ q46:]~k]`!S}g,0BgL0pvꅉ?M.n/T"7L D?V=b4'>7N9z}A%uNH'RA |"쯿.:k@tPDxR,*i2z6q軷/`dB"Yh6,pe>OeM쿪W,} q5\ic{3 ID h?:h'܍8fХlbf]꣢;.#v^QG(Xx2 ݍ~ɋ0EЂ '(^Q- 2z] 櫼b22F<" #BXܦ1!`+p}0;G@kE]Rα-n+6Gl_4>gy;K3߷^<d[9̣?zg^fxݘfF jTUj{p|UU}kmmfY JRwTSlQ|_͋o.xWϣ,Նdf'yM˅i~N&X4qtӇ\47P#ЛrvWA^:sۉᬓi6I'+fڸ /u<~?uDޚś|Y[b>O NKjvRbDz¶WʻX0C kMܞkA(/cwŸo0iFºt(r-oUwƻ41Gpv*Eʂr~Q{HSު=yʤ[e#?vVA(GjJb=vJ habِ0s ]!\C+Dx Q.#j]9Jkp͋+=%~/ a }H0-pD d)#xbԘ|o`\S@\NR.__+9:NCp2ߖgk h8+{!~_/E0Y??Ŗwd{XLJ}FY2)iKƇV\)@3E9a0Ύk̎s;FaYw4M{v4"$!p&h}~t(5 ɞ]يCox:Z㈧\}bZw"J߰(m DW=F5z סէC)U JHbL]!\-C+D{܌ QzEtBJr]`qP{-P tBl%]] uX]!\B+DkOWft5ZDW0 ]!\.B+D+jGJ]ƛ+wj7 -?EUB]܄ ǻBCkM QRt!yWhá++%j~ PjoZHWXـQ&B*+Dkd QZ]CxـU=ԇ6Ղ؉ZyU=7\ rDW=qӾtcWtB] ?Ptpy0thkOWR ҕ{M@tЕěET(t%ct%0}6ҕː+lÉ]!\C+t(K']_}@t `lm7R*ҕq6ز` ᆳlZV;գЕu\MVcT,ߏParxz.,ʽF֕id({Gx?B;& `x eGhmw43(ױ;U. BNCW5>5Qrul#]yom<_ЋGБK0~Z|tљ|Jc-ߋXD8 ]tގ\g7o:;ٰO~^$_E_U6Wʞ^tfg/bE:>Ҧ&O+?,1_mnx͠K'ia=̔ʽ~]}TzE]x+KOՉ *Y-ppBٽW􊰘n=^|ϙ*n$&T?f?=Fs *<^b*meeǴ٥V2V{Ytl櫾чl|t65?F6 Z,ΪSۺ\U:5Fɔ;Jc0/x>`G\U29rH{\.xT84>Uj{p|Xvi6oY)U-5j@-{,Qur~|O)KWٲu-58v8W1{_͋ł~N&à|ثq cW()7Bq-dKoYٖ\B6'X_z/`Yxy;im];JQb۹_y϶麘cG m"5[%Hm]y[':ˎk.=SiZ7rK}H̓(7[Ln&@jznrZ`5sAxps$U#ʆ]GtHtLkp+=%~/@ a w1-pD 2~^iL@X2CqS y>'cbFY*9VuVLU;jq]tqu;SQOݦTaӲf-;4G!{ÚAܙca:Juy u(~.:o-M䪳 I>g2z9Mf8?&_^q>D`@{Uwg TOX˔>3N.:wv [ϛm1g`H]HQl & pU0'!c4Q:41 3DWԧׄk}(th-kIc]=wۓU=GԃOJZ{"51^8ծCϕU|t`(tH6ʸ;\??E*X4WGH*?g1 :n6GԊe8ǭv{= ZI_-1~ bGE#Wdyи%Of#7Yޔ䶣5=jϊBKX͔n.V~F<3筵{~J` U->4aȫV6reTF܇xr}t(WWVk꩗UZJܽ6M}]u ^ӫ^rVLtŶx] =+)׶ ٴ]DK+߂~e~yqVY5-.A_.W_xݤnx9M/=s@du;R^J,x2 rQv_d8|F'`5`qey"eq~vFKϭ ;ɬ#JADVV P6NCpo }β䖯^ gs zpő:_zkōbmq]*7ۿʩܗʩCz_գQn_CvE] 6!ĸm,ڷzX;nbE$:V~)tb5sS'5 {thhMENg"JrkYKHCF_ qGpĭ.T7JüqlMѝɽkSb\B;r!W,6Ub_kǟzEu{?r\4iW{v'=:,9vkvMgl/M);9/S)bR j}hN㌊Kkfz5V뼞Nl뗍Z@v[KCkBDwPOZM2=w@2ԭ5rq=姣x$DO(g$\3T(9OQMk-1zcmWe쎚siԐX1rpCr^$8eP))SI-$k5RHنԸd% {KNoTy`Uhuдk/^huIp՗F9vXVgf.\L*W ÂJ%PӜ%R*qY}G0CFiP[$|΁$b?( oBm.\f\ )ZKy նpCЭFSF =-b < pʡq yOvM'q_YW*&OSkn$d^?jK8*M,Ugԛ4K2w?^Y{S5MUo+~+Һ')Õa`&e0ph?gMy1B2Fl4N{hr]0 $_^I29ѻϳC.7`k/ེ S-q"G4 j3QB9aK2M++p?Hb6[o Ќe>6~!!q@K6L/ᦟ #a~5x:?L\Eyst*"[ ,HgI1jf$gBA۲& @&ˬM3h+Є)Uy0tqbTޗ9ތܟ"R]Wfrâ`矓iu">MHJqh98"OW|~9ĩhGo #lЉ4J_}LGEQ3~OVx/uGy:!T͇^~wŇz߽@_~x h Ry 3w $?nJ[$͍ iM7IW3kt5c7aB`ݸrF} ǚ}{Hړu[ kW=N]KC@&Z&3#N$O±o%P[dRe2 d!;uO=:`¸|Waw1l@P.qסVK]y,_>pJ s5EJ](&\J]"'/3dXQP{*B9}׶['VFȆ" .ע=\I@|~T|2tOll[7g(̀]u+Ӹ$EˡڗCIt}YoAw6ḎߦcPןm:ee k՝ ժec+D) b=˔ ]mҐ+]ZJ!l K;CWWȮ QZ (!b֐ NW Q* ҕ *!+EW *rt(JŘ]!`k:CWSB;]!JՏ]"])b ;BWXpBBw> +M95]RWiB3cWszt( qqY9;b ruw9YmꅱvIp`-; q(>`s_PU ,,AQqty ͏[yT4+u ˫oy%H Y[f1ބ3vqSee ]v/.+vcYrp]w_ˉ >.ᮽfwἛܖp5hۥ#\źңA{PKXߣ9eY! V;틮l<觷ec߰limɡgC+aj++]ZЧt+]!`Bvpc+DYOW'HW? sn:CWWvF]!Z +LPB3th{_t(JpK]`lg >;]!c[ՓЕ\.]`)tg j \&:ERB %u]mW5G?3(=] ]iI6OpȾ vgCJC{:A2RN]}w$+th=vB_wuteKcWXt"]+[]Y{t(M@OW]+aыQLïj >BJ#l؂DOW-zӞ+]`Iig rB;]!Jcz:A-mv0QWW PVNNd Ԧ3tpGNWP׹lRӞNV6 "CZy PZwO$֥vTw:vf r&)K |vp ]) ;]!J{:A[ek{? !2#Ks;tV.+ )[yޮ\[Uƚds?6$/F30ZGDyl='c`+l3pmg-G?(yOWHWrM4kBVuyRw+D)UOW_ ] ^. 0VWQOJWۡ<t%{zhSk*+]`! QZ GEτIJ9?L&__ݿv>wڹ@gO' W3gB&4 Q&8Ŋ`e %|S[|JoH>cYHR;06IPk2*mFgDg(.Q)n&wpIL;A2MB*(XY\H%d`UD\:\hINc.j7E-tP9e0E4:RVflХ+pλg}M y?WOf?80(Pq02hN9~1@ߍOS%ցSB ^&zV[T7§8aZ\WFZ8Ccia'!JkwAEtE#Wdy AXu'yh`55=R]gE]YAUd3[U_;3uf[kwW#hji9 C^iȕSMr' ̗Qȯ\ւ˰JUS*wMs5._׮.cv8)pŰ?ڶ}aZ?s5~@W%oY?}VZ8´lG-t䋸E}mbs/nR]bOl?y|l,=r H%zGn\,ӐGsx '*=h%}W0EkhXez@/=Z3Go()0tATh5ht*\iȐY/ex4ٷʶ%Kzɷk%ȣGZ~pb7֊nr;(uy$<ւn*>Or^WhP]Cw^ !mMcC6,6ҋJud R0˚Z)CxbSYFxjZyRh?R}*h6i;3…#n9>^y,knڠXI-tN4qFEʁعfzVNlh^?7~`$*> Cd6N0Ic|ƍjɉ}t-an]N%yOZA#B)Vc!H5A9:?~Y^=k`@I2(1߻t$^821DF[(뚩,6{$G1ע@>lj!pU:>P*+XġZIkU jǠ)F-UKpm*.`Puùoxh+T%rM4ODgi %x᪸]7D_u$Ӽ /%={{LoWs/uϯ)%./zS̥*R4EFieRbK$]mDZ( ~7s$hㅝw:bfo{|hƖ91k7!#7X2F" _Gcog8w(8.V繏z%Xq*kR*|ˢXSΜ}ds t]ˍ/-5}u'q~f+5HYjTZ\&pW$Dc1ra˔Ҹ44+(cxeUz]9Nf>۫t.~}K3.drzxn$NbE;\43"?ד &-R̿`>3{?>r9{g>?٬n ;ξ{*X-IbUbqY9_WenLҽw݋|}{nR!~'GqN$K+_ty)ׯNioE:ٛ>&ϮkYw)r{zBJ-6b+}Z~zCÏ.vƗ'/,VaE~mz?_ք ^(|z2K>w|J X$%-5zvtzrSܰճN:WPO׳ oNg6֯h&9U2g~YcN?˃t7G}MLVTԝLٯL`lXŮhsCW-<],ޞ{ɟ~y?|_/iWoȅ¸o A`~0>|M[3VhiyKӮzNv~S>_׵j ӗkkjn^l]N6ayh~Ng77w`~S1IC?(W6E彑V8{7]~heu4PXĠyYIBL ZBxߤ5U VӶ{ّ3&ܙqdy@Z.Y_(]~%:ǽԕ5JG7ڨ4O&Uz[(u^J%7jPld}m%U7F 5'Gtu׉|ĈZsZ‚O./D{Y_Iz{/ﭨ:%'_[wJi}Fe;뤮u/ZˡV-BXml#;iGliG~O;(/'=vd4K'>ކ*jp=ݢ:{n7ZwgZp {KOt@jt+566gOkhž󚟎횟Gy +!_Ǣ+e6֩g@F z@ƗHWVka `ycYQ.xtQzHWNʦUGs+RfKW/vAѕמ\6Z~5dgt*T S Ms+$ު?]5*l՝G|3] zZZφόnCO9VDWdq1tyS h?P:tJ6hs5~S誣骣@W\zMurU|3_eӽfᯓ_{{`ug{g2!2 n>r}?Eܛoˑ/aytzq"y/NGcVzSPщl lJ4#L0GN_}i-]&$B} iθ~j2 IxQSJ CQu| gObUbbiͻCsEdת?Cf+ .Xx!=-VW *=I{?RObW1b\r%sZ ,RSm5ޟJq]ֆysTM.1ڜTE*ɑ F[5jւZ<mVG싦H)V 9GZGHo"eҗӍ vQx"KfԼMƳ*.zʰ!kՔŠb3L c~LHX*H`T˒B4J(+٥$j S2](!m`0`%0حhPB[ Z q9mxAYSGqȩʃMtS1A_ L %µ[M`sjjQ٠+. UPhc]V`TR\m/3*pY<`NCRg-a)-%Ն*Zgx@ً+ ٻZ8Z+c:!.H1x7&;KnR2`XXq26LA6LaLcL,͸P B@D!ՠPwSAW(c)(`! vjVÊG3!v]̕]Ρ'8TRp 3aź*Q \BA2B@ APQ{se*̙D)(J8I`_xT˲r /Ez@ߐPQ_ ()'NWW6!; u1j{TQA}n  R)ުӕ!!a_N3@yP" %cȲFH{ҰWmF웯yk& ԙ: n܊ ~ }Ƭq@r2|b̡:C8'_ h YS=p%tXT,}V2z89Hmi>Z. p1% n8_<*X ~a-<%EHiD(2 #(C V8Lta,;Ѹ<{*E%!Y=Ec<ԭH!3XhGWM`QչQ,TGת>^_E܉xn[QMuY 'QOmpxO#O檢82sSߧ֧^R,Hcޣ.dԟGsP6DR. ˅#8Bd453E `ZӨYBm.FT LJuXutxY+ jw,w惌%"F,ipZr- ^c8PS2Б%Ys EpQIbD&U(M6ɀ!rP+ bEEVtA W~T}BT ~X5bl^NP gZs,ߡ@gHG,Yʡjd Ԁ,. JP)z+R@L iorBeH@A} Z3P\LCRjZM`X9n~uy6ݩ-NۙߦM9^D1b``&Mf=*kpp)]q׎*'AŨRa֘u} rDq )Ek bd(",2<@4-P9nJȀ5aĐ 9 ]QnDh-&?Όr9VALvЌ R!!K HOO]PBz q^/ٻ6dW}"#s[?x&@ b d#xM IɏzIq$JlD8ÙW:UY^V>ᯜ8 M"Dj #ZCNFZy[X c,`c ;*#ƋJX~S%:M :# !Cѐ0'W22\{XzO %Xz L$$Hkb>3ZLD:R 5wV-9}YN!1mXv| o2H BS ЫpLP,mP4B %`Z*&s 8 ˳L|gq\|;z1As i~TPN5zlק4AR ]RPW c$nE*G/Kuդ-x*K4ٰV'hz&dw_osz|R뭛ݏU*Qa{qח VxSLk^K\cIӕ8a`Of֛ c Wlʩ 8X1ITfi$3(i IGv$(2 t$S|)?V9?i4~\Uy:/<_ı^U7$gW"Raօ+\Զp9|NE'YCC j)TC:Q]@~k.޼@w<='w'4;6ޠ#ji:Uc?guʜԥyݭlwv탚6H -~g Oхp,:Rf߯}@Idl>?ß;Qc8D[|d&W`hXRL,ir1kWwNq9qN@s1+ұNOΠ9!ߓO+?菵T:\zWܾř~?͎O[1.`4.*IRߛ->Z5uU2XTk䁦;EpltTf>9,^:]?*ƽ}ާf͊ۯ*.b>]f:yiyܞ`Ȟ'zH|ȟ?G|oS,l?}r,%Oxi3xM ^P)㥉4Vڅ3,i%]׶`y aB>-=Fdl0ܑaXj1[Mu}Sj=jImZ6$˜NUjW h"锨OV3X`V~OZ)dJy0;`j׋ݶ6%:M{*AuU-|y@Pn1m2Z5QFn-ڱ8+n YBkιd-$3w oeud|0mҝ /F03/wÑMqG,_Q$MnOO}TǛo z ) .5p"Q8\Tzj!oÁ=snlco$v%zהDt \.]97zO3Y[5flNn-wR;+gR%,.m.4Zh<:m2Y KͤH# '8 guNs-d.V hH Ĺ.󽁱0G DeIz~C&wv޵x078;'+|<>5 6IMVh-,:.f ۂ, X`!IǴXP+5w뱚P0qb?N b]XL`^s\0T'/`:VpM@:YJ:$4c;9ښ6ZHT2ZH%%20,I) "ԁY*uBjv!՝8?.k3)kuTR^obg<.³9+n/L\-,;SSF$o]q:Nv̮T*o%ϫiU8'c2rێ8" ;$gF'ЎFƠ/ckuI"lc'oMDg#!zq`j.nSiE_)-O 9t.mH=j"h|qDu+r=ZѸ7:=]Ӥɼi簅 P'6MLnnFr>Su|EpF1gX sa?w=;-77}(j\)9yb M}tQVI>Y,`!T: z0bޗYGnٽjFF~묑uokeoe~I%Ó?*AWd= w\T;˛e|K&Ecv~W:Gńu봫43~GSK&M';{yoyw㛷?{}tw7sװ e}OAE}+d^7n7jx\ ^Hv&-lr{^{ET 8 PENiM<_P*fp$YNg` nwiwI8Y[u~5.BOj/ihmLn%^He޼ȕDե;ȓR[lxM! 6ݝbRAݙ'%YPF,|, JVˎLlQL!6oML-29D}ϔ RR,"3er95Of.!Gk?-_r X~1Ibqf-VspDHX TyKE8W5:Qi-=#*ER@R)eRTKP1@CN,+43(!9+e6"YZ41q?8\_>2޺ EnTe&+voxn}*26W:6eS$Aa3#!R'deTjƕS%.Fܧadښ*h_H"Tϕc .9f[tSj p p p p ׋TZeP ]2F1e.ct]2n}#aa jֹ:X\`}y>hC$\"A ,tHeQ: 9MLj9+Tk%ikuk Y3G7aI'>S4]xk3=/<.nmЮ ށs~Pre2JOwCF1K)`_ EgdLe[MK,c9sޣ0*h!W; ,̤H4)EyI aڸ9`XŠ`me4Fx:$3ur(z`eܸxwVH;8VaEƕ•05?:4._] u}'PӿWm_տLL3yFQ)ImEUIů7GPk m YE,rk@N5"Zsbn14kYK{`<)peX.S5:"+N5h@\ٱp]uR 2՟kCkVk׌ =.QgqX.&yg_L9TP #ob|>k|K|ztgD KIc P k5Ax;483HNJ)@ڋQ->kA؇S"s %H)o+CE A+jMI:U_ߤoZFK˧~υ9UsV\byFa`S`#Q $6!Y9R&Na=ּnePB]輨Q T3[k0UTuVg8JӻߺO Z*~Pv 3ہKW_}R7_[\ngc'MaNv"\f%4Bd\^Nb;^ђ-{mH-OX"DJYf]| Ayy?ӝwE\BwSTwśO 7}3|Z`8^CvE=99;TZ$t \v&Ïx H^g)sd+3|Fzسwϥ3lfH[|8t"ֻv'}L;w'*n񲯟F./G|تAF;h|Jbv, o<[ɫa0dr趜 R\?ogа+2's|rO9>9'%ɤ2dRL*L&ɤ2TFZ!-$ǝ#u)a\tޕ[AI)n'+ڽ|W:=K0e&#QHYu2LwZ D佖豉hjd ix4lJ]Pwzn m!ƠWZ֣q+*p<}ޢy ƷI^?R0t}Jܾ*.QvYk!76N'_}ym\t*w5>1࿋ 5{^ޮڟ.79?PGެ 7Kxjkb7ur/?\46EKonև!\fEpYę@w@> 99kն;;XS>?2h̑W|ji??;C@.%99}I\!b_V~EK?ׄ0GČ &cu?%2@JF/tJÃ&V|+yMXs|92#;xmXJ@;@*[Y`\>K?(O9-J-AƒEjC uub燨` ;rhtϸ6&ie!9 qQKk[:+M9DŽRn&x9)XGEderbrX+?B"Ξ~ q>t\M%#QEc66f^81C_:A1s_fs 裐"9aFC9aes\9aesHchMVZSyeP+}'}7fN$:"n4heVn\qXИ]XJPF,*xMpy144mj14aL#FFlr3ΡV#v:b\qs47˚^p" k@>bR{)ŀlc>zĘK֦W>N>çv>Bk@{/$j Y=m7^K1b|)6Ws,P='w{d#rc Eh#MVجЗ9qw+ʄ*^$D]0ٿ+bzr7YG2q.6*e40ٕȬ}ñmP{4WwwFhFEl\nn2?bt^-e]{(|vwQa,?Ϲp5BGʽc#= +Uys]oo+ttVs/M ?h#~4#xj {`0-An!* z7pcHjXF|0arroDuǍQoYc|0οքh-J B))&`[)Q[M*Ffd j0DNl,ٺZZם&"Xl{Y-Ջ]r ~TA\sa!T5{^jj4WSfYYނ!0Vs`ŴYGuzMHZgl|Utf6*>\*f W,G bXC+T(>jȇHfvjmOSw:| 13nƠOTW`+Q ͌ >`K#S7PiۉNs-P _y'n<]f́7T|j@繟}`1l%GQ - *q]L|e$ +*Xџ'q]p8ZD9&r"VFuY v9lalaD h$]\W<7`dR@頧زzTFX>_]S|ndPy P #AeTfPAeTfPAeۀJsP&3/8!KI ˟|! ;]!NpR}}uRۤW^|koo"tE>!\ G ,qOfW=O}roigKvRE8']T DATKaijE#yW*NkDuZm?xBޅZ֫(<ﺓ y9)`QlhycQׯe ?>U0b'LnFq=Cś+÷H p)N vx3Bznɿ>􂪥*ueLԕ2QW&D] %ux+ueD]+ueLԕ2QW&D]+ueLԕ2QW&ʛLԕ2QW&D]+6[IMD]+ueLԕ2QW&D]+ue{6$_6{H~0>{ 618{kH,W=CR#Ç8#lqUnoI`GCKѰVc&NUWhrR0(m)doT|:rCac ͯWYV-q':8,XO`0r̓tD$s✿uve1ҬgV=h 8S-mAJ# $^f_3 _?e_3ֹ ,hLrmtB ֹ.`,U|ut_fPRָM\evj]4a obЗ6cUU_~+À"޾BvuoeY"T-;Uui{+Sy"l0[; U[3Ei^X-z&Nh9/zI[<.T :Ѡ5kU&k8i툐L/n|ӰZrocAjMoPIY T^Zv؉ /pݮAI 7 \\:EכMHe3sfε*R̥Ҽf%zqſ*{L4 WGR|աfWT<7iWF㖶gv5,*S>gvڞ{mi]A1luD[eЗ)$BMo3};y37őhhP?fqN2m >i!N<Ŵ@70!Ph9Ѡ^B} s_r,Q)HIvrأA1-Znӛ r3/HӖ}fW? t/qڛUvJ@x'q6r.ߡy4ӲP+uKx+['.|(dymT( V0b2>U!WI! tFM C!fH  ƶjSsQ b{U[NLˮEm,Æl0Je>l% mEO,m@r!໫궾ܛ_aML_#&`#y40_²ņ0,Zef%zeDo3g@(Wnms{Z.\@ev^3R;xUONIM #Me|%mW+jՊ~$h]]V+0껄oر oFW镴f M%3጗W*5zľD%w"ZnqfPhסRf 5wyePΞZ{Τg`ԒPnG:%ԡ64.(ԨcˣC@M]2vA7-Y}RJ\]nԋ՛ Yl×%7&mkU]aHa8OtPls ttmW }ߜAzt]ޠ830˿]V0#)6gF @yjow1etM Vӽ[ǟܺT.|Tdx7֭{MbvƒW.\W*r߶:W;.i5Ai 22һVY R*uM3މToSV:V?It}SȆ}TESe%`ŧ6 TO:O3e&#QHY5uaj3j84&ye4zl50DmQc͓(2s3AOnƣ~X"E9p%Umw0 '~x+uL2JcCƳbC@n~ơӨ8Idl oL<փ݌͵݄޺_ofz(rj[fV `ee%fҼSumVr@h54ͥX, &T[wsm}Iu,.&OW=}U5ǟ^y|~>|cMnXb)ZO WpQV.oڅ (YhdQ,ֹ XYqgw:+uWΕ.@.eM<[ǡW]5<{%Z]8/٭!v]gهۯ >+??H[LبӬރ̮&i)|`2XuU&e^D* 1v x2ry?_W6k'lcvN9emsInt9NN[t8 h,ZURX )gP\x}5=HP">. QV>BO{b(rjƟ:Ipem%]rG(goێ1$5AZX,^vGE`.0 $- T)Jkk&N˺?<^P2zEx#x+8>jK ^Ȍ ^`A Ɯ\áNzy-t=H6[6n#^gNE#K>OP fh/#hHHLBDN&dAɁ}AY15S@u4H`&$>rhAst< 10BB2PPg($^2#YobA% 9Pg H5c5Nˎ:N8MN;jk)= կmn%"=I"\#hsx^>] 3k9`|b+Y-O 'X`T9F$㑅H>0+Cʘtc[T{*?O__m7ݾkoxdul01VYE 62( F刊`) NPLjħV9/ZKώD֫qMfg=QNOG3256y 1sY7~Ky6sH[G“wXTg!_ ̒0A0ɵ r B($<3cĜuXSw H>ŽV$^֟Iw^!yT0ejNNkRS;zamYebh4kI?}\'>4q<_ϮQVQ~f+/b9nE ny;u8Q*gXJ󈙧\+*Iж[}@棥>SK"|h/ u= 1> Z>U՞Y1bJ4u>pJh` i֎HXG9(+&  GJII 7 DOo_g'HYn Z)7,x$Ḿ|64WR)>٨L^7ҷrPn)qa)xпO] ݼ]O9)VOGQ"+h4,cihgNr)r~-3_>޿}o]=LT} }r@``~>.ݧaTU˼%y%!K_::pYVS ^9"UcT֮lIG1f %9T ᢒ`޼" vὺ7Y24h)iL2Iɶod"ixTZf"3"_dFБCXk&2䈐OJo۩,QhHZH7́²ohI@tJIjlu]ԝzwPK!: >jj"14rQ ~3}LBr5o#͉r`r Hڣ8gGQUbMEEZϞbER#9F-CB &/N]M^M4 $Ro O\ &)V)bIf<8K$8 nW1@pM 1A)x FcG#Mۮ&Zwv$eU/< =nFKяաsL#}SM'.֠@uRV`S8N@ pOe9*XMrJۊVt0m VAK|2;bP"]&|7o&թ`MM;ӵj  @{% CH:`9bzZXCdBʘ8 TR$ω'+ ,6ׄ AOñeq e2#sO= pgD yM9HhW&SV\XX ENrs횿gy>e:}ЛtEeZ1wAl"&$ΘY'-uDE?Wu.P(Z8_\"{`g˟`mR$vㄹsp M8a)5ȔQZR]?&ߊT߆-r'9vMWWlꍧP6-]ߣ9oߦxWڙ[1dhZoz|?+t uWEZ]ۻ[6͌E mo:IOnO5.?ߍʻ7Mw4:g.͢A3Sj6|ƫ{pɛ۫nO\~}xˇ.+Q~uGU5`U\x-9Y~}\ a=<%ѴSýd{g%&G }Q1!{fa4&(As(Hg8 G^kjE4[أ88;&6FՑ) _;hXJ{Hu.HE4*2;\3vh])2 5 G\Zd4,2J'?=Evft\BD^^F^J[_*d>}stbSrg߼ڙ<5 Vx 9#ha'յҞW1pK Sy@ s{P^ y iIZ~pne>};^`4tYd'clr45rE0LT+]+Cse^"a&H$LiE`Zs"N)}qm?OUiAz5]Zs8oJCbnE*$h1蔸U1H"%Dp@ [ѵײ8ڵv-e\COoS_qm~yv:m@ylY m9GvI%wEMP "&Qn"72W v_e_%!N bJ8@YϢe {I#^Z JX64lmױ^^M0@3bTiʊG%Q,jRUІ5lZkZiԴNkENlyv`o}8aG^nݍGp|_F+x?w6ή)&Gi@u~~Am> Jyֱ"F*]d@3iKN* ;d9SWQ&u lz@GǰӍN&g/8r;7-ֻ<.p])]@K-J۝~ZKV.QRpm6=$h! ]Ylӣa􉗖1jybi<;="V3E1o؉52,ĢGPMO8rq}08;;"Nd&㹏kwz\ Wٝ~5:?gtz9X}M. Xd?S")mvz$ 1\-㡆O0ϰ[]P1'I+<_<vw$w0$==V|^]-d'!3Y4J1"$α> %7jRhG~/@,|:L{NĈv!Ă2S5ZN,X[i-Y]/E8W~rX=~EU\\df7= %N]^е7D$52B1ȸo5EThR /%HBt|R}!gZg'[͐>O 60F"SGe<׿?PػxHh}HYa/?)+bՙ* 䢶Ib*Q}+?Wp\63&sf13̵A+ `S"`$Kou(Bd]Tu܇vnYbe"('=R03cX kCwPk`80|F$K X, XMca"ЂG[u^Dk2C"cER#-CQ$T_шGo$74;WeS\(K%1E ֞#!>+ y// ࣇg5 ߳0/ס337$(Ym*bd#|g;CQGƓouXwӯ0Ix`w /?[P5( BL@/IIS.z0kRw8݇׽{_/ jE[ؿLAC[_1\Rs; `+lCecQ퟊Aʯcz꫃3{_CYb;瀮?1W?`6&j˓ſsEՐ ;|oE p qb6Wݵ&ݫMA ò|ɽ[#ǖXƒF JSi`B3l GwmI Ob8Kpbo~J\SBRAUσ(II3]S]U]U+"A]C%Wd>j7@E^G*#R"%1,`fEwZI%hDP4@2;vW = RKz\eGZ#F1 iZ.4Bߋfzii/ErD!y0<BRÔ9Ӕidkc\y)%y& Ԯ"z^ Ǧ9(aFIfd;MR<Â;|fq.6^} 7٥) I%_\Z('LbI(8^`ȺM6 ^ZqKHQvJ|!NUlvE"]ŃooV 0ɂ y8y[]Iޜ񂹣0~jƑRM#]4 CqUayxR(`bZ'Mrpd3 >jM6+f|Y{ 9_&SY{8:Uzi 7ُW3޼a^9\i)JB<&Ӌ2c>e5&5nyU||9(W?7}6|/?7oߝbN|Z<8XM$0h(xw7??m CSZ6ߺb>6_aܿ+ Y.|aBsojuׁLVc1mo~L#nlɝ'1f %;Tta}PvF_ob.O6ƾAF=GJQ&/S3;Oq; 4 1\ qj! )6Eȣ*(2XA :;R[¼Pbɼ@.(bV"B(ŒG@HDB싴 ( D F߄je >۸n71ҚLJ="].v'e+uB;B^:CxC:ocOXS'}>oVnG|1#QHYu2Lw(ZFL!N3MY"Z[#gا͇tH=ab>) { jnkVZÍ[p k -[Lb};R2j޿賂ҝгkDn%1O&9B ?Ȁ.&o;p PZeCm`ܝ]v{lDwJ0LW? g\2?Wwt\U^Ü|孷$*Y~t߄h3$__|뛕uHjS(EW\"]<Six`,ڨ#۠Ln bwIqzjCd["*^@1D9s1.Kʛ1ƈ(0gij|5*Cz 8٧\9`K05r(lfp]nVվP^=V՝a}jp/A-n5qq콣gTy`'c/΢^@lVC r`ypPh? /Qs^E׫HIV R *xn9x6!W<QX~YqҀtq;ȣ&@, HeGv֕ܯl5%!Pwp;]QOM<¦x*ta~[NG= Xw4 K^;k}`3H%uUg;&6H8Ki܅-J5:K[aE~z0YOt.7ܾY|f-i2)vȀl` e6r.|κ4:Jw`6r̓iWNƝޏ:"cegv=J7G1 ɿB^);u9G8jgV~gmIlv򽧠t^b?+AO^oV-& ʍ;m؇ e>'㮜~[~v)M{N>tatY}4u]H{fޓ!:Y)-lHup&1g}R^9GA{9N-=JhykFcj:Rjsfȵ1'Hk# l=c`=>>\wߍCʥxǕ' ar"-v S {y>?e1 6dI0ygshk3|(pYv wQH<btG <[>] h4mnŶrv@,B.=L3`p0D /IgiFc~ނ%7́K#k,QV8d}ݣO a' D=(kcdzaHZfGUЊ)V U;9&r6ȹMY:2,"1rbrX+/B"Ivn }vūi`)k'gBۼ>3Tq!!WUE\0m̱|><@rXo|%g 3wZQ"}jOL=[#͈Y.a)]gr T0(T]~_Ά\qXtfg8k1bҹVʼnuXTS}QwQycL,1 r B($bh՞FY^‚S~V;[w% 14I]-19HL[ӫض3zq:.Qm|ZQqɢY~.m21ɵѩֹND,*xMpuTZvu@.k8v˴.p6-jlE3 ,^h^`O2^sktOw[[>]`4uw*~W4'0@ll"  Hb}5@tAt ._/~K%yƝ+h%ހ\IF`鰹E8^Zr0oxv-wr0bQG"`"D֚2""&ZH0<)c"ҭY"vNeWj\h^mn?fĝ[]8y٦g}k;>ӘG"h@QAsglrD`p'%O'yl^>;ZDa6e.%sHϔCRLepB1X<ϋvfnE-?dB!6:Cxѯr+de4 +\R캈ʼoo??Kuv&@l0Y[Aȁ 94!&U7NrXLl 6PE %8SQ"*°6a<ڡ6'm:q8.;c:JG2cU`k _͋1[A+у?XhJNny~|a/ >MAzA=SWr㿐']NLVոQ*|,ZUӗ O>YR8S%=K1_G}Tz󑜻Ţw&`.?έ+Mo(rlgJK¬ز^oz9HE*L1OE"k/%%T -XZ 1wDðq I | i|^v$kI-wjx񧼾0p{ܠb|YڠWYzxbX&z#˷wjcIK.6na=lim%䳉潕;$+-&˿7KZolyWn+ `NT|oZNVZg;ݦ7`DPհcUv;ΠOyӁY)&Ǒ2PW][vuVDDI}=8Wq؄tQ@YН8i!7/Yl­rǍB|7%kg *}/9E0%G^)q\E[U:F[Xc5GcO<\%j?ts7W@s%CzRW?; jz&qqk03t~\D#,B\8(\.!8g֨\r Yi8_ ϬIQ |N g@UBPք\`b>Iڂ;ٻtgl Ql($B3)S\[/1c~Vh**d?MvLa]2$8I!G+EyX ٟ&x'L考RʾR}Χ7*IvK^? nr=C[lפKdv/%v/ nSW3 ߼{6JIC4WvtalʩB6 g𿐢yM "9c@0%;&r):g,Qԡ;cJzgٻ6$U2RGnH:F?%)KRuW=!Cǀe3쮮ǯB0&'Bmc`]Pl==Lv.ݿL2J)( `w1Q­E>e@$8U@Vr lNM{|s,QL4Kbӻ q.wf|c`Wpʀ,:ʠjkѨdb~u~]]7bXs~+kNގ LAsY $cLK21g ~S(fNtq{4e\ ,EF6>7oֿf;%AŢ.t=Z8Q5")xfuʺTlɖڸQO-qJҒ| |8h $m?Ԇ[4F(1KHD7U0m)fqDm 5e`Iy`ꈪH0Sy'fokަe֊}W_|oujzY8&MpRZK*)r:yᮄJ/ۤWe5ѕ@vw._ȁN6oM;+5{^haՒU qVA3A5m*m=&!h4=7%}->C6x:d#hMq`_/d@ėSUH^VWmx@@Gb .q{1K)w#1u77>" ,LǡK7y-R1wjJu@J_`mPA3u7ǝMr2Znۘ=b-v>JQh`vAb&js;Uh 2WZI$Ř+ . 2J\ե+pҘI|rs,'6W! vbt,j);Q2gQ K2WJwjߩ: s B/\%r/\%j:wsT3W_"DDW7xw MBޟ~釿tazyM4&6Zea)BX5o nt;SitIiv0Q  kh4 FKJ raO `~x$h`Bu ?bA|i6VB ؘ͛_ Z$͛YDVdI5YR\a:2p:wycC'ZňHݒ,iFT{V i5jķ]"t0'!4vr{_% Q6ʼn 4"c gb抐†p@C2NZdq0y8[8\} av{ſ֘i02,X3bs) S!C S [& sGBukH'l3Bb@d)4$VI"Wاŋ-L] 2'.f;ˤ@c3se?)2IWM~JC mTr3g6(Yx"ZYt } zYW29PuH]$Y^"JɎW7E~g39͓,S>l&b%v nHEKnFoLaysF.s1HIʃ9'#0giD[cR;ĬW0jfZ{ށARkn`OLNbkiCJ~ p%KͶ}ż@[G6%IQZGZ[_fQ S2y 4yidl# 4mSLF6,0Bjj>}NEK181`#\+p`ɣ20 r$08F&&Vcp2(:o;`&ye4zl5yBZ"ZFNs}N7gm8skЕkέCZWWnDksi5I̋^r?x(}",ϱrw@ZHmyXncc㨭A@*YMu`\#Z!,/2]T GeYwLG@r` iCÞ:tV)dPg($^2"YobA% 9Ij$f.2iCjB̝ui155JbNP_Xm{a`Wvb.hR;uqSXgm_oFC7jR'/_ߩ3`.F,B0#BXYL)#"b1h#2&"%OQ9P%xy!],5;9xT%/vx~01VYE 62( F刊`) NPJj%\dfibZl3/g[hQ&ퟀoR%XsLv=7HO9IK`'j nn5^Ù?0}i^Ldv{Cmo~ae.&;&wKK|.z A_(m_?%oYbjzN37hfbN1V֝SVe`Ӊ'jDj$WgԶas&לȖm6.Myiʝ[ wno/!`qvg ܬ>qɹv%he$\BB O1 9f2@#A@䭗ȳ)FNӦ 9?j ak%( \+p(_wrxw.^K 7ɀ5ӲҢh l92P EJ4@&Z-D* pۡsIOS5Zކ\$zwD@4:&(fEwZI%hDP4@2Y\} & CQ{׫M{k8bs0̭B/)mT;?OrMh@*1r)'s)& c\y)%y& uԮ"S9\SA8o\ Fv~q &?g_Z$`5{)"rfC9 1LXޟ<f^J0"oS fiȥ|Ǟ":yɣ}-&Sk Ǯ@ N!\ՍwWf: wvx5 ͟*MkB¥CR2F\ 黏y\vx1IQ 42> C{c9ߖ-۷׳u0%Qt5z ,44d$|mJ|=R02۽f_p?'??xWga1wKۂT/y[  YfݚႹ0!T=1nH}7vS> G&.&O>KFO}ml$h{ z+f4m.RGR/KMO~^_+Uzirw 7K_SB>yL<◶ 䊒}ϗ_F| eM֕kN5C C=߂?û!|?~?~ٻ6W`G {76vCSk-oEq(Jǀ)i]UTuu^&M$01 oDpo迿xЂ54WM[ch|jW 8$v\w C< 4Iq)ɼF&vW{%8cJr4 zg?Mc?Ff gexPQC{NjFgȌ/3ۨ 4|Q\LԄG8`dD ܢTcc f, uԒ0/X!zH@.(bV`"Q!05`>)ؾH;0NR^`MV6R5w 6/ y+7ي".J(.wl1jS [%ފXJ⭪#Xi2p%a߃|zS"KO_8*>ώp-\yeބf#%c['eo]* `3rKwGͨ[ٌJncL Q2HՂ %xY2p*.AEp!W poVM 5-G9ɳi6Lf PI:ޖo;+;?+/el;h0.) x:+L ,ˠ5;pvrjnz=~%o`z ^ Х:d럶p9\!rLn)19"vwhxI@캙S{Rd)ZB#>eңBY&w$1x.8YdJ[+4k1 G e9oժڡ=U2q4<"|Ba6flhW>j3&vQDEix{ J\XlAtꤳbv^.dJՒ/Z76lXA%((q{D! )xێ[)Pʍ)[ r8X[: 'RHDc5rƈ ;Т|:ZL+~R9zuQeC{n,j$L5Lqύg2* \0̱|2@(X; 3|%gS4aw tW:VaG'h!{@k~iǃ{>-@in?rKkE2`u_I; ,hLrm:Wu"`wgK{a> Ls^Yǹ@,:%H>jnx}x}=_¢yhrt+%3LdE AHRVa S띱Vc&ye4z VhnDVf~?GGNUF+6J#Xa%<L&h'C8``GMqQlh IMR +,)vGES2~]`FʩR`?ZHyEŨ0\%jF)@PJNtTn8G0`&A[9L"b&H%YlS 2Ād;K*Ljd<)ViʈhA #(p2&"fk|0ףᗛCUO?7zTq+_НtrYKg;]~0)Q:"ui-hYĔ$-U:-aEDa3{ͩm&NW?R!JsYȕ!4Ox@Yس#S;RrGGy'v-?E 8`{ӗF=hr鳁6Aoy$Dl֌z[1ԶԺjc0E;ܻR/m\-'W-adY7$l-#&)A7 w[`i 8]>S=nyokATi(y5n2w| ,wǙK8c̎1ǘ uǡ |ugbLqxZګ TCD=J {aaGْmC%n&;$)$2 0b;$ʀ|_\)x'XR[+(\{f}m'ʰDPp$U! Tzxy?pe1ocJDGz웛k]4m}=}͗,NV+uY;3CQna{ےwOqeV\#(w SБi#0sNqg76hq#"uWF2i|L%m_kk㹧k6oZ|`)]~B9〉Y}9Oy^ux{I؇#@YpT|4JySpF\zz}|ҴNGa3r5{r#j92ufT!a|QP&(hgJ9Sg-NHfP0Å^6B½eZ5ܶTp$Ϧ:HݖTȆJ>iT|Yiu]-~Qݬ@f4^}tdeLRޗ_/dek12Gp΢QbkP8z-V] NH1 5 dY8ý`L{C"@;GC@RH'Dz.w}[ :# ?S^fվ+մ#9"< H%FIiʴIkk1<< [:Fj;>@8_;f =qI8^ߗ";1yD.$~."B $1V>0k(1&8O˥S"/{HŒ/#"![0&k1V` $"a5:B،oWǡ^S @E`7! ^1I3r LHϝ~yZEa0Ic,>f緡g&υcɮM!HY=+$dRNĒ(Q>xjC/{4d4{Qs 2zBb0!LUl܌gtrY]x1Ζ&3b̗ {R};s"Xj7I f3PHT4򑮚!8by`X(4IzǷczU'գ4jӳhF:N40pR1ȯ% >ٰX֩ҷr=78E~Lrø.+ ]B(OM~_nca2oc/.baCI&aX<`l`'ic, *0wH?{Ƒ MFꮾCdY: aU1E*$e[9TI ICƀuᴦkz!FgMN$6ȥT-HE3)i=(([]][)2β7o{?ӷDW)S! *UhOlM6}T粥@Lv<(Jxp^Y mJ{b ")eW+ZAǣ= d< mע ̕r \|zٓɧ#7Y;YyNf)"sΥ%I~E^FpV"sOEijTCy!OW U+kso mg<}cUY80oPIk '܂#hIΐLV>Wy~Ȥ9ʧ@\dKwN-s030%2h䌚o@RR2sÌ7`SAQ xf1ZA.{g#TiUGaJL7ic : !g׍K\^ҳf[kq}i+ W"< pBJ,JBTJ|eD p ءB- (:u Օ%ee頖hF]r>uU]-{ MS:uuJ^$s]=`yE|Vu@jR+Dllꁣl;uU~(P^]*:u a+ ]QW\!E]jqUr@SWH]!]ʬ!,0"ͽ9uG_KW`c? o1%2!**㸨J-dcyzxj9}<9]M]a9gINj@Ђl*PȘIO9)r,-0-OAv+ D!20i&FaNHvK( BodT,SRM0*jLGJ1EL*ƶ\BP"L}i,NgaR,4KVCI-.t;%! -#6*S.-YbS xIqШP;K8R^2jEl Yuhʠh#M*i ENZf" 2:,Ȅjb3}Sz߱m@kLqqܠœAIunΧ;mw2/m[԰Wd4H*uGU9INZ;4rP%\a`*s}@i^TՉu'yC5’` l "": Ix1^A!8gaфaZԚq/Dsѯ=.I`Hs}}=*2})o1_]|t5ykhݎRᤔ]95vKL*HBٿIz|H+)? bݦ-Rr6-UJ!ae?.asY{8$O15eLJ¹MwRZÕRPklPAfE| )DvI(&nUld E!l,BBOr(Gf&]-r֋h:Ő˖;]a:&'x"zG{+^ VޠWjFet1{ϊc_N'"F꘭ѥhrFST2)5ck׌Qʳ8㑺PutNu።JӋnh5a$;6ܿ F_h60"#QXL39sFJ^aFcf6U9̂6md-muz%a6$,K]F9D0L &aɐ P !}|2@Tε!awܗu2d̚ʧu(fAD<AM*) jB Q C jC:I$]JAKjqp`uWEEt0c\*(Z8 anD|yA3,Gr1zYZbՂݣV v/Z-xpℭGÐ!s>h]Fi#A)'[=mȽԞ 5ʌ) 0>& Y\eVq<9#T`ƹl˧-aػPG߶=}ދv i>>̉î~ȟonV]r٘O ;|Ow^5 h-Lف†r6贌V`A ߪ 2{U(k$Йde(Y'nrr52(Hr`e n|/zr\-䆳k+_]U6l+:y,KAM2ݷ#7ɎWjG}n"Nex4 (0oZ|XjE7&B@'9ćn"bgF}Q̂}}p80!3|D>ܯM9V)ȅޡrl9=wl:f/4fBSf"g֢lZpI$xLA\mM;- 7-[2!N7::)k(9QDFLQik9`C 4eYw W,XPj={IG %*e2c3K+]h'o`i*-mօ6I+Cs8AEIw;mq҂o~\48mL3!L*$!#cKM/XkqEۅM{7ެe)]3sh=~E>լ7춡AFĊh**T0 iMj@t}D$}; D'`Ku* c]z4g`e"[ (d9Q*E)`˔"T(@L@(3eΔB¶4YJ95- .wB&V&De[୑alM*(>.y %PJ+?]Cj__^_yWJ7{T mW$;'dd B9ˀp[x! )%,8IA|* .&-H*KdUZKj5tE$"0/Q7cPĸuwݺ+?DkaL!o`!`0ɸ*dtL>? O.ښ6URZ&g L[zI?o\WjS&5y&O'~2&kB[{1PS9;:JÑLҙ\MώVkFDX9XOvcɻT'I6~r"7^҂ç/;r1Ş Sw9]]HD+n~&zxzv~QQ.YI; )Ѥ!գӟWrOP{7i>7^܌kʯ./{A3?̃lm^%GӺS7ZACƚfq&t4 oFa41QeN>. =֟z8ݼuM6=+fvr2R&>CńW1Ao4;}^ÕƩo+?.0?'_.΁jxq_e+*L<h|Vrü=zpkbC<4.q9lcn-‡ncQ=dE7-pqV?@} vw?5}sdb`VFfA[}~&F?uuIR`K*lߙQlIZ2G09geC9Ψ{jBV=S,$]ޕq$BMGex8$/,ЧD"e-/"%I#d4_UסQDRi"(1s}\S$ARKF.5$EÂ>4m6WQtv^kouj lcoʼndffo5q;σy^3d` Ŀ2ZNzIIeN<7 ~Wvk:x-'1GTVbGOVNy12I8R[葫<ș۽Ois2:nbÃJ"|7n]/[G φ廩0fgkc&_/:Vo,NUɛD5}۩w7b/Ϋ_TQ7Յy'!`b=촃s،nVN9SD4yri׀hLx @YpBfw_D̼ v8|=ɲu〺7OipER BPyK!6IP&QQ)IBֻ!ґ67ۘ1;OzwɻWv2ظtRV?N>f!RLw"pUql7o#w"ljқR_^6_dw޻_53xW~YbK- |۪cIF/zYS/TRLD sE"@ HN:$YHG :{R3mҰ1O)6m*/hD$HXoE&XBH e;VP,m#r!(\X Sp i1 A+R2=u(ujb)2aJ*S1rcD>$=m"w>_]6ⲮCB=AL ? }jsvA!+-y/'WVDTu@Dᢶ 2)kQuzg:(hj;9Mn D\9=";lJT F!VZ5  4q |2 Y]1r6Hur!}!k1iMb*qJyo|_kS7Uڝm!ñEG{BGR9{Qڋ>1.@lI\q)D0*HD ι+)"r<P-2{B8MF.ٷ_15?JЯcH l 6הeV32y)nz.vz|\>1?bi},~ʏc;8q۝OX]^SJ7^o\CgkNU5w.m4y=M_UO㏩kN uAȊ C+Ll9h0!\bT8@tjZy"Ӕ@1ldT6F)D;MM.xPHY"!(OaTɆupmśI԰+aE] ?Z *\JHUfSZ>_֙]S޼sj鐮MXuȕU3uŵ@h}ޓztdO?6MkKUmgv4Z;ծ5n"ߴ湒p0XQ$huf(O*vqe69覡8SclխyVCk'; wȒ{ 5p 2 # ׄ JPEbI(CfVDV2.D0q7 {O^y|W{{w$w#>쎣(m!~ R?mzۨ=8,Mxz)aY0yY 𷓞80-: [[k[[3[[ Ak y&rp Nh4Ѡ 9- N֝w\ ʎOsWPPp/wlۀyܞk-JQA$X0~413%`VC11\ sL!Q`ɏ'$j&ȱL\DeWJ)[v+ɵ1DŽ$7d)vG2:TEnîPFL.®22ʴkdWQw{Ur;#ni$v8PcuЏzoA:9#eQQ-B <S\b{:Ã~9E$t0 ;9{X}?M}cl/tozO_nAaU33[kN:MBdCi;ML,CmYZD OL389̩~h1lM&x $$WӣI%L%@+|i1s!~<\]ej9tv4aWté4a{gW#X=Gk{bWRbW슶SOsgW`]er<vs_vnϮ2sc-zUUz"۾y0|?sG՟1KW$K?'^PB$W2"^b-sRi\ Qy_s|ڎ-$FBX V+O*pPT`N h2ZK[BD$>Fl5@H_חO<[SrK= e&G"Q1FO:sO/O[` ;IrS&hR&IMДiI28<d!QWb G*b{e j{k}=Fp9ίt$ .#BH0hC& z?tE_#zPD|6ojnIH<9шzwصwؙwGK1&eRHs2kɩ zF"i 9딥B6{# ݑ'S<5mn^N υf "1} XC$b'D2`5)nF6-.<=ۤ7eճ~Ț-3Uk mW \A暩-ID.J=W$DCt i b4%mMS¨bJsIR"gfL45Ӵ _H2Brpb+R9|ui(;˺G 119aDs~D .^戀Ġ NqZ$ bl>/?d(D!Et(W.? <񨗏adjı'\kI: &볬K1Dii%JPX1X<8-E!2a.(a9&ȸFPݣgJKٕ1aDN&Mt`P k2}u0ο'U.7=U}8'x׹'Lf(ۂ)( daPN59PcԎFnOGcdWgɻ_-Ic,Cl$8Byt\N.@uEZnq@/*Rn6:?lal]xo;׶d$Q‘}94D/^ ULKAƽѩߨ69YׄFɬkϊK6;qZu~}?|}59KxV(ʋ:a`xUwG?bwF*b]MW%?.v^_&?=7?o?O9Wux e}ΛXG$BnA|MKT޴iaiדߤ]dkvėl_hOuz2@\MB+e}2*)ud>TVq F_@y p.ƾIgmEx #њ VY?wSSaa"i!3ܡ³lHꔒĂu]+Z;2E?d(i_=J5hA Y{9ʾxO҂`9$P\Bz[{n=V1ؚXP "n+6GX{r݀ZYt6Z[{[1f$kj1bŧԄ9 PY_b]gRh[s֜N;OQhbTQd(09BwOYpB%ldbD I&ijb (w[ PK$GDJf[#PsL6#gC٧͇dȯ=ޖ-a]zT FYq\xA($ݢ|UFtZ(NcvMՓSxN!ۘȕ},Aoqm3Z]u"SxtEUzu8њJgeؠu,+Ē,&ujW  |_iJ^`E}ӷww}di3]k*課 wYg^PTΖӢ*DkǾX,K(??&!K[+I6u鳅*?(vBqj@tʊ cu݉(\Զp8mE'wOL,XP"Lg?r¼n tY9m;lJR"\TO#ur)\ІH?{Fr al΋-bxH-IYW=R$!)I1`IiNWTU r`'+1` b1`fgO>]VHiM0DoSu`L konT"{l8xfl7vqlщWZrmˣ"}g^Hor:ʚkޛQ`sϦ,DG`*WG̶_o7a=_x:AZ|&;chVX|(C'.:o$Z{w_hY7pcr0 :Y3s^M5.mND:~FU8MDRH4 G A{ vxynQazð`>ɑVr؁D) !5(O+LY}hr-"Ƶ13.ϳ2uJp8VLY䥵JpPY}/75|CZ;byEȞ q5W/v"KMi" |K+û^ h;kH)`񅖜N)bxЄނo=!3h3'Oq:_d ="wfTT 1 Ѡp \ {[Ƣ3uLc h3ł9 ˭ʼnuXTuyt8ҁ(wp)B0cF@̇t\2b" tj$Yy r7#>פS)26nw Aôq޴^p}4=/_-13aSy}3, usr:,#Zׂ:IrsS>iqFw+"ä4A;ABC^GMqQ{!IJE.h*_> ̂HB9UaaR"rEf rk?-5MH혍OvJx*xԎM>/UKVERg&J0Q:u:{qI,{ǽE8R<0泥(o2d>Z#?{P=1\-&x#32x5sf"(ܚ195zr]3vԅeօՅՅ I}-X\u<O<7#RF<^xsZ`QD`m(e37DY`g 0QN l2G{7Hcr8%X+c2FHHPbXk$^2 Yor"JDr&$53rXM+ivHI;k ;Pd̍~?gx]Jt`%By{Jl`z斋i`"Y *bs)ZjRG7nDak0fp a[& sGBukH,Z0fGb@20bQ@G2Y+냉4eDD FQ4`pHiFΖ=Ck\q7|7V}}y7~x:rOT~|3i3W㘰?^(:W,bXs *XXvr^saJӘG"h@QAsglrD`p'g}rjĺݳ{g+_mYujbbaey"˒+u$2O"Rg#J (-gPЂX!aqgkGP;R7@OP$qJ^T@h\ B\¹=T3NzS}<܈s/3w7u[1Sumnb`qn|[ӂ?kʦm9սm[6Q`tY+74Rnd =!"}*X{)) JnHb0$#9&-, |精pz˟Fq5 AhAU)JE fnƭk[rnVTYnqĜ|_^{ӟO|oqZNZo~:46TtOv0/)0G!/?u?j9?,< +0EI@(Kt`u%\Y(/3;W'߯\E2Å2.$(.ЪbTx}>ٴy$MTJ=uɥoJ(љt5J,v) 0E]\'2ŒF\Ơo/BOսo%Pe, `MHuڹtH\V(_`B)__%. q"ca6;?c:PF]~ _:wT$,̾~Wjׯ/Ƿw $D R U+rUC HF VܤR}ëG(Ur: cZw&S|,uN+o|w?.g+ah',5^LǹsOg 7 &+{b|uOuݐnfYރ : f0bZޗDv'맑NYkXA5DZR`lHm0t#_9/ &+a~4NXJ?t~];rd×i sϙHu ݴ%&~s4N~+gj=QQ^yD3 rL43&Q6Fj"HoJ o[dw0j4&|rzC=Ŗf|7'odgo|tYy%'^ VLjڳED ,Z^(E!,Z1*6^lT6^کhol٦T%((q{D! )xێXFbD9&rt"V-:hVN,%ј#f#g3F ],h&E;ZL;kutn=j>-<H̥B_._(ָv}Rix`,(U0xa23X$ŤEgCWLaJ*,EL#FR8A&FQ`΄,+ C!fH  bgiKz\Tknlf-^.)oa-V G<igXd66cy[آ#|wuX~&A$ø5ɑ݅:Rj fcv9L0&[d2"!6%+i]dCYԀq허e3 l4@{gv8"/;R_ouI`žaU",E$!))ذd4kz^KAuas2އnz.vz|ew8.x"XG8u:M ~w'm!d2hrIMoa%1ǯm%aJZqB>`ڹ+}fpTۇ e[GuV4JS^\1ɯ6ʐ]\Wl2ĺ**ϴ:WYn6*8Ҩ n(rF ;FH>񔐎Pj:4%PYQ(Ebh18o )K$)P5ܢV*cˠ8;fNx1y ڒ?ՙ yvCgw49ZC&x|L5}3cvSxziC[ A X[V6m{~ԣ3Uf'Q;kɞGWћ쏶L:[| ~.Ni{|?xSeeFnƇ`au#yGi{r[&jRt1?ǕP"cǟ ݵGs[1/Дن@X۔;d֚cJ4T~Q/洝Bxx+f**$T(CfVDV::L2Z݉dG,N%p"ՁD*\g tFa,odŴX*gHŸNzIN P/`{ڌA:QarB)$HPi)hpRٗE/WWH!f^``V~_pAl%e?[@_x8$1K$$gBu>ʋҐH9*RHxb*'y`NEJC_U:C:L2t &&%ȏ %&2` ܙ!渏D$4**7F+Bkb*cSs6fyoϳ2n3߯kmNV;8^&Ց%Uz\"SPVkr/5V>VH0ڊv{5Mg!~؛kdfX9Vpnr b))lDQI8N6X@J)yJS`L=Z` y[0L9y VZ<-(uHr}ZvJL<9`Q#wGheUJS$uAmT5GTPEAUg)jd z:P|q F*G0Q4"QEC AJS ĩv!ڄzdHfEy:鉶6S1T  wǔgނ a]?gz-V7ɘ/%|/UgDCKq" C`R6Qȁ'V`ƢR7T$52J1hښ"uI{ST)Ƅ"Q%Ola .v ^+g}tk@{'K_m9M: ֬W0fr|BsXSd, / ++m 7PlWGcn脯LNyp! UEf+rnMf./M;#/C/rYu&Eޔ:i }z]00;6YEحUYz{S_EY#,nZ9J >P1@CQjR $%Τ Q( s,PY |K 9onar@<]CLZ8b{wfryYZn)s esEn)LX-Q0QZxSJzz8vQ Gଛ@:mKLA-;&.bo9எ_IPޔc 8pˌwytV{~l Mi,avjiC|2v#9*e4Ai#|<ݎmgS<wʊJ utJ_O;7;CSրUIQ?-d8)`}ty|V- Q "I`67Rnz߲b+Bz݆o g 5'uWakN<ZmSsQP*{D ֜xꫧZk`'Į`ɰ ®2]Qma-zIB=^Dܛϯ?O.t~7 PgV΂r_q!&S.!e^$4:TweDeacVP*4Aj=y<(y z ,z&y̛61u.~Gv]FP=Ş'y{ؙVC Z(gR; &i *j} <:m\ac9sPX18(8ผgE2D&L hHb7)&8 O70/hŝ4cSUȖŦ@1] *gW`kY;r:N.r,h-\!YC(e*Ilr>PL(RrC-4" .xQPg^u3%#SJAiP^V*|CG>S]u=kBĭՓZs zu炡:yA/uރF&dNHÂDzK#TC[li'BLFD%x(%!p060$B8gHNSå&! 7g,.\Iq7^ZbQˎlබ{9I4V;?AFרd|&`FO|2SM_usmkF2%g՜͛⥊i)H.qwK_hB#dh ý(51zgLX1K}㇛El 0@(se?w׽-7{`(BT~y7!M+qٸ+]5-Ú.,Cj;w1~nѣ5WSUgٴjӳ"k&:a@qF#yaoFȿR'Ǯ3LM=z!o>?vG>q"4sKPW(#q?t+ܣj}:~::*! *S9ynrחH$w~z/?~_?@oo?< e=WgM << !~KK^ji\o4װҊ|ﲮ, }u]*,D 4A.S>&Ec,M<.П?#UfT5H$z g13aPn&ط CY[uHĈ@+þk42;8OJ[jfѨ"q4A YTr}\S$ARZ'T'QejÉ'ej,um ̈Ιkd63 P #Ʉgp;PcVHSҭe׀ZզZR%f:qqTƛW;Z8Q'఩>^'noM$_׀!m\;h;g@$9iXޚr\RN:CˏZ+˭= Pp~XgMdz ! cMNM"Q1F;OtO = |pMDgyߡ2]@9kdDq JL{ שWFqrvFYSČ1Ms+@㹝&4m69 wZP@-C َHЋL{#(D ꨬmX6vyX:jY0Dl>+ivM+)!RQh(ʃpeDG=: M\㴎ӎZa{`녵⪮eq6uq{ hT7Hfb_wIkK]\ D%0˭ZtT(I]DYQv$n5GQYJ$Vehb:oQi Q'b,M*HT@ AJS ĩv!ڄdHfEy:鉶69Pc5qrM>O_@E ]pi`лEڙT!(?=.Wb3軚&bfy_܋FF2iѡm]][#ìȲo:9J >P1@Cn@ jR d%Τ Q( s,PY|| 9o)abM<]MmZ8b{zrc65Rb9x;|3;pYwZWf)^):x@S I諾P;YnI*it; i&d\SyN[;$Z_* ?o bt d>V-?6{6e cJVd2oҞ|1œiYSiEYg N:s u_P*+j8y:8il q3B[e0D.:ֺE \F\epE\eh:vq4WoP\1% ֈ+WqeGo]e(WK:qĕRgoWE@y?"DCYa5V>H0YO:DIL FP,&90EIZRh `Ca_F:HXMtr 6w(c ߽ej׏}| ߽,~b4pfC.WKYG;f9e# r&2@1V ΩeqP&O%KPh8]1ojJʙDqrzϢR(/mJE`(H H)Nz!pa4W,7%_~=\d?BD_PM~k8 ei }4VMێg%+ppkIԹBJs\QT~, Jss)ˬ? H8 :N^`jo ŭҰ O\BGNwqjډjv1êq0@K{A-7w]9(BRvÅ.;@HJ\֮.aH]fX>!#82 15h8] zzf?۴9UFF6dݪu{EL:eGIr5Ezv)=QUIc󋜥)oW7? ÏunAhѝ,A_UQ&aWz䲺O{~>cU߰UGz2~?1ٿ~>~:L}O?o^fѣ: E,Ӈv_ZKs ;,H#|ue[Y*!mueFSQcl[O+_PhufT1J$ˡr, 5("b:$xbD ZsaO"_m!޶SywBhH޵$By T `&. 2/ }I IY>xxm}$[dNꯪ2RT ^(A-d#39ge):9Ω{Fտ^b+*.[.PPt8HC#[KmHBU9JArWWâEjϢ{tdG]vqZMʴk0`WvCeZgeV԰VkWɧ}UDb  *E l;_pG?}D{f-!ko,;K3vwbɚD@Y9 q\ "<( 3d0{eASfMff:'Xw%Qk,PX:E\WȜb0OM;< UWYSFO&k'0_H^s ΥM8&,B%B{:孢3vI{av7Mk]Z(qk6ο9̰+F-IGqגZ=WmTFQ84G9(뗐[_}+]"SO2F {IL̊(<^SN)V]^Zʧ/+;I؁ԏGqJ2mMI.8R)QdytvK+d\邰L,(cT( 5ƄJSkҢ&+[}s[Ρ74徺㭽>{4Z|e\S%<3y8ZHhKysv@.w iC*T;:\ *#l5P^b9a cR(s, 4 e[ H!Lĝ7ӵ*bB-۴T G?66ާ^X{MǕmw^o8ɍ>ua0|pXϛ뱞G|c> U]~Gj%X_Kmܶ?,m+/7i ii&5V˫}<<G! Lk1K'Ind2ZOy] y2F%5gZM-*0d )P )ELJhuf"GwV8yJ{*򪉳c^6_R-4d-Q;V 9@^~5AxZ$+p덿'cS~tkXs͂gKv/=4Yk{; Q0|HMBwmt3~bO㵆o &82LGoB@P68ʼ߁m!>}{8x7 !>`EdE,D]DEmV)~6"SnD,Z5\KTk=3mYDkJI`?4QљL*!8Ʉ4-f E&YW r]R*9.vwvm7W}tyX*V&WG;O2 0pcV)2G_g8s#xʈڐ`R+x~B4 j`/AwEܤ8bKtY`uI3/ :DQ)&Go'Cʂ2Vq<9#3@s>j5qv$͍ ~4|ozM}s?4m77伄׺Z-Y"q阥b ;?}~׋2iMKh-޸+ӗ\g X7Sv2py^W}3_>;r80-}5Ep׸ %nCr/fɖuzE Ӑ 1NF9 FJ+8jg3tb<;`.zS6NA+~榽Qdyʟƈ}FZJۻKTaξ`M >gǝ^^^65)fSgu-zusf╕)0ċ1,]wNJ זZ62{YiaNA!M:+wIVYtv֚rCs#NWvJȏ7#yB3wmmG$m'dx5WǜN `#-4xbfI]쐖 ھ,.cn;baLw_6||;}asN]Y!j}f37C][A]BQt[)pR!6e58mTS!0aT6 , FS.TVɟY݈ !7&Rڸrua۪f3Op_R; ~D< Lߧ·~X@TY^>Zy ׇSq/9}'e'!*ڇ_JyQ:R\˟~:FZIs/HԋF2EJV3FϚMyEEװ߼ad3PZA\c u: o;NbZ0'{O p jl=׻ӗO@fT˘AMig&xj|~O0rz^.YgqEmWca~4vyԛm?z2sE՘"VbPZnLu;4WZ2 ,sUĵZJtsU43Wߡ2 +2WFJj+s҂nPJ\}J++-qjbW(S}/)e`s:3*,{Vʣ>_  71]:ndߍ'/7YmpDoG>0mfo?/Nb?->n3W=ҋae?BtcOa2lʛ e~m0i_??6GÏ~ f_Fcב~Zd8Ԑ;D_"H,>BQ3߫/%Fx]Y'$| SDT B?d{yj u2Rs+߮v˰MRĦi[K2-縑kwNXࣱ&'šZ5Q!Lĝ7~oB|HɄ^G\T|4rzO4 AaИ4Mp\7cjXo{O$7v?toۣ>> fDĀoLM O? t(B\6<͝Ҏ {R7T'T;wq^҈@ o+(OU#8wWz:@-uDfWL[J`}}0RU>#ic\Isr pgJ{F_!e > l;w0 >mmt8YIIlR,J& ERb:BzC0nȗy9D Oݠd? WXS~1+|] VEEͧ8ZR.tVXa7/8+Ypm~7P%̯Wc~};8K^Lj/a}{33"s_s5w|_sܹ2w|_s5E_s<1_s<)r5w|__s5w|_s5w|(*w|_s5w|_s5w|يP{tce@Y0g]CGzTp[z.>tZF兺6 ,hLJm*úT҅2bTk3A+꠫w5Pmn\Mk x$ )[X1c2b@Dk45[!-e}j(?wAFw+"ä4A;ABC^GMqQluO$0$5AZX,`RD|0ay# T)JkyO5rOp6*3)ŤQ]iMN}?^UXV%8w 1~-F>UwnFSn؉}~_rV[ϖJ|(h"uQc)&\(HŖ"ҪfJF(R>RL@*J%bdF/cL%նfl~(5x.e].x,8u| g9 mjQ3 Jt;/Ze0՞sh(& p kgG)v;(!D%$ItY$Y"#a:0- 9095q+Z5=_6H<>D"Cr Kcƫ5`"XBf 4Sݎ[ iU3b8`X('8=r`TiL5rʨ&Ÿ+5xF-kD5bֈIk ͉'.MHmn0lY>e)ο4WF[XMn׵#"JSɰv3OVTlGnLj@6heExܝ9:c\=GciyC-D(KS|12pJh!6.x5RZ;Ҷ2]HXG"\`!`BLYF@8,,Q P#&mkh|gp3({;W7QuЅ&P=9|ǤyW"uCzǫlƤ" e-W1Ym` 1Jarf4€<DRQ2DTcnɜH:(]rn()9F*#R"%1dNF03,*͸Jb,aF*iH& w>VGEɺcrM^cG/1 iAZ.4BF8ﲴokiZi"9"< H%%D!ᕓ9Ӕidk1<< [Djw"=$wF8_9f9qL*]-D`%Wc򈜅q1J C x'>0ktQ 02'̨wez=E"M(C nGP0&kK{c 9SD/nhz{u8.'禿@&5!%sTnr#rFti#&) q"caƝ_T᪸0"p 8{24_ƽC%aKGiЭt 9]f6 .a8Ok(UIg\]ԗ'7#|ug^W'_'^̃;a1=ӏy[y;(Qr987sjƑ{^0yw0}*L\L+߽2zt=fp?!FmzVPYGi#LH}0t֗#_H#|ArMc︙ƩnK˯egR4>'W.[$}_W `ثlwKr)<I}_?b_.\&_~/_J?=?_>uo/N߽~k_ཌKH`lm~[. ?emkho:4Ul%j' _e\#7JBu=_a@/*}O4?^FZD+Iz%8Ub,JziRӝ)=m ?m{%AF=iR;FgЌ ._?Qh$ 0Qqj 6fRGUQ(d l6j²vCurA P틴 ( D F>oB S=bZ.{`^:[Q;5o) ) cfVu8CBo >9;KO_8~T钓p%ë&mස_SI@둫뛗kQK5,`dzPü(eiZg"@,U`q8ul"8 ҄p:t1nz9T}ZǓ3V)/jz3*;aT+cxMQ^EW7ͨY aX\56#b8IN׃`:; *9{Ar>cBP@8?5{JJ%2^L'O-qQ/X5Sz∫a%V2mp%b!pZdu+gJpo^qV_˭ rxupufsѸjcUO Rq2Ym7zo.:Tո0޽Ekt04"~._T*gobܬoYcŮ<6il?'ꧧs]tE蛗'+u9(氱$`IyhVȕFJ$r+1*gG`kiV 81O:,}PoJ]]rTϘPVZU(Wcೃk9KxvҌ ^k@--\sΣO̜9R!"ZZJL)R#b+hsutDrR 2Yĸ$S4C,2IacA8d H`(`-6[8RLPsCWrZ_uir.LpG[a="U0( yRC+Q <&Dx[ERwi>y*4%K"2h^ (vc]6|=EBRڲK"Vf?`նn^0urW~;-#`ۖ1\hy|R~u6"7T&P2[ئ(HIVRPTE6p/ހn E@R:7Tܩ nPʍ9hl `;lWEJ)$m9!:hrACXi_!~:Uosmuo SIƬ<%WY׮l[@TR}jmTMBA eb^lamoamanU9F bD9J?pZM'[`΄ӬC!fH  bg)֜z\Tknl۽"g9ĺ$5ì.<$[n`3WP97Z^)}8_zq{υKnj q%4:sR J,˯^6ʮR@RGTRJmɌ|,&VȈ7WwWSfzH%P } 6 ֻȗ5>%>SBRW=3$IInDNYU]w)Ge I.,Y&412I7t2!D!e!x030{-#cԀV*%cAglYi4nLFU (E.o(|8~k-Ky35a~Zρ\ x>Ryk޶~45>0࿋I%mwFWOz8 lhKҳOF0lui7O(+[̼0r=>M/868vu"w 7B`kjWPsmkd$4n}:4damhYs[M2%j(NNNN͒VÄ,ԑ^^^^^^^ɺus _2͉sbNω9?'gh*q7 J)fvf\UPBE)S1=e\Ҵ\0sn]wK-# dRKJOfxフHO7X˲1Gu\Yܓ4J B/hCqMR TT+cγ,$|k}PiT2ULU/7QoXFEJ4x4ETz :(9nU8y{QPxkV ]5#,bM7k b;֬|/|[R!- ƥՄQ~EU~ sZD)Z2li%% ,Y> ܷT|jt?Ƶ1 3p$- qzD!8>}+,[rNJ?B_uKk7^3oKXcy_ہ,/,L*mrBU`|BUTJUBHVJP*r}UTv\=۪\Vvٮ#@)==:Kvs?HGCǠz*n[."ro0CJf((xA㓽D#~drs;Ux}pO:-(}SŬNX!9[x,f3G 0 ҩZf/EǸ.qՅS4e,!]Znu '1bePYVj\9+UXp5n*<Ƹ/ =ܻNZ2Y2RKϏ%~9ء-n,H#L젋r!kl孺tQnƤF$;K%](#K&8v;>Y!S,.0i4PNFY)i95(xǾɛgƥaw+"ä4A;ABC^GMqQ6! bD <* s]`FʩR ˽)u^qWl&p5@ 5^!JɒViS7:ޢ=yIJZ=8IEvʕ\qA N:D+DrwLPT#SZ(GkQ2 | .yEoW12#XP1g&3vfθ^J9/ 3 9 w *-N]|5H-ÍǮy;̱#RF<^xsZ`QD\BA"Qʄg$E#cHQk %mR!`V0/HHLBDN<g3ǶÁ_P.kwGµ2l sC*2- wIꁁj AH47B)m[Bg 4d 1@ @vZd G=rTiL3p6J~ Q/N z;ÏrD1Gę#fTKƼ@:z-$D(E=$Ɂ8zR:Zad(Cb%#1-(ANhP΁&MFbR<\3p6sOj |q<*~Lΰ|vI拙/x0\3RZ"W=&[!Q9Kuz$'d`]T\s+2{fP5 hr틁S/r {"pgJ:霿){ lOǭWo?*{OEB^9T̓ԭ/t(Kp3"5SDc)Q"!@%KAMh6tR&L  >Ӓzo0wAKӵJ#O"xZ6 c~i8e%g\A03a0h֙B[bD17 exS Jo{{uqtPOR{v300%n=T1q&Vm lM'ć#vWuAȁ[D6Ǣ{ Du0z0l00tk@ԤMCZ7&1~;mfLQ)MTǾ_:nx2t+*-\ .,}=WjNnz^[QAـEIQqvd86q V\H"T#.tdS 1@bD ňHiρ.0EX;ÍS/ܭtj}M!?Ρ W ?eï( \Nxˠ;>YAC袂L U WtV]K̲^N&ިb$  `*NVQh3t9*}l}=ׇ<;Al74nӯ*Q'hZU˘.0O|fr7k{ o~UI;*"Y;\u~V6û #Q\5Ƽ+:]at1A{bg% ^ kWZIq1kE%A˛s3a @_ڃiP bO>dSh%wY,Gj1(' i`\8%L4z%(Ό֎bu+Pa0!S&F | \KIw]T3pLedk B@JZ Жn<\K}^u\;ަL[|-K3;|gSnS 6F r*ҌFl2ѺT("AY@$p(]nV=`lYHwD@4:&(fEwZI%PD@(@2fw=+Ľ PKA^bZ#FVXSQ紈n-x!]Nh8A|?pu5mEH $F[!+'s)& $ǸSJL0lʄ-!=+w!o]?bhgϿ-X\0=TmwM-D%oD\I1T٪B SlLָiP"-(H6q7M.'+ G/(\?N,v(ɖF"rvή0;3qYy6 pE7! I)IV3p=H]Vne¢0$1g S §ʱp!b$)ɾj8|0%Q/|KN_5d4(~wu2R0XS3'uP=x}7nx(!LfܘϽA볥z~ "@w\pa!Զ+1zeH2.,} +'N1A~zdd1~&Vm+f48BÔFoÑKSߍ"9.tդJX%܌TR7v)~궮\xs8%+Ux ?Hg4޸p+לf0_{qz@$wo|~߾D]w?[^&uuc//hyK ҰtiKK宛f][u%!uݍXy@]m_=%/ueY#-KOv/v|(%qX0P5ib@0j*7ls)2K@~V;GJQ&n5֏u+uT 5QqthZ+Tl낺7Ypml4dHb=AgSʨfJO9# 2[$ 1%EBێ5T%K\8s)IB8J+R2=uhujf) ϔT:b1r눈->$mQ-拝)*p>Ź >ACd멪y_TgR13Ϋ#ʰ+z<2KLgX@*5Q]aAiԭn\M7}{7|"#LNp1s c8k<ϧkViaYcb9&l_׋kͩvp'շ`k2g̭N j2-޼ܼ$$47 6WܿO`Ԁ^PrWW"E*RHxb*'y`NEJpD*Q&XQ2J *]%T2өD/A%))ʠ۔JlJ*F; $Rq]ؐ|{lS @n( {f۴K6A C" 8 Q*5"VZc򈴍L62Fr*]T.K<1#+$XvvT?,[jWOq*g ie \ej9;tT'{Oj"C~_?73lsL~=U0 ;hj{N.d +"DCYe5V>VH_SZNd'p(@u,X%P>iy@%˩5**i5Ŏ!G`}<*+ ƧfY_O)?N)eRT,S32L眆߄;Ge^K?x@ r"hC& SoA )G&S^ 8E##yIRD#5謴{6XVKzjRiԎvP0 )KDHdp^Xixi\mHBe_ $a"gSׯfo_#̈t] 'i e6Y`2.%Y^~UGmQFGυhJ 5Y-6:%ăFRpc)Tn[VF$vo]62(O|"\RϨWOBhA) a^$#]HgyK"<%F ʰ\ !d|I\ʎh>B>`U 5)E/,Y/,_VcTrhosCI_OvJ;v2n۴TkAhm,ȓ-"rlMOTsnɔyQ$&③WporMApB24EVEFU !9Gy:$NBOHP!V :%*I% A42#g=2Uaa1 )GEynZ>l1חk??mP`uЯ'9bBe xrzC\?z3fxz"?_JG3w7]B R.B]vMg1gAߵ h]NVȗ=cڪ=_Fff~\AO.kYC{%,䦓 qCQPL(q&e(%*GLdZJM;+ h |_jzNF ,wג)XMvi5`Ť_Z3RN.G.$N0O0X֝gN(X(n:˷7t*tQ[m[O 6vez~Gst(wH(wCvNOAu)U. Ȋ |Žvz}lXvǛVef Y5'W>^7=upYyj;Bzݹ?= RQ-m$X?Zڙ\}4ZIԡ3TR3~QQL oy;׷?VS72ȋPt)ND!pa4N,G=^܌xC!H# Etu*C>#^`GXmd%vnNn-w_x"AJ<^K,ڇRFp h3mi <Mq,F .p΃"\D"FM2H!tʹblʰ󃉨Gݭ%% Sn5j>ϴ>D617IC[^˭n.7If"K% EBW(iff|ʹ(%7 _#sK5<}e1q-tB ԛG{?ӣ'mD>9ksVSGDivdea/M>{IsԕPBC/wT>YpB5,0J('䒜)ɱ{b:^I=A8S͞kI"nc'oMDgc.a*'JLJ_5k\?mgpɧT~Zί#b0~m.^׽Klpe?k͛ד\'Z OFpyMMh$,6{{y2>QwkNֽi|+Lԩ678֨}@[?~n~>;\ϯ+|sx%h+Vxzah|1=1=?!tFpWLU"1b8}sC?w>ROǟ^|/q:i#`oBA͇8|ha)9قo2,9qvw\͸"4QMh;ﯶq2jnky;=lYF)dj:`8w+oWOmc?$p gmEx #њ ~l_G˿"6"O9&&?Z!<ɡhD T&#h:OGHX5!۪$>jfh"q4A Y4r}\S$ARK"PZ'vrWX*]r}@ 4%UӔ([ip)hp ["a88 t޳ۯE6 W7]CWpzPMxW/ iTq29a05+Ld2]un*nTf/^#ײъxK!\@nC<2लњ>yL9{; a n@Ih[nrw ^ϬdSf͉4NMO>LHyUvd_^LŁ€2Dݽ՗qNzeXNXTBnꎧ.J B>A9Hʥ$UFvr/4JkJ鐔gv*HhOLZ*6G  2#BQk̜zxllh 5 !|ty"ix-/ ѸO׏Ԁ}*tADX\AYϋCwCwGC/:/):H UXxbR]M1&]DUE2$M)Vr yዀblR>#m&r&{"Ct1jfN^o°^LSKj-),?T}{>^y޺8vvt"vd:. ߆1w1g4 e.e uzR*air3_aF_{{Ko,Y9?`I3Kgp_P`)/AUa;?nʏ(O1>]]M_=rb>4񏫫N ǺG?{6=ׄLV'vKXwoJQPkikܶyLDLjלi}>QcN%f @׍/BM5}scs=T%(j$% VS"Ƞi\LȂ+ B+SE6݃R&鄋|)9"S` $24(M=f6s~W}m&sV^)¼}.]o9_45}}i]wmMz8#ĉѡz˞[OkaquS. "&G-BE|$P<0!9'.^b݃=DH ɵǻ88ԅqLd~]ǚSQ*c6c-\NajRجZŹ6V fZ;|c]'C1Gc}"1ɛA: "1D9I"EcأQ >:2Ija#FWEer;tQd3r4g_6ɾ|BAj[xÄG)N&7GN4h ShevĎwb j7gu v͈12P@ԑyMUDO-IzP` ڞΒ664 { ِ O;`/H G(~!L:Nu!4fléS<<ܐ b/"BcD7K"f#h$s$Nڀ|XglYD[ l"7HQecP=WB \%ZP (Kp)\&  ߘW73g0ȇpp9Ǹ+܈o{zK-r4WKٔffIrק8>gl;_cZJUT#Ph^K ra'Mx VgUFl:MM( x IGjRdTIEI>Cifg*MӚ]"Ǜ>\WtD3rIgye$i3nG LeMP(52 ܳԏ\E"cWIT[?܍FyZIc.`@Ty4YeZ١! =Cx1k fv}S|Q|}OM/=#/%/ @!vmq4$dfc?*dUMbw_pDcy>)Xt`x/D)BAMLdB Ne;gtB!tf+6,2|iD=ޣ|VI͓SIyF`,'%) Z9;`6wKge|>P^nզ-\oQ-6 75~_E7GADfWҀ.yAE+SL)S"b!xT%Kq5w0 U%9,',ɱXl,(ҚL..E voIj ω12͋lPP0tgQy<|߾* vK嘢 ̥ۯƓ-|pw㠫A6iH[AVۇq_NfR+ohA8_8ěN[2DG-Wy-R]qQ۞|2%rLR)`df$ѠJAwecAj< O><5m58xsN`^XtBP^3;t "XBDBK ^Iy|.$ƥѷgU6'{rjϩ>=D][P6=/5c#"%~uj_ T$>HJ%-|JY=2dv饷Xd4CAC'0ս+M*#EE1DHeXF+URudoz@U9sMQ(OQ>]]M_=rD>˔0quSϾ^=ٳiU*zp݈{iڝ>E^Xܫ,HhObst"AK$dGZcپ=@s״՝ x9E "z<&bb:bΣS;?uL+yK{ /gJHBKv]}&^dw" _L Iٖzx5:ȡDc:8Ɦ_WUQ.pP=$n9>>y&o.2-E>%ƗAn_G$+/i3r<Q+v]86rY\Ԕ)&2Jϧ@_Yg^#g֨^?ϕ/ve~|r5ѵ$t~uޙ}eOE׮K0Gy:ӛ=oU`W{3fz1Nf,bʌze{ѽ2eaw9/I=>.{sW/@A dw^qq{&b) t B6۳ &vr\-ZDlpZpNSx_aMG;g +jeG D:S:R\v@ dZˌd{c}IL R`)TCA8QȤbjD|Ԗ,05rXj;cP(释ڶT,z~>AC󪩛x*!-JekDwA 'Dvhxr$tB M,=feJ7SntI)aIp̑JB:}Pń4̲VGX.E(Vk-E ta6edf&H&E4j;Acf(bwif.K>!_ }`r 9}q(B*z,J !cga0tT[Ӄ)pR!0KôWγhZWxEZɡ4$dJ xv_[|7%rzK%FXIxI 'I4Y0yYì2Ęh*|4\.zr5f?kry9{$Fm+A3뤜%_GP}1!ʽrwgo{W8VDBr0П8wH/[QaaşhF㳙Ǽ?#nCWi⎩*Φ.BOW?|}_?oNWyuK DeBZm"qcRm@C/ZC Z6g=_ڜrø9jpk1C4,WIhkwH׵uCVp~hEB={1ؗ ~t0]FI6N&PD22c c79y4: Q9F]tDe6RL OV(A$-d#2̡ |g=wdU-wC)d֐TbKP]dޑIIUlBV%WD,vPBTj|k]CZk"1\ X{2ɯ(T.תVwn0Mj(zU ino;#`%mhfNq:#X*y?>4c}iFQ@M^!$jB%.iLڳ\)i \}+V:\*yczWL0iEqJp8j5<\=JWp:ZY{W`c@H`_PŮU; mԷ)\0 b絫B%bW \ J^:GAO޹O [Śtz|W"Nn /#qHl0i_f{0a(~ǿ}b.ת?&/>3!ZF.঎XM0tO6\cR*E4%!Yi,(Cb7e޺|;/?ߏCIdr0m;>b1o1AڣlDyho.ԢPy7{J'Ǿ{7M0Y&6y6 "An!HF-!x_exjU0nH]B+6>Mӫ9a @`裱&I#6$lT҃;o|jæWϖ~VWm8Ək9^s+!T|JesMuj?jj3E:Q|ZDeR4p0maƈ wW,Wh;o/rS'QsC$upfT>pɴGLڒ I1d Ϣ( +G$&Q7Q+8;'^M3|)/;OUZ-wm_?]Jb7Nq{ [ _b5t7u (?jI\Ɋh+WP LINNP)w_]* ]UA I(H|Ϥ,RB`Hό$\y`%cxw 4y͎U錇> 6i*BFg}D2r,E%r˪`kܯ qy4@fqH+Qjݥ(Oج=v6}oT UхϤM檣=k2Źm]|Q5;ᮌ9ջ^kC Jl奷Ud:m8md=muJpB-IAJBFΚAbVZE݁.jDh0zLAFeАkUc޳V,P;u֓NQ@Rm#ck܏*aak&J2 KܦE3n(->X7`Ҽ*7>OĈKC*Qʐ tvf0m4MG߃"C{.r&J^aF" flhZC,hvt[܏~4+&殠vkܱ#=^ڝǔKOz&fIƫwd&[<P'IA#f0MRid!fHHxLMP8|9,Qs-ak܏QhQqW5XeD;DX"1*fsOY&Ef܈XI%k4JŸ$x=&ˮi)d'H F ɘL*r" ,iZ8 J[mQ^'\Ts%_gk\..qwy%@|49[1‰Lg>gkXI-Dmg=r T6jS,H7VFmb[^@Fd^P; 2%mv:2^ZdrbG9r/R(L@QfLY1LUnwȓ3,T`ƹlny`ػPVrB׎ qÜ/]OܝR״Bמwb.{{S>}çsiJh-!x; dف†紌V`U d;P:pI3'nr*%eP‘P>m%V=+h4 gWV.n)ffs=#4-X܎,-BX!B2{istΎFHsv٤rx4vKlˈݻ؂;gbt 2*p_'`,(T LZA\dbaA.k&,c /fmK׶9~T͕t_o޺/53;t9 -OAv Oc)m0A#;^p9]o9WwG 8eg/ &Xఁ-D=9~nIl$ۭ0vnlU>˅hE#lE#<ƇT(Kb$U'8(y0뜝h2ZK4υA6[*H\bӠb"F 1438892\w*G* dZ5GglP}G*ӒW hexޟ_k)~ޱ%«մ5{}9:8ahh2Z(7Ddxpe{bJ6`Dh(a׉3Чj'WR?ҹ3Ņ9KNc(>%kO]Id޳WO/Y5eoE`H j|U0:vm/I%(P"|,U?hF ppkyvZDJ<^K,±`BBT<:m\HЈ|)c$~<8-E!2an.(.qڙ8a'pAE+ZqW#%L/E+m?1MfCkYj^;orO@lxECNhqq4kE$69F(%7 _` s?j9(YyUߔ0)%x4SUR oTኈ7tD P כ!cc=palJb!,M? 8q.[=xmM#['aA*!鿗?2w5m k&bYhF qh4"gjpxǩ]HOqRdJDkrc($f18hE :RD vl{n~5Dž#l G{?) 8]\ncis70D#?f̸7].+) x?S%ԻcLf8W#)(Krt$'ϏTkZXv!٘)DL 78 }R{m'y$O4_C7K[+(Ⱦ5mDoO/?(4#]Mhm½,3NnT=s@}-=^s{v3⟵ʇWl.̹A[nnvQSmUVgv8ADɃ=q{:o놵w#bËZGQ;pp3mdéaddulYQ5g磇8[#c#J?GwќM<Ԩ6z}?8u|aK씨gW.A[<f_]Λ2qOxH #8+ULGasgH?/~??pϟ~pÏO('?^'|/q:j",|?o޵]uͺ6Z.z:+|~z_m<*D#xh OsKsVHhq#J\$QH^Ŀ=x;NL$R:$5`A+Ү=3R!zMjXew hߵ2jYQFi\e.#og?u511lͅ>;#?A%ƍ8'WVDTu@D墶 2)+E}zg>(hj;9Mn D9v;lJU咤7)ĊY˵pAb"&D;cB>OZ#U!2bu&w+AX˅ [*)k-?,݌ 7]Q>O-1_ ::!8: }{:!)?HMd,*8ú)H9wX##TSg #O(=4StGk Hz(qr?H΅.y8Mc^bx-ւݻΏW߼ ?/-!YO%.̓Um۞=^nac0u\oҚϯV֠v pl RҨ9Қ)d|5W?Yj"+ /GʯpEͮ.h8n[zMUSOzx݌9yQgcj45LsNt7~ ^>n~U_'ש uE~+CLQ1x(Q D4%fh1D  Cޡ%)NS|smmR$v]%4n%QMn*1AgluC:-xj!dtS߷7MvCg/o֓5dzit}A5W5AlN5MOw͛mNyҵXw\)Z.m3|#))fLYCͨZGCڱ]lvk.g-/ܹz>..V7Wm:77+Gqiwb,zEԜ=MKqQlJ׽'NkG kAXT;dV\{r%*9+٢0[flQ- E)^-JlQ-/٢0[flQ- Ea(٢0[flQ- Ea(٢0[flQ-J ~b$@Ea(' Ea(٢0[flQ- Eax 3lQj( Ba(٢0[f1evsͥvs\j7ͥvsAZ[@Z*#^PfX1BYZv3JiR,%ܡ;sRn.KRn.KRn.JTZ3CYbZ.n UlQm$@]qZLww9 0[h ε;lJR"\T:Ovbgqۧx_GX:U2 똵8Mvp&nus,hOj!m2>BH࡚.]$@-"P%>V::`m8pO"~Y侃1 ׃ֲHH[;<ƕQ8"1U Ȋi+U:V?ā)ExԮLjg^& Ak y&rtM!h4hGrNKAd_v}yW /S#|j3rn1َ}V8Xzx,gOЇڔJj ߏnrV ķ#L`*pUI_q Qw+a 8X!' ItGe\B(фRW2Nϩ7),Ih`Ěd QV; Tx)Pc\k\ "8O`%Uc(ؙ886zxb9' sՁa!e0%n;~Su6g-uܖWW9h[s5g|T5m y: ih68KH!UpJL09uAztH/HF 9+z.|BI <XC#w&@9+QJъPD)qƽ@=i;b__G9E|Z~(a[<`/PoOPO0b)64 ^99Iy<:)l4^2 yTe9?/@vDo&l&bb\YvK=_?RIlJSvɮ KY 88IkhEF)T&O); F= xzqk INE,4K}i-svjyz~^~b^'j{yT߷ܧ!U:$jͪ׉zvQtPZaGC} +t 5j 3+Q auLmdM%9Ja ȸ9q=@+ca>`b᭢N&3^v֡9xVk 6GwF17^-ꈒ/Ѥ<0sSv ^"VW_ ib4GbMؔj}K.74s9դko^˜݈moN /w jow@]`[0TgW+Щ=r(朲)Mf.=82BLP`aHsTdwZ##0 k#4=2fn<W}J'0 "ED^D}G˳]c4AN=k!K냓^Zo2QLڸ/ XZnLf=vZ+IJb_Oe]x\<*`OkycqQVEw.;{pJuU)21k\:k1耋OyǾ<V,^TՎwv?=~j08lMs6yM+5R ĺfR_a&>UmWWCvm5fvoěܒn"KZޣ~ky/vm[a$nuŀO [-4[ɱ/6RlQi"a)}.>)G#[ؤನu/b9r/>`o _:G̬9(+pH|) `OӗnOU; 02 g7'ViVjSB:YF,9!e(g.3ib(qgZ7RD9(?np7Y8Z=;%G Lo^-k؁+͌e \*;Ɂ1G?r{݇ڈJSsLzs ,GOY/Cˮ~JGr8ZG80 ,CMYGR@pau~Sf:<'>@{+"X)ڈm|qKI/h%2qT0s>=ڇ6>)j ?KDJNAu>8?t^=jc@_%]xHu|g77)n>,VOjWi9ޜ9Tfqzqf}YPR~=U\Uk /,0qk/5p͋xZYgrPOV?1#e!q].}܍??xj~LaR?ȹk{q^"₎Q|_,'k #jXǟT/6mWv8 #l3kOˣ3$?̕.s]lst+s}a%vQ2|&I⭷_ fQ52`,)7:bu! "ϺZu KLH \)bp%PYV\IT?u{"acj>&gShpLd35wA=mE_RDuR;3qˡ}3,yaד6y?\m.fcրE#SQ/t`̭]@-D>&u`eCK_7`"\,*[e-@m XO y4"E䄴llH_~Uc! ~< 4I*D&rZȼ" x5{L h]=E RI>sU(ح9HwJ6 K2YЩ֏Cٱ ԽKmlz^=; l ƻhptz-mE)dTBR` 7Xk xh:RfV.9xXŪ:"R ZaAc,9^5K9EAMj!ܱ-:(t :6PS3%c$mê12 a&G(Ni<@BL2"\ecm\gUi3U k. P6 VhwțEcp nX(0UA  giR؅l-h5wmY4H[~ȇAA=%)R+RIJDRԲMnV:unu7,Nk BԨ@ ›W PʘK㧷 ^EXY7i$afs$ yp7$QBeÐ%k&nk=HcvyŭN:MM<34g`<*k02 jf B][ MW-$#hd*B5 lPR0[֘g1$Z\ 8,h4 61uP tw*3| gpb7xSArH w$CK6ڹ N;%f ECw F*P=BA# ^G5>N(PՀ`[j0&i` ׳ Xa\0r\8pl oat9B pjqN @i9 Jqr@"'9 DN r@"'9 DN r@"'9 DN r@"'9 DN r@"'9 DN r@"'9 dON ,}@W8rJ9 DN r@"'9 DN r@"'9 DN r@"'9 DN r@"'9 DN r@"'9 DN r@z^ z+~;yS'S)y2[wף3L @LɸdS^KWo\ZqK^/}Y"`ӟku:]!ʥ{DW}+3OW{Vm5l %uJaѕ݃,S_=w7tp5}+Dܡ4J8+]`mo ZJC+D%ҕ #G]!\nBWVB]!]A`Ÿ]!`ӟ+z ZZSDWHW+B}+D١Kp]]s=+ ko2Do?`PrGSHW y ;ٟ+|_ Z~t(DWGHW&Ӻ+ 7tpyo$uЕջ9_߄`^p9{幫ЪW9AnrDWO}sX ;B7t񿃧+)ҕJ#ZU/tho6ӇzAU~[SME4JxU]עyXx\d(g4ѸU)L]EµSSk}Z({=Wwhv`Ulk{ r4d3LW;L* Iב>=>-)6|;fM U F}l@¥FJ%aPhIzgy5VF7PvI]SOuft{)R!K"V=R. \Gd,L9IFOf1pjT vhC)ǥH{ pzqPBKϳ+7-z# 0Ok붋6v;ׅ1?oqoat}fW-_nb tݛ,ery۶FuU̦E WҫBجa/Rg{ eIUЗd!C۹d}%YJxd&*={s5OQw5*i(̂o|UJ1RBQu?ŋFOgXelPryL>P36"[UbmiE<uQU 3rH_Ӣo^!8A<|a{S wrkc׎o>"2\}GQprkQQb.v%0d^HiN] ~9C*u . %j Z!t22@{pgpku7Џiowzۿz 1E;%WSxydq.#4Gg^'Rkj1s0*ÃgFK`+W&K ƣױ`!3 2_"nnCi{?^/.-W\\=8h#riҠXlx_{i_V4N/׋ՋflݥNLd9> LY$81‹0y3>5-X!RS> RYd)4z.,SbVUiheKegWNxRhL)]6rrC3t}ip}F7P@qKk^ ͈l3lx \] r|jfWp]"hi)E X 8#eɂ{VO73gC1LheKҊZ+HX^[/jEˊDŋ<0IA#hYd%bҬ@TÙ"~g"e4`_^!z}gnb:κ*-6]GcE]Gv8@.H;Ef&7Nǐh+t}|[ m߶>=4]$翝|//qUfOlz58h@jƧQ_]i /6EʿF)g.cc0ޏswgzoaxyۿdx x8H_pMgspZEHR昗ϳ_\/QΖ-qN-mk~k0]&ar[V}Q\(\OmF~)lxKpRةxxo'v7XnMsS[Цܭ5osSȥ^hY"O튽M{[|p‚pAMz~E6t/!8-%m #|,98MKt2-19e4[Fѷp0Lxiw3ˤNKLDW)%S`9Rk(VD΄DzTϚI^_>:m²ga@L;m}b4С.9,h)D<r,b<32!8RD1q+j,!qCg0S^^ʁDcd$u~(#_t8'eВV4<,MڥQT^e&③áWpoRkP:/$ hA8iaAjBHΑ-S}B:%*I% AidlL؜663š9,4-n=hz:ʖa 8 ]\zk~*P^o}uO)'WAw&FR2 =L4e= 4K5OÐ=12lr'!(fe8&10L>&n:xS,Gl79sWPٱ#.O/ł.$uyCKdWgO 3:4έ@ '{`{QÌr{<t02ѭakpDBj>ڼGo6YOS@Z1hc/cZX~!٘~qq@ܾퟝY㒦ʏy`C /Nѧxݹ%ԣ#{2DGIArCR)lrR}]E&Ӯ _WV̔*g}Kwy-~/7HE^ sav=;ƖmTr5. ޜtp{,j˅-EͰHXdbPb`Enà߻ hf;~sj02pN.juQ_ɬ|':R`sW#į;Gv!}j3ñ)r;9zrUDS,A_<&_TJ"U+]8SUuۯw|oݛ?>O?})e?ޜ~x¿{GJo$pzӒ747ikXiEu=૴+vy]ج]JhCg4ވyGiMvqaoKf6nJtK@E7_$1kꓴ $:HJB{NQ 76'֊yXy4֧ Yxgǃ nu:5cW1LÔlK9_H]qP8l!3K9-&shVʴpM 86lb)zX-8' (BNTDgDH$ Cpuً/7bpyҚu4\e¢?YZũmf/:l|>}Oww4,H @QkcD,PyN. V36KAh@ݛrjU|˴RV*=Ony}87Cb7qZW]51DVS3ǧy0Ou(ۙڙG?,lZߓo.rr+~ގ>\%.~K 71ub-_ fΙ~zBnWͤb1E)$%˭t\k qvJา^~Puj>"hKgih]lu鰼ĕY4\Jln֯ίGONUq䷒oaEMv*:jU{ tS5qu#Gp*jޫXڸ[BlostBhWF_q6N 8\YRaι(\Զp8eG~E@SCή{@N) ^rMRs`H*@z=Z5  4qJ3&.ߚ֦.T;_ެf2WJzC=ʗTE:[~"c:5'ݟ]cbV)Ojڞ%E鹵&н iR/!mYZw=-K9x׆69Gp3bP\ؾUVJ pJg髿w{Gt;#nuj'{G%!TebrԄŃ~4dP.35$x(^ڨ_ 'hЋ>!sGuB <SO{wcL̮D|(Ò30dl2:OK>M(Fj\d,(Ѿ&9}"nKH!)pHL$̩7z´e]L_ JP $F (̊hX꣹G7Mw<,6hh^Gc|6` \eqUhpY՛CL?zR=3\=N\[VQY<·ӫ3w_h|тAZr<AE128R:HH`,rD/n4w9w"ۇy0Bn룻z9Exq{uG[o8ڱŏ7MAy*CQFPUWq* q$QIH r/^8t ) hjI`.;R. yp*5K3|TR&yևXR:,)u(uԹuI'w-@\P_E_~`٬O=EfZc$$֟ח4g>r_ $0VGhaEJS$u$%[")1-IضHJN8K)@2Eg@e'Λ`lTZ*64"Q0`)kL@[ q]6 &eF2 ,Z#h3xM4LRҘ8+ q-qz=:.ϵw]vv9onPثyUf).AR$M2E$EqmLU"ѢFS5Pvh(Xj#l o5ELhR /EJ؆51qkvwǙ$-ʘ=[j)2ȸߧ83-C,{hmGr~ #.]>bvişds` ~sgJ@Rk 6 IFae%@pef.Y &X`1-Ds~,jK)Ghc5Eç*Hƽ>h#.J"7Co7C"v7=_Á|=A_~s4 q[K~E~R5-(A2$ԣHQin7 8ϒR  -t |ЦnS7a0ޕ3 YV.˹^'fS0O aC ?Wi2r3|'a/('0߻/L~+d\߶IDZhѓx87KrMgӓVe](CEY+{ㄅ(:;#+=t},ӷ=8w8l%(?yܾ?LH4Wܰ eڎآA#`٣l!MClס|KzzMZ=۪zzFǪ |JbQmwUNV?8f|׼ͮ*mKPI ׵\&(_]Q{sEڶ|>˵)ؑ~=e?x>i;=JvYFev6s<;R~ou[+9%i ] ykoUTwtU5A;pw Vؿz6^~_wePW*i;qכԚ&'N+DfWk*rח[O5,4톑ʺc w"`Zr6UN*^kL"P6' %9u 5A#Gz遹|?g![z6ռJik%p|"eMCFGH<%L#<C:MavuD+f-67Vldkm{L/.Q)Qq ƒǪC~a~qGQ1/>8:.eDTOMc 510 Yٿv(=:B%p ҾC;@@FJ"B8K{E\)!K<*$hb'|4X"9,F05{ F -5a:o ؇߽wci4u0,:8=vيխw`^$13;{3\um<12rSz!6SkJz_1-r}5WņZa}ZLi\B jAg5=! D}Z o&IᾸIrAMT,:O53x|!APŨA`Ua=-{U8%3xC-@rfCEI]rgW3UضYL-<هt#3+kӟqmB!F S^K% ؠ $j$Q+"%BAe*АHd2-iW x(eZ /bl^eo&uMr8˛C9(6O^h6|Z~};᜜NK~0r/U]S}X2O펣_J'r @Q0>@x)IDxLAFGIFHb$zl4 et1yO[  #u̩ bNVp;Jp A42g32*Ű xDpN wUf' 8wvz3 m.p|=LgpN\G9@6JVQ 5L ey GՊ߁s(8CV\d{8˰ɕJx%f6XcU)@"Aˆ]L͈ǣb jڻF;`I9GȻ4Nv*QyT"1+gPHpGfHnaJ# ԐRXZ$0TNɅ=|L2J5hW:U1q6aj30 """-GD|X"HqkoR3H80#5ZHE$D5jvEњI ԍA N~hʣ,eѣ2 p0!&̈́NS[aD,&fD>|qq:*xHbdW\qq޿ y€Y=5IO8#eIzO"Y PF\<<,6;nxp_/%'w(\ޏ_QExqhWp)z jCkN{ȥ59%@\k%pG' ;!|(T|ԓ7ǂ8  FY:@e 9ź5GKM  · q`UBRYOq_Gb=gښ>£ &hF,>K.d:C8濍"X )Xz!$R WA8ͷ:v5fMwq1:rkK]Tvpdj뫑ⲿQ )IrU)lr%I"E&++γwnPn?V+}jKWMlfjr6CØ?꼓]vu `F(8}~OB}lZCu{c5r`OXX oEyb_X࡭bJ+y, g?.F* .q׿}8w??=珿?}S~w[ e!G(A'˓ k}ӊjۛ7͍آiMJmU%wՌݶ.b !9nDX:_p `K֠6&\$ޡ2sQ$6pCCRNHx<)rsP͉VͅÆzsLSNg;NMkxч0II$Y0ɦN`6.k&>V~0Len+UͫE)y*|En\'7O؈ 1jݟ~ΑR(fP9HD{˞Iwor=E7=z@w=EI .mjhrD'FR V2LȮl:d.%eKEvJ^,6!iEPZfNcP-EVT#g3GD !E;[^^6)>;rqn_NuC5#J*LszړdžgT}r$z֙Ȟz^ )p/Ӝ#|6t]mo#7+|]\Y|g|,᰸%j{GI3ɯbw˖l$۔- όEUŇOŪC/C3wrKr^2M r Koxٺ'eZx]ІHBb.*l3&7龩8€s6/yB ƉJs+L^O#sy%BELP7FpFB7^kjE4[Q=q Rjqzz.D}"S.[wFq|G@hh θP8(X; C>?!'IGeK]BCHfg2NA%,$4eC>L@V"BeGָ-yV"_%1H Sbq#bُnj֟;ٷ>~r;S:]×mW?=oln{+SkUBu5S-WY•ҰSrt $iȡݫrp=5<{+h *tQspZšGPJ}G>C}$` 桧/&臟~0b`gm8_;O~]_|:nfc;&()9?tnI}^^挴6;ZIpӯ7G?Nr<[4#rwXҾđŷ{=N(u}b1_N0&o ryiK{9Jhj/;9_M_;}Xeg^D6^kx a5lװc 1a8>Pkx a5lװ6^kx ?ZʴYkUctjg9 gDގW Deޘ jQfޚC<%??>YݨSALCLP $6S׈h&I쿮pVaYo{ >vkY]bB^^bln{(>= -Ǘ{)@9Ft6ԁǵf=5d*YJ֬5d*Y砨Whū|Rb=tᾏUvj[}6E3boq\ymWM-saI8OkaSM3 <*J ,)6JkFτV()5rn DL! O,h*Ke+߷7 :kwWkifcqL׶W\نZq[c3E3ӳ/ATm53F+21 곩5Uu`t4\Y'jY$˘&#8s<˟/0?J\?>q>x :V )JhE2^,K62INȉ2IG4ÿ  R>rDQ&ֈ"NѾD-,q~t3g_"v|\mf=yr_.(EiQhdE@C;lp9.vm4v<^/_j #vw- ;I2#A%LB T#/aZQgE\%!;{6PœfdRyɝAQP-È6rDҥsIRg7b$ǂŬHPWQ2.$upȻ;t^E?5*Q3f44@51\C:IHe8z "Ev4D^dK$Q Bxt0Sg7NN_cAbTDˆ+"n=H:ڻx54WB@dbq x4&(ݤ%A>))A $]ʇA:m"9y+}D8oQi āFQ) Sǂi)z@mBBk25ZN )\ mm"0.Ξasg}3O~;'!!wĕl6+qmgx7!ʇ~^^tmVY^~z&/+ErJ`1-1XMDɂpLlšbF0o,R Wt{"9Hc,ȸo5 .n)fWR%8'P{?ȳLjO5 :)&Js乲g1sDnJObiRD7Irg5SoTO*Y6 KcnnQ$;@;Rj=f=yce_yZ J"(W&DCtԡWi SwǞ+0js(PT:# 2[Ц`Y 1WuDIrR=sa=7H)$(#<#:j>142)OmَB+- i}2hϫKӱ?;;u :l{6&꓀Ak tPƖ u)mw=+z|n#0pG"ՙ|diNɯ{+.^rb$$\[j$clycv. v(X*Qm`|0تDF^_'YfhNX0sG]Ρ=Z~k{(ZP ;=@`GG0 /g0ϗ}@lCM]Yϊ"[]X] pؔO[^dv&fpQ %.OCnFQ.߿GIVyZ.6qLU.z a\q!W!W5@6$.H6D ϗ?z0,^op?rdAhqq4kU-ycZDij~os`KCkLMʅᙒ)%x4SUR o!,h7tD P pBN]0:ɺӀz͍sP /uރF&4NHÂT>QCrIܩ i-M;BA{qS,: ;Ô]bx[Ūb8ȅtY>ߔ'AWr fA~72Rοpؽ:.[dbZ NNwa+lk&4L6np%Z_)b坦wyfZ77oKTK/U\ُA..fѷK"T7jͥ;ll -]55Ú̲É`X( AӬNj6{WɦSݘJljuM6+fxq,9Rl@p c_;gz7jcsS=ߋo>~3>g|3V{Uq-AKQVA(+E]yƅ>>*u鎆W֯ԗ ߻I_~xwݛo~ޝSZIz$EQ4o~ܽijۛ55д"mwiW }]آ]*чh\ eE6_8P<1ND0;ܡgϳ:N)IM,XGNu@#3RVc:9$>jf;2AX, .3ǟ$1kꓴ $:HJBQ{n?g͊l93|W1ovs=~U#,9pZa6A&!:ttI%@!֕QyEJ.@w(4%f*FH>4;g,N`8M ]%ld "1MP(@G-J#NYí-#m1[Ef0} .WiG u]lΆ\[a>[3A2UxG'3E>ҭ@ׄ(x-mvYQ S6s w-ݭnico)Oo{ ϛ]^Tv;pv .N _`\ꬬe2U^z ʹj^vJ\V=V>tO8/#?6N8XYYlRhW܍/;WVDT3t'tQI]qAQK?u]/dupȋ0y@$*s:+%w:ؔ(D$ O#ur)\ІHmMt }& ܬ{ZOLkN9~p}jpnr9ű;(u0(U~ )7h$. k(6GHDJι+>RHe{)x:1;mO߫Hos5/k^XɎ\ݗ16Z_.cU*b5 qPLbb:qZN ~w=4:x?r1'ޯvˈR׾?+YsZ ^ZZhWmi=D˅ƋC˕y\YJm:Z/SzQfG0a/v|Mv,)vTm+9̈\QNau:Uw!ą^$׋|gP>?sJJ%l̅@Sl/6֛,0dq&K豯7Yʥ ڭ7_z"K۴Ml}ZU*y5#!Ih|тAZr\iFuDN&9 ݺԊkyRN( __,2KSx}d5P6}wե*fdǗ? 39C9L@,ULQzn|f}}2ִuNv]MV5.Moubصa?PQYOR <S^xHy2RHڔN8S&&VCK(!gGυV D; Ĝ˟ăF[E?Bo' ~W>xs:BE9>?OG f^\?ۖ;]yْŁjP_U`\^IBEX@)LTFt ε;Ms7M!^Nr:Hl>q\Nȣ*YքcНlf>Y{$~D!IL:%xɀE2HPE/t'$SMd dhIG4ϔ4+D<NŸZq"8m'%/FU2q~zZӓzVv<ΐw_TV|D$WN,b]" :EafLP5RDHb " !aFx)S&^;Q+%Qx@.Sl,R]3"JL }gt1(aHÂ1rvem%z KMf H]PnwpI!d^T`ä753Ow.xPQĨ<`QNE!0PIXOUq@Hx0#@O YvBvӯ߬Тf? m#Qiͮk ?ȆRB0S ybb|r݃t (|i೿߲8\gW(N0wq("ےF:ÂQog{!}(dPsTPUy?wD~inE{.˩=sPYy! |żjwL1'/g^]T[YZA 'bY3Uc(&s{gūbXwʤs; ƻNսI)oӠuwͱ'͉Yb£z\?5Æ^D@?uYL/aâRedd-ѵqr|~@ baRTPh{$9fI@.`io~`\4׺zcɯqU[Ny023Ɍ]e;+YbП 9^G6˂?#xZXSzTPZQP?`szU,kӮ2R+D;F\ȴvsNqg76hq#"u+z]a2JVxWti:i9bzEQJIm|3G+5k<>z#Jngk >qO,,Gar+)#fr{]}|ᢇ[DO9zxO= 'DήrepWy" F?ڨ,_x~i'O^SNj1ZI<ёHVk*HyJRqb-Fjv#cŎ $F+$5F"LSM"k3<< [USڀpz?^pU E?8nW||m.E"L-@r $kl`ɣ+5Ja<%oo ye=E^AŒ#&oe2|'3v(8F"r6DgRPޝ8\xB,t^=T8ƢN[Л քTȝK e9 m IY,WLs)IJfUzY('LbI(e 9s;!|WS?pV\vUGP&dKsZgedfVNO>O'R bdN̵^\[vPeyåBu#QQ;G#X06A!`*|z|?f;49YwxӨI֍Z7W~tf{i0rW̒Ep6(&u+`~%N-___ݱKww[&Ltn:}g_JBHvpWuNri]3Ur~8s%l?~u>¿{Y =ߞDO ݇aTYW ˸)׌긒qݵ/D h)Zͦ9ȳW(>4Qq#N- {vҝRGUQ(dv UmqI>NpWZATŒGրA2S> ( D Fib<㉽`MK#S7Pih7o#R6Oi>Ma/ ![lrxW- Ԛ_֢ ߶1pLB7?/$]myO~ h1#h1hy| R$GЦ:B m>IS{.ZW6ImXtA%((q{D! )xݎ!#XFlARʍ)[ r8h[: %RHDc1rcD>%(GzCַqLGG^߮/u[߁K| &DcUu|aC.4Ħ[ք=UD TΙC!׌؜QPZzZpaGC |L~ kfMVyUfmf?l~ՙ8wd|u< >A {8 Z'X^8TSG>6V{&Տ frlġ]$6Qix`,(3xn23ŃzߋpRvRGs 25"*ͭ1j QTJ0#*ob#n"fZP0vY4a#A2sVk/wE%怤| *۰u1rv("r )oUojbt$8dG:bv^c Ň-]sч%2E_&M(ڜ-#'uDcEFOAFPfG>2[;CYtz v܃&[K庀۹~"=_x > l҉y 'E,Y-{W_~Mpo?pz{yF j`*p.m iIi.4 m/2yg@u1aլgESl`ݪ6׮[vMMr3? ͞n/]{U>+{aW5Gi>(C>J;$fF9e9$7ih 5(¿V8dnԋ  G 0bqmLL/#I,C\*hŔE^Z M*Ѐ)mx>p` }G[ l3AvռcZ!DL.Uuά&2  ?{F_eg6oȇl`fLv >_dɑ$ղl^NHb[d*ֻX+l'BMÎP"Ԅ8'Ҽ(}T60$_=49M#yd* 7CМ6f&JUZX RL("8OJWI EY#K$'vqn=w=^rԠ P.{ʵ1 hLFt]`f'BL+I^5|ki7:` - 5f\E7;#w\݊eܨgpt,M7Y**(p'[|I)~Ju6ݽu)6Ctc.]) oz>F|/狫+mlX 3MY?Z5)6XSS}[yNY^*QB8|H'Tޏn/7A~-n k&N&!7qO,RfS o 'CyNz#(['F`Դx s@SLʨJdL#2d ɰ 6hxx5 CKdMͧnL٣oq"PGP^$1=_8-<0aPѤ=xI!lxdvG}v1UP ZQ:wrQE.&sUP`} ,:m2 Bd1$0p8͐\ֲk_Ԕi<^)Vȏ??95^!' jVќX%5 Ҭ!e/TbZdhE|i; u;:a2xdJ GݠEصؾiB2,h2.`Rr'%)& eDA4_ebw $ZBjrW2COeJDkPLIh F0nlS (shKN), Te Kcg黩޿ 7SB\8 "'|L +T2~~YzQ;or<)M]׻1Pbx;dfC?@#0BOɹ 'v8{s2!<]SMגDƒO8ԛFG,'JrNprƍO2^[߆.CN?ǛΕ-xpg_M~o,wQAE$w?x3/-S`.^^z_7lo{ ڙX>e4~i=b~ɣp.曼L6z8bjq~)Y=du֭uY:qB#ybt.^N=xej3c߯CM޼_^"e@o#*9_ 0J?,Î #e+*S?TKU;ɟٻ7oy}_wπ7}sx e}.QrMo~_jɚL4_cjE\xיW6yvw^El^JQx Pg.cj*,輎RGڤ n֏u(dK I J,YwZ'蔒r )VQڵa]|n갃 /FoؑGMm2&&Pnq "3x'iԘZN#a]_mEklh ' dMSâǶ3GeҞ9˩W6΢sMKfZC9]iup5,+4(Q[ # FQTM&vV5pGPM4 $Ro'spE$"HnnAѨC*& t&b(pp@9<hojZc<,&!OV4NN_kydkw9iղ˗1oES4 T'eE)T$Hpkc@"~YbN& VS-6BchwЀjJrV vUqA] Dm=7\1-oBoXUA̖nan@ip(xr!B;%un[oۭ@RYm3ː(-"$Vɹ g1%EBݎ1ؠ,v#eC bvc9,rZ%O-dfֳꅋG+bIĔY0yqq9߫ Y"{a\,|hdϒ'LKLD U‹Oz_ߑ@_v*ڜLm( VU(mw)[p匣p5ՠ?ezT3s9/_{sq$pqյX(͔D_zo'ALAp\SWh&IbmC+Y$ns=-u~/}Ll5&#)6f-')2<29U\$Be25ZN,B. 6薓u~#x\,9Y|aN煶ZnwّtaA{5ȱ}UyNWfnw`Gٟdz%ZJEhQdGQBiZ(7)|!m)v"[OiߥX1h*LkO  XHE)1!g'OZ#P Zݣet`jYT{Q膒ǻ:>^gK!'qK.߽&;l>>z5U% VXvhrlg"h_'I(#7ѷD@tA@ Clv[xEJUy8DQ l!%(.QP"Gtw(Zy >lsXQD5N18o<M$&g (Il8 1p4WBS^FEMɆEPՆl*j, pZp/ 샕+18fz| ?tjTVgka[f:W6 BΣ~e0.rz{𸝽Ny[focl]U{o>soƇAb%[fk^^%Y <qٝUN3E'[2;;oKQ1wȚ !#vX違5p,=Z p=2#Bn~̤@ ]YBQjqX-D.j[8"*8 ];+  :': sf ϝbgܳBI/D$Ap oԜ5y؜v!=‡Rٜ;_[Q'ص:.OLnZ{9ܞ9F{iNVHzm t"G{L1 `Jzj-[eG9apqlAe:19:8L?Ǜ?%'LV0kąe8t!ßO:09;yG& .2;+JZZED*k\//%opSP\~o@:P) 30T1P1 *I.Ⱃ?2!'I4o@&l`H%Tib1r G4HT:<o20:J) ie&Yj^ Dp7t42Lk\}i ?c\Z\|,n}pe1.-m$CdxSGJ &:49ɑcuJΥ*^T\tv{Le n5u q#y(pBc=Qs{R7qК=? J߇6`Nqk꣢ڥB y\<|M8uS΅PD)72 P[t*: G̒c4hL`RX̙C̅AXȫ@ fV,6m1$ v+={I jϋ燒87޶y7|k.Xe.G !1¥rw*vң +}5j)T-5DB+J`(ݖ6^۴NkD|>L&9H 1M@EŽ1FZ͹qbMa y q2zDB%S9~VH-WF%P/]HgYI"giw)PAmOXcf9 LXޙ OHdIhaj% M+{Ƹ ً$,zƞ 0(Ym2N HaDDc"^lJYc0,qǞhmAz#u/\I 8w)ã,کQD18cV @CfT5MT;-8YD!"#^"pDXģQa1r6ƨ]$OE#m5"ֈk^#H: ڔ/tj1jǮ&uwǧBD3UrG6A:%@( _P7ȭ rۑVwtă"f4qeO(M06*-U8p( BR1ERagT\޾'܇g㓋nH,&Z Cmϧmn{jPz 2V䒇D}ZbN( A65*`X{"9Hc,ȸo5 .f)fSR%(Tɓw(,wW hVxp\?{E hPmSj,ɲTJjRܲ] 녌k#yY#&S2Ԝ0Vsc-e5uRbގBHwv~DV`/0e IZ:+!q l(p (xМ! ]ͦO{K| N܊A[pn}餱ӧlbdh $Fh*j]'4 Aw@0% {yf6=$&bR%15m+O*0@`WA2ZKI5a̔ 5W155`R NF΁rX&s.^ܭ;+½i>^L͢:Vʺ[&Kjiݽ`͠+3gWxHJIJ<(2F.| ĚEbc1P0$^Q<8g5k<̃Das)qSd̓Y2 e1G+CX~.C2{_Q8d0Җ݉L 6" J< pv*WƍVGܠGMSOF ÉB$q YƷ z*NdV7FLxtZ}ҍA.fSOi4>a~:_ekxے$g # j a::;EG ])N?"<5iG'sƏN '=^+4y7~vDg-,~nPMPAxP7|2-͹Wh3C6c?Be.ų]V- Q ρ"IxH({-%4%"(Z: T^Qz}:'qz~/YKRi>bo' Dryݝ_}̓%̇򟫃D?Asw8?>h~C]&Ob'?'|t4zFG(>q|'}8j)YdTrP& C[*ܻQE7rlKCmCM,~_"jAm ]5]^ݶ<_IOg8nPnV%`w Шήjj3 wUo7W'c'T{gYuo*jt>mƣ 㪃uKO@_KbU&W2FB*^]}ꗤ`/]ery)*S+ȾL4Օq4F<p5;o;|`ι>a7AIɹK>cPĜ-"2SlVsObm8u&hod.я--'ک'o)QANh64_"964'21$)h79D@#1{EH"-96 \Y= J)ieuNAq ! LKM`\Np*ցRH<'N=bZy%_f +K,ZC Nzb@[y͈B;URBMSLYs(月Emk'9*y};2~7A#[IUm2A>6/q(׈N{,Y{XOǠvt).)^=!i wAb" Bd;cDł>ipBiW=9HrC!SPR٥vCnO2Z 0_?>g\q3tNY~pjTGqZ!2Ԍ1Wsk$beS\{d-2E҃aHomZ=׮}5@{hڀ~ryg}?(I 4:j0{@.㰚bc^b hAM?UK߮OOp{e58n{XL*{DLꪱ}f-(`P|;s_ m@[ ຽH뙐R%L]ˤ )ɣ 隀M׆XWGQ+%]}g:L{.Zʟ19 D%OqTJHG@{\'nNcb1J&i01ǁ-RG X `YyR9* #瞅㯠m:͹ Sx[̒&uۺ`ʌeqX95lnMﭞH3yz~UFW7{9F0mf%:qW$:#Ž,:Xs6Y]^-3i+/Eb[Ԅ]zlsIg>BH.9٠ J:HkH:="ʭ-q!RO^ HW+|q./cVCTR  & 4'7raV$k/:ixyymղd8L(.aHSs 92aV&hH7:+w6 ߼o{KuyJ,ld>8d=iuL>n.-6{[ܿVrpxp7: 2\ ,$xȱsvj1}QE|pK!V w6aM:S}?㓈}z>{v6YTc$3~& ^}:/?E8*o!N:)i[VJ:I[+h39aOɟ7Rt^u󘃓ii'EŊ$|0G4-FTkVsK\5IIO7" S' 4Apa͜TDBBh 9-'uKN&PƄHeg8R)h})KrJA,FKΛZځwia>b =Slݛ`o=ݧQXl|uجJQR:jC `=ٍ&m5c;wۊP?eC.!ȁh ̣$ST9 萜洬IA@xQL("8OtN"$K`1rs}*CӳZϲ#~!][u%kU(1u ǫσWq'В*,cya.zغ؊Cv{.Sut(52^xj.#;Я`^ϕrX39ŝ4hM,(8EKh5rM<ؕfUT?̛7I/OaRN*ׁ)[c\ZP)HwumH_ae7RSڗThH >Z%=~rB{ɽԒ8O :b}fՙ=wqǦ Jt+]I&bA#q6&ǡDw аTh6 6i|Ʉf3 [ a42:紷ROu*ƺm04}> pZjS` TSi^E%c<ȠH^Tڙf* 8Μ2.V%GgAҦc'kglI)>oyǁ-w{tY:_Q0t\ #@LDJdPڦUYR1PZgd[t&C`N[NcA] C YzI@IIIu;#~؝ҙ.3 X^T^˒;?/fL9Ŧ_4h40>}k\BI@%bi0tdr8:ImhY2g'D\66Qˠe2%" f]ҡt:G,uhzWܯd\}ڝqǞhmz[wq Y2*ebIk䧢p+^F̞eNZ6UGVI2Y (Ď'dE"QH:>!eX$ȹ_Na=E#vՈбF^#kl)f.'2BEf܊ĕ謑J3n "^`좚ewFoJ ÀHF\LHIHƉ@#y\ X]GwE ҫgˊto"3ON.R$Z9Y%ce4ЧJa{XQEX0$#{Ps#/'mx3d. DGc&A9Az3nv3EgWlGڹϦËӮ'[57HܴBדTɢZuȧ4J/>~az[r K BrT\ċ7*9%djN%8:P9FQHoѩwܮ@nƹZxYw(so9.Eas ٳ]fbagbV}%~G~}TS*ɄXYEŃ6DgG~~;?>" R~Jd43N`~s^MkRwL.uR'SdT2YəO Ҍ~.zS'|i(-lw}~X{:[{_l>np4;) ɠ%K'2r.qrLJGr^#D#^v"Y"cR>%ob6IHV$- _Y5LxX6BS`f%sNZ%cW$k9]|^]#쌜-z YwZ½i9_Ȓ'hr\kxWt"ӪI/Z3HxXٚd+5( @h*Q#Jcޡ̗ 0qp yxQ<$UK<"VLBJ9K opIKĺ P^tyL-q5dzġȒ'4d!*ɓǯ}cL9C&(^;^[i"]O2~aYO #{ %^it zPJƝ59 eHG{ѐf9C3'C`*TTȲT!D_jC:{O׋clo&i };཭% Jxr8|2,c|m)#8 GXS" 'ggQzT~5xR"?`[<!Q٠h UVꆘbOfrk3Ï )Yo\*jOuƟSP`QgɚͰ4.> 8$4vzM&;R\1k_JF45:R/:|]ó^aՔHN1-'ٻk `v9Vԗ׭э㨵q;[e,$-"KɘYSďӨLʘo.XKS>oYS5oD{Pb7j q:O/0[Md0d6`͏IA ;ݎ; # z\5yl6hF*` <]/o]#oڰl[4_&yDoʆl -vWNidw"6]ן-ӿcafdܝ옄\q@Z;5ztw;&I+xoY}O3jަDzlbylC=w t.>w=\Ys ]g5=iVnsa[K$Mz0Ff3d xyϝdƐJkQПtr|>3pB]up(bYaBE 1].B&Ɏ1Z=K>`yLqqDl_zuϽKv}wGLpJ>>3G;n<>z)Vʹ*&[Ip F@A&-EҷaU8ᑳ7 ǩN;{x=Cw2"[-d'ދDWڪRayۤ--Z6U̗ۜYN(aHҼc@v[Rb-yp}7#sdkNTqt>+=o(eΟU\pDXnd4X^f뤊C7,E{AܧcPĸ8.ՠCn !99*&9}QF%NgRAiǓ61C)l>"hq]r99,Ќ f*I*,!1pA[Az!(DQޕYYfƳR |!S9dKe$!8$`[Jp>p0 +נj /Z p8ț+DZ)&|3*0rs}|R=et W'M䉂6L@%gN1 ghA?i]Op i1Gob R]ȌJ͊tI 2GKx@weC}Rw/gV$*ʾ\Wk?Tqip+YCVhkSSHfMs+NJ]M 4Դ,cɛIv@jMdWĜq Ok+ڮ=/||ďW;KXHB$Nۆh::%AL8:Vb'rgWcO&gڗ`u9ɶQJ]uVp)#5248<('pnbS]ƕƩ,-Q}&^t ò&AZ~eW>D9MrJ,-Sĩpx<ó뗷G?͏WG_o^=zGo߼o^i:h#AGP_Zt54ZXІu9fWw9q9jxjbɆx!(x+(ObW(4^l,{Q;qk lY[팡wa]Z|S=G^ir%Chc%\Ă'7, (,ĬT\34_trWxAG\Q>zg1L[V ӍI 0} Ic6ÍVӿ `0+)Be$ J&J}͇ mS5m ֵTGu2;ٰgLTYGJz "k];C1ײƩ_Ɉudd9C%]"`>v\Ol/m@Iـ0g}ti/=`]6 ගz4qrv´{s|4f~>/M$T`TI{?R@t%uȌ"B!=`u %Y$ʁ\,,VZԵ[/Q;ߑLpթ{kzwj8pvIgVdc =ΓN}(|;m2V5&D`l"M)-5D`M6&D`l"M6&D`l"M6&D`UNHFDl"M6&D`l"MFX0j5QTVSjZMUj5Yb*05U@VSjZMUj5UTVSjZMU}Wm{'? =NE J;Z/m'%S]@뻠V!7كPi 1JD2 rˊQ` Yv{a-@d2$Ch$bp.:HQlur UB ]v ;9X[|7tT_샭s?޺H1,]_qGn:mdhZ45G{WoI~X}4${9nd5ӹ?%q/Xj).bquBI4A̔s\JӀK/|Sz[/RןtMwƷUd<xstnӥh<-Kh]drp\MO7~~row6dz=-Y-0a<8k7g*κ(41Tތk pm@uȮ)W|q5'?upCv T٧KṡMV5F}pn ce8 "3YNpQ9f$Mt``c hY:9KcȫnaDU`! A=DL`6[՚D`iҾ4u1KYm`w`]~پ?X9]|6n߹ oNoe4vU_gevGUUO/}6>'7rk{ RG;@͠1yӚJwT) WOs2QMa+얣n{j-j^+}<̦-n`St&v9|uGŴ] q9-rs7CsUѧm,}nsu(|W͹{} TJK-[.56.̑wR]{K>D*d@;WO֎<#oe8<]džAWTR8Z>\6 :9"/Fy.;gБzTesh9UнT22+f )W4Eg ,VU_%VP7vikǻ9'lPnV>{s5Wb'?6JG-sj{ If_i;#xNq;!B"*~(J7Cc|(u<>O/pQ @fEu0)Ř+=u).{!.P!pw!I/U ٢} 2LZPZs6#B7=#^&ۏ>Să3MG+`xHv1jB;py36@׻V"ob^}%j"wK*ptX|T#ySd;>tEYZ ;'>]c:}ZxBݩ4bS}z{*@ہj˱].AT.q5%6!=qupAP?9-ܞu4|wrQcZwλ"&bA#Bv{pkѸ*a3`u Y KP`9x!,FFVJ 5yL;a߾3ivsYDK*dcl>)Kc^jEfEz,mFʤi ǬPp"qe9:kҌEAX`!^ƴ~H )3 1'ͥހV;Vsv[3W.W*ǴYm.v7xeAZ fe"ӈO Z󐳍'fw& X,g..=6:=1$FkmG?#'g?xFqeV]P Y#v_ 9Г%%0\2YPFC-v;Aa1wDYQ F5'yi˘!s!<7C6o[+7F4FؑVe#vl,}3"yn|}8?nG AM딅dR"@ KSґpYX2+AA$1k{1 !2&/C& ibk ;k ϘBK|pu>,4U- fVsr.#)VӼ `/Kvjy0hHY-}ds:J? dzSu<=-m*Jl-[:5?O݊>熇圭I2\<@P Q dcțcxYjgiYA+DN ! %ÌBD7Y8Y 4Kĺ PfyX}qzOm<0'l9H#4!xrIsIbH^ 0_i'6(Ϧq|Ew:=9gb:R8* RnvF3c,)_g1,K}>0XxV"o|{i狳!N/NK%Oj>ɀoAztrtٹEl^w fCW5K^Zԣ'X=50teNǼY?{rӛ4M:/C1k_J5g4-/oHZu_VzQ9.nDpU}8m3he1D~]||>zva~r/$=D\ ~*3?I3)\}sh0\~yƋo ͍(L9|c7|};Ub} J兾8ȧTßnɊ0_#XK!f>+(Qؒ!a'И9G;w"Vԟ6g>>7 4fZ{8_!6r=n d? &F5X34EɦԊ%G6`Yf8z.lv$!_t/Onq瓴SsB4DR(frE C]{qb3s-]:DU{VdGA]{pOk\nb}ljA}vhYw/lq۹VW!x꯹mMe,'}^il&&IRP)Q̙ů,qAw[%%3ͅL֒Qe'$>yB# 0;}UgcIKǔv\c*z|ȥ,JaJ0euLz^X^O5o<]i{FTJOzZqG rVᕳoggM73?Mڕ'uk7~ӄ (4W+"g?2y1if=j=/CfFS ?uHƃ?|n }e}Wԇ5X0zmFM_#bN^<@m/&9^Ln4o59ϯOp$Igȓ9A(^S34 mvQEGl\yސ65:"x 3USEuzZ.5@$< &?j8-8N>Kmh8j"O@;>_xhl6Vu3Жq6(?7X\%8# *EV}/Ff&dvoRWĿ>CAsVt1C/JiD=]O]%8VɉxƐƌ[Z`od|Ï|}0 $3+haa0xndI3m҉EMàdmx򾵛 /"kV%/S#6/It˫V%/gx &z^.sw{W;uCLǓ?]!WgIK~#C}*_#,$Y:hZ@*;A^Қ'#߯aʲ2h0¦R0%T×e'ˎa[ʪd'rN3GApVu+zLw2`wz2 ޻aNH&X1|*`U-A˜GP$%j@͌,11Ŵto/`?7O3e ~\}w.N2Q S- RK4Y,o!c\4ɼ?!sn5%Zڢ >O5`l˰nj.4bT d> !R7Z!#Im@R]Tv~_~=$C5cFUw|?nsC|XI`͊4?(d_jzc*x'Tɬ!Z[aԮ0iT*}o J5P"V*8^!5{2B_On]SwNχtM{ޱIwSo;w}َodp cֻ 6E4خì6h:2d 7su?kk+5ɍg>J/Vh4 x"C7F 2<=LۨuvvP#9\tPpl|-n1f8-bq4r j=f&} ]3RC?\fM'pg7;|9ڡKJVB$j.zW[HGmQ9GK{s@*~slc+o򈑼]i)8mv#9omn'C-fvtWg#F>Ƒ۹ kmms{n>GD4h觓6ax-FB? L>  JqAV6iģ?e'>/dɔ{xqzѓ>jCW[3gdӣHÈu gHϏ&ϗF0bqԝ:FCE+N!z,k8'=]MqP l(6ƺ'| #;TQ=}e'>Z#GS>h=ܜKl+v\ÔhQkg lNM\t~nKJ#Kyꎱ蚽:9QA[J|,$֕=̯"kS6Kof_,~vq9wW~w_y'iQ+2j$%}..:L6y(edXq9?.ыVgGm3ոu?fYȾgngܟ-gᄋ\SZ-O4k0$V8j/ϓoNRt*&6W]:94q4#iW{gզ--emkwN=}enG?=EQ7-]2ጒ|b6_ߢh&kǨmϗjo)tҊ׼˟xrqv1͵TZknJnxant֩M^ېzxY?6{.V=i[3KKmŷ-Yu}[ߥ([]HGsVRV} tvu>j>LǘKAoa u}.{Aʖ̵+D焐,dQ:r/ P7w\#qSgfy߆̣<7_cNͧ'g)_K^yd񢓣lK0 +)RSAqϔ$`ērVif[ߦ`>YP!QJCnJ(CH"&)čS1,io}JH4Rq]Ғ KzLiM>gx] V!Eq)\} IT&+-TksuzwK(SRPEe̕ `R*r*sb%>Of,Bho,*jE))HeJ 3m;ǣ9õnSOy2ph` ڮi !M";P`* Ci6 C#z+ƓhV\{X8 /;G 0stK~S&3k\D%57k(Wچ@.,NHgpרܰ i|8peyYβ5@8`#ΎqU]tvwa"H9dr c% (xԂsH]Q239&ͨn`DSIW:UUe HO "q3F0[e( J Ȃ@ B\*HE"" Q jB02?%( a"e21PI1YJxT _XGi0q3XAfp "u-Vr!Ȩ)و2 10J8I9WXT5pCjx;x)J6􅲎 R@AjTMuR1B^vʺ@J"@9&댳$"/do5"($K6nc/ETpPR)-@B;@ITvÐ6#VZ8'23{ 6ke#]'J~sǬCr|bH.+C8qB4s}w lL|0uYLU-bE|dM$(rpbP Cكg  .3XLW[(Pu*KZLtp1 %Hv9 /UX~5<)xPiD(2!(C`$ (g8Lt_X2ĘQqd/  ƒZ`2V#FP܂̠m@MGEUgFeSY_ " j+xi{%7>8. ŨuԆHTzښE-2.FdQu ((BmLBQCX|,mt{TdUuW*vJڬSrF,JZ8-7%B̓9-J "S<2 g]~ar%Ψ 2qq2ow_u*ī:>I 2PquSE 5IptdmT<{(xg1(YXVjQԱVo%z#eoѠQq5(/N=4o |T.q[pP!(Q^"f (TɻQErr!Bsһ\6rԠD>+T`=@g\{ʑ4Wٙ>6b6 it:g ^cMd#ɎA-],ْ-˔-%:|yX,EC2!0 h@0JKpz 5LR y^ e*H ' eq :kc@%6k60W5:GhVbba#/+(@a\$$Q%AdcV@,,x x̏@yEBЎ$=96Zqz"+)!ƪ:*=(`(Ke$i1iCTCx ~wͽtU8_w䘙LC,(%8mʑB= g-!ox}a  mvbap˂G-k5Aұ.DS7 <]C ciؓ[BOnGA$²GhI2p! v}a4F6󁃯YӥƖ"B[6=f9GD$|w]W9g/.4/@8/Seu@x'Q<(B ^%]?nsq`0a+b# jh``i)Mu@pzB&BI8o2?EԳќ-:I1I B9m\W2?'6|Lt<M7;94Hmw1l<DI %rqox447":\KbO?jS7EHuc7iܹr.ԱkisW@qP*b&~SAyUqa  Ž- lI (UH}$uÐBI $BI $BI $BI $BI $BI $BI $BI $BI $BI $BI $B%D 6pH  =Jޓ@PJI H! $@H! $@H! $@H! $@H! $@H! $@H! $@H! $@H! $@H! $@H! $@H! @eyLpy8$W!zI (E[$(%[$BI $BI $BI $BI $BI $BI $BI $BI $BI $BI $BI $fI JHR`kjJ*lIRJMI $@H! $@H! $@H! $@H! $@H! $@H! $@H! $@H! $@H! $@H! $@H! $@ t?GIٯjZjݦקiB PRM;ap-1A͋;LaJNР_  ;au ˣoлLn~ڙN9K/5:=M64EcDHu.2p9s?v.tgKPf*}y&(7=*}st~/{'= -ݖr`:1k1|\+؝Mv0ܫ8&%A{N@1Ϋԇ`&E(D.}Zѿ'vZK%S) o_m;]ߏA q1*[27@U8'u>sOc!A/Flq ~j$3zO3wٖڝVM;ڃl6ep×4;, %4bk:#׏~pݾR_~14Zq`ܟ;wcG܁>5ދ, {eW4E/=^z2ppu`*hK}L ^ i>O#VYZrDgN!R ャP5;uFT̰B&_)Y=;}sd4,Fp!0{lɗ~V#s!wKf޲e\L)|#Uu17p9B'ZVH-d7F9)6g(:oe()Me&x-x21&ieD7.ZtÔ{ !/B@сG)eY> yE߾{yic+kqp7\N+wiF "~(iRHP~W/Q}ZYfZqË;f}x2bbZj?9PbxZ]&XiО .:G9rݶ7_bi<^P~SdaqI\pEߦYͥFmg:]R&Nhb:(]<&#Tvq~}׷.8Z3ػ>_R||oנ\fvYf[$_ٱhvCQsw|ֹljOl)Š~]WmkFE6oU3ɕ͖6#jw'ۭ]*_co\Ze:tcn7wt?Y^oek|P$dxg \,CAY,6thOCSe\(Wn q51VL?Mz*(_|Н/gO2`]e…7,J֋w^>э79r95ݕU7$¬ k[͈Ͻ~ob{ Q*xyc]əzi xAj5IDypZߦ<8nEir(}ZS$\Iivҽh5d`aD/*(ON";^E^#=ABUj?7XAr̯-amW\.*'BZe"$DZɬ 4Ic6a<{$^e)L]Sco敋sGMw E0p䞠jߗ;U*=Xqɇb4sg!R4#xguw/K4ӣbh6Jh 畤vW>Slf(d45Z/4Rdr7p.8"%Q;?>EV7kUR_ +;1y@S'LyJB;*>BU5RZqSG p{3%P.]^ʲWOe\(q\/^Zh_wTc:՞)b1UFcJ珚|kinߙͣYhw&Kogx¼mlQbt]0)Ez,;W׫%<[n"l݁lէj t|ݺ\'/JmoޅnWMwz5xjkD~ڂ*ٷ"WZC.6S9_zGkB9sğhRFp k"eGzgb ;^Iܜ܂v3=ǟD.!CW a`cz3"2FCSm ռQ$e9Nev')u!aY<+MJCglW tJ%ZdC&,l Q'$;]@IQ 2bאI!l8[ 4$ᡓh%7kMLQ>۽@a!bm/nelQ1KMۚuKpwK&4_rz;`s0ꛭ~(AUxF$Y扦6]Mdtl$F&a 9=)3ʘ9ezR((zZ))*T IՊgUj// w %*u_e|H45QF_ {ݗ=v.{8"QER6JVQ"L et Њρsu3< x %s=㬸MtP+5e9163a*C6U=B/^uז6*؇,^2Yݥة;S 1CN@c+ gJ"dPT@AOXLLP G,# T'\eX86/n٤X>G#>,M`Z*!5KY&x&ȤaX%t Ȯv8I)MyԂg|L&@dBai*BIp]l}=Q3\4YJ_d/>'a>QasTkoMp3X|,K9Y$9}y(#w^evp.lf[ԛ8c䣳 ǻ7x| 01qA.wl&TAw>ntv.?kIʳ›p66j9ZQqOo4 rÈvQVeT 4c`eMh'Bqꃍ%mv:OQh)'Rwag{rqj*j*mbؽ~6gXw+ qwcKҫϾr֤tm%%2-Yj%ˎJύ-&l*/Nh%ցU{pTfobRu%.IVgvyØɩjɠNR8:+`2v>рruL10YRdO\ !ǶGmwm,i|?9iz[-h`) %E8pCm$+EN6AX%ppG&& Q3B 9Bf=`NFH}rzzBTz,=4l 7Ga&3G&H 69Tx$q+F#.CŦtJ{Oz} *P m֠WB#_MX{-l\ph4jGcAT+eP8 xR=ʅ%$A$BnN&1÷" !MʳSXe 19|%Z)*!Kn,RU3"JL ugt1(aH˂5rvHxd$z 4\38$./k!d\MZӼj`C3מӯ\^%iEʢ B`yۀL[< vh2i=$uXpiZST 1PP"r tc]`wɓq{;;q%dz_1: HzϬZ Fq}bpȢnշO0 -bޅ kvWpH`hQEmkv}@(DiWV]W.Ƶ \6 MF-1KPdZAN?^s驵ym~KOx#>},NFٕ*VpFaMdz&G.}[RKgX0Zޚl w+ٻh~P Ȧ}NWl]}[Dh? 7-jm~~ϫ7Z eW& *?sK]ǚfRJL>m{l e bV^\rSx4U*|I?ni$҄.pE]e1go`Ei"Q/Ɋ w61_Nѵvrz|H b | 67KJvy cŢ EƸՌ%'rVbE y0L23ˌi?/Ί Z]4F1T_gOk&G]JAoT;6$OHqs( iVvæ<=B 3pwy`*suO@Q^;NyzZ3DIŠ7*\_Ten1mӒMT&)sZ,x4k\YQm3:{;j3-mbhVJ̅RcCXk8͐L9 L)&9.ϧ''D>?Z8igwQ\H"T#.td+93b4Ĉ8A\ zā.iǧի5x]Ņ'/Nƴ|c.J?'uhA>~?^q9h=r˭8yʵJuw{DoIoE 7gsepWy"$~|#A`/C,MƅSDp,s) Vi2@G: _h,LJ`F | \K䭗ȷx5rvָOAErZa䧣ąCfHx)+UHsK(d~h;59l*wIAS$XtS +hy.֥Dvqjl8붬9$ ̧)=ZYĚ)")9IE$s2JaQiƝVc +"PO+@2]E>  ap :=SX?T[5Vq|`u9eZD M.b|Q>qLSF{N>LkpUԢч!0'?|BlsRbr7As|X㵵k~"Mr9o)> 8пf$SuC,2I눅#Hƀ2o 漋ێ:i&oI\v7~->%WYOT`Mg1̘FDX4R=AzVnx+vJk Uo TÝ"ǻM Z[5mlooNϋ I)=ق1hd<"s-ϕ"[ATreS]#2} i,Ux@ ,Jm^0!чhH A#6Ѵvr 8ȹ@Z* S)PKA?"D4[#g;F ߑЧh1l4#=GG^/uB)K|t AL ?=Vy10v)@lݰ6)`:GPu f / ^ Z1omlwͶCZ_it :Һy6\t Sry24ykdl#% Mn{.TcF-VGe 4%a,@+I`*2MqLM d&#QHYu2Lw(L"^ˈiDk45Dlk=$J|J} g"SC2}{Ta<=zNg<ߡ-A˄Eέ:~-UWUƳ*!ݚIU(bocs)[<\Los~:mj*[L[|uخVixԼ,F敒a2m~=n;hwMXCU[* z1pWz!^ڂ5Ew֜l׬"q7CEr;tH_F7 |!A!mMѧL:3Lƹ"6Y[(`*hD@MGPWx0▱y_::ay3R-g HmTFzU#RbHeӫx Uv 8r 2`Jʯ %UaJvOr~|2+%4e2,qˠp, h>YS@J WL^1yrb>s˭#:F IX7˭ǯraD"FKZP'I "ˠo˛LЂ!2Eȃ:gڱ\1rM5} CŒ/;V9L0Ơx¦D "i̫FH(򹖜N)bxЄ~َ˒,X4c;rl&QDZe '?([JƩrʬ=ǰ8q{V2@A.HKA: ]$"-1z(2n-H`|:yg!ҡ´ڰEw/-#M Shw,,-hfs8h{ hg+shKG+!* mZc!ĶK }yrWծV BiSp\*a)'Xs29k#aJͨޕK\HV=rȆe_ѡjflg4Ӆ8cO]hDc].|U]U@giW_Ѕt~==[,q]W)h6+1U8,ab }fJI)mnC*:{!KbJVKVG,TcI-Hη)hE;gy(Zw6bڣdݘX_h"].՞k~*9]$ 4U[@7ՇXR4{lfB";II'ˆ#( \Ljr!4ևy^NSlUǾkD5|-Q|wuK W,TH+U*W TBﴱB:Â`K-EqM5"z:;1*;-K31Gv&'|JIKTpS}X#6#gS⭎Cu6}jG|hO!s}1ޫ s|YLeNt kCч͸PO kغx M?7V?Uo^Xٿ{Vxl)kuW[{OύiE%U tJQ  4ʇԕ" 2-m$kMS"Ą9o]pYD8*$FC-4$+fه5OyMyc+ݠ+.\-OW섞(:3vNҡ50f]@2hdcea/mx d4B:JpB0%XU&UHnWd7LBU$ UI;fjWl|HbT ߋbH!`31vUƔ|MuP{k*aHZ_m8]GkGEVQ{9% rE`_,'%Kltqk0C !ǟy a*z'&500هn(x=C Q,6) DFģ,EcĭX/yT=g<ߚQ F:*'Kт )B(,bE dyhj4|ry EOi v/z7GqR.瓊 g4[v'Y_⸄_~[j6m#JbQʔwz|w.}K= Z.C: "Bun}[NΖW(+: ϧO>inLyŇ>X-jT L|wh%]״GgwzNwlT.cY?NO[쵓ΧgS;2Rb|{>i}yAik:hQNT7Sb xI#In K-&|G'2yym>:3\~м\5ЇcwfyZLꘛoX,j||`X:A\XM[}>{+b UIXNBu&m"%Bm>˃Acdr3Wkzne?j?vV>;`pr<=y2G{I<ݜ??$||P.-8tLr@z"T!rP^A'f+v8 osPL7 d1Q%h}+n+xX]ϳ3%7{n»~b=+6>6աOO qx1 q+-cfxz6'g9j}&&ٲ؈"V @ (Aα.M^Pe ^]7(P&iP(Ţ1AaN)y!R)d)Ha Q5QzS98^xkGtik:it=GJ?'_>ZYGUЅ3YJH]$;vl1z;\)zu'xO^,iv`?߯yva{K Y|U@>cvUvQBPYSN5QIϴ8D DpZEIJAP2ܺl3rv,Ѧ3ͺ;JYp_Y_~eG%1!ͅBڰbh6νeQa%%1=$t!xZ$(r,5n3r^* wQXHXR*"S(STjHֹB>¦GNg3+emGay aaF>b7/4Ag ^p2"`N: /̧)!d V2?_tw[_|k??Lۼq:p܍{83'3c` c?y`n),61kcJ뒴$'-ttjg$&bW+2:cGG"6S{te^njloSzF}-,@V;ؠI9g6u@ 112] ]T*yвziv5r7?t^GLȓSM|_=,LԛGy{߽mjow~6 r9e3Rɇ _b'r5|&Akxjzuͫ؊+Vq ?z_3V3MM<[R#ȩv\aұ+BCK4v6BzdhǽRgj?"A/p$!H FIIQtqR0j9¦]RDYS*t5+mNR)@YccZhLc ">JJgPua3r߱4!5&qk82EWy뤼?OqHyfM/dbڊV0(guOgYmY P-Ia'5Hj;C}bU} ~6('s^ݓp8ڔ@Za|6߮I-@YQ*ۆؼϗ>&6hvŹF *nf@QQm񙴰E Z-5NUo˕5Zحo@{whO-|W־Dvoڨ0-,0}Sʠ´0`,LIFTu>*o6*icTOc32[e#&EuF75,LѦc۔ <f`[yQNd/@P}/))KZ@gw}m(B ҳC@*`պQ̃uٹwlo7,"JgI:gTIytryA$D OL}^ARy\] *E}1aݩ33_+1#ВO!ZƢ֥gpd/c衕q0f,%:(IUKDT^FPFAj+H~A3BCMc4*i0[ Y)I:k!zx4fUS4tuE$^MӺ&WdZN5! ]keͶMUdS=5j\c./?ޓS<{Q}G|1?c1Y# Ԡ @ yt!NUt :OG%+#b >J* #MBP $xKyz\!79KIʞf}w:1$LftGӧEzxe~^~xy~q|NjNzkD})tvj7oj3fj|ڽli}du{ ]ʿt<]0V@O o9fezvrZ[k{'@¬6ϫei,f?k!FRvp$m|h9<Ճe f{+XWfc-^ ԨI:4WVv⧍ ޵q,B,vwu9©f1i"zqʖT̵ĤJsu]%$tdt`VspeJ9_kwt_|)ĢI@^`^M0)NX8=& LWftm ]CZ X̦u@udOxߴ2V +(vi훧oPN웥-]l> ʍ@b  (h%, xlApxJz )(!D#$3 ]ݜS{ yd*$p!4Cse*f0iB' *cjù#]WH/j=jॸ`Õ6nNcI|nun=:z"tJQsz$+4Z %qKrF$d]fM)S( abPEі_շ=wyy=NbX¨6+f#'BZ,?g?xJ翚/Ϲ{Q0)1QF􎦞=g?Tp8aѕ0VU >4qU䢔QN1yU}p#0b ːpRr6s9x'=e?ͶA;{˛ooh 2why0=Krλ<|ZCoiv9z KzKBJ4_޺kQGG%_=#_Xhb[nmDpmf=JRT+ 'ԍ5*8ab#!BԲ7ck}FVfk}8Fa'IW,$!g8ׁ 6B߮ss5(QHw 8@!p!q]dɻW:ȠT0N \H6ez t}`eNĵ\]b+΋v`J:`UsY0l bzDxJ}%-2VmaϾ}mluQ5@;a7'a }E`BHե4Vi=RܟӅLZhRpC!Ǜ53LjHطބy"KOx ("c6tjɰE$KZ5@xN({e \|I&gYfUmLE| :R= RHet1ϊMJ':=+(O^Eg Yha~YPPς>6}4z OvL&D *(]!c5):PX!<UO\H}cbYP]VdI/(X;Y`韰2kSU펟Z[tNiě)Ix-Sz[tzwcؼ 牠V^p!; ۊ=w84Pp]~wk___Nw&5I+Za^ȦT~!%'i9毨LX}I Ғ<67~*o ]͉~]}G6Lw[Eqҕ\%ևɃV)or. qu)J1=ZnG;zhLG h=nC."Pt~O9s!٩zS?g z!/SRgewhIJEpʤR٤rp:d@ Lƪz{y9Rݔs3t)ͻ}C6zKg}&0̝~~߁btRZ;ۙ?GT#Ih$B%#&U&5Cj83ɓo|Ldd 3@ڪtc^kvWQ80t۠j[vpd^^t `|^jj+#dkh9M^YDK0C ^lt!Jr5i'N6x2O4;~7ȓ٠pC<5'yMp2=>FM:m)=FIVj4N)Hµ?s?f>/"?a[9^pcxGgvn7)POVY~ Ųf>f_Ȧ%|T혪S}sB?[I񵨞zZ;I:ƶɼh?&4uzMZ&E=B`־`Y6iq!~+(˜)`zl`^ӊ.SHn/,aW~Z?r[mSqVҲNj<8Zlsǜqx"Kgڹ+ѬϻZ0Kuq*ֶP-W|q!N~6)w5˨\K\~vN曆ͥs-pP] yS";6+VcKأaIٰ;Gi=gOXS6}@2˜[׷}_2gdi6[<.ϿCUctOm.Z,t.&X=;X6qg٩X3 j z_}}!'JxsY?Φ YdGDn@KVPF1-%%GZ.PVާQ5==bp*DU8LΊ=s@cIirgZ1FѳQ̴`LLX+*JH`'g~6wuW+cx=}Fٸ%seƒU\ 6^3J\>T&{@QqׅYhJB޷FyV6&ʬu6`gst9=*)d#?~iMʃihqW 0$,:eRUu!*3!J] J=8%AJ&6Gs<'ǒ Q gǤ6= dT֭Hͅ[]l) Z.q#6g&eY(M6FBK21+tGgTE$cɂe  s.6 p@7ܴb9Ţ!et"BX,蹷hKS~GcD))/},P_9l*dr`Ω aJ$,Aÿ*b7d7)xt- ݔ^nʫvݔq+wfhq[3csdTttˈz{ˏǓyCJvI`Nj2*О7UnTd :& -BT"X-eS[VꔞCFY 5I2)G\mt\,FK|Zp}nJ (Vս'=mU1Ω/J^0C)xl]v{4PO3WZ6y.f Y#X kFl27-`r$Ј;d~؉,d8$# ,%JEsxP^"zDCLAqR!8rY!Z̜&9x' j gGV}Hq׏Nۈf:dc0j+k=X|LDQ,Y 53Q#gBEŬ̙@!" ּЋ#@ YxI bZm8/yv3!Xۓšs>M.o׷=˕drfyg&9-Qӥhwܼ\jp #DNh#!Lm ePFdg9+oTF)@@ 䬼ds|&%U[3V횱V I=uႹ.t.;*} T x{dٻ޶%WHKTok;  0z8HY,ۃߧp)q3Ք("4{;_uW}iF \W35"8Q訢H)x Y NیqDi u)p"{.xQB+ۈݲ9V2 NZ;5㜉y(Ztޕ >Lʼ4 ;hzvjҐsn$4 `X4Y*Bqh!#2d "1Q"7A!pD84QsalևQ?t/P4b5U#Ոk% ]C9ZRe\dʍ\ ZB(ijDkJ[ h' !LD <=Gc2D$J0-f[eX;[x%|puHgդdW(*EV/u $ޚ,Ndx^)s6H5wVČ&1N[x0PMAm m6{'q ~3z}؍q ~|$G&]8C8,lPn4-ˬ1酱E Wi: MJ1ec. (Id(i3]0nr*2(pQ@kߡTg}r0U!gOv})mڮh-G,9ImZt5INhdJ '`$3BZSrOѴv7hK* G*EQ.rDP\HZKA$o-iEFә6moncJeFw>ԕp )44g)C.P)39D[16>' aغmO1 *q0pьi\ I4i/s_4K$-9źGpDhAR@;0q B*RY3Z [o?؊h dkqȁg E%+C(QBBI|(BvqYaL",yT[lg) tSZ4E!x 7+\V[;0:??=ڙcˌ㯺8IU4 e2&|'g?ޅwPY 42VnM=j>i\2x'u=\=Qsޑԏ35ށcf(23T+Շc7st<̹^l< ZQCgK@骖Z-jnFf5.ĘFG Y,;4Mhit}AjuXQ5ӛ&=OPʝr|ŽYg0XV7z08MB~~|>}Yp9%h+tVT;q!oTLJ+kPR?.뷯^UǷ_[|/?^|/ .ܻ ܫRݣw7DiMkZsԓ ߦ]Us+vvylTA XB}.[HXk#-/`tU#V,HS)<{Qv11YLYyX;'I}0U.Y= pAC(m(Z%\ޡ1 )М6BV$WDhluTj2qhIvavȃr3óV;㺃غKrS{_'e "1O UwD|[2!!Iz2Mw::D#. 9a3)tT0G-wSN}@Y3c4*Mq*Fc#H,\7%入 ]eV:4f* ~^6J\vÐN.^jw 7 hhH ]l(xfoTɲTeaJghZeP" 7 Unfk8Au_b?러?߆ӬÁ.o:-HRTȀo@[ DQ` d*l_@%8H02i$CJ҃Ar PD!4 VA o˙1V2L+1*غ.10@GcM9O8c db_;xlvWr;0" X٥^;17Lj ae4@,u$hڃEu=T]abp3-sSfGhO`2XrCtW=)P )EL$3Dep\U>֝-=$&#SSBy"MYdPd޾zdo:kAx r Z_uc7_>uK7/ɸq: l`T% ߾u{›+/ZM-j+^JM'62?fkhR'ُnf<֨Ds:imyV8L-n<I[/GI'Xnm8"h-ȄQJ EZ2$  ~ %rp࿎ [4O@+cf^ocuqJ[HHŨېK$˳,%Z:Vn 'yNhƟ^Ikg=H\̨|jЪ)m%r&YӨ7J蕩zRqB_!8d^lq@Xm}'/ɽv&BG>BkΫCF)qOcnc#Mb+IxX-ُˇ!\><R BHE˗g0j(cS#dDKJmIa62DUF:Yb>2l`BUf 5C!^)*+Cj"j{g)CϭT,s_}yp9S]u;o/=+\zlt]07oSG«4gLiğcC/u &I+61+.m [Fu bKsk"{x{㟫BڅjYZſ TN*Gs8Z%!Z{. #t^S.%XCk'..A*~9l SӄQ[0Lޠɬe΄#;>bm=I]ZҢbz{;eyk^-T "* )>FoZܽk*e2s~$ a7 8t^9[c6#8>?ru`3|qӈFf.wd(+1=Gwbg ,׳)Ā49x~ .f>U7h\;ѳ퇣?cٞsK4{5z)TcLCXsH"9!*GC^hi!sg˜qT&AzO#IGuNdoW{&SJ۲Y]PlCڢuDS|uݓ?[m|(U m.mv=P{1% S h0ׂ i<>abûݱ{=+cpZwuwmI_Kԏꗁ|*KoF?e)R&)a!)45(g X8~MU]g1C^nNn-w3:[TZd}.G>b x>qʴv&UkbA<8-Α(Cd(T8ACvnm:;ZdHF҂)%xhYU>5 Ia5X䈈rCG}r{ ILk"ڪ͂}vi@ƹ`N^@=S: ŭ;9)? gcM0ṔWELJND,d4&Dh*}eRfIh]FjQz&T"Zh3%28KB`l0" $Pk-]fB 6f}?C7?cni;þp-MjSmq3Ȏl bxͺ($)OSOJ6Fרd2ſ?/fO*FB ~]puzbqh/jT0J('욜]*ɱ{fxwg 7_-Ic,Cl$D~n wӬ26"ʷІ [ m9k[ .8m1˗aӏJ* ڙߩ9҄)F wP6ҚyvVᯕ_?LO?Xgk&vT{]]0 ~T~܋v0~4eH ֏t4 kFB01QyI>f=w՛lZjFF6:[dӨMkEL:n8<@pT٦o(RڏSmK58U-~ \*bxu"Tн%ul̮V;E}=*"{SuU߰T^boP?.|=2wW?*|ӻ/~zeş/޽'< 2>'uhGM@GL}CK8lhais e\긊qUDF h\W Fփ#clԑ&%i֓H\O|^MGNʊ ̳T$̀1DJ"Iȿ/}/}(hTv-$ Z3N(%w:ؔ(D$Hov+,=q,޷X8|%b>x{stKP|T6pd_x@& G4L:|q/1" ݐ,3"jG#u o.gͮϐ#FO;mW/uNs:k> tx*SCE&4cSPW9opDj<Op+ ,$Cr#r~H ĖYXIuX=}y:oϵ nCC5)-mſ \,Z*jqm$ aRr;3x)ͬ2 E2̤(D(hn w.(kx"Y#y2PD LWл ?*OyIUaY8bxTpk)?Z=G3ɨqQDnIZ'|,%-Ugs A[eᴦ(ku ^:-1A6ՄHt4q*AAk/ .鶯6e 㻁唘r(nw6t)g2R.gJ')y`NE݅|SfֳO3ΒF 9 ZD@JJMd3-QJъ]n'C.^7o>P.+os,Ojf~^.=? m[pwPɠ}lwx\^] 1w{ij2US(b5EAMRˢ,4@]BC˕f'?yܝOӽcw |ڎT_cݤD /OnG?),xQQgd$|I<bI999gQEE/ZZ"Bvj%J?/Vm=F+I&zd^m;YR"xEq㟨gB0>są *?$|Y,p8ys=@ m+MYv3*48 beUzJ e&g w9Q1F;@YMwج}Ɓ Y9Nġs am"dr UQG%0.wYD7p;bF'1\Ry<*DJ"ZPFG0/@.zA<%nL 0bCѐH2>9_-Gp$qF)DpJ.t3jU[P)lNr@.Wx*yk~`<Lx565.'ӏCFD:"},Ql"軑2iJrZ*sQYpQ"qeyvH7 "a+SZbL8+)ÐHFRq&`2 r"UF.dCWJ=޳8X\/?|N{iɡ(E~#._3I'cEY+(D5xb{gB&gCisCq> xB7:)$cW6j5|@C|etW(),J0s]2v\%4ȃ{4ȃf e4$1&PBtl6$((g8IOٰ OV+3DQv&= "4&VK0\2Fp\+awL b) ^$08{WR8"vwz^պsy3MbzV)WwL~puupϠ.ozsy3^fϿsak^I@ eqY uRĘwP rXGxy:BO/{CI+DI ! R$% Y&')e] [oAR< O>^T?S1p~e歾dhh$0xrIs 6D2b:yDGb<&! BV??7ƺI-zlv_1qdhP"ECGv[@XBėjM7 4ˋ}ʹCM*YϕKin^99SheB f6ㇼ)1 8]_Y4rM,=L+RRyˬHKbKOKO l/74yTfwx1ɗtrYy6?->.OMu]y?.os[jO/HKqr;Ms7?ӫQԈGQ Kckߊv6Y_qRvYHWO-~8[՞fvNޝu[YA5}ccdzs6_ ut+m?s6] Pf-X&Oo4wxo%a$~Nr%k|h2} )uz\ P9.VCrzlgoF?"|ݶ<~p39d[Qkr[qLb@3囚cw/E3}. Ъ+ةXlU)I~}-[S힊5-(݅F<ȧTjdE&7NI6I7k`!ڗXrR72{Î{+Oj`l /i^r7zdEqGj/49LC/hcZ֚@~>p7r=MD_Afv8Tjdkz;gEغާutܧ,}m[ګ-}׬ *zZݪIiC?lmV7" ޴^U(]hUt6iPg*\q;'wN&@= ;Eقh{VH*M5({?ݎ:Gnqݭ67y.޷E@ݍO,RW7jYҤgv*pOxjbBeumaăug* C7仙[seW |cV*UI1ؔd̙tpWK-xT!s'1.di4<?DP% 1@F<.7* *J)t ,;hU,Kg@wA@j`,gy^\_=Ʒ~dװcs.a/q'P0>/p, `XδX^L.ǔh/S~\ke=c<(ȏ.2VcDi.1x{6}< cr<>j0Dӝ}Z+ŗy9QWtlUD,gg,RQqٸ8vg# nW=?&ͦ|Wr x9Z\}k7uO* 1j3f- FQzv9o%:ovzUunuYwb7@'Jԫ<ul_ /NDҸB0+%%E.yʊ@Ǖ}4UV%z 8SܸpU+`}*6ʇ>q8e!nE$;uģS1:.q]-2bqE\xc)X;V*;.:|c"w\G :Y~.ɷlwI Y(Ȼצ@T.1q:k%ڌnث|œW'SO O0wӣ<Ly{]"gq ohʡჺ7 ; 7yPnu[-$e#}N__,] ZyFgzOWAчE1hlqo (}+Aonz>t􋺓|u}Gġ|ux֟z r߷l-$hW>Nzz{lwX{cq_ 'l;n~nrF|ܣv4#U\Զ0x*;3#@Q]Hnb6cSqif;|>=Yiu6카F/WdV$(Pw 4Et$/$.Yt=M'ۧ.%5\NH;g9YN>9mXbO.SWz]{9*:9m `VW'cAE)!I9$%|mߦ`>Hq 0-|2>cր̉d1,i%֮=ydTĶO{vUZYsV!592,VI1L!QީMWZL PEk*>]*Eň!2)T6ҩLgƊsI*1c 皠EE07"D^J `tdeJ 3mGC ,9nN=CdDc׀ U0u 'hSKa}C 1caUJTؕhYȰ\zXbv:(ǜ<ٻ1oYc}$HKd.@P AOc9Ę(ZqWVɰ3TDW `$(_" 93p&’ V Ev LЁoU{Ces CIq3F0[e iHadA VH],ZJUk0P6 WX,pDj`&.\&C(41IA1Δ2i% ~a`z$\! "u Vr!Ȉea*3((v50苄*9JfZ 4}!PS}DW:l/t)! $k PdweeEeFF|1kU-Q*j`ŲT?!0Н[&~hټLIE u,FmUolSwq`yX-$L>b`aJ{a2]Wҥ`r낄AKF00U8&#q'~aE!t^4DL2Ӳ ვEJTOUaXhEYR:[&@8ڂ d P?`yrx*֝fE ;-&sJ qYI J ;/w&-I,Lyk9TX]R]`QD \{# ]Rb8AbC8*f7K|Y] P6BQE\HP_&#(= ,^BGmhv)0VcZ8As^]k 5gV6QPT`$lܶQi< LCPHݔ@Yc_XxH8KnF|d@RP0;H-2bZn〰<9yCpYTep!ƺp",U@k69i?`ay!td1b,F%i@dB޼m`JUUuY݋Qm^ @̤F@ xtW/2cÐ% & k1Ҙwr NQJtikȲF&AWB`b[@7j!kKx`d*BmtLvhRpu.֐DMypYmhq=(oF~?|Aao𙻺 <%<`:39\^yВ!=tQ˥P4LT@=B!*|0z;[ j056I*dO-d_p"\x>\!N.@Na}; ?& Z1 `S!/B԰(!RXFUTk z),A2WhâHYa=+NF"5R&T?R_ 5bAP/6'W+O(A I "t$ ~1X?Fra\[Ἠ[~xYk}i3QbT ciberffOqJf:t `+߅X4SpFX#Jx>h9D)Z6l!& [[m:_E.cdfVMV YB#D$6D~ӅK0p9/ [cy]jFկ+Neg,,B#W=ld^Bƶ"7*j`VSꃰ6s9H([!DӲJ14NHu~EIVY5Ez {0u@{._yyٳg @[Ԇ{>yq>+E'a:j*MrI%vb9`pxE*4kD4iOkmߠu6$NW*vP.v }un7)]v#D_T-[0 2 ĸwx~r;jnS:ڊ|[arzF}>=vQ{GyÙ`\M,.4r a gxpo'Ӏ&~ve~~/n۞'l~Lg%<_>aqzK戁σ4/{#ŕ;w'[rꙎ xQ9TaJJj d+dgM22}-I~scNaũs_eI]ٮiIy~3-->VYnݭc˛:? n &/goEmC/g 4;s|~FWKJam=TB6tV|촞 #/ݪ4uyt݃R )fV>nf@nbz#.rIu#~ ynO}E(D'>QO}E(D'>QO}E(D'>QO}E(D'>QO}E(D'>QO}E(Dv)~4}72`iַc 4mS@%p/!91J"uq.*w}7u0(=:~T>tez1˹|AAGpi{ey%y|tV6 _sRb98kge|2GʝQMdn rgq_x{8U?0x- uw}7\X7w|xBut0Bu ]~m-GCls4k͙8Ĺ&5qsMk\8Ĺ&5qsMk\8Ĺ&5qsMk\8Ĺ&5qsMk\8Ĺ&5qsMk\8Ĺ&5qsMos놵|Akl˩M]UpVͯ[;;f)]øŧsj< )6 P},I,kV墏Y=|ퟮ}戮eXo8Vp 0֝Md{c>!2VU^B/bкCF.`]ϩ${`I|r2#M×6q4M\~>\d^?õK"ly4=$_;^z㻞\λ>1jCvEM҇25m~pƤѺAkfpNmP?ãu Z6Km=C{M;_+G>L'ZGK?B灶V.&fp[}Ts;Efa6ɅI_&rx͍{ͅyQV;9f.ܳJq QR #UVRgϭz%,uT]s~(_cċ" ܶ@l)" `u#L]J&O!x%y5Iw!@Kd'ʋXCդQx?c v&6t=Sff>mrˬj+bz7RTO4R=Z;]vTfM8ǸilTFhXGXTԗ(˗>j/@|䤥u%SڇP5Rdf akE6x\d97h!1&-L7~nZ甹>,si+|fͼ<"3ً8f `;A>O} c}:ج΂v!;΅эhoT¤FJ%/T)?Rzwz[#,CA 5U'V8s5HdN&daI5c6+R ~CB9.?#\Fzۗq5t / ~Mv1Ɨ1gp^Yt0-mCxy?\2 fg:e9U]'6`kY=e}IrDaU˥.sDT-"(+V 'L^Fts<8w8Ok%dڋ]_T91Ţ*P\AE95<5pr5*uVd7GtIG<Қ(&e* B6" 3'őhs9{]MF|acqsk1\g5Ǒ2!vG֖v]oGWԏj!8I~aZ&,(%')ʔ=lQw~WS9w ՈZWcFJ`4//: 磐so?l䃧u[Gz{٫nԀo7^Uך#knI1ʹChE@.mDݸwb$I8WF<0[9Jcnq3+r4ѿi3D쮡3Lpd@7H`mQ%{q$ɸIgk&{/^6͘k@6YD5嚎}aģNNkf%?֏,&yeZEVD9C"6zQd`%eα#RӋ[}gCˋ@@>q<2 x6 X_'YSJhR9D*%*E% (r~zS9jft~qDt㻷q-6t|L}trbo@yBz˪}ΤԷ @}ן4M +Gy2 ,ddjjQ Jt1X3w_l^ [ҷbGp?~LnVoPbXy Yi,u|E@w^8WE6&miq2ID0'F Lxe^dYE퍜 to,tޣZw}_)yٝUp;tq̏?E8nØ[`֜y/WTt"-IjuX5=_@MO]$fr),"ΗjzҔ@Z%(#{o(-rDȚP`DrcN!F/H"*HdhPb:xo|ol_+rGܹ?,L,.ZcXC S LDv)ac%$)JJ WȢod_L)]1YDsSό93*ytVsB{Svz/.K]NǫDIx}g;kR$ =BRzM79.xP13&>t>BFFy@HOW 3@-i?R\Dt۔DL- T($?'gڔ׮_zeղߛ%oor^j6yZ[QCC&ΪgG$1P2BI!PYͧ0I<<<9ct? c|dI*qI8PnIn wwtrVYy[]ڷij0П]NC+HYfkD^^Q*Ί5^ .o=54LX~}͈8{w<7vtcLۣս[7?O~h|{qub@H$$|O|nkj/[xd94XՓ+{sO'QWfLi+x,֓|:9\L|iTb}AuXY4Ϊ}L7Mo7 ߽w-u9TQ%Wd$}w?o޵}u-tD&|~mC^uJ]Nr+BuUqo-cb|卜6MmQv~v"ؖ5/LAc1uU)Yb)Ţ`]Pv8,Krϧ}t:9 P⑭m J,b`+CƬ_IL0 VVBLRSudmJ0o>!2MZ=q˝ bOd8Uiy2;lz {>,am/6ElD`.QF>AFFV̎*~W>V3 B[ҿdqk@q>`9bvp֔Lن`W0~kcڄ=k +LcL2Չf) uBuyXGd ܐ-z⚲}>{]tzjU>O7T1Jż8#>Ye:IӏLaR涡#_h> ZPp4 ҝ>" b4gI>#:o>fS+5TL'?8ϴ]ͫ>b`owOVѱB8Q?uFWGB2k5kEk/2i/GWݓkǼ(^>OG)98w0zˏb5߽S<<yIjuUk$r[w⡶/>;#SdSj@xgI* y"J7,n\/Z4AJN =ieVI+\N\u5o3+#0k{l YӖŰ}0rM!5P*o*:[fʄu}7v)ݺ30z&/ܖuX۱^a~aߝURaYi}r)|Y.%7HP־Atc.v| x-i*#ktye/dbv"Rd%Y :%p0 7$+F:#AP!\( /U ZfUJ:^7I^DD  @$u BH%;̦(m_ef!`zwBD`zdZrj5 6Cd{  ?1gXVC[~pRL~(xÒ+?fQs,u[ }`^\:$Xߠ]lG-rI+S$9Y$K Y#rjլ>lVԇ \naj|p\NԻQoFayKgBFE86M>CͿFGQQpVy#EOrk[[zyTghOYiY&lAAaF aC4J;ha2(mS*y,[U(-ٙA0^U2jkrJ9V<ޥoI`$dIR0ٜ74hTmX RMV#ebedFemT?l܀#Cw8N7>p0};3+< t1TZpq,L,p[Zi(I֭,Ȓ{>q&B)Z-.ZN029bfej_b("QljHNjwl ū0k%+=I2Dx T42s|܂+RDthlUy6>@L!HB$ $ YmԤ8bi|HY%ɨF}eyX 1u`<X>PY"B';Z"DFʤɕs)+2V35Riƍ"FPD YvU%6lcTRD2"gB dLFRNXy4U9 UNYJ\\t#Hރ1٬^d<59xb{g"p琋'~Q:NE3UAw^6g^oWq Q4Auя/%~$VM_=|QXRE,.A%c9 hAta0l4Nь;0$#}Ps#/6i2KΣxX =dY]<6JNkؐgRh\gNs/9z+HKLGfU֫s`$ؿ?ݝPaH;j2Y xb˘zv1jK*7nYưR/9t%]H2C)= +$l./ި䔈Γrptr:BzNEodv2øXL5USjߡTCg?oճ>dzV"d寮O.rM10β|!RU7ZyeiIU#S їJbz.:A=h;vKl ނy+ðm6hHWn ]7)m`% okm9x@/$q4mmZMmxgB,s @%"pyqZA5TzEU5YϧT6G6fO6n',U*;͑<y<<{M{v/R'wn""=HpƿFeSK{{wTKFQiA1(QzK1m@tgOjH`HDEPY1e:D,i|4"pJ^m4][O8 :!Wqcv!j淒'c㕹5}dtOדmEx_mm]CJl@t@ (/xɀ=1PyǀD'!.ܒUZ ^w&_uK?71J(FFeN0Hr"a_8H4< >" SH=f=!] Br`sTLr6K2y$n-|<)^tFB xi5C:r46W&2cq;RrB7ц y}hq4k" A_?ƾ.oիx׻ YwIի\0jJx(r#lҔ%jiΊmr-I6ִݛ/5qS<cbul 2_!s? Fy8p1[vAB6ч~!M3 :TglOe41h(y(On9ٮřmlEڶV~r%4D>?'ʽ ~g}ۀZw/5W+OןތXL#/^(*DNgJ{H_!eyh ӽ`Η4yڂ)JMIe%H*R]lKbUdd/2##|N{ WT:䭐&=(o/;GoǤ ]8z]}xQۆ5k߼P|cd7ir<1|`zbg h/4oTt fFQ̞١!8vj/\p_6Ǎ+*K |*i"()(sfFN)oR*:eE`@+ ZD",_aa{#g3FnYcBfwX7vTj |חn(|| @;:f wcFк9! ovhnQs|l9..tr5wx¸ٹQ#05bY<7%uFέ_НTC4]~6c?ɍ36^'6O- QfG7YnШ쒍h[A;#;qlOGwL*[k1欏N3sKjw/QranF1!B)N1^ڐ z_jNiB;n:F"0\jS>D3I%X]ltYi!AaN:35Oo@jwܯӋ\[evgZe->Q6Y.6J}.=xJi"o&EVe;'>g>g>U Kغm,g!%²Zeɗ9g* erdsqϋ ِO18j6Lb- UzvEK\_ :oQsl0-D8Cci=/KYjMs #A齡t?Lb H|͟ɉ>jJ@ k!40R*M}IQ=NukwZ\nϮ?n&Úû?~>MxbKʥ# 4DaUaip25Y:lyvF9+7B`4ҁqU"-$/~ۛ}Za S .CpEa)9S,0O tkpe6ۮx&m%L~Z]~Gvj1ko -EgmRKQϬqf7M/rzy^\yּsB/&enwI˳nx0^rsK\ qemkxY$6>d:H}5rbʾ.ȼ$R4Ej>RE\27Edѐd&Z+NWCml<2{; = d<6nyl.}q\t;>=W:@TZis۳*lW/yx>Q(OFX+,4Ѱ,IښCښx'弯tB1(Sּc<+^RY%Պmo%Vȵځ7W-(,]O4~)kwy!c9LoЧ5ͬ]!sU#Jr 'tl$ A9+9|(u)-CI؉ l|P_ p(2g2U`VFhXO+\*ʒ7/\@UVyx!q]@ JޅDT ZC,=C.T;>^ye<:٬\^jtڵm/& =lz.0w% f/W=TwKifqqq>~!7rlu O8uIغZnl_[n174zPk*,w|y՟ܚ Uvlei) Ry{79NΆ I{nJ JNh-E*r) #KI:M}%h1SeV6!3[0⍱\]ٸ.!ny1.\ljԶ3zgLhdG|;9O_,[ V_Q%`ۿ~tyXףOgm`-Jg5cVү:k^p[%n?kH 4WJ[oiQ3ƪ@T")|Hg]Avпցrγ)5Lgt5\β*MTj̫Al0> ŋ{!ӀW#[6E,m):ctb'6aGΞ&gNJlD[Zv\[_Ϭ̞bI |X),,uN'zU#1ߐQڟ4.e{qI:N>i aO>ifdVYZYzW-ycOrc?df$FZHyt+mC uäHgv^*rr(Ebs,ÕDf,+ZǃGk3HZ HpgqU]tvCv' (&'JzAxcx.Ob ZE%d H*'X,'Fp #„\ tZ-$$˥_D/K@B]N*е2|3.cEg c)SU&!>j?J,tvD``tY.a &R&LGȺUx׃x& /Q`%'؛SΠ)uK` _%f`!`~BXO)@AN) 5B6f;MRqJ""f瘴 >%Ih^68VjD&@Hym$k"*B()$Jd $Zqap Q+cfXwęV6A˖9K.%KBIՈDErٰXԉsB 甇]GaKf&ojEݐZc.bR%Kl.XRO]y*jA0`xY.ufH9~ꪤQ0tT0h1ը*$#liӫφbUUtDt0K1at"J*4Jx$"Ȱ=tbɾT\N&aF t^\ר"!w*;c75)Z`j?Sj"7vōڎ AG 7QQ >k3TG\}Ā6B4-R.CㄔMZ Xdb?.7us-V?'ɪ&PUY0@‡F>%HÞVrIJQ t@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJW 1J Q\i֑֚3TyҤ"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R}J 1gg>%PVZKJQ @"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJW 8sRaT(`F kyJ @`H@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H (knLŏjZK/7ݩ<||v\y5pT퓬 aC]'3Aq>ڬ˳\ \_G_XuۅKx:|K,'}#*&M~gäD?LǾ-]3:k_oTsvz2j]k =wiO9Ӛdܜv>  /okd|:~wÓm3پ?MJ}zQ( 3^^^Ǘ|17#8-^ MG+>,gyZD$ˍ@"9Lk2彂у, xSΚ>+%3p,}TF,FyxSf/iC`^)|t-[NZ߲Miۦ(P'H֎0L0kSn;A}j@/VmU- ާeXpZӪ)Wpՙ\0^3HQr?*ͽqŔ~~}mI/&W9qbxvidr?!gIۂUNB}%.u];ʁpY5QZ6*UqAFDc1 ŜYr1_r}^qw_;6`=Y =VOlxם ; ~Z-+чVe%t\*L-'Gn<~?exc1BXFx54]svQs>pˎ 3D|O*1w׆vnQ%'/WJ.Fra^mG_Xm6INru?L.s鶭0Vͻ!mz`|>ӤfU.[QK[wH(sw8ёz^$_ێdg+ܵwPk9;w]{یeYV;ì_ںY9=q)C8lC ov8Nѝ% G۲6(wCV2.0:Q@EAw1F~6?ߍ'Қſ,dۭzGP<6uk -+O/7E?mva۷}'Ώk^6]r>bNKS hJHi׫g!jIFkѬbNOjF<4ϻo/Wwin;?h[l>h  i+.m6HKH6Қs͇*1KƵkʩ,TY;=IJ| V݁V_;ط6Dznkmm2OUzxK[ݦf3\4q"U U^UV ؔl#{dslؼH+8]QgB|}x!(|u6L)=\?׍)b'n&*ۇvfߨcIJbUhIpa~Z=|;ɖJņzs=QřvFdBO&ڜ$O^3]ncC$'c to^;O@{H}DۻrMk=%m'\O {ǿHu=d-l `f3 tvgްuS>lG= (Ew}rWxsC|x}qm"UUn== W羴;LF/8{T&pۖPP c$FpN\zl!ZdhJ&O>xs$y=9Ӎe9h<QH4k2G6%֨*MaIGKm4OyOelgfݦΙH3Sn6mu\4[FgLR%tG+Zƌ *ש>/q7sLN"F>Ǚl[>Pcњa)# PEfh\r)oXh)ox`>pp)1 OJMx[4ϳ^߳0,.`z29=c]^ů ^ OQ)'^Wz˟^VK^|V>4X2 3f7s+'.6kΞ='q[| Ǩn뛮ҫW?-'!W7AjjCxۇzW.My 5u|vk7x<@!~ <ě\ЉdELһ]8JY QZ\1&6S˺&c7|4 _χ@]J.迅`s58&ٲ&qf/h }9?,Xtbx>޽($=P!U8b7E L8>Lg7{?E=05SBDG( kS\~JHpae.1Z <;ż^MyG`nYϯju^d ݿZHq -+%نho ?~Lo@8=?뼦`3W˵W H^e} >Vkq~2ǵ_;g5 UP'U Aԙ:^_7.#fu]YA5BzK95ȶinc-]cf[-j?2:S-#nYoϟͯرN,z19AbFV f#z gr(_r&j)d`^sWLLR;C)\h5%IR B倶zfϲefΞsΓIXAG{ ̃wAxf(IJA( :V;YVS0hk "nZk/v {gG}}3Nz\N_o]8܆Je$ZH|D8Xnv* рkq+\֭q{aZ@ԴLZ'iKrD.%eR փUI*OSMm,SBsKNbEwmmH23 wL{{FgҁK1ƖܖxtMGRdʖCEXUX1yoR&wJ(4}Zh)bYy#{®z&rHL1cdγRu)d]9_* օE )3e+l ڌD&ӎ蹷hKS~GcX)) >c 1<)0 owŏyЃ DDn2CrSAA̚N1:Ή]YBm77Nߐ ޴Yx  RZ5\c[_9-(~qBn}[s/3CWcOmQ7UUDXz)*: 3I&(K͗JkeUhXbz$.6ҽ'hϳG.Afy{~(v;|o꽐svf?kkX`\ŴEyWJq׀d^{nF>bMKY*ɈKqv,OJEsxP쵫w b,]O6T^YSe"bb ^f09uT*f`R eQg寄^DQ+ENZvZlTp&'/Kvߞ]J_!e_~7^޴e:H>B ?agz*x`՝xZoUʛahY,$D%VrAY6Rs2(@3duB9+oTƒ٦Z g-&Ík3)ښ95c=[.;B*B½%Eenӫb&i nFLn&q^By4m"'fK~D6$cpI똍ta\%\+A$-Cq|L (jS F%+\"dP39bfBbX٬xqǁhmz[8^@V\\L!WDHHb'2"Q@4d8bI|HY'ɩF}e}Xu rX5ǮWֈ׈F\{,!qwT2V`(LLB@ViÄ$x=s n1cƨd~X.U"=R g q"b#r vZl4>EI3CuV]E{m[d 񉋬1" 1Ĭ ׋Ћsw>1TزnG=rc4Upe5,ǚG%Q1&SҎ^ޟ39=jT9(iTo\,%%{a}0ȹ: rT `sكFE)U< Cty`+ϳŌnao^^ 3 lGڹ/.s"Ol3?!ޜ`` ݯS㯦6=﹍ϟ;g],LiMjQZa^Fcis lޏFHwp~$W6}-'swpԽP\3} w,]D"#x?H ǔyM$'5^4B9Mb+oE1xȘr+o6!f[rdrkyv`3r^>UYZY̬:`p%k$R]6pR"{}C5r6+QH5}=-q,xYTJw))M=mc>ft|劇&@0\cdxpC- Dcb;< %"d^Q#qhdI$KFL*@!y57>r&b2Lsr\z&X䀄Enk01/2j{vð@jX d/PEwA8~r.yCARY)Yg9 V0S@o0)EB6wT{m~KO;O(x^x] 1~hOlrkF݅p}dA`@Y,7baᎽ~u|67Qی#تq'1uMCtv"6?}-lCRmJHG vl;rco ;{܎.։ =eҤ4oRGKTyv}[~{^XIOuwI@U0n:hq-dmxeG1vstwTzs=hyA1h~:)>Fx\AYEv*r/+J|;xB![Rc3y d0,iLt|{@ yCgau_ݒ4rjF(Y }{&a쮡hiwj6l~tk/ L<&|[mf WQ{,&d&0oS3fc?yn_8q+9+Fa&ɰ*CJ*"rz(FF.yRr/bu*aɁׇtV (îuL꨸tL2;GƄd%i@ `0{A+Hx]G50h4VN* yNָҐxRJ(#տzݸ3}e|EA9HϊQ zHƟhI 籟d*0L.w%p{jX~8tk{'@B6(ч~\3} X+Gs$Wt5ƨa2;ĄqhgPdƣBOn&gQ.&5j\I5۹NJ2҂O}~_ՄWH1?,&Spz5qqn|?Qp˙tH.!_aXģ4H hƗQq*B!NA^_?;?~|O'o~xwۓw?/L1vuERO&'x4e54ZbhjzیkjNck7VD?$5E~/|~\ڐΠ fZ2e1$ΉCFvIpX(?1p#"yX` AIJ RS\砀Fxb'31O99pP`¬;/ђ[YK9^ns}VyhhM7`/t:%VX/hӝPG[Ưe=w*0aȘt):E\e.eAen`6#3<1R\>m\6aewһpwb]*Kf]n?7Hz">&YH5E9ݴ $)1!gHm2ɯ2$ AH酷έC' iPW6y{Wߴ2xϬb[f}7/P㜪MM/Dϊ{f`+x13M^h2SQ9R6 <{1{*}AgXg(msBxjedrɋDdLwmH_iN;|e 8co3s_n78Wnɒe夃8VS"YcYָ`Sll\J8s@$)FZ ҊT!92g8Nc,EL+Mx*##~^Ccbْ=hV'{\s* ?AEadVռ| *3FOU5]/GvkVcEc\q/r v] Ob ZFCRA=|#׌(w/Z+D1^0փ" K3婄2lԒ0Rʺy4RR?Vr}z5/{+A ;SHarF% *qE* &rMCR0/vZg _Pɢ[+,Mv}GiڀvE)=.Ƚ?7qԠںjGdzڠ6Su@0EAUcg.{\şPVϪVLW;A盎:j/j#<ܪd 8z|[᯶+J)B3mP N?&0y>nAP쬚D?n ɝhC[n5$cCe[dNM.p[|pɿZ>jmUFO:?~`ʾ(mf̀LOTmQ p: Wqͧ*f_VMY"io ZuiR։VI[Nr>%nn.pXm;\7W.A_ ' .1n78%!\]쒔J(pQ:y̸ >.G (=` r &҈zSd|S]4mCݎИg7eWi!DEF]զuQgzN&IeeX<.Y oj v#0Uv3nYi?}B_}1/r}GWA)=! >ASRxXd$Ԡ\fVF`ǰiy")MFq1M]H'DBB6SoŽ{˽kXdYАH9wj!æQ+ˆ-{HreHKHE7nP;qJ2s:G2'lXǬ.r9w{7Sψuu&_q}|jQК rԼbԼq/{?mB }!ɈT>hAϰGnhHJ䥕ѤI.Guix&wFjkߴ?(]ԄJS | fhV!59Ş3.E͙OůY.ppi?U&\$/iV /.T+#f@qFTo0G5Z&jP&$4%aw*cJi5sQ_PTs-$%^WY+z~N~lO[/뾻;;ߏ1Mͩ۬-3OJÔu^W} &GP_PkMIʤb?LM=i"8Tɤ&RUMp {wuAD/,Dӡd 0eDܧ )ڄ P2JR2 EEr=X-HO}QA[k+ǽ%Ɛ6bjQʙDiҚL0KKFbRHD-Ǡ2X2D@#wpJx0UTI`4OWzh5-r}XܫnB= 2 dm4_kϋ hK^[V'i7mKwUyT= V\[rU-㋎fQD Bmu*RmbF@MӛJqlMYd~܍.\P 6{D|>'b1y*F-P-[KhcT21C3b NP"RHEE,p )X3bx(ߩ* FIw"]ϨaidBk3U¼L :%Qd#WDzJRQAh`9BC" |\th> BD'Ӛ_XPsƷxSTKvh67euG(3~jJ\!n,_uyrs;|==ecs2[vO_JnCu^M"05 {HIo&IeB2XH2B#0hg:NXQdTFs$x+cB!DH8Tb0) cPTqK6!HFb܏tbXX3BI c! XxP,\R6ʚqOas xgGov4\.f1[AwIqhfhHhaB0\gM+{΅ lDnx8 س,&W+A@ T1MȹdVLcAbq$-ɀڃ5; 1)N:Ի42;h:vj"11p+QA̐! 5E0IwFR2jT !G;"A+22*Qu!@<Q[[s?6F"qcAb/"ˆHDqD2"k!",\' `oTi hED0Q NQД-X8Cc2 `&4  #b1rv4>qqVwuLgŸd_\q 8V{OqTh-ځInyB%ʥw,,hDȀţ (ǂ|?<<-l7{佧/l('g~?Gtm]]Rd FqDpqSJs|p2 ;Qw8 7Hg)APg, mO41N6jՁ8$&S$e"VdnH;Ӌ nO0p އ^6;+~xOgkMrJOKD9P,b@faTxZ<`T&gr *Kmš .f3)fS1(Tɓw(ȹ_Գ>9N&ԗ"d\sn0|&R z1ef8( ؑkjA869qssuBH8:;<"F!Q6¡ח4<8YEPQ9 o@UD2@$@(n>lJǤt鷾х^~x9_nA_h,JM_ =Ǘ&`QPIhMРAKR ; 04LRy4;6>Hb2× HkU*927W( }P-Wص1ÀBК3ə9E[,xΒzVs?0]p܍p/|}fE58zRHeۻO>lCxꅧ3R2:hf(*TQrUB@qc`Hdt~酞y>7*I$Os!㚥$HLsO"nuUZitF)g.?*o]j H++ ª<#g $1s8=`|b봞m~Ǡ'}Fe /q{~?a]s;'Uq6v>^ue9H8i*mk{Mya Wkw%' i=b0r{1Wm7 %DMNBRl}IB4{v]ݦcmN}AyoQ}ۓP#1Y?L?5ƹbCzӷ7|ݳ7OmGokJ~O./?e{|gGrʙ;]G#~3 !W_=cʆ'{Ƴ77/KAEvas`U|ZV;ȖB<6$gzRȬޔ{ᐺi_&G]x_'?#KJ_͵˷puK|YL [( ӕƳ{\os.`kQ&kx`t#H Z/;6a7z*ЬeɆ~a߽SeD8q=fns덆˴EtZdB>Fh@P?2D0FhI"#bPK_w#`$Mr RAiB- Ŕld)Ɉ4Q4 W2d*/KY#H C)Bov=iq/ft?ꉨK1 I­ٙ^] oG+%HFC .< Yb/W)Ҳ~CIQeOXg]UUMu}4qK!!# `7X RZ;Ҷa`u+@|Y1 9f2 #A@Y/5omclyonM/5vsKA$I尥ȅ"<_"u,M!=mJ&iL`ә7ML"F`@(4hx.%+8A569؊;)ʶ]^?")9IE$s2JaQiƝVc n' L<p傂{I"0ǘz ʺ5zUt1?r:ǜ2 \^-"[ ~0BF]O۾kWmISrR۠JaD") $ XX!.%L cH*.?B`t:GBRCo$r4e$168cO) 3U5 [0{l`_摆Sh &)3;t,IOlOA lp3B $6b?))ݔy=,+e% xOVţ]dEwyqHqb@ w(BvoO1i|v E7! ؜K eȹ&.he qQ|v~h~)#Tgm64!Y{%:??gI,*n;i{+C DF ׉mty7[ivw6=2Lʫ샗l0 a2g>v۹-Ջ] R1I '*j{b|}OnH}7vX^"6 fpl߽&Xuƛ'맑ͽVn`bE JGSX> F.5=)~,E)ӘF^رwRz1c)\zgz[)ceܖI[QZ}+`+}u&YW9%tPLO/g3%9y˟_/ͫ~}|//_x.`:GחlߵMu U]SŶZ&G=mM_V,u=_b7 P28GjW&>ObM_hL\҄G8iu"JE=V1Fa"1O:t2.y=1p GZATŒGրA1S; ( D F}{J4jlMhtsFvSw/kSwVRw6uiRw$cT7406"g4`+$T}K&Yiۊtɚuuiyvpu50l$1xH qȔVh&*b# #ʼ5m3y`GU݉Q*vn,n:I >Ý$( /u(|FaXSy|i &X&ɡXgk!Hf5FO{A9s(r: jC @@3@A>tlg~w=I(>JZ8'ViasgJf]wŊHPrl֭]86Rix`,_1sɭ`)$q}__`_["*1j DR8A&F16 1f]1F: lp6HF7g)z\"G XevABܭ%_S>q[>ͽ^AMqozp#e6R9#|\aaP'0?zM(rIIQN)93ZF1O~dH-`%6GC"錯Iܰf<{v!ͿuQj|EY˳tA}nET;7ip\=6ېuiugc/7#P;Ͳ߆I. |w54RwM_?peydldluc{ ;w:,rQEZ oMhkio-z"ŕ#ޤ Ϧ?D`zJU0O&r̓tDmM= J>UB=d<7es8hɵ" <,mAraVxt"$qބf)e&#QHYu2Lw$L"^ˈiDk45dD4elL-s@+kr44vd^%S=Beo]c;}7^d'%ڌ` Ssu٤#Wst v2͸LNW#%8[XsÄYߔb xb.Nr>J㨓[@vPl6-({~Xb~1&XSE'@*heM׵P>m/}~`2?k3Nًf :RM$ȝ[4 Y}z[y0[6KS E+ e|h0;߁]L^f"銟8ݝOnz;ټNj.7izﵼ ;7`lֳuŚ]5ą~QQ/zCҬ_ݺm%z mn6of!cEm4R#&'^%|à Z<(rf8Z< -x0 [cic9m,_ CF{P }RFnIZH$ԃ2:$!5'5r(joFI*N/Tj4vT @)ı|}N&x~5&NL44R'F4FdIl%V A,$-T0 H$L4|cFlh/++Y>N` Ce'7)6mQdՇ6+8{`w +! 3~2R'ߞuI$;> ˪3,An6i(g6uϦ~k:X2qzkz0a6kO7Ekg_9ϋ@xAVX:xlA-Y_@\S|Q'L/eյ yl2Dop3E'uϊwlnY\x7{ IF$_@>8bW#ctR3O:SCn&=FoD9]? 9DtYFϦ4 달eJ"Ude'.6uIBI<"|-cZp " FLTtS굂3B>4c6Z.vNQMs-%zJB#Ӡ*Ō]褰~}$G~osnYq=&-liJSr+ Rp95#h`ٯT`8{(UnL):^|ס$'MDgmEǺY,}WZJr8$PhҲ.$Vb V>FG4d}sL&_hc-l2D# (QG`X#l=_4J^)Fjճ6?>4qly( VX`'55"O|-l +/5P*e+ bϲ۳{7FI)bv fr%R90:ZYK0X`"k"X&HS*X5u[M/#_A`,[ ж.L.҅KAyژn, | X0pB?۔t>[yNGKEYrlS۾}0f?y>yW^~|>pr6V2wp'pJ]koG+}͎bdq :&|~P$CJEV )҈#iVAÞ~Tct*r-)E"e)#p Ğ)*$+du% BVW Y]!Leq0SY .du2*du\*duKQBVW( Y]&6q,;1=TnW^WČOSSvVZ \!+Dp"B#mLi \!+Dp"BW \!+Dp"BW \!+Dp"BW"BW*Dp%"r"BWL}L*9ׅP"BW 0}Ҝ"R "BW \!+DpHU-1+{5eUL n^\t7RItG>NL%TeMoq&-еH3okR\ZAA!+dUx z^zn1jz4O+z'Q9$QJ|(& OK^kZc$Nj|/G,ǏwC^ l({,ҢPV?-?jf7SxK׼c|TI/?  FHP)|\qjeR)bG80*0>cӗss U9j0[&7G&Hp hŜLx*Cҳhuj#xe=F=ylJǔUw;S]xd{fwyt>Z>JwGB?Bxi0G81V+eP`wq"ĬA$Z InA$˛a#H@ 0BL+Oa]VkUKRBA ,԰&b"4!"}bDehpPNJ7bi1bQ N֝Q-mA~s_>-,y~]j'G(^&=\ש|? <(ƨWpIb0<`QNE!0a#RFYo=:;a8$ch[eFv{\v;^I6_t_{JkglHcFqfSq)QV3#bL >'JJ,0mw!ɶN;rEcK{k|uJyo&>oȎ,bzm$S)EbAki|3sS=l{ڒ!!hy{Azԃ;؁~hV u$Ke_S"avl] i]ayoz'}jcGiOht%cmWooyn{ږFxZIz5Koi"-}\$l!6)k6D-ЪL qRF`y_ fZk]龨:sYfn2aATzzW<.E%Y~1V%Ŋ^KH]ty7o3xO?xDAt֛O͹%{ݟj2:-#X9 [8o?#] l>~{]H{&h5-.-3~v-&hъZ4M$G{l> ^4htYsS'r9Jg8N7s˝31h>ô@[)YT3VKE&#JIs+{,)"AycD8(TI] > G^QX?`zWא2@%fc, @z6z7`3~v6ռ0_i|*9 TGM*rDf[=_e'>skuMfgڭ|ߌyYگvV$ db9zUf]-yMGĦʢ;;( ӄ1t o^ IQ-jm|Y^+K,+cew*X3tmu$XLjAʌʇ߾qKԌ"_kY"M{-2g,>-)\>C:ۿIn~[vmL~L?Lzd3ܕ]H_O ebA1w}l{쎞n[IO\Eg֎ϴvx>|wܿv1ǵ,!:mH›ą Dȵb4Zi~s`l}[L&.YCϫMROh֒z!!4K1uEj))ڙS7|hko2@LJͷըGg\5{[<`qkJ) [8mZ3O ofh͚ffoXٹ>yWka;Q%7!MdbJ>!TlB|!fu<((׊ Cjf i"f0DA; ӌfC5cI_3zss:%U+8),!/^U!T-YB[\t&õBV3γ'Xu)dmB7/q,Ꙛ]f)! ѵúV"bT \QdBsH9gUDQ9~?hjLBxc$*"ktҁǰ PA2F1-Kp'C0!B*Caj3jX$D'1hFs+%okgB1Uvlzϣ$6J]Ɗ+0)M.ݾEC^GMqQ̪!!]QM9N!,($SE,VJ\fE֝KMSSaX.kg/|t%]{daG&oZj].S7=rT^p&$]"áQc)&\(Hqd4oB}%wT(#'8 >j V*Ffd j0DȘ;ۑ1 IƎXB]7+ƒ˫d%nlNZ|Zj3O?G˯#RF<^xsZ` "QN#ȁS10[wa/9Q0 "f]gFD\ Qy .Y" k$!@ ( W$R\ "9(YXgJ4o8Ą3[p&`Р\*LFbsfD֝5)Xj X=S3슋43.,0l3RZ"WA=&[!Q9KG]'lݝoonw|m.߷t{6J?~K])N9-A["NlkNR+"RƋkN6`jج4{Ƒe /3 Rcqu3`1POD)sQ2[,'tId7Y[UlKć% !XT͑bh] 4LQLՔjY =287[FcHZ]j} fN1gb,4<`V.lJš@|8y-Mm U۝}5}lKYg_1$r47]u\l1Vd̵.B5QH~wg_d9Zzxz\xI'jD -R5+:[ʼIM2Z|q>YZ|߼x&wo%,V];#V P_.iqh~;KS2gl0FH(>7bl$POgݑKjp%z[2vsKG}ͨfmf]Y^Kw>x^#rgmN&gu׷d_}}k}q.mPO5g_mę}'GM_Ju0WZ_~:|;e%~s —\GiGFӣyl_ɻ`tAS9d:<ʓ70/{˧~@*:xg/?E, ټuY_$uwࣦOo顚F{˦5[4Ő^ muCvﯶ땺h7͋0v@onWSVK;H'$yU?9O퀼kau :)f[Zs$Ikv<K2CL:vzەLۂO.&i>mqƄU++jVm\7[ͥv尥Jk@YӖd _D^bNh bAZ*ٚXK-+L Sx+~v:slݘO6~kks5ֈp-T T#-yfKJ#\{֔3mFFf7Q3D u-.PGxzy۔r恚ܯ)[Y:;޲CR%w*?>-;vA;E<~>y4kŷMܠq|;IK>ˌݤk좓 ɸQ9rj-?u!Ί?OG=SG,xsVx1 Ϙj#}l[/s{0\q>ȫafyǣ/]yݤ8m\XKEo6 \st(})uӛۛnICbƬm{898جp;:|3U@஭n>9E0Ǔy-Z&&' J}ɺK&xڔ짤?wBBk;T0[9Ѽ7|R},Ɔڗ|\SZ>1+* 8 ){aÇ8pis&c|cFR-%f}dWYfzDM 5*CrZYjSheQIӂˈk'Q.˨ .uG$,CH ׌~؅]v^Jכpjx2-&:.L%lxBZ^cKVFa6(ad7H=ڧ|P)&#jQBC(cZ"8g1rъ3׵HӼ(.m JV̓rx5|"㆔HP"j-xJ:1U*tߤnEƖb1|RWZ -TEGGS^ۣKYKRf›3@Rzl&V!ZIRoK7P1166c6]Uk%i22Sh9(/{wPfQ'uj, CRBTG܆m^6G^4IhPQɔ<ߦ7QVeu,o94w賳Y+5/2ࠅp"0n̖N(g'1])f՞9#W˄'xɐu U\.gcu.o ٪f*1ڼ~NTeM98dJLAF4?tMaV`v=I'q15Q|$1F&\>%d)`1jx;E+t :RH>E*DGmU7_*KI!nxj#^&\*>a <[*;;%`jvN~AV T HDj@yU+AN_ z#0Es^8@~Ha ~ -t*h4XZiɐ̨@lP`A*\42 ),2B}2ʦ_9eri]cY`+IzXDk(@B]A[ pI E5a q2M!,'PR!vl6 ?50 lGPp d_!L680)(Й5d,Q kUr4*~ΎJ-ܛWӀb8!fL/<͠#kRP6Z'Rᠠ&zmǮ\B(^PZCAQ@ISӗ5 n\K`* UdZ"4 -J!ZHd{K=b`\ĕU>%(M2 7[HEΘ5j`>f.*P `NR'_:rН Jm]xcPY8+I< ]4-f003)

[~/[ Jw+ a:H'`>oiҒ)LĐ-XmFxڹ[gWrcgq}Zq'M`&mHp"=j\Oஹwn~SA33seMF)AhS* ZB .fa m{5bap˂G-k1Aұ.DS7L<C cigjiTꬋʷk,"" >7MJC "C }O bp`- ktp^;Yj^NKDD4xoQڵQXjm*r/+<Nuf^F50ZǴ/:5!8=BR,BzWy90{ UT &y8&V0ۧqy2L fţ3tٵL@B&2 L d!@B&2 L d!@B&2 L d!@B&2 L d!@B&2 L d!@B&2 Lg 1>& P0L;H) @ `B&2 L d!@B&2 L d!@B&2 L d!@B&2 L d!@B&2 L d!@B&2 L d=_&A~Jgqz2L{)%2%`) L d!@B&2 L d!@B&2 L d!@B&2 L d!@B&2 L d!@B&2 L d!@zL Z=%&\=&Zw&Hk3@Jk @PQ L d!@B&2 L d!@B&2 L d!@B&2 L d!@B&2 L d!@B&2 L d!@z>L[덼!5;^<*[MKiuslC Rl}Uw'hè]2ΧaR5nS]g3r(TNݘ,+WEEՋU*( 4I f|> %]9D= Vfz~AӴYޯ3Z 33o]9F?(ofT%Mm&5LΪ_X 5Ji>Fe!։:4Fd5zŵ(?nl4xX-81 -{~5Ȇs .`1$jOa8VP˝.2=U\;Ǯ %De(_ s<>:{ Z WoWo)W"B Pt@Z<s1-H_/n?ޅNS:>r;5RsvEguԂp> n^^-]@vLn+??rdm`3RY93 P\6ӗ%9@YI yNmR"Zܺ6"ѥnK ,}>h4vky7`D烸댩 ɸڗ0Y̻.SDZ7EWlto. [o^!9VBqˆ;MO\e/Kݹ}3^v84˳RDeuՉ磛B-4Fi2]/M^v(]\9Qԓxqy!{n/ux݀Ao6 Rگot+?2"22bܥjoicQLOQ;azV:TkkzCYoi@ 5@ھ`ئ,}v5(w`4hNhft>7'i%90>ާ- ~$)Wf bYJ.^r?-֡чb:+H^֠66kqyxKqS w1o7 ÏYD{ p]li=;Қe~OӋ{SJ$975$b-q%ch0ѴN8 s~Ю>wop">⪧dc98󀒂ɒYe$ư`pN /X*(o1I 5Gsp#6NR_ $5 Q-sȅGC 5IJ$LӜf!Rz]]׫\|!GToh0>|\>oQk)L:e]iΛa3 ,MNtXv0vѬ|z~~*RJ"O)e GSBAutAbժb}?6K7鸚nZXT>hW#[ķv-;1eNܴ[);ZYm|ro!d.`.E+^G3\+P,yz-P|z.`ඳBvyCDm])@ם[%&aTx)As]g{풞7swhp7pů@ 4١u*!WpY;6b_BnNt+aK hB.5574gp෿ͶA;뙶{77>`C+wnׇxmhM&|_2Y74<ۑyչ+.iTsi)r kp7QU,S՞ dU2u@Q]g[3O.:\ȶm)u& 4ecєs$"ҳRCnG0Es,̆^Sg}1u֛8=hf'p<(C ϙx _)*RhOJX(2SgLgrjwliզkN:}:i`pʤPաzyqR2&ozTz4MUcF~S5gɴzUAg^|UYq S:hy^ّᆱVZ|+mvv_"˗gS ռ x8W&WW;^3aw;75qM9*4'%S !dN^p9' A#@m  VY9ܲ X ȑ {9qDt)bɯy:LHdBȄgH}abĨRT"\>|5ٴ([S;%~Ñ<@~r"XU3cx-w$^Ah+,'QK\rˉֹQ$9">%Ah4l5xFI2p))\Kd)n( ~Yr$k{ Vz^]M3TW.mx݀ftز?} zPu+|n|%+Jnsl6}1+n镺3rJzGROm_>I2%kLsa)Ipa iH߅4/؃P(KHZ]`g,J)B̔1p$Ě2K#IqvH7qv(f=35́J.sW/o_>鸭s\5aH>|г 22 d:-5{]AUP+/#[ew/RǦxjb?}~E*F-3D4~ھq{.:'283(#nOQ T68䉫kL7.g9-0pɹxB: UO7]詀4"2FC]6q[yILH!a`9"9 C\Y@Vv$!HPGM1Gm,!D)o2K t,'l s[Wjs7QNtmH&!5ev:Idˍ$ O(u"oL<ǻaG;~<nCglvE/5HT8gRڎr9ẕ*~:^|hDЙױLb-dz2J&֢MN \E&TyT ֱ W2蠿K 76#;yjbﲍ>JQm*Լ:rόr1ym b~K*`cJ뒴4-t$&bW1i1Mu}HXyoGz 䁬=ˈqzaZR3ocYʞeXS%m>7^M=w̜[6fݹ,e}?z/Y}W5wabq)޾&%k(ha>U!)hȎ"%#l RV4eT¡h#" &2+LA&H#3(YDs§ƌ ݌J}VK~cp7\&ˍ\6dv=򫊗o*wXrjP|rIvBGc}Rg;3*6zWr.1@Ts].YѢKQ`U0:R%c3tvKvJ3Y،2F48gwZ_]W(nM{yf_ FpAˬ\Y٘TB -LgDi%sRJ۬Mv^PP1J%T%Ǐ*HіDE$[װkn|<݌:Dj1JQ}MT/Z@4.N%iRzJ {^жg1օ2kNflEɲH1Eb\a3tvިZee>=،>X"(GDޝ ,6NR1*Bz;m0#=$уih~8PiY91IX9QSMک@ֺ6[+t4>#~Ց[31:QɡrQ5r\|f%6O=Ex*Ld1ZK), "H1琋G~ъ:EPDض@]+:ȝWq ~e3 v#n~|GFcǐX:v 7]]_ :;7[T%j rfDfRSwDZwyP3g l n.6C?,m!5FQ۬lN±2ĢR,@+h7?oJǤt4zo=mA_|WC_Msp$ F@ MCvY2`$fɆF8_%³HJ[A&" BcB31WkdWS5AAﭕ^ !j~¶ItQAkGERQ{9% 3|E`[,'%Kl,GǞ(t/[>KnV-N+RUvTnƈv xǁ67.xD)Ż)'-X]dIY #e (HY\^Q Q5ɣX#*Q)uTNR* TdQkƺQ4&#6mH dV`c_(lE1r}3?&G=U-xv_P]H}Rڲ_d#} 薋޺ 1HJb9xkgJQ[JLy`/Q0x?${[m~Ϥ'_ikxϿ~bXCb<4q}%/>m~;̉eAj x?~3b6MͣG¿~cAk*/}PH{a"߃A}3Խ)ba/؏`>q?lc,Ub jqiJr_G Ȼ `;uu^+IܙPn߾t{Dca ?ïW>Y[~BZ嬀 P\]ny}C0**3Iȹ.կlG~ҿ>chlʳܩb9D#05z7t6_yw=L2>3Jr#LZM;RxLޏy \_շ,|~baŴrHs~C8e:T@0mwqlE49]|:<ZֿR$jj]iV_Y|{C|縻xCoP8LҠPEcœR&CRRxAHk"*r-X}5饭^x3ә Ҙ7ByRyps1" rY7c5!d=K]{J9ϗAl~3?s\ĆIX*Qp:{w@AJ(y#(!,eib9"/,UiRr$ET .:yJc-zR@)$ZQhZiY͔iWny|^݅UAU^e>_3E:n[@{FhE&G_+6Ubd%B,SqRFl یs.|hH(}mm$Кl)MTԶ\;EL^vNȢ%CΡP,kዧZAtY3t䳞 Ƈ`oK{}@f)+&T,RlKWH~ɐ4JY% ||+{p@,m@J|yA`!D.IC,ExuY"Ӹ3jh'vҖyx#(y{dCFb!I6ޅJ\! IVX4!Qp4Gu׌&:ڻ='^ޭU5H2Nj'9U0gGA bVՅwb䶆K/zy,׌De>ݘ6 {55g=t)ldLQD@To1T'o٣{lzn6C4=:AucMa6tmlKPzEl`M(|`M*iݏP2 ;Aw0u,ޠ+2k1~r7yeP@:BPrQFuE66yJHPr!Q#Vucl-lLxiOFӊK:Gefod1;dQ]gM4=&n^7/rD"@P%f:VsZ@dcE(1A4QzGUw㭘'E|K%vmåmR{JHgI:gTIytrylx&g1G O}<@K'(&SWXcc`ֽ2ӵ3/-ؠRB[XFcQfRsų^8<} U'0c)DI^֢EB*!^FTsYedtqT36m( <,hTCE`ƵRtVCq4f\3PEf4Yĭŋ4YI"'aU=t5Y+0BĒdԲ9uPen=t5l>CZ`U>zJ҇b1eQ>K'<)#$PB\SgO'a8|b!x~B36?XI%adIbpGWkϼ {7?R/_vuJVC}\^WWͧ5^ûw/uR;p[#j˖(N{f<,vI⌷ni.~r֣'ս{7?VwϷGi:/jmެ- UO> hۭ=:3iLCi6ʌie]ȋf7sNϧWO=^F)uC=+fq˚I+1uq*%˯2wp*1oXOV`z?k#G`\Y| }.,{`aQ4 {=шc$oQ߷ڪV]s]+[tD˧^6ږ<7uR^^ c݀@̃,)TlЦg "O:MuFlN|!|n:7:lU#Wl|)/6~ڽ?+O WTCa dlM7yؕXq~7,d:yѷ6[xꎲȖ/mѶVFీlP:L$dxx?Rb:/7yB9_f4ΪzyOo)|1K,.ze !P&^hr@\tvZMi>3`4j -.V(5ۉ<ls~ Y3׀d^MF`X>B.'A_r%aGY< 6P!!>Y2JΚN.Jh[BR)NE1&:,cj9M:Y'Tis7Fd ]cBnwXvn)U =y>EƜtnLi b>t(|nTpgaNYV<}-;u^/VE>9>PO-Wxb~ Dަǐ6&HBcR 7rFHKʵM^(YuJaiS43JK[@z+Pp$0r3ih3qt0:!u'F?/VN~b" xcT-mqm|?I"[5>g/m/L~u+2i+J_5>죱ms U)WGCW WBg^,ֈ0; nx4צC 2L A?A?gyݼ?Ŵ4hG;7@b19X(BJ'%\%z sV-Eb2@oJc:1;%n Y?kkVJ[ԓJT$ F@ 0zЄ5K$*s?*,EP&g">ih:Vо4bR OE1A$Lb>%hAJ*aHS~ic' Yw,|f0_2O%x6ffjs3b#J)2R0NZ0TQ{x2ʒ%&e!'e=r=;'9)6gqrz'3z&iaT+Dт )Eu'[k (j 'd'7WW'&5g MHVJIYGbV߆Bl&2@a Ŵꛭ&Ѣuh1f-CCw u/PԚŻCFx x}­ENp5P׋I_x.xHJbⷂFǂ6;o%F-FJ,X_k*%ksenyT^{y=Sd]k-9{}Z/~=9(Go*5~_VIy1`~+ѿ{;ǟo4HծR& l%ϋ|EwlnӺ mem `I;4?rGA@.'r tqr_|MNCZ~m3zPoN+T^mՏȯU|k^.{em}ycCRq9֐Ob|u#y!*qA8=9}PjD|qnQ6oEfgQ{ +0&#뵫p{wV<[ۼY=ݻэe,PY5ɻ {Dkw5V6߳'='GKm3]߸g>O<0@|c]Xr5Mp7QI,9i^w7q֠qRW!Z&ɮ6'm:[MtLI}"C]PAjHfJA^ WHQ;z924:jpϸ</P&iP^ DJ $KF 64~RzM^FUͤةW3ho< qu՘1S[`c(j({bB :`^ tSMHS,^/4cyiX(p ywckJ!0K_Ȃ 3hTF ^bV<5%t[t ]צc,TBBiEɐ RBHm29Z^*XRR EHEo DDHޅH6*(2E!X25f3b2x4 l2ŢR12EU&:W,  8rSl1k;-},fw6Fg/~ьBP:d"Ed0:J+ymrJd V|_!vLxq#͔&؊.!/9Co<30'ȼx{;;{s4[i]6{oZ i4:]jYE}WYc2:)C#!6trkLr͊,_}}'kYGb@r0Z6u@*yRIJv.Oզ*d2[-H*&(M*iFV ETW $ґ_g-M Lj3q^Tw' Kƴl,&fs/T:$=ʕiMMط{1/ϧ@L?DMnTsJ"`1$TLX9@4yfTm>U\bmsQ YPbaGZe VCa%Z36n4ӅfƎЈƺOp .OUd<j؛ Y-l9<0->̎_.U Zf͊(EI*o|a3D%'ͺm}hJBB(jSYVg'}fԞ%QacL5v\9lvDk{OX,Z@4qk S" QJT0P÷t(FBO43dFY$Ed8BdT >l&'U1E#6jDha҈FX`NT0doq!%UU%ytX!`xpjD̍h ?,UL%'0` tn5b3q4>$~Ց<טlfɮzQ5֋rҋ^>o)p.z,Vg|YLVR0EH!"^܇^ME1}vӇ{Paw$\%;1ϸ#]VnNp#y?jWz?tqjf: 7>ρ.[cdQ%j C"B4IBb*c*'w'4B{rޝ*t{Pq阱s"!W15ߎod `B_0J[=^3|`5y}@J$oR-\vEa U&IV^T%g(Ĺ{oY_:>Y,NPkڿL(7s|$fi-(ߜYe/U1t+ۥh+@4e/([R ^,O"{)!hTш] Ԡ-fy6G`$40e0JۦrNkm0V-R ^ K$ilٻ޶%W,~?v3\ G-DG8DIEIi[ILRfWbu/NiN!bklm>USTPT VFj\P^R$:/&RIN9 HgQa gq{4`~{|гD21PIT#ZfDVI*Bp9'n5LYBq"5"oք%P2$ &!ZUo[޲&ΖzVo@j~5OKQ@2Ⓟs;nЗv:N%AT-sW+g!Hj1 ĺ蜧Nƍ Qa`2Ȗ3̾o݃t𓾒L\3\"(-`@NH]N4)A@(t G+ch [Vx3jLKbs D1<5K!_I DSmdm-FDP̓zo_6EvU>ѲeǶP%}dѫDJDΨ\Jr=AkN5O>-hO=`ld6Cvٲv1:5n6.)g_?G$rIpH6ydk=0 #D@46hh\Z5K>Wҽl,EJ^3e<҉* 4EG5u"k30}nskz 7܇@jMp@A\^0 *z,WT%rif4$JA)nMcKs;w:qxa=q LkS0-NP~Bk1רvJ5@ 6j^8eDhLuy5GKM  Z<[/eQh01dLYU[M0A[6ڔf׉XF-Wq)XNB3$xN"4]EuT` ,\h$\XŔQDBH# DJPo0)By3a# {R1G ?t44;+&+G.`\ 3m -IC~ EYFWdxcjT1+~+ZqVޗЏUQpi/15 aPN9Њc؍Fݛ`ER&B;lR WN86/,8O:.B,\ч*Bun+W(ȾWYӓqdPɄΔ;JK8luvY֤Aֆake.9<_@j}Vܺ^?03/'فWl0P6>u۹-] T>WL.]>]4UcOB kFn'TAep(|=ZunN7##{oצ{EL:ʻI֑ܱ`rӓگTH =`tU.X;anqf9?+~|zُB.zSb9%+tR'q?Jc4*"ìUUZ_s g?OOsuJro~?߼<󷷯_2s^y#|.c`I!,@p__}׊5Ww͍آkMڼـoӯjnE+bC^cB}I7z+ftM|$M}~yne3W0VܻFgt5}d&i'+0IU&E=fV8(&Ľ\0VmEygmtP {`/:63+Ql֬ySFz 9>lW-x1hϗZeibR brfy$s`Qx{,, YMxf>D'FRV2e*;:vysNIk&(# ԣi hdi OmcٌB,>E;Z/v^SVʄkϓ#ϽK{(eqZO}q1)T1ѡ*2jjX^u]JVh6CEat5%KV0_Jp@3|d?{~ JtLU!hb VD16 ɨ%NɯhIXD\[@o4^s- ~^TT^ʀjwk/iʏ4ss9k,KHsaF%AɛE* &-#oZy0G]y9?러oCۇ=U;]l mrT[3GT?94J붡bNBҔBYz0JHa$h2Oe, /fr)*G,W*(Mt)QJI!Q z1"|4X =hSƜc!DH9o|5q6y;6"uU} 3q# tq{rGnѧn\FGCG.88iOѲVXViH9pV!F~K10!=FKK;l[e-Fug|YLՂ5ŻiPuc.Z=n(pڗVrj ݇HI7V6CRcG%1^Wk\"Tq++߽izg߰/58q[{'FFkM OW-]>tvCm|~-mZZhoi=R{.>*?S)J KBe4PZ| :0#hCdHp?;18󈒂IYe$ư`qN I`@ (=kG8%3أ z༷T,NH]r0OeіcpBJn)b@ji ƨ)$.nZ*O ( 0IXNS耭.A,U4;,5D)!ŻN?o&0J-<: Qsh x2D.'yq} O|tru1)y~UeҥH" Ӫ2.o )VGl[C0#8+fÂ7 PŌ}r @iU-TxWA3JLgpPb2 | }^SKQ ߳{P`;oQcD6:{=@,i&[ݎt n:lmN6/ GM;ooj %L7sBpusx~0mpCŭ[487 D71]\qJ6&k˦qHkrm\+h[U)\vk7[}oP ){lP?_]N^<g5Is׎f ^Yln͍g+х-/]~> 5gח\6s vf} O۬ܝ^Sp>5ۯ/ݶbCJ-' !' v,ͬC-agI-yE]:QxK}"%@J H^zp}`k TT}S(w' /#$оH,y4,xQQDV2Z lt@4̫`]TN[HIio 4pzb, 2r5q`yIy@ȃ2g۠]z"ED(LpD(ˋ\bWT FE & t / E1yWyMa4*c\] E7+? gŨ pUGUt䧟~8я늣wGY> AniԪ(g65O~lj޴1u|z8`2kOEk'˜7FP_; yu}lDhɪ %5cM1R+sq7?~ώ~oQݪ(Fq5kt/hpb]YuqXzm^~Dnj+;|nqx8JA3Ƽe7\u?ғl1*[g*)zv,+O!Zka*L8J{C?|(p!"wDkj!t61br\ qd78Dku>9Ѣ,P>L*8;2%%Ghbl&YIT &S džq$EݰɅL-e96Qy7M"os/9 h"" êpI O ? 1/qf,gjdtK,:rM1WJ5FI5٨V3&VvyֻaoO/υ6i!ͷΑ٪Tc?G ?l Ssݮ9 **{wdغM&Jrz|nt}Lꑼ2 MHT8ZnH)c_m/7qޖsq!gK.i[jP9 nᢍ6){E6*-XRsdy5c:nT+ J3gc a8r5bmh :ۇ''!Tz͵Nlu%NnQΎN ~CnRn{`Ďi)r0kևU驉{񠕵Z2y2AZ/yh(-T^EAF| q 1-rqcb$w=deǙnY gJC;hMJ^Z[][Rk|:AO.׷ wi{yxow*i/߿yךrֻ e킘"$c&;kUzE/?6ZsUѻ3@wgڍY+Oc,?TyC,lAmܔbW/.~y4~TL}x~}c$ IJ{iS$%kz...) jCTҰZIJ\fX~Κ |0.`rKˋXG?M~ven_V|5-\/KKώս ʁ2shxfv?F[lVi3^ޥAJ4:ɒ߿i\LI˪N]c Nқfzw6:Gƴ6v5ujI]|v7k !>//J!dKOVC?8'pnvp~eV9HA/W궔ū>"oR-wh`* ɵwO̧2wSq`5?}Sez{sT_necKB%Yk8]5&k#9r:]wvE'KyT_vX}laT{d־^t=k4о[*/jqt\Kq)9ӓ.JE]BrW9&`eYMs4tؙzIfUkLNZo}|ʹ2VMݫRymGlb8ОuV P[:V,0ߎI]eF\ Ɯb&2ܜRkPTA6jkbDMԴ0j]^u9E"#MJ 6FKUrc@)]ꝹlS5Dl37OԔ1bOB,=ն.'Bu͌e8 m\\\L5ёBE{))w'pFk(3E2UFiCCmF#ZC =`FFEz~~moRiYRѦH{qX -ϰ1'v!g5}0'9"豫NTRYΩWRXmm ;#ף(]'p߬ 1iK1)$ZB[7p sm vMX^T6"d#)nI;mgŖbSsHFu)[_c ;7% }Z\*@L3. =5]j~;4F%.SWۄ L.,’- Objb Nka`:^vζiPTϪH Pt]vuL%,`,S.kjcoue+yJXl%d ą%duk]=^1l[(QP*lRXNi^uո@Ѧ@ʨkU (JU&A'Sʔ\vd \:Wlh&XXvaBm"&sOs ;G;[vT v^Do[@B]Ѳp>3|3JE`HʡN[0 QVagaGܬ ӄBl0 D7S/%H t:% ia `Ʈ pq#Xi! TLgf(F<_`*xufi\e-{@Sm0PSH\+^IB1YQ@IQVK#gx0[Qo $2%筆ke"2!5s eq:+aMhǻ =sJ <{#[K{5zĬSs|#bL0Qb h5!؄9%NWm2vjl{Y] lCN"n `9:pإi;ؽ3 Yob8AbC8*KtK|CP0=q  LEP{Pc^3m>%DmIJ XqP(.f)aiyV blj31b 3 FdGj[?-,j$fj"ٕA~Rܠ}8DDTp}y;hޯjc,״G'$jGFyƝMNÚY`liCm F u@QIpXu53ZS' ՓFlׄ `L x0ӫ[ 3樒:!,hogݔr%5v B`L^a Pwˈz#a]aX ?:~8Mc'7d G8z|W,F,(qCa b!-F8j#Pwus kUP?yڨlJ%cT-،znzgm͊ڸdXz=ش=ɗG5l׉ d, UXs`mN^OϬ~~3k7!q*Wf!wU<fD0ianZ302يЂ3_@2q/0"Q2'Yc9*.ͪ\,lU<. q ]%xǬ<̬ddMZNWKnHҰ5 6u4nW,4<,B.8 P uH\擓sc`̴kLuoqazFjС8ŨH`UˊeQ E?k;Q,< XZ>1 wa@C<Rnt&WP>&0 $L a H@&0 $L a H@&0 $L a H@&0 $L a H@&0 $L a H@&0 $L a eŽbqR?L 0 L H$L n&0 $L a H@&0 $L a H@&0 $L a H@&0 $L a H@&0 $L a H@&0 $L a eA=b 4 {&E]g )&Wr:(a H@&0 $L a H@&0 $L a H@&0 $L a H@&0 $L a H@&0 $L a H@&0f&w%mY4 iC cv`&$,G0SMMJxjZ bwTL %e%1@k>ĺ3(\@}&P g@}&P g@}&P g@}&P g@}&P g@}&P g@}&P g@}&P g@}&P g@}&P g@}&P g@}&ד 0􀭡g]P4*>,7]37@ɝMDEUHVNJƣ3EfSrʬ'O^*23Yn2R7b۲d;w܇:MI}QQHRU^[Ƣ#ٳR9M\yrA$g˧0/,2v|\@& S (mnmIlMDJ?SͫdKЬs8 mΌ9 R0&ıj2pl|W!soըȧݓxS/Hy\T{ɱ6$Qs-J9:AZEE ꃴ3HsaԈh<>e9 _.dv|3_eKaayIuqH砎i+ʀv:[8JB˴~)0q㌿q)}J#>:di]1_.pe/Fz*=fԄMwD Dk+S&,v|4vm/rl+bw*`fۘ @L8Q*g er tXZW"<u[~,C3| ~{mo φ{7K٭B855#vǼhĽo'5­3>ɯw+̟-dNO.YQ%) j71bM&^MW5?_Nς>39iӺ4jOn97Z:M՚ʚᨪ3ӮmxkYQYO70V^eJ_l~]%B94a٤\_7spy[AϵO.OS"At5u-ܙuvbE.{6!V򉿹L9ۺC4qNǠ#¸CGoOGF;}=2z 2"SjkYZ[ nlVTVgٻ4S?(;Z#  AAqP 9)f}{4~8mȪʕ\"BtbS|R'Yv0.e6.}6.>byX 8q5ۭ@b5l`=xc;q~mG#ac8t?|ʥuۂ;a#gBZ;GZ9#88%XR#̓ɻ\#̰(74S?OF61\n>}rV/W'3*:2=-EATD[=R85QZx")^X YKl6.Ovh댜u۔ZX0%. 5_ r@v"$qބw*캚]z:C({犁n4=~U(l*+oӬP4 C} ?Mi2 N+OTv6Kyn9I)$9%tJ,4M^COb8[aB|z;{WAsJ! '\i! +[kV/@_vA$d Z**"3b| 'O @WW,Ɨp@kT`שHvF])U@=G鰪|Ocz1܎Gu,Xll&yjy~T{^>ߡ9߬NJ @\Œkl-kǮ6v&r#iDw&5N~P ns1il}(:Sn{΢fp`7f!̡epn=/oI<ۡ祖a2m}V7x6LG<ηt܄ԡ)u2lcvm͝osE^Eѭ*Sc u8x;xoZ#es&=DKF[;ikpnvL6[֔/;]W֬78< yP"Pgy@[nN=f;ЇNϸ>;՞7>皭e:c8Hq¥ܒTBRdzFGq$"$DNMg]ԕt9p$ ##DPI{%V ,$-b ƣ#$6w4UV)/<|10t(d^\&ɻ,/^7eb4n&Mr߶?30p\וuﲥ_^_me߽N~>o *pn&*o@I~`n{eq]MMۓ[sV!tRlY~T3!t<邭* HI\ +(keˌUF>KW`脢2EXPC9z[V;ZS? bgì̸?fzN曛'A!05}+s_qQMY*_"K=`kdG:_*f`sKa !|2F'᩻AEkβӎհ=uF]+uNutsop^Sn1*۴U%9:{ <%9o# [fLvh?O_)Cݬz_ͻ>o:߈C Mw-X? Ib%ְˇR`.fPvlu;#g{Ț hrE@пvIF*z.! IXt& !`)$A B8@m)ٿQʽb<%ѯ\ƬuDEDf6P8@eVd01edlj<@޽_/kOQ87_uC[s;Ǒ L>CrڒOv'Rk2$k^ ʗ@L>rk&=>SE _X㚣TF_k0&f+sX:lX'Ü9Q [cI" Pmԋ(%?w .JpqmLt"Hj"M^~S N\;ωۮOٓdOl t `j23Ub*X'].a!`uwBRj\E?Ş Ϗ8q/V^z& *NS~lt݃-'g$F%@T8%Qq.-Ja1*0pKALBBX1J9Qu}Sg쇟~jklRή§:NWwX'o/)/h$'+hzU^ORД .0s78Ŝ9s˭#:F 2ʁe[!nKE.PHx3jH NV%VF?/XXB15gҝuY2c)RLr'|DʏzX2j| -̰bf2 ,hLrm:W҅>\/oS .NgX7CKD<ƫ˓en{n,cDrP}Z++ 2((MT>T` :Q)  Mf %)Ca+P5L YJP]Y߀_f/W+_u冼B1 q̮{ !c';M;n`*gX h{5h/n19ؑߩΎUa)?wIھځ_<\,C+5tA۬lN±2EgA`R+`7?IP=֗q:=NjseIg߯WSq,|v8ZFȂMe|M/MYaޏ B6L*c|}@'Aˏˊ@${DB耨fmc-^吒NNཱུB! TUh!.rQ5ՁA{2dN Ed[,'9;T9;U+&Nv(6{_I!U[lJ{?E|w`CM{gz=^zϿq'44Hu d%KMB` '# ;QQ&#)lF-&) e+dcP6`NX=2Bq6>I)jmMvrw'ī/}g]~01Zi_l>3;Q6zF#hO}0tGRo] e,4R FxDƞ$OrǠ'B74ս?s#.1'8ͼ&\lrAgknt5An{ltOdsXѹYP_ޮU5:ZM#$? $1dN5aڮ.h\}IEEȾ§ ƃGPJ)HCoAO(X|GJn ce3W:Lc0&ռ. OKqqd(^Nn'K8mҾ6 F(d豠Z4fx"uV&?ip@iF:< ({]8dPfE>*((&o`%te!TuÇ Zؘ`F: byᒎiTl&0B L`uEf옯yŪۃGE6J"i*=бZX0J6V`IF;*R*H4ݴEx):_ӫ9VU1UR5UvN8/5%>hgܱD'Jv{A13{V1!NyZs$#4/AGR 0FcdRgp`/iRO.D 9.5%AQ%x ($Sox!fBA ^Q ,YgC-P.hPA逕LP62+f?z~9Lf4ŕ#q?r&ymմFX՝t930BH2>FUldp㭆nϥG)y1yJ#҇1Ml Q}漴l MRH%\wVqs:7GK#: Žoo2VRI!$E$ظ$j nEn w]h}]j(i߮fo3p4ߺt3:H3Exq|yv~|5C9J;!_zh3bh'Żi㷣ϴh=m^cM/.O^W1gdZN'OnVA¬6&])#ݖ\;6G:w1tamN9̳rwcN&O#ͣ.oԾg 5jy拉tYĘ}~]՜Nf(;谱Yi7BCӽvHrg3y/AG:lQjJ]1ڠiB6pY(#Jcm2&^̮L׾ z..qVջgϝo0ϗ8fšyNP֒FLR),h2zP.:;o]ho$xn'~K]^{fۙ5BƶMqBiaO]+&KaXۋ"b6+у! ^+b{=iioj|4Ts&4j2BT(I8k vй`KWmi ewm%+?=}Ad1H0_z˒xMR:)ۤ]&ꪷ;P;l `GUꌪ3F2 >(ֈFD^|wI-`灣+s.o^IOŇH̯$1`]b9Αӣt$xͲ65 ^PQmd5p67C'_F>k̯һ*Tڃ1h1'JȕѬ(MR N{gݼ t؊ec!ujkKڰ7LK6=pbJ;#?gñw$!x'Ysj48ReYGZ]8đ85=V}]ҪAhE' 2bgeMcp3`o7PͭNѪ%t0v3 ppZ߯YvQJ .Cu.zziY<\}Zj/˩byz`5˳U}87V9r|kB4pt˃!}_:Yvei0?F%hiŋ2/q&ˡg,<(o8#GfOZXkC^! sFvlq==r^yA m=US; UN=,uJ\2M3 .88ZMr %wwiCAdUͺ]n]TFT\6&Z{rYf/e62Y?c̿Ѭ[qͺp[Y/ɭ;;'λl1$;z[C[)iqN[9$yӅVFV.I彝GٰI#Gmժzqj^߸龸+ͱ?j#7n$c(eҾ۴ ԻWm3axaA,8wj8X7'xW6V"Ch䬎m>^&q ?^Age/;ó |vߜ|x3|=\%??Zګ),+7uXG"RS%sޟwfx"D_8¿ 2I[Zx֗~N_^xNNH{hh''"^Z}2.|DkX8<:-F$ڄy_Ž2E:.XQj "Ro6>;KG󋓳8a'n5v}5yÛMJ:Melv ~y 6Ƹ8zgєܭ-!ogk]BBׁ`՚\j|0ɪS,2OBQż@ 8{?[&n:Ӽer|2 x7C,91ryRO3d(X޴&g&oUn:#ﺏI%cFUjbm{dn}w]vYchǵye.%ozE)Wǩ=.RP-(tRΒvlxsk&L:6*х:kUjPθrTkFjAUS<ʩX4iug=^kV&Zj j9jnblifT]^u9MIkѵ2[֑c _ݹR>tV䱡k5M)ՙ1'Zj3zɌax oxF\L5Cё26KNIP=ÄGiyRl$UCɨ|S!랺 V)YDPѦJ,хڑkj Rh FPɔ2e,Ac؎~B܂ѕU] 9uF!E.(;Z`T j{z=XHֹ^0EV \*\=B)O;`| 0s'[ U:}"Fa"7$i;OJ$HH !HH !HH !HH !HH !HH !HH !HH !HH !HH !HH !%<A sY 4#'`#JH ;#$@B $$@B $$@B $$@B $$@B $$@B $$@B $$@B $$@B $$@B $$@B $$@1 đ'?$̥!`H XLo I ~rdHH !HH !HH !HH !HH !HH !HH !HH !HH !HH !HH !hI ޒ _`W&37Kmg-!t- XH'@ B H ~" h?I/8{G}HH !HH !HH !HH !HH !HH !HH !HH !HH !HH !HO#<>>xXj:Z\^n/_/ VWGW qKg{8W8d~?t 3H"xH.z~U÷!)jdQ6Ú_uWUsvIRTඖe_xLNNsYgYn _oZL _ zd;r Lny?Ufd֟f2t9㏗I=0|]w )|(y:czTQHCiI~I^A \m3)pVL}fXvB.%.װ9Nјܧ)~bO:ճ*]\4*E//vtN-ën{v Nf !γdۃ]^5nk{?VM? xRSܞq񆍏_ū=gMwKqS֐t@؜Vh9q2ϧcGDaQ[L2Ѿ>[-Z߇UoKoƿ&\_zWe޹ŜIc6v@2%YCb090{H[ be{0jTu/tf(nfڍN`4O`鮖c3BMCVga%<ٷ7.DG e,ԋ~g^/1>O [>Q,g'6m\xޑ&Q Wro[_ZCa)I~;kλ9 _z=0HB&؃#ӽ/CAv(Ӏ{X) J &EzbxMf8 VbMQaތ Qi v(8h8ΐRV:܌p#\` :}jvmͺӻE':TJ篚<<Li,Y8LZqZvZ%* 9«ӭ՝ [w98瀭pF_rX׫*n_c:WZˑwoΚ*i-FbZqy:ojj?5>Z.o\w{h=hpu*K1Lj0v*c϶xkM5sk0-S <2i1V:.cgbrޒrq>p z~i2_} ,хy!Fp&D.LCISzS]k܃FR OTZ#D'G iviSyZޗ+u|_vzO^ϗfgwsMXbd4X /R,=ü)AQLaZ:o#?ŭع9Mnpy143kW rMlYMUzpMr~D#q|AV'g/~l"<%8}s.aex7=\N?[Ķ VPoY=R cbl^L+&TviexMŬ8\NO fN8AE!tj*e`o&p!!Ҧ3DP=AUup_W݉Ƈ٥7?LUu!̇a' зiϊ Ccg [诺8Av̚ 0Q`v ! Ɏt{wj e'l9m0!5$Y,sgklzPx1?k'r 8MA ~!暏v۔^#qz寞_ O'8{0(uZ8 ;B,~h?V}WTWZjP܋\OD=xc#z4g֛p ?O_ թþ;Lg_ӓNm2h flC{ϡVV28Lo4mao9k;M(=>1:E̚X[-_ù: l#p#z30(pk&z $Z) t6c R创ސnPv(5vvܫhp&`K|vX R,&,0)x{I !QhҊtdi;Ymr06*aq^ՂQbau@Ut"i„xҜ Vj%eN^sp9' 9x{#=A׾ua“ Hǃ:HgK[#ljd#$̞+ ёpYJ8tay'䭵,MG%#!vkdtr F^w 6?޽?x(WXt.n\7_9\oGb(V \9ui'sMaCy_q*gWMH\ 8LԖet%$kRzT/L=ڧ ӓwdvTj׋ZkY.)5)puu}) `*!{Eģ *ɭ5xYOY .\ף5i;ϓ8[{}x:5J -V#{!gϖI? a`ZM _m栚zbk_KɄVش yrMX!@x9Wi"T& d#$12 ~R~c2&fNzZ))*T Hm[cgelf [=m$-B{j aWOֳ4i VwUgY>j.0OKb;}$v4<h|-fyxJss!,Hʉ@mRYr@1c]M*-O+ ={;# [W9 HGmT(A^ %*: }hȓַx)}ТAgwmCۘz(j۳ ƒ8`%ԆG>4 |H܌[*"hR9'$Z5* k"Xrp9Y6A N6ff|'ga}i̊?*fN6(+!3ρ[POl>5#5 t!1aM*tl2Jk׻w;(pH]"ZSf߫dX!?zW}=b0=7Slk$iwbO5p-\&AG°4{RWP; &aimNpHq7n䣉e{2(sم`tWO[+| [mhk\IamH ٛvf L`cK˞oՒeGem+I'iȮ"*֋j{,qdD'40wf$9^\~ZxmGJqﻎGM=9wWhmF=Us Ws]?z6#vп~nx }^QPeaf ?1 N[ipl2{ϻmn S >1I]*EZ=/;6y7zb6g玩JY2Bilx4kXQ+}ɾ6nGioJzpQplE4ѱZ'"iw?v`ɚ2(9> 3rubo|sz+x  Σ mH$),,)%4)"H"mB9?lzlzS9mftu}q(zϽJCRQqB1PeڔQSUyT#eRFPͮQ7(?]y%!I2Ep(es4y_TxjC~%:M"G@Z$Wi lj,}4%E/vs1 !I"z0R~Bxwq^w7 JW+9NIĚOM({QH FNGm gs@y稽eiރ" {),Oq>5}/e-rAmB݃BJoO1ܙxӃ-䠁}&$])o}1oEWoE*8oq8|!+׬bCJ1u;/@  IdZ DR)QS&A 0jmfR;>!ͅ&U ZUJ:7I^DD  @$u BH%;,(E 3b2&гFc}fa9$auE^Rs0QLAGN~id2+e<\R.7! a[ @+éDAr`NCTV109%/%dV1?!V5bz/KVYA5'2}ֹ{:C<3^0־㵟6]kAP-NPV?`b%CUۼti% C &[*(i")9+_6F*MH3:Ř_2UL ddRdr ]ƈ0FCZtjߒe>{yqP*j~ >L=Aܵnyi\.ϘK]K^G)7_.^-]'.}̿su}9͹7zFs523uTGt|4fz}$ϘQ.Gn˧vDH%Fld\ 5cgIͯ!}##6|Ջ.σߙ>qX}{'-wrqIoSrcH  Hg0`Ra P*>IWaYW8ƀHezUI.mL lѬJL (C Qwӡ/@Wܯ0m[տO+{]) r_d0b5Zdv)*RWg-pU \Ur\]? %IGe(nRodL6GH^ĂɔR7))O᢫m0lIk0z!RNxɀ@`R;r/rƟu=`"K0/BV>!”LsԞ;⣂Y5Ye"#xc A < ,w2'mdgۛ.,_rv\xoH(*IBH1*R[IlJՂbrNie$\T։jbr9`?%U DmyݳFΎr_-oaZ4@ѓ0C [ ^xB`0"1d"E:hd]:B`]'C5WA߾6N^$od$DC!ba ALNz_dVoP+] zҠ'}&y!$|"mT|gGK*_$geu]x'i"´]_RSEEzc}zE8aM*GJ)BLJc s1`tO<`IӋC8d_0=K׳zpvܕ#nҴM4-( dz:gRPF]=9$UdA>5ɦHxNsȜ^Is8!rȜsdFtZYEM`lBJH5t2${W>!szӗ nk>te("Ƞu  zǒ $ )Ze0%c磘_u`Y^Y/VeÒrr\q(`GVKW p`A8+DtJGI NyZXXtJn^eR s"?)8:^ |$})ѩDIQI_#ִpa!dIQKo, +H{]G%e:C3,ŠpGS|d$q0eZd #6?euإ4.ϓ-- /^YSFדU-t5~h3[PKdĐ~0.XW0<ǦœRXrKҙuiFE_7jDp$jJ :Gj>ٌ7#:Ml >J*G֛H+6. \^q:iC&'h 5]b 6 (&PB7n$*~݉^::8^QShGiV+B'PpFwԵ͛Wؚ1}Wb@H$<>j+fɏu#~&=ɷ Ħ8#oFmaV- bɴݣx*֓|:9Y,vjeTbInu\yksxRZH0uzj3޿ʨZM:E0;NۭF?^$e>*X ! 6I"Ć\RhC!Pݑ=wh?iW1JXQ>jJiFRGFIM3!0 I0ȇUA0(Qz))%"rHzZ(1$Us9=i/#$;Q Sҙ DÍ($&&'ײ)Rq1%&8TwmḃNM(>k3WeZP5{:ՃNm5~Wm%I AZʦ `]LRB#2Z&}=vOb<6ލeWy>nUnS\O4j]Bw ~}޾j=S L2x6OPgo\!]CKxAEKwQPv6ĘGluSW嫯8՝ahykߍft'M_yuyZ47&.o4}ۮ&loȅz.fCIr#Kvڱ߸]Jlʢ۪Ejk dR*2 nu;ں=/d@6Go]ͶCnp˦ݻݶ|λ;r;?\N[[xn>#C_*z8hG/ь`Q۳)N1RAj$U%+Bh*D*ι:RHUf2)x&4}4#֩VvW6/|t3:}V+G*H_yhwAZ^Z#WM Yӏʹ~#nxvOmc Ecl?'w+FM&eiw|8u{DZ9hi D.5WIջhn;,i:d.6P6[n6]_ZX?]'{#D-G߂gFNG(a,@ժy+K+Iz@7y .)xQcܪd >ɈR' riadj)ƥk\I\d4UΩtB$/$Tg:L0>`T+{+}VVV+ۅ4E%NJ,Jp*RTfe&k%aK,\e򭊥U"Ut.9*?Qsyww:չ¢ˋ_*{%.J :ڨחVJH+crj' I<'}'=kt[]l[V-`__%Zʅ!bRpY0ks-M]H%&XJUx41K6 !yqΚ7t)PwSІ_5I$9 믳ru|k}z_5 W}߼ۋO/!o8?zk,~L{S{ c I8:4RӠq IL2%9c1'50}Mǫu|^G/??^yeǿYJ#ArXx}Q٭ig vm?/ǫ]4%w=̀nBwm}pg>3\(<+p=ɷp7o<| K5!XEPQ#Z o@KDr ]Fl HE2I2xPJp&285.D8<\0(C k&1hźӳ%j _nC`8]G^SjǹN;, u"LnJ]V1JBuMK'sS93r? e#I`^ U38# Q-!2KpEj¦0h9BК33%-}vm9KVPYa`(֝.#;0/5yޏZ㙽=ʉ c<*Ͻϑ|f!)%Vj  * QS Z3Q SxfWaN[CD8B̷)rINJS{QF'hTa@"ȓu'Y6=8}aw}(&٧!(b$JTh+)a>(")҄&#î/돐T6}YU=W(8deN=I)*mٍ\BR }~E[}P:Rs}yRao!5=yMmvrhn9~]RnDФkwu,?FkͷѸnGy#Èe/`MBt'_75}z-]MJګ׿FOٱuէuMCնK>cGyG rjvyr?CXxS mkz'},+|1}ՏV4x4̍Pլ_!k[b;AhUΚ m(6w^-peW/F ^lyUwxƗB%l-`C{?,tz3Lx>4$m6οݜ缑˙+"t0U[B)1|-:gƩNۥg3bPۺJ0Se/>ًi< )4ėZ`wO\cGWUZ}5_O%Az]]'7cwly)]͝?9odz:B5=1X&vrqnOm|6괭ݹR?Zȅn=Zkjlm5qUD1Af7C-/\ WhX~a1t/#_Ρ#oqhiutoWggRof-dDqb~1h,[81[|JJo>qMPQP}72)]~.՜GL?]GV`mFh@PPk$2DgQpQV+\EnNG"ʾBqb9nbi#`$Mr RAi7[)R"ii@e :*bzvӛ4oO+H_P//us]?+uw9.&| 8]s3!ZE8\cvZfmO?Ng7w%%q+ѐU5SkX?'w=`}&=C)lM]_K̎n﫵.ǥH芍tv{?UeT=j>5_? ja'uh#\\QD BeU*RmbF$Yd36;&(D|>g6b)UW8I_OIK5L9-Vhm#S%J^,O6rE$ݨQ r $Q&%e(nx^Dq287jFbi|q|Q8|ŤdW\q 810Zdf %>PSw,,hDȀţI(%ǂ|7<<=0g}Gaud !/N-E'Y 28s"g8B+T)V9>A~a-\QN3v0ȝR> YJA}P1=˞6&ƩFm:Gͨ4")o e(q*xr?̑ko/0mC,<_S^)rgkM1,@h$KJ d6Le=XT5\L.R`\XAzk,$4a̤M5ƤWJC)֝wV=Ǚu5ucb%ȎK/>% !W%ײ@,pAN?w{Mžr@yt`8E&ё &#hj~l8ZamNJST3TPv0 '.es \ QV`Zш8RCx;LKbs D1<5K!)62kr[m2|.ALƖȾ'fyڱCIY*8Z(/|3*'{O x͓gK}ͱTVȾ!=:A'X&eդjV@"ǟ`g 61p8tU' 'R)DFDEYjPLf=i.Pp縧)yNGNT(\): Xᆱ^s,mqCp.Nj˝g9J<rak4g E".S0!Q JqKnȋ_YAQN|OmWP< " h+d s(7 Pl+qˏjC~{[t=q]s}Ԥ `BMphЊ`!XPz,JD']ϲbՇVh zѦ5'R`'k\Ɖ`%}e<92RvT` ,͜h$\XɔQDBH# DJPoj0)wΐBy}!.?\x/bC ?t4;+&|+{.`\ nrٺ!$J4xFFWhdx}Ͽ=t5̯T>+~OZ#qVޗЏUQP4S0J('_hűpF#hBw ][ή_"j,Gr5֓+ r-NH'WyAɧT-f!M:yŕlpf_WIӓqd K(I)S+r[I&[‡ ˜sxRsOŴyv῕ՅWl 1P >v۹-˹]qR>W_]br!ڑ\?RnV?ì.7ȈJF Y̚|'z7ݜFF6:ɺQE3Y(&9%._ÐP~"{O1-:5v)* SuY~[grJH .,A[UQ^XW Wf]J[UR9NzKds󫿽9}޾>syuKԅG\r:b" wZCs#vZ&z6Ọq?fl9nJm̟.{wk6HwgIqkNGyY& '{U+sǂ" cWVhBITDU`D' FEJl7"АH.vNYj񊐜)POR@ %' a9a_pD>1cfu LbŃƽ7b?ɧ˯#3I;Y\TkQPcfs6 'e ttqj条C/ M倪9d %g֢ʬـo[n޾I@繛mخtb"%ў/ĤJG brdtkV(6CnDnt5%KV0_JzO@>rhg~ =3%:&MԪ4r"16 ɨ%N $,D\7-uhe\zߩJy`/N1.@D!qc*MR8#Tf٥EF_X-`%6Ϗ3,&Ja͖inBZ|8Լ Ƌ9lej?xs3 9G^woir᾽b1SnҍP8Pqs bƼA:w\w\IA}Hݣ>7~mԸ<;6Z޸nWxarc¨6j[:kŽ5}n[֊*i=B4ۛt8IM*>rnJBev4PZ| :0#hZGS{ \N^<.)<`ds1,(C(sD#6JxϚ-P @ bp[*PHV'S.ygCrц}p@Jn)b@ji ƨ)$*7-'$@$,) Bt֧;* o"䔐]w|o&0J-WT:9BwTtEUe4 s_<^BB.'Y9U'Ztru1%yvUeҥH$ *3>w"*Gm[C0#8+fӂ7KYPŌ}r @iu=TxWA3JP&'\q`vIrHV1{UNtSsXљSQ߳{P`{y ^"{L Ykd $H/=8Zn5S TTCSIwQOrlKFB"NҳDyDf \dG!ZlhuH J\3uQ9m^X"%1DҀJhhîiKbFDj7Z Կ%ɴm]ۺ6/ǷV}Gz,(9/~UģenUQ,d98AϪb~@D>*D1т.@7NFZ@_qݔoZ!hh] Ӥ[T_aҙ ܘ_4x4—浏+wߴA/tGnJ%9۰RU#!F$ *&V#]2i y( gRHŝo.%"XWJ}Qu=Bje&;to߳obԹ@?d\܋Gjḃ=Jc[|y6i+8[XsCӓ9Uh Y 1+Z,$dQ&QE0pz]DkM>ֹkg` A J$G"I5(Q^rL:L9m-jd?{WǑdJ_vqyD!@X`lc Vk̬ !򲹦HIy$ ˪nI&YH%@Gu+2222~pRs߻DIMX@ &*I1FU+^f8N:zKt*.<)76_jX\Boo}{zhR-*Hi\y;},**clIR"eS^E8nՑu2m_֪A| IB>~򟷯_n'u0HEюG2-N8A":l%YU3-sFpw?:+Soa UB*!oԔ<ljTT7A*zV算O6 ꩽPpueK7ysY<+fplv\bHɒ[ h&ڨh,+R-L6bP#37'nAϙfqMvGa> qFꃣ e áx^:iDхo߆ F+weI%1ec&9ItH}>h;?裵 }}8Òޟqþˋ:X?Vɍ蠮iP;^Zqp8L?݃fXЧ=Y;~jzc7?}GWC cojqjj6x)w6nz=^«,rkbň֟.FF;9<]zxwhw'CxB?Jq~/F\] ~YQ˿W7FqUS[RVq8e3lL5dfZ]O6kN6a7먓FIަ`!Xv'Wnb*䣷oSqct{0ȳ50/.-נ/#{仝M/|+Zx3?wW`&pQ^qog Dg_,xQW{ N!9.6\V|_ܜhb c9`-3JEg[Qr3{Y76x% $ ?׋hȿ𖃊vBx+B/^,cֆ[=xFo튷l+_ۻ9A]~b1-ß=hl(؃Y2z%ϑ#z=;ve爣fDOoNfjvifY^~:c.V=lwodo}i7AWa@ }z)(liݥg 4,?ttǰJÝ̔.=GvΖX5q0&noƭ}!L_k=+6`o+P޵q?GN>HJ̳U$rT,=l qK^s{9V}?W8u2l)ggu{2*\-_ag{yꭎZ< J=/[$PD8ƽu@* :e}s-W z,É1 ؎]lAmӎN`yzvtYᝍFV-֐"v/UY%}x?88W Ӳ0w)u [MΌil8 kl]nlʦHSj-'|_{X1S$swL Y1ti; n2 +KšLu(#F0U0 ͐oRhx[)rl^;qV~Bmo6oҲ)i#%yNax'5/f 3tHF5 N7i֗XDw"L ]Y5n)`R5(x"vD|) c60Hs K>!(<݊5o;? 5\jviPTjH@yUvɕJ<, `,*OVFE`lt9M G Ѡ#. x2@lk[RXIi^uȆT߈*@&r?+җ] (_ FeB419O EP)l kܸXe+R5!bIeM2$,n eЄڛLF$v&7 m v^Dok,@B]Ѳ~>r3xW9+qS̡L[A qgGܬ5wT( a2U9Ǟ`@gL-Ia_ ` Ʀ pv@P u&V )n7),X LӬ%# 0!}R`@H<+Wa2tj{VP1oh R )To*BBHAI0P(kV eq6uҰ&T]k?VZ}p.hdȀg9bn%RYd:cLz;;i "&8vM~5bcdx=<߳my] lBDԂ'.UW(qi:(OMPd髒l鲸҂X\:p1'8$;৅eE `A-I$RDN+k2] P0ⴰC4f%4Y %#Ax$ ې t`u~V$ 8U9a;VGVɒ3L0)iaM0 v?6׏A[x1+ 7DR>Gw@&1´d7@ @>R{0cYm^͢Tǎf<D /B!s%N KȳbF$ƌV F}K< 3 e#Kr0qح$fVYr@?x=+.#"Ǧb ‚y0+ !D>&vH526Љ@k>9ozΪYvL#ʃ55,K޺1 B)BMsVTEK,}2߬te \-y+x V,AbK-SfqiZXym|.7}uOy\iOyu"76VF4llt FnZ(nQ91,f J6]Mr.)FiA͟MGo=+̘DŽ rY*mFAt0 )a䘐A `H9@u 7od{4v '~ +AqJcin;0RgXy*ap@ c:-BZJOz-d&c` Vql]H΁?!\$^OkY؀ҁ $쎺Gm)^ä'5$(ztMǐ2)ףp}>Yf6 /v>xT9 s Qz0Ex.T1'X؋N}^>2Egɰ|R N\!EU`A`xG^=.,`{@eW-ޥwD* ]Ph#-&&˭!kZEyK=$ _\iiiZ,=19,[EO G/?V.:لBRjd$XfL:Nb'?Y8 D(tg,ɅV60.SQٵUtѳ)[f\Xr<,0ъ=KDït\w>^\]$w -~nEgew?ȥv`.f^W?O%ٲH),Yj AdKD- oz"оz41"@pa7Ov~!{ηOc3yпt~^=;[|њ͓kטjg+!?cp#.钋t{- s ω:](%4/By;TfϬN/-)$PAZ56|6~T5֧ dV˷_`\~~o-Eo!v>xkg 9w֫E֪} +Oku^V\mY5bSi=i:U;"q:۬̆D/]"IJVȧAy1yݿmY.q!HxA# >,tĘ`h_(tdDdd~yLڂ=Z3@f^pq -u_7JGP;q(ZTWca!a_-Au@3pctW|X.fu kq͠\Z1oaEi',[64rm?,֮ov+H,ym@upb3l`xk5UÃy#a}80 ޼]_GJ@[+ ຿"z&%dEH6feR" )|nMWϴ(6ӱ  Cm^4*θ)\ W*$KɌ2A8 8Q'!Q3訄s#O1Xn'nKh i4c/i_*GEKa8-KQ0l\Ḇ^7aUt;,g3B_h#?|<~Q %0 `I;LUt]MO/PiizO; AV"%V&Җ}mPwvk (ĴהO8ΚK je+5O/pm{ Is5s-j%m[\$ߌ('O꼺e v8L\WEl;[ɭUc]>ɭ.:I}a/ֽs(]?ΛCsBe~ n{kd%v/v[ޓ,[x{w+%xEi~~UHUBܰ eūRkqɖ̃`xgwٖovZnq=SM adT])";8s1p3?ړ\82K']HTcS~TLcS=RsYwQ3AԆhtJ$Yt,qsa:$D#c7uzђ14r4%,c_d+%u "#Bq%dY(J7Θ{vUV.<~e`0iQrL_|߾,̎Ɠ}UtΠX|WeqV cd4<_g+_}ꎬW_WJz:?D3"MÁ쪼*Ox{XڨNWv?v neO>+hz~/Ki-5xb}|dkVĤ Uf0BL';X =GYvDMOvtmN8q =V'p `6m O"BaV߯NٕۋS՟'AN0v~pniγ46J䚭Ht׃ah3|Hy)YJrwqKPE= YbVUrÏ'\G{u&of))*ZFL5)/%;٨_W+>0wwIS@uPu0=ZbG4^_[=zv,tt\{ Pz(*4>R7 ?eܪozV4ܳe_7C{2{UohxG s `g?fa%{$Ӄdu+jI {CO:cSnL`s8tмw:xrJ8t/}Ǧ0}g+M~x~D?/["hBy+|+z(Jay[`ƀ$++IIkʔZYb[zC}\V#M$s4\L1֕6e 4%te)##̇H Iq`_'ٌ8 !#{2! yg!Ή >v,RB5`r X:b8=(OVxr5υ%.wֱԋHU|\'Ket'BkLi2It霒 4JZRGv$i%ipJE=]+HY+=[Wޗ8WY{[A>}S|~oT}iBm^LJԑRd`r S>\pX/D#9L&);u-vԥ((|,l0}+((фRf2SSo$!FY*ӥˌ 0G(؛:{dcm!v>ή&s$RuR!Q 9;:]('{`O.Yt]0Mo ^}& JX҇'}n@zZr-+LV5ztЁtK5=P6.ϷǝM:lmcPMpKۓ{)YchB6q,-h#a"9&=e$]Rouܖ.)Wouco$/|8:ϴwY?d~hхy&FY.zS uaAPnV!\HsmϙS̅@:0l<9jY~ k\, VR(j Q^W"ҙD Z0bD*Bzg8<c:vb0Á+hm{~[fyXxxWfxۯ"KOPIhiTi(%9)Gp0F4 a$FBi3=H).xZ) H|e2?q&B\hQĒҕ"1L( $LNΔ 杣YiөGT`v'uN]zS{_$qնJάޭQJ\ݾ`KF7.?OR.VLQA(M\E>bM,kO+ O?g<]4=-_pO$΅,eܓH EMT\,H4NШZ^,OӸ&=q̖2 ,A $P_YO AIg9YhדuZt}G!]J= n=dOx#&p[x@7J),hT ID!u T)` 5J4x#a 4:p)KNa2Q.5|L9suOL(=~COYIGb.(E{dIOO8>ͿW/7<[ߜz[{Fũ`"^{+pzu%Rҽ~;w$}>ۯ9 .ٚ%{Ӷk|t5쑊Bbil2ճ7So\=(s Hc=N[B rt/% jjBHgzXr_Cua]`1b9 Ι} ²5%e%ߗպVR";mnUI&?8N[1F'3XlO$nF7w0xY?K?|GaOg:5:1{5,~[FoonG*Ub|?A&CO;#Ւ͜>.j2t6tjt؉I˓yu/j.v\> =|ꦾ^;ܷF{YBA6 Fa,$ KvJɛDJ JF mB9^7/gݜ_Egƽ1:{ԋ˾>f@XڅGos cE[e^5ZZp4`<*O {;_ nbq~ty}CUzi]n;2R2u+mAcӝMt1Xoe[/Ǵ 9q8TXJ],:`H)fPb@ :Ag>6&mrVk, RpW3da9&auEӨR)$\!UaQLAGN5Y癕2AX!,Wj1z9h0\VoOT'l>+k拝&`_-c$Y~oZ9 7Av$tvD2*&Pё<ݵ_Vjok&JY{I/a Π'SY:#fXKΞuF#z˗f݇?13j~x_zNJ)4&Zi:z:}݇6Ӷ[7NOg[LJmo9 ` :~s}Z=V26{wL%l1wrKK L=Vh”Xt!/M0Ko,0.u'؎P@b]HKyʨ` [oyFo= }*(EDR- HHd ABD<FM"kG/(G:bD @RQ{Ɇ %M.֏0#'U1܇I1*Zť^5@jHZd ZQdA@ (M@8TmdPDDxYT 8kJ"𩱠6#yAb*K4^f|ҫ[W76Uѹ{Ffbq<:?Rz5gaPmv$;>&NXɃ9'Tx0Vm[QYJ%FQ#\Tz(11傑 -ʾV) JflF󚱝L6=uV;].<8;zQ--˃}Ě<]OyB&x~5v^*kmD)b6&JW&f/gDk}l]Kxr-J6uBd$@[oƎibnGm8hł51SD34~*9( MJy Z3{V^ xb 2$M,v*`/O GQA,;Bh>5:[UǢǾQ6ֈrЈF|DY2;X2Z0NQ1.By] w`*8U"{vM5"z)u1U$6?ف*Uٙ$KlhB)p:8 &5ֈ!QVG֋51:qɾzQ7֋jЋ^|>!IIs}1.Yb*STY8E>DORA/B/MJ+8}ӇPa4SX6g/11g`|4> r$!GuG@oU c AI^9.Q])ZK㬌Qi?fJ]Q^+v4Ƚhi mƤj \TdQZƄ9o]pYD :%!⨐hopXW({G5 s?bl )zB<>c+tok5V>}o;'*#ai!pE(Ѩcealx d4:a`S#6Ҕ+1@'4)ݮn򅪫IVDg(y^^צe؋>d \'Nl4,yu έm`b5!A%ۥVY=Q̱*y<{C c$U4"dہxY˳9BB)+$BQR6.:y$l`Kf|D!P)fs@`،JhX m˸0[l;`z_wDP϶ m"#Ɨ A Im2ArInV8_rvZ l`R-JX*D!ŨHi&6ؔB  :SZYZV*rfT1/j/r֌嬗}bctNXwr !1"ɘEuО} { 戨B Wz=8Im@JF=˂C2e29l~iacZz2IB%o#`DolH; ! .ԤL_*KPdEGqTı SLB22$Ssv$PR$ᤌuaDi[\|F E3WavkbV9P21( PF1t4N<6Qd5}4#A󺱦0^t)ݺ* NHMSC.51NmLVQe߁Աz]A*CK>ȣA6ޓ"%C4B+ hc$Kɐ( D1pm:<@tm7bXEH BE//BfcI ($+XJLZGyGEJV \Ccފy;tiWx:qxS^Ά%%Ary xгcX"'J KqhJ˦lC#DžuyZc/F)C )-nusų]8waF}a$XJt*>Q*WJ ' ;}lDD+d 8*)K fEy6UԀi0H,֊ |e)RooD8,nO{i;xF9$/C ʩwŴsH]fTXb?1* Kv2M=G_ya}ܓiciFEMsS5qy 'ҳ3P2BI!P'Yç0Eܜ̯Y ^ձx+$7 A1Pa\(!7o}!E4ڵv1 (Pejxe^[k7oN/.O~)p5V )?vi3bsiW[SpFiuhyz >L?~>_}q1 s>e2>;_۵ aVyqw|{ŝgbLn 3]lFmi2 fz+XWf9.m_F%u}fvk5)eNl}}vЋ92Sy"Geczc?Nџ~7ͿSdY?c5}KW׭24wmH: TWYI]88V9 0ג(";!9C<#Oq!A_s4]^VvJr)*@R$.{ ~ѫcNW'=D`/e2_`txn4.6ҫ8֫}WSjZmR]Nxm/W2ova Q!]c[towͽ6]4e'F_hNR& lw{FnM&cnwZ靈@wY:ם/kRdʧt <>\:&@J$$Hs. gi*iE>v>m9<;NY>J d)B3V9:"#H le65}Òάe;GZ3E4VJ xqJ7slFIn62LpIG&9̄/nV# 1T3323˯ZrixJ* :&A/i-NZ-[%5jM l[2R2еJ̚{mg|!Fi^CaS+bINXD\$") {FwV̎(خ(FTQ xMc‡rKm2í`L[Cr|CR) .ؾ3n]o1j@(<0hQp8XNT RP2/B"o#nm8cD $(ΚkUřҰ?Avj.b~mbAT<U默\KHM%/?6ѱQ :arf$MVe:4u| CME917"isX,H+iaҖq5b*(`&/x5l? ~^}R^eىjoRZ o8CY#)V وa*+9U>`0rFRiM u%{z__6otb"lkӻ6K0T-elWsz28/`}3ͥM3^g沗"a ^` ṛȾQY;Վɹ V^箫Yҟ>*XoO\*VlGqz38j y F4{kjɚ)A3$H(1yP{( Ձ6Ũ%f0 zPW `aeX#k 5Dδ#i&1H)j|mg}z8hR7{uߍ,ݗW&gÑ8^d~ڗIycgb_>gvmsT 0Di8s2&> 9Fj-Jy߯e<\񢧰`WB뫓s3q٧&ƾNR {dٺgE|*<DU +ZE tWu R/ʔ)Sz;C;)7ɬ?Xk8C`oW[Ϩj\"')2+ UX՜_.AZ|t16A5mjQ_BךJpVU}jG@-dhmT1$kJgb㒮R?JMDfrU@t0 0}t뎐ig@D1W 0bj6 ]=|: vg.BfH `\` eHh>+R)S8Knw'~9)Jzal8NAg?臿݄&7A6G?5ـfx4vm4\/6azc(|Nn2)o_ź^2o+̲_YZߖTxuʁIA9e -'*MߺmLO1ɳi_$|L=hb:PVo?]fbQW88 +4" 1PyvfL y&3yug  Am@H;r8̍aW@ a?^>^Noa~v8^6l2l_'1~FuܔءG`Y~N2ьKhߕ)10\13V)mmD0m2=9:ɆB9ŋaǷ.:μAL(gTĐ&,32)|L'snߔ6k0G~ u^O̽|]^E#T<$yX"&WfrV }M,࡝p![CL7g1qgՌ'xY8>Pšn7lȾ.8AmI*e>u~]pۦ"9޶nxߜWYɔ?zw'dޯÉ Օ>Ə"BG_vp꽈~u &n*4w IN|˅OG&BKh[ GA{RHWg%cV ^߼,a֕~< _Gީm9 |sd&a2+3jdJu #!K,rf)uJg"iA._t创ZQ zvSkp4$J79Ɉb#inpf-NB 3VPMVaے9.X> (и+¼6qb9mK.1j)T[1ˈt0tے{ ޓb<Ϟt>< w @y`)|&f&gEwxKht"K-*$0bD#`9wJJApR]tJ ^,C~(̌Kdy3.*MTΉN)[,K Y4ELfSTgH9ҔnC7 ؂]^w65S˶ѠjnR FraH]+'XN0HV0L3Mq3I2ت&P%iԨ7>^5)TVJˋ:kbU.ujG -.#u78 [ˊASy$SHMRJp N> KfQjBGQWG;}R!]tm8 `LwtGnt]XbRU<.Jdk>UYws(v ô)w 7l1Lww7a5/VNH%~Da~ 5ȗ_hVJr:m4SpOdS2ieY;W6Z&5]$k^WRwRԹ4{sH(oE U96Zsn͵eӏ[~ #JdU' ͬ` k+ydM.=}Nh2rgj2#V9f5*f0cV񥨉Q`EbĬ?JpwlKbDWTO2iMA5IhP(Ze#0]3(KuЮ.Jz9ktb1.0fQErb2}sk̩(r5ۜ(e5;b]J^*]R9K(:NSR4s)jI:3N:0wH㰤ix9ZDxr&0e[ 8nF{1vN.'HF&𱃹1`8K 7`jOS?IwqqTJYk afXXAnJ`'yG?:44Y:ζјb2]XJ!?ОvqΑ ȀI6r <-*ߺz2f8UR6M.f6MqR@8JQ.s^ `4{g4I&Wso–r!Tj\"0}JhDc$04 C{}voz8^b$*tcR +gƀ(S~> 0ޓqcW0X 0AeuUn%KWI*I,X:Z3Ab$g.>zȥ59Z= ! C8{|Kj}>Dz!H~z8{k$>L^ƋJ Z9ZdSmoJM/z-^~Cs o{q6X_D%;/UU]~7I5Y< ̀GY|h~<{3.jCMk} с/&DILoKT!f\JNnI`iYukqʄ $H֚2w%Yq ϓH~ph n%ns?PVmeg0 c9_cm4 c Wf*ȳs/:448b.z3{czUZa6v/Ju,Sso;SKYY-\<)"5~<{i152mo x7~^)' %+ǿcwMeY05n!{UQ2l2)@c{lj[]4O-֚kV k-\ўUجF=AX0B**bI״.Lqw NwNjDSjL|8^;R4Ȍ E0.@`bt-߭P0 )،&B)p-A­%^sG)Z6,%,#HkX"%ikAOm\i̩*cJ;B  =P(ꋺޗE!)~wJ.ŏ-tX#~b@M ۊ/^/2*vya2Dz"5Ȝ[ہ dh.JY0Gd 6#3q1#j DEnhɂw.&3 f1UE]ffV#,tq{މ}8Cu_LyK#-nZe -SӈNIK4A!6BFX mh2K^ Pjٙ7RUq<{}164﷧ޓvMePe[ ]km5AԢyv@4 !N%ZЈ͝GfbPN,P(-VB)fFW+d]~a];~44y~g^3+$7 yy8ȂHsBa541-8vou߸#ZN^O?5GM"h?O7ˋ9-<'No3 blg2ߟc}sP"PP*Ř< 4^"j&\VrLDd_u'rjv u"?sm*g6Zs6z6YrF7Ez CzkGr'{(EFkkF/U>6J|wL!wj0l&Y6 D[{!F;QI @Cmyҩ|+nc.=B٬o IqQ5\![e|DIuP7g%|MCjZxLs&xwIt/{5GoPâmO ao$7Nb,I$aLRN4\2ke|Ͽ28b|wm}ICqB#AHmծYx''(8]@rqweߧ%ϓ˱Ƹ#˚bxP8^6WӖT]޳1SKSo"}pw-1QˠzS_K"ILJJي&(TҦ-P8_\7p <"B <.Y:ݮ{Qhz_<`\Hr=͆1_ x(-X6Y&[lZ|J( 0°?9.@{"AL 8NUl0%-(%W(ьsuO?2kjQ痴e]#Fa"/~`y+*c6]t(7@OVo2U3ڑ @0.RăgVnjXh!aԃ8ј@B{X)Ch_|AۤO b/RD|sI)$<Â2ɛm! {&V"}ms u^mo# ӈȮ>G'>6 *;vފ};?/\t9Z`2{@Ciu$oJ*B7퉖iR"Nu~/#&vϑ% yTi*Uyuׄ5ЎhM-[D~&dFKFSqeS7KE8G),(*RM83jD$k<,itJx\p6V{C XɎģ'][J)'kgd<89fd}?dU)5r_ r /S&ʿ)" ,ǗR(X RJhCY*65/UFہێ=5"ٳ,tUB`v.+/A 5 ;J<ӗAxUN t+Ovrui*qC/Uf]k2J2-QkN*;g_zȗpF-G.;UC^t;xB؎= i\G.2^9j+eEW^I'!g(OZuz%iz֡Ԍq 'J kT^G:%tŚqkBwxJM+|-Eoy%F>h;%-)GmRc$mvOG[ùRH&?Ē i(9O4v꠮ܲc!;0m ļ ̮oxSF(q@PϽk>0Nءv>.k#FP{X^ NU,4n6 h ?ox{|cHq˗O^I@(UT0qtϘ=-N3'F޻n(qQv6okB`<ďn8NYr0 0(!,s@&4Q$4X[0 !MvГm mkuAj'p} Hϡq8ytT8JIYGb\\`\"T(ix]U%76BJVECJShEO[0 ̂qlDBK𜌽cIDW?ܚfk+a 0SfT0z>m=gckyl\Ϙ{+W]:.yV/~v޲ҿX_]<|nA-JRjjt*>^:Q^Uݨ 4',QɩxhP`Iݱ+ rDFD\#]"ꈒY%˶Ip8SG+1ncz5z W$Vi$X)%_ga0]22b*eNi .FQV"m%0BS RbJ#Kw$*)~@F$ M @ ԏ/{N`4m#j^qHE$!8t-etњjKPI cK3ǘyM@q1νf뒭k> Ӧ;oөzqJBtefxS;e7H z˼|ؗQiD=Tz^:LSYI< S!HBq臓.eع](jmoec\U-NOZ%@FQl<͆-0#68*'n5ɌΗk "O?ٓQxثB2-L$ +(W8PPr+ȉq&DgY".z l̿~ BT17[|Ƅ"*} Jꆣӑ :zQDR.[I: dE~⛋lXbZ$(m}Ƌ㷤S™b)#a%E6raЎ<-jɌ((hN\\:%B'-L)zRjWۑ*挭ӧD.̆Y k9 Wrxx[=U7>$$B%;Ӱb1 9r Q()H)1Jfa!hcK)baCK6Y U$MEخ";DsV7CR.-OųmJ; cKCvӰegu/ ٮNQɂ_VM$IT PpJ\ 80$ߥd3q-Uعx*}ّ =mHƸ9JLuHCV1z7@-L7kQnS=C LJ;$QJ(i G1g1P/萓Ck@l82Z܊ ʔ@]B?;S_w(|MxٝQ;8u5PuC r9/ VꆗU3URT=JV!Fq0Jq{<ɪ)\o'\C!@H;|h ODQ"F"-.^@G ("p"`|tZ/4CEv!>G)hA:~jjuNGS!ښaNtC2(֊B0s}x3uj1%>fVN:Hr?HUaA|eɻ̵͓,<0Va~RӐS [el _cB@ ^Sx*"Ăq "1Ћ@D3WIa^}F+ 8 XNRC6Puӊi5yo"$-ϟYiҒ&qc%q j%A#"SsR%7eYE c2 ͖!*XPXXP'n+ xϾVma·1ICԵ}*k1䌝@٠A "PrOxJ.C%(U]?'B?,mt 7  U) |;}N:a+QG5s{ p X`U{RrqkV1n&e5b>'6U6uLJ:AjkrAG:Jw*<ɪ 醊"Ҡ }G hlZ *3aNjy IFoÛ鯿bͿ,Ch N9<~!rp ;4b;~ۍ{꯱~ut[m1˦ ۻ$˫ɰ:v OdDmT}'CwE>ҾN*ma/(9Nge՚t0j(oݙ3W3nnU{.:)[z'e gmHǓf c7"4" !:~~Pq/FRJ#iNIH%+[YT{%,OSVw cݷ^9Z+6ޠ:y)7^{NJ+rf 2\Yч)(F^Niu>joӘ6֚xDBϧO^)[1xϥ bq`I:uZh:"ȄSA KޝvF#t ϻJΤJHM}3J`p3J$kqH2 z~ zrǘ]2#,$rˑ^OG ("p"-6c΍!i}穙Զ܌ԶrY(`'-dTEݝç$a7y)VS=^ʪ07Y֫0'nv nURzBx/3_g@TTH Nq\RVۣ+Ƹ,FOon Dcm_ ^_\9iG4"ġO>"b!9@A~h8N ?@էLJF@<=x-Ni(Y-tLt)65Zm\۪jQ dꥹS.H" pw~jJ_ΟG 2\a,8l8K}ů3sEy'C #4GhR|3T?ufxf6D+aY Sd<˗Q$X~_@79y ^䕺HD$̿q#NS !kOwB&is 8Oyi$¥$bwm6Gi#HFs &4͝. >(cƅCĿ\.$+7wƣķѿb.bTƗNS.Gg2]n =# bOls'PIfp ,IfԗıjUŃSN >/LN };͔A͵te}yE#q 'GߛΌP[3f'Js =^~;S <ɂ L 0RKp;R1Ov:R`3<{v D:.PcTH)U:b >3P<\"DE[P麓0Xi3N`=O/_DxEZ&rVuvz8y[eF c ZѲɂM{f6q&á V|n߿Ym iКFJ8U\*ίC}1gu^ߧ-^*J~\Tak &oinm(y5(OtV+,^%ϣgq*ݢbv?}7˜w¶ׂPq0pk.C$4[>eA8VWΰLA;XDDbƧy gN@y!e<O8m |YLQg^[3e(T8= [s߰`:ag}m# ㈩`L~;FWԣ"C:{12zd& NB2mn%'])N ކ 6r9k1N^{BKIa9uы!U3_h$~q€8c`s`wb00P a G܃B\ V: * +Ov8ZBAm>ĵ TA=h2Q}d ԛN?f4qskh,pL_DbzB<"73 -ap'“Cg^39#6o#V3g !W_ Ζ@B ѩVf4>!EٺɨIy͛r$ = t0O%zZ0o1{v[9.8Zf+=soa&ms*C] ƽVk6oB)v)P /n헑ʥد;4_-4=,Fz&_5+ȯX~W?Unz9K wI~ޕ'Kyդl!j:@ &(Xk& vMɹReSn.mO =i Ѵ'Γ 5#+VJ8a1'.+Vi?[ ]eu?{Hr2 Knf&mْ,Kbrۭni,cY"c6$g<ꤣW際;.1G/wq|U+iTbt$I@PQq^Uل(jABj^wӬ"D%acxigOٺPCrG.o S ݂\aRI"Qpl.TZMs>%n8#'JeÉcf!Zzẅ6m#6Wl n.7łgf"οY^ץQ4,_ qe[2-d g.pr:i2ʌwaAy&ڀbO BeF9|L-)8 qwVugV%) S׉<;*v.Pde^ v:5$al`Ң#_MrWwþZW8~s`.UK7uo_V_'X?`wcQ!FC&ҧ*%ނ"'R%0H(gPSMʦAvIx?f0JP*5E<1 r X)*20wفc9,+#Q1WPqCiRI 3 M1pJ)Wi$t|lsui&3NX\%W 75CKQ.$>kNua päT.5'J(3^%Ul1T}-yNkUjw;T:N%"^Hy*"奱^(F_CyGK'yQ>ɡCmZױszqjS/z .$4`!@\2t% |qAcy/q@Csg/6_.NRZ\͸"Ja PRQ,Ai2Zz-ue0HFؕ2 ޏ3LG(t5j4Sذ3Ih ψw yqV-JpBS%y BVAIҠ=}՗h[ c˕4 _Goyu\Kv=)JYcQro9FGy}m5K W .'ߐ1& 8Ne یe4˴B""ۅ7I(%Z<zwDP.s-3z5Xjֵ&Ǽ6nٵ =q,fyӷݼͼRLrBn2A8P2F7]+jV:@ӉʗiR"u$(:k@r:.=g@xs=^RO4)zjכ[AeĂߥYQClܝh?U{≵F!y{ab|̟04W(QBmVړq`Z~4C963j䞰]sC+ZpeqfN!u(00C X4_0NPh.'4_翥`oqO:.;⇳emэIN8pGI#sgkΟ;vܣ]ꦏR;IC1/./3玟 ;>+!;+fnzC֙./_t~g2/EZIU=%wLк_q$ju4LNSGeΕMI"Ihd 1bI˒u_ޭn^2gn1T BVj[%p~ZF$k'ϩ0|gI̮Ԓ/%dSԒP`4 u`(ŔMcul&aV"WKd GB'9x1ZDgh=ֲ>sK|{'6mҮ #oX񄚔hbE œ0=I!hr1<+!ЛAJ9T-M bB 9gX p#B22Ix^#WcHrXMxLS|^6P%G`m3O+`i!%;Mм\>C,aHf0}lc2q )B0Zzp@~)srN N_]R BmbBWPR F-,ee;TD b u(5( X7>8I rs9w7~Ρ3wzfQ r*NWc%7CZBUR$IABS|S$v\%,NY^X*~#H9g O'9b9^ Jh ds3 1?`Py>Q́ӮF[;uA߾1AGqߎɰg`'S"NB(*KL-:P׷ܳ14 ^VtchVЃ%W,uh°8xiTI`f8՜C; kn0r}Ɣf 0S7L+rEKtyVEpkujVZӞHʷ C,ppJLxGa| o՝,А XQhR1 xOUEljNI B w8Y7tFAuwka*-+qMᾔL>nc0v’v%형PU$T*{*@C4rÄڃt6N';g%t 7͇`oLS ë͜^1p(Fvbgob_-L-~ӐU)7#9206p)q0}L}Ӥ3I_DV:M%n JmPVgizaZ+`%/rϳG yA:b ;R41ޢW2/fSʵ`; T Y6>k:e+2pcWŷ*/,ķgU~>|n!}i5\$V^zՖD`b.i®.]h+BKq- 8eˌ^o<,j)֦ ,Ϸ(jT!,#18libx9R}b14C\b6.}ױ',{/%|KLL8$n;"tAcW(\*^*8S3̻PE#skv_Pd滑4R $O r˺b}/iiRV|.@*]&H\ 0ӉS|%B3J8, u-cdm7u;XF ^"S .77=Y;F/b53cqV*DzM,L4Hm^6r ؈0Q?L X5u5taz &̋fT9ܠ*'M湧le dMf BS$ Rː2A*YPbeԘ0sgS9K|:ٮ IWV0\b#M>qY< l* xmTpBP2"n˨<]!A N&8K'47BH"n#7n}^L6r~f/<ע h-LKHÀţR@R jz rЏ4VsuMGnR^@ixݢ7˺2כ+4|Esb (vC,#1=f2JT02pUʑHӄ`Ow_XgaȍGK[g}~CDz>DlҜoT8MKO5/&+Q@#dqHqCܣLz%`E=h󚤬هA Gc *B6?6UkS28|(N9x (򏯷t˩G@|j\;e@_}{T2C" hLQ3J(x> EU]/Ih6.d)߉Wt2OE/x6lK[)MFmTB SgT]4*IzwodTBB[KMGm6lB |МPZm!=?áN-(GX K2l%lS3c p;*R"ރcfZ@L }>$оT*!j+/k #[4ZinXmJ}Q6w%*Ȭٹ$J Irj":  R`$I]oAc"2{)!60?砌fQ YANXfZ٨4ZYH>Ի:X :L0TRudK{ rg] %pƗ 0S.y_ɧ/}vCIOt0 /{u}ɍ&:Z4qjv42^_ۯ>Q*hp_n=:~2)I&4m844[sPzBq"W( M.V'8^:Y ֯0?< ;IAeXrvvJf|_8τ1h4?~>_Y0L;KR\\߆)|;ا ׹锦(3Jaqm,1 f TfbmYqPyWX"U vhGsB{^&z9q/nnj<疤6ĩ{4c"q&SΔ~;DID֌f[ =Uh]Mr.W.'3}_[i pӡ@âgάB;@ ȣtr*ez-)[ ?nΥ1=eQftﶢuˋ>]R//ݼ/7o%~us?鯣$frzvm"xi_Y_Ngvt /:`Н:V&u=ljiPtfk?q7"m`Swpz=@ϓC#Ҷ8 5`NCʐ]էdK.뤜 ϑQHkL}@Of[hyeAI؍(k 2 Xd7eyzD3OJg/#j` 66:ö)lkx<Dc /"ot+ḻ^wQ[h\aP!,ku(O1!>I+~\{%+Lgtʀ ~˗rWƤ>V]ETq*$26ͺ25Kʼ0=40; 2@Qcei(S@S`(sx23u3>(q^%\[["'("ܛ,P!ep+=VvD}pEermqJ\%쵴Ob}mTGчw,Zǀ:ZWe'}lݠjQ~^~ewǗB,GOJ"3/ւ=MFgvJm>!)S/?槹L,2A"S80ZV`[]0in jDQ?J4Μ_>McI/r6VPK1EbkffϧʤJɥRw3KΡbE4z.)0/KעYN2Q@`-#(4ޙv\ (ljG4^`թ\&ġT|~ 3<+ϝnr,i"Li㐁]]} 7R %{wֻqlcY~[fg}'ާj"^_]hzTl@4Vjrj:ڍ.㛯]u9 #dW `;}Y[4\t6ZK/MEٽˎY"EAfp@nnk Yuh\]dfG+2G;#r*T4i5!l^ذ%0dBC{lFi l|ϱ6\]ذ]1v)2Owt=yl2R,vMخm;Ab:"/~f nNf"KE0le\ԆkOٽa1K$a - g{aO†E+c\eBX1mA6Wbm[pƶ]{q̆K[-YNetxG[ B(7WTS~/ WI 2QxBjXFkeW3+[,))^WiL]yK1.,%YvPnZt2`~-:D⵴G \QcaBq1fEcǛW] F([Z wsNZ:SoF7 זTE'Qb!+HE&cv5h%ؠRZ,#YtTN+]wW01(F;M1Ȋ|F*%K?XO X , hM7w($`$tEt4 Kd\]NTv3ǀH{O]X:k)=D Ut%Im4@$ y7?Zmu egB?ŵ4;;vqJFwͳcndw<;HP)pNo+On!ٍO 8 ^LP"fz@nvXrԑaxߔitJuId{)tdh%ﴁЬb#J Bb͂ mo#ͺWjz7joQ*(5g hAm ׶{u]yz>n A|~vjCN 8fma-Éֺ\~^eP wȝ=J#^$`K^kQq(j\8!90 ]]8̵Ń{dBfmƀ34}:̻n"0p%QLۘ[ueLr]c ӫ@DR )9vh6 ]2?z˩,>fR§DX8Bj=8)R5vقOQ]Ά-FPyh1OZ/9=:#,D=Jy{!t*e[xUh91CTrKWnz){t1> FTC}J=s+F-ZL LSb}!g#ʕ+r09JsJ)wO)ʡ\92}hizAcO 0Lp`o_pL4N'7PO~6^$˧?nn?漝::`#RA$u|f&jPx迗yU3w1_?acT叨>@-goxx$ߪl] 5{XM.okƺ%rmVSM͕ʹz`aMGv FDvi.fas, UJӬ}$&95W]lNs1bt &tU @ڔx5\bKA,`(i"Wc?\mfa!Li^ȂE۷nnf^,r>I>>xΥ^YS 94HOtӧF="n=| lzK3<)+Wn0AʔkW9Rر̚RմY%@AV{]Qd r쀂jk.| :ԇFP{$5̫{|~C5r߳1_3#0]#T\D> K~J1ft`<>5|׉=/~;ㆊoUq8lҼkIԞJBLZA:ê&MDmk*JL#h;5޶cv=!)epws=]4hz BdʺOh# <I] J0LVґee R)H/:CWvR Qa-4N1bZ,E?)j^&̭Z> 3q.$8(E k^5./BIOb3$ # [BwnJEͳ?*HBAv>P 9{#G 08S;`PaI|WYH*jTGUt`pD4@N>D:MZg0@>BYEW?|&,&ΩlWJr-oN d5V 9XcQX/\Gb*ԷTJ) *WOJ~D>LtNar%a^2dRZԜz,'[hfm0B\nm[Ţjt<7({q]O& Gl:"9\`V\KfMev57d;0xhi5uz7/[tcYEV6nRhW2rRIt{Ah^3MXXc휂|ݲbq ŔA/LEFA(\oDW BY^&Az׎Pr(q;\Kg ]YA( Bq#oȩD!o?_>xўZbl-\_:FkL0FϵK*X1QQZlrӚ}bdz ': \ ,S5^WNzqYPy^0 I/LBa/_sVk[tS1RZ0Y;mjokeù7]<12~xQgz_G_ a.,a14ɦe[eIE][RKV7e' U,VXUjHkLͤ*G {33(a QX謕lwA Jmm%4H Beg5EUDr#7lB ^K1[ܑ0Qj ŵe3~drms&Y A9Hh O1:Bc-ql./4"4*ЪqvᤀlX$u|%}|^sdzgpz]Cpx7}gyv;QV6PFin-j.d«m&R;[I/IuS4˳\ S3o:?sTY"gsCiEQ+~+i/ TI.Eh%B[Lg!,E"h<}e#YM?)\~v=~w , H42y-9g  ^>>\\~A+5n/| T My6gݦjv|ѫK~h"7/c%&A~zSk!b`=ĐVl aiJA WeP2,x wrږS4crQj瑡7ӭB̀]TEƌft1;)7q7<=)\2p;bo3̟1ɜJgd#狛 kuIRlUX)ܞyYɕu0UQNҩƄ9Q$-Q@c|B&U*~%5 FKՔ/JS^+j-ܨ ۨ KPBt? Vo+O*h%y,G 7}c}] u5o<$,[&tn +_K6d; +#м ܶ2 ""Z vA,-p9A{VUkј!Y+4E)ϧVZNVUpy}帀 ]*ڻOCQJЎ|bKqD&uri<ʀйccyRec+[d RHLgb{PG^*4HJMZOR+ΈFt{ZkDTg]xW+^jߔ2׊3S8DL4Ub Vw#ƺZSYz$^M5 @bIχnXeT9ʢˇg]D@7rE8W!{ΐ{y* lH"Ĥ^}{3-8rF%)NH$:BcS:Ec<|eTA2]CcdD\,aEzG.jhKދGVo?,nmōחr)yX^q_ +PBLXǫ^>xVj\s 5Č_rK*Qo?x-R­(qTt2"gr!ӊ>lztsU qi22 %pj*吾*ՠ [JvO[q.g1+l,kr[.oڃ.-sb6Sn(WԔڰ,TdcAydRؖC피n-NBd?;McE d"s8-gyid#{lk %^:{\.!) wFb8(IwL~.A iUNJ GhldVbFr~AB))R -T W3j bB Ĺ!"u|Z87+eS ֻzMh1WvJPZCjaQ]Y?~OrF#E4o[pnr0%β,pNsWb.e:fvj`ypX ME*m.]^$הuY~p@?j.x:": z1϶/.ub|`EpȜrj Vbb:Bc @;h/t]ia~3C}j} Q6eiLU- `c]G OfC dH>"⚲yK{@g|>N m"ꣻt 3!Y[EiMqdь~H'/Gqt%%SEc4ä ə“٧@|$HB%BM,VidҜm^U~X6HɈޏ,"pfBL0GHRW0Sh:~z_jJRV]n5!X`v4J4np Xc僢Σ}\iDդᔶ"upbIh)ZU2ʮO^}|O78FaD Ut7'M43lǿAcqpFMq.Tj]vUr4o 11e4 ݴ4^5R/6F>4+h߱A{&Ԛ@ˊJ!=F7ـ&b)|WU'7KG>YL2AXP HtH0("ٙ=;IɝMIUCp('+!k.ʒN]>BHe/\Ր~0} 5=_–V&HAR0i$0y|:ѮmB)D)ѐ&G(V`J S $a g{$CymI ww!4AI.UX|A)hT!4.wjDӨZO^h8~O8J @#߮f91ܘ~xx+ĭ@(hyhj&u}ʢZbC{9{*&T99h9;(Ǔ䌉k$gxWհv13B>K9^gHLSyѳb'DPY3rXu#N\9 VC  AN VV$D:cyFKsN]uhG ʤUN #4vݻ[Ty<"w)[.:#H69mXF,)?`{0P7`RpeF3Qk&(L#La X`k7jkqs`f 4 k Tj z~3V |,9sSopJ~Hls U+$e UW62R+SF2voeW6ٕM7I Zv3z׬&Q3$aC)d)z8nEc̴Iќ;#B/NUԏGI7ؤ,Y?o xdFPsq5Vh3 h<#j͏Jc{yjmfEs%YjZsSð(fXv#XQ:_zY6?~ cC.o8efi+ *WVʭ*uo>)e ,ϥl5*Unw%&X`2D;cNYv{Sxo+vcv&(W9wpQ`a T*ȴ8F$)i4^!J y5gh 5Ɂ%jK$1 P1otLI]+Y5Cz[<'GX>^wUM g%x$rV8GW9ϹW^):* FX`[{8D mrOewVX S,4D12gv8|2뙗<./c4jʢ(lfu%:m!Y쪊8'2.Ws6{Td #F/TDh\@={;9u!;h_ۇw`$uؘF =5& Hfۀs6Ҷ*UpKhQeRh'm9Ǥ%j *YnZ~\ ZY6rl[5C_<^dNʵ(OQ2Z;~ol\8OYK;ҎO>Erʏ'l+cD?zV‡c3yo(C2VƩl0ŰH=7OլFƀ40wF ^a karb PN9fG AɿP0 }9 5#c t̠R:(Z؁nӼw,)2u&]zs$k_jSQyV(+U ()U "gȷ9mBQͩm5xDDkVD"<#J^F[C"j;Xp/! \$6y+ tFT"rasԵP`(V{QOu#mH/:$p1&fhY -kE(yhZ2q36TLUU--PÔ[/7~L0J~P${.6ɩ1nwqW kFIE?l۵W=̴M?}&xw{ݖD\]9~|`:oyY{/O]ՋI~'^XcϮO ̫x:_<ۿEs(_t |a&#t;ܾVvVɏw1ŎA/睌߭h/Ov>QWϰ7sx3ajy"U-ƺioe+Taw?X#VZԲ4o G;Td6'D>{2A#>6 Q/=އ HdƱdΉ>(cqbt8iz@ 6 K`LvKjC֍QMnmJ)"$ce\e2YLKWOFN@O7zћޱy 9xfpm*M̴V:<5|Yu{p27#UZYHGޞݕ;#Qo-^Dz?{RvwomE֤5)S/X _@h ln灂^zyM\wiWz9dk _40l~ǧ &se1z|kc5%3ڔl=9h 0MVkwxmnsDxoqaa~?)wR[(')h@@ፉȪޔ5JF >~*=IkJ2ٕ7†#V[驤lFC-HI@ DAQQH:=5j}*9zSћA㔏?p)aG4m4y24FA6Fk (NJVʬ4rkCJXUL.m4ݷ$)xU*go mћNY3FY<^Ox嗋ӡ]!GTly?7֥qߍ V 4o.`@`&W /~xKk$H5ˇzj>-k憯C֍ep=۶/u V甖64`F!pXڀHY㳕9kQ-lxbnB4Ru *mžȐC?8~Fn~*қu8=e4KL76y{}kk^2}gNNMʿynûdN=(K^,H{WwwSN7Iq덩nkHCڗBvW ݕɘĒCJK Lf-,o4o>Z~:)tNQm\yc l޸RDn`m*ML恍u6@^&UR9yZB c7Ҏ|D;OF;8:XU6Z a&P1RRPزDӛ{Ûe$gvEa*؍U tWKPZ */X͏#O j ~HIr0QX#d K#I{<ƂX7ڭ'@"7@ yzذо|BJjtҌK m@†[UUf'Hɘ+9ibII7e4I'ctl@qf7o4ю }I7 t,wkYܞvmV[h++]9`ߔDi賭gW9 JO$}> vAIz ʹVQӧWXblMQ8\_ƞd!mo~vd7!/)s;2x+W'Q ~?07/O{!n#aSgg0#/OS?Ϥ|3xuW篮Nyl7zP Njw03y?mYy kSBtRەn<{˴ҟoZwß9` }Sa Cp Н*¶R8Yo8={3*c"U!k>8n6Uq=sSZUfДsִ.}} =2@)>:FKf}g{(lCoݏ /8ټ}OE/z~{iY#F*痜% 13ta*!Tjg 3ղI暔Ƴ,*C H$ytΎ YRwx[(zEi) I1zD]&p-z b=k`P_NwW淭MDom"zk`Oiky2k2"sEyQyozs4=i\x+yqd w5kʥbO[&`u$3=Ѥ*ailt!R^-9u8õ&_OԒ^ Sث'QXVzh"t8]CgrR.f+,I3P>Riߎ*ã!:ZrެXa2 ;&7ag6IG@W?\J*<6"F>-v%|[&zEiWiE, pHs0Z~qOoDϒJ *4{%7KyeC"ry &_O+;Ñ^\=bw!(3ϲ29tFZ4$ .USZѮP@D & @+G[b(0$@zI~ƽ e}.st.N2fm`a-W,%FttbӪQ^c9$&Q`%tzgz8_ v_.W۱8_SRP&5/1TW}_u]@hUϨp?}3V&{6E<@]-! .Սmq#jl}j&{ÛyMf݊7O 﫝\MߍVB>sjwU2) UC84W-ZnښlVc-wZ0[He̮q) 㞰lG Ս7ILcɒ6E (X(xHVw69uނq0b ?50 Cƕ<5AĪ>ZAU8MNADLD9 df/x  (gyZڛ8 = :7@=D@NbGDϣTNRJKЩ@aoc33;CvIGAx4NQYo߉Luk(ǠZ JIlaEf?FN0)K9B^BM*K'YB$sl1]GUn!+zUʯmEoآAQMRk3a*n,zݨ-2>‰CY\`$6d[G :u5Α?JN|jДHz Ҿճ7J$hGŃ"(5ЀbS i@+kDJm$u1=2˼pߨZչzԂ=k.y2E> ht6(zklcׂ=ἡysSN$aƃEz^ z$O7рـKRzOO!;E͎l|F3`~bFیB\#7O{>#W7X iP )[Ða 2&yLc1Z`B if'!w$N%OYPS]Q=5Yw r[ldKae9|#I-hzx}9|olx&kglC:}30 F%c^X s pv浄^}O40QZ$ndFl}|CD97 !ylWHꯞ/ |',kt 1Tu5Z ѭlf{1FOgV$g q&AE;{Ub0$F툕P7=)Ñ#Ï#Ь㈹?>C;êp3\:PYNNCEJTNv.8r {q9uʦGĚlLEMU",rdy?4s06i-k`Dk ]bU?x TrӋЄK]l[fò5s]`Qp6:نvEWxay=i@g3=j);룾=u&pr_ޭ#7Y8"l=QGcDژ^|渰Ɇ'.yp iH|CAą}'|>tw[3h#K\‡&X95ć"xmCt#7-1+`)2s2H?9X@?kh7e߬Ǎ5l*sus֎jɈƂ'Rýz"m fp~41c:k1N\qk/^wO!Y1.\\B1}9dF؎w$:{Vl,y"U9>HVRAt_] rc෡8ï֝tȝ({Y97 ֛jtWNj!m1IwMT]uV ,Sڅ`|lLkUTBuڝ$~l?a>Cxmi|'.1LrTY' XŌrKdL7+始WWQbU9ْsjۃq%'BvJCjsMaÑ]+lhmQCdĥ:*4k^Z6@cW 9KPo{=! ei N]" \3#KoZJll4w1V_-dt-EX<=L#:NܕMcJUrRmVNzfΑeob30E8qƷ }r15j=as81Bgܠ!pt(ׂ7~ 8qMk"z)R;Hٮzժ>c4P_x.]JoV%8]U=lz@+%l{rcid,f IwCDS {GKvvG0r vIZC3S`q$`VAXᚓk2=RIj/'%R5l5)fL2&m[k?ft.N:Y=P͋qq$ I(A#9;d OTAN $ :pxdP *1/0IA-`cOQZP&TS2jZc+m;Gw,FUVzR_|3f1G X7Bzds 0ȅ?lVenEK7M yrE C◪yNLzfl2kDr|5g#cc.sP 6460(so}X|6 UltL]p ܐt#[*2z nY,XWި?r#+݈l@O?>]~tſ%ˏ?ze407/H^mݻ5?3(y.qW4]UA}+b?#^k}{_:s툫d>l' _.`I_~;/y 2|bN;Yۜ 1Zr*6Du.%S-&K)A#4kĵ:ijW !s@ d58^aj {=޾#)vង%O'+V{D27` 9qM8ov64(B۹Fmؿ1FQY'5Ӌ;v3=DU6 0\%s;9Ellʆشjڿ L/KȹN]ݞ!_QYуΈQQU UKPvxc8Em1kMWg29&绔::/.\_-+Zfimv v7rm%k暲oҘy;]^1_œI/FsَWSd;qWc2jЭJ3Mn.Yi38Cwn\*6d!p9esaGJwSX&=${3l1}йc~_d׋N&{ñ&pО-l]qӽ!SBbWk̡C̠!1ӖIJ՘Rc ܚ)&0msK>[Ƽjf[x/I2Us^а_'WkL&>[Tؿd{f0Z'hp HHb,0A[A^3h,IAYy4fu|^,y:z jJk¿V[9^s 3J@W(o qʁUpnsl ]ncbt,ljڂ:R Ғ (p\L*`qs٬un#5aH1zv(^O󅂌rPG}-)Q廫D2J LsXAI y!Hᝫ69Khs6(qJ0>Cb=})Ć0 O<ƽh18a.y@6i<.,Qy} o`UW1D"e]^TRJu璠C\ u0ԤQiNnpΟ4Qؑwez3R<5R1RmNF[BHNV 4k5[n!RsD'W@>lQ".zz9 GGٷ(Lˀl#'rZO[=]'7a?]dND(3I w:/!Zgo2{ h2{ul≨xaFSh7j%@{Ćs {<1Ŵa !q۬x*%ONqD |6Y 8p5œ6΀U.-v|*o.x`fNQ yDi2wem$IzYb#"hk `ƻu|$DJ}P H{֓e42Jl=׷' F}*@jiZ H;6-}{MjZk8U oPs50MlQcpx&xpf+siYuY[_!;$ی`t!zHU KI{9$)u1TVKo23]J],QxkK5*^  3++nRҏ̍ No^ÜRI$Dܧ `!C!@*lPHqO aL:[ȸSHv9;+U6$ݜ\Ep!d0)'ɓvʺ@BȷZ1r-Z=t^'64Gt[d߮.o7:e4)2VNB_!SYۋ!g&ZoHt}aYQ+,k+Ns*2;IB{AvH(z#kmDYHڨJ<}lB$ʔTޔ!zh(XT;fxV:J:\Mnz"Ct6I3hT` IWAP tzd}+evEk`ZF]=PH:@ZOgLa`4 *Fg18Qc,e -CxBleyfO7==ğPcq_ǘA420 QՌ6Fk7r$@؊wxR7 =&U K=%*N8"L3_vcQ{_$[E5Mb*Z!-1n̵ksrr.+riZfATٰ[W[9eҪZ\y5 e( D)i9!*ej)v.2f]2pjy5&iqmgQGr(˗W8(O%>wn&nŖbґVr%??PH;1R܈Alԫ mB-W:RюE5qEi@ c}?tUϯi/RRժ6ɇ|5鮜FH.%{$hd[ʭHۭ7&M5v@^ Qpdk]>ܺ׼GR%pJ2j {y_=MA]<[03+6T+XlpJ6΁Ir):Haպ>]<y5.oM[5&@_[*I:ZcrՎ$W;ϗ\%w&=İOAEY11ѳLNk Vc.WL"GKrr Obwe$YYv?^tJ#(,5I嬇V<)ǫ{#z]NP'FJ~A,F{}F4;u,VuĖ7oc Yg1޺Z ^v B5Jnǧ/ kg&&YDV?U(|עo嶢 "Q>LX$P9uI[#~;:8L[`W~ ^tyE;/O'f'NMXL<\>Hmp.v%,4D$zwzNd/{$NTcØ&تKnW/ 5@xA(۰khhZ32e'C 2*L#:&W*JeP BLK]]޳V.@"(N,꘧Y1w=@+| XlԝiGK,o)X cYN-dmDf~{jQr=۶u@A8dN1ݸo/v[|m*? &}hR6!9RIi}X7n{Lh7lmP| eUsZuت6f`Ãig- 7j!zѾ9tݝ-N׶utX{ HK)7'\sfH%~|^9$cJR3QgVh+ I$H]ƌ"DBɢPg#Q)UܭI6][:}pqrMl i,YpX%0 W 1M@N+c$Gf'Pf NE!QJiս-tߐ iinZgG^I2e,n JVlNp>tt<M輝^kָ~ry:YgA`'UPq]N,yX+Do6}?־k;ͨ1{JssRʈ^H u ]T]| }7Mݤw[^K6K$`5DV X3%z"th^Uq8]>bPkjbOOX d:H-k+ :`c1H0bDBkpރI:M>ph,YۅCq@a()wiM jzd=/¢w4HH:>9N('b TDΣ"+U,*]Xp#&24e 1-X1UI)Ҏi cV_#V*Yzuhȫ Z5p2YM>⎐Ⱥq}ڨ "-O$cK&c1 d¼N%ubp&;L&A*€dK]BpޠI>,Yu)q;ƀv俻{fN:j zy|Z0)ΤH@ZL[4KX!LXDu;! ziUdcبTʑщABiOdԦ[fj.Om<ÃO9NYNAM0q/3CIct- 7fJ,LyIWݑ`,X-! aebkFc0M!W`cUL]ٕ f`V¢Qmֆ\K;zG@;A^GCm իHaȜx>jrz2u;szFsç)3k|m*^j JU!DflμqIBJKKTT VC }Dt\0)hvVTrH"!M9V-M;_9Bth&/ ! wd:0^gor l>&⯱/$bL"BwE̗%yآ+Lƺ1dCNi'0=9agwmqz K0q!0*Ѽ \ʔ|w8;K,iQ9uW972xv8:uVbѣNgd2;ƃ"ތ16/Q\Um7w~ FD9hDuB!f@Wzb֠RCfZ*c܎LVX}Q%Z]tU_<'rrʕ|-9j+iʧUעt{} 7Ξެ;CsdU9&.ncNIcj42;&gEHҐ)D1թ\=(JN;<4J/ 5Os҈A[=@>p럲gX?/9O%q7: MїaS~(hJ5?550"?d%Dd6r;޸ٺ qjV*q5o\^jؕYJrmU ̱Y_@P kRգT%$wO5C5vd,;9 IܒtUm^9kSh&PbmESk"Bq.KQ)('h[0M2`B>d1IYG9!RZǓٿV!A+lu 56ېn]ޜth41GI$\/IIWKԬ& gP&LbmTCZ*/ a:3*y&aECk6m*aAMgm)/wCf#+6W.jfUJ\ŌUm؝V\QƐj㨣S+;&g7$7,X\C2E,Jk/\f=P|zX-5RQ8B tI`ng'ObJMԓ-V[R{~[a>֏&k/DB*k6(&6RԆV]VP+MPL2`%)G-od )UJ-)d-y!2Ȗ&{lWOޮ 6F8Ws7 mCSFcI' Eŕ7U]qQ>oµ&R0&)vJwpi*Lxz4m/P'#*B^lV[a URA/t}sb  0dtɁ.OjFdUfsIG(?&E-<,㤶{ͱ^ \4'LfyRۅن#sZ $Kd-,cSmoQɶi-$inSo#0g7oWG5*G~TM}nu60w[',PhC(afJdsImYqLC5 A?{;Eac|Eo[9(m Г4Gc ALR!v$yFh=E jd:hg?nM{W߷`lw2[NTg\)ϵ#}eg#}~@=|R4×C W_0 ibNr" H_)~M{|S9 Fќ5SƜ*?f;-PT?/jrNy Zj?wv;To ~v@#M¸kb/cVwl-av5WP;֕Vށsn*F[`*ye#g4a@:aXIMe#Sf۸DO3§18^{7@I;H6?Zv`fg;&Ȫdn5'LyT;Tw2j.p{Ì`vHfO6@{n|5/;rg!6#ɜ. :R#UT1EjO۔{Dj됩` `_Qj`wLkP)鬟:5˱G,-`*uwH)r7j~5pr#Q nfϻq( i{ {l4/ ۱Amľ'B)왿 MbsYYRk;о*cy4M1en9!)Ud~_1+Ƞu6D^-/N.^dRUju!"iIAd5G$4>~h Mit@h^'ravbn)\sNnzsr$ކQ+V0j J+L]#b{FF)E)o AO3wμ(VyFKrߣP)gv8vw1U zFm^'vU2;V@ G?OGE e׫#EZYZm0H%>j"3IZb.@]O2hs\u3{f9]\܌Ț 7\?I~L3o®e9ZH=+8 n|܋Fv)TW6DyK[ֳ7ϺVPzSa[!~W߼x|JR}+>B#UqZCffvtAD"tdE]e ȖZ\zLbubvK/U\dѴOYŒ uĻr_]sI/Zk]nKH#t4c+[͖}HXV}f lJ(GBAFHttTWlKٛvFz,JZw˔jy&u6Z \OߖvmQ8. B|'O!{[,0+#|UT++U]36 ץa5d'c|"i9e x;5ش_y:f5䚣u!S~3ǗMV%PdRj3lMިȬ2 @g&Euvt 5DaT7SJ qLXn|ʁoRZsdG+OݹS4=3:oNZέWZ f3$mN:ߧ,Ց6u>1,Onb]!\V%OjOwCɟڕ Fpı28>{OZ:RN)/~>͸O ZEj]Toy/dޔoMJϽWciK Y(ۤnuOdy Q}L9V{&bpbE|W^aB@{ :P,˥ X%HԘmp:BZ3oEd.1PtM*;vQk-3>1 =Lj66_.a+Xx8!rx{]%CHh,Eł])}]osJNTr.!UG%րA^R!{)j))Y5Ft-1n26_}/QLf Y1)4&]҇jIL ]IJ"o`FX֥øLO&/ JJV{MMQbd754\ۍ CUC3, YdO6S:@T7ǐ'kmKՒ/zP) :\]q]~1 p*xXFtFmh LC6%Q N%v+Ƿls}VG3)F귷0ފAu0mŦ] i3gopfy>+5[9qU' `ƕ$n#89RT_6i2*kь-Ju2dV3>W4ubm0H˳a}3gD^8{zAAC/:>2lv?4嚥-͎ Q|G,!b>>':?ꯙL.X ~J J0kɓ#%Ð!LY Rt-Sš3W儇DK_[\wYw H/yfh=N%p>hgmnHe7y^=UÕͥ6Wq퇋`$FTxk$ESIٖf=p)I7s 6ǛT=N'nD0Pn#ȔdW3 `|u.ՙeYP|G @d{U1d]8?q"TCv=eAOOf-wN(JKƓm7:;ꌒn\[˹ r|-B#ؔB%B"q#S wiRMl@-C^Qt%rdvWn0ѥaUF%*Ox,ol7"nMM8eZ1%%n $4b= `)sc?37q2t NVa&n">`.keh*Pj+4;$~ete<Ō ւhj@j#()fbF I*^qZlK]AĶʍyno Q׷OR]C:5'JWRIIQk@4Dѐ0Fq:sH$:uJNUrH_C*,uHŬ0BBXÀCrHCZI)87k4 _.]CJNaN ?fMrAEt(Hi K"TPA#3nSO䂮oVWafB! pt})v@9:Gyq%2qM*Qt\8#nbn{g-UUGR!pBu2Khdc.kO>|GkϞ`r>{ڡ8\c+ObySDR6F8/=plLhDD^LK/x4h"N煻hFW;Ww}XN+ P X+wjN՟*kjI=M˶.ls%nvtrS+hT b#QYv1ސ8 -̓YhKL[ATf9V#JuVYEyWCpb lG_PӔb1tg#?Kbo>\bpaQF=B $DŽ[4qceR {RX3S0:9VB%O?q,.Y噩}&ޭ12~}U1Gap3̊.Ь#JOf?Wyfpij" 6-]BsyCa<*`Z2~|tք $kqm&ѷCF腤Tq5v" (Gtt9(޴yi?Dfо<5W&ݸCK4Hd9Ͻ`Ld 9%R)F24$&2cj{B gxf We\qDaΕu)sC߃xSL j[Wkmeӈ?F"Yx;";ÁfV+=C0M 7&H]Fwjv:1OljV*w&inфb$&?hhG_zbF*np D9WB#/hUhL :>#F )Z@뼇6r[ȭm[P Fk7h"^ j飼7J`jTBP1,SQ Sճ-%^Ȓf<퓒Y5AtE~(aVYCW:P@+^Ʀ9q H^Ufn C[qrMIR .Z054 anjKCI!v X8W(oRh'vV14ôT[豑L8 *eE(%Vm }TvkHM][ v).RjORQ1R~;BNۆ@FZOaK~~K\hF/}r 0IWw,ao6+b ͢)qq_ߛ(a/D$b253n҃NlB; ~T(G[Ep)s8_ ( /܈ L 2UJ2/9B\!4ʳ/ZbL@&MjPZMF)JX%aQn˘B'ٳ~)=*\"5֓!#ooP)Amaf !w9^0,(Bs*6@ z%|(S)?A)4*ʶPv|6o,eW,1VZ>i4ڱ+>ruDu{ l+-3 -,$E}n.cV fAq$EH_ɜFAL*IY 6?.&ؒZkSXeQ3fPmQ#s*STQ$7!NN 0!y,&Xx R\(t?<բOs:h+8O.ѡ9ONâ2VE6Bjo!y khoBl2S_jbLҎQ=aJ#eGDBVTϤ_svkL'PcmT;/1v֭̿IW|<`w5rBys4er0.`2Ȇ/N W(.c;/d/lʶǸoO\fq) +PD9$T#$>o6AIGʼ_.Ϡ/dm-袸 Hgk~|~;;}o:m3('vA?\񱼰i'd4CueY}⋶ZBR. FLDKt(iRLd?@ WL'u@"RaYjsH1.Q#RLd/c"]"⠒CO̅`I1Ψ"D/&ANZctVۭn~\{F{%{bV3䅗RߕY(ÅʅTdV'Iq)*f’DG}û9axR"(]~wsy0yT=$:´b* }v8F7~;@|{6dq;n:VB卛1ؚ d OUo!+Smt߇d-Id+brH(1J3^ +rioPV ( 6T2g4\.Ykasd`JW^s_螢AE y!y2NHPNg!hFaY8)Y_d*(V 7b(̑2B壨 I~%*j'eQ_`VAܱaDsz+6Ibsgh9A+`\ f: ȃ޵md"蟻0EhAp8c+mU~{z(IYZH3瑆D zAV \h/ }NlGtM1V2$vRLgiQ$D.)ӑ^F1vc%FpqZ4 CGKYblK*0R"~ V$EW*٣f ]^. ?{w>6 YBq~u?kw9*#/簗Ѩwys uPT;R D6>({鱰i Ab) 6;;mLl/6s;뛙{|rƓ!❵o=nМ#d˄ 뿼͠2qb͉567K!\k38My/Jo坯2 |EMR54Y{c&+i~AVX; ߛ;wj]!sIw)$7ĆJ^Cf9W1P" Пh0DtI4G" v`mmriTMx>pvC?]矉ŰY~'5vL7;&Xa,ai_+ .7xBہG~ͯj? E&Z$]=Zh[xc2ݰ#dZ_w)csY +nŢbQkbVz@ GDqbQ){7//gcM-HCG L'\rp#IJRDYՇ}aU~gkmcpň,1ߤ)#T>[UHk%\J>յJCd3fUEu.rL! SYI嗋4QǵTTVn] =jZZyżOzv^myr!&KI8@MY崷+B*$1va5WRSQFL;Y }fi 2+Me^_49`y'HE!Mk:8ZM<8-' vH$p}B-* ۖJF# 2R"ES)pw)b77>u^~Ƌ!kjS\4v#[\K[f/ NJq3>U|d |X5#*9P$%KsDc|$G>,4vbaAVѴ[9\j” Ѳ+7sJؑwT8G\ZFF3e"iJk6S*=Z۟v[Ira~KdH6I +1Jl8.Sk\iArUW*1nl{۠wHJqe'yR*ۢJ٭y%U7NWh K}4"(1 ;F>Ͻj}WXcCޣ3m%W%a.w9"QDXzL%}#\_mdzڔ!b C3o~qY -fɍ0+3N&¿?/~{URj|K,R` xR³DYP~ CꗳDWjʌMIj'e]H3mWC7=1nA͊Hɒp2!,I,a\gR`,;eM=EکM `Y2M+ n]D?>UB j@Ұ'VqqJ3*2M gKx)i-acS1bMHEJǻ5C\ uTq`B bǁ4e2ʼn.laIip.e|b ǢBHb ys'6͠cZ$^]_N) u빹{:ym`bwlĻw{vzŔ{Yc[ٗI~K\ǃW$33Hr:Wfh ˡ7) nno?U3Oӷp4`27mS1NŷNIYY"':,DͩNlJ YX^d Fݩt՞y%*xs>$Fܫ Y%U͌# 0oĹсG qtaW8Rz%*+ݣ|,iM;#L; LYi(4vѴ[}R1C1@0!v9 LV *Nݱv‰I4ힷiRGeXfK1OqEL #Ч(/]L,B˾37A*酝7|8umm6:r鳻5j5Q&I?J|;4AD4Ծ[I.?[!3}$K/vi_Ïppy:9B2T&6!_g "2\Zb-vػ\sթ9Ijb.<%*.BhjlqS5Ps lt>H{(\~|700`swӬq]E@\V.F57ߨJ MJ"sa le.k)Iq^utYe12JH2ȓ)uXiToBZH<Y62H u8 ^->w:@l>W@;JGA6qLLE=2|ӈh>w{xL(ЩØ/Ry,$ =>TU>u@dXB0NZ6o5؋yK@T`tͺ%neBa4s v,Vefנ.7+#ό%cĬҊ+geg@B%ho`C0q8NO1'gJju:]~gl+>7Ύs(lCpPyeHS2A@':*RP8BW JIEԒE b|'@aᗼm{YU&89dɷ`[0t, n p YhLK0Cɜ[Qm;;c=v=mv掹@[ b4/ _5P!XMk\΁*s'/r>pvj;~`~նx||mo:[eMB]=F99EڸI@Kw]d݀clŽ|+6p=3)J)KEqʬ. S)" +07xWx}unzP}OKJDhT2^op7ZOxo:Yݳ]Z|Q "0i]% pH]{\Q8*KOFuO Xi1#&O$шX^j0&BKaUi0`Ò[i{Y=6XK5hF<2kIsjS/Ӳ:﫾Qxi@VmP'%غ,, a>~Y 3kiXbeXx4)*I ̻ax&g ό2T{IfZPD)nK*FVK0' g1 %F9Nny!e߇]tٺYl(]v%9Pֆ!]v(VVa }c&cvf, ŧ=n)Y8^9?yDs\(Ȃ&,G*54`h~冂 *LFhܵ Ԙwf/7T>%J[hkqKdbacp|MexJ1-UU&ni9rerZסR:c=?3(qZc/݌[j1cE?8܂#{0sp<788v0BaJIIdM9! E~dJe*;9iC6.\&`|"M€<:l<#~էH l)î>=OySV>ebӋCf'˃}]?g.o.N\):4>t[@2ec,KmvEs~8D3 |hX8rR;44BHm6:K48{, cZBh{SDo oqd:ݫ'r+ܧ//nSЙ@u\SkŔ}t$v9jG᰼ӶwŒDc+nQs8PogO!A#IbHj%J}q%ұ ։ 3;ܸ^tb^V5!Na~̺g@6fM~X<;oZ4h (8aK龢T* qJ c0OJv.gJ!' N9YNpJr8,NK$~b+Nb<}+?I18 s"{,9e1_3PI ]Y鍒TBikU9PSA[]UuTmyU2Qe"զjl!NjՖG"&MlwCY 0 H_G.qb]޴Λ܊7U7zobokPm؊>LqaWgͦ8h$D9b; :c5C"QrCuӄ)d3f3{Gk';uCw$1RW[Fȏ2"$ ^\K y$܀h׻ 6W^% Lɓ0W~mpUi5,e 6ĭIU쳄0y ,=kԚXR;I[-ogC_m;ۅ*;pẺ4+ p H2ţq`IoY~u2e6b*ƋWFtF[L8-X/*DDlxO{!"J4"n|Kk7F8;bWgÛ-}xO%9 Cm7@1&pK% tȹaa'fnw \l@XtYPJq_IJxԲR9 /^=ޭ+!za^ŋӵTEw{Ъ=h&mB(u kIY_8IoQ+hPRPܞ${ 1R흈UKU9En 3\(* 2Yf*jZ-+0ĭÃ$x((uO\7t  V|-Øp<~۸.kh5U87OG:}YmC^x&^!8S7gt,cv߈$M^'"M,%p|X`aQu4eAP>=;ۛɉ[x+|#68ק,׭=ʟ˄n2Sm =v1'<՘D@aq8 4C8dxX-H%1Dtvm3/{3o1tuͭ땭Z*s-~ fAxwɺ/]Svwh)}WrI|(}V_;U8NZɍ?9IѥN{1'e5g/%Yt9Sדڔ)dBF.Bqsyugkg&ڒfhf2T6 A0xZs'hLH{Wݏ*tM^lęzodtxݾѩZ`E5{>^E~tr) ;{b;*$6r^c0di(=z Hp_)-`%Xh4c0MI;շWx^y 'uڪyP.GORNwIvSk5.a.q].$j,$B koioUCxcfU?؝TwAxM!!ʕ'zJ5nU|GѲ*E[GVc;-[97whYx=}D(reJ"+Zg1z d~bCԎ–;`-?aw@'Ѹ+6`Ѹ;=JwgL3;#T*[z|lp)'MLR Ǐ?lq! nb ؅T@'xϣ1*H8f1$ 4q !x@OdP> !BS_ |Nj@w5Le};e[]^&ka{лԌkr }q.5lct>g?|jXVD]茲Kc܋'VdtÚ ~ޢ)LSX7JSgL4ICq>, ^3^@iza8& M.siGn$R*BnVA%'V`tω7׃?M|ɿϟo{~p#0e:9=~3L?氘_>>S_ƅVNgƟIlPҫt=1+y~0s~?笛#|lNyu T:|p:L_nǀ8dM<Y\у׹l/:"C\]^EN SNVנځmWШ<DtaL%~'W0oW(|ߜLy|f4HûBcJA9@_T7/⨠U:}=6 ϳӟ0BK Og4|;MOʹ1C<,u^`yY kzœ5R}>ѠP.IKl'WwSH)0O~S( ]>aoX_ ѫEzRŘ/5/B%&K̪XͲ\NY4(^:,k =>_X Wc<0"BĽc)OPx" 3Xmsia&cՄ\ 2_ c.%BHEƐri[WM]L6X**F؀%,9`:9T%,Y K|4KR & 05T&(5œb$G@"2`H,KK| (mD^o9{G<NjM"Z$Wl*!nke`X}&Vx0"w!R /e{uϚYҎ_ɉ~RϗMz-!uz!2]Oz2ꖅJL%Mt4KsV+K n9 Z@bIR0 8hDr5QS-|<(W)QI;kǯrϟlLNf6O~ZuW` U(PlP)lERS bnG᥀&Z{ ʎu0=,a"aT 'q /a9g>zuډ..:A# 3'{7l7rݏE" DB?QH Ksg2`c)@C?%P /1o9괚c=LZM{ӞߛaL:ܛýS[%Vc"(!Pi!HLi342ȇGr9,@Pa, $`ThP"kȻE(r@Uxo>~Kp)P~[A P%'H&9%J5QI+\J -,0T+->%Q4S#Dd,Riyax{^&zbm3`6&;ׅ@7Rp].'*p=WGjG?[_o s.,@؄&)&P{ @FpjgSƲȲ\0wvH %&yy4?Z/e!$y y($\ )᜝._ӣ 1©" 0BkAbK]g쉨R3hREi,P)39ZɃHzaEG!(R0F+ P&ZkiRh$_W*a?.pjs/~{sJnVXp<"n^=mGj~DB b%A+TRDR$ CBa DB0%&rlT^Cc%Vi@) %?Rꬆ7(y(PyfǺKgUTA }II;pb?(CbN`; &Ô$2ތdDb 98=(xO/#a:DN 8 q)X':(?Gԩ:_ǭl v!&#(Th)cLCSR<Šr$,Ů3Xvm /f*&,Vw`B 7)NRj Nb1I59<_n0mi>Ws4<yw$)2H""eH8h!!L.ZíV+ )bɀcS%B"U5I HA"&``0Hb)Rydؕ+Pj~ S^FD@j[¤z)qck̈́$ʋX Qj`Č)PeO^F J'r:"?K{NRcxm$W Քr@Equ-gܢfEcS-̆l7v&_Su|Z壀Cm&;ZC~ofad' izb=䝟 w?{Sb F(Gbr;B^$3*SD-ߌmwf+0GN0Rn={O$%d=<gnxLA֞|.gML at.)/g` D\)P#gF$@Q.*3"?PA o"JM߬1nBk1y&J@Oс܉--K`MqK5^,['Q}PL&l(Ę-ףdwsl" w._µAg-mLCmǂ6qb6'L=mՔ u|?@x"Ry4Sz2֚hTS~h>)%wWBiSg%7 :p)&^ Dj"#ᲇy9cS,v[+*"GSϺ =ýչWJW\Iq~9n:潗'7ia@/_vl&e#%dN)ӦUy9%)]kT[;N?|m+cܵ8m*xkx7SF*jCTN~T#"^x %IUN}ͅEHq}oI1&'9|>|7xA3{ӋBed>t.("@R%ΔO-}TqL[ulfNOz'jpEv3ڕc ÐӐ 7q*3"eHYit4 ~Vg$=9)ze hc9tes38kC!5S|u*vKZ.RAÕo5.Uk[j\[=o6 i·*qu:*DI2ހPrҶ@*Bm&q5\a\pLۦ6L(' H$qu .y7 WAykx7N$!v˾<:\C R}.'g RFS!`~lN@;OԩH(9bOg×I@]1 7`ٞ>q6`Åer+Fl8!cܴE;_>,Fn>ȪGLM2+4yƎގo&l#cg3ΖbOcj'];e4}G0p3#T@ZBpsRڄi߭djDѵDˇCLckf\~V^7vYPeܒrmdSJ[cVg8()u֯p2)>M {Ep!9̿y FV$f[!>R & eCV( 7X\i D,t@#eƦJ2,='RZpN9!bR+fC a ~/@eK[ld{ƣk28dKUO%Tedx+E}sb oTk\@ZSL'sy/"o,K&ڬZ1+//Zt zXO# MT́ʇt*І(`ո$ߥx{V([IrTu ˡ f gCkOy /?dyy,, a\٥we0wYRwM :]^E_B"ݝ ;age =`K B%fn鴻IhTިjNAI'7Y&ͦ/Bf:{gȟ! ӛKar$4IeS:uR0]dR,ļ(9҉7kGOX>L KX=X?oQ('v|*Tv|#ߪq ~'+Ue) ґѶOYꇁ`Įr0=e8`Z>k{ sdGmZY*<*$ E2n Gv%‰1K9!oL22_0Xoצ|L:u0)b91T cb2)Jyje*0Ŗ`SWEuc]Y7+ ?OxF=a0yHm}UV֥f3m=ؐZٙ/n2{( nu.Y/> 4Do9 Vԏr2ؗ;&bjr4ٮ ZO_Q^w.zhe/]^fvqeM-\)qĤ e4 YvF uETL!;˨9j>w4V[A$B;_yt@;;w7gM3cKV΀aM"F;߻$<)#x5"oHgB=Q@o-/HYh?qh9ݽKF_Կ~_v{av]YkFPMO3gU1C /crXCevPRds J(,׼0rZz?naDhfIgH,K# x>q^/#lW9zP[VCm)aYIޕ/͛|u}{uV')6,ĠmƧY%h1eJuߛ8WbՄ{2vԔD`$A`z< 1جXd錎IHNO$˼TR-0صk%@$It&9omu!ET[x'k$tL" DRP${-h#B 98ZJ#YF>A  g97*X8Wi:[&8/x^ЃH"(S*4ƭH0|ƝEX.p0$Y6]6`vrc$~~*OKwf`w{WmKi_}7/7Q\>SP2W>yywy4r͜J߁NUYڢAϿB6I6g++XC@ݮN|WϞrx5gJk|6) jm۹lCӵaҝ>ck}XB,%{z bmg=kW.c.8Je2e!ZWS"=W}!:JII NW::Y^/PyQ'z.b .%|/ aIn P)9eF h$2i.ڔe=VvFf7H]q{JR 9y zR9DX!u DΆIEXanFuq6¬z6,cNH28y0m%8veSd/X"E z g+@1kL)pt KQ|g{g:{儁 )DJ#2CoZQ{Apԇ}H$dґ@-V1}!ynR|w~ԽWVC* 澊@%2HG{肠6&]l(Lb-9{Vh B"/J1H6i.BԽ⏭[ t"؄!JbxboI 9YGٱ)cB9wL!d `lSقҸ$132zH]ZD5z8;N3%Ql)㓈dE8c r`"cz-> ]Li0 C'FQei*)l 5idxNqW&28#Q0{AB܉"9+rB3(`U %q޺W`c;du4ʱ(ؕFbf+( g_jR+`{읖QEj?+rV!}΅}n_߽̋7dI杄YY. (f;+gQPY&J B-'-p JX\BfX\,J&A!vqa@q$au`=3KQ5^&A`d B!dDXm;ڙʲ7%h)4ި`V4o 9Bۅt6H98A&R}fwd})6GNA*փKb`p&#Ton?Q v{y@tg)[Q!X7@'}DߏGo>q:?ٿ^rƗ3]4w]nwϟ=k ֤7'dژ8G&-C 14XM@ҹ&Έ:w|6e8S9%iGS21yN;'gCc`cCV[JVVJuIZ]5 iNCJ;N %LkL;T5>4CԉNWF5; d!B3m/5I'&idLwQɤ-D&3{os3œ|yhqn|K<=9[.7frVHNn5-glU8#(i7K90L:p=%F>X7 ;ǧQ[ə4mFYwfms-(ڬn%IPhބ!:Kyg!قSF2d!IGk"ؔdiĐ>o$&w/#w[LH-FzJ+YlO?۷'Q"8]ךtkH޳Yd!I62'$V0Ď0YC񺨷ob2V>I\DF g+!oDZN{'*jƵ`"miMڣ&}BJ^jfKMB36p̽R 3N$ܓgU3R-'ÌH=ClՒ}hxiz"DNc9$*xW40߽\P޼goD;?S›WTS'qѐ|7Q~xsL{js 3OlR`ç?KgOz&*S^_j +ʼn̡guϞL֚U.iI[ʐ>Fjq: u뱺 ]\n˟{ܓAH[:%8I J#E5qS9̟W7!z%淗|Rr` 9ڒFxl`j]xu5Me,_n5 ȩgy5Rh!"S/I)ޔnE5Ƥ#sbRKnGO8FH`TP%eԫuVaʜi`IؠxL}D#'pMysTቴO)hהD> b!zkU"bfcdd|8H}|vl]fɇҨcFF>>vB/pd11륰vqi.F%RhmkGuJ<|1"J5x#licl)g/ءcW_Sgzti$ϸ{ f4FmtG$-Jv,4BXJ +Q)qm=JePppZ:&/-ZtgmV]d+fFXyNS{eLS*piRLJT)IDNwѥ6ta[@{[<_B$ /4A}Nu)Ot;>^wo®E!]e֥|Pf{lΰ jW8cMQn AtQb*, Of+1pcKf=e}`UFx56_sFz6T6^+7Ô'?6dR#ZV jDc#xuk&ֺM!!rm)/?c8n`Z1on:n̈́Z)$O.M2y֭yfO18Z ;D ;i 3%{[ʁDpW:Ac& FmHjУ?a{D V36v?J:),p[{L]peay`%|,A ? pBpRIԁȕ)`A`Q#BXSNj=Gީ$k-ڢ?߿9FT@er/vKלv4 kJ jќ*:G:_?rǍݮqu%ag>xUoF~7>ggO?+Txv9%.L_" ZvJ,?nj]TmY/Uәk*kVB"#7kְ߳\zPpBn|wtOQ'{F_w8|?${./3lm$9;WlrK~E-ͪ_UHփ|'B7aH{d!|# Yj >΅!y,,4]i3xMYf3,\6F3M#H;NZ@l]߬7EY0pETƽd(xe1^;nsi5Zage9JP 9`:d:_0W,=}X-wNign>mB !`Y_HtA=wY z3˯\`y]㧧Ya 6j SmkXlmýمAJmDOwns.w퇟>,[o ّH $q{ ]*\'2<H:LIH='R)Kq$P4ӀuC/K r ` ''`fxi?N[2 DIAd:uE5ކ5 1o ^- %{؆-MJevH)p<5uMlI$ YCXTn)!2tX=[ITȧȏ\-DFc'g!:1]6M%R=*#%{|Xi/jf Nm%3b*u_5WIoBuDYQjZ)0DF<'xe)ÿe_f3q3$IFxC*INs~g+6yJesr!r8ZtWBw5-jtQŀfa8yЄ|')uaqFo:Q<̑ }J밪D/,7%OV\j w{R=eX:<U\ $PPM9QїY|tk}WN`Ⱥ`˜u I$L(\j 2THqi\8k1w='^I5&D$QbVz>Gӭ R8leKҪnq5-JD5 ,9]kUULeORq8>%Buw%fY}Qb=!& .yWRiR(B2qW} 0mpGBb -lp k~~ n%[fn]sx<} ާ0} ާtYΩ9Lfi("i]F1 [v$R(d1qθ\X`r~Eo)_=Kd ߞm><<śH*Nmtp4&:ḩKX-YIubpEIVz~-~UgLRp<\fp4.zD5mHzއ۫Gz]>Gr%v{ž4Or&:W +5bH9ySݠ^@PuXzcLRzIG^j5뛋ϕ%-v*:cH-RnPn|,fcNŠ׭o7ے+w1[>n7oϘ |jR>p(j_\rQ,3ogx1n0y=Rj b3t89)i'&9>jT'ׁ$XNEO9z۪{ (*A8XvƫJjQL!DB8TA ՎĜ xP zʰOJ*Eжi/JCق{ S.t@-~ܶL*ǩ ,`&ԡxqS,5; œ5l۪{ S23JԩΫU"P"WB Qq2x8r4턆0);D Bm$D ]٤v"K@/(H۩?e\-49[*BAT zm@(Tc( (%Bx'BcW5ݤPXp$x!4&pHMwI#W$ňf|h?1|9|zgK-Ï$|*S54>'R_9Q}׽ñ2XzrON\+#k>L 3) X)M-rBa\(d76rk{rʅX5"H[0TDY2+o@*e uTF.^ %MYy| rk4sRTGISX.)^ek3-f2-\i9Ar8giG7y?E]Yd!uǑ5V޼ | ˾T׌9$+`cSdFS歋 tE)e<شP_MLz:MWN_6IhZ8 Qva5ʆn~>ڔx>tvew>mRdOgQ=z .uZ3hJ%?uUJ'2 h@UPƴBH! mK zz6ATj{vŽWٖ+ƋF mcp6l^=Ucx) \7 G@@Ԥx G"a?\5UY{Kvܨ!S<PPwxYT3>H;ϖ4fWt]+K(CoZŁ}[2w:DlUL(-[3(^ɗhp+KCGX v8ji]^ӐL^ny\jy7{ 5U}n2V7~YRxf?͏7^ml 7/-xNWHn늇L؎<tn/ݺ,כondTnKw`nW!ճr=ފ)X ŭtvۂI\DSd2Y䝒V!4l&(L;LBѨE:ሳb*bLN3̊c%J}ϊÑb2 ϰ@GFJkKQ<ۊs~v׹7u]+~{#IxuR)dҨaZi5̹Kkφ7kgWn^"zbBr$,h롆EV.BXyB )t}~B}O7Qq e0Qp!ʩi19ycN)( 8Y%b j-q,ژ\[XXˬV OA\HU7^\uyÚ#ݰҒn?=yE F6|8X K?0@Kuݶٷܭ?}q^\Uǥ .&Ё;cԚ Juj;U#)[kٕNBID.!'$:Ua&١+sbSSf1g/_6DTBc:E2Q֩Yr.(p `Jcv*= =)$1t (u( KBwy ,T%sݒᘝі 'lT.Oƒh: 3r9@g d;;MBM)A(zkEY/ ̘RH2aqXmjل/u4T["u6Der=:Y KcooR>K25Sؖt:ZFc`]EvEQͮKLZϓ0%&2Hz|-o6MKy5(ꦺVsS{\ELz? spJ?ló=g5]EvmacXF3g-37DOI1-:L&Ejʰ&UΏ ǂ&""Cea~<./CP cq+ o\~̺e-.&mÌFdKVܮHMwvV7\|LÂ蓵lEW_6r,v\ڒy~:U35/s3sD QfprƳ[f`Aѩ͉FERJC' H 6RB#N-kGZ Cj8`ʾ0;T=UKBa%[o!z"$sҖE8gCur}ˆ=?{W6 aα#};6evM%d }IJ.II0,]Ua!?̯F/$V[e|_i "R0p)jbx3y2vqj8:h{b!4QXtbiWeJFF,%©JOj;3atfk[-jC&}@Ƶ+Xw!q(Fymj,&aC⪖xP@)Ehge(Qk׶r,H fH}Xzx3O^lsn#6Km;\ j.⽪Ð"sjY [jhu/Ey1*ŝuL=m۟~F!K` HbB-@cSkK`8pcR8*:#nI%1xd-lC]/ [۴z]O K7~]!p0vD˸7>|wmnDSWoۿ}}uKnMJV_| }n{:X@ }WL!w5\D7 z]\6LM^^Ȏ ]_O|Kw?-73ֿ^>Վ'gzzsK{OnZkri=X.+=(3s~e.lJFb!\tuDE),q5[IE$ d$fEJ3h%1:}e܊P2A12ħt_bJ}<t{g0c MD75ף.ADmZ&2!3jsJIb Jc*ycvq+_ko%ޫʭvܳ.r_uA}w! ?.$`g :6=u1[oB$s1㙵w1<#2,>fjĵmG7썧5͋38e$hK![d5'/Sx1-i5 `"K9s5wgFm>_ ?_MwtS-*@{}t)n}]pRrN(Wgfk#NYwt)w%!q  V/JUieL V宍9#Ie&Ao020|pz}hg3Itפjsj߉\4[h*28qD/=]s1x8qt?%^Og F)?q쟽cG 8s/W*"?SQ䗯" !L"0+Yo3yV hQr:+H9Ӕi㢘\E 1-,da_2|&Gmcn6^"ߧ0C C4C. HK-W,Xb@-XHT b{<"Μ ɢHvUtyt%YM bZk^s~?OznJv|\W,gf!lA ϤEؒg.^T*{eWnMTy< xZKr )Er vE?G$FO|M a65gJ:' 52xȤ+B⺢9_Ga-"UJē/q5k6g鍞v R˴Kyp]x,|#> xs;3W}U0YvG`aRaJKEiDbO8$*͐x>C"*ﻐ%ÓL]5Ql RTo—LƙDS¹.g5` "оF#ABBYh]WaTTJGQ I5-(qBfl-t<ƳvŶZx$ޝ{2&N^h'~ fe]i8 [XI2Os+m$=wu!xL7HH6-բgWqdKM% {WO.t;w[q=_%vEV( y*ZFg?lЭ,N6 ET&=tVмV׹l d058VQq,@-Lîy̫y^x铸gS{=N6_B┐V%Z2$NZ0I»D <8 t~(;3MI͔795ܜsSUnJ%nRDVi׼ #f# ̄6E\%&qi|N4va7Iҥ#522qݻ>k;`@ss$'75^7/v$ܵ6zқ|wgO?|r/[7֌5q}ח9z+?UU<|3xvVu֧؟ ?FZm0Oz{ȋAK0l@0VCg֋v۫-L5 ]CW6T{qebu ën_޻`[yδsT6-9pkUnAl?nIR=5eh]IZ=p#I>[fq'"V[3?/X=!~:ߎ%0CaVfx{G/viF`kOǣ~xX6yCqCP`rRIv8/jSJ63Z #cQn?i^M]oMV; pA}?o<@8>K~ }Xq2p8mܐ)˨I7zwF܌\lo@-xgkN x z'+ᠶ$v@$2S}E$- Bi'RI_T..my0xn&|w,?|1=IAH_HY+}ºgeJJ 176}v o\}-=7x2g6y1q.dp}|{<ѱniu51aKQ4i fWwU axyC~($ۓ:#ƅAY'{Ąs<$>⬍l Kf!ZDI髠߳SψDQ#J-m&N[锯 5`&Ԕ{ǓR`?h ZԦlW#r$w.88D|Ѻgc u/0@m9L D[aF2̿x -BPfɘ z漜 z[kar3q|Л^δuںrm]mu&^`KMqBymS͖P W,%% ,[ʧN *'bWYΧF KjCFH(MD`. Y * JrLc,Tݡ{? 캱:;p \ᶮ7][FӦUH2<˱!B3~>K]Kq.nH~+60n ϳUxag't7unV[YB')~@}ػ~hG1)wխ g #'x\d&2vv4rBhHmu?Ö 1nr0{.Ǣ>Zd7N'm' Zƽ|jB1엩 ^>S0LA0pu~=}'@z }V2 Pރ"1372v( \h5|4"%0HE$ d$fEJ3h%1:}e܊gI{7+k}! ޾A&ɗM`MV+%Ȯ$wmXG~h搄D@#LL12 F\(JOG޺h4Z.MT1rحuU1-BlVY1!D  oVDi)h0 F_|c|+)!yiZ[GALLhmBFObv0_$4Xys ~r ş_4Lw0ܝ0 w:?%k4\fm}qffEh U$x!đ $1%N)_#_)ɏZ諦GUqǢ⹌im1'Zd`A- B {b'Ϳ=#Xc֌`)A &!boGR$6EHkeA"DEȇ A Rzn q` )BX ǔ1Xj.qؠ-Rݫ܀ hdvF;:ʗD3Q쇁&( eM)ɧ8Q ũe1O~1Ua[ 44qh;fWErpOeVn=C'svuOe 3#IUĖVn":V~M*(VU Itzi~nU9 @6_mLqh -ԩbC1lNV*9!Φ.ՉUvJk-V@qmsSisH\tp|p=[*vB faSQo 6b'[4DZn%Q2HIXtCDr, lhkJ@=z,Uo8ٳZtLMGPZsY sDȋ\S="h.RNZ p5 :RW% Rb,[e`m5O4 m4_X7Hm}۪deKmJ-&4K d*S:%G`'WsrV꣌6/TZ\PLwmPa\8O:D OJO k2t#,h L!\mc#[ ?Rz:\>s?Md˘9dM6 &_OmUH0ore0`==ch=@•Jey|Moi(`YP " <RӤq 6L2ؓ`)A= q>VH6RP\Ojr$Ш3em'wRJ jv 'l+f$seDJs0׹s~Ni_c-̆EGiE;"oˎ),wZ<Ь}mz2"W:LyeڦMa8.sP֮^ŇrV׎:MZg)YOu6yZD ڐN3[tE8p&Rm|dqmX8C2Cx?eZ< [wȂ)h\po34nS3󂹟4^xMΑov+)Ƕ' HQP^ry!8]9~]BK@=uNS.T}Dg-]!k]LXd sJ8׿Ls PVaؓLÞC=uNkDI1p9uN MNaiTj[UwJʦdB?{ď( ~GvnqUZ֦VeZNɡ T=Kq'Ǖ7N !TUSR (._NZ6O @̡t]⸢Sjl~)t(Q:蔎;KsJ,2\R+ŤAju\I邤D-B B_Gi)`f(R N#|idM?\P& H(EL&dYԂZvA:3iCYfs+GD &[fƲ.(|w.zĴ t@d́\G,|p}EVaʊlVG W+[4w鈭)d/PmgN˹6ӥL3\J 57o_[/GpK\J=!K% 덇)ԹKى-+%t ҄Yo^CW]3$~k "1OVs D)pW\MsK]c@9|4xkԝ 2&G?#!Ͱ1=O֟H>&GVS8&r>4<_V؁糙 q =W]jԪ-agRtAdRH[ fU'Yp(W3^Vt[,){֕LW76:RbN9u4``zM)ct],gK?>dhp:SSJ*V;(5L!vqd^ ƕGը֖E,`n{ l r8BQvW=eNFCg_6: ǁL Gd}\uPRB wU0ȿAIr(Qɹu?5!]:-U]] i@pvmY9r'#8o:J!d?GeC3ÿUGF!1q,L;#.5Ycì ¬eu'i] !eL0r(}f +yX Qpx >Zs4 4ĺ/:|?cfrVr>+[KXLcʀJvU_aAmQ2KӆW!:JDyD{h<2¼bXtDƤTsL]fLas 9B-N^HB@?2e60|FudDbT-.X6X + o5`VɆPt4LcOgX2C~{,cV,(.׆]sNeq JhX)9ps2W*$[J;Y(XV"t<{^l\o`:t`,\f2s5k!8l!֏/Ѿ[պ'uym»=UW&V00UC߿?G: % =2~~6%:yL:1Ӌ HL Q(Y-sueȎ5C~ɰgLGQ*0~=3EƳrx$'$xhޥj6bbWOt H>̞VR*hs-peg@&<cV^1T' x=p-L>Fwh5kg0b1>9lмI642'3xj3ZF%< G1Hq,s=|J S|O&6 >&KZF!%,ԇZ,aaIurX"%6b=\D%$$J?d9.qgIĢ.3x\|Ueͥ9NU@JL R l!#(vRn,Ǭc*Z9ܭB!1qwUa=7"$/v9fh-/`VR.ΎCmظܑ^;_}U61UFJvߨ%Қ4 TK&Ta9lֵuR)[e] H_@YㅔHˬ;Sj1AˋQΐ:0ڒD~0'kZquZnN֫ݮWM6B2*Ꮖ{vYۜK =lv %R[f6)m,QLɃėG$a>0E/.UgɔDcLeR4ʨ0F16$4/;|LS a X= eK>|kn!E7~~@?"y;X콟n!m;(w^P P&m&%~x̓Eh>_mĀEiB x{,S2S~`|%Nd~dᇿ|¾irPyy{Rh5! D{WB=}P;5' 3,\~^BxKl Q0& =g8E H\BE>j$;Y}Wiv6VZU?M Je10ƷPܣE g{oU\F\ⷩZUWHAq$ AKf̜oR -.*tH߾JR.7(3Ƹz`3m!<ec1Uy;mr~Qn*s5M&7P}CEv"wº޺wm_!$@h{} )ӺAIaP$%3(ԛtۢ0ZfvvwqDyb);*OjVΦEURnoax UzzWX_EXY?Rg_ʹơ&FU{({ƚXOZo'Ń2B?s8(_>v5PIBfHzvTTņeJg';SC ޷qgJknN0LYL;/ߺ!ԡ{(θ%c݉N*:a*s-'N00y cdv̨)h[pN3u"I{;m^.ch;ݮh^Nc7s%f Mz F̞ݑǨQ:FvdcHzCճ+= ǃEt; |xOu,Hc>s v jG,6I/Ryb\"AFhYD[Xvʘ-oUh?Dzɨ*n"&ⳳa5! x3u=Q{f hE@T8cօHۘή4iE2H .f]x!$ H̢5͜;U[0K?I42wޚbF? ] d ]9ߒȼkJM2:5[jZ A]vh ”>xm;'2 *@edEP볍vcƥeۛmZOHb. Iv(Ê(XE3n\`5bgE\87JN::ωGg$f90`gcɘJ9rkwP(ަ@P! #eɽVtY@rg(]A }B; ٨cna͏8LosG3<*|3X9dv9]QU3l\Pṁ]CPwWѠt_52ivsei8 ! \izT caN1<|3/0DȌ->ghgp0G hfRam3k 0Aen0\9H/1@ii,Co1l o;x_Fq'UH'Hhqe/&iژ_%s Ӌ cm^yAT^b;X9Qzŧ̀6 g09l6UHپYi2|-dP:7az(4mc29N|I)M3n+kaE:1G{t0aEs7Uhg -S@;4`XAs`b!ް5??gr4N^>uhY4Mt?P^}nK}'z~v n TFyuVms%'r!:7G?3BwFa~ $3#`€:nҼl)h΃= K/ 5GY^%P,(fQnpB\Zz$!u&+/U{N9pR} ' W2ISOĆgF[  t'a(p0 ۫tX<HqPY T#*}|j@y7t9'S,8|Q;LHg/B &QV^Tݼ> bPS n**8" _ut{Ca긆ȓƞW(n01}7YP""͎B42zџN"~? >Rddw\)àIriX|'[N~^91vp˼}R&HY[&2enKWޞ/'-S| ^ 9 16$< ƈ~ >dW|#>d̙w@׽f4ܼ[qJ.If EdsX^(o&@]_K$(yKՐCtOvO޽5Ȟs2/RcT;y1\^t`H zHn4H5nOp9K|}ۛ׿\|yAm誴?{wozw7n߽jQuYQs{e\뷫!EnJOgSx v?Qj+najN[6_WDOtU\:Tϛۯxf`۔$OY㺭VL2HvY+E(E$Cu f[/.(]n*諻W^ո?X5t7Rèw&UO[U?H{7a6ϛ_eM-l $Kx៻zg> ,L0 ,L&awՐ# X~,Agd~X zXazXQzZ[Ϳj90 r\KʙGFo?ʧ^Q*_R(c3^Ogn|:q>}zW/-">7$V}} tM{7#+Gj`-kae2Ԧm 鿲+/W &a>}w~c̺W8m~Uip֍em~|rŭֿi ѭ?u,]K[[{ۨovY~iMJ ~Q.בgX7KW|@Z]fmu{`&p "&>PoG$C8>,PI/$B@ PMnraM{X. $ a]!e8Bh\ضT9OM{e ?u_;*Wj{Dgd&F}GX&PbS]ozm뵭׶^zm뵭>=l ^C-[oݾu[o7 i/~Ku$lhsr]y j䇁Cj?,.2b[ak8"uFxL՞ Oct`I 5c)!!6rW!ABQR62+-!SPilA2NfgPfK?f~}‐NmP5HIjOݣ[3'a01TxV~Ҁ 5 wMfpK vo5Tq1ļ3r`$cO :{rA.=W^ը7gI8iiB/&K?;3CACJ2hdC0 mH vZ+xL! H/g} =g`k0B @ƻV"I܀|DVs 4:-gRAX!%pUPi9 0L3TNHh.H#WM[C>:Ó` !`vؼE9 X TNcQd+9t:6G9-I9 Me,/O/Kg3UC <%$mcϷ:G 823yȑ B uWgzW#Orh@DyojkN ]aBa(̰ުF2<z\"CT'E? N$52.r-T OgHyrFWN*TcNky؛s>"NwG~' n)[@}juw ξs<[Q XeJx)~p0%_ү9qIg%DeNcx[|]lFp_Ad](C32O9 |`(,,O<@o̢eH#az(< MKY $KLќa_nBKDjՕQZ|2"tfwƄ!>B L5oer :X0J;<>yI@0hz8ӋhÕq# o]fys*JFq8u+l^^#{`{?2Ѽ7(xzmmԄ;L!|z(`^<17`zL.S6hc)I͍1$;]xs!>ͮy-EˡX+FzmgԼ0:; s_LLn-v [&'q1ZJIEדh}Zv, YcwZv)D5.lNم0s^nfeKNIP 4TMag3/NZk7_Xך'w۵6VVZXM1J6v.X}+v&Lk/_w^ 2$<"NB[`{lfrK>ɓ\赌(=QoL:u21g0%s#۳w!T N"~л٭(^hj0kn/"wLYє;+wY r* )@z*q'>biF@[Ś_T1e ׄFFPO Jl^==Kg=+jͷ^ Vkҷo O̫TӞǠzH13btw ?|fCWL|RJsBB mIAQ=;P;C:a|;L*RVkя1P7Hx ѷa8 kh5VE_ܨ;0is!yeJ[m5Qb ǎtJ!#0 茥:H zOKEalցOlԢQMpq-cJhtD(= P Vy yGJSN68X5jn7/*#=J4ʭrZV^1 f4q3P/AsözVlGndj  Q^=OuwՑK9?o`~`B_aǓxR­?ƒR f.擱Dux|̄Ύbc#ƍj~zoQID9;FyAS LG-I&: Z1Lh- bs懊&W춪s.t5b|0Ywj5Q[Tk4r,j #{[<2##%q:9`Niu`gp ıb0 T&-G孵 {IR ;>n؟ݪnf P'E~v8%ݕex4h|IC\ }!}X3|gStd xE#+CpZ.y 2RS fhlD%?U|uQ'4z,5[tLh#w{Iv)\`N{veS•vUMト'M+lL)BXƤZᇂFybҤdY$M .&nX8w>)-^sxrqƨ4pp.<{,v2qdPܰ) (U_&+mF [sx!Z' A1 3_ϮvtisJQuă^լUzÇjosXy(% )*."zX7B`4NK$RVLb2BR#>-R M5V[ZqHmSeks`JXWW/ZCK:)n$ms~kkQ,hox*Hۛص=[_P<Lw<98Éڝ_]q}17ռ ,_K}c)G1ƈ;eS,}o"/6jͲ E\ ip E'zP5!RNa'Xjػ|kNmYz)x2^'')||Y䕒PS_" XuR>zޢ~@c2ϟ"][|HoWJ7~Bl}jam寔^SkկV\VDWvdůNڈ\ 32r0@>E"˭~o` $'^ 9!x! Nv<~q=G\/wW$!ȕVF_vx dw')܇#ysX &u$kxѻ }W U:t_ؓe#M-%`=<(%ƈP8vb5{߼q~?J(3<iZB:N9*N%"NDRHuBA/RT/R ɐRϱјslN0(D!׬n1 Sl tEf:??gav,1]?2})i֜ T յ;;2-`V=5mHD R\r` L/!W筍K(1h\;\;\;\;sm,&lp4 e6*%ȸ0؁c;ж]Uڶoq)dks!V:7c͢V4v:N<3p uT%ʃ<44IA ЦL)>f{<m>>0[pG(OonS>/}x@rz-wEN Л ɽBw*zX%²Ye VEPd?(@y*Qywmۨ\\1+Z?}7~2\"E%(m JTPAփFմj.t~i8O ޅz~j{]lRЊ?9hZ G `e{bOLw9niM5F!e!!L Oϛ A/&.fQfQf<njLQ"aQ(  Dj'FX=1$F0ET&RiPHŏf{8-8|J( xU H bM+ ecƥ,dcFfۧo[^ܶ5JrIN:I 3R(ϋG^kjE4[poh "!hΡnjd0醴/8Y"}2ӎP&as[#!!)x Xc?`} J:`[PI3Mg괠YȽ6@\eS1`90D)%Dy.IMYynwiߣߣߣߣ:%'I)xRس<=j8b"F.ʣcdI]5rۅZKm?w9oZ%bĢq!->QSp B k(Sp;or+r=U 8X ؂ ɕ+ETx˫.6jF3$3#ϭ6zk! G>v_Q]VC}㪞+$34گRb˜g<=sUζif,3~baH'9[0.@+rVa=4[`OD:pn$&)Cqd&˭ E-;ڍ~\ܾ8N9Q< ~4' U; EHV(ka ̝dBX~`J~F6*W ^5Ҙ؟M0ǽyu]dq3Yܜ[sY8{>AjעE:6g2m鄘vB\}MJ643m1'g4ZT##M5ޑA]9RAE" wF娈!* E3{wkY`7DnWױp:cy{n2 S*ge}hT[xo5xS p[vh]^=8W/c-KP ۗ1G/;Ճ;Gۗ1GG#PX6ܾ"8>cIi%`)gۗ-L # ԣ$7Z %7-ǐ5} 1g!L`ܹ"d_pt Z\+8D3rHP) w&"a*ck1"BE䉽pMsS@=IMR7)#ǩ[P=R"DJ8h~%tҭ7TpS,1Qfx)1-Q$\R JI(8 \HmG3rj_VӘDϖA;,n'G#vXR>g\rzNg0 E9 ZZl|ri.:ד(-aj5aj=:&8$MD\qã |z;)Ւ7k1JCEc"Krq-,[ݗ7Z09^xs4sZ4 9@Kemr靑/㯗cMZotNko<7 po Cth VJ8Gxt&(e<.4N# }. {k:͗^+ZYsbf[%Zƿ`d_VWO2Oԍ~0h;"(OB.pBZQ9xV`1K :B`%HG"h`Q&B5:A$&W1ijLcGi`s@ (#w1ziAj"0Tj9:& =Cܡ 8m]׃gܡ cO|;tAk}V:oN/̥ 1[O sZדYI>lgfr X\ӋpNMϟbbϗUEtg9-g.g۴̕ԦuǗI’ZXr钞{ f/~,-,`I.cɰ$laIm[,i9]R?ŒI =iڑwl%Xę?#ؕ}F,1ư,]1.q,wʒn,e-&a Kj}%Z Npu0nqGli1-}uT)~0WG;E1}n`bmaWouPζpf͘'ʝbjFuGNGCI;1ꎚp,I΀ E!/HHm$rx*DH9l"JrC W L1n2g@( Zx${K3g^N[ƾ/o#ʛD tj$=e' ؃DmxA02`1LJ>^ة"LDnuКિKq@;"шfoo~氨O]?&9j>n>bIIj5C&_TX[Jg(UY)J?'wh$jU)_D+h)u3h-QmnIWN4y‹Tv^*-μ%a BP s ELMK4WJd+۪cr0ꉿY:*{&|sJ|;X|2-B.,  zӛ#nbŐ7g}Fkt@޾zrq B3unWۥ42STS&ɂ]xsy Jdݻ݊.;PS2>+lhocq>jց3׋AzD>p4san2ҁ9҈=:1E"$ qi|U-鞌#6Vvb4,#(1g,xIK ~b.qgd, <@` H놉{p S** *-"B3XHP*6R8-8+F{D;5H=H 7kΣ6Pk 0sBa8gs"̅90p" 5P )V@@tsuC]ms#7r+,}I9uC U6_Ur[v|0@W+*$~I1CJC1Z-j~݆v MNɒvel\ zh )$r$ ev2**¥NJz>QRj#rڽC0qO9>%2q*.sd.4W(S\+0>%8L XҚt5+:G{ lGtlx(z$JF3y "7?IsK͋/ m85TC 7~#(Ӓ{ p&y ē @rώIJ$ H6 poXV(2V EA@cR 'hόp7Tkn`GCGY n滝\r fgN>kmږ^n+Aw|8Y癖c Ec0)=2A˃Qw [̲eakK$ BXx!RQH#/,c@FL~(9} [Ji4Nⷞߚ%4#G4{󲾿t,3-~Y8ᖳb?Ooս]dse*n"h?bWc|,2"R^1&` s J`&x9CΩ af]]i8>kp <.fG׳"~ 6Z:˞;r3V;PߠRt ^p`V~|{[~/#[xX/q7dueuy֢l9ooO6˯@({W ӬM *dj:˷=|}U8ş&d'I87eŸH3c#:(O.d$GTz(L۱AQ<6 w)C cty&%V,W_yRֽ2>t:3~(~:v] oϣp / ~)^볃E‹r b(b1%̐E>M<_兯*~nnr7/×GUr~CGpTwpN/7(,t;h"tqY5^װPtWA Ne[pF:-]GF> 'aMJ sʛ& m_SHXfgX&};=y2Rj!H<_]L2u} Le]+ՅveL׋w7ޮn~#( ^Љ )xOS#3 nDw088P}z); ݇cU6.߮R-jl_JOa&[|]U£wXt`6P}5y ;V}2:-=&t5IU%9IZ߮<1LxX}phK1= BI5`(*HOT??jܡ zr䯳>Sim?䇋2=5vJ@ TE-dk?3r2QPJ#j..,)*ր%l^GA7*p{v>[.2|5i1#)[}kV'q/S.k;DEJ1DdDU챶 zr4bCW.O|@ ou9p)I %g1bN3=i#o_׉^Z\_.s8[ݪKKsbm#~d}9;uNFA.,UgbhN%*<,bu ;6kO6z +Gf&v2! lZFcx9M$N"ش4 :?pAA8;x!DZɃЇB{Q8*moIusWx"|Bǹj-Lg 3V3f7͖67!7sjbrռu)q]J&ԪyTzj&QM;:Y\7)s L ,NK#ԯ~>&e,wϺrׅz3U;H3|, jS;e"r6\|(fT/2c8'))/oJa&rA䦧bv; n}>tq_ҭb鹀vl2[/mC5@b #DfAYLSPha'`8o=uSpa^Y.7OebՉPbcfU)&\lM8zgxT<2l@0 239冨e1e3&?^=OxmS, (9Pt'7?gKh6$Ҟsp]E4I|;r رgo2KFj+7[֟B}]&D-#LY՟x&g~d#d$/qINDb\`_&h&T;0s)9QR8n5jdp"g )m53*ŦC@[\h 46p<f\E 8 X+%E` FAnvA{_-ΊfQkq⯧?,!;VKτO;{ nNJY͵TlH+`JKӠ+ip#wuWG'(Sc,NXcl g4D`"xBC$cǥL{~6nwXq1ZJ s\ {OU8׺zpS$IMy`ϭҢ䴖tuض]lۯ︡Х!#&ȷ6 ӣq?$7PTPֽU`(mQAfXCЋZ#w+jCܥ]'j=:oUV6@UV%4YbM+|#)| TuύX ߗ.Re?C&M$N ƫu{g:e$^)5 Z597x .4:hI 'nX/>7e@MnK!i<;5s )ԎznDnsڜS/{*\Za~%^C;A9 o"BQ =yӅ }D éKSb?9b@!459X$1ؑr;_>Uk&87|8N$R3b 1NK))cN&hLՒF)ŏQBgLOɕ~6yq$+]RyY)w )T/s9\{ )#=Zn\tV暐XO#Zg~19zz43N%K%MޓF,MR|#45X#F r-M]436*XQ%ܭv^,Y- yҨ ։6\=oѱ4*%6n@3A4dHX;ѸZ\5eQUbҒZL%4Mi2%(Lʎg!K9KTfF۝ۍd!Vvwm5@ߦ gd]XE\^^;}U gP(hs;ڶy^uu.'yIM)xJ(hxa2'Y<8`,:gVXonAՐxUx <8Ő 3RRCFq󘽕NimӢ])f)9k ڽl:<ᆄK"Soͅ9]rs3b5FKjcЧ.>ެ~I?$!x7ٴ&4Ɔ&LB:^҆vh^Ԭפ*檩zq7'ښ*RKaJ0~hFu?u7~dn?TCLck򡫫K/ O[]kaCGgqqyQd+خQEJ0/q:&Jq 98&a+]N+Ta9Y2XcMï)ڄR'p WpXR9ǬXZ*&U(Q!$ ϔmrr8Owl1m 3_Q8$;c˪ D!uNK~g/ gЉ<[|v;Fe )o$s_`p` /g`2K/etA#X#Ob"ΘOU0}alVe{_ ]S/5UV'Xi^K1AGW#e\XMs^9#p6 N*̈́=XYX2A!8,1#V{3}9VVNkXN"JR SjǚYRE|Jľ`  rZ'=CrbnBbVhKfgYdS#mU h Pmej2nݼ9DgVPD `^wE`9`T ,z;e*G!U=qe}T6KeBDgIφ`thGI s\9b6F1S;Hi\+ۖgd"ހ2V4fgz &Ή"&ᨈ/E?Yϳbr(g* ׯ~.p_`BӥҫDc) w/7yG[Ҵ=1)ŽI ܄zr5NdŁNK>xyw!"eNgn0 L0Mo^*ҹ7OyU.fXŽWnez$&t^kkq?A m0k^_V "4\L{ 0O-k# k FխZ|7-J4ByG$O_gw7@i7jĈ>a\ptc1Bh(z(VS9\1tgy;[s;?3Qp۳yh \צ[ϣXt$o>`v@=*7s#Uk;aR'˧93M^4X' ݤgYDr$`oysza`&iWvyvq]=_]g7(?9}_r`y{f!Ӗe-藵[젥{|,{l?@HUy~YՃ%5F }/~ I[ǯ~UyНbyjJQV $ehR"HDGQIaP*~.40%KEA|I9F>r%P:Q˨a<Od9xj ?t˝05nR~9+#(Ucf 3`V! rfZ}&7됧umҎ6N#&0ZSBkWO±Yi:؜)ȉ$cNZ4d#{)2d[w(q9њv1ٖٻiʾg'ɎP<шN,tmV3f=w)B0뎕ӊFM|,Gx4vi:I^t}@#B6"}vGy`HuC捹Hh~|PJ.$2&ٷlnU6Sdw5҈1Nҵk~|M:I6ttѰ+}EN H2T''Yi&'BwUaJd$RQ.uR*Zcr}3zeL'jȷn^~J t?42%;L+S7_u.7k^wUu(tFi@%2hsbEiU/F??^*ÐW8,=#JW?1C,H,E^Bzm*KqI\iRbNQ 炀9 ^HJcQh#r0(N)6D3$%AD)$ ٤y0#-:m* (D!jY紤Gf;eSpIh,X= !ąG( y x.-5 sS4WIs[*`HY(d(VΦ RF,[%h$8U GH:V%sdˆ Am"PGNkhJR SjǚYREA#Jľ` W3BVC 0l1 ;VrAa$qc1",bp @A/9oq6}?b[mq, vÇf߾zx'?L+..AծY/&7xLt?`dܷL]1Ț~ލ~h.fMG@4ZIӔ`/+L {0E2{/k'dWo.՛sa\ykPNӛ,;Z_ $z(_\Opn~1!lbu=/&_L(?[0t~1z/&#*J8)_X%;_8ipf8ipf8ipf8G^fS/Y?L. E$fq?^]ZkWJaƌ9;c/-Q+)\Ы];dg\8x\Źդk* *%+8B3$8XA͖Iv\R%NhCWfq?1'Düq#l &-m.4 ϣ nxH҅,h^/;0Q(]cbr֋]6mO?nZti@;8_˻ٗX*(=KxQwv54]Ώޏ~-^>Mge/W +])mz^ z+u2_tSUb/Id~ő3 K6ڍQhFA mv69mR v OUPŐ3 :ֻJiB1h":mnr)δ[x1-ɔ$o(/\blʴ;NTwˍ4]F3djS~n46h|MsGnjr"$SxSڹ{FA mvwj":n%jr"$S"ڍAB1h":mnr!ܙv / PŐ3 1CM^bDtBhŁUnlne?jr"$S>pC-&F1^8U3^طv!!g.!2} vAbDtBhrƕs*;3 /PŐ3 Rw6{->ĵǝ BǵoZS<@1Qg1C/a&te7{͐b+ʨ3yjY?B!oË:⧖z: CYl>fkJ=C|O 1z]^bAX3IyryqC@ԙ .Zy8c9BΘwmm A`ks$U@UM\^8zHICJ!%zqlU5X6=`6_cV 1scnTǗc^9d19&4+hr̰6昛w9fMecnsjB9Gu49f%ksmQM_YKscnT>BX96ܠ&L1tt9fEm17 13ĨnsmQM`]Vcns̍jr̰9f5!st9fhψisjBXr +,scnTԮ=0s@9fF(Gm17 T13scnB"Ǘc&}QEH%/Lkcns̍jE;GQ0L m17 RcU:n} WaZHjL8MƗ AIE0B>gҜqre͉ )bG6g39rxNc>O|if2ML/Apał^e?'c%ʞ`Pr.!]xryiKKK) 4tËYIm&I6 =<0 2CKY>e2 ymfIk^Z u IJc Yy[DGⱤ;hA#, @vkRV ܊T:"@r 68w p X]-=`$%3 1Қǩd (ĸ$7{&ՁkNnFyzJyXx|פnwu,B%BZp#CEbdS4U4zPxBGڸȥgw d!d]M!Rm\UPڀtDLY`FE$M1R-q-9!$܂RS/}qC #DQ#I*szb (2I3#X8gBQmBT!- sb e#^BR:h!tj!AЂ "8E#``'S Jbx:"x=߀-$ym@^(%:5YH`gSFuaZBCn0s.#VY@fr@ FQb9FpAF/BYBՂRDS0*o@qP2,-JJ3yPc\]$$+ƒ1cCHp_d|injzv`|:ME8)){ (.ubۃ_F q̙af"~ UL6 _0,85t>$Apqa tPL>&a8td?]yI܏ܧ0q">"^}ق`ؿٌТh4˝ɬ`yjkLFpiMg)0 wAj4\TP9gbfi¤&6 g4!-To9E/;j8;d߽55r4)ub9& s+d̝0g6}wUb_<6=CXPh$6p:ͣcOFIƏ c%sgcNX #qeXcҰ!0{3Pgkv-oLwkUo8qw'h8(!^8a_g o֞o@\MQhv E5.J?<P}J5X ^@I*ߎk~uOb#-J:z>/Jۂ#7 -.p2|瓚c^wB(-'!FDwչ\p5HzB &(3Q#P(/({ˏšJUfdavf\t '4:OnfVi Z\b5̼dMocL\:k2oGB5Zu5,K#yU0K;Ӫ&/%RdQGQן%eӯe^kM?c^y~G`LҰa1-3*L{+J'9[H6';?oL[~P;E|jG]IH*'޿}g 孶yUOmES}FBX|~iLExaVH~O*Y +-ZouriS:#an5"7>sB5IM7ꗶ\FV_[_5(r= 4MOH@HDJQHJKAZ#4SdJ_Y†`..pG>~zsSHZJK]F^_)]0-Q,CկUW8G4 (W'RqXFO;-΋?B/?{f:~y=%`{hK )_]M)7ygRgԑT?ܐ৏TX?I!9S_\LqkƉ91(wIカ&9l=P像\lI-'MEEI,]#bDK^~8ykbM}ߥtE,T}a兮Aj%"z@A<]GrIի!`Á](c斱Q٘o㺺14KU+:Qdlky$eR:N;N?DrPgF`H.6+knC0r.gY'\^"╦ѝG5 7M_g/g/g/g,'I3YPF#9bPPT$I cJj0"*b[ϻ /^fXnbF2{h([v/˾O˰ޫk&@NLr=e_E^ UYUBuQ\ %|HEII8Fp^P#ǩ-+d4XGPam)J#aq\fw<9DAX:9іr"NyfA"e0JC*fA$:;S-SRM8σjZje6OΎج5jT#|$+rG(mBƵ'NRbXe想VƐQ$icpa5U점:1De*u|M>XE fT>i=]ҡ&lwLZuZku aj'hWH6Ѷë#U lr,8 \Ak0ש@Tga8gjf?tl"/kAl9*bHZ'g,ar9uHG5Q(.7볟8rgwܦm78vt%|"qE)QǛ Ɖ֞JᦿLo= ugWgrLgl1j&,Q];2 c# Zsö& pxY&Wo߿ ibxÂRXeMb#WұvT1Z{SbO<-S.e{:h?ֶLcKt|f>ގ|岎 *6 |kR)& y\~d7xawXOOJUZxdUj.-!Mr]cIсY Y!nr5ͽRܐ b(]RU)GL6fmk,XcQ:̬N~A^C*RqQp#:(죰&:&Vj .TS!Wn<]iJ?ٮ .ƌ9IiliA)P R[jld+fڱL=2s{N4|&|gnxmC)%}kȆ<cqR=at1SWwְRY/`H jx"X\ƸvǰUQi"!N aZ}kT9?C֐"O.SxNtKyq3j>jAETq-$$$D;UPb̨ %lZJʣ5)lsd*mqR((*x;Pu_UQUa, W5kw'SN*gVigJё*Osf1mG T3آ$ 0^B@!83ƕ*t!/Ѯd !mp+s* Kg&kMXITP0 JA+V >|m0F`[| .:Zv2[3(%^ֵj@VzR kE~k /-ey_k{CQ2f#R _IBޫĆ;\->^TJ3Uj1流[=.tqFD"M%O\JB>3j'Nj ("Q͢f"Gـu}MKO ;4n:[v=muڒU쨪EnonoimJ]fFd󄋷c>vsL@MYMU#weڪ}Y8)~{*%(e鿆1n(6`;$G)YsJru-ObP#Pm:̫L %y0Jy77T<0 9{QmLōrM؀U}%܄d짙͍'8e CD~}d#s2Z&w1K~hRF/!wS&K) &?x{ۆa#3u MrmyRݫElW4= R}`16@F{R k[y-ɚŢ U&ifTI|eVl3߳A Yj{r ޾|8yiBA-kDy P#zrNrbXQ# v꺿:=-u" BcYf9wJɱ.)uc]N.)jz5#b#2j@@zx@`cs3r+ v̱NVn %-jD6-pv=~ {7ִRBi2:%Zz};+TOX| *rWd$ "2KGNd sӆa`6a$53QI&CEsdӘusް2}Ob4yyn V}YH^۾kꅙ昪nP}p]l`-(X>5Ui ] Sb,|dieH.ߪxi*nU"QS #WO1\ z8汇xDb|=ؼ'*G%'TB??$c7/F}iq^}+1WD[[xOehM&.8=ƹ0bWCn ElGd`gdRP^:b x\%?ݹlg% 22K*9Ǽ`&dnI]*FYyܲ wM܁'9H&x$qČ!wYZ2'`t9Ϳ|LVDs|B#*P}OǷH5> Hpri\CYeA+.tz}gPM% L h=]$Et8_DαCas`15 ag*4]_;t c\\Mߟ&@ 'B@ѾTjHj+H ֵTZUΕ_[#0a~v-lPū8i\hzMK': ˫WUz_կ/]f* % >o9Ҷu.r\y99W|[Qq1Ϫ*!РDonMdEHrAf.96il4'?вv]TOlPrx6Lɺ)YW1%*d]S EY 9r2gg`8+(@"ޖ 0@Tm'jm'?L{^Ҍ9}tq >q\eDbXF^F.eaIEʹԌK*@}l`!~YY71űDı&~|Tk!UߙD]0VReo=P#Nr";ߨ͙v* Uʹ%/F%!Ǒ}8تk)4+ rˉ-'?r3<6,_JضF&D`?Lgڌ7eJ3ޖ| 9sƘ % SnHá{-U rS0.dE.Sx ː>d>h 9i }PĠ/ܕP v .iI",YHġ0 Ƶ Eո*R.L&–6#SVYr[?r^Aƀey~bGN0vQԿ$ XYa;E]qӬq˘$:ѓZ]t"jZ.}aEArad moMqڑ> {7{m?ͦ>1Y<- kF첣\mIy9&-D` 1@)_ d_a#K`%sPp:Kˢ=i6B#Vmwh[^}@kM^ 5,r ^zڲ]r֜F|ɉAmC~@c?vŴBBqKd1)XВ- ˒8.=5|ZІApźZPۭ5p r9i`^km[h6dJ(2wrE-iRBsnYA$~}4=đL<1Q-t8rj8G`Z}@`YFN{-(jM҂T(DӜL'Z@Z%hӰ#Ta$ҫ9 9ϭpaňlONbo ބǓ]+\?Tpw$!,ů|kr⍑>/T~\vfJ]ni?8qxJ wl: qH?gw^晴+M B "Ꭴyڗ!W6G,22U~]-G9z#}L}$|yخDk<hYsErv`<+1P@آ4B HVp|3-jpmLmiwu%>~Ͼ$\(Yzl S ?VBI 9㍭9(@ \`T^ e7B\scs^ZaZ"8y}L&}Pw :)R$)b7o_VZFP=:m (+ ,jdU D+B^8]@0*!+ˑHTeA:(x  呕JH_qB L<{{5O=SDGR:I9,z?>Pxr8{&J![rϻ_MVk{͛';(Oqbs{ JϢZ"^܅i@ٳѧ2ʕ3GȮNRs~_|թxզ`sOx+6{"< #h+2x;ɦvӵ^&'nmq:mhNhMkbڭBS['g< 峷87[g[[N9hSzZ/4Uu!r])Sj4wN /sJ2$D]vU'7EJݧ릞̮U|B\쁏A0dޝ ˜MIBAo^0&)cק}Nd&翿6'0ݸF79U#-arc"->S< 3k|e&#8(G"/ _6If fbY&`u7Jgb8E6ai DKK]inv֒*P !K(=<$DJ1c!ӽďA~I0e$?t ps~' SJ0v' ױUH%gC*t!Kb ӥkzR S?w1H88?j 9Z;@L`-ƛňpC:lJMċ:S 9*Gc}%뭥Kp Ao]] In4,Q^6x_^~:"p͡.%SK{kD5S?βt.kbzcMVMq^A?ȼ)!4mZE&=iZ~g'xT,T⋧{GqPg! z>e,v2暀F+&-{ЗkC-9zmCq$^1K ?*x9QBN3y,9lRK" I9F&;5b1O sIN~Q)Ր\6H"+FWpG&`>]ȧ:y{vѕŐ<ۅ6H"1.Fp"ꏫ<+g]ħbBSݤmn`-A~#gvp$ qvkCr=ڧLC'*e+ۙO"Q^1*pLvFIhvw)C tG,4!F j-`ASDsԏ4WkjA0H\0ۯM<1J[W*Z2;*e¼L:lf+ dhXve@Wa$QrPD2 EhHThohmF :r!Oq,}/o*(?6-޽Fn../lɬɸէ߯M؄= l =FZQc؃=J #S !k5S7՞o2G %Hsdq8LzO>3 KnKh2W]31<<;3&wy*IOzwAEvH_(,}I0`J{5cWsjIZ1[΅CrHF*12rX x(w=, DԌ6RZCC1I&/D24Eze 9R2cl">XSǐއ ֊i-CSPbtc96Rfi@#1#^dCo6\bi3[(!dx' sb ʀVa XІsi-`CI 5؂D#n슆2qc7Á˧ލn}z[o ^iw6L$df2 g+"C8 ri'LqE~gfJW%/\GЂv~8c4XJš)HV;޿S~? 35&ҫ,]cz i]=ߓzO\vfe°w=:~IW]~]W v?ek{H?Hoɡ:͵3g͗ȋѝfzNRI3-7cǣKꌴ3%uF3v{Ƴ1gr+ߋø-(R P\_9Sج?;~<ɁL@*L(V]j2-FKp`F1aRN Ah)m~f 2ȥKF6t.ۥ.p~%=r.+^{d@d7e&-pJH D]Gfͭ$YpB;~.bu[Ԃijk-4֔k )r nbӀeM6֓XzBЏZ L,OA[j4֫o 5Q=1[ZLV!_GvhxIj-WĮUOQ**\ @[fZP1LIL eDPA7B!@ j mhK{Ccvh/1%mM0AHsYAfygly$׍H{Tp !HLCE FZ 8W3Ҋm9c6V͖jr9d"*cL;@oZan~oəL )UnFwJ4p0_C{2pozcbHzx_@BcFӷ!c䄏ӽC[>2 Ć]^la?^?ү$3OuE[OE5"8r8vL}u,K+_H_}9W܅_)tģXk|D}Iꙹs^N:;6WeKe.-~AqS;?LƱ֞\5+y76~Y' % ȈK}ȽYPxϔ{Į(Sf2G狾ޮ LtAPf8-"b-ɊFTmRc#krqy6&F zK};PzYI;WF_laU.G '_A_jQƦHfN)]v/"IYc՞o.W0ptך"A⑇RaMO]InkB6`āWwzBh50T%okt*PI).{)FP"+ΕGnt-hWhFFg?:nz]u/;/,K{,x<"Gv3b]"dɲCC?ԁ}R,DOb_4~z{c ]U@[o6^ۇĻLh| }sc-H~'yonCI6HѤV@^xHlN"RI_ο*ush)"L :B<,1R^,߹wT$C8 l#q8$tD6W'Ce!?1ض$% RÀ 6Ȉ1 R(!fAl?\v+)&Bb |9GM<$2@`LXL+. yy1[-VTb;X_OL+0<˿䤚8ʔV0͑g+8f)/=8WhTBXj),%BU6-[LxU=H"z*IW!MBETr[^5EIMA3Fa$]'蟱9 Ž!\)/6?o nIkGo'vrW?^34u# B Ӝ?B94'5 B KMm|;Yҏ~G)sVʩ}O}hJ\ȒQIXPvN pT6 aC t4vas#kt0nv6ӕ ͹GhDSٽKNtU6`[²X+?4;Ž;or,^6H ?Nmв-Yr$`:Xr@xYqQ&| +NKUEb튔ri*ٻ6dW=KrbbqA}ńzHCr(pfHfıDTU]nJOVـ`FZW`e]@5a(4CKxrQ yam*⦠QV5ڔ6R e34[  g%Pdvva<9mP*EˎǠM16`nKCb'FRDZR59@)nAAn;$acSӼ(mY]D &Њld3"Vd"{G7ƀWZKQɎކoRmRꯇhQJI+e8#kcCzo=d0kE!]R ՠyĮkx/j1OAOC6NΕ$^yu ЩTeC9[o"řCz6\wlz >`³ɺbPq`CEZ1֢"o5A V^^ܬ9ץ:+b{C>,&˞ p@ԧqrgCvjٓAݰª}V;q*Z c2C£ { I!q}zk]l^~!Ձ=y%/Tr^YsM:qT;.UOP zza4,!u[d)Ax%P+6%;LAQ ʦ]y-[Κ65|`}M9gKPSα6pxq³CK;tYy/L?ι 5;~UY?CP\Dq@3Qq00{ 9 V0zo7؉ecdy=]5x{|ѳXieWTzM&2|tP6/c$)~.iohw;:LU}1VېlN67$E=;4_?rPp CS,V{E^,la;Hqz3eEOc+ܾSQ #|HBn͛_c9IkM{˼[;)A˖Ҽ[g1\sFpsv]M>;˷ÐJu!d%ջZtKJݵ.m0>:_F0"L6A4զKmZk-¥N}3[j:i.eqKkQ|u[]7$igUj zή b@Jy{o(,Yu%R O",aV|gRSU!ɂo4WSʶ+5*&NhutEω=H^AM Lhկw5NLт<ގg_Ɲ< 嶍נzJTct`6J6[lyCp 2zh!xmldXi;E;clzYkH=Zcިߴ kJ9}W|؋Y'Vr드 S1\9GPQ&oUz21&e< Q+$F5̹ ڦuUY12oނ B .igr˽,3Job#PkvAF%=C7HUĔݖR :Ox̧g;׋ǻG>dPHE19o`0#˿.7;1Ŀ.ӫszcN~Տb? K&ZaO&ӨmN FNI3pr"LpλPIgi\ܔ>r.=D䫌iy3Xzz0 k|{@eK5;[[Bn+)U]["T?KIc$:-]Ax!FS>#U4a\q]P3|. 9|F V!^] ?.ѢE@pufʜ<]D&K,(Qsn#,qR+v(%#j&s7E0U3RH"#Tr1̨6y :Ƹ!Fh%L fT$@pf(BTQ `Q.2}ii}ԚFT mIz;.Qk4[0qnk^j0ITNm@⑩8:5ս] ƞ+A1s|Mm>_bH>]ۼ`Y¸%/|*W 4{;O~(=oD$bV5K#~xLƷ.tX; }u~nՏL`m0xE1D]̓^/:ǰjIAz3AI=ѡ$/N;q8'^2b ν`* 10Jh Pۜ![sA-h^*&LZP.jS(BtN̙ %t(jJft(D :xPp(Q C4>=cAP"~"5tNT6\X2*Y5a M4$fpQ \DKvNȜ4 ֢eD 8F[s̴nSfj֬P="eruqu.6.ogq6#"h-P ٛy'.zyw=7fb^w: @gݳKm{ Q Wk HIQU!2$xn7 ?v҈R.ۿ |m㔙MVGCB+b;Cc63M6,!NtCKJ/}8DUWrjA{[unQh&gPq- DT- Ke2 TE@rw\n<[_lMїCbGF޾ǻ^lg5+Ene}'G<7E®^ĵ@D\B^z j%!=vB_Ëj]6Pkؑi>#B+Qj욐_J\hweno^izwj~iBv _A7qB+5{qF:!9gӷwot"qg~~1cŘzbzG7vG?L?o?߱㋿ +me'bby2}Wp n C]:5i؟jmD麤խM5g \U"|mڏO/֭ZSqqnmQ Oͽwj^|p] {&Wo~7{x̸Zٛ3>mT6ڇbnY~If^jȰhW{Fsx͖48ڽ1Ih"%!xFb46*9@M6O-hX~DPn? ?3'kW_Ha.!lď7f3Tdm$I.<"0Om`;;^Yl]CGϠ}#T,AJa83#"9;]^\P gITb'4YGdJ$ltbJ$<S@AdzD•GH%0Ozt|*Ti } O)gn|bx5>)DUlGWOW"ܦd(>_y_vK%0c8FֱB2jxSU=ȔqnW`GJ >ٸ8YM41`_/c~9;8~qXՋGX@ 4];I34V|T|fۧ}w.!f7@̽i9&V<&tJ*{-#Y'D4䤗A^i;"O mc+MYinЃOa\sP)0Z@aҘ2v5)O&;u(m fGehS;9 `+abJY&L+qBԑ &  6d&RZ {"7m# gZR`CSAMbg.8`T#dm3⻷V$Ќ}7h–ٚ[De8Fȼ4ZC鬉ʤc`<=ѩKƊ( 6Z@3#aLuaLUzEkFOWoŖM}a09Fr3^,GH,WBPwV(.?^&Ĭ/*L\S=b+&xKzvۣp;=Uᨩgڥ#^u|Z4 /-oW}zJO@&sK3kdHeߍ{-t|WD"=DBnn &fQZģ3v]ZLi^!\YUiB&xlfݢQwR!݋ ڷyDn=i[:ԆAHlدd s(}H|Apo!'d#ҿЛC:WY C ]9GQSWcݙ/nj+^7TFvnQI9ezLoo=Isbo=]!t P t@i T2>^!.i2f,#G@υ(IqH< .RÕ0d1w'v3`MtPY;R]I'f DY']_j⤃T[~>_63`"; |~YÇīa4se쏋Iғ$QS˜K7V_r2"e?M*Dufo s=~ңXkOӕMBHј~rlD>"Jre(lv 'H2z{ϑ]-Q,(JQV91'ȠO}`J ˼QAi^mzmh/&RF[`*(La,@el я>p*<(+!K)4%5}M(-~˚IʕmS|bT*Ը=sJ% Rb)3RS Zm$Ni>؍R)TkRXA<cҠynTbQ042ZFʃ_fV *3,7J]v=pƢJV2Z4s/Cw^ER ؊ZpŘ-nSf9A"3i=U!!-[h3ͤ~@XPr:3) |:+dlu&vj4hxFEE7l>;9q%3ҝPPCg@eѦ[dPVH h%cݕFf"a'VsS.Ȟݞi,bø`3@0Y-3;P[)m2j]EF:K)j˘H"5K%E۞GKɐBK9ILe f.3PNdrqqP p3Yi+jxsJ qRbB_:; gem~Q֎>h/.go3`WN/7_wޤT؛VqxZs󹃠哇x쏆YǰS>3Äq>J)$ASU`d,qY| t6$0kZmLo?9r}kO>˲fLA?-x=z2HQst'2Jw0~Hu!b8&dy!η7[Q32{W*4;so5F꣇:ԆEBJS=adze*YU.J)?ޕ\۴ӹ6;:g%.o?]:y¯i7?])Zu 76<͏Hi45d!mI ޒަ׏4ӿ:LK{ŢCJ w!=CE7XH IBs_Dt&m@6 +y(@VRyq_AZx]:Oѥ]:OѥftY2iΗ^ȅ3^z/f`rJ!Ѩʖ ultdVphF^4j/z"Ή9fUxKfΡ2+[F m(:ZXjiPDvHQBYv FB cO#$5D6OlJ'8*ɏ},L"dd3F0+sCپۀv<ʛF%V??[.(mЕ˧Gtj-!!,|­2!W+${I%̑91Txom2rB`J}[Y4S ۑ@N>VuHTXgbޅRHAPPydAKTH3i+Z1ͽTUUT,r BȺhFeD.KiWr>7y,U{Zj2Nob XM~ؙR"2lX}}tWU eTSF- udL} _^Kd"*&sȢWViz3) 4UB8vlL0;F%gZ#q_ >nD76b#JvMm_*3NJ)Wcag%A)OB]$P='Zwz{ JNd9T)I{#IQJJِ=~(2FTȔ Uhdu UƵWanlSSʶ1-MûG޷^ @ʂ,J/hNckqKߴ_M叻ӳI "i)I画x7a0iPV  rf#XӠlW؁1^`Ebfsl+lFQ8>W.j :TyJԿ=^~Il:0tD7xtq ?DM0*ZM^1Z0.NI>kn;,7J[̓3M3J)<o :B>Lv~߿œwbK^q5=eaϖg!M[_}ִUEU{]d1Q , P)$ ! "/7S/)fKYNj@}AjD0}zy'nz>D}S~=5hG>]ŏ1U(j$/vv\~8("AiĢ0,nPVX$za&CJނ~<'3H5u"$ ; V,4_7DVO.ؖ0k)P| 1dڐcq:\PsVݯjܝ5glf*Nc if<߶qyfOd3)47W$\u3B|*Icu8y[zUH6*S(ǹgݬR+XTĨNCۜ*fDZ)47W,衺ɂ%6O:<94OVe >Te?(^3 ޡD{ξ^"":w H>eW偏krڶ L[Jļ?~/fJ(lw7׼)<`P4)AEۿVi{Aa?~>\V4:S{]6Io`)dhe4Jc@|@RH9Hoo쳢>`協-MC17'VǸ㱡6ψW~x&$zE3vqPXr^;2ࢩ*Ϻ;V @jR5mf?B`ZtM?^n"bf-)#g)Ս1Gp,CcpOnҭ\B#]bݪ k\C1]Lk8[?F \up9;S@vTCo AuY,F&WKif8;д ERάӶB>空=[u'nqR|?O쌳RPٽ~B(%Y30"^>O8h¡䰍ѤVk5lї]̡Z>_GA QDϝB67]W¹upœŮ960Ɲ߰pc9UCP|!ƄI6ԞguAm/QO]Nz4i;IW iM6e₣*g}s@R!MXZag/$~clamH`V5ĚJ{V _v`F9G; {f-b[M(T׌d-܌'C"웍8(ܧ{r9 htAaC#΂iY+WRJVR:9+#N )Gu)8>fQy Q6ksR7WT7ZMQQ<$W" ?*OADi,-$ bfv[o .WjKY=}Ri|.RVOӘ F@R2Mn8{.hZ8]x(LW#a^_-LX}dU?gwMWK3uLZ6yk5 o4zwMuÂ9ykh l2B៻(†0iSjK!y39f-LQ<~gܹEEEkR+#%~&L3HJk骺vC@Պtm֊p HIW") xAf9 nʁ6;*>zkrDedPՖs4B9v`1{#f G^=1Yc}> j2 2$R PKVf@U%t`G138pmSܞxj$ŠAd8$+2@HZ^sv`5zLEHng4їڜ 0ax)eay Bmo#&q cNতqU a7 In 08.< c\=I\8(6j.7_ 7_89cBW;U{*`x zS;_>lޓKFlͦޜ #yd&'1Ⰷ)*+γT4T8G~pR}Յm3c.TpwpSVh6d6BY"ᥣ.syER[6a!4h@hWn J ſKKh_ɷ?OSzWw 4q|)9ª+ ]'l' =Du@]xVYa# 5] JnVѴ%3ap9=h%SAܰ#rl 'Ȱݭ+_ |82I]%wvV^pH|$U4)D:szsΥ$j 訑cByU.QnRbeOJDȃί%:\4а9GV{{SOeHX˽Y3 gt⯯6g  +X8wmXcjkJUN1^Hc{LeObu4 #h@CD`mà Ft=w:ݺ exxc8fG.aQdǗ/^8T(|PhҍTG\4,OlKl{&cץg.>WP0z뛇P;VpJ``h Ȑ ({1@y=:t}BYН(HQ`C+T{?usl熠%z/,X[T 8 ާ&-cРQ3Ԛߙbe!AhREyyS[K7e2TirNmS1[7K'%+v`W=~c?o3@G.؋ o|< W5o—$D](#ف1^Mܡ{Tճ D_8jI.>sO|ɹTX+eW`Zzl)`lFpdqN3V7ܺhEfyDpp{?DǗc\cX#sWZNjot"ֵ Lk:Kɶ.W\P`qi;VFNX\.Azr[q\A7G`tZ;V[R0xH| 'y-G^e10KCʇާ:/쇴4-//[gиTXF5Zbd@ qCsOjv Z\W;brѺZU.OW >*gJU%RU.u]"kcIrBp.^\^J䡌ț:{)*@@AŷԔPPL.9tܲ^]^n aWr6KJkӂ-xy;[GMm}XS2”h$dCϗtcc@n@뽯LT$H䬪m@y{*G~l.uE9c%mhѢ|Z#-eE+J?{Ʊ?Iɼ[P H\bjF@X~zVxV+8j=}3Bj"@()uq 2x~l#pVyE VgWOJòLl%}CgSK0+"G?'줒Mԫ.$кWF>7}ƂٌAOQdZRYNО6&*( 9ykR~c٤2+~5p%[ KWd1C47hX$k7iy7%/m*ll: .ݦnm{ܽdƓ2 3ºG!vO/CʍV,ɅS!?sxa%4&'OQM^^8@w$inx4+r,Bj|/ֶ7kJP'ex3eޏ(LI,na?M32 Qq5ȧGXbI1Z#bITMceqR\U lwftOvṈ"FIg-͞C$JC~ɀ Y^ ̦֬ ȦMw!kt sTS RԯКfz7i?S_22_]nm-sX3$3b& ˣMgggm]sʪ[ֽ^=mb䅵2fP .̹w卮a7fo@kY}: .-@@J%8ۖ uѪ_:X^4*-YbH]ø, zRTH:pcOX+<.f ib, K\K9l8tMBLy%gR65M.6*Mc$\cirzrΞn JLu,I5_֪&P,=|raqꠜJACj) f^WDσ~__Cb=I~uYߞ )NʗL\l_iIHSj4 u0dp8I=;g'U+{Nn~FlX-"#:R-%XG^¼.0TrcNrGTuf?L}iGM+vsGrU. [#3:b ˆ򭐷L 0rwrju*0<-t/T:jqA$"ՒL0{LH+?x$;"P:T\^{ab>H{D ;vh:0duq;F\Q%RX%s p" #KiP28%w[AE;4”U;6^wHUнCg" Ƭ(dG9tfl@z 1FGM"q$:+e8^c4;*6RQTpy .ic1\C<:: ,oDSHuWԇ4e"ТakEz@:4ts#pS:|cBDgEL),S˥{"sc,)aN,P%LY'u%|f£)"t?ͱO%WZ+i]ȻGԜjLWïE;Ƞj[1Xf q :R֘ TÓiHo7LZRvVf0l,,3`TilxCPЏtz?xk;h#hr҉/[, .&rlqj/; m:7֫םͭۯ&!}`?ݭ7Wj[[;{ m?<;A3/׮>{.iuqeK^%_ {dz_֏Qp_I(EW~Y2= v]匋;y3ݣ7'.GctA%Q d0~w Nz~h? @O#x_UͿ5b87ӭ'kݏQ_I 4_s2h.kͿoa~6cå/by>/;N{>̓&r= CwwGy.t{ssblzV<0]߄v6ƒ{ ~<^tO?<:?wz1Fѫ^?Vtg@Ǯtl`kLo0? Oº \nl }YAxc7C/5hеnIW C]Ej F'2а{>Ca .'u =Yb,s WxelNgw[x Z -6C PK};̀ȁ E)wC% ׬''p4UY˭r7kyD]pYI߬ݐ+sוJɔ-ES ƯfRHU-DHV_A ʼnlyyJ&7Pɀz#Θ5̄1 \pʐ䷼F@G O1eVr@tZin;ENyJP8*=!H)eTF4&0ƞR`,/ƚXc/^"&J Ea!C64"5sKNfRGn &܍ ٣1ec|R)>sU[KE]׆(t{Í;ݍΫ[ϧHD0qHvC~cSx?g _0@_U,]ҊEoFchuxU^2+v߽4'N|A  *-E3|eY_Y_rP3~WqPޙvLc:aa|ʘ(33"Ҡ .a` FX Cf!up@{wwxZE Z"Ug;MtYJtY1fVM YEzmfkEEfJ2nG-rUbB<@ȴa &Hn2pQ Y#e%8_M Rٝ,RٝXvcXrt:HY̥ 0 Yl C0/ &4<@pT+LINZc32RIZb^qGsaS=0ˀ,:,`Fi%Ax%<.12x*pdM|⁝ZOTKXE8wmk:3/{Ǒ0Ogӓ*^O}] Dilorpa.RZYm#[]G#,NfJQxiv1%!l)b¸^gB hiyd$-ߡ𻑭5Hd"[a/j(aR"DΒP6("dr|i"> W8hZp+B˄.%kYXպWZnQ:$LHP`ש(ZJQJ4IbK)L0'C2_~Z>G򃺻m<ڠ$iX=[MfSDΦݿmEKX5XXX1Lb40FsdrEҗ`"c. i'%lӨ5d+\==i!Hx~ w۫)Cbx{5P'U$MO{q"FyH, fb{%&%8 l9k3S\{b=sA=Mg1wm+fuHs$bsSOl/ɒ 3ڢ\PrG3FK*h gphd%o7]]ף{5c~R2ы݋zФ1 tj*܋}rn=U)asޞ=l>QwZd{ۖ}Soҽ~iTc$}wVګ7po>nzwzOPܜվiO}]Ӻ㣿il.Bm',Bβ>+Y |Kh;5צ_Pmzr< ;8HV|dRsT(r49{Vqҁx' ;w |K[ϭIpǧ*%GW]NlkG5'jr7{&[o`>uy0QueL\/\{8R&[\7I&pPYwԪ?_Sd̃n+&c[7 =/KkQp2km6@ Rm:y_/% v03<+R^}8CH=!- Ve͞/N9h_]`pڒ04^5);f29-s SF\h>x@T^mR^zwTӧLŧ0bp fzeҁDcY$}lode Im.RDWL3y@FiQD;pwNp_S{a?s;'s;'st >Sv&}f|w{Op?q=_6՛׸}3TY /ͰuY_k:vXt"6`4_97)= [3~npBRK-+3.w sj9eY٬LmRtD6s}Ms}5>ܟ~잘#ʵ~vܓjUw|(P՘Ș9KC#Pr?+}=BƐv/HDeZ-vT=t;e+zpzrXm')l/S`Jߐaoon-^J"R?7 n2͝R^[f["g-aElkrmȆ F gu1kźZ0!BHڄRF|\_|]̼5v.xk:wݮykw&;=_%!”׼sBtriwk EuJ8K+u ֿ Cվ1/G(:h0sdqKR|V(*(dzǏo&ZU%oPA9§G*U9u‰_\=.W̐L)БйLk|Iw~ QbDcl-肁ZQ%>h**l͋_Z~/ܑ:us7FixjiXM_pQ6;{[kTΩ }W)%2>XCb1PX*يhAs tR(;5=le+;n$\= $>q1@Bm{Xh85(fXCւ6\Z!+@YŇF!F-9oIB), qNT`t.!O }Iq n@;f/iX*IrOAb̾K}xy 05V80 Bgj7H^`XQ{00eZqXV¬9$0:bJ^ȗҨs=bA"I<6D^A7Uսη&k^X `bR! rDif(Y#?>CeqQ]vUKN+(гgUm&0!ՇߺtrU>?O-IHI(TyS|d̽jzd+)&lXSA bb!2@tdmdR.d (\cC 0CVRg% VHnz%B'g鑛\yB?.n'[ub&h:V\B)bַ1>ɦ*7 ZkYsVdmH,5C596eu4iX='i2a:bܦ5:۰G2Xq!Ґ.C| ڤTRFUTD؜fƔdvpaGΞ{]7};Htyuyϫ yśQ"Oq}#O] ̓Aǭoh³s/ZC|5v U^ù|h5Q=\]cW'v4%/ Ya@hG'r?ǧǗzxGZ{ӊy ʸY-ήϏ򬶰] =-Q/$h]CoŻ㡤$*9=IbH?.iZZ22em"''GR z#]˅=zS. K0q;7PǭqS'z`u,b7*{/ 6n(y/PR 9Z%VvҁfTz6ȚvKQy#ϔljDǀY}]¤ I65+lb&vn) 4.%/Iѩn57 YzIYH)kĤ$~D74A/ɶ,S`'Ǥ7Q7$ޕFr#Rjއ~0bgÎ!LR]vR`Fd(UeId^Fn&#cL"7Ijwm%npUGs8a4vsBL9i0Q3;} vr(CaH֖TBR_i1CB+--d;hd7U- Unۘ4%LˎG%i    Yn[pa]Vr.~I+)cMlQm"@7m&7݁5Sn'`w楰!>K Dbv̒K[m%[,Pk,I m+@%j1ZĤ%YNKQyܶ1#)Uo)Ry.G@ rvڑ\{ Km(q"fm|2䪊RHC&ܴggdSoˠ0(cd=P"mXv;6uJmQP~. \FN ldɒ!Hi_f}v[\C={RS>ap0xPK=AV W )%䓐 V@ +H m-B2(XHnIr(! L÷@9&Q,|9'GZ0(-[A(3zJiq!WPf*n.T)(6(QOA@ r~=xԄèT~M:Q )`yH-i,6*?UbokV,Fz[EwNiFv"W.0gp&ZB5~kJ*uSڶekIC1[+%rXW3Ȋ/XЭvƴAF@zߚASv=}-qReC=ѽ%5-^W0a4W$@ 'R]55+tkbpʼnze@j.fn9.ɡJ݈ ¸P NP#Mɑ!XrXvf2KoK U8mo-T{C^{4t$R*w2Z@Gqo9ơX`JIC'lS%i;=ۧ7S gBI=d}%j*0fn?O#Z< N$qaKD @X ᖫ\H?eL$ ?AiT+7V`銩2Anq):k'Ob\h52&S43'\qaGjM2)QK >sXh=6ݶ[!M2uԿ/yO+~3Ay#}gxy>Jq oMubnC5[DuŲi[E]_z{;g>n:egM7;fѸ w R)eLWQ*1{y]nWMPEZkZ_sd_>x7\G ͇#:H1+ ?~,$ԊQH0: 'dѦAGEH 4 D!f;׎9kGR]\*x "Y5ĨzeĿg֟C8;#J-x׆MvZׂ}gz$΀;i*kx}Mb8k-y_q] I{ڤ[eӈ_TISgQ$3l60} . Oм2AP|g$KFr3Z?qB6U}my fkAjB^h1 ,Ak`h XbHy̕{?^N0>HUQ$``90'NĂTi hא;"Vtc4`&CdІ5%^%_C]{UH5dU>pib/ {H&G 'gg} }$J+4HUA(f cT,4-C,BZq-3 &<{FQ B!D7Um %25>8~XŬL!5f5teh 1*9Q fۧ7t. 0&/Fji=p Sᲃwk%. SAXOIRs@䱴LI HmdCH ,PӤ!4~XI+z[YwX*ĹCJnV1z()0p ^> X3j 4`ܨ`.y+&5bLiePIiTi2^ု`èUJ veY"8&RVW78%Rر A0IŰUo렐MJ)`Ljq?Y5S`ȼA ߥUgFqgwZLb\܆,|զ`jl{=Bf!2se fMk)B "(dp)#+frRco?ƌ&A^[j1EaP?ꈮVXaݶiY^s VJ[(k$f3>I[(M.pQ%h(.9l?rH8SqޖA%!|Kd&FkqgC#!tpRʚ1 4)%8 S@i/MЫ=\j ZSc!}>EF.u(eo G^C! XД nLd14=O#9 ^51AKdxJUNԀ-qģ@`g.|JӪv *gR|jf@qB܄ 󓷘=aN [5]-׫8^_^?G\vy<2='Af;: OcX}ȝbGYpEE^J޺^@Rsƍkxtu3M3_ǣēY *\6UJhV{~toIpS53)H3 TuERS*,2yWe M&\37 u˗KY2uKEHDLTVgޔ%yLo7|7Jj8 gÁLI8:&wvw=Npw=Nw=Yd $<XW[&AjhPQ%a IfIwHf^,x}23`iƊe<^qY*:A!RN:DF9㒨#$BZ@ ]iI83ClrhӞԞk.!h^'ia%R0Mco-l$zͤw} Zx?"E2tU 2,B=3FyVr3`SXM2< ɲ́gvCQb 23CgCωby|K*ܾ7jx!^}.;=V7R$>V[,HJ[6H_Ni\0uCiɭi2"ܴN ]}[{ȭSaٙ{ T?#$a{Z[O4UұDr")$!GArz9g9 3h wlac J? D!2 %[\%oR &ԻԻʶ>w!>w 7%|7y֐-zT+ɴftv !_HvPf\1!XC,qJD0YgI1#wZMX5)HwL(_{V@K+b} q ;N\Uqe#/C>k #'crQ/λׄ2Lry}g{Ɯ1({Ot߿RJN0:&w\W As\-)E5߇'Gou }ޕ*"Ȁ An03lȍVu ]dU`3H ]l щAzyg[V b"1*uqO |w"m2<mV&wfFLNI [eEԵMQ{k1{#b6kq|mm\bpST6.Z h ^(rf1[=ZMa3.fUWj 3%ݝXX3k5qmoZ1eg`[<ŠLnY]Sy>eͯlFRjA @4m$iXLC@ rw&Hvg> GDlv#mREK4)sX")vfuUuBPa)7 !ņ1䀰[ܛ9qGڞMc (ms7`6#%{keVJEol[+{!9~tn}dї~u;/$țUyCT+-ꈈq>P;:ҿ}v9I-Q΍0ѺK6 -.>mCnVgBZIiކ݂7[UF m9wK`Ky-!ΚÏ 6ݐ #FdX{=@!QjL85cH̾5 6k%kHl)wQTn"!B Nm%"J(meCJъ%^褸L RyoBy? `ى2}R`HC!x UvA@476&߮T_Hڻd D@*1H\B?FZ ǥG4ZH&5>|r.ܱRP*+֒:.)ސYX b\&&WDTYO [ Kz$OU0l؟EWI\No+vzi8xWR2KEUPQL \QďUudrO^R]GH!, u:Ŭ$竛;Ls6-¯fn߾[hWͽ F(SLa5)V^C說Z{xdg}5˩Ƿ O;,Wb G  qm\Zww aV^8mÝ[Km$C{JxlMxci?WvqZH.I#=%bQ"nDA]oFڣz i ޒJ@"HY&ސ R3 ˚{ } gz"/; &yi)S ,|MS*}tF''LFL`=ud(}RmBw^Mւ_K9I3JIKQ,@~śJQ n$xN28@ym&&$.Qz `CA!r܀2SB PO1vd a]bPb%HXg ablO!VwTÎ@CR?~N!7`2Deñ+3ᑥW~Ɏ0فpnC֘x:$_V6⊉[u\uk9-Vv2!ܚ X;8]yjG&ގP Nk;6SBږ,#ւ R|9 5Na>̅h]C{3gMphd[ѹ Lp{JP m Ȕ ZR ׃6b8乶Hlǔ$X.vĽPyD&? ޸Z}\K|qh@y:vAgiY<|TP?qR|9T\ħ%ax2`Oy,:ˑU!azC  }{RrP"TD.>,Gc~ gi[4&nU>&nr Z3Nn  T6O1H%AT3Y-`-IILK7vߔu4CpK|Vz' N:t&8ǹlCZ6ΤDʚR29?&J, )C%hI۹wi^!Lr-A]JD?gJ7 Df;" r_JC&8r p$PNJ&!6jw!Vυ. -'8f`$3ru˘!BFs*_/V[(Mtqnk>\Z+7q2sD,'V5t7_1Rz(0kꂢ^NM⦆B#zAV;0;W!  եj}RBi2ҝz<P2s#]*Fns!Pd`r"I~g+s͈ct! 5ȐҡF,Xo=6>W[CF(;iCoZP!r_eQXv% +=Kʭ1no)*Cѳ%t|Df/_xw%)~ԫl/FTZ8%5Z/~4bmXJlar*[Z4U4o\/,EGug3vd2ӤwM;jbT蕎9-?MSqBvĵ׷v [2먓 '0dqܐ\~>:IŶ}A_`,s هrHp[K2\*O@ϰ! Q{q9R HКyʹ0pnXHj }ՖtW k$YGbeK)~>89HG@$]O! ,&OfF n\+>5HP@(=(,\H8СP O` wX7*dF\)s![w:vb4'|qᮝ,9F];ZP[3a&3Ud̄v ;{ZL"׷_%j$`2HA3h.s 9n؎G U6:>K4ZCw֛(%})⒴-8H{| ޓWVnW{4"ҽnenC.EZS{[6ԻvsC TޯV*쮴mqiM=f+Hczۀ8vq!d9=қfC7#3~¯ Ď=Y %5^[G}K2D#=zc -egIMcp(ݷ}>(0"["%2N.Lu!-`sM/d3IzkśԔ ט7Ԙ(C.7).UXHP|ozkG[~6 je(G10ZJ!jI㤮RZlurrV5.5aM;6 -hEQg1ֶ,.:cJd~ivR~[S+nS18&]hE{T5;XP3`VU zMxHeqj"(-q[g`K HVN=6DE.U)F[uD('X= KsX Wz4V |2q3Y@N4 B8.DDqVix56󻓷N' J r@q=DCӓ~RtիϼrWA\AP.u$S !Hp=%@ Rz'?*@F9Akr=F דLHkBe<\<a"/ K P8ѯV\fbp+C %刔Al9.݉`O7n4Ktk~w? A,ks$NTO8OܧsIB ;hE{.OKPC繡 J]}RKrQ _ HcvlgEMhr16:p5@y*l4-dc %vnT|~~]?LK78N ^\?SlqYϤ,o7?+uohI.fڣn.gME-8Ңgv0ӢWoSh@1ŋD(&S~`z!t씾t|8J,]D먳(Yɧ?zڿfz@ ͧ~ş/_Ow7j8 QHF4#.G8Q`Ľӟy>ǽQ GL׿clBdgw r?h0Dӌ h.]н{qAß~h.߽.dy7w4C/ntxv.gndNzA]s\Q0ݨ|k0/N?xʽȬ9 ^o4a9,kZ_Y1y왖o GCl-V/H@wDShzIл(T7<_5&H/ӒsN|n%s;~o÷SMU)UZ>\ßVܽt\/!Z {h!~/޽彖z P߮RjL^8N'dsƓ{Co=oZ޵u#ٿ"fgL0A8c $5cɊdOvf}oVm^qb)X|oUs@_kx{]} ez?S?@>}jU؃=S~a^>}oG޴͏?]1_.ֿ~~||^mWKM?CýˆR$wsru}V4罚s}G~%]r.wifWv_OoHMʉb;l8K^_M_w~Ř<~fn]Smj;낿9wl6Uns49'e-ɞ~yfj6+oo-ivmu05Je]&~6+&VRiGlc9hFoVέhf,ݛ/R/_j#_MΘQ"xVSɷv—!aVUE1;CUF"+Bbjb3)G AK-s+%s+/߬ɇd^5i0F.fv gmd"mʝ 6F(UP:X+GJm͵P1\bRIASuR#IΈkdɵIG$GurKEm] |RaeCVJr2)ɪb)kO\0-j[ɤY)^*=Kj+_/>nc^]) dv)8$%fzm'Gg $bż7ũ"HL隖 5ⴒ"Gx\9hޡZc._5٦8[ >\zN؝H>øZ''UkLuKkғP3 sU8I Ȑ6Bk|ZksTTR6B}Ismk3 OIlgH[!9iFn\ߜ9n`i;gIK+;%j&µk]pZk:1S0[dZJG5X|7\nT]]Tԟ_|HcoD*O\1EemOޡ+1*ݖ:n2V(z@|4e1jz[#l{Yw86EڿoT<6wSII ~4Yn~7fmfG-N-FSfZZڝێRȽß_{Gr"kM@@HC[7W{O蚊&E}V"sTh+Z+Ii-Hm$J>R3k+বJVT$t9lV@tX.,Mгʊ]o8jy,Ĝ~O'7|jf[!B{QֶHyyIR8 IzsJ<ݱ mqƊnTjV^BNdbt)P㳉*\YEܬj v\zq"]i}~*%BH_-IEaQf&n<÷nJxnN`0/勛ۭ/+idzv.1 2Pmo=C;{ӏ_ճˑh.^ 0w>:I$NJJ}2v{y AzFy-m, t #q|n ..PVPym5T1ŵ EhA{5%ZbXST+(,txHI|8Elk7a{?AtCMư2 Q ׆frzICBM:0sCR'A,ڮJ-CJٚ P>xۗm9t{r߿ݚAIx?hnxGopq~a38pXX7| M)Gjb6}bgٲzԶijes:7cfv~j:?m瑝gW%^__ݤa9>Zz~<|h3_{vdm3 ZpݎɃ5SR2ᆚޮˌfInw{`1\?Yyp>ʚxZ??^)z,Ygf{}w\?F"Xa6>:GA=غ yk t+)h{|kS !\3^U `LJ,SVe8vۻ\'T81 -V\nu[081?Z 8F$Pi5A۾Q Fa 9G:9 G!A hzlhZ< hG2MдhƎiG8A#<ܘ6p yRibPiAƉ{eZD<p"MkCJٵj$wZҜLRڙ=m!~#3rW-'ƊBd'D!lƐٖ eVIR19'!fL$V!FU,L}UgiU- md z,9tF?ta!wOn*s m!>hv}uJIC)&6EFixYw),AѴ Q6+2R6)U+4e-̐*I'R1s#+,֥n~jLhۈvmpUh3mF'FAbՀ6-, f|ant:qYMhmF LC;qKAZAMfBh8, m;]o4n.-Gr5讀 iA&U R`Sۖf|&i31$iZ^S$M#ԐP(`Mx l; ]43R XC FN8,wsqTZ X; pcӹ`)aRE Һ/9v#9f/3;ex՝5ΑG)9KYSd2]*۳fH7N̨-~n 6V8dCBpLT#E%jIf" Evy;k㕜`seC -|gAkOZ7kEhD[lxn|m=)rZyP5/X5'-Kz+4mo* [IyY`܀wD Caeښ_Qrv_\INrf*d?LtY` g2;~zI# b 5AH׽zu7m8Gb 8B`e8L&*G ?%|#k,en oM`.qeL cf_bJ؍s%| A/0Q|Ģ kPfoyX|5ǩu ^u 7:B!@"1j$ yk^SyF.kI-kj5`tڧlڪbK'{_-ǒ -'>1c(Oċ[*TvT\̎g>ks< s}:CS?wiZ괂lYi aWtdqW0MZ8Zz8 9zkݧRš1)8̕U7wY Z6Oئѣ?ז"c-GwLy)ye3Ok j'6x JQdt ΀`kvf stj$fْtRɵ~é_;/#/eVw#Xـe?y7I6έz 尜 m1*'\zOv975~,;]r:P+B5ĕQ@!͇_]QҒ68%›[ ,3$j:']auC_bUXc%0_B4LJK~?)/!W'y-]T)ɤ&⃩4ߙ/h؇xaA=_<3#Vmb5b_q]|K2A"T2p׈qde~NgLXmr~q.Cr]̎K72ip#%AuLރd*ۢTPDj]#2%{N!"$bv"7)@pv~Q4id\#< L{_xR("Uؔ +U-MzǖlM\[_c $i|fɵ[a>qtJHLC-ӅvVގGh /B4=Btep7@2%bHVW='{=/u/3;xOzU`!OF W+vc%RTv&dz+<{#{$o}c>特FNC-W7`9w(ʃ<(ʃSʄ*8lA8,pIQ3^OF~w1k'  F hw yT}е.^C 4Vhg o_digV~|pTL93txM9]oT2srugo}׻.uJ\}]rLys}P #FBΨύR>($1!FТJH!)w(E5XGH<z!UJ Ha( 00J(N}<`*4 DXjq^-5:ZEbd:c;fil2Fd86Z/ Sg*53]v\)%ay.z:3`!~pHd)QP$€ađb[Uӂ>>ߠ `$}2T,p?}*V L:_ kCa"HP:l8r abK.-} A>fч2- qh3NQY߽ȅXSb k h;[f'PBOs{<1Vwˆ^+t9y-:ү錶jwnY3ّ?nKKkJKTGů05Kѻ&eN0CXQt5) ɅZq# "K^XVF`4=$:grHV7?" # FxTB86\ha czs0U@zJi7wlh2&&C0u1 >?̧I#DItoNv—D qow*Fj=[\0oq jϾS'7=puë?痟o{ <}Ǜu'7X2,#RH*Vvo '}ؠ^M8Tcp1PO21բ-\\PB5j#'$mG ub@&,$TŃ5-i`MiM1vJHށ88(4TfFa2BO>R2tՇ,Ħ|bflԠYĖ6Inj٣h&{#޹a<bM/GܿMϰZ[]ػ%TNAXB滸>ñ5}Z5G+Q5n LZ҇Kߕ|~UZg c8 sW#r?p7`ԘEG +2#v R!+x<קMMDVPs|'6m3-WԌvzS8ỹd+~~W}Ov8}2SfxƉayc݋Wv *z%('(Ƌ6q * Qz& p,0:-%ӗ|l~sZ3)7]jyzq1҈.*J d#,q`:pWWM`TidSi%ڣYd{FmjX@+)/URá~q|T$)e5,)8QG~(3C*_ H`9E!"ZJgR4̸Ph }<_I^g;\0Is hHA)JGZ -l['B^} VpEw5&t[Iy843@^d+,,U0[e ~]o.Q_}mo/}0wFO_+d^RZMf-[Mtg_ F{Bmr@WeHOw6%jMr"ZI2і-&q#j(ݎJ9hZUQP!!g.˔T[@*9'H@S:7!R3Wk38T*(¸(E1"+ Mly'SSv@ѵq?c)5j|!9n# XX~;4US]N8minzWp_߼СUMQ۶H\]%ʧK6aVA$cMSvгגl*s[1>1JzK ˑZiia%h&8c\m }7qgí *XF9(+| ͡kb7u,y%^m҇grIwem$Izm`K!/Y v6/kyږdF!U<,/"##2xodto_]ăZ'=+E-7Zl71q̄(zՠv'nʻUX {~|wmSN`\^ږ=ǬVDץVJ|= pe!{ 0"!?%bw瞏h7s ѱZs50zg-F؍nOY|Y)D#B^MgpH aLW-+GPM0 s)rF}[r>),m]Q$DU̡n8D%Su޳4'>l샄?{Z gcPk.wO ^uoy>ڋBgU6f ~xtHubJ`ZfB^?A+*k_c ; '-Y1}.[6'>964'h]R ~縧M켁bƨ'Fw`۫<ߢL#& *,#R«;^ϫ1Lڦkn#;bB!җ:O.B|֎SȜUZ']埞739қq>bE]'̝2:Fl#5PIgۨ $%ky?bb+E# Ѯ>2+ʘ9lgpSB0 nVr˦⬱6 DǖvEj%ܹ,HG]S3vYn:OnJ};x2|hK$RH V;LIfIBG3KK^#*p&Wk\aժqY5ƆD%[$L>4;iwFc@jD(xvҁJ/_0TArRճOj;\+Vp<ӑsSTqLDme29fB)"B"D|yCYL[@|È$Z{ 2)''LiGosG"x)ː&CQ"0 K-APJTScP2EAӁNSɔ-J#. @ǦƏ l؊gW09jE_NʌK0=,nvqJZg 1+Ù×R`ԻuCG}bk8CϜ6{qXhg8p뉹MM,{O)mdXOE&X`Nx-}ޅYarILiox|s}SIE,jļ?Zd2dad-̍DnXf}gp}`SVЪRyHL)a5mji eq/c#^z9<+sx:w>{.8WNè#q!Żs_'mqś ߝnJ=zN%nE$.n<7+\=yzgG ~Iu<}Ϋi;ik OEZ#qoID$CͱslSHЬzvΛC(寜vMn;˴kр>9{)ЪW^IM%INW0F1;܉[u ]- cLB916+ی7*}XDIJ>YGOuX<\H &%}bNSClTgqx݈ݶP9iTY$aυžy-wp.Zoqe?7>BAά:ym։ ﯣN uGQ ٸlO8t1|&;-KD"Ύl=㉧{nVL5Qo3}zf2jӂBSS{>HT߱gH1)Ijz՛B^,Cx1mfN^"$Fn@o }`Տgc,f9/xLyǟ/ӏq5S4Mr| 2keS4Ra7l S߰ao `,`d9'>F sP\p"`#qYjI4S.\Uʹ`BQ@WTxK >^0 tm t/|gP%D"tϕɣH(͏!=_+ФI+,B)tnB)tnXkC ᘓ'N75N8`V8 FlP:PV(|wVRa|$4j&\ !XyI&'mar҆u' sFV:+2!XD)X8*P6txM({TJ/Q)湯8L0+b]>!B- 2uJ5A^YPa:Z`LD Va6+G;+= R7 ?^ ?_\(^x|1q<57ߏC?+iu]1W};N\Qڔ:=uy x][ש䀩2Y}LM|yJdw1{vv]ʩTfC?+R[">sVX.DԂYWT{jcM꧟4NH:*:I5lĝy!9σrCl)C=]x1۫^`3hXꗓrjnHBx=0}Cupz+Ak)P=ѰxfH).YakymPfleaNR:Y-Rk% GK|]-3-+Xj <Ǖ^* {e";e)F ]Dm" ;Jk<`YF.BErDx&c0GsB ?"E쌂AVP"Tco` *9A3FZA ZNA qAES>TVxEe4` !jd  &A x`/PR`{EJrͼ2˰D~GX.EX2-ls{JUDUȾG]1tL -&(}+*#k;G$Ak2(5TOBW4 G"({㵇ՕR=i}?P{@ V]N.4Ľ:PĿq}xb.OI@/__<~>{ɿn xZuDK_ h<Ƌ W߹5wemju}N*u%gTRaHL(!(ي* nXD {zzfz~+{F9L7!n-g *Y,a||9?tg 4PͷyhʹInH!ܢDKW;gE @*\rJ` %r{L P.#pB1´fC 陒S"p*6v[(aUq?ťK=6\1%PL!MǽIsErT˜,FDoV97x6$a%gtTB󲂴qC1"(%*HV$imgEj=U~vG`TڂôR,`](?v_OXƂ-mwspW:rsXȃwedQ{OU{IJ,ng{JcPILPyhLgݭ\:f/~>IոT24l4tjy4]’.MaBG:xgKt d}+w5 U@ҧ{ʸv@r h-*,Kj%[=Ϣeɩ! lLv*i!eɩ|^0- /R epBg DedX)[>Cytdoi^1m g$`'\0 LHxم"Jܵ IAx:.+% /X":%cUR;rJ-#u!;rvFчФeLU)t=]X-o@qC|pUFUHhnR5 TuFE 7 ,DkzT=^]p-Eu!*l(1RmI\QcΡWQIaJ JDBzOKNiV"0aqp=&X9\GeT7 'mBVQeFdfyv> @J<+(dV@wC wD3yɈ&GUu`tZ5X)N+y?= ?௎&0 Pd::?ߤ5"l(+d#v1\ݛlH۸Bw,TãIK`_%Ӄ6. pS}!+}ͳi4̯VMe$xL Bkr"SPP-5JvT}GVִ[厺j:$\D)ATq ږf.~~,2_3%>DKIHZz6WNo߃,i'} Ԙ'<ȼBԦۙfMWgI7DI Lcz֟ I;INwU<)nu^8v^t%V/Q+(Vr|*@K[sKdKׄs}s^ 1ø5uLF]1 %8.앷3+QzZJwoҔG$_$t$8vuo)FDCt՜szUHV%hwv[@Pe#^3 Tr'mo;;<SS;-8d…>ĩE*O1n3Vx[Vuj|;BeLX7b3f.isp-VNo34>h2i{Yf2'g4 9gf[$7(m5䞏(mݺLFeO檷R늷>m"aߕdP($z>qp^Viuo1pR'z>ϞfKOκ5AbڛL,+ni {O WYKƔQcڨW]QͮR{RɂaЄPT]%: m#Q˸%1b )TKޭtN:vNva4^]L]Y<bX Tޖws;<ީx ȩ{W%v %viWQ^;G H׿`)q$ ab2k/R[ DXiU<*Ǯ.RXbTKd !ׂķC0)=db KyS!B&2t]~m8e[~'M.R hWG GU0dȳIŵβyO $_Q aZ]-$NK;:qO~lո"qEUթ-WJd/e{^۟|.X{T|%7Cv-~0@4< щdٛ M wVY EOߜVzZzrT4'20Tg3w b2-쬫1R~v? ÿT!Ex*3ʄ8FcvL #ƃeخtu2ls'^msl)k\aMߗxDw}2b ),#ҳ`4(‹}y!$guEkU"%]ݒ`3\W@b(I٥5GCo8L"O(,8S5g'\C4,=`x: /`,LXJ-1r1)Vz /0B{ASϽsgC=%@,=T.O8WYkbUYgSl],/10'};LK/$|x@20m&k֚f7wLfMgb0g\Smne5n.yjfRҏ"\Z8ǹܖq9SntInujJD5*Ś 9JW@9ஃEU7{nU~n-yu<]Ϩ_VTHo1+JR@`/TRQV˷ 0J1q(C#SCPh}f_[JBk+ibCs|.c2IdAadq~תK]yut06L|K >lX3T,򴐱DZVPoW/=2$7iVff:w,fޞw{,CV=IK;^_L1H6),XvpĻ,ձ0ڣB*:(*bHfE*O9,B`dVt<5BF3CBD  mp?J!:Lt//4;]-),'v)*,2KwB4=yj)=? PL)Q ^!*F 8QM xn XA($&:)"2N8ȭc]&qp8Yb+BiKNPѺN6-,hzh[P6׻爒}_B dӉ?O?ŚX&C.v֟x!mQ222Ť' dHԲ2.Rf'?#2[wWca9(ۇυYH.XKnd/i]Uȩ%˒,/d!CF#xYij<,fyEeRS @6 eNމִf_V U8BQ1e=yܰgbpGc V]N,'XQ) `ݵi~zÅL^Mơ`]CN.o^9I&'Q_/R)xh'›w~49~;>"wN/}vO9{ɍ>~3Vn{}7}{a7>{_NM櫐f_7IZ ׿O'|$_zz޾'Sש _r06غw wicwbIO֍?id4/)%{c`9,'IߒF3EwpiȿMBoO &0x?|6Nò O׌0̳ ?ѴuʐH7V`04QNF70N/a4Ɂd)Qw͚oL $b³2Vǧd)d@L O'CA ouHK.YAξۗv3.V{WNĘK;yXlx얍/D cUWH)iUs)u0@y0j2%ŋބ6'0$ .:fK9$c[S9x`]~f?(92pSz5Yve [Xky;<[&8FE\(ls=yQr6) ׭ZIR2qcЖqvm a- ta$UaX.$[>o`U6\vX"4!(UFQz6bA/?iV]>Ǝ|cPJc)E\<^%HeTF$)\+t4fV;Ev>*UwI[P'v ^`, x} }a}//(BO8bXbLlUPDxl{ąf+5bq/:I _l2Oi6bbŠ/u>d۠M|3S 4}ggze /M7QW B/ BÅpa+4\+4#GB ıԗGBJr1`s?|),U_TI>|oU|ى"e#Hb`t &ٻ8$W,v+ aU]U}-"AVWE~qdDVd'A8C2hQY3F !h.@(bki]04Ed AABd|ɰE?AF#^qr1۵]ʟp\ -AZq,B "y 1;<Țm& &p;dp3@HB(Uz" cZ )p7>̥1gyԊ.{ _1^i51qF 7/ NHoz~wSӞon_] е*i5w )Ml L(C䜶:$X-eSb!9Vo4C鬭WGU]햢рUz?K34(R]/;ݍ7%We xĘJX(-`$)[VhCN}ucZUIџRNzrS,sه_i>xuCg6L>qV Fe ۲IYK8S&Ffkbk"ͺ[{7!cuvwmvrvT[}|5$nu?xS_KGS)W `Y?(RpP_)&IlI[tHM Jo_l:w|kIBk'l,apSN;; 6؃D1(c3`!@˨9 #1{2c.$`eX5lf imtR)F-tOaܺł#V!^_C]C\{} XmQr/+ú>8NߔqO9=c̻&?|VLC7Oh%~[3 .\Bv;dXBvx[*cDa)ckop:k,>8371VFw*-ѳ\w-`:H :x7$-POG?_bv=W88I`8mtc3~Z;V rjm?ضh_:etNY9e~ {Zk7#t>E0rFeGme`7y?Sע8p 8IX'lti/ 8Ad/:(Iy|QtN=^s,h<t_nå:~H%Gly ])1ݩX Iaqv8H81%ĉMfg܏hѵ {INdB;G*ԛ_W5zʎ OЦ7}#*-_T$t2s_QG5g2,&/%U .׆_n><W[S~7U7Wê/Wn9o>=䧷cZu-UÜw" YޤꎤZݒ^R.wئΦ ]Øxhs67jfouvac֟|eu ʥXVLF6P yf|'ItAA vȄb&FORಯSlb_\4e?^z32/.2+`\{I=W҈=H%u6]9Z YĠ+FvD$D\Q;KUR""]?ôsklZX r>Y {KeW@ ,5ƋE3QadcA u-JŢldc/ F,o%GPX|u`2#k9-FDǶI& D$S J:6Grz5:A9uǫAYAШQ*pQV&*Y5kl(BbDuҕaֶx&YO*@"(,YG*U$S ='9n`(XyfVp,`n Ѻy!@cT B>`^"ۻicYC8BGV$o=\LH%y=8Gzұڻ~S!u{=!X8JQv?Z9xuC'OÁe[CQ1e&l8/( Xt1sB+) | rH@8@ 5iLlVR-HX Qd5_9GWMtITǯgw""i]121"{rVU ^MC MP٥ho l^U (xY|RGfX,O:Z@R"z.>0N0)8` ) +@β3LtY "~TֱLX}aȤ(S0`XٔYjB̼ F%E2|̹Yo\${BT @f1ccg-I'\-&:`j 8aq ZVB>,NcaqzSNsf.Geaqx $c  D}) ]͙V # %uaqKIû񦒌4$C:.Lo/ oZ%t^>(Z{寺5itqoW;r`uӉ-in݁yQ&] A,+i*/C)}9 m% {z #7,8vzڶþ!ؾLPm=9^yGygF`JNM0si t"T,dM4@<* 4RtT(ŒRVdEq=;j nO pEVJ1z5jKKe&D(sHDLː ' .zA`Ĉ/ ]%#A$=Fsn=2EGAY"KVI3Uۆؠ CTα묗I('ͻl6&_7p$_֍+lKɗg҂YF΀*s꥞IÿD!QjΩ7 :k36,N`K *KVqZT1[ASKb-a62e0ΰn*dhEupvA2k|bᚦ@nA =vvCW׵Ws*S3 ܿY FgJ l2L`Y T; ו̿ڟhJv|`FtHb=iE6j \nX#*<gݠcّ~^\^&Cxe;e-ِc? [U)Gʅ>&&eC|X zQWvpu 7Qڷ}yl۪U75PBsVa6i?>u?"?Jgele2 -x<",ñ.x86D00 _7:'b (ϔnu҇F͎dH x Cb9Mݡw+@J3:Յ(@M3_r,ڍlE\|P,)mbJ'ldb=[nLh(('ĔlYCtlyƊ2-~uyKJ4W]هg1«v v`z{^u<˃|ߝ9v-8I ?{]9#YKtcMmjpCTġ6j'"w(^ TWVӗK,C86]՝PNghEO9N3GRY NChjqXoH??%IPis vvtx:;@dq'yYMiIBAi2b8jIi|6jt,o[1C.Yay)A زKCL*QeB^BBZlhڒX3AʄA+1b辄 Y]LuDMH42)ᥗ8[6ϻ)J𿷦NdJjJifVn{YFF V?JM!T$j\GKaѮ#]r1Io{kwHKn$9(L*2 m,ܼ[!5vt^Mx]mo9+,{RIvq]$Lf -{%ٙ+dfV+A0NV|bdUM@< 3Qn>fc-~Hw'ÀZjUc&WwWNO⿣DI;ͮ.מⲢ9R3-' #h;-coH/ fEBr77EsC"D)V2]H?@&ũlل\#:cf/FјaXBdu OWJiE Բw u!/4{-=LHޚipoOf?Aeʩ9Ob`9*PkEJ8-MS2ݟ5LQj[ro}w8+|I37p)m'9kL` &ԊPxǁGjMNk cOA0UݬԼuzȮTj,Z)fBP*'_|UO^,!~H1y^ a0 K@! '!+ck,ƓJ܆^>I 1`/[N2/?J_σnpk} _.\ 9WBUsUSs V*h5!A/Ik%~0)8,Vk<# "M܆$8%cetD"{$O.Utp]PiZG- bA D++ R`eRcLS8P .dH6%ҹ.Қ53: >O3[΃IbGHܔ\BksU`L Cd,g D9 bË5,(3+25X.[ɪោ޾JxDQp(ȄNT< )C T{%eZsÊ0RJЄ"@3$Jr6#(/b*Ee 050vdh=YpndFFW) }s2,Q∟,JeL1eBx+9,pW0c-V$Z!\0cT4Sf>%ڬB8y#cmi9[^zi) ?0sЬ i}_gUx7 ̒Щ1r`T_;URn$Y܊dJ/%`d9F׆0`|0B)SFY)银Wu240(1 ,ðYA98؀HLaXO=sMYvػ_Te b.MJx-ϼ$;4ˆ 5W@yAmFnJlٵ*8Ih]r٤*j:w6 8E*S78O ,:f4ƅd;@P@|,+98XdwzU7͒$Zxb0n9d@%"F:hhN"er荏h0)$1 OJb`-6@Ll#uÚϬ/0"FK^ Ҝ0]ߘIf~.8)P.5 $/wp1(^^r0 2+Oe}E7ݚٿs7wG/B++\;??RZ8`~ )~J; Ak|szdV^~͌]*,t4`gjZ3 r8ZGWoNûׯ^%sΘpm #V[BQS%"}(v'?ODҍ迚Pc0 A{eZуIӃ Ai-d%:%S딥f;P89m&1&F hvI*:EnhOvgXo) ӦԽV\c0bEٶkES'm5zk0ѧ $MKU?VR""DMLG)R=t=T"ZTPw) :֬n@?t{_,hqsݜi0dޅ-= ><\'wpŋ>vx2BQxMa!ג¤{L!FJa6aqӝ/1JÌ2:dp+ ӝk8}ˎbdNR[mmGV&~U Я>AZ^i g٣@>*{l/ igMLKB{G rx}y9>$T ?BZ' }Քt<0^""vN&ɝ+N h_E8gs=V!qyt7Up>o&$Mrkoʼn/efL(=wb7]<ݹٳG`;ddFOwnmUH }腱h;]'"F0[HՕ>r@|!$~/o# jهi\OEE{vָœ6 8̵۽h gDh I~WpHV坩VOaIHX]AJ] Nz/DaG* L+|SA3tPaIhfݚzUv @\q{ϽnfQN'|K&^->'{ Nyx^ȖNZ/h֯X a$voލ7)eӫhI-B ѠD6Ry-!^ZYA(ŷ+P٣G'xz͐)ƣk-}zu\T2,X9U)rQPXk5 iZ^ǪZ ү:uLCK\ؑ̀P9P1č;8g簩MT$nb[cV1ѺpjQ# F`x>׉ZsFB`j#@ҡJ d81ݺqq I%^!L``C8LiG(4 ۩k)M3YM(~v咺|z&A2X %I*)`}={ŢÄ`g+Ĩn//"b0| em#!1"X~"4l[4},!4x_RE^ 0/"$rE!9x1j/]»ç.Y1M~|]" V06ݴK^㻍>V[31׋)0z'@n$77Ϣ*G{y"5g;yuo8zo#0|у='ݺltk5ӱ-kf~]ŇD4# e"(bIёJuI1!yD1IQJJ*]P1ȩTN8[PEi~XtplRTM݄rK]Q3 .>ݎ` !pIJj_bPWmaa%Xk͍/DQtbe_^go:C588Gud;BJznrE@_2=WӻnMow#gwzR*n^-f#yכo&nz{|kfգW˦_ő'ulp;e|k/~:C |'f($5!04b,4 \LTbR(c>@ E;j֯dԸЎ6N(7{vDՍ!Jbf=R?A*j%q$dɐgm̾ES<: ^1g[8m4roev4g5U*x3ଓj3RH8;xZ<bb{32裟Iu .YO3-hCQm&dv3Lu x+Yq?G#ܟ&g@6LT!vaGTSDR>.JiO/ / 2k<ND 1ƭTǛCY,!UAvFDbDž+1Ӿm<s%ʁN~ VB ۛ,?r t"aFKj.U8 X5J$(F FDȈ!RrO=g yl,6#|u(#/ J %iͦ&tG厫 ξ=mf̰ZD#drqt%Rc&R[ĮÞ827U>N*t997v |7`Q-({ĵwUٻmW n~ 2٢ؾHФ_$q[l$iؖ7j$HM4c[<~y<Vɨ7Im)O,@Ānd1IێQNBqP:j6.Eh!iܞ˚&+rt\pn# f?ggxZM`Adnf2= w>y-s}5즋&W/_,ۻW{JIK3/BCF!- 1լhLUBJg6T"Za O']Rq|<2AXr ,("ڣ-&D;]'GlBk.˷-'xL2ydc9 8S>8m__PI\P?H6򂍃7_r0|rd2H2hc]xXJ{t8=9_ .v_RVZHj$P8LH*Du#u} |}B9(+ tWI[|p QC=\L@Կi/$-! bGAʹǽėsuwNǑ5>?[̟~A}zuʱ^NbFG 2(PxiԙR^B](d;w\ڃY* Wf0,U䡉Z5kWr#V/]z<0Rljm͒X`\|V ;- ;>0QSm;%Vjj낊o!:MhD=*/@k%t}];ٛHZVa0vw/RB;CGTtrdDqQf@B0PFMY2o CP]h `_p R(Ha)Al8eOI!!}85~)D0! (*l=BSVg@ #]fׄ.SBtb 3W:-D24$ɅA5?e,[+?{8O/3.Gz0{IiP"IH$ɡSeiyA 8`c-}w}?FQe&<*oz:9Fг* ?ԵiKr담 h{uş@W+8 r:qݲ6/"A $yjt!.ٶ9DX >B ,D,fv Jt*BUu>Vu` TEbiD[Xh L> dDR6.EV]Fȷwo"4h_[{ra5_cP Q0r)(DXAĈYc͚ñ o|9ț3Ɏ9Lj,ׅyvF^|xɤ#k4PDoTM{ns m$ iAn8)@M>~ÁESYNx,pV9PqA4M(Ö#k41Uv tD}4vע3M81MogV)s8>di i}E@ҀWnT&= ) ǘF]hpV2Nh/nc'MIIKf˹W>ȅ/#8>Pqmbt0 PQzWxKd?%tVl!Wt œ|Y7O%|jv^A[U%Rƣ\ ^aB^ںg^7wqaKޞQ{/1sd Fy݂q-*s fzv5wB])9+v8"DR&αEJٸy*IejONg叕2kx!vY%BMݰ^ZR%z˛~s-wft0o^>)Xnye"ܾ\%}ZeeK1,50f\L62N8ET f1FcÀEWJQg 'ڼE az]> #}*)ҭ7ԜcpzQ,>+c~ޱ_NW_c@5]) L1hSe^+8?ֺ >*.t oP3i`de fn6:8,A_VgH˻dm Q@nls!p]qKN/4^g?ծ^j X$&sލNJVӁ@=J9˜qs0~{£Wg켷#`3Aֹ [%S-ЭHR.^`dA>IЁ""VE(3ޟqhJ0N ll<|4'w -xZxYzUb9"e2*<YO"`Z9(Yu>N!n p2o& ,p3=!`R0[Bu $}a[N tMg:A*VG>Xg-@"{Sg􇿄u\b 򼶺ƏmZZz u[,rn =2֧Q!>F܂CxcAϮ'_LJ݁J̢nafOZZ}@ܳq o7xcd @co?ξ (CsI*( ^gb拖$9IשAirԼ{^.4sby/nWYy_ԠjGOV~qs.5k](r>ZzIl%:}9pv "pIo~omn7 8.U37c"1[IwGʥ&" 5ɽyed0ٶZz9~ʆȋFanqdFK ,z|>PE?Kb0, ӬBޱOf߳cJ࿅Mq ;GO ״ӡW 3{3疚0^GQq__^襳ɨ7I%m?pA[Vlv—F2E]2-B" k'곙&J  谺_7/o *!ZZ3~V`PLPYcg/l/bs1+ouiVc:gQO6 Xh!. 0 (g{僣gN8'Qny0>)ɀ؏/@;Wen>` 栙@d&&w)AmgkTwJ&J&|a׎byTDr { >/ e;- T/ q%]lY.9/v5zM ` j.R|j<`#3CҶydBB|L<1H.mEAp8h 1D>g>i)3NFYI a麨4A-PȠ; L-rpdYw釫7)}?GlϽuOb5T*œba0ZIXL??|FWaʵw1&$(XJ1 #LtLNB +uG ʹq*HP0Hc"C1SC4KjlU>U gо,e܎EkF%-HG kZ-TSUy1%f Ժ1FD+ͅ!J02E(S-p=aaY*S-Y=v]б+Ib.T }fO*q{0JY)Y :Z͕Ea9*ʹy nN0P(Xj`̸2jmdp҉ĺN9,D"T D:,tm ̶Tk~\r7:VnU`2黕!+/oן~ADݼ^ϼa|k {`d슣}@,8~Q'~KAlW}*{\-g?3 ƃg&8z7;ѶϛOF A$VZ~C8zvN l0&u.v6/ ՁZiuV[`I=Hj0%UM\le(gXkФV,cP Ȉ c"\HaOPĺVY׈)H!YU&ZadiNJT*+H(*_}/1]few.ts=Dw,Wo嚽ŭ,u tş}1e}n1lק7E5ʸ[ ?}" ̐du)<{.kVEAZ,`x Hp,VVҙtl!禞iykpk]>XU,Vy(- P/%ZT|v|QVE% Vu^2ae]*ZH%EKm!U B\VK^!kE!/EL$!yh=WZtkb/'!{̡31]kUpz;yGn>3A- V*VsV螶/Ȋ@ܹ;ѡ<>mt8c9K2W1js/z-0Tʙ$ z w?6lvzpQh:u8E;tAN2)Nz2tlnEl@DŒ{#qb ""dTN} xQޏbc#uy҈sTpVߩϧٝȌ j9c&fOymAq%( x]&'vw򞙓̜Y(Oմ.RFBdDʌbP? UQR D&$Z'cq2AO2-'dCI z蒠;]{>聠* zIRG*K+X5h4lnYN*GFV׳aƅ%TUY&kZвFQ%irvYBfnM`XK%xDUwI"T'B*E T+{RP 2+T^]nеj9b{_qZ+ި؞$ùK$L"qf3HmVot;sǏn6FgRBǼ "FzV6Hʬ_c B+_&k;)KMO)*TԅfA!.ζC ̘6KʴPfyL\Xe\#8%Rh!K׿l܇ QUJ?P EI FYEi(ӞGUNnpy*g|,xN%M )R±㥌`]8--h||E"m8iS4l;FP&m\o뿛%}_՘.V2?-}q?!U9 FX F5?sqB(+oƋ68zF6yQF˵R("T'KȭJ5d)+bįӿR+h*NA$ekޔ5gsQ!MaRV@Y1G!|K#$  ;3|ջC}?sᅠwNZ~9t+l(\.ﳧ~^EHU"}|@C3B4>Tbuitu-FKgr;=|  |-R團k~&9c폧g {֚strŘBsv%hk&V$,6CCdtlNeN&O'Br.d[Z1u?~yh4kdz#txA]ת|+z~_jVfZ;o. o/ 8 Dnb ±(f(!wM)1e;OB6|(T׈R˯%Zr;kH8iv5\!L(1uBh omBThgG㱯4ϯ9w7|:r,<&X[ެO[%:Ov~xi~ϟR "SXs mEĖD1E\tt vW.|(^=sOC^QU3jENm33U_'lfM|+x%}?1[;O\^7z˧+?lךr)pvnUA%9$7f ˨hS>2 HLӑp|vȯ1ao*Y+v:˼]2'dd%5.!1IlEIa"q1"ՊbHL`OP#%ڊÂ_LHF\KeGh!:BxWq>k9CYQFn2A+U5u[}uw>M 4>/4N:v ^2OsCŢڝ~QVXj2oczV../֓ղ.:VÏ7zQ,U0k0c r_m eToojtܶ|5oD`fM?a|v_7P;%5u]T`د&+ 24r z؏ r.Ģc@:K+ `ݝ{ah] N'(yL3P2s%s4Z;wpS!--eǷ*ٺ p{H r(Qs8 }YY)BMBV9TU<5l~l S5⁩mk=KlEM[zMP@*ϯ}6X!{YLD4mw"S+{1kW Wl~ YsAY"qoJ ]⇩\{Ewm/_{Z=wsM|CFO#YXYM@1 Av}`OW8blҢjDO|#Z'kw _' r*0h;db14BaZ5gχ xhixNRvz+s_A=w }խZ1Toog&j_\y(Y̷UA$xY3U9 L\ E zgv͗܎ } 3j7G#8E 2"%Òc,Ket)1+3Z2pУ 0)/#yL-P\JDﻥ|-ECrPSx\Ј`tnMwyߙixa<TAw;$t$܅e_ ŋzͫ/Oh/@ckW?ӏo.m勫]?^o)(LO`R`zb'So޿Mjҩ-˯Eke7`&t5~ϦpZK.gʝWʽt<1aN/Ocj4e7] Qu׻ &7a.ǣ`ؗ37K-R4YmJuvoNYp:gђ-J|CQxHNMvNyI㶖3l[~~0~ebƒ7N_xyqw鋯nྸay&LoM|z6G7C/Á)<RCRUYn 31v0`!ҫ_N@Am|1%^G_Qyxj<-ۖnK> 9z>?/A,*q:Od]zq+tJ4dwR)LOo꽿N_~4<[\ylF;}z]70m[xQZrde2'UP \8rEC{2ο|ezZ MVd@+w?9zK)pA0|r&K_]Ecܰ0-7R^HA^a0‹]7ޞY;y{rQy0{-;1sjeuڭx/' qc9}1&p?޾}ȑX D* @Y\2JK$ iÅG!0QSTR=j`XHaE"xXЈ$D\kcS]jt j@5G,y/G8ir P{A>PQZ`gF VBT6Gi8jZa87sP48#XkhA4k@#c'-'9M?G$B++#*X{?,up!z̽_p|Lm pIg9X,8!~5eе\2D]AVѢbyPPPwK!ؽ nSEʱ0Ɉ80pXkk{S/zzM_T>'Y}0/af}c4$Qa('טrB]SŹsJBUi+0""A0$j &`f֖/j$M_`9̵z3Hm!kqٚ[?Yf ֜b'22:Ҕ 2.6&`Y > eL<(-&%ԓX }pÜq +FŌ)-]LCЅ v&eGϠu(Ϧ>[bK16akB`LUp;}5[,,,j J>GTi%UI qͅ &˘@X "ԥhCsa0P+r^.(8E#I}ic((*&LڕEJ|`f)5[=[^}8ݳ"}Ğ-+JdqacVZkg*8nb "'MGVrju/>lDS)G`[|`  "},*?L6הĘ E+v$+?7{5m0Kk)j-ui:) fc&@}jw7o>hVD]i<'֎Hf]i}NqԴSgkmW)V:(Oer>*j&w&j1-D 68=GH R 6bbE2x D`6@K+Y]SN"K6]J.iifk~gq6ĩ 7TsS-\t $z`Ǩ Ζw h-{ Q RXxUM}Qf;dENg{;rI wMc[Q(ƏG%>1J` $d}QKwXw^lISWkL,d[_w]x#5n(7 kޕ`x}iMvZÅ]#B;YT-TjũcX趹%)=vE-jM?rNGl!\T%4yI{qT`,/gW>Nƣ91y.{ñfvGV['Xtfd%QVn f!AH|m]X2W/v%;d{%> n=4fhA{'DmZl~KDŽhw!'5? 0K;kOZ8":Eht3)]/x,M\/tæS/Z?! :2l~piSL>%wS({eI]CQs'quB`t՞{+K|v2?R7yLP\>#pYin]t;Gt{<ÌJhk잤˧$ٔe!bV? k :c Q;uADoA4F*1Sre6kxexV{qU|-C5h?*2aAsJc*+W>3ޭuZX}C7F]A*fϿ|[BgQ@ɚd̯+'1aoQ4ݹs=Ɇ:iƕ7Xq>_^P#y>x7J/FlzC l3Y٢M=ڌ/6[1]lz׌-vdˁ۰[e 8'1y)5=+Yh7&&8 [XQXk֌Y1݄F KW-=*!nAk5$oܜw4E<%ocSشto@7V$[ PEȁ~Np]67ьh;culice@$Xvk5lB((VC jԅ̮ +}GmBwWdvrs%"۱RMq0,7z6̳6)Ǜ,g)kqh;40cl"hf߰qf]+mH/Q} 02`gFOlّ4v[5)i1Ķ.֯J}rD!#'NL3J0rBZ"7eƋ^΃OMsz0HZ Po_ObvjYLCDhO׫I^ RNTL *-S0IҺ0̗P 1F~-+7&aZhZgxudf+qF )9^<:rQ_ǘ^͵jPKCs<5pD5)pxeT; ? 9m}㏙x_\i]G`E3{H:7*[㛤+ߝv&ڙiMn`c& t{ǵ[ݗ;.778+[ W߲PC*@X8lZ RGI+:/g:HR VcY5 1eᏺ辫[V\2҈Ρ(ߞEffbAZe i{]v~],>׸ϐsTj#/ f[hNWX F7.0~\> %@vdLnu8ʧ,2w3ehӝ%<6 j9@>,wkc&6267bܛ>y= 5և_Eu~!0[H6SUccZ? X *. K qg Dv\7ڃ0ќIeD),C, D0Em*fT-BLeX☂y-z Kho:Qָ;J>?˖Koe!f0( rc*XSmp6hWw 2Ő8pTd5|oګnk ƽg" B/ݛ{2hʔCpJQݗŽmY ncTq`+Ud(T^e ~oZo?zx (7 >TbjwCT0gV4hm5B)MTKb[Qf_]ѿ3[ro7Hq~+ـMRco_*Zi& (A#N$X 1mᐆ JC/ QRA);eEpz+$Mi9B 8ڢaP%lP ~CЕҪÏ1A: tC < -43MhpA;(~v님\"AҽFyAK.@3N$b!-,Ph>շd9_e$c[Nj[e42@ ARѲ5N:K1k|)QA,gպ-)z)TeVqh}v#NIpɴgpC{XWչ0X,/ "wj=(5k}Dvo,![+:{.m1Ez?:yFkV%Q2dxڻPGug[^"Ӆep"!IQq|V+lO )1*ݔ1|-j RAɏ$I~F@+O2랆%SCm,>Cu`Ьo1ykCpEb5Vs5kSAfÕ!//aƨ 91C1$Jn f[ϣo/ߎ=⠥29ՆѲk}c4i}j D_#L` `]5QJqһXǚC8&1ZHҏNYEџWV׿X 43I(>tr(6Pv.!v]l< B'1F;VƋpMf70m%{8]R}0˹ .˳ԠK4rVri,)DJy;QyZhrk$ &P߯[2yߖ_e"߄i'~Jy*SH"jĩ:1)gW >0eIyƕL"RPcMFd9d',ߒ?N>f4O&h,,IӅN>.|FGpυ0wiqd)V=_|Ƌヌ%;7՗wFzô)u S ;1X昒BNVЅR48g'A{(9υ0!ze 5MvyC`b?St^z3Ք]I6zc']ɤ@qi9洜csZα9α !A#opE6%5xM"@!F<$~PVlkP#kۘ*N_v}ok΢ dqNjZQ6 NvqdGP^n P' |.}Zv>-KݥOݥsM(#bEc3BԆN ]WpkHʖ 5/~\R¬D4UE~XogT"yfr҄l[,ϛbPq!3GW%ro_qvY-(#>͛2x3gVpϓ}^|O6@S<"Z_&߲exP@ž<:DSdÎL1Y)ݹ)b[AsaȹNWؙEhW_'͘uvtit8?YJ{3_zޟ;|R祕p.,~[RRD"D!bSIU2p>+ý'4Tgȁ'4{/O^[G&4O5&/'*;SBV=5cK!T-enVl4dGk%muuϙC}n ␙^+_hC xpr#@ n xFQΣKR (WQKf#@=oL4n k[*PQ懊2?TTC9 lQi@Ć@ϼet'^аg&svg3ňU.Vk9ZiV5g]ˇ,Qm 5D!F)8 H,q) %'m Jg!QB}hAxw&a*5n$<(4\%Gi hPCh;%%kqm{̊eV(E/Y5#7&Z@KV.yA^(gu4ˇe)3_T崒ZzW79{QgγT?^fEdkkpQ4%.@IJm~cDOr>??ܣW $H**\ Trydn/ Τ jyZlL㎗[0}ࠋ@͠1|HU ~Xc(|+!Vw{w̍`?0n&Ίkqz3 ,֬1x&i24} 2Ʒ r@Q=i-6o^-)%^UlTDWCF}Z TQy;] n<8 FtFh C·;BpKhxn]hĮh(%nL1ok¹"}fQKi udNVJT]t~csweut\Pe$ͮ.'WܝWBO}ʌ?eȯ/qY!5!Ygמ['bqo 6fE:}-zukވj;ٔS-vk_5 8/A3r) S1[eR >kU9)rh]ݍ ǦCEH\-bq1/JU9o(J%hIƐuP%&gϷ58gfA£(ctIgAQna?ce'P tP BݵkzѬoyBaY†D3WHDVQoGlLNh%$[.(ˊhI3Cڅz0b W);CLXmEbsA`c#)Xw;x= ,d5xhDM\pa YRCmTqwT${C ⻫ߑ&,Իy/Ll{4Z5zJyTs%"I>˾'TR݄bK-B̰ǃ 'EC>'P!)Y4WQ<>O6hp4?ЕXP-q7d1fPQh^IweqI|Y,M)@"^i !OaEM`F6fuwU`d_6|Q(\px #lhD9DKrJŁǒQ; }߸}~D?ޗšg|ӹ ]?~䂦{?}y#~W^%z5E<*Nz7.-#5~ڏı҂0P#Ar4WF5Jη a2l ;KǚG'J>'w9׆|t&1eww9Lgt3s 0ifLv}b˯Z" Qk<-c(2?PF=>?dM)H9GQセEX+M ǜ vmCq๙m7Th@5܎jVo-?J w!D8>XW? 뮭mZAl? yv zW?>+ku;rˤAl==ꇱU V{2݁[jl  ? \LȘO42k3m`O'y5@#;8/^9"^~ G?  m;y{`y{(l<̓tkf̓1P\`g!Ԡp4aT8d9GGz? q4oN `mZ%`9W!zţ6l@pL!%R(Sڱ6PQ>WV+64"S&'!tJ|cv6⑐uY'/]K>ˤTltS*Y bLVއTZu3Baq$q9&a+zP͕Gڍ) цѥcfeAufjA7 *a΂N id1zdbZ0<X6J.9W ^I&ĩ0-H46))rrdΓIVL{Q6d鱟I[\e+r߃lj9Ғ݇XщГT`Mxk$ZE;Onśw. AKc{EpG=G:^zlsMW-WL) zb,++mR~4޶9ϔ&!}3\J(ҍTY[BXK6׬?PX--]=[P`R2oȍc ^m^ㆰQieN2X :,eZII9ԺdR"d[A-!(ugo r2roû;5XcQp]A,v?iQxapM{|@CeHv:TVmr4vo=tYԡJ ו)uwJ Fo^ nIfyF*[&V^cE: BU&K,hKW^ICwVo= -b8J|ei8ƐeNJrtK5Iz W6pm=ڑ[)0(!gBtwKD<:m{]Cd@5`4l(61[/ci} I5r2l3Tۘ>tSBlN㲚~K.vJ?g^3v\8{Qcʐ% aNMTib5#ɵب1]yOk}t7ɑT&!k`f ˉ0&xZg܇ (YIo4<6@ћ̌4#ؠA@',zX(K*))TtF9֪D{Uȯ˥?'Ws{[Tdbڗw=w?Lb.jۂ- w7<̻iCrIVr5howÏߟq\-F >;iBWƙsIA;(Fk +@'ZFu[Rlph{%{eOf_8yý -*FMTrbx^S?䷬@UR'<6ub,UN}3N"O-3Z1cπ'Ax2 w/W#fg! thɡSzM6'S;ԓŖoҊE' işmܧ!Fto5щrUqt@$A{.~]5 5;^jn1O0@Z3>^O ڲڽH:ʼ {jmzК/etڄQr3W??@Kog~Qd-m1%YbsLÂ|Y~e^cp~ZPC\:ts:%w\> QZC'AmJA_;v W~˂E*g%ojN=w{y^x0x'?<c|~,N6Iҟ# ,K22d%dW0/’V9Ǥe: i@ ֊=#xIȕǜM#RW. c&V.h_.Z-LЂۺY4jbaQP@ʮv ׂC%cK:qQvROT[eT@AtK`M8a ;!iF Qdo샤RLoUڜ\j1P\]i]43"DIOoHZdY94 1-˷8g@rrɀpICAHRȵr P]62ϖ>Mq(mx8Nb&,nS`e $h)W8W,,K)(p_D (g$ |:뽋0c&D ʑrL { 꽺rz~r.gCSbb HuSPeI=s}?ΝcFX #N)̩߮L+QDYA<TjY0!QGK.Pr;\\΋BWRd)s_8,kȬ-0U7\̮j=mv3Oݗ]7m~o}mn?z hvhkNg(kmu;rU=򢄷L|BqՍ&l`}Qmp\1 ][o#7+_fg'yC 4Ia0V"K]/YlR2HVQn!m:< 2#3 *E#oțD} (%` ˟dfiri@CLrQVPd+J>V׈[lVx]?'u-rn8<_socJJ!i!Ufs@<8i^9 b_n1Zr\9v 8obGIěB<3{##Z?X@$*1c#m~E@KeHe F`670 ] 412ØMZB|h2 frp#ic+k5u!:l*&zƈ/b:4/Vc:u^N'44ӎ7(+%;_|;v3A G!wpz ? h&ndՇk.ymZ]\gVO=f>mѯ04y4L[vNk8FIHAx#:sZ^F='(rסtܻro !bD4;ZԒRH דs#z'WSc~\"L}(B@GmЭKԥ Ϸ3gPLL+Y,~lrk]E"B[},r#Ҙ"C6G#C670i xD1Gpݎv?`)BXN+]+ȜyD|Db92`]3n%gex"R8 ȱ3P>6pyF"꠴>>(.DTt@bvFVF3s PؘVK@܇s4 .:auu2}B0F:z)[Z-^K)!Dݽ5Fruu2) ^J I]}'$ $q);Hإ; )eEz460ZRuфZSKj֘_>C (&ԼN=A;8VdCTPLqHW #MbQ0jn2jxvn9A<9}}ynTlzSHJ$yJH 9&$@-s ,ҤrPp7$fd%fSY4"/Sg tucz>Gueh(D\C4!)J(W <Bi&HHRTf S~!@!ʲ$c 9 .hq.Owځ,Pm ų"PL~c5>"Pnkup>bFvQzLa~4̈́SS)C(EBHKP(pƼ@(DVzEP7aWՁ|'J̾*aV[_=ȗ8NS3I]]asY- 0̟>~޾||jrw /OaCn.@lq٫Swr4/ ǻe7VV53N拿ߙq9MӁ%ۃBKA@Lq1,oA= `dҁud=)RGVaV!Z'CX3 Ef="`X/icHP&LƑ*Ch! l r2z Rjfbn 7o[%1Y<7~1+~ϕbʰZxVMOoU+ a}L& L18N%ש e/ nWyun3n\$ulz:{~G6/7owl6wp߼Ci3]Q=m2Fw w'.;lνCA묰r~ܠtr~wbPK 'P0,Dln`t|C@3Y9k~hgnj 8ĩXd$S7/|7/ZYCz, *օ0v)B av rFă(\ y6=,>%-c~!?P.}5ψrqD],8D%U  S]7O̓x61Q\?S":讹HsѺΉyx4(D^zL!Cuh@B!qU.֛Q!upS&f۱s:{qEC~K!PƄHnK 旜u}W `mb{CO"M )Vv6ň-FmpGLjScg 9 8m].V/+C_5XHNMˠ0'@PTzToKjuA T%]_E!we:>.T4X'ݻOa.|U;!R90+, ))HQr6N}Ш@ vhuYtYގoIF$;6Վ6}W!p(zc[7PW1Ĉp0Iw=, p(+QMmh3$u$Q;'2b}=Y.AFK?@ z'9ʼn F[nHAOR5JًV܍ۊ.)MFA|vc#0ll߈X<0b͉HFϮ~,F  +n(.ei7( Pξ”3G1ݘP\9gy'}{sQhf/lkbf;c] .BiBLܶBi\m7tb5z[}K?>(uy 2ڽ沛 2-O9.uNf;3WKxXꙣ~A9p>i?pY%"(jZ2$kunfZ/Âuts|\B,!$}$ki)X6k<#1TH"QEi6|XI{޾e.(%C.E9$NԍJ8 1_C~Ʀ>ElP y,)PȻ#J7ϊ{t~VV4[sdYm{NuA/g-} tӢ]Ңe%-ꛀ-L U^ƘR=3թHͲ̆-iQ+ptN!ߘu_Tg  ExNAXUO?V܊̂U 0[N#-R$JbtbΛL*b߳r 2ꌁ{ߧ2qf&iag_ K8' BJp~|agUA}Ϣku~ p${{p.`8慂r+h 7x63j|/(4Nhq1j9 `k>/_ 8{BҚBN)nEAVdCG!Tm%? o xmnv}f]zvEcx3{k}3'>Ijpg(H#:Hs+s؊T }[.got yQC匜-)ҥR:~\|40㾱v{t[Gh \i+/ث4pNhEޜTgIAo7zV tTAa;[/YoE h\H\82fL8M pP|Rq M`B$}?sV aN2C&IJ;UL$[KpkBؑ&}> n浅44_Uˢ ~.U66K(1>聕6bb*I2ee`k+^2~&Xg)=C@xcZ_W!EQ25$q+^Չș/F=w@zO=+Z}uc^^$L(;{v Z@_eE38|GmЖ3ps 9ˆa~&H+Ͻt}`=3T_փ;wQ̫lak{4ОP#`u.@,4X)0ŀI t 2 SF3QBRH(`)8y5>ȉ]ޚ ;j`Y D:`,nv+j^ULan\$ˇtpWOeWYB8 NPE.󒪫c2ҫNlp~qM)#EjMoYյ7FPD}/L6k|}@bu:ÜHEʄR hVI%<\=)ʋ?ۣ#Z拵rlVeٻFncWXz9m]R'9΋]*`hPBRkIIQÛ4ùp)-r8n`3x`$o@4:@ˌeN</iAP-f8}w\lb"p`eG Fu8T}4݅|ZY`fp9xS/ߎoGx߃2?=^YK=S~.\e[JykVR5sv0j`a Qe\(L:[Aòcg!d<OBQԊ}2÷8. ]CxǿU~X;u^iO95*:'N5z!A>|߹mVC"ICɃsiz#C#*%U/ 5={STt%0 jcr|{yl>}\z#CRvNOVouR7յ(ʰ&SC/B7|B4RjW1+~JRjhy*'!P|?p.@,l<ٷTBjPI}>P5^] *Y~0qN"p;R7mZb>\%k㥖Qi4Gy@,Zcgb* }"hrAR1F1 9`?H wE1su1˗fRq'oƠv~1\5y j~Oo~Bu< m# fXi!%e߮%i, >81m~aQ`Rmn¼eOC3O:sQ{kdC [] LϨbNvfo JZ&f #g0Pz~` ejS'V#FC7}[g [,ZM(YQjGg@dK(t[LS>\߆eAZS?xpylx3 )U(ж^&Qxwe2 eLuɣJc`4L0tDD(䫰K]z)?Çݧ?~ٗ2G.\a:VB0Q6בUQNۖލymIlGS\IşK0Ư)NF?ϗ͖~O3󐁈2ک,$feDSU M*!1#C<7Y>{Ep?l*zAuV,Ƨ2uR`;=f]jU UlYo-hU8ͲT6:ww5eF:]F+{gwR=5̜C!M8U_Rj61M[~s bzr֔}=9G99Қx=H,/&*aFWQDzdHLbMcw. L0[a}=YTa ȂrӴax=86k xCz+;:;u܎k)./P+s+Gǟ?-tBIZeS8+0 es'6y ݿj_D9>-:ћpF5*˅wh,4$5!D9w7}\( ^JG8fOch)Ĝ( h\HF!3\."Z,p|!Rh 2*qRT1XtIcXkX-Cd{7 6 &ӻϋNou 1WT,GHgWoнY=MZ֥zڽY={A%%$( %PB4?YW!O1⒰m3ҿRU[2XI h0?3fщqu{SsbܳNIFi).$Z9B1 x BJ''F*І0.H-C*o\/nz;X5[eɧת`r9l-Z^gk,MZW()Z(PZnsjcNبhEt$[.bWCFZlcf2{R,5g-jS(YO?&=Y 3*s=kXcshväeQ-R򽼵r!yk!]l,2hV6)Eov(aqJj´ykL]Z~4Uyk}7UI޶$l:]YqKkTEND NŐ3 ͙.NKNSSD 约ƜZW} "0A`CAX_H,7\J &@} pJ  =jOѰ \ɼ1mt5C$ 8)p)$ۆH,4TRث@ g,FZI32rIvfi=1=,3m'eZwy(`)c k^"TOKG@GmS3+P 5MkqdNIkq8OHkݑǗ6+RwB_E=옾ѮH_Ua7Ww>w_q**uGĂGvGvȶp+PKUh/V<ҺזX];dU(5ǭE (OOjEtL5HbBUDADA_2¥sVD ykO*&)BVS)e-U<B ga 2OHn@6S:>FCke۞pЃ MYj=pdRyΌ]/L7PKEfTHX;~$n˕Y ~OtIzdR9ܥ"V12ˊi6v)aNڻ7e?h̏ K][Q d!U˵j p$iԴ:l7a49>p+F;16 Ŭ߆i: *5Ut\1Үa:] kԤl/4ݠ" }e{=)ȍQ}#{'0B%_/@/_{Fۀ9)ƮkC A)ؼ^LD2}az$Tӳ.3I(&4Jʥ 5.:tիwV7J]\qRU(4e20hxN߬TlPō" xYْŬܺsFR33޸'Г(>o:[CڞyDx.@> f,WI`;b,?xM$gFky3] 7Yvo4eJ@=f^fx-!NKƋ BC)B<p<'~ i`kyfSdVF'.Ivt2 @r\ax)p#~g 8N{?d, zdFvU6V8p<ׁ@{|Q@ b::TBMn)G"$cYVK5ap?YqUH˥y.uw!_廅Gv0q9M㠿r{OVpП/x$qa&qxh/ɼ>ʂB+:-BZr[Dn{C3^65V 3nͧ P#pEr"x/9A0 VD&r_D ۬U+bvGZRwںtWjRq]s8WT2{5C|UĹJlJ2nT :ȒW2l~ Ea"Jr%_7@|[g[O2zr>Q@=5TkCP"Z J1yBx洦4iiK#0m&K͐5T3) .d=N?ĘRLit{u]ϟԗJ俯An.QX(65y$g £W< 脰N脰N9b;HPŁJrt`R# sQ=5.ra*Fg?C_q}e}aBd=ؑaQ@A0*}˩I˷bL)^LP>L?nhH)`vr\&F*h`Q\d;fALO¥P-2Ł``\!IÒ}a:-eTmc+S)QL1Br1፝iN~ۯЄ͏bX{; UDRN[xbUXhZ_"M\H6z / >t< "Drȏ~Z~x'}~3.J\*NBoG H>Bۄ *&[wW?BEK-ɲq\P 6 ۋԔ#p nfc11OX "'D37bqʅ<Ͱ<g!F0|M R65.YN#| HTBv.%|u#MbuZkuњ{wQ9 R!ܠ֩)!dz8K,Zlp8<#TQM]P PIMFe{kdq`4aerkP zF4ȚggۧE* &ufyAb& Srg .)cz B@,iJzx“4"AH1_1[\B'S}t>zpC6Dtڰ80A(/ŃR3`m~R[o]ȅ{ V{ J8:KM'F:kuH˰GcϪe tYSG~S\:Y{?\I֗4' Gx%@o J2nm^NNZr K#Od ^0(w!5Н meővxf?{;go]~L->f#6f ?yI%@;1:NcRMyFؤt5 Fkj֪ ,rqC tAyC/޼}7Ơd]u_hiރL[wV8<`*wO^90kYPD>;L.[?XlE Cm/'A =3“ &9N,jDڭY1#joAoϟۗ7~9]-`~Ww^}7:_^!"Fڭ:n]{˙0M p&QQ>&hz{!L`/%Sju(ZQm,A=$d!c=.?h=y]]N ~}/)0Jڶ7G~f6>Q!U4/3~RGCa".(|錿߇-7HM/v~ovgZT"vg=a=xg?L|aP/GñTYC& ^$czhRVFUיe LtӖ ֫9c8V<[glu*4I8οklODS ىu֖c[5~DxB+[v> .PN!Rgq2L`.z-ȧm(Nlx)H1YKEdyT<\&RHΐ^@|Pa( 8j%ޡp%]½cF T= B9]I8Ïd>yGr|R;l+VkuZ1b*Sٮ[^Si7Xj6-rZ#KJpTdeV3-^ȹGwn^!RQM*{,J|' =hy M !W{erEr=Qф<9B6G>Ʊ21{C~=/UV Y*so߰=J, ͉Zhz؃h\BA2QR9TL;kXJR{ƌIM5tR#%T*A7M9*` `\P@٤M:1),Q,KOpyMJ^3Kԡws*Az \oZGE%\*T7e# Mif]MDtq Y!=GvSp. a9Aq^.UD8~"fR@,cZz8oݬc#5蹛in%LQJj #Qa%˪5 rϣرX^|dbv,Vh[#tm:6޹±k#KS=_N9]l#Ž QBG>D&#-ـ;3;s&T5"00}(0󘑚FKlsXTM ds54-yRtϤޙ0ͅ{rLyssFȾ]'IsF*#XЊKvۏK8牵%"擩۫ F,w{LQt]jc[Xpzr⋺P:t>/_heue:)6)_s^;CƂ>vAd5E^[?*rތ|"ZK>-FWǠb#:uQEn['Ti`ڭ Et{҄k=X7@i<|b+;Gek;EGí&&+Z:UŚX߷'NUse!%.q UG eTޏB e 6f/3o„h,bp"ru?;7.k=1&ϻ"vP}Xe / C1D #'.b!H 9@9[Áӈ0t6K;ZnlP_-5ÏlNӧԒJRӌxL5p% '\ cF Flw'bb2-C魽&QvŀN}pN u$6*/;HvFXItzxYO=ő 8)*ǔj绢xQW"rܺ3A)/.^ `x;qڰ%1?W,W? vaZֳku;<'lp&" Wlqt0bQ- Aۇo ~k }{]#8mDQ'@ uMcr=6yyB'}|p 0 `j%9+iT>UrX#Lg?u挑%=;[ՈPw/gwsxS[瀠Tutւqv,Y 0M~.?$!>৏LY!,yv0<"!`H) Dy3 _@M7Œi䫯+x4f\T9=є -us!3 5c6eJ a+?I)AriVVk0)x|GoIHƎ+WI߻uW`R"c%?5P5r`?kB_/w65lx5yFYkx a}i1:7ͤ4F\niƱi&d4Gr_ "0э`x ,"oHE:RVޘ5䍮{9 ﯭQc*ƿZRFYe2=3Ybl`fvdz jP9?yk~*_bu{Ws s}Ėn,U}囘Eo1<Șc,RV?:u4őfXpkRFWJ"Fr6O6 O3ap!)3#=5@)L3DqhN:q֤"d3BSQx"IYJR#0/I ȠQ1O*";/.@SNv=`a eLrOP5{gdgZa8S^rx:!錙L,Ag*GK*0Sˇ9r >}3+6 Y=mr?"4)@x}W@z /"]jp*D.ߏ~Z~xxߌ6\IB-3Ԧg0Eݸ߇u1<ޙݺS Ufd? u`c /s$P`HF-GE_\7ҏt%~%¶eAO *@%qQ?U8B[q!lmKtֶs%IDKJwby3ܭ(H=!0$){I6ːlC \잵&t }O/+ ,)9%mB$Su;]%NaTp]b>,[t|UDX:"W ' cg=L8gV): ؖ9!YV:݂؏kf1 `ŷhNP+wDcƄ#XB(kpVpH)$@Jr)r#F"7uM=*]R=D^O2hiRK{*'?bhJE)@87]Ke8(jU!oL^/@FmVZ/0pA(JϬq"Ij,yz) NH>Mw$a~I z;jz TfoHpCNK I;"ċP!yKIZrg4 w@Y =ibąSqE+ ?jtqhbGC똠8OߚL6~v-U\49i벩 8m]srdU[M,l unYBhrX]@ZgCͩdӻ?La'iio=Y*[vG2!7NraX"mfOk8fWQ8Cz S\LC%/eN geJan{K^CM4ɦwp-Iw;߮qTμ[zFHքqM)Jx]t {R11:x3}<;r?һ5a!oDlJޯ-Iw;紘e,һ5a!oDO)%`יy 22-4˅vYNYiw_m?R9fx˾w_\gVxyrEF~S)z]=EJxSKE__E#ߕ?- HѫUz콿k\,&0*,AOas׺ <w7ᆰ0]8u=ݺ;/' =Rl\c U .k 5݁؟T+x6IRmՃ]M~E-n'zmVIϞqc їtNRy~쾄XFօaS,]Жu~ BpK$J5E袵4%@@c9@2A>DZN*8S㐖BjC,05cYgL:9nR8Db@3B'8 VIx Dj|ofml\ &c}qkhK9߿̧U`UOy8ѷ~ϲ.=0DVh}٥ܡe)}Ͷ->mh}ݏO @,8+V?p3pl6ϋh8lwwFj6i-}_/XҶ@GЈۃǻ  H ~g$ kB0҄-JG`É,'C9siaъ+9%:OJYbNbdZF *ITF™t089JО٣Yy1Hΰ@bAZK'8cƇD )α*ze0HA#NR5B܇,7<\$K3a dqCz,r)Ebg @iAѷF[>jK֩M}cTc0Ee LI26P #{\ken Yls?;\#R„z BbA(h"ǗG(zr(v[$ ~A880c> k+'_%g:^ǿ:-lͦW)/g[@#M2h -:zbӤ AܗM*rSC ]y | _= 8o.NB0u+5m .HwY"v"r!($( bs؟v=q] 1ge*X Oh-pUgjЀ$cyQx#y;JBO"c O=<&9ŌrxV[nMlͯj@h_b8"7\H)\8u-:9W1+N|/x-~E1XQ릱?!ӴrF1LP& yd}FDK|? ~%PeLI)DgR`ћ:bp%h9fL`YgsŔq6РѮ ml:ĂVNh'!gĕ %QF(pJd8\K=g/Tzl%;N2cbt`EFKLek'׊A2Xat{5zeh!*-/~(DdO-<_2aԤyM|cnZ鳇G ͞|WvJ>_m])q>Ȯ8K8' B\g4C5Zo(QtY+\RM\O8Ai5[iT]N$e$Џ*.203xOSs2>^rAy{C= Pvd , ~?.9F`Ϙ_fFSD ֐|4~s2 鲧i*[C^J(xb>b*0hO͢ J(u4ŧu>* l T| D{3_@[?ޞA.%bܰUch!5hwώ?gb`ǐ#VX_Ww߬86.IO W?K3R~f\/v]z;[үyl;Wy4=W|&0x:Tұ xbX41WrX^sJ˛q *2)^6ːe:B/Z;=?گCm??؊}VΏC v(_j^hL_t‘d=ئFYWn#$O;ihs\I:'ґ|CO+e~η=8g2g8kchft7).^Qi_X2o~vvә}Z~We9O_=e>?3 ٙ?3;CXCa2O_CAdGoBpySJV?ӫ߆?2M}t6_џp|_t'~o=xbn}H A`fRNdȫI0MָBʞ&;f=RA?yw- %e:I14 7<]=V9iidFY.r)6\AFLyCLӚ`i`N-9Pz"w޵SM JqޘB ʼnb0ChJ2#9R F77G`._Hbd4"i:8͹ \ΌB~CĀz'sz Az-(opI>t2$43';ֱY?4u/en R $jrH+sg}޵q#"eqb {^6b9ʱdA]?Hh[![/u/!fc@yF#y _>9q%] 7.]^P /4^gtŸt8W o _*HƯ?[ged1UP8\;Gb"5+t; f&y#U?%#^T%|ZngS){mIə!$s)ҚFs&B垪=aUg9_̐R1iWe:3Wp+dQ#LǦ# o5,9p긓NξgL<'T\qGX&8gLjMW-(8 n٣=F&Nsڨ y!QQp옘-Oc2tT<8gWsԩ+=QΦm#lZYJeҹlE:}˘'=Ǵ.,2p_shNZ+9RYj27Dٱr3gq?ݹۛR7<>u bۨ뺚R{Ӏ8J|RsJ@|:LO5n/,~)ݝۋ_.' 34 Ɍ:v4fȌ)ŒN[# nXt~_9#輜ˮhe}=J=hݟnTOXw[vi?eǹy*\\(<&j rfn8mb5^@4"ko tO7wM)e$໲]Эjdп]C&]l5@  nTR~:ЇFﴥ{`~YcQɐ?wr&ȁÞG: Bq$.,sq<*- e~`A[$K/4|C2K<^_س_^:xW^ju5ز }Œ㜪5y7ZdPWf^|L)Q 8dLMG[uޓ$+$rj탄aKM/ΓCrS u&-3@bvv8qيoJw?_o7onhKdO_UP \d]ROLSYήwWwO=] 1XF1CE{tU3r:\=xsƟ;#3B;D6Y f̙k" d0NEUh9S!CH:Z_th}kZrT6v}eTPatgVhfë1G1*j<2nn'X!V 8۟\>Fg|-X8Yj=XBܺ4fY6wqoM#FuLHTmCѮ-5g-0bvn(toh}0fAq& UUxl[hs SjO.h0@dV=qcF*4B80L:]H?}DLuǕ1' זIM~B8Ǭb8x]Ҙ!(B`C :іsvlWX%J%ZVEߟͷKl>&W ӟ>\X`5JYf}wwzmVu|zwFNӅȐ.>~;ىup3.v›Q\(atoUNwhZȖZ"*Vxb;l8 ZJVz NF+B=(vhYgEi}->(~E.~9k[{xiXFWyl3OL>&bH*QmNRg]T'3$X%PI+L;)N;I$PE:RnbAǞYIb.2!-?@y:!V`cו\b8I m1BI)di1(8ADڈc 5ZfH [b6DrzNfqJ$ Il Zd:(c9Iwt -e6hUd1;Wj~)t #p 䘅.2,L 7(ֳ`iKOLB%M)V'ΗRh* YERGn~_Igi}KC`>\DK߿΀#+Ϥ_D;wyM%`r4U<3~wo݊ կ߾c,t?Cg>+_FR710O>\ZVo#~~#[bnP HWzlo^>t[񲠗,hWꜵ_V`2q$F'YE10A%)Go8ڬހf Ec0}ɠ2"I%hg'' \0(QDy`efz}(: UA`j B&m &CtK$/ 82OwzTAh8'Y>L(=]it~H%rurj|߇ؘY ̦c,>3'Rs_(|&3i ݼzZP ʚլAUX$LdҴ+ pYh\(G >HM V^Bkf&u:!$ʜ1c܂Q-[Ug nm|aIbb/G5+TFlSQWhK]qvgZ!ۆue򣯛֧FYYs< b iˈt-<6 {9)9l -h\rl&&@67!R^SL߲d+[גiF? 8@1ZY vUkޣa1K-.8=hWu:E;KMi`8N*+ 2TjoKv yJ+0+ǂe I)B&tyz>(P׌;bY,Ub3h@X,Q0ZTNd$:h1#`RD)Zg)WҹRXyÒg=^c3-ZUyS A0RQQAbC8qXe zY;qz,l4FA 2(T<$yCb"mXh#z.LƠ(}zVƠj#U|uQ^]U{Mx|UU'_[-%\PcHBOKF!]mx̖PrOՉ|c!&)8|,#*iGKK;sԂ"TGVQhD+OG0N -MI5X͑%ѝNZ!]m< Ŭӣ9KeNJ"OMhY! Wiq;Yad)\z+I-i]X#nMP'pHVd#N!#(F:FL 5MjAd%Pе~up}.{t9p,vYnɪu FCn)>珟1t;';7wֹuW;߹IK <aó blqѓ3s;/qA9{@w<үgOpL{W#\AbrӉ[N#M">n}#oQYU~AxG"yɴx1Uhxnj!N(d{ u/+j]˄#Fvə]k[fs+`MTb;gt-`*36; wVhftM @|tFM'L]ჩ?D9Cn5Q?X!M2>fR}Lp.хw]J^=(y#^+ }d"FQh+2m!FN8P ʖUvXͳk\z;≇GyF:sLJA{ w&f2#rjVY5,mJQ#HԣA!9v8*$8eR+ߘ}ԏ28\ _H@R1Ii2('Jh@hb)"{zl@R'$ Zi@QEٵLw4{ِ eG"Kvg=6Yrٻ6dWdE[96{bI! {(RŎwT)qxf3Ù6`YҰꪮKS0D#JhK^0Cs8T8a!P cޣ4ĦܧHHcᔃ9SЖbe8c1z߮RZ5ꈃc],F2pI+qd"' >Śy9עCٝ>9S4hMv5kdaD?5Q:lM$!Vޚ!ƿ199VJSg!$3CIljq00RG d.{(( Q~2cGBmUn( lW;?~";¤8TS kJOMR &kY&> E10Ä:3?K=lVƺ3v:PmH[%\uQ-#y2Hxhʤ\@2X. \}Lk ],7Wϣ|: L;,3BAイ2V[3^*$lj %>")= :a1pWu/='G0&|=%BL`0L!:Ҙ!4IPXP #ԃ>;r(ePt#䁔 ̵F sQ88PI,"V.m=4(F:Rv-e 6Rf{ˉI>ؗtJ%\r4˽ξ2t=)F5/H0 K QҋlAE>{|?TSӻaFbm34w Uw.|.ZGju}Ҽ:=\\>0.c_-+nc&Ws?je}ۗbk{h}:^nn/pυ؟f\ ^[XqȈrkDڼ,b65//oЈiOpx.ٝԴlD] *KG;ݫ+zq&tU\[WԖ{uGB+UM2s M3DH?ljCm[ OmlVD*HUoTP!sAHl3M{(*^2+|$>' E隆 : ADa*\㧆!O;ڜVy|)F*VAEHwFqPQV$T;P/Y$RKY#M\BRհ0%O.I=%jJb ,,.b 8}7hTkHWppDLmZv/'lFtFHY+ym۟nzrf֙.85 ?٪CڔiiF%:Kܯ9"VHn6!z >hA磻}Vt`Q.\]&|d^ ncu #\%j(ԋap:"X7Vu ;]gv~+`WVr +crE]>G$lPՕ' T,(tIJM*Nh]][1Wgp\0 *T,!4Ggϵ=]ؒ;kpTE~mU76&Z5Ċ[{u>=Aֲc=XqC>=8{LT@Zs:W71̛j6>So)ItR1l%FDt: {mehʹ3|⪽t6Mm`z&o#+bdLe>tJ e6Q?ݷH)*wyi)_Z)B_l~ sMow7]l}Ad7^6.yud}֌4dJg6g: 9nz;]|u v~n{;]ݪyph)-RtVBľc: ?FEki[/t2WMcgXKXn91>~߫ŸD)܁'a5zEE4Jr.r[3['9[$D'٪O%nNEnH+(Ңg{ڍ#۠b1(":('QS՘v/-ݪW.12A}q+[,ED'emSn[E4F]+j匦 P#@5_JWŬs9rJ/m~߯ a$'\JWc?'Tdl3~VKp*X  q|.~B+ ?ҕF ?ʕ1@p*] Z0t.~B+Ab=?Ak/~On%0 }\w>.ú&t/E` Yد"`&ïLW} c.psuID] fj._YdN|ߊ|ٷiZ1e,bUElC,h⽯P\jyr5o,ݼ] ȗ-gHN^V(w%pyv . ۱7xЛ>elts泙7W }| ]9fɋ%纡ϐVpo8wOg谘ª}1 i}o×pz̠ O?_pq7}q9 r4 i_NOsR1O%.S FT:ͭO`xIWG km8ώ귮ž躷.g"aU:$_E}ŀ pDƲ_ZXְ. (Setg8B+(;1PU,5@Rz lKe0,aX*RlC0{$y'ѥ$"e]&bI$2>Elc_~0&I.::Y@ETgd,{2NqMF'iөqaǕ%jP)gcKK1gZx"B (K0A.$RNx*e (7DXCo"Q#bbu-ZU/zsYպ~j9xwc3PJr&Č^W);E;gV 3(TPNQ1)Eh.hKц'$)!KQB $#I`,U6VN3#ֻZ |y{m8 ]$z;a}ͦABax6yw735e1[nv7t1JG6s^N5R~&z?lr7` ›#F*D+E  aVwfЬ|]A6ڼU,ERrRf[:1 bIl] ~疫ɗql~| HO%32a #<-@-7Wdh?_Wz?fz~ǡτD|`6GHPWfn4?=;HDAKNS@OIUor21 B&{(Ø6X|5ôv"!YQ%.uGiWX"ҤR Ѡ!4ם۪֔UQH:PVRD *E't~,!I;gET#i"/O`xb%b/ 2Sm|B9f:Iɚ'F&RH@:̓HSH 2 -Y$HQv>+ڐ&&aIGiwL"Β?'ZE {ɮGHIEj6Te($dT͆/lt-͸ i6fC-H *8큲8堷X4PA<)&9MduY,C ȭcyCcF2T?8xҾO/`>_~*>sm. ]0ۅ|'dw&|%Mrb %K IѤw麟eTUhs@7bGϧJ@VhwI0xǬ}VY©(wLO'Ɠh6>藃@UҩTj Ĵ49%`T'`P48u ("S^buWcA j+ H~gO =M/ Sҭ^\[![NŬDf͐kj҅Ť@r+}3Uݼp] j@U}hn1UkbCdmH?,փ-FY b[u7w/ uuޮ5޵6r#KKf$=$y`cX,&?dKZRf7X ůŪb*0F3oɏwwnF&珟gw__e?{2=k40sihR8Ԇ xeOj2^b3*rԹ񥫡GfHƤ)XhWACϪa۪y}fm[@bó34ڱ}Ƴwɝo~Qnݯ~ѳ")&2AcDp G4K$B(twnRQ.5^U`%0cڎ7GwG{ߜ+hy昗FQ<;(yTuw.c߰n@[vq>!gXYWQb jD'[[Ey:āVg&DŽcSx0eGߣMC>fSm,x'ͷӽsO*8Bܛ:6C`,tPLġ!85`H_+Ze L~_Z/m)#*Jij/wv&P7#ҠD3K30oȫ׀Q qNo!OKxi9M7,rC" Q "9X'\&?Ed57;Oг+$1a 帶Nا;t:ɥ7?zZ;ی"T7p{`MLQ5ё,Ff 43$9U9KQfBTdVc ,MQdgu9zPbdPK='J>A0?ALQPxs$,R H3,R0hjk(;IĈ*G6TSY:렉G,T Sx #lIA=U=Dr${}O3_pZ2GQv>,3߳+q{)e+Cmz)pYI觗p-Dt]9J -ɜGj{+>J۳,Wm4o ̥N߮:jŀHIr=7_엵ޣ4?}ήxx}Rz.LbM&P:3c8*Q9ԚyX"I`A)2U21Z讨 +bW` s8d4#h/M3J,%(_\)͠#[SɺIL,^+m.1ڰ]$P{^>Ei VjhS̗a~Yޢ/ +"Q&PMwsk6֔;4C+~*Ω|,Rb6قzqY^oOYQT 1rLk-U&lw/#}D$T:aKqړtt5Zox뤔B"#md7> O1W6-Jo-Apc~>mv &cP=2^~y-Ҍp8"Je%ןbtbB:Blj`u[K޳fSVq)] 5<+ lqx*XBsɮ~txۨe+PeTφsh7_ $l]JϾ.Ȓp[ٳQX:o y8P Ղ5p<ͭv+5yO{?̓P3K^n/.|^nA*/7cGM32o^|'/rå:N/%RMtѸ>1WH/+qK֯]E?>/zJ^>g>8Inʫ_7~E{S^78a1ђ撥R1YFĊʔ(fD#- `+ER+?OEUa\-ܳ$$v_=>$ǥu[$U1fA5[&,zpqO>oOS |MC]]$m, D wɓB)B|)\3Z!TXTkr\I"ĭ)\z DߎI#&Ɉ5eiXF5y3<'`6hpjWIѢh)A9) ;$K3:萤 ["I6NrHjR-bXT Hs'9V#2)pƌ)F897GP*kBab@vrIjHVEp}Qˀ BK;; GL""W[aQ*bS@P)e_ŠXz`v9 j]HAoK[9UyNj„LL ~C`"Ss!)V`!>,?8#6nannrn꧒;`<;ಘ,_M\4˫fNzI Νy\d\]_`Wɯ>T~uy"_IW de'8^ؓE>I1 RO-Kk兛Q^__<P'/Ak.xNWK`|)gIS8M-7o}x;Nl8skxi i̍n=[aPĔS;SZ>ޭ+%ˆAA7yEM¸nJ OFs' gy ΅R ⌤*'6zS,s25Z8ٳH U9&w3뭩TZL; /Ε}kõ3<{&w]{=+vN>;&nٮe9.-[K±bƼ`<)yE$}fp_<[.맀j#8fw^GA LB΂|૘բ5'i\;6_;.fZ&1Z *QӒU9 (kSkiS0mRfLX)EL2IqFTP~Ao{2˭\1zEG[a:dnWjUANq{PuteOAtc$~!FjA  U @9ʼkb'Ug,̕)F0 ] !lE'%6σ@Z"2|Y|ĹCA*!X}y\Qb+ZL62Fh)gAŕ1B ;hf1>L Rׁ:1nZW=[Z„qYu޹znՆxsOj2^3*msT_x{gm!FP|gIgg>QΚqΚuXH0#t=z=k D$^ E3"%&` gd;EhD v !u@x\vZq3`v7~sц#2ߙ:C22TD|ḆL m {9:|oքJTm_ DNkԄ f9y\TKȍ?,5E*`9L \0C>VXM{_m)$FJQ-QSkLoh?weE=n4],.SM:VФ/&.E2۽@WC*@aQQpQPZnY CXD)BL}wUib"`3dYWoM-V{Mdq%noYWv//}wCko1֔ ]g oh" v=SvŲVFEpmm ݝSuUkɟ*vJpuu$䍋hL|h7Bv DtbǨ f+BKn[E4Jn=ScnNuۀ8L{nڭ y"%S?nX[,!;F6`ZŧWj6$䍋hLN-;Q=h1VWk8?OOgϷ§D;~-F'V \׮Q]߽-sV1Ԓ!,JbP 2G=n皈 ' _} DP_N~KïqgD=E猨W$zE`Eந:Z x"+aB~hѧGҠS5n $RZHJAq?27#tϚcj#j(.^/ˋ%4HgQTYVR+{xx sclJ㨗3[6l,CRmy;[KDڶ(ްv yYp4-z6)U 0 dK %Tc| F/2"|_|.J 1zb?NR_&'u%KMY6opbŠhSpfrr ֐,)JAPa"i4hO~T`}s|IEn؁5@\MF̲-HT3Du꠮c|V"CLtJL )@M0J &ʤ!{V,KXI X*2 H3,R0k+1R/5 ⫞;#od$s{ߊܲ=1 2P6!(x(438[5}XMn`⛝,tW_r>K/7C_z-Tr3.I'OLrY*s %e(5q)pR}U$%Sfҭ MԦ)JY] X{ӠӠSLF8DT;%!,at|!>UiE3ү>`G3䵈qo>Sp믧is i~zӨ'̮4 z8隌z Tc*iZiZCJġVzbǷzb`D 52i-72-" joVuk=u;{5@&Ng(H9\3LrN*6sRL3# 5m7&R p NMjmEUvm3Z4=p,wid;ErDnaYϒRTg* hqe58d'iNJr勖,SHn0`NS3 -yJT.#y<D.fr~W`RBq7v,5֝^b@ Ì2;G7ȸ㜖#9֚U'KE,ɥo}X_Qf0҇ҷ7pߛr; L'Κjck&GxWcXև Rt!hI?*\O `Lqptx~{UohyxhnpA$~ @]ʕR[R1:;rpQXp9ƊCG!2r(MS,4H 1D"$ϩ KeZJ@Q ΍9Գ@ؔfy1 caMr 95)Hsbz)0"̸)le;vƀ]jH|bSBCb޵$"`J_od3@I. k֖LvTSMKIEWu =be)gֳe?ZZ#4uG}rzKMfZ(o؀F|R=T pwsS|~prcʲ-虌 f}gGpB?4{޿ovRj7+zz>O<{um Iٻ19ۓQ>͗r2rvҤrZ|Lg[b^tUNT=><:b6@-b>(UvʱS TZԆ>穂 JpUuTj47 Dlg¶[ u1'-Y#FؠVLik#MDL(TfO2voeZ*!^1)cfL4+%ZԷ$zH&kn"Aa+"yϥEaU{b !dG9'FRWv )BS`I.$|0AXg8f$EXNX$RŝCƣv|?2/49r4pTؓѲ4 z0gFlQ;:uٌ|o Dcv(K.+ѯ }+Kr_W,, wgO6KG ?T1[rioG;ɕjδ4JZ&0a #]WPhq4 ZCXYF{vZxBj'pςAڥ^yPkIV(BFKnO܈dnd2hV3jqp^'WYBLAS<}&(cicӉsFl%?iw;P3ơ0Dp9#6$Dˇ⏎d2a[eݘHpgC gjBx]OL`Ԕ=B!sΟr~Y\c0_ty ` U: գ.= rB[_-B,%?88ȾGHnHbʖ`p74z~-t]ޅx?';9$'xlgE gfH3PriթQ"יyyfsedqSeځriI3OKIJY>jDJ;> g%.KfG755(|7 ս/ dT}UOΜNN3M^]ށ+4X:&D@\<[PӒ^S T~J'cb-ۋ-(S2*7a@xv9$Q~$FQ乐")r+HIo}>ޢ瓩C{5fvlps#_Ř/ m2M\`J|; n"*G$ͅI,}e 5S|b06E&Q!8Ǥ+Ǔ58V4c5c,fwᾋ3({hwnV}\^`dBx|i}*oOб͗ xwy4)F_g.t2zow)Ѷ`yxztu# h@i\;SD(^-HbXՌ;^G슁jc)j*Zܦ^Yf,9%)z2"Lx6ϙ]M0 Y |PCL.Iz$,3;!AUwi~֙"y#|{ yw]?vqr.&3R_HaPp(4:Ш5Q!B1g0uHɻi b ѲO̜6ށ60r bYፘԠ}cv73e:4j!R` RsYָ*?Lw}>ORN}wCYbI@'QN3۔){164Te8A,5nH1Jgnz b >V9#xM{>6Z lgCDCMAfTV90AR3ryjrHءVK@$![2B[IVXS%'ӌIƨMFhTjkP@@BO0 :VJL湑`z%w{7:p|'\M/\3?=xQ=6g+_^Yqe$>}BxTfƸ'BJc .'X' d5|d;\F0[BAk >X mU2ͼtJ5v50w<)eJH) {8+f8N0q63 wfE4o%}O\0#&+ RGdTnE 9:l E8V*n0&9v5zw4'WygDXF`+NE`R<[)sșBjqDHs.RLΥJPQ^&hr\1E(U%l[òUCѶQQiĘߠqȸ`TW!73!DH(nbvoe3('`é:3a!J@(GKJB^㿅}# EVEVj>jFD9"&#E `;)F6 ߠu^|<@v1hTzF щDZ_]2U_n8g+cg(*ZسfC`0,1"bu) Ε1f&Πġ$ʋB`HzQ޵NXSL<[ Lj5M:H߭Zi6t.g#w ^/I.WA'.zH,GeWx7erA .Dꞛf@(*8t#O$ tZLЙM˯ cLgIizL$۰!IFRr/R)sskҗ2#҂C<TR\PvNE[x2g%Am.%%[# R⸍)׺ @V!?^@mDNBZOdg*zZHOd5KFՖn"P:ڦ}ЋM=a&ۿ&bϧ!~ԕl8# Sy$ӫvq;s \ɧ$5cܟmGt?ZA "ԌqyL%㚷ߊ(Μް(9j GH9?Bk\)>y:pqAza|Q2vL6F-Jb׮Q ڷ1aTmtÃ[JRY YQ^ܶLɧ:+nԶEQȿ$swq#X7E՘QZ)xE*99/qsy_^Nt>No֣[:>og?šj2dZ/kd}hg޾6>.tB{&KUf(028xdkՋux4oueYu%2KF5c`5ϱn9YyR>EOj2vQ J' jyYt猬ajV ǛyfL{|ܢDj2 b1nyr|ü=ǖIJANIgbVƧ5Zv49V$-(]} w;|}=M] wV0*ڃBs(7mFa%>ᐚu /1?{HaD2Y|6 .ާ lmW$yvؒԲHCYn5"zj^Lt$'{9}I7?O`8@Š_wij݉SJ y:qEeCE_Hx>|pL0"%#S .)&׺knBt^iUg9n?g[iA530y,θbd%0*NZQKXUyK(%5E=WZRmf/M=O>t"\?_f9og9$e:M~k\#GHY(EUJ, <-9!CRh(+Zx .ßnMiAa׳ZcގIkl^GdOcs[4dZ0qnV+ض jEQ`R(LPVK&;]on |$iIqS t|^I5Jk:NqCR( HiWȤᐢf88Ӄ3x38 JQ#@c]AbȸA':'heo0a@Ot+ {ww^tao7dr֖ii;h4qzv-&5 }5~z< qwݗ[v9@12ʽǹѾy==_DٌXyM ?{#w}gPa:f!O^)yv7%+ܼ*;Mv΢w-\ndb>݆Mm#/h_~x/חw1fi%2;ԛVMq]Vu6v H~+YZԲjyКn]E_3_t$IKLNuz&:_}3x'mu/<7LC,ieutseI}mL .A[؏@׵_~@V0\vLZ\4J}%7+kW/^|:/o(ѬSG9P~ BjA_7CW-P^ȓUwlR/s63{8K%0ĝq>\} T7rǿ|#%{X8It_ x;r8G,$H.yh~xi+bx8#C&e28sWk㒮 tiNF78<>LW8mH|˫]|b=EC'%_>~IPc6=>l;yS6Twz- 8;W8-VV'u`>vRt/L ӻ~^?煛RGJK! . Dң^Ԃ`!h ە=7n$"2UԌPon!_{ |ptHx Zh7)Ww~>q~e}^GL"nN!\p?aE>Lr0񞇧C.UbT8)*2 .'|?>{3qJ@ڇLx\/rp9((>?!);[+]XqJVz7V{z.x#ücF8Qx2QxH$꬞7nޮY0/S`:UmY%¥f/@u O^F^NxA=C9 ,%QcN){J>X)HJzlu<ܫJҔSӹWdtΝQZ9A8SdsƺӹsqqtyO:w PCO} 0ut#^QtALpt"8: fFq#KIE'[RíD'V8L9C!/fpvYpZ|sV&J鲐[wk3ʂ N Ɇb+ '2SU s)[54N94H>ЅۣˤEkۀ\O:.ypHwa(3bi60@Nʌ<:M2՟ǘZLϑ2ҭS_&x.V1Uż^XT 1~. :z\ⓧYWUu㬛S޶n"n1Qr5N_ywդĵ2t_ϫ;#゘Sr/F9DIcҹňTA5)GX˩O7b+OPؔwO v *H~AޒgafUwľ#F)NvEDAT\^oV+Y*1ng<پV0C˱}zU]сJ)Q;s l,e׾.9yV|3Z\^Ml.u6!jF͛wϠjz\f4:BCb@&&k_3/&3ß.)I)g#*X]pi@pVu3Ǹ" ! C j#,IAk# /L=Z48eʛ`ޔFy.) Jxl/):5ix2Dss$m;6M5Hq =k:;jM@G) Smt :%v12IaxOx`kOyC&Vꇗ$7\*K-Aj!w^RN J/TJ8Zì-Ŷ+1xSi VLC(}$%@%g@l-Sx񃆂 \!m@콱v^h^z-"JkdQZͳzz9±ᥚf}w1pUߺsb\Eh7B h"qZ3Er%Y(F: C!]aױ(T0$,s([5]ǍpW!2  tGXujջ\,+# OV>?!+͈e%Nj[bM %$OV /u*A=e͘&R^eeR y={ͤ "N1Zu*3nѠgK}s*l6^+,Zտ]/{$b B;UJ) &Lyj SCBiU/wr+<ͯ/.J q<!qzwJRxX ꋒ!RB[PJn5}@+pAB>p;]4nJY==cr_p:!?@f@dH6RHkIL p) H ̀fM%8[RKRҷw&+޻5[퍆SiN+0~?PxsﺙMG5GlP?Ilӆuڃ+O Q)4-aPn1 z~BmUꀠKb 崁'?P_wFn˛:^ (;,еܖBPƵH-&) :`$9-G%^5}}'Ki0 #5uE0B.![~(̟ٓ1΍1(J fՑ14[t^\.d~.ЃkZ߆jn I|ȁYئjo{#F% zF2{T|S/JʤjbDlSaMRRJ [PhAVsm.QR08A䙧;&XB܃/rmg& 37gǫ|Sk8pulۛR^YX9JAF҄(.VhekT(BQ0{ 9 f*a_^׮:^ŚBΜF[rQ-$pp o pq^X)֒ ֋KT6a_>.SaIJՊ:f .EzS^8fTQj|{u`+EtFJi$q 4XJe .8j 0[*e~+7܇2g4u7yZ s3iz* yA^Kf(aev D¶[o?Z@% U-* U*?k-zjxYEUF+Ae,,h y!<Fv{iJ;Ndj{3{wJЙM[ AIFK4|gfp$dӋ6xҸ }q)W3F%RH.!Wܩ",vl0i^ 8i/6.9%^emxwKvZ~g:!w&6P./ \ |^ةeFG"iW_ ݻ|lKuNwmtx~zax\ kZk0BWkMp3J]Z1pp75/ߩy-d~/2*'gQCnkM6Wot^>f9k(=izC՜%^~ƵZ 7~HJO30?\k1nֱ|o٬O;ORW*k3k属g|8z-农;UO~V%aj=}YDƆ)DcEIf m6-r/kZv}: ئv,al]չSMf+w'FK`Vl,VAIצ94;)`Ӗ3tc"W.U2eODWm%mb%E=*6``d;`usJ& !\Dd Ա]QТzN1݂$d3)63 B%= j:Xۅh& s?r#5Ib|gvnMZJO8bW}Ìɢߎ=w_F=];\gUrP9JRTp\ ͝,)!'<("ZAcy:l~-@!Zp.@2<.W;`&ܤf*Q`JjgiPEYr2za-y.l#<st3Ag)ke K l l鬣Uh)&JY!z˽sMR( AltɘIxj-]no5ۼaHWW|F_^EE}`CKשr"m| Qf/BF PYZZc3,7<̕Qk86˦ d٨~JZRi蓙 d(F (Լ/8/7M,/sJ6Kqc&K -Y蜭n`j@ #ָ[_-YӮbՈݙcɏv{Dvr`Ԍlv 3 s=5kl{N^jԠ&8J`ʜGθBގ'ڛ%K>ArɁAEI_ڣqXEW3j:8.(k`p[;]O?4(ӻ(. p8j],KէMѷ-Ң!gFo~> }؟O|+Jk@5'l~w!Uf7 =c0kѯbo?]]]^_q9׌O}\~8r \=l[SM}&+gu8y.K%7g7קkyuhxꜨYA0v}Dwݠfd(2L6SCuBMôXW1sM| wT+#&X\AI^S]Rg6<8Z] Wzׇr_ Hkz:Alk]fGbs9 J2ub՛}s=t{jjk4)4Bq쯇`:7ꁼw+A+"f*Oֿojvo:Ls/&z>qj.9mQ 'uֵuB8L:l@ {hliQNᰱ>W|aƻiEM3"C;Zޕ)%8 s<^Ѻj/{檝 7N(a V3\0"\CNvd evst6\ȡo)΍7>\M>\M=jb>;WZGP+9YffEY'jnlTV Qk1L69wρpZě s){)8.\-kl# < cqe8alHywV}20؝ Zn4/^rg8^r&z2Ə~G,x<|,[DD6 6&fmSdjR_H,T"Eʡkپ߯ U,蕲%'n/(&hXq{1`m Das7; uR i؇Ң=0I3)j/} ḐN*2{|/w.1F۞ !{x!Ӈ ̑9?ѻ*ϔ)F[vQj%cni{Yq֐_2r]O6xqc#g(\;N&dڌDžnqEпL`8؜%#?"7q 쐆C`SE!Pv"g_Nb|WYpƓZO&td̐as'`+S ]yS_WErZ 5u@'R_US~"Q_:)>pɮ~hhnU¨[OUzdM;5EW[nBw-SE\u̹:#$/2ɋdؤx@|Dw1kg^aٗD$cӰmO0j"/ wq)}qe13 Z)YNRsEfO>nu[2AX!a7Ņ9<3F2q7/Ha#gbldU2h>=\$֖g8vY3vI\yAfr>r%ftߘʹ Å wF HQ'㶘'NlG~OfF[ G< d}3ڲ*'}}v/>(5NZ4?E@ 59'/ZdG`?2aE؋mƶ#w.HaUbIgB!$DžmvI4e1cH**vrDz|XT7sqQH8on΢%u<+0$}Ж>ilrPJRyX\t, dT`\L#fb*A!a0 2x,w71}˪ (mr2"]51R9GJ,q2[ sn tyu !TN\FO_{} fg~7 Ie+;hA>P$t>.V Q Ԝ'x$q͠q}IU-tGT%RLhЖʚv1 )Rgh "T0YD3PbGj -&·*ѤsRFr=KP1ttHYެN{ع֧4eDny`K2䃷"q$^*ohcMk ™eB`%Rt_M'Gdp 5 X=x g}J~8K t82G@@#pkf(@U-_CApiG>Ž n\ٵ+du7}*,Y"5?ƕ|`R_Ϊp3ّEƝ4 %hVvH^,?VhExocʐîb5t-^&kǟ4z8-}u~-wJ9 ߅ˏӐSTQF~z}dpO&!'Z9‚'>ײr~SxͿ-FeowʩqIulsyp*=UٵkD9kfF'1-jY`;>s6)ߧ?.we#z<0|LczO3hHvˣkU%͎ wGIU$ Rc"a?o_N@LeIImppLEA)OPU\'eǯNq1sJTF5a\I@^;28<!Ԕ 2);)CA`o%Eİz(%&@d0&$}I˳=LK`~ϷϷ3A@tPIZ!Z]$o,@ 5ȚKmK]FHc(Iʈd,ޅA>`U8åidnK͉h5-%uR EtnX ekRKA8Q3>Ǐ Lŕz Lnn￴JJ4^g;GIO .h+_TShT?|E:_|Z ː/"u9#1E?2Ǣ];!ƠaGjJq ^n{Wl[ ѿqM˷hA}Mb(tȉH*듦b52j{gK[p=oQUO2q\ܔ.6Iwz~"95~F?TA*H@m$U%$\ R3q@9XK/hyulѪRF+5R@l(4m Wާbj&,a3 A UaɌoSJBq r^p̈́Ϙ> / Avt֐5^e?ks,/xsa!> #G' ؠsh‟Y>x녴$tR >Z=r<.~ 4E,'B-O## WC(^~b>}b;tqU^<#%Amj fL؊ΌLU[ׄY36 ;iz詥 rH/܏OmXv;nO{#_@V3IAӬ"dp0$!nwJ/ObX=aܕ ̭6>W3vpkBL`o+f9xшFS_Xù$SA7{'Su{rZRU6} ZqF"j6/p R%%ހK:e贛! #8oE*8]y2B700'S~{\a?ڴ+ͥ}g+ 5@RtV (f 53ˍ҂J`>oB>7>؏!5ٞ^l}\02鮃äI|< a7sO#^u8h2eLF)[R- g_NmZSZEkUs^릮ZR$>?ZH`P JwXXT + ele4u VTJ T8h1c+qzŠQ /J LCuaWp.i%pufp,E %z$i)b^V T%eTINmJbDV[UK3:+7W(JpVV5HiN ^McqXikҍֆX,ɔ+>AŝC6cʖ>WS^{ (??l~ں`Ň;sWx arF`gN~ӓnX3Ez'[e1_e]37k~띀dC&(RgO09JHOO~/շoEMO u$T םQX0ުM1*@T*71+R{D/*MLIIɘh5EO)PE.J732t-k^+Τ-ZVFZk&X;6 7 ''41v\Ŝ4%3Q!h;hD̙3Swd:ޘcF50WBM9VɤUU1 ?nfv]߮mfƫRR!q=[,%&Zu]-U&C,Nh'#IѷAdVVf&Zf&+Rp͉QZ+˺,]Q0\׍״h*5.!j-4> }Crejջ}Z{'LZe Rx/hܼy+s7m4QU?ћ7;L>nm4]~dFif1'!zO0_6җϘ4 f亜8Ț5#U 0qe,uȨnn:aP5š{pݽwucn#uoޕaDBR3vRwYvy["g wG]BN[띬m̓es6E/ݞlqb.)'"/rG=P*"I1x֘BJp資* p0g\'@0>5/tuy眸sipf>F%({tu(}~:)#"f>(‹sk@/"mM?$#}V9W0MSUvI)Җj VHZV R вmid W åPC(FT4uEPTBX 5BFPCjtP8ݦvQM%Zadlb* 3F) ܏Rk:K7bP |>OqѲ(w<[АW!:Z1pMI4`GuBǺ Tgn/kbh+W:hOKJ, MAKO5CGSP//pFXxYv4Tww<\`\Obq BURLd¤]|{IXoEc24B_zӭ(MkB4.堵#pZhJ]Pz`/{uі0^V-v7Z>Z% /kRӎ\{5&M_D3 2_NrUP=SDY8mUF+`c lYhWe%T0hU6xEQR#j!Ud-h+g_Pg<&c z×t{ukۿnNP o?Eld؎m{(|{r#ݞ"d#2۫- zƸ{szHsW8ez[i^nĈևQIsM'?fkԧپb\z$ghp-g/ RӢpgN"p`H`bpAIڑKswDv P$H<)}>ǖ10aL<$BORuqGN5f—-r )IK[ԦUmcӚ1_ʂƘ9mN9W>NDžb~H#Ѿ-;/Wuul:ڗ¢R< o=noPwf C^- Q[Q|g|~^/F Ȋ@Ea.X0.r >!2 "e2 jzvZLQIRrQ#T6: 9ZQ<6w/t*@Wu@C˃y4{1]Ah7e0qaNFv"bsirKv8G_czgM2Mg;Nw*>(2\N Qhr$gzȑ_g06Y"Y$e@`َrt$~HI(Ybwx,~U,W\ef/F"Y9W퉒 E0y ׂN%d5WhUyþ:O(>XWaJoG+.뻄n (H^̤LZK0غ6t{pw>/ >"C ,:;Q=ܷ/3xza" =mtZ P.S;q:}]H3`t{-o2FaVWӃW܉<㩚%>YhS #)ĿCˠ1h_Œh)%:LU7`ǻL!g,A1Cf_ 50B5fRk%]+^_BLzDZ!yF9K턷|tz8JǛ7m{}zHֱNܬsXbW:v8q8k~+ngWo=7Oh1{P|Lp~jP|WC7_ #DW_뱝Ex뻋"!+뾛bƓ9E}Xx؂drjwplV2ݻS" $0Sᗮuq A"%؊BECOr oʱ>׳mf&cg.4#PnezύxyN01d_ݎ|_k33& jpސxΊ[Ly]J% N0jvZBFbv5./8ύqrqxn =Y*e2jf&q6bGv8IaZ5jX4ep*[$'I[Q$[O*^Йk}wKU*,)2 ZfV3$ֻ:Ĵ}r2?4(!_'QjzWPtW) '31J{xTY뚄_G4iR$8H@F$&dF3#I{MMU4SMڈ^d-CHkTВ^߱: )IkJ2Q_ԓRwFKPn("JB9/AC}ꢼvO,,P'~LGI k53\eňZcjqľLϥ5^|Y~|7&fLm4:I~?~ 7 fX=ߡڹGn7/{gB3Ѱ3͸8vVzoGtSo hLR W.[ 7͍qml}"["upa*o2눴[bsR3Z %,ʭ␗ [BI~ck%eY؜ZLI@CR  E2nI$|RhiF "HQv,0'Lqsn[H1 7ᒗYr:0vtT#I7EYPnUj!85(Qs^HE)0Jy,,w>+yrLMLq7GqFq2%KAz-HM2R+&m99!%`h>V uS1cu;3ւOHlPx?1Eܕt6{aE@y.\\]i걐DlJލOR116xSγqњwK'wnuXwnG۔"aOU4?/#`))HQ7,Yb]nE[/,,nN1QպYa71 %B)kEzA))V~0z2^\{PAan(EIA&  NP3Y+f΍x:$6oԱk, t;/'㪆I`,yS0xa)!%6POǺ$_?'=\虹%vQE<nqqY icX 9+1:<2Q8XZłiiQ8xǥ`ΪY!(r8+1]Eިڭ(w mlpD%"=ЮU~K5iSlgR |N|qI s")%tvϔRLO^z0& ߽xۘ %xt̤ |!^ZX bN2XJC޴޴޴޴ic\2_Disi!\w@yI%ڔǑ5B+˜kD6D0і+`BIul3wK_>pe d CdF8PJAq_/ ]gY-䀊Gn&T34omBHOQZ]`\zJlWd^OZWBl37 th%b1!҂ a)W Jܠ"pN I3 UaǘP%Ȕp;{s% si&3Eɽ;sH-PHb,혘UCZ(X~r7 F!``Ö>^s\xzɺ-Ye[Qs]%"#u)):Br£TZuyxb4M0qKY9 Eؗq|Rn} 9~I \=],"FgGV^gO߽,Zu z\dt?a~lMG8_yR5΅~EfƹzDS{I9w\[ra=5PU_MmG<ǛPVsY}ءATJŌ߀kmfҔL؝: :-݁Vhj &XH2>VEZ)[M Q I1T g(3,S>59`Y},/ac‚NG U԰'vG z#AzzgH0ÄO%F-[}'Φ ot{'gPyEb5 ۲a&yq[;?vp|W@'^WW(BQkBrG D{ˏyG0ZrI1:An2A.We plB5J$3Z,3ᜅ9-H)/r.CdSSO?T<l`si! $3HKBb .-gpA钀EJ7GM6D6Ѧ+$ SK|WzQ,K&\sms(FфY rKݛA@d =L]Bn'fDB֒T1i&4W c`v;K5e:AGܺFgT"T)\txþLJQZ3qu> ut^*7)Ui5d0yi$sT2VVr tcHYgL'x# 8TH1ڟF|B)p[>t+&XRi4R㥀uJn1i0 qrsCQbί~o󢿴Evӷ6כ(IPdHAg_@|=]ͅ*?^2!Kќ$\fy (vil4]Tמ݂v6n3Bs_M 37hXXףsgެ+{rO=΋Z;=.! mh^DMm I_&;p&0#'bw'nY%b"-GiL 8ym]KK=;{W wd5tUn/z 6#̭@x-.OVaN4̷c+ .GJ ިGnG ]LdNջjnT5vkACd1Cb*j~'@IO#&81-V3ܐjWȢʀ\0uLkcq,V^m3X/_,V0TD 1s‰"QJ.m\h%( B[:nIpp`~+ B<Yb\4ꑕP`'Dnzbmoo>ŧGeo lvWc3l:U h}׸Uxru?"|TErऴw-V5 x;| :m tڃ >›P=3v4ٵug󾸟Kfǣ%HDde[Tg!ƹe1IƷ$:qh0Tr;Nޓ(7d+5{y&vh0=ĺk |tVÏӂҭ9|jYŭ#QH >&|EAFa+aƛW‚ Ep/j ՁU"hr QHiǧ3}GaP!H[Zj(b^ON9oIXYL_0ryL,"2`BM$D7D:93GECGĭOcD ʞ`OI](gԢs* /(T(Lq֤B p::9S3 NڣD^;R<aOjP晩{Fj8=K!L1 ?fnl !`[͐ozJ~Azf⫌c4&m JvH!h!Dtc$|Jvv_,KM\06D6YlT6GDAאf&ݤah.{,̂:a4r8:1m*yisC~<kt¿s|LdH1&|uѦ p@V@avݎzX0Vdcz!!֦w #mIܙ%>A0Ko6¬b @s [޺S xnn+}sԶ]x@SN&7tq P:=U skWV ՘+VբWk.$c_h Vm2P")35g4X- ZpL TȖf3I{aS ;n+9S׵.e鿷J<\@ab%^aǒ\^2%ӣ${nی^!)0Nz9e0a&VQGFE_p{LPLUP#= Pz +-u ;\tb4S3*7Z\HnTr=#I0?\R\ZP"ConXSE?] [)vQ8 WLpE५Pd"8|8W 4XBML {iM Aa'g|g}?Xk -@(UnWZhȒ^/*ˡtCuaK}/Uvۨ:1r|]FRi^u+-ZǭDKYTiҋcF\.S;C.)X+v-qfq)cMSkѠ =eHQ|ƆεhŒ[AY,xLe3$i` #3QQ 2rhan\ya%$BWZB*Qh6wYBYVOLM QOD\}υuVU:1y);Nc>7QZQoKd'cxl̙< l!(J&t`I%dE,M*<7,wٟTax}Rߤޛxak*J,r^lC}͠17븬%K@p`?jt<Ȇ#(wM שŬKǿ&,ȗYfY~QE{FbX/L7~vpg{^<~4K?[ERshx=(ub:c+?J;rnMԊWSZVdaJ4W;0e/s 榬0;.q$R4~^^g=wGxig6x#_?n d%dyZÐ"/[6*eZ tSnx}pQ_tT-7#w woOڃݝpY|="]\R;i/vwarsOk;|>_2y_/OI^ <]v/{8k ̋~'5ϓPwAfi}]|&Wpi$O>yƕEjzFN#v>axwW;/OmAor1Ɵ}胔w:l1}av2|0_r;ٰh&3`$h:^n1o9 c}8{0,Ngy@6@zBErd&v7%p ~8cˎ$O& ߯ؒ-dٲ۱v <,uA֫ʓ\i+omuƲ˯~"t_S?y#?ͷgM ;] dtף/&d$q96_o|U3LE/jv{^[ }=j_2tg!0IkLɟų_]}1/ǧMy2^ρoʞU5͕wMl&UbXb׿כjY6&g\aII.߮s+]Fmȶ%No7MO^}^v|3:ςI_S;q:r bt[f#uZIcyj{fLI؄Ae4=VH W"J*xYXAf 2l4ZkST KrJ uAV"J` r)M1Цzͫ bفQ*|lxM)VgNX0L űcZD9r+X/عŋ8.뤘|ֈjC^K}ZEDey9,6wzBU_:kYߊ:7~<tw>85aM2%AiIӞE 0SIVd)k]N^vI8+܅ۘbʘR 1} YJk=B;R(j!q-BLr?!lz bY@d6R "bN#%4Jijki]n{*\=f!D Uаm1:eTM %VJPH D[+jO^m K, S"eŷ.@sMPk|GkAF|#7"۴dBY2%GYj`WAXfVYel>{".JnI@HF^|?0:YIAYo Gu}L?]"k"QXƒ %xֿbbÅZO&cp0lkf[.)WQbo [ Kc(A!^^^-[.źAdbf5*1yȾZBe9ؔ{9 /Vn;F_Ju +b/ΛزDXq IcUʌ\p-xM2,$vt?ah|R)zEX= gQ,YDVr*$J5=d^ ةa7W|ivpuܐ+ZEZdɚݱ4R |≐=3rjgs۬ހ>fW8,pT}Ѝĥ6~*ck d j|ݰL)%H䝽b5C]rK?j!.c\q+nQW5S"_Ta)q+q2N$Xw=fWǣnnE`Kqe*[Gbz}ֆuEaٞnKïc$(cO1tbQh;W` tx"=>TjmP GT2ڣ4_?^΋=V-߮7zyO'G(jq-MY{܊:|w9$1X&ihZ|z81~>~@ǧd |dMD;؍;sΩ8sj ؍SS&z".pvGx/S.7x7+jv ) EGadb~͌`#BET V`! r@A!1*#}Cr%S \#Lݦ+JSċ*]γVUףpuVsv̡c:fyͅ`nػ2z:6u%WG0%pMcAFi# yYG;_:Glƍ9y-|F=v./G%zk,6VbH /vL0:4M6"$V&QZSP\t=EZ(pGX*Г5mbmTKn.M2MrKaв-Y,tV-ÀهpM.oJ GJ]t?O~S7b[*Go-sjǹáuJteʟKT]@0jղaV/`>=o}ur;>=E ӷ)|Ok7?vN ZZ]葍ZܹYC? lDEklCCIlqŇ.ޢrP7[x|y~ΎE뎓fvSQ6%Xvj)UTKau[ j&x ]@b1>j&y}qk.Ⱦu\WJJJJJJJJJJJJJJJJJJJJJWq'^V»Yq@%}24lߞ#z@n}G+ݶ5rR]^9-i|`Jd`}JV idlY},FgGD?GKU̝r/;p2KUV:p#sdԆUc**+܊icI3:/Ӌet4#{8`]88osEŰm>2hƷ}q z_6C踃MZ:$$ b~v0>;6ܒZq=Ȭytwem[ <-׾C pcrBbEJ=9\ԢHnRRRuw2Ғ˛왧+ͳiFߘhtK\=c*_/^0b&7‹y&кUaN^oy? u޶j^v]Bi^-[)g-?X9_|(I61hT}V 8X+`UV@nPB :Oc! P8 C`( 'r]ؗ9l%0T2u0QeZ={`yc,Hjӄt4 'a1NG"Eqrt[~'k[KJTwGw>T1,~[{kl篱>^{dSoMTK؏^bt>xoh"[MSjDv.7LyDD` UqPTZ87w nZc|/>xJo2yu@]ӓz:$.^f+'9i?U6JftuddR@Tٖb,{*Km뵲d7`t*B4zyEX)] nKܽFZ.ýJ]h"SGu#}WרjXțr},UwFֵl[gM!N>#mnwq5`X0^i\kd:Ҍq0CDWqHKDO0؀iغmf4 0M:VNW0/Eyi{車4]:P"8| J%=&UePA-E~y1H #[PiL;.AӂKE JO`Q"J`JF"Mwfрi\DKI[`=ؒwiK-i?`u!_ǨG,naн i:>ir5°| Z^t c;ڪjCPp?}Ҧ:@~g #G7ر4zOhB{lu\3z)8BNsE|R@0G3,V*WwmV|S3^K \Wuuiuy;QwRxMjHgSKN}jֶgE ofG8iti~Y Z@(S ?] #B,J ?Lms88N_?֮߇]/L0gKOl.J FQs#Ί@h)v(ǑFH C*a"؃r =ѵ\SZ>|:m?jvW ̷ kl'ںY | 7XBBPF{1( L $q "=/Iu*iJV턬5Kre]q/Aj^䇗ƅP`+6=;xW\HN;"= Sk%W & !>3K@FHHzAAd#gEW RLjt%>9!avrQϚH$$,"N4hBlSZS8*aYВH[ Oaf*UQ0ˁe>.F}0k.Y/"1P Ӗj9⷗CC@$n)ndF(hP5>YYj\ 9VL10?A{X@n] <#LBqFL2ɱo\r4S~  Zvw&IDUNj =iQ?mHMҴ?-J>N}Y] aP:4tX~X!|[E/.~U=۟Jjp~+~5׃٫(F.=37P]]*+_JGW&¾Jɍf~9A|w;omDR7]%w}w%=3H(pdĺRhD|gVe+]28b p-E,SADSI=#S#_LK,;ZIӾ 4J/i~΅x )-'Ļ5'=W"BF{qkunLOzMrc{;`]YҊR6= )æćԩDa灃sC]X2$B8a_~J6AF.Z8 sKLq$a>h嶚/IAqo3I]ԯE 04X([(:FV(K&DK$ V(R]K#lS.C$qH1ƈGdEPUV,~N4ǽbkuڝ+K9xWͺ]\]VfR4SVxVvo?\Vx@yM!6tclJdE v $Kέk4ʰB-?%kHol3AHeGBjFᘥ$Dۤɢr<;?s5ۙ!sF lDߦaӍ<0p|Gytd;G}}ז2 ÉǣOIO/;Ùq}=<?./5[ajw^U;fS;u{w{Yo&5Nٷ3 oC #uzmzl˰:{n:L?N4*;tu?7&_t!|<}p<iPzWhzڼn;#׫L><|h0ԟRVF'_$m4%vJm-]Ὅ&ٳ¹_zx3,Mnlۨ}ptadx3W6ڼz`vN7تM (wqw/PVxⷻcc}Pw>7o_]4Fݞ*bepG,~ףdX䍤 2<=il>.$,k\̧GV~9i3_R!Beu$e5&5Ke~4wf ^YI?eQZ3µ<5<s·3{cԠ)gxjRt¨hrd~q B B1!>U.9ˤƍfl@Qa=-Z5ᄣWvG^-uzŤtH|Jbq?!!XzM&J"hU4fĮ%׏~;D5ʒft߽OdRws[7@JL/Oi|_d G^LQ54p#W(.ܞz".HqH kiXNvlyq0U34k4%j|z:$%זj#?-M@_і9h& umSO:tQ0h?t?>ZAGg5$/4}w3Y@7tY W N]/Y^Tu;Q(A>˳͞xlu Ưth 9|8WY IT]c.;n%wG%BEXI1ٕqKK1rt ݴ{Rr,mTЯ*(دgJ<~hD$>H8(:UDH%N$֤[e@J:8?E+TFҭriqל抵0WˁҰ8;.ޝwiڵ''=~~&1X?{a#b̕x%W $ E>FF!H 1.f A #EU$kx?9xɜ)gߏPo qѳ4" 'k*/^\ȧ:OuoWs|ҧ:Ous^9}uﯲ:l"ukjK?9|acOusX9|s80^PBéwdy}AuE9dD9\uy;"B*(SV[wkJΣ'n˅z p%e{f|IZV]_ .u@i FuRzt] N5mm#;B\QaGVDBPpqQiPJIKMh(k/TAF0Do1|;(( "c,ƺv1`fFVFIAV`d9F+6M_K3I&0 jzjɍN~M;ʘST%GqQ2ڥŘ;ir.V}₅eg`16z<ǒyp|vA4b%僐ľ HPb fi i2CyP+dc]4ŧpC3Hc׬(5B6ˆ4b.t+ $*ڷ:i*O] .7*m vevj¼5ڭbՀeLۀXy R~ӔvN}Z5KAN`Xօi jP[Tp_$Q׬Cyo0^ss*DݨN^7jFX{|eVilQ{cV $5X5nbz T(SլF_^p(\qTp:?j)glu i{+WkcSl,7jPW6[P,sٷ|D2PjR*7֊\p/_rGT,ތm7_zdD흿|k5:KlblϨ Kiin>E̐w*~DW~+ً[v~%{xy|EhI 4(ԸI4BC^@ֹ ֎_@B@EWGHJfp]J{K*ɋ23׮|@Y7#O}+bNAL#iZ=P ) _|!FY/l=u(Ceb"*DD%D6|23(,"  u(ZUW=#)uS^*#!0X[`Xq?@49#? |n+#p8m2_I'HkPG23\ 9pQ`ZGǟiɼZ2UESE,zeȄ!<,JBCls}ܹ[`f6\c5`m#4'ou5 aezSW:-``shsQo %V_sc\r9y=z{ͱDZOp!Րnj39s59lچj{D3ǭWtcƨW$Dgrpd@ɵ)'/=G0v$tHxkamG'݋G֝/nT`6$RHCź`bR4]%cԶ2k v`i>)6a]gț=fHgi9ָqmJޗENZe&h.5@-_\F?^eOSfv;ΕqՇ=һ>e>Qj5(Y iea&/. xCb( Q!G)k}N/ReMZBVXjCkj` hUYQ^?=y祣eD05Ekhi<;Қn 6ϋEy!ԋENi iYS\H*lD3"|ц@騴.zd>I+8y 0/ 0@Lh^kAǺ8exJ,iu_Lձ?{ƴɨ(VPQ <0,U#>"YKg2}l-hS)8'%鴜g9LZNl2y/[Ak`rY jcT EN,scbR-E +wEm H2Ե+j2}jmɿPLp-h̄1D$AdK`sw~S Ԍc 9\W+b-iM?2(bE pIބЧI !#M%\ DCzIaCMG jCt"$E& ȁ)9#eIfZXEkQ\)K%62U@Ee)(pkk FZ0*|GA~!a` Ц7Pm ד~+-l'|19[@؟oYOgvZy59[ֺZAna,O?T$0.J0+hAHr*E敌 HG,+]^F>mYZ6Uy8W !#3@ JG!lL j%1$du)&'vc҆v]%i{IL'e::=-v(P;iT̄gadbd{CAߑ4mv?9MbYVjY7(>EmEe⛷Ƿ'4L V闡YZL,_@zi|+FnG΁v5$Gf_7:de2svfEZUt,5o]9I|5I->"3h6?"C9"zٻToܳpUw )ʹV_r}M$M]ĚܳdCْϜE[xJW}AbЇuZv8J֢38tkOn)݆ϜExʺgX:@Aw/ɮgR_tо:ױSfrڂ q. àajs)íejN ?zpZ?mCZD M4N p>RLB?[p)H/8 R(ŵBI# ,x✤4D`.E͸80Ip7.2vv|hY)PyVN[1c9DS;qIuȞh OuOU&oJ7B1HV N;#gkE 9tkS !9ŅP~_|Nܮ{!(tf%01[ ~.$:9lל?vs;Օ'=g:Jt9tAPNieO3TW`v{8BWhu1 ĐsK5\6Q牭!%!rJ2E n׊7l4oAU\ 1&E%@ +xiOdtJdPZU@"#)$?*Ѹk2 :mC9#rn )u(1#Y&e?b͒LY;Ao5D6S$OFfju"X2(`K jʺ1Y;wso?yiY7;=Un#M'^rԸ>yz[=Y[NjVCC߮aF?jf=3[q+Acb.˧o+ 9|[Q}.1TgPT9mHŠw;! "v9\Loʵ<Κb{^uuI4A0]a:eÂM (Zp-|TKȢuD.2XާH%pSӂEz匪Vcp< 0yiw҃ ébMj\.B_t 痿֨_wrC2!HlPkteOx[-I2# v2{9}@W5Jŀ} ?No(:Y!gU]ϫyZʼn0=@`&?? ЖlT >t\dZz^ۉf{ n}?O׷Y:~wZ|onJqm>[jݢְ噁V>{L}Z&ӻvTo(t<7(ؾj'$W{𷛬WKM6$n*ri,+_̄_Ds%G|= muytt9|i-sH5<ۿ1:AOel%8NMƌI >`j cGL"g)f 9X w"jeY[ࠆ \˸Y54+~5ݙއi| E`9PY6:<ӬF1}QsnQE!J^`CC,.rDoN6Ih.B;>CF Fs<8  ˦iT( Qэxb:;r h`U|ssu.(zG B.{ Vd{E]!8>Ĺ1~ (""}޲{=R~0O+A8e39"QÈ˧Ggǧ1~Q.k\o  ^Єsc<^߼WliWsQ3'O'k^RVj빏>NY ~&ZWtw^-yUw:滟~?]].n ~5 :ѫ !7Ѥpa72G4~v:2*GgA VreMDТ#NRtL_Hq[V?\#}G9hҠNd60ƲiJbkmH HrX~7v$wmbIJ/~|ġ؜ I5gꮪ##a:gtElFӭPbyQ'9r2PC @N,=<B1P$=!Iӭj*NӪ .Io ~RrVC-1 N* ;E-i*Zi94-.l@7IqꅵQ*!IVjfdpV9Ntu$N;"SSO6#SIm8@K`)ZfSGy5J/\{ B1y1ڬd?afj~Sb;x3yW̅dZ('EU]˯ډ?:v^gsx7Lx}5XC)FO7ʘTKA]{rYxmUk7cҙ\#8:nAH;{ޱHOJTz{Q'BHKog}&x95[a\:RXy(5FHT3UbNr1{RK V@YRA:HǞpl6Y-J_sce?moaƱe`{E" ZfOg͈ɡ^#s:6_C)0"]CRq\J+A)8(Խr. ZD4RK45b*:uaSҋN>wͤ({4r-޿(Ǣzl{ o^o{v4~-!#/'.?CJYE,w*e?ނ=l$D[+CY);Se RPpU(ZG񽟁6|'w7&G1KeZlU˛7 2WwkbOg_Mݣfӳ٥j_B<9i g#* Sbs|_ baϿ6ӷ# t,6y.rʉq@ШQD'%9~G8_d ۄ^#>x9b?}_^|}CFhࡿOa uƋhSN`]Kx)  5)AAІx`3ӂ_K)$f)>&o>Q0e'o'ZF)oI{Q$d8b'q˦;jYtH:Fœ0' 03\-yis-=x1wbOc(*1(LiZ)RW_q>~2__^ v'?{Ƈ_5޸_^Տ_u;5 dY"ESim-/u0%#TC*`9_Ym >^Fֽ6~О8\`($֚Rh ,b[ҔIT(V'hu?Ac@^~qÁ-7Y߾Zĸ=KCLsPKK% N#cKcT2ұ]jVC F Ą+5'Е]Ѥu/6njr5\"@AFKl!aMq+1< sZ2#Q♠JKx\9ʁ@S#s:9fqӄsRXQVxqH2ڴ,3Hl`>(I5WH²`AyATڳ||H^.H5DݦH>e4VMu>3 /]M.6lC]W n*= $RRp%({mz_DL${hu9r$VZBKbL5t[kׯ-y}s{ x'磙xOgbۛwLj5mc_ ʤaBDvOq ̽c'޲6t b)ZY]d49B(K Rۘeh R =M*/a)5a(õҺDFTrrJ<*V485Qt9Ea EUjIdK;ƌ2YI): :K2}pu@>.R=uv>i}i{V`O>p;NhEێS +[)uS(}Mζ!*貸6OmژǔVz+j6L`Oi% ~['7婴T{?o3`uKB>\/PE̟KٺV&͖Y4>_I%4[.)cv_v셦nmpEsxU$ B9lw kbWFh("JɃ 6I ?bgSdX,T3ڳ*c*)8͋haŃ'nR~Y IղkxvذuVyO޽;V{1huw1ÏFk'ACF}ؑõѺʻ& ]])fzUNYOQDϧ"(.2Gu'!=DX9**7遛'^N\,"@h;&H):CNqbn'u sMsKʸs*qS\9@N'is_~ Izu^rf@N[5:oI_?qY}EvVX͚b{khןМ  DH@.ܷ㘐%<M hC~,ژbvޒM .׏Uc8VQ{L(}b;AޙT)#EI.ٚ2bRe/Sy2.:ELh>&_{u @%@2f N+#^:Go:9S-d8b{`mOOk3/](b]),-qphfJEGA?# ;iÅQU"ؠA)*mVKnǣ|5?/FP,|]zs=oF'~4gz丑_)N!@by ^c`LrY}hJ7XUݝu33I&53/" C}Y荟x"+zt+^yG=S*=]ww݄JS{N HQ)L]uhpi( 0贄7-b%Q\)gxXI}-‘dTĆw^\u=}v~_ %%Vljⵑ;,–ͭ3#?m \㋎&4x D7-ѲIFit7}{Rv)t^ߴ'<*mHMBz4]MF>P X%, rt7AA"Z '9! <+Ig5⯿!aJL )WRVxV*uIy(Š^5Z`-Bbw7$ !MIb]$> )7%5ӊ{BoB W8$SikæuEcސxRk{8 tI2gȘǼ!J%#ؠFecz+i+|σjIL|GMc |lJ:0׸l'27 qqu. )!N$>Ec0Pڀx]/wGUD]3Sd*.Qcz5m ݶCtDd>!Wb!}u\#ty(xOjY>.eϸ}xwYwa2++T6eG`v68vgznz)Ĉ Ҳ\~x":n;8OO#U~jJ5bĦcYLZkty<08$+re `@HX :ǻL{xyn1&8; KpɒG=m b2SFd'A}F/#a^v|Rsv;1^wb< 5,g 3/U?3oòe w_,ukDXÙFpm%yޅ ٽLjηBSj;ϝGBJdBq2.t58Ἄbebe8c7!} h7h \r,s_?]W]D:Z>٥l~۶^Bf_=Zxv<}=c1S:tkmeDZ(YctaL J: eD*X+[Txl')NI 4 4ܹn KBׄe2fE aɜ!XeP)K[!E]GD7g妝)<^٪- k$,XڙL%M7I4Q'{aG[ LNNQI%w +N$@*#]"ffY wh5Dtm5Lmj5tD,tpq8DQjs J9";v[ Ŕ+ jVLXKag#Y/="<k҆-Au҆8DTaXBkt+C9Ϛ t Mc< (ICqFKؒ<5UpZ̄44-- e2+Zjp ƘwTc^rϓkKCp^0Z'fHLKڎIG(⛴ip;l:k>"TV&C6TS!jXEeB_`ߟ蚴LmH0-F#Ղi$ w-ltRJEHnaI)M1To|n$Vjp ]C@%ѻ`G*, V `_LdX `>σ 4>j<yh|)8Zz1!h*2kc3fg_%D4͑d-v#`ǺNIǷGxEz{ZP{͎z٠Id,7J<ԋyzzM=!QM=5zdhMCbc B)%^I>|^&U\Uۅ|`-Ц*|I ^ny_?uE=Zܼ3MDȈ^d76VBbNO`:ah$2Ni\T <D?vd|9t7l@OOX+-d2ѫ=8'ȗ 'PH 7$q"@PANg*0c,Vmka*Ř=5`DHdP 8 s͞α.SFe^a 5b^C:nTʆBh,8jDh%٠}9-׬̊5j \=OEZ6Fm:W Y"&?~9"3ߨY5jQ|VQ5"]DOrϿr1xFŐcDeu܁= ",m|S(_^5ȌP~z!IR9else+k>t>a0&Tf0ʓJl9ԧ筱*xF9\# R\E@p'푊ר/㕴r7*1fheRoYOPDy?O<q>󍚅R\Ӷwvj۫[*6f^)cq}R?8ga QP2kbcBˡ_ uEuP!Ҝ ThQ7jV\JF|0jtwܳK0TkDIO,ǂzA{0(ep"Bu@SPFGP&#$aQ簅vˈ].w\Q]wsF^>]'\܊Rzՙ(jpeYrY; ѽY34feJbYGރt|cd\ 觢1#H>͡po&u氌.1bWiMY6h\k1x5۞IO6&Do|lη]'Q~h^*ڌM25mp3qLV4#.eN ډmu_n=GIU;Uz9*= <㽸Lkňdbϵ^?=5K?ίOIʹ|$_tخ-(%vWY?>-Cp-et{pon?t?p,~|XOu~Bm%{&%wei qfo}^nj+rjv٦ jE[FS]+]R4yW7VHȿN)zRcJ*w.9fIvtq4f8 #Te9]ܣ:=8* وC>cmrpCA"3:j\Hxe#^ی6ԝ>@!6觐'kçq SkZv]z+vڎ.~Xsz3NIjL\;Zܢ:1ƕbr7#XXuv\Y-b:/d%%XE@=YpFl9%eGA|)øFE Up>_Feƪ}]|&\gQ5BRCQ!*D| K:G lɡ>5}b{Y"D/Գ {TVuqM1SN J_thɠҢ1OV(qɀЉ/1~>ݬ*R% l ʢVX,HfAVԠ&YII.'zh˹`sV8[^C!'e%ʔ6yreѝ\w;P>PXk].Q2: !*Q2m$QkKTR肥KFmGgibi #o>G+K $Cysв/E.&J5ղ2 bK9;Q0%*; XXh,#IݨIFU7R ✫G2r6/&J"46r3ʍ A8q_Ŭh=_T΄ÜPl:)f>^t8&.koTƒ]\ sUqYZܽ:1,dWHzyxh@y2}uãPg&~m7g w 'M? %\ b*8Τu&pKK(U5egueG3(<+~U2~>K(-PM2ɆWN[VIʔɧ9ә,ύ/뜛HqJPMVHZS*[). nvS)Emi)mhqZEj2/텊1ROI(s^Qu(OV'[5Mv2iteǥ2.9-V ~7 *3ߨrY:JxQJ]ZRQlUYW=#߼wz IОq.ffLBESF0gG$cBP4d =7ˎ304%; |: 1dFLƅd$A59JfDMaܠ&g?Q=ndz-zyn~ _'f-V>h]g>"(8hzy|ul_9U^mzG}]LR1!zHDCD󘷽-\YɂǟG'oGgiO4W*M4ϒZ^Z=*#5qkVosm SDMz]q_"wjln=&QBb\1DaTŭ n4EĬ&Nʱo=O$珬Wj$[x3@gFf1YQs0`VT#B3/&z s)QYPp-AYIt2'R=0S&y1kB@Ld$&k,8aHk%<ڼwWk.W5ʶӁ|~%xrlp(e@w_, kp?.~^o77s"]ctXj[["o^^KGX)a8a5qێ(R L0/ \y*, oo0zvf<_>w?<<|x}r'NaMo~8 ȷp9u6X,8j)e#XA(8#l`L2+Gq8$4 NjTJmH% mF^d9,zl"|kP9jPu1{ygdBp#P3>:F5U#*Nr.3[Gvr\bų1ˊsR=d=g}fWWcGfY?2g(C ߠAϬS,{s0fFLGj泻L3WE-hJWMTNjߦ+Y^ճ ѬHUh9n$xtj1i QB=88fpbHTKho Ä`9Xםg$l#y1V;מ?F{;0aF @ю6'8Ya.]ks +8V"/F+?Ph/&%T!\iiӪayzBoNX>=tpç Ex;E/ن'СY>E(tC!IB2(u`5<>)l`9h>a$X)y>5߹v"A<I4aI~ܤyxJf'fb0nUUT_k4N3o0G<|b"6NEp976hBS:MRKcDR Ƶ㒥=8&A%ɵCҷg_p4wE^Ƒb_/3&/oZ᤮oLPJG.{IϑrS+<ӿ>$~68PP}Tgba+& U y])LT$,hV I: QZRsZB?E$j.~X%d%࿰V(9FxnBI1\?tg^kgFW16uvPNz)o4vT!O4etE: c}h#mU7|2e{߄)KC~,ZS\ٗn1HR rXH&݊+ݺ9T|t Q `4RKr DwtÊS 98`-vQC)NJ ? [Ms8̦i˝!W@?Gfv "8Q,vw<81rNS)M#WLML))҄qc-/+gXݡ;.**U0-@y vP2 fQa|TFqʄl51i$cQXrڧ{;w@ULkISFqbz'١SY٪dVJ6jլt-:v"N&₝8 D[ c k^j@X[68ZXAj)FJoKKfM[:*:L:$nťZItY]qBҞ"0qs鱋(Pѭ#؉Dx5U%b f-čʓczC&1n*2b~(}9DS-\u!s-)M/tN;Αn?“sBYNjsFbB +PCxQVVE탌sFbRs<X wxUDhT1U' F OQ>c+8Хxy%Z|],Υ ^K@f+=cN->c.Oz;,QV!P$LrV"a O/>g.hY\|,^ FhWn *mvG %'gq(8J_f,EWڽ掳7ߚfqWB\/'rUѧ{hJYD"{+Hszt]WԪMgaq;@f'WI߇;,o 4.h)Y1LH gU,&!U$Ug@=Tᚼ|Ϩ`wrk$p.YykQcuY}HKL%;FDk 6PcDtސ{3׏{( @rv} =v1Ⳮ,ġ(99p ^ %i="#i,- Yʋc"( <:8 apJ(Wnjmo/&WS󸉳m}a}YvaN.Bj$$E8r8. ΃8SP *AD6Lhf%gWpB%rKaN_>28#~`W.βzX[g|=Kf$_#[n}q<[#TMkv 6)EC,+MҖ=J]2k٧ÈzT'sjO ~Fi(`'?j㣂isBN?]h3%/zńvӻ@^Ql|_+&quG6޵u2ʟf$&8d*;(l7hCDe26s4^$<'A`!^JNAm@ tR}O$%o JBbSz5Kۍox ]CFz\vWi>jl>YK2Z]B*}9.6j]ߙY hOQ:lЩZF.su 9h*Ur@WEJAi3O H=(~]0x/:|QX"YTE,SADT#1"1*vs)f )FA )_HqQRQšAqT0# UNHR.mA D8O B1ӟP/@B#83":yA 4(O8# .TJEr{%6p EcA0BYbWFcb"zA%ED />/b+. SwV@ y ˜Ya)9Rb(&@e!uZE SpkGF;b$evo' SS_hII* 3RPA @C9SR1`6FP D20nyOULYFMOL p^A@ kX+AG,zH#JH%1jF+ͭf7TfE~njº7lܼ>}7 2$3$ mEݣӅr1Zu79l}4Oc/`}-rU{vtP1GR0ZC%/ E jqt@QA@~O6j";4 iMo7a,Cag^i9*oH(eB?4B#ޞ (/7 oAhu) , )Ņ8zcAܕl#9GV30ɯӄ: 0jLF8.'ȧ,#[!&8?sǍUoyyʴ'0zŽ;Alt:*/`2<8,d*S)捭w^Ѵեyx&G {{cQsњSX@l^غ%O4ֺА\E-uxɨMuuy\Z+1N5F&dZHmQ p:)_ע(t S- ?ō[ ۱V80p Gg1 S,ReΫPXT + A42WSc87kӕB3O>pi$~6[w%иZɶ8ôGɆN.L18$UN s ;xM?IL=m(Iz A?L'I!cI6< Ϝ7q5\}p__]ǮfYc&:3O2q 9niy7B5S3ߞ|rL^s_3 7eN%|d<%@Rۙ-nS" }NJE2i^mD$- 2q̺ދуٲ-K rt> \C^_OD}!S-q,NJ {4'z͑ MyTEYAy/q^A&>u_b"h`)X\>6 ⤁РKbOeM YDb)+eϾl=\o巐= .lj‰QO>'!U/Aa7oQJ3V1JjNtfuuAF)1:Y]&F58&=\Es U4r iD9mx3$ɬ2Į!Ydl)~6'{=m8n ɖ;P2PD{zOȗO27s*s>'g|gpRg=Bf=B%"KK_Zar 48C2;)H 7EML.5$ )%5C?Np: KY(cdPvOW`1SeE>aYԳuUi:zʭyS n5á2LyHMCN˹q"gw[ɜc^]^McQC7`wŨǨ ֌.Az<ч(9 Fa(^}A8I<#QOJ@"5&ҷVRzz i5#d:ՍaԏDItMDXDTJS&yu=u\9H$ĸ</&iuy` E(1g6,DmGY,˨Ʊfs;&JrfTČTz\>U2E[;n=pepo%}1KV}q%ȥ-+:+[\,[ Ikvqj\lEmAJt'iI !KK EZv9Kn}NkVl+WSh-8F.X{3 +B"hCL؀5Oy e#s$ߝDZjUj_>ZrHb NjGS%Ԯ(*շlP\gzVؒ͊b'E{IK.dݢ0ABZmQ5!(&,IU-L'\[*,.,U]/:ըRY$4(:i;;ƱLaf]b}z=y|TRҖ2֔,NY ĽȽ.5r͝``sQNK줣ԩYh5Ug5J_dT~Kc^À&Rk̬BHZm02;`hIi"FiG]oҠM=6mD4p R Kʩgf`A`88Rde+ jg91Ş zor=9W.eB|գ_5aXZs5, :r 2@6.X6B8=aJPOa \@8bsl0Ź5nù#3+vM byfr QBTR?ZPn< y7au{} ~2^~augph}n5x]-_97,FGI u{޷KpOD>#>nWp$_oyiQ PG fptSbOmzNt洫Κ՞*SrU<hN6[=Y7JFnĨNjSrWA-y-|*SR7ܷn|TrI-1IbA[raACsM)&ٷn[*1՗@ z-y-|*Sx0jIHƻ1< QhkѡjY󺰻I~|*1f$);݂d:PzKrhwI:5vK! ' kNC 3#RtpBx#iP:gQSN7HDX800f VM >{Ľægƴ R !y/ǔo2%S "hN ޼M޷nZ7aA$1ɺ-Jyn4;W$3}ܘOt޵[ɩ(|jSl򂚨3W5ishwuJ!>.x6;f`DP|vuwsHj{lx*__%>\|p__]Ǯmzmc)쓏 81_1y=yPbEd}Ǵ]O9z 6G|!1V415/ 9&Cf`g]7w 9NN/5HW _5d]?/^Ƨk2!ߝx+^=ږa?f1 5?o{t??r&nVW!}ى-6QM1t0Q0Z&P@AV)RY&.?Kqύ0 O5` n7~ж nXsl8sƅ֪wJ8B0LNyHa`Cv.۪b!&C*Ih <'An4XJ "yؐCĚTva8JbI;-|= ,tB`bTzG8~* [:QPU1Ή^9- I߃o:ivw_Eu.fæ hk%~c E}D]ðZ^*F#O?UQE'a,(C)Gn*{wo.mv(kUN\ހ/r_F@6F062Ғm~8Ґ=#?h4?eoc$|L|s\FGWVM$v6{y'͋!{PNR%+'a^s*;G0.dTv|e s(5os "JApM~4ʧ,x6|:ĎeLve#/a,K+;Ư\"Э)ʾ 3&bb&g?U"dՒ;[6YŃvݛA&Ƣ.(Ln%ȝ7oإDf[_ /1|._M;@ Je%A]] 2n-s3aL'QOڬ2V s v[HAr@k q5 F@ EUo0^dۊ!>{#,fË5V7sdI0)($/8ɩ h [7dDdEu<7Е;"-Ѭ=V|$t(~0*?m(FB>~,k@k3mȕ^ct_ؕgi(Tj8ΐnJtVK0 |=Z{YDPmFr]ڌj:߾[\% U;@ooqhmlc1Gz!s9x 2BhB̆NtbTuDk`H9gցMRF=E1 "3rRR}c_Fn9@SU!It1HiO.bʔӰbv _9d%.7:A.DnjQ,lp%q F+l X+s\[B-ZnQ2:&5`Vk.e r2$u(HvQ(5ϵEαVVQ))rYM @5 BFY>B)6^ن rP9!#I=EoSva4qk7JHwQ `J,eCPt'Uwm%q7K,L= 54Yͭ(jgzSQ*[* ATt|(헯=2ڄNI%cxTju^f1)zT6YLj銗[p6 ?8 MpD st3>rRPE_*MP:9\lCQefz$Tˬ[j#O'557)QnK HG2Fefʤ*hrzSŃrV1&']fQ3&Y{%*mәYP&Hm29jQ!tq SB^1Ӻtu5-ҧ-ZT]nJdf*8}h-..Slq-B+(&U,([ ;ÑKEFTj$3Ս,AP; ҢЯztZs ֪& NS.ݙB =HEX܇Jf+Mo _6cWb8 iܛn:/]qZ$:I^mfbJ[)ɞr 8BzpF'{;5-¿?2t Pn߿ChpXx1ou#_a:,n^ߞS|?_tm[6q+1nyna|5pFBh+ OSSH8X/ZV:ۯ ;LSByFmyǓ Wa~*>m@8XRsǜHd7$aB] x[|PQ5QJK]VڑIQR8i}ѐb;G[6d@!:ڲaM1xxVu)ld718Ǝq n|q`z-_ew9G%a߷l8+.B3>(apr%b,hXd|]vi . &HI "E6q/I>k..uC{S>@_^Ϧ/GΌ^;泗_g~;(v ogٗ4x`'x-^8:F1>{sıd+^ɳͨ捻BJSDTx y+LT$@STLw@3_9HOTn#Y:etDINEL"f`$xc /𓿼gO0N)1/ޘ{+nEXsNLrɢF)B L Wù[UhtBo(^iIzB/}v.摯ߍVskXЕF ?mUTp2ŋ2Mo2+$e`U`M-%.`)F(CTi0^] J<^PdOӪ'B&Ul`1«??7܋Ag A/o tQ$@?,dLPJ1$cJ)^H eB |U Vb7X UC3{2D\Z10rI et5^q<3^ېn(C}s^WNގ˒߮vA=y]5`|L C njvZ) -0hah}]0Cp.rTYt-Rdȅ|.#WGf&| Q{gK7ؗә^%UR(n5+r8&EV߿%o߬Hzd:/x=)Y;qTfWa3jTXXjmB'R%鳘H`#r)1VxEf<ΉJ~7`k>R ƃ{4#sGPFN O(P# S++%O>?x+-6N 3:=xtGT'TzfJQIށN k!2NVjeV*;aoLk㽽٫AjݎK>/k'_Ϳܴl8,Bɬ~Rf.ƂF'x/rdIВ -O:.:*@PE/B৐`EC4.cbBopi:_}9Xt)OALYp_E2&$c *mTy2Ys.ztu7]^x{tAٿw4/EC9}J4u /n/CJVQyymfV ^SXź6[W(ϩR|-dp9\_,Xͧ0ˀpv(H~abZ{/3e&BU0mF۳L:6//ef{woyCK߃& 0)mmuroL>9)GpRvCZۢ23© "]D ‰:S >$3fL7酽1,`'b'G˅^a;[p EF_Uk2@F\ǿ͒$*ۖmdd{;E\z.*u'o;l"ŐhJIm4OM3_K}X[|Z)X "~Qy5SeM}9gIV~H`oX[36]q]x݇G ˃Fe?N&_|Y~t}1,>E_ft OCpcm?)V|k@HJ]!ԆڊXGGJۈJ!;'GU$>MњlUa8>tW1INc&ѱraћGL̈FyH7!F7ߛ40sA@gC+]=J<,hHd*|zԴXc+kz⁤G 5F%⟍̻oc\z Hc-\YR0S$AZV{6p[9=Z/Ur%IK'%-e$ ((I]RY  jɘQ' &-֑șPY]ITsgT$F(K%RM DP͹Jq&p"wzq ՜ Նu s VIӠKIm(2H[,--'S0[?vJ읨wp*ޢQa Us/y΄S_XYձ{ޏj-+rFT @($)|sIH*J5Q`L 0b9݉jn؂4֩!S")!d$5 Y6aϜv4xfR@#G*ޤBFE@1JVnEaU*CwB7q)0< *أ1@:ޕqc۝C`kM`)bIH꾽8^R/2ZN Dvwxs y.R1"0/hɋ$g@g(skgJV3đ" I A*#f 7,%AXژM-Х9RZS( ZjKBsjul0\pbʣ 7Oh)JǷ?w)nRy`zkHE8ԁr*>x0ͩ2Rzp=MV%a( %9aüH9\JQ`E4Ѓ15)s΅Z2~py~}43/GB!8A7 Hx*ZcHe5()ŽTmcE2D8BxI"hXvaS (ǂ93*e_q :E_CjGb[?40|K;nZjbZ$+6e&/6墺_EAbxʿdϧ2a8%Xn.]o߾~ྂ)sjlzߛ= zRls=l%dL3h 0nnT![ BFvЧHunъ[ 9Dc0ŚROb7-!ЉmuC$ ;chEC-BQj瘵Ma9vAtbۨnvnъ[ 9D00jFwAtbۨnCVٶ-ZPvK3hF\T-v# bm6[>=D"!ݱ[!g(LQ5Cvc bi1Zs*O nh-BQdA Nbm6[SQfhEC-BQl& "Bbm6[4Oew!ъ[ 9D0/fX :ma.-y E+n)r”EvØbekĶQ}*[F+n)r”b0{nA*D h6Z ]>biT!VS9sFa_?d#׭`X@':S*;chEC-BQx$m 6Nluح!SU)6E+n)r”҃lrn ckK|mԺg_neh=9s6ǔ\j}2 KYoet]dߺ [paօeVIefmտufemmƓ9{wo_& zT]/e}5 Ao=?IY%J{zAw M5&]k@ýt UɱsF[ %ګѾlYrrsofaf}|ͪ[.>HǷجcf{s5Y U U}/o?UwFo}YZnoœC2I-V6 @Q.MX t5sᗸQ״\-)W7Owb*RQ&$8*&+-WA>/Momk+|=KAK%g*l*i * B] v0ՔXiy &P n,HDrAkօG)Jd^rYJ R~Ϯc"yx nGQ"| >ݕ^O3hx|nR6;FͱhEK#$@șC4 SMK#t8<`7N  :ma>Wtx*ZPvK3x Xj,w%"7[6_l :68*cEMu3lp6@hbb F⭯ho濹Xw9wŭYf|#HFbLS$b;@aXs[o'3G ft3Fd+OQJiXTgFSQHiƘ I ̏Y{Q)!ęhcJzuLNjҨGd9q_vq afwЦUٚ=,63 My2maؼll`ç_ 0|8 WT [mN!;q1R{S&d|z'#E:F)J7 =((C p'{s}lAO5@+K)\s7f-$W8ր_WC6 Kosawe. Ls!FXRk (WƺD%T] !5^ …0q6\lVpm)@ rÔĔKMH hLo8\KQ9QXd@||MmRP;/ {/f) V̓1G#;~?yY7^(4gc6"lˆʝ/{wطab݇8b1y_Cb'ܙYC}:GQR4 av,;GHBCt¡Kd%jϤeV8#IGR9J)|Z.?,t)gT&} I}!IG%KS_.y' 5~QDq9z83*z. L4ޒEv$ؐHl5+H0կ[+q-|0+%Dsi/>?(7J|dn+5&"B}Bw|KS8Ԛ=uh>_LvO8G"XnOpJJHe/?WB9;'0(#x$e2)RIߏ2 mzhyT6Q&WI Fd@Yx)xB<& Ɂ(䀹=ۤ؉G$ah pKV9C{INeS7N P:$ i+wI@($I:0N+.ϯ;u|W/+E{JDpe`<y-vIceLQ$,EL"#HeBFZJVXD'MWn:eǐڡs- [m|#bu\t$+c&_p VGc "YC0jEyV:Yxտx*T#/~„L8p.0"&HWC/@*IX%ckJ%KsW` K˂YlsK"HHrLH> "d=Ivl$3† IUm]G3*T#Q mLsO\p (evm5*$ʌZWJiRe޻rB>4ɯ7>j7;Aa91wZҬo77.`nݞ$PL2뚎߆Eo(/S#Qi)" T seAa#܁R:^޸7Z[o{(*cH5NzV% "$'li+VðehW2[bSy)/1#[ҳmk~3?vt{|}fɥ}zT6KZ^uE EkrfF~*az_kr@ADdfչ}чR/[[ˋ(~{tw5!;XcKT/9}t(TI vn~{Xka͏Oc{.Fro_3ﱗ/N~r@a뵏V$Q$Z T*ǫMm]@LwV9X#c_rY64XpN݅3 0:%~yVK66Js)f#%u!|s%F RS^YX(!edbۨS?>f](~t!gL 4TNkĶQzH+k->$R !)&?<`7F*ibzdX@':*uY7>f9R !SD"R8#u~/@CpQ9ѻ#TC^;XRܒ&UdY۰RYې7(kD35F JMߙS[ӤE6bM =6d\4fjo1ٓ;Ǿ0|&rp-܍l' \.W1?Z]n;`VLhigOGg8\-\q=e2˧#ɞݴwݾ[l5KyfI˜L3XLReeI)R<Q{G߼[ix T}A^[3ΒqtnݍIr!1) B@{7vPEs[)?k yݍ}WZu=h ~=fnHi!"1{BF5 /c_MHSzDXӘڕi(Xp))\jeIRG #kA9‰ CҸz)%=hWNZ\Fב $Nr=/msE@!pPxwg2[8>"b/z\lvBB.yiPuwI$r,XZb !q͔ZKR5W185])wUErwKFD5[ddSb32) BYSD 3:GJ73a ,֕٠KɥkqӁa:]̮5ׅrh42B!ES9*(v2. D +kJwZ#>,.'gߨ0?@5^ }o;իx~? lW}WuH@7`JcPq|_7`b5k__Y%"mEl9mlERyb[w|$61d|^_~@u-nZ1V~*B<t RMW?6`<[ޛs*iHIsbT*ߞ=LD%Pt|4 5g dAcnK^#F Kt}wb a):]==owD{ft]`50o-gZ*:1&ؤXs!(2C){7dWacx5k>aj]퇩YOe]6ڻb%TW[?o\fr>}2nCp/ ¾=kۏd߼ep6Όozӷ^}ܖ7XeJ/|[OѝxzB_x?qp"-84!F{ܻKQǷoQӰnzܤG!"pyfn A`nen½j3'E#nܒKҍU_Ec _];:x Tu`E,Hc pCZ" 0}鍺?/on{KAZ| /I}0bD8 p\G>c$BPz}/ ?t:O/@v)C\ZkPkS8uXz {ıkDw[:ȏxv(Ezo}'fp!n[;0~eűЗoo$x>^F{wŷśawNEk kV/F'^CE!h_) Wctx;W/Gh?C~])G۷nOm Fs-MѬAv)S>ncnke0Bcpa:sm}:u;C~])ew(=.A]-lKߟ)v<6o=;I|{#4B4os*_16crc \`@pC܅j~.]yҥ+0xCʣ>ލt5g>Ezcm ã`1nQ/x, , pA$=pLHjBK-v,r.2kF)ұHcs dj @-5 %F !,Z(>*t-x,c5'XbaKB}Z<=e*tͫ,>}nS63#ƢH¸)9{q'J\N\j DN-rۈ`7G jѨS*>R$^<{t:)gցe[9)W$|67?fQW*/L_/m-4'?|lb䇻Bz4# 9+Pf5䀛X B Mٷ-ԱkV0߷ BpHC&&ښ˅h5CAzTڕo8yOgep|(*/z;koݴXí.m4~*Ah i4O:eNQ C'α_9x2G\z6_*_ڞKHrvt)P ݍw%ټtmi10.)w,{w>Z9 K\)LF'R6y4ƷfDrwW i"i NГ"i<2Q[+T4qR؏jEZyx$ށ!-`ƦYH[pSZ]'֍jK;/ L8m9R9\*ڳVhM6bTJ$c@O5JhޡB@F $&m?Hd #̎@5 ݍ&I{fƀTS.kLD A R&9k!WCX?/6hڡT2)pcoJʹD&5 2,&6R`O"3!Ȥ(Y0)l.@pڑ$_ˍO5gMO "qD d99O%+y1r2A;X!=gfz9BO?$qzerD"}1hHrIiC61RɐI2 b*6%R-}FQnjbĩ"㸣@YZ!(xR*+8  Fՠ qRT6TK|RuΫtiy9ACywTKrORzR NJ,=)^' RKRyHUhVx(V cXN( <0PYrMx6ITۂߏj":Ԁ*jb-rk*++3"m'ZΕ3~$h deu,ǚ^!s2BGAk. L 3F֚as#0 j1tZ! éI&m@%(4N!g[TJlN<ﶩ.>iߣ־Il0ooT:)mڜNYJ,EJiDp~3q+1zGHV*ɿ C Eh͈HO>I`^.-O#䯩J<{hm):kcct"b[_(%d{a)( 'icKߪ O`$f*[#8g ~pZ^k#ѪVq!V r!wTknO(GKU}*ׇmYZ9k :$2Ʉ04Zx/*&(3rkdJ"w~Е|->`*nleA1S%}tEt\%Uh"΃Ed1ؗXc,., qVPvnˑdb!SQ caArC~tX ѱzz QUBtӊV\!:&,] ::1`~ gtϥ vlj!:nϜʩ(!v*Տ"B0Zi`mRAsa.s'\sR5ox޺slڤks+qW@>rpo7.>Z1Хt; n&bKV)9pۓj+5"Mvĉ@ {0>@H`r@4Th:N#4TF8j EUKzaBJ҆j=IQKTJJ|B{*8f)ՕpzURF!R]YNG. VR zD8^ qpYJ"pUC29րS2 z&+2VsD7.MeRYgͦ$'.oAL5J>f&yǫ'S;2[3@5ecdp=I6^3ɻz:kڲnD`lF=TP[L Kr^f] b .1^̄]sP3\L*H00Q&ih0ʙ䧵68#C4 lu:|cSV˸r\9fߦ|#KtW?)W2߬Mt| jkNU,*_cQQJ^lt7>k q3.LZD ⋠*j=N*nlG xYDA!lByɜSI3]t R7uhծT,0(m&v fJ&#-B]?]ک{L~w? .?SWOST:TwZQOYwbNroasDηF7tv&SU9U擥Eԁ`5d ֶ^J,JUZ@>GIB6(1G OtܢT&%'T!l I^:(,C")L_)#ޠ<`Ʌ_Ed܌t\OmZm1;~Z<' n$d3׎,y1W g%_fZ45jwP<>Z3NySjܚ 2\:ً006:ל>W&;i.+b;VN.%C9 'hFI(}+<T=rET`Ls a,gIɔ؄]Y`$b6/wX`f2VcdkAg|rF.&4ئc=E긨 0-@bNZfܡ]`hY͢DI#A =/|^(yі͜SB%sZG팑hmzI$ ),Z] /_^ǍdЁg5!L2Iay[7R3 -1g5hx%Rq+Q7ktVP.nv#͘mUV^6'bۂfo)@0f;o˻ύwE=m7iW[tMRDA#JVs=Ea5{w6OsVWeCܠzuPk:C_zx硟\x WJl5@zu"_ zzU|1 4(9=}o\&_mߑX|S3:ʤJwh69{;!L#>8qɄ,lm"cvWݨ9cYnv,5; MCW? Q{NXBтbIh'pSmDct447՜{M55m$mL>sNA^Oꚦ+5.@U\!ӕ+^yP~n5Z`1MkTZSxa]|F ܼoۜӿ_s}AV1݆NaӲ|ʑ$7.Cdw -ɗڭ ?G653t.mAYVnˑ*A2eB A vgBɴJ6EtKDLB嬹*lS1EQixAm4RLrBAgΚ;J<伭7 uϡo׼)2)Y9y K8R")|/gץwr JrV3K:ۡ.nҐ1VJ*m_,oU%sԂb5{Ms;z"|7&tW`3xzON'k"~Be8PưO "y^Υin<ȹA @p]~%"WB3\%o.əzH24NdlR9TXw6$X$qN1}AK%@{ BfOKN[V5̷^DƒeC4>Gy>|%f6y- b؏/ՠZ e;YUzأZjҥ=RʇI:JYITIP%CAzjEHܣJY s8:lː HQۗV,Ma%ӵM_c4#_$jk6E i5KfboIϹ}n7,))ǰӃJ-K-[^?>#6^0!f)3He^^ 􍙨;8V^2M*xcb%J<^Le>Ԇa$]3\+vZj3P~H5m[z)LDdAX=T0BGe)w[_ Y=R #p'CN>FC#qƃ A-s4HΤf,ghlPdb(Zi& `dj)vA3 [iN9qPxI !";b&i9ؗU/+`l |6$3lkC&%&($% XZ6\hk8」V`Ns!%H*04@L :z]vnV4۫OB++P~ˏwy˗B!;?>+"ueaO?џi\-@,?B*˫[ufۼ;iT 9>Q I$ ᫓3{u[t2݉ap@&!W'H<ݫt?ĞihZ1Iwdl13x6x@* ZHr}$}[J7Fj @R|bm@"@58%I06x .?U}S=V{%tt C%Эf? A`_ܺmӼ}m(7:Cy|N-Z`e͛nY,.nD=ioEzr-c.2\%\Qڍ@0Jj&s[F?s[Ǿs;c!o}}UK/V 5t\c lA l>./50ֽ6ۑZY4 WM3J)y%6];9MʦL7nN3JVIM>k$ӻ a!Dl nJ[-9Sv(B(&nm y&eSs(R*if'U[O=c[ήH]&OC(rGMK*1pz3m)MNao)v/i)! #p&x>}4MQs=R|J;}vq+MINa .b`,-tkZ12 l5jhy8ҀJƾ"#K6-p~Ŷ0"lI` )H"4e`kXû力1yfP!۠Ƨ*ƐhS!|0SA7uKh d ac@I2"y2uz/k{s}Wˋ_傷M" oEi<ξE(*6MY`B^%ܴz?~emxu ÅkۢzmyۥZwzvyqq=a 2\?$Y\EsѮ|\߷kw LSTC"ETQ4Uhlh<#݀:?][*d}ЛRY*l%viSMFNn{.bAon#J7>,G$Ja;ğ{GWNhLs{ "s$b D'J9z6jiwl az ks"E ,'jdqSj8VV(pz0=+ƐpdSa?1YmNg=CK-Ap'uuUGuf8Bj^Zjyk R9uf蛩=GhkRW0Cۢh3FBJ(`j+18]N;ɵ_A`iT+2r˺wRk8#bW#.յ:I L!,䙛hMuȔx7:ɔGnN3JCS>t}MBwBVٔ&n[-9Svĭޭ4ӻ a!D{ХӟP%t NO%o9 t)83Z|~T8;%y߭*]u"y0+Vs< 3KRwx_L藁?4yd'GH/Oj` f]%Lޮ̲.=Y1OF1AN٤/KL,3ShBe RƞpW[[գu27Ȇ[v/Vvq$n9P;4K A%ϠɅ/"ڨ} S_ K ٝd¡ ږ1aW%Ur__bt!&TI!Q!URXfυB砥Ɨz]X|=\# :G8g@)ɥDz'C,d?b{ߛwK)P\vX8hZaKƿ~~bs_$Qo텲TU#WZrwx*jq*V38]Xs"ԳoP'S#uzKt"bĽ陧QBTAz\iݩ31!!ej hy!&%)wI~Ϩ2#1Tf:Bv6KEk&Qbr3 x W*FȸD`yT ,*'7A4q]LP&ĉӀ1sAa\29R5謐}@g%%\2a #fn",0TV=NO nFL?<&}|Z1, &JZ *tk`t u}HHJJjj-u;0L I)J@P+'C*qTۘGƐy4)BMS!;1|OR*dqA)HYua8EK# oTN4$"x: >V\%sEe z&ݲB !a'/@M_149qMϾ=u.)*9d=6Y٠;cA%T1_yb:2O7Sݕ\<gڪ76ago.\_=1PgFS޿J8 L%0Ҥ/ҝtj*+-1NHRpLk0E,>]WǝV 0B7,{AYc{ 3iL. N^{QF*XtD.P(,j6BA-H=kU_j͍I#"oI:CJ%} V-SjǸZrQ^iӪ+6(Z"/?#`!/.{ZZNN8.dd<07FEP &YwRgYPyWbœ\"*yդSi::w!{ y&ZeShf?Sz7hRA>ĻoZglV=\6^_yu 얐eoLZ^ٜڳkdўN!\epͿVj꯫M7O[g{pBvjf!gb5,bIsb'koo[9^{WS e颽. b_2'^ ~4L]b_2#^8gߴSR!IFzo-_QI_?q}u a୵1j(<6HVbI+ۖ ͌r)@{;:n6";p@JqYEzuPθ1ckK6whbfgʚ8_aew̑qY&3W93Jr0LX%`$]b 4K3"7FQ"cHF"qա7GQ, ]MrT$M1  C^7٭Ar٥`y٧dB.dX?w!,\s~Rg{fZZ6Auj**bĚDӃ R##|F~ ;%ُFtym!JP.VM9NɱPWU2.(2Yu^pWZd%_yu-(նiԶz-\3pe@tWpE\3'ON#5?; $Tpө  Wt!r~Ө1M4UMJ]Msdx^AV7M7u%e3YRuq$._<(X*&9o^*ۗ1"? 79F#۟/Pދi!Og"uL[Go|~Wg+9DIsm~6]mߥՓS tm[\zzv5>M$k4j)'/]5S[8Fάq>-" 9S` `Iq4wc#8=A3?~h&j2L+$!I*˃k_Xa`Sg敼 ̔.d^וBD`7W]Ȝ+N2+ZCGaܔYiCdxnV* 7 . B#IL D* ULbtȸq}Q[tWPg+UX[RA:PfP %QX)@6TK"Y)Ҫ'"YiCuNQOJU!*K5e5yFVo窰fwMlOP*?TGV#&rpv[ yRF08TDՒAg#\1hmOI@p\,k},hXVo+nD|ki`B%2;b ꬣa:1wmY7tZck0i!&6fc߾;.?E(}XH1 2q!": arD<j'OpUxQ EI$2Ep!cnWO1eq-zs"cm/qcYKpe/M~ ;K?!jW h0Nʏo;BVp! GAQjч`\5.bu=!LM ,^WڻTA=X#ʠ" ,yÓ`H2y *6Gp}.FtPSKb&7h%yUʦ>*ycb[BU 2]kkWukK2+ ИsV*MpӶR,e+B90U?P 6䩰aM3Ȳؚ'gg%ȰCP+q! o 9`}@_¥>`}@_b×#p\y`)sy@ԚU׭UFd$-xf V!.f*j8h^u3L74'I{͆M|6\,)jH!C` 9EZZDp1)3wģ.g,nBOM _4m"itȨ}}& Fn-pc.( 6ĐEC΍m!fҶ" iJIcٵ#Y(D+10MUz lƠJa%c oٱYF JoE|"׊|7ÅX\zzz]zSKo.-Y1IM/JlAқYKқ^¡+k5E׽;MR.(8 YE ˾KoDu(Ap6+ҳތ'Xzb!TU+b!)YvwM,\1jueQC5dWNN:=KTdR#BZndv $4Yr 7凬XHD󎥝4w"elQ-)VC޹'0VZuxnrie&u> ;m+-LZR4T/ 'p RFZҮQ)I8F* K z_0ᓯ!\lP7o!+l[piqUl>CUZG:)@[_+4GxM 3Gx4?&8 aOQHuνF?%VmEC;BP{28?ދRa*1qVv&^z`ՖoSg2qk4"6@QQ(r ocEԂ V @xJ92w) {sI)p@7<$L•›޽79}c7~Tݻ_ś85*M\[.z[v+aL@5My'w41糰"΂t'aS"9.]j^_sm|w{g} J*0*TF[[!3LFPmV++y!g4uSyQ-*W|9%#fu wg^PYq4׏gI 3]Ο>ьxMZ0Zgmf0d} N&!7)k ݼ2)c)v`'u=.rUJ3mEFZ DÐԽ5ᅇ\- պz(慊DF l3\Dɜ&qQ <LOz☹ im0VTx RʊtDCpȵ!9bQk 'Q"g) ul֧oEٵ"*!-B5(zt}ʂAqAY"G Ie_^@8 x.IV$cڼo⊏ԥzX_WA@?jMuJw eob_ XdA_Y]2{Z!?|~~?ɲ_ݫ vō]Kx-ZG $e Z?~|>fl%\(OHh//~/]ɗPSReNL]hګ<cR=~8C/wB񐀃k$@^Qut2HRz8y46mǃ 6Ow\E.k岹|uho9Eދ"B*X.017퀌IĿ>^RC[2 @M-NТv|jSwgzi_7Nt6- ƋgtLuw6eܡ>|bt(WbkT|k>!hCtYv.mlVȕ{G(ZƯ"aV!|yb]yPQ+Vyӻ >s|AmW?2ܛ׏]޵q#"eŞc",,6%/lGHGl-sdlM!.~U,ɪSHFԭy̿.]?#é$ QAA ~'"qK8»4 W@5~I$;CjCsՑNN] @F mC"o3K#Wz#xqA[9-;DvDw1RދܗJxڃCņo{% ooVnn_oV߭kW߫zS}RT.?ܿ˟/"ǂ߭kj6+3xX~Dm*8Ko6 ;֑%(?rڧW-UnzWnY6ehwvd=w trݦNczJS"[M4˦}yݬE얋A~#ŻM ۢ2ڻ姕Gz>,䕛hgN\ܙYʠ7 ṥЬf($~Z'Q}ζϹF'&p/K5#ЌԍdGDl&k[(=a$a64Y s| T%W ’ad߲pQ"4C< xٵ.$6gFfzJ[j$ݷCnt2J׹Xv.ǯ?Kp7ЌEguґ8KMSNey_mE5xsCnHyRL~ pbƬ18I-eJKDV~E]A4JXiUjB UAeaA>+])w{YknI`1'R\}:3X|w+0?O~ SpڮufaPUT(}c*'K4S;²c)E oa<߯Rqy1$ӈws?tyMB%lm'n|Ug|> rG 51~ ,fg(̳[sH Pp mx!%T4\Q 8$aͺuHNiB*Yz]քl4HEb@_Qj׌btI!Ic<(Uˢ*yrU2<3f6yz];ʆ3ǎX2'P\j_kYӌݶJ*< W5N lIda#aK)dbd=1 'x H $eG 9(';2$GxZ[B͸@eKp4-Z^hQ4jOws(1\JAY&EPuLwpDK?=WwV2 ˈi JaIڔAYSBU6kc#*$.= gqDgG*}"}F#~X:ҼUUBƵZ/eŽւ۝ŭTᨇGjw-?7ObtQN!ce;O;Kӟ"gtͬ.Giw1I8jjhsPp[z\BUːp } 4,!N<^}aclgHm= Hv R+ BN} T]Pn˳a%ٵYv"zw [\}=͡oF.1M r7c!UU2tߌ(oFM4˦ڌ3]ΣY.16aeyѼ[~ DwB^ؔbwnĘN7Rۄ)Fz>,䕛MEJ\\ߩQGgSA#p*"Nd[o-"[@% Z@ *Mߴ?!mvk~5h}44K8WKix^J"E#k0,FJɎ &tLQ{M?z}WpK&:GgilޙxƱk ύs"jWq8ymi۹;B|z"諢(8 7#SJ9 (Ve?R K,De%r-H(A-i5ҋ"ƶ6e%VQ%I PoDn0ڂvJ:˚.V5Ʋ򟦤 _ ղ!yi^oUd}K}DU^[_TZwtwڿ='}IѴB%2]m,Pn>NEigF¡Jl+ijj=R ZpzoQ)}i_) gL+ZC~" >PaC=a%g @U8[E%7vQYTkB(yCٞ3gJ-{ k^Ȳ*Hӊ$ypj>kx`F;iuhwkh;(`qfyxa#G؎ |8p&L  xQҡO c>\B`#a)8Ԉ%2 Vjo{Q֒OG##S5Oqa Q8'B@ǽv:nt^8W0пCC?@ܱ-Aw Dk h^[F8ez8z#FO:rRX>i }H' f׺Aw3 N=O@NI1bۅ4o0cBu埂; f yFlfwN)#Lשk0AkRg1j :go&1]v0 wݥݦc|n Fw'yu/, #*18'6HVT) \S.Jz=>ean/߁|UZB6R,|Ar{ZTߒIZG1_xe"!5Ar?sH_D7^ H5oIՍB9J׻}k\^XR1=kV6?0okuNZ72r ΐSo%$aǘY%VXH 8}xـQMe{(잺F zaaFna%@Y~2dQkd m ﰮ,4"aTKbQO5pZ<,Rۺ`[1X4h QQy]*dKc}K31 /a6foF:C$˜1C7^`e@:-Yږ<%**ֺ/QV˪q*k5?F90Zl \yc埰d ʎn\Gk՟[59^!딈\#!W* SKiבБ՟[M>>FsV%T~35G_J{$FFaݚr|dž)“hww~180F*ׇqV GW\6!}id)Hya"+#>h!L$^OW31#uvF{H4& kiϧr`T vВJ'8b!@Wffqq+ն(*ҴGo]\f]Uj>11]_ ?F0߭!j>^ujvyIs^G rou;V% ?XOugjrs֓rͰ)~;n-& [.16a4C}4WnY6e#C- [.16.ЉѼ[~{wB^F.aϬWqGXbpJ_FS~dvFU^T mV(`܅n'o|(1. #1<騛H )fxM'h$<8HSg2:zPx7*&cM%#]#  ^ե,KW7R`KQZWZsw9IT @&.Ee=3 E)<,nutu7l)$U<@~ ..m @&aн8Lc7VqyFe,TiUzDLc{dcxEqkWc[pJzٹGD# \Ok; a7/YExL -F/'yq8Z1nD5%0X h29r(U>j|ҊXÚGxK{^ܭ-OzϮO8OEjK S(~Ԣ Fs>\9 To{~\ HrCa!V.1HTaqvs7yn 9n 9Pwz?%)`({{ML<= _ Hk~zf)ݤfBw_åmy1y-ݧocL;Okvw˞8~"'xA*H rJ[8{Ķ,`A+Yܷ?m@V\Y43³eHYpUDJkPqkpѥFTPW EShO܊찥B89:lyZ |I8 -}u921iYk)<-E1FhkNBK_bjvl~1k F뤐8:Vai]D+R:[K+#1@V*VN9ܱ- xyVbu=L&\(ZS@]p-bs6~Ѓd7;AKck\d bl%Q0VbXڀ[1jy gK+&FC{Ϯ©̜zȻ FԂ`Zb`mf`W塨,,*tP]Y'2A9O bItv,:㗩X8XdnW*sZ*2ϐkA7>NZzZHiiDrFf128QB`*O;_V6+Jrj\+C |1m%/֩'yfSP -g'N#TcYUiy%lw<-}u S%yk)<-ИRyZZS yk<- < -5ոb3RPyZ *eOhO×D:(aʞ[K3qɗ2nOBKijep3RYf T L!zld2BKj9!Բl+T%9}BǠgBǘR5je9q[LJ6$hAwKfyX;b9B]/+]Ya!E(2X)\{K7s^ #!&uJāBl:&uj\}Qj{6bd9ZVHjjNfAoh.Η%#1p ZT"DDD 8]MJjk+>j$?>g&j=R1eX&eLdj:EkeD<3<wP mX{>%GljNvS͇>|v}HzZoEr?, 9K_|.]O_>~f1&^t6ԟ: aB԰.I?!_J仇w̍`zk*aUyw% yli 3)77-X4K}E"<9 9Դ.x(,jc2U撥' 2Pv`<۲ʋ"nwaDk$.ܱΟJ*RݥǟϿl,͍J=BBeOJ\E;h8.lbE|(ڬ(h[IԦK(Q dL}٠HK߄hDaQ}ggr /XBZ60Ulp/Pq]X)|! ,u(9zOYk;z_X̧z)&shձW8 +UoɄRA=&2=' Q)(%t2hNʤ:UR*d -Tm-h-<$~Ih)fV?ͦ*R/6œR)򴴦4SYKmf=n.STbz%r.:җXpgyZ yZZS \MŘXHaNBK13TzsRB97q=g+?^8lKͶZl¸:0d=w á0ьX'as ejO,y[qd_Aفc*.cpP,MKp5)I{6}c311D}[vO'YˌL0`#f:-6F`vxَ^{ظ040ɣfqE)f%*P7_sk|eb;,0Y_񶭠c5^$?_K,c*Ϳ@=Hyb0E{QRPx擯j2u,"SU(ڕk18 m''&Ork~E0XÁ)v2<&8!hrwIݴNl5.hdU*o`Li [Ѵ*+}2%ɵnJyL'(^IWx3 sƒB $UCwΕoO6+}P\?՞.]xnI몫~|'?ɇO>4K5 Ui@*tBa`,GzUFi+6:3W?jc-gR1wTp7K;|yM8PlO;gCuҥ*+Gd?* ;'紊9o*x(ݗB#g@AKF*cʆ71QTڛ=\=#~mdG7cde/fyd&F)c$& k8ޓRb>h5?d>,,Pmo{q}ʋ׬wv7̞Ę0ךA}o WfM/ظSo:k6R*T_STf"SKmDU=T2^W nL\1ʼpV6å ]dy{_\6.MG8o9Iߡ>39AjTapucì0g_ll^r;:N]lIlg#uۋ.F.do/VjЃuu^sutPf`l?3[fr (U5|c̃t;>̞Ybz.[HYܛ608 9 9 EӃAf%8^\7`e}x1B(;0,۴w&_؃lkzf-|~d^Y؊S8Ee*S6JlUMb9p?F"von $pK-M\xr,# J,PESTs2q+1yŽ'f[\-Yq3vDFBR 9GcȸV-YR(U"`cIq ^|E"D+,Ҁ(KjT-X9jҐ/VySХK]U`h5F],\uˉD$E}i i,%j!~3IZu F5TVҪTVPb)SAWAYPamݪP Jvš|nݪh[wj/֠zǧ?iT)O*m6EE; Sr8H7/{}lǒӜw+[dݕl+-pH<:tJrU8h.['G\-BwKϵ׼BF ƍCn|6}qS(q&Q8⬗7Fvt}9aV?~ױ ^<ydOSeI}s8}Hٵ{(d) vc]'m߮+&l&W ~*t7zxZ5w*w>$8?Uü-i Էz".>LBeֈaJgP !Q[aŒV[x#W7]]z°jYCէ_|HmibFnӣkqTi{#c7\ѵ`)5ƫQJhp4OgsPL_cLjRElۉmۉX4{7GC(ض(ޱv,$dւPkR3G1PAN΃HZ KΆ>nH?U B(ˮidy? -+\EN+5]L~bR)?w7eM[1} gjXQÜwv^h%ԩnЀ綠sE~o}L?#*좡' y&dS F2R11ox^rGngޭ y&dS >@nm~@A~#ƻpӃ:f4wn-b6fL5=:Dp$IʵY/旞gHWhvS9;/Ξ}ޯ|?%6bt]6j6bGh;Yy{-/p<^.?ūUf<5OmY%00%xhd'*YU21Vu~Ve<*9̸fM*Ur(V)`&r'*i`nH=]ZW/K%-D>iU;|9;LeUy*B\%0PPI!lkv4dѼL3J=^q2uBd..`7"(͕0ʊqE.h3iM RqG-l/>>n;ΨJ2pPF>@Y:)BV^ĕ3o5׍dBF3f Ri0?Yqp BY%/uoNOhu^_є\GLQ)}=F(p63Єq@#hͤwKѡ ͨRC4'8ߎX D{f"D> zǶ |( Az4f@ö3B +;\AK0׸d z8+(jF̨%U$W/7q!&FX=bF D58e@&RUut/i 0ut} ,S6K-)`3NL3ؔ;bF:s[$S8򆭅#49ڙɠ+%)`F\pI(ڈ,Z 3±\Eru$53A6-eG](|}an  .(?*sW2!&N=Fgc:E[uN šPqqWϋŽGjF?pj s.UEq?YG0Wԏs&MEȔJhO)6>U{ֻ⋳'l%լ&c(s\ rStezpnNzđo^ ?,>iꂝC.s^ d 2#ٓmT?襧-HOO˒~lKA>#OS{ET-95fl<Mn`[0Oj2U`S[r4ǦBcJ[SԶU2š;پCK*0,H9;%A*ה)h[R1OS&B+R ӄ=&Nk> wQ$g|Ёj ԅptnÅ5@ҽąXSRZaD,4e#7L3FInהp,yYhvB`eB[_wA;OсF4:qL&Ϙe+878(*Iȭ%NmAj $!#a22*o#>,JëϷvQ&{.z P0^zhJi"+J8|SdV:p6e'&;{C˫Kң>Rіm+W)g>5r3ғW xޣ CߑVk) {N 8&u,rnO'͝6N [C5"+hDf+-Z0ϸԒTYQO' d(#NPWr ϓĠN.Z_G/?s2\N"oƓһK..ɛh^*'6J\15Q@N!j!Ĕti)^S͏9|ѷ22Z':,;q[msӲm7.?}ziɅ91  PڅK"eAr^R=dcXkΘ:)/ ]H6=B c/J.:i&]Fן6\pKn:@yV\#/*q u-)%hÍ٘S`T4 qxqE6z;dgTITeE-XBwTG5ӀR0EePM=iJ_H@P0>Z07ߧ"@>gӨRA%\@{wy(&c c`KeA3[RR`P%z1kֹ-UTq!u :fہ?0ęَ!8kV}^]-=Sm]5=a)n[jCjʧ!+#$q(o#N;"XH K[K Pb 8xYZ٘&F3#F3:Vۃf(8""Wv`*͔27QaD_)ԌI "Ќoy!Y\рwKQ;쬻&gEŒx'3 P菟)>[5BX:Rz9X=/Č0BZlB"|߮+>&W ~3Pܽ;;Gl(?U-i ;`P8(gA~m\)*(fm06ۼVTVG_v0jrÆ|2%xĒ8l^2=x؀G 8V[ҕ xaAr=8uwLr]?=)o(*Bֽ\tfG[NCo?7gƅl߸G# f$!WPnT)#eG r73UK-|w%aXah4y:K liٙJuHuT,Ϝ-I ?wЎ‰CoH]fY!w JzWm* B0*VU\Nihd hT?E՝FUp)E4TK^XJQFobE YP(p/rThf^ Y/ez.Nn_'!j΄zH[dX5mTK%̈ArotvޛFA`(s6AHۘ;D}7FbٜJ[NA8hjX)LasIDQo, Tڨʊ`dܯ,ɔ-n"KIE18R,SlRjYdY3D`* 3R-/$-}$F[Y9igNOb9QCD#H=0JПbL4`r@(.9 I{O5 H #A̧tF POw:@7PYwao7a*=IZ>5و;K @M#:"AJFK= 0H˥m9_~~ܐZ4( D7 pe+u[RQKt_^y# R)(?%,ܯwUPC_VSi{Hz T[{*K,䝛hMs+wӌcnN7b/ԩyt@ևsM)Ώ?nnN7w-!8Ի`!D[6X+#Z`sLK2!oQ+#DŌi\dC|y^|m|7|Czt/@ylp ϫ? d-.1)!׊ {պbhjVNsd@Cآ}xVIJitE-3Cb!Li;gAW?35hA5SYIJi7SaJ^UUq9(צV[RJ5##}1X0k C>./&I Dդ>:/ "?X cL16ta6O)+V>{w$SYA\P|"\o^•S8 `~"[ڶͫmHj;sR= -E`AV[(  Yy}} dZ xM;( s̑C .'%=U%gԯ 0oNЛ/.?)<' #uctb"Zo?222}!gxB0 DxHʕ !xM$Z\Q"K9|rFhpO< J`˂4 |k+λ^ޖY9]-azdUGS5L5D=aڒKI{УKBtTI # tc*%M;WZjbؼSu@h%ةyÍk,_mr&hש<.oa>T2S[ܐ TkKQIV;/+wC p5<||Jr`umo)qd*8R4Fv H c! WU C+^Űe\㢌rVbs1z)ȢP&es)G˴PN.B+A8bKEWMR1fJf]Qu5WTdr}/O֯Oyw5T)J mZDy76WlF]ᎲJ`m/{-j " ^֡'m M&oAuuS2r.Gzr͗NzOe}Q)2*J~ۙyM:~<ԦԺRD|.{bOlϭVT>4~+V#_DcbFS;@@kjx{jμRk r*5ስu piw4;G 6Xmstfm>'h'L)y3ɚwÙ E誃,t|G#6m*;C.׻?9D;aJ_qy7]w>w몃t}G6f$y4LBmYL^e\|IZT;RT3UϺkϭvZ,puYe}zgd5g=GΑ܊+G2ߢ<[+k-W nU0*My,@޹&ni/V(W2MwߢL*;fj2ˮwr.LW5ds:YAҮWW_uūZ^.jM`jՇGpP=tyL[vNtja\ q,E:tBN/UPtaT1m*xR5xfŇ+bU"0'F1^AЇ+m*(P.-)7 .Y;y]]P*k~ lF/ 6l8#!R TR̎dN’RcSE~U٠j(U0u[FJV2Paj7X 6K+lj05!=@-2-.SƺmCBihY8J좨XWf/1Xn\|mr @̃j^ߡ~Ά+t'3f xLX@-\d<⩻1Q'ՂR[$"PQFK[&qnZ~: 4d Fld9OOieZ*s:D+ʅ:FD^8=ڿf>0&D'!$s/ZO' O}i>nEL)&r&QL^|֞3t 2UjD :\Z[Dž {PJԩq<Lcyc1GH&D+✣& 22%ږTjСY8L%TR"kaˈ-.AWpǎ:țګ\^y?65.Yss1z6D1;?v҄3 m'smfW.M{v]nN"` Əea5XSJ`+zّKtjU cޖ2%t7?6 SVѯ!dbMjs;+ulPRÍq (J>7_e?E0sXE&9W0,JIn3ò `Uɖ.1D9G!0,-y"QR ư! GXT1,+1a9o`X2q.r.F fUѼw堃2uZe*G;a=iDIʰܻNSj"Ւ'X[R0QD! :P%!+0[X9@5(blA PIq-9 c3@%#|(iKijz!5W :nL2"t&fL2{I?U.phݲ^ʪU̓mcBuCukI<1}t1B', 쳾tcuE\!fDduxzD#>~+;NP1B[v@ܗ#mڪrj}>TvJ(؍hΗlVp?}[{o\Dcdav'#Q[, BD'mۈց^?)[|r,P!!߸F8k7B `|G+mtRHM pj~8t-THN@pT(j^џ^ë5aSKeuc=>GPK ~Ex$TS·8/wwMWк>u*4ޖA}ѝ (5:K^cXkY,fQL5> IyC+'S\!;w[p/r`o|[c_x]qd_z_9w}Я>rοeԡ.@7z637(J|8Mn3q"i]bOe!% w9vh` 1,ߪ/ j4$Ҥdq:Uɑw<{|EZ"u]nVAd!uuT丮 o:J1yv)8#D`\nnʄd g5qSrLn ʴ`9"WD\pS4$hØU)6fL+: 3:h jc Q ƊLqK(+ڡSߤ RQ >LJ2'XB#k3H Q)*41d2,c Pq8)uža7ZShr#{- U xܗ#HD:e\(Nz#b8*}`n#X`afU{+;Ɂ`ENઞruUzšU=%T3Tn64 $UpnĶ#'=(q 'B6mKn y.;s*\b\b F/Vr"z%Yi`߯b0¯riEDNK.JB84L %dW<̑,P9N)$ %W͒YRLF&3t.1 s0䱜"ꐫ͓}(`*vgw6uՂTn^5s'OAx(zיSH)7K3Yls<^_yAL_K1?/wd'GvNqfkoV5+6IȿD@-#o{WY|돍-Fsψ|rOfY3WCɋ_7uRqZDrK4u6UJ+9ZrZc3?[\bbj(AK`.۴6mGdI.t۔9^RLT-RleӾo% [,2xj XdPS.+.:q&LP4B}jV\ \}9k1.ԤZ&d*i4A,"'GޱX%c}l{&2O,^]j.SCXt`f GH ȖCstvKf98Uͅ] Tt9ePt"\Pz&.|̇ĔWT^ΆVβKCh\C45EC_$ b#-SLsNej$cD( rR,9 یTe"4"ͤnV57R+Ϭ_$U֑twXg|U,u/M}jO#0@~}ʔS"T;aK1(^ EcoUp2)~@-/%#sPDF$` !d;IlӐMG Qp"G9In l?a{+=0d8ֺhn/Dyθ~s9zPYˁJS\q:§hEf<$K`Bc"$",9d2*3-"§RI|lƂDB!jVWc~p{.J"~LFBNj>s16K9E~*^A| W z^ ɬBX~XҲ\ C '*i*rI&ؼ/r{=)洳Q=}}A1:vYF3SiY?K6Kf=` Sse[slk)+噳 7ɰ<ɰ\=8$VfUqP;yy g$b:b=!^k(\+bN{?`,vS\H"Zj@2$j $Hp`H+ԪT'\PPDB*2Npj̐'鐒˩OŇu2 UL]C5}檊^m']mj_{hm# 處X  8|= c`p ;t p0rn»tM̊D~yE:]:'5)kYP[jZ܍n,Y\y*}Wy{TLb<Řp1c{ L ܰ 뀳1f,CZ^g9>[ԙ&!ɥRԱ̮*hrYTU2hO)0~>Nق[)F :4*oxjAn8y`~oQD+<~?s6d(L?Vmڎ 徫Z?6i4۵Ѭ`/j rةdݡ׺;U>ڱj%SIUR1 M|jy˖&f"v?f);zd/ ~%ծY#{>yaq҆޴&NoO"&N␊?Kf]ceIBqUUznLܘ-AVmD/BZ-zsvABq,SaNљ] `$伏hIV SDhB6 Anֳb2aw)h؎}4 GJAq`\yw|9RD\ ;^.eRּ W)~nϠW'!(F-: Gaq]76ƊLqKKcc.ʩfc 6Hdi,~Cp> Za5U_g>aYu3ռg%VF G—#ԙ".X/%-%GK(Z@pg5(\1@(D>-\hf}WuF) m!؝:q$3HJ-rS26W2Z0ckd\R&Rple끇A |ʝ]<9iP6SL7 Y D*8#콄g#vj8vPƁt_.7fHj.m++Onl:qOCvZ.jM=P6 w4m~DP8foq:k|y e84|VۻEpi$Db::KbX"A(`I)]bJm  "P-![Z-1ﶎ@Jn\yk/|ٱ|f|()N|[A \{Y~IOt:ÄQԸo&bX86bn޼ĭS]#7rW%ɝ#Y" Il 6<6Ҍ[V4~J0FR?Ub*0 ~j__~2^+IbA 3c#!xVG7k9 flhmN2DFd#7R:ɦlЗV/6<`^UEe$x85G'sS-.N~3FT<mX]P(uFP:ǩ***0,n4hюvHRhnhke;WB# qIxI'K8ʬ<3m=kxsiZUb{4/CA!a_~hA0>xytB ^C4 IgAg7/s2麹fd YT%@@MSNd)>I u&E!ă--1'HkD,@̕Nn (F:HX]FrJj mכeb/*Y >@0JA59ZÅ0S(=apwQY೵Aɑ_=\IS!ϴN脜YTrma gَw<bc&=[&#KqN\@tt"Pm6K8[ XbZ"iEƒiNN^9-eárS]䀜 ,}jd.!S&0g]:MzP7jRApF9j8 08+M 5Ѫz/潷PpFmYX5wl3Żׂ><4 P ܥ+mZ 1WGzrQӻ˄Fd#mij1CqA ˂nMB"RڲQ,zuP˛fc+lͦLʹ1r0U͕ UVC1|,l*Jًu@{|KI7oRLJeU:*_9]KwU|?{x[1s$_EL^꧿\^"9ןh3f fuɨ,wk|_չvO2(t?"H15)P-S/j;:u>1Y=G Y4u ;A#CO^~Nt>,#|OfaFc󄙁Z]";Ђ^Bvԍޥ#zPm0Uuhh hRmcr+GD[0XV}5!`)}X,ޞ/>GU:=ۻ?Wʰ_ksoFoVAT_}S37mXV_K;dj6uUU"/RJcWEhݫJ{"]M8@hZ桮ϭ{n}y:}ƺ?dlֿZOuBCpSx3oޣu˃6mɬ[ʀm US6Nfg_/qL_/JrWiLYv\WG&U-ȧۖoXڋϗWds?Ί ;LJcwX]:{Տ =KR!pR:h7ؕFΥ>F7!դDb vSy{-^U󄭽 Pj.SmJ+6Yn#Qlsּp זzn`kO;%DTQ<5Xt2jh_cˁ qPʎc $Wmd3m5􄐱VL=r. FZy %2>dΆl\B:`(ۤs;[KX L2:}A@-I7z;:,)tQ; RM.j;]ᏧеZ2Poi}d`׀koN5E@󎈝ٺSmL=fPg&3c#iV-{NN52QK bVTsڗ9hvP>i`~O&a UNunC/{m`P'P= :LiSU{W!_)<|u [_NwnE:;u_g>Ӻ !_v)m- 5:k~kXCeQOi+dw.zyOv旋wA+bFrCbaej/z_wݵqW 7q9"~whbo6po=Si0QKSa%s]맪+V}O?p*!<I"!|ޢPP}2 EgJK(Jcty0kb> 3EZS}gb5kXtȾהyŔ2= aS H"'T>9@ym\1fASVwU<( $>ݧO+~)/0^/ius[$IV-M7^!-1MW(&mjdBF7냴ܶ}:ߦ65{%AܚװKV+V66E>7%y N Y",b&\aUB )f;7ki 9vl` HMO2FFR [@dI6q<.?29bcqbf8Th&i(*S͊t^(^VXXWH٩ҙΰ"uG!D&)*ō͐Z `F:KA^0+B!t$6 4o8U2Φa&q.ZM6f Vr步,G Nٝ9׹\yq˕uqn抌Ar6IjrI'K.d3&8asiD2 աOOhTnKl'Gkpꐗ2Bp<+fyhAVVbF'X[rPj9W3 ^LpbVfP'p7xr䩉l&-,DDv&meS#Gitro/C[-:ə7iU=-qIo"w/-`^+AxHhI"U[1 N*ό L,!fj 0@_vyu\d WgP}dF5JloW96J)eq) }.>C[Dp@gK2uOyyuFeI_+i~p1 ֟x&A=rvEN߽,B镯!o9 VG:m7Kf+~סm+pV t'_ {ZVNBԅUPj{Ia,uࠠN{FT;V^|/T'f QlTaێҡ2W' 4mgJjm <8KTCf\a-ȄVT 9ͭc[y2iZh`-}&ՠi8n-ei;{WkXVT#'[zZjIL9ZKK8-}gT[8RTʖҗxFQpңRyZ JҊj]IKZKR)=9 [MIKXKCW9*.$Z]->xWhR8HKh*y0\#rA5=t} \G>I_=۶-["vkb9Dք[1YYS X(sewyWDpv _yio¸гx-'󁛵Đuoյj04dMFcm'$NAe!hQVKYxEu4f2iC/ye8{.ڎz{t}'84(J ;Y5ά)P-p3Y.e&++T_< 0uZJۍ= HK|A z̦:ʥJ F#}bu+%)e5ݠ2oB1@ͭj'ۑ-cO-EҊjTf2Åd T<-u*=qZIr4]GVJ72M5j&T<-ZR!򴴢ZH%NZzZ *OKASB+T_4FNR4yZfr/iݧT@sgqFWwLk~,fqn ${FRԒ}z4 [ƖحEX"eީWI*VrV "/Jzl (n+#vȋ=LBf<VS`up,Cs+#aB}5^Ȣ}6f4LhKbmPU#c KedV!`Q߹+6.tAYI爂8_TepW m &=biԋo|tO fMKY-WIQR,\2xP/$_ H=DRDv#P6 e3k# @*Q!kj"I@ZU2`Q,fFALֈhY XI"EMF u>jK-XUA!F!V3)-a$N,@@NŠ1H(}(D $>S<$[X t( -"OTH IՀN+0@JhʙQB8^,$\_E9 ,B-bd9"f8 @MA;M49f(XHc!@`Q@V]2b:)d|ܒEg͂YBR5ǠvYj .%{W` !FWl5f ƴ,Ɩ @L q, Yl0+ml901+KzYcn2XZ̯a!cSur=v7Lo B/iCNƟ-9ǽh=;|]e6s3u:G;)J_Ô UuݓJCyWJh4{H9@SPc(o Q+EpwE֬La,mPmnCߕv1ۖt˼|ŻK|y֩ZW\6㭴\o:h#`[ق.{hGBf,%:=:U?vnv+0=0Wu*BGaQvLWxЃ)1%a_ȭ)r|@ 1 2,_bXpVwB4PqI)U1e\ᴲo`NS%e3Y_qjvM.|nqYJg& 0c[+`Kp Jp<\g+O:%&)D.wSuI1[ wY+ 5!u9KAxQμcsEd&8Tlñ?vɏӬlvp7y5ޠ`v8xnۓI={Que mrtr0:)K[VZtwd\&Y!)AՏ1vR e] ʗh5[U,cnS=MlMUn nu|VZSy[Q~[U,cnv%/-^Y7`7d&#/eg;wos8PVs0]U:CzXV:1,"Ɣ`]SY{N#1)0o=s' ۱gdACs0r (ǐ#PR}(9 D`kuYt=۶{C/k6i | f\X(H,@cy~˳Hn:J8>WUT<5 "c9Hd$d)|-[ a CiݎvBS@LNwܝ,B*p"Ivy0$YFFW"Jэ}3 \HU> \<^!Y#=*@͹d5&,.'h)@ĵu]\A?7aN\V]~ B) -m-%?ߨљ4a4NcPQ-P sJqfɍ*!Cz.$lP9B@#7CE7x7L'?Qa$apKd=q( tD\&7UvJ.*CՈR(oa](gb[˘- @ 4PЮd@o8I+.z@oD:` Zk2:\8cMIc))ʲViLb(L00= 0W,+o0 8iN1Zt0]RPޚ~!!W235ճh0 ۚg%ge\"D (iDַЖA3t7Irf! ',5=R 4p?{9xA.~QPضg=5w4?H̤B<HZ(#}!A 0_ ]sFHρD%##paE gF߲c8S!E | R s9H/^y|sRQM }׿/9EUs%F G  TBzr" ]v'Ѥq7pո"k5 7' KDDIe(B1`-0I1V}7q\BH94Öl]cAb.X?*ex~xgY7#qfO'ىP&Q=;f`䅤74SG)>K)Dٙ'TfBaH0\5Ɖ@ gV?pܛ95p$z 8odz_7Oѿ~s =uG=f4r tϩ>jlNNfUoyF߳.6=g޼]*c׍ _tl" H3%e5cP4Jn@iM P>ppZH9] OH-|yk^e}*}XW7w.gCα#"(}.151PD1c5'DH!1L 1f(!֛oIo0ϫek|EnAƙZo)Ǎl-.dVh^{O'_fy3 d&O3wrCu/;__ϦZCgqoP,:/{ӗ@N/[jyO˟~3M7xPw{e5Ο^9s| d罛d^{Ow^|_ y1NS{5\;vzgm=#uMYs'0_SbW0&=~p0 Fev;,ͬ-٪"=ے\Wusr s?m7魾zK7v|d#9;Qx6Y~-6gA{SQ}JO_{xz1>^{x.mta`Nj+oj:HrX ~^ żS:9]Ï.ShsDCe]}7NG3P:Cy ?E8df%lrϟ}Koms(~(ht1W߾8rb7=91 L;f@:-5b.%A<+!_\/iyW|(3kMdv΃'u ĝb`3'HBtԃr5Am8 .o2U|~|{d#[Ζos1p`r3cc {`W~1K.-q$Bt`EL:"1FK97p'^è0sԗPcZ0> ħE!s%?H`aB .8T!`T$%)jX*PFM-ԊbSãB I=PPyɪ[3x;nb_[vI&G+>a*WR")Q~oVS"ueŜ='lk% jVΞRŐs.Hs['f0jSLJSv.=51f\Nr՞UmQDfzIc( C 汳fΒX[؍rSlY gNqVPy40x5NjxM/):^Ds o&#%iP\xeGғ6w,盛}O)>Iq_}:Cn<1UBPA5dN"pKqPc#1S?{ȍ_{{!ŗqsa 2l>f^I"-nIV?$ .*A ,Q"T0 rWėUmrqXa\<=T_RdF$O^鹔 $*kcKE Z:?dq)#-,JmSU^$~[bޗVX[_V6#jzU+ K'Dnļ1N'l?)@s.t&9W_ri"'ֳkBVSA*nr*݇EaԎo$XjXus# ˲PBR A@RY/ٺ٧(=7"O.BV"vD@Mymt$M65K.Sh9[.5%9 2Hύ;;OfM j)~teb/V.7F ޔjv[2n[@]cz5a8Mt2. %p4ТDWe)XJy57X;Yq+K€Lmctm<Jآ8R3'~4I˴ %'ޜTJ.Pt}SApykJh]mN O4c!ed-}HϧWpk6a9jiMV`1^3*@@ DX+I)i13"O v\-߁ea5Qb!KuDIz}"R+:UP-ՂybPECP̗Ze]O~GDxx|Gݼ17U4/}9s6kv煊!tL3uc,gD>h4}54tp(jwa5<Pj3MT9kqdvrUEOjoaui@3a2k[.ƀ^T'wMџogVSYi ˀ?~3z, ^;*BݭV&tv0)T.*-5e~y1I<>h#jjm֓OX9 h%fM18|M.ݸf btھt۟y*V)_fJ.V+_֔oqI7AAnm1Y:mD!Q{0>%SuB^]YS4sMUs=T+ggk~Ϣj0oD1ѳeׇx@)fgTnjި}d5KKv3,es{5EkXsU߂tŁt}ْοjF$Kuxxy+rUonjDF=mE?ʊ6>fClmLšRO XgRrr^2PJCQ2zIA#/aU䌽z!ѧqXXqg] &b͆w"X&ؘh6i wP45C^Ɂ=Nr(+;'v?/ͩy q {F@5 %J MhD*ِUw'"JC]'r]GK@@hd"^c2)&FפIۥIg˝IE(/6Ċzm=w @#3q܌B'N\B8k!y-؁x^@E}_J޳pNֳashǫSqR?0R4Ƹ$SnXN{dAJ‘;6|5@tS~N^ςl[}kk}LЭ[ym\R'0fDR!Fe2b.s˺SW!TA_:a Iɂ,=wiP@*.U!m_J}yPD hbjo@ipɞu[Ћt`gQǏD:U[jF(~ ks~m(L^)])]"YPS,*PߊTIxߤU٪ӶFHwDݥZH==MN"iKɭa)|Y5ўt5jQuHt(w#?`r~,~DL%Mߣ3V^r"p1`iF G 1l MBI4o)2غUWYU7V;qIi`J@EiT @σ4 R:jlT\WP/ر=~֝'|Κu^dO>! x_$>'n;ḧ́:kEh/8*aTkn0br;JH;n~Nr7^˦x?ev׊pdz;ɩ@* iBIQuqO`x,(GB/suvW&2Eq9 <f0u 0&Gő0D=F._:ˆw'{8lLsT B ^(D수ce?]:(}D Ѷ`es7,N#g{#` 9;Լ&ᤷ4 :JD'^qa%Iu $ %-_՝)L6+1U#"u)"~(pH߯̚?Ό4Z[Uw^Sh$Ao<!xڔ80HI q^)=j'KZp –yUn JdW#Vf'P3 CG $Afì:i`o 2%3@KJ,BR AYA*ǃT+a,X5(ԅir,ȶ5j oC~LЭ5j z@%vk].,Co+J8jIjF1 q qVx3D䨓 l(=iBqCA-3N+LQSʥPZ_cH]>)X S=]9zѦ(F׷go\tzw?nH_a2"i`K y} lҞN8vߏliF׈HA2z_=IVˋfXOs]l#U~ ntlR[uIN{[i@Y,k(ኅ?~.YEA2?.jM7}qL_tu7° X̙ 9o.Osp}}u"~K!osm&̽s!6>D['c_AI>TҕF/fw}9L_ :U2$Es_(2/ L)C\kR?J(Ez_"01e$2>D->J A!(*PJCsl;P+-Ph\S\Em3],~ZCgAM{*h^:;7*9 z_:9~KiFkEl=K&FO'Y`W&@/ pY8my0Zn, nf˂z* E'$Ÿ+_N^λ7ui#+5[0cnlrʴ:\u_5fImbAzZ=|{IQSqZLV;>iܱZ~i@6`8k$)8qxaDE 5,&(.obX]VLbNzRۤKNt ng_ՇsKWy e@΁9 29z>/6⍪ ڠw4UyB9౯@onqO.=[3ض:&1]kwBw uPs1] ՈdEmfnՄZQ۸Q،7Ղr*p~&Y ^\ݙ a[dp[t_E1(@׶DjcMgy#,#\[ F-Ha+e\gaBg8(r2r0<INkY&o0N9)tLN׉Eavr;Y22HSqy|Qw*d162uwSu T U7ɼ\^<P׿ {5;2Z`(fK8BU Nb'בs\5{"UR\hebL )i`Xo8g9!hWqAP\u+!PkIj~!O.O-0xI3mOԹ$m `3a,06$$uVDf4hYfZ+viWT βtnmF*8_Bi6(, X*`QE9jnt>Ѷ70y!߄_U@@7O#E^y8[ԪY9Og8nx,l3)2|NkzᐔHn&"Y)P()K$QwBG J ZQ'הrFVܤr(U@#ViFy357T lc.j-}K N)(]կ_? p8Ÿ b'߽g]9bI&D[dnQNGOam"ZtW.[hu s"Uҟ+:倢H 4+,) KEP{JW7od9fcH7bE-ETY'NťaC>TSdoS)N6AN6+|*O@MŎ0*tooVOfbG9=A+򦩳Y&%nu[VZrEb'o~ݜ_ӃGTl{h #fx=tl#c8P-ـ}{U\?'[pTk9"li]Ifoesi>5 ŮUv?ΥG%ԻͲ?{iN\.xG*ZOȣ!W(21$\Z}#+허[lA>a ݵ],:wj-.tR'}Ji[mJJ8C "89JJ NLEf9is?QݵOԔ m6 S-k.W$(V\\i8T( )@C=pԀ`,X"@Wa?RRBIJ5E 뻏N;^5Ԟo&1H{8GqH,F8˪֯J__ZcVn;wA1f M//( X3bfIƷVJ^ՀN~xڀyԂʇYJ6$XJ:T))hgݒ~\.Ǧ73b`1&G*|M'ݚxw}Z!).Yw] *]5T|H%MCK6}~ge:~Ift*c*n˔ֵFD t6eylHw9]e8Z8փb,aSP@63hF6;<:^=D)]ؽ5h 38 ؖf41¥`_<ܗս|JZGw0JF\]$#;N,7N[Zk?ѢsZ)|dr&zҽ`SC[񀳗AWKX s 3 L5sO_1`}\07f|eJ[Tu lCP]RՊSS*UQrK *2jk)tA݅a-3gMӂ&.r 6RY2Ɵa(EWj%<7j4M׭)`=WҐ+RZ#3~R F"g"5]kw|QQQQ]?QyaV Kz`*O5Z4`S!5`;~U϶G)!Jnؙmy3 A){F,3ʠ!߄2`8DLTɴߞuMySo`ە}?ǐ|k=jg-%4DJUjB9(\QkX(Q4BY/g k᚞WՊ O!,W{Xv5玠xtd\cttGyϞ>[f\N)#'+L^0*ܳ9d@枯9Z|eie !@+Z SArQ IdSQA 8[{Yڣn -OT꽑gʚܸ_a- q+BIk,*pJZmτ$xXbQ–(Ve}D&2J2)ir)Quvu037jS[ܶ@VBrkY @lll=7A ^е\d| |]>c.Xɰ8l88kXzўY f;w/*jj^bA}7}tq g* gtl/-GOУ#7MJ&v&J_p?TD!B4An {9wkMh~ˌ M,M=Oء bw6ft{i'J[,vM/ !SsMLʐd+Cfm09YM%DZS:aYq8P!P耥 `X#F4IdT HCǩ4kwM_BPw9ؔi1rZ#$+̾~vYw^B*YdNv2\/FK5AedNz5i!K@MH2+Ru IQ Dy:Đ dzp(,P p#4ue[֌*EVh^nZ8o>,=m.1~_FWb"'ZI?4zE-VMS73")!dL+ɼ `ASכ^^^S~!x$9 {L+)0AJQjp &ځ+kG"i>MH-L7"p0b%?+qz>%⭙j w7Brֈx?} “,HkxgOE@_2`.o^>D_??L6Jя5VIL19%JPbQIX;srz0Bub8DFy qxuy߆S8NFM폯dhGrGuʍ\DSXl"3+*+t$qgl{͍j(es ȯQ 9*\T?R?:H.uՂ@v kF)?NԀr,up@|Hm ƃ34DrDldBGF\eC$..Y 47Z%%% {֩^H9B/{t*LON.!Jh!u6)]XBz3 jLFwhځ~4ߟܦ5h{soŨCRDPrf0xPȣ֛]˛?n4߫3в?(tRe5|^̡vUw=vIf.w3 ؏/է8ԮRb!}yK]Ϯ{{vpncxR\.WmoWmʺxSզR~&mUZB巑 ;oEm>6…Ub7E9@*Hltڋ;gs۸_. aQ/OFU& #]A RήbR&ig.&f[Ic"xT|TVuDhA88 ۷ e3g~>eq#u ֆB 4w, N\v#m™, QŎCˀ6C6^.TI7qpP!]QhoP&V|>T$.Ƹt* H\?Y{?E?y]mF?i{>x&Yyd_R% c;Qcz팧aF+Aςyq.oč{bbK.t+a"格`1- +,wd1mzD URP "(&"Є\juq^kfhv([Ш"^ױGhę3~fbӇ[=LL.CuFΨ;pNpμVJN(UUevXg&?ix!- qsO< ;5 Z$`70>Uܭ';B:R# -r<28PEEp,qAsTQ;@'+5E+FW8'y}nR nkIm^zN$FM!'hI4 H͂[H-)y[ ,Ը W%BB`[.֑ N3Zr ,WzE$֣tsR_O3C$HFw~`8 ΪsBK%U>/}x8>]]%{א@Inp.5T\6(iԐ\5#א Ihu^T*vEpPzgVHtV 1o54vF]^z.+ o[G`_I.5=>? rpmҦIгggK.6}9 t7"/7%Մ&` :1ۛVvx7i0$m_jCq+M kt@PQ^IJ0i z5A2,*(`x$%velm6/~< E8kJiߝ)vmtIKr6sō-42] ({͔GЏ @ݑW2`k lDvGEt,B96fv5,ΧT'92RdV_ p'uEg9&G~rm֤cIRc"DQV,\)oh pʎ푥!zH&C͝'S])kȹv:!B+YlWz RdzRo6\/)ÙS]/Ylx1Q7?6P%Atr4+eޞ|\H(VꭾwNNfZ-NMcJSd6]0v%.) w.B} S EAddC۽,r,tI]OkWt{f}SX.yEHMOvīgAdn[!K間cBZ3= `SNU`\|dcF"wn|Өpv6;O. cͮ P`>g#l~ܢyVݞ9I@k< @4xҲh 0g5>]\}1g]L ww)>e6mp=$m#@>GҎdʭuN*E=Zh9s?ڝz+Y0%ۄJw({>zXKl#NկCTo&/ zbyIOiG=\ hX/b_㾍!,7Y_Y0Pه8Wl3*iixws#m_~FUf NI)Sh,gAh BQ#m27d4O[،dBPٓq*!ku~ GO) Vmv>5SeNT-s:bDž1Sk杂|{r7pjWV7_@'*'E:,mQo Vƚ0A1#@* KdC*Q*e˧R+Ji{o%9'LcNQfP kf1Xį3wmmYzEX-`ILv^&ITHjf=դfMʖؖԬ>Ks)%8 ֊jҁgiJܠ3rW$5Yk!%XRCX MQЎ*!8QRH)%V8B5DR/9UjRM\on*uوYu7~1 }1㤬]qh.G.K=v[_%UN+[6;h:qYDF; ]4̫o75_1#gS0 ϔ]h~VT|_o_OfvlEeqݐOaQc|)~HW.92E9>yeRr -}GvuV2nMn}HW.e 6dyv?^}{9ǥR%\[NqLMBzy>V m0QFMxFzp'*w',YP`& zUϗ;2a)#ޝ!JZaOwVrEÅQA- 'ɧ4M&NgAqn% 3.9HA`yQag($ YdՖ~䭞* 6S<`eW2@ PocGl5YPq"TA;|Ly<5=/1^E@t nQM4f0:`u }ymNah^ :e᷿9Q!S ƹBi+ % "-)w/ܔ4k/A߇cvk~^Ĩѿ(fry̩%#D:yI~you" Ɂ}y{äw;Ձ)3SzJ\7nZ$2P6*qj Q{N">'-_/f*CQCU wzzeW`S.?~Pd GP|~3Z@bx,T?.awgQ Zԝ?"5/ChXڧGwcI`Iq657+;<P,)nǘb3l4]a='{<%`8ᤐHj|BwנZK~$KY 3 3°Fߗg 4i\_]"^EҘ57.n>=9ΰfaԓUő)D{r쾙b cl*A{lw3L`3Ռ>FI) \gd}^#5*<{e^NjduLsX-b&"*֒j#'R9&Zw\kq'3C7\j~Z?R?%=/zZqRJkݲ$[>=8wS{(n 6 8 L4ュ٠єu;md 88% s`&~ wVԮ{Gw+x7* 9jm[QN5M/t$db<# -Idѩ]z\DdTW-&> 떋A侣u;wrRz%e^=JؗtM,igc-$5<{F@(Fnmy"j|a6+cSJJNCK 1XA%qci`g(7HQjVxKmp1z!W!U8@$4vy 7MC} aEOK`-/],tIuw9Kz^goE_3;kV^LZ]JV%mLqDZ:XS)-$G+] 4 ՠU0Z h 5B``4H |00Cb4D/'z .KRgV{){P{ qF49bPM )qVA] !3*ҟ 31 #g=5vܗ3sngw6bB4Fcs#2W[JMl.ک&3'zn F6KT-,Y'g1c2$vʣYv8jʞݎ‡dLB(Vq*o[/P?>s9] ?T/<,CSs؋&g¤3c7[dcp3PW(@zn ?.+?3cyPtG3/&ХTQ`x.g\E00Xߋ>N{TBlvhD!D iGx$"yӉ7L߸G6ք?y߁7G[zug~hT zoGza%po'|$Q>|^Gj7L@l [ |3&X\* 4VjegDK+G_1Jl *!HȊUj&4Ӊ&> 3E%bybRΞsQ<{|h"tR˜`A-9r sad 60" ΕD )1f삐x@K4&mf0ΦD8bX+8di2}* I``Z]c"1Ew^,^%Ugé]/dG9FH> ,mAa<>FkE8{.@%>~K2o֌.cr~ziT r_7ކ׭gE'XjY74SF[0}6ré 0g{cJ)9TV4cD]v;{%$ZBƚ^]wf61v#?SuդZ^IesrC<% pHy]Ԣw`t>tukWVXU-S*֤U_R_\m-5IE^>=8wK-r+KʮEĄ9n UU=y8p(,t]۫&XOpp~LyܻcEI7!PGIԺĿ-{XZ~՘u7OqB=;5; @HuvbSw]?{Oƍ_aMx8]5GRd狳. LdICRN 4)?{b.{b-`¾ |sX;+ Z"˒1]z3JŃ,3mb ySqw0񕹎e0 -[J46>w׬ً7' 4+8-0QW C 0Rq0 '9 IfіR* J̅HJ#<ʇ%{1=  KlD’\~d+IVeKYtD/i@(Pa-ޔV!(Uzo (UJ.jC'~fjUٔ#H$" VжϖL19MyL,s7}|%ow,GHbpmrCZ+3cZr 9٬leSmu]l[<30lIݽnӃq3}Oj` O4j1h5qϖ.+%hA##lxQ/v#Rz}hCX +QSѤg8ƮMD 7ݧRtNG(+Z vs 0 "%΄i>>۪;Ɂ") 19_< ӎЗ!~é!BtR@CAe,N%L~{چ &t#.KB?%=Zt=%`eA q>ƚ!"d8onQ;!<#秼aSSNYȗBi+/mA=W+SrhOyqbc!9(2dwA}ՄZs)fȒDgFh͏CJwA}-$T$Js{ǎ@rZ%B 8qᰑ@zD؇(쁴(31I:c\zE< T,"]:BGW,81:>j)ԇP_5l8KIK)4)Py|ABJ1Ii5Pr-it|.J)(tWMC~-I)%QJG!jB |rRh : )e4MJ+)9{RXO&O 뇛7}$Z)jy2=(iYܹqks2XCFxkz |ˊ ͤ6(=8Bxb@Cjs`tf A)XeH[7hF$Op0:\Sbg2=[1KAE"X. KG`l%5ki)ui"m(HZyԍGڷ7x#("㐰K_Qڎ"~R.!O]%n'YԫB7 BH ac#JYidJpa<,#f/9qCGpmdʏwDx[?g_]d:/Yu- "=`5(p ,.Fz6flXE A%Sa on«"^^×щ3U| Ɨ i@A%.IWOnK$O<);ZջD*UCҖ* =@3HUmޫ1w74.N]Ko+2١LU6տ#c :XA$/j]ٵXq-CA6Y"l!5%7x[nV23+C4fa0 OMe,leu0ܷc7g=Z{3Q wVQhO9I{nXoO8$Px9g={YS(͞sG8\JB$;<1ĵ9v)2Gyǝq2Ŋ ĴQƤ9# f{];1:cwI{.>!!CIjm9BJJL TlS^ !:2RQ11 9IjEϔpRZ@v6ޕXP+PÜT>Ȏ&`T:'8\]lEPb.1R t mU'Dj ' p&8W2uTNC%'Sr%=P#^;CC]7=( Ab*$!860P0sKE @k9@4!2)Uk1a lU^Aq3LLhZ!1vA}Մ:skǠ;ZmĖGFɊ [~1~}>Yk ]P饁C%Uy JM2=tߡP#/QV&4iLY7䞲YϓI6uգ!q̤s`ZR S~2O?WmSܟNW١*4_DbR"(PB@iMs,kS&HZ%`8?>S$$rn|zbY!樖Io|ZzWU#WEǜW]\O`TچB֓ŎCx+/ &}idAecD1)ŃWj[l Ճ,{#% ..ZKڑP"t$6% IY5ٜρZ q5-ˮ7soCI0GmpG"tZǪ~){&2Ê *Ր\5& W1wxS]^,/Gf˛NcףοZnfg{{].:Lnj!ʥۋW[zxZ w^&[7~hCi/>-U}2uiljR1r="*ȏ7ݼYOhL)v~zݺ&hri":kn"Op-?E2Ѻ!!_Ȕ@puLCZL$N6=L gݲ1R[hg*XZ蹵aکֆ)oڐDULR8wA}ՄW$NRfE "s>Ӊ.MsfZZLԺеlBf%P,qʢcAKAX ,QhX򔝕Yb9Y_z TyՁ]]R>lե'Jjf S!T2e#=YJDBϥU*쩬w\5oTweP,(ѝ#NESώ9'Zv5D `¥F%C ''M{ l9)daĉ8BR:C!y&b^R:3/ui}c9b_d?|AVʒ֗J#JpHpkwUWӿ \!d;_|T(quڐU"D&G'$+M:&\r] L2p^P@|0ͪˮ?o$qm'gy΋|}S=)˝4?Heoz`hK*G3.v)˲T)PdngAmb˕d(J(&_Ur6d9qMV$OgcgNi d)Z*,C!i2|k(m<_5c*.Sh'BN6 8<K>]VsoJ9mo>ԮgxU֌J{f4ށNWW Htj|` w{W'}H*=-}Ӣ,#mk4wYYNyüa^0/|w$}sPZ`8mRV惎1 =ԧOz/}s%J-WYSv>Ұp3'֗EQbF2M(x BhUM!P#e@Ze/lkɖ`=OԮxiA|_o& yeK S76PO ߀d+ʆp΍febZܐ0+юH~EݶCiuRyC(Xde8nusC.pV=\\ '(?Y1ʨZa9votgrs<=ڢ5b:m=Q-/ RXO[OTnzU7ZF2tJ)gK+vRjx!Yp\AЙX`rQY`2sH4g܁bR[R Nb";''(p1'Qrt25mZ:89н..N6wakXK(* se[,TvUt;dd0ϴݦE㗭Yݥ4p~A]GC4$R ]S;'^(1 'ۃX)07Ɗq {@ #Qii`\a{ |}ˤQg`Dxl0_)~V/N_S>gVM"6BDhLpI=Eˬy|c>aY:X;WO'ݘGi#=kLB"a{.ۯ YaV)XɊy j]9|Ot촃 n|Bw$XZKΦ+WqI٦Q`,uW\2 "9`+8HP/B PDS)TH\Y)"j|jb 0D_R $1=}Kvд.[rNޏz0cԥ\Kb#ۻ`Kٽn*1 =֧2z~]B{v&K!|چXƺ[ |O{sF3>--8 #ul<&XYO5˫@vrw2ΙJ y|ߵ qwb`&gθn&8Ƣa;x7qA <*mn1c>mQJv9 SJ S2V2>nqIiײqJh,?ws(]Tξvgi^~1?7$:7_^]ѶKg?u}vvk|òR.+sq1Yy嗾oە'k)> A oe+kv"p S!F?]i]D(ٱa * \L>LX;# e'@cAA=$vK=2R4"I#㙩$MC0e)rYh9@qʿ |bu w^_p ;nXOJFN5PQ/Cp^1,=GCGeM5l08˹r61uD=K St+Y#H}N lnlּT5ɢ9b@8r:o-GEx0+K(2vlْx(ՐMJr@U *rͻ׊uo>h9n+(Rɭ4 n ! U QV0<8M!7BoSCal 9DfζY؋YFqe)N$kq,Dﭲ!ڝ1n,{]uBMq4 Ii "EɠXnX>@?>xy7WcbH6f[e| 37+8"60Xf'0fM *ՅGӄ\N&JG*sm׉G#MiAo[ow)<ݦN`, brD&}NX_;9I7!ƄpJׅCӓya2_|O~`WUFd~rVseN7| ^ 'ڌyvVJ7Tt~+8 HZ>Z  31؀-#OLNʞ: Ց'& 薏|dv$O)L~łiuĂJ# ~JS <'X@Wjb{ @zNپ94mj`TxS 0 FZIYMyX[RI;b F}7p;N?(=7ǛFRemnCu.i9RHt! L1$ e9:쌉:Md/C.SsS.ZI+#6,QhF*#Lz3lTrSv3UPHcF=S&H>pS3ԕdC-?e%no=ȯC]1mO[rD y.tƳ2Jt,"d4>;{)䔠%3Z؁jۏXOOI,m>][8]^-"m-7w~_-3(|M:gV[M/]r㯤w)GW-祅n 7 >0h4"Lay1V, <6Pu>%ɗױlf޻w3A1$ ?ԎF};KZ_ރo-܃kMDw4dJ9iЃ4:oX=i-293Hrr8![GK1/qļnѵEkÒG Ą:JOC$Hf9O`&+/|J-;u3%ǵؼe-6/kyw-֕m"B QZfH,Πq[AGdɨ;u|||kӦ6Nz iwIoq=ڇv}o_?zs^uW՛{So蓯ܽOLFQ*k\6 !L+GM: :$DjtCoꁘ!g ) fFF-1 1 "DHERL8&)I}O!|ΖG5K$`ň*-RYĬl%cd/V]` r8p{|,RȆ!(JNXr)$x)HjB[#g 1Q9 Tr&3(xBB# 8HGu1B !wbO |ݤp3r:1%(L, Fӳai[_`K \_W\_h#9y" ԧo}7I"pFQYl|Sҏ8R!4r#-ӃKLoU x)c?{Wȑ Ey Owxmޗ yԖHTHJ<*HZ,fF}GFF$(8" *`$)Я44Zb`!,+$iE L1*OVRҰ" j}kd ?`8kLbjJ 9+oG2gL,c# ]}>1'X}&I*Ll4|H& Fc0ߘu6 `ܫum÷q͠Y⪁t}aFc`?zEr{8HY[\P\3FfYZdEi0hvrFUpHp1l>9FgkGAz%C{-qLI/4[ $'g Pc1*ۛI\9u\X 悢 `S#ĸ1vi jo~0}ɘ0`r `C2_dRE1n3>)%zWçSp@.5J2#k4&r6ưRqB6J礹`ၯG^ 2V^ X}yYnPCZ<t)0l %Bp&Y<!DjID.6{xdU Υ1뜣yG"A-RN[#kPnj॓9<c; Eg;O3&F\pOixy"XDdN!ތ2soȔS-F)Lz8]əd =O/Bpyfu_ON0#^.RL*Ib$S^{:x3m/G7˝KGg$ce[b`лΜa#SeSb*Lә]m$.mmwOp~%ݬ`x&Fs^ds^%YOb1CVxK6iGgfQQr^,7J>YVλcSSg,~zYeM~ WC8FFъB]#KѥnBTE_&˵nkޏUһFUŸېRO d W_:W6.y]υ/ٯ"7Բ.ΥrC>g*=8^~u2NH C)X}w!Ai)4g-)JcG/L$4iViMZXwZCG*i*o{ld2ǿo\'U\bb0ɚer庚ߕqQ%mf(+qa(M`Uwmݪ_jpi^Vbh2/)+rdf -T-2/R yeqd'#^5h-~5nH TӣO()!T~J+~R~] J>,SBsXtkz*CQj&,P3J.ِ$^hI0&.h`44,"zls)!LmL _@ITV) ݃H;^d+{Q;y%y" Y i F8ڲ@AtFw5|g1Zm)+ oN\`#NOD5|!?#pN[G QjnYB[ E;~;<\{U?loh#^18]otl"cӇ3kPۛU^'w稠^mS^6pq</_&1"AZlbc3mi0OIud̛\{sv=L|>dԶ%s 驎]Ѱ7k7Ӽd1<{ ATcH ҍ8i1-'~nL/^rf&D m!q`.pI3#T&'y{f?v6Wvn*Wl~|c:)O &ЕX~W.KW|-A&c\lU?)~Pd*ss2/f>Ŭju97^Nϒ1ƏT89ʝͧ>o^59Ͻ(*aoE!:K_5S [!G.J;ABC/o$U/f񼏞]#dѲ?LM]W 5 ћ~9L> o<nJ"Kb OA<ieJإl(6L>ɤl2]f'ZN1Ig2 p `Οs}0;%uE#W0FŠV =Z/Ρ4+3-\ U7S%jӵ_&^!M\P`Zʴ9wg8T \2 * $wX{ A,5$ c,{U_x5.G6&Hf_2Op) FbB#+p䓖Q]{YsXM`\b' (+X0:?_qTC[{y2%_ueM.B|M~뭬L{?]?_~xֺ뫻/7(f^ۃKl/.O$L9uet;=29Fgw٨ @bX ę% ,[5Q"OLJه둳S[[pvyF/m7XWf'^0d܌N+-\HUI~RBu" EYʥ\69^blOɈyRBݬ=pц9KNԨ;˭=z*O7J Zp(J\Rh?G@o'8"K|wXp]k < 3Z#Vژdze$R(jC bu<0`ГASDAٻ:.!-ɱz XvL%1,^XQt@@ !x c2d֡IqYڊoqYڀT@o'3?0*ȧ߄lEtk蝠gt{7iDVu:UZ;I7M(n :V~@0`-p-wtS*U=,RVe܍Xs)$w}J\03Bn͹98FoC?ޡu5~y6r68ZU}sX/F=nt*|U VKd ֈ8h"՜qUmK Ek,m iƋWþۻp7 `SU>8a!Xb `:2R&S5¨M-։K-yS܊I4ž&q3Vّi*q}[{8նes|`>PXIiRLZK-X* U,@E0qEH7D$4`Iܲ~9%#my}JB@xσnȺ"Ƌ/ʺ|'Q.52$ăh꓊QQ1-!$JŜ+*i:Q=M'_ ; P;(J{^e!OTU4jJ-|/jnS/ZZ'T m-}0`T}UsTw~.h_W(ʗJ|W4UYe001C[v[J t?>X/lYygj&{YU;SJF\CW8qȘTiuwls:|RlԠ[NdVy>tׯ؀ & ƃyH` :A(,KJ2HIj $1 wk+ae=J[|V{٧Zn `vDKSQӈYb $bU^zx(\HH$;)VjiGH$E$r!gw2B>B#$V Rw5hDTr@i< q(E%X!pJ}Y$Mt"nQ276TJM} >^oGp*,)DϏ+%k y9Eq͘䶆.)h::Pgt#fϝEu2†NP`(PR`Kh$!H4siu:vsRyIqRNoid -Sx*?ƨg$K,:AT|d!0OUM1~}<(ED4;fl'YcJ9qk1J7dT)g-<_;Teb.]lse4K FQ! @iEJ~@Iiן;{Gy22eE)Y{*=~aQWlbϏ*DVLٺ~3)khغ ׀_j}(@C`_YƷMDyӯ5dKn$Q}%R;,N}cERE-нaET6Do]݄])Yeq/9-t\A%lD$d>ЉexrO!|D7YR~F*tbԂNm{Ņm) 2dwd5:g)pW4HIZ;S``“J+RlcyITyޒ@?{9s8Yr6$ /s`8o N=VɦV/-ů" ~^#u\Dp^#OZ*!,ʞmi6jZwN6fQ{0Fl4C/%l-Vۆ7+Z `4Y;|~V@O6@^F-Td?TΘ ƠwZ/[(e1d#gOǵ&wNfAi:QN3/鄵:++mIj0$̓B&ǂUGүT)˓0Bvl+ww)MݣkSQN'ދ EIK)+;$MZ'j^dt m!7//V)/V~Y{[u%{*'/9†,^}MH_;| Q 3!~gBR/~H K]=n-mCu &F⬣tD)Գn S߇"?$57j-fqv=nn>/Jj0#\Prdc;3(.$=%ha}1B[q4d2r=8;SsE=kMN8UL ƒVdݭUU8ñPLGPΞ\脅2Uq88EktvkT"dwGdRR!t9+8@8^z2>?pRU`dDV7&ބ֛ak; ڊ DuBsfעCWqhh0:*@WDh;A IQ6S2bTanb4?>:sqPae6Wle*s޲?MxEN3D$Pu={9)Y!1Q丙+ H*M聧.NcfiB[%FF>m\w.NL괣vh]]݁Cf{;ce)QtEndIhѳ6!%r4–qEzesZ%4GN3U7h߂eМ,v/ST*1B =l۳* ۚJN+gj .7W68ȬTK/V+ {Ea^ _o?=Rɖ6NQ9(rf6$A]Y/~,l=}O 1M8 ǜ7}9_ .8ںp 6JK(a O%mxA; ,K]QW_CŌ=VK9 }L/aމ{cR5?hҴyr[mshg}|Z`ne m4a(RZW k ͍r\[mɍU8MYaͭU.z|65>r?Q#$H;*[eƞ%4p1"m1eL͔UEo1ca7QQ\#IBN yN,vC̽ RUbK *$cB#ÝD-QrReaG$ lYV6Y9%^%=j Rk͎m$B}TZIІң(PI8 0VαHhߏI4g ۬+d|Q):b(g$wpM3 nKl42?_Y3RWRQꍦkOvz4oKcLRj*E/%n5/kߧdmZsofi a{ow\7_ S_fQRP5p(ϠY AkR^ 1+>\_?t駽01~g+^UV/r*뭂&痳gӵJIZʛsН0%B$xI`o\kkiپl̂NBeIm2Á*D;λ(RjDIPT`U9е{ȼ]FZd9V%ZwӍUAT\iѹߘj 9{;`"xH1?SŖ6^Tc:_{|, rWf@o 6Vv⭣3vRA+ɱw6×h6s8wvοp,L5 cHaf_?\k ӢIiyݪJGPv#+(CNf^Yv$q ?{Wȍ0_5" S2Y,6|I`t3J9c!|k2^nT4Q6Il{l4ĕ)bJaJM_qTuhc$6 KU` xbVIII1@xqՓZ-[6Nt7T7 O yñ~+6.:~ +FDôlvZ" el65'Υ"#;ks$0RW.^`Ԡ^dp n^1Qlvmq(PCœQyzq)ウdsgWPʮu3kͪZݮ^\/V)(]0kמjWPn? 2p1> ڑ@%"k`wHTL Zvܴm rXuk:XM=NEMhZ؅.6 U S=)zNytJ &B+hM-yq_Y iy%`dvdMqRByhwkwS1$+},(SSo$K#LP@0 ?ɰ!Eq @{N,x΀^ul҆&|دxF'pu^rYDa" "$<{@Ize[)d*^J'd.vw|xSLL&VN&sMF Ș&GG=wFC+ǪbTA )b rmKYhqF P(2Gi ttHuIZ26Jf\@k(+ZJ H D)!GȉKNq訫UAپm_G9"o oR@A&ɜ!x>Q"n8FAT$39 )JcPTWp,1nBMTP_;ZV۞]jZJqvx``νE-qL-}wMcg-=m-s"k PG6r.G}/s)?%v`JKÿ5{{5=T.ڐ͚Ԧtn5??fWka]nr'FP5`nSk?#υDg,& 32gJal{]C5o]ȋz:[%#/Q$qd2mvRh5/8yhM8v .P祿fY"`HJ9kXRc3N[`eۻw"Aŀ'oS)L26(H.mJxiLi+hUSpJrތFDZ]ޝOPC2ʀ>*g!R4O&yVy/`몫؇YaVrZ:}~^ze}n`L"QCsn\@F+ T:Ur;~ y1uS^e: =M%oz47Zǜ>ڠXâL Kݥi45ww,!Sj]ܢ5G BkR fT+*pXIj7cӄS,URC7(Oc)!8l)!kFƹpFs㡗*SMɎ}".DP:n.} KGֶn֟l߿|*zE0ў.`ZgwB] p1Z.oYEN(n2*ݪ RhqoV{)(\N!ȋE/?_O g/DDP_"!:2K"M|Ml(:82ष<NYu &aN̔9q2Q(NNy sAk?!Kk_5;G5>u2Dٖ J3ZrIViF1n-d. >Jd/tU *Klΰp^g])אR]BDp`\d4LP\RUYr]bF#%3~|Q@KHb(1˘ %E2 QBs0LH%>XQʱ(TR`H_q`GEPpy82EEBkh]zּΰr; ;HofO˛UfYRqÒJw` = cŨq Jň®?|>BffTɻMZ:q^>HCXϚ7}هSfޓ{_+2ϻwo^\d7o&u7~ku(! F6z- Ͼ}H>xt yħmϕsIv7S &3W 9P6Tg7_3:Z͘oوq1"Smp {?泿^_7Ǯ=}zl,Wݓ)BIbe9*O.d<%TX>XʲRl0T~GnTM,eBڝ]f8QT$3^+ '2) )8-y_ZqP|zbH@ M@ƐvEߞjACdթI2Y2W)SMA:5VkC=_uks϶yeۖ [)g:}=w~vܝqUR]͏ 뷳"/RP//o+gݭ&z L@eT[QWm?u{3rh6`W5Qni+W$0ouӵ#nN1x nnuKu!\EStp:ysouMbuK DuRLjYo'2--_y|jQcv;ww|Y)VqM2[`qKLV89 kf> G*y cjGgIK+wX#(99*-(B.!a%nD.ep+_VWl6|zwM8e'U~rVs[$WY=$%ͱVğ'iSQޝŇO]8`H4HS`Tab_,u*ynLx*c͇5$)ԮH 8:daLJሿYV TfzJ07#kkxhO6~Vj{ۨ{/?K{3r(e$wy$. Tj-@uע A A >13.4ȔEV%dܔ4˩t[- @H{05Ȑ{0kTU`[~@A;2 !%枠d Ad:bځ2kmMN!#<- 'g9~ YA92i HbxЪ/iͅtj6B^7^!dQ7X+Zw3D!t^P֜1(>"sD y+ų/]Kz;"p1Uv`ZNS]+>[6fF8'٫ h8W8c}pi}?7 p  u'|~w0cVػ$sg~_Y7&2(5!Rc#W8d[y@d>~y1{oOFNMbg݉`u_mc>EYR$@-m }۟#щVs*nBMlgpS}f@ڷ}=^^HMY 7+(R|msQt(!5)(D+_ɦv%m{)U݄dhU88?]{tYhZ~d*P'ᩘBӹXN{KV&]َw@u F Mh-*A"~G =Eflk=|O!x21"d^:a2:*,%Qv|&X9PΫb +UňtpFؿz0 f] PdzS JLкJL"ZᬫVE[WaUj[Wa2%Jwn ;p [@o㴝Yܼx18".Q-4ޝ`DZg>:Y4 +q_qݹ~]|`)+xexۦɧ|On΋n+u(Ix3+D$KEFw:Ϧ Vj&Ɛ@]iVI'6BS=ՉݞhE΍ <|㳽L!VN!ZJz%zs[+G(w$7_%˵]*I\]fzX-{$ٻCcIz]~ek1Xȭ.֏,"YTFn PK]v:o@w ݗe/3_L!3G4τ/d3<]갞+SSx 5!/Sw_oq$Tq̡T|Mg.,nW&aqu<ωZAB;JhCZ.d?~-$K:npy2@"Vr}F_N$P*ˆ#\?DJ2RYsA$p+,)ƒs/Q+xAC>[Pq&/ DZ,k ,`5Zoj%M.T܄(Wd]%lA"QU2YzFɤp$zZ!Vu:pfJ&u /e3Prd2!af*KԹ%-Czέ]5Kh[!48%R;&-@+2h̽މ ޺&Ue̴|27 tDe2/\s, ՘#x6A!&d!"x3dEJMkZHƯ]@or 4aPnd8-}1?a~& 7+h{z3'i E_:zN&B2}=͚g2}=:AĹJ8 %ndv;Be7w/`=R۵]6z/OPt%ղ͙PBf4#:0\ˬb4 (ES{>!uJW7__log~<=3 o,dR<N k FdWne9/q&e9ӥLYWR7ĨXJRzPuUʲ8|{4`2xHycqǨn;umof1NlUe[}+?uUK07 ʾu~2g2Ŭl+4 &MHo@9~uMLShG R bDv݈HF I|ڍTNGsLQ/Z`ʻ걁;2jJJ73,*f3tj("u"e(f}&Nm1z-1*eafc uyGv=V>̡]ˋj =u:ޣu-2ٕ|<%G0Q{.M㰵nH~B=ܕ7P)d 3kh{M9f22 39&Ӗ[@mWqB{ߵ"矩A\>4|r[#-cEGwuLQҀx$Ehj{Yтzy>9'QmjJlD;n &r$Ud*"i.p3anj*rHAeIgcC5ʾҁ<#O3jpApL} @P,$=c?uem)-Jv{L~wb4"!xp+}6CGbL>گ/G5}uǽ6V5f[]atA>.UG@R /f\ ߑ W- 0ޭ/yH;ԋ+'k71#g]mҠ=!ȡ佾_u٠~ Ph c +ҍn!)Rq N:Ջ,Ok! R~$ (U?h/$S\_\]x)2Z8hj?3l^ {v/bEڽ(gL@>é1 EG B{rp͂eG|Yt+Om}f#A:Q;S~!o0_9_Ҡ痣ov'|R5Ph$^aNO86UHv_hp\`|έgpCжρÆj3j^yj"rCy#cBtq u0SdQzQȘ)$L P2"E^q0R$ F:)+ ߼%%Ļm4B-U(oUR %0q X„Bϴ`4#ة̋ ܰ&4XU KԅtxTd"(l7BPx-y\ ^~_OƳ˰/ΓґN^ QBQϡ\pUH[.Bk&=1tPA9oI#GObi\68$V^A 3;ED"vl/SDet"q5~X鑿[;_+,ygOJ~nO^ߞ0S3³rXT|ʼnƓ'[ԛψ?\V ';7+~LH.ڐZ3%R`2};aRt$в F^r+%?K(wOC-?qżkeT㮗j4Aѷo @xlKJq*d %Rx;$ F7,=@rKDYJ)td MCBN 0cpX48*hA8AL`[f@QdZD𮈮BRd"R=+܂(3[x# .Έj-]0R7+ 5~iMvt;erڒǏ~ע+xEi+H\]9iUn#.dܾxD =^HYp>7p {c<1*!@03ֱ BK-dǑyXk#YxAm32PB`T#QTX+wf( j\B.bj甼i-1hOMWð;Se7ʶ`Y+Ii _  tt;OeBuEXXZ3JYvn%)]1RhߍTZE-/*{UP_[! ܪ Ck ȓ.A dA37S/gWiIXW1\5T/a;k駛_'k&:QndnO4>x=/&Ud~:<=-{~uOBk6RAxwnhyـ7Kcl̛yq"ƼL[wf8|hqUra=I?@ ìL>)jXLFEѽ./iRث )86Li @.O_k"%RIp$hKк>PĀ4UZ;AnkA'JL3(,+9!؏>^w{wQj- (yԌ_Uid-ax27~̩#?EQ>;l6:F_W[~9a UA[ҾL/sKOXLx3{!U)S--6aiASnPPkoA?MǷYrg:EًPEN%zʳ?V@Wv4aZdJ彨ѣ-3c${е\կ5~И5=k`r+WM?FA!aw1|)qOO,ZRF͞(?9+ ߲{e"%+SiEEWMԏ>ucQ  R82ƼD̨̬t -I[@^,%.,"^Cۥ*wcY&mmv}ǽu}qyaPl#J;'͹A f?92o;q+)cZ #.DJRj%<`W4D,S(W %SͯHry?/꿈L`fY%TZ^ȃQ `CINLh8P;jjNH' )v7 cx 9%@@B{`:(V D.fCH*I M%nL"HT \ xN2:c 0AAG8'InIԄ@w L{F^Jm_31@PZqF 0'+ Nu^Y#RAeJQJ ph$NkC#[+>>j Ւ)pPmTCQ*V#8z)PBM^% B5(ڊNUwݿZuׁPEX孛S:F}-.tڹF'YC % {8" WNW*n:D:OXH'=\v`jPrndM\^,hK Z˽{w[kyongP+tRHiïebt.|DAGlr3_2|4/aqm6m<əəS"xkm/59ţ0ǘweH~Y)^]3lnyF.Hr1>dJRT40Utf 2AF|1z+< (wqϞrݍüxe5_iUJy^6?O`<zĸx|t=a"(c4-Gw^~`2~6q3OϳO+L7 @R!X( ^(i1\P]xoAjVeYfFB>wBH)fccVY7/Ϟ&C5G날^w{wHHzIy$@.$:9Cg4ǒ-_h9hv`e={\C1A78ޕ[|R.' r5zpB{mggJe贞Q FQ :Ay*$t@u3V26hT\+IO&J͛Xr:n@I8hDD5sG)JBCEPr|֘2-f:>^f"ѹ&r*@] $ep[Lݯ&" ,d"7UX;GjJXm -U)j ++c98T)nR 6\5y ҜyA[M4 ׅ[սoW ooH2k&:U`>L x<ks~L(]~=vhQYK(c&޹"B .j,FWEln֯rSN٣(X'&_%k Ek V.)[5F$8I}%V} D "I 2TPN8^jR##87KQJCIG+CW?Qη0c] h(Sg/YЅR(zgb66 h&R2J]hZnVօփ06)[ZZ‡Naǐ/9t0tmaٽ;.. *1-MYBQTGg4gΓLIIy24$Ӆ-%(":/m hh![Z\vhW\3Sa}`w@,Y -xzj{}]s$Mz:lxfʟ.ࢃr>u*EVN2 :yIID@\kw|x*%;|@X{ F\6 $Q{Wi-- ))%R]%w`B4U`DݒSד{Bu= s3O^GdE+e%R=O>nZL~g`0.n>X`Rca6_rŬy|bgH'F^S!W,TX}-k\X=vi߿" HOw'Y&$edkISIeed0jЖyN_^:dxD͸aF̕n/dB`.Z nKWĜ:|.>Wy_a4QZT}A}v9){7*NXw2ʆ/Ro.;{fK@6-VJUndھeq}w7 y_6F+I _"-芿r0A ԹpG}m6]S2H}#8"1#L3E91ZnlXYLk7*!㜶r5B @uTx%Pb1SN(-)P^E&w?_Jv<7D熈$JT`σmԌk;-ϲ&k4(M(}4 >ݏkalC?)_VǴ?1|'E:8c;4V ^n 7Z[#zAAD<˃X~|N'%ُ[+4DzD{ͦk׼`~+u<]ȍLt8ZK9:00!J44͊3;ƅP8hRE `iE4+ B)tIu  V߲5eY*[mMw?䶸PA5Mp@SL|؉CiNm\VjPʚG^&u[f{Ur q12)' 7mxʴ8S "eOqgA!vh\2&(,pz#AxI1q 5DN_͐c|Cʀ!W6ye盫T>^^:!LI" ^38c̨A1ۢC%ECX>SUpREYoLIKQbt,x5:p*qD JD0; !Z.+J̖S.WݛpZx!vϿeWjB)1ksȡb$ZvKՖ(zbE)4!JФ%ft7/%rIӏ K?\1-9%yAiޔ}DqLӒVYB  )&W1~}捈Nz4iH5R(Q`FbKŌXXl&&fH[5DB[ZH R\RG^fQ爛kOy`1)\)1-sqBZs@jM1\% ~?-7Re \x ?Kdqt:Bytn@dW!ir~p3%u3n'nhIVO1z3 ~cx^&~VSfY)),mho6~>Xҿ4w ~^Q"zhi|SќM޲n8u)Yf4׺1C)`N N]gU6me;R%.$@9zlYdLig ׮ Bw,ae54?-7 =8g BYbvB0DY}}O/{n/{d=j!FK}jh`Tvѣ|̾B,@[VB{VҺګzWO b*1Oy'VIGb^wu|zH2ҸVJZP~^#NgG1@pS㹣:TΆ:ZuVGJrk~VLj(sfVA4:]#8pi产8nFtt?6GiYSnDL"H(9,T:~QC\WǗUi㿬v?5xlZy]Kߍ?L.}(M*S?<ϟ7W%hSWBѝtI~0O'ۋǸ!,J$f~ir'*.}_ޱizdof55 & D  0y8[C>wy)~|?_\wE\u]0 F LElK=L?}O0U}R19_JfI P= OxHDxƼ;fdFctcE71v`=ri'c?;_EGJI0p ͘.$B9򬨼i/\wo21H)u4 ̣AkY9[oBT9:]˒de)j7u AtbruqW.&(}<7mbUv/v*{m~ϋ}%tP^R-HIC`H==,S^6$xYjiQڂ: 3lCVpsʱ X\k(j 2j"jzo\ Wogawz16PUz8' Lc5q(`!D #^#b֠.x ZL}D(l~_4nJU~;I$IߔH/!&G1%жQ4>,Wj.BdQ\pd@ cO1a 0LJ&+Uѳm<!@_=4KQVһʤSdwL$ԗaB>uY&[ vr|y N4'r{͇FI+7v7E%H MJRsŀW[+VBr7'`R2.EGo}rR@opr9҂FƉB'8 H-(Bbm1sCLU-%uT|R4Ʌӂ(Ƹxm:8EAU{o΢XO,N(,ۻ,&&9r_ք•~ ŭ|C?}/KRꡝ A/ty3L-+eK->Yy^s*as\3Ңj `H _-m^ ŞoBrJ󹝧t]2`@vstÑ$! 1/pq 3O(3^8sdNW^1c A(rދ Sc%XzSFc.XqF|H&dH1Vo28)ehicfPXGD(?S6DHHe䞋x4Ŏ]" #FXn.7sB [TB c,e^^ŐD;%JdJEKM"߶} #DtN]/C8f44s!%2sZ\?zn=ׅ > sV,tu1gK:y~~\&•}1ZE! {.^*`3 "D/@Acs>j9H!C+.z#;Aě֏z\J)z5;l^K <fIp.BVxjiTR%j~S^G g "_|-ܒ:æt鞭 К qbpSH HW!:n0g/4A㦀RΞZ 8 )P{᷻LJڌ WLOxAv70Is$9+/-V& A`DjAjt2O&2 u\yA3;?g<\ȁPpZ8 D`a^ރ)߼d*Q) ?H+Fe/E C>J`mxKF7ɶfk5$FþFہc ƩVCh 2n='D 8KRWߓߙ%3ZU3[tlb8"r  ʈ5\kt-%K*T-:?/m<(ݮ]T. bQxWwcp% ;H51/r&0ѷ+3+ˏ+4*&nnlI&S~Mw'?a|'?a|N'(ΔZ!2LaA1G [6qs)4ǂ+׏.CTžj4[fnFx0 8r+L_I|N& !OZ:%;/.)D(Q` ֦ <$9Tg"d193+ 4fŠ=kA5mĿgrruȟ8fFL39U#F1¤*OsLa %Ynfޅ;UIʒtk#їeUvck D]^eo%QQ85z"d8^<;-z=4ͨVCI+41TĨHQO.w,"2¤YpJ~s-Y-5Cczѩ`z9"ò aigQYpg9g5{*T1@ТcMθeYl< j+Vv}aKU#,eY.FS5HU^ խYJM-0uTOg?Eld4ˆ(u힟/Ly"W֍Gz~[1[71org˓gZ\/?s7uľcyf~nݤ' mpXzQ[G{}s<,v`YX,֥䋳nZҐqҩJP֍r6AľuqIVf[|@քq)-MiCnN;XS;n[U4Bx`;Ckk;wu.@{Ej=ͦ#u7]3;R-p$t* XE!Qk Hz 2!Ք^7'/ #EIYlkɫQZhKU[0eX](てHHն Nj̉ e+*uvv@ZDzp/B Wi:M}a>kaQq>ϫ{A S%اā"K%S5QxߦAm\z)DB&#]ƴʩSy0uX#?t㖇 p`wxq`}2/dRF L egz(Oڬ]bnRqJIחw 5רzTཹad}3iBxkH~$Zm [*Hw)_rM*GZ=Ѣ,b4QE5-h(bl={MI涶SM}D`4" !Nٴs5 iIEsim>K^t=pBǒULƤUpҀt B`,/(Noszij vtb:(^StÁ.O__`iRHVt@X2e(Y9.͂舢;#dgۇ Lj#PǸb:Mr'q"j{S-lSyB J&֦%P=Z5ĖLrbDO7krŠ]K_\﷟'2|IiJ9f4O2&`2c\WBYٯy޻U ?b;?(L@Dc蝆 NB~ Ca#V>5ApwDu3>6z2Lutd症's_%B 9T9p qr(*wKd& H&߯jIbv%W-/~qu0twoOuG] ^D{%Z0]1Umq^F]ʖP?/OU[8n:TMZΫ^k t^"er㻗g7z|>L/׮ܷL%JIuwy(i=\-_&>V]cdu Ȣ XtC/TuadYF4`IBX D[<s)p-`ƭE Aq{R q߶o>*Ň̡S₶v|ZwPSf}7Ni.wy :)Nt] ceݓ"Bq>ݏ<Ǽ7~R [{K-E-_ug=S[|>d **x'!:q=z_)::=Ea<8Hkwup !l.۝3&Mj֍]ø3D55~0hE{"|F½x> xBm8 5{:"R"pǢrZ,rrHccSL},-kxrʈ#aW[GQ SkRAH*EW$gl Ps=LI%(uF\NMuY 3`0SSќ8 _$ډE,-+3IK !k) I,401r>=a))"S)Z\DH /%nWI~B: (ɥ(˙Siaies'8%9jkΨ81# qSbp1Z07vZrD_ytSjdV9p_JɕA˔bIRsМJqbt\3+vFHY ^_A~A+:)Aar 3XtPJ-E6 K5F^M\sRȵr2q|gSs! Bd0B9i- R/EQ+7XOItkjB BsG;DO+ h(@#O4|z S' ~0_>MMsf%2([#RtKq1 KE.~Er~bَ)? T^~ \DDhRMU_zwktfe>t}3BS]:~;[{~q/ E6χb_|{tZ_2ŰbZH%9L+n@Zf:0SoW"/]ڳhjG ͶPE]Y3Kbi%VY;z\|Dm|a]xm%ô/#VVRIJQkw{ _bNLTy헦_zHWDqz(7z۟OkX%7b ַ-Ȼ:4Hl%n  Z۳W"P@[/,9H@9j .u Ӄuƣ&*i; Vɥ~ .g)YN3dJ\r􏇇wW_}9xyP|G/Onl/{ql)=0}ITaoAɁXBOIfɺ2[h kSwNoGD Ahёhk*M,Tw;74:)"=yI)xIeGf$s~ )GC@n6vۦ ]0âpYCv7w6=]ko7+ v%q2  /de{WԒ }T&WU^s^޷Be_:KBӹnzN[/dt >ȮsA'u V7?3ǜ3Wp::BCAc(K'5wT}IƮ4옘2v&b'Ch9Fj&YVK:tb'ax64ܶg'u Cݧy{}XU`hn..Eq(o?AW{6d>Y1J:_6A{~H>Q/r<4]s|5{x}VwXveCYԨ7_؅#W۴];?OYDRgc˕"C4+9ɖ[.1m[GDV"BN!u)}v%{&`9=Q>I&F}z(u`kL'@zI8ӭ#Zn:w9qX)o2.o!Z@8/y(e~v:T./??|yE8ћ׏߿yU?b.6,v*]>+A<>2(_O({S:<>!F@}zywAO( t9ْBr!BOn!=ϵ;XO7'?gN/pv~N*9Q!K&vh$ *RNe.-s pÙ"$V0ji5HB`!1X`Ԉ d7S8HEf"J&.Y+[4nZTҀާ̕ok5$M&kc<"݊ٞ|5]ICЬGB{6F1)Da3JhMKOgQ0xƘ!\D" ፕ1mV% U(h+ ~3TN|"m K4Q'-&GGx/z>*ʋT;eCER<Fցfb{wv3scBF6E'篢IaccHE8u$BpD$L/zK/t 8Ř.SSϺe.[XzqI.~@siRDwM dP{aQxoq QH|6uHF A="I yv=&WDv?;~Ž„ˎ'0*9<' J JJat|%p&٘E L+ZZƨIŤ;V @cq1Oa)U>Qx&Q˼8)GP: >MW7)U>^}Q~J`G Sr H2V(o#/qL'nGD;G(J(VN(kI„FV{LB`脖vu9:]ywCDLŔ>΋$p:S[)*ҤE9@; 3+}3tͨSSZ2ƋWs!8XiCrԷ~!vdbCcd%:1W?^|#d_cyy__aG;t1ސ`K^~=q'9B^ AxzWaj wQ5Liv8d9K5J)ù{hsߡ]| [V:rbm fn}cALe'1X "h.}WFZH[đ@4'SS0d1y |]z)! J"mD`x ֔ma~_{,Ji#DXdPGC5_JvJImMCuHőҸ91걵y6tֿ]=>x\% '3r H$v}|dnFo?Ӻ`6Ň7_nLo5 Ѫ.oD0{\z/Osͪowл\Kv-_ fWA7|~xɀaow7}qMQ4Ɉ@D=s聹YXۣuABGoH?s GJ$JӣmC%>X= Y}bnܿM#̻om6n]$i|R$TOJ (N; o6ѵqͱB4Jh?!I !Ӕ#3 ˰Z0IT96(c`jԭ7k)V6'j,95`|8 }`?7M>;ߠhy 8W(l6΀V|n K40~5)J^h9J"Pð#4.*HS7V2(fcȉR؏N>P'> wwufCukUB}6&,?@cܪ])@`<A&O 1C.:S k]8_:fX%A+FHԟ*>Zkq?.\G]0*VG"sfM;]vY>\{>PyV<-.? ѿIs92'-^zw7=?x?)ɈG۸vH". :O)ؓ` wv|+-AXQ8zt {^t 3lZ5jRZղ"a+*e()IB# UYVS2b{K"MCR$ R12VPS䗈DpMH B)l})uia)'ґJM=LZdy[vкQxÖ{׏K'clLH%B(bܛ嘳A#\ A(O8<8Bo6g=8f;UrLGv&#;vQu7YL >EEHOт-*Nd&Iu+raE _z=>|?!ѓ:/58#GZʇ:|&XN.-Hˁ_RA'ykP ]f_l} VBQl~yAHk)v6avoN:9%e>(e>*Hv=z{(:n[cEF56J+G+Ck*-A\)-v(7GH0UŌk'lhDcSЬfaYB ׈23M"C9YS0U%A+ȱЯ(Ocn7c*FE>GR$N.!9zq)$s)8Sg@"(o6WӀdlE^~eϓo Zz)=. k_n×4{-kd.qaʆ D_ Xnc# |Kq9X-T (KM#~ ?5Eb T˦14ߌ@eCÁ 3}[>y᣿E sϧOqvϵl'K'&؊똃)63dž,PFktX.&4=8*@hPJ3ṅw]Yo#G+^Y'0vvms<DnٲgIRRG3YU,dYdF~E?b6szQ' %oqI先؄ eqLIEr s(CJpT˶ GX@e.$ICK)$Ϭ*(MjY-^.Qgv `p?M}wRp6YUBKfY7|msvwk%CYCEBzzig<|[>Y'_.~Hr&O" n@~J']Υ} hJs+|> M6ԁB )'#D cejM>G5<{T؍YKvi߱x h ;xp2Cz6Xkyqۏf:ĺ@tTmkY‰(k|j]'<縑 3<#r1㺿{OTv Զi5teJjztcT8rý%fN0Gx96FMED[ɛpE_EhESSϊJ=jJvm. 0\×wqӃv1:jNIpcך ^9ˉv!AIC Nemytpmah*1Eܭg֖c"gLƤ+ҁE GB}~zܧIAw^\5zQ=)t^أ]s"ss)*iwaN6XHq "ZZro _5&Ս? Vҽ@\-QH Go%3-֊QGMt'_ l38 cyJY$H)8V=xj886^4 JNw+lN>qMnH!efOI7`L[Oި=%m(2V"M`90!cu} A KrCX^}6ν&eLYd pTl 4#sFQ^OqtR;*"55)7T5.hK QׄĘ<,,,ެi{Mb}*<İDic"yI @nY'qD}j"|>tsBjfÈWהZ6 lL!v<҈Sk%I}8c!DڈHRn]HA&JdPT1R0v6jťXTԦIQx5Eu%vFEQU+ wl;CJ. ZP0S0d5p" 5j-A4Hd; qۏ* ya;'LU;J %@HE!;g6E"(W(bCvb &m44Чe]kbc.;N0{6DzB$~j!SϼRm5~z7o\4D]E2O$o%R4$kXreUR,%.6~sv1w4T6O,$O@BC{,%_neݨn 4g]o0FPվYO˝rܺMmr0V cr#qb1wزHku7RpҨX=hsyx'ڱldJz8Tc+mmc(OP#SjwߩWГ.9"-(LN>h^{kpgc^DoN_NaԾW8$%`*!nyPR|5FwnӧbijE:G_Rm@:]謣@N @.E#6H,W*Д ɔP1\ <ιG%_'nRkÍRІI>֠@hDĕwXq::4&2T2p)Ҵj铘8g{hbREcRkTL|ybfM^Uqbz)&iQyߝe`1x7뻯L!=__%OS(|oʿrw苿amp狿xa~7o:5=k-8>ӿ~/[NFXckrYOґpXw x[-%S;FvĨu*`ޭzޭ y&zM)#܈g˓lN_\Ffr}{a9o_=8-Zػ|o./ܛorZtf5.<\+ pk a9,JLPN@5J9cL&(qBj؎䥫R bD[qhY,y0KycB h kAxкiLi&eR z:b+hX\Pz)ޯ˧E Тt@7U!);*..t 7&Aܚa p6tw _&f xKҐݺT%Vjop4kw\ OmDaS2My??f SLcCtVQhNHEr%c|Nr9yqDsJJ4 MMA Eu\ v88ޱ;gA NMeZ9۶u_ޗQ_ 쾦1}CKf| /.YX4M^לִc^ʀ S窯H&85gMFILsOmhi|(1u:)JCYn>P]ɠK4?M+1>k5*%M{ ȊA4_֘_#VpJlfTcO.'%L/s}ډ"ӁN^Nձ`[4EO652}{ JB. DIa`jT3(N>Aa}`};q<\}lID;(!kL&B 袨WױMSF uuOt3,u>v65d6}G]IfGj!AwJ>6FYRi1ꑜ 7[W{?r}O&~K& {_дf//||]LO"Y.2"Y.25WzxC9*PFph'j@Q.=%, kz;Yi5l X}\z-]q( hBj̷$ 1bD bb+&fbb+&f͊#`%x!%M8D1Y>5O"-=(uS1u`ݿKJ>!7(S*_z-jFw3rfL!˘,cY2f,ixp#TpԴ K+&Ac*DjGv*`ƌE6Ȅj:O\\O'FS̃W&V `4YFƚ '͒WBi0Li c vX!L٥90E ԵoJ͌4^J.90K/` 8@ai0ohFԑ3L/~|ª!EN vc"6 /׷ܵj>xgXJ1^.|~8Y}q__뿱Uڛ_gj깵<&R{(EW9TxF*2E.+-sKGsgQ8+.c>[>U0XM FS!qEsI `΅I~biZ#aNcXU 1G1ù̫O9VȞ0sYH 1iDrudLB)kg"nLsy_gv`r (9fk_q72"5cGk)2F[r?&p=K_^z[oro_iC+ ϧ(=WW"l|WˏwpE'kTA¾;H&oW% wLs=ޗ#ڻW?@.|N]O#*F. 5* P)D[6R7"2bXr_V0tPcU6XF5֞h`YHsUA.`uoV>_{wRe=)ВJ:N)̉mk1!AFCwЫ$)ZĀ*<ȽVaYJE\̠wqVaYJE[ڈ#rK ^>!98dI@0@yv1?PFm*]>=]chT5{5]>=x]5ʄzۖ-Ơ#x z8Li݂.G;2dq2; B(>n30t_]丠*k!͓ye+^Mzv5LlEO t~p f>zv5c N&Sql\:GAFNJ3 #566Ԗ|8i%]DT '* ۳ "l@=h:ohRwmmzٳgN;iijX`Lfvd'XGdْcJ,,Wd]Xb =l3]2P v%*#]]< IV0¤ \*v#mt`'C* A`s?E~UaXFHðJ촳*K>vvZe g/N]֘S|D(S񒣕8HGgWcrbc,5xaJQ6ĤjMGE|ĉvGb۷Zx]5Կ#Q׆H[XRʣ'E6( ܻ3}y; ?YjEQu>?R9bFOZs(Fc7.# xyӯ{J @T.z@E^vjZTMUo+*P{SDT$*m{E P5Vom[Q@i*k I*vi}EEŻճ (l^Z9F0GJd֞U#Lau̼Ug}dzokL )\ . 'J?d=:y^{sd pz\y3]LLLLb{!༑!-@c;Z[%Qaunk[3lp1ܔI@gpt7P-~4/>hϗK[{𝅙)<6ozk~Poχ0πd.s/o~Lg T~e?s͖S?s`w@:<*/ڊ=~z0>q,,0Is.lZRLB/RDcLKKR "rU?7e1.>w,N]:ݻIzHqE!c.ԪubIP$i"qv9~SK?5GݏOe͙,'dV^UV.ͬbh&Kr1S >AN8fsS{1S lecL"Y&2F+2zt1ƃOmEepe'[ah*Vh7-d L qvws`}C~[!8Egssssg;V-Y3pJ;Xrc1[vf `&GgT+:Sw@`Ji&QЖjA԰hTT r}^ݼߝTO`bzEe4̗pr’&*UC[Db. eŪ1sQZ~1QMNXĩ0N9z 1F}rlf^pQ7 ChrZh]d0bi.50p[UZ\IorRZ'aUAQ̻+dyt') oS5v^1Ec=>-L¨vCb4l)(BY D /V;mnmży)zs>OSb|r75!]~TȺ]s/ V"'4CGdoE?+1BlUUHD|qD'1ܤ!&+C&frwŃ{^rmF \&7w7p[{=94OÜOSp{wt$>` XF&afp8UJ@tP Z&d)=Ɇ1T.M\)+iEзZ FXo%XiGf9$!kuQZ߫m`[ w?$ʁvȅTX|+[)2r/j$Պis\X-ѡ6gViY8d,7Vr\"'Oo&T8T i<<*,'61H:pFs.s `\)uFFg߇0L2{7=|An[{e }{ \ B`R՚k+e^)T(뜏O_ʁ0P^A:A͡ɻIzJ=8w6#{<{?=RHvi<%^峊ٺdEvCS*K)kjR9cDJ%)Ȭ5*L2f2o%h0Kp0/R_4N=B챊|$6RχuhKd0aWZ϶8$,9)5ԟ9t3twBHf{nsEwd  ݝZƽٚ~#1dKSq/z.2ߥCjs;t. )P\0s"f(+½T c,AUoƥfn+q :[>;aLߤ z.k_|غkf/'i,U<<lkiSi^[:.7{ M}NQ I_}2 L^2OK*TxGDqdo;k$$:h1[\E{G I) ǛJC09ŒQF<Ăg P ]c>,"R^yh"jPP%k:e;ǐˆ,G2 &,ӂN؁ю!4K3Ep\lrzy&;Dž곬'M_Ikl7ki.8}I1tu]2>}R$Y58TxI#勵F'8D%Xbt7 c ~%ޛJ/|PB1h{N{6:^ӱ>gq妼]Yw';eF4i _xJPEE'o_N>'}C]c:g ?qTý1P^4"ы3MjѢ+S} ql'OKwI@MBu /_kE]lWȮA@KŚDj#9J"| 䅆d$qc;c` /mEs@ǯ$OmTۢPE7g031CX2ySJ3b4~?_(0;~2-Ĕ!TT&>ϷA174+\X XFјgZ\Q(1 /_CSJ``b_$][%- cZ箇El]ws?ljd cW/}2IAhDYCm_C#MÓ*zqǝ 6;*t:{6(׼O@Uy :S p6,MvK5k. :j2dV iG+ jѴ|ŭP=X#J[:N[1Pm#hK@:'Gy:T!ɔb75Կ#A)`ݓ` jL36^_^oGA >YqlE[wRANW̸뜛lo2.PeʻByQ\eMfC\"7[9,ϵ+rV<\pB4@, Jsoa>K௳dŧ~oxϴ7' n:@u`ocf]4=A(|<\p9`.|h^TEXiT΁X1N N5 =ϐbyT;P-;eTNZV c F*&D %a8{ E)kuV؂ MV|/$ Klp+ c$$Vc$qe%+:SuW@[w,*/%ڴ?e#1e\x?y_ogIU%7 ̓~.t7o{6kcüeMX]ӵǤlcdKi>*3J Rĉ !9&ycpS"ʵ%weH 2uꗶD*;I"'V\kA)Z?d Paj2 k-_N",/r^`(/ [UlayƔ:-9%^$uZ|1hT8wοfn2ԯw [&nnIU]pr+CUB_FLWW\,?_y3`LJʓt>EY58+\,>6WBLJ?=B'ܯnR ]KqK9RR"5l: ~]yȣ N?;vq$]5 %0V-H}'%nܞJ>d92SRz2 e(^/qd4DYjLNbdFcS/&9UwZǿV!B!N`Kz@GP&8y/r(D0m(B*ndwc9蓐v%^M`Ֆ;@/\0 @#&+̐dB0\(SWua)kknVYJ_\-T\<%b1-ĵ~CJCC 'Q(puht-L_J ) #e9W*E! 4'%`L9Z#zD0 Fjq@:/f / dXyK`ZXNе5*x)&10°O=+b I# ('A#G3tp}'_0JXIDZl!f JPHX 4Z/*J+QY1&R%[q8"x+Ťj*0Kkc=%0 2F2U t%0U7zH&AwM^<yAlkTZ1`ƸYXPbxyф?Wi-Uq@"!MMp[RqT8=F"4Kv+Ba .lQ=LRc1S\;ƌ1+XCۃKa,My~R?KĂ3'}fo^%)Hej(u6Q@ak@}9 ;ôK4V=!"p28)bh/]A'onclz㭷s<NдߝM{>\\?_l.aXCmVՏdF>. `>ܜͣUu-zwk1eMiBQmf'{Nnsr=p_|g2ke?Yw 7<b@V|寖w #o?|AB&_@C u>W{ߪkɮN3v (hؽrD:t"!M+M)oCAx\^clAjÔGf\oWw!ݼObɖw%sxpS$1d0@_ȉ=U|UHOsa5/J"PJ0;uO$Ѵ:j)(.7"We T$ƥdo&ƥ|"fo),,f7%=dnB eMNck R0QgͫC\3~r嗲3?'e nB]YIgϖ͐ Tt6I[@x+5>XJPP a JT`z•; s.X>ֲ},3ܮ͠Z0yTوl O|mi;eוӮ-<9k]yxTheYZ7s+Wv\Xb8!xt(I 9 q8!-+Xq$s߸ {MJ,Q>)&'WwRM#.W~<};? @=r0brp $tߟT{K=na "bځovOoOQUp~?#ޔ߹4w/ ,PGkh+Qa &^~{tu7pw#?ccC&R!6@rr#(ݟj /”y%uu9K/o XKDFQA}=6nUw>e,ttAvwscY8M|wG<򊩺(Di[-KCv3ruN`,/-p\z/~VaG}uoa%Hбy@5{^6V{ g`SM[ӕv??D*fZw(-t(/ǖ1u$f1#bF!ьfT0Z`FZa:Zx!sy#5ЭQoYWX5dW3MJ0ÇM$u?n"TrQj=2QQٓ:^)+̕}Hi#gK3%\z_P~,w1ec2?ȸ)b6xIk2T=6? ߄(-bRGgDIZ#s1H 徣y#n-^{nCB^f߽6\(8v EtrD"-$j6$䅋hLUGr/ӯ>JR`S=uO_ Ua\RG0ڤ2U0RNm7p2N$ " Ռ:?$|mzs _vaVr,i5ւk<3w<]V)n7bE?&ڄzy,`;6ѱC8 km|I"!8DUk 9匁]\p)d>1i%OГ? !GAC8{uPEB6, gR26G'b0_c3懓qH(R"If 3alC^akfi"@Il bJ*G/qH:-P.Fe
fgfPQ{C瞩 ;Ȧǃsi7LT{lp?j̻c̻ygDp3հr9 ,x|6|og(ɆcTWpTVp|y{ 'b2A0MkʄH6cД#'L)ѐC%w*?B'Ya` 髠:nP$,eW9*Osg <kAn/t~yBٮ{ojcA;B)+|%dIQ (1 ׄx^}Mtt͹\-jԲg2Rcƭ=JGUo9P__WY I!G kRM-.=:C]sW_~1?,yʘwܕq?vW&A%1dei(*m6EM^1%-hpN }z禈0vnc[2r'pTP%xS6Oxu=:8#HtBFAM3#0~e(FvRp"<ĔVW^KϊZ^[ AQPL:ҳJgei(8]To3TG\l̂K\VVrnYᠥ/.L) p!$U`LWP˄˴r՚#7X&^+k̜*Q [ |EQikxxٻftnj\ 0^_// 0ѳ +6r=S(9s]rhwe%[^x(8$̳[nz8ݠWЁ &\_mbYFp-wT Z+8>b&~.1iιcdxa߅|"^Dw%ڥU5] .]1M/>'D$bd@Zu/("Քd<1󥞣gP)q`OķƆ6 ՒRN?Kq$1@19[_CDz"l_JP̰M"şLdë)\i{8FSJ"Mxmhn^29ԋ+*4Eie`xe-#UwiĽVRi~.5nP^Q S2h w? Z ͺ9|A, Xhe/W)PY*scYlUآp12)K**l/ >|}J< &Pݾs:²8{A;A| _Ys=xk5 Ii/Sb1(R(& ˽t*`N+] * Xb01XTX=BLF"&q#8R8KL$'ChntYVq欅O ە`Rd }tG+]( OBB7gmP|a4*ZK]AlΔ"ЬB#D47} Y%pJ䄔*7v MxbJYKIP :q}_}.#piXx߾pK3ն*{5^9h9 +G4s*ґõQv͎󫛋6ڣUx_ݯg|s^_P`T-}')[΂GsۋsF` ? A& /Yzw)CKAs:61 3ࠩznȝX% 1@)9ڭ-]nKƅIOFZkU + +sjXk¸^ڸ8+9IqɓLr\ b+ lMzj]jF=Tv.&qHњ_RrtӃ30Nਠ 65ײNpNr P\M[&w&Oa˱e,ʱ%bo;Wp_kهY_಍]BdR\1uľ2vԋ?_?fuf.krJLf9~ú.Ź=QpxU|A mgmg8yZ`n ,XSV <SVJK*&) sx HRGL ؀HA`EBsj'LͪJIK%eJB|QUzye'>X!UM# j|?qX ;N3)Pu^׸)2!X}:3YG+( 3Q VY\4( 5ᔗ]2_q6gQ'g8$%*nK7DST,Chz( ʋJBtB ΑN/Nb͟dowzCv͉di?8~DAl\ OiI\clԑP;z 2Q8='2틌DK-$MUJ3\!1") e AUPbQp\&(}4ڧ}z܊;[QdtSF+@^ԕSuEarZ<$)'P N6P, cgX}bH Li  tsx˼9g_7&]@ 8j&zD:.%W=w(O\l P]JA;eJ ى_IgC!W~.dQt6nQU= ?tĞ4#'ۜS 'mٓ ELQ"N%ƤdvIi);96rI4 Mة(1+nǜ]~ т}rv:ѮV.?4$v[xi1 tnEXTݎEc18c]2)]]6gܥ3*B#Թ=36:IYu˞t))H]m 0n/$r@= BnBq?rK|y{ ߜJ=!Z- Ҕ$4HXNe\jܝ_,_1R{qWzBQuQ"s}^~f?tS\p6Hn" `-e94"㾁l-W0_p@[-FM4䵢!*%( c C C.qq+])AH{_c]lܑj,pSfԧc!F U%7p OdtdQ\|b-Ma1 QB<ض \]4 k2u-SkWy)!=j6.7_w03 z[NM2.zDcңשK!3U~(&ذǫVB@9FA'jO?:I2=RL*ԓ+-! DQ>OPMD.?G!0Ɣ!M("RM<&&VQ_spF`O#ѳ}zp&cxGPZ3-=1r|G9  ɴ<'gC{ic%vg(Nǻ8Y_d&zz-=#upnJC|#-8uo-IR@ܝfٮ:ɿ߅B!S2I4X] ~~rdZeZekF1E*A[+-EbBMuo*S@%EAda7} y:{*l1j۝$@K*0Ϻ iTfܑ؍%бa P*KZ-64PhI^!Ԩg(p5~] ewB-}[#B&֔׽;r+M y6o/ 7W/Wgg.Ib#+>ɯ8 }'1o8ٍ3y]Ofqբ|:<.lj }j1oh<'qdA bQI %3^,e2\,eɲ~YROf.*-*M+$50,nY- xV XA.D'3̜;x{r=XN)uB{k, 2+˺R\-mebDsVq&]T\p(/+T;k||~@Y;xkE) cdµɛW$<M"70BE2GYJOъU 0 )491AnA! m$+.U-g8n!FcFv*̮ug힞7B ۿm(Ḙػn?S{`yks4>Cu\/_7Wo..P?׌w=nZ2/Rj2_0<ذ@U$ӊ%EK;~,mR-2BӢ.Ϲy2?1;>g-O9u!,uRv臮pP&8+*e0lc#jϘ)Q@D!>U H`/E81fRqD 0a֙7YRJfԧH;$'8A97{d-D+aNJ+A\Q)LʌX4\ e^ˠH@t A %`LhM1XH(o K=cNg# V!- t=J3۝R12-Vp=\h͑i! Mˢp"-AWY븊cZk˻ =Ղa%8v*!?p3GavA nYF)45QQ!$0sL|'Uw&{׋y8Hm=Tg` G&swye~l/ o:? A=K4?Ʈ3MG}G7|>N7wwo y>~ؾyLQgě`eTAy{M>I})-lX;)W` 7}w @ {4@2AǝB V{H8M鎺'/&'`bL jN&Ef9+Rik1wF,Z^.Hנ֌ʈCfupq˳Z#g8 qAy" Ds83{UvTzdbB.tiB,u0{ԇg/t_n<a֌ 0*\w[mpƏdC,pG5| :轨7&rOD–zn5[ߖ@5-<~aQN[o6Ӹv[/{bԸN qG)b.|oz%,dZ]^={Pt^L02Ze<FjC&:abY_Iqf.Z/*%T}H/&= < p/V֦ ܕsE};}Fm3,ZYK2 I.T^eL<ڛANsGM0Nvۡu 9֐7O=F!vY9qe"s^jFG{.5'if I81aL=S;J0%"R)I Hx4iDPdh(b.=f^9TR%9*HP4es=T+x[#ddntRϗ[SHë?]Y(L*oςl'[Qct%P{vg:BuxMHx1}gDbLR:wD#N£+Ƞ᛻TReڤ !&;!) ̧d0!=ʵp Z&% _át aLwkS|+cfe<:B3Y| l Lzrc|u_ϵk8B̷ NqſTUS;p(!/^ԕ(\ Q ȓ&A\:ȪQv& U5cK, Aagg? /4ci9Csw\z 8dFik.~F =pq-IJD_^I K[isDf}b3G$đ783yW0 H꬛w8k6O'ax{g>¯>WCp46Zc؈7>(RV3$qpDccD wq".q .Tj%s;H&?*ie>/ÇoE6(g{XW2j°hzC8t~MߨgfOT/Ə1x9#G3D47DL\9_'Jå_yjAcT@}|vfa>Wwxcݻ/`#ך#}!(#m#LHՔ2ʭ=?+N tFQ&6|^a<;i/w<9Ȥ*$ }"KRU(ͼ1@;9ֱ؊yklMѧbј s 9'2J-oU n%$#t+rfpZњCpJzk=8"5_ ʭ/ (Fi.;_gنBzm!bA f\[m9f[]ajhq;rR8^⋪iT|r=$[}r&g6 ǣI;_zizdݻ3{wWڏvxvag=Zx̐׎cK]pqD7n,8Um :[BL[<Rx/MahAaTዦPRD9ψnf%SQn4ҭGf(&h][ɔ<\]PDc&!~VG<1Hp1j/Zq}_͘?ಣ +p:0K1YOoBW]%((j\t ;VEˡXXm»+׮R崼{YC ӂ(lx kȱdlywН ݦ^ok sL:>4*t~Cbr{x]g8fn t:zsw!QRGgo:o:l,s X#po6 9='~2+\Dw~bC^=Z\}w @-^5j;D<)#&1D[)C4;g>yaG7= fmPS .Obi`^oTJd{]{yEq .Q*K;+ @S+o.h.礼wX(R)p5牶FXʠ3)83n8E =05o}k) U8Ӻ"(αqW@(w\`m!$dpRPϷ.N 8,;/v߼W [}ޏ|vl&!~w3,.x <:b38m>Zڥo$/.Jk**;rِ}uӤfIgՆDIEg.E*'UrCf UfQTp S!,9u)Ԫr.1&u{Z]djPfi0\_㴂8(6i*VRYV.+XM:-6O y{'v?1W:s#)h#DdJI>~Wb ]aGR \)"#"M2!\{!T\+C+;lS-tݱHoPngwDqF ZGf =p2M 8F~]T GZ15,p֔8_]`B )87mPY|-q݊RGuzeaOr=zcuo!|fɋJ|KHV*đhXs6?>nSTr [VySAr(]ӷbmGBqA .PyM* Fgc#(wܳ*lٌӔxB%*aU(*s2'oYam7_|7%HKح#M=DMqBkD`fc,<hAYLFL8BJ#%pE¤SG3L9PIPjRUBVjct!:  LclSRVAM0Fl6T oӱ @C ~~\vQcg,9G(oޓ,88wϻ3QoÆCs#(%b]!nzx2]a§8tz)~LX$.08 |0pmsB;R-ݹ2臮pP8bDlw$$,#= m Z *f^{& DB%5&K\@LhNIKĖSԚ.r=`qEDZƐ>˪!jZoz-r^TC {[MZoqi2)Dt cAm}FFfzYR$/OO%/['s5I^aTSF%cj["Q"! wWKhm $D_;vvwr%ؒ\qVYrZPШ~-V@TwV4 54ńUWnyQ8xu8] h\pխվYdǭ\J㗳o?#I_6vg`gccg ˞yXO08& . j@dQ>А )DhteU?N bɵS1J^||WAhYajGO7a;TW:p!T|;gS*m/I%F݆Y ȟ~r%%y#aqbaFyZZY<.*)97ߘdPGfj1Zmsf?M"kRLmLpQ&R+0'z.VG,F*FtT 7wB娽/;j,i ۖ8L^AaN)B< "G8R !gNvGg*rT"UVݡŒVVS= -:Њ&`T*pHxbŵq&jy!mвܖ$RbC&hl G"&kIDj( Yq {Oa/tҔ C--e1&y )Rmł+τM=Of԰(~_֣R=d8RuQ5y5ԛ.-3M0w{luʟȀZ ["5:)"F!(U!ʝnNՕb^ 'Ā=,va)0KSh;GJY HY |w哽`ⲢjBmZjV4jeK>[c ?9% K]LflfL/W.՚Ȍ:MBUR aޯ-^cjQ/ aڥB[!h Yדc~LK[.Gw8ٽW/{O噞SY[Ro>ylG%)z ^'K^*!HŽU.]%*٘*"7c̛y/5_-l(u0gVa#B 64my#bI)f=IrDxjּdδIӖzv񅖈u[!땉 jCt y0H1k +u=ۺ(Z'O :tkEup[P~R8xK!HBz۲՛>",``a#$:v7=z*0{!/Nv9-ɅE{l `/O*<՗y +x( $ɮ=N\H, FdF5^ ,0gM# tAR$*-heLw2$+ho.N$0 I糧9?l (5ѹİ)tˏ.G܏fi21cjb9W|FH̴­<\J4CaoH{i߯,qN<ƍZv T|"x֒hMY핖5t[<}KokNE=$kYV6-u+>i_тW E"H}{L3'Ojy!,L雋OfrugC_%k4j9rUu;j PIDE?Y/`ڱu5p`] *!y4WEq~Fo::Dl-{8zP+1({,#,edXŻ# [ (Mܢoo@So p7<88z鋿n DW~<]Wp*#xqL/Ir|Ы+DKYqLVjvGe&h${>,=UTGs,G͊"& 3!"-4[Yxoc3$19p8΁K@^x9uqs`AF!`}WAxc=vBd\RV9k?S\Xn&K Moަ/օvy/`/mz3aVa"H% Ep`NvԯR-do޾aY'Oa:m6ym{i9IYI:[ mJ!<.4}0tz6y L=1TdƘ>~ >; jsFyO,^ڂtPpB֚ےm0n:V\٪blA x:"*?{1ܣQqnv31oarKlo72V3X)⩜q.`AAn܂V%,eYZTec{qE~.|+>h i+XԨ0^3lSQ ;,(FOu4>S$K}mD5̹֕ }ʤp@g/gBk FKi5l]2_WRwĂmWh5jAKF/(Zp&Z ,|2QU|f?>mND:[>;Y(C25A+7ہlbuāZ1zek0 R*5j`:o \)=/eW;l^SĶEdz;4|x.9!8{= PjN/nS5*ĪQ=6i2q[~yMNIc~oTrCtۨp(&dpgTD7XN7Xj1P-4H1Jj=lmLa<S4f9T"$%0)NN8ã1Ϣ` Bb3jKm0VqATĔ`[5o%5ƭS"g 3јS(K`:9lZ#Tf "@۫ȩm)N5IM4JH%+擕|"j/~-fB5w$<2^ƣY6XG}f|.m ~t&Gjw6/Z˫G~p jYJ?oĴ y㋗ v .ADո? 3o/@N供_\(.Uso?-qwK?> +S)WÛ?p%LR3s˾g7ܓ)@djxgs97x.UҚ@?{WH/mY|`d19!!;uv,CJ-eZXX{OUbS^Xx 5R͈Ъw-%Ѱx_ޕrƀR:~5.a- iX.na]u#>OJ=q&ӱXw[\-\źF`m*ջas kyw|"kl^ɦhK$uyikv)u7 ;B ጀACPFZ/%Tg}цoBnY~Ҵid6獅l|CX+7^6z%g1Štޣ6NZi٭nCX+76Ea/ؽyEoGZ!]ǎ[E1kȕQDЪwD爭$l&"\FghND8\?ggYބߧ<A˹(͈'#>:";+eʲҢ'9k0^-\[Zh)NtVpV1"gpV虩JpES$(fv&rF;0uCxWs.29r>_.FNIHwyZPàJNiaҧUF 0}DhC8N <Gp11& eu b*r`:6⢐!yZ5iZU iU+# DوL=N*kxU}V]m[Ȍ,HGZxD&w$ҥeT4VsN[`qmL2L1Bq&Gn-m!wl$ZH+ Ӡ|z)\4 V&̝u,ZHphn4s`Z8rQVCjɥnjTR4.;Tk*l8,RFI"L׊1DҮLj8TlAMh },NSSVr [Vq<ҭM[k>lJr֞-=j?mz'١ 1h薼_Z{Em=;Q^wpe\| zJ0Fj>6=5Q4k?ꫤx`{\b%`^JAX~qDNADsW::fiZJrʎ"(hq 3W HXEè1ң`X,̦70t(RYĞ5XQx<ELh`6rJ) 6nd,`za,q,BĆ&P+MbSa$6vi%boIŒS5oGbj8D@4M,i]j ^_۪UD1DA)B{3ʤ,}v t8*o u!SK Yy5ѧk]؁w8F ${HzӲ7|L^zʣ"y`X81lVyMEjDae( N՞#bo3,&jsæɛashn}Ѧa!-1 O0P0w y&ݦTϨh";_rV.n[]OM3B&B"ϲafR)Ͻ*»3L;WWydE\1T;s6Q/|^rөn^rmyikSn*VmM=V˶5PPLSШ9:'5*ݗp*Hd$?LidrYz2I}e=r5m/gu2;M/cu JiHY;^{_\ϧ)\\jsȢE X%P%*F |Y~I6H.+iw+U@wYZWZ2BS(|xy%LL\2՝KYagi4hр$JB\ɸ \麰7E;R,zb!!G'5apKeJHY$M"wW X!j-M1X#Bc|cITd]*IRٮ~Ll%j32Uy];]Y5%Zø:AъeSFZsV9P4u92>X80HWXF=XnOj Mɠi{LB#]+T3sCtz5U-iǡ*a tED%io&}fSR\*W8/^ɹMc>agH\6QoHnr2w5ҳt<5唒NJl( "si&-\!8w+ULx}4 [?Bt8ΙLv}]h:lDfJ[J*dSWQ.ԙj侚j^3П\}B- bj&%.Rpsz%Fn ԂK2~f]J$\S(3.S4[=L i;*]kP2)n /liGrs̮ʨH󭻴pgٵq6mL;Z~[>SkywLw\n<<]:K%oa~).yZ7o_9'93 ΀wJJ>"K`@[דgY$ $!qwPIȠْ9+6`ׄ͠Zcs½{ (/nUҜqLWA{SX)XJjsmjw懴"m4㨔€$ìr7_  Աkc҅ SSo u|x{lTm_#C)2 wwy{!z!gɹ#8Vk[7įj s:Y+d'Pe+sK>c;йo{H瓞}|P#mo3璴&nL ȋDD1^Uم`X7uX.np.V#7eeY6CSޑy.ӞgKV6,CLZoKO>emS#Ybc6_uiB)MA f|$IA1F)D!hww'1 I;YA! l͹T7;o h6fN@G jk:z]%L}#cOg5 #؊<9/~U|;I'> T@U#J-tHՑ1lvՆZfw=eee#e TE\P\^ 4hwD&;=]=_O>FX- Js! ,D@U4s@N )3pyF0`E:meIQErf(B>lB-XQem]jI˘S6DݓW.4P!1Ey]j4ͨi~ pop&*&@%cs[ ,cr+6`_]Q/j(kc{#U #tRpr,ZaTAdV=195׬5!,#K[}['#*8sNn^OvO(D=-z@G>t_FsA֣JX%i) ~ Ov=#+ٻF$Wf#j@OmؙuϼhUa] ),Rd(2uH"Y"/: ?5o.مa/|E18e/UV.v/1M#={C^`HC7^gދĄf wG/YDsB tKt.i1X,@TGN8%QIRUȧ;*%LZ^`9 7Wf< sGsgiG)7("Pk#d6v^H8S nJq%O3:V ;˭KG e )g 6#.'Q|6F6 dX Fɝv[^NYc>~xr/iopl+&Ҧ|n/8Ekl/qൈqď i_Wl*Luᒼ_wCsYBY||UI~c+ g=<\DaPo??\O%' =zs'o}#;Os~xy| GᏥ9Son$Jy W3oM5 KrF‡OeƢힲ˓"ׯE_^ޕ^]A\"2BY^ߦA#;3!6^_ ?G+l}c9_>-0ٺmJD2wzTox9_r"v"䭑 ,+}(up^)oWՏ(gBQ}8cCSY!!!Otv*2̖&=s:_= J=S=+Uatqb5gHFzBy;6NE!Xx3R3 LL/gfZwI]ÜSB;zѩ<.JV]nn nUf;mQ&Kbp{FC&>$ܑ t0&v5ɒ'Ut*]X˜(yV$SXS?.85/̰(5ګҊJF].@C?N? ȥW n1>!v<й:~z6!hasE{nFܪ'0к||F\KBI,aރժ~zbJO@L%$\ߦ A4w܊gaBCR$- >K?o/lؤnwac@edzLx&dv_=>?v쭒&^`oG_%\YG=.몧֊TQkuRfǰ_pmJuU]]]ҰF Rri".-ΐhX +R~޹^ܱ _lݥ6.=B'ېk JtL :2v-\ٰ%*}'g!>=\mxGS$>;~Kܽc絊 `[P\vZm=&{IY jSMgg*׈)+[ ,0Ɯv}ӿȠylǛE+Fitwe Ҟ ش _vJPa0*|䑕Wh! D#W$AyhE]@t+Z3Ʉ6GkEf_HZidV*^RobC )L?0,n[ԑZRq, S&p3Ϊr4HpɄoq<Ùf3C%SFZ% Fb<qk)cE'r(``e'rXJN6g ;XB@DQ,438X[S 8 +e" #15 ynB?SY'>|DBHI/+'~ՙX$LeE{A?ٙ=ha}M=.0mѕZ\%qMf3p~bkrOӸU ;?NnǙƾQ3cǛ`y-h2u*W)}qqƘ܇ ]6}?~+rMy]hPcf=3SCmKU67nulQZmn}bb:uQet;uJTg[톦nMX7mAW਽È;^=b-$)>t1C\ j2hj;9`s RnaCZt< S9tsD$w!4đd/kihb{ 'zNgN8L;kTvY*Dp5AbZS4^5 m%*,O&FK,ɺ_uuU{W:]F@珗W5Ci[6¿4{~r?/B(t븚 ^~.@0_(?`n$p9H,d#L Cupi'(F`8 *H 7 C)< k3ebwNFr2xby!I[tOÝ..EJXux9HND 7Vsd52FOcm N+VN[_w.u5IZ7v0I{!6j%=Gz] -VtگF*0{0:xwZ+hRDwq}]4 1sO}nsaR3E;\L *ܸ*ܐc8/eǫz^cnm5ǾEG^|ȴBr=pc =SfN]18x~sk6+Quxe ~iZ6t #N 0dxIF2g,q%T*]r7^vwbK&.ܗ]ŎP)KҒ dT/bx,K,U%N&D"OFo@AR2`[^s9Gwtӊ> :e 3@j)V\./ayRNU zRD^%|ڎ_x'/զq_g,D+CMaRsg3ـn\Z۩aRJ1/\t1E t?J!1MAbCcZ#?K.Z]PЀpX˸Z;K+; *o<!W6o!*Oy64'!f)`]{G_}?3YXR:X,"8lnіkhc3ׅ#6b/8d]"4nQV&Hfv=Ymt垯7edQ~c=dNQV48$sŮcKsg6Wfm/YKoY+7"n֙[)9S>mTQyL6rڦ :V C%gR Y6YA6H"ĔC#~ )xٟ#9h8&/k9W%ҩrE;A wk? БO%GLGT&@WnVݥ$oL,ajf#X'j~`4iM'XYP| 8ZFfxP^iEի}6Qs释K?\ ፴CUXRd~t$GGcP Xp5KM5S"]#F2I|t@ Ϝ "֚_׎s ( ä9GDoB fZS8dK,ذ[/%&k=ЃL.1YDTH^0Ws2Zr^d:&#-]MP!xpùwJc&mm ( q4<R_t.v= 9Ig}tiݎ744i #Sl3|?m3΍BhnB8>.Ǐj髛 'ջ(~>ctmn=`S lU3Z?yA.}kys>w&&Js U$6bB~ވ̍r ,Ϛ8.]+5'Q+9{踏J IE6T-j!U-qR'(mGϽ442XNKObwQ1گA*y{iiii.%o+28$Rbl!u> & qR:m)5wf~[ДaxNk* b!xkU@aS RC;q?@ly>+ MB!1Bĭt&YJtz6>Pr[)3gsȕ770ǟuĔbʉߗ u\Ѫ- ReBׁ{FnGk DM r.̮LD2 $1NQ#F}!)2!F8nM5S53 T Vuy,QQ`\^'  uqNi|+\J-50A3}>Xxħ eڤG!ztЫ8t4Js`dիfjK}ѕcTp| Qi) a"%噆K%ǫ$Qn=N@*4+YmbLqDi1~܆oRm}լ݌9Q?:V|O=7)3v;yupGլ8rf_*tĈcsmmIpLQ/=r\ Sc^ZGvO306pxu3zlm]J'G嬇W-?z|ě{jKHWjn;r1itʰը4y;\]v?#Qr}'U zl0hT@(dh(&)}o;?=gH= ;N![.R?[h)>^j=T{+….{WKz;(Zjͨju?wѕkFÔ~|j)03!h y&ZdS~~?VA_7zO-J:ջg+%B^ؔfp cvޭ)F]g;j+#m y&ڱ)iSe ^!A^yA4bR z+5T|E,dp_K}ѕ:qg'\5~>+aLjJJ!3iڑJjHO+k(j7W?Y1Z)<+E1MT.7?4%=>՟Vƺpdmviw5$l>]==Kp > *NR?}ϮN>7oOCXօ}H) mD-b3{٨H)!]Eohk#,ʳRAhڙq/.ۋ,jpcڒ ש> q{Q0G:a%ol*ଞ|s&swuąc/ ^[s;GjY%[YTD|>E'x!"ӻg[pYGr*h xZ{V4w6RNܩ!*RW 5Be%2Ent)]8|=KF>Oy+L܎y^Wnm(`3r`%zJw+;[W -RM\Ͽ02ѭnR,)b6I*"#7LP;+N oز#uwG"%ns~8Ni%#ڜrH#!}JX@\mx8R5&TKQ*wX5-)ɰŻ3_CF!!8*zJtWdܘ+ .<]kH $S`62=r Myʳtlys->ز };H#QxS$EZ¾e{38Hk$C#}ʕ,GZjo/Vph[~yST#Jw= UACdMI_~䋍R_^jBpmf?uEUۺr_qF9#}lKj5i__9?meрV4iO< yS(?+jm\|#!0De)D4><8N-5OiS" u850A> %/n f찄:^VWB^tZO{WGpk@_I^&)ёh.Q./ed5yQ$ѫ`JHRS _idڰFw_}~"!x(֢/T׻IY"@dޕ e#]͸]\+b(I̻ʊp'y+< HR=9I 5bLwxff,/Iˉ#63Ap9tvY#I`= -U5< !^/]޺zA7՗V^AV砒,n&$#R+z1IZa iZil^Y'4ik΃kj)֞8^0 TH& qؚX사d~|{3Bc!)#-Π2VAK[B@O&Њr>׍vrԄ\O:NGR%OF7G[ꄰKC1$QQ#P`(u\Y 0]b5t7rscCYEhvnw0SR2R" Mjɘ yr12QU?e.F֙ɱ0:!Uɏ<[jYA04 gD$\0xQ'M(~4 x/_֒A~5(J_MS7M4zvZVs@J*-xQqu!{__}};A3 N8et7s9=! T`ic6,KWOW/lZi5̵}DDI[!¡~ÀUj"2+ׂHzԡMw`k!*03$B*-qxb*wa8lxAOjAⵈcV? BY#P[mƏ;vJY*JXul@RpA6:]5.n/-jU7[ 6%i.~}M ݾk;gUKY;P`BJGQyijOmtWo?+{sp Eʙ; Q)eu%6 > 2+LD/J ]|@ i hCZPS^(i&j8Ϙ`L:Xae¾ x3AIiŖOP LJ?! J&El\@x}s@?{WFJ^zT zX,nx/2Ir(oP#K'<*ˀ`|`o_bކzzX8{__/ӱ|~~f W:[./ιnv}#kZRH]RLgGW#2#& Ikq:wxʘ/'bCFݞ̖ņsvqgdd]I54* 4euu{9s+9KIMD9 Dp~80=]`iKc0{‰ҒPkUxk!liB# wxBhtMpN;'Ջs2<3?./not<>?<5vW9ȑTx~˝}Lh֗Ř4/P"x쉃o־z.ak>^4}^KO- E`WT|9G#_RWy~BͳϚ32ސJ!1B:v"tM*n=nb@leXP[Ŭfa2'&@J|C 9{'S:^>qx+ރr,Rl_>9oذ|ɛ<5+%4 Ro2Պ6N5<ˌC`dB$k[wss?DZ)"Q5|SsH\_|XuX齹w?~>σN@=rÂߛjFTL2Te|oLZw*u^ sZIg]a|LlYg>ո $ڣEPBf ($ ;`v<g]_9q>=ٝV @JF٥5]G-uV9B%"TVprްɒ {e'/"+jkeL9ƆY$1r0N*ďk-Z3D=3+L' WE:@6ՒFk/>p蹤[oۆ+dv]U(M4*@ܪsu9ڪ fDY)%a@VEUH?qK95J0wRXfN "U$\]߷a9P'!5k )[%6&J<#u# G?awzt V8 s=% W.;Oxtc)Kx/ds-7HTׅzpF>[Ϯ͒kюޖ T#-M@ X8"AdM#͂hD'jyˑw^~vފBb,нlF\3Km$RSNKm/j>\ٻ B[ggazm HWmof*k BTxr^k%P\jdPP* CװֳrFFx1=Diy*Df5Z HKZY  ~BR/!vswI ?^\py+1|槳-tnޜǧk#&I ~P-v[5~;>&;Jo1&5MgpEf&Ÿx쉃o־z.nk>^4KNjx_-=h -Ѝ]\Gh*F)(I?jlHԜAلhMӇPƅ c#Zga4|3#rp^K;s}p -PҝA}{B_%ù|w{=]%0 n]^{O??Xlo{ bV톯Bzt=G!"e%2r:H)/_@ǻ0~a~C7ϋ_o߽YurD5rȗ<|??nCh( ?+rAq?>03 z{7_gWpZ;ԭR(hwN4ހ`(GK\3mo hAE«rDE01в"A' mvWabIM2(o[cn4SâepVRދT->ݒKAv>S(4ۧtC.΋Lj7De}Whjό "JXO+F%QȌޙśvqR3|?#}[_, KyR,Ưӛ_<0,;Ƴh|gaWlCEz\ Z1B$s}Ir!#H_VO.O1ɿ@0893hޯhn>L#{|g"gJ3^>fzxSi 2?; s f1bV* |'N;eCuHa au"$jXmlͣir~{$"Oz2(ږ2/,w_)y'h'WXBVJv9γר G%ֳk19A=̜0EAAcom% h%7WɭV#iO_BB MSas86<7[zTwil B$Zo~i6ɧ#LYUcvs!N`RUBs^@xe6FuM+}Y׸Xŏ =*Y;X$;^Y[X2tčJ۫EC|,6 C^4@>x59>("Z廖MȘZJ|Nba(KӣIaC}I(5&Pߧ7cs r 睚S-b{A4*zig`|**ΥtZ56LՄhɁVeAZ #y\Ӵ iANӂ6d4AI<1c4.xSB}ThVn)4=@kFEd{*^ z_v[ݴK93{.Ґ}k5=lbXEiƺ (%}@ ȗ RT?ΖBb1[<.V?5O>2FN/VyTl<_Wy<+cy>{wWŻ%$~~wkq^MWH&79@9פ4pc# */; a s ʍ-I((we|7|p;Oo'Ǽ܆PDV 1h&FO#Ť$16\vjD$L;Θ L;MfJi 5tMҷ205 bMV#vSܠ$WS$l: z͢}U1eEhm 5Rv!6 \`:s%a8X.7-I: ^|EI ᧉ 2 l8('pRwOѼRnt+ QEjy WJFaz@ل 4 l 2s z@W|iyXGAs#;+QѵDc1LJWKjU{IFY!P(&JPSFIt\K#.QHR(!ʷ0 l.9' ')2\veBR417 v2:# K]6;!7]q%ff`J< FߌkK)#q Èѫ)Ѣht2e&t'ۛrN 3k*7FP,2Kp)OZBV;yCrMCKb)pE%HBCf=6>%=S\|yGs8Sx*DaټI.-ݒǻA~x_/S,$z~,kgjבǧ %9?=}<$ŲٯW7;!C1t<3u_{~V;rJ3#o>9QsHk |uv2b?Q1V{\5;]c/_|[a`_n{j}י-э>Qqlt|7LqYů1ƴӨБsDT5'MFz1T))9/opEE\jnMA:?4 XFFUIU8ەH ~ g3WJp=~ޘH(t7Ӣl~5K G> ] sjq^iD=<$sj88(GDOqpa&Ar`A9!aJOy6=Z偎ŧ'IJ#ԀytXyu<"3@C'+i&^zx^T?@+#l[\==iߍ8'$%c{mk@;,{M$ 6sci{xr2LiS7_|]ßB#ˌ_vDmO11_$j☽Lk>U, h¡Ke*L.nl_:|\.Wz`\Rto;,ikG)1M]dA)wN0 qP3nEe:}koP"_׭ 0$Q^q(qV(TnOUyH q#JxHt9)6=:h'Q[S9&h$nJt2 {L3E[)6.p=$dv,`Y:9!2k)FYAE}Pmp$^7V%RϥQ8eUϓ]evluVJ"W Tleo,hȊpGc7Ε+i$ge4UOiAf:>齺J=]eVBep  Z4AKwZ_AȥI<7tp=rjlzgwίۋ qwF8ڎ$nsFowFSzS>etp]6ў(nt7Af 2 Y[;e2үeQd9,+I|7qm?b4$xWuK[$;ހ$ 0z,D7%sq=D6d';.=.fzd2$,_{:74u J OQg2@CpfT4&%2dâC +teAV\2*@~LaQU^PT7I0 O=F9C`B*-3o91܁-ZIi J m"^*pT2w#aù]7~ uY*&r!0WB Ȕ (qBP8!X 7R2ly$'j6EifAhkzT4G1+7: S'i6QhXtU!&@,ZqpY_1dJ(9bkPϪ}li]4BcMV0ݳ p\rgreƭt8/TEѵ^mkW:kTѠ,7)ef:vnG( 2Eq8J! Hlsc? 3 2K0}s Vڃi1ƦUnb};9RL07ݒI- ?oZD$$srNbX0Z69&r1BVȭ(+NXlOȓ7`2+fg?g!яf-CЏY> d'W̎sz\rFvw! ~xC%X|, sO+;S??>}R?=}>3׉Ų^U'^݄ngPegu&u_a=M~iw^^iG|R3^u^rHw% /!E&lQf?oq#4J!r\">MewR-{w"׼ȫQ)39_WXS-\+S,7Fezly]hT 0Ls[Z+' 7Y hi+߮E+-Ufiz(%\w[ˋiܬ|`()[Tjdw=3*Me0NR9٪tF Qq091*XrUGj$( ЦڴFo[m"  X ])8~S)](G^Di Tgż"V-(s+*XiKQ;eu)ћ%Üu.z3 aB0 +/T#; Brw 5&8!rsAFjLl{k@ )B̜^26XD 洛4 Ji'tRfKx5 . ~[xew~u=m8!mB}Oը߉ך1#`lH֚g5ʅ1ضDc}_;z"J)Uc7lAw߰9Ԝw#mHt;_O?o2}X 0El9#d|)"]Q@)fb8og],wQ&$!)I~G3K򁥅~CKEӣcIn$"wA;}z.`@i#zJQE$^BW*=;d( f]jƐto( Bp3#0J ztW#;y>QXRL|{-yli-fT mu8fʋRK+.fz yM6^{sLܕaZЫ ^z^J5JC2SC3+C]?o=q4,Etrn&yG#7tVS?0r]^7"htdzaoL_t}C~myWh\LgѴi~ߘ{ $wҥreOuiwkb5)PΏ$!G.dJy>gڭ+JD3h=1F(vݺE(n}HȑL NXϭO>+Æg\c13jAEa\Vw }5Ռ~RHinL9tCXu7nvSNP'\VJO~^ Y6X|~e<0bw`7T&% Ci׽ wJ+\O >4< EQCH3#9Rddl E %OS[('[i#5[a>OVP{HPhz0؞ܟTzC'#g%$dܷFik?B `8lVH9Z^U̔XQ_YK0$he|Wvb&Jy!|L$Vm諒/(aP 3,J\%em2606@!3&p%HU"jFrN0ORQl אQ4#:0.úƟbGƠ9x{c;.nMS'JЬ@Ѽn\`8mĖXEcR[]*ΩT^ ED(Ȭ眔Bd-`9LrH(0P[VHQ䠰ea &cTTF=S[\6mpū(!0J"eD+ʆVt{^TXrb`JC<}\1n/:Jգ`{/:J17hEYGfiãF ,!4pk jkӐЫW~]/g9DƷ=q&n޲:ޢ7M]OSx$VJ\Cu@f"-(dJ"=Wͧ3T* QE,FMLtͬ[~0܅@عzT鱱_ I_N2GYsCW+J;Z "nTǭ%rb4!UT9e`>M&ސ?LtXE׷{QHFXI%RZWyVxRBw(T OO:l;@t@d)f+Bk0؜>jk7ƣyCl @|2lxd/giʉdAEflfSUaf7՝C,x<GPb@u4'{/{}^WLIXu8ea,WN5Q(J>|}]3Cgg"j?3&BtB ?UwDo;Wv:za9~fﮯL:&9y&GI<2 ڙ|\sS B1x{z1e,iZD&|kX:3Lku"5EVL*C#TL0"s&"FB!)CZڏQ˲ә/A5S<%ϻ8V)E!FEMP1!8 _ޝ>:F2SXjZFYwH Aֶ^1J)h%7sW*_5d`@K}7z|;3PʥœkobJމS5.&DKt%.;jHB0#j%q kdvrlY*H% )ۊ={h]UcCڰy+Z Ppt7t?@ 2f;( O:@ ce4pD.M_n,%>Gi0<2j9Q97 =ot̩ɚ`랹Gm(La*zԌfr+TT_.Z'4vWMdjMLv kܱ JȣHcKٷͽ.ZRjX}7dѧOG[d=w3& h%UCRUX@$gLIVX,0FպfG-]eTQ4-/zZa܏]J@CGuhJ!8(8{)b te}BLjV!cU-aq].?LJPO,:^98;4>1kO~Rm\E}W *]p6X_޸8(<7oLx;!ZvqxF gM~Z6)"+M߃NVy; 4z.P8]´K<- H@N926-1qYf]Ee<$Ҩ6XSzNPu0`pi}TI PSMw0W/cs |a%Iw?)P-Z|d4PR$G:jW * ~,LQXq !Kr<͹{VTH(TDr=!=aCO>P0 Ey i9)ꍬ東x!%lW>X8ń6R_s}y:UpTS`S|Xs$<(T]D{w<6ڀzK~;wyMaTPqQgw<&OCE\=u:?U zE{A&PR/Ǡ㜑!^^y CG$DջT׆?=|'Gkb9gZ"!2ŗ㰘 { 2d(e,e7w~ŖlZk#FƖQUX*VAc_('lI_IEȷ%tbER a+14N1*Oмjt+⣘'KuD`,T{>8yN^έ%WYy`Lckql ~޴}w>\ LߌtwoU{-IJ@x; 9Wzg|/mO4Ҡ:C:1[ލ(1fSL5 rKд˛A/!"-lzwKՂ$3lˆ l1mY7T;d $VUIa Q\4Ҳ -;DlVjmڲs|V7gwt$w;,Æˁpnx&ÃTB`G?38o{!-a|ݧ/Q:y+<1K4.f(euͭ,zw6=%&-6TMT!L/yAokAU6N_Ξ/q-G6~u N18j8q%;IXqt`z:lЁM)qyu\NW!\Qݚr MJdǥ,TҘǐ {}sK 6VSx2F˙b&_">,W(v:22o<|#dcp)&R4ƤNGxwJ5?dh)&Iy]IszIs ꧼu vm/u:=Guoa.ӡkݘ55Sp_i[.j%Qu;tݳX`͡ġ1p0d+Ij C^; ^ZpTr!C()޴qpB0ry 9Sc9" ^}Q[|a% 3lW娓8G-P'r !Eb9()T W;;]kê4>8@'\:r$5X-ʩ+ B[yנpРm0NJDfJ>1('[eIKԷ)%*?O=7I/`z9iX\|h^(QAkL9145ʒ hDMSǃ2$rJ5v]Sgq]c9KN(Vl5Ёs_8ɩv s8 (([o/x 9vldnV,ޣX8M?*$ShX mA5- rCeV#W尸0t*Ib}fF>VWbOwO͗t~3 Z{T &_zwswyYU&jxw6*&vS@@<;DyLڸ ѵeUl}d?4pm$;ݮ~1ⰶ=T!aσ1p@Ta.n{∈q SpCӸnTJt/Sd벷1Gؘ"Ihs ɀKt>I. r4ك_h5.[oi@PuR3IVV'm5Ds` b*=l4؄jh{U_vP>Nsi?GgܢՄ~7"?li*U68P(Lb4< %?jxыj7P뻏^݄R>2u?OLo<ZeDҽφRFWD0VI&30,YtVJFQA5_"巳߫Rz:@]rD=@Ҙ'Tα27gj@b`:Q*ҮsCq.0muID=kx E(_DRg# )Sܨma`UBљSx޶6 #@ռ`[\|uK]LG-.7:TfI B3ò'1;;?5CrP:$J ųRmSmy%p%c$Lۏa6)~Ko|iWTLsqB\ᰙ΀ysYCqَw٨;t<,$cw(w8Յؽѳ;;zyv_O1!\M/8->,KOyBkNR ܽx52@DuEx d_Q"|{60axO~ouD+틨+:k˧m}}|;5RT:##SUwE%=6ksT͵ډ60R a@wJ%ÀaTK&.DŽ0>c NIm_u:%O;R{J/=HOqb,Kp:/߇OpAF4,w#I\y5M?bp*bpZ= ی"4K|c|ihCIE FqisՄj+ H>)S ba6C*&SI]bs=_b%oCykXo2_~K"~[{֓ąoF志gʸ7y A0#$-{'it}j{bg(vBƓ"+QT߆1PI-epmE33z-)ZK-*IQ MEDhşPV)vbgNO}i   H/xo˵!,2wZob}FIT1cn w^#^-`"/s3C |(DvYz^H<6o{] 5{^4+F\?{ֶsObjDK RhIIկ<Eأ#=g]$swKDݹ٦02Z (p/=8nMbE6i0&@Usv&feydii>_W(p]褀Q7]%ͿK#U(B]Q%aM(4LsAq- IgE_'3O_NsNyŽ4ׄ0-NB"pyf8<[$/> obKC&PH=QNvVh_eFA#Cy-Ĉ`Z+//F?QڑB(ǵ)R.Ѵ#:hM{"M:=S>[#͞:C~2$"7N]}zG#yS8t&3ds2eucM WO]Q}G."YY@w1[߆*1jLT&{{ڙ?/N~vfC."iӐNcM}X{ޱ!\ٰo\[x Xb2ȸ̝Odp6d|Azl854jQ=kdI#O5 ~51g8|8A5(_։(Z JС$|$rAƄ1{5=:Pٶ{@H kV/Ld9Q<_. 6/=`(!>0p‹By$&XX W.!C()޴`HJSUɰ(OL1'hU-ap )Oϲsy?45D"j˺J\"8&癔2dZ29gyar &8Z2fh$ sGLzi{4ѝly{8VrY"σyLp"MP(, Js-4"8.8gH&'|%o5FH*r5ud"l1ø@/͕iKGl" u k FqJ}??'hz}S<ׄrih,*Ff_ܷbd,|,#)AGdqqJzDUQCj7D M ^/NG’SYpFfrCs  5fx B۟CHT?j篒(:tr$qMM"b#"Nh&Ѯ\÷s^/H"AӂCxQ4 <f^\6yC*{gO'rzhg`b5Uj<7e~=C%z_w2`wg{X]8?IیR2"jqhKķRVJ'4w k+&kF1gtSUx(@) FRPTϥ*WRBz0.ʴ OA.~F;SwaeŊVLvm /:`~cL&v4i'U;KGRD/OWKKEqTO7 t? b$XN JBu9A^;pZ~5!PmֳOwv{7EM0L$ۧV"0q4@H%9Ȃ0@xs*ȥ6HEW @Y)iS"\3[0n{IeE (!|,1sׁ(,"`Z(X3iS8IP[|k4kxE7#lk1^EmnQ 'pʠ:GAroߤ?(RB'x]_HLY"+ i &@AzrJ ;¹0=tJ u^ PYzu4y]!: [T*wNAK6i)5I5Zi_{{'s-èR= HfttvC0jʖК*׃Pz^Narrp$.Z/L5Hɿ֓0xm%->vg ZdܧU:j2?/l|3ttW:|Bpz2IәxtM+hqo{gO?2jR2Eyk XgU&+Z1t@t@WM㷣U||kD+,=-l?$KGXӂu߽n ;O6$L {>6֠K LpBz@%-dF-L`<3qo[@vR^6ZD_.}z"o)B>}'q|+coń@!Ԇ.]O:-S _2Õ>u.4Ӫs5yp:xğ2ˆ ntက)54ox " #:HS%owRsʝ\[]nMtחl̶VZa hٝު11$zD(אWJ Sn{\V݊&۬M$+bcTإPdUgH%4E/ 4 o22_l.U[frXпz݈E^yM/姣. ϲ*mХ8뱘g4[^K~ Pm݅TZ΄^,|쫧ŌYLNR7DǾ觺<0H)IMjq©RLv:I+*I[~=]=˧Dk={͚? ރpW+-kÈ2f/7' fĿ'v8^~u(-TVK2ab؆S]U8|qK>7ODnRpׄ D[#s Y!-dx_|RwF,@|x;OS Xn!r4Xbm Q }Z-TUʺ>Pźq*}F}nmUDk\}!/HZz c`q 4zu%3ɖ  >̙8S Xv')CAPL2N#:eM`dPu;L3%/@f=-"/,Rj\1[JLFSO1U:Eo(ʉI 9'Y MTι3#%JsmfKЁJ*:wn4qtoO00&yK΄x*gQZzN,Rd`04z=Jߣ&~G>C K%A.Z=PUE:J$eCALSBN,M/o͠|6Afg$"+uK9>zKI`qzR`޷rxQӶx!~/#7s$5LN1>ra#~d' FJJVr!G,:$`*\ֻg@R!ps+^Ќ"9p G6%Wh&g,#~+ /e񬢦B&azg-g)oT j;LNS6tr&P1>Gp3?/Lr*}!\tY|r 荶6}SO23F5?HOb5[r=h[D_$Tt5JN Qa;*cGq›قq WPBxb4ƹ@laClaŕk&f0jJΟ:keuZJIqS9XhcqXh#B݆׎Hrbj%biNԻ.ݽl"!dondixO̐sf 癶fQlT` iZ̭<'#=O`V8HMIb'O;FpΈ ޲o:$.Ne4d^{S$ȆIsT8YL &S ISJk= Tz.`2^.}F$WP̢ JeB.:\4 #C H˫ H*pav.>$q= KNiY.OꔵFL8ЙVg$3e,^^*4|TsӾIF~W[J"^cǛkwsJJL0{C3MpE28f7єT^hCR_SIT|2|8Fn8T|d оC\25Xܲg[vrjMRZ R.Ji:l>} H03hXD6֗%@EBl G? i]cKY=屑pu$]c˭YVrh#Ѹ$Y)떕K/gL"cj/ Tj޹n@QG3A*QssLq␦1ǒ8=o>4Y(;GPlOq7y|-/,FZLj̗[Ǘbb u|K/lͿ߿^7NLnXAr̙ W2^<'c!pje ƒ*g&"n-9uP XE./bEixŝdꮢ$!& 8vNx o\<DmI+8x]o\ǑW~ 6#@k/1>IZ$g<3߷͐|s9S (iտC$ȨCHsn) Jly4*TL*_d%<%hE .a喢F-_d>!@Y9codzU4|%4|kZj`g>LZ} }"l7=f )ZYGqO=jX:C-ݸ3t>%:`G|Uw3;7:7/7:n/:n]F/x6ZG,nSaFˎ -N !fOma:z[tå{ 4 + Rzkv[d n G?7n{ËħDZ'lGƧmǯҶZmx yQ@>;k[D퀀I%8bpWSLM)yjchC(j%}Veco9zz r2mG)IzyKB{wFޝ걟ջVbKEuT@G"2 A$|\Sd4O^jGpUN;D-먋G9}׋[;p<_N2-e#%VQTSu-TZ=PIr2A*((8rX7joFŚIs|t^'w|T' VG M>c4ɾzoOo4k&۝'1$Oy?9x;8?BOp:;yv m([q>ŕ=œckj.ͼgo߁6(e3}g VN7EL5G^n@[_ JD;h9VM_j!$w.+2 eQF)/)d lSn*򴭐Q0)!I#If6J@ $b2~.?|ڸ`Z?^oj[ox a"j)ܠ=}{;cV !o=7v:jqNRzhշwwx}&pmYDt?IdÍf8F)~#*PӄN\iZ`d}n."h2egqaY򃧱]==Noy[8|CL HǭTT .9-F<5  8JQS('{P#Bh7q SgUV &pAN?j\NDqIݝjлH(@.)\r3ftĆFPd #ZT ;m84xwM|c*'c,% =`jkzʚD$Ky1t3=FYt.ܢZwā`t#B% 8)njQH8#D(m`V&ODB8stp*ZC(I]B3lw2R/'kz):JŋmÀr0kfIJ+zlU : Qj$4\UߓO":obɶxF=,PboG8ͫwT Onz(>RK )n,tPϘUГ v/t\qAJmATm:`D:Nq!Q>Z ^n|(.?raUӅBϲLͼ'kJuUv)76WK]>WPdZfWX%mC=Ibu)Ml1ӹ<-!ǡu=WB%*xuMz}$aP9\UjU\lr@pft6\Sqmϼޜ3#{{h1гV1o\~( ИuhRxPP)pP-:.4]ԥH5 yT鏰D nLE*v-FvG)qoF97M<ɳD򫵧?1c.|-zw:^Ϯt`=fhsʗ'QSw1Iv5yE^y]]-u%zNEǴKAjGwB/׻/_wQL@m8*TȔЊG=x庽sq%ts>T^Ce*{w[RX1i\TϏ.SHI/$䬦^ҡMv%FF*h&YhDp(L>z7Rs'ǷQ"kl=1kFOtt[6-Fr] -`$H$E4܅|ML(\qB@ zPYҁv[k2]vO"M*xK%/7DjBed)ke97@''R){%+T:"bPmg2OSwm*6BB~"Gd#X-vLG.eO zU./)b}ZK9EL:zs]F[_ JD;h"UlZ݆߹v)zIde[_J"+y–$JR=i?62*o b;CT*0U\9Ѩ×ǓBxԩ _Z@kEbg_&QCr>oKLeDFkEgfK'V,>B҃p s9g:32 (SѤhONzPѿd %$$KǙM%FAFY4r9G@]3gg.~^G[PւhFP.KYH}EKz7Ǩ6u9y fm[\X? ^Mķqz3/^g;=_p ӫlڐ3F=Bf]N΃^s{?:GSu}wwwٗ %Csף;\) Qꋺ&{SV_FV U=Mߛ. [{'[r6סo! R-1*?gYTnoǓeȗXֈ'j㭽sJVQoY| 'Kr$2d>FzFVM#α0ߐ '>| N/n a%4bŒռ,QQ`,> ,i=,yJ9O b[`K~/&3hcwcƞH_gmT/6pq@6&i6&7pձǪ|DV½GsX4rt$}Fbk͜ e.tw"cn&t秵\Nej1?LϾ{G.gTt.=UB3>]VN]^s%?G'؃3?"]JMFYbtm`Ez=<1qr>ol&3;Y&68ViBB+a$!394Y%Cz#۾n[=GxdD\_Y$7xFbxs+o :=B Aψ&6siHs{qɷ1M"R%㿸y L7}" w||B\O KNIQ#qݕڒ9BZs\v qg[AX>fp-myayea2Q-A(p6 o;fn }Eiй7r ™(r/YTvHA=`ݧxe&\Nw̯^_)S+Eљ1HV5R*/²ҦZ#~f 8{.@4G& 7aLWܓޣ%!؜.U! + hH*$✼F*"' ]/m\L-}ZG05 -185 jS-:$ol@ $}k~ kCR0hLMu\GW4=O@hއ;orΆ?m t8> y?c)Q)N2;s(7bc@Ȉ1TrYE wzP; =ll>~Q̴|h۵i0W?x_cdjY=+cWo&0p $UE "5VhO<_ 9k64Q3'Sxʫ:$!쏋Jv00_߾'$bǁVpو@xqpp`W1XW6V~Mb-ɨ(˓ ORaN>K&91 )N8Ey{_?VF%VOsq8B4?|X|VNj_rR2'w~+ez-|mhy$wp= 8l̮֔F JD52Ј܍.m.9NlOk*UTNU#4 JJs/'K-xXlưڗY~hBF}vhSZĜ _,Cvΐ=زd૥n+ pýWe!;BU;Ωʁ0lKk%"vsj !_B[avxVZ,z) +lsRSe"mtBj NYO\rx @VU$3M|:yޱ0 gS@zS,0rY8;.[. $ʝ6gِ2搝BȎ`uƘ !fyeD\N(+ncv ivJw}DH2)HJp"'oCN=f?H`xv"sp0)B@T#ݸZ9q>U aALz_:V9r)М>Lk2$Q$8^gezB_M:&:RZLfjVi6=\/``xVuOBZICj,2gYJmI1:eH#}" 7KT7jy|\4|ZT ):kSۧ'f69kw )hA"GFR7 "} 3wv@0y6OM~5ݲ7dw\oOhln4Sw=Z=Y-!T32-G~9? bm)U*< `~yye oEhIDW nc {\lq7ml3'tF{U^Lj+er^O>;ZYkٝP=so8nԻgҍ}J/إ#bo3v ,o3vzxUyOd=$UG^bOs0?F[ J.\8Љ!zg?v?Љ+gO ukJYuwYCqþR1hJ&UQ=v8j ~Vjiˣ[cV_Ù^5]Tjv06TMG i,l+|fkNS)D{]$9Bjn1qPծC +^} b_a4K}kڐGg.Zc3D82A_O ^H^EtJ3kJg@5lnP+r}-&DDdz.KNnLd":E K?c@q5eJsNC3f]TNVQ!D50LommER[rn9bW- λ >L`ח+v06߃+LwXíu 'lbaQ^k=b3*r󍚷jFDeMcGi<!5|#p#Zڔ9g#g\&&)dDiE"2vJHf((`[VQ:1t},gT},>gS-iw2vR5ZԔIu&[OԖ ZXA'6ͦ:(!=ձDC5}zb/냧,0F 7X8Bm+QT\Bx`2ª ux)ƅ#4@_vZ?ȕ@~kh9eH6Wwu2w}*Okl+{u_'׳Umn!]ߜ$S{3Yob]'2IR*[u%" !ui{َYBNѻW\xfmwq/[ ?Z\b## 7=BvzQ%(P7?>Kgm7.Q}}}"BF gI9S  2+(n}ڕ˅g!v=~L|ը +K6F-wA?f6A,F훡! y&@.3Fv@d /#^#d*޼sO)&S䛲qM Rpg(o'.' \YQqmnQ7 Wd 5:otzEYY C0 |t[CWTƍ\qq%%Wvؼ̞_ϻYCU)PYyR+`}bTm22mkm- Ҭ:>g'ِ{/.>U:1^߆%f>(;2֏sAi)$ԃ;%zf>h)V.͕0bc.m(8|S򽑾Sܰgoyokx zSP0j{=QDt1YWqʢ285 գykrs+.04sٚk.ȝ%c{YͭC0tRvb5p,HQ1fb ΗVP1oW Ԕ4m3-?e]kXG=.46z'hNׇ[3cZbl5n`)03_-QpW:`&?:F/Y|5[3@5n}֙h_LT{_ba\ `*[?՗;TJ17T[ÇTr| **i&4r3 MiĽDvсg $ &PPJBzS8wjHS2{hz Ax%" =FGI H b) p3MАLH V(U'߶ :ltPGs[ $(H-mL.8F*B --6zdPTdcu.2/06HK :%SȞ5l&c썵YX EXy?D6p6u] \U sAO+x%wZd$0bGTQDFUJP+=D0 3Qsͽ_Q*V/vx"'.ȝBR 4$AvsfHJiW~(u>cܒD%]Τξ$f%RO ;C9.!íLs )ILJ!TFcL&7P4 "5.R;( ٿb0J-7L^vAN8CarC%3wwXI6>t*dDFeTRmPjh?Um2%4\A),e$3dF2HAp|+e 2)uȠҍA C ,F]Sv{$yJY,0J%Y=cZdRɢd]).,5Jc[ĆQZW6(Yc6d€FY֤ Ym) ~ ѥf6K,NF?Rr$kzYJd˹mޚcUoo}GAM'K 0{,YQtjMkJ,ǒY1+xtR`AzI'bAv] {C:)KA"#o۠,]Svne ¯eIngInd9IBaE! JSqhpY^ʹY"Qo\ :j~ .-x{hjrNbs}Y)\ y8RӜp bJB'&>foސ2F[fޜ$,\N>qd-5c.r/\6-JXse$?3^٨M3e>JcT#?cfm^HG9ž7wINJn΋P&4VhBKr[S%VD .UHhLIuAOztDDD}e-"f= kX ,' &Nȝ (}_cVO_θbe69O1Fu l7"eȬms(t\]ccvN> dڢ ۏbT<T]I|E*m=;tu5'{?(u4+wnNbG,pJ2PV6eYE︤? sg=~NTH1&e<)aQt7[xQHbY+*}Ԇ2y2>f*쇻@+%Ggvv+$%zXIJX&A 4Xe(|7d/v5):~G\?I5ŌN(Ѕ鯗4/:|?IW/#v^=@zKixRJ\.{~^tˊlZ4 ɴ^՞Z_Cg7vg>37%4 L#,1Y׆pq)WLISbM4i0`k,uL V+ Ջ@L/b)ZjZ%b.Y^n뻲%Z=Æ?q ?E Yu."  \\n\a8^J|LVݎJ}m{ n2RÁ!>(Ub=SaR3."/؛\f S@W>迯!P:{=߽]3"ƫVv#T}+;OÕYC7`qs / n`:[2k ѓh&ճ0Z鈌zpi6pU[ Aʆ u|{ UBݴX Wݺ~c5HJs)<6"֗H Jlb?xLxan ɮ'btelv:]e(p-HDxʣ4:K%vq"LoO,O!N,_pRb z (d򁫬LO.lPQ|M.YJK~)`6qo+P,:>>љ(ǬlA313q`Fq*%mKLr-k䙵E>@?˚=Idx#G?i;;8rygek wbZӝiˊW$7~1á<Uߌ?\/##B+{M}>BJ)n(-f%Ɋ]t Օ>V0Yȟ:.7E^ 6PQ 6"[!*qX $#pnF2ljq֘vi 6UdǸ&qVÉjBpwގt=z [)^wrQ0%zZ77A5T݆3֣[a #z+xE$`Q2XEqYO/`jTm%Y"sa3KY9ň˘PYA}{aב#@S^2V*_/Zu1^x6Ոn9*AuHQn~'B϶fw@!U(Rph%V/f4NJ )2t܂+j%͂h;[nd}/ǁ2 iH$k0>ޑ*B> ߋI<%>8jEɧZ~An'wyp ˦ uw9O OMsU~nqz|s >Kh>S39atBy354nvB)JehX)qr_#-t4n:]a ݇| ïeGߎ?Z!d{o;yϺtv7ω/ ɲ=J(!:aRD #QFL) X2ڎt/Uw<koYَ1RIy@oeRE{@R):4!ʲ/.쀇ҽ13w%吱b;fKO!U7N{KDP}{F=Pa90+BFAxӰrH{J8~P#'DRltz c*vդKZ h_g~tO?2PgE\V8 &x"I&gcCsr;{ڛ"'0uYrbMHשwl1%+̎p+0/T,,` X?k,C> E;FR*uI(=xlp"[U}S⛉7$w;o̫8 ,XqzQ JQE5' p؜[Ł!w4 ;t׍@dPckgNB)؞M|6<`Wq|j'׋mRE)fL5Dx9,1vF43tE;䠃q붃~PΤGD4iԼ˲OK ʧTȼ[TO,"R$eqR'D2Qm.?GcnLF=7J#LN4Ro0Ɏ1#ԯ)&1#mULz}ţx_(EydTmQȈNd/G3ISr-3vT6iRC`%%*C-HgHclxm zy2&̰6VƏ]Eq~8 PxcIw|Q.A}g; lo]goU%,x8 +4(Tς] '%lGǾJ6!!DMjsjP#1Qb<23oQhg)G* aӼ8ypDsvx2֊S*eb"1k5VqF`=~U m8Mka[뾕48U2+c O/qت€ TWͮ5ΰ@t)roV[yXV^`-8`sN+Le$К#WnۅafѯVZSh5'fkZ}^^<܄4BXoR_&.Z]`?+Z"az5yj~Xk]|ջonoB4OMD37C|QwxjoqR7@?W*YSh[=DrD\Q\aŒp,NWx)^sb(*|;:n&C ~`7"Wq01VC2?Ùpk/y` ^ >[9_s ]E? ,?V {UnGP6a4BbSVh h3E(-Q#xGr]ubEڍ8vT}>Ə#3=i-1 ͅByhjR/P 0D qWrC̉:.ևy7G`OJ*z|'"wb1G%J81H2-f!t+)uBI3c9 A !s Hi8fOΗR*r>|!4>qFXu 1E/xXwS4֟qC2Kx`LNt0Ay,k㈀P$qNafOfk6KڽXN~ > fgqP 7qGaג s&l.Yױ=қ@+殫s?wM䄤ء_;()ЅWk(k!@Kn"!98RRNZ9'+O:#4$@JkVTbI*7A8o#xԣ$21K$%#Jp.*0!'P~ZP-v3ĢD@Ѫ4 1,(CD >JD&m@$Heeźx[hvBBqm!Sd[[ڍ 0bP":mDЋT*LhvBBqݐTGh%* '5V2$7,)!龓w2RN\C(Q,fIc)VX0,&>W6TכzOiw a'q:_Շ{oqo?>TR~v?u`OL~Ҵ/c;NU02nRwgTGsY|@Mo6ڻm^alfjU2ƏJz}sTc yV g;!{5: ^Q!R%m8G2uo8@c9J/>c1@T8blBs:m'2-Њ{#9i" !BM7lSބsW"&B!sVߢ=PfvկAҥy T2I,7ǓjApJ#V#Erc(a)۝c*3X0JU*m`Ԑd @;ɭ{~Xh/Qnᴯք ojX =AF!{ڕPw8٬;6Mo$u|zW>"m;kiSXZ#yOnYkO?GWt:p*({)`'gI%oZr+h8SLb-VlYa'eJSZ$q.2e>Q]F2ḶmM*#BzZPCmqɕ՟r':OQXzyˁ+ukTI{t'ٴ `klG3Eddvx㢿ux+g)Ӄ F/ʣ1ղQN+Ec#j فuЍWKK R؛!i#pH7Py[9:bT (^t|mܳ1:jI/I~ud?{9C#[=S犡q{qjv"_/|l'y՝=EOoB 7qQʜ4Bw3/o/:9{3yd яqݥ.DFU.Ax;Fb-pMd֋sM+}a dwA{Ж\Z SF#]g%&kdmT)Lsi SQ)-X98NT*H'H>A}~YރΐdV|~ D0$O0N9(YJ0*-(:ckG}*&8= WF5ʼn R:.FڂK^_+jQ!7-.߿_ g?WsYX+ \V|N]Vς++;]VFDv9,w\8ouvYX+on ?^*fY[\D-/}'_ސdkfGӻ/_V.Ng1!fR+YeSOY?eq9J0+)@qTkEd+!K, rRHjm#$*\=97䷛syAm_ts,-(%w"ij۸Y8xZhUS{JWP2zJiIC[o94FIz:F,Im"vJdCҦ*zxp ihxh5Јj}UCϰxzpMFW>C9G3{a-7qUHѹ4HNHP䖱_*<{zT1svU0ӞF AzcMhE >8cg8Tqc;6@U"H;T4it~+@ڒ!LB[!Jϙ< {񮘢|Vu F8YLQrTC!0"g ZMϢlAL5Է6c:mHfNޞVqOGUJA{j| P]?T7Cn#ގ+LZā0 8g׈[3/*p.2!"\G.B Q&HJG0A1Ǽ(39 ڻ<]Qn)RHm3#TAp#0}F*,J2mB 1%|ƀ fB~BqWm}6$ݵu&VR\yqMIJ.yiq}!}F|sy%tQUr 3YZl'dou$` 6UJSuhQmZB5mxӶEY|A-$PuMi^*z@%l [s5 1-2U@PseɸzD+un@Dƕw hncTRhS!w9jjпjhֆ s Ϩ;yB"\7{9ۉgfX"K±T͋ "I'Uu.՟Sέ\XGO-fHa+0U\i ']`2+ BChkp^K2JJqwar3\Z]ZIqրlT4\8l~ ~lƦEI;.mqAYǿ F֚Wk+qʩh™/8~@ g75JS:`l8p 26KƓ K$c%rXKTKIYXF9] YCTf0?l36W–z `$rߥE{5 -CR8V=TjDM/z(>/ {Jʞ/z <#*zNz$vIg u\"VL# XfDpH;޵6r#"Kph /E8]Sη:):):):ik9mubE.側 1r圹w,̓rR3,灠ʼn;ǻɇbqQWr}ѻME 0 e~M+ NmwNR`oAj RSY(;&[og}|K̷\  QHc $hޚ¿vSJ_lxW 5 \ߟ%WbviE/VvKRyKͥ0Y1FB7\o~qgot]9yX0voI&$f}Ed[x5C!a`v ~o}F [*-kg Ff bD\qeG2i;bqJ-TRJ&FzP!!ZAdl,@@L \r1~rPA%f$%Z&[8+L JV^'hse 4v.s  ZC;kȽ.&{<-] W21(0IX\&&BMF=/g~kJ-Y@1RV') A\ZiQњҹ$Sƚ!m[xF|~QQ)HN˴[m" |P9g >j]1'+R1뫒)3K[P$B 3$Tpډ)\2ALN1`S*wnzGc(Kh;u\m4Ĩ@PW(w~ S $=C)[ݔWfDO=zM󁪷 }>PFZkwo5cY_+.WGTF6G|QZnjp.h~sφDxnj TкcD[o*!5$]51CkxָG #iMT= ?BL 5v˂]fO,{ԝC,{T=GkMp7㧻t㖱ȸu:rsF=颏jMC!4{li1Y3p3Q( j=tz'H\[csnX=xߙ#cwf/ |g/ }w 7{@1m{t7R.,Tu\,ܩ>6l* ZQ뉷Uhj% h㘝 Ŝh@x,{5#l߫<2+ <00:| v2<>$P7 ,4`jcPG6J:P=f;F!gǞURYt"yc!Bp"85$.DB^$u^8g8R*DD"Y>%sĉdVdUJL'T ׫@b1~ sp>TMց`0ח'yd>svzoN56rƅߕi*Ky˽sp{+=/ h9+OC0g VTa8\60}(&&gIOy2 p-7Vb;X~ dؔ4k@Өp?3OzgVZ{35.R(FtK/T.Z5wT+œQ/sH856 4V6B |4[5n=J+=C[Rp=61´ Ge}ips> [qF;V!x_iyT,ţEZ14f\/J.S|=Ũ(d윆m%FCiA!? 2rؔDL`͖wܚO2 j^4؋ 4{:}n^I02ɋH"j,?߿)H,X޿ٱA2(OpO~}水=i ijT1-j`**UN^vkTx`JɝTA]5G hJ Ld `vC֙Gt92 H[ڕ!%*Si2/EJY5moSEbTs4‘yQVjf2Y$sX/eLL_<(Nɫr&N(q-Sjp9#T%8P"0iy .!Ql&dD 82sA:BiR3D}ТMx6Bʕ0 LJJG7fbJGK#+FZ%dvQdHT:j[`S@ߎ7>/BQ~Aj5H]bJyDK\@;eF)./~I0Hq*q&%NjOXaIیa˽yƼ:dl̹ dj*3o[?:XI8'8*Hda4adPtKe,?>^}0081۞FY`hx,eL`LK;rZ($Svpk)$Ù߾g]/N ,&e*sdMi0&uqZ8=jKj|}98U{r{~L8@4j1Z^X !Vܗgʶb;;+T#ѯF򹻔 h$M V"εee%je<>%,$|qZȞ^tkhh|2ˍ\ш{; knky:W{kw܍D=<%hZ{~bɴiQocwqi s eJu8jEOwdt}k4GOwnL7=GA,utoc6dZA0{{I{uh!{ kPUk7T8crgI@Mk.yvON0u\]vlgCRކ'.ar *a֔Ir+o[Au%/t$ }:f0 0$,W Z*| X-u9 M$+9TJpg1ONwQ; {ج"%QkH}ehǯU(|֊$/wҖ.BR钂)$+ bEnx#5sټ$-QR[pƘ4[u*h. ;):ye($pTUǡb|- ,ȋ;;8:;: ŗϮ?1|5 )xŝ ]l0y FORx\ "tb!=SUXiHB޸&_֬[*1:֭Goy3h,#4UڹƠ^']Q 3p26S^L*z:$9{*zYF)l`IN,yHb}2IToLl0Tn NSS雔<7{Ņ>p%hʓE\.S߻23W'"}*yFkD~_sYQ om\&<Yn"Gj=|eTƼb2#2F]^},Y\^ǏFu"-q9`Ⱦr&GȜr\}XtIf>Ykx.p-ql]+ ߎs8{A׆6iU]@wrLVVB|i܂,5<"QV ~#+'I`uT9sq+LÂPL<\X s-ua砹:'6:%]Gx~M>Ve$yB6j+,!51tPGψF6sۻ/6sxGO:v)_cq b%%X\3yPb#àd_f):%Bjt*AІ35̊"7Rf ψy!dƌ-0~Z)(E3 VfBeffB f(vO>f-g70z&F:JˮPUrb I5"W/zLĐ̩^uPS*q=>'>O9: .]<?s,C$~ Í jN|w!? Egfo>_qIY~7F~:[2LJ-ON ={ts ôNg-VlP4\g΄g|<[b[+Llx8 {u.)x|M*ʥD[U4mƑf*)}AToJm R-1XV9MpN`[5$ԢԽa\~x%[/#>CXYCidu3<)ڴ݂90lܲrG˫꿼`ZdA\s74Owd4GZ%8]^uOH1,n.<;h[cAӝzp:OG3k gLز ,n7#R_(szQv}N]Xk4f*e*/ 1׌?֖VNT/=p%+w2m% s mV>S*֤Rɾkwk U1X\BadgPXIpI J4}~+M#Njp]ýmc:O2EդymJ ( \տٯF$SeBz)dB*ȴ2cZ]8Vd[ޱ^yH vX~0ESztvwUaǦяc각BAw"G>';̡G>fW{Yt7e!96 QʝGF|vzcLP .S|9-b-;i?͹-!p8J0ZfB>p$ "3DF9/yNlhK]wFe;5( kbVp Ucv8騩)ĵUjx5\3b\18RVm;aNrlwsnZ$8b:lKO Cl30gW2h -eӽO fFӃ 8ꂓJH-*@%7|};"ТoGVh)yt0+\)cx_k vƯ -f}\XfQog/5|W~jk9=يH[PI>(:&"\ѪSNf܍XS>P\Z)[`' Xӕ''ɯ&"'h\vSN+FcMg,.ĥ!=yr{19f}%(DCfR4 C4&j8VD)f]6/[{+osyw ``mpBR#hi #.G\a^o>Ͼf}ڋ! @]LR?ք :gN::^_9,ә/drzXNBIr\2ZR D3-4%^@  G`5pp̎s= Myu[\^v,PnG?a:( h#?>K}]/?XxZҞc&8s/Ə C04 C04& U(#D*d/g`T 'DaP*K09%qQ㱢j*&j*&Պ|Ιԃ #h(5(s#U%P8m0Z=!J-/Jԗld5铨ԡv+Jգ'= $xғIOtUl.KQی3ڣЌ84C[b\1Ph4]ҡor\.-*EIjV e*y/xUl>/KX푛ŗ.~|U:q;|#X?-P~>NE? a[hKb1hXZ+!+7 r`uOX=wQs=~ |EZ#)y3] #srrut5uLK"W=y߯f A hRvz/( uzUdGT5KSГy_m,抋R7kHZy@הG/ޢ[n8X5Mcp8eP js"pw<%\)UeiJ\,TUq}[-Iq3dQBn!+/ xk(HH"Z RZ4R=ꥡL9d!0b5T3TqZjzzpFx)=+S8̥QǸ4i(3Q tcK,I6Lg'׭:#=6 Eu1?X>=*f{ ~8O?||Sxq|wᣟnp3no>3KFxV=~7F~:[5/)!\컥%6{h (ۣo$> wdы>iP''%eخamRE&fCoshM7r 0%JT*Y )AI&c-x9˝ƿw?E)%pN^̓v+m#Iˠw<"/zh`F !O[!)yrW穟W|ea!y`Gh}HDJ$PN) q*b$lyNeCEEPuQu&,r-C>V*iT'U&pI9 lV.ޝ6Z*^&?l4l@AO3 ܯ1o0%l.qyo7b[I(L9qH)W|f|FsZr1 jyWfdtT>NBڞ]їh_տ3Nku7\ih3)S)0^Q͗5N+Rhyt$'l}-gTqڱB3*Ot .;<1TH4q\j-UqRKuKQfEQ:R"XjRs0%{#-W-͆kQ*jr]+YĻnW 'dNT7:Cpt%)R4{(b! e1ģӧC|3^y灃0mNR44Bq MT9L:tlS9-E5ƻpX%|TQ`F#^;P)xôITcĀ@mc.`KXpӤJs,6ޏ(b0<&H|`#uG<$ˑ*$ OS#de;Q%> 䔈:/DP*g^ %SCII(ʗC[gM–?:P*@ 2F2B^޿Eijv}yvVI`1Ⱦ4kpq1\#C>qG#F9T qF<,jMO.Kƥ'f^Vdp5-oSU1j?ߠ#gOM9c J ~~{=w8P⯫B{HbJGƝH=RN`$ݭ3)csC,KT tm~T{ns7hKedh"+H%0/2:*I;MuN(ߏꄞ29NQ xΟ^\-`0)U$r'H2S.=Fǥ I\@룗BgY ctGY]4!K-7PޏjP$/ 914Ҩ'zIXD3QQ>5{Ją*\ls\F{!c;ky2"RIţdy 0 ceڢB.*:R% :Jq/ b8Q`Oչw9Ec4Rb [Ρx8TsDm j5Yp=͇Q>sYO=67K88SE5%@BR)a((JJD;9D4N944( 2IbEBIM{ԋ&6 ͮ6O(E` %( ]0P>c{]ˆ4jΔ͍ $ bgqa }4J.! =`^Qsf.n|oŗ.l@6ZԡPDPq&TW1%cVX!hmUnjsoqAXNz%!)w4pf.榒 X2[EkT U<Y/e`d# )0DI&4:] O5UT u^DlXZ}[j6UV1(gZ3 :k PP4W9OFhr(k@AQjf_חőR%v)5nUi"kfPy9TMT@AdȀ&D[ KA@wX=hUa=hU(T{|,ҴÂzUwcĔqw;_Y ҸG b\N#zf%7!w\ XԊJxͩG-s'4QCkv5:⛽nĭ~8KyuݾlcjJoy~b.-CX^\zi@y"S7 W8Wp~t\\OFK8 Bt=bʍ |w=xmC@8mʓsm'%H=nm)_yww*.i 6Z7 e;s)8y1x/0}c"[AnJq:o`.z{?\Ap&7MUt[ BXBBrS`^Jg-<]tfDWBNoMLhELrgz_$(Lsb=V%P,}PTU'毕zPRۚzMP̀J=ی5.*ܟj-D-ykS~-i|{_~:߭"^, ƓE<b9-RZJagi)ݨ0I% -RVrQ5_<^5=ru*fx7GuʚE\ߣzk pt/'.N^=9q܌`ZI+굌:ߗgteWdz3vQ?EAP3Sȳa6ɰ`3];ȅi\r9K+9瑟6DpbZF ׻k[|=h;eeжq/I2G 8Qֶ= hDcNjPuij%.Se䍦tX"YlH^UʚRtCnH]-L\⢕BZJ!ɘ 3(QE9j2^->$LSC^Z*F+BW6$at-, B Qs DkqKҦ"Z0װqТ!BCI97"c2@$@фhi5 M:E)WC"Yy5 !jkŃQb  w(j" _˓PjWѪxW5!eY*d%&571P\pƄG̱4q(K‚V \B##/u^>!>\jLfIZA 9Oq9sK.su TJr.pO}ӌEo\Y (AwL&1ҘlP+& ‹g~~{6HWt!gR(Id{ϒBF )gk;j-! ǃSǻ]`(72[7^fhXx.$ qvX}kEa,ߐ dM{-t=/žյqnMT'WTKvV54 4 \qK7?.G)" +:zI,Rw{11$9/|b |{SԣX4-VIHB3h{zӄ@tuG2vc{b*) @p:iܧ }5'/5Ȗ'8od*tIa( Î퓯2†36y< \Dkg_,<,uȓp+g y"yE%PH%yo*`=6@;¶N6g&ِF7voR[!RRK9  C#Y{vݑJX7[io@R"{;;m.${O%* JzL}y4.&nIxޝK(M\9YY< 95IC<6)@jw::ĐӨa'vv> ")%qEj܀ZbiT,%1A#=yT&[X ,g`s4y6Al/XL%혔'j#Tข$E}ڒe6&?afGLҟB }5eVÐtR"Yxs}1"…Kx=ufƒbQ :4cLIvU]>i"Лj%7@4Jjݻ;N:ItRpS ;pA곾K0oE* @TD+>'wTdt)fKPz(:WmQ\4 A^61pјdO^7_7O6(L^;/08oP~/]XXO,!Ȃ\_8HM# r}vTA&5MyHN^ʞ0f߼;l(B~l?qK&F?KSuŗӇH4Ԙ݄Ɠ᥺,3UAt0'f& tGs+#dgliKniDa~,Aрuay`N!je>:1B3 ݡoQfN9EQa HK3mF* YW  Ց.M$`Lכ/nn6%xXe1{oia"p=oyCdL'nfgpI3wdD.dtXk"TeDb>&e gֿ3[ $OIA|ܻ@o}I&ڪc#iۂ=VA?xF$ j$8~9"[-l5"#]vS.px>^v:}[]=3vm9 2Q0 MS2E*IJ,+e,* [5D *::AE2PI5s^祮Ȕ<M~ܡ27q}/^Y?8IX'&~|{e8Luvߋ_9*΁X3fW:r1~zSűmH*t07nWoz4a( R/O}paq~ CY0aY?,CW9xd_}Wv6?-p~%( -I+F2wQ`ڭ)bDuۀV `ڣvk-%ATq&dQ'l5ð(:ixۛa*_!|lV2YhLV $θH5|8a)NV;ol>N)" q^/&;$=U i};Ý2D꺭]2"zM^;ea%sx;TjAu=1, & g_4Nuxm\7?jjg`dJb:IdA; E~mְ.oxZ iyʪ'yW$~zAV'!론gfLzĠ&{[eI3]㒩6RBt#m_ /_hP F>i{3t_%z: J,*ZsT>ݜSV@CH:*N OϏ)4$UگK٤'L)Dv [^-BnTF+;v~\,֍LAz leQ#sP>?ΰD N 8xیOɏ,RNpN) =ѽ`tF$XueIpQy!Zv$OTV}#1y4a(sSlS+ahl%FBg(:&㞤l>"ȦD`Ts~ -jJ>˵T"uGˎ8ԙ-_w̎[_]wreoHT.H;3/qzN +މlBB;baO: UYqt8]CRaY,͊ep,c \'x$6Gc!GKZ {Bai(tM>I:w7?m}$s$Ý:}BT1.pp󣭋ZDa|9XE6wQ/RLc:ü%Lvey(ޑ\ܫ^duŗ\:8uH|x;Ǩe.shP5WmM{s.+i-p2-;Ź1ƒo5|@#*%jDnTv=θ4C 3b!z2G(J5Y0.;ܘX!D>:$ s{&aZ U.?\MT p ʲհ 0A\5a~ԯ>)Z$U懠2?!ey] C^ÌU氳qbﱗ;9'^P"(F(]Jӌej09R+$EuhH'6)ˀJpK !_CuhpeQ.-{ϘҡT:`)(B2/qLZXc˥a{1g0XrF{M(|_J2%H!Sq>5ʄ*#c*w':둰)FJS8>r} SWGT Š5܀L8*(:hp$99{TWW̼#-uxXɚKǎz<8,k59ڗ*MX1v}@[=ёSc~{wVcRadg{cۈZv:]Y.%P4`y_q)cb3O)@rS2F9] GOond ͍Lñ72-CTv߹s$ Z(튂&hTeG0$c怳qC!BFaИ}bD=7NJP+ XZQ9J4fN.u[}%Y +7޼idRXJ!rkqWmT5 +hd S 'r//W^"*VԼ,.OoOEZ1Hw6-^3?6MJTs$p{]`f-93E!$gI#FPFR9E 5 _r+5j?DZrE[,qݎO?fR=ifu>;g*CYZT&1脦3 &1=ޯkA{+ u $WgS:N[q4"0cYe?8mZr,h10n#Թ@D c:PmZHFGSbK+\qb:Hxir_QE@s1*Uz!1'ɮ_R:;~NBCc ,@uHH^iW+#9쩧H8! %rC&bfEbN7E'fl̇ݴ}+VJ=/E"^`Wk?s4zO=I`tq{-k5 RWoer ΖÙ`}cSӢKGݗZ31A D=؃6k 6Uq#iY]/o]%w9}k?jt7P`0b ҂{7\߼/ kC?tuVC455Rlxa'oFO6~ O n *{*'fM罿p:uj|U:> Jc=+:8zT8Arg 5 f䶳f< ;J;kqdUgՈfB#8 ^@g1f3&\筿(C]wE.|וb6I[C.z)dfVXm) h >CtTΤ1/5g] YZɲA*P$[/G4ezRFQe~ҋV87K~^n"k-cH(EP2_ I_AI0H<8%x5)kXH$95,>XzBK43~#ʃi ߮z@umr) &7c=ꗡьd#Dg:L5Vfe&u>ֈ2V*FBi@)7$t}.fBSXYvt2%ԗOP[lΦ2qBJs3Gې +j=zIH_`^K\ PFI𐅼h'E}/@R8.ri T532l=qשN_D)j~fadvײm%K~H1XR;v70Wsuw &Ɛxv?ܖ#*[G>>IXgV# 6/$צ$%8yWOa^>w.>c)8 Nt*:cmx7;7G|`L߾>zĎ 7hDp0|Խ:,;#Y:|6>|' L:ݘs8iPy$67 v3Qk;^Gkx#p+?_;8\T~cs8&jΤiv>N>^].̭/ʩ6{B_8-Dn2+5nk129yǛ/i9Ij>z5ߏA魴q .^.~03)/O7y{CѩH[/ rpc5=h۰y>wX;HB~":JS:rҠFt~v;< nS$++Q2suҠFt~v;dA޴J6EtC\Lqwe֫.mi\r{qswarz\L~UxrPu2vJ3c+6G2']霤|Iʪ2GSQIJN!%ԗ[SRtN~#꣐RtP|SR:)%(Rr5YJOZJ!j<7F*7U׫4q~wX刴#l>~֯^F{{&_P;>sV~V1"\Y₩>N'w)%:]|CZ?_ix%(rPy!lւ 6x?[?rP8^ʗϥ5n6L =Ȃ(\3)I ت9Xm+`Ki*pXID!C԰*aDR([p[{A 'bu/ӫ.uqE^KQmȀj|O8Vr+$K8Խhz=֊n"cnfq%owQTG9_ rPx6>]?{B3]Gک pTcUH fOJ5䭋 EKμ\g2Ӿh+O +Ln V}M ;$;ljp j676n#K'~%۷ ۶o[e^|RҲ`dSRl$d#kްm4i}QK/7Q9V=(\|J$>[Q}'FH mGc,GBڐQp#P[n`߶J-TlZ-;u^m0cҎ eٕ-v[QuCgoM&21!Xbz1J" oҭ 3HQr۹酚N-2iO9[`8 G /DdYJOZJRix,j'E>a?|ehts|Q7[6mZON uYj l}ŗ7<\G Q8T5 %*U\݁cyHStR NJY/ĥRr( is-vSˡ:G9"f@ɁGF=ȱd)54,Hg&R^;'ZC&j"'[)2笀ɵ+]s\7Ġ5ebomX J;40Gmqs3o%{~3lk-Y X)MfoSkdBB4.s4mcR+40""?YmSrUzl,QkN&7F 6V/MBh$djmtTN-5dd Yܪ<9Ӡ.|K]\VJx._(_s&q[eK$/$zk 'buoӫuz[KQgSf{,}zkb  ;҇@;1zu$D#6K1tWo ?~zIh3 T^iCKRaNY5̩Rz{ٔ7\Xеeԗ :5KSU*#FqRNJyiK)T0EJ .} &j {|]ZnO:e^l(幅.=JNSٺ~Z='➶ Z/̈́Va7B Oz?V>ԹG+ӕM ϱm[Ώ4>m068_lku6S&@{M|Sgcz.t{O+ lpM,QnlKְFI%|qP-T.~?Ga=KV+OzR3[25ZovSyGȻez[4 R^m;Q-ϚO|J)-}m@J%`c ['\@y\w8TYUۢ{ކAAԐIFkE N,٥l+b2[%1ܐ"&Z-  Xmƫ+jv"j>7 q B7v[{0Ǡrg S {~JY XP2+D}-dMJ?[!2j+F $/v0a6&i$kdUIʺ$fU:djIY`TmJR1rMNUJ'S;E#ů: JkE79ԕt%0DLbĊ k./j Y1b>5!^- 2mӣjr흑[s{3J}~gp-ٲ/B&[ h5߾8ՓlVɷ+Fkkӣ*c%#lպد)՞1!pBofwkV8Y.PRq6ow_PőfCu:z*nG!CD!ၐ&+@7f=Dž#S X8׸ˏz8.%F|Y\u6uB77}D'.sӛN㻢j|ZBjg@1-)HC]><=>*FBRRpo{ FFVr{Dnao"u/MN?'72ֽ z2>`V~M;^ K59[.›j. 5;>BTߛq߽L̹W}nɋQ2๤1e6T>zp3kU+Z|,ϭEuMJYj tMj7=D٤4i{z9-xbu/D]ظwu79O,Xj"*IW-ʬxsa-u#Ȟtbu[?yZT=`x @Hޠ%骆N S"=r3){/"k)P[TYYi^Rv#5᎗2hHiJ;ˀAYB %Մ,Ub,Kh?J;-VCU*݁QE¨ƂRHL!d9N7 hIgGP җ1+? K *Qk+7m LԥDT0 XFKa,G Ѝr;Qd.veƺ6WpVCɡR0&1UXT$uU2WA LXbLQ&jΑKtc@Q%)ߟ,sOt>y9ubZ kncӜ=Z ՜ pc:69QJJblO{(9m弶% (%Zȅ].4ݟ~5[?{ BH3p?˲"k%<t~=`D_\tj*<=\ Q9Hd3e=P a:idLX2OsCV ~a[&K؈D2ft7t Eߖ;$*v:)n߽KWv4[{ܚf"Mm 1WKԍ&b? vcp v6]MF~,ʅ9My?i"h"h"h碩u"-燳wX- ,@a+ϯ"Qk9؏ǭQ" .SNDx.EO8!pO~[;r+mП5taSPW}bc|^1-bCLBhmO?w=T;"2.א5??SL@WuTvaJ>J3>ST":JwaW%Ɍl$4Je0NtB5|oq;8 WOXx`2)D>m Ad#][ -PBh n9/JLQs#tŁwAǰ'))r[UI^S 0R$gT.Բ'9y*A -b]{aZHN5:, Oy'>?λilwC;m6ԆJ~r Ȣr~ KG]h Xd6kNrZZ+NI%9hUm(E)9ӢApJʹ($FKHy+~k3e: 'S# @sQHG,+Jdb"e*ᠢ 6,* cPׄmc#y9 :G (Q(%ƀdJ1^t(,YgKʔwFcPc坹aUJ+*w oe pGwAqk&cy =w݆"k㰬ңLm -ԪJ!Aڅ$Q] J4 z;~8.f 'SK A|Hޚ8n9c8zS;4jPRyؙ)hJj☆TLr5J Rr{SE[}IP)oh䒇%O;.jNr&qq]K⃣Fr]npm>+rޟ}'~1P߹jJ[LnsAG |^kgYgG-S&#xn>Ɔ޽/ڈ_sNSI$<7 3%LYZA.C9[uG/lNAM\3SV'9gQVxʸijMS@q-װC}Lʡv Їpy}4oMyYxZiTKųӌ [;DcBCub:Tp8T\TkݛA0BA?cGDx=o H#?ѝ@N#CCYmwzmJ|J_Xnȶ=Yjiqu.öԂI(bC6<#ǛN`<#qO$H y&dS Swf[*16xq咐Xғ#[ y&dSMS8w8D&WRp[zI,o y&٦oCf8bZϙrN3 {TUr2J4?JwaW%dTR1?L1{lׅРb&rBI4#M pG` 7+Q+geLxy3/"2 3w~a3_mޔ!ن3ST7֗e)/u ҚnyD Czr?)ٗyӛANS}z:Yh< ? %N|Xp0* Ӯ1E 7OG+Wj߶|$M1aLPq 􏣯S(׌Rs2ZefV@{Zp߬O7Ixz:=.NH-B3mK*i8`]3&R;!7r rCseW ^J&AL1Q> )*_;77$lo7qٚV(,5Mq#ƐZrt585CpEadj^|a|߫[6_e%@Y^RaM{7(rC>)+^'p6t5ih\d& z i m5w6PmO ?Փu&pݧhVӃqԭ5dN;dtM St R 􆄢J]; :494ɐUF?f[xɽK霄|cʕ`?l1 B|;BdnK -K m,sL)N y?y\d4ፋ$R  Clk%JZg!FʲNTX){ynMEpnME@#Ac[w%h$p9خ1Rۧ)w6:;urk\]'ȴNnQD ̾4t[Gq2g{@JI]JO<4r,0{i h`re@vrz{AFbv.S(G?@H9HY y)Swt4)2/w.*a_$* *Rwe͍G mc"$W >uԀc}Ѹj̈(tg}U%["eev-t^8}U$Vx2i)Oh󊨶-1 - "EXJA4sp*> w$~e)aV2hd olí& ULV&ۻwRM]b1_L`O?9Ļ-&r~@?`&3D/?.,fi Nq,ktn)wPvoM_~ڻ5 5lL3h(łY *b4ak'1_4sӇjuA2S O@\u=-\~{ukE$Ҋxt!nWY0&rWXa7[XspW̝X(왼Wj0IVnYiu=gŜ'tjq\R+E7 aDyt>@1)¥M]1K8s *S \i`9K%4=L g"DZYΗcę|5vYy #V*}! 3"*0L#EFlBr[ :LYW(@~0-V  YT1…Q#`I0r0c@ ]Lȇ-b,`U {v  `![U?ު2[[HPjPBtAb!{v?:yi~o̷U_,csoG!Z {od)*A ߤk џH4!Ѣ; onLgF~XFfQo~'4s1Ef02V!KD{TxgxVءY&4zq*QL͒Ng=P8DxgVe9W@T7ϡ(3B?ϳ8LkT=nd:뷾9-+he%``y>)2aV0taa+l2h5Ȫ/]wh8I#}6S-Lp0pF]?rOO#%dñĿHism=J콍|!۠Ȁ_V.JtR-z~);- ІT |R;ys[TB;NwxE3g]fPtpVlﮘZ"a645?IYpe~H!YVD]10~q@ʰ02&Es`8.E6}}L3$vGO5y/ʪZYZP!tNc  aYrGD"%ZK}W=J^SkUvC3tYNV񔽝v l iN$oew^ͣ|@dKr!Ƙ#n1OߢrҒfe`$}3 &M7`<O\3-5x:"QEkp A"h ^`X Tq Q('3jÊD`d8F9$4p(5$25=FUUiqjnb>.lZ{:~ˡץ}+-ɢE0^氞lMQ-s4@p . O{ b5SMs.DU6/k 5RIe@p@%KdsRRľZ`nS-5iܖ=okMvc]ZU5B9;B._L±l7Ռ@* :tp(|`>,??D dyV< O+m7s™64F*o>x )81Ix_[y Nrt<7GF+e : w&3tzQuD*a8hp ߅+P@p|/'Ьyn^\P?yZo::OGTPN;}󦛳tÄRqO68S`"!k.< gX0JR4Rk-g`dyBKc'u!B~}X];b !)멗kBaL vi7o.?6\J39l>eUck\e?f53Y I-J"$:XCLnL,kqZ\(U7IUWX9g0։^ 渁G 3<7Q&8Pp!ͧQI%ŤK…]:,,O6aJ#+]2=S4Ulb6{HRgRMqmT!lCNzŏ59@?k*Uҳy5k4Mϻf4n96! zWSSeu{XVNmiT5 9 7-zW}<Г*=Yڱ =kTS9k|^ze}hir .!yk\*wmnP]-j;&8*R;zD~+0Ug wck%.ekҩi۳W*B5R5,-]@-W}$1h{[eTӇ7臖Rzw8 ws{NF9qmdewsә6`i76!>D͋wjI$+V7 K r݅N;2+M/gX{Ips9хFZx-P+sܛXl25HJb{N{zN߰Rۑw|*#+%Ce_QA|B?meIGRvERҐ"'I,ÞUWWw׃ #KT`& v+Y@>60TS= _Qt\0'F.7Ea|{$[ c)~c0F/1P3du=F5)BzuO'he ȮJUx͔Yr) ..̩Hkc,2-8B:5AibņH U#giP](zT;Pμ~ULjǕ~jfИ$TڳJs6T`/$ZWBi#ۚ.צ],E9/Z_}(BHT}<ߙL\|. 8}$R\Y}4(!=1L$ t6=Z:`:?-T}(t}0ZS7="PT)x<rTM'$B,>,4 &q܀\FzFMHEh qF"M:ԓT U{R~fE _/RʏQS*;S6\V^==qwˉFl~ۻސiN#srg|Y̗|)t D7(gp9/VpWv1#ΐ=12̤ɕ/_&3"(UbS!WgC% C*9^|a%K*).Еt*%&scu5̒4R N:)V>l6є t~ m0&oLcJW.V[K+e omV`Ed hAyADqƤ*5Z;3QdJ'kL))RQ"if^ L$$2(z2kIR)YM=|XJ.(f|2Ecx) *hN O5G c*6&DCq?DjHQR\i]ح7^C;hQD?.iUF?kkQT~PxX>xN9Dg<'8+xd`P_H)mJXy<#G%d9H ,ۣ;ru䛳Tr JϺ8Gر{BHC hh ۫N<5 `S F`d8wׅ L-rĺq|,(N\+.Ň~\6>gM}i?g|1Zf1;,nl.d i,tʁWf^kSO4O6:aݯ2_]Xֿl#`݊y cj~VAVFOrO -Y*}%8~~0JDL|~QNFGL ʐb[sI3'wzriWp( y*ھ >bM!dސ3:34DV q};|FC5[_7RAفTkPCoEՂj(`BJlX4R<䷩}zw4;WN[U6$o^}ֳWWկVWwaQV[znYS :X3J-n ;E?+zg QP[j?2QD'W-dՈ6jC10 O! {YTMZ!@iSX~sU;(MnЦ+߼l9P xs]ވVJ^!kG%zjgV2{%7I=W{N7 &N^:ŕ"PD 1I}ܖk(v&9iVnr^5oT_ T!`ĨFD${Εi CZ"]ofl_щŘ&ߗ!IISk4)PK$"L^ȱY9yGm~C3Hu#vޙ%sz)LB{\|"udDƨRhq0 Fc4JKCԮκ"L*m$BAaT rld)"\>eyV}׬iܦp!Wͅr|zgW:Y\-$(A'KZac^ŠӹRK%[t\, :}/Gd឴ sSAFD짗}'[OI0"ɮM ,2 S;Gu0dP )9.`G_3Ž,O,0fZ+dI蘵Sz],nu|FNw}ZDՃ?DAU=.7an+G1QQuYHB;l9nO3`7MyK 2)p6vى?Mşnsqrf|J.B36IJlnk|,ױJwr>F7%ey~(:(,G^sd5 a Q  6SA! ͨn-; 6QyZJTH": -G&Tl$Դ2#]M%f:Rv P e0tp3'60MS>jsYV<*I5IsU:{rn|j>V:Ź2hHZJ+g}ToRMq_ᨥ5ZsIK1gG>ߤJȨi)oF+}ToR7G-=5-ʶ/E(<--BL[j=RiKKAQKOYKyf_!?TΤ;HZ˼m m[TJ `@~5L>\xMb1㵉V8 $\R`!60'.}Sxwύ#Lء,QpVRb;]I&9MPt式5(H8~rGlaW!3S E"}QBhJJ!% ;=Sai9a(WkA"!ڿ/jB单<pBΈR*UB!6ɏ>|%a<\=zsQ/Wj]-=Vm3S B O}/׍ ӚV;_)G2 M,ilTw5BК@g6Y8%ԠtX9&: khWgmHyQǺ872S n)w\Kz Vl5\ds+05SOߧgsSdt_ Q nN:0Fsϥ+!} s|)d:R9*V“P J`|m ŀeĘT*_%XE A猰55?̯"`!u?TSܜӎ- {LR`ABK Ғꨈc4ik)yZJR%Ք1വ2~OMj)ot\# J<+,˸sBZ9u2flV ZSp̠A)R0EK$[D&Y kNq$3+\q584 9vAͶw'+ss/+U~  _6vY{nZ6c܇RʏL|j%Ii22BA#17poLDV&J?w9/^Xb{r=~ml6S,x1\қGusv;_q2+ n.Mɧ}RAXU.kZv^6ozx׽^j].m-%qL'K+ 04,͆ hdI,dt&QQB<a{|3(Q8vx;Q"dhDN[(vS/Bt>^So>lA %p,Zʉv8SorCkc,2-8B(8MJx uB'7=X<3zl\w1CJ0\ntedADf`^PEzY@;Y5B(ْԆ`C%:,q3M`BR˛0 JuZ%w_4Bc iEĨ]nO/:TzFt^牤*!ן[ kGqm:*V0͉{`4$H%jr~3Hn~Q͎g.u)rH_M]w\~^UT'YZ+EZY.rVR{N}~6i=L]ks1-k3$JNrAYہYmDds8 WܩP ]Bah!aY^Kך n֣tJqHFXKIU^6,v 9da|ًES`E To"G)8n`kЍneC=3 Asq}`Jnל~!SP)] dZ]x;DL +E*z)P8JJM/lNuY@-1pSF6`{DA^j-px-[T Oȓ"&_<5[Lf`J@\{8E°`II47m@r^+kb1wcnߙdUk-_/ݏOUj)jsrcgGHļ̗|) #`Eզ~?~~{6 bwWW}*tX@oO>&ڵqJ߯mOz$If%ﮯXm*yQ]8"x gA9ژKQO&ՔOxUbBKQ}Iu\;kػHn$W,,K[,vvm㝗]y0BԶg0}ɬUb^yvKٙq1 Fì78 +/}^zlZgSQM/൭Uk1'뺮4"w H@h CQnS?r Mv1r*]'(G>ct$܇%@ ۺR.`q?TVXԤYUx%dhQ AuLi}nL87H3J)|܎28ネ!, C+cpcFbӫYaq501K2:3Q׆u5l3L'|"C8G.QB$J.BLx; gE#`hUT*^=#fƚNf$ ujУܕ63hYF9IXй@Fc yV4Va36R \M&̈#4a_b`8MA9 +E ҆jx,\=l+2JyS,pVe6TShhy=E1{L[q%jwNUgѢT,U+J\|F.KTk$MX[oQ}jÎ;a-3漓뀠u;Oy'aw'imtB (q ZX"Fyk]tO8\}?LPSi(v״k+39 U/JZGG;{Uѳ}/FwNFG71E:$NiQ8%:ֺғHCB#Fcqe| 12`fjԫtz٫pU3d؛ 0%#- RFܥ0[wѩ@ #he:qn3t6(7KyZ7@"K$P=%iȠID o[؊[:PbdkV8hHf59Epk!COgxK&Yx4+%Fҫp<8bnrQ1rwEmA2,29X"+YIks\ ]F˰uͭNoС@pi+o@xSb>><vT^HA[BFܨW[G:.mLTVRor"(븺v* H`8)ã RGjg1B\o\q:aAR\N)i׎vAĺ+s}%~ j*S) a {.V:T#!iK.AZ*v㉴rVa]YBjT9RZ2Z.`9T\y=(Fk'B3wxb``'gE" "E-\ %G\44(aN}Υ#/`azK"Ҋoyv_YR~޳4R1Vʜءxl rަbjJ ݜ3fL,hVT sCrV&JAylO(J9Qa aw'RZxŅ.)Z+  Z,WEY;rJd$] ޙ9wך鎀nr4yi/֏Kqz3`'!A&Rַn%\}*e}⏬qr8 IV5hAg"er =۟=Jꄼ~+Gbw@3^J(Oh7_QfWza+D*>,JZj}+X+.rޕ0!ܪ%AR?1Ъ]mqՆځIX!flԃv8B*)!ִ#U0Lh^J[dQWVj%RR^V+jkli.8{k=$f<$)4R_'B11RXKūB2 Uk25Y^eQZhPrU r^)? EQjfTTp /JRT E|!kU5{N ,[:PP,ݾ,\p?t_ֹ M 3 (/Q Wr95/nTi++8ٺBȐDDmNT0F ^#<\"&,Y1ܾ,X14#Fsw. +iK\WAB,苈«*<-LQdts'M;f8<7cnS]ߛ;Z#$ Lk'5}zvFCP@X'~u\>q2+#JtS-:  xw5o<') JVfTj \ԅَo<ޓf+Yɢ] r:KG܏a稛tCCf $ͤZPH6Cl2)y.Έi͙%W2SSҝ-ӻ&T*;*u$/(G>S&$ ܭ.> 0.Vupop<*eUUU yYHD@Ju*M;@!6N8 NlO~̯c4)rvPʳx׋/td@wdm(f q6BG=A_x9˧W0a #P%]i9,t0b@'$綒ޓ_KNZy~ZvDy__*=nŖbGkFs{~׋eӿ K#idw;9=-nOCg}o])dPc?J+D2wIi3G"g/j:_.RO^#FhG~t_?O_@L%Uar[JU3濽ݬ?*V>28 yO9G)gZq"W%[|U#⎬iD:U&fڝ&70l.waα1ϫRjsG݉՞u݆Mk;V9Β<\0 7DKq_:> ʰގBvW'U?q(f#(rԾ.agU)lZ%ޕhz[ɥgO.EiKɥA\y\)F \0\tu3]>=)o'iF=jНLy[Y1NC|np$`x-Atc Zsk/й!NRK#^ ġAl5MDMK6VhPGur{YRB(AADHugũS- N~]Zp]DؾX6ϷO }n+}:ş>uϽm\u1T |3ߟ<ym~s>4?r1b2Tѝ͖D0k26\O;_sevfL{6g~*7eY7nI6Er{+&wKA~#ƻw9"gyޭ M4ɦ#Mn% hcLFT6)Bqo{ݒ'MnX7n6XkLߪ5 yîZCȼi 5?%HvףD`CgY;srd̴DzʶFreAVm4(0Xxsн|(Iʵ^(M"9XX9(u/ࢮӫPjUphQג4jT:+/WVշt[Šbvv^\E/ZOnGo vě;u5j(J.dW',J,!=E}t^=h\ݛpF&%ZX)7a޿ЪFM,%B*J5 ,'OHdR#S8n̏b4vC!d ܬ{Mnyxy柗E~o4W\ԥanרZD ƷPáXޓ34'+ h7UO;FE_*:)U k e:Vu k@УKPvc?Ǡcw\RqĀdZ7bI'~vw |Q~YU YM߳ghp}Uf}Xit7^z{m7󫻬fΥΡȒ(w*3SYumPBƠV Yɓy~~nqIãLOlaWPYg᷂QfB%N8'gWkt~6̻@,U Y N= 8TnqTٳ‰ h!_9j{V4 U#(^hGDMLsdtL8mVi}ןA-MICK] ƹuN+#,QCSeI<7R(( A(=OLclLR/JTRB*ׂXB%5RJ,UiqRb*lK%t?N9ݖcA@0!9lOj$Iӓ'5mLx % 22$|ըyjRc/0]d/V"jINeoOKc+Jj\ —KRLc2aq,ȧVk9^ɲxB\nXVjDiU0,%(\2|zjBu|nFE6Ibd}ޭ%ƎTl(xrx=r[! =1=hr= PFduA.hrh/uR#_5Wǐ}(Hzck6\59r2(xZ%=~ מּPyg]"lLR0ϊ+ԐiqܾBZ #;+' 8A~T)g\p=dϋhD8ٔ/ҧ"!p,54¢u?F?jgUb,X>u5SH.h䉹y׵~q3ɔals{='3oLrI5X鬘eM9?2RB7m7fnn> rC }^)GֈoLP5c: nQgzrEF/n&oa)o懸4?ĥ4(htz! YexKS 1"RGbgeq6!"N8n-TEWF0a8{O!7Bü~h~/m2M\zE:lA]OJJq6Hʀz@WZKs5^<d2+^+n+H FWiם3kB*^3CNk _9E BWJmX~ 猅ȕ KFm`$=3[?%b~{+A:iZ<=8z( 옶ړUöz;OPpڡVv1Ej E{74 E>A^jיsX:;cN59 VrYz/zv]MXsӟj{n6oR}ϩ޼bxdjH #3ZhkhC}b>1`$p.CHi//EzuxkG423+A-h-9A'[NiSRHH0PQ!: *$1ﬓhE+LU&N-kV eT3~a WZTHo?,cx]Y[$ .|gix_qw0RK4aO~֔˨#ewM>0 hJ?gbTMnn&Nx|ۣOȃ1čx117 IAl3x8ݒ?<ކĶ0D +(l<:RofwɃ>w۠<:.f"3$R]nyߵDp;iT׿p!S&z.6 A7%#zȪt3eT ʋҙ*ZZN&ۚdWp[˚^ݽ18DLަ<eFva$/Yb 9=L tapWذxGH=鵟 \ 1%!=}\jqCLWƖwݦVʲL7VXkڲ,(x%(Ȉ^^Y-J1W:_S.lLlĻWSr-x(7g|DWV^+u%&lg,9\)3JU<|FؾsL]4r.OCnXI@`AgQ+ %DBc+UejKܠ%(PZwj<(&N@wF(ciNԧ.èkq#ǮEEHM^ C0 &lKnMԭ^I3"=dI*IȢՅvu=|\CYq+/TeYǍ\Qy 6*&lT(z\Q 3J =|0EI^F rtƹ^9%[P2s `b9tXyDW'H5|,L'2 (<,nmPh=%ckIzSD)|6Hc`|-ALx,O,7Tv҆ a[޵:!h2nEx]J20ƃ/$ Lo#jE2~&:rw< 8*7Í 9Ov l7Fg`߸=w0W {FnOTx k@8tDx~<|0c‹o-k伎{huoB--ot̤a-$Ud3zuWqDah2>)v]µg0%z%z -s iUc%0\P:LZ6:JzZ ɰ^/{n =jpaTW2-A ߸=XJvbw=1?䑊)C-n}8*k0Rla`]T}L}ra]@H J&y@oɁlT)Ez? CSd KPUҬ(/)1J"RjKGGySS\;|?qKS=urT5_ӝ曣}ohZz-$+,;n&͜eSŃd K0D,J~dvؔæ0]o>tb< ,5o.??EoNڰ9n(ѪiVHYkFdHT ŊVPHc($ 3"xOn7ӔKN))rJ8Дu/n38("tuϣLrh61D`Ŏ3MRڲZC TӶcS) ֽ^ʩ"xt? >!lЏ9!X2$^PJoиNd4MlUj?35+ʑdNŷq !xΛaПv~.FOʲ%jb \cO4v ^nV2ZyL)_pyD\K;TkwlĦnTY=|lseo-(1s@KGUS_?XWSq3v@T 7{.:bĿ1vpXE;tF9Q`N XVSdJ}I27t\N>l1>uчdnRUhv yY lva5R}`/o@ۻS?ou+ZC/Z<%( DIh7'#sá7jaT3%LТCOؐ ղп/ϳr=`(yP] 4-kGr^-?!r@BdMyB즼+ysW&ֽ~aZZPp;!+z]+E=n>:p(Em г(Ey]ԫOa|jI l7\fwSJni[=wѫPq/e3qӝ~U=OstypdSw/n663G VGwo5Xf>qv݉;i ߌݣӑ#'bG@u %v%ru~;JA Y<)Bt%ΆW H 0Rڮ)m7PvA&a۔ j~bXV!L 0oQe1m?'R©c!xrq@SӂΓ"lnna,]%A@`Ҡ+} nM[uLhDqPuy5Nƙ)t%R꼽_y`[\Ul]n(Vo0 S$r@By2l @i58Nj -%Rm:Z%Q9H -y{khKѻ5 :i"v(fҿ&ɒe\U(O@f`g; سٰ0LQ+(p%ge\4zuh gE4]-+Nd@o3n@oۉo0Y:iV^t1 9H5Ø/$5 @AJDi7pCԳ%5*ǁQO jXM&\bJ 5D8IsǙqe$QpTjo&`<)kwSlT~,XY,ͻd$uiHKQTh?[";WR4ѯ&nȕ4^ʪITI*&8(AU 3*/$R I 4]H5JۘE5?W>Kkj (wkM'S^NhUNV7~jJN>:KQf{q^:׻#2Ɣc#T¶ڈfRLCfa⤍${_ʅޥ3̒ 'D|8ޖ\\(n:vW:tL8EhW:F(v'>daں6EL[)\s5.K\̆!7NZ?A 6<(VrC Q*DmcZqC3n(`nC4S nxu ` I aQ|c677P3gEB:]"0#9mJCΉ} 8V:Nxr#%*(=wndN0B!jj~ʹe9{_&2KhF"<>P*>Z%f7AI-dmn0In#ͣ-b3B>yU\hW0/'s,0a`ܞ !$#a̦tar=U G64lrN1Pcr9VDOyZcWr|+'Nn[p%lCzO?)0?4Mf!wo%2|VY#³{fAH-wCEd4ƒro. l<$b:N;@ŤZ<4kNN+mR`<=z)aw|srKs]UJz;'(LMFQ$!76<^ @ i,E(X4,p*Ԕ9 $>&ǟÍ %J֯ D[]~\@H]\;[ xZR`g\u-0@eC7rEQ1tCLu pJ06EZl8ٷ۰>p`jNHQo8hQK`բM¾-\vD#3MXc6r)Q&kTCu))P #ײddD|J [O$r'k%7 ׈IP%Bgy'Ԃ Q)S#,3@U<)0BJpt1hۈBG:H!:Jz1:r‘@PI:!B:PjBBKE:ֈc_'TdqBA㭗4=R%9#?vB:!V)j%{ u ]:wBXzZ;zN9 CB x}"9 ']Px_ E'WI`>O&I TB6WCW (J)I\ӵI&,=+rsI93?`xUty[EՖT.zfV.Z?ft\(C(znpճ?ܳD kXJ_Ob||ɠHHtJѐ (ΒBhu^憓$u3?ho*nVEW?K1>I,"$Qxhǂ Tsɨ`*.~o]S+޹%;i]օ(ZXup^uHopUG (HN)g}s$_Fjt`9tP8|GҺ8x]tg{2:gT|>P?uVt?'-64IIO<vD+ py-^AwP~z*n+R'oi}5p]I;R=~)__;75u hh~mrnһu>sd E0Aֹ.nwKfBg,^//OL|`TC}]WSl:E{g'mWj+jwLɖ Ѫ\v|yGN?_*ݙެF(FqFH&Aqs׷*V QϪb3gLذԶ==u/xKa4ExksCP_ǔd(ӯΞlΪj\Vu<[{ȹtb}e ժ-ީ'qVΟxsnq9h|-v{V\D KjDDvRG9b\eU0cEMg[I6>Je8 &Wm/Y6:D5] cU跉<VI7/=VН-HGN0 3IYj?Cq4Q7p`ڏC\H$5LjгF0zvinFKV5 R҇h RNJxXZM&bb{wzߦPz# qQגVPgy3I!JUJe5Ha+&{əBܔ3aT\R" )3R0' ey]gBa+ueUZ r&mN=ΎN9 XдҖIpdG_?z3c HL$"Iѵ)E!}z֦RaˎndsEtz>8e D${3 bNF "lzt]Ys#+ :d&/P٢GI{fXE:Xb#Bf~@"ȣX-&BuZ=0<ۑ1RSp'AcCQ?8ܜn"W-Qr44m RFVfD.WΨ^Jb^w WR͇nRͫ{Go7-jAiP ?ڭi$9$Z] @jR-k~5’9_éWSHWSڥ54} 3e } ^0IJpHk4M??\0&dKy.?:9پ|mR9k!1}K@8g0qJ]ƹHRa >kF|送4J_眍(-ЄTU/i09c&[n[jyޢV 29Uf4YdA0,2<Rsȳu!ټ;:U^' ^`{DpSL^{ s9ȉWԎƀ(o7w_3)*7oT|dH巬ܙ=*PF}Y'~E(DB Z^$Vk'uZCGBJ'K uov׀*mD o2*x6h4à Fbʐ3%{0eVhP[>2^hNr2WvqJ ]dq"L'RrHF,g>iQ&5!B4NBż ")T@3ejRB2d M ]vn$56 XU'ĺJ)4B}& Hjp@R0 Zn^ljqVWHîNN)D&JwQ-@UXOR;_jеȥ+B=7hi{=+֠-yvboaQ򶮳è%d{d ՘ gg5q%s>&#ֻ;h/ZtlB$-u7'٬ڒȊe%S=ND3Ԫ0aK[Wv7#izڙtb vk71T vs|(i(Jc̡g|;)c5.]tkz &_ݗtk#0>`Չv8*C/╆-.m 19TiӸ}V UBvpPW)g۽G(R Y槩tn=}3L(FtCxCu{BE&&XS+#ԥE$t*YuZI&!HTegC**J;pyOgya57>~eczt=7Ff>zX=y$ 7cOQ-}MeU2췂f- 1녩Afխ|Ȏu`u FE"eO㜰c=:V]Y7N`퍧8/QbPͩF6vQ)1Gn5z{1C3Áאu͐m@PwN#[.8t`-ǴbLӀB<#1He:WQ7Y`AILX$15ϒr)Ya07"!şjÈ8o/*X$,XWKp.)!SryD! _8HYRIoaIur3@%hߔKi:E:1F<. ͲXE*2IDRȮ6~2OLpJIϱ”ߖ~e\`K"u9cN& ~-V%+iP3e_Ki9K{(Ni7VmwiC9 u>77+k$y綧2[(]6Njzղw-l6OMmx|Sxɕv?0~f׼5~&%=e⢈g7|r|x\^oVr 5eVF_.&,,ӇI˞3E?.\rEEg^4oĻdUv9uq$5h6)J ڙq"„:* ;II$"Ei/o,kdy6ci oࣛY7yN;O] R Z:EۯFnn'Հi&&q-0ڑgg+pـ%ZeJnmү?$y0VIZEIdXh4tH0+g>lw 5a>cD[wBF^n#cV l[/>q:ؼqu q)1Yhrtpھ5_[Hhox`9gQӵA36!߶׷Rsi A`vvuװץ;Z&)9֦Y?%s br >+7̾r3:/HHYU%L \1$\dXp ֩X2Kcf&sTr>/e o}T^6L5Җs 1/GS嘓u^ly=eߚs)\^S~]}kJ~]ke&7R=E<\{E\=;h-b.G7oҟԙ>PZoի(N9!c'~z&XwᆱW Ew̜dn;ܽ([E] MLp.qOpExO dʠ/I]ARo؝W?>LnAdJaU-2(Yd(2EG+2 9,>syg_,9_}r[Ӑk?湋r8&*x&* Y1QDTYSr.Woq|Yu+$N+&hRLQr%Vz2%&]贖_.~x~iJV-& 5H+љ4Dvw)6"eCD@o 兀QoUn!<${^>B 5p!z^|;}xVdt{i { `6rHx{qn7JԃBU3z?_wk{Gyw`?DTכjGS>>AYUGO{<;{IE7p:mX`Ot43Qjt٩$6ջ8T=) Kgzשpxծt;5iFn@qJp5+X t%7l|$\ojF rVӇǫJvO*hRc _mG6IvN?Nb*Jp խ_4઄׀#^oSK ܁RZ[+i3`$.p.,nvo~24WRߚע^$iV %Ap,_P;nk麩/_LLw*(p9$q^ˤl-qƍNjHUJia!?(fDBd]PZOV{||s3?:8 F9QQ^C4֩Uy"F1Z]Bg^DSz9׎?%+nV5??ns;} 1/x6)fr77ϻcQBJr#5lr/\|}zʔF`Lq~-cGBSSqbV >yI^ B2LopȆ|jtj'Ǫ[z ,W|LkS0*bHXZ3.UZJWBɤ`KMdHr\5XVJXj׻ncJaί"dwk&/z&=ޭdxz#VVC9b9*M}2&PO+Tz2 ,*yN},bSEa (,rcU”TdY.H1:߭Lgr׼_TvZ&_g77}ppe4L=KswiGi,$Y`gpiznAB)W\.͹-.#/Db}ۡv"nBK@w Yޮ gnl.ÖmA{S`hnuOIakc[flˊi$d~ECQ4cx 0j=!W|5^fvrͻIDC ('JU<ѨBF"?Ѧmd\#C4~0nz ϏNG\9k_q$'U^_LXDcaoj_O>ğy[^i@Ғ`sƩpŝ2`JIK[3\F/Oo# Gt*GC}8[Ϣ2 j<[]f^4B d2t[ʹxK[< &fx&VqG. KahӁ{FZA sJό2)mpRyv&H(9-BmQ@>NMi3T Ufoq$#w <1!'O?=scX(J=xIKʭ0yoA n_Bݹ8lS6cu[hJaO*?]e )n;,@0#؞u;>[֪\Va~RiM˵]Ju-Qp\*3 긹4Ϛv)Ypi5ZqK9RNTZ΂K_ 5C'.=n.U\TTDNYݤeWfWwwwv;)3j6*bjZun%v5ǂMM #fq-(M>Mp%&~b"~Htkw_/=M (!.!bj9 ̿$QSdh^-E=kbEKOG$ W~sT{%j*MD?Ғ2M8, Dk5[SC{\Ж H](3`kU-T4t PF PqZu13d=. >9<h'±` ցc&0μ!F#)) *(D=ui0dS mg]g # Նaj,S Ȭ HJK-VFSZ*"boJ]0={(!"(F橋sU#>ꧪ˺lfIDS/uȈ2AS"4B(> 10%/%$C7*2 Jhq({o+P$/x"BToCZ(rOCĪAp\`,AkMPd\35A"Pf0lx݀_$w7 |)n_-o4"u i"sI0%D,%I@*TmMoH5gaڢ#wTҸ@5L*!Y$ejV=%cBfd٘~gj C-SîA~ D=Pȯ ]' `}\Ge6 SF4gU ^JZ;,Nz϶6 -\)GܖK 8Y;Tpi@˫ZߚXn@ @}PIaё"K -j 6k'9u[1H垜hԍF"RDNJH"Rjlh&t!dɜ0&g#4C^ķ`)'Dџ%J*B5I-\[< ıse2:>\߯Uf39\[s3%Lⰵj#J!5Z%ϋf:aLh(0n1RήۖM+&E: ,mRx4!QoaIp>'塓I/pNoaIj I~n]?*QP+:B}'5nA-A>"7!8Z`3 (Wk>jADF.Q`y ^pi"ZSRsb$5?ѨqhӉ6\3,~o.&쮙4j_~~ Ņ%C7\/)_U|S&}†U(#2VRğ>}"Duro׷ Ci@&tʼnR+ɀ , +9hzɠI|% ?U'Sv ]m*]v]n⍿l*TWWYUz{/kLVhvIЗ[& ?,l!$S" /@X#DLj))xI%:H@@D+XeKHY"d횥lf"IigPֳOWortU0lYwc| J;dOodDJ6*5CTɶ( D^%;aiQ%9H[8[.()JlN꓎Y*cA+P_>ZSˑ'L. dܼPW9-溞5v->e&rQh2P юFt_ܡ MHRoy΋w^=A G&eePΣ%*y{jghN(>!yN:tैy C.@0X6\.w1+`SN*`KZG ԼoBQ?>_ֿ} v7{ՈdaIo|Enۉ.J[4Ѧ?M&Ž Q _]ݓm^0 #O;Jp)'J@:i| e#Qd)}N:Iz` 5#5vIRCz\U*sjVSCUJ_M b`,(T`yi_(T=ǵA9 J/umPulm^.]^szG2Rms liG˽S:ʙTȹP,8f2$Q%0{CCZ*,nUܡ^ДhjWhyM; Ls>^Sy?OOkej+y,lYc(g9Csj߆KgWMGë wWCt9cS~k*!w\|t_f g MCZeƘX&?Rv[7#&;"lj՗HnC..3 ~K>29_ezJ21]~zrcʈJac%''>@>ZG`(~|WLLjW94'S>b6ػedA52aSq!'e65Y|55mRQLi${Ԥ́0 5Sx,RYmQx%oz'Ns]KcMN6a㬶 YQ6)P6MVޢ7IQF D)Ju!Aq:4s% 8R&tKĈ˘tC1 ф\#!fct0?{ȭ /'8C~QEJv$yKcJdHj/>Ɛ)a8jE;No%^:㈴Uw!Ԅ횒:h5΁#MMPZ1:cfkfNRw1Bȶ0-z0qg ֻȚLp^3dpCPIFZ]0RuϮ\&ukhXMF0Qy|e(Vl2X80Ť$Tu (*u,1RHR]H'Rh%8S3,AvFT!iUdLu*VYCLck#vnRR7P}rBsR`I3k9yFAA25>x8 p2^ &OsES3쑃{'o[.8zO F':{J1H NS_{tΞf9]#P^4gdkͧ3~_ ךq'_\ n,oUs_nZ5XlP˱iy&Z4b5[~>¨\eb9 v&G,WS~@/gˋgt5}>ب t9n0<qR^柷%RFHNFi.j$RԈ^JW$y Vɇ2 `%b?gOci9U$Y6&r ڛ2\?ǹU\YHSީe8U6Wraakȑb? r-Ecw+L}:ރ#GɁ><81m|7q?\儼D˃$cxvdy v7WޚL e ~zLFM<LtF=|Nt>RВjVkVndT1)le(W$jJs'阆YZUeb6X 4hVGC͒,qF!Rh/c S3072G$8Gc-X**3,5$ `eȱ8KV㩆O u FTZIEOyZX W[w>1™o)yHd*TOu<ˬ#Hj^gS@*2?^FHϽ'T b0zr˃e]G҃G-Sg6g:V J.ϫlgnao;d㻂lh])Ci!ڭ h%S& Oݪ FtQFre iVyڭ LEޢx$ADsT_RMp!L~ӥTIjYa.{7q&`:*73˛g3?ӳug+_t{g VXjB) Ψ.؞kDe*T[;f!*J/]nOVg4&VaȌ|YXuG~b7k *q? _xK,&ɦ!?u*uCZ5d b2 Ɔ~]/`3H gmN0-7Ɂ>ݢv3`P<mpaqZ{ZaH%_%Buy +Y F,B:zJ7>e^ P> cgy@oy~ϫ WA>|^s݂!dZ(ɔ m Pc$w2CJGQ!ioa#7NSE&s*#ٹ+G,lĵ Zr븵2~(BhQk?N`ZMY.|"Lzz}A7N5/F={EF-K!*%a~R. G20aXӷnb"!+NPrЦBM`ni;VePY="Y~_ٓg Ҳƻ 뢔7*Dİ,K721R jԹ4}>-o$2,<7D3q"2MZI|3s|*RSgʴJ < HZH{#c8 ؿjVvEwJ]X~5zk"g'pBXxQP|n ֽjY\򅱪nv_yO</~I3<dItesfV{:2)iN4[亓d%5rǝܖ*9UPyǚSve3q Eu:[.EsDZsbs'Y EmQ]{e3, u.rw:\{Nמ쨇>^eu)`B}.V bϿ:ϓ:Ʋ$XWXӭúbM *iDkcöeHMPytUz37r {}{wv1p}E浦/y]!JhY5Qz0 Π[DmUEh{nt" yh Eb"7F=tʆdDA'[ĉULہ>2wX!-+䵡Y2RjfKR7.q-RPpJ7E*~6bx }"k7P\h{@#265z?@=jQ,[O^]FI.t :&$nȣ&8uC' jM *&KL% LCF(d` T3dS8U! Q! w! ͚,8[ үM~$8~W(:tPL_]pCvWO&SʕƳva4sfsPf>s)aJ ,3B@1<`V n#kdک'#s6\SXq)0|ơљ" x(aTTRi3yV*Q..rj~vb췫܄iO~͇tUOTfΛO?|xh(aB/$XEĻ˿i|렓<$Dbw~?n8bFdb; 6on/ȈS X~.0L{lDTL`F5 (?A[a` ShO=q1'p,͇FۨZrhdmWmWxy[3-gcݙu؏P\M"zm%^k(c%qJZVxND\Jy2Y/vdY9jSne>a$'׸Csz򹼯t.pI:ΩՍdBCOteťtI+ៗ ø VRQ+tT(x s\PƘPɰ&0RU)NIMνԴXDj-UQhFpw9V U~1BB"Dž->nCTwIIQ؝JM'F թÃ٨,8M`b1K`B4g01",PIS:F=VTJ@FQ :Fqb4VVݨ3fg@9!+-,e5 !0c ('&La&z5jED4SHN_cAbyl% ǰpI +7 ?J)|S"Q)H Zy| =jc/%#N S>TS!t#s\bFI)טfs?{ ԐHpd%{.qg/h]mv[/oֳ[oO5A֜?CQrE Q`eDXa5wIɱG$̱60$xIPMX"TX$I t #'_:zqwO|qS~̀{pxrkoF=Ttzu^\ͯF^ݮVo1c%.gWoFnb>T̼]=>Lo ڴQl^d}IUN=[^oÆFoszR9<]Z"g9™vReEqMG߼%G\OA[\!]Yؙ@T5-)uT߈^FѮYg+/~[j9xȌ>@,G04ʃ5L &Жd/2|v1$I6|ސQѝg{9M8QHcwa IF?n gǣRD鳾. ]&ܔoIq/jë{4x ' >C.-zqXw~mp>ɊՈ󩦢Zts}к!s);ǎCqjz1< zY1;^ے~X{oK<͌-.gkǮ=0]k8fToiqD(>oaJ>cB5ŝXKMw;NszJzz{FV.}#jN!ٮΗf/ ý'`Vѯ0 E}Nҋ7##wTcRii Ւz9*6: KqG/7W{~r6`|3?B $𻝯 9AWIA&K~CwA%偡7nj0 G޻*MmI1a0\)]!IM_}-KSRȍ~(=ݖ4+W$4jq#CnĨNM岘AfDZ64+W$^3ԴɁury[*1:FvTXTY֭ U|;$Z>NE8KZ^Ʊ'3=TQNuyU[}7+濐~r 6,_/`|k"_jއ}(y_-پzʌ7?Go,=Y#Y (Fuʣlt Sy~svA!C>Q&.l@/Gѻw㗋0J 8Lhz[@C;%@. \O5DQC*9Uv _WXuzxu ct ޥx+a̿ x b ^z8_<\7S@*'{ht"w_rr"@ԛ[ߏl =Psu,z;u:j$c3O"'3l+TFp,q7/Y{ah|ݙ!/,T6:T DNwrp.U "TcIwTK3A̶cMv/gUARx3ކ|*Se+uL!XT bT'u&rWi|-sukCCrM)= uK Nuhf.EH,nWD[t UtOhNѱ¤0l{\l{j,[$ts-9[r6M-9mޠ1ej/NoߞJMf\j"q -29ZL}O-L'j0/d?++AdSag&Chѷ?Y.'fw pD$n3ذT% RaӣPR8"JW b9G6IB-3Bc J2Z)k$QATB^'KB8Y"66j_ŠӺDXpܥ01r !BN h@X Z XjI8n[eH1!P;2 XeFM"pNDn- ʍpp AB8Wb;w\bcRLagFq(ُ k7̠&[fSDr>FZjUق@iH>bo! N+9?dtw[BEwʅJFtυ"ӚXνύВuwry?S}*YnvcT%!pyK%p{ po9xRvwűL6-$WdMO5d)J07_|jKRڍ{aPR@Öpu}u9xI.&ǘ6%8CF_QߞNOF!_3ƀ!gDm%oh*xw7͛(yypajJNo4 T I0ϙZ=,5= Xcf3/0cOPԷF H?H<ݻdFzR*(ß7}Qӗ2$l|6G|P®?ʸE[GHeZ06$>fYkeDR{o9II.N*oVwֈo]7=u֞UxJs~wKB]kL@]%wlKt4pRZZy ё1JB+ S^`iH T"fU8qZ) EfFF:(j#֛'sĜîqjO6hw׫|-gh^*~G9vB'elx<&_W8Nzdó5tJFD\jJd{M.=8噮L#axRV{ ⮮vVgeBuAuf߇Lg%kf/MuB>5b?H">V^T3 2?@Fd$bo1:̀9iAyƜP&\[xr{jd*e^1վjU QRܭ{RbٻFn$rەG5\ݽٽ/ H6{glrf߯ؒV[ICmG`r]_XU,V?mF3$vTA( Pa,a",= 6{*;_o/vէ,oA0LhIe )T}Ays լvrU@O~iu>YDžOqTʓU+ (''`ݫB<ѽC(h{ɞ? IY&G?}PNv|gx)ִ͡y|?< }@:nD7ni$ѽ") SdEQBa`Q-RZ Ҏ9< R[ %ARL牍g8qdoL_Xq08@8X̭Ph7^0Bhɤ5THD0[EH vBI7~!}_?,6f=2Qc6.6\^?=)W[A[+?< yAg:I@mEd=ɪfjsKasVf=ɋG4֒xvy5bJmPu_(a+3*fM2P5Bۃa ZÕM]^)1\yRM$ˬҔ, Ҝ$w=ڕjj' 菓/TGm2`l i^Y_+?=&VV]%!Z j?Vs@g"M,4 6o#Y$BpXC8e~{BOrXUYuT1t՗M$ZJI)ۊ OƬH:ؔ;͊<k94(=AD^ҙBQ=HO}fĥZ@;1- ? ЁK^(QrIG#ϵcFt@:OCW*`mʎ7>C Tsmg)Pi+'=B/\ "ksInzzpdI{~AKzLe8l⒅NO'! ^}C|+V.ם72O9C3֐k|{_.ߎ_KV0fܑ읖p̜7K71yCDŽޡ楒?&̛Qly/au上:nR\=xtyu?͗M% vВ#3a\(~mfַcSSkcT#i>M8bLbg//$ZGMaD.Fcw#"/M $;BjZmZa"݆*;|ScU>Zf|UK k)D-NJ @/WPGE~2;ցx$!C2qI@m/G0bD|iuno¤}[ )=?Yhe56E Tdgv"#:s2*14(9Y5@¾8Cj`FɎXɄ؄$6U .M}HEIB]<ߎR;LUgh &ЧSz!Y*ϓR;I7SStYJCZ.P+b9o eXHYZJϋ=eq0V2pSqF(iEJAKF_gCЋ}P6Zµ6IA3GK['T>(+\LQjgTiCA3QX*nTse8DY3\ct+Q}٤3CO+_uJaҤTr (t՗My ')}}RߑEjž-F?װ2#zˏM*Z͘χ͙ᓐt}Tꓸc6vJKtY,; :#AXMN}4-ל2I`c Gn"=:(5|j,sSKbYYDq~sN0!9~0l4R1 6\y؉OD %ǝdS&[{>/+URq4L:/w(!ǔr,xt67H/hp<^! @81eQ8ʨw:M F}uv3a|d3j_7vԮKI/+KFwMpx^(` KtB-((h=n8=h 4L]ZQRVɌ֮5 jvYR qҢ (Œ?j|4+(q<gŎւ1ƍʈ>Gm |isǑf zUmz$-HZuK. SmVe\A"afd]%"fi-`u`Y,*=\iÙ\|kj\WcMz84NJg!J9Tv^tŒS09M[q:w@(zvm&gv*"g4슚WGaϗl_ų@F]-8t/?o_DLIsJNe*^3YY^Ur~pRk `RX ^) JKjnQ^jsw~yyw7ċ&Z7S}٤M^A>K)Lt(t՗M9$ZJy.|A:J7dQRևE(GS#g PNJR{U*pr)*Te b@5i@CZ.fzk;čW^@tL(W(R/5:VxXHjP@ٝ;_ 52)E\ dW'%t/?sc5$ruQ:mKnD?f AD :zf8Zl=k1B׿"H$xr:c}T_6|UB&31(\_-ɝ6. DW-\ vI% (\zAF-/ޤu۵ ϩh]~k7Rl@ Tv[4^fL>O8W n"(Fg4 d]YKGm [_ܣR 3 n$vl d;eT9+uԗ&Չn#z{؆]ܸ٤,&"w(wU.<<Dz?Q~*ׁP7RaN&-t{O=t|H]KP3zZ^{J_՟nܴ f~`eVdE8[.VrCOoޝ6%V;(J saϘ1ރ;B<->orzcCcd7tlFۛOOolmTIﺱ}4g-8x2ܽÕvӃ *5ډeQl0 U޽<,_f fPLond2$ 2E>_{s_ `>gD|5Xga׉=y,cN0[N"#TsZtXorB:L&ZDL5ZawjSOezu㫪:?Y_D1~++h5]Lf}UwoU9-ɪyߚ5OcFAw<y/n]ZCwԈi"CƺTjh~{ Vgk1,kP:߹.S;>}j| ڭ-RD;rl/TJ9'v&j.$w.dJ[ϴe Etھ#Gы@LLBݺ߹6d$ʔ)TK-EZLaGE1-(N^J>lRvIlqAN:Coy;rY w{͵G)7YsUU}s iKgwUј{HwKJ--F[&751p9.YZt|e {\|39Ɇv+/ҥ'nX= 8-z;ua]pQc`(K|xyΡ<9y:5͡Xg&="|LD\m-j!~}`1Fy;5?cTtTKىCqbGb'j] (^3o[q7D z,g_)s#wކ̫WAS=2[3gq>FkG-nqc33΅Q :9)A"%AJJbR2P AY&p[v`, ٳcVJ^;lٵu뙧LЪa`WS}٤씸WXbJB(6 )e4MJ+2N\Ju)3A$մW[iygU{'?!P#%Ό>- Ә:iAij2w[}ێwd@5`}YjVPbެ*(uoi>@sao1v5#Y݃\нrC $t\HM%raM]pPn\ _DyD9#xH$'K>3Z.v~]+@pHDGYme-@O9e".oC D(L($qH~wm~Ydwe~`I`ęI`IG\v~ݲݶ%}oٝ%UdUPERIM s"W1) X(OĬ"-VkNSJf*Tm|x|۩aZ)^2%Np4ss o_+{"SPa6GUg YP!Kҟ]y&o8YfˬFJ6R}S.S1%)_4B]:`" R=웖걵H)$@JC(>A˯EV(g[X#%*Xs[?lm1:Prz攖 0߄@:>>͎$Ltt1pgyl~JeɎ~4W{:\M/؜%.Eމ~|j/S?TR6!88WfCʯT4R jA.rC"tVVfð|_>2A֔}jO8僭ww> k15Wӯx?vGr=C̟4W' eL rt6)L Mj&&셼.eyyYg\vG|٘(ѿtNa2?ԬS$O63.])q<C|Jwu3wIDw?LU{aHHwFw1tiNd}ûV];gZ[ Ix_3ZJZ ^z)vQ-b:fUy.Qw2Uw[ /ItnLbքݮj^d'sxtz&ZF=NBxt:&YJШPk_* e1pNr}70̳U8xjۍXP*jHUf #o%8^[$BQFzroJpeD<*fn͞x>FL E>Z!xbC9+ l5ާVX Ž&~)7n{f0f8&cb*߅2צb+]\'1s5rov}`fR iyHg1LY 0 JOQ@hc 9=>śl YhD@'XyWmux7ǓՍ-}\\Ó5U6NtN]\([Zp60O{ԇ) VDݚp徫ty7y=|n} /Ct[v3ߛG귣Z CU0B&WkPp4 qM1},q1HB|ub2 m ry)ջ7tȚ۪.V;٘}Qz[y7Zju=z[hXl5mu] +݃y^j2;ZvKk0 `EFH-(aqΈ #H,e8tv ͩ0Jfw0D9R?ky^-8Wj &!I |Tc A"EB |!A(R1B167cw.3$՟⓿uaN{ ݈F+LTbMfmD*(I{mF  #S EbJ罐"HEV)Nh@$ yB̀R8 Ԁ41 Ri㩴#4e YL5;ڸUuLZm9תZvb}w%[>W 6LugʊFAEP~OTHGf|{J1`y} 0wnr&@63xl~Qa@¯`aX}}۳3M7o;ɿ^&wRl"7jʹ_\׌p#S&Am˫ש-Ob̚iʛǘ?)?=^1*Nz}wHOH []"Q@hc+"Ua엛8}s>ѬէNDo(,Jxt\ItnFq%`I#s<}}9(٫K37y&7"o /7wMxɭBs.e*o}7y) f/| Rc4 Zn~ )M<&FM*. %*ԎFT#BmtF_"$aœJ lDDa,"g-Qa08(YUs34 dۨQB i,^^xPOބt: }~ǧI~H?No㛫Ǎ;TV!R;GHeM)tXVJ 蒴R4C6uԫBpBЦTRtWU!EPK+n|iUbٶ0W_rq:h&u8 `y*Zq=U g5۵N O_ҔQ0pcW|!S(Jvqhxp9S"n<Nn"C78 /ۭFF$Z%ʢE.\ O8!$1.$8O֫]E8Ë7|ryAaI{qI)RHXCV YԚ2޾zGC\#>8(DZPZ{dV@lM3RTwBW`p'kˤSiDbsSJY VuS㍞2%L…c6mlx܉idߵMАxz X%l\®bG,1 Ӫ: =rLE:, ܨ*3t | ՝h*} Bw0*9>W(5X5h<#Fjq,fˎ"*Wnh+H Q:, +IvIZA%Rё0n"=W틗c xh/hR㜽59`Bp.UN[Ec:=nZIJ#LdKX \틏 PՃ -i&A}wZ() P͗sS @\d,WkYTQX.THNN!R,jRdegKYaVk{PVx%s>$ x\k°7B)p:`9Uo>gI.\Q\zrq)UZ[4$I9!yA&4PXj6:W4.[գ%lgy,Tn:8RLWʽV D姜lڮ1x8 >SyC'J҆WǼSPWNQd5mQjXɛ [J:X#2X(@dAC>ȨP%zJ᪋RoXI%MЊ"X}PkcY-q|+JI™n4'B l)JZJVO' lV7_܈#m.餺OMo_ hޚ5#|B2,K{("P H+T`,An!_*zc;eR82y#&֦Cיpq'̻rQO8ؔrPn5?BGAe΋c0ׅCV7j#Ŕy rĀ7ŘLļnLl5ʠMq/g/fC]}{34&t8QчdE~G 6-錵 RrE4G,CjF"23V~D$?s[{35֊\?j{c w3<5'YA\0=Cc$̦s%h%UC.P,fn>O1VewҌ( ,-!-+\1DwyO_{&T!غ }% D?Yۏ| Wh͘,!xu14OLwm`)IXXtNqiD|BR[ J&3) 9~|5y\M>ЫI'^Mb`!gde9YF?'3Wʒ+w {`1יnc"TWi1;[r`Os. .Ǐ`Fk:q}bjFNƦ)"R EC6R)͒+B4)@3M%M-SBRu*8EAjU&) 9I9z"jnToW s4}}f:1W*([ݦkftlfiА1}-w4l -bHg`8j*qg?{WȑKv^Gzٶf !OI!)= ,Rb",VҲѰ[b1+Ȉ83Og:\ Jaz%$)kvb&I '5J7O!6PG HNHǫN,9rVp q\p9P hRq ˸*{QEktd\*)o BpIYVSk.( [ر2RɎeH  shAI<ْҐ|#+V0b.sRm#ō3{V/RB"<08Sz PgCI9 nPͶӎvz!5;,QOJY;u*{u&2VS;Y~Ww2 uI-է4nfwmS6)%iIKۦLmJ5PͻEq}Լ̓| Y%Qo\,Uǽ|,;𣟜Of6ij~5{sg!<nr~isnlP'1:̣Z=.C6&ڥ2C'D^xp4ٷӧr8O=noC=g|ֿskGNJ0E釧{|>6VUET:_|[@227n\< KR0m|[mee,̗~Y^-$h&-UB F(a:Kua @ !pas`ma-lՊVrj |Zy~T2* pb7ʵZ:O5=RiyjwnR睧pK 6NMyfWF*jh: PnW&1ВR\ٴ+5Ѥ('D?z?/jٿK/;3BDr:E? :wj) ܥ~?twdłCW*r';\V& OM/''E/AC=dBL\9͗u}XCͬ~^K[HфNRs48,EA]tIlBIW7oEڂ8X*&jm6;j/Dڥzٜ4- ewCZ.Sh37b.&U%FՊf^ EoV`FmfZ] BkBsٮqѸkjf:BL=:r_-4趪Өrwx\~P ڕv𫳄:ldف$C惔52`U."b`ҽ/@VJFmI/]2Dn5mԨ:RBy'::rq3:֖W(ɩ0 W$UңV1-sJUD!WrMmٓ 3yDX%*G8=2UܘNp@:ȤP!"!.qCK9"J ǖ6O|  V,S?\jGO%қ28 i NJ>vhcC@T'T_9M:5J"CAU_i0LJM犘0SJ1a4 a7·@ 9OILlP#5L[WܺW yABUWX+5n܂A5QC\v嵮sz=7|'oΞƝ?HDr?^1̎evL/czYwLBXkSH $ s( sCrDU^/E۾|A>D芁"*KaR+f|M }gwmtÄ !:}X[ k[70\fIQJA.j4[R"g6:: 845 aq?ՅHdj%;V6x6!*-RUڬXsH^WE\߈ޘtkÊͨ 1.]V -GB5W3WD g_T|ʓ,8#ø/=cJqt;d= #n2O'Kz<~%FՀ ǼxA*ܾYczeO7h ^>!8ۛ*YpM+ WsN Eaw촂Zc)i`o4jqPPꍅРFSSfZN.0I74BjJn\C, z}bj-HO)[WRȃ'3{D@J2 nBg ,$ߑVB#k6nhMH,@e870ڒcK2d`OdB2MAE;AQ}H5ڠd ʫz=%´/PybۛWӸ>qWbBnHcHn.P Gh3k114>^&<]+q LS<>v3jk1)̪W}i!9Hoq8"JYR).< gbF٩HG12v㾷?" bgVnOUϺW6ωʕSW hI0Ä&uZçWrhYAя9 KZKɣuQW9g-49|[ۉ͝ ïz\ζY#b_Bw,~$`X FWmo,J޶k03'펓|"- VvUpmc~+wTI䴽m# bK<:i֑!tGdot6)[Ku&QBi{ڳGz{79mLFZvj y؞d,tP*P~׸ńv Oi$耠B4Uo;ߒZTjG'.zQK5a`hOB$AeBUͤa]sio`l5eRt˼/]}鲾/KR9p]Txsc$%hVھ| ȽVBY h9go_i]Qq|0 \*I6VQqZ4T@< Q&BH)pIVe,[Cg `5;` BýdiR^ q̹[BsʴzbZUqg9OOR悹Q0K ZG܁id[C3w O_ nc}VSneӴco4~]Ǘٿ!·͹.J.z\N:9Gk^v6b+87WLQc쪁n}4y3yٝA[Gbi(DB (aLusXRhWZ{JhQ EV(ŃGg(dÒFpÎӧ?i -@QKAp7K8\߲@rDk@(7(aY<)Y8y9ރ7^#i-^fSmnͨ''뙡Eݦ_snݐ`UHB]{juK!5ۧP) ouJץ%#Х00gފ"A"B.4;E:HaUX AVT e|\^.l}J:/V$(b-d7ւI28%%" Ɯ-C@o2;}۰9 (Ր 9 6pef#M\Feu!P t #`aTsOeϏY#B-?+rQߠb|5zңRFˤ|I ?)eLJ+'ลq\(qRE$G-BIЕ֪SJQ}H5$G.PJ̀ЌvT΢o1\Y˨ҿBXO"l=ǧZeUG\ نr Rʀv .}GY?nxT^:v||ڱAe`x\=Z /3T>qZOgK[F\$@n>F%c7v/X$ bwZ8pQ\\1ʥ wqG#+ Z_KKڪTQQy݌o-Ļ~163Āa-f_uʽw̖W7sXNE!;*{a\*'6lI;5Dg8P N- ;oi7.`}ĎJܧ]ԽznӤiZ%^yB>1\j˃SD+a'C]-|LponrJ|U(PjCBe-Ȝs6U{sS0Sp(WB ABS0t%T 1*-br:IUxC$ #VaX,WB!0 6e'Z5@8v]ֹ ̭̐w "̋`wDH`񇠣`D#N!p8$%' h>bPcOz9)ur!g{NO! ys:0jJ(ϟ_?琢x4jOX}]G{<$>m-?g?d8OՒTۦLp oݝ{P%ePSf<(unH` haO8VزRrSUx.EܛTI\<3UF0R/Ph{:pX~-_\wV7^u4:=_G{;yh6e&ƨ FvJڃ؇uPMzR |ܳ?aw @RB g$@9!@*aػq#4,s,d J<~d2~EV?dSoyƃ kbWbUXE`mkn$׃h #vOXw7,"q"#3i+ WeIwcɬ+2tI7-錓27sA)(I"\kQ|D1w4\- J MKt.8}ДD{s;WY䡃Ǜ' 0iƚvdo޴N*,f!Aݢ0hKF}cż4ZJx+&QGmj"Z#\ x=@]"l/ ^~*`85ʼnDj.!xBP58**4:5\/TDqA{HJAʁѫjBKr*_:^ k5P].R[(hzUU; n0*Ai}y1ՒHP-} 8 {'јjEPkGTS*jdqnAcP/nBďu}Xqq5r}~u'MZ:.JwQbbqtfڎ?//Onz3yz +JJOWn]sσݔ^Mek)8HBqm%SO~k=n%tDi;F6yz"ާvk_%Pu!!߸)-'oqh7NMA DtڎQGEP^[/ݺo\D˔6F2jZR8 xY2ߗ&|5 =k\CxMrHNJ9IOSjVNzѕ 0JIH)&1`(Bp834Gh褄$&&Qcp\i&T,9dYe9>[ښTvYBNO!1)3>cv\j$ S)gi| X>EƱ;D["R>Vzl Kdw[ $߰)2̯,Y*&hT jؠ}swsCvP'~ T:ߏ5蚹"o;5@o y=LdQsu^MEORgGu }K+(7?O2JygÓႠcʐk_~y?u)a_r.ן_ݗt;Ύ1aԁ3–Ɇ2ıj0O]ut}M0s&g2YNuay:O|F많нaqKH^:e\y˃N= u*ϕCLs]n兺neuBÕwO]L(ܭ16ߥ-{.m=UZjO&+g j:=!=s_(BAe :sԟ__o՞wAA@|!1C/_#4/w锸K箤 GW'~ cSS^ɵa!s.7":YRݤ7N`dS8װ Eg꺪t2:jbX`”, ]. {WK]psG!qzʒut ;W|3zEIQ2GW4qzT`yW}ib4lT0Q' +J0u+7?Wt'Jݠ4浲zHQ%sVt[Z2*45TT$`Ai,ӄSiFfZKLai8R4ƧNC~;]ɕk־;B*>mG2u]U8̊B!忿B)J]Qk+YU=AIgL)fs8t!7ϫ5,^-qޕQxss52z<߽{׮  F7EI+-c}赍%R$cumʾ] 1h%vZ.%\nZ LRoek辉xD0#= wV׭Y-o+mbI]_xj9oT(TwV81wikivDp7,_ .: DgO>˶֭mʣf<ⴼMTɼu'Il~?ʛx#Rє8 :'12>fS`CrEZX?f?lh<-WS]n"iuW8ÃRya.U %LD+%NyZnulԉH5*nY>΂)"Ah)DdI3?٥}0Nd+21`RhMrG+kPU5|B@ztxc9 9Gu& }VH ՃU%:[!J6 41I&Aj$z8IAF'HjqTQ4$x],u[jcѠs˥踕zvm&唷S81\OZxz^\oN< ]ҺS+cЎ{=Ik ;8Hk'" fߞ򆳽% t)k'( ܒ+q"CmW%^s)b$Uq.V oVu9ѪӼ& ;&5WXX0'pc;wɮ 7v XI,"εt$TZUVXrC90n$^%] 7̏n7_"cuEE 3ޯ-|WV7xCV^U USWCR˫nC Q̕׽O0fNzЊN: ͉Դ$0qekp)$8S:"b˔)Ђj$7c*n3k geV'0ԃ%wO{45Pzie9Ko4cf;!N8CftfqxAŜ#?"%1tLTB 5m]b@AMIiDiRq?Q($d,zJPyJs1R9jOIױqpu\;c(H+EZPb%MRv\JF)jXkF_GdF.C_? u,6zS`[9.燇W]w._Ȝ~ 7]/_:$Ȉ|Gf?zcȇwpUDwv{G;Ba֨jnzNK=69O9g,?-wHwi{g#pO'w-Pv|!gߩ7@crB(yv  ,N.b{ٛͧ/8@A&sY. ,&^02`f-9-+hHGv{|)ժ2PT9sPJ⚋9@Ux,5VD+(!&2TQUKK Xj^@6G ^4 ebι Nod՘@y).opO-/UatRI+8Tnķ3for (%9/uù/;`Rv+쿰Nm#%MfՕ!wozoUT%T֒?Տz~~jF<u%Vi.U@y2:I<Ҿ sj@Ju^ eP8SjbȦ\ Ώ JgFFMAjaU 0D@ԛ | ]gOÐ~N!@aLEV%4(jQ:k9Ք).N7A/cPݤ\Uǥ Tkf |L5ץKh+}s; <ٮ/,hʡ v>?,WhAj^ ,7iQaoM[cqϞ?h;ƾ6yt#4֛hUMUe|o-jqfurF?_ډ^u )DtnSbeZܑ|"B9D 'b":mǨF([ݺo\D[ɔ&8Ob#jN1h1Jn)j.$2ť`ƾx46|D'1I@I&)3i YJH5;b7[taƥOkA 9)$ZɲÐhvυD[{ p"1 K-`TFC:vpT7TԎ8y=gY'CZ'LŚh7'Bp i L1\}g DL ;!ب9M4zWjzTv!xQs1 ,I4ױUQi+_XƆ'7IR]HjD 4#PJvzVXF[2U|-Vr5 ꗫM4"'kSpA%451'3sЂ5~AhDP.ߧҮ8n*X3/e3X`'W 0mxǠ~9Yn't sdbM3I^jfKhSYϼ$$}ٌ Yn(O )"% iq)u-k_fJv-p=DhՀ@BN%/󿝬fYejps>|%]"_\=:7SmVx`cx\ ndʌuum0D]{PþZbZ%uf}55)|E"Q#]AbuףnÁCWzV*a{WYvڬ!VR9hg5o~߮SxF=vŠ$y;oKyvw {?v)-SFvAt}GWi塵[TBևEL:qDiBuŠDtvۧAiڭ 2Lo#cuEis%OR]kErN }L$KWޒ9^tEQ:Յptnrry~|犺Y۹.@n:?]U.8ͣNa|9(o?Ŝ޸wqnOf7w7#=fׯ^qa&h,Z$ߜ/g_Әn:Uu $ ?Yӻ7f  !{5 *TjR5 V)IɕѼ\ÍO^x{w1[fiBEhP[,$? p+-g_,@6K[_7v+/QRҞA$ ʝ˚_3C3͏u]_)_u뷻u+턃@?0|M>6kc}91%!"1RH[rrs8٬~ŊO..rGֆj<5@ʓ±Ryۙ3fIl#q؜<zMAV|%GgQvTKa̅hƓNgb>Z3JkmP9M5WΧ3~vm@m-9NyWZVXPxIKb:PA`CZ"W lk>rdՖw,lcXX^ʬ2P44 @JlY1G{UI[YTsL@^zx!N*P?#:܊lB]LރdZvA6H% [d,'O8a̪_^}[ݸ:' ~Rni|/77 [d220%5hsڲبZ&՚b`w+XŨ &!wB/4Z:\>>H$ 8 qh#6&W%K6"t[Q-QU0vwH0b mPz{ Z \!gYR6tRh*b`yֲ]ۈf\)Nj1c-A N+/E :Ug;\D S;_</y–0~Y7A+:;9mD!Da&6H)T}$$VPU~@%b1hF90nE2-w<].9Aέ=팞;p-yWS _$%񚑶׌Yp׌ t&\#C3Non~Ūn†Kj-S;vqc ڕ[If&%;1a,70AZ6*ZLw_vmQ%=Nc/| O޽]0 `I-Jp[bdh!PaF4}|4JꔴU<-EzK"*B2;^7DR]vVA`$8R#vEdj~GH<}/pΐ؈Bń)eA,˒ ɢ.]ǃ 4'FܠCj"b$}sE2mzD˵8ȋҌt4|T^%BWE+XXc+;aq5pB .Epv8[^ل+/n]BjкDKN܅Z$CpV;IN0v ͣ[N@pXǭ)L\DLIzƹ̨rߘBm򍄾  a#j@d)*t -NVdJK *@A$4X ]@D le n`~6u9uv헁pJ?荘FQ)f%j%ŢiHd]޼9LɪiY迓w?}szΧ%;~77qB,RF'veNq=IicϳjJڡszrDwWW5 3Ywd*6bBn𒟖T"3n"= UQUG޻ek4:e}EwCNjO~s?d7YVoLֱ6B?=J B;J*R>JzK.-V_Nrƌ2Njl%$##0TL p.㚞cƩ$&.LJɠT\k PCX`B6! ^4RT$e-J&"z˺@h4|PPJ--ӛc|AaBByxᱬlLe})ZmQPrVhb] .& IoCQs5OE\3ߟmE"Z򠝯Ѯ"i1USՖw￶جR+xؠEL,Ov,GLF(bӓ # gg430ڑ"'G}֎3iWkn(ƇZ'GMUٸBjrqBj!N &~5r8k_.#¶6-C6#\uyۋs"۪cꋸ}{/\r>\.וv)iH1EH.M$r-#<󜃬%wύ_aK#4 `+JUU]]oٷ/raf02I٤IC1Okݮv4u@?;懂o&a^1AonyC&4Jڬ TM[^I{7Y珷fo,SSZd,KD\XH9prlj T8-A@\xڝ@Ig0pߠM"($*dƔfrJEb4Bj$ &%\p$ߠy3&whj7rƸleZ*rB%ʂA\LH$I '{֚j$JKwfv=plG WA, cSO9WS9w!%g u(IF) Ӭ!'6p(\7] `:3&H߮zae 0Lz(iG&ܮWQKT .UT|ҼȞT|2XI3^lo@ +6X7"/n_QI}/0Ò$UOft*׋3(  !/ rӚ@xg oެY9a)1嬶@Zk0׉w"DT &B֋࿐[s*vEim8 ڰLrg/;_o 3~c'W1pԁh翐: GO/_ϋU^pe͊weH${x *;ɢd|JWlxV3/RY~܌ׂvq4Ͽ{g<?Y^,A"aO;c |1Op_.[F >#^ 䢐 w78]B; G`e y4a&'8)\% 8Eb,G<.[`Q Lq+[\nq5߬:\6iC{un #$n t ~s :~d3EOo~(!Ͼmr0f Ճ.VTԪ]\&j9[u.8+^I5:><➫ f)@CΗ+0H)jMrѴb*mGeٴ F1&Y5+04,=Tկ7/w@=7l8D@I >\f[gXNsiQoV|Ii/K=){' (5Yo_{}=|ztLFOcp{x0cF;R>"4_ fTWѼ X[5" '# (nh['gFD$5+56;frѹN;R'j\ NKPuW14(8:32@.޹c):̋ Uji3 ?ˉA(+؆ ݲ+P #5LZ ٞju\KߞjZ}OfQrQ {/\}W$Tk ^+>êH1J̙T_*ي$cZv% wl/&sQfnOYh%^͏1ek9Av[& uƫ!<{pG ׺bOIv W\k/͕_+LIB~"Eʡ7M0vڭ+1M|SGn[nDj>$w.do<&m1Ziڭ{cHևELnoˎ8}u6Hax]9UN^tyqv$էCluar^}A [|->p48ݫ9ݾ9<,JXԒ>>ۙH^̰7h+ZKx:OX{Ԡݽ6*uw6^e6}_,q-_:v*2"K/aK8iubJ-pLil zdKJrE/2։##GKx_(7Ŋf)Rm?/oҚ~=N9g1p/:S; nv阷7۝AfԬ,F`Rd0f db n$z^E?N` =nUvqAgGÑމSOIjd Oc:g\rR&:ee6S2aDɍ %BP.KFvGVŌQNgIU'e4 pJLBJ=(iՑ@7,%fTK\ м:Xh53EH q=*b=iUc@5ʈ:*Ѫ`,{sG @ZpePx1j=gҨGi=Ԁ=)}j$ K,Foo/˧GdUQ )z/.&s–i˝=NT6:rʄ=g2obް-&6hj%l՗T2(lxߢňKT_zԟ NJVDqR՗T{'NRzR*T ͵9)}ue,$(KJT)ۧb.-FT,fN\̶ݺ4[>-b>ߝ{!O *4 rekWp߆7'6/ڤU^$eI{] Z^0"TW;o]Z3 | Oӭ};/|?iՇ$5NI?;lvΗV?t7W^izcWӫD]knİhegA' ?}oWk{jARMQn1Jye\>թ7Zet7VwYVƄUT#e&xԎGƈ1L~7xO6ᑋ!%%{76 BrUo҅ >>ڼ."`IJ>L;7W3ۧ巅QOտ&V92׵['&VR$<; ¥)cG: }L'iNi#z%)28ŒW7eQ<ay'e)rpH0G-$c S'.8*z9TM&)8闟B?wRKڹzz=5 `~uC!yW%\~Y?Zo8w?.1=&5@S&uG}dzI~3_,+|z*fnBx4ߞ]р13u^/Mnb*,tth]}P ?wm~)Cއ(9邾_2.nz)'Ri\6"R Cr^Amwc,1j|,DFa~՗Tz+qQc⌡GWk&/wxƏHQi3^Jr{GnOW)qJIz8zVAF)8*dP4fi R ~Qr N8꼺.b)E.f _{Y1Sp oN,4a- Vl=唹"?$)&3 xN#q{T_S-r9GG ~EdVSAq [Je,Mm>TBTH)3FȶׄJyvS !},GPXRjd6M(5)K)F)X)~wH?:T<#[Y-+CS}NFXBM!ԬrjYpcR!Tș2e{R!㤴[J'2T҇a񥊓҂jR?jz Հd@f%hiCG5Ժuta^"[u OVؔh+EXMوpW',u:4&I^atj%) &cM|cgRja1kSY ƪ;yM' SwM+ gBA!,RMevt?i]@d@Aϴ[Um|=Pm ȡ2 Pؚ}F xw©wo]+sw/SQKNaJ[2jd:⚌Χo|TD7'Psvv*XG(x Jfy={!R1kʙ8fP{aWN ++ޫ+'}-tJ8 jLIwn &Kd"ABC #wϸ`@"T/21ҴvDRP vhPH  @9p!198!TA\sg4@[Fk2 ]w !o2d&y!w(#DB,i9&hE1d\~adOp`I)8gu*8wFm<,%A4$_U&-)zP\v?.@G(C•A\G:Sp>`3`kMK-zwD,TJs<7L}Pwz $a5b [|8?2T, {{/FIn.hP{;o~)r=8 ^3 7dkkC-p[^*TUʽOWzM%8bJ;|zt.n&@fjE5lދPGP8'CC[wsut}m[LQ(][o#+¼^x1v%gϞ E^6' KdՒlvm%j7bJ.77.7@f+D|[F+_Ad~lo_z]r'p%d é`+8,Q9S1.([|D >򡹪XonԚ4ŵ Z(Jf}A-܍E39l*U{ogY)Rxa9 /M#,Ɵz*OC4Kꪦhfᖷ(rV¯x͢t>eKLM4ɦ:5z˻ [ bL'u>mSV#[ y&dS#-&;-U1:F6.B 1wK/.n9,䕛`jӊ(}ݤ<><҇VA)2A݈<=pY%05v8QPIA5Fkm(UX3HKT%e#,P#Fu!αTS %}!tV8T'mTY)]u(8+mT'+}VWEWGGVQB9YsҸ 'hTR qVP dJY)Ro%$"{ܗ2,Gk6ήiB]]=tŹe58FMwhKfB);hmR˭QfD 3uUg3 _f~ԾRE W+J{qXcE0љVl!MHS*? ;TҲTҭ?NQjQ%X@B[iz*R[P5p{N.OoBȍ'6 JAB4 KE7o[?Y_~bDtS?8S 7$}c* S񩔟t H?2]hmFC;Zh5/ jE@,2xfjqW4[1%> 9at4HG@_}0HAZ{cCh$tN" &_l! y+Tf'2c32k |lo.|t(^{}~ 1`~]{KIAK2q&xP]"ղ6bFrcC35ǣ!k*u:g3.W넳gM;lsLְΠ7b|V\xK*K@в] Q%TBG38Va)*L#%Zs@f.tV6G V4W;AZW| c1Z'Fsh+E1 2yҲA@'-8L" 9̿VLxiHB(.ihng.DNY`W,'˜ۛaqkwWD"{o'FJ`s.rxeÅ]zziטFQbȂ]Kx,>lרQ/d߇b'~YHhvޤ|utSKada1z֪m$ XP*<@e$r=SLYBceRZN!RSBoY,~<󀠇2SPǺޮSꪆ8KhE"6C DMʦe+:0ՂKtE4_~̃_(24l /]_-@ .$ܐp ++ֶ/O=U د˗=HBŘO]p1sVV '$tsra}uCtsWC̜tS̲?́zDHv:K'7psʅB=&=^*ɳ$0ĘU(%!@MM pcHM%D:Z0Q )T[3374_eIQ{6)Y'Ȑž ڨC&i 2SW5c>^~NQPfTYkƒ[QqT#@8GeJ"5ښԵ 9˜T©J9?" KjnGp9)_LԒz@݌T \I4Q'Fj@JLYRZ)T\HSU*2& b FM)݉@0/؝ (Ao f)&J>L,/Fc3t_' P,6C7L8L 2tшwK!9Dn*+(0K\,?eSы-DW!>24suGM!rheŔ1bUQ _5d(_&éT]^s緹U͵,0[]sTF~u@5azyE@dAWzӭL3O~ w3?.=gнo7|cGfAHLe1ڷs"0>?jyi6onxsL5:oϽU*O+SVueJ*Sg# wEGn22Ш ,{}Ѩs~‘e=2DG>CO"3܇=eL7eV vɪO]rt;Ȼ4s /6TJNugiéRqdξjo}nK@ByӖpJ6TKFi bZKѧ:ǶNiz4rM)}wSu9wKA餎ǻMyx-= +һ尐WnI6?öwS([ bL'u>m]Vݶn-rlSTz? nxpe1uRz]|RWQ-XѵΑbFx1^~*bBu'(Ղ \d5JE{x Bq>R&I7}uӊ3HO]5./.M=?6ݯI&=p.iɐZ&=29k:&zˀuC?;Gv nՄq}WI9#Ƽ|(o^5E?ޙ`i~ mUa8\K^U"85㍪JIAok(=1pQl>`5VR:K{H,WkHgׅ:l(%>^k޼΋V1^f$r`C%rVZT܋M62\)QHEI7zĈ ^SmqK1VȒV%;dٛVO(ճ쳱 l4uyu#4 8:[ ok\jjs{;YA|Cf&Ǧip@.+Qx%VU[U3Sh[LMׯ><7Ni$3i /id7#)C5 x! p0MɵZXt ӡqTT)*/q(2Av۴n16]?>_qw ߛMYuگF.wz,Y Pw߿~ y7 |{~DlϿNviywU d@t`~d,[;Gާ!;B/[JSfv]( |}57c_z//D*)"C➚#OzQ,o/?;sG"7BJxUЫ768wo>q0xπ|O܍Vܷ 'v?es_}6{@$\:s‰?;F&̥߳JV;X\-%5QXOim_XUkaDA?QY&)4ASTRȉlε< 6I9<^r?ϛ{Rgf|f~!_^gU&&8{69\1};֑zN #wrxNoLxe\,]cJ9 c,䕛hMi8oۻy -U1:F6a4#h!ڻ;FzM4ɦ}nVDŽ-U1:F6.딏K@"[ y&ڱ*ĹԚo&=?0|"hjZd;歹]^fQq~eXAW?o}ȎkjY;Ja{|Ok:^72 *Jvﭔ2JrRK73I)CM2^hZc`#&-PU˅7!7Ɠu}VY_UjsUK4Օ_.)*>dkAz(0STMA?JI'sOȸ7?{׺Ʈ_ea>%78`ڤu.ݦyC9ک3ī( 4v2(^$Jq :fUVWS#.z"Zx;eζBenoAch"6i[nJ9N%?s󪹱➟Z7Ao  6xMtW !o:7-D9F [%Rɱ5k7C_7sL[x7 ?7 |^ jBSDR,F&FO&*_Hmrs wnhr>#HDY7y{,y,O0]z}=o2z"N{>bhOДaL2sЌz`)oX0p]ωeq&\"" ƀoB:u:30] /ƫU@? pC]pT(Ԧ8TAsQlF9?6< &V0mQ4;痢j #_%pisU:.E㬮{dvΓ"p{=xܗCz$.Zz}|nϴ3Cb6M/mj*];Ȣ5'oe/^-Ė QSQxV!I :w/ UsjvZVD3 eY kI-[\ɆnPEcB]*`*h+ԧA g^x?teL3D0i0<8ڛR|  c+m_as&Wo쇻ye9hq4]ًhaysΔwY^&&NZ.~wFDߏVX4L}*Ape 5i.~Ή;:hKt피{8k5[xW':>O48V"UͩH2Sϼ㗡܊ׄjA@:Q9z>vyܐH&1lfN#ÛE = x[“CE\-f˶1?k]*1A YWkuնSq qS߄Mp B@FJYC0qt!,f ~wAmnp` 8o^P/^Pw'kݻڼ3.#nCz- 5h:8 ]. 3Ketu!N=XG}$ʃu̦$I/|KI<sf1f8vv)0p $wdؾ]88ca>caC31IA3N4+N同8Gw8; k:jht-ԩG,cԺ\~XCp}cg,<7`?p~ZI ˟Q<|RB*h%nV97cfৢO @ͪ<mYq_}fD }ZDSMuQ@hTP(Sr˫a7uyK뙼PseP;9]wNA[qDfN b NuNQY%EM+`۟<`0[UM`X4.hg|X0hzg0bV`leN&śt.z], cIRlV(Anͷ ъL sUHvia~\H9.f*}\})(DTgQ'ײw8? :БHp{uIɮ'sMݦOmJ'rdUϢWWpXO{:H8]<@3gl OjZk>\Nau(NrƊ(0W)qf>T/bDLK\ϗWbL M,~c"NBVfw 8oػM\N?ޕ`  hoսc/3=vcu!ZK Cr#d-Օ`1xưΌl(x|r7/kP\1qQ*UFDI삚;Ƣ ڤuE# f,7BRr2ʖEH*B5Uc"ϟIFmL pnX C d(2:hK h|0m"T^Bx,G9f엛.뺻//(ۛH(DCV<"BJ<5Dv,U_cQjAVal RAIyU ֊)RzMԫ\.곒BiÅ| :fEQdU KrHf?!t@[F$mS'"Ol~|rya2}`jr)gP9e53ٜ|hcAL>,e,ת~;c=X%["TXke`Q'6}/x!Z7Ř~@:YW63(ƌʤbc*B$T!>B ?U6 ;؜:qA[qѷ ޥY.a"f",Z8h=Y` , kXk)^$(9X\gM1І|ș[=h3m?zW`׫"6{R %ppb8\.YX{}^dK@5|e gV-bv9$SM8x-z"gԙ櫸eSg1hKN,|Hc y){Pka07nHMNP}ݪa\%ʅ 8bUιr,yz~9P*lIQs_JJb$l?FdCO.8}/ϗ޶m` ?U Ӫ~ҽnW b13b,hs:x׮1}|9ރjDtA[!: b[=L,Ի^u`\+CRGPMG \ͷ֣>zn,6Gn͍+LՆC/uWQ`W(5suXd֨+DQ=ƫ?~ax9ExHGr_x.Vd# b*A~Dtz:ZVGcvg'dEA{NWs42{\[]߱(ȹ񣫟ohA٣!~yuqe]Uk_~z|:)E(:,e4h7S]o6G DvKGo,{Gãš{ q)^TbRɥUS|Ç{V ,0m8 =_K{lCØb}Ϧ?Z`b*;3•޵3nnן>.Lf5Dc%񎣩I^[5bU# f'h:5d=xgOp`$'w,@t,mNig)Ң_1k*AEĠ*!5% 9`D7 Oi ]}YO^KcuB ##*)X~\g'loc2Oo bpkD439-qpެYɟrTie}f"|* _ᐼNa9ғc۴>j~}^X4ut5=02(`ZdCA1o0stC/ԽZz陰"#R\2Q?Kv? R?8)m95WWz~){z,.+r#"˶b^jz(6Kwɧ*5ew @;2٨=P#+?miÀ>_lbq>%ڼCxdj lLӿ:Ⲵ,X֍Zm,!!lx;PFtO'x>t*|ℳR8•DbJr%]V$`|h>M AsWh&邜L%dSI}^DwS;*!Jj;b"9] $FǺϧf TJ+Ķҹ} =X,B-(`Eg:]k}lɶ،aM@$< Nrhσ.I\)Q{ZeXwx^r@avx=Ngǡ&,$l<+=1Dt"|::Pk;ڹ-q'eA0f 1#x\gSq9K+.V W4&Gi]")pW&y%7؂Eg%X>%"F)k\q&ɤ0.PR|Ƽ*HfDX8/F< 7WlB ҽbUV9:)JDYu(n vg̝6c/)8ZnaY-$IL)HG 7c3;5XS+)$q*3Ū `!X╳`ѳ9a͉tp)7jAijxds1nlOY]6ԏi]Az}F>Gg+DTuqI6b><%Xf-saYZ|fEdā3׻>gAeԾy,Qunvk6>B.}O }{BW 3ޤq[@R;z=ɻ޼jͰRp^}sp9pyqЪDiuRKfX8I1zj,E^f_c[|WK wpk`VJ/`W4 xkxJIZ&fW u%G"J -Uu]3Ϙ]ֹ99}Wxěbk JIjP^0pm$^&Ip2)h1 SzNbCu7o /'dz6!'ōpOsWAU6{k[XqX6\:w]tXQF''[00(vSJ-ުaARApylr(7>ϖr^mA` f7Q͹ʌ{:h>j{!#<),Ϸ S*sX*7R@հwRG6ޗrƯ[53{OOU[}EJ8S^XxJ`^ Cr%&ŢBaIv/hYLK2JKG#iȊ(c_r ZJwbp7鯋onL'YEc^%SXAX3l.ٮyM ~rIee"sN>T6G_$6)땟 ѹ8\^?z-]ۀ?O'ø]CN&șUOpZ< Y3"XxSio婢V[YI ӧT9{"A8ꚽRl/!{vU&2*-R3)V#V7-$:8xpJڛ+|Qx+zABWQ)TaD=2#H v]uo@?X'+DY^q%,&9<LJv!|3LuQbe!FOG 9>=N#u9e1:@˧FE c+XO-EbP,Tc 3LISEgbFt95㶲XHO80%{B`T P~;vP Em蝕1/L:p hz@(eˈ@lbva{|xGm1NdRlٮ,,y걑8^rC@(QfayW@rJ܅`Н+̲yT<>(<|p9\[ZAuqN>9;9"[Kh < d\f3A=G` Ivcyo 㓳 syL7p:)a w< mx`c_.qYLK'Rx}c[LDyCq.ve{S|g~Op4"W9y7[WBqǔc10>39ܘ/\|O?*49amtp'p=F+-hXQ>>2˚3 돓t:pt-ẏ}?5TqѿځkNq?\܈lU]P+ʷx ^52_N€ YL•V hir`nSz#eW?؆Jkn/uń#?Hݏ޼˞__v֛"0&텒w5/ p}06J̯ORtsE? W4)Cz|֛NpwClԳPB}?^>Z5]h6s^x<'zvW#|JS\ַ;VumY1/=o ?2Ћ'4:[Ɯ"3+%J[=X#|}O,[,յwfE:bz\Cl)gˏ$ _O T0/:fƸ}6ƻL77҆3ҍr؝\_:Γu\  y3-V 9}Pܨ \ma Ch6R̟O@0'7RjUJɍ-Z&͛_ꗗ?}zW.6B9Mu)%Q,*k.˜ 5-1~c80)& -Ÿޛ?~I/+j4Nz=0$xa*9E&/ [3ixP)[7 B`wgc@sVd{ufob`(0{[z\ջ4vntڨrÈջW^r?(op3d;{3&K`uc-8nog Ǡr^3|H\g$ʾ L_kj::~"z˥/Tp<[yf2~ 6vkCf[tM'<>x5퍴`g\3<"hz6#YJZg9tJ.(P z?O&߽Ahu<qzO?Z/[prq|'nu!r_?96^QcI3[񽼼' < ?h ˔: ;WIo2M^wOL{ ߶$ $n>`lM/8Qe؉kH 'Yߍkpn$4yL;m-Ɲf`Ŭ}'ZOt=Vܸ.D)q鹮g sͳj+w]mzŃG|~؟"e:GޯnsF \%1iëyۚo/n|6kx@>h .xRLj6CW/"yBkUQ= I#g_F[/ CmU Kf<0Y&03ܟO;m Z)?4~cjjm Oi PbZBm暁s$ąI>4B |ӣ6‡N, !-]5v0OZZxq5˓<6zKrp]Kw\yZzX5A.>3HIzOF׳"/|64=NW8_TT{.d'g?\\5t6*,wa26ݱw9l>a;,O4wP| u 0\0e\W9$yz=-}( sLXYg|qQM5@tr/#ZP :<$k4w,L=Lڗ*G9jMJq4KK,; Q8By`ݞt()mVܫΑhμk"𧏢zݏGP1['xey|Jλ'^=ыmh4X-| 8u+j1,j=o,g,3 )Ouܲk{yynȯM0}9Vy(rێ/.e:f6Ww1Kl%>M><ԑm#kgN>ײYڣ=]R6?'P@L͠8f<:E2d((kFI\GY!])sZL\}1<,$c_mt0޵rͅ/޻;aRgտ/|A9Bvd5 "[_ΐ{sM^,~xwu5@|Ͳ2Pr E-̀%`V^k5h(tl\$:[HÃY #KA('I|nkh/[u-J'){\{~#ϰ=j~x .n]3uŻڙeVonbAF,ȕIn1߃|Cly3.~?w!t-<Nϝr ݕO/>+_'X⪻GlZlj:q:՜%i<2obےҙɫ0p2zae 8(FP0 k(b (Y}NI~7&EWtw:; UQ 2rQ9O mHI)EZ V[={gJ k' sZjx0^yLǧavQM~n; ;/(Hd0I6d#e2j)]CBL?LsC` H쳼r "z :Yp29k!TɑDU=žal'B&a`Y4қ]Kg~6-YɊ"¬](Es ?]hu1Pe)e2a1%W% ]C~]q!JR&l8-Pi%j#vVF!r˫tOy4ꅈ=,a!@y=*9);AU'j+ߝN7%4R+ꁈ=*k>T6*@C%$حH:aSb2n~AH~E/!آ!-{7~r)\*GeQ}]g ]Yfͯ@mmwqtM +9kV Z?k1<ԆA)Y_~xWZQO|]E@%hC`)@kHQT߭НTb/kx:=a,ÏD -Vz ̗YIF$|2s*-/#ÀFHв8i*r--MɁv&Qyp&)N AlK\M>D:PAPRpzW㚝7mob-7{YN1]h3z{II<;џӼ‘K|QYܵz\3>ӑ(sc =\{LEe)b!P$jDo]b)L-[W N>] 5ZEN2syb%6Y9Vjͪ,B[hB"L =qXL_z7Է*[-{{:_no7:뉟W>G)5 ^U_vٟ7+rՈ{=xPYFk)hU2vomGH{%S'Qgblv\RPI( dI-a5@:1Ôn?ꝙhMƞ`l- VQRގd od^gt["ZD>ʔ2}(Mq:0\01012Ja6vky3-A B6y_wf nƀGFUF w3bٿrW^X%:hB&RUŨ6AzA KY+*>O+4jCaÁ(%qLmc#j{JJ2*.܉R8 -B x}܂ ##zIaUGIɝ/z>QM^8>1eu‰e\%Sr4ޔ\YU^ Q["/8@>2'~ %] 82Gi>.ڬA*ёBI4sܧZ:y H[K@9O[kP]t>I@ȟGtn/݇OrP30A<, `%N#"=ebOH>22BTI#3܌FH[ѓNkwS5P=/@N(95`NM'YH:d@q!u L]R"ssq8pTs#f gAE䒊`tӓ6oYWp#yJcz$s[Z9='B̲BnT[Z%>M{,p񃀥`=R++ղmӪu/-JyњX#C{_4Ɂs8Vk]򵝀w/Y vVImP6؎X;_c@uF 4rET\mZv=GG'&؇UX\\3myz~զQvW}yhB4d; ;8۱ȊS AW-YzC'^b)⩦h 3J4 }TYZ_?JC Ah}TYڨF[D-U*GQZzo֩vBQK_bKsɑbK?/htG[TzS(xresw?~hƢl'be##e޽Rg.bޝD|Îۦ_7OoθF8<ɴ#=MJj%9J+9ʗEg g2.3%䙖 R4fHyNѐQ(gn><_YFy[3F̄ܪ$ڦ,XbuNRsZ(Ypd.VQ.* 2=kgN.w.Uot<_уlQj֘ůyjK:`цGlfE)) [wyt{CQU oA>K]I}QۃjKQ*z1gU "`cBFڠ?.&mmۄx&) m81֕+o'r 뤁q\ >y;X4F&){Vt.04hT^.ХՅr!ܫ%<"ܣ SEQa:`&'M(gTޑ4Ӹ{O>SN۴$v.1)K| f8~2ñ x8f V.s,pd{Repat]̖7LGtX\IT-8?}cEx KrN Ct4A"Z.ft~-3Е@>vK3MY< ](8bu*="D;w;df((M7FZx-`W? ` *21dxQj9 p7uKK^xH^t[dJrd\DI,\Ce1Tie~ /k6OvGTUyD;\sriQf2N2ǥHNKRS'> hĹӃ1d Us'yĤTD"*0,BgZE89XҜ%yu ji:P誑ӕ(œS>[ɫWXӕqPPs!͙KOK'!(R VyF)$4Q8"Z:93%@XcYS@lrrAB!DՐ7 6E_t.f#zv1n_tnr{aK; {;Q :ԌAȐ]̩z>~Y~css]znQ:iN!PP#}AL{r.zm/>h5h||kC{eH5.*zxqfˌ۪qJϵNݫ/nۜzZ"gB@qe4!,u3[dm Wԁ̸HHv]dD{}(~E*2-2As/}hm[`33A%gox-g)9åͦ77^ޱ\V)14HOS+F/2MkwZF'ij(Y(axo%r[j[*zVI ?+P @ZvzH5$r&hj7*sVuE5,+,/"W p7u2q|9"4O7Yo:T/8UƤCٝB<Dɾ\6u_sv{ku'\ߖ~+ |,S^amH79 ]tAnq.O|,ڞ$Jm=]b߼2[`x](#q 'zm 6Rݳ9fxRB+84?ӥZunjrQiATСFK>5%< AQtUp/D 7pZC+ ;(>!Uڏ+{7*]TMnVLgkƨT=iY5yPoA|=V_\_xaȯ}lu_l"XrI^>GMsRGse bV> ~R- rHJc&i\˹,b,`UGk|d3ڔ?A52*~Q;cIˇiӊ6NWy&(љw໇;(\Q,J|}ȠߩïQbgs#2?^ΐNwBCkVIy*8IM~lٔ#b9&h ~vyWhWo{9\2Ы2jWX? Y:PT0 JDx-7JYo܍nTNH:6jib8&g@\*c%5QVd$I9 S$$7Fdgx" X?E)wJaSMf4kr>ѴG;r2O4.Q)M#Jcv-$)i?.-R34g@HʣRLL+>ROXwՈKs45`fVI|$<ձRx+T&%15絍šV(SPPp>[wgƍJ)%)؅-'a?M{O;U;HnhۚJ,F)ˀ=S9ďp]bA秵+ds<.M5(HЦMIqT"G4ژ}܉JgY{f)چHx>SE _z%7?4Ek<ri@ž҂GƗʦhLDoA*I?)_|g"g=4.SisNvU( vw=}·}r7,r@ɳI9KinmkH4q؄CA'hA#6zB KlVLzjޏo`Ɉ0/kkT lEowc/(aOH ȋM<QϤeB3mvn܄wKtLp2jM\k#4(_yϺYA@5}@=]1_vO[;ӻ~s1(poX<͐eˋNfK<7/&l|3{0 d˫NvE< ×Q?Amm2sPGsF;tnΙмZt-_xծ+im|hᲔ)`Sq192.168.126.11:17697: read: connection reset by peer" start-of-body= Mar 13 14:05:15 crc kubenswrapper[4907]: I0313 14:05:15.573585 4907 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:46124->192.168.126.11:17697: read: connection reset by peer" Mar 13 14:05:15 crc kubenswrapper[4907]: W0313 14:05:15.657170 4907 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": net/http: TLS handshake timeout Mar 13 14:05:15 crc kubenswrapper[4907]: I0313 14:05:15.657345 4907 trace.go:236] Trace[431240111]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (13-Mar-2026 14:05:05.655) (total time: 10001ms): Mar 13 14:05:15 crc kubenswrapper[4907]: Trace[431240111]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (14:05:15.657) Mar 13 14:05:15 crc kubenswrapper[4907]: Trace[431240111]: [10.001747778s] [10.001747778s] END Mar 13 14:05:15 crc kubenswrapper[4907]: E0313 14:05:15.657385 4907 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Mar 13 14:05:15 crc kubenswrapper[4907]: I0313 14:05:15.712619 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Mar 13 14:05:15 crc kubenswrapper[4907]: I0313 14:05:15.824473 4907 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Mar 13 14:05:15 crc kubenswrapper[4907]: I0313 14:05:15.824544 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Mar 13 14:05:15 crc kubenswrapper[4907]: I0313 14:05:15.896557 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Mar 13 14:05:15 crc kubenswrapper[4907]: I0313 14:05:15.899549 4907 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="fe6cd60e8033e3165c90914a90e482e2378e469fd98044ddc2488e020f75e14c" exitCode=255 Mar 13 14:05:15 crc kubenswrapper[4907]: I0313 14:05:15.899603 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"fe6cd60e8033e3165c90914a90e482e2378e469fd98044ddc2488e020f75e14c"} Mar 13 14:05:15 crc kubenswrapper[4907]: I0313 14:05:15.899745 4907 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 14:05:15 crc kubenswrapper[4907]: I0313 14:05:15.900566 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:05:15 crc kubenswrapper[4907]: I0313 14:05:15.900614 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:05:15 crc kubenswrapper[4907]: I0313 14:05:15.900632 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:05:15 crc kubenswrapper[4907]: I0313 14:05:15.901293 4907 scope.go:117] "RemoveContainer" containerID="fe6cd60e8033e3165c90914a90e482e2378e469fd98044ddc2488e020f75e14c" Mar 13 14:05:16 crc kubenswrapper[4907]: W0313 14:05:16.064732 4907 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:16Z is after 2026-02-23T05:33:13Z Mar 13 14:05:16 crc kubenswrapper[4907]: E0313 14:05:16.064812 4907 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:16Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 13 14:05:16 crc kubenswrapper[4907]: E0313 14:05:16.068329 4907 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:16Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 13 14:05:16 crc kubenswrapper[4907]: E0313 14:05:16.070169 4907 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:16Z is after 2026-02-23T05:33:13Z" interval="6.4s" Mar 13 14:05:16 crc kubenswrapper[4907]: E0313 14:05:16.072793 4907 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:16Z is after 2026-02-23T05:33:13Z" node="crc" Mar 13 14:05:16 crc kubenswrapper[4907]: W0313 14:05:16.075506 4907 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:16Z is after 2026-02-23T05:33:13Z Mar 13 14:05:16 crc kubenswrapper[4907]: E0313 14:05:16.075575 4907 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:16Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 13 14:05:16 crc kubenswrapper[4907]: I0313 14:05:16.079317 4907 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Mar 13 14:05:16 crc kubenswrapper[4907]: I0313 14:05:16.079374 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Mar 13 14:05:16 crc kubenswrapper[4907]: E0313 14:05:16.081474 4907 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:16Z is after 2026-02-23T05:33:13Z" event="&Event{ObjectMeta:{crc.189c6ba3243feae5 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:01.709142757 +0000 UTC m=+0.608930476,LastTimestamp:2026-03-13 14:05:01.709142757 +0000 UTC m=+0.608930476,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:16 crc kubenswrapper[4907]: I0313 14:05:16.085921 4907 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Mar 13 14:05:16 crc kubenswrapper[4907]: I0313 14:05:16.085982 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Mar 13 14:05:16 crc kubenswrapper[4907]: I0313 14:05:16.716480 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:16Z is after 2026-02-23T05:33:13Z Mar 13 14:05:16 crc kubenswrapper[4907]: I0313 14:05:16.904581 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Mar 13 14:05:16 crc kubenswrapper[4907]: I0313 14:05:16.905484 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Mar 13 14:05:16 crc kubenswrapper[4907]: I0313 14:05:16.907864 4907 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="8cc2d976ae1bb0e2999831b35473eb89231620ed94a8938258a0fbbd5a909a33" exitCode=255 Mar 13 14:05:16 crc kubenswrapper[4907]: I0313 14:05:16.907941 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"8cc2d976ae1bb0e2999831b35473eb89231620ed94a8938258a0fbbd5a909a33"} Mar 13 14:05:16 crc kubenswrapper[4907]: I0313 14:05:16.908047 4907 scope.go:117] "RemoveContainer" containerID="fe6cd60e8033e3165c90914a90e482e2378e469fd98044ddc2488e020f75e14c" Mar 13 14:05:16 crc kubenswrapper[4907]: I0313 14:05:16.908156 4907 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 14:05:16 crc kubenswrapper[4907]: I0313 14:05:16.909351 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:05:16 crc kubenswrapper[4907]: I0313 14:05:16.909407 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:05:16 crc kubenswrapper[4907]: I0313 14:05:16.909428 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:05:16 crc kubenswrapper[4907]: I0313 14:05:16.918995 4907 scope.go:117] "RemoveContainer" containerID="8cc2d976ae1bb0e2999831b35473eb89231620ed94a8938258a0fbbd5a909a33" Mar 13 14:05:16 crc kubenswrapper[4907]: E0313 14:05:16.919706 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 13 14:05:17 crc kubenswrapper[4907]: I0313 14:05:17.717764 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:17Z is after 2026-02-23T05:33:13Z Mar 13 14:05:17 crc kubenswrapper[4907]: I0313 14:05:17.911489 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Mar 13 14:05:18 crc kubenswrapper[4907]: I0313 14:05:18.716252 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:18Z is after 2026-02-23T05:33:13Z Mar 13 14:05:19 crc kubenswrapper[4907]: W0313 14:05:19.571090 4907 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:19Z is after 2026-02-23T05:33:13Z Mar 13 14:05:19 crc kubenswrapper[4907]: E0313 14:05:19.571211 4907 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:19Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 13 14:05:19 crc kubenswrapper[4907]: I0313 14:05:19.623301 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 14:05:19 crc kubenswrapper[4907]: I0313 14:05:19.623510 4907 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 14:05:19 crc kubenswrapper[4907]: I0313 14:05:19.629160 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:05:19 crc kubenswrapper[4907]: I0313 14:05:19.629257 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:05:19 crc kubenswrapper[4907]: I0313 14:05:19.629285 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:05:19 crc kubenswrapper[4907]: I0313 14:05:19.630273 4907 scope.go:117] "RemoveContainer" containerID="8cc2d976ae1bb0e2999831b35473eb89231620ed94a8938258a0fbbd5a909a33" Mar 13 14:05:19 crc kubenswrapper[4907]: E0313 14:05:19.630654 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 13 14:05:19 crc kubenswrapper[4907]: I0313 14:05:19.630729 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 14:05:19 crc kubenswrapper[4907]: I0313 14:05:19.715482 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:19Z is after 2026-02-23T05:33:13Z Mar 13 14:05:19 crc kubenswrapper[4907]: I0313 14:05:19.918473 4907 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 14:05:19 crc kubenswrapper[4907]: I0313 14:05:19.919295 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:05:19 crc kubenswrapper[4907]: I0313 14:05:19.919334 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:05:19 crc kubenswrapper[4907]: I0313 14:05:19.919346 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:05:19 crc kubenswrapper[4907]: I0313 14:05:19.919921 4907 scope.go:117] "RemoveContainer" containerID="8cc2d976ae1bb0e2999831b35473eb89231620ed94a8938258a0fbbd5a909a33" Mar 13 14:05:19 crc kubenswrapper[4907]: E0313 14:05:19.920083 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 13 14:05:20 crc kubenswrapper[4907]: I0313 14:05:20.717247 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:20Z is after 2026-02-23T05:33:13Z Mar 13 14:05:20 crc kubenswrapper[4907]: W0313 14:05:20.974674 4907 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:20Z is after 2026-02-23T05:33:13Z Mar 13 14:05:20 crc kubenswrapper[4907]: E0313 14:05:20.974799 4907 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:20Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 13 14:05:21 crc kubenswrapper[4907]: I0313 14:05:21.501035 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Mar 13 14:05:21 crc kubenswrapper[4907]: I0313 14:05:21.501230 4907 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 14:05:21 crc kubenswrapper[4907]: I0313 14:05:21.502567 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:05:21 crc kubenswrapper[4907]: I0313 14:05:21.502637 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:05:21 crc kubenswrapper[4907]: I0313 14:05:21.502656 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:05:21 crc kubenswrapper[4907]: I0313 14:05:21.520139 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Mar 13 14:05:21 crc kubenswrapper[4907]: I0313 14:05:21.717311 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:21Z is after 2026-02-23T05:33:13Z Mar 13 14:05:21 crc kubenswrapper[4907]: E0313 14:05:21.871662 4907 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 13 14:05:21 crc kubenswrapper[4907]: I0313 14:05:21.924306 4907 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 14:05:21 crc kubenswrapper[4907]: I0313 14:05:21.925443 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:05:21 crc kubenswrapper[4907]: I0313 14:05:21.925507 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:05:21 crc kubenswrapper[4907]: I0313 14:05:21.925520 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:05:22 crc kubenswrapper[4907]: I0313 14:05:22.334152 4907 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 14:05:22 crc kubenswrapper[4907]: I0313 14:05:22.334240 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 14:05:22 crc kubenswrapper[4907]: I0313 14:05:22.473667 4907 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 14:05:22 crc kubenswrapper[4907]: E0313 14:05:22.474499 4907 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:22Z is after 2026-02-23T05:33:13Z" interval="7s" Mar 13 14:05:22 crc kubenswrapper[4907]: I0313 14:05:22.475084 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:05:22 crc kubenswrapper[4907]: I0313 14:05:22.475126 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:05:22 crc kubenswrapper[4907]: I0313 14:05:22.475140 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:05:22 crc kubenswrapper[4907]: I0313 14:05:22.475166 4907 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 13 14:05:22 crc kubenswrapper[4907]: E0313 14:05:22.478179 4907 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:22Z is after 2026-02-23T05:33:13Z" node="crc" Mar 13 14:05:22 crc kubenswrapper[4907]: I0313 14:05:22.714989 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:22Z is after 2026-02-23T05:33:13Z Mar 13 14:05:23 crc kubenswrapper[4907]: I0313 14:05:23.716990 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:23Z is after 2026-02-23T05:33:13Z Mar 13 14:05:24 crc kubenswrapper[4907]: I0313 14:05:24.718011 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:24Z is after 2026-02-23T05:33:13Z Mar 13 14:05:24 crc kubenswrapper[4907]: I0313 14:05:24.855113 4907 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Mar 13 14:05:24 crc kubenswrapper[4907]: E0313 14:05:24.861316 4907 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:24Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 13 14:05:25 crc kubenswrapper[4907]: I0313 14:05:25.164996 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 14:05:25 crc kubenswrapper[4907]: I0313 14:05:25.165315 4907 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 14:05:25 crc kubenswrapper[4907]: I0313 14:05:25.167190 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:05:25 crc kubenswrapper[4907]: I0313 14:05:25.167244 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:05:25 crc kubenswrapper[4907]: I0313 14:05:25.167258 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:05:25 crc kubenswrapper[4907]: I0313 14:05:25.167948 4907 scope.go:117] "RemoveContainer" containerID="8cc2d976ae1bb0e2999831b35473eb89231620ed94a8938258a0fbbd5a909a33" Mar 13 14:05:25 crc kubenswrapper[4907]: E0313 14:05:25.168158 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 13 14:05:25 crc kubenswrapper[4907]: I0313 14:05:25.717342 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:25Z is after 2026-02-23T05:33:13Z Mar 13 14:05:25 crc kubenswrapper[4907]: I0313 14:05:25.824482 4907 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 14:05:25 crc kubenswrapper[4907]: I0313 14:05:25.934521 4907 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 14:05:25 crc kubenswrapper[4907]: I0313 14:05:25.936036 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:05:25 crc kubenswrapper[4907]: I0313 14:05:25.936084 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:05:25 crc kubenswrapper[4907]: I0313 14:05:25.936096 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:05:25 crc kubenswrapper[4907]: I0313 14:05:25.936825 4907 scope.go:117] "RemoveContainer" containerID="8cc2d976ae1bb0e2999831b35473eb89231620ed94a8938258a0fbbd5a909a33" Mar 13 14:05:25 crc kubenswrapper[4907]: E0313 14:05:25.937052 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 13 14:05:26 crc kubenswrapper[4907]: W0313 14:05:26.083159 4907 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:26Z is after 2026-02-23T05:33:13Z Mar 13 14:05:26 crc kubenswrapper[4907]: E0313 14:05:26.083261 4907 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:26Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 13 14:05:26 crc kubenswrapper[4907]: E0313 14:05:26.085706 4907 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:26Z is after 2026-02-23T05:33:13Z" event="&Event{ObjectMeta:{crc.189c6ba3243feae5 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:01.709142757 +0000 UTC m=+0.608930476,LastTimestamp:2026-03-13 14:05:01.709142757 +0000 UTC m=+0.608930476,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:26 crc kubenswrapper[4907]: I0313 14:05:26.717760 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:26Z is after 2026-02-23T05:33:13Z Mar 13 14:05:26 crc kubenswrapper[4907]: W0313 14:05:26.775387 4907 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:26Z is after 2026-02-23T05:33:13Z Mar 13 14:05:26 crc kubenswrapper[4907]: E0313 14:05:26.775494 4907 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:26Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 13 14:05:27 crc kubenswrapper[4907]: W0313 14:05:27.179577 4907 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:27Z is after 2026-02-23T05:33:13Z Mar 13 14:05:27 crc kubenswrapper[4907]: E0313 14:05:27.179658 4907 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:27Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 13 14:05:27 crc kubenswrapper[4907]: I0313 14:05:27.718161 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:27Z is after 2026-02-23T05:33:13Z Mar 13 14:05:28 crc kubenswrapper[4907]: W0313 14:05:28.384764 4907 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:28Z is after 2026-02-23T05:33:13Z Mar 13 14:05:28 crc kubenswrapper[4907]: E0313 14:05:28.384931 4907 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:28Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 13 14:05:28 crc kubenswrapper[4907]: I0313 14:05:28.715385 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:28Z is after 2026-02-23T05:33:13Z Mar 13 14:05:29 crc kubenswrapper[4907]: I0313 14:05:29.479311 4907 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 14:05:29 crc kubenswrapper[4907]: E0313 14:05:29.480642 4907 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:29Z is after 2026-02-23T05:33:13Z" interval="7s" Mar 13 14:05:29 crc kubenswrapper[4907]: I0313 14:05:29.480875 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:05:29 crc kubenswrapper[4907]: I0313 14:05:29.480966 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:05:29 crc kubenswrapper[4907]: I0313 14:05:29.480984 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:05:29 crc kubenswrapper[4907]: I0313 14:05:29.481024 4907 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 13 14:05:29 crc kubenswrapper[4907]: E0313 14:05:29.484399 4907 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:29Z is after 2026-02-23T05:33:13Z" node="crc" Mar 13 14:05:29 crc kubenswrapper[4907]: I0313 14:05:29.717671 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:29Z is after 2026-02-23T05:33:13Z Mar 13 14:05:30 crc kubenswrapper[4907]: I0313 14:05:30.717229 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:30Z is after 2026-02-23T05:33:13Z Mar 13 14:05:31 crc kubenswrapper[4907]: I0313 14:05:31.717138 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:31Z is after 2026-02-23T05:33:13Z Mar 13 14:05:31 crc kubenswrapper[4907]: E0313 14:05:31.871807 4907 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 13 14:05:32 crc kubenswrapper[4907]: I0313 14:05:32.333731 4907 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 14:05:32 crc kubenswrapper[4907]: I0313 14:05:32.333821 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 14:05:32 crc kubenswrapper[4907]: I0313 14:05:32.333939 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 13 14:05:32 crc kubenswrapper[4907]: I0313 14:05:32.334170 4907 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 14:05:32 crc kubenswrapper[4907]: I0313 14:05:32.336181 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:05:32 crc kubenswrapper[4907]: I0313 14:05:32.336235 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:05:32 crc kubenswrapper[4907]: I0313 14:05:32.336264 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:05:32 crc kubenswrapper[4907]: I0313 14:05:32.336938 4907 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="cluster-policy-controller" containerStatusID={"Type":"cri-o","ID":"1f22becea3c68103ff0b2639c7ae8c051b76d19b407b0463645fc2e4a0682f23"} pod="openshift-kube-controller-manager/kube-controller-manager-crc" containerMessage="Container cluster-policy-controller failed startup probe, will be restarted" Mar 13 14:05:32 crc kubenswrapper[4907]: I0313 14:05:32.337184 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" containerID="cri-o://1f22becea3c68103ff0b2639c7ae8c051b76d19b407b0463645fc2e4a0682f23" gracePeriod=30 Mar 13 14:05:32 crc kubenswrapper[4907]: I0313 14:05:32.716756 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:32Z is after 2026-02-23T05:33:13Z Mar 13 14:05:32 crc kubenswrapper[4907]: I0313 14:05:32.955536 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/0.log" Mar 13 14:05:32 crc kubenswrapper[4907]: I0313 14:05:32.956372 4907 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="1f22becea3c68103ff0b2639c7ae8c051b76d19b407b0463645fc2e4a0682f23" exitCode=255 Mar 13 14:05:32 crc kubenswrapper[4907]: I0313 14:05:32.956407 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"1f22becea3c68103ff0b2639c7ae8c051b76d19b407b0463645fc2e4a0682f23"} Mar 13 14:05:32 crc kubenswrapper[4907]: I0313 14:05:32.956448 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"402c1ef89085d840edb4313577f5b2fb967072cb3db00c34c98572f00699529c"} Mar 13 14:05:32 crc kubenswrapper[4907]: I0313 14:05:32.956548 4907 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 14:05:32 crc kubenswrapper[4907]: I0313 14:05:32.957805 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:05:32 crc kubenswrapper[4907]: I0313 14:05:32.957830 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:05:32 crc kubenswrapper[4907]: I0313 14:05:32.957841 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:05:33 crc kubenswrapper[4907]: I0313 14:05:33.717062 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:33Z is after 2026-02-23T05:33:13Z Mar 13 14:05:34 crc kubenswrapper[4907]: I0313 14:05:34.715856 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:34Z is after 2026-02-23T05:33:13Z Mar 13 14:05:35 crc kubenswrapper[4907]: I0313 14:05:35.715491 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:35Z is after 2026-02-23T05:33:13Z Mar 13 14:05:36 crc kubenswrapper[4907]: E0313 14:05:36.089429 4907 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:36Z is after 2026-02-23T05:33:13Z" event="&Event{ObjectMeta:{crc.189c6ba3243feae5 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:01.709142757 +0000 UTC m=+0.608930476,LastTimestamp:2026-03-13 14:05:01.709142757 +0000 UTC m=+0.608930476,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:36 crc kubenswrapper[4907]: I0313 14:05:36.485456 4907 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 14:05:36 crc kubenswrapper[4907]: E0313 14:05:36.486653 4907 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:36Z is after 2026-02-23T05:33:13Z" interval="7s" Mar 13 14:05:36 crc kubenswrapper[4907]: I0313 14:05:36.487360 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:05:36 crc kubenswrapper[4907]: I0313 14:05:36.487613 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:05:36 crc kubenswrapper[4907]: I0313 14:05:36.487747 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:05:36 crc kubenswrapper[4907]: I0313 14:05:36.488055 4907 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 13 14:05:36 crc kubenswrapper[4907]: E0313 14:05:36.494148 4907 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:36Z is after 2026-02-23T05:33:13Z" node="crc" Mar 13 14:05:36 crc kubenswrapper[4907]: I0313 14:05:36.717377 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:36Z is after 2026-02-23T05:33:13Z Mar 13 14:05:37 crc kubenswrapper[4907]: I0313 14:05:37.715238 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:37Z is after 2026-02-23T05:33:13Z Mar 13 14:05:38 crc kubenswrapper[4907]: I0313 14:05:38.717197 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:38Z is after 2026-02-23T05:33:13Z Mar 13 14:05:39 crc kubenswrapper[4907]: I0313 14:05:39.333847 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 13 14:05:39 crc kubenswrapper[4907]: I0313 14:05:39.334113 4907 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 14:05:39 crc kubenswrapper[4907]: I0313 14:05:39.335763 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:05:39 crc kubenswrapper[4907]: I0313 14:05:39.335848 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:05:39 crc kubenswrapper[4907]: I0313 14:05:39.335869 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:05:39 crc kubenswrapper[4907]: I0313 14:05:39.715245 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:39Z is after 2026-02-23T05:33:13Z Mar 13 14:05:40 crc kubenswrapper[4907]: I0313 14:05:40.533492 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 13 14:05:40 crc kubenswrapper[4907]: I0313 14:05:40.533776 4907 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 14:05:40 crc kubenswrapper[4907]: I0313 14:05:40.535528 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:05:40 crc kubenswrapper[4907]: I0313 14:05:40.535599 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:05:40 crc kubenswrapper[4907]: I0313 14:05:40.535617 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:05:40 crc kubenswrapper[4907]: I0313 14:05:40.717608 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:40Z is after 2026-02-23T05:33:13Z Mar 13 14:05:40 crc kubenswrapper[4907]: I0313 14:05:40.782077 4907 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 14:05:40 crc kubenswrapper[4907]: I0313 14:05:40.784267 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:05:40 crc kubenswrapper[4907]: I0313 14:05:40.784364 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:05:40 crc kubenswrapper[4907]: I0313 14:05:40.784385 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:05:40 crc kubenswrapper[4907]: I0313 14:05:40.785293 4907 scope.go:117] "RemoveContainer" containerID="8cc2d976ae1bb0e2999831b35473eb89231620ed94a8938258a0fbbd5a909a33" Mar 13 14:05:41 crc kubenswrapper[4907]: I0313 14:05:41.275945 4907 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Mar 13 14:05:41 crc kubenswrapper[4907]: E0313 14:05:41.282523 4907 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:41Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 13 14:05:41 crc kubenswrapper[4907]: E0313 14:05:41.283954 4907 certificate_manager.go:440] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Reached backoff limit, still unable to rotate certs: timed out waiting for the condition" logger="UnhandledError" Mar 13 14:05:41 crc kubenswrapper[4907]: I0313 14:05:41.716545 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:41Z is after 2026-02-23T05:33:13Z Mar 13 14:05:41 crc kubenswrapper[4907]: E0313 14:05:41.871986 4907 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 13 14:05:41 crc kubenswrapper[4907]: I0313 14:05:41.985495 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Mar 13 14:05:41 crc kubenswrapper[4907]: I0313 14:05:41.986823 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Mar 13 14:05:41 crc kubenswrapper[4907]: I0313 14:05:41.988787 4907 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="da9239d5489f168f1e0695a6eb1d540c331de494e29ae89624cfbfed390520db" exitCode=255 Mar 13 14:05:41 crc kubenswrapper[4907]: I0313 14:05:41.988842 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"da9239d5489f168f1e0695a6eb1d540c331de494e29ae89624cfbfed390520db"} Mar 13 14:05:41 crc kubenswrapper[4907]: I0313 14:05:41.988917 4907 scope.go:117] "RemoveContainer" containerID="8cc2d976ae1bb0e2999831b35473eb89231620ed94a8938258a0fbbd5a909a33" Mar 13 14:05:41 crc kubenswrapper[4907]: I0313 14:05:41.989135 4907 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 14:05:41 crc kubenswrapper[4907]: I0313 14:05:41.990708 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:05:41 crc kubenswrapper[4907]: I0313 14:05:41.990749 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:05:41 crc kubenswrapper[4907]: I0313 14:05:41.990767 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:05:41 crc kubenswrapper[4907]: I0313 14:05:41.991558 4907 scope.go:117] "RemoveContainer" containerID="da9239d5489f168f1e0695a6eb1d540c331de494e29ae89624cfbfed390520db" Mar 13 14:05:41 crc kubenswrapper[4907]: E0313 14:05:41.991835 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 13 14:05:42 crc kubenswrapper[4907]: I0313 14:05:42.333941 4907 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 14:05:42 crc kubenswrapper[4907]: I0313 14:05:42.334051 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 14:05:42 crc kubenswrapper[4907]: I0313 14:05:42.719009 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:42Z is after 2026-02-23T05:33:13Z Mar 13 14:05:42 crc kubenswrapper[4907]: I0313 14:05:42.995198 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Mar 13 14:05:43 crc kubenswrapper[4907]: E0313 14:05:43.493261 4907 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:43Z is after 2026-02-23T05:33:13Z" interval="7s" Mar 13 14:05:43 crc kubenswrapper[4907]: I0313 14:05:43.494288 4907 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 14:05:43 crc kubenswrapper[4907]: I0313 14:05:43.495846 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:05:43 crc kubenswrapper[4907]: I0313 14:05:43.496123 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:05:43 crc kubenswrapper[4907]: I0313 14:05:43.496305 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:05:43 crc kubenswrapper[4907]: I0313 14:05:43.496481 4907 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 13 14:05:43 crc kubenswrapper[4907]: E0313 14:05:43.502468 4907 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:43Z is after 2026-02-23T05:33:13Z" node="crc" Mar 13 14:05:43 crc kubenswrapper[4907]: I0313 14:05:43.718588 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:43Z is after 2026-02-23T05:33:13Z Mar 13 14:05:44 crc kubenswrapper[4907]: I0313 14:05:44.717476 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:44Z is after 2026-02-23T05:33:13Z Mar 13 14:05:45 crc kubenswrapper[4907]: I0313 14:05:45.164807 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 14:05:45 crc kubenswrapper[4907]: I0313 14:05:45.165075 4907 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 14:05:45 crc kubenswrapper[4907]: I0313 14:05:45.166802 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:05:45 crc kubenswrapper[4907]: I0313 14:05:45.166851 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:05:45 crc kubenswrapper[4907]: I0313 14:05:45.166867 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:05:45 crc kubenswrapper[4907]: I0313 14:05:45.167798 4907 scope.go:117] "RemoveContainer" containerID="da9239d5489f168f1e0695a6eb1d540c331de494e29ae89624cfbfed390520db" Mar 13 14:05:45 crc kubenswrapper[4907]: E0313 14:05:45.168222 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 13 14:05:45 crc kubenswrapper[4907]: I0313 14:05:45.717731 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:45Z is after 2026-02-23T05:33:13Z Mar 13 14:05:45 crc kubenswrapper[4907]: W0313 14:05:45.741811 4907 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:45Z is after 2026-02-23T05:33:13Z Mar 13 14:05:45 crc kubenswrapper[4907]: E0313 14:05:45.742208 4907 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:45Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 13 14:05:45 crc kubenswrapper[4907]: I0313 14:05:45.824161 4907 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 14:05:46 crc kubenswrapper[4907]: I0313 14:05:46.009176 4907 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 14:05:46 crc kubenswrapper[4907]: I0313 14:05:46.010444 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:05:46 crc kubenswrapper[4907]: I0313 14:05:46.010510 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:05:46 crc kubenswrapper[4907]: I0313 14:05:46.010527 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:05:46 crc kubenswrapper[4907]: I0313 14:05:46.011328 4907 scope.go:117] "RemoveContainer" containerID="da9239d5489f168f1e0695a6eb1d540c331de494e29ae89624cfbfed390520db" Mar 13 14:05:46 crc kubenswrapper[4907]: E0313 14:05:46.011618 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 13 14:05:46 crc kubenswrapper[4907]: E0313 14:05:46.094122 4907 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:46Z is after 2026-02-23T05:33:13Z" event="&Event{ObjectMeta:{crc.189c6ba3243feae5 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:01.709142757 +0000 UTC m=+0.608930476,LastTimestamp:2026-03-13 14:05:01.709142757 +0000 UTC m=+0.608930476,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:46 crc kubenswrapper[4907]: W0313 14:05:46.637210 4907 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:46Z is after 2026-02-23T05:33:13Z Mar 13 14:05:46 crc kubenswrapper[4907]: E0313 14:05:46.637645 4907 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:46Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 13 14:05:46 crc kubenswrapper[4907]: I0313 14:05:46.717081 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:46Z is after 2026-02-23T05:33:13Z Mar 13 14:05:47 crc kubenswrapper[4907]: I0313 14:05:47.715099 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:47Z is after 2026-02-23T05:33:13Z Mar 13 14:05:47 crc kubenswrapper[4907]: W0313 14:05:47.755471 4907 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:47Z is after 2026-02-23T05:33:13Z Mar 13 14:05:47 crc kubenswrapper[4907]: E0313 14:05:47.755618 4907 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:47Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 13 14:05:48 crc kubenswrapper[4907]: I0313 14:05:48.715291 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:48Z is after 2026-02-23T05:33:13Z Mar 13 14:05:49 crc kubenswrapper[4907]: I0313 14:05:49.719397 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 14:05:50 crc kubenswrapper[4907]: E0313 14:05:50.501185 4907 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 13 14:05:50 crc kubenswrapper[4907]: I0313 14:05:50.503160 4907 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 14:05:50 crc kubenswrapper[4907]: I0313 14:05:50.505011 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:05:50 crc kubenswrapper[4907]: I0313 14:05:50.505057 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:05:50 crc kubenswrapper[4907]: I0313 14:05:50.505072 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:05:50 crc kubenswrapper[4907]: I0313 14:05:50.505101 4907 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 13 14:05:50 crc kubenswrapper[4907]: E0313 14:05:50.510249 4907 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 13 14:05:50 crc kubenswrapper[4907]: I0313 14:05:50.718379 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 14:05:50 crc kubenswrapper[4907]: W0313 14:05:50.762574 4907 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: services is forbidden: User "system:anonymous" cannot list resource "services" in API group "" at the cluster scope Mar 13 14:05:50 crc kubenswrapper[4907]: E0313 14:05:50.763752 4907 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: services is forbidden: User \"system:anonymous\" cannot list resource \"services\" in API group \"\" at the cluster scope" logger="UnhandledError" Mar 13 14:05:51 crc kubenswrapper[4907]: I0313 14:05:51.718749 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 14:05:51 crc kubenswrapper[4907]: E0313 14:05:51.872157 4907 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 13 14:05:52 crc kubenswrapper[4907]: I0313 14:05:52.333931 4907 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 14:05:52 crc kubenswrapper[4907]: I0313 14:05:52.334044 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 14:05:52 crc kubenswrapper[4907]: I0313 14:05:52.717612 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 14:05:53 crc kubenswrapper[4907]: I0313 14:05:53.720533 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 14:05:54 crc kubenswrapper[4907]: I0313 14:05:54.718708 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 14:05:55 crc kubenswrapper[4907]: I0313 14:05:55.718515 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 14:05:55 crc kubenswrapper[4907]: I0313 14:05:55.790582 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 13 14:05:55 crc kubenswrapper[4907]: I0313 14:05:55.790772 4907 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 14:05:55 crc kubenswrapper[4907]: I0313 14:05:55.792312 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:05:55 crc kubenswrapper[4907]: I0313 14:05:55.792364 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:05:55 crc kubenswrapper[4907]: I0313 14:05:55.792381 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.099110 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c6ba3243feae5 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:01.709142757 +0000 UTC m=+0.608930476,LastTimestamp:2026-03-13 14:05:01.709142757 +0000 UTC m=+0.608930476,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.106400 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c6ba3286e93ab default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:01.779309483 +0000 UTC m=+0.679097192,LastTimestamp:2026-03-13 14:05:01.779309483 +0000 UTC m=+0.679097192,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.111643 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c6ba3286f4326 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:01.779354406 +0000 UTC m=+0.679142115,LastTimestamp:2026-03-13 14:05:01.779354406 +0000 UTC m=+0.679142115,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.117962 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c6ba3286f9543 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:01.779375427 +0000 UTC m=+0.679163136,LastTimestamp:2026-03-13 14:05:01.779375427 +0000 UTC m=+0.679163136,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.123371 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c6ba32d9aa280 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeAllocatableEnforced,Message:Updated Node Allocatable limit across pods,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:01.866082944 +0000 UTC m=+0.765870643,LastTimestamp:2026-03-13 14:05:01.866082944 +0000 UTC m=+0.765870643,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.128972 4907 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c6ba3286e93ab\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c6ba3286e93ab default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:01.779309483 +0000 UTC m=+0.679097192,LastTimestamp:2026-03-13 14:05:01.884013463 +0000 UTC m=+0.783801162,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.136161 4907 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c6ba3286f4326\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c6ba3286f4326 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:01.779354406 +0000 UTC m=+0.679142115,LastTimestamp:2026-03-13 14:05:01.884035894 +0000 UTC m=+0.783823593,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.142209 4907 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c6ba3286f9543\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c6ba3286f9543 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:01.779375427 +0000 UTC m=+0.679163136,LastTimestamp:2026-03-13 14:05:01.884047345 +0000 UTC m=+0.783835044,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.147541 4907 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c6ba3286e93ab\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c6ba3286e93ab default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:01.779309483 +0000 UTC m=+0.679097192,LastTimestamp:2026-03-13 14:05:01.885015553 +0000 UTC m=+0.784803242,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.152604 4907 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c6ba3286f4326\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c6ba3286f4326 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:01.779354406 +0000 UTC m=+0.679142115,LastTimestamp:2026-03-13 14:05:01.885029624 +0000 UTC m=+0.784817313,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.158359 4907 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c6ba3286f9543\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c6ba3286f9543 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:01.779375427 +0000 UTC m=+0.679163136,LastTimestamp:2026-03-13 14:05:01.885040344 +0000 UTC m=+0.784828033,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.164743 4907 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c6ba3286e93ab\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c6ba3286e93ab default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:01.779309483 +0000 UTC m=+0.679097192,LastTimestamp:2026-03-13 14:05:01.886466179 +0000 UTC m=+0.786253878,Count:4,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.170773 4907 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c6ba3286f4326\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c6ba3286f4326 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:01.779354406 +0000 UTC m=+0.679142115,LastTimestamp:2026-03-13 14:05:01.88648071 +0000 UTC m=+0.786268409,Count:4,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.176120 4907 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c6ba3286f9543\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c6ba3286f9543 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:01.779375427 +0000 UTC m=+0.679163136,LastTimestamp:2026-03-13 14:05:01.886494331 +0000 UTC m=+0.786282040,Count:4,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.181539 4907 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c6ba3286e93ab\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c6ba3286e93ab default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:01.779309483 +0000 UTC m=+0.679097192,LastTimestamp:2026-03-13 14:05:01.886610818 +0000 UTC m=+0.786398507,Count:5,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.188043 4907 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c6ba3286f4326\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c6ba3286f4326 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:01.779354406 +0000 UTC m=+0.679142115,LastTimestamp:2026-03-13 14:05:01.886631789 +0000 UTC m=+0.786419478,Count:5,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.193670 4907 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c6ba3286f9543\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c6ba3286f9543 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:01.779375427 +0000 UTC m=+0.679163136,LastTimestamp:2026-03-13 14:05:01.88664583 +0000 UTC m=+0.786433519,Count:5,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.198791 4907 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c6ba3286e93ab\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c6ba3286e93ab default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:01.779309483 +0000 UTC m=+0.679097192,LastTimestamp:2026-03-13 14:05:01.887948038 +0000 UTC m=+0.787735727,Count:6,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.205215 4907 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c6ba3286e93ab\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c6ba3286e93ab default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:01.779309483 +0000 UTC m=+0.679097192,LastTimestamp:2026-03-13 14:05:01.887966539 +0000 UTC m=+0.787754248,Count:7,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.208592 4907 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c6ba3286f4326\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c6ba3286f4326 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:01.779354406 +0000 UTC m=+0.679142115,LastTimestamp:2026-03-13 14:05:01.88798662 +0000 UTC m=+0.787774319,Count:6,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.212601 4907 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c6ba3286f9543\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c6ba3286f9543 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:01.779375427 +0000 UTC m=+0.679163136,LastTimestamp:2026-03-13 14:05:01.888002191 +0000 UTC m=+0.787789900,Count:6,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.219290 4907 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c6ba3286f4326\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c6ba3286f4326 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:01.779354406 +0000 UTC m=+0.679142115,LastTimestamp:2026-03-13 14:05:01.888032943 +0000 UTC m=+0.787820642,Count:7,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.225559 4907 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c6ba3286f9543\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c6ba3286f9543 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:01.779375427 +0000 UTC m=+0.679163136,LastTimestamp:2026-03-13 14:05:01.888069705 +0000 UTC m=+0.787857404,Count:7,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.231340 4907 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c6ba3286e93ab\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c6ba3286e93ab default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:01.779309483 +0000 UTC m=+0.679097192,LastTimestamp:2026-03-13 14:05:01.888267746 +0000 UTC m=+0.788055435,Count:8,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.236067 4907 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c6ba3286f4326\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c6ba3286f4326 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:01.779354406 +0000 UTC m=+0.679142115,LastTimestamp:2026-03-13 14:05:01.888279387 +0000 UTC m=+0.788067076,Count:8,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.243524 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c6ba347c65aa6 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:02.30515575 +0000 UTC m=+1.204943489,LastTimestamp:2026-03-13 14:05:02.30515575 +0000 UTC m=+1.204943489,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.249702 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c6ba347c69067 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:02.305169511 +0000 UTC m=+1.204957240,LastTimestamp:2026-03-13 14:05:02.305169511 +0000 UTC m=+1.204957240,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.256445 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189c6ba34828b7f5 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:02.311602165 +0000 UTC m=+1.211389844,LastTimestamp:2026-03-13 14:05:02.311602165 +0000 UTC m=+1.211389844,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.263537 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c6ba348d5c81f openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:02.322944031 +0000 UTC m=+1.222731730,LastTimestamp:2026-03-13 14:05:02.322944031 +0000 UTC m=+1.222731730,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.270918 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189c6ba348f41381 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{wait-for-host-port},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:02.324929409 +0000 UTC m=+1.224717118,LastTimestamp:2026-03-13 14:05:02.324929409 +0000 UTC m=+1.224717118,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.278679 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c6ba36f0ea4fc openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Created,Message:Created container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:02.964204796 +0000 UTC m=+1.863992485,LastTimestamp:2026-03-13 14:05:02.964204796 +0000 UTC m=+1.863992485,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.285517 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189c6ba36f0f21ad openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Created,Message:Created container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:02.964236717 +0000 UTC m=+1.864024406,LastTimestamp:2026-03-13 14:05:02.964236717 +0000 UTC m=+1.864024406,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.292993 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c6ba36f102c73 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Created,Message:Created container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:02.964305011 +0000 UTC m=+1.864092700,LastTimestamp:2026-03-13 14:05:02.964305011 +0000 UTC m=+1.864092700,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.300031 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189c6ba36f1bf973 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{wait-for-host-port},},Reason:Created,Message:Created container wait-for-host-port,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:02.965078387 +0000 UTC m=+1.864866076,LastTimestamp:2026-03-13 14:05:02.965078387 +0000 UTC m=+1.864866076,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.307506 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c6ba36f264b72 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager},},Reason:Created,Message:Created container kube-controller-manager,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:02.965754738 +0000 UTC m=+1.865542417,LastTimestamp:2026-03-13 14:05:02.965754738 +0000 UTC m=+1.865542417,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.314417 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189c6ba36fca6fa4 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{wait-for-host-port},},Reason:Started,Message:Started container wait-for-host-port,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:02.976511908 +0000 UTC m=+1.876299597,LastTimestamp:2026-03-13 14:05:02.976511908 +0000 UTC m=+1.876299597,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.321323 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c6ba36fe206de openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Started,Message:Started container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:02.97805795 +0000 UTC m=+1.877845639,LastTimestamp:2026-03-13 14:05:02.97805795 +0000 UTC m=+1.877845639,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.328607 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c6ba36fe78e27 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Started,Message:Started container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:02.978420263 +0000 UTC m=+1.878207952,LastTimestamp:2026-03-13 14:05:02.978420263 +0000 UTC m=+1.878207952,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.335148 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c6ba3700096a8 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager},},Reason:Started,Message:Started container kube-controller-manager,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:02.98006084 +0000 UTC m=+1.879848539,LastTimestamp:2026-03-13 14:05:02.98006084 +0000 UTC m=+1.879848539,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.341598 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189c6ba3700e41c6 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Started,Message:Started container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:02.980956614 +0000 UTC m=+1.880744303,LastTimestamp:2026-03-13 14:05:02.980956614 +0000 UTC m=+1.880744303,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.348281 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c6ba370135a3a openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:02.981290554 +0000 UTC m=+1.881078243,LastTimestamp:2026-03-13 14:05:02.981290554 +0000 UTC m=+1.881078243,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.354637 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c6ba38202bf57 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Created,Message:Created container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:03.282192215 +0000 UTC m=+2.181979944,LastTimestamp:2026-03-13 14:05:03.282192215 +0000 UTC m=+2.181979944,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.360943 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c6ba382c72e2a openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Started,Message:Started container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:03.295065642 +0000 UTC m=+2.194853371,LastTimestamp:2026-03-13 14:05:03.295065642 +0000 UTC m=+2.194853371,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.367536 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c6ba382e0cc04 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-cert-syncer},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:03.296744452 +0000 UTC m=+2.196532171,LastTimestamp:2026-03-13 14:05:03.296744452 +0000 UTC m=+2.196532171,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.374604 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c6ba38f25706a openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-cert-syncer},},Reason:Created,Message:Created container kube-controller-manager-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:03.502569578 +0000 UTC m=+2.402357287,LastTimestamp:2026-03-13 14:05:03.502569578 +0000 UTC m=+2.402357287,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.380797 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c6ba38fe96641 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-cert-syncer},},Reason:Started,Message:Started container kube-controller-manager-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:03.515412033 +0000 UTC m=+2.415199722,LastTimestamp:2026-03-13 14:05:03.515412033 +0000 UTC m=+2.415199722,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.387456 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c6ba38ff9afeb openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-recovery-controller},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:03.516479467 +0000 UTC m=+2.416267156,LastTimestamp:2026-03-13 14:05:03.516479467 +0000 UTC m=+2.416267156,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.394107 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c6ba39d27afab openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-recovery-controller},},Reason:Created,Message:Created container kube-controller-manager-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:03.737597867 +0000 UTC m=+2.637385566,LastTimestamp:2026-03-13 14:05:03.737597867 +0000 UTC m=+2.637385566,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.400540 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c6ba39dc723ec openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-recovery-controller},},Reason:Started,Message:Started container kube-controller-manager-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:03.748047852 +0000 UTC m=+2.647835551,LastTimestamp:2026-03-13 14:05:03.748047852 +0000 UTC m=+2.647835551,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.407073 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c6ba3a1379f21 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:03.805751073 +0000 UTC m=+2.705538772,LastTimestamp:2026-03-13 14:05:03.805751073 +0000 UTC m=+2.705538772,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.412730 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c6ba3a14c8c73 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-ensure-env-vars},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:03.807122547 +0000 UTC m=+2.706910266,LastTimestamp:2026-03-13 14:05:03.807122547 +0000 UTC m=+2.706910266,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.417563 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189c6ba3a192ddc2 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:03.811730882 +0000 UTC m=+2.711518601,LastTimestamp:2026-03-13 14:05:03.811730882 +0000 UTC m=+2.711518601,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.419792 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189c6ba3a229341d openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-rbac-proxy-crio},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:03.821583389 +0000 UTC m=+2.721371078,LastTimestamp:2026-03-13 14:05:03.821583389 +0000 UTC m=+2.721371078,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.424759 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189c6ba3afe0c887 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-rbac-proxy-crio},},Reason:Created,Message:Created container kube-rbac-proxy-crio,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:04.051718279 +0000 UTC m=+2.951505968,LastTimestamp:2026-03-13 14:05:04.051718279 +0000 UTC m=+2.951505968,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.426152 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c6ba3aff60fc7 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Created,Message:Created container kube-apiserver,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:04.053112775 +0000 UTC m=+2.952900464,LastTimestamp:2026-03-13 14:05:04.053112775 +0000 UTC m=+2.952900464,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.432105 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189c6ba3b003b64a openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler},},Reason:Created,Message:Created container kube-scheduler,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:04.05400737 +0000 UTC m=+2.953795059,LastTimestamp:2026-03-13 14:05:04.05400737 +0000 UTC m=+2.953795059,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.437527 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c6ba3b006033b openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-ensure-env-vars},},Reason:Created,Message:Created container etcd-ensure-env-vars,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:04.054158139 +0000 UTC m=+2.953945828,LastTimestamp:2026-03-13 14:05:04.054158139 +0000 UTC m=+2.953945828,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.443377 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189c6ba3b0b1c9dc openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-rbac-proxy-crio},},Reason:Started,Message:Started container kube-rbac-proxy-crio,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:04.065415644 +0000 UTC m=+2.965203333,LastTimestamp:2026-03-13 14:05:04.065415644 +0000 UTC m=+2.965203333,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.449748 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c6ba3b0feca2c openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Started,Message:Started container kube-apiserver,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:04.070461996 +0000 UTC m=+2.970249685,LastTimestamp:2026-03-13 14:05:04.070461996 +0000 UTC m=+2.970249685,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.455787 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c6ba3b11202b6 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-syncer},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:04.071721654 +0000 UTC m=+2.971509343,LastTimestamp:2026-03-13 14:05:04.071721654 +0000 UTC m=+2.971509343,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.461537 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189c6ba3b12f7f7e openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler},},Reason:Started,Message:Started container kube-scheduler,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:04.073654142 +0000 UTC m=+2.973441831,LastTimestamp:2026-03-13 14:05:04.073654142 +0000 UTC m=+2.973441831,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.467936 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189c6ba3b140496a openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-cert-syncer},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:04.07475441 +0000 UTC m=+2.974542099,LastTimestamp:2026-03-13 14:05:04.07475441 +0000 UTC m=+2.974542099,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.472141 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c6ba3b1a728cb openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-ensure-env-vars},},Reason:Started,Message:Started container etcd-ensure-env-vars,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:04.081496267 +0000 UTC m=+2.981283946,LastTimestamp:2026-03-13 14:05:04.081496267 +0000 UTC m=+2.981283946,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.477943 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c6ba3bc954357 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-syncer},},Reason:Created,Message:Created container kube-apiserver-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:04.264872791 +0000 UTC m=+3.164660480,LastTimestamp:2026-03-13 14:05:04.264872791 +0000 UTC m=+3.164660480,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.484335 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189c6ba3bd1fb389 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-cert-syncer},},Reason:Created,Message:Created container kube-scheduler-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:04.273945481 +0000 UTC m=+3.173733170,LastTimestamp:2026-03-13 14:05:04.273945481 +0000 UTC m=+3.173733170,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.490272 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c6ba3bd410ea8 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-syncer},},Reason:Started,Message:Started container kube-apiserver-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:04.276131496 +0000 UTC m=+3.175919185,LastTimestamp:2026-03-13 14:05:04.276131496 +0000 UTC m=+3.175919185,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.496431 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c6ba3bd57864d openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-regeneration-controller},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:04.277603917 +0000 UTC m=+3.177391606,LastTimestamp:2026-03-13 14:05:04.277603917 +0000 UTC m=+3.177391606,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.502928 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189c6ba3bdfba96a openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-cert-syncer},},Reason:Started,Message:Started container kube-scheduler-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:04.28836081 +0000 UTC m=+3.188148489,LastTimestamp:2026-03-13 14:05:04.28836081 +0000 UTC m=+3.188148489,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.509413 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189c6ba3be0c23cd openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-recovery-controller},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:04.289440717 +0000 UTC m=+3.189228426,LastTimestamp:2026-03-13 14:05:04.289440717 +0000 UTC m=+3.189228426,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.516638 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c6ba3caba6eeb openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-regeneration-controller},},Reason:Created,Message:Created container kube-apiserver-cert-regeneration-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:04.502189803 +0000 UTC m=+3.401977492,LastTimestamp:2026-03-13 14:05:04.502189803 +0000 UTC m=+3.401977492,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.522598 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189c6ba3cad24630 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-recovery-controller},},Reason:Created,Message:Created container kube-scheduler-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:04.50375224 +0000 UTC m=+3.403539929,LastTimestamp:2026-03-13 14:05:04.50375224 +0000 UTC m=+3.403539929,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.528784 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c6ba3cc433def openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-regeneration-controller},},Reason:Started,Message:Started container kube-apiserver-cert-regeneration-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:04.527932911 +0000 UTC m=+3.427720600,LastTimestamp:2026-03-13 14:05:04.527932911 +0000 UTC m=+3.427720600,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.534820 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c6ba3cc5aaac0 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-insecure-readyz},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:04.529468096 +0000 UTC m=+3.429255785,LastTimestamp:2026-03-13 14:05:04.529468096 +0000 UTC m=+3.429255785,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.539017 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189c6ba3cc780e43 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-recovery-controller},},Reason:Started,Message:Started container kube-scheduler-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:04.531394115 +0000 UTC m=+3.431181804,LastTimestamp:2026-03-13 14:05:04.531394115 +0000 UTC m=+3.431181804,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.545413 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c6ba3d86f79fe openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-insecure-readyz},},Reason:Created,Message:Created container kube-apiserver-insecure-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:04.732158462 +0000 UTC m=+3.631946151,LastTimestamp:2026-03-13 14:05:04.732158462 +0000 UTC m=+3.631946151,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.551143 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c6ba3d972c8eb openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-insecure-readyz},},Reason:Started,Message:Started container kube-apiserver-insecure-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:04.749152491 +0000 UTC m=+3.648940210,LastTimestamp:2026-03-13 14:05:04.749152491 +0000 UTC m=+3.648940210,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.556947 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c6ba3d995e9f9 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:04.751454713 +0000 UTC m=+3.651242422,LastTimestamp:2026-03-13 14:05:04.751454713 +0000 UTC m=+3.651242422,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.563870 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c6ba3dfb3175c openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-resources-copy},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:04.854030172 +0000 UTC m=+3.753817871,LastTimestamp:2026-03-13 14:05:04.854030172 +0000 UTC m=+3.753817871,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.570000 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c6ba3e890b760 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Created,Message:Created container kube-apiserver-check-endpoints,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:05.00277232 +0000 UTC m=+3.902560009,LastTimestamp:2026-03-13 14:05:05.00277232 +0000 UTC m=+3.902560009,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.575512 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c6ba3e95ca04a openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Started,Message:Started container kube-apiserver-check-endpoints,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:05.016135754 +0000 UTC m=+3.915923443,LastTimestamp:2026-03-13 14:05:05.016135754 +0000 UTC m=+3.915923443,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.581010 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c6ba3ece1107b openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-resources-copy},},Reason:Created,Message:Created container etcd-resources-copy,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:05.075146875 +0000 UTC m=+3.974934574,LastTimestamp:2026-03-13 14:05:05.075146875 +0000 UTC m=+3.974934574,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.587164 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c6ba3ed9af87b openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-resources-copy},},Reason:Started,Message:Started container etcd-resources-copy,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:05.087330427 +0000 UTC m=+3.987118116,LastTimestamp:2026-03-13 14:05:05.087330427 +0000 UTC m=+3.987118116,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.591622 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c6ba41b616858 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcdctl},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:05.855309912 +0000 UTC m=+4.755097601,LastTimestamp:2026-03-13 14:05:05.855309912 +0000 UTC m=+4.755097601,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.596831 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c6ba4263cdc2f openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcdctl},},Reason:Created,Message:Created container etcdctl,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:06.037464111 +0000 UTC m=+4.937251800,LastTimestamp:2026-03-13 14:05:06.037464111 +0000 UTC m=+4.937251800,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.601685 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c6ba426d74abc openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcdctl},},Reason:Started,Message:Started container etcdctl,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:06.047584956 +0000 UTC m=+4.947372645,LastTimestamp:2026-03-13 14:05:06.047584956 +0000 UTC m=+4.947372645,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.608160 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c6ba426e6b436 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:06.048594998 +0000 UTC m=+4.948382687,LastTimestamp:2026-03-13 14:05:06.048594998 +0000 UTC m=+4.948382687,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.612647 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c6ba43179cc54 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd},},Reason:Created,Message:Created container etcd,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:06.226007124 +0000 UTC m=+5.125794813,LastTimestamp:2026-03-13 14:05:06.226007124 +0000 UTC m=+5.125794813,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.619250 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c6ba4324778f4 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd},},Reason:Started,Message:Started container etcd,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:06.239486196 +0000 UTC m=+5.139273885,LastTimestamp:2026-03-13 14:05:06.239486196 +0000 UTC m=+5.139273885,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.625955 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c6ba43254cda0 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-metrics},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:06.24035984 +0000 UTC m=+5.140147539,LastTimestamp:2026-03-13 14:05:06.24035984 +0000 UTC m=+5.140147539,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.637690 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c6ba43e729543 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-metrics},},Reason:Created,Message:Created container etcd-metrics,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:06.443638083 +0000 UTC m=+5.343425772,LastTimestamp:2026-03-13 14:05:06.443638083 +0000 UTC m=+5.343425772,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.643824 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c6ba43f6c95ed openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-metrics},},Reason:Started,Message:Started container etcd-metrics,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:06.460022253 +0000 UTC m=+5.359809952,LastTimestamp:2026-03-13 14:05:06.460022253 +0000 UTC m=+5.359809952,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.649958 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c6ba43f811e8a openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-readyz},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:06.461367946 +0000 UTC m=+5.361155635,LastTimestamp:2026-03-13 14:05:06.461367946 +0000 UTC m=+5.361155635,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.656330 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c6ba44f92e19b openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-readyz},},Reason:Created,Message:Created container etcd-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:06.730967451 +0000 UTC m=+5.630755180,LastTimestamp:2026-03-13 14:05:06.730967451 +0000 UTC m=+5.630755180,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.661330 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c6ba45065b402 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-readyz},},Reason:Started,Message:Started container etcd-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:06.744783874 +0000 UTC m=+5.644571573,LastTimestamp:2026-03-13 14:05:06.744783874 +0000 UTC m=+5.644571573,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.668310 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c6ba45074c74e openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-rev},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:06.745771854 +0000 UTC m=+5.645559583,LastTimestamp:2026-03-13 14:05:06.745771854 +0000 UTC m=+5.645559583,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.675620 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c6ba45e840115 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-rev},},Reason:Created,Message:Created container etcd-rev,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:06.981650709 +0000 UTC m=+5.881438398,LastTimestamp:2026-03-13 14:05:06.981650709 +0000 UTC m=+5.881438398,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.680022 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c6ba45f41ed95 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-rev},},Reason:Started,Message:Started container etcd-rev,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:06.994097557 +0000 UTC m=+5.893885286,LastTimestamp:2026-03-13 14:05:06.994097557 +0000 UTC m=+5.893885286,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.688323 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Mar 13 14:05:56 crc kubenswrapper[4907]: &Event{ObjectMeta:{kube-controller-manager-crc.189c6ba59d8c577e openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": context deadline exceeded (Client.Timeout exceeded while awaiting headers) Mar 13 14:05:56 crc kubenswrapper[4907]: body: Mar 13 14:05:56 crc kubenswrapper[4907]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:12.334129022 +0000 UTC m=+11.233916791,LastTimestamp:2026-03-13 14:05:12.334129022 +0000 UTC m=+11.233916791,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 13 14:05:56 crc kubenswrapper[4907]: > Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.694803 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c6ba59d903f8e openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:12.334385038 +0000 UTC m=+11.234172767,LastTimestamp:2026-03-13 14:05:12.334385038 +0000 UTC m=+11.234172767,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.702965 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event=< Mar 13 14:05:56 crc kubenswrapper[4907]: &Event{ObjectMeta:{kube-apiserver-crc.189c6ba65ea21eea openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:ProbeError,Message:Readiness probe error: Get "https://192.168.126.11:17697/healthz": read tcp 192.168.126.11:46124->192.168.126.11:17697: read: connection reset by peer Mar 13 14:05:56 crc kubenswrapper[4907]: body: Mar 13 14:05:56 crc kubenswrapper[4907]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:15.573559018 +0000 UTC m=+14.473346717,LastTimestamp:2026-03-13 14:05:15.573559018 +0000 UTC m=+14.473346717,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 13 14:05:56 crc kubenswrapper[4907]: > Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.709571 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c6ba65ea3190c openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Unhealthy,Message:Readiness probe failed: Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:46124->192.168.126.11:17697: read: connection reset by peer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:15.573623052 +0000 UTC m=+14.473410741,LastTimestamp:2026-03-13 14:05:15.573623052 +0000 UTC m=+14.473410741,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: I0313 14:05:56.715823 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.716259 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event=< Mar 13 14:05:56 crc kubenswrapper[4907]: &Event{ObjectMeta:{kube-apiserver-crc.189c6ba66d978dd1 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:ProbeError,Message:Liveness probe error: Get "https://192.168.126.11:17697/healthz": dial tcp 192.168.126.11:17697: connect: connection refused Mar 13 14:05:56 crc kubenswrapper[4907]: body: Mar 13 14:05:56 crc kubenswrapper[4907]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:15.824524753 +0000 UTC m=+14.724312442,LastTimestamp:2026-03-13 14:05:15.824524753 +0000 UTC m=+14.724312442,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 13 14:05:56 crc kubenswrapper[4907]: > Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.720327 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c6ba66d983a0e openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Unhealthy,Message:Liveness probe failed: Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:15.824568846 +0000 UTC m=+14.724356535,LastTimestamp:2026-03-13 14:05:15.824568846 +0000 UTC m=+14.724356535,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.726532 4907 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189c6ba3d995e9f9\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c6ba3d995e9f9 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:04.751454713 +0000 UTC m=+3.651242422,LastTimestamp:2026-03-13 14:05:15.902518846 +0000 UTC m=+14.802306525,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.730245 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event=< Mar 13 14:05:56 crc kubenswrapper[4907]: &Event{ObjectMeta:{kube-apiserver-crc.189c6ba67cc80036 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:ProbeError,Message:Startup probe error: HTTP probe failed with statuscode: 403 Mar 13 14:05:56 crc kubenswrapper[4907]: body: {"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Mar 13 14:05:56 crc kubenswrapper[4907]: Mar 13 14:05:56 crc kubenswrapper[4907]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:16.079358006 +0000 UTC m=+14.979145695,LastTimestamp:2026-03-13 14:05:16.079358006 +0000 UTC m=+14.979145695,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 13 14:05:56 crc kubenswrapper[4907]: > Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.736474 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c6ba67cc89e45 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Unhealthy,Message:Startup probe failed: HTTP probe failed with statuscode: 403,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:16.079398469 +0000 UTC m=+14.979186158,LastTimestamp:2026-03-13 14:05:16.079398469 +0000 UTC m=+14.979186158,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.743754 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Mar 13 14:05:56 crc kubenswrapper[4907]: &Event{ObjectMeta:{kube-controller-manager-crc.189c6ba7f1999a2d openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) Mar 13 14:05:56 crc kubenswrapper[4907]: body: Mar 13 14:05:56 crc kubenswrapper[4907]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:22.334218797 +0000 UTC m=+21.234006526,LastTimestamp:2026-03-13 14:05:22.334218797 +0000 UTC m=+21.234006526,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 13 14:05:56 crc kubenswrapper[4907]: > Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.749990 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c6ba7f19a748c openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:22.3342747 +0000 UTC m=+21.234062429,LastTimestamp:2026-03-13 14:05:22.3342747 +0000 UTC m=+21.234062429,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.756182 4907 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189c6ba59d8c577e\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Mar 13 14:05:56 crc kubenswrapper[4907]: &Event{ObjectMeta:{kube-controller-manager-crc.189c6ba59d8c577e openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": context deadline exceeded (Client.Timeout exceeded while awaiting headers) Mar 13 14:05:56 crc kubenswrapper[4907]: body: Mar 13 14:05:56 crc kubenswrapper[4907]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:12.334129022 +0000 UTC m=+11.233916791,LastTimestamp:2026-03-13 14:05:32.333799643 +0000 UTC m=+31.233587372,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 13 14:05:56 crc kubenswrapper[4907]: > Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.763136 4907 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189c6ba59d903f8e\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c6ba59d903f8e openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:12.334385038 +0000 UTC m=+11.234172767,LastTimestamp:2026-03-13 14:05:32.333862304 +0000 UTC m=+31.233650023,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.767665 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c6baa45d248e7 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Killing,Message:Container cluster-policy-controller failed startup probe, will be restarted,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:32.337154279 +0000 UTC m=+31.236941998,LastTimestamp:2026-03-13 14:05:32.337154279 +0000 UTC m=+31.236941998,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.773527 4907 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189c6ba370135a3a\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c6ba370135a3a openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:02.981290554 +0000 UTC m=+1.881078243,LastTimestamp:2026-03-13 14:05:32.461056377 +0000 UTC m=+31.360844106,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.777761 4907 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189c6ba38202bf57\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c6ba38202bf57 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Created,Message:Created container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:03.282192215 +0000 UTC m=+2.181979944,LastTimestamp:2026-03-13 14:05:32.706689196 +0000 UTC m=+31.606476925,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.792702 4907 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189c6ba382c72e2a\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c6ba382c72e2a openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Started,Message:Started container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:03.295065642 +0000 UTC m=+2.194853371,LastTimestamp:2026-03-13 14:05:32.720223796 +0000 UTC m=+31.620011525,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.800635 4907 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189c6ba59d8c577e\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Mar 13 14:05:56 crc kubenswrapper[4907]: &Event{ObjectMeta:{kube-controller-manager-crc.189c6ba59d8c577e openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": context deadline exceeded (Client.Timeout exceeded while awaiting headers) Mar 13 14:05:56 crc kubenswrapper[4907]: body: Mar 13 14:05:56 crc kubenswrapper[4907]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:12.334129022 +0000 UTC m=+11.233916791,LastTimestamp:2026-03-13 14:05:42.33401524 +0000 UTC m=+41.233802959,Count:3,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 13 14:05:56 crc kubenswrapper[4907]: > Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.805714 4907 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189c6ba59d903f8e\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c6ba59d903f8e openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:12.334385038 +0000 UTC m=+11.234172767,LastTimestamp:2026-03-13 14:05:42.334087382 +0000 UTC m=+41.233875101,Count:3,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.810407 4907 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189c6ba7f1999a2d\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Mar 13 14:05:56 crc kubenswrapper[4907]: &Event{ObjectMeta:{kube-controller-manager-crc.189c6ba7f1999a2d openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) Mar 13 14:05:56 crc kubenswrapper[4907]: body: Mar 13 14:05:56 crc kubenswrapper[4907]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:22.334218797 +0000 UTC m=+21.234006526,LastTimestamp:2026-03-13 14:05:52.334014601 +0000 UTC m=+51.233802320,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 13 14:05:56 crc kubenswrapper[4907]: > Mar 13 14:05:57 crc kubenswrapper[4907]: E0313 14:05:57.508461 4907 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 13 14:05:57 crc kubenswrapper[4907]: I0313 14:05:57.511494 4907 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 14:05:57 crc kubenswrapper[4907]: I0313 14:05:57.513279 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:05:57 crc kubenswrapper[4907]: I0313 14:05:57.513329 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:05:57 crc kubenswrapper[4907]: I0313 14:05:57.513396 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:05:57 crc kubenswrapper[4907]: I0313 14:05:57.513476 4907 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 13 14:05:57 crc kubenswrapper[4907]: E0313 14:05:57.519134 4907 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 13 14:05:57 crc kubenswrapper[4907]: I0313 14:05:57.719700 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 14:05:58 crc kubenswrapper[4907]: I0313 14:05:58.717296 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 14:05:59 crc kubenswrapper[4907]: I0313 14:05:59.339825 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 13 14:05:59 crc kubenswrapper[4907]: I0313 14:05:59.340089 4907 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 14:05:59 crc kubenswrapper[4907]: I0313 14:05:59.341394 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:05:59 crc kubenswrapper[4907]: I0313 14:05:59.341425 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:05:59 crc kubenswrapper[4907]: I0313 14:05:59.341433 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:05:59 crc kubenswrapper[4907]: I0313 14:05:59.351140 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 13 14:05:59 crc kubenswrapper[4907]: I0313 14:05:59.720208 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 14:06:00 crc kubenswrapper[4907]: I0313 14:06:00.048830 4907 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 14:06:00 crc kubenswrapper[4907]: I0313 14:06:00.049920 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:00 crc kubenswrapper[4907]: I0313 14:06:00.049953 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:00 crc kubenswrapper[4907]: I0313 14:06:00.049963 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:00 crc kubenswrapper[4907]: I0313 14:06:00.718477 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 14:06:00 crc kubenswrapper[4907]: I0313 14:06:00.782384 4907 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 14:06:00 crc kubenswrapper[4907]: I0313 14:06:00.783592 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:00 crc kubenswrapper[4907]: I0313 14:06:00.783635 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:00 crc kubenswrapper[4907]: I0313 14:06:00.783648 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:00 crc kubenswrapper[4907]: I0313 14:06:00.784316 4907 scope.go:117] "RemoveContainer" containerID="da9239d5489f168f1e0695a6eb1d540c331de494e29ae89624cfbfed390520db" Mar 13 14:06:00 crc kubenswrapper[4907]: E0313 14:06:00.784528 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 13 14:06:01 crc kubenswrapper[4907]: I0313 14:06:01.716711 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 14:06:01 crc kubenswrapper[4907]: E0313 14:06:01.872401 4907 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 13 14:06:02 crc kubenswrapper[4907]: I0313 14:06:02.716998 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 14:06:03 crc kubenswrapper[4907]: I0313 14:06:03.717634 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 14:06:04 crc kubenswrapper[4907]: E0313 14:06:04.515662 4907 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 13 14:06:04 crc kubenswrapper[4907]: I0313 14:06:04.520067 4907 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 14:06:04 crc kubenswrapper[4907]: I0313 14:06:04.521659 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:04 crc kubenswrapper[4907]: I0313 14:06:04.521714 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:04 crc kubenswrapper[4907]: I0313 14:06:04.521732 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:04 crc kubenswrapper[4907]: I0313 14:06:04.521763 4907 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 13 14:06:04 crc kubenswrapper[4907]: E0313 14:06:04.528424 4907 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 13 14:06:04 crc kubenswrapper[4907]: I0313 14:06:04.718175 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 14:06:05 crc kubenswrapper[4907]: I0313 14:06:05.715954 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 14:06:06 crc kubenswrapper[4907]: I0313 14:06:06.715471 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 14:06:07 crc kubenswrapper[4907]: I0313 14:06:07.717082 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 14:06:08 crc kubenswrapper[4907]: I0313 14:06:08.717806 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 14:06:09 crc kubenswrapper[4907]: I0313 14:06:09.718513 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 14:06:10 crc kubenswrapper[4907]: I0313 14:06:10.719468 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 14:06:11 crc kubenswrapper[4907]: E0313 14:06:11.523942 4907 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 13 14:06:11 crc kubenswrapper[4907]: I0313 14:06:11.529074 4907 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 14:06:11 crc kubenswrapper[4907]: I0313 14:06:11.531109 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:11 crc kubenswrapper[4907]: I0313 14:06:11.531151 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:11 crc kubenswrapper[4907]: I0313 14:06:11.531163 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:11 crc kubenswrapper[4907]: I0313 14:06:11.531196 4907 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 13 14:06:11 crc kubenswrapper[4907]: E0313 14:06:11.538168 4907 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 13 14:06:11 crc kubenswrapper[4907]: I0313 14:06:11.717069 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 14:06:11 crc kubenswrapper[4907]: E0313 14:06:11.873335 4907 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 13 14:06:12 crc kubenswrapper[4907]: I0313 14:06:12.718498 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 14:06:12 crc kubenswrapper[4907]: I0313 14:06:12.781746 4907 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 14:06:12 crc kubenswrapper[4907]: I0313 14:06:12.783039 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:12 crc kubenswrapper[4907]: I0313 14:06:12.783079 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:12 crc kubenswrapper[4907]: I0313 14:06:12.783097 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:12 crc kubenswrapper[4907]: I0313 14:06:12.783802 4907 scope.go:117] "RemoveContainer" containerID="da9239d5489f168f1e0695a6eb1d540c331de494e29ae89624cfbfed390520db" Mar 13 14:06:13 crc kubenswrapper[4907]: I0313 14:06:13.091671 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Mar 13 14:06:13 crc kubenswrapper[4907]: I0313 14:06:13.093921 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80"} Mar 13 14:06:13 crc kubenswrapper[4907]: I0313 14:06:13.094094 4907 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 14:06:13 crc kubenswrapper[4907]: I0313 14:06:13.095098 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:13 crc kubenswrapper[4907]: I0313 14:06:13.095138 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:13 crc kubenswrapper[4907]: I0313 14:06:13.095148 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:13 crc kubenswrapper[4907]: I0313 14:06:13.286092 4907 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Mar 13 14:06:13 crc kubenswrapper[4907]: I0313 14:06:13.300093 4907 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Mar 13 14:06:13 crc kubenswrapper[4907]: I0313 14:06:13.715373 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 14:06:14 crc kubenswrapper[4907]: I0313 14:06:14.098876 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Mar 13 14:06:14 crc kubenswrapper[4907]: I0313 14:06:14.099696 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Mar 13 14:06:14 crc kubenswrapper[4907]: I0313 14:06:14.101730 4907 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80" exitCode=255 Mar 13 14:06:14 crc kubenswrapper[4907]: I0313 14:06:14.101782 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80"} Mar 13 14:06:14 crc kubenswrapper[4907]: I0313 14:06:14.101827 4907 scope.go:117] "RemoveContainer" containerID="da9239d5489f168f1e0695a6eb1d540c331de494e29ae89624cfbfed390520db" Mar 13 14:06:14 crc kubenswrapper[4907]: I0313 14:06:14.102034 4907 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 14:06:14 crc kubenswrapper[4907]: I0313 14:06:14.102951 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:14 crc kubenswrapper[4907]: I0313 14:06:14.102992 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:14 crc kubenswrapper[4907]: I0313 14:06:14.103007 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:14 crc kubenswrapper[4907]: I0313 14:06:14.103708 4907 scope.go:117] "RemoveContainer" containerID="1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80" Mar 13 14:06:14 crc kubenswrapper[4907]: E0313 14:06:14.103944 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 13 14:06:14 crc kubenswrapper[4907]: I0313 14:06:14.718620 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 14:06:15 crc kubenswrapper[4907]: I0313 14:06:15.108138 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Mar 13 14:06:15 crc kubenswrapper[4907]: I0313 14:06:15.164624 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 14:06:15 crc kubenswrapper[4907]: I0313 14:06:15.164964 4907 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 14:06:15 crc kubenswrapper[4907]: I0313 14:06:15.166169 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:15 crc kubenswrapper[4907]: I0313 14:06:15.166219 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:15 crc kubenswrapper[4907]: I0313 14:06:15.166234 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:15 crc kubenswrapper[4907]: I0313 14:06:15.167044 4907 scope.go:117] "RemoveContainer" containerID="1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80" Mar 13 14:06:15 crc kubenswrapper[4907]: E0313 14:06:15.167245 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 13 14:06:15 crc kubenswrapper[4907]: W0313 14:06:15.302390 4907 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User "system:anonymous" cannot list resource "csidrivers" in API group "storage.k8s.io" at the cluster scope Mar 13 14:06:15 crc kubenswrapper[4907]: E0313 14:06:15.302501 4907 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User \"system:anonymous\" cannot list resource \"csidrivers\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError" Mar 13 14:06:15 crc kubenswrapper[4907]: I0313 14:06:15.716786 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 14:06:15 crc kubenswrapper[4907]: I0313 14:06:15.824392 4907 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 14:06:16 crc kubenswrapper[4907]: I0313 14:06:16.114455 4907 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 14:06:16 crc kubenswrapper[4907]: I0313 14:06:16.115353 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:16 crc kubenswrapper[4907]: I0313 14:06:16.115413 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:16 crc kubenswrapper[4907]: I0313 14:06:16.115431 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:16 crc kubenswrapper[4907]: I0313 14:06:16.116408 4907 scope.go:117] "RemoveContainer" containerID="1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80" Mar 13 14:06:16 crc kubenswrapper[4907]: E0313 14:06:16.116691 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 13 14:06:16 crc kubenswrapper[4907]: I0313 14:06:16.716383 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 14:06:17 crc kubenswrapper[4907]: I0313 14:06:17.716098 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 14:06:18 crc kubenswrapper[4907]: E0313 14:06:18.532060 4907 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 13 14:06:18 crc kubenswrapper[4907]: I0313 14:06:18.538671 4907 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 14:06:18 crc kubenswrapper[4907]: I0313 14:06:18.540201 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:18 crc kubenswrapper[4907]: I0313 14:06:18.540292 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:18 crc kubenswrapper[4907]: I0313 14:06:18.540355 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:18 crc kubenswrapper[4907]: I0313 14:06:18.540398 4907 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 13 14:06:18 crc kubenswrapper[4907]: E0313 14:06:18.548688 4907 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 13 14:06:18 crc kubenswrapper[4907]: I0313 14:06:18.719389 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 14:06:19 crc kubenswrapper[4907]: I0313 14:06:19.714977 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 14:06:19 crc kubenswrapper[4907]: I0313 14:06:19.831796 4907 csr.go:261] certificate signing request csr-db9vj is approved, waiting to be issued Mar 13 14:06:19 crc kubenswrapper[4907]: I0313 14:06:19.840220 4907 csr.go:257] certificate signing request csr-db9vj is issued Mar 13 14:06:19 crc kubenswrapper[4907]: I0313 14:06:19.935138 4907 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Mar 13 14:06:20 crc kubenswrapper[4907]: I0313 14:06:20.577576 4907 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Mar 13 14:06:20 crc kubenswrapper[4907]: I0313 14:06:20.842022 4907 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2027-02-24 05:54:36 +0000 UTC, rotation deadline is 2026-11-24 22:13:07.011008384 +0000 UTC Mar 13 14:06:20 crc kubenswrapper[4907]: I0313 14:06:20.842099 4907 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 6152h6m46.168914302s for next certificate rotation Mar 13 14:06:21 crc kubenswrapper[4907]: E0313 14:06:21.874204 4907 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 13 14:06:25 crc kubenswrapper[4907]: I0313 14:06:25.549656 4907 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 14:06:25 crc kubenswrapper[4907]: I0313 14:06:25.552031 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:25 crc kubenswrapper[4907]: I0313 14:06:25.552122 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:25 crc kubenswrapper[4907]: I0313 14:06:25.552148 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:25 crc kubenswrapper[4907]: I0313 14:06:25.552344 4907 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 13 14:06:25 crc kubenswrapper[4907]: I0313 14:06:25.564478 4907 kubelet_node_status.go:115] "Node was previously registered" node="crc" Mar 13 14:06:25 crc kubenswrapper[4907]: I0313 14:06:25.564857 4907 kubelet_node_status.go:79] "Successfully registered node" node="crc" Mar 13 14:06:25 crc kubenswrapper[4907]: E0313 14:06:25.564927 4907 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": node \"crc\" not found" Mar 13 14:06:25 crc kubenswrapper[4907]: I0313 14:06:25.570517 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:25 crc kubenswrapper[4907]: I0313 14:06:25.570570 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:25 crc kubenswrapper[4907]: I0313 14:06:25.570585 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:25 crc kubenswrapper[4907]: I0313 14:06:25.570609 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:25 crc kubenswrapper[4907]: I0313 14:06:25.570626 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:25Z","lastTransitionTime":"2026-03-13T14:06:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:25 crc kubenswrapper[4907]: E0313 14:06:25.590615 4907 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4b182bae-7f2d-4b54-934c-85a9a277d132\\\",\\\"systemUUID\\\":\\\"05bfbfa3-f04f-45ba-8233-a355fc774c24\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 14:06:25 crc kubenswrapper[4907]: I0313 14:06:25.602231 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:25 crc kubenswrapper[4907]: I0313 14:06:25.602296 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:25 crc kubenswrapper[4907]: I0313 14:06:25.602319 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:25 crc kubenswrapper[4907]: I0313 14:06:25.602348 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:25 crc kubenswrapper[4907]: I0313 14:06:25.602369 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:25Z","lastTransitionTime":"2026-03-13T14:06:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:25 crc kubenswrapper[4907]: E0313 14:06:25.618805 4907 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4b182bae-7f2d-4b54-934c-85a9a277d132\\\",\\\"systemUUID\\\":\\\"05bfbfa3-f04f-45ba-8233-a355fc774c24\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 14:06:25 crc kubenswrapper[4907]: I0313 14:06:25.630098 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:25 crc kubenswrapper[4907]: I0313 14:06:25.630206 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:25 crc kubenswrapper[4907]: I0313 14:06:25.630225 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:25 crc kubenswrapper[4907]: I0313 14:06:25.630253 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:25 crc kubenswrapper[4907]: I0313 14:06:25.630271 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:25Z","lastTransitionTime":"2026-03-13T14:06:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:25 crc kubenswrapper[4907]: E0313 14:06:25.648941 4907 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4b182bae-7f2d-4b54-934c-85a9a277d132\\\",\\\"systemUUID\\\":\\\"05bfbfa3-f04f-45ba-8233-a355fc774c24\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 14:06:25 crc kubenswrapper[4907]: I0313 14:06:25.660313 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:25 crc kubenswrapper[4907]: I0313 14:06:25.660365 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:25 crc kubenswrapper[4907]: I0313 14:06:25.660382 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:25 crc kubenswrapper[4907]: I0313 14:06:25.660406 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:25 crc kubenswrapper[4907]: I0313 14:06:25.660426 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:25Z","lastTransitionTime":"2026-03-13T14:06:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:25 crc kubenswrapper[4907]: E0313 14:06:25.677815 4907 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4b182bae-7f2d-4b54-934c-85a9a277d132\\\",\\\"systemUUID\\\":\\\"05bfbfa3-f04f-45ba-8233-a355fc774c24\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 14:06:25 crc kubenswrapper[4907]: E0313 14:06:25.678116 4907 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 13 14:06:25 crc kubenswrapper[4907]: E0313 14:06:25.678166 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:25 crc kubenswrapper[4907]: E0313 14:06:25.779115 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:25 crc kubenswrapper[4907]: I0313 14:06:25.781498 4907 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 14:06:25 crc kubenswrapper[4907]: I0313 14:06:25.783174 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:25 crc kubenswrapper[4907]: I0313 14:06:25.783227 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:25 crc kubenswrapper[4907]: I0313 14:06:25.783246 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:25 crc kubenswrapper[4907]: E0313 14:06:25.880036 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:25 crc kubenswrapper[4907]: E0313 14:06:25.980763 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:26 crc kubenswrapper[4907]: E0313 14:06:26.081324 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:26 crc kubenswrapper[4907]: E0313 14:06:26.182324 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:26 crc kubenswrapper[4907]: E0313 14:06:26.282802 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:26 crc kubenswrapper[4907]: E0313 14:06:26.383859 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:26 crc kubenswrapper[4907]: E0313 14:06:26.484172 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:26 crc kubenswrapper[4907]: E0313 14:06:26.584689 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:26 crc kubenswrapper[4907]: E0313 14:06:26.684853 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:26 crc kubenswrapper[4907]: E0313 14:06:26.785798 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:26 crc kubenswrapper[4907]: E0313 14:06:26.886248 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:26 crc kubenswrapper[4907]: E0313 14:06:26.986944 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:27 crc kubenswrapper[4907]: E0313 14:06:27.088081 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:27 crc kubenswrapper[4907]: E0313 14:06:27.189170 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:27 crc kubenswrapper[4907]: E0313 14:06:27.289973 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:27 crc kubenswrapper[4907]: E0313 14:06:27.391020 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:27 crc kubenswrapper[4907]: E0313 14:06:27.491722 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:27 crc kubenswrapper[4907]: E0313 14:06:27.592345 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:27 crc kubenswrapper[4907]: E0313 14:06:27.692461 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:27 crc kubenswrapper[4907]: I0313 14:06:27.781478 4907 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 14:06:27 crc kubenswrapper[4907]: I0313 14:06:27.783030 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:27 crc kubenswrapper[4907]: I0313 14:06:27.783171 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:27 crc kubenswrapper[4907]: I0313 14:06:27.783197 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:27 crc kubenswrapper[4907]: E0313 14:06:27.792993 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:27 crc kubenswrapper[4907]: E0313 14:06:27.894064 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:27 crc kubenswrapper[4907]: E0313 14:06:27.994426 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:28 crc kubenswrapper[4907]: E0313 14:06:28.094985 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:28 crc kubenswrapper[4907]: E0313 14:06:28.195571 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:28 crc kubenswrapper[4907]: E0313 14:06:28.296159 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:28 crc kubenswrapper[4907]: E0313 14:06:28.396934 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:28 crc kubenswrapper[4907]: E0313 14:06:28.497341 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:28 crc kubenswrapper[4907]: E0313 14:06:28.598460 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:28 crc kubenswrapper[4907]: E0313 14:06:28.699406 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:28 crc kubenswrapper[4907]: E0313 14:06:28.799511 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:28 crc kubenswrapper[4907]: E0313 14:06:28.900016 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:29 crc kubenswrapper[4907]: E0313 14:06:29.000613 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:29 crc kubenswrapper[4907]: E0313 14:06:29.101693 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:29 crc kubenswrapper[4907]: E0313 14:06:29.202784 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:29 crc kubenswrapper[4907]: E0313 14:06:29.303481 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:29 crc kubenswrapper[4907]: E0313 14:06:29.404537 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:29 crc kubenswrapper[4907]: E0313 14:06:29.505447 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:29 crc kubenswrapper[4907]: E0313 14:06:29.605983 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:29 crc kubenswrapper[4907]: E0313 14:06:29.706798 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:29 crc kubenswrapper[4907]: I0313 14:06:29.782259 4907 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 14:06:29 crc kubenswrapper[4907]: I0313 14:06:29.783482 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:29 crc kubenswrapper[4907]: I0313 14:06:29.783529 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:29 crc kubenswrapper[4907]: I0313 14:06:29.783539 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:29 crc kubenswrapper[4907]: I0313 14:06:29.784152 4907 scope.go:117] "RemoveContainer" containerID="1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80" Mar 13 14:06:29 crc kubenswrapper[4907]: E0313 14:06:29.784324 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 13 14:06:29 crc kubenswrapper[4907]: E0313 14:06:29.807316 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:29 crc kubenswrapper[4907]: E0313 14:06:29.908361 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:30 crc kubenswrapper[4907]: E0313 14:06:30.008731 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:30 crc kubenswrapper[4907]: E0313 14:06:30.109269 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:30 crc kubenswrapper[4907]: E0313 14:06:30.210094 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:30 crc kubenswrapper[4907]: I0313 14:06:30.242507 4907 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Mar 13 14:06:30 crc kubenswrapper[4907]: E0313 14:06:30.310929 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:30 crc kubenswrapper[4907]: E0313 14:06:30.411961 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:30 crc kubenswrapper[4907]: E0313 14:06:30.512805 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:30 crc kubenswrapper[4907]: E0313 14:06:30.613477 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:30 crc kubenswrapper[4907]: E0313 14:06:30.714336 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:30 crc kubenswrapper[4907]: E0313 14:06:30.815461 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:30 crc kubenswrapper[4907]: E0313 14:06:30.916484 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:31 crc kubenswrapper[4907]: E0313 14:06:31.017589 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:31 crc kubenswrapper[4907]: E0313 14:06:31.118500 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:31 crc kubenswrapper[4907]: E0313 14:06:31.219423 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:31 crc kubenswrapper[4907]: E0313 14:06:31.320328 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:31 crc kubenswrapper[4907]: E0313 14:06:31.421346 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:31 crc kubenswrapper[4907]: E0313 14:06:31.521736 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:31 crc kubenswrapper[4907]: E0313 14:06:31.622008 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:31 crc kubenswrapper[4907]: E0313 14:06:31.723118 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:31 crc kubenswrapper[4907]: E0313 14:06:31.824133 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:31 crc kubenswrapper[4907]: E0313 14:06:31.875551 4907 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 13 14:06:31 crc kubenswrapper[4907]: E0313 14:06:31.924468 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:32 crc kubenswrapper[4907]: E0313 14:06:32.025478 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:32 crc kubenswrapper[4907]: E0313 14:06:32.126188 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:32 crc kubenswrapper[4907]: E0313 14:06:32.226920 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:32 crc kubenswrapper[4907]: E0313 14:06:32.327746 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:32 crc kubenswrapper[4907]: E0313 14:06:32.427945 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:32 crc kubenswrapper[4907]: E0313 14:06:32.528841 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:32 crc kubenswrapper[4907]: E0313 14:06:32.629911 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.686265 4907 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.732184 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.732226 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.732236 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.732253 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.732263 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:32Z","lastTransitionTime":"2026-03-13T14:06:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.777817 4907 apiserver.go:52] "Watching apiserver" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.783177 4907 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.783585 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf"] Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.784101 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.784136 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:06:32 crc kubenswrapper[4907]: E0313 14:06:32.784358 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.784145 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:06:32 crc kubenswrapper[4907]: E0313 14:06:32.784407 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.784538 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.785489 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:06:32 crc kubenswrapper[4907]: E0313 14:06:32.785609 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.786398 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.786797 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.786456 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.786792 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.786923 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.787010 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.787068 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.787115 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.788290 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.790360 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.817491 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.819368 4907 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.831038 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.837427 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.837506 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.837519 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.837540 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.837552 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:32Z","lastTransitionTime":"2026-03-13T14:06:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.846022 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.858939 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.870287 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.880785 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.893851 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.907236 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.917631 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.917668 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.917689 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.917709 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.917732 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.917749 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.917765 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.917782 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.917801 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.917815 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.917830 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.917847 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.917864 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.917902 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.917919 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.917935 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.917952 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.917974 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.917991 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.918007 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.918024 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.918043 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.918061 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.918267 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.918473 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.918815 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.918861 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.918904 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.918840 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.919196 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.919364 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.919399 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.919423 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.919448 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.919470 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.919522 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.919499 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.919871 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.919917 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.920137 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.920137 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.920155 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.920165 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.920173 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.920301 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.920331 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.920352 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.920375 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.920439 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.920730 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.920741 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.920737 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.920769 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.920803 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.920826 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.920855 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.920902 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.920925 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.920944 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.920963 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.921505 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.921589 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.921841 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.921941 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.921960 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.922040 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.922274 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.922278 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.922383 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.922407 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.922584 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.922591 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.922805 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.922978 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.922938 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.923031 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.923063 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.923076 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.923243 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.923277 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.923328 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.923285 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.923364 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.923455 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.923484 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.923490 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.923509 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.923532 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.923554 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.923544 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.923576 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.923660 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.923706 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.923740 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.923776 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.923796 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.923824 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.924210 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.924270 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.923826 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.923967 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.924998 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.925019 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.925039 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.925066 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.925104 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.925147 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.925192 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.925242 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.924181 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.925265 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.925297 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.925334 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.925346 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.925393 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.924618 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.924685 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.925451 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.925503 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.925548 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.925581 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.925453 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.925618 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.925491 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.925673 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.926063 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.926115 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.926166 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.926213 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.926262 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.926319 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.926367 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.926422 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.926469 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.926508 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.926555 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.926609 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.926661 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.926711 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.926762 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.926813 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.926858 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.927049 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.927110 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.927159 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.927209 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.927254 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.927290 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.927327 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.927376 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.927425 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.927472 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.927512 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.925796 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.926841 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.926995 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.927301 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.927551 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.927560 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.927662 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.927773 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.927800 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.927822 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.927934 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.927957 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.927974 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.927999 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.928033 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.928033 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.928057 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.928152 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.928202 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.928242 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.928281 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.928315 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.928351 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.928399 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.928436 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.928593 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.928635 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.928671 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.928706 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.928735 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.928767 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.928799 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.928831 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.928877 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.928931 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.928963 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.929002 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.929041 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.929079 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.929117 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.929149 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.929185 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.929219 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.929250 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.929280 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.929308 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.929336 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.929487 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.929535 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.929572 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.929609 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.929642 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.929677 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.929712 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.929748 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.929784 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.929819 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.929854 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.929945 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.929979 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.930014 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.930045 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.930077 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.930112 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.930146 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.930186 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.930222 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.930258 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.930299 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.930332 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.930367 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.930400 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.930432 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.930472 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.930505 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.930538 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.930571 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.930604 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.930637 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.930669 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.930703 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.930737 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.930770 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.930808 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.930844 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.930902 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.930935 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.930969 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.931005 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.931041 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.931079 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.931122 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.933255 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.933946 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.933996 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.934034 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.934123 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.934169 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.934209 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.934246 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.934282 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.934321 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.934439 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.934482 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.934522 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.934561 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.934600 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.934635 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.934673 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.934747 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.934859 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.934919 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.934943 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.934963 4907 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.934984 4907 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.935006 4907 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.935027 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.935050 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.935070 4907 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.935089 4907 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.935109 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.935138 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.935158 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.935178 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.935197 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.935216 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.935234 4907 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.935253 4907 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.935275 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.935294 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.935312 4907 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.935330 4907 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.935346 4907 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.935364 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.935383 4907 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.935405 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.935424 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.935441 4907 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.935462 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.935480 4907 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.935500 4907 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.935520 4907 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.935541 4907 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.935561 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.935579 4907 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.935599 4907 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.935619 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.935638 4907 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.935656 4907 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.935673 4907 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.935691 4907 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.935712 4907 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.935731 4907 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.935752 4907 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.935770 4907 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.935790 4907 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.935809 4907 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.935828 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.935847 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.935867 4907 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.935938 4907 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.935957 4907 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.935976 4907 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.935996 4907 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.936929 4907 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.928281 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: E0313 14:06:32.941998 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:06:33.441978151 +0000 UTC m=+92.341765840 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.941980 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.942184 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.928230 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.942207 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.928551 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.928910 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.942955 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.944003 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.943000 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.944700 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.945106 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.943218 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.945632 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.945929 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.945988 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.946174 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.946255 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.946314 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.946409 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.929033 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.929415 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.930138 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.930459 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.930779 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.930786 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.947483 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.930811 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.930809 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.947524 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.947543 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.930982 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.931289 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.931482 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.931563 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.931533 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.947588 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.931643 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.931833 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.932099 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.932204 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.932592 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.932591 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.932678 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.932922 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.947976 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.933047 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.933423 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.948080 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.933611 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.933945 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.933981 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.933980 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.934660 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.935255 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.935465 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.935380 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.935534 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.935582 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.936114 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.936373 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.936341 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.936936 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.937094 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.937297 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.938290 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.938545 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.938650 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.939502 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.939519 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.939766 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.939813 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.940019 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.940045 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.940350 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.940384 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.940632 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.940929 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.941024 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.941062 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.941336 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.941521 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.941594 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.941828 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.941839 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.941856 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.942245 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.942237 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.942251 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.942397 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.948597 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.929008 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.946637 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.947358 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.942673 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.948783 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.949218 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.949535 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.950078 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: E0313 14:06:32.950795 4907 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 13 14:06:32 crc kubenswrapper[4907]: E0313 14:06:32.964238 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-13 14:06:33.464202375 +0000 UTC m=+92.363990064 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.964426 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 13 14:06:32 crc kubenswrapper[4907]: E0313 14:06:32.964493 4907 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 13 14:06:32 crc kubenswrapper[4907]: E0313 14:06:32.964527 4907 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.950832 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.949759 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.957970 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.964590 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.964615 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.951036 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.964638 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:32 crc kubenswrapper[4907]: E0313 14:06:32.951422 4907 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.952158 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.952661 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.956679 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: E0313 14:06:32.964567 4907 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 14:06:32 crc kubenswrapper[4907]: E0313 14:06:32.964677 4907 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.964655 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:32Z","lastTransitionTime":"2026-03-13T14:06:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:32 crc kubenswrapper[4907]: E0313 14:06:32.964707 4907 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 13 14:06:32 crc kubenswrapper[4907]: E0313 14:06:32.964728 4907 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 14:06:32 crc kubenswrapper[4907]: E0313 14:06:32.964766 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-13 14:06:33.46472099 +0000 UTC m=+92.364508679 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 14:06:32 crc kubenswrapper[4907]: E0313 14:06:32.964802 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-13 14:06:33.464777672 +0000 UTC m=+92.364565371 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 14:06:32 crc kubenswrapper[4907]: E0313 14:06:32.965041 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-13 14:06:33.46502194 +0000 UTC m=+92.364809629 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.965232 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.970512 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.971172 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.972479 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.976743 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.977541 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.977840 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.982010 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.982150 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.982242 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.982833 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.982910 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.982957 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.983060 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.983726 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.983938 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.984997 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.985026 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.985109 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.985692 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.985692 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.986000 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.986065 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.986078 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.986097 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.986230 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.987354 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.987934 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.988319 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.988594 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.988710 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.988798 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.988798 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.989144 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.989211 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.989248 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.989301 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.992102 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.001379 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.001842 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.002524 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.015236 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.037261 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.037351 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.037351 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.037408 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.037422 4907 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.037431 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.037441 4907 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.037451 4907 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.037460 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.037470 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.037481 4907 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.037490 4907 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.037499 4907 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.037508 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.037516 4907 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.037525 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.037535 4907 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.037543 4907 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.037551 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.037562 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.037573 4907 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.037585 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.037596 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.037606 4907 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.037614 4907 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.037622 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.037630 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.037638 4907 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.037646 4907 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.037655 4907 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.037664 4907 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.037672 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.037680 4907 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.037688 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.037697 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.037588 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.037705 4907 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.037760 4907 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.037772 4907 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.037783 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.037793 4907 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.037804 4907 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.037817 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.037829 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.037839 4907 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.037850 4907 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.037859 4907 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.037869 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.037897 4907 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.037907 4907 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.037916 4907 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.037927 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.037937 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.037946 4907 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.037955 4907 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.037964 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.037973 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.037982 4907 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.037992 4907 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038002 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038011 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038020 4907 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038029 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038039 4907 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038048 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038057 4907 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038067 4907 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038076 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038088 4907 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038098 4907 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038109 4907 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038120 4907 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038130 4907 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038141 4907 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038150 4907 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038188 4907 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038199 4907 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038208 4907 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038221 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038231 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038240 4907 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038249 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038258 4907 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038268 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038277 4907 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038286 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038295 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038307 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038317 4907 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038326 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038335 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038343 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038352 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038362 4907 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038370 4907 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038379 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038388 4907 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038399 4907 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038409 4907 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038421 4907 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038431 4907 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038440 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038450 4907 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038460 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038470 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038479 4907 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038489 4907 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038498 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038508 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038520 4907 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038530 4907 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038539 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038548 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038557 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038568 4907 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038576 4907 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038584 4907 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038595 4907 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038604 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038614 4907 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038623 4907 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038632 4907 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038641 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038650 4907 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038659 4907 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038667 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038676 4907 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038685 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038694 4907 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038704 4907 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038715 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038724 4907 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038734 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038742 4907 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038752 4907 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038761 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038770 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038779 4907 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038787 4907 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038796 4907 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038805 4907 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038816 4907 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038824 4907 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038833 4907 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.066988 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.067031 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.067041 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.067060 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.067074 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:33Z","lastTransitionTime":"2026-03-13T14:06:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.110584 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.125416 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 13 14:06:33 crc kubenswrapper[4907]: E0313 14:06:33.125823 4907 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 13 14:06:33 crc kubenswrapper[4907]: container &Container{Name:network-operator,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,Command:[/bin/bash -c #!/bin/bash Mar 13 14:06:33 crc kubenswrapper[4907]: set -o allexport Mar 13 14:06:33 crc kubenswrapper[4907]: if [[ -f /etc/kubernetes/apiserver-url.env ]]; then Mar 13 14:06:33 crc kubenswrapper[4907]: source /etc/kubernetes/apiserver-url.env Mar 13 14:06:33 crc kubenswrapper[4907]: else Mar 13 14:06:33 crc kubenswrapper[4907]: echo "Error: /etc/kubernetes/apiserver-url.env is missing" Mar 13 14:06:33 crc kubenswrapper[4907]: exit 1 Mar 13 14:06:33 crc kubenswrapper[4907]: fi Mar 13 14:06:33 crc kubenswrapper[4907]: exec /usr/bin/cluster-network-operator start --listen=0.0.0.0:9104 Mar 13 14:06:33 crc kubenswrapper[4907]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:cno,HostPort:9104,ContainerPort:9104,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:RELEASE_VERSION,Value:4.18.1,ValueFrom:nil,},EnvVar{Name:KUBE_PROXY_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b97554198294bf544fbc116c94a0a1fb2ec8a4de0e926bf9d9e320135f0bee6f,ValueFrom:nil,},EnvVar{Name:KUBE_RBAC_PROXY_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09,ValueFrom:nil,},EnvVar{Name:MULTUS_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26,ValueFrom:nil,},EnvVar{Name:MULTUS_ADMISSION_CONTROLLER_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317,ValueFrom:nil,},EnvVar{Name:CNI_PLUGINS_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc,ValueFrom:nil,},EnvVar{Name:BOND_CNI_PLUGIN_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78,ValueFrom:nil,},EnvVar{Name:WHEREABOUTS_CNI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4,ValueFrom:nil,},EnvVar{Name:ROUTE_OVERRRIDE_CNI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa,ValueFrom:nil,},EnvVar{Name:MULTUS_NETWORKPOLICY_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:23f833d3738d68706eb2f2868bd76bd71cee016cffa6faf5f045a60cc8c6eddd,ValueFrom:nil,},EnvVar{Name:OVN_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2,ValueFrom:nil,},EnvVar{Name:OVN_NB_RAFT_ELECTION_TIMER,Value:10,ValueFrom:nil,},EnvVar{Name:OVN_SB_RAFT_ELECTION_TIMER,Value:16,ValueFrom:nil,},EnvVar{Name:OVN_NORTHD_PROBE_INTERVAL,Value:10000,ValueFrom:nil,},EnvVar{Name:OVN_CONTROLLER_INACTIVITY_PROBE,Value:180000,ValueFrom:nil,},EnvVar{Name:OVN_NB_INACTIVITY_PROBE,Value:60000,ValueFrom:nil,},EnvVar{Name:EGRESS_ROUTER_CNI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c,ValueFrom:nil,},EnvVar{Name:NETWORK_METRICS_DAEMON_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d,ValueFrom:nil,},EnvVar{Name:NETWORK_CHECK_SOURCE_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,ValueFrom:nil,},EnvVar{Name:NETWORK_CHECK_TARGET_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,ValueFrom:nil,},EnvVar{Name:NETWORK_OPERATOR_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,ValueFrom:nil,},EnvVar{Name:CLOUD_NETWORK_CONFIG_CONTROLLER_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8048f1cb0be521f09749c0a489503cd56d85b68c6ca93380e082cfd693cd97a8,ValueFrom:nil,},EnvVar{Name:CLI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2,ValueFrom:nil,},EnvVar{Name:FRR_K8S_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5dbf844e49bb46b78586930149e5e5f5dc121014c8afd10fe36f3651967cc256,ValueFrom:nil,},EnvVar{Name:NETWORKING_CONSOLE_PLUGIN_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd,ValueFrom:nil,},EnvVar{Name:POD_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.name,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:host-etc-kube,ReadOnly:true,MountPath:/etc/kubernetes,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:metrics-tls,ReadOnly:false,MountPath:/var/run/secrets/serving-cert,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rdwmf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod network-operator-58b4c7f79c-55gtf_openshift-network-operator(37a5e44f-9a88-4405-be8a-b645485e7312): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 13 14:06:33 crc kubenswrapper[4907]: > logger="UnhandledError" Mar 13 14:06:33 crc kubenswrapper[4907]: E0313 14:06:33.127247 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"network-operator\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" podUID="37a5e44f-9a88-4405-be8a-b645485e7312" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.137544 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 13 14:06:33 crc kubenswrapper[4907]: E0313 14:06:33.138410 4907 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 13 14:06:33 crc kubenswrapper[4907]: container &Container{Name:webhook,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2,Command:[/bin/bash -c set -xe Mar 13 14:06:33 crc kubenswrapper[4907]: if [[ -f "/env/_master" ]]; then Mar 13 14:06:33 crc kubenswrapper[4907]: set -o allexport Mar 13 14:06:33 crc kubenswrapper[4907]: source "/env/_master" Mar 13 14:06:33 crc kubenswrapper[4907]: set +o allexport Mar 13 14:06:33 crc kubenswrapper[4907]: fi Mar 13 14:06:33 crc kubenswrapper[4907]: # OVN-K will try to remove hybrid overlay node annotations even when the hybrid overlay is not enabled. Mar 13 14:06:33 crc kubenswrapper[4907]: # https://github.com/ovn-org/ovn-kubernetes/blob/ac6820df0b338a246f10f412cd5ec903bd234694/go-controller/pkg/ovn/master.go#L791 Mar 13 14:06:33 crc kubenswrapper[4907]: ho_enable="--enable-hybrid-overlay" Mar 13 14:06:33 crc kubenswrapper[4907]: echo "I$(date "+%m%d %H:%M:%S.%N") - network-node-identity - start webhook" Mar 13 14:06:33 crc kubenswrapper[4907]: # extra-allowed-user: service account `ovn-kubernetes-control-plane` Mar 13 14:06:33 crc kubenswrapper[4907]: # sets pod annotations in multi-homing layer3 network controller (cluster-manager) Mar 13 14:06:33 crc kubenswrapper[4907]: exec /usr/bin/ovnkube-identity --k8s-apiserver=https://api-int.crc.testing:6443 \ Mar 13 14:06:33 crc kubenswrapper[4907]: --webhook-cert-dir="/etc/webhook-cert" \ Mar 13 14:06:33 crc kubenswrapper[4907]: --webhook-host=127.0.0.1 \ Mar 13 14:06:33 crc kubenswrapper[4907]: --webhook-port=9743 \ Mar 13 14:06:33 crc kubenswrapper[4907]: ${ho_enable} \ Mar 13 14:06:33 crc kubenswrapper[4907]: --enable-interconnect \ Mar 13 14:06:33 crc kubenswrapper[4907]: --disable-approver \ Mar 13 14:06:33 crc kubenswrapper[4907]: --extra-allowed-user="system:serviceaccount:openshift-ovn-kubernetes:ovn-kubernetes-control-plane" \ Mar 13 14:06:33 crc kubenswrapper[4907]: --wait-for-kubernetes-api=200s \ Mar 13 14:06:33 crc kubenswrapper[4907]: --pod-admission-conditions="/var/run/ovnkube-identity-config/additional-pod-admission-cond.json" \ Mar 13 14:06:33 crc kubenswrapper[4907]: --loglevel="${LOGLEVEL}" Mar 13 14:06:33 crc kubenswrapper[4907]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOGLEVEL,Value:2,ValueFrom:nil,},EnvVar{Name:KUBERNETES_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:webhook-cert,ReadOnly:false,MountPath:/etc/webhook-cert/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:env-overrides,ReadOnly:false,MountPath:/env,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovnkube-identity-cm,ReadOnly:false,MountPath:/var/run/ovnkube-identity-config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-s2kz5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000470000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod network-node-identity-vrzqb_openshift-network-node-identity(ef543e1b-8068-4ea3-b32a-61027b32e95d): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 13 14:06:33 crc kubenswrapper[4907]: > logger="UnhandledError" Mar 13 14:06:33 crc kubenswrapper[4907]: E0313 14:06:33.141144 4907 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 13 14:06:33 crc kubenswrapper[4907]: container &Container{Name:approver,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2,Command:[/bin/bash -c set -xe Mar 13 14:06:33 crc kubenswrapper[4907]: if [[ -f "/env/_master" ]]; then Mar 13 14:06:33 crc kubenswrapper[4907]: set -o allexport Mar 13 14:06:33 crc kubenswrapper[4907]: source "/env/_master" Mar 13 14:06:33 crc kubenswrapper[4907]: set +o allexport Mar 13 14:06:33 crc kubenswrapper[4907]: fi Mar 13 14:06:33 crc kubenswrapper[4907]: Mar 13 14:06:33 crc kubenswrapper[4907]: echo "I$(date "+%m%d %H:%M:%S.%N") - network-node-identity - start approver" Mar 13 14:06:33 crc kubenswrapper[4907]: exec /usr/bin/ovnkube-identity --k8s-apiserver=https://api-int.crc.testing:6443 \ Mar 13 14:06:33 crc kubenswrapper[4907]: --disable-webhook \ Mar 13 14:06:33 crc kubenswrapper[4907]: --csr-acceptance-conditions="/var/run/ovnkube-identity-config/additional-cert-acceptance-cond.json" \ Mar 13 14:06:33 crc kubenswrapper[4907]: --loglevel="${LOGLEVEL}" Mar 13 14:06:33 crc kubenswrapper[4907]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOGLEVEL,Value:4,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:env-overrides,ReadOnly:false,MountPath:/env,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovnkube-identity-cm,ReadOnly:false,MountPath:/var/run/ovnkube-identity-config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-s2kz5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000470000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod network-node-identity-vrzqb_openshift-network-node-identity(ef543e1b-8068-4ea3-b32a-61027b32e95d): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 13 14:06:33 crc kubenswrapper[4907]: > logger="UnhandledError" Mar 13 14:06:33 crc kubenswrapper[4907]: E0313 14:06:33.142359 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"webhook\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\", failed to \"StartContainer\" for \"approver\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"]" pod="openshift-network-node-identity/network-node-identity-vrzqb" podUID="ef543e1b-8068-4ea3-b32a-61027b32e95d" Mar 13 14:06:33 crc kubenswrapper[4907]: W0313 14:06:33.144872 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-53d2ba13c3bb87827b87557f503d37f81ee3023e9e9df3f8373171a31ea9559c WatchSource:0}: Error finding container 53d2ba13c3bb87827b87557f503d37f81ee3023e9e9df3f8373171a31ea9559c: Status 404 returned error can't find the container with id 53d2ba13c3bb87827b87557f503d37f81ee3023e9e9df3f8373171a31ea9559c Mar 13 14:06:33 crc kubenswrapper[4907]: E0313 14:06:33.147009 4907 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:iptables-alerter,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2,Command:[/iptables-alerter/iptables-alerter.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONTAINER_RUNTIME_ENDPOINT,Value:unix:///run/crio/crio.sock,ValueFrom:nil,},EnvVar{Name:ALERTER_POD_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.name,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{68157440 0} {} 65Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:iptables-alerter-script,ReadOnly:false,MountPath:/iptables-alerter,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host-slash,ReadOnly:true,MountPath:/host,SubPath:,MountPropagation:*HostToContainer,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rczfb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod iptables-alerter-4ln5h_openshift-network-operator(d75a4c96-2883-4a0b-bab2-0fab2b6c0b49): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars" logger="UnhandledError" Mar 13 14:06:33 crc kubenswrapper[4907]: E0313 14:06:33.148185 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"iptables-alerter\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-network-operator/iptables-alerter-4ln5h" podUID="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.163689 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"53d2ba13c3bb87827b87557f503d37f81ee3023e9e9df3f8373171a31ea9559c"} Mar 13 14:06:33 crc kubenswrapper[4907]: E0313 14:06:33.165185 4907 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:iptables-alerter,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2,Command:[/iptables-alerter/iptables-alerter.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONTAINER_RUNTIME_ENDPOINT,Value:unix:///run/crio/crio.sock,ValueFrom:nil,},EnvVar{Name:ALERTER_POD_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.name,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{68157440 0} {} 65Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:iptables-alerter-script,ReadOnly:false,MountPath:/iptables-alerter,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host-slash,ReadOnly:true,MountPath:/host,SubPath:,MountPropagation:*HostToContainer,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rczfb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod iptables-alerter-4ln5h_openshift-network-operator(d75a4c96-2883-4a0b-bab2-0fab2b6c0b49): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars" logger="UnhandledError" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.165303 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"21c666661c2637edf91b1da9c29671bdf2ed4f25e257c5e4d8f5b72db45fdf0b"} Mar 13 14:06:33 crc kubenswrapper[4907]: E0313 14:06:33.166348 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"iptables-alerter\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-network-operator/iptables-alerter-4ln5h" podUID="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.166587 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"37ca6097afd5a084f52dfd5153bb04aced74584bac8e714df3e5dd76919c5caa"} Mar 13 14:06:33 crc kubenswrapper[4907]: E0313 14:06:33.167217 4907 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 13 14:06:33 crc kubenswrapper[4907]: container &Container{Name:webhook,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2,Command:[/bin/bash -c set -xe Mar 13 14:06:33 crc kubenswrapper[4907]: if [[ -f "/env/_master" ]]; then Mar 13 14:06:33 crc kubenswrapper[4907]: set -o allexport Mar 13 14:06:33 crc kubenswrapper[4907]: source "/env/_master" Mar 13 14:06:33 crc kubenswrapper[4907]: set +o allexport Mar 13 14:06:33 crc kubenswrapper[4907]: fi Mar 13 14:06:33 crc kubenswrapper[4907]: # OVN-K will try to remove hybrid overlay node annotations even when the hybrid overlay is not enabled. Mar 13 14:06:33 crc kubenswrapper[4907]: # https://github.com/ovn-org/ovn-kubernetes/blob/ac6820df0b338a246f10f412cd5ec903bd234694/go-controller/pkg/ovn/master.go#L791 Mar 13 14:06:33 crc kubenswrapper[4907]: ho_enable="--enable-hybrid-overlay" Mar 13 14:06:33 crc kubenswrapper[4907]: echo "I$(date "+%m%d %H:%M:%S.%N") - network-node-identity - start webhook" Mar 13 14:06:33 crc kubenswrapper[4907]: # extra-allowed-user: service account `ovn-kubernetes-control-plane` Mar 13 14:06:33 crc kubenswrapper[4907]: # sets pod annotations in multi-homing layer3 network controller (cluster-manager) Mar 13 14:06:33 crc kubenswrapper[4907]: exec /usr/bin/ovnkube-identity --k8s-apiserver=https://api-int.crc.testing:6443 \ Mar 13 14:06:33 crc kubenswrapper[4907]: --webhook-cert-dir="/etc/webhook-cert" \ Mar 13 14:06:33 crc kubenswrapper[4907]: --webhook-host=127.0.0.1 \ Mar 13 14:06:33 crc kubenswrapper[4907]: --webhook-port=9743 \ Mar 13 14:06:33 crc kubenswrapper[4907]: ${ho_enable} \ Mar 13 14:06:33 crc kubenswrapper[4907]: --enable-interconnect \ Mar 13 14:06:33 crc kubenswrapper[4907]: --disable-approver \ Mar 13 14:06:33 crc kubenswrapper[4907]: --extra-allowed-user="system:serviceaccount:openshift-ovn-kubernetes:ovn-kubernetes-control-plane" \ Mar 13 14:06:33 crc kubenswrapper[4907]: --wait-for-kubernetes-api=200s \ Mar 13 14:06:33 crc kubenswrapper[4907]: --pod-admission-conditions="/var/run/ovnkube-identity-config/additional-pod-admission-cond.json" \ Mar 13 14:06:33 crc kubenswrapper[4907]: --loglevel="${LOGLEVEL}" Mar 13 14:06:33 crc kubenswrapper[4907]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOGLEVEL,Value:2,ValueFrom:nil,},EnvVar{Name:KUBERNETES_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:webhook-cert,ReadOnly:false,MountPath:/etc/webhook-cert/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:env-overrides,ReadOnly:false,MountPath:/env,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovnkube-identity-cm,ReadOnly:false,MountPath:/var/run/ovnkube-identity-config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-s2kz5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000470000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod network-node-identity-vrzqb_openshift-network-node-identity(ef543e1b-8068-4ea3-b32a-61027b32e95d): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 13 14:06:33 crc kubenswrapper[4907]: > logger="UnhandledError" Mar 13 14:06:33 crc kubenswrapper[4907]: E0313 14:06:33.168027 4907 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 13 14:06:33 crc kubenswrapper[4907]: container &Container{Name:network-operator,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,Command:[/bin/bash -c #!/bin/bash Mar 13 14:06:33 crc kubenswrapper[4907]: set -o allexport Mar 13 14:06:33 crc kubenswrapper[4907]: if [[ -f /etc/kubernetes/apiserver-url.env ]]; then Mar 13 14:06:33 crc kubenswrapper[4907]: source /etc/kubernetes/apiserver-url.env Mar 13 14:06:33 crc kubenswrapper[4907]: else Mar 13 14:06:33 crc kubenswrapper[4907]: echo "Error: /etc/kubernetes/apiserver-url.env is missing" Mar 13 14:06:33 crc kubenswrapper[4907]: exit 1 Mar 13 14:06:33 crc kubenswrapper[4907]: fi Mar 13 14:06:33 crc kubenswrapper[4907]: exec /usr/bin/cluster-network-operator start --listen=0.0.0.0:9104 Mar 13 14:06:33 crc kubenswrapper[4907]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:cno,HostPort:9104,ContainerPort:9104,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:RELEASE_VERSION,Value:4.18.1,ValueFrom:nil,},EnvVar{Name:KUBE_PROXY_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b97554198294bf544fbc116c94a0a1fb2ec8a4de0e926bf9d9e320135f0bee6f,ValueFrom:nil,},EnvVar{Name:KUBE_RBAC_PROXY_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09,ValueFrom:nil,},EnvVar{Name:MULTUS_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26,ValueFrom:nil,},EnvVar{Name:MULTUS_ADMISSION_CONTROLLER_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317,ValueFrom:nil,},EnvVar{Name:CNI_PLUGINS_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc,ValueFrom:nil,},EnvVar{Name:BOND_CNI_PLUGIN_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78,ValueFrom:nil,},EnvVar{Name:WHEREABOUTS_CNI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4,ValueFrom:nil,},EnvVar{Name:ROUTE_OVERRRIDE_CNI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa,ValueFrom:nil,},EnvVar{Name:MULTUS_NETWORKPOLICY_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:23f833d3738d68706eb2f2868bd76bd71cee016cffa6faf5f045a60cc8c6eddd,ValueFrom:nil,},EnvVar{Name:OVN_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2,ValueFrom:nil,},EnvVar{Name:OVN_NB_RAFT_ELECTION_TIMER,Value:10,ValueFrom:nil,},EnvVar{Name:OVN_SB_RAFT_ELECTION_TIMER,Value:16,ValueFrom:nil,},EnvVar{Name:OVN_NORTHD_PROBE_INTERVAL,Value:10000,ValueFrom:nil,},EnvVar{Name:OVN_CONTROLLER_INACTIVITY_PROBE,Value:180000,ValueFrom:nil,},EnvVar{Name:OVN_NB_INACTIVITY_PROBE,Value:60000,ValueFrom:nil,},EnvVar{Name:EGRESS_ROUTER_CNI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c,ValueFrom:nil,},EnvVar{Name:NETWORK_METRICS_DAEMON_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d,ValueFrom:nil,},EnvVar{Name:NETWORK_CHECK_SOURCE_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,ValueFrom:nil,},EnvVar{Name:NETWORK_CHECK_TARGET_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,ValueFrom:nil,},EnvVar{Name:NETWORK_OPERATOR_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,ValueFrom:nil,},EnvVar{Name:CLOUD_NETWORK_CONFIG_CONTROLLER_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8048f1cb0be521f09749c0a489503cd56d85b68c6ca93380e082cfd693cd97a8,ValueFrom:nil,},EnvVar{Name:CLI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2,ValueFrom:nil,},EnvVar{Name:FRR_K8S_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5dbf844e49bb46b78586930149e5e5f5dc121014c8afd10fe36f3651967cc256,ValueFrom:nil,},EnvVar{Name:NETWORKING_CONSOLE_PLUGIN_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd,ValueFrom:nil,},EnvVar{Name:POD_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.name,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:host-etc-kube,ReadOnly:true,MountPath:/etc/kubernetes,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:metrics-tls,ReadOnly:false,MountPath:/var/run/secrets/serving-cert,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rdwmf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod network-operator-58b4c7f79c-55gtf_openshift-network-operator(37a5e44f-9a88-4405-be8a-b645485e7312): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 13 14:06:33 crc kubenswrapper[4907]: > logger="UnhandledError" Mar 13 14:06:33 crc kubenswrapper[4907]: E0313 14:06:33.169182 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"network-operator\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" podUID="37a5e44f-9a88-4405-be8a-b645485e7312" Mar 13 14:06:33 crc kubenswrapper[4907]: E0313 14:06:33.169253 4907 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 13 14:06:33 crc kubenswrapper[4907]: container &Container{Name:approver,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2,Command:[/bin/bash -c set -xe Mar 13 14:06:33 crc kubenswrapper[4907]: if [[ -f "/env/_master" ]]; then Mar 13 14:06:33 crc kubenswrapper[4907]: set -o allexport Mar 13 14:06:33 crc kubenswrapper[4907]: source "/env/_master" Mar 13 14:06:33 crc kubenswrapper[4907]: set +o allexport Mar 13 14:06:33 crc kubenswrapper[4907]: fi Mar 13 14:06:33 crc kubenswrapper[4907]: Mar 13 14:06:33 crc kubenswrapper[4907]: echo "I$(date "+%m%d %H:%M:%S.%N") - network-node-identity - start approver" Mar 13 14:06:33 crc kubenswrapper[4907]: exec /usr/bin/ovnkube-identity --k8s-apiserver=https://api-int.crc.testing:6443 \ Mar 13 14:06:33 crc kubenswrapper[4907]: --disable-webhook \ Mar 13 14:06:33 crc kubenswrapper[4907]: --csr-acceptance-conditions="/var/run/ovnkube-identity-config/additional-cert-acceptance-cond.json" \ Mar 13 14:06:33 crc kubenswrapper[4907]: --loglevel="${LOGLEVEL}" Mar 13 14:06:33 crc kubenswrapper[4907]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOGLEVEL,Value:4,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:env-overrides,ReadOnly:false,MountPath:/env,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovnkube-identity-cm,ReadOnly:false,MountPath:/var/run/ovnkube-identity-config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-s2kz5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000470000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod network-node-identity-vrzqb_openshift-network-node-identity(ef543e1b-8068-4ea3-b32a-61027b32e95d): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 13 14:06:33 crc kubenswrapper[4907]: > logger="UnhandledError" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.170044 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.170075 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.170087 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.170103 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.170117 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:33Z","lastTransitionTime":"2026-03-13T14:06:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:33 crc kubenswrapper[4907]: E0313 14:06:33.170490 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"webhook\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\", failed to \"StartContainer\" for \"approver\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"]" pod="openshift-network-node-identity/network-node-identity-vrzqb" podUID="ef543e1b-8068-4ea3-b32a-61027b32e95d" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.175209 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.190147 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.202917 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.216466 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.227339 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.240253 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.254548 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.265138 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.272120 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.272245 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.272314 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.272421 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.272510 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:33Z","lastTransitionTime":"2026-03-13T14:06:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.282246 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.293629 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.303212 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.311868 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.374911 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.374975 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.375000 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.375029 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.375048 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:33Z","lastTransitionTime":"2026-03-13T14:06:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.443279 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:06:33 crc kubenswrapper[4907]: E0313 14:06:33.443494 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:06:34.443455948 +0000 UTC m=+93.343243667 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.478106 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.478158 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.478176 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.478196 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.478212 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:33Z","lastTransitionTime":"2026-03-13T14:06:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.544711 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.544791 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.544854 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.544924 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:06:33 crc kubenswrapper[4907]: E0313 14:06:33.544959 4907 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 13 14:06:33 crc kubenswrapper[4907]: E0313 14:06:33.545055 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-13 14:06:34.545026103 +0000 UTC m=+93.444813822 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 13 14:06:33 crc kubenswrapper[4907]: E0313 14:06:33.545118 4907 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 13 14:06:33 crc kubenswrapper[4907]: E0313 14:06:33.545144 4907 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 13 14:06:33 crc kubenswrapper[4907]: E0313 14:06:33.545163 4907 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 14:06:33 crc kubenswrapper[4907]: E0313 14:06:33.545239 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-13 14:06:34.545213198 +0000 UTC m=+93.445000917 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 14:06:33 crc kubenswrapper[4907]: E0313 14:06:33.545326 4907 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 13 14:06:33 crc kubenswrapper[4907]: E0313 14:06:33.545374 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-13 14:06:34.545361663 +0000 UTC m=+93.445149382 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 13 14:06:33 crc kubenswrapper[4907]: E0313 14:06:33.545451 4907 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 13 14:06:33 crc kubenswrapper[4907]: E0313 14:06:33.545471 4907 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 13 14:06:33 crc kubenswrapper[4907]: E0313 14:06:33.545486 4907 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 14:06:33 crc kubenswrapper[4907]: E0313 14:06:33.545528 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-13 14:06:34.545514568 +0000 UTC m=+93.445302287 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.580963 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.581037 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.581055 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.581087 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.581115 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:33Z","lastTransitionTime":"2026-03-13T14:06:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.684003 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.684072 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.684092 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.684119 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.684138 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:33Z","lastTransitionTime":"2026-03-13T14:06:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.787582 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.787631 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.787643 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.787659 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.787672 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:33Z","lastTransitionTime":"2026-03-13T14:06:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.788355 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.789592 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.791748 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.793042 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.794819 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.795405 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.796077 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.797575 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.799033 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.799845 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.800525 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.801913 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.802564 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.803741 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.804435 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.805578 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.806332 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.806824 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.808031 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.808762 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.809368 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.810633 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.811223 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.812494 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.813097 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.814189 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.814793 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.815254 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.816179 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.816642 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.817419 4907 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.817514 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.819077 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.820026 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.820394 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.821788 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.822995 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.823497 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.824495 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.825124 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.825543 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.826464 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.827377 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.828044 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.828910 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.829406 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.830272 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.830975 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.831735 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.832225 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.832659 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.833542 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.834078 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.834904 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.890265 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.890537 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.890805 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.891041 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.891185 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:33Z","lastTransitionTime":"2026-03-13T14:06:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.993908 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.993950 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.993963 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.993981 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.994026 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:33Z","lastTransitionTime":"2026-03-13T14:06:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:34 crc kubenswrapper[4907]: I0313 14:06:34.096008 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:34 crc kubenswrapper[4907]: I0313 14:06:34.096062 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:34 crc kubenswrapper[4907]: I0313 14:06:34.096079 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:34 crc kubenswrapper[4907]: I0313 14:06:34.096103 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:34 crc kubenswrapper[4907]: I0313 14:06:34.096120 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:34Z","lastTransitionTime":"2026-03-13T14:06:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:34 crc kubenswrapper[4907]: I0313 14:06:34.198661 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:34 crc kubenswrapper[4907]: I0313 14:06:34.198704 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:34 crc kubenswrapper[4907]: I0313 14:06:34.198714 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:34 crc kubenswrapper[4907]: I0313 14:06:34.198730 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:34 crc kubenswrapper[4907]: I0313 14:06:34.198740 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:34Z","lastTransitionTime":"2026-03-13T14:06:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:34 crc kubenswrapper[4907]: I0313 14:06:34.301279 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:34 crc kubenswrapper[4907]: I0313 14:06:34.301592 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:34 crc kubenswrapper[4907]: I0313 14:06:34.301840 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:34 crc kubenswrapper[4907]: I0313 14:06:34.302105 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:34 crc kubenswrapper[4907]: I0313 14:06:34.302311 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:34Z","lastTransitionTime":"2026-03-13T14:06:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:34 crc kubenswrapper[4907]: I0313 14:06:34.405694 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:34 crc kubenswrapper[4907]: I0313 14:06:34.405997 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:34 crc kubenswrapper[4907]: I0313 14:06:34.406081 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:34 crc kubenswrapper[4907]: I0313 14:06:34.406150 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:34 crc kubenswrapper[4907]: I0313 14:06:34.406221 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:34Z","lastTransitionTime":"2026-03-13T14:06:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:34 crc kubenswrapper[4907]: I0313 14:06:34.453362 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:06:34 crc kubenswrapper[4907]: E0313 14:06:34.453747 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:06:36.453711167 +0000 UTC m=+95.353498896 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:06:34 crc kubenswrapper[4907]: I0313 14:06:34.509865 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:34 crc kubenswrapper[4907]: I0313 14:06:34.509957 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:34 crc kubenswrapper[4907]: I0313 14:06:34.509975 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:34 crc kubenswrapper[4907]: I0313 14:06:34.510003 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:34 crc kubenswrapper[4907]: I0313 14:06:34.510021 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:34Z","lastTransitionTime":"2026-03-13T14:06:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:34 crc kubenswrapper[4907]: I0313 14:06:34.554979 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:06:34 crc kubenswrapper[4907]: I0313 14:06:34.555283 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:06:34 crc kubenswrapper[4907]: E0313 14:06:34.555277 4907 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 13 14:06:34 crc kubenswrapper[4907]: I0313 14:06:34.555455 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:06:34 crc kubenswrapper[4907]: E0313 14:06:34.555589 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-13 14:06:36.55555414 +0000 UTC m=+95.455341859 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 13 14:06:34 crc kubenswrapper[4907]: E0313 14:06:34.555395 4907 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 13 14:06:34 crc kubenswrapper[4907]: E0313 14:06:34.555679 4907 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 13 14:06:34 crc kubenswrapper[4907]: E0313 14:06:34.555742 4907 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 14:06:34 crc kubenswrapper[4907]: E0313 14:06:34.555851 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-13 14:06:36.555831238 +0000 UTC m=+95.455618957 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 14:06:34 crc kubenswrapper[4907]: I0313 14:06:34.555987 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:06:34 crc kubenswrapper[4907]: E0313 14:06:34.556134 4907 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 13 14:06:34 crc kubenswrapper[4907]: E0313 14:06:34.556258 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-13 14:06:36.55622818 +0000 UTC m=+95.456015909 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 13 14:06:34 crc kubenswrapper[4907]: E0313 14:06:34.556384 4907 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 13 14:06:34 crc kubenswrapper[4907]: E0313 14:06:34.556417 4907 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 13 14:06:34 crc kubenswrapper[4907]: E0313 14:06:34.556437 4907 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 14:06:34 crc kubenswrapper[4907]: E0313 14:06:34.556496 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-13 14:06:36.556471568 +0000 UTC m=+95.456259287 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 14:06:34 crc kubenswrapper[4907]: I0313 14:06:34.613508 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:34 crc kubenswrapper[4907]: I0313 14:06:34.613561 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:34 crc kubenswrapper[4907]: I0313 14:06:34.613580 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:34 crc kubenswrapper[4907]: I0313 14:06:34.613605 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:34 crc kubenswrapper[4907]: I0313 14:06:34.613622 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:34Z","lastTransitionTime":"2026-03-13T14:06:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:34 crc kubenswrapper[4907]: I0313 14:06:34.717708 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:34 crc kubenswrapper[4907]: I0313 14:06:34.718687 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:34 crc kubenswrapper[4907]: I0313 14:06:34.718756 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:34 crc kubenswrapper[4907]: I0313 14:06:34.718798 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:34 crc kubenswrapper[4907]: I0313 14:06:34.718824 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:34Z","lastTransitionTime":"2026-03-13T14:06:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:34 crc kubenswrapper[4907]: I0313 14:06:34.782391 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:06:34 crc kubenswrapper[4907]: I0313 14:06:34.782480 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:06:34 crc kubenswrapper[4907]: I0313 14:06:34.782391 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:06:34 crc kubenswrapper[4907]: E0313 14:06:34.782628 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:06:34 crc kubenswrapper[4907]: E0313 14:06:34.782796 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:06:34 crc kubenswrapper[4907]: E0313 14:06:34.783001 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:06:34 crc kubenswrapper[4907]: I0313 14:06:34.821462 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:34 crc kubenswrapper[4907]: I0313 14:06:34.821505 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:34 crc kubenswrapper[4907]: I0313 14:06:34.821513 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:34 crc kubenswrapper[4907]: I0313 14:06:34.821531 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:34 crc kubenswrapper[4907]: I0313 14:06:34.821544 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:34Z","lastTransitionTime":"2026-03-13T14:06:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:34 crc kubenswrapper[4907]: I0313 14:06:34.924266 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:34 crc kubenswrapper[4907]: I0313 14:06:34.924333 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:34 crc kubenswrapper[4907]: I0313 14:06:34.924353 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:34 crc kubenswrapper[4907]: I0313 14:06:34.924386 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:34 crc kubenswrapper[4907]: I0313 14:06:34.924408 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:34Z","lastTransitionTime":"2026-03-13T14:06:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:35 crc kubenswrapper[4907]: I0313 14:06:35.027438 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:35 crc kubenswrapper[4907]: I0313 14:06:35.027495 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:35 crc kubenswrapper[4907]: I0313 14:06:35.027513 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:35 crc kubenswrapper[4907]: I0313 14:06:35.027540 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:35 crc kubenswrapper[4907]: I0313 14:06:35.027558 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:35Z","lastTransitionTime":"2026-03-13T14:06:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:35 crc kubenswrapper[4907]: I0313 14:06:35.130529 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:35 crc kubenswrapper[4907]: I0313 14:06:35.130567 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:35 crc kubenswrapper[4907]: I0313 14:06:35.130575 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:35 crc kubenswrapper[4907]: I0313 14:06:35.130590 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:35 crc kubenswrapper[4907]: I0313 14:06:35.130601 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:35Z","lastTransitionTime":"2026-03-13T14:06:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:35 crc kubenswrapper[4907]: I0313 14:06:35.232673 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:35 crc kubenswrapper[4907]: I0313 14:06:35.232711 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:35 crc kubenswrapper[4907]: I0313 14:06:35.232731 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:35 crc kubenswrapper[4907]: I0313 14:06:35.232747 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:35 crc kubenswrapper[4907]: I0313 14:06:35.232757 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:35Z","lastTransitionTime":"2026-03-13T14:06:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:35 crc kubenswrapper[4907]: I0313 14:06:35.336036 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:35 crc kubenswrapper[4907]: I0313 14:06:35.336073 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:35 crc kubenswrapper[4907]: I0313 14:06:35.336085 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:35 crc kubenswrapper[4907]: I0313 14:06:35.336102 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:35 crc kubenswrapper[4907]: I0313 14:06:35.336112 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:35Z","lastTransitionTime":"2026-03-13T14:06:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:35 crc kubenswrapper[4907]: I0313 14:06:35.439123 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:35 crc kubenswrapper[4907]: I0313 14:06:35.439208 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:35 crc kubenswrapper[4907]: I0313 14:06:35.439226 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:35 crc kubenswrapper[4907]: I0313 14:06:35.439271 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:35 crc kubenswrapper[4907]: I0313 14:06:35.439311 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:35Z","lastTransitionTime":"2026-03-13T14:06:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:35 crc kubenswrapper[4907]: I0313 14:06:35.544061 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:35 crc kubenswrapper[4907]: I0313 14:06:35.544119 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:35 crc kubenswrapper[4907]: I0313 14:06:35.544139 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:35 crc kubenswrapper[4907]: I0313 14:06:35.544164 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:35 crc kubenswrapper[4907]: I0313 14:06:35.544182 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:35Z","lastTransitionTime":"2026-03-13T14:06:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:35 crc kubenswrapper[4907]: I0313 14:06:35.647575 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:35 crc kubenswrapper[4907]: I0313 14:06:35.647704 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:35 crc kubenswrapper[4907]: I0313 14:06:35.647725 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:35 crc kubenswrapper[4907]: I0313 14:06:35.647757 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:35 crc kubenswrapper[4907]: I0313 14:06:35.647774 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:35Z","lastTransitionTime":"2026-03-13T14:06:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:35 crc kubenswrapper[4907]: I0313 14:06:35.751696 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:35 crc kubenswrapper[4907]: I0313 14:06:35.751794 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:35 crc kubenswrapper[4907]: I0313 14:06:35.751815 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:35 crc kubenswrapper[4907]: I0313 14:06:35.751848 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:35 crc kubenswrapper[4907]: I0313 14:06:35.751871 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:35Z","lastTransitionTime":"2026-03-13T14:06:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:35 crc kubenswrapper[4907]: I0313 14:06:35.855667 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:35 crc kubenswrapper[4907]: I0313 14:06:35.855712 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:35 crc kubenswrapper[4907]: I0313 14:06:35.855722 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:35 crc kubenswrapper[4907]: I0313 14:06:35.855743 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:35 crc kubenswrapper[4907]: I0313 14:06:35.855755 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:35Z","lastTransitionTime":"2026-03-13T14:06:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:35 crc kubenswrapper[4907]: I0313 14:06:35.958723 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:35 crc kubenswrapper[4907]: I0313 14:06:35.958788 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:35 crc kubenswrapper[4907]: I0313 14:06:35.958805 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:35 crc kubenswrapper[4907]: I0313 14:06:35.958833 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:35 crc kubenswrapper[4907]: I0313 14:06:35.958855 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:35Z","lastTransitionTime":"2026-03-13T14:06:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.010271 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.010343 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.010359 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.010380 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.010392 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:36Z","lastTransitionTime":"2026-03-13T14:06:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:36 crc kubenswrapper[4907]: E0313 14:06:36.020190 4907 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4b182bae-7f2d-4b54-934c-85a9a277d132\\\",\\\"systemUUID\\\":\\\"05bfbfa3-f04f-45ba-8233-a355fc774c24\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.023440 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.023479 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.023491 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.023508 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.023519 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:36Z","lastTransitionTime":"2026-03-13T14:06:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:36 crc kubenswrapper[4907]: E0313 14:06:36.037836 4907 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4b182bae-7f2d-4b54-934c-85a9a277d132\\\",\\\"systemUUID\\\":\\\"05bfbfa3-f04f-45ba-8233-a355fc774c24\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.042318 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.042458 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.042541 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.042631 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.042838 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:36Z","lastTransitionTime":"2026-03-13T14:06:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:36 crc kubenswrapper[4907]: E0313 14:06:36.058143 4907 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4b182bae-7f2d-4b54-934c-85a9a277d132\\\",\\\"systemUUID\\\":\\\"05bfbfa3-f04f-45ba-8233-a355fc774c24\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.062597 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.062632 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.062644 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.062664 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.062677 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:36Z","lastTransitionTime":"2026-03-13T14:06:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:36 crc kubenswrapper[4907]: E0313 14:06:36.078289 4907 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4b182bae-7f2d-4b54-934c-85a9a277d132\\\",\\\"systemUUID\\\":\\\"05bfbfa3-f04f-45ba-8233-a355fc774c24\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.082863 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.082958 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.082976 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.083368 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.083421 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:36Z","lastTransitionTime":"2026-03-13T14:06:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:36 crc kubenswrapper[4907]: E0313 14:06:36.099053 4907 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4b182bae-7f2d-4b54-934c-85a9a277d132\\\",\\\"systemUUID\\\":\\\"05bfbfa3-f04f-45ba-8233-a355fc774c24\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 14:06:36 crc kubenswrapper[4907]: E0313 14:06:36.099303 4907 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.103216 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.103303 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.103357 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.103446 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.103507 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:36Z","lastTransitionTime":"2026-03-13T14:06:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.205935 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.206230 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.206348 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.206510 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.206628 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:36Z","lastTransitionTime":"2026-03-13T14:06:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.310227 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.310265 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.310273 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.310289 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.310322 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:36Z","lastTransitionTime":"2026-03-13T14:06:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.413273 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.413334 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.413352 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.413379 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.413398 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:36Z","lastTransitionTime":"2026-03-13T14:06:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.476281 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:06:36 crc kubenswrapper[4907]: E0313 14:06:36.476485 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:06:40.476461712 +0000 UTC m=+99.376249401 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.516601 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.516674 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.516693 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.516720 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.516742 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:36Z","lastTransitionTime":"2026-03-13T14:06:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.576991 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.577049 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.577086 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.577117 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:06:36 crc kubenswrapper[4907]: E0313 14:06:36.577226 4907 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 13 14:06:36 crc kubenswrapper[4907]: E0313 14:06:36.577243 4907 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 13 14:06:36 crc kubenswrapper[4907]: E0313 14:06:36.577262 4907 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 13 14:06:36 crc kubenswrapper[4907]: E0313 14:06:36.577276 4907 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 14:06:36 crc kubenswrapper[4907]: E0313 14:06:36.577283 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-13 14:06:40.577263753 +0000 UTC m=+99.477051442 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 13 14:06:36 crc kubenswrapper[4907]: E0313 14:06:36.577281 4907 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 13 14:06:36 crc kubenswrapper[4907]: E0313 14:06:36.577307 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-13 14:06:40.577297684 +0000 UTC m=+99.477085373 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 14:06:36 crc kubenswrapper[4907]: E0313 14:06:36.577324 4907 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 13 14:06:36 crc kubenswrapper[4907]: E0313 14:06:36.577350 4907 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 14:06:36 crc kubenswrapper[4907]: E0313 14:06:36.577352 4907 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 13 14:06:36 crc kubenswrapper[4907]: E0313 14:06:36.577385 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-13 14:06:40.577375608 +0000 UTC m=+99.477163297 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 13 14:06:36 crc kubenswrapper[4907]: E0313 14:06:36.577421 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-13 14:06:40.577394898 +0000 UTC m=+99.477182617 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.620179 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.620234 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.620250 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.620273 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.620286 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:36Z","lastTransitionTime":"2026-03-13T14:06:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.723077 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.723111 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.723121 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.723141 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.723153 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:36Z","lastTransitionTime":"2026-03-13T14:06:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.781848 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.781929 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:06:36 crc kubenswrapper[4907]: E0313 14:06:36.782074 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.782089 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:06:36 crc kubenswrapper[4907]: E0313 14:06:36.782250 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:06:36 crc kubenswrapper[4907]: E0313 14:06:36.782349 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.826000 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.826059 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.826076 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.826101 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.826120 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:36Z","lastTransitionTime":"2026-03-13T14:06:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.929321 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.929393 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.929411 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.929435 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.929452 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:36Z","lastTransitionTime":"2026-03-13T14:06:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:37 crc kubenswrapper[4907]: I0313 14:06:37.032149 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:37 crc kubenswrapper[4907]: I0313 14:06:37.032227 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:37 crc kubenswrapper[4907]: I0313 14:06:37.032256 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:37 crc kubenswrapper[4907]: I0313 14:06:37.032281 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:37 crc kubenswrapper[4907]: I0313 14:06:37.032297 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:37Z","lastTransitionTime":"2026-03-13T14:06:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:37 crc kubenswrapper[4907]: I0313 14:06:37.136416 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:37 crc kubenswrapper[4907]: I0313 14:06:37.136478 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:37 crc kubenswrapper[4907]: I0313 14:06:37.136495 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:37 crc kubenswrapper[4907]: I0313 14:06:37.136520 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:37 crc kubenswrapper[4907]: I0313 14:06:37.136538 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:37Z","lastTransitionTime":"2026-03-13T14:06:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:37 crc kubenswrapper[4907]: I0313 14:06:37.239975 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:37 crc kubenswrapper[4907]: I0313 14:06:37.240037 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:37 crc kubenswrapper[4907]: I0313 14:06:37.240053 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:37 crc kubenswrapper[4907]: I0313 14:06:37.240077 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:37 crc kubenswrapper[4907]: I0313 14:06:37.240094 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:37Z","lastTransitionTime":"2026-03-13T14:06:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:37 crc kubenswrapper[4907]: I0313 14:06:37.343165 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:37 crc kubenswrapper[4907]: I0313 14:06:37.343259 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:37 crc kubenswrapper[4907]: I0313 14:06:37.343619 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:37 crc kubenswrapper[4907]: I0313 14:06:37.343669 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:37 crc kubenswrapper[4907]: I0313 14:06:37.343697 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:37Z","lastTransitionTime":"2026-03-13T14:06:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:37 crc kubenswrapper[4907]: I0313 14:06:37.447277 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:37 crc kubenswrapper[4907]: I0313 14:06:37.447337 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:37 crc kubenswrapper[4907]: I0313 14:06:37.447354 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:37 crc kubenswrapper[4907]: I0313 14:06:37.447379 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:37 crc kubenswrapper[4907]: I0313 14:06:37.447398 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:37Z","lastTransitionTime":"2026-03-13T14:06:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:37 crc kubenswrapper[4907]: I0313 14:06:37.550553 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:37 crc kubenswrapper[4907]: I0313 14:06:37.550602 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:37 crc kubenswrapper[4907]: I0313 14:06:37.550612 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:37 crc kubenswrapper[4907]: I0313 14:06:37.550631 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:37 crc kubenswrapper[4907]: I0313 14:06:37.550645 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:37Z","lastTransitionTime":"2026-03-13T14:06:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:37 crc kubenswrapper[4907]: I0313 14:06:37.658622 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:37 crc kubenswrapper[4907]: I0313 14:06:37.658978 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:37 crc kubenswrapper[4907]: I0313 14:06:37.659064 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:37 crc kubenswrapper[4907]: I0313 14:06:37.659148 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:37 crc kubenswrapper[4907]: I0313 14:06:37.659223 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:37Z","lastTransitionTime":"2026-03-13T14:06:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:37 crc kubenswrapper[4907]: I0313 14:06:37.762169 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:37 crc kubenswrapper[4907]: I0313 14:06:37.762232 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:37 crc kubenswrapper[4907]: I0313 14:06:37.762250 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:37 crc kubenswrapper[4907]: I0313 14:06:37.762276 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:37 crc kubenswrapper[4907]: I0313 14:06:37.762294 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:37Z","lastTransitionTime":"2026-03-13T14:06:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:37 crc kubenswrapper[4907]: I0313 14:06:37.864532 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:37 crc kubenswrapper[4907]: I0313 14:06:37.865212 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:37 crc kubenswrapper[4907]: I0313 14:06:37.865331 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:37 crc kubenswrapper[4907]: I0313 14:06:37.865479 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:37 crc kubenswrapper[4907]: I0313 14:06:37.865596 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:37Z","lastTransitionTime":"2026-03-13T14:06:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:37 crc kubenswrapper[4907]: I0313 14:06:37.967301 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:37 crc kubenswrapper[4907]: I0313 14:06:37.967336 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:37 crc kubenswrapper[4907]: I0313 14:06:37.967345 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:37 crc kubenswrapper[4907]: I0313 14:06:37.967363 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:37 crc kubenswrapper[4907]: I0313 14:06:37.967374 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:37Z","lastTransitionTime":"2026-03-13T14:06:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:38 crc kubenswrapper[4907]: I0313 14:06:38.069322 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:38 crc kubenswrapper[4907]: I0313 14:06:38.069384 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:38 crc kubenswrapper[4907]: I0313 14:06:38.069405 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:38 crc kubenswrapper[4907]: I0313 14:06:38.069438 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:38 crc kubenswrapper[4907]: I0313 14:06:38.069471 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:38Z","lastTransitionTime":"2026-03-13T14:06:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:38 crc kubenswrapper[4907]: I0313 14:06:38.171405 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:38 crc kubenswrapper[4907]: I0313 14:06:38.171439 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:38 crc kubenswrapper[4907]: I0313 14:06:38.171450 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:38 crc kubenswrapper[4907]: I0313 14:06:38.171466 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:38 crc kubenswrapper[4907]: I0313 14:06:38.171478 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:38Z","lastTransitionTime":"2026-03-13T14:06:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:38 crc kubenswrapper[4907]: I0313 14:06:38.273271 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:38 crc kubenswrapper[4907]: I0313 14:06:38.273301 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:38 crc kubenswrapper[4907]: I0313 14:06:38.273310 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:38 crc kubenswrapper[4907]: I0313 14:06:38.273322 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:38 crc kubenswrapper[4907]: I0313 14:06:38.273332 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:38Z","lastTransitionTime":"2026-03-13T14:06:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:38 crc kubenswrapper[4907]: I0313 14:06:38.376486 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:38 crc kubenswrapper[4907]: I0313 14:06:38.376857 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:38 crc kubenswrapper[4907]: I0313 14:06:38.377084 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:38 crc kubenswrapper[4907]: I0313 14:06:38.377243 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:38 crc kubenswrapper[4907]: I0313 14:06:38.377372 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:38Z","lastTransitionTime":"2026-03-13T14:06:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:38 crc kubenswrapper[4907]: I0313 14:06:38.479931 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:38 crc kubenswrapper[4907]: I0313 14:06:38.479974 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:38 crc kubenswrapper[4907]: I0313 14:06:38.479989 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:38 crc kubenswrapper[4907]: I0313 14:06:38.480011 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:38 crc kubenswrapper[4907]: I0313 14:06:38.480029 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:38Z","lastTransitionTime":"2026-03-13T14:06:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:38 crc kubenswrapper[4907]: I0313 14:06:38.583139 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:38 crc kubenswrapper[4907]: I0313 14:06:38.583188 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:38 crc kubenswrapper[4907]: I0313 14:06:38.583200 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:38 crc kubenswrapper[4907]: I0313 14:06:38.583222 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:38 crc kubenswrapper[4907]: I0313 14:06:38.583234 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:38Z","lastTransitionTime":"2026-03-13T14:06:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:38 crc kubenswrapper[4907]: I0313 14:06:38.686716 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:38 crc kubenswrapper[4907]: I0313 14:06:38.686770 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:38 crc kubenswrapper[4907]: I0313 14:06:38.686784 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:38 crc kubenswrapper[4907]: I0313 14:06:38.686805 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:38 crc kubenswrapper[4907]: I0313 14:06:38.686823 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:38Z","lastTransitionTime":"2026-03-13T14:06:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:38 crc kubenswrapper[4907]: I0313 14:06:38.781343 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:06:38 crc kubenswrapper[4907]: I0313 14:06:38.781456 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:06:38 crc kubenswrapper[4907]: I0313 14:06:38.781491 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:06:38 crc kubenswrapper[4907]: E0313 14:06:38.781647 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:06:38 crc kubenswrapper[4907]: E0313 14:06:38.781812 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:06:38 crc kubenswrapper[4907]: E0313 14:06:38.781984 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:06:38 crc kubenswrapper[4907]: I0313 14:06:38.789803 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:38 crc kubenswrapper[4907]: I0313 14:06:38.789871 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:38 crc kubenswrapper[4907]: I0313 14:06:38.789938 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:38 crc kubenswrapper[4907]: I0313 14:06:38.789975 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:38 crc kubenswrapper[4907]: I0313 14:06:38.790001 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:38Z","lastTransitionTime":"2026-03-13T14:06:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:38 crc kubenswrapper[4907]: I0313 14:06:38.893748 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:38 crc kubenswrapper[4907]: I0313 14:06:38.893785 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:38 crc kubenswrapper[4907]: I0313 14:06:38.893794 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:38 crc kubenswrapper[4907]: I0313 14:06:38.893808 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:38 crc kubenswrapper[4907]: I0313 14:06:38.893818 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:38Z","lastTransitionTime":"2026-03-13T14:06:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:38 crc kubenswrapper[4907]: I0313 14:06:38.996811 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:38 crc kubenswrapper[4907]: I0313 14:06:38.996919 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:38 crc kubenswrapper[4907]: I0313 14:06:38.996934 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:38 crc kubenswrapper[4907]: I0313 14:06:38.996955 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:38 crc kubenswrapper[4907]: I0313 14:06:38.996969 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:38Z","lastTransitionTime":"2026-03-13T14:06:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:39 crc kubenswrapper[4907]: I0313 14:06:39.099961 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:39 crc kubenswrapper[4907]: I0313 14:06:39.100037 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:39 crc kubenswrapper[4907]: I0313 14:06:39.100056 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:39 crc kubenswrapper[4907]: I0313 14:06:39.100084 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:39 crc kubenswrapper[4907]: I0313 14:06:39.100104 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:39Z","lastTransitionTime":"2026-03-13T14:06:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:39 crc kubenswrapper[4907]: I0313 14:06:39.202411 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:39 crc kubenswrapper[4907]: I0313 14:06:39.202492 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:39 crc kubenswrapper[4907]: I0313 14:06:39.202515 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:39 crc kubenswrapper[4907]: I0313 14:06:39.202547 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:39 crc kubenswrapper[4907]: I0313 14:06:39.202570 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:39Z","lastTransitionTime":"2026-03-13T14:06:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:39 crc kubenswrapper[4907]: I0313 14:06:39.306156 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:39 crc kubenswrapper[4907]: I0313 14:06:39.306241 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:39 crc kubenswrapper[4907]: I0313 14:06:39.306265 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:39 crc kubenswrapper[4907]: I0313 14:06:39.306296 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:39 crc kubenswrapper[4907]: I0313 14:06:39.306319 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:39Z","lastTransitionTime":"2026-03-13T14:06:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:39 crc kubenswrapper[4907]: I0313 14:06:39.409048 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:39 crc kubenswrapper[4907]: I0313 14:06:39.409097 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:39 crc kubenswrapper[4907]: I0313 14:06:39.409109 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:39 crc kubenswrapper[4907]: I0313 14:06:39.409131 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:39 crc kubenswrapper[4907]: I0313 14:06:39.409146 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:39Z","lastTransitionTime":"2026-03-13T14:06:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:39 crc kubenswrapper[4907]: I0313 14:06:39.512481 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:39 crc kubenswrapper[4907]: I0313 14:06:39.512545 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:39 crc kubenswrapper[4907]: I0313 14:06:39.512564 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:39 crc kubenswrapper[4907]: I0313 14:06:39.512593 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:39 crc kubenswrapper[4907]: I0313 14:06:39.512632 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:39Z","lastTransitionTime":"2026-03-13T14:06:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:39 crc kubenswrapper[4907]: I0313 14:06:39.615822 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:39 crc kubenswrapper[4907]: I0313 14:06:39.615916 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:39 crc kubenswrapper[4907]: I0313 14:06:39.615980 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:39 crc kubenswrapper[4907]: I0313 14:06:39.616007 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:39 crc kubenswrapper[4907]: I0313 14:06:39.616027 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:39Z","lastTransitionTime":"2026-03-13T14:06:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:39 crc kubenswrapper[4907]: I0313 14:06:39.718930 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:39 crc kubenswrapper[4907]: I0313 14:06:39.718977 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:39 crc kubenswrapper[4907]: I0313 14:06:39.718993 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:39 crc kubenswrapper[4907]: I0313 14:06:39.719018 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:39 crc kubenswrapper[4907]: I0313 14:06:39.719038 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:39Z","lastTransitionTime":"2026-03-13T14:06:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:39 crc kubenswrapper[4907]: I0313 14:06:39.821512 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:39 crc kubenswrapper[4907]: I0313 14:06:39.821568 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:39 crc kubenswrapper[4907]: I0313 14:06:39.821584 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:39 crc kubenswrapper[4907]: I0313 14:06:39.821609 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:39 crc kubenswrapper[4907]: I0313 14:06:39.821628 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:39Z","lastTransitionTime":"2026-03-13T14:06:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:39 crc kubenswrapper[4907]: I0313 14:06:39.924734 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:39 crc kubenswrapper[4907]: I0313 14:06:39.924820 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:39 crc kubenswrapper[4907]: I0313 14:06:39.924852 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:39 crc kubenswrapper[4907]: I0313 14:06:39.924918 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:39 crc kubenswrapper[4907]: I0313 14:06:39.924936 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:39Z","lastTransitionTime":"2026-03-13T14:06:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:40 crc kubenswrapper[4907]: I0313 14:06:40.027268 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:40 crc kubenswrapper[4907]: I0313 14:06:40.027326 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:40 crc kubenswrapper[4907]: I0313 14:06:40.027345 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:40 crc kubenswrapper[4907]: I0313 14:06:40.027369 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:40 crc kubenswrapper[4907]: I0313 14:06:40.027385 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:40Z","lastTransitionTime":"2026-03-13T14:06:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:40 crc kubenswrapper[4907]: I0313 14:06:40.130438 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:40 crc kubenswrapper[4907]: I0313 14:06:40.130487 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:40 crc kubenswrapper[4907]: I0313 14:06:40.130498 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:40 crc kubenswrapper[4907]: I0313 14:06:40.130516 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:40 crc kubenswrapper[4907]: I0313 14:06:40.130528 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:40Z","lastTransitionTime":"2026-03-13T14:06:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:40 crc kubenswrapper[4907]: I0313 14:06:40.233396 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:40 crc kubenswrapper[4907]: I0313 14:06:40.233540 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:40 crc kubenswrapper[4907]: I0313 14:06:40.234001 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:40 crc kubenswrapper[4907]: I0313 14:06:40.234081 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:40 crc kubenswrapper[4907]: I0313 14:06:40.234110 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:40Z","lastTransitionTime":"2026-03-13T14:06:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:40 crc kubenswrapper[4907]: I0313 14:06:40.337178 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:40 crc kubenswrapper[4907]: I0313 14:06:40.337270 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:40 crc kubenswrapper[4907]: I0313 14:06:40.337304 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:40 crc kubenswrapper[4907]: I0313 14:06:40.337337 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:40 crc kubenswrapper[4907]: I0313 14:06:40.337520 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:40Z","lastTransitionTime":"2026-03-13T14:06:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:40 crc kubenswrapper[4907]: I0313 14:06:40.440564 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:40 crc kubenswrapper[4907]: I0313 14:06:40.440627 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:40 crc kubenswrapper[4907]: I0313 14:06:40.440644 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:40 crc kubenswrapper[4907]: I0313 14:06:40.440668 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:40 crc kubenswrapper[4907]: I0313 14:06:40.440685 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:40Z","lastTransitionTime":"2026-03-13T14:06:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:40 crc kubenswrapper[4907]: I0313 14:06:40.516271 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:06:40 crc kubenswrapper[4907]: E0313 14:06:40.516567 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:06:48.516527977 +0000 UTC m=+107.416315706 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:06:40 crc kubenswrapper[4907]: I0313 14:06:40.544296 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:40 crc kubenswrapper[4907]: I0313 14:06:40.544355 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:40 crc kubenswrapper[4907]: I0313 14:06:40.544371 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:40 crc kubenswrapper[4907]: I0313 14:06:40.544402 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:40 crc kubenswrapper[4907]: I0313 14:06:40.544474 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:40Z","lastTransitionTime":"2026-03-13T14:06:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:40 crc kubenswrapper[4907]: I0313 14:06:40.617555 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:06:40 crc kubenswrapper[4907]: I0313 14:06:40.617618 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:06:40 crc kubenswrapper[4907]: I0313 14:06:40.617660 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:06:40 crc kubenswrapper[4907]: I0313 14:06:40.617705 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:06:40 crc kubenswrapper[4907]: E0313 14:06:40.617758 4907 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 13 14:06:40 crc kubenswrapper[4907]: E0313 14:06:40.618014 4907 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 13 14:06:40 crc kubenswrapper[4907]: E0313 14:06:40.618054 4907 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 13 14:06:40 crc kubenswrapper[4907]: E0313 14:06:40.618232 4907 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 14:06:40 crc kubenswrapper[4907]: E0313 14:06:40.618101 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-13 14:06:48.618078581 +0000 UTC m=+107.517866300 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 13 14:06:40 crc kubenswrapper[4907]: E0313 14:06:40.618457 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-13 14:06:48.618436433 +0000 UTC m=+107.518224152 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 14:06:40 crc kubenswrapper[4907]: E0313 14:06:40.617964 4907 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 13 14:06:40 crc kubenswrapper[4907]: E0313 14:06:40.618551 4907 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 13 14:06:40 crc kubenswrapper[4907]: E0313 14:06:40.618578 4907 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 13 14:06:40 crc kubenswrapper[4907]: E0313 14:06:40.618595 4907 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 14:06:40 crc kubenswrapper[4907]: E0313 14:06:40.618679 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-13 14:06:48.618639509 +0000 UTC m=+107.518427228 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 14:06:40 crc kubenswrapper[4907]: E0313 14:06:40.618707 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-13 14:06:48.61869281 +0000 UTC m=+107.518480539 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 13 14:06:40 crc kubenswrapper[4907]: I0313 14:06:40.648432 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:40 crc kubenswrapper[4907]: I0313 14:06:40.648472 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:40 crc kubenswrapper[4907]: I0313 14:06:40.648496 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:40 crc kubenswrapper[4907]: I0313 14:06:40.648514 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:40 crc kubenswrapper[4907]: I0313 14:06:40.648526 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:40Z","lastTransitionTime":"2026-03-13T14:06:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:40 crc kubenswrapper[4907]: I0313 14:06:40.750868 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:40 crc kubenswrapper[4907]: I0313 14:06:40.750989 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:40 crc kubenswrapper[4907]: I0313 14:06:40.751015 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:40 crc kubenswrapper[4907]: I0313 14:06:40.751046 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:40 crc kubenswrapper[4907]: I0313 14:06:40.751069 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:40Z","lastTransitionTime":"2026-03-13T14:06:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:40 crc kubenswrapper[4907]: I0313 14:06:40.782155 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:06:40 crc kubenswrapper[4907]: E0313 14:06:40.782319 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:06:40 crc kubenswrapper[4907]: I0313 14:06:40.782414 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:06:40 crc kubenswrapper[4907]: I0313 14:06:40.782556 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:06:40 crc kubenswrapper[4907]: E0313 14:06:40.783980 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:06:40 crc kubenswrapper[4907]: E0313 14:06:40.782868 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:06:40 crc kubenswrapper[4907]: I0313 14:06:40.800966 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Mar 13 14:06:40 crc kubenswrapper[4907]: I0313 14:06:40.801235 4907 scope.go:117] "RemoveContainer" containerID="1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80" Mar 13 14:06:40 crc kubenswrapper[4907]: E0313 14:06:40.801556 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 13 14:06:40 crc kubenswrapper[4907]: I0313 14:06:40.855688 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:40 crc kubenswrapper[4907]: I0313 14:06:40.855733 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:40 crc kubenswrapper[4907]: I0313 14:06:40.855748 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:40 crc kubenswrapper[4907]: I0313 14:06:40.855770 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:40 crc kubenswrapper[4907]: I0313 14:06:40.855784 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:40Z","lastTransitionTime":"2026-03-13T14:06:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:40 crc kubenswrapper[4907]: I0313 14:06:40.958610 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:40 crc kubenswrapper[4907]: I0313 14:06:40.958668 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:40 crc kubenswrapper[4907]: I0313 14:06:40.958679 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:40 crc kubenswrapper[4907]: I0313 14:06:40.958701 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:40 crc kubenswrapper[4907]: I0313 14:06:40.958714 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:40Z","lastTransitionTime":"2026-03-13T14:06:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:41 crc kubenswrapper[4907]: I0313 14:06:41.061317 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:41 crc kubenswrapper[4907]: I0313 14:06:41.061347 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:41 crc kubenswrapper[4907]: I0313 14:06:41.061356 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:41 crc kubenswrapper[4907]: I0313 14:06:41.061371 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:41 crc kubenswrapper[4907]: I0313 14:06:41.061380 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:41Z","lastTransitionTime":"2026-03-13T14:06:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:41 crc kubenswrapper[4907]: I0313 14:06:41.147441 4907 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Mar 13 14:06:41 crc kubenswrapper[4907]: I0313 14:06:41.164526 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:41 crc kubenswrapper[4907]: I0313 14:06:41.164616 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:41 crc kubenswrapper[4907]: I0313 14:06:41.164629 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:41 crc kubenswrapper[4907]: I0313 14:06:41.164647 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:41 crc kubenswrapper[4907]: I0313 14:06:41.164661 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:41Z","lastTransitionTime":"2026-03-13T14:06:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:41 crc kubenswrapper[4907]: I0313 14:06:41.187182 4907 scope.go:117] "RemoveContainer" containerID="1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80" Mar 13 14:06:41 crc kubenswrapper[4907]: E0313 14:06:41.187750 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 13 14:06:41 crc kubenswrapper[4907]: I0313 14:06:41.267187 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:41 crc kubenswrapper[4907]: I0313 14:06:41.267262 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:41 crc kubenswrapper[4907]: I0313 14:06:41.267279 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:41 crc kubenswrapper[4907]: I0313 14:06:41.267309 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:41 crc kubenswrapper[4907]: I0313 14:06:41.267342 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:41Z","lastTransitionTime":"2026-03-13T14:06:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:41 crc kubenswrapper[4907]: I0313 14:06:41.370336 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:41 crc kubenswrapper[4907]: I0313 14:06:41.370399 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:41 crc kubenswrapper[4907]: I0313 14:06:41.370408 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:41 crc kubenswrapper[4907]: I0313 14:06:41.370425 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:41 crc kubenswrapper[4907]: I0313 14:06:41.370435 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:41Z","lastTransitionTime":"2026-03-13T14:06:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:41 crc kubenswrapper[4907]: I0313 14:06:41.473713 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:41 crc kubenswrapper[4907]: I0313 14:06:41.473776 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:41 crc kubenswrapper[4907]: I0313 14:06:41.473791 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:41 crc kubenswrapper[4907]: I0313 14:06:41.473820 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:41 crc kubenswrapper[4907]: I0313 14:06:41.473835 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:41Z","lastTransitionTime":"2026-03-13T14:06:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:41 crc kubenswrapper[4907]: I0313 14:06:41.576810 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:41 crc kubenswrapper[4907]: I0313 14:06:41.576914 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:41 crc kubenswrapper[4907]: I0313 14:06:41.576938 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:41 crc kubenswrapper[4907]: I0313 14:06:41.576981 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:41 crc kubenswrapper[4907]: I0313 14:06:41.577003 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:41Z","lastTransitionTime":"2026-03-13T14:06:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:41 crc kubenswrapper[4907]: I0313 14:06:41.680647 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:41 crc kubenswrapper[4907]: I0313 14:06:41.680751 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:41 crc kubenswrapper[4907]: I0313 14:06:41.680771 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:41 crc kubenswrapper[4907]: I0313 14:06:41.680801 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:41 crc kubenswrapper[4907]: I0313 14:06:41.680818 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:41Z","lastTransitionTime":"2026-03-13T14:06:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:41 crc kubenswrapper[4907]: I0313 14:06:41.783199 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:41 crc kubenswrapper[4907]: I0313 14:06:41.783233 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:41 crc kubenswrapper[4907]: I0313 14:06:41.783241 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:41 crc kubenswrapper[4907]: I0313 14:06:41.783254 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:41 crc kubenswrapper[4907]: I0313 14:06:41.783265 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:41Z","lastTransitionTime":"2026-03-13T14:06:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:41 crc kubenswrapper[4907]: I0313 14:06:41.799455 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 14:06:41 crc kubenswrapper[4907]: I0313 14:06:41.808448 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 14:06:41 crc kubenswrapper[4907]: I0313 14:06:41.823976 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 14:06:41 crc kubenswrapper[4907]: I0313 14:06:41.835202 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 14:06:41 crc kubenswrapper[4907]: I0313 14:06:41.847060 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 14:06:41 crc kubenswrapper[4907]: I0313 14:06:41.862797 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 14:06:41 crc kubenswrapper[4907]: I0313 14:06:41.880529 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5d2b054-0ed8-419c-b731-f14d3598372f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4b503aa3b13268b3688d067f3e456c86afac1fda171aced0990899c3eb89b79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be40a1a33fa482eab363537afaa901127d560afbc1b5b287ef11e66391a5be4c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d1b539bbd918fc50428fc18cabcc4c8a43cc96f220b5692d67cf1c8dfbf8ed8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T14:06:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0313 14:06:13.202871 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0313 14:06:13.203175 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0313 14:06:13.204427 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2471728650/tls.crt::/tmp/serving-cert-2471728650/tls.key\\\\\\\"\\\\nI0313 14:06:13.581160 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0313 14:06:13.584012 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0313 14:06:13.584036 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0313 14:06:13.584064 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0313 14:06:13.584072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0313 14:06:13.594504 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0313 14:06:13.594891 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0313 14:06:13.595323 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0313 14:06:13.595363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0313 14:06:13.595370 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0313 14:06:13.595376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0313 14:06:13.597226 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3f2b6756078925f8562b6053b3b83798efb72288f5fd18fff680a4ab0d33feb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 14:06:41 crc kubenswrapper[4907]: I0313 14:06:41.885477 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:41 crc kubenswrapper[4907]: I0313 14:06:41.885525 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:41 crc kubenswrapper[4907]: I0313 14:06:41.885543 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:41 crc kubenswrapper[4907]: I0313 14:06:41.885568 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:41 crc kubenswrapper[4907]: I0313 14:06:41.885622 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:41Z","lastTransitionTime":"2026-03-13T14:06:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:41 crc kubenswrapper[4907]: I0313 14:06:41.988695 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:41 crc kubenswrapper[4907]: I0313 14:06:41.988761 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:41 crc kubenswrapper[4907]: I0313 14:06:41.988778 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:41 crc kubenswrapper[4907]: I0313 14:06:41.988829 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:41 crc kubenswrapper[4907]: I0313 14:06:41.988848 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:41Z","lastTransitionTime":"2026-03-13T14:06:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:42 crc kubenswrapper[4907]: I0313 14:06:42.092650 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:42 crc kubenswrapper[4907]: I0313 14:06:42.092713 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:42 crc kubenswrapper[4907]: I0313 14:06:42.092729 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:42 crc kubenswrapper[4907]: I0313 14:06:42.092757 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:42 crc kubenswrapper[4907]: I0313 14:06:42.092780 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:42Z","lastTransitionTime":"2026-03-13T14:06:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:42 crc kubenswrapper[4907]: I0313 14:06:42.195449 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:42 crc kubenswrapper[4907]: I0313 14:06:42.195517 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:42 crc kubenswrapper[4907]: I0313 14:06:42.195545 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:42 crc kubenswrapper[4907]: I0313 14:06:42.195573 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:42 crc kubenswrapper[4907]: I0313 14:06:42.195592 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:42Z","lastTransitionTime":"2026-03-13T14:06:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:42 crc kubenswrapper[4907]: I0313 14:06:42.297866 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:42 crc kubenswrapper[4907]: I0313 14:06:42.298002 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:42 crc kubenswrapper[4907]: I0313 14:06:42.298018 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:42 crc kubenswrapper[4907]: I0313 14:06:42.298046 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:42 crc kubenswrapper[4907]: I0313 14:06:42.298064 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:42Z","lastTransitionTime":"2026-03-13T14:06:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:42 crc kubenswrapper[4907]: I0313 14:06:42.400965 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:42 crc kubenswrapper[4907]: I0313 14:06:42.401017 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:42 crc kubenswrapper[4907]: I0313 14:06:42.401034 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:42 crc kubenswrapper[4907]: I0313 14:06:42.401060 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:42 crc kubenswrapper[4907]: I0313 14:06:42.401078 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:42Z","lastTransitionTime":"2026-03-13T14:06:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:42 crc kubenswrapper[4907]: I0313 14:06:42.506228 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:42 crc kubenswrapper[4907]: I0313 14:06:42.506298 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:42 crc kubenswrapper[4907]: I0313 14:06:42.506316 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:42 crc kubenswrapper[4907]: I0313 14:06:42.506342 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:42 crc kubenswrapper[4907]: I0313 14:06:42.506360 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:42Z","lastTransitionTime":"2026-03-13T14:06:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:42 crc kubenswrapper[4907]: I0313 14:06:42.609363 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:42 crc kubenswrapper[4907]: I0313 14:06:42.609430 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:42 crc kubenswrapper[4907]: I0313 14:06:42.609447 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:42 crc kubenswrapper[4907]: I0313 14:06:42.609476 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:42 crc kubenswrapper[4907]: I0313 14:06:42.609493 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:42Z","lastTransitionTime":"2026-03-13T14:06:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:42 crc kubenswrapper[4907]: I0313 14:06:42.713188 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:42 crc kubenswrapper[4907]: I0313 14:06:42.713271 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:42 crc kubenswrapper[4907]: I0313 14:06:42.713298 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:42 crc kubenswrapper[4907]: I0313 14:06:42.713331 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:42 crc kubenswrapper[4907]: I0313 14:06:42.713355 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:42Z","lastTransitionTime":"2026-03-13T14:06:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:42 crc kubenswrapper[4907]: I0313 14:06:42.782391 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:06:42 crc kubenswrapper[4907]: E0313 14:06:42.782576 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:06:42 crc kubenswrapper[4907]: I0313 14:06:42.782682 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:06:42 crc kubenswrapper[4907]: E0313 14:06:42.782816 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:06:42 crc kubenswrapper[4907]: I0313 14:06:42.782390 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:06:42 crc kubenswrapper[4907]: E0313 14:06:42.783040 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:06:42 crc kubenswrapper[4907]: I0313 14:06:42.818790 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:42 crc kubenswrapper[4907]: I0313 14:06:42.818932 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:42 crc kubenswrapper[4907]: I0313 14:06:42.818962 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:42 crc kubenswrapper[4907]: I0313 14:06:42.818999 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:42 crc kubenswrapper[4907]: I0313 14:06:42.819025 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:42Z","lastTransitionTime":"2026-03-13T14:06:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:42 crc kubenswrapper[4907]: I0313 14:06:42.922294 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:42 crc kubenswrapper[4907]: I0313 14:06:42.922366 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:42 crc kubenswrapper[4907]: I0313 14:06:42.922384 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:42 crc kubenswrapper[4907]: I0313 14:06:42.922412 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:42 crc kubenswrapper[4907]: I0313 14:06:42.922431 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:42Z","lastTransitionTime":"2026-03-13T14:06:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:43 crc kubenswrapper[4907]: I0313 14:06:43.026925 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:43 crc kubenswrapper[4907]: I0313 14:06:43.026996 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:43 crc kubenswrapper[4907]: I0313 14:06:43.027016 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:43 crc kubenswrapper[4907]: I0313 14:06:43.027044 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:43 crc kubenswrapper[4907]: I0313 14:06:43.027106 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:43Z","lastTransitionTime":"2026-03-13T14:06:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:43 crc kubenswrapper[4907]: I0313 14:06:43.130044 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:43 crc kubenswrapper[4907]: I0313 14:06:43.130086 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:43 crc kubenswrapper[4907]: I0313 14:06:43.130099 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:43 crc kubenswrapper[4907]: I0313 14:06:43.130119 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:43 crc kubenswrapper[4907]: I0313 14:06:43.130132 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:43Z","lastTransitionTime":"2026-03-13T14:06:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:43 crc kubenswrapper[4907]: I0313 14:06:43.233271 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:43 crc kubenswrapper[4907]: I0313 14:06:43.233338 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:43 crc kubenswrapper[4907]: I0313 14:06:43.233356 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:43 crc kubenswrapper[4907]: I0313 14:06:43.233386 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:43 crc kubenswrapper[4907]: I0313 14:06:43.233405 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:43Z","lastTransitionTime":"2026-03-13T14:06:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:43 crc kubenswrapper[4907]: I0313 14:06:43.335933 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:43 crc kubenswrapper[4907]: I0313 14:06:43.336015 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:43 crc kubenswrapper[4907]: I0313 14:06:43.336036 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:43 crc kubenswrapper[4907]: I0313 14:06:43.336065 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:43 crc kubenswrapper[4907]: I0313 14:06:43.336084 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:43Z","lastTransitionTime":"2026-03-13T14:06:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:43 crc kubenswrapper[4907]: I0313 14:06:43.439198 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:43 crc kubenswrapper[4907]: I0313 14:06:43.439269 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:43 crc kubenswrapper[4907]: I0313 14:06:43.439287 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:43 crc kubenswrapper[4907]: I0313 14:06:43.439366 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:43 crc kubenswrapper[4907]: I0313 14:06:43.439393 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:43Z","lastTransitionTime":"2026-03-13T14:06:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:43 crc kubenswrapper[4907]: I0313 14:06:43.543454 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:43 crc kubenswrapper[4907]: I0313 14:06:43.543547 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:43 crc kubenswrapper[4907]: I0313 14:06:43.543573 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:43 crc kubenswrapper[4907]: I0313 14:06:43.543608 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:43 crc kubenswrapper[4907]: I0313 14:06:43.543704 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:43Z","lastTransitionTime":"2026-03-13T14:06:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:43 crc kubenswrapper[4907]: I0313 14:06:43.647448 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:43 crc kubenswrapper[4907]: I0313 14:06:43.647532 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:43 crc kubenswrapper[4907]: I0313 14:06:43.647552 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:43 crc kubenswrapper[4907]: I0313 14:06:43.647584 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:43 crc kubenswrapper[4907]: I0313 14:06:43.647607 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:43Z","lastTransitionTime":"2026-03-13T14:06:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:43 crc kubenswrapper[4907]: I0313 14:06:43.751713 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:43 crc kubenswrapper[4907]: I0313 14:06:43.751834 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:43 crc kubenswrapper[4907]: I0313 14:06:43.751856 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:43 crc kubenswrapper[4907]: I0313 14:06:43.751957 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:43 crc kubenswrapper[4907]: I0313 14:06:43.751981 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:43Z","lastTransitionTime":"2026-03-13T14:06:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:43 crc kubenswrapper[4907]: I0313 14:06:43.855398 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:43 crc kubenswrapper[4907]: I0313 14:06:43.855483 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:43 crc kubenswrapper[4907]: I0313 14:06:43.855508 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:43 crc kubenswrapper[4907]: I0313 14:06:43.855541 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:43 crc kubenswrapper[4907]: I0313 14:06:43.855566 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:43Z","lastTransitionTime":"2026-03-13T14:06:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:43 crc kubenswrapper[4907]: I0313 14:06:43.959374 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:43 crc kubenswrapper[4907]: I0313 14:06:43.959442 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:43 crc kubenswrapper[4907]: I0313 14:06:43.959460 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:43 crc kubenswrapper[4907]: I0313 14:06:43.959489 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:43 crc kubenswrapper[4907]: I0313 14:06:43.959508 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:43Z","lastTransitionTime":"2026-03-13T14:06:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:44 crc kubenswrapper[4907]: I0313 14:06:44.062635 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:44 crc kubenswrapper[4907]: I0313 14:06:44.062715 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:44 crc kubenswrapper[4907]: I0313 14:06:44.062738 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:44 crc kubenswrapper[4907]: I0313 14:06:44.062770 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:44 crc kubenswrapper[4907]: I0313 14:06:44.062842 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:44Z","lastTransitionTime":"2026-03-13T14:06:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:44 crc kubenswrapper[4907]: I0313 14:06:44.166447 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:44 crc kubenswrapper[4907]: I0313 14:06:44.166511 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:44 crc kubenswrapper[4907]: I0313 14:06:44.166532 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:44 crc kubenswrapper[4907]: I0313 14:06:44.166562 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:44 crc kubenswrapper[4907]: I0313 14:06:44.166588 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:44Z","lastTransitionTime":"2026-03-13T14:06:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:44 crc kubenswrapper[4907]: I0313 14:06:44.270129 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:44 crc kubenswrapper[4907]: I0313 14:06:44.270235 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:44 crc kubenswrapper[4907]: I0313 14:06:44.270263 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:44 crc kubenswrapper[4907]: I0313 14:06:44.270297 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:44 crc kubenswrapper[4907]: I0313 14:06:44.270319 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:44Z","lastTransitionTime":"2026-03-13T14:06:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:44 crc kubenswrapper[4907]: I0313 14:06:44.373494 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:44 crc kubenswrapper[4907]: I0313 14:06:44.373578 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:44 crc kubenswrapper[4907]: I0313 14:06:44.373607 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:44 crc kubenswrapper[4907]: I0313 14:06:44.373646 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:44 crc kubenswrapper[4907]: I0313 14:06:44.373671 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:44Z","lastTransitionTime":"2026-03-13T14:06:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:44 crc kubenswrapper[4907]: I0313 14:06:44.476731 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:44 crc kubenswrapper[4907]: I0313 14:06:44.476811 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:44 crc kubenswrapper[4907]: I0313 14:06:44.476829 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:44 crc kubenswrapper[4907]: I0313 14:06:44.476857 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:44 crc kubenswrapper[4907]: I0313 14:06:44.476918 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:44Z","lastTransitionTime":"2026-03-13T14:06:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:44 crc kubenswrapper[4907]: I0313 14:06:44.580185 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:44 crc kubenswrapper[4907]: I0313 14:06:44.580265 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:44 crc kubenswrapper[4907]: I0313 14:06:44.580292 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:44 crc kubenswrapper[4907]: I0313 14:06:44.580325 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:44 crc kubenswrapper[4907]: I0313 14:06:44.580351 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:44Z","lastTransitionTime":"2026-03-13T14:06:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:44 crc kubenswrapper[4907]: I0313 14:06:44.682901 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:44 crc kubenswrapper[4907]: I0313 14:06:44.682969 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:44 crc kubenswrapper[4907]: I0313 14:06:44.682981 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:44 crc kubenswrapper[4907]: I0313 14:06:44.683004 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:44 crc kubenswrapper[4907]: I0313 14:06:44.683017 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:44Z","lastTransitionTime":"2026-03-13T14:06:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:44 crc kubenswrapper[4907]: I0313 14:06:44.782241 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:06:44 crc kubenswrapper[4907]: I0313 14:06:44.782379 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:06:44 crc kubenswrapper[4907]: E0313 14:06:44.782434 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:06:44 crc kubenswrapper[4907]: I0313 14:06:44.782481 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:06:44 crc kubenswrapper[4907]: E0313 14:06:44.782649 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:06:44 crc kubenswrapper[4907]: E0313 14:06:44.782983 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:06:44 crc kubenswrapper[4907]: I0313 14:06:44.785614 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:44 crc kubenswrapper[4907]: I0313 14:06:44.785651 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:44 crc kubenswrapper[4907]: I0313 14:06:44.785662 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:44 crc kubenswrapper[4907]: I0313 14:06:44.785680 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:44 crc kubenswrapper[4907]: I0313 14:06:44.785693 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:44Z","lastTransitionTime":"2026-03-13T14:06:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:44 crc kubenswrapper[4907]: I0313 14:06:44.889042 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:44 crc kubenswrapper[4907]: I0313 14:06:44.889088 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:44 crc kubenswrapper[4907]: I0313 14:06:44.889099 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:44 crc kubenswrapper[4907]: I0313 14:06:44.889117 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:44 crc kubenswrapper[4907]: I0313 14:06:44.889133 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:44Z","lastTransitionTime":"2026-03-13T14:06:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:44 crc kubenswrapper[4907]: I0313 14:06:44.993707 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:44 crc kubenswrapper[4907]: I0313 14:06:44.993777 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:44 crc kubenswrapper[4907]: I0313 14:06:44.993796 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:44 crc kubenswrapper[4907]: I0313 14:06:44.993833 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:44 crc kubenswrapper[4907]: I0313 14:06:44.993856 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:44Z","lastTransitionTime":"2026-03-13T14:06:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:45 crc kubenswrapper[4907]: I0313 14:06:45.097290 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:45 crc kubenswrapper[4907]: I0313 14:06:45.097358 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:45 crc kubenswrapper[4907]: I0313 14:06:45.097378 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:45 crc kubenswrapper[4907]: I0313 14:06:45.097405 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:45 crc kubenswrapper[4907]: I0313 14:06:45.097425 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:45Z","lastTransitionTime":"2026-03-13T14:06:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:45 crc kubenswrapper[4907]: I0313 14:06:45.199756 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:45 crc kubenswrapper[4907]: I0313 14:06:45.199858 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:45 crc kubenswrapper[4907]: I0313 14:06:45.199913 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:45 crc kubenswrapper[4907]: I0313 14:06:45.199978 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:45 crc kubenswrapper[4907]: I0313 14:06:45.199999 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:45Z","lastTransitionTime":"2026-03-13T14:06:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:45 crc kubenswrapper[4907]: I0313 14:06:45.202232 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"94b045cf2fed56457a2ca14f489bb69470198f159f16c207fe061d034dba1555"} Mar 13 14:06:45 crc kubenswrapper[4907]: I0313 14:06:45.202419 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"bd28f7a5c35822582b70ed41645294aaede10e28ff59f3f3b5d36e7da8237b76"} Mar 13 14:06:45 crc kubenswrapper[4907]: I0313 14:06:45.220613 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 14:06:45 crc kubenswrapper[4907]: I0313 14:06:45.255436 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 14:06:45 crc kubenswrapper[4907]: I0313 14:06:45.279865 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 14:06:45 crc kubenswrapper[4907]: I0313 14:06:45.299046 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5d2b054-0ed8-419c-b731-f14d3598372f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4b503aa3b13268b3688d067f3e456c86afac1fda171aced0990899c3eb89b79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be40a1a33fa482eab363537afaa901127d560afbc1b5b287ef11e66391a5be4c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d1b539bbd918fc50428fc18cabcc4c8a43cc96f220b5692d67cf1c8dfbf8ed8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T14:06:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0313 14:06:13.202871 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0313 14:06:13.203175 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0313 14:06:13.204427 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2471728650/tls.crt::/tmp/serving-cert-2471728650/tls.key\\\\\\\"\\\\nI0313 14:06:13.581160 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0313 14:06:13.584012 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0313 14:06:13.584036 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0313 14:06:13.584064 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0313 14:06:13.584072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0313 14:06:13.594504 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0313 14:06:13.594891 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0313 14:06:13.595323 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0313 14:06:13.595363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0313 14:06:13.595370 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0313 14:06:13.595376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0313 14:06:13.597226 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3f2b6756078925f8562b6053b3b83798efb72288f5fd18fff680a4ab0d33feb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 14:06:45 crc kubenswrapper[4907]: I0313 14:06:45.309510 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:45 crc kubenswrapper[4907]: I0313 14:06:45.309542 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:45 crc kubenswrapper[4907]: I0313 14:06:45.309553 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:45 crc kubenswrapper[4907]: I0313 14:06:45.309569 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:45 crc kubenswrapper[4907]: I0313 14:06:45.309580 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:45Z","lastTransitionTime":"2026-03-13T14:06:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:45 crc kubenswrapper[4907]: I0313 14:06:45.321677 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 14:06:45 crc kubenswrapper[4907]: I0313 14:06:45.337093 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94b045cf2fed56457a2ca14f489bb69470198f159f16c207fe061d034dba1555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd28f7a5c35822582b70ed41645294aaede10e28ff59f3f3b5d36e7da8237b76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 14:06:45 crc kubenswrapper[4907]: I0313 14:06:45.349826 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 14:06:45 crc kubenswrapper[4907]: I0313 14:06:45.414343 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:45 crc kubenswrapper[4907]: I0313 14:06:45.414412 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:45 crc kubenswrapper[4907]: I0313 14:06:45.414432 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:45 crc kubenswrapper[4907]: I0313 14:06:45.414464 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:45 crc kubenswrapper[4907]: I0313 14:06:45.414482 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:45Z","lastTransitionTime":"2026-03-13T14:06:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:45 crc kubenswrapper[4907]: I0313 14:06:45.517854 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:45 crc kubenswrapper[4907]: I0313 14:06:45.518243 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:45 crc kubenswrapper[4907]: I0313 14:06:45.518437 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:45 crc kubenswrapper[4907]: I0313 14:06:45.518708 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:45 crc kubenswrapper[4907]: I0313 14:06:45.518929 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:45Z","lastTransitionTime":"2026-03-13T14:06:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:45 crc kubenswrapper[4907]: I0313 14:06:45.622545 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:45 crc kubenswrapper[4907]: I0313 14:06:45.622624 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:45 crc kubenswrapper[4907]: I0313 14:06:45.622643 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:45 crc kubenswrapper[4907]: I0313 14:06:45.622671 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:45 crc kubenswrapper[4907]: I0313 14:06:45.622688 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:45Z","lastTransitionTime":"2026-03-13T14:06:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:45 crc kubenswrapper[4907]: I0313 14:06:45.725860 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:45 crc kubenswrapper[4907]: I0313 14:06:45.725950 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:45 crc kubenswrapper[4907]: I0313 14:06:45.725965 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:45 crc kubenswrapper[4907]: I0313 14:06:45.725987 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:45 crc kubenswrapper[4907]: I0313 14:06:45.726003 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:45Z","lastTransitionTime":"2026-03-13T14:06:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:45 crc kubenswrapper[4907]: I0313 14:06:45.827952 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:45 crc kubenswrapper[4907]: I0313 14:06:45.828186 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:45 crc kubenswrapper[4907]: I0313 14:06:45.828277 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:45 crc kubenswrapper[4907]: I0313 14:06:45.828389 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:45 crc kubenswrapper[4907]: I0313 14:06:45.828476 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:45Z","lastTransitionTime":"2026-03-13T14:06:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:45 crc kubenswrapper[4907]: I0313 14:06:45.931671 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:45 crc kubenswrapper[4907]: I0313 14:06:45.931954 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:45 crc kubenswrapper[4907]: I0313 14:06:45.932016 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:45 crc kubenswrapper[4907]: I0313 14:06:45.932085 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:45 crc kubenswrapper[4907]: I0313 14:06:45.932143 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:45Z","lastTransitionTime":"2026-03-13T14:06:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.034468 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.034668 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.034751 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.034810 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.034863 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:46Z","lastTransitionTime":"2026-03-13T14:06:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.137740 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.138375 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.138524 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.138643 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.138732 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:46Z","lastTransitionTime":"2026-03-13T14:06:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.238040 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.238115 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.238138 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.238165 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.238184 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:46Z","lastTransitionTime":"2026-03-13T14:06:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:46 crc kubenswrapper[4907]: E0313 14:06:46.262399 4907 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4b182bae-7f2d-4b54-934c-85a9a277d132\\\",\\\"systemUUID\\\":\\\"05bfbfa3-f04f-45ba-8233-a355fc774c24\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:46Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.268181 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.268255 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.268276 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.268305 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.268325 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:46Z","lastTransitionTime":"2026-03-13T14:06:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:46 crc kubenswrapper[4907]: E0313 14:06:46.290953 4907 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4b182bae-7f2d-4b54-934c-85a9a277d132\\\",\\\"systemUUID\\\":\\\"05bfbfa3-f04f-45ba-8233-a355fc774c24\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:46Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.297268 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.297510 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.298147 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.298559 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.298816 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:46Z","lastTransitionTime":"2026-03-13T14:06:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:46 crc kubenswrapper[4907]: E0313 14:06:46.322585 4907 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4b182bae-7f2d-4b54-934c-85a9a277d132\\\",\\\"systemUUID\\\":\\\"05bfbfa3-f04f-45ba-8233-a355fc774c24\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:46Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.331437 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.331709 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.332601 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.333326 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.333584 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:46Z","lastTransitionTime":"2026-03-13T14:06:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:46 crc kubenswrapper[4907]: E0313 14:06:46.357546 4907 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4b182bae-7f2d-4b54-934c-85a9a277d132\\\",\\\"systemUUID\\\":\\\"05bfbfa3-f04f-45ba-8233-a355fc774c24\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:46Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.364683 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.365100 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.365492 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.366159 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.366394 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:46Z","lastTransitionTime":"2026-03-13T14:06:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:46 crc kubenswrapper[4907]: E0313 14:06:46.389841 4907 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4b182bae-7f2d-4b54-934c-85a9a277d132\\\",\\\"systemUUID\\\":\\\"05bfbfa3-f04f-45ba-8233-a355fc774c24\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:46Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:46 crc kubenswrapper[4907]: E0313 14:06:46.391157 4907 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.394541 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.394594 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.394613 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.394640 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.394658 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:46Z","lastTransitionTime":"2026-03-13T14:06:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.498306 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.498384 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.498414 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.498473 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.498495 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:46Z","lastTransitionTime":"2026-03-13T14:06:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.601818 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.602175 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.602278 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.602380 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.602515 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:46Z","lastTransitionTime":"2026-03-13T14:06:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.706526 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.706973 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.707059 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.707249 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.707372 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:46Z","lastTransitionTime":"2026-03-13T14:06:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.782354 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.782408 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.782355 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:06:46 crc kubenswrapper[4907]: E0313 14:06:46.782560 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:06:46 crc kubenswrapper[4907]: E0313 14:06:46.782707 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:06:46 crc kubenswrapper[4907]: E0313 14:06:46.782871 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.810838 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.810940 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.810964 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.810992 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.811012 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:46Z","lastTransitionTime":"2026-03-13T14:06:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.914355 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.914441 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.914464 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.914490 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.914512 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:46Z","lastTransitionTime":"2026-03-13T14:06:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.018651 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.019070 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.019214 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.019357 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.019486 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:47Z","lastTransitionTime":"2026-03-13T14:06:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.123679 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.123758 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.123780 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.123808 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.123828 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:47Z","lastTransitionTime":"2026-03-13T14:06:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.210370 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"afb1d430459aafab2d6bc2507b4b4059529401e51fdf8e9848b0d7bb8651db68"} Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.227364 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:47Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.227443 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.227526 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.227550 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.227613 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.227634 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:47Z","lastTransitionTime":"2026-03-13T14:06:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.248275 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:47Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.268284 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:47Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.286511 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afb1d430459aafab2d6bc2507b4b4059529401e51fdf8e9848b0d7bb8651db68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:47Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.304511 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5d2b054-0ed8-419c-b731-f14d3598372f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4b503aa3b13268b3688d067f3e456c86afac1fda171aced0990899c3eb89b79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be40a1a33fa482eab363537afaa901127d560afbc1b5b287ef11e66391a5be4c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d1b539bbd918fc50428fc18cabcc4c8a43cc96f220b5692d67cf1c8dfbf8ed8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T14:06:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0313 14:06:13.202871 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0313 14:06:13.203175 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0313 14:06:13.204427 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2471728650/tls.crt::/tmp/serving-cert-2471728650/tls.key\\\\\\\"\\\\nI0313 14:06:13.581160 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0313 14:06:13.584012 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0313 14:06:13.584036 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0313 14:06:13.584064 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0313 14:06:13.584072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0313 14:06:13.594504 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0313 14:06:13.594891 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0313 14:06:13.595323 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0313 14:06:13.595363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0313 14:06:13.595370 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0313 14:06:13.595376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0313 14:06:13.597226 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3f2b6756078925f8562b6053b3b83798efb72288f5fd18fff680a4ab0d33feb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:47Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.322779 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-l7zg7"] Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.323363 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-l7zg7" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.327298 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.327663 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.327814 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.328648 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:47Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.332566 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.332763 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.332936 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.333071 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.333186 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:47Z","lastTransitionTime":"2026-03-13T14:06:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.348319 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94b045cf2fed56457a2ca14f489bb69470198f159f16c207fe061d034dba1555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd28f7a5c35822582b70ed41645294aaede10e28ff59f3f3b5d36e7da8237b76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:47Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.368219 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:47Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.385037 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94b045cf2fed56457a2ca14f489bb69470198f159f16c207fe061d034dba1555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd28f7a5c35822582b70ed41645294aaede10e28ff59f3f3b5d36e7da8237b76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:47Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.400222 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afb1d430459aafab2d6bc2507b4b4059529401e51fdf8e9848b0d7bb8651db68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:47Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.414188 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l7zg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d105034b-f041-4fb5-8df6-6b1bb7681fda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szmh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l7zg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:47Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.435739 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.435777 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.435786 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.435802 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.435814 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:47Z","lastTransitionTime":"2026-03-13T14:06:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.437420 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5d2b054-0ed8-419c-b731-f14d3598372f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4b503aa3b13268b3688d067f3e456c86afac1fda171aced0990899c3eb89b79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be40a1a33fa482eab363537afaa901127d560afbc1b5b287ef11e66391a5be4c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d1b539bbd918fc50428fc18cabcc4c8a43cc96f220b5692d67cf1c8dfbf8ed8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T14:06:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0313 14:06:13.202871 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0313 14:06:13.203175 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0313 14:06:13.204427 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2471728650/tls.crt::/tmp/serving-cert-2471728650/tls.key\\\\\\\"\\\\nI0313 14:06:13.581160 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0313 14:06:13.584012 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0313 14:06:13.584036 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0313 14:06:13.584064 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0313 14:06:13.584072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0313 14:06:13.594504 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0313 14:06:13.594891 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0313 14:06:13.595323 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0313 14:06:13.595363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0313 14:06:13.595370 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0313 14:06:13.595376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0313 14:06:13.597226 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3f2b6756078925f8562b6053b3b83798efb72288f5fd18fff680a4ab0d33feb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:47Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.454373 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:47Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.468939 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:47Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.483229 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:47Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.490927 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-szmh9\" (UniqueName: \"kubernetes.io/projected/d105034b-f041-4fb5-8df6-6b1bb7681fda-kube-api-access-szmh9\") pod \"node-resolver-l7zg7\" (UID: \"d105034b-f041-4fb5-8df6-6b1bb7681fda\") " pod="openshift-dns/node-resolver-l7zg7" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.491194 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/d105034b-f041-4fb5-8df6-6b1bb7681fda-hosts-file\") pod \"node-resolver-l7zg7\" (UID: \"d105034b-f041-4fb5-8df6-6b1bb7681fda\") " pod="openshift-dns/node-resolver-l7zg7" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.538990 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.539238 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.539354 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.539543 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.539668 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:47Z","lastTransitionTime":"2026-03-13T14:06:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.592554 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/d105034b-f041-4fb5-8df6-6b1bb7681fda-hosts-file\") pod \"node-resolver-l7zg7\" (UID: \"d105034b-f041-4fb5-8df6-6b1bb7681fda\") " pod="openshift-dns/node-resolver-l7zg7" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.592662 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-szmh9\" (UniqueName: \"kubernetes.io/projected/d105034b-f041-4fb5-8df6-6b1bb7681fda-kube-api-access-szmh9\") pod \"node-resolver-l7zg7\" (UID: \"d105034b-f041-4fb5-8df6-6b1bb7681fda\") " pod="openshift-dns/node-resolver-l7zg7" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.593178 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/d105034b-f041-4fb5-8df6-6b1bb7681fda-hosts-file\") pod \"node-resolver-l7zg7\" (UID: \"d105034b-f041-4fb5-8df6-6b1bb7681fda\") " pod="openshift-dns/node-resolver-l7zg7" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.617820 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-szmh9\" (UniqueName: \"kubernetes.io/projected/d105034b-f041-4fb5-8df6-6b1bb7681fda-kube-api-access-szmh9\") pod \"node-resolver-l7zg7\" (UID: \"d105034b-f041-4fb5-8df6-6b1bb7681fda\") " pod="openshift-dns/node-resolver-l7zg7" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.642549 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.642773 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.642942 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.642680 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-l7zg7" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.643264 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.643338 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:47Z","lastTransitionTime":"2026-03-13T14:06:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:47 crc kubenswrapper[4907]: W0313 14:06:47.656307 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd105034b_f041_4fb5_8df6_6b1bb7681fda.slice/crio-43541ca6821c8fc2f866fd8525a0781a2c00d4fd50c8c0b176be0d71affd01db WatchSource:0}: Error finding container 43541ca6821c8fc2f866fd8525a0781a2c00d4fd50c8c0b176be0d71affd01db: Status 404 returned error can't find the container with id 43541ca6821c8fc2f866fd8525a0781a2c00d4fd50c8c0b176be0d71affd01db Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.697857 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-xtwf7"] Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.698152 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-p8ptc"] Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.698782 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-p8ptc" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.699165 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-xtwf7" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.700245 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-hm56j"] Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.700535 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.704738 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.704793 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.704819 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.705016 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.705075 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.705099 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.705224 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.706315 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.706509 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.706748 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.706865 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.708315 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.716967 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:47Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.728990 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:47Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.740751 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xtwf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64ebd338-0158-44dc-90ab-e7c285e87762\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6cwbq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xtwf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:47Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.749605 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.749652 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.749666 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.749689 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.749707 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:47Z","lastTransitionTime":"2026-03-13T14:06:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.759709 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:47Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.769658 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afb1d430459aafab2d6bc2507b4b4059529401e51fdf8e9848b0d7bb8651db68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:47Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.780829 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:47Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.794440 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/64ebd338-0158-44dc-90ab-e7c285e87762-multus-conf-dir\") pod \"multus-xtwf7\" (UID: \"64ebd338-0158-44dc-90ab-e7c285e87762\") " pod="openshift-multus/multus-xtwf7" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.794504 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/57e4e912-9a98-4419-87ce-f16e178ed921-cni-binary-copy\") pod \"multus-additional-cni-plugins-p8ptc\" (UID: \"57e4e912-9a98-4419-87ce-f16e178ed921\") " pod="openshift-multus/multus-additional-cni-plugins-p8ptc" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.794533 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/57e4e912-9a98-4419-87ce-f16e178ed921-tuning-conf-dir\") pod \"multus-additional-cni-plugins-p8ptc\" (UID: \"57e4e912-9a98-4419-87ce-f16e178ed921\") " pod="openshift-multus/multus-additional-cni-plugins-p8ptc" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.794556 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f3fb4213-144f-4b01-9d78-ff3315761bb3-mcd-auth-proxy-config\") pod \"machine-config-daemon-hm56j\" (UID: \"f3fb4213-144f-4b01-9d78-ff3315761bb3\") " pod="openshift-machine-config-operator/machine-config-daemon-hm56j" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.794579 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vxwbn\" (UniqueName: \"kubernetes.io/projected/f3fb4213-144f-4b01-9d78-ff3315761bb3-kube-api-access-vxwbn\") pod \"machine-config-daemon-hm56j\" (UID: \"f3fb4213-144f-4b01-9d78-ff3315761bb3\") " pod="openshift-machine-config-operator/machine-config-daemon-hm56j" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.794603 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/f3fb4213-144f-4b01-9d78-ff3315761bb3-proxy-tls\") pod \"machine-config-daemon-hm56j\" (UID: \"f3fb4213-144f-4b01-9d78-ff3315761bb3\") " pod="openshift-machine-config-operator/machine-config-daemon-hm56j" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.794645 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/64ebd338-0158-44dc-90ab-e7c285e87762-host-run-k8s-cni-cncf-io\") pod \"multus-xtwf7\" (UID: \"64ebd338-0158-44dc-90ab-e7c285e87762\") " pod="openshift-multus/multus-xtwf7" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.794664 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/64ebd338-0158-44dc-90ab-e7c285e87762-host-var-lib-kubelet\") pod \"multus-xtwf7\" (UID: \"64ebd338-0158-44dc-90ab-e7c285e87762\") " pod="openshift-multus/multus-xtwf7" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.794835 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/f3fb4213-144f-4b01-9d78-ff3315761bb3-rootfs\") pod \"machine-config-daemon-hm56j\" (UID: \"f3fb4213-144f-4b01-9d78-ff3315761bb3\") " pod="openshift-machine-config-operator/machine-config-daemon-hm56j" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.795025 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/57e4e912-9a98-4419-87ce-f16e178ed921-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-p8ptc\" (UID: \"57e4e912-9a98-4419-87ce-f16e178ed921\") " pod="openshift-multus/multus-additional-cni-plugins-p8ptc" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.795042 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/64ebd338-0158-44dc-90ab-e7c285e87762-multus-cni-dir\") pod \"multus-xtwf7\" (UID: \"64ebd338-0158-44dc-90ab-e7c285e87762\") " pod="openshift-multus/multus-xtwf7" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.795087 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/64ebd338-0158-44dc-90ab-e7c285e87762-host-var-lib-cni-bin\") pod \"multus-xtwf7\" (UID: \"64ebd338-0158-44dc-90ab-e7c285e87762\") " pod="openshift-multus/multus-xtwf7" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.795125 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/64ebd338-0158-44dc-90ab-e7c285e87762-host-var-lib-cni-multus\") pod \"multus-xtwf7\" (UID: \"64ebd338-0158-44dc-90ab-e7c285e87762\") " pod="openshift-multus/multus-xtwf7" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.795164 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/64ebd338-0158-44dc-90ab-e7c285e87762-hostroot\") pod \"multus-xtwf7\" (UID: \"64ebd338-0158-44dc-90ab-e7c285e87762\") " pod="openshift-multus/multus-xtwf7" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.795183 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/64ebd338-0158-44dc-90ab-e7c285e87762-multus-socket-dir-parent\") pod \"multus-xtwf7\" (UID: \"64ebd338-0158-44dc-90ab-e7c285e87762\") " pod="openshift-multus/multus-xtwf7" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.795197 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/64ebd338-0158-44dc-90ab-e7c285e87762-host-run-multus-certs\") pod \"multus-xtwf7\" (UID: \"64ebd338-0158-44dc-90ab-e7c285e87762\") " pod="openshift-multus/multus-xtwf7" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.795312 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/57e4e912-9a98-4419-87ce-f16e178ed921-system-cni-dir\") pod \"multus-additional-cni-plugins-p8ptc\" (UID: \"57e4e912-9a98-4419-87ce-f16e178ed921\") " pod="openshift-multus/multus-additional-cni-plugins-p8ptc" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.795384 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/57e4e912-9a98-4419-87ce-f16e178ed921-cnibin\") pod \"multus-additional-cni-plugins-p8ptc\" (UID: \"57e4e912-9a98-4419-87ce-f16e178ed921\") " pod="openshift-multus/multus-additional-cni-plugins-p8ptc" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.795413 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7vldr\" (UniqueName: \"kubernetes.io/projected/57e4e912-9a98-4419-87ce-f16e178ed921-kube-api-access-7vldr\") pod \"multus-additional-cni-plugins-p8ptc\" (UID: \"57e4e912-9a98-4419-87ce-f16e178ed921\") " pod="openshift-multus/multus-additional-cni-plugins-p8ptc" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.795437 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/64ebd338-0158-44dc-90ab-e7c285e87762-multus-daemon-config\") pod \"multus-xtwf7\" (UID: \"64ebd338-0158-44dc-90ab-e7c285e87762\") " pod="openshift-multus/multus-xtwf7" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.795462 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/64ebd338-0158-44dc-90ab-e7c285e87762-cnibin\") pod \"multus-xtwf7\" (UID: \"64ebd338-0158-44dc-90ab-e7c285e87762\") " pod="openshift-multus/multus-xtwf7" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.795480 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/64ebd338-0158-44dc-90ab-e7c285e87762-host-run-netns\") pod \"multus-xtwf7\" (UID: \"64ebd338-0158-44dc-90ab-e7c285e87762\") " pod="openshift-multus/multus-xtwf7" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.795500 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/64ebd338-0158-44dc-90ab-e7c285e87762-etc-kubernetes\") pod \"multus-xtwf7\" (UID: \"64ebd338-0158-44dc-90ab-e7c285e87762\") " pod="openshift-multus/multus-xtwf7" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.795519 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/64ebd338-0158-44dc-90ab-e7c285e87762-os-release\") pod \"multus-xtwf7\" (UID: \"64ebd338-0158-44dc-90ab-e7c285e87762\") " pod="openshift-multus/multus-xtwf7" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.795543 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6cwbq\" (UniqueName: \"kubernetes.io/projected/64ebd338-0158-44dc-90ab-e7c285e87762-kube-api-access-6cwbq\") pod \"multus-xtwf7\" (UID: \"64ebd338-0158-44dc-90ab-e7c285e87762\") " pod="openshift-multus/multus-xtwf7" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.795568 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/64ebd338-0158-44dc-90ab-e7c285e87762-cni-binary-copy\") pod \"multus-xtwf7\" (UID: \"64ebd338-0158-44dc-90ab-e7c285e87762\") " pod="openshift-multus/multus-xtwf7" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.795610 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/57e4e912-9a98-4419-87ce-f16e178ed921-os-release\") pod \"multus-additional-cni-plugins-p8ptc\" (UID: \"57e4e912-9a98-4419-87ce-f16e178ed921\") " pod="openshift-multus/multus-additional-cni-plugins-p8ptc" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.795633 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/64ebd338-0158-44dc-90ab-e7c285e87762-system-cni-dir\") pod \"multus-xtwf7\" (UID: \"64ebd338-0158-44dc-90ab-e7c285e87762\") " pod="openshift-multus/multus-xtwf7" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.796226 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-p8ptc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57e4e912-9a98-4419-87ce-f16e178ed921\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-p8ptc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:47Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.814734 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5d2b054-0ed8-419c-b731-f14d3598372f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4b503aa3b13268b3688d067f3e456c86afac1fda171aced0990899c3eb89b79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be40a1a33fa482eab363537afaa901127d560afbc1b5b287ef11e66391a5be4c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d1b539bbd918fc50428fc18cabcc4c8a43cc96f220b5692d67cf1c8dfbf8ed8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T14:06:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0313 14:06:13.202871 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0313 14:06:13.203175 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0313 14:06:13.204427 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2471728650/tls.crt::/tmp/serving-cert-2471728650/tls.key\\\\\\\"\\\\nI0313 14:06:13.581160 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0313 14:06:13.584012 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0313 14:06:13.584036 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0313 14:06:13.584064 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0313 14:06:13.584072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0313 14:06:13.594504 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0313 14:06:13.594891 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0313 14:06:13.595323 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0313 14:06:13.595363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0313 14:06:13.595370 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0313 14:06:13.595376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0313 14:06:13.597226 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3f2b6756078925f8562b6053b3b83798efb72288f5fd18fff680a4ab0d33feb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:47Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.826657 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94b045cf2fed56457a2ca14f489bb69470198f159f16c207fe061d034dba1555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd28f7a5c35822582b70ed41645294aaede10e28ff59f3f3b5d36e7da8237b76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:47Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.836207 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l7zg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d105034b-f041-4fb5-8df6-6b1bb7681fda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szmh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l7zg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:47Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.849001 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l7zg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d105034b-f041-4fb5-8df6-6b1bb7681fda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szmh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l7zg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:47Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.854635 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.854673 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.854682 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.854707 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.854718 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:47Z","lastTransitionTime":"2026-03-13T14:06:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.863835 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5d2b054-0ed8-419c-b731-f14d3598372f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4b503aa3b13268b3688d067f3e456c86afac1fda171aced0990899c3eb89b79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be40a1a33fa482eab363537afaa901127d560afbc1b5b287ef11e66391a5be4c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d1b539bbd918fc50428fc18cabcc4c8a43cc96f220b5692d67cf1c8dfbf8ed8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T14:06:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0313 14:06:13.202871 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0313 14:06:13.203175 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0313 14:06:13.204427 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2471728650/tls.crt::/tmp/serving-cert-2471728650/tls.key\\\\\\\"\\\\nI0313 14:06:13.581160 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0313 14:06:13.584012 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0313 14:06:13.584036 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0313 14:06:13.584064 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0313 14:06:13.584072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0313 14:06:13.594504 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0313 14:06:13.594891 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0313 14:06:13.595323 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0313 14:06:13.595363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0313 14:06:13.595370 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0313 14:06:13.595376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0313 14:06:13.597226 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3f2b6756078925f8562b6053b3b83798efb72288f5fd18fff680a4ab0d33feb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:47Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.879378 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94b045cf2fed56457a2ca14f489bb69470198f159f16c207fe061d034dba1555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd28f7a5c35822582b70ed41645294aaede10e28ff59f3f3b5d36e7da8237b76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:47Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.896495 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/64ebd338-0158-44dc-90ab-e7c285e87762-multus-socket-dir-parent\") pod \"multus-xtwf7\" (UID: \"64ebd338-0158-44dc-90ab-e7c285e87762\") " pod="openshift-multus/multus-xtwf7" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.896548 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/64ebd338-0158-44dc-90ab-e7c285e87762-host-run-multus-certs\") pod \"multus-xtwf7\" (UID: \"64ebd338-0158-44dc-90ab-e7c285e87762\") " pod="openshift-multus/multus-xtwf7" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.896572 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/57e4e912-9a98-4419-87ce-f16e178ed921-system-cni-dir\") pod \"multus-additional-cni-plugins-p8ptc\" (UID: \"57e4e912-9a98-4419-87ce-f16e178ed921\") " pod="openshift-multus/multus-additional-cni-plugins-p8ptc" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.896592 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/57e4e912-9a98-4419-87ce-f16e178ed921-cnibin\") pod \"multus-additional-cni-plugins-p8ptc\" (UID: \"57e4e912-9a98-4419-87ce-f16e178ed921\") " pod="openshift-multus/multus-additional-cni-plugins-p8ptc" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.896616 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7vldr\" (UniqueName: \"kubernetes.io/projected/57e4e912-9a98-4419-87ce-f16e178ed921-kube-api-access-7vldr\") pod \"multus-additional-cni-plugins-p8ptc\" (UID: \"57e4e912-9a98-4419-87ce-f16e178ed921\") " pod="openshift-multus/multus-additional-cni-plugins-p8ptc" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.896638 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/64ebd338-0158-44dc-90ab-e7c285e87762-multus-daemon-config\") pod \"multus-xtwf7\" (UID: \"64ebd338-0158-44dc-90ab-e7c285e87762\") " pod="openshift-multus/multus-xtwf7" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.896667 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/64ebd338-0158-44dc-90ab-e7c285e87762-host-run-netns\") pod \"multus-xtwf7\" (UID: \"64ebd338-0158-44dc-90ab-e7c285e87762\") " pod="openshift-multus/multus-xtwf7" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.896687 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/64ebd338-0158-44dc-90ab-e7c285e87762-etc-kubernetes\") pod \"multus-xtwf7\" (UID: \"64ebd338-0158-44dc-90ab-e7c285e87762\") " pod="openshift-multus/multus-xtwf7" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.896708 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/64ebd338-0158-44dc-90ab-e7c285e87762-cnibin\") pod \"multus-xtwf7\" (UID: \"64ebd338-0158-44dc-90ab-e7c285e87762\") " pod="openshift-multus/multus-xtwf7" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.896727 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/64ebd338-0158-44dc-90ab-e7c285e87762-os-release\") pod \"multus-xtwf7\" (UID: \"64ebd338-0158-44dc-90ab-e7c285e87762\") " pod="openshift-multus/multus-xtwf7" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.896748 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6cwbq\" (UniqueName: \"kubernetes.io/projected/64ebd338-0158-44dc-90ab-e7c285e87762-kube-api-access-6cwbq\") pod \"multus-xtwf7\" (UID: \"64ebd338-0158-44dc-90ab-e7c285e87762\") " pod="openshift-multus/multus-xtwf7" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.896770 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/64ebd338-0158-44dc-90ab-e7c285e87762-cni-binary-copy\") pod \"multus-xtwf7\" (UID: \"64ebd338-0158-44dc-90ab-e7c285e87762\") " pod="openshift-multus/multus-xtwf7" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.896816 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/57e4e912-9a98-4419-87ce-f16e178ed921-os-release\") pod \"multus-additional-cni-plugins-p8ptc\" (UID: \"57e4e912-9a98-4419-87ce-f16e178ed921\") " pod="openshift-multus/multus-additional-cni-plugins-p8ptc" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.896838 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/64ebd338-0158-44dc-90ab-e7c285e87762-system-cni-dir\") pod \"multus-xtwf7\" (UID: \"64ebd338-0158-44dc-90ab-e7c285e87762\") " pod="openshift-multus/multus-xtwf7" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.896859 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/64ebd338-0158-44dc-90ab-e7c285e87762-multus-conf-dir\") pod \"multus-xtwf7\" (UID: \"64ebd338-0158-44dc-90ab-e7c285e87762\") " pod="openshift-multus/multus-xtwf7" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.896897 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f3fb4213-144f-4b01-9d78-ff3315761bb3-mcd-auth-proxy-config\") pod \"machine-config-daemon-hm56j\" (UID: \"f3fb4213-144f-4b01-9d78-ff3315761bb3\") " pod="openshift-machine-config-operator/machine-config-daemon-hm56j" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.896913 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vxwbn\" (UniqueName: \"kubernetes.io/projected/f3fb4213-144f-4b01-9d78-ff3315761bb3-kube-api-access-vxwbn\") pod \"machine-config-daemon-hm56j\" (UID: \"f3fb4213-144f-4b01-9d78-ff3315761bb3\") " pod="openshift-machine-config-operator/machine-config-daemon-hm56j" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.896936 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/57e4e912-9a98-4419-87ce-f16e178ed921-cni-binary-copy\") pod \"multus-additional-cni-plugins-p8ptc\" (UID: \"57e4e912-9a98-4419-87ce-f16e178ed921\") " pod="openshift-multus/multus-additional-cni-plugins-p8ptc" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.896951 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/57e4e912-9a98-4419-87ce-f16e178ed921-tuning-conf-dir\") pod \"multus-additional-cni-plugins-p8ptc\" (UID: \"57e4e912-9a98-4419-87ce-f16e178ed921\") " pod="openshift-multus/multus-additional-cni-plugins-p8ptc" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.896968 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/f3fb4213-144f-4b01-9d78-ff3315761bb3-proxy-tls\") pod \"machine-config-daemon-hm56j\" (UID: \"f3fb4213-144f-4b01-9d78-ff3315761bb3\") " pod="openshift-machine-config-operator/machine-config-daemon-hm56j" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.897002 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/64ebd338-0158-44dc-90ab-e7c285e87762-host-var-lib-kubelet\") pod \"multus-xtwf7\" (UID: \"64ebd338-0158-44dc-90ab-e7c285e87762\") " pod="openshift-multus/multus-xtwf7" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.897022 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/f3fb4213-144f-4b01-9d78-ff3315761bb3-rootfs\") pod \"machine-config-daemon-hm56j\" (UID: \"f3fb4213-144f-4b01-9d78-ff3315761bb3\") " pod="openshift-machine-config-operator/machine-config-daemon-hm56j" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.897055 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/64ebd338-0158-44dc-90ab-e7c285e87762-host-run-k8s-cni-cncf-io\") pod \"multus-xtwf7\" (UID: \"64ebd338-0158-44dc-90ab-e7c285e87762\") " pod="openshift-multus/multus-xtwf7" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.897073 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/64ebd338-0158-44dc-90ab-e7c285e87762-host-var-lib-cni-bin\") pod \"multus-xtwf7\" (UID: \"64ebd338-0158-44dc-90ab-e7c285e87762\") " pod="openshift-multus/multus-xtwf7" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.897091 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/64ebd338-0158-44dc-90ab-e7c285e87762-host-var-lib-cni-multus\") pod \"multus-xtwf7\" (UID: \"64ebd338-0158-44dc-90ab-e7c285e87762\") " pod="openshift-multus/multus-xtwf7" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.897108 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/64ebd338-0158-44dc-90ab-e7c285e87762-hostroot\") pod \"multus-xtwf7\" (UID: \"64ebd338-0158-44dc-90ab-e7c285e87762\") " pod="openshift-multus/multus-xtwf7" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.897131 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/57e4e912-9a98-4419-87ce-f16e178ed921-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-p8ptc\" (UID: \"57e4e912-9a98-4419-87ce-f16e178ed921\") " pod="openshift-multus/multus-additional-cni-plugins-p8ptc" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.897153 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/64ebd338-0158-44dc-90ab-e7c285e87762-multus-cni-dir\") pod \"multus-xtwf7\" (UID: \"64ebd338-0158-44dc-90ab-e7c285e87762\") " pod="openshift-multus/multus-xtwf7" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.897253 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/64ebd338-0158-44dc-90ab-e7c285e87762-multus-cni-dir\") pod \"multus-xtwf7\" (UID: \"64ebd338-0158-44dc-90ab-e7c285e87762\") " pod="openshift-multus/multus-xtwf7" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.897304 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:47Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.897645 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/64ebd338-0158-44dc-90ab-e7c285e87762-multus-socket-dir-parent\") pod \"multus-xtwf7\" (UID: \"64ebd338-0158-44dc-90ab-e7c285e87762\") " pod="openshift-multus/multus-xtwf7" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.897680 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/57e4e912-9a98-4419-87ce-f16e178ed921-os-release\") pod \"multus-additional-cni-plugins-p8ptc\" (UID: \"57e4e912-9a98-4419-87ce-f16e178ed921\") " pod="openshift-multus/multus-additional-cni-plugins-p8ptc" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.897725 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/64ebd338-0158-44dc-90ab-e7c285e87762-host-run-multus-certs\") pod \"multus-xtwf7\" (UID: \"64ebd338-0158-44dc-90ab-e7c285e87762\") " pod="openshift-multus/multus-xtwf7" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.897759 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/57e4e912-9a98-4419-87ce-f16e178ed921-system-cni-dir\") pod \"multus-additional-cni-plugins-p8ptc\" (UID: \"57e4e912-9a98-4419-87ce-f16e178ed921\") " pod="openshift-multus/multus-additional-cni-plugins-p8ptc" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.897789 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/57e4e912-9a98-4419-87ce-f16e178ed921-cnibin\") pod \"multus-additional-cni-plugins-p8ptc\" (UID: \"57e4e912-9a98-4419-87ce-f16e178ed921\") " pod="openshift-multus/multus-additional-cni-plugins-p8ptc" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.897861 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/64ebd338-0158-44dc-90ab-e7c285e87762-system-cni-dir\") pod \"multus-xtwf7\" (UID: \"64ebd338-0158-44dc-90ab-e7c285e87762\") " pod="openshift-multus/multus-xtwf7" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.897901 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/64ebd338-0158-44dc-90ab-e7c285e87762-cni-binary-copy\") pod \"multus-xtwf7\" (UID: \"64ebd338-0158-44dc-90ab-e7c285e87762\") " pod="openshift-multus/multus-xtwf7" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.897926 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/64ebd338-0158-44dc-90ab-e7c285e87762-multus-conf-dir\") pod \"multus-xtwf7\" (UID: \"64ebd338-0158-44dc-90ab-e7c285e87762\") " pod="openshift-multus/multus-xtwf7" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.897979 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/64ebd338-0158-44dc-90ab-e7c285e87762-host-run-netns\") pod \"multus-xtwf7\" (UID: \"64ebd338-0158-44dc-90ab-e7c285e87762\") " pod="openshift-multus/multus-xtwf7" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.898009 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/64ebd338-0158-44dc-90ab-e7c285e87762-etc-kubernetes\") pod \"multus-xtwf7\" (UID: \"64ebd338-0158-44dc-90ab-e7c285e87762\") " pod="openshift-multus/multus-xtwf7" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.898038 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/64ebd338-0158-44dc-90ab-e7c285e87762-host-var-lib-kubelet\") pod \"multus-xtwf7\" (UID: \"64ebd338-0158-44dc-90ab-e7c285e87762\") " pod="openshift-multus/multus-xtwf7" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.898049 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/64ebd338-0158-44dc-90ab-e7c285e87762-cnibin\") pod \"multus-xtwf7\" (UID: \"64ebd338-0158-44dc-90ab-e7c285e87762\") " pod="openshift-multus/multus-xtwf7" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.898079 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/64ebd338-0158-44dc-90ab-e7c285e87762-os-release\") pod \"multus-xtwf7\" (UID: \"64ebd338-0158-44dc-90ab-e7c285e87762\") " pod="openshift-multus/multus-xtwf7" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.898129 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/64ebd338-0158-44dc-90ab-e7c285e87762-multus-daemon-config\") pod \"multus-xtwf7\" (UID: \"64ebd338-0158-44dc-90ab-e7c285e87762\") " pod="openshift-multus/multus-xtwf7" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.898519 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/64ebd338-0158-44dc-90ab-e7c285e87762-hostroot\") pod \"multus-xtwf7\" (UID: \"64ebd338-0158-44dc-90ab-e7c285e87762\") " pod="openshift-multus/multus-xtwf7" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.898572 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/64ebd338-0158-44dc-90ab-e7c285e87762-host-var-lib-cni-multus\") pod \"multus-xtwf7\" (UID: \"64ebd338-0158-44dc-90ab-e7c285e87762\") " pod="openshift-multus/multus-xtwf7" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.898608 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/f3fb4213-144f-4b01-9d78-ff3315761bb3-rootfs\") pod \"machine-config-daemon-hm56j\" (UID: \"f3fb4213-144f-4b01-9d78-ff3315761bb3\") " pod="openshift-machine-config-operator/machine-config-daemon-hm56j" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.898699 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/57e4e912-9a98-4419-87ce-f16e178ed921-tuning-conf-dir\") pod \"multus-additional-cni-plugins-p8ptc\" (UID: \"57e4e912-9a98-4419-87ce-f16e178ed921\") " pod="openshift-multus/multus-additional-cni-plugins-p8ptc" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.898745 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/64ebd338-0158-44dc-90ab-e7c285e87762-host-run-k8s-cni-cncf-io\") pod \"multus-xtwf7\" (UID: \"64ebd338-0158-44dc-90ab-e7c285e87762\") " pod="openshift-multus/multus-xtwf7" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.898769 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f3fb4213-144f-4b01-9d78-ff3315761bb3-mcd-auth-proxy-config\") pod \"machine-config-daemon-hm56j\" (UID: \"f3fb4213-144f-4b01-9d78-ff3315761bb3\") " pod="openshift-machine-config-operator/machine-config-daemon-hm56j" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.899171 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/57e4e912-9a98-4419-87ce-f16e178ed921-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-p8ptc\" (UID: \"57e4e912-9a98-4419-87ce-f16e178ed921\") " pod="openshift-multus/multus-additional-cni-plugins-p8ptc" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.898247 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/64ebd338-0158-44dc-90ab-e7c285e87762-host-var-lib-cni-bin\") pod \"multus-xtwf7\" (UID: \"64ebd338-0158-44dc-90ab-e7c285e87762\") " pod="openshift-multus/multus-xtwf7" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.901671 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/f3fb4213-144f-4b01-9d78-ff3315761bb3-proxy-tls\") pod \"machine-config-daemon-hm56j\" (UID: \"f3fb4213-144f-4b01-9d78-ff3315761bb3\") " pod="openshift-machine-config-operator/machine-config-daemon-hm56j" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.908602 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/57e4e912-9a98-4419-87ce-f16e178ed921-cni-binary-copy\") pod \"multus-additional-cni-plugins-p8ptc\" (UID: \"57e4e912-9a98-4419-87ce-f16e178ed921\") " pod="openshift-multus/multus-additional-cni-plugins-p8ptc" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.913172 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:47Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.915857 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6cwbq\" (UniqueName: \"kubernetes.io/projected/64ebd338-0158-44dc-90ab-e7c285e87762-kube-api-access-6cwbq\") pod \"multus-xtwf7\" (UID: \"64ebd338-0158-44dc-90ab-e7c285e87762\") " pod="openshift-multus/multus-xtwf7" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.916348 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7vldr\" (UniqueName: \"kubernetes.io/projected/57e4e912-9a98-4419-87ce-f16e178ed921-kube-api-access-7vldr\") pod \"multus-additional-cni-plugins-p8ptc\" (UID: \"57e4e912-9a98-4419-87ce-f16e178ed921\") " pod="openshift-multus/multus-additional-cni-plugins-p8ptc" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.917782 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vxwbn\" (UniqueName: \"kubernetes.io/projected/f3fb4213-144f-4b01-9d78-ff3315761bb3-kube-api-access-vxwbn\") pod \"machine-config-daemon-hm56j\" (UID: \"f3fb4213-144f-4b01-9d78-ff3315761bb3\") " pod="openshift-machine-config-operator/machine-config-daemon-hm56j" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.927869 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xtwf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64ebd338-0158-44dc-90ab-e7c285e87762\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6cwbq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xtwf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:47Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.940868 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:47Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.955201 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afb1d430459aafab2d6bc2507b4b4059529401e51fdf8e9848b0d7bb8651db68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:47Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.956749 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.957176 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.957249 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.957327 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.957409 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:47Z","lastTransitionTime":"2026-03-13T14:06:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.968555 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-p8ptc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57e4e912-9a98-4419-87ce-f16e178ed921\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-p8ptc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:47Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.981577 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3fb4213-144f-4b01-9d78-ff3315761bb3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hm56j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:47Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.993473 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:47Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.019032 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-p8ptc" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.030256 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-xtwf7" Mar 13 14:06:48 crc kubenswrapper[4907]: W0313 14:06:48.039455 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod57e4e912_9a98_4419_87ce_f16e178ed921.slice/crio-079e22966c2667db80b9a401ae908cc814dd5201d0acb692a9e8a44a1136c095 WatchSource:0}: Error finding container 079e22966c2667db80b9a401ae908cc814dd5201d0acb692a9e8a44a1136c095: Status 404 returned error can't find the container with id 079e22966c2667db80b9a401ae908cc814dd5201d0acb692a9e8a44a1136c095 Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.040664 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" Mar 13 14:06:48 crc kubenswrapper[4907]: W0313 14:06:48.047580 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod64ebd338_0158_44dc_90ab_e7c285e87762.slice/crio-f25904e76871efac8ccd4826678dfa90aa051973756ce2119ad33d1abda0b3b3 WatchSource:0}: Error finding container f25904e76871efac8ccd4826678dfa90aa051973756ce2119ad33d1abda0b3b3: Status 404 returned error can't find the container with id f25904e76871efac8ccd4826678dfa90aa051973756ce2119ad33d1abda0b3b3 Mar 13 14:06:48 crc kubenswrapper[4907]: W0313 14:06:48.059069 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf3fb4213_144f_4b01_9d78_ff3315761bb3.slice/crio-7bdf72caf166aa15007523a91a3e3eae10bae5574b7406aeebd6c8d935e0be6e WatchSource:0}: Error finding container 7bdf72caf166aa15007523a91a3e3eae10bae5574b7406aeebd6c8d935e0be6e: Status 404 returned error can't find the container with id 7bdf72caf166aa15007523a91a3e3eae10bae5574b7406aeebd6c8d935e0be6e Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.059489 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.059620 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.059638 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.059654 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.059664 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:48Z","lastTransitionTime":"2026-03-13T14:06:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.067519 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-qxlx5"] Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.069254 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.071269 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.071430 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.071692 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.071825 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.071937 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.072074 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.074492 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.083362 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5d2b054-0ed8-419c-b731-f14d3598372f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4b503aa3b13268b3688d067f3e456c86afac1fda171aced0990899c3eb89b79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be40a1a33fa482eab363537afaa901127d560afbc1b5b287ef11e66391a5be4c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d1b539bbd918fc50428fc18cabcc4c8a43cc96f220b5692d67cf1c8dfbf8ed8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T14:06:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0313 14:06:13.202871 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0313 14:06:13.203175 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0313 14:06:13.204427 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2471728650/tls.crt::/tmp/serving-cert-2471728650/tls.key\\\\\\\"\\\\nI0313 14:06:13.581160 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0313 14:06:13.584012 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0313 14:06:13.584036 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0313 14:06:13.584064 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0313 14:06:13.584072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0313 14:06:13.594504 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0313 14:06:13.594891 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0313 14:06:13.595323 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0313 14:06:13.595363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0313 14:06:13.595370 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0313 14:06:13.595376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0313 14:06:13.597226 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3f2b6756078925f8562b6053b3b83798efb72288f5fd18fff680a4ab0d33feb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:48Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.094025 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94b045cf2fed56457a2ca14f489bb69470198f159f16c207fe061d034dba1555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd28f7a5c35822582b70ed41645294aaede10e28ff59f3f3b5d36e7da8237b76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:48Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.101891 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l7zg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d105034b-f041-4fb5-8df6-6b1bb7681fda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szmh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l7zg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:48Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.118565 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:48Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.132519 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:48Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.146615 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xtwf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64ebd338-0158-44dc-90ab-e7c285e87762\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6cwbq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xtwf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:48Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.160162 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:48Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.161818 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.161863 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.161876 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.161915 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.161931 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:48Z","lastTransitionTime":"2026-03-13T14:06:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.174111 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afb1d430459aafab2d6bc2507b4b4059529401e51fdf8e9848b0d7bb8651db68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:48Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.185295 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:48Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.199136 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-p8ptc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57e4e912-9a98-4419-87ce-f16e178ed921\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-p8ptc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:48Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.199505 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-log-socket\") pod \"ovnkube-node-qxlx5\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.199592 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/1006376d-4033-454e-8a23-2b67882bdd0b-env-overrides\") pod \"ovnkube-node-qxlx5\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.199736 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/1006376d-4033-454e-8a23-2b67882bdd0b-ovnkube-script-lib\") pod \"ovnkube-node-qxlx5\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.199793 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-run-openvswitch\") pod \"ovnkube-node-qxlx5\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.199815 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-systemd-units\") pod \"ovnkube-node-qxlx5\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.199846 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/1006376d-4033-454e-8a23-2b67882bdd0b-ovn-node-metrics-cert\") pod \"ovnkube-node-qxlx5\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.199867 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-host-slash\") pod \"ovnkube-node-qxlx5\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.199898 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-run-ovn\") pod \"ovnkube-node-qxlx5\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.199916 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-host-cni-bin\") pod \"ovnkube-node-qxlx5\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.199932 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-run-systemd\") pod \"ovnkube-node-qxlx5\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.199973 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-host-run-ovn-kubernetes\") pod \"ovnkube-node-qxlx5\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.199991 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-host-kubelet\") pod \"ovnkube-node-qxlx5\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.200006 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-etc-openvswitch\") pod \"ovnkube-node-qxlx5\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.200022 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tm9lw\" (UniqueName: \"kubernetes.io/projected/1006376d-4033-454e-8a23-2b67882bdd0b-kube-api-access-tm9lw\") pod \"ovnkube-node-qxlx5\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.200039 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/1006376d-4033-454e-8a23-2b67882bdd0b-ovnkube-config\") pod \"ovnkube-node-qxlx5\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.200180 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-host-run-netns\") pod \"ovnkube-node-qxlx5\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.200232 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-qxlx5\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.200251 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-var-lib-openvswitch\") pod \"ovnkube-node-qxlx5\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.200304 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-host-cni-netd\") pod \"ovnkube-node-qxlx5\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.200328 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-node-log\") pod \"ovnkube-node-qxlx5\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.210672 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3fb4213-144f-4b01-9d78-ff3315761bb3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hm56j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:48Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.213977 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"e5b4819028f34db81cc43a04e4854c30488ed876c63b8daff11105bf61546c94"} Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.215584 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-p8ptc" event={"ID":"57e4e912-9a98-4419-87ce-f16e178ed921","Type":"ContainerStarted","Data":"079e22966c2667db80b9a401ae908cc814dd5201d0acb692a9e8a44a1136c095"} Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.217129 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-xtwf7" event={"ID":"64ebd338-0158-44dc-90ab-e7c285e87762","Type":"ContainerStarted","Data":"391b8bc1e975982b172175fcf4b4509756b018cbb99c12e48f4931ff472cdae2"} Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.217155 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-xtwf7" event={"ID":"64ebd338-0158-44dc-90ab-e7c285e87762","Type":"ContainerStarted","Data":"f25904e76871efac8ccd4826678dfa90aa051973756ce2119ad33d1abda0b3b3"} Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.218493 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-l7zg7" event={"ID":"d105034b-f041-4fb5-8df6-6b1bb7681fda","Type":"ContainerStarted","Data":"21a207f917836785e320f4f79aca5d6449da2fdeec3ee0765f34d227d5617fa1"} Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.218524 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-l7zg7" event={"ID":"d105034b-f041-4fb5-8df6-6b1bb7681fda","Type":"ContainerStarted","Data":"43541ca6821c8fc2f866fd8525a0781a2c00d4fd50c8c0b176be0d71affd01db"} Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.220155 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerStarted","Data":"770c882d9e9f9b2e9bd4f135e591952fa15b3898c42826aed8a75d953c9a875b"} Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.220220 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerStarted","Data":"7bdf72caf166aa15007523a91a3e3eae10bae5574b7406aeebd6c8d935e0be6e"} Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.230811 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1006376d-4033-454e-8a23-2b67882bdd0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qxlx5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:48Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.245352 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94b045cf2fed56457a2ca14f489bb69470198f159f16c207fe061d034dba1555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd28f7a5c35822582b70ed41645294aaede10e28ff59f3f3b5d36e7da8237b76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:48Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.255778 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l7zg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d105034b-f041-4fb5-8df6-6b1bb7681fda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21a207f917836785e320f4f79aca5d6449da2fdeec3ee0765f34d227d5617fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szmh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l7zg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:48Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.264539 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.264580 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.264588 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.264606 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.264616 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:48Z","lastTransitionTime":"2026-03-13T14:06:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.271361 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5d2b054-0ed8-419c-b731-f14d3598372f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4b503aa3b13268b3688d067f3e456c86afac1fda171aced0990899c3eb89b79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be40a1a33fa482eab363537afaa901127d560afbc1b5b287ef11e66391a5be4c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d1b539bbd918fc50428fc18cabcc4c8a43cc96f220b5692d67cf1c8dfbf8ed8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T14:06:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0313 14:06:13.202871 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0313 14:06:13.203175 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0313 14:06:13.204427 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2471728650/tls.crt::/tmp/serving-cert-2471728650/tls.key\\\\\\\"\\\\nI0313 14:06:13.581160 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0313 14:06:13.584012 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0313 14:06:13.584036 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0313 14:06:13.584064 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0313 14:06:13.584072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0313 14:06:13.594504 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0313 14:06:13.594891 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0313 14:06:13.595323 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0313 14:06:13.595363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0313 14:06:13.595370 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0313 14:06:13.595376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0313 14:06:13.597226 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3f2b6756078925f8562b6053b3b83798efb72288f5fd18fff680a4ab0d33feb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:48Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.287539 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xtwf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64ebd338-0158-44dc-90ab-e7c285e87762\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://391b8bc1e975982b172175fcf4b4509756b018cbb99c12e48f4931ff472cdae2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6cwbq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xtwf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:48Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.300330 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5b4819028f34db81cc43a04e4854c30488ed876c63b8daff11105bf61546c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:48Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.300863 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-host-cni-netd\") pod \"ovnkube-node-qxlx5\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.300924 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-node-log\") pod \"ovnkube-node-qxlx5\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.300943 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-log-socket\") pod \"ovnkube-node-qxlx5\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.300971 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/1006376d-4033-454e-8a23-2b67882bdd0b-env-overrides\") pod \"ovnkube-node-qxlx5\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.300987 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/1006376d-4033-454e-8a23-2b67882bdd0b-ovnkube-script-lib\") pod \"ovnkube-node-qxlx5\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.301001 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-run-openvswitch\") pod \"ovnkube-node-qxlx5\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.301019 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-systemd-units\") pod \"ovnkube-node-qxlx5\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.301150 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/1006376d-4033-454e-8a23-2b67882bdd0b-ovn-node-metrics-cert\") pod \"ovnkube-node-qxlx5\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.301168 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-run-ovn\") pod \"ovnkube-node-qxlx5\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.301181 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-host-cni-bin\") pod \"ovnkube-node-qxlx5\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.301206 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-host-slash\") pod \"ovnkube-node-qxlx5\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.301229 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-run-systemd\") pod \"ovnkube-node-qxlx5\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.301109 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-run-openvswitch\") pod \"ovnkube-node-qxlx5\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.301254 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-host-run-ovn-kubernetes\") pod \"ovnkube-node-qxlx5\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.301533 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-etc-openvswitch\") pod \"ovnkube-node-qxlx5\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.301041 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-log-socket\") pod \"ovnkube-node-qxlx5\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.301553 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tm9lw\" (UniqueName: \"kubernetes.io/projected/1006376d-4033-454e-8a23-2b67882bdd0b-kube-api-access-tm9lw\") pod \"ovnkube-node-qxlx5\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.301637 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-host-kubelet\") pod \"ovnkube-node-qxlx5\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.301694 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/1006376d-4033-454e-8a23-2b67882bdd0b-ovnkube-config\") pod \"ovnkube-node-qxlx5\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.301751 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-host-run-netns\") pod \"ovnkube-node-qxlx5\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.301780 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-qxlx5\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.301825 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-var-lib-openvswitch\") pod \"ovnkube-node-qxlx5\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.301070 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-node-log\") pod \"ovnkube-node-qxlx5\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.301083 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-host-cni-netd\") pod \"ovnkube-node-qxlx5\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.301967 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/1006376d-4033-454e-8a23-2b67882bdd0b-ovnkube-script-lib\") pod \"ovnkube-node-qxlx5\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.301969 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/1006376d-4033-454e-8a23-2b67882bdd0b-env-overrides\") pod \"ovnkube-node-qxlx5\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.301981 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-run-systemd\") pod \"ovnkube-node-qxlx5\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.301654 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-run-ovn\") pod \"ovnkube-node-qxlx5\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.301129 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-systemd-units\") pod \"ovnkube-node-qxlx5\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.302044 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-var-lib-openvswitch\") pod \"ovnkube-node-qxlx5\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.302051 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-host-cni-bin\") pod \"ovnkube-node-qxlx5\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.302057 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-etc-openvswitch\") pod \"ovnkube-node-qxlx5\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.302043 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-qxlx5\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.302018 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-host-run-netns\") pod \"ovnkube-node-qxlx5\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.302040 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-host-slash\") pod \"ovnkube-node-qxlx5\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.302072 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-host-run-ovn-kubernetes\") pod \"ovnkube-node-qxlx5\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.302063 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-host-kubelet\") pod \"ovnkube-node-qxlx5\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.302608 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/1006376d-4033-454e-8a23-2b67882bdd0b-ovnkube-config\") pod \"ovnkube-node-qxlx5\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.304471 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/1006376d-4033-454e-8a23-2b67882bdd0b-ovn-node-metrics-cert\") pod \"ovnkube-node-qxlx5\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.315500 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:48Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.318820 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tm9lw\" (UniqueName: \"kubernetes.io/projected/1006376d-4033-454e-8a23-2b67882bdd0b-kube-api-access-tm9lw\") pod \"ovnkube-node-qxlx5\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.327087 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afb1d430459aafab2d6bc2507b4b4059529401e51fdf8e9848b0d7bb8651db68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:48Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.338566 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:48Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.349340 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:48Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.364570 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-p8ptc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57e4e912-9a98-4419-87ce-f16e178ed921\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-p8ptc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:48Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.367059 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.367121 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.367138 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.367163 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.367181 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:48Z","lastTransitionTime":"2026-03-13T14:06:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.378757 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3fb4213-144f-4b01-9d78-ff3315761bb3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hm56j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:48Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.384634 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:48 crc kubenswrapper[4907]: W0313 14:06:48.397852 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1006376d_4033_454e_8a23_2b67882bdd0b.slice/crio-4a8bfe852be3826bb96e2a956b4a0cd7eaff7036e060420d181a2514210039c0 WatchSource:0}: Error finding container 4a8bfe852be3826bb96e2a956b4a0cd7eaff7036e060420d181a2514210039c0: Status 404 returned error can't find the container with id 4a8bfe852be3826bb96e2a956b4a0cd7eaff7036e060420d181a2514210039c0 Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.402663 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1006376d-4033-454e-8a23-2b67882bdd0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qxlx5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:48Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.469725 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.469770 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.469782 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.469802 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.469814 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:48Z","lastTransitionTime":"2026-03-13T14:06:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.572569 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.572915 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.572925 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.572945 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.572955 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:48Z","lastTransitionTime":"2026-03-13T14:06:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.604094 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:06:48 crc kubenswrapper[4907]: E0313 14:06:48.604282 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:07:04.604249229 +0000 UTC m=+123.504036918 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.676277 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.676320 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.676330 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.676348 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.676358 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:48Z","lastTransitionTime":"2026-03-13T14:06:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.705268 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.705312 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.705343 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.705365 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:06:48 crc kubenswrapper[4907]: E0313 14:06:48.705500 4907 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 13 14:06:48 crc kubenswrapper[4907]: E0313 14:06:48.705516 4907 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 13 14:06:48 crc kubenswrapper[4907]: E0313 14:06:48.705528 4907 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 14:06:48 crc kubenswrapper[4907]: E0313 14:06:48.705573 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-13 14:07:04.705557956 +0000 UTC m=+123.605345645 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 14:06:48 crc kubenswrapper[4907]: E0313 14:06:48.705868 4907 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 13 14:06:48 crc kubenswrapper[4907]: E0313 14:06:48.705984 4907 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 13 14:06:48 crc kubenswrapper[4907]: E0313 14:06:48.706031 4907 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 13 14:06:48 crc kubenswrapper[4907]: E0313 14:06:48.706048 4907 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 14:06:48 crc kubenswrapper[4907]: E0313 14:06:48.706112 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-13 14:07:04.706089972 +0000 UTC m=+123.605877671 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 14:06:48 crc kubenswrapper[4907]: E0313 14:06:48.705929 4907 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 13 14:06:48 crc kubenswrapper[4907]: E0313 14:06:48.706168 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-13 14:07:04.706158684 +0000 UTC m=+123.605946393 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 13 14:06:48 crc kubenswrapper[4907]: E0313 14:06:48.706277 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-13 14:07:04.706252997 +0000 UTC m=+123.606040686 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.779475 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.779711 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.779807 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.779914 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.780009 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:48Z","lastTransitionTime":"2026-03-13T14:06:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.781689 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.781716 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.781689 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:06:48 crc kubenswrapper[4907]: E0313 14:06:48.781800 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:06:48 crc kubenswrapper[4907]: E0313 14:06:48.781899 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:06:48 crc kubenswrapper[4907]: E0313 14:06:48.781967 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.883386 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.883421 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.883432 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.883448 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.883458 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:48Z","lastTransitionTime":"2026-03-13T14:06:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.986740 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.987068 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.987165 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.987433 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.987511 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:48Z","lastTransitionTime":"2026-03-13T14:06:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.091471 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.091827 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.092042 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.092342 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.092556 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:49Z","lastTransitionTime":"2026-03-13T14:06:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.195557 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.195623 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.195642 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.195668 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.195687 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:49Z","lastTransitionTime":"2026-03-13T14:06:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.227261 4907 generic.go:334] "Generic (PLEG): container finished" podID="57e4e912-9a98-4419-87ce-f16e178ed921" containerID="05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f" exitCode=0 Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.227353 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-p8ptc" event={"ID":"57e4e912-9a98-4419-87ce-f16e178ed921","Type":"ContainerDied","Data":"05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f"} Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.229266 4907 generic.go:334] "Generic (PLEG): container finished" podID="1006376d-4033-454e-8a23-2b67882bdd0b" containerID="858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279" exitCode=0 Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.229428 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" event={"ID":"1006376d-4033-454e-8a23-2b67882bdd0b","Type":"ContainerDied","Data":"858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279"} Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.229565 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" event={"ID":"1006376d-4033-454e-8a23-2b67882bdd0b","Type":"ContainerStarted","Data":"4a8bfe852be3826bb96e2a956b4a0cd7eaff7036e060420d181a2514210039c0"} Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.234525 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerStarted","Data":"ba1140a2462713c2401b47a4cf1f231bc9d9ae7867a7d0b1b597214a1f7efe07"} Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.253605 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5b4819028f34db81cc43a04e4854c30488ed876c63b8daff11105bf61546c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:49Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.269094 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:49Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.281647 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xtwf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64ebd338-0158-44dc-90ab-e7c285e87762\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://391b8bc1e975982b172175fcf4b4509756b018cbb99c12e48f4931ff472cdae2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6cwbq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xtwf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:49Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.295309 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:49Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.297800 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.297992 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.298017 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.298033 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.298043 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:49Z","lastTransitionTime":"2026-03-13T14:06:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.308662 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afb1d430459aafab2d6bc2507b4b4059529401e51fdf8e9848b0d7bb8651db68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:49Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.320629 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:49Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.335975 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-p8ptc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57e4e912-9a98-4419-87ce-f16e178ed921\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-p8ptc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:49Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.347123 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3fb4213-144f-4b01-9d78-ff3315761bb3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hm56j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:49Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.370295 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1006376d-4033-454e-8a23-2b67882bdd0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qxlx5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:49Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.385534 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5d2b054-0ed8-419c-b731-f14d3598372f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4b503aa3b13268b3688d067f3e456c86afac1fda171aced0990899c3eb89b79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be40a1a33fa482eab363537afaa901127d560afbc1b5b287ef11e66391a5be4c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d1b539bbd918fc50428fc18cabcc4c8a43cc96f220b5692d67cf1c8dfbf8ed8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T14:06:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0313 14:06:13.202871 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0313 14:06:13.203175 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0313 14:06:13.204427 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2471728650/tls.crt::/tmp/serving-cert-2471728650/tls.key\\\\\\\"\\\\nI0313 14:06:13.581160 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0313 14:06:13.584012 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0313 14:06:13.584036 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0313 14:06:13.584064 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0313 14:06:13.584072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0313 14:06:13.594504 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0313 14:06:13.594891 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0313 14:06:13.595323 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0313 14:06:13.595363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0313 14:06:13.595370 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0313 14:06:13.595376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0313 14:06:13.597226 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3f2b6756078925f8562b6053b3b83798efb72288f5fd18fff680a4ab0d33feb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:49Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.402687 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.402744 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.402756 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.402776 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.402790 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:49Z","lastTransitionTime":"2026-03-13T14:06:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.406204 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94b045cf2fed56457a2ca14f489bb69470198f159f16c207fe061d034dba1555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd28f7a5c35822582b70ed41645294aaede10e28ff59f3f3b5d36e7da8237b76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:49Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.414622 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l7zg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d105034b-f041-4fb5-8df6-6b1bb7681fda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21a207f917836785e320f4f79aca5d6449da2fdeec3ee0765f34d227d5617fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szmh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l7zg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:49Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.424663 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:49Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.434075 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afb1d430459aafab2d6bc2507b4b4059529401e51fdf8e9848b0d7bb8651db68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:49Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.444775 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-p8ptc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57e4e912-9a98-4419-87ce-f16e178ed921\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-p8ptc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:49Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.459774 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3fb4213-144f-4b01-9d78-ff3315761bb3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba1140a2462713c2401b47a4cf1f231bc9d9ae7867a7d0b1b597214a1f7efe07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://770c882d9e9f9b2e9bd4f135e591952fa15b3898c42826aed8a75d953c9a875b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hm56j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:49Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.494305 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1006376d-4033-454e-8a23-2b67882bdd0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qxlx5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:49Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.505438 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.505479 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.505494 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.505514 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.505525 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:49Z","lastTransitionTime":"2026-03-13T14:06:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.521457 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:49Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.536763 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l7zg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d105034b-f041-4fb5-8df6-6b1bb7681fda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21a207f917836785e320f4f79aca5d6449da2fdeec3ee0765f34d227d5617fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szmh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l7zg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:49Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.552084 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5d2b054-0ed8-419c-b731-f14d3598372f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4b503aa3b13268b3688d067f3e456c86afac1fda171aced0990899c3eb89b79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be40a1a33fa482eab363537afaa901127d560afbc1b5b287ef11e66391a5be4c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d1b539bbd918fc50428fc18cabcc4c8a43cc96f220b5692d67cf1c8dfbf8ed8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T14:06:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0313 14:06:13.202871 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0313 14:06:13.203175 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0313 14:06:13.204427 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2471728650/tls.crt::/tmp/serving-cert-2471728650/tls.key\\\\\\\"\\\\nI0313 14:06:13.581160 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0313 14:06:13.584012 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0313 14:06:13.584036 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0313 14:06:13.584064 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0313 14:06:13.584072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0313 14:06:13.594504 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0313 14:06:13.594891 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0313 14:06:13.595323 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0313 14:06:13.595363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0313 14:06:13.595370 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0313 14:06:13.595376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0313 14:06:13.597226 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3f2b6756078925f8562b6053b3b83798efb72288f5fd18fff680a4ab0d33feb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:49Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.565365 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94b045cf2fed56457a2ca14f489bb69470198f159f16c207fe061d034dba1555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd28f7a5c35822582b70ed41645294aaede10e28ff59f3f3b5d36e7da8237b76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:49Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.581227 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5b4819028f34db81cc43a04e4854c30488ed876c63b8daff11105bf61546c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:49Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.595182 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:49Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.608308 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.608352 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.608365 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.608387 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.608402 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:49Z","lastTransitionTime":"2026-03-13T14:06:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.609545 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xtwf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64ebd338-0158-44dc-90ab-e7c285e87762\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://391b8bc1e975982b172175fcf4b4509756b018cbb99c12e48f4931ff472cdae2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6cwbq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xtwf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:49Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.710187 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.710226 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.710239 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.710258 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.710271 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:49Z","lastTransitionTime":"2026-03-13T14:06:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.816194 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.816599 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.816610 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.816629 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.816643 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:49Z","lastTransitionTime":"2026-03-13T14:06:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.919498 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.919548 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.919557 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.919571 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.919584 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:49Z","lastTransitionTime":"2026-03-13T14:06:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.022274 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.022321 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.022344 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.022359 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.022372 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:50Z","lastTransitionTime":"2026-03-13T14:06:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.124850 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.124900 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.124912 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.124929 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.124941 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:50Z","lastTransitionTime":"2026-03-13T14:06:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.227224 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.227523 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.227535 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.227549 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.227558 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:50Z","lastTransitionTime":"2026-03-13T14:06:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.241457 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" event={"ID":"1006376d-4033-454e-8a23-2b67882bdd0b","Type":"ContainerStarted","Data":"0fff87d8ea572615eb1329164f9e536e6ee45cd00a9e71be0aa8a9b3ed2190e2"} Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.241499 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" event={"ID":"1006376d-4033-454e-8a23-2b67882bdd0b","Type":"ContainerStarted","Data":"ccef92e41501fab91189581d7e8875515c02d637426b329d2fb80a356c75b637"} Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.241507 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" event={"ID":"1006376d-4033-454e-8a23-2b67882bdd0b","Type":"ContainerStarted","Data":"bf6a3dbaa8dda61f461a906538824c8946d36dc07d4fb95bf4d8420ad0105600"} Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.241516 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" event={"ID":"1006376d-4033-454e-8a23-2b67882bdd0b","Type":"ContainerStarted","Data":"3c165053fd9933444b372ff40efe33c0e193a7f39e6f531a4f370e971eb41e6d"} Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.241524 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" event={"ID":"1006376d-4033-454e-8a23-2b67882bdd0b","Type":"ContainerStarted","Data":"fea41c11bebd2d9d622ed4b8bc50e03a106dadea6ac102cd60512ea2a1f8b8b0"} Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.243036 4907 generic.go:334] "Generic (PLEG): container finished" podID="57e4e912-9a98-4419-87ce-f16e178ed921" containerID="e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302" exitCode=0 Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.243778 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-p8ptc" event={"ID":"57e4e912-9a98-4419-87ce-f16e178ed921","Type":"ContainerDied","Data":"e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302"} Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.257184 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:50Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.273017 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-p8ptc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57e4e912-9a98-4419-87ce-f16e178ed921\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-p8ptc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:50Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.286598 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3fb4213-144f-4b01-9d78-ff3315761bb3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba1140a2462713c2401b47a4cf1f231bc9d9ae7867a7d0b1b597214a1f7efe07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://770c882d9e9f9b2e9bd4f135e591952fa15b3898c42826aed8a75d953c9a875b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hm56j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:50Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.307645 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1006376d-4033-454e-8a23-2b67882bdd0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qxlx5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:50Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.322683 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5d2b054-0ed8-419c-b731-f14d3598372f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4b503aa3b13268b3688d067f3e456c86afac1fda171aced0990899c3eb89b79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be40a1a33fa482eab363537afaa901127d560afbc1b5b287ef11e66391a5be4c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d1b539bbd918fc50428fc18cabcc4c8a43cc96f220b5692d67cf1c8dfbf8ed8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T14:06:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0313 14:06:13.202871 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0313 14:06:13.203175 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0313 14:06:13.204427 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2471728650/tls.crt::/tmp/serving-cert-2471728650/tls.key\\\\\\\"\\\\nI0313 14:06:13.581160 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0313 14:06:13.584012 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0313 14:06:13.584036 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0313 14:06:13.584064 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0313 14:06:13.584072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0313 14:06:13.594504 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0313 14:06:13.594891 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0313 14:06:13.595323 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0313 14:06:13.595363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0313 14:06:13.595370 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0313 14:06:13.595376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0313 14:06:13.597226 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3f2b6756078925f8562b6053b3b83798efb72288f5fd18fff680a4ab0d33feb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:50Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.330629 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.330682 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.330691 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.330705 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.330713 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:50Z","lastTransitionTime":"2026-03-13T14:06:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.334055 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94b045cf2fed56457a2ca14f489bb69470198f159f16c207fe061d034dba1555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd28f7a5c35822582b70ed41645294aaede10e28ff59f3f3b5d36e7da8237b76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:50Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.346156 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l7zg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d105034b-f041-4fb5-8df6-6b1bb7681fda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21a207f917836785e320f4f79aca5d6449da2fdeec3ee0765f34d227d5617fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szmh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l7zg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:50Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.357993 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5b4819028f34db81cc43a04e4854c30488ed876c63b8daff11105bf61546c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:50Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.368529 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:50Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.380633 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xtwf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64ebd338-0158-44dc-90ab-e7c285e87762\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://391b8bc1e975982b172175fcf4b4509756b018cbb99c12e48f4931ff472cdae2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6cwbq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xtwf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:50Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.391943 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:50Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.405468 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afb1d430459aafab2d6bc2507b4b4059529401e51fdf8e9848b0d7bb8651db68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:50Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.434475 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.434523 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.434534 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.434550 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.434560 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:50Z","lastTransitionTime":"2026-03-13T14:06:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.536599 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.536632 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.536640 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.536656 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.536666 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:50Z","lastTransitionTime":"2026-03-13T14:06:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.638995 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.639040 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.639054 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.639073 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.639091 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:50Z","lastTransitionTime":"2026-03-13T14:06:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.741214 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.741310 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.741328 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.741356 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.741379 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:50Z","lastTransitionTime":"2026-03-13T14:06:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.782407 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.782508 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.782421 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:06:50 crc kubenswrapper[4907]: E0313 14:06:50.782623 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:06:50 crc kubenswrapper[4907]: E0313 14:06:50.782780 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:06:50 crc kubenswrapper[4907]: E0313 14:06:50.782911 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.843756 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.843799 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.843811 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.843831 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.843844 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:50Z","lastTransitionTime":"2026-03-13T14:06:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.946965 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.947033 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.947052 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.947079 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.947099 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:50Z","lastTransitionTime":"2026-03-13T14:06:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.052802 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.052876 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.052931 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.052959 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.052977 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:51Z","lastTransitionTime":"2026-03-13T14:06:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.156307 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.156368 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.156386 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.156414 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.156432 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:51Z","lastTransitionTime":"2026-03-13T14:06:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.250743 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" event={"ID":"1006376d-4033-454e-8a23-2b67882bdd0b","Type":"ContainerStarted","Data":"a0ee0049aa4a6d49a6bcf7d347183d9cc8b85171f988662803a005e0cdd97395"} Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.254072 4907 generic.go:334] "Generic (PLEG): container finished" podID="57e4e912-9a98-4419-87ce-f16e178ed921" containerID="f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45" exitCode=0 Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.254159 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-p8ptc" event={"ID":"57e4e912-9a98-4419-87ce-f16e178ed921","Type":"ContainerDied","Data":"f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45"} Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.260171 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.260212 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.260226 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.260243 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.260255 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:51Z","lastTransitionTime":"2026-03-13T14:06:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.270646 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94b045cf2fed56457a2ca14f489bb69470198f159f16c207fe061d034dba1555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd28f7a5c35822582b70ed41645294aaede10e28ff59f3f3b5d36e7da8237b76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:51Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.283827 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l7zg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d105034b-f041-4fb5-8df6-6b1bb7681fda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21a207f917836785e320f4f79aca5d6449da2fdeec3ee0765f34d227d5617fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szmh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l7zg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:51Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.302345 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5d2b054-0ed8-419c-b731-f14d3598372f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4b503aa3b13268b3688d067f3e456c86afac1fda171aced0990899c3eb89b79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be40a1a33fa482eab363537afaa901127d560afbc1b5b287ef11e66391a5be4c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d1b539bbd918fc50428fc18cabcc4c8a43cc96f220b5692d67cf1c8dfbf8ed8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T14:06:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0313 14:06:13.202871 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0313 14:06:13.203175 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0313 14:06:13.204427 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2471728650/tls.crt::/tmp/serving-cert-2471728650/tls.key\\\\\\\"\\\\nI0313 14:06:13.581160 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0313 14:06:13.584012 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0313 14:06:13.584036 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0313 14:06:13.584064 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0313 14:06:13.584072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0313 14:06:13.594504 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0313 14:06:13.594891 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0313 14:06:13.595323 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0313 14:06:13.595363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0313 14:06:13.595370 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0313 14:06:13.595376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0313 14:06:13.597226 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3f2b6756078925f8562b6053b3b83798efb72288f5fd18fff680a4ab0d33feb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:51Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.322962 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xtwf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64ebd338-0158-44dc-90ab-e7c285e87762\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://391b8bc1e975982b172175fcf4b4509756b018cbb99c12e48f4931ff472cdae2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6cwbq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xtwf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:51Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.343381 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5b4819028f34db81cc43a04e4854c30488ed876c63b8daff11105bf61546c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:51Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.358931 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:51Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.363635 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.363674 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.363687 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.363706 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.363719 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:51Z","lastTransitionTime":"2026-03-13T14:06:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.374264 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afb1d430459aafab2d6bc2507b4b4059529401e51fdf8e9848b0d7bb8651db68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:51Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.389273 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:51Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.404785 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:51Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.422236 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-p8ptc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57e4e912-9a98-4419-87ce-f16e178ed921\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-p8ptc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:51Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.441427 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3fb4213-144f-4b01-9d78-ff3315761bb3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba1140a2462713c2401b47a4cf1f231bc9d9ae7867a7d0b1b597214a1f7efe07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://770c882d9e9f9b2e9bd4f135e591952fa15b3898c42826aed8a75d953c9a875b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hm56j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:51Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.458873 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1006376d-4033-454e-8a23-2b67882bdd0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qxlx5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:51Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.468642 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.468708 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.468727 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.468755 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.468772 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:51Z","lastTransitionTime":"2026-03-13T14:06:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.571097 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.571136 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.571147 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.571166 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.571178 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:51Z","lastTransitionTime":"2026-03-13T14:06:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.673150 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.673179 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.673187 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.673201 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.673209 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:51Z","lastTransitionTime":"2026-03-13T14:06:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.776284 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.776339 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.776356 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.776377 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.776394 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:51Z","lastTransitionTime":"2026-03-13T14:06:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.801712 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5d2b054-0ed8-419c-b731-f14d3598372f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4b503aa3b13268b3688d067f3e456c86afac1fda171aced0990899c3eb89b79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be40a1a33fa482eab363537afaa901127d560afbc1b5b287ef11e66391a5be4c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d1b539bbd918fc50428fc18cabcc4c8a43cc96f220b5692d67cf1c8dfbf8ed8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T14:06:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0313 14:06:13.202871 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0313 14:06:13.203175 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0313 14:06:13.204427 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2471728650/tls.crt::/tmp/serving-cert-2471728650/tls.key\\\\\\\"\\\\nI0313 14:06:13.581160 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0313 14:06:13.584012 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0313 14:06:13.584036 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0313 14:06:13.584064 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0313 14:06:13.584072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0313 14:06:13.594504 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0313 14:06:13.594891 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0313 14:06:13.595323 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0313 14:06:13.595363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0313 14:06:13.595370 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0313 14:06:13.595376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0313 14:06:13.597226 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3f2b6756078925f8562b6053b3b83798efb72288f5fd18fff680a4ab0d33feb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:51Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.817400 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94b045cf2fed56457a2ca14f489bb69470198f159f16c207fe061d034dba1555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd28f7a5c35822582b70ed41645294aaede10e28ff59f3f3b5d36e7da8237b76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:51Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.837495 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l7zg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d105034b-f041-4fb5-8df6-6b1bb7681fda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21a207f917836785e320f4f79aca5d6449da2fdeec3ee0765f34d227d5617fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szmh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l7zg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:51Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.860627 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5b4819028f34db81cc43a04e4854c30488ed876c63b8daff11105bf61546c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:51Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.875078 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:51Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.879545 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.879601 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.879626 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.879655 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.879678 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:51Z","lastTransitionTime":"2026-03-13T14:06:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.897813 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xtwf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64ebd338-0158-44dc-90ab-e7c285e87762\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://391b8bc1e975982b172175fcf4b4509756b018cbb99c12e48f4931ff472cdae2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6cwbq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xtwf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:51Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.915609 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:51Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.934358 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afb1d430459aafab2d6bc2507b4b4059529401e51fdf8e9848b0d7bb8651db68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:51Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.950720 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:51Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.973145 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-p8ptc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57e4e912-9a98-4419-87ce-f16e178ed921\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-p8ptc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:51Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.982765 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.982823 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.982841 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.982868 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.982913 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:51Z","lastTransitionTime":"2026-03-13T14:06:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.988826 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3fb4213-144f-4b01-9d78-ff3315761bb3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba1140a2462713c2401b47a4cf1f231bc9d9ae7867a7d0b1b597214a1f7efe07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://770c882d9e9f9b2e9bd4f135e591952fa15b3898c42826aed8a75d953c9a875b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hm56j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:51Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.011410 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1006376d-4033-454e-8a23-2b67882bdd0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qxlx5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:52Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.085586 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.085647 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.085664 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.085689 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.085757 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:52Z","lastTransitionTime":"2026-03-13T14:06:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.188842 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.188935 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.188956 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.188984 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.189002 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:52Z","lastTransitionTime":"2026-03-13T14:06:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.260104 4907 generic.go:334] "Generic (PLEG): container finished" podID="57e4e912-9a98-4419-87ce-f16e178ed921" containerID="5bdef68f401b5ba2c7cf61551eba416e74f295ab2281433a5115334d28094818" exitCode=0 Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.260156 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-p8ptc" event={"ID":"57e4e912-9a98-4419-87ce-f16e178ed921","Type":"ContainerDied","Data":"5bdef68f401b5ba2c7cf61551eba416e74f295ab2281433a5115334d28094818"} Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.272726 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5d2b054-0ed8-419c-b731-f14d3598372f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4b503aa3b13268b3688d067f3e456c86afac1fda171aced0990899c3eb89b79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be40a1a33fa482eab363537afaa901127d560afbc1b5b287ef11e66391a5be4c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d1b539bbd918fc50428fc18cabcc4c8a43cc96f220b5692d67cf1c8dfbf8ed8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T14:06:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0313 14:06:13.202871 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0313 14:06:13.203175 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0313 14:06:13.204427 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2471728650/tls.crt::/tmp/serving-cert-2471728650/tls.key\\\\\\\"\\\\nI0313 14:06:13.581160 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0313 14:06:13.584012 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0313 14:06:13.584036 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0313 14:06:13.584064 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0313 14:06:13.584072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0313 14:06:13.594504 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0313 14:06:13.594891 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0313 14:06:13.595323 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0313 14:06:13.595363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0313 14:06:13.595370 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0313 14:06:13.595376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0313 14:06:13.597226 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3f2b6756078925f8562b6053b3b83798efb72288f5fd18fff680a4ab0d33feb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:52Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.291638 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.291704 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.291722 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.291569 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94b045cf2fed56457a2ca14f489bb69470198f159f16c207fe061d034dba1555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd28f7a5c35822582b70ed41645294aaede10e28ff59f3f3b5d36e7da8237b76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:52Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.291752 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.291960 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:52Z","lastTransitionTime":"2026-03-13T14:06:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.303943 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l7zg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d105034b-f041-4fb5-8df6-6b1bb7681fda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21a207f917836785e320f4f79aca5d6449da2fdeec3ee0765f34d227d5617fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szmh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l7zg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:52Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.321257 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5b4819028f34db81cc43a04e4854c30488ed876c63b8daff11105bf61546c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:52Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.331990 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:52Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.345509 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xtwf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64ebd338-0158-44dc-90ab-e7c285e87762\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://391b8bc1e975982b172175fcf4b4509756b018cbb99c12e48f4931ff472cdae2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6cwbq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xtwf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:52Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.361131 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:52Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.372407 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afb1d430459aafab2d6bc2507b4b4059529401e51fdf8e9848b0d7bb8651db68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:52Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.386176 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:52Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.394557 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.394603 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.394613 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.394632 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.394644 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:52Z","lastTransitionTime":"2026-03-13T14:06:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.403808 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-p8ptc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57e4e912-9a98-4419-87ce-f16e178ed921\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bdef68f401b5ba2c7cf61551eba416e74f295ab2281433a5115334d28094818\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bdef68f401b5ba2c7cf61551eba416e74f295ab2281433a5115334d28094818\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-p8ptc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:52Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.414401 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3fb4213-144f-4b01-9d78-ff3315761bb3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba1140a2462713c2401b47a4cf1f231bc9d9ae7867a7d0b1b597214a1f7efe07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://770c882d9e9f9b2e9bd4f135e591952fa15b3898c42826aed8a75d953c9a875b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hm56j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:52Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.433192 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1006376d-4033-454e-8a23-2b67882bdd0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qxlx5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:52Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.496519 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.496813 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.496821 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.496835 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.496844 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:52Z","lastTransitionTime":"2026-03-13T14:06:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.598647 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.598680 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.598689 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.598705 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.598715 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:52Z","lastTransitionTime":"2026-03-13T14:06:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.701768 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.701812 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.701824 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.701841 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.701854 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:52Z","lastTransitionTime":"2026-03-13T14:06:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.782324 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.782343 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:06:52 crc kubenswrapper[4907]: E0313 14:06:52.782538 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.782357 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:06:52 crc kubenswrapper[4907]: E0313 14:06:52.782739 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:06:52 crc kubenswrapper[4907]: E0313 14:06:52.782902 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.805752 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.805811 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.805828 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.805856 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.805875 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:52Z","lastTransitionTime":"2026-03-13T14:06:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.908952 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.909007 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.909024 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.909057 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.909075 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:52Z","lastTransitionTime":"2026-03-13T14:06:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.012217 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.012265 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.012277 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.012296 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.012307 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:53Z","lastTransitionTime":"2026-03-13T14:06:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.115488 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.115537 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.115554 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.115577 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.115595 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:53Z","lastTransitionTime":"2026-03-13T14:06:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.218646 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.218694 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.218708 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.218726 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.218739 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:53Z","lastTransitionTime":"2026-03-13T14:06:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.275687 4907 generic.go:334] "Generic (PLEG): container finished" podID="57e4e912-9a98-4419-87ce-f16e178ed921" containerID="3f8d1f40744d609fd964572ef5e323f0283d2582e98745602754165818f17847" exitCode=0 Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.275805 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-p8ptc" event={"ID":"57e4e912-9a98-4419-87ce-f16e178ed921","Type":"ContainerDied","Data":"3f8d1f40744d609fd964572ef5e323f0283d2582e98745602754165818f17847"} Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.284329 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" event={"ID":"1006376d-4033-454e-8a23-2b67882bdd0b","Type":"ContainerStarted","Data":"4d60b5b774f8dd4cee16b9784f80d124640199a85331a2940772029be4884751"} Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.293177 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5d2b054-0ed8-419c-b731-f14d3598372f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4b503aa3b13268b3688d067f3e456c86afac1fda171aced0990899c3eb89b79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be40a1a33fa482eab363537afaa901127d560afbc1b5b287ef11e66391a5be4c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d1b539bbd918fc50428fc18cabcc4c8a43cc96f220b5692d67cf1c8dfbf8ed8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T14:06:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0313 14:06:13.202871 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0313 14:06:13.203175 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0313 14:06:13.204427 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2471728650/tls.crt::/tmp/serving-cert-2471728650/tls.key\\\\\\\"\\\\nI0313 14:06:13.581160 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0313 14:06:13.584012 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0313 14:06:13.584036 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0313 14:06:13.584064 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0313 14:06:13.584072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0313 14:06:13.594504 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0313 14:06:13.594891 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0313 14:06:13.595323 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0313 14:06:13.595363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0313 14:06:13.595370 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0313 14:06:13.595376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0313 14:06:13.597226 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3f2b6756078925f8562b6053b3b83798efb72288f5fd18fff680a4ab0d33feb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:53Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.307039 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94b045cf2fed56457a2ca14f489bb69470198f159f16c207fe061d034dba1555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd28f7a5c35822582b70ed41645294aaede10e28ff59f3f3b5d36e7da8237b76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:53Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.321541 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.321597 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.321615 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.321641 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.321661 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:53Z","lastTransitionTime":"2026-03-13T14:06:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.333700 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l7zg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d105034b-f041-4fb5-8df6-6b1bb7681fda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21a207f917836785e320f4f79aca5d6449da2fdeec3ee0765f34d227d5617fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szmh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l7zg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:53Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.354248 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:53Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.371571 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xtwf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64ebd338-0158-44dc-90ab-e7c285e87762\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://391b8bc1e975982b172175fcf4b4509756b018cbb99c12e48f4931ff472cdae2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6cwbq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xtwf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:53Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.399353 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5b4819028f34db81cc43a04e4854c30488ed876c63b8daff11105bf61546c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:53Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.438185 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.438237 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.438254 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.438275 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.438290 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:53Z","lastTransitionTime":"2026-03-13T14:06:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.450953 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:53Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.467980 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afb1d430459aafab2d6bc2507b4b4059529401e51fdf8e9848b0d7bb8651db68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:53Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.483327 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:53Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.498139 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-p8ptc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57e4e912-9a98-4419-87ce-f16e178ed921\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bdef68f401b5ba2c7cf61551eba416e74f295ab2281433a5115334d28094818\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bdef68f401b5ba2c7cf61551eba416e74f295ab2281433a5115334d28094818\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f8d1f40744d609fd964572ef5e323f0283d2582e98745602754165818f17847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f8d1f40744d609fd964572ef5e323f0283d2582e98745602754165818f17847\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-p8ptc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:53Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.510611 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3fb4213-144f-4b01-9d78-ff3315761bb3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba1140a2462713c2401b47a4cf1f231bc9d9ae7867a7d0b1b597214a1f7efe07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://770c882d9e9f9b2e9bd4f135e591952fa15b3898c42826aed8a75d953c9a875b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hm56j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:53Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.539864 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1006376d-4033-454e-8a23-2b67882bdd0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qxlx5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:53Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.543970 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.543991 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.543999 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.544013 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.544021 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:53Z","lastTransitionTime":"2026-03-13T14:06:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.648018 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.648067 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.648076 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.648102 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.648116 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:53Z","lastTransitionTime":"2026-03-13T14:06:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.751405 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.751535 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.751598 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.751630 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.751682 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:53Z","lastTransitionTime":"2026-03-13T14:06:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.854843 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.855214 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.855362 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.855469 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.855553 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:53Z","lastTransitionTime":"2026-03-13T14:06:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.958801 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.959457 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.959479 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.959518 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.959538 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:53Z","lastTransitionTime":"2026-03-13T14:06:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.063743 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.063824 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.063846 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.063873 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.063939 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:54Z","lastTransitionTime":"2026-03-13T14:06:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.103917 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-745zx"] Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.104507 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-745zx" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.108826 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.109029 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.109241 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.109452 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.128351 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:54Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.147359 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afb1d430459aafab2d6bc2507b4b4059529401e51fdf8e9848b0d7bb8651db68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:54Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.163856 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-745zx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58e11265-2fd9-4b9b-926c-3343bb84ee53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mnz7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-745zx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:54Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.167277 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.167447 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.167532 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.167631 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.167717 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:54Z","lastTransitionTime":"2026-03-13T14:06:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.176276 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/58e11265-2fd9-4b9b-926c-3343bb84ee53-serviceca\") pod \"node-ca-745zx\" (UID: \"58e11265-2fd9-4b9b-926c-3343bb84ee53\") " pod="openshift-image-registry/node-ca-745zx" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.176387 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/58e11265-2fd9-4b9b-926c-3343bb84ee53-host\") pod \"node-ca-745zx\" (UID: \"58e11265-2fd9-4b9b-926c-3343bb84ee53\") " pod="openshift-image-registry/node-ca-745zx" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.176453 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mnz7b\" (UniqueName: \"kubernetes.io/projected/58e11265-2fd9-4b9b-926c-3343bb84ee53-kube-api-access-mnz7b\") pod \"node-ca-745zx\" (UID: \"58e11265-2fd9-4b9b-926c-3343bb84ee53\") " pod="openshift-image-registry/node-ca-745zx" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.195427 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1006376d-4033-454e-8a23-2b67882bdd0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qxlx5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:54Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.214833 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:54Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.236525 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-p8ptc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57e4e912-9a98-4419-87ce-f16e178ed921\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bdef68f401b5ba2c7cf61551eba416e74f295ab2281433a5115334d28094818\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bdef68f401b5ba2c7cf61551eba416e74f295ab2281433a5115334d28094818\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f8d1f40744d609fd964572ef5e323f0283d2582e98745602754165818f17847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f8d1f40744d609fd964572ef5e323f0283d2582e98745602754165818f17847\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-p8ptc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:54Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.252475 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3fb4213-144f-4b01-9d78-ff3315761bb3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba1140a2462713c2401b47a4cf1f231bc9d9ae7867a7d0b1b597214a1f7efe07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://770c882d9e9f9b2e9bd4f135e591952fa15b3898c42826aed8a75d953c9a875b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hm56j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:54Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.270537 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5d2b054-0ed8-419c-b731-f14d3598372f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4b503aa3b13268b3688d067f3e456c86afac1fda171aced0990899c3eb89b79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be40a1a33fa482eab363537afaa901127d560afbc1b5b287ef11e66391a5be4c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d1b539bbd918fc50428fc18cabcc4c8a43cc96f220b5692d67cf1c8dfbf8ed8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T14:06:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0313 14:06:13.202871 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0313 14:06:13.203175 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0313 14:06:13.204427 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2471728650/tls.crt::/tmp/serving-cert-2471728650/tls.key\\\\\\\"\\\\nI0313 14:06:13.581160 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0313 14:06:13.584012 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0313 14:06:13.584036 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0313 14:06:13.584064 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0313 14:06:13.584072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0313 14:06:13.594504 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0313 14:06:13.594891 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0313 14:06:13.595323 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0313 14:06:13.595363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0313 14:06:13.595370 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0313 14:06:13.595376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0313 14:06:13.597226 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3f2b6756078925f8562b6053b3b83798efb72288f5fd18fff680a4ab0d33feb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:54Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.271100 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.271214 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.271299 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.271395 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.271493 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:54Z","lastTransitionTime":"2026-03-13T14:06:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.277645 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/58e11265-2fd9-4b9b-926c-3343bb84ee53-serviceca\") pod \"node-ca-745zx\" (UID: \"58e11265-2fd9-4b9b-926c-3343bb84ee53\") " pod="openshift-image-registry/node-ca-745zx" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.277737 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/58e11265-2fd9-4b9b-926c-3343bb84ee53-host\") pod \"node-ca-745zx\" (UID: \"58e11265-2fd9-4b9b-926c-3343bb84ee53\") " pod="openshift-image-registry/node-ca-745zx" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.277787 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mnz7b\" (UniqueName: \"kubernetes.io/projected/58e11265-2fd9-4b9b-926c-3343bb84ee53-kube-api-access-mnz7b\") pod \"node-ca-745zx\" (UID: \"58e11265-2fd9-4b9b-926c-3343bb84ee53\") " pod="openshift-image-registry/node-ca-745zx" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.278377 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/58e11265-2fd9-4b9b-926c-3343bb84ee53-host\") pod \"node-ca-745zx\" (UID: \"58e11265-2fd9-4b9b-926c-3343bb84ee53\") " pod="openshift-image-registry/node-ca-745zx" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.279624 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/58e11265-2fd9-4b9b-926c-3343bb84ee53-serviceca\") pod \"node-ca-745zx\" (UID: \"58e11265-2fd9-4b9b-926c-3343bb84ee53\") " pod="openshift-image-registry/node-ca-745zx" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.291844 4907 generic.go:334] "Generic (PLEG): container finished" podID="57e4e912-9a98-4419-87ce-f16e178ed921" containerID="4c1ab2e148eaa177b98f5883a3f323d12f5b662bff8cb2d072a97be18ab8e120" exitCode=0 Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.291944 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-p8ptc" event={"ID":"57e4e912-9a98-4419-87ce-f16e178ed921","Type":"ContainerDied","Data":"4c1ab2e148eaa177b98f5883a3f323d12f5b662bff8cb2d072a97be18ab8e120"} Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.297693 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94b045cf2fed56457a2ca14f489bb69470198f159f16c207fe061d034dba1555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd28f7a5c35822582b70ed41645294aaede10e28ff59f3f3b5d36e7da8237b76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:54Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.306733 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mnz7b\" (UniqueName: \"kubernetes.io/projected/58e11265-2fd9-4b9b-926c-3343bb84ee53-kube-api-access-mnz7b\") pod \"node-ca-745zx\" (UID: \"58e11265-2fd9-4b9b-926c-3343bb84ee53\") " pod="openshift-image-registry/node-ca-745zx" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.316148 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l7zg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d105034b-f041-4fb5-8df6-6b1bb7681fda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21a207f917836785e320f4f79aca5d6449da2fdeec3ee0765f34d227d5617fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szmh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l7zg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:54Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.337385 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5b4819028f34db81cc43a04e4854c30488ed876c63b8daff11105bf61546c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:54Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.358692 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:54Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.375748 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.375820 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.375839 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.375868 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.375913 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:54Z","lastTransitionTime":"2026-03-13T14:06:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.377288 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xtwf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64ebd338-0158-44dc-90ab-e7c285e87762\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://391b8bc1e975982b172175fcf4b4509756b018cbb99c12e48f4931ff472cdae2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6cwbq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xtwf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:54Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.395435 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:54Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.412244 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xtwf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64ebd338-0158-44dc-90ab-e7c285e87762\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://391b8bc1e975982b172175fcf4b4509756b018cbb99c12e48f4931ff472cdae2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6cwbq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xtwf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:54Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.431263 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5b4819028f34db81cc43a04e4854c30488ed876c63b8daff11105bf61546c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:54Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.432788 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-745zx" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.456291 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:54Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:54 crc kubenswrapper[4907]: W0313 14:06:54.462605 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod58e11265_2fd9_4b9b_926c_3343bb84ee53.slice/crio-f83bd19b71e3bdfc0d53ed197f3b3f54ce16f8b3d4028e9e8b9b0cc846eba6ad WatchSource:0}: Error finding container f83bd19b71e3bdfc0d53ed197f3b3f54ce16f8b3d4028e9e8b9b0cc846eba6ad: Status 404 returned error can't find the container with id f83bd19b71e3bdfc0d53ed197f3b3f54ce16f8b3d4028e9e8b9b0cc846eba6ad Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.473662 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afb1d430459aafab2d6bc2507b4b4059529401e51fdf8e9848b0d7bb8651db68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:54Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.479072 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.479120 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.479137 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.479163 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.479182 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:54Z","lastTransitionTime":"2026-03-13T14:06:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.487372 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-745zx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58e11265-2fd9-4b9b-926c-3343bb84ee53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mnz7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-745zx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:54Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.503502 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:54Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.524685 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-p8ptc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57e4e912-9a98-4419-87ce-f16e178ed921\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bdef68f401b5ba2c7cf61551eba416e74f295ab2281433a5115334d28094818\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bdef68f401b5ba2c7cf61551eba416e74f295ab2281433a5115334d28094818\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f8d1f40744d609fd964572ef5e323f0283d2582e98745602754165818f17847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f8d1f40744d609fd964572ef5e323f0283d2582e98745602754165818f17847\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c1ab2e148eaa177b98f5883a3f323d12f5b662bff8cb2d072a97be18ab8e120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c1ab2e148eaa177b98f5883a3f323d12f5b662bff8cb2d072a97be18ab8e120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-p8ptc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:54Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.544549 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3fb4213-144f-4b01-9d78-ff3315761bb3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba1140a2462713c2401b47a4cf1f231bc9d9ae7867a7d0b1b597214a1f7efe07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://770c882d9e9f9b2e9bd4f135e591952fa15b3898c42826aed8a75d953c9a875b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hm56j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:54Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.572376 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1006376d-4033-454e-8a23-2b67882bdd0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qxlx5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:54Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.581201 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.581241 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.581253 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.581271 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.581286 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:54Z","lastTransitionTime":"2026-03-13T14:06:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.591159 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5d2b054-0ed8-419c-b731-f14d3598372f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4b503aa3b13268b3688d067f3e456c86afac1fda171aced0990899c3eb89b79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be40a1a33fa482eab363537afaa901127d560afbc1b5b287ef11e66391a5be4c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d1b539bbd918fc50428fc18cabcc4c8a43cc96f220b5692d67cf1c8dfbf8ed8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T14:06:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0313 14:06:13.202871 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0313 14:06:13.203175 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0313 14:06:13.204427 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2471728650/tls.crt::/tmp/serving-cert-2471728650/tls.key\\\\\\\"\\\\nI0313 14:06:13.581160 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0313 14:06:13.584012 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0313 14:06:13.584036 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0313 14:06:13.584064 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0313 14:06:13.584072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0313 14:06:13.594504 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0313 14:06:13.594891 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0313 14:06:13.595323 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0313 14:06:13.595363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0313 14:06:13.595370 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0313 14:06:13.595376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0313 14:06:13.597226 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3f2b6756078925f8562b6053b3b83798efb72288f5fd18fff680a4ab0d33feb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:54Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.606368 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94b045cf2fed56457a2ca14f489bb69470198f159f16c207fe061d034dba1555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd28f7a5c35822582b70ed41645294aaede10e28ff59f3f3b5d36e7da8237b76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:54Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.620418 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l7zg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d105034b-f041-4fb5-8df6-6b1bb7681fda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21a207f917836785e320f4f79aca5d6449da2fdeec3ee0765f34d227d5617fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szmh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l7zg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:54Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.683677 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.683711 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.683720 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.683771 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.683807 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:54Z","lastTransitionTime":"2026-03-13T14:06:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.781983 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.782028 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.782109 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:06:54 crc kubenswrapper[4907]: E0313 14:06:54.782408 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.786454 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.786488 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.786499 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.786516 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.786529 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:54Z","lastTransitionTime":"2026-03-13T14:06:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:54 crc kubenswrapper[4907]: E0313 14:06:54.787124 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:06:54 crc kubenswrapper[4907]: E0313 14:06:54.787403 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.889246 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.889284 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.889297 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.889315 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.889326 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:54Z","lastTransitionTime":"2026-03-13T14:06:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.991449 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.991515 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.991533 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.991562 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.991580 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:54Z","lastTransitionTime":"2026-03-13T14:06:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.096205 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.096275 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.096297 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.096324 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.096343 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:55Z","lastTransitionTime":"2026-03-13T14:06:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.199139 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.199199 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.199212 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.199232 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.199245 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:55Z","lastTransitionTime":"2026-03-13T14:06:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.301098 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-p8ptc" event={"ID":"57e4e912-9a98-4419-87ce-f16e178ed921","Type":"ContainerStarted","Data":"38f113418d62dab1728f0ba57cd1d289e3ec41621a338017e74c9996a8449b1d"} Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.304200 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.304285 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.304604 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-745zx" event={"ID":"58e11265-2fd9-4b9b-926c-3343bb84ee53","Type":"ContainerStarted","Data":"4ce7ad32fd2bf2332f3ee41080115b083c34326b7dc31c950098dfba15f58422"} Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.304664 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-745zx" event={"ID":"58e11265-2fd9-4b9b-926c-3343bb84ee53","Type":"ContainerStarted","Data":"f83bd19b71e3bdfc0d53ed197f3b3f54ce16f8b3d4028e9e8b9b0cc846eba6ad"} Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.304311 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.304861 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.304957 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:55Z","lastTransitionTime":"2026-03-13T14:06:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.314974 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" event={"ID":"1006376d-4033-454e-8a23-2b67882bdd0b","Type":"ContainerStarted","Data":"569b1c6f63ead2d0fd069f4e3b144e45e7bb6f74d210c9a276df8d5d48cb51a7"} Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.315384 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.315409 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.329564 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:55Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.343641 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.350384 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-p8ptc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57e4e912-9a98-4419-87ce-f16e178ed921\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38f113418d62dab1728f0ba57cd1d289e3ec41621a338017e74c9996a8449b1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bdef68f401b5ba2c7cf61551eba416e74f295ab2281433a5115334d28094818\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bdef68f401b5ba2c7cf61551eba416e74f295ab2281433a5115334d28094818\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f8d1f40744d609fd964572ef5e323f0283d2582e98745602754165818f17847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f8d1f40744d609fd964572ef5e323f0283d2582e98745602754165818f17847\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c1ab2e148eaa177b98f5883a3f323d12f5b662bff8cb2d072a97be18ab8e120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c1ab2e148eaa177b98f5883a3f323d12f5b662bff8cb2d072a97be18ab8e120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-p8ptc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:55Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.366753 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3fb4213-144f-4b01-9d78-ff3315761bb3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba1140a2462713c2401b47a4cf1f231bc9d9ae7867a7d0b1b597214a1f7efe07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://770c882d9e9f9b2e9bd4f135e591952fa15b3898c42826aed8a75d953c9a875b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hm56j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:55Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.394552 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1006376d-4033-454e-8a23-2b67882bdd0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qxlx5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:55Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.408466 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.408543 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.408581 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.408607 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.408619 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:55Z","lastTransitionTime":"2026-03-13T14:06:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.413727 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5d2b054-0ed8-419c-b731-f14d3598372f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4b503aa3b13268b3688d067f3e456c86afac1fda171aced0990899c3eb89b79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be40a1a33fa482eab363537afaa901127d560afbc1b5b287ef11e66391a5be4c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d1b539bbd918fc50428fc18cabcc4c8a43cc96f220b5692d67cf1c8dfbf8ed8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T14:06:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0313 14:06:13.202871 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0313 14:06:13.203175 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0313 14:06:13.204427 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2471728650/tls.crt::/tmp/serving-cert-2471728650/tls.key\\\\\\\"\\\\nI0313 14:06:13.581160 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0313 14:06:13.584012 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0313 14:06:13.584036 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0313 14:06:13.584064 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0313 14:06:13.584072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0313 14:06:13.594504 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0313 14:06:13.594891 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0313 14:06:13.595323 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0313 14:06:13.595363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0313 14:06:13.595370 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0313 14:06:13.595376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0313 14:06:13.597226 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3f2b6756078925f8562b6053b3b83798efb72288f5fd18fff680a4ab0d33feb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:55Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.435446 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94b045cf2fed56457a2ca14f489bb69470198f159f16c207fe061d034dba1555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd28f7a5c35822582b70ed41645294aaede10e28ff59f3f3b5d36e7da8237b76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:55Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.446855 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l7zg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d105034b-f041-4fb5-8df6-6b1bb7681fda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21a207f917836785e320f4f79aca5d6449da2fdeec3ee0765f34d227d5617fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szmh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l7zg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:55Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.462324 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5b4819028f34db81cc43a04e4854c30488ed876c63b8daff11105bf61546c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:55Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.476456 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:55Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.491458 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xtwf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64ebd338-0158-44dc-90ab-e7c285e87762\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://391b8bc1e975982b172175fcf4b4509756b018cbb99c12e48f4931ff472cdae2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6cwbq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xtwf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:55Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.507460 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:55Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.511668 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.511713 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.511727 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.511750 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.511762 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:55Z","lastTransitionTime":"2026-03-13T14:06:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.521418 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afb1d430459aafab2d6bc2507b4b4059529401e51fdf8e9848b0d7bb8651db68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:55Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.534419 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-745zx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58e11265-2fd9-4b9b-926c-3343bb84ee53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mnz7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-745zx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:55Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.549055 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5b4819028f34db81cc43a04e4854c30488ed876c63b8daff11105bf61546c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:55Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.568008 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:55Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.587471 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xtwf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64ebd338-0158-44dc-90ab-e7c285e87762\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://391b8bc1e975982b172175fcf4b4509756b018cbb99c12e48f4931ff472cdae2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6cwbq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xtwf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:55Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.601639 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-745zx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58e11265-2fd9-4b9b-926c-3343bb84ee53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ce7ad32fd2bf2332f3ee41080115b083c34326b7dc31c950098dfba15f58422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mnz7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-745zx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:55Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.613643 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:55Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.615366 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.615402 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.615414 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.615434 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.615453 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:55Z","lastTransitionTime":"2026-03-13T14:06:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.629207 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afb1d430459aafab2d6bc2507b4b4059529401e51fdf8e9848b0d7bb8651db68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:55Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.644678 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-p8ptc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57e4e912-9a98-4419-87ce-f16e178ed921\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38f113418d62dab1728f0ba57cd1d289e3ec41621a338017e74c9996a8449b1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bdef68f401b5ba2c7cf61551eba416e74f295ab2281433a5115334d28094818\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bdef68f401b5ba2c7cf61551eba416e74f295ab2281433a5115334d28094818\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f8d1f40744d609fd964572ef5e323f0283d2582e98745602754165818f17847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f8d1f40744d609fd964572ef5e323f0283d2582e98745602754165818f17847\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c1ab2e148eaa177b98f5883a3f323d12f5b662bff8cb2d072a97be18ab8e120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c1ab2e148eaa177b98f5883a3f323d12f5b662bff8cb2d072a97be18ab8e120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-p8ptc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:55Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.659258 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3fb4213-144f-4b01-9d78-ff3315761bb3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba1140a2462713c2401b47a4cf1f231bc9d9ae7867a7d0b1b597214a1f7efe07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://770c882d9e9f9b2e9bd4f135e591952fa15b3898c42826aed8a75d953c9a875b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hm56j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:55Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.686266 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1006376d-4033-454e-8a23-2b67882bdd0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf6a3dbaa8dda61f461a906538824c8946d36dc07d4fb95bf4d8420ad0105600\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccef92e41501fab91189581d7e8875515c02d637426b329d2fb80a356c75b637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0ee0049aa4a6d49a6bcf7d347183d9cc8b85171f988662803a005e0cdd97395\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fff87d8ea572615eb1329164f9e536e6ee45cd00a9e71be0aa8a9b3ed2190e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c165053fd9933444b372ff40efe33c0e193a7f39e6f531a4f370e971eb41e6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fea41c11bebd2d9d622ed4b8bc50e03a106dadea6ac102cd60512ea2a1f8b8b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://569b1c6f63ead2d0fd069f4e3b144e45e7bb6f74d210c9a276df8d5d48cb51a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d60b5b774f8dd4cee16b9784f80d124640199a85331a2940772029be4884751\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qxlx5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:55Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.699764 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:55Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.711008 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l7zg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d105034b-f041-4fb5-8df6-6b1bb7681fda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21a207f917836785e320f4f79aca5d6449da2fdeec3ee0765f34d227d5617fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szmh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l7zg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:55Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.718643 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.718715 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.718731 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.718752 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.718766 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:55Z","lastTransitionTime":"2026-03-13T14:06:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.731614 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5d2b054-0ed8-419c-b731-f14d3598372f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4b503aa3b13268b3688d067f3e456c86afac1fda171aced0990899c3eb89b79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be40a1a33fa482eab363537afaa901127d560afbc1b5b287ef11e66391a5be4c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d1b539bbd918fc50428fc18cabcc4c8a43cc96f220b5692d67cf1c8dfbf8ed8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T14:06:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0313 14:06:13.202871 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0313 14:06:13.203175 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0313 14:06:13.204427 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2471728650/tls.crt::/tmp/serving-cert-2471728650/tls.key\\\\\\\"\\\\nI0313 14:06:13.581160 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0313 14:06:13.584012 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0313 14:06:13.584036 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0313 14:06:13.584064 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0313 14:06:13.584072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0313 14:06:13.594504 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0313 14:06:13.594891 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0313 14:06:13.595323 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0313 14:06:13.595363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0313 14:06:13.595370 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0313 14:06:13.595376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0313 14:06:13.597226 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3f2b6756078925f8562b6053b3b83798efb72288f5fd18fff680a4ab0d33feb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:55Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.744542 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94b045cf2fed56457a2ca14f489bb69470198f159f16c207fe061d034dba1555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd28f7a5c35822582b70ed41645294aaede10e28ff59f3f3b5d36e7da8237b76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:55Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.783143 4907 scope.go:117] "RemoveContainer" containerID="1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.823586 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.823616 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.823624 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.823638 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.823648 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:55Z","lastTransitionTime":"2026-03-13T14:06:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.926658 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.926690 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.926699 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.926714 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.926726 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:55Z","lastTransitionTime":"2026-03-13T14:06:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.030598 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.030639 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.030653 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.030671 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.030681 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:56Z","lastTransitionTime":"2026-03-13T14:06:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.134911 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.134988 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.135001 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.135023 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.135041 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:56Z","lastTransitionTime":"2026-03-13T14:06:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.237924 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.237993 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.238007 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.238031 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.238044 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:56Z","lastTransitionTime":"2026-03-13T14:06:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.331650 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.335490 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"93d8c8c79be9d9ffb72de66fc2f3f9821936023071e95875582b45ccb319601f"} Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.336834 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.341716 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.341786 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.341816 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.341848 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.341872 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:56Z","lastTransitionTime":"2026-03-13T14:06:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.363657 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:56Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.380371 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.383643 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afb1d430459aafab2d6bc2507b4b4059529401e51fdf8e9848b0d7bb8651db68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:56Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.400364 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-745zx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58e11265-2fd9-4b9b-926c-3343bb84ee53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ce7ad32fd2bf2332f3ee41080115b083c34326b7dc31c950098dfba15f58422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mnz7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-745zx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:56Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.419744 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:56Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.436029 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-p8ptc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57e4e912-9a98-4419-87ce-f16e178ed921\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38f113418d62dab1728f0ba57cd1d289e3ec41621a338017e74c9996a8449b1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bdef68f401b5ba2c7cf61551eba416e74f295ab2281433a5115334d28094818\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bdef68f401b5ba2c7cf61551eba416e74f295ab2281433a5115334d28094818\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f8d1f40744d609fd964572ef5e323f0283d2582e98745602754165818f17847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f8d1f40744d609fd964572ef5e323f0283d2582e98745602754165818f17847\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c1ab2e148eaa177b98f5883a3f323d12f5b662bff8cb2d072a97be18ab8e120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c1ab2e148eaa177b98f5883a3f323d12f5b662bff8cb2d072a97be18ab8e120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-p8ptc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:56Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.444719 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.444785 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.444804 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.444830 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.444848 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:56Z","lastTransitionTime":"2026-03-13T14:06:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.449598 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3fb4213-144f-4b01-9d78-ff3315761bb3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba1140a2462713c2401b47a4cf1f231bc9d9ae7867a7d0b1b597214a1f7efe07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://770c882d9e9f9b2e9bd4f135e591952fa15b3898c42826aed8a75d953c9a875b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hm56j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:56Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.472198 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1006376d-4033-454e-8a23-2b67882bdd0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf6a3dbaa8dda61f461a906538824c8946d36dc07d4fb95bf4d8420ad0105600\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccef92e41501fab91189581d7e8875515c02d637426b329d2fb80a356c75b637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0ee0049aa4a6d49a6bcf7d347183d9cc8b85171f988662803a005e0cdd97395\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fff87d8ea572615eb1329164f9e536e6ee45cd00a9e71be0aa8a9b3ed2190e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c165053fd9933444b372ff40efe33c0e193a7f39e6f531a4f370e971eb41e6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fea41c11bebd2d9d622ed4b8bc50e03a106dadea6ac102cd60512ea2a1f8b8b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://569b1c6f63ead2d0fd069f4e3b144e45e7bb6f74d210c9a276df8d5d48cb51a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d60b5b774f8dd4cee16b9784f80d124640199a85331a2940772029be4884751\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qxlx5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:56Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.487984 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5d2b054-0ed8-419c-b731-f14d3598372f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4b503aa3b13268b3688d067f3e456c86afac1fda171aced0990899c3eb89b79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be40a1a33fa482eab363537afaa901127d560afbc1b5b287ef11e66391a5be4c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d1b539bbd918fc50428fc18cabcc4c8a43cc96f220b5692d67cf1c8dfbf8ed8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93d8c8c79be9d9ffb72de66fc2f3f9821936023071e95875582b45ccb319601f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T14:06:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0313 14:06:13.202871 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0313 14:06:13.203175 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0313 14:06:13.204427 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2471728650/tls.crt::/tmp/serving-cert-2471728650/tls.key\\\\\\\"\\\\nI0313 14:06:13.581160 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0313 14:06:13.584012 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0313 14:06:13.584036 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0313 14:06:13.584064 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0313 14:06:13.584072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0313 14:06:13.594504 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0313 14:06:13.594891 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0313 14:06:13.595323 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0313 14:06:13.595363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0313 14:06:13.595370 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0313 14:06:13.595376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0313 14:06:13.597226 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3f2b6756078925f8562b6053b3b83798efb72288f5fd18fff680a4ab0d33feb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:56Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.502806 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94b045cf2fed56457a2ca14f489bb69470198f159f16c207fe061d034dba1555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd28f7a5c35822582b70ed41645294aaede10e28ff59f3f3b5d36e7da8237b76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:56Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.514594 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l7zg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d105034b-f041-4fb5-8df6-6b1bb7681fda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21a207f917836785e320f4f79aca5d6449da2fdeec3ee0765f34d227d5617fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szmh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l7zg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:56Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.529380 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5b4819028f34db81cc43a04e4854c30488ed876c63b8daff11105bf61546c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:56Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.545531 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:56Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.547697 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.547756 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.547776 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.547805 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.547824 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:56Z","lastTransitionTime":"2026-03-13T14:06:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.562479 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xtwf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64ebd338-0158-44dc-90ab-e7c285e87762\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://391b8bc1e975982b172175fcf4b4509756b018cbb99c12e48f4931ff472cdae2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6cwbq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xtwf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:56Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.579832 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5d2b054-0ed8-419c-b731-f14d3598372f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4b503aa3b13268b3688d067f3e456c86afac1fda171aced0990899c3eb89b79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be40a1a33fa482eab363537afaa901127d560afbc1b5b287ef11e66391a5be4c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d1b539bbd918fc50428fc18cabcc4c8a43cc96f220b5692d67cf1c8dfbf8ed8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93d8c8c79be9d9ffb72de66fc2f3f9821936023071e95875582b45ccb319601f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T14:06:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0313 14:06:13.202871 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0313 14:06:13.203175 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0313 14:06:13.204427 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2471728650/tls.crt::/tmp/serving-cert-2471728650/tls.key\\\\\\\"\\\\nI0313 14:06:13.581160 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0313 14:06:13.584012 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0313 14:06:13.584036 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0313 14:06:13.584064 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0313 14:06:13.584072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0313 14:06:13.594504 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0313 14:06:13.594891 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0313 14:06:13.595323 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0313 14:06:13.595363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0313 14:06:13.595370 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0313 14:06:13.595376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0313 14:06:13.597226 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3f2b6756078925f8562b6053b3b83798efb72288f5fd18fff680a4ab0d33feb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:56Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.603415 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94b045cf2fed56457a2ca14f489bb69470198f159f16c207fe061d034dba1555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd28f7a5c35822582b70ed41645294aaede10e28ff59f3f3b5d36e7da8237b76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:56Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.629617 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l7zg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d105034b-f041-4fb5-8df6-6b1bb7681fda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21a207f917836785e320f4f79aca5d6449da2fdeec3ee0765f34d227d5617fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szmh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l7zg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:56Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.642212 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:56Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.651795 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.651843 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.651855 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.651898 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.651912 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:56Z","lastTransitionTime":"2026-03-13T14:06:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.660671 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xtwf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64ebd338-0158-44dc-90ab-e7c285e87762\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://391b8bc1e975982b172175fcf4b4509756b018cbb99c12e48f4931ff472cdae2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6cwbq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xtwf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:56Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.678826 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5b4819028f34db81cc43a04e4854c30488ed876c63b8daff11105bf61546c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:56Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.694367 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:56Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.706812 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afb1d430459aafab2d6bc2507b4b4059529401e51fdf8e9848b0d7bb8651db68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:56Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.720104 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-745zx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58e11265-2fd9-4b9b-926c-3343bb84ee53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ce7ad32fd2bf2332f3ee41080115b083c34326b7dc31c950098dfba15f58422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mnz7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-745zx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:56Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.734824 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:56Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.752545 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-p8ptc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57e4e912-9a98-4419-87ce-f16e178ed921\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38f113418d62dab1728f0ba57cd1d289e3ec41621a338017e74c9996a8449b1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bdef68f401b5ba2c7cf61551eba416e74f295ab2281433a5115334d28094818\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bdef68f401b5ba2c7cf61551eba416e74f295ab2281433a5115334d28094818\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f8d1f40744d609fd964572ef5e323f0283d2582e98745602754165818f17847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f8d1f40744d609fd964572ef5e323f0283d2582e98745602754165818f17847\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c1ab2e148eaa177b98f5883a3f323d12f5b662bff8cb2d072a97be18ab8e120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c1ab2e148eaa177b98f5883a3f323d12f5b662bff8cb2d072a97be18ab8e120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-p8ptc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:56Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.754575 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.754629 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.754645 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.754669 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.754687 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:56Z","lastTransitionTime":"2026-03-13T14:06:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.765018 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3fb4213-144f-4b01-9d78-ff3315761bb3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba1140a2462713c2401b47a4cf1f231bc9d9ae7867a7d0b1b597214a1f7efe07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://770c882d9e9f9b2e9bd4f135e591952fa15b3898c42826aed8a75d953c9a875b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hm56j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:56Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.781500 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:06:56 crc kubenswrapper[4907]: E0313 14:06:56.781650 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.781739 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1006376d-4033-454e-8a23-2b67882bdd0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf6a3dbaa8dda61f461a906538824c8946d36dc07d4fb95bf4d8420ad0105600\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccef92e41501fab91189581d7e8875515c02d637426b329d2fb80a356c75b637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0ee0049aa4a6d49a6bcf7d347183d9cc8b85171f988662803a005e0cdd97395\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fff87d8ea572615eb1329164f9e536e6ee45cd00a9e71be0aa8a9b3ed2190e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c165053fd9933444b372ff40efe33c0e193a7f39e6f531a4f370e971eb41e6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fea41c11bebd2d9d622ed4b8bc50e03a106dadea6ac102cd60512ea2a1f8b8b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://569b1c6f63ead2d0fd069f4e3b144e45e7bb6f74d210c9a276df8d5d48cb51a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d60b5b774f8dd4cee16b9784f80d124640199a85331a2940772029be4884751\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qxlx5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:56Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.781447 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.781962 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:06:56 crc kubenswrapper[4907]: E0313 14:06:56.782030 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:06:56 crc kubenswrapper[4907]: E0313 14:06:56.782100 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.790171 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.790227 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.790242 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.790260 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.790276 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:56Z","lastTransitionTime":"2026-03-13T14:06:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:56 crc kubenswrapper[4907]: E0313 14:06:56.802359 4907 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4b182bae-7f2d-4b54-934c-85a9a277d132\\\",\\\"systemUUID\\\":\\\"05bfbfa3-f04f-45ba-8233-a355fc774c24\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:56Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.806368 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.806407 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.806419 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.806437 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.806451 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:56Z","lastTransitionTime":"2026-03-13T14:06:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:56 crc kubenswrapper[4907]: E0313 14:06:56.822399 4907 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4b182bae-7f2d-4b54-934c-85a9a277d132\\\",\\\"systemUUID\\\":\\\"05bfbfa3-f04f-45ba-8233-a355fc774c24\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:56Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.830946 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.830987 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.830999 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.831018 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.831030 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:56Z","lastTransitionTime":"2026-03-13T14:06:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:56 crc kubenswrapper[4907]: E0313 14:06:56.841614 4907 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4b182bae-7f2d-4b54-934c-85a9a277d132\\\",\\\"systemUUID\\\":\\\"05bfbfa3-f04f-45ba-8233-a355fc774c24\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:56Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.845192 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.845224 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.845234 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.845252 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.845262 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:56Z","lastTransitionTime":"2026-03-13T14:06:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:56 crc kubenswrapper[4907]: E0313 14:06:56.859001 4907 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4b182bae-7f2d-4b54-934c-85a9a277d132\\\",\\\"systemUUID\\\":\\\"05bfbfa3-f04f-45ba-8233-a355fc774c24\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:56Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.862315 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.862339 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.862347 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.862361 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.862370 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:56Z","lastTransitionTime":"2026-03-13T14:06:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:56 crc kubenswrapper[4907]: E0313 14:06:56.873302 4907 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4b182bae-7f2d-4b54-934c-85a9a277d132\\\",\\\"systemUUID\\\":\\\"05bfbfa3-f04f-45ba-8233-a355fc774c24\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:56Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:56 crc kubenswrapper[4907]: E0313 14:06:56.873420 4907 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.874704 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.874725 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.874733 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.874745 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.874755 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:56Z","lastTransitionTime":"2026-03-13T14:06:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.976705 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.976764 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.976781 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.976806 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.976824 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:56Z","lastTransitionTime":"2026-03-13T14:06:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:57 crc kubenswrapper[4907]: I0313 14:06:57.079375 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:57 crc kubenswrapper[4907]: I0313 14:06:57.079449 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:57 crc kubenswrapper[4907]: I0313 14:06:57.079467 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:57 crc kubenswrapper[4907]: I0313 14:06:57.079495 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:57 crc kubenswrapper[4907]: I0313 14:06:57.079518 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:57Z","lastTransitionTime":"2026-03-13T14:06:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:57 crc kubenswrapper[4907]: I0313 14:06:57.183422 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:57 crc kubenswrapper[4907]: I0313 14:06:57.183494 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:57 crc kubenswrapper[4907]: I0313 14:06:57.183513 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:57 crc kubenswrapper[4907]: I0313 14:06:57.183548 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:57 crc kubenswrapper[4907]: I0313 14:06:57.183566 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:57Z","lastTransitionTime":"2026-03-13T14:06:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:57 crc kubenswrapper[4907]: I0313 14:06:57.287299 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:57 crc kubenswrapper[4907]: I0313 14:06:57.287357 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:57 crc kubenswrapper[4907]: I0313 14:06:57.287374 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:57 crc kubenswrapper[4907]: I0313 14:06:57.287429 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:57 crc kubenswrapper[4907]: I0313 14:06:57.287453 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:57Z","lastTransitionTime":"2026-03-13T14:06:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:57 crc kubenswrapper[4907]: I0313 14:06:57.390823 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:57 crc kubenswrapper[4907]: I0313 14:06:57.390904 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:57 crc kubenswrapper[4907]: I0313 14:06:57.390916 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:57 crc kubenswrapper[4907]: I0313 14:06:57.390938 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:57 crc kubenswrapper[4907]: I0313 14:06:57.390951 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:57Z","lastTransitionTime":"2026-03-13T14:06:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:57 crc kubenswrapper[4907]: I0313 14:06:57.494017 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:57 crc kubenswrapper[4907]: I0313 14:06:57.494069 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:57 crc kubenswrapper[4907]: I0313 14:06:57.494083 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:57 crc kubenswrapper[4907]: I0313 14:06:57.494106 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:57 crc kubenswrapper[4907]: I0313 14:06:57.494121 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:57Z","lastTransitionTime":"2026-03-13T14:06:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:57 crc kubenswrapper[4907]: I0313 14:06:57.596617 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:57 crc kubenswrapper[4907]: I0313 14:06:57.596660 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:57 crc kubenswrapper[4907]: I0313 14:06:57.596671 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:57 crc kubenswrapper[4907]: I0313 14:06:57.596690 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:57 crc kubenswrapper[4907]: I0313 14:06:57.596702 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:57Z","lastTransitionTime":"2026-03-13T14:06:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:57 crc kubenswrapper[4907]: I0313 14:06:57.699688 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:57 crc kubenswrapper[4907]: I0313 14:06:57.699941 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:57 crc kubenswrapper[4907]: I0313 14:06:57.700024 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:57 crc kubenswrapper[4907]: I0313 14:06:57.700098 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:57 crc kubenswrapper[4907]: I0313 14:06:57.700186 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:57Z","lastTransitionTime":"2026-03-13T14:06:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:57 crc kubenswrapper[4907]: I0313 14:06:57.802791 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:57 crc kubenswrapper[4907]: I0313 14:06:57.802834 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:57 crc kubenswrapper[4907]: I0313 14:06:57.802849 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:57 crc kubenswrapper[4907]: I0313 14:06:57.802871 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:57 crc kubenswrapper[4907]: I0313 14:06:57.802927 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:57Z","lastTransitionTime":"2026-03-13T14:06:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:57 crc kubenswrapper[4907]: I0313 14:06:57.905135 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:57 crc kubenswrapper[4907]: I0313 14:06:57.905603 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:57 crc kubenswrapper[4907]: I0313 14:06:57.905614 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:57 crc kubenswrapper[4907]: I0313 14:06:57.905628 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:57 crc kubenswrapper[4907]: I0313 14:06:57.905636 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:57Z","lastTransitionTime":"2026-03-13T14:06:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.008054 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.008091 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.008101 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.008140 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.008160 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:58Z","lastTransitionTime":"2026-03-13T14:06:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.111040 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.111080 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.111091 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.111110 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.111122 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:58Z","lastTransitionTime":"2026-03-13T14:06:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.214462 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.214526 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.214544 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.214573 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.214592 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:58Z","lastTransitionTime":"2026-03-13T14:06:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.317309 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.317392 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.317411 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.317439 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.317456 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:58Z","lastTransitionTime":"2026-03-13T14:06:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.345702 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qxlx5_1006376d-4033-454e-8a23-2b67882bdd0b/ovnkube-controller/0.log" Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.350322 4907 generic.go:334] "Generic (PLEG): container finished" podID="1006376d-4033-454e-8a23-2b67882bdd0b" containerID="569b1c6f63ead2d0fd069f4e3b144e45e7bb6f74d210c9a276df8d5d48cb51a7" exitCode=1 Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.350362 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" event={"ID":"1006376d-4033-454e-8a23-2b67882bdd0b","Type":"ContainerDied","Data":"569b1c6f63ead2d0fd069f4e3b144e45e7bb6f74d210c9a276df8d5d48cb51a7"} Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.351783 4907 scope.go:117] "RemoveContainer" containerID="569b1c6f63ead2d0fd069f4e3b144e45e7bb6f74d210c9a276df8d5d48cb51a7" Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.374607 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:58Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.392628 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afb1d430459aafab2d6bc2507b4b4059529401e51fdf8e9848b0d7bb8651db68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:58Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.415725 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-745zx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58e11265-2fd9-4b9b-926c-3343bb84ee53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ce7ad32fd2bf2332f3ee41080115b083c34326b7dc31c950098dfba15f58422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mnz7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-745zx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:58Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.423288 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.423320 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.423329 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.423342 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.423351 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:58Z","lastTransitionTime":"2026-03-13T14:06:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.429303 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3fb4213-144f-4b01-9d78-ff3315761bb3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba1140a2462713c2401b47a4cf1f231bc9d9ae7867a7d0b1b597214a1f7efe07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://770c882d9e9f9b2e9bd4f135e591952fa15b3898c42826aed8a75d953c9a875b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hm56j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:58Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.449696 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1006376d-4033-454e-8a23-2b67882bdd0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf6a3dbaa8dda61f461a906538824c8946d36dc07d4fb95bf4d8420ad0105600\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccef92e41501fab91189581d7e8875515c02d637426b329d2fb80a356c75b637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0ee0049aa4a6d49a6bcf7d347183d9cc8b85171f988662803a005e0cdd97395\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fff87d8ea572615eb1329164f9e536e6ee45cd00a9e71be0aa8a9b3ed2190e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c165053fd9933444b372ff40efe33c0e193a7f39e6f531a4f370e971eb41e6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fea41c11bebd2d9d622ed4b8bc50e03a106dadea6ac102cd60512ea2a1f8b8b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://569b1c6f63ead2d0fd069f4e3b144e45e7bb6f74d210c9a276df8d5d48cb51a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://569b1c6f63ead2d0fd069f4e3b144e45e7bb6f74d210c9a276df8d5d48cb51a7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T14:06:57Z\\\",\\\"message\\\":\\\"ector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0313 14:06:57.947520 6737 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0313 14:06:57.946905 6737 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0313 14:06:57.949275 6737 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0313 14:06:57.949375 6737 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0313 14:06:57.949412 6737 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0313 14:06:57.949440 6737 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0313 14:06:57.949483 6737 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0313 14:06:57.949481 6737 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0313 14:06:57.949508 6737 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0313 14:06:57.949515 6737 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0313 14:06:57.949523 6737 handler.go:208] Removed *v1.Node event handler 7\\\\nI0313 14:06:57.949543 6737 factory.go:656] Stopping watch factory\\\\nI0313 14:06:57.949567 6737 ovnkube.go:599] Stopped ovnkube\\\\nI03\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d60b5b774f8dd4cee16b9784f80d124640199a85331a2940772029be4884751\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qxlx5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:58Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.466690 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:58Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.488030 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-p8ptc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57e4e912-9a98-4419-87ce-f16e178ed921\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38f113418d62dab1728f0ba57cd1d289e3ec41621a338017e74c9996a8449b1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bdef68f401b5ba2c7cf61551eba416e74f295ab2281433a5115334d28094818\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bdef68f401b5ba2c7cf61551eba416e74f295ab2281433a5115334d28094818\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f8d1f40744d609fd964572ef5e323f0283d2582e98745602754165818f17847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f8d1f40744d609fd964572ef5e323f0283d2582e98745602754165818f17847\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c1ab2e148eaa177b98f5883a3f323d12f5b662bff8cb2d072a97be18ab8e120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c1ab2e148eaa177b98f5883a3f323d12f5b662bff8cb2d072a97be18ab8e120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-p8ptc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:58Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.503076 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5d2b054-0ed8-419c-b731-f14d3598372f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4b503aa3b13268b3688d067f3e456c86afac1fda171aced0990899c3eb89b79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be40a1a33fa482eab363537afaa901127d560afbc1b5b287ef11e66391a5be4c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d1b539bbd918fc50428fc18cabcc4c8a43cc96f220b5692d67cf1c8dfbf8ed8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93d8c8c79be9d9ffb72de66fc2f3f9821936023071e95875582b45ccb319601f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T14:06:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0313 14:06:13.202871 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0313 14:06:13.203175 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0313 14:06:13.204427 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2471728650/tls.crt::/tmp/serving-cert-2471728650/tls.key\\\\\\\"\\\\nI0313 14:06:13.581160 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0313 14:06:13.584012 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0313 14:06:13.584036 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0313 14:06:13.584064 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0313 14:06:13.584072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0313 14:06:13.594504 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0313 14:06:13.594891 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0313 14:06:13.595323 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0313 14:06:13.595363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0313 14:06:13.595370 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0313 14:06:13.595376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0313 14:06:13.597226 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3f2b6756078925f8562b6053b3b83798efb72288f5fd18fff680a4ab0d33feb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:58Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.515510 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94b045cf2fed56457a2ca14f489bb69470198f159f16c207fe061d034dba1555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd28f7a5c35822582b70ed41645294aaede10e28ff59f3f3b5d36e7da8237b76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:58Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.524740 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l7zg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d105034b-f041-4fb5-8df6-6b1bb7681fda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21a207f917836785e320f4f79aca5d6449da2fdeec3ee0765f34d227d5617fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szmh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l7zg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:58Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.526177 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.526228 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.526240 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.526256 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.526265 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:58Z","lastTransitionTime":"2026-03-13T14:06:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.540655 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5b4819028f34db81cc43a04e4854c30488ed876c63b8daff11105bf61546c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:58Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.552621 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:58Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.568144 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xtwf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64ebd338-0158-44dc-90ab-e7c285e87762\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://391b8bc1e975982b172175fcf4b4509756b018cbb99c12e48f4931ff472cdae2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6cwbq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xtwf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:58Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.629253 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.629301 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.629315 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.629338 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.629356 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:58Z","lastTransitionTime":"2026-03-13T14:06:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.733141 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.733209 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.733322 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.733354 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.733403 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:58Z","lastTransitionTime":"2026-03-13T14:06:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.782028 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.782136 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.782049 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:06:58 crc kubenswrapper[4907]: E0313 14:06:58.782257 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:06:58 crc kubenswrapper[4907]: E0313 14:06:58.782475 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:06:58 crc kubenswrapper[4907]: E0313 14:06:58.782597 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.836630 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.836700 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.836722 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.836753 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.836773 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:58Z","lastTransitionTime":"2026-03-13T14:06:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.939331 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.939373 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.939382 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.939396 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.939404 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:58Z","lastTransitionTime":"2026-03-13T14:06:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.042613 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.042667 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.042685 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.042711 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.042730 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:59Z","lastTransitionTime":"2026-03-13T14:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.144823 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.144871 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.144914 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.144936 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.144953 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:59Z","lastTransitionTime":"2026-03-13T14:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.247210 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.247260 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.247273 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.247291 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.247303 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:59Z","lastTransitionTime":"2026-03-13T14:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.349993 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.350041 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.350074 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.350091 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.350100 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:59Z","lastTransitionTime":"2026-03-13T14:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.354642 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qxlx5_1006376d-4033-454e-8a23-2b67882bdd0b/ovnkube-controller/1.log" Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.355386 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qxlx5_1006376d-4033-454e-8a23-2b67882bdd0b/ovnkube-controller/0.log" Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.357922 4907 generic.go:334] "Generic (PLEG): container finished" podID="1006376d-4033-454e-8a23-2b67882bdd0b" containerID="987fce3ea49cd6a0ce7b37820e7cf26ee21b0721b0b5ffda8e9e0f45bb0d22e3" exitCode=1 Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.357982 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" event={"ID":"1006376d-4033-454e-8a23-2b67882bdd0b","Type":"ContainerDied","Data":"987fce3ea49cd6a0ce7b37820e7cf26ee21b0721b0b5ffda8e9e0f45bb0d22e3"} Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.358021 4907 scope.go:117] "RemoveContainer" containerID="569b1c6f63ead2d0fd069f4e3b144e45e7bb6f74d210c9a276df8d5d48cb51a7" Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.358726 4907 scope.go:117] "RemoveContainer" containerID="987fce3ea49cd6a0ce7b37820e7cf26ee21b0721b0b5ffda8e9e0f45bb0d22e3" Mar 13 14:06:59 crc kubenswrapper[4907]: E0313 14:06:59.358956 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-qxlx5_openshift-ovn-kubernetes(1006376d-4033-454e-8a23-2b67882bdd0b)\"" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" podUID="1006376d-4033-454e-8a23-2b67882bdd0b" Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.371836 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3fb4213-144f-4b01-9d78-ff3315761bb3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba1140a2462713c2401b47a4cf1f231bc9d9ae7867a7d0b1b597214a1f7efe07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://770c882d9e9f9b2e9bd4f135e591952fa15b3898c42826aed8a75d953c9a875b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hm56j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:59Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.399490 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1006376d-4033-454e-8a23-2b67882bdd0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf6a3dbaa8dda61f461a906538824c8946d36dc07d4fb95bf4d8420ad0105600\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccef92e41501fab91189581d7e8875515c02d637426b329d2fb80a356c75b637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0ee0049aa4a6d49a6bcf7d347183d9cc8b85171f988662803a005e0cdd97395\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fff87d8ea572615eb1329164f9e536e6ee45cd00a9e71be0aa8a9b3ed2190e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c165053fd9933444b372ff40efe33c0e193a7f39e6f531a4f370e971eb41e6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fea41c11bebd2d9d622ed4b8bc50e03a106dadea6ac102cd60512ea2a1f8b8b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://987fce3ea49cd6a0ce7b37820e7cf26ee21b0721b0b5ffda8e9e0f45bb0d22e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://569b1c6f63ead2d0fd069f4e3b144e45e7bb6f74d210c9a276df8d5d48cb51a7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T14:06:57Z\\\",\\\"message\\\":\\\"ector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0313 14:06:57.947520 6737 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0313 14:06:57.946905 6737 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0313 14:06:57.949275 6737 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0313 14:06:57.949375 6737 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0313 14:06:57.949412 6737 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0313 14:06:57.949440 6737 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0313 14:06:57.949483 6737 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0313 14:06:57.949481 6737 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0313 14:06:57.949508 6737 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0313 14:06:57.949515 6737 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0313 14:06:57.949523 6737 handler.go:208] Removed *v1.Node event handler 7\\\\nI0313 14:06:57.949543 6737 factory.go:656] Stopping watch factory\\\\nI0313 14:06:57.949567 6737 ovnkube.go:599] Stopped ovnkube\\\\nI03\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:54Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://987fce3ea49cd6a0ce7b37820e7cf26ee21b0721b0b5ffda8e9e0f45bb0d22e3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T14:06:59Z\\\",\\\"message\\\":\\\"r *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0313 14:06:59.164308 6921 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI0313 14:06:59.164611 6921 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0313 14:06:59.164770 6921 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0313 14:06:59.165234 6921 factory.go:656] Stopping watch factory\\\\nI0313 14:06:59.165255 6921 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0313 14:06:59.170923 6921 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI0313 14:06:59.170953 6921 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI0313 14:06:59.171016 6921 ovnkube.go:599] Stopped ovnkube\\\\nI0313 14:06:59.171061 6921 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0313 14:06:59.171151 6921 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d60b5b774f8dd4cee16b9784f80d124640199a85331a2940772029be4884751\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qxlx5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:59Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.412066 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:59Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.427522 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-p8ptc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57e4e912-9a98-4419-87ce-f16e178ed921\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38f113418d62dab1728f0ba57cd1d289e3ec41621a338017e74c9996a8449b1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bdef68f401b5ba2c7cf61551eba416e74f295ab2281433a5115334d28094818\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bdef68f401b5ba2c7cf61551eba416e74f295ab2281433a5115334d28094818\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f8d1f40744d609fd964572ef5e323f0283d2582e98745602754165818f17847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f8d1f40744d609fd964572ef5e323f0283d2582e98745602754165818f17847\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c1ab2e148eaa177b98f5883a3f323d12f5b662bff8cb2d072a97be18ab8e120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c1ab2e148eaa177b98f5883a3f323d12f5b662bff8cb2d072a97be18ab8e120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-p8ptc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:59Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.440603 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5d2b054-0ed8-419c-b731-f14d3598372f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4b503aa3b13268b3688d067f3e456c86afac1fda171aced0990899c3eb89b79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be40a1a33fa482eab363537afaa901127d560afbc1b5b287ef11e66391a5be4c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d1b539bbd918fc50428fc18cabcc4c8a43cc96f220b5692d67cf1c8dfbf8ed8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93d8c8c79be9d9ffb72de66fc2f3f9821936023071e95875582b45ccb319601f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T14:06:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0313 14:06:13.202871 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0313 14:06:13.203175 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0313 14:06:13.204427 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2471728650/tls.crt::/tmp/serving-cert-2471728650/tls.key\\\\\\\"\\\\nI0313 14:06:13.581160 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0313 14:06:13.584012 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0313 14:06:13.584036 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0313 14:06:13.584064 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0313 14:06:13.584072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0313 14:06:13.594504 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0313 14:06:13.594891 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0313 14:06:13.595323 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0313 14:06:13.595363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0313 14:06:13.595370 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0313 14:06:13.595376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0313 14:06:13.597226 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3f2b6756078925f8562b6053b3b83798efb72288f5fd18fff680a4ab0d33feb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:59Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.455050 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.455084 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.455092 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.455107 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.455118 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:59Z","lastTransitionTime":"2026-03-13T14:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.467867 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94b045cf2fed56457a2ca14f489bb69470198f159f16c207fe061d034dba1555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd28f7a5c35822582b70ed41645294aaede10e28ff59f3f3b5d36e7da8237b76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:59Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.481604 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l7zg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d105034b-f041-4fb5-8df6-6b1bb7681fda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21a207f917836785e320f4f79aca5d6449da2fdeec3ee0765f34d227d5617fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szmh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l7zg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:59Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.496925 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5b4819028f34db81cc43a04e4854c30488ed876c63b8daff11105bf61546c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:59Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.510637 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:59Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.522188 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xtwf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64ebd338-0158-44dc-90ab-e7c285e87762\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://391b8bc1e975982b172175fcf4b4509756b018cbb99c12e48f4931ff472cdae2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6cwbq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xtwf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:59Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.535023 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:59Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.544501 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afb1d430459aafab2d6bc2507b4b4059529401e51fdf8e9848b0d7bb8651db68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:59Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.554106 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-745zx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58e11265-2fd9-4b9b-926c-3343bb84ee53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ce7ad32fd2bf2332f3ee41080115b083c34326b7dc31c950098dfba15f58422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mnz7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-745zx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:59Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.557826 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.557865 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.557893 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.557911 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.557924 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:59Z","lastTransitionTime":"2026-03-13T14:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.660945 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.661025 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.661050 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.661085 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.661140 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:59Z","lastTransitionTime":"2026-03-13T14:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.764732 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.764798 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.764819 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.764844 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.764916 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:59Z","lastTransitionTime":"2026-03-13T14:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.866936 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.866994 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.867011 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.867036 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.867056 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:59Z","lastTransitionTime":"2026-03-13T14:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.969383 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.969718 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.969841 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.970009 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.970132 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:59Z","lastTransitionTime":"2026-03-13T14:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.019302 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qncw4"] Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.020223 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qncw4" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.023117 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.023182 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.038063 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afb1d430459aafab2d6bc2507b4b4059529401e51fdf8e9848b0d7bb8651db68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:00Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.050278 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-745zx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58e11265-2fd9-4b9b-926c-3343bb84ee53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ce7ad32fd2bf2332f3ee41080115b083c34326b7dc31c950098dfba15f58422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mnz7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-745zx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:00Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.050389 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/e1d7ee39-09d7-4c0e-942b-5e9d8c9f63be-env-overrides\") pod \"ovnkube-control-plane-749d76644c-qncw4\" (UID: \"e1d7ee39-09d7-4c0e-942b-5e9d8c9f63be\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qncw4" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.050440 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/e1d7ee39-09d7-4c0e-942b-5e9d8c9f63be-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-qncw4\" (UID: \"e1d7ee39-09d7-4c0e-942b-5e9d8c9f63be\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qncw4" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.050494 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/e1d7ee39-09d7-4c0e-942b-5e9d8c9f63be-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-qncw4\" (UID: \"e1d7ee39-09d7-4c0e-942b-5e9d8c9f63be\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qncw4" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.050520 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6bg45\" (UniqueName: \"kubernetes.io/projected/e1d7ee39-09d7-4c0e-942b-5e9d8c9f63be-kube-api-access-6bg45\") pod \"ovnkube-control-plane-749d76644c-qncw4\" (UID: \"e1d7ee39-09d7-4c0e-942b-5e9d8c9f63be\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qncw4" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.061057 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:00Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.072615 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.072655 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.072666 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.072683 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.072694 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:07:00Z","lastTransitionTime":"2026-03-13T14:07:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.073453 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:00Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.088154 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-p8ptc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57e4e912-9a98-4419-87ce-f16e178ed921\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38f113418d62dab1728f0ba57cd1d289e3ec41621a338017e74c9996a8449b1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bdef68f401b5ba2c7cf61551eba416e74f295ab2281433a5115334d28094818\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bdef68f401b5ba2c7cf61551eba416e74f295ab2281433a5115334d28094818\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f8d1f40744d609fd964572ef5e323f0283d2582e98745602754165818f17847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f8d1f40744d609fd964572ef5e323f0283d2582e98745602754165818f17847\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c1ab2e148eaa177b98f5883a3f323d12f5b662bff8cb2d072a97be18ab8e120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c1ab2e148eaa177b98f5883a3f323d12f5b662bff8cb2d072a97be18ab8e120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-p8ptc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:00Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.102193 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3fb4213-144f-4b01-9d78-ff3315761bb3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba1140a2462713c2401b47a4cf1f231bc9d9ae7867a7d0b1b597214a1f7efe07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://770c882d9e9f9b2e9bd4f135e591952fa15b3898c42826aed8a75d953c9a875b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hm56j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:00Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.122829 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1006376d-4033-454e-8a23-2b67882bdd0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf6a3dbaa8dda61f461a906538824c8946d36dc07d4fb95bf4d8420ad0105600\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccef92e41501fab91189581d7e8875515c02d637426b329d2fb80a356c75b637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0ee0049aa4a6d49a6bcf7d347183d9cc8b85171f988662803a005e0cdd97395\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fff87d8ea572615eb1329164f9e536e6ee45cd00a9e71be0aa8a9b3ed2190e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c165053fd9933444b372ff40efe33c0e193a7f39e6f531a4f370e971eb41e6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fea41c11bebd2d9d622ed4b8bc50e03a106dadea6ac102cd60512ea2a1f8b8b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://987fce3ea49cd6a0ce7b37820e7cf26ee21b0721b0b5ffda8e9e0f45bb0d22e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://569b1c6f63ead2d0fd069f4e3b144e45e7bb6f74d210c9a276df8d5d48cb51a7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T14:06:57Z\\\",\\\"message\\\":\\\"ector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0313 14:06:57.947520 6737 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0313 14:06:57.946905 6737 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0313 14:06:57.949275 6737 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0313 14:06:57.949375 6737 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0313 14:06:57.949412 6737 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0313 14:06:57.949440 6737 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0313 14:06:57.949483 6737 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0313 14:06:57.949481 6737 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0313 14:06:57.949508 6737 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0313 14:06:57.949515 6737 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0313 14:06:57.949523 6737 handler.go:208] Removed *v1.Node event handler 7\\\\nI0313 14:06:57.949543 6737 factory.go:656] Stopping watch factory\\\\nI0313 14:06:57.949567 6737 ovnkube.go:599] Stopped ovnkube\\\\nI03\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:54Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://987fce3ea49cd6a0ce7b37820e7cf26ee21b0721b0b5ffda8e9e0f45bb0d22e3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T14:06:59Z\\\",\\\"message\\\":\\\"r *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0313 14:06:59.164308 6921 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI0313 14:06:59.164611 6921 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0313 14:06:59.164770 6921 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0313 14:06:59.165234 6921 factory.go:656] Stopping watch factory\\\\nI0313 14:06:59.165255 6921 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0313 14:06:59.170923 6921 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI0313 14:06:59.170953 6921 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI0313 14:06:59.171016 6921 ovnkube.go:599] Stopped ovnkube\\\\nI0313 14:06:59.171061 6921 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0313 14:06:59.171151 6921 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d60b5b774f8dd4cee16b9784f80d124640199a85331a2940772029be4884751\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qxlx5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:00Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.141229 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94b045cf2fed56457a2ca14f489bb69470198f159f16c207fe061d034dba1555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd28f7a5c35822582b70ed41645294aaede10e28ff59f3f3b5d36e7da8237b76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:00Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.152049 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/e1d7ee39-09d7-4c0e-942b-5e9d8c9f63be-env-overrides\") pod \"ovnkube-control-plane-749d76644c-qncw4\" (UID: \"e1d7ee39-09d7-4c0e-942b-5e9d8c9f63be\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qncw4" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.152262 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/e1d7ee39-09d7-4c0e-942b-5e9d8c9f63be-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-qncw4\" (UID: \"e1d7ee39-09d7-4c0e-942b-5e9d8c9f63be\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qncw4" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.152347 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l7zg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d105034b-f041-4fb5-8df6-6b1bb7681fda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21a207f917836785e320f4f79aca5d6449da2fdeec3ee0765f34d227d5617fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szmh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l7zg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:00Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.152540 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/e1d7ee39-09d7-4c0e-942b-5e9d8c9f63be-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-qncw4\" (UID: \"e1d7ee39-09d7-4c0e-942b-5e9d8c9f63be\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qncw4" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.152661 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6bg45\" (UniqueName: \"kubernetes.io/projected/e1d7ee39-09d7-4c0e-942b-5e9d8c9f63be-kube-api-access-6bg45\") pod \"ovnkube-control-plane-749d76644c-qncw4\" (UID: \"e1d7ee39-09d7-4c0e-942b-5e9d8c9f63be\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qncw4" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.153178 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/e1d7ee39-09d7-4c0e-942b-5e9d8c9f63be-env-overrides\") pod \"ovnkube-control-plane-749d76644c-qncw4\" (UID: \"e1d7ee39-09d7-4c0e-942b-5e9d8c9f63be\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qncw4" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.153268 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/e1d7ee39-09d7-4c0e-942b-5e9d8c9f63be-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-qncw4\" (UID: \"e1d7ee39-09d7-4c0e-942b-5e9d8c9f63be\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qncw4" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.160363 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/e1d7ee39-09d7-4c0e-942b-5e9d8c9f63be-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-qncw4\" (UID: \"e1d7ee39-09d7-4c0e-942b-5e9d8c9f63be\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qncw4" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.166225 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qncw4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1d7ee39-09d7-4c0e-942b-5e9d8c9f63be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bg45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bg45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:07:00Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-qncw4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:00Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.170506 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6bg45\" (UniqueName: \"kubernetes.io/projected/e1d7ee39-09d7-4c0e-942b-5e9d8c9f63be-kube-api-access-6bg45\") pod \"ovnkube-control-plane-749d76644c-qncw4\" (UID: \"e1d7ee39-09d7-4c0e-942b-5e9d8c9f63be\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qncw4" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.175408 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.175575 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.175661 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.175757 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.175860 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:07:00Z","lastTransitionTime":"2026-03-13T14:07:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.180911 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5d2b054-0ed8-419c-b731-f14d3598372f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4b503aa3b13268b3688d067f3e456c86afac1fda171aced0990899c3eb89b79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be40a1a33fa482eab363537afaa901127d560afbc1b5b287ef11e66391a5be4c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d1b539bbd918fc50428fc18cabcc4c8a43cc96f220b5692d67cf1c8dfbf8ed8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93d8c8c79be9d9ffb72de66fc2f3f9821936023071e95875582b45ccb319601f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T14:06:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0313 14:06:13.202871 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0313 14:06:13.203175 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0313 14:06:13.204427 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2471728650/tls.crt::/tmp/serving-cert-2471728650/tls.key\\\\\\\"\\\\nI0313 14:06:13.581160 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0313 14:06:13.584012 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0313 14:06:13.584036 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0313 14:06:13.584064 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0313 14:06:13.584072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0313 14:06:13.594504 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0313 14:06:13.594891 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0313 14:06:13.595323 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0313 14:06:13.595363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0313 14:06:13.595370 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0313 14:06:13.595376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0313 14:06:13.597226 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3f2b6756078925f8562b6053b3b83798efb72288f5fd18fff680a4ab0d33feb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:00Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.228352 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xtwf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64ebd338-0158-44dc-90ab-e7c285e87762\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://391b8bc1e975982b172175fcf4b4509756b018cbb99c12e48f4931ff472cdae2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6cwbq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xtwf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:00Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.243055 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5b4819028f34db81cc43a04e4854c30488ed876c63b8daff11105bf61546c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:00Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.255977 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:00Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.279026 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.279104 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.279121 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.279146 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.279189 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:07:00Z","lastTransitionTime":"2026-03-13T14:07:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.342783 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qncw4" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.371583 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qxlx5_1006376d-4033-454e-8a23-2b67882bdd0b/ovnkube-controller/1.log" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.385403 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.385443 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.385454 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.385471 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.385483 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:07:00Z","lastTransitionTime":"2026-03-13T14:07:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.386705 4907 scope.go:117] "RemoveContainer" containerID="987fce3ea49cd6a0ce7b37820e7cf26ee21b0721b0b5ffda8e9e0f45bb0d22e3" Mar 13 14:07:00 crc kubenswrapper[4907]: E0313 14:07:00.387099 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-qxlx5_openshift-ovn-kubernetes(1006376d-4033-454e-8a23-2b67882bdd0b)\"" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" podUID="1006376d-4033-454e-8a23-2b67882bdd0b" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.405936 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:00Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.432685 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-p8ptc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57e4e912-9a98-4419-87ce-f16e178ed921\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38f113418d62dab1728f0ba57cd1d289e3ec41621a338017e74c9996a8449b1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bdef68f401b5ba2c7cf61551eba416e74f295ab2281433a5115334d28094818\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bdef68f401b5ba2c7cf61551eba416e74f295ab2281433a5115334d28094818\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f8d1f40744d609fd964572ef5e323f0283d2582e98745602754165818f17847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f8d1f40744d609fd964572ef5e323f0283d2582e98745602754165818f17847\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c1ab2e148eaa177b98f5883a3f323d12f5b662bff8cb2d072a97be18ab8e120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c1ab2e148eaa177b98f5883a3f323d12f5b662bff8cb2d072a97be18ab8e120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-p8ptc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:00Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.447980 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3fb4213-144f-4b01-9d78-ff3315761bb3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba1140a2462713c2401b47a4cf1f231bc9d9ae7867a7d0b1b597214a1f7efe07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://770c882d9e9f9b2e9bd4f135e591952fa15b3898c42826aed8a75d953c9a875b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hm56j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:00Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.468541 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1006376d-4033-454e-8a23-2b67882bdd0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf6a3dbaa8dda61f461a906538824c8946d36dc07d4fb95bf4d8420ad0105600\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccef92e41501fab91189581d7e8875515c02d637426b329d2fb80a356c75b637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0ee0049aa4a6d49a6bcf7d347183d9cc8b85171f988662803a005e0cdd97395\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fff87d8ea572615eb1329164f9e536e6ee45cd00a9e71be0aa8a9b3ed2190e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c165053fd9933444b372ff40efe33c0e193a7f39e6f531a4f370e971eb41e6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fea41c11bebd2d9d622ed4b8bc50e03a106dadea6ac102cd60512ea2a1f8b8b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://987fce3ea49cd6a0ce7b37820e7cf26ee21b0721b0b5ffda8e9e0f45bb0d22e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://987fce3ea49cd6a0ce7b37820e7cf26ee21b0721b0b5ffda8e9e0f45bb0d22e3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T14:06:59Z\\\",\\\"message\\\":\\\"r *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0313 14:06:59.164308 6921 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI0313 14:06:59.164611 6921 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0313 14:06:59.164770 6921 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0313 14:06:59.165234 6921 factory.go:656] Stopping watch factory\\\\nI0313 14:06:59.165255 6921 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0313 14:06:59.170923 6921 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI0313 14:06:59.170953 6921 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI0313 14:06:59.171016 6921 ovnkube.go:599] Stopped ovnkube\\\\nI0313 14:06:59.171061 6921 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0313 14:06:59.171151 6921 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:58Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-qxlx5_openshift-ovn-kubernetes(1006376d-4033-454e-8a23-2b67882bdd0b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d60b5b774f8dd4cee16b9784f80d124640199a85331a2940772029be4884751\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qxlx5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:00Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.486194 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5d2b054-0ed8-419c-b731-f14d3598372f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4b503aa3b13268b3688d067f3e456c86afac1fda171aced0990899c3eb89b79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be40a1a33fa482eab363537afaa901127d560afbc1b5b287ef11e66391a5be4c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d1b539bbd918fc50428fc18cabcc4c8a43cc96f220b5692d67cf1c8dfbf8ed8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93d8c8c79be9d9ffb72de66fc2f3f9821936023071e95875582b45ccb319601f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T14:06:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0313 14:06:13.202871 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0313 14:06:13.203175 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0313 14:06:13.204427 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2471728650/tls.crt::/tmp/serving-cert-2471728650/tls.key\\\\\\\"\\\\nI0313 14:06:13.581160 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0313 14:06:13.584012 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0313 14:06:13.584036 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0313 14:06:13.584064 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0313 14:06:13.584072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0313 14:06:13.594504 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0313 14:06:13.594891 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0313 14:06:13.595323 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0313 14:06:13.595363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0313 14:06:13.595370 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0313 14:06:13.595376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0313 14:06:13.597226 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3f2b6756078925f8562b6053b3b83798efb72288f5fd18fff680a4ab0d33feb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:00Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.487953 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.487991 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.488006 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.488030 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.488051 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:07:00Z","lastTransitionTime":"2026-03-13T14:07:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.501385 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94b045cf2fed56457a2ca14f489bb69470198f159f16c207fe061d034dba1555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd28f7a5c35822582b70ed41645294aaede10e28ff59f3f3b5d36e7da8237b76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:00Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.512980 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l7zg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d105034b-f041-4fb5-8df6-6b1bb7681fda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21a207f917836785e320f4f79aca5d6449da2fdeec3ee0765f34d227d5617fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szmh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l7zg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:00Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.529299 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qncw4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1d7ee39-09d7-4c0e-942b-5e9d8c9f63be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bg45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bg45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:07:00Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-qncw4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:00Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.546540 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5b4819028f34db81cc43a04e4854c30488ed876c63b8daff11105bf61546c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:00Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.561515 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:00Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.576539 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xtwf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64ebd338-0158-44dc-90ab-e7c285e87762\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://391b8bc1e975982b172175fcf4b4509756b018cbb99c12e48f4931ff472cdae2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6cwbq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xtwf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:00Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.590564 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.590609 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.590621 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.590642 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.590656 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:07:00Z","lastTransitionTime":"2026-03-13T14:07:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.591073 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:00Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.604092 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afb1d430459aafab2d6bc2507b4b4059529401e51fdf8e9848b0d7bb8651db68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:00Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.615469 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-745zx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58e11265-2fd9-4b9b-926c-3343bb84ee53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ce7ad32fd2bf2332f3ee41080115b083c34326b7dc31c950098dfba15f58422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mnz7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-745zx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:00Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.694385 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.694504 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.694517 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.694535 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.694547 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:07:00Z","lastTransitionTime":"2026-03-13T14:07:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.783459 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.783535 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.783665 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:07:00 crc kubenswrapper[4907]: E0313 14:07:00.783731 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.783916 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-tzmvc"] Mar 13 14:07:00 crc kubenswrapper[4907]: E0313 14:07:00.783953 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:07:00 crc kubenswrapper[4907]: E0313 14:07:00.784195 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.785232 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:07:00 crc kubenswrapper[4907]: E0313 14:07:00.785329 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzmvc" podUID="f797b1bb-e769-49c4-84ce-93a6fad2df93" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.796953 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.796996 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.797007 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.797025 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.797038 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:07:00Z","lastTransitionTime":"2026-03-13T14:07:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.809863 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5d2b054-0ed8-419c-b731-f14d3598372f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4b503aa3b13268b3688d067f3e456c86afac1fda171aced0990899c3eb89b79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be40a1a33fa482eab363537afaa901127d560afbc1b5b287ef11e66391a5be4c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d1b539bbd918fc50428fc18cabcc4c8a43cc96f220b5692d67cf1c8dfbf8ed8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93d8c8c79be9d9ffb72de66fc2f3f9821936023071e95875582b45ccb319601f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T14:06:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0313 14:06:13.202871 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0313 14:06:13.203175 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0313 14:06:13.204427 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2471728650/tls.crt::/tmp/serving-cert-2471728650/tls.key\\\\\\\"\\\\nI0313 14:06:13.581160 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0313 14:06:13.584012 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0313 14:06:13.584036 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0313 14:06:13.584064 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0313 14:06:13.584072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0313 14:06:13.594504 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0313 14:06:13.594891 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0313 14:06:13.595323 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0313 14:06:13.595363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0313 14:06:13.595370 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0313 14:06:13.595376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0313 14:06:13.597226 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3f2b6756078925f8562b6053b3b83798efb72288f5fd18fff680a4ab0d33feb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:00Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.831877 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94b045cf2fed56457a2ca14f489bb69470198f159f16c207fe061d034dba1555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd28f7a5c35822582b70ed41645294aaede10e28ff59f3f3b5d36e7da8237b76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:00Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.847386 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l7zg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d105034b-f041-4fb5-8df6-6b1bb7681fda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21a207f917836785e320f4f79aca5d6449da2fdeec3ee0765f34d227d5617fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szmh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l7zg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:00Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.860303 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f797b1bb-e769-49c4-84ce-93a6fad2df93-metrics-certs\") pod \"network-metrics-daemon-tzmvc\" (UID: \"f797b1bb-e769-49c4-84ce-93a6fad2df93\") " pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.860352 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tdzl9\" (UniqueName: \"kubernetes.io/projected/f797b1bb-e769-49c4-84ce-93a6fad2df93-kube-api-access-tdzl9\") pod \"network-metrics-daemon-tzmvc\" (UID: \"f797b1bb-e769-49c4-84ce-93a6fad2df93\") " pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.860678 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qncw4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1d7ee39-09d7-4c0e-942b-5e9d8c9f63be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bg45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bg45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:07:00Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-qncw4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:00Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.875781 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:00Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.897958 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xtwf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64ebd338-0158-44dc-90ab-e7c285e87762\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://391b8bc1e975982b172175fcf4b4509756b018cbb99c12e48f4931ff472cdae2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6cwbq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xtwf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:00Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.900145 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.900180 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.900191 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.900211 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.900228 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:07:00Z","lastTransitionTime":"2026-03-13T14:07:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.914277 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5b4819028f34db81cc43a04e4854c30488ed876c63b8daff11105bf61546c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:00Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.930075 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:00Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.945061 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afb1d430459aafab2d6bc2507b4b4059529401e51fdf8e9848b0d7bb8651db68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:00Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.961656 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f797b1bb-e769-49c4-84ce-93a6fad2df93-metrics-certs\") pod \"network-metrics-daemon-tzmvc\" (UID: \"f797b1bb-e769-49c4-84ce-93a6fad2df93\") " pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.961725 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tdzl9\" (UniqueName: \"kubernetes.io/projected/f797b1bb-e769-49c4-84ce-93a6fad2df93-kube-api-access-tdzl9\") pod \"network-metrics-daemon-tzmvc\" (UID: \"f797b1bb-e769-49c4-84ce-93a6fad2df93\") " pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:07:00 crc kubenswrapper[4907]: E0313 14:07:00.961933 4907 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 13 14:07:00 crc kubenswrapper[4907]: E0313 14:07:00.962076 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f797b1bb-e769-49c4-84ce-93a6fad2df93-metrics-certs podName:f797b1bb-e769-49c4-84ce-93a6fad2df93 nodeName:}" failed. No retries permitted until 2026-03-13 14:07:01.462036991 +0000 UTC m=+120.361824720 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/f797b1bb-e769-49c4-84ce-93a6fad2df93-metrics-certs") pod "network-metrics-daemon-tzmvc" (UID: "f797b1bb-e769-49c4-84ce-93a6fad2df93") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.962673 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-745zx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58e11265-2fd9-4b9b-926c-3343bb84ee53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ce7ad32fd2bf2332f3ee41080115b083c34326b7dc31c950098dfba15f58422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mnz7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-745zx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:00Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.981509 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-tzmvc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f797b1bb-e769-49c4-84ce-93a6fad2df93\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tdzl9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tdzl9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:07:00Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-tzmvc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:00Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.994977 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tdzl9\" (UniqueName: \"kubernetes.io/projected/f797b1bb-e769-49c4-84ce-93a6fad2df93-kube-api-access-tdzl9\") pod \"network-metrics-daemon-tzmvc\" (UID: \"f797b1bb-e769-49c4-84ce-93a6fad2df93\") " pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.003389 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.003451 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.003466 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.003491 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.003518 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:07:01Z","lastTransitionTime":"2026-03-13T14:07:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.003619 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:01Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.030469 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-p8ptc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57e4e912-9a98-4419-87ce-f16e178ed921\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38f113418d62dab1728f0ba57cd1d289e3ec41621a338017e74c9996a8449b1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bdef68f401b5ba2c7cf61551eba416e74f295ab2281433a5115334d28094818\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bdef68f401b5ba2c7cf61551eba416e74f295ab2281433a5115334d28094818\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f8d1f40744d609fd964572ef5e323f0283d2582e98745602754165818f17847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f8d1f40744d609fd964572ef5e323f0283d2582e98745602754165818f17847\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c1ab2e148eaa177b98f5883a3f323d12f5b662bff8cb2d072a97be18ab8e120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c1ab2e148eaa177b98f5883a3f323d12f5b662bff8cb2d072a97be18ab8e120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-p8ptc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:01Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.050895 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3fb4213-144f-4b01-9d78-ff3315761bb3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba1140a2462713c2401b47a4cf1f231bc9d9ae7867a7d0b1b597214a1f7efe07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://770c882d9e9f9b2e9bd4f135e591952fa15b3898c42826aed8a75d953c9a875b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hm56j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:01Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.077680 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1006376d-4033-454e-8a23-2b67882bdd0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf6a3dbaa8dda61f461a906538824c8946d36dc07d4fb95bf4d8420ad0105600\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccef92e41501fab91189581d7e8875515c02d637426b329d2fb80a356c75b637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0ee0049aa4a6d49a6bcf7d347183d9cc8b85171f988662803a005e0cdd97395\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fff87d8ea572615eb1329164f9e536e6ee45cd00a9e71be0aa8a9b3ed2190e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c165053fd9933444b372ff40efe33c0e193a7f39e6f531a4f370e971eb41e6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fea41c11bebd2d9d622ed4b8bc50e03a106dadea6ac102cd60512ea2a1f8b8b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://987fce3ea49cd6a0ce7b37820e7cf26ee21b0721b0b5ffda8e9e0f45bb0d22e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://987fce3ea49cd6a0ce7b37820e7cf26ee21b0721b0b5ffda8e9e0f45bb0d22e3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T14:06:59Z\\\",\\\"message\\\":\\\"r *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0313 14:06:59.164308 6921 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI0313 14:06:59.164611 6921 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0313 14:06:59.164770 6921 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0313 14:06:59.165234 6921 factory.go:656] Stopping watch factory\\\\nI0313 14:06:59.165255 6921 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0313 14:06:59.170923 6921 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI0313 14:06:59.170953 6921 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI0313 14:06:59.171016 6921 ovnkube.go:599] Stopped ovnkube\\\\nI0313 14:06:59.171061 6921 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0313 14:06:59.171151 6921 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:58Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-qxlx5_openshift-ovn-kubernetes(1006376d-4033-454e-8a23-2b67882bdd0b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d60b5b774f8dd4cee16b9784f80d124640199a85331a2940772029be4884751\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qxlx5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:01Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.106460 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.106497 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.106506 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.106521 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.106531 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:07:01Z","lastTransitionTime":"2026-03-13T14:07:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.208612 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.208662 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.208675 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.208696 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.208710 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:07:01Z","lastTransitionTime":"2026-03-13T14:07:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.311682 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.311717 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.311730 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.311750 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.311769 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:07:01Z","lastTransitionTime":"2026-03-13T14:07:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.391562 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qncw4" event={"ID":"e1d7ee39-09d7-4c0e-942b-5e9d8c9f63be","Type":"ContainerStarted","Data":"540ca207f31c9a8a9892f80dee153b3fb86f8de406f38360221512f424d8c884"} Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.391624 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qncw4" event={"ID":"e1d7ee39-09d7-4c0e-942b-5e9d8c9f63be","Type":"ContainerStarted","Data":"07df71887196cc7098ac9606a4a15b89d45219048c31280198ae75022bff9370"} Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.391636 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qncw4" event={"ID":"e1d7ee39-09d7-4c0e-942b-5e9d8c9f63be","Type":"ContainerStarted","Data":"cdd74771efe08d942155d906ba03e386875af19b9e9f5488d227dd5f4c8b84ab"} Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.415518 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.415562 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.415573 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.415592 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.415603 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:07:01Z","lastTransitionTime":"2026-03-13T14:07:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.417010 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5b4819028f34db81cc43a04e4854c30488ed876c63b8daff11105bf61546c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:01Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.434744 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:01Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.448689 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xtwf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64ebd338-0158-44dc-90ab-e7c285e87762\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://391b8bc1e975982b172175fcf4b4509756b018cbb99c12e48f4931ff472cdae2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6cwbq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xtwf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:01Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.462132 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:01Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.466776 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f797b1bb-e769-49c4-84ce-93a6fad2df93-metrics-certs\") pod \"network-metrics-daemon-tzmvc\" (UID: \"f797b1bb-e769-49c4-84ce-93a6fad2df93\") " pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:07:01 crc kubenswrapper[4907]: E0313 14:07:01.467017 4907 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 13 14:07:01 crc kubenswrapper[4907]: E0313 14:07:01.467131 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f797b1bb-e769-49c4-84ce-93a6fad2df93-metrics-certs podName:f797b1bb-e769-49c4-84ce-93a6fad2df93 nodeName:}" failed. No retries permitted until 2026-03-13 14:07:02.467099639 +0000 UTC m=+121.366887348 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/f797b1bb-e769-49c4-84ce-93a6fad2df93-metrics-certs") pod "network-metrics-daemon-tzmvc" (UID: "f797b1bb-e769-49c4-84ce-93a6fad2df93") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.477244 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afb1d430459aafab2d6bc2507b4b4059529401e51fdf8e9848b0d7bb8651db68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:01Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.488979 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-745zx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58e11265-2fd9-4b9b-926c-3343bb84ee53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ce7ad32fd2bf2332f3ee41080115b083c34326b7dc31c950098dfba15f58422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mnz7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-745zx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:01Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.500864 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-tzmvc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f797b1bb-e769-49c4-84ce-93a6fad2df93\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tdzl9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tdzl9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:07:00Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-tzmvc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:01Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.516118 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:01Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.517802 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.517827 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.517838 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.517853 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.517862 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:07:01Z","lastTransitionTime":"2026-03-13T14:07:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.534640 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-p8ptc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57e4e912-9a98-4419-87ce-f16e178ed921\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38f113418d62dab1728f0ba57cd1d289e3ec41621a338017e74c9996a8449b1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bdef68f401b5ba2c7cf61551eba416e74f295ab2281433a5115334d28094818\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bdef68f401b5ba2c7cf61551eba416e74f295ab2281433a5115334d28094818\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f8d1f40744d609fd964572ef5e323f0283d2582e98745602754165818f17847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f8d1f40744d609fd964572ef5e323f0283d2582e98745602754165818f17847\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c1ab2e148eaa177b98f5883a3f323d12f5b662bff8cb2d072a97be18ab8e120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c1ab2e148eaa177b98f5883a3f323d12f5b662bff8cb2d072a97be18ab8e120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-p8ptc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:01Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.549424 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3fb4213-144f-4b01-9d78-ff3315761bb3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba1140a2462713c2401b47a4cf1f231bc9d9ae7867a7d0b1b597214a1f7efe07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://770c882d9e9f9b2e9bd4f135e591952fa15b3898c42826aed8a75d953c9a875b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hm56j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:01Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.575028 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1006376d-4033-454e-8a23-2b67882bdd0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf6a3dbaa8dda61f461a906538824c8946d36dc07d4fb95bf4d8420ad0105600\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccef92e41501fab91189581d7e8875515c02d637426b329d2fb80a356c75b637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0ee0049aa4a6d49a6bcf7d347183d9cc8b85171f988662803a005e0cdd97395\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fff87d8ea572615eb1329164f9e536e6ee45cd00a9e71be0aa8a9b3ed2190e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c165053fd9933444b372ff40efe33c0e193a7f39e6f531a4f370e971eb41e6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fea41c11bebd2d9d622ed4b8bc50e03a106dadea6ac102cd60512ea2a1f8b8b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://987fce3ea49cd6a0ce7b37820e7cf26ee21b0721b0b5ffda8e9e0f45bb0d22e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://987fce3ea49cd6a0ce7b37820e7cf26ee21b0721b0b5ffda8e9e0f45bb0d22e3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T14:06:59Z\\\",\\\"message\\\":\\\"r *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0313 14:06:59.164308 6921 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI0313 14:06:59.164611 6921 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0313 14:06:59.164770 6921 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0313 14:06:59.165234 6921 factory.go:656] Stopping watch factory\\\\nI0313 14:06:59.165255 6921 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0313 14:06:59.170923 6921 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI0313 14:06:59.170953 6921 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI0313 14:06:59.171016 6921 ovnkube.go:599] Stopped ovnkube\\\\nI0313 14:06:59.171061 6921 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0313 14:06:59.171151 6921 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:58Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-qxlx5_openshift-ovn-kubernetes(1006376d-4033-454e-8a23-2b67882bdd0b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d60b5b774f8dd4cee16b9784f80d124640199a85331a2940772029be4884751\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qxlx5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:01Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.592789 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5d2b054-0ed8-419c-b731-f14d3598372f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4b503aa3b13268b3688d067f3e456c86afac1fda171aced0990899c3eb89b79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be40a1a33fa482eab363537afaa901127d560afbc1b5b287ef11e66391a5be4c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d1b539bbd918fc50428fc18cabcc4c8a43cc96f220b5692d67cf1c8dfbf8ed8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93d8c8c79be9d9ffb72de66fc2f3f9821936023071e95875582b45ccb319601f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T14:06:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0313 14:06:13.202871 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0313 14:06:13.203175 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0313 14:06:13.204427 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2471728650/tls.crt::/tmp/serving-cert-2471728650/tls.key\\\\\\\"\\\\nI0313 14:06:13.581160 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0313 14:06:13.584012 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0313 14:06:13.584036 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0313 14:06:13.584064 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0313 14:06:13.584072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0313 14:06:13.594504 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0313 14:06:13.594891 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0313 14:06:13.595323 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0313 14:06:13.595363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0313 14:06:13.595370 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0313 14:06:13.595376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0313 14:06:13.597226 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3f2b6756078925f8562b6053b3b83798efb72288f5fd18fff680a4ab0d33feb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:01Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.608329 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94b045cf2fed56457a2ca14f489bb69470198f159f16c207fe061d034dba1555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd28f7a5c35822582b70ed41645294aaede10e28ff59f3f3b5d36e7da8237b76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:01Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.620406 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.620528 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.620677 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.620756 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.620852 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:07:01Z","lastTransitionTime":"2026-03-13T14:07:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.621310 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l7zg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d105034b-f041-4fb5-8df6-6b1bb7681fda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21a207f917836785e320f4f79aca5d6449da2fdeec3ee0765f34d227d5617fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szmh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l7zg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:01Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.634664 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qncw4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1d7ee39-09d7-4c0e-942b-5e9d8c9f63be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07df71887196cc7098ac9606a4a15b89d45219048c31280198ae75022bff9370\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:07:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bg45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://540ca207f31c9a8a9892f80dee153b3fb86f8de406f38360221512f424d8c884\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:07:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bg45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:07:00Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-qncw4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:01Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:01 crc kubenswrapper[4907]: E0313 14:07:01.721250 4907 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.805785 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5d2b054-0ed8-419c-b731-f14d3598372f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4b503aa3b13268b3688d067f3e456c86afac1fda171aced0990899c3eb89b79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be40a1a33fa482eab363537afaa901127d560afbc1b5b287ef11e66391a5be4c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d1b539bbd918fc50428fc18cabcc4c8a43cc96f220b5692d67cf1c8dfbf8ed8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93d8c8c79be9d9ffb72de66fc2f3f9821936023071e95875582b45ccb319601f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T14:06:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0313 14:06:13.202871 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0313 14:06:13.203175 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0313 14:06:13.204427 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2471728650/tls.crt::/tmp/serving-cert-2471728650/tls.key\\\\\\\"\\\\nI0313 14:06:13.581160 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0313 14:06:13.584012 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0313 14:06:13.584036 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0313 14:06:13.584064 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0313 14:06:13.584072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0313 14:06:13.594504 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0313 14:06:13.594891 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0313 14:06:13.595323 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0313 14:06:13.595363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0313 14:06:13.595370 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0313 14:06:13.595376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0313 14:06:13.597226 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3f2b6756078925f8562b6053b3b83798efb72288f5fd18fff680a4ab0d33feb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:01Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.824253 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94b045cf2fed56457a2ca14f489bb69470198f159f16c207fe061d034dba1555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd28f7a5c35822582b70ed41645294aaede10e28ff59f3f3b5d36e7da8237b76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:01Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.838173 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l7zg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d105034b-f041-4fb5-8df6-6b1bb7681fda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21a207f917836785e320f4f79aca5d6449da2fdeec3ee0765f34d227d5617fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szmh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l7zg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:01Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.851769 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qncw4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1d7ee39-09d7-4c0e-942b-5e9d8c9f63be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07df71887196cc7098ac9606a4a15b89d45219048c31280198ae75022bff9370\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:07:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bg45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://540ca207f31c9a8a9892f80dee153b3fb86f8de406f38360221512f424d8c884\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:07:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bg45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:07:00Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-qncw4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:01Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.867058 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5b4819028f34db81cc43a04e4854c30488ed876c63b8daff11105bf61546c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:01Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:01 crc kubenswrapper[4907]: E0313 14:07:01.885603 4907 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.887164 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:01Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.909593 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xtwf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64ebd338-0158-44dc-90ab-e7c285e87762\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://391b8bc1e975982b172175fcf4b4509756b018cbb99c12e48f4931ff472cdae2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6cwbq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xtwf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:01Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.927869 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:01Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.943949 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afb1d430459aafab2d6bc2507b4b4059529401e51fdf8e9848b0d7bb8651db68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:01Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.957634 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-745zx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58e11265-2fd9-4b9b-926c-3343bb84ee53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ce7ad32fd2bf2332f3ee41080115b083c34326b7dc31c950098dfba15f58422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mnz7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-745zx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:01Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.971769 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-tzmvc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f797b1bb-e769-49c4-84ce-93a6fad2df93\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tdzl9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tdzl9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:07:00Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-tzmvc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:01Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.989087 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:01Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:02 crc kubenswrapper[4907]: I0313 14:07:02.009639 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-p8ptc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57e4e912-9a98-4419-87ce-f16e178ed921\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38f113418d62dab1728f0ba57cd1d289e3ec41621a338017e74c9996a8449b1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bdef68f401b5ba2c7cf61551eba416e74f295ab2281433a5115334d28094818\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bdef68f401b5ba2c7cf61551eba416e74f295ab2281433a5115334d28094818\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f8d1f40744d609fd964572ef5e323f0283d2582e98745602754165818f17847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f8d1f40744d609fd964572ef5e323f0283d2582e98745602754165818f17847\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c1ab2e148eaa177b98f5883a3f323d12f5b662bff8cb2d072a97be18ab8e120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c1ab2e148eaa177b98f5883a3f323d12f5b662bff8cb2d072a97be18ab8e120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-p8ptc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:02Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:02 crc kubenswrapper[4907]: I0313 14:07:02.025140 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3fb4213-144f-4b01-9d78-ff3315761bb3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba1140a2462713c2401b47a4cf1f231bc9d9ae7867a7d0b1b597214a1f7efe07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://770c882d9e9f9b2e9bd4f135e591952fa15b3898c42826aed8a75d953c9a875b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hm56j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:02Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:02 crc kubenswrapper[4907]: I0313 14:07:02.047479 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1006376d-4033-454e-8a23-2b67882bdd0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf6a3dbaa8dda61f461a906538824c8946d36dc07d4fb95bf4d8420ad0105600\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccef92e41501fab91189581d7e8875515c02d637426b329d2fb80a356c75b637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0ee0049aa4a6d49a6bcf7d347183d9cc8b85171f988662803a005e0cdd97395\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fff87d8ea572615eb1329164f9e536e6ee45cd00a9e71be0aa8a9b3ed2190e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c165053fd9933444b372ff40efe33c0e193a7f39e6f531a4f370e971eb41e6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fea41c11bebd2d9d622ed4b8bc50e03a106dadea6ac102cd60512ea2a1f8b8b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://987fce3ea49cd6a0ce7b37820e7cf26ee21b0721b0b5ffda8e9e0f45bb0d22e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://987fce3ea49cd6a0ce7b37820e7cf26ee21b0721b0b5ffda8e9e0f45bb0d22e3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T14:06:59Z\\\",\\\"message\\\":\\\"r *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0313 14:06:59.164308 6921 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI0313 14:06:59.164611 6921 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0313 14:06:59.164770 6921 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0313 14:06:59.165234 6921 factory.go:656] Stopping watch factory\\\\nI0313 14:06:59.165255 6921 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0313 14:06:59.170923 6921 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI0313 14:06:59.170953 6921 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI0313 14:06:59.171016 6921 ovnkube.go:599] Stopped ovnkube\\\\nI0313 14:06:59.171061 6921 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0313 14:06:59.171151 6921 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:58Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-qxlx5_openshift-ovn-kubernetes(1006376d-4033-454e-8a23-2b67882bdd0b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d60b5b774f8dd4cee16b9784f80d124640199a85331a2940772029be4884751\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qxlx5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:02Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:02 crc kubenswrapper[4907]: I0313 14:07:02.479581 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f797b1bb-e769-49c4-84ce-93a6fad2df93-metrics-certs\") pod \"network-metrics-daemon-tzmvc\" (UID: \"f797b1bb-e769-49c4-84ce-93a6fad2df93\") " pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:07:02 crc kubenswrapper[4907]: E0313 14:07:02.479760 4907 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 13 14:07:02 crc kubenswrapper[4907]: E0313 14:07:02.479868 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f797b1bb-e769-49c4-84ce-93a6fad2df93-metrics-certs podName:f797b1bb-e769-49c4-84ce-93a6fad2df93 nodeName:}" failed. No retries permitted until 2026-03-13 14:07:04.479833493 +0000 UTC m=+123.379621212 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/f797b1bb-e769-49c4-84ce-93a6fad2df93-metrics-certs") pod "network-metrics-daemon-tzmvc" (UID: "f797b1bb-e769-49c4-84ce-93a6fad2df93") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 13 14:07:02 crc kubenswrapper[4907]: I0313 14:07:02.782103 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:07:02 crc kubenswrapper[4907]: I0313 14:07:02.782211 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:07:02 crc kubenswrapper[4907]: I0313 14:07:02.782248 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:07:02 crc kubenswrapper[4907]: I0313 14:07:02.782280 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:07:02 crc kubenswrapper[4907]: E0313 14:07:02.782248 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:07:02 crc kubenswrapper[4907]: E0313 14:07:02.782387 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:07:02 crc kubenswrapper[4907]: E0313 14:07:02.782542 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:07:02 crc kubenswrapper[4907]: E0313 14:07:02.782822 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzmvc" podUID="f797b1bb-e769-49c4-84ce-93a6fad2df93" Mar 13 14:07:04 crc kubenswrapper[4907]: I0313 14:07:04.503963 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f797b1bb-e769-49c4-84ce-93a6fad2df93-metrics-certs\") pod \"network-metrics-daemon-tzmvc\" (UID: \"f797b1bb-e769-49c4-84ce-93a6fad2df93\") " pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:07:04 crc kubenswrapper[4907]: E0313 14:07:04.504277 4907 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 13 14:07:04 crc kubenswrapper[4907]: E0313 14:07:04.505325 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f797b1bb-e769-49c4-84ce-93a6fad2df93-metrics-certs podName:f797b1bb-e769-49c4-84ce-93a6fad2df93 nodeName:}" failed. No retries permitted until 2026-03-13 14:07:08.505276072 +0000 UTC m=+127.405063971 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/f797b1bb-e769-49c4-84ce-93a6fad2df93-metrics-certs") pod "network-metrics-daemon-tzmvc" (UID: "f797b1bb-e769-49c4-84ce-93a6fad2df93") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 13 14:07:04 crc kubenswrapper[4907]: I0313 14:07:04.605836 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:07:04 crc kubenswrapper[4907]: E0313 14:07:04.606193 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:07:36.606147345 +0000 UTC m=+155.505935084 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:07:04 crc kubenswrapper[4907]: I0313 14:07:04.706953 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:07:04 crc kubenswrapper[4907]: I0313 14:07:04.707015 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:07:04 crc kubenswrapper[4907]: I0313 14:07:04.707041 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:07:04 crc kubenswrapper[4907]: I0313 14:07:04.707060 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:07:04 crc kubenswrapper[4907]: E0313 14:07:04.707148 4907 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 13 14:07:04 crc kubenswrapper[4907]: E0313 14:07:04.707202 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-13 14:07:36.707185103 +0000 UTC m=+155.606972792 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 13 14:07:04 crc kubenswrapper[4907]: E0313 14:07:04.707381 4907 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 13 14:07:04 crc kubenswrapper[4907]: E0313 14:07:04.707419 4907 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 13 14:07:04 crc kubenswrapper[4907]: E0313 14:07:04.707468 4907 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 13 14:07:04 crc kubenswrapper[4907]: E0313 14:07:04.707497 4907 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 14:07:04 crc kubenswrapper[4907]: E0313 14:07:04.707521 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-13 14:07:36.707488232 +0000 UTC m=+155.607275951 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 13 14:07:04 crc kubenswrapper[4907]: E0313 14:07:04.707594 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-13 14:07:36.707563214 +0000 UTC m=+155.607350943 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 14:07:04 crc kubenswrapper[4907]: E0313 14:07:04.707712 4907 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 13 14:07:04 crc kubenswrapper[4907]: E0313 14:07:04.707736 4907 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 13 14:07:04 crc kubenswrapper[4907]: E0313 14:07:04.707751 4907 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 14:07:04 crc kubenswrapper[4907]: E0313 14:07:04.707810 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-13 14:07:36.707794622 +0000 UTC m=+155.607582351 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 14:07:04 crc kubenswrapper[4907]: I0313 14:07:04.782362 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:07:04 crc kubenswrapper[4907]: I0313 14:07:04.782426 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:07:04 crc kubenswrapper[4907]: I0313 14:07:04.782363 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:07:04 crc kubenswrapper[4907]: E0313 14:07:04.782530 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:07:04 crc kubenswrapper[4907]: E0313 14:07:04.782708 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:07:04 crc kubenswrapper[4907]: E0313 14:07:04.782945 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:07:04 crc kubenswrapper[4907]: I0313 14:07:04.783059 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:07:04 crc kubenswrapper[4907]: E0313 14:07:04.783291 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzmvc" podUID="f797b1bb-e769-49c4-84ce-93a6fad2df93" Mar 13 14:07:05 crc kubenswrapper[4907]: I0313 14:07:05.164991 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 14:07:06 crc kubenswrapper[4907]: I0313 14:07:06.781625 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:07:06 crc kubenswrapper[4907]: I0313 14:07:06.781674 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:07:06 crc kubenswrapper[4907]: I0313 14:07:06.781716 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:07:06 crc kubenswrapper[4907]: E0313 14:07:06.781789 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:07:06 crc kubenswrapper[4907]: E0313 14:07:06.781970 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:07:06 crc kubenswrapper[4907]: E0313 14:07:06.782090 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:07:06 crc kubenswrapper[4907]: I0313 14:07:06.782364 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:07:06 crc kubenswrapper[4907]: E0313 14:07:06.782606 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzmvc" podUID="f797b1bb-e769-49c4-84ce-93a6fad2df93" Mar 13 14:07:06 crc kubenswrapper[4907]: E0313 14:07:06.887840 4907 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 13 14:07:07 crc kubenswrapper[4907]: I0313 14:07:07.018928 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:07:07 crc kubenswrapper[4907]: I0313 14:07:07.019194 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:07:07 crc kubenswrapper[4907]: I0313 14:07:07.019294 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:07:07 crc kubenswrapper[4907]: I0313 14:07:07.019375 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:07:07 crc kubenswrapper[4907]: I0313 14:07:07.019450 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:07:07Z","lastTransitionTime":"2026-03-13T14:07:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:07:07 crc kubenswrapper[4907]: E0313 14:07:07.034816 4907 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4b182bae-7f2d-4b54-934c-85a9a277d132\\\",\\\"systemUUID\\\":\\\"05bfbfa3-f04f-45ba-8233-a355fc774c24\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:07Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:07 crc kubenswrapper[4907]: I0313 14:07:07.039812 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:07:07 crc kubenswrapper[4907]: I0313 14:07:07.039913 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:07:07 crc kubenswrapper[4907]: I0313 14:07:07.039940 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:07:07 crc kubenswrapper[4907]: I0313 14:07:07.039972 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:07:07 crc kubenswrapper[4907]: I0313 14:07:07.039996 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:07:07Z","lastTransitionTime":"2026-03-13T14:07:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:07:07 crc kubenswrapper[4907]: E0313 14:07:07.061282 4907 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4b182bae-7f2d-4b54-934c-85a9a277d132\\\",\\\"systemUUID\\\":\\\"05bfbfa3-f04f-45ba-8233-a355fc774c24\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:07Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:07 crc kubenswrapper[4907]: I0313 14:07:07.066506 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:07:07 crc kubenswrapper[4907]: I0313 14:07:07.066580 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:07:07 crc kubenswrapper[4907]: I0313 14:07:07.066601 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:07:07 crc kubenswrapper[4907]: I0313 14:07:07.066629 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:07:07 crc kubenswrapper[4907]: I0313 14:07:07.066647 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:07:07Z","lastTransitionTime":"2026-03-13T14:07:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:07:07 crc kubenswrapper[4907]: E0313 14:07:07.087295 4907 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4b182bae-7f2d-4b54-934c-85a9a277d132\\\",\\\"systemUUID\\\":\\\"05bfbfa3-f04f-45ba-8233-a355fc774c24\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:07Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:07 crc kubenswrapper[4907]: I0313 14:07:07.092382 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:07:07 crc kubenswrapper[4907]: I0313 14:07:07.092435 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:07:07 crc kubenswrapper[4907]: I0313 14:07:07.092452 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:07:07 crc kubenswrapper[4907]: I0313 14:07:07.092479 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:07:07 crc kubenswrapper[4907]: I0313 14:07:07.092497 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:07:07Z","lastTransitionTime":"2026-03-13T14:07:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:07:07 crc kubenswrapper[4907]: E0313 14:07:07.110339 4907 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4b182bae-7f2d-4b54-934c-85a9a277d132\\\",\\\"systemUUID\\\":\\\"05bfbfa3-f04f-45ba-8233-a355fc774c24\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:07Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:07 crc kubenswrapper[4907]: I0313 14:07:07.115638 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:07:07 crc kubenswrapper[4907]: I0313 14:07:07.115698 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:07:07 crc kubenswrapper[4907]: I0313 14:07:07.115710 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:07:07 crc kubenswrapper[4907]: I0313 14:07:07.115732 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:07:07 crc kubenswrapper[4907]: I0313 14:07:07.115770 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:07:07Z","lastTransitionTime":"2026-03-13T14:07:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:07:07 crc kubenswrapper[4907]: E0313 14:07:07.133013 4907 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4b182bae-7f2d-4b54-934c-85a9a277d132\\\",\\\"systemUUID\\\":\\\"05bfbfa3-f04f-45ba-8233-a355fc774c24\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:07Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:07 crc kubenswrapper[4907]: E0313 14:07:07.133360 4907 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 13 14:07:08 crc kubenswrapper[4907]: I0313 14:07:08.558695 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f797b1bb-e769-49c4-84ce-93a6fad2df93-metrics-certs\") pod \"network-metrics-daemon-tzmvc\" (UID: \"f797b1bb-e769-49c4-84ce-93a6fad2df93\") " pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:07:08 crc kubenswrapper[4907]: E0313 14:07:08.559017 4907 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 13 14:07:08 crc kubenswrapper[4907]: E0313 14:07:08.559169 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f797b1bb-e769-49c4-84ce-93a6fad2df93-metrics-certs podName:f797b1bb-e769-49c4-84ce-93a6fad2df93 nodeName:}" failed. No retries permitted until 2026-03-13 14:07:16.55913248 +0000 UTC m=+135.458920209 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/f797b1bb-e769-49c4-84ce-93a6fad2df93-metrics-certs") pod "network-metrics-daemon-tzmvc" (UID: "f797b1bb-e769-49c4-84ce-93a6fad2df93") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 13 14:07:08 crc kubenswrapper[4907]: I0313 14:07:08.781970 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:07:08 crc kubenswrapper[4907]: I0313 14:07:08.782047 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:07:08 crc kubenswrapper[4907]: I0313 14:07:08.781970 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:07:08 crc kubenswrapper[4907]: I0313 14:07:08.782024 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:07:08 crc kubenswrapper[4907]: E0313 14:07:08.782210 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:07:08 crc kubenswrapper[4907]: E0313 14:07:08.782344 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:07:08 crc kubenswrapper[4907]: E0313 14:07:08.782451 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzmvc" podUID="f797b1bb-e769-49c4-84ce-93a6fad2df93" Mar 13 14:07:08 crc kubenswrapper[4907]: E0313 14:07:08.782561 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:07:10 crc kubenswrapper[4907]: I0313 14:07:10.197344 4907 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Mar 13 14:07:10 crc kubenswrapper[4907]: I0313 14:07:10.781383 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:07:10 crc kubenswrapper[4907]: I0313 14:07:10.781455 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:07:10 crc kubenswrapper[4907]: E0313 14:07:10.781570 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:07:10 crc kubenswrapper[4907]: I0313 14:07:10.781627 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:07:10 crc kubenswrapper[4907]: E0313 14:07:10.781763 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:07:10 crc kubenswrapper[4907]: I0313 14:07:10.781834 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:07:10 crc kubenswrapper[4907]: E0313 14:07:10.782083 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzmvc" podUID="f797b1bb-e769-49c4-84ce-93a6fad2df93" Mar 13 14:07:10 crc kubenswrapper[4907]: E0313 14:07:10.782660 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:07:10 crc kubenswrapper[4907]: I0313 14:07:10.792450 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Mar 13 14:07:11 crc kubenswrapper[4907]: I0313 14:07:11.805134 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:11Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:11 crc kubenswrapper[4907]: I0313 14:07:11.824358 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afb1d430459aafab2d6bc2507b4b4059529401e51fdf8e9848b0d7bb8651db68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:11Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:11 crc kubenswrapper[4907]: I0313 14:07:11.838833 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-745zx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58e11265-2fd9-4b9b-926c-3343bb84ee53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ce7ad32fd2bf2332f3ee41080115b083c34326b7dc31c950098dfba15f58422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mnz7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-745zx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:11Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:11 crc kubenswrapper[4907]: I0313 14:07:11.855745 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-tzmvc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f797b1bb-e769-49c4-84ce-93a6fad2df93\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tdzl9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tdzl9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:07:00Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-tzmvc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:11Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:11 crc kubenswrapper[4907]: I0313 14:07:11.874521 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"584d221c-be74-4a39-8173-6c8e3f55f686\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b782a795257057929fe616938d1483a9accbd7a93f0cba963f9791a3d4ec4749\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c826e7b9be5e3a0ecdaa6dfeaf2454535e542ff86acc532cc3c80acf4a95c1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4333bdf047392cb3f64e29e677fe8d2f21c7ebca1a7e7097167e70c6785b8d46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a4dea84b35a82ce16f0e0d38057dc994b26240cf45dcc9111ed0de573a62881\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a4dea84b35a82ce16f0e0d38057dc994b26240cf45dcc9111ed0de573a62881\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:11Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:11 crc kubenswrapper[4907]: E0313 14:07:11.888627 4907 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 13 14:07:11 crc kubenswrapper[4907]: I0313 14:07:11.897750 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:11Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:11 crc kubenswrapper[4907]: I0313 14:07:11.922764 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-p8ptc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57e4e912-9a98-4419-87ce-f16e178ed921\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38f113418d62dab1728f0ba57cd1d289e3ec41621a338017e74c9996a8449b1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bdef68f401b5ba2c7cf61551eba416e74f295ab2281433a5115334d28094818\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bdef68f401b5ba2c7cf61551eba416e74f295ab2281433a5115334d28094818\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f8d1f40744d609fd964572ef5e323f0283d2582e98745602754165818f17847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f8d1f40744d609fd964572ef5e323f0283d2582e98745602754165818f17847\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c1ab2e148eaa177b98f5883a3f323d12f5b662bff8cb2d072a97be18ab8e120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c1ab2e148eaa177b98f5883a3f323d12f5b662bff8cb2d072a97be18ab8e120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-p8ptc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:11Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:11 crc kubenswrapper[4907]: I0313 14:07:11.940808 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3fb4213-144f-4b01-9d78-ff3315761bb3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba1140a2462713c2401b47a4cf1f231bc9d9ae7867a7d0b1b597214a1f7efe07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://770c882d9e9f9b2e9bd4f135e591952fa15b3898c42826aed8a75d953c9a875b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hm56j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:11Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:11 crc kubenswrapper[4907]: I0313 14:07:11.962835 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1006376d-4033-454e-8a23-2b67882bdd0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf6a3dbaa8dda61f461a906538824c8946d36dc07d4fb95bf4d8420ad0105600\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccef92e41501fab91189581d7e8875515c02d637426b329d2fb80a356c75b637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0ee0049aa4a6d49a6bcf7d347183d9cc8b85171f988662803a005e0cdd97395\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fff87d8ea572615eb1329164f9e536e6ee45cd00a9e71be0aa8a9b3ed2190e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c165053fd9933444b372ff40efe33c0e193a7f39e6f531a4f370e971eb41e6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fea41c11bebd2d9d622ed4b8bc50e03a106dadea6ac102cd60512ea2a1f8b8b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://987fce3ea49cd6a0ce7b37820e7cf26ee21b0721b0b5ffda8e9e0f45bb0d22e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://987fce3ea49cd6a0ce7b37820e7cf26ee21b0721b0b5ffda8e9e0f45bb0d22e3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T14:06:59Z\\\",\\\"message\\\":\\\"r *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0313 14:06:59.164308 6921 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI0313 14:06:59.164611 6921 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0313 14:06:59.164770 6921 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0313 14:06:59.165234 6921 factory.go:656] Stopping watch factory\\\\nI0313 14:06:59.165255 6921 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0313 14:06:59.170923 6921 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI0313 14:06:59.170953 6921 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI0313 14:06:59.171016 6921 ovnkube.go:599] Stopped ovnkube\\\\nI0313 14:06:59.171061 6921 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0313 14:06:59.171151 6921 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:58Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-qxlx5_openshift-ovn-kubernetes(1006376d-4033-454e-8a23-2b67882bdd0b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d60b5b774f8dd4cee16b9784f80d124640199a85331a2940772029be4884751\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qxlx5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:11Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:11 crc kubenswrapper[4907]: I0313 14:07:11.985052 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5d2b054-0ed8-419c-b731-f14d3598372f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4b503aa3b13268b3688d067f3e456c86afac1fda171aced0990899c3eb89b79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be40a1a33fa482eab363537afaa901127d560afbc1b5b287ef11e66391a5be4c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d1b539bbd918fc50428fc18cabcc4c8a43cc96f220b5692d67cf1c8dfbf8ed8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93d8c8c79be9d9ffb72de66fc2f3f9821936023071e95875582b45ccb319601f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T14:06:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0313 14:06:13.202871 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0313 14:06:13.203175 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0313 14:06:13.204427 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2471728650/tls.crt::/tmp/serving-cert-2471728650/tls.key\\\\\\\"\\\\nI0313 14:06:13.581160 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0313 14:06:13.584012 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0313 14:06:13.584036 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0313 14:06:13.584064 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0313 14:06:13.584072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0313 14:06:13.594504 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0313 14:06:13.594891 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0313 14:06:13.595323 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0313 14:06:13.595363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0313 14:06:13.595370 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0313 14:06:13.595376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0313 14:06:13.597226 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3f2b6756078925f8562b6053b3b83798efb72288f5fd18fff680a4ab0d33feb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:11Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:12 crc kubenswrapper[4907]: I0313 14:07:12.005003 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94b045cf2fed56457a2ca14f489bb69470198f159f16c207fe061d034dba1555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd28f7a5c35822582b70ed41645294aaede10e28ff59f3f3b5d36e7da8237b76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:12Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:12 crc kubenswrapper[4907]: I0313 14:07:12.018181 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l7zg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d105034b-f041-4fb5-8df6-6b1bb7681fda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21a207f917836785e320f4f79aca5d6449da2fdeec3ee0765f34d227d5617fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szmh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l7zg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:12Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:12 crc kubenswrapper[4907]: I0313 14:07:12.032943 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qncw4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1d7ee39-09d7-4c0e-942b-5e9d8c9f63be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07df71887196cc7098ac9606a4a15b89d45219048c31280198ae75022bff9370\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:07:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bg45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://540ca207f31c9a8a9892f80dee153b3fb86f8de406f38360221512f424d8c884\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:07:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bg45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:07:00Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-qncw4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:12Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:12 crc kubenswrapper[4907]: I0313 14:07:12.049438 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5b4819028f34db81cc43a04e4854c30488ed876c63b8daff11105bf61546c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:12Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:12 crc kubenswrapper[4907]: I0313 14:07:12.064630 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:12Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:12 crc kubenswrapper[4907]: I0313 14:07:12.089048 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xtwf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64ebd338-0158-44dc-90ab-e7c285e87762\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://391b8bc1e975982b172175fcf4b4509756b018cbb99c12e48f4931ff472cdae2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6cwbq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xtwf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:12Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:12 crc kubenswrapper[4907]: I0313 14:07:12.781624 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:07:12 crc kubenswrapper[4907]: I0313 14:07:12.781697 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:07:12 crc kubenswrapper[4907]: E0313 14:07:12.781784 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:07:12 crc kubenswrapper[4907]: E0313 14:07:12.781844 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:07:12 crc kubenswrapper[4907]: I0313 14:07:12.782402 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:07:12 crc kubenswrapper[4907]: E0313 14:07:12.782726 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzmvc" podUID="f797b1bb-e769-49c4-84ce-93a6fad2df93" Mar 13 14:07:12 crc kubenswrapper[4907]: I0313 14:07:12.783016 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:07:12 crc kubenswrapper[4907]: E0313 14:07:12.783281 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:07:14 crc kubenswrapper[4907]: I0313 14:07:14.782319 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:07:14 crc kubenswrapper[4907]: E0313 14:07:14.782470 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:07:14 crc kubenswrapper[4907]: I0313 14:07:14.783052 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:07:14 crc kubenswrapper[4907]: I0313 14:07:14.783057 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:07:14 crc kubenswrapper[4907]: I0313 14:07:14.783113 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:07:14 crc kubenswrapper[4907]: I0313 14:07:14.783149 4907 scope.go:117] "RemoveContainer" containerID="987fce3ea49cd6a0ce7b37820e7cf26ee21b0721b0b5ffda8e9e0f45bb0d22e3" Mar 13 14:07:14 crc kubenswrapper[4907]: E0313 14:07:14.783177 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzmvc" podUID="f797b1bb-e769-49c4-84ce-93a6fad2df93" Mar 13 14:07:14 crc kubenswrapper[4907]: E0313 14:07:14.783287 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:07:14 crc kubenswrapper[4907]: E0313 14:07:14.783381 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:07:15 crc kubenswrapper[4907]: I0313 14:07:15.170665 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 14:07:15 crc kubenswrapper[4907]: I0313 14:07:15.194157 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-p8ptc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57e4e912-9a98-4419-87ce-f16e178ed921\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38f113418d62dab1728f0ba57cd1d289e3ec41621a338017e74c9996a8449b1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bdef68f401b5ba2c7cf61551eba416e74f295ab2281433a5115334d28094818\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bdef68f401b5ba2c7cf61551eba416e74f295ab2281433a5115334d28094818\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f8d1f40744d609fd964572ef5e323f0283d2582e98745602754165818f17847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f8d1f40744d609fd964572ef5e323f0283d2582e98745602754165818f17847\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c1ab2e148eaa177b98f5883a3f323d12f5b662bff8cb2d072a97be18ab8e120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c1ab2e148eaa177b98f5883a3f323d12f5b662bff8cb2d072a97be18ab8e120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-p8ptc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:15Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:15 crc kubenswrapper[4907]: I0313 14:07:15.206089 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3fb4213-144f-4b01-9d78-ff3315761bb3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba1140a2462713c2401b47a4cf1f231bc9d9ae7867a7d0b1b597214a1f7efe07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://770c882d9e9f9b2e9bd4f135e591952fa15b3898c42826aed8a75d953c9a875b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hm56j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:15Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:15 crc kubenswrapper[4907]: I0313 14:07:15.223596 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1006376d-4033-454e-8a23-2b67882bdd0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf6a3dbaa8dda61f461a906538824c8946d36dc07d4fb95bf4d8420ad0105600\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccef92e41501fab91189581d7e8875515c02d637426b329d2fb80a356c75b637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0ee0049aa4a6d49a6bcf7d347183d9cc8b85171f988662803a005e0cdd97395\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fff87d8ea572615eb1329164f9e536e6ee45cd00a9e71be0aa8a9b3ed2190e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c165053fd9933444b372ff40efe33c0e193a7f39e6f531a4f370e971eb41e6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fea41c11bebd2d9d622ed4b8bc50e03a106dadea6ac102cd60512ea2a1f8b8b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://987fce3ea49cd6a0ce7b37820e7cf26ee21b0721b0b5ffda8e9e0f45bb0d22e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://987fce3ea49cd6a0ce7b37820e7cf26ee21b0721b0b5ffda8e9e0f45bb0d22e3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T14:06:59Z\\\",\\\"message\\\":\\\"r *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0313 14:06:59.164308 6921 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI0313 14:06:59.164611 6921 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0313 14:06:59.164770 6921 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0313 14:06:59.165234 6921 factory.go:656] Stopping watch factory\\\\nI0313 14:06:59.165255 6921 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0313 14:06:59.170923 6921 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI0313 14:06:59.170953 6921 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI0313 14:06:59.171016 6921 ovnkube.go:599] Stopped ovnkube\\\\nI0313 14:06:59.171061 6921 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0313 14:06:59.171151 6921 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:58Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-qxlx5_openshift-ovn-kubernetes(1006376d-4033-454e-8a23-2b67882bdd0b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d60b5b774f8dd4cee16b9784f80d124640199a85331a2940772029be4884751\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qxlx5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:15Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:15 crc kubenswrapper[4907]: I0313 14:07:15.236720 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"584d221c-be74-4a39-8173-6c8e3f55f686\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b782a795257057929fe616938d1483a9accbd7a93f0cba963f9791a3d4ec4749\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c826e7b9be5e3a0ecdaa6dfeaf2454535e542ff86acc532cc3c80acf4a95c1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4333bdf047392cb3f64e29e677fe8d2f21c7ebca1a7e7097167e70c6785b8d46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a4dea84b35a82ce16f0e0d38057dc994b26240cf45dcc9111ed0de573a62881\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a4dea84b35a82ce16f0e0d38057dc994b26240cf45dcc9111ed0de573a62881\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:15Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:15 crc kubenswrapper[4907]: I0313 14:07:15.247749 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:15Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:15 crc kubenswrapper[4907]: I0313 14:07:15.256161 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l7zg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d105034b-f041-4fb5-8df6-6b1bb7681fda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21a207f917836785e320f4f79aca5d6449da2fdeec3ee0765f34d227d5617fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szmh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l7zg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:15Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:15 crc kubenswrapper[4907]: I0313 14:07:15.268100 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qncw4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1d7ee39-09d7-4c0e-942b-5e9d8c9f63be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07df71887196cc7098ac9606a4a15b89d45219048c31280198ae75022bff9370\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:07:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bg45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://540ca207f31c9a8a9892f80dee153b3fb86f8de406f38360221512f424d8c884\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:07:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bg45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:07:00Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-qncw4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:15Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:15 crc kubenswrapper[4907]: I0313 14:07:15.281123 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5d2b054-0ed8-419c-b731-f14d3598372f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4b503aa3b13268b3688d067f3e456c86afac1fda171aced0990899c3eb89b79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be40a1a33fa482eab363537afaa901127d560afbc1b5b287ef11e66391a5be4c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d1b539bbd918fc50428fc18cabcc4c8a43cc96f220b5692d67cf1c8dfbf8ed8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93d8c8c79be9d9ffb72de66fc2f3f9821936023071e95875582b45ccb319601f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T14:06:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0313 14:06:13.202871 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0313 14:06:13.203175 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0313 14:06:13.204427 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2471728650/tls.crt::/tmp/serving-cert-2471728650/tls.key\\\\\\\"\\\\nI0313 14:06:13.581160 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0313 14:06:13.584012 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0313 14:06:13.584036 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0313 14:06:13.584064 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0313 14:06:13.584072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0313 14:06:13.594504 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0313 14:06:13.594891 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0313 14:06:13.595323 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0313 14:06:13.595363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0313 14:06:13.595370 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0313 14:06:13.595376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0313 14:06:13.597226 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3f2b6756078925f8562b6053b3b83798efb72288f5fd18fff680a4ab0d33feb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:15Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:15 crc kubenswrapper[4907]: I0313 14:07:15.297620 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94b045cf2fed56457a2ca14f489bb69470198f159f16c207fe061d034dba1555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd28f7a5c35822582b70ed41645294aaede10e28ff59f3f3b5d36e7da8237b76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:15Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:15 crc kubenswrapper[4907]: I0313 14:07:15.324958 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5b4819028f34db81cc43a04e4854c30488ed876c63b8daff11105bf61546c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:15Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:15 crc kubenswrapper[4907]: I0313 14:07:15.345021 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:15Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:15 crc kubenswrapper[4907]: I0313 14:07:15.358780 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xtwf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64ebd338-0158-44dc-90ab-e7c285e87762\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://391b8bc1e975982b172175fcf4b4509756b018cbb99c12e48f4931ff472cdae2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6cwbq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xtwf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:15Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:15 crc kubenswrapper[4907]: I0313 14:07:15.368894 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-745zx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58e11265-2fd9-4b9b-926c-3343bb84ee53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ce7ad32fd2bf2332f3ee41080115b083c34326b7dc31c950098dfba15f58422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mnz7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-745zx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:15Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:15 crc kubenswrapper[4907]: I0313 14:07:15.378922 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-tzmvc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f797b1bb-e769-49c4-84ce-93a6fad2df93\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tdzl9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tdzl9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:07:00Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-tzmvc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:15Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:15 crc kubenswrapper[4907]: I0313 14:07:15.391263 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:15Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:15 crc kubenswrapper[4907]: I0313 14:07:15.406347 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afb1d430459aafab2d6bc2507b4b4059529401e51fdf8e9848b0d7bb8651db68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:15Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:15 crc kubenswrapper[4907]: I0313 14:07:15.454374 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qxlx5_1006376d-4033-454e-8a23-2b67882bdd0b/ovnkube-controller/1.log" Mar 13 14:07:15 crc kubenswrapper[4907]: I0313 14:07:15.456750 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" event={"ID":"1006376d-4033-454e-8a23-2b67882bdd0b","Type":"ContainerStarted","Data":"a99cabe30a7abc8da854bf86d79628b94a6ffc20c28f828b695080332b319db2"} Mar 13 14:07:15 crc kubenswrapper[4907]: I0313 14:07:15.457166 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:07:15 crc kubenswrapper[4907]: I0313 14:07:15.472594 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1006376d-4033-454e-8a23-2b67882bdd0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf6a3dbaa8dda61f461a906538824c8946d36dc07d4fb95bf4d8420ad0105600\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccef92e41501fab91189581d7e8875515c02d637426b329d2fb80a356c75b637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0ee0049aa4a6d49a6bcf7d347183d9cc8b85171f988662803a005e0cdd97395\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fff87d8ea572615eb1329164f9e536e6ee45cd00a9e71be0aa8a9b3ed2190e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c165053fd9933444b372ff40efe33c0e193a7f39e6f531a4f370e971eb41e6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fea41c11bebd2d9d622ed4b8bc50e03a106dadea6ac102cd60512ea2a1f8b8b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a99cabe30a7abc8da854bf86d79628b94a6ffc20c28f828b695080332b319db2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://987fce3ea49cd6a0ce7b37820e7cf26ee21b0721b0b5ffda8e9e0f45bb0d22e3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T14:06:59Z\\\",\\\"message\\\":\\\"r *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0313 14:06:59.164308 6921 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI0313 14:06:59.164611 6921 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0313 14:06:59.164770 6921 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0313 14:06:59.165234 6921 factory.go:656] Stopping watch factory\\\\nI0313 14:06:59.165255 6921 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0313 14:06:59.170923 6921 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI0313 14:06:59.170953 6921 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI0313 14:06:59.171016 6921 ovnkube.go:599] Stopped ovnkube\\\\nI0313 14:06:59.171061 6921 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0313 14:06:59.171151 6921 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:58Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:07:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d60b5b774f8dd4cee16b9784f80d124640199a85331a2940772029be4884751\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qxlx5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:15Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:15 crc kubenswrapper[4907]: I0313 14:07:15.482571 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"584d221c-be74-4a39-8173-6c8e3f55f686\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b782a795257057929fe616938d1483a9accbd7a93f0cba963f9791a3d4ec4749\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c826e7b9be5e3a0ecdaa6dfeaf2454535e542ff86acc532cc3c80acf4a95c1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4333bdf047392cb3f64e29e677fe8d2f21c7ebca1a7e7097167e70c6785b8d46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a4dea84b35a82ce16f0e0d38057dc994b26240cf45dcc9111ed0de573a62881\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a4dea84b35a82ce16f0e0d38057dc994b26240cf45dcc9111ed0de573a62881\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:15Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:15 crc kubenswrapper[4907]: I0313 14:07:15.492988 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:15Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:15 crc kubenswrapper[4907]: I0313 14:07:15.504909 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-p8ptc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57e4e912-9a98-4419-87ce-f16e178ed921\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38f113418d62dab1728f0ba57cd1d289e3ec41621a338017e74c9996a8449b1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bdef68f401b5ba2c7cf61551eba416e74f295ab2281433a5115334d28094818\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bdef68f401b5ba2c7cf61551eba416e74f295ab2281433a5115334d28094818\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f8d1f40744d609fd964572ef5e323f0283d2582e98745602754165818f17847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f8d1f40744d609fd964572ef5e323f0283d2582e98745602754165818f17847\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c1ab2e148eaa177b98f5883a3f323d12f5b662bff8cb2d072a97be18ab8e120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c1ab2e148eaa177b98f5883a3f323d12f5b662bff8cb2d072a97be18ab8e120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-p8ptc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:15Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:15 crc kubenswrapper[4907]: I0313 14:07:15.513763 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3fb4213-144f-4b01-9d78-ff3315761bb3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba1140a2462713c2401b47a4cf1f231bc9d9ae7867a7d0b1b597214a1f7efe07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://770c882d9e9f9b2e9bd4f135e591952fa15b3898c42826aed8a75d953c9a875b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hm56j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:15Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:15 crc kubenswrapper[4907]: I0313 14:07:15.524850 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5d2b054-0ed8-419c-b731-f14d3598372f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4b503aa3b13268b3688d067f3e456c86afac1fda171aced0990899c3eb89b79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be40a1a33fa482eab363537afaa901127d560afbc1b5b287ef11e66391a5be4c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d1b539bbd918fc50428fc18cabcc4c8a43cc96f220b5692d67cf1c8dfbf8ed8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93d8c8c79be9d9ffb72de66fc2f3f9821936023071e95875582b45ccb319601f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T14:06:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0313 14:06:13.202871 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0313 14:06:13.203175 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0313 14:06:13.204427 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2471728650/tls.crt::/tmp/serving-cert-2471728650/tls.key\\\\\\\"\\\\nI0313 14:06:13.581160 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0313 14:06:13.584012 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0313 14:06:13.584036 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0313 14:06:13.584064 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0313 14:06:13.584072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0313 14:06:13.594504 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0313 14:06:13.594891 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0313 14:06:13.595323 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0313 14:06:13.595363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0313 14:06:13.595370 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0313 14:06:13.595376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0313 14:06:13.597226 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3f2b6756078925f8562b6053b3b83798efb72288f5fd18fff680a4ab0d33feb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:15Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:15 crc kubenswrapper[4907]: I0313 14:07:15.535604 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94b045cf2fed56457a2ca14f489bb69470198f159f16c207fe061d034dba1555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd28f7a5c35822582b70ed41645294aaede10e28ff59f3f3b5d36e7da8237b76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:15Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:15 crc kubenswrapper[4907]: I0313 14:07:15.544981 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l7zg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d105034b-f041-4fb5-8df6-6b1bb7681fda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21a207f917836785e320f4f79aca5d6449da2fdeec3ee0765f34d227d5617fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szmh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l7zg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:15Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:15 crc kubenswrapper[4907]: I0313 14:07:15.554480 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qncw4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1d7ee39-09d7-4c0e-942b-5e9d8c9f63be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07df71887196cc7098ac9606a4a15b89d45219048c31280198ae75022bff9370\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:07:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bg45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://540ca207f31c9a8a9892f80dee153b3fb86f8de406f38360221512f424d8c884\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:07:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bg45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:07:00Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-qncw4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:15Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:15 crc kubenswrapper[4907]: I0313 14:07:15.566033 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5b4819028f34db81cc43a04e4854c30488ed876c63b8daff11105bf61546c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:15Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:15 crc kubenswrapper[4907]: I0313 14:07:15.579681 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:15Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:15 crc kubenswrapper[4907]: I0313 14:07:15.591803 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xtwf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64ebd338-0158-44dc-90ab-e7c285e87762\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://391b8bc1e975982b172175fcf4b4509756b018cbb99c12e48f4931ff472cdae2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6cwbq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xtwf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:15Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:15 crc kubenswrapper[4907]: I0313 14:07:15.605904 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:15Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:15 crc kubenswrapper[4907]: I0313 14:07:15.624545 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afb1d430459aafab2d6bc2507b4b4059529401e51fdf8e9848b0d7bb8651db68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:15Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:15 crc kubenswrapper[4907]: I0313 14:07:15.637722 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-745zx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58e11265-2fd9-4b9b-926c-3343bb84ee53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ce7ad32fd2bf2332f3ee41080115b083c34326b7dc31c950098dfba15f58422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mnz7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-745zx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:15Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:15 crc kubenswrapper[4907]: I0313 14:07:15.646721 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-tzmvc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f797b1bb-e769-49c4-84ce-93a6fad2df93\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tdzl9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tdzl9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:07:00Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-tzmvc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:15Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:16 crc kubenswrapper[4907]: I0313 14:07:16.462228 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qxlx5_1006376d-4033-454e-8a23-2b67882bdd0b/ovnkube-controller/2.log" Mar 13 14:07:16 crc kubenswrapper[4907]: I0313 14:07:16.462812 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qxlx5_1006376d-4033-454e-8a23-2b67882bdd0b/ovnkube-controller/1.log" Mar 13 14:07:16 crc kubenswrapper[4907]: I0313 14:07:16.466845 4907 generic.go:334] "Generic (PLEG): container finished" podID="1006376d-4033-454e-8a23-2b67882bdd0b" containerID="a99cabe30a7abc8da854bf86d79628b94a6ffc20c28f828b695080332b319db2" exitCode=1 Mar 13 14:07:16 crc kubenswrapper[4907]: I0313 14:07:16.466927 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" event={"ID":"1006376d-4033-454e-8a23-2b67882bdd0b","Type":"ContainerDied","Data":"a99cabe30a7abc8da854bf86d79628b94a6ffc20c28f828b695080332b319db2"} Mar 13 14:07:16 crc kubenswrapper[4907]: I0313 14:07:16.466975 4907 scope.go:117] "RemoveContainer" containerID="987fce3ea49cd6a0ce7b37820e7cf26ee21b0721b0b5ffda8e9e0f45bb0d22e3" Mar 13 14:07:16 crc kubenswrapper[4907]: I0313 14:07:16.467907 4907 scope.go:117] "RemoveContainer" containerID="a99cabe30a7abc8da854bf86d79628b94a6ffc20c28f828b695080332b319db2" Mar 13 14:07:16 crc kubenswrapper[4907]: E0313 14:07:16.468120 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-qxlx5_openshift-ovn-kubernetes(1006376d-4033-454e-8a23-2b67882bdd0b)\"" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" podUID="1006376d-4033-454e-8a23-2b67882bdd0b" Mar 13 14:07:16 crc kubenswrapper[4907]: I0313 14:07:16.488032 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5d2b054-0ed8-419c-b731-f14d3598372f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4b503aa3b13268b3688d067f3e456c86afac1fda171aced0990899c3eb89b79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be40a1a33fa482eab363537afaa901127d560afbc1b5b287ef11e66391a5be4c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d1b539bbd918fc50428fc18cabcc4c8a43cc96f220b5692d67cf1c8dfbf8ed8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93d8c8c79be9d9ffb72de66fc2f3f9821936023071e95875582b45ccb319601f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T14:06:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0313 14:06:13.202871 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0313 14:06:13.203175 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0313 14:06:13.204427 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2471728650/tls.crt::/tmp/serving-cert-2471728650/tls.key\\\\\\\"\\\\nI0313 14:06:13.581160 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0313 14:06:13.584012 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0313 14:06:13.584036 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0313 14:06:13.584064 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0313 14:06:13.584072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0313 14:06:13.594504 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0313 14:06:13.594891 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0313 14:06:13.595323 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0313 14:06:13.595363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0313 14:06:13.595370 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0313 14:06:13.595376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0313 14:06:13.597226 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3f2b6756078925f8562b6053b3b83798efb72288f5fd18fff680a4ab0d33feb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:16Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:16 crc kubenswrapper[4907]: I0313 14:07:16.508064 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94b045cf2fed56457a2ca14f489bb69470198f159f16c207fe061d034dba1555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd28f7a5c35822582b70ed41645294aaede10e28ff59f3f3b5d36e7da8237b76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:16Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:16 crc kubenswrapper[4907]: I0313 14:07:16.519291 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l7zg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d105034b-f041-4fb5-8df6-6b1bb7681fda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21a207f917836785e320f4f79aca5d6449da2fdeec3ee0765f34d227d5617fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szmh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l7zg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:16Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:16 crc kubenswrapper[4907]: I0313 14:07:16.530704 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qncw4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1d7ee39-09d7-4c0e-942b-5e9d8c9f63be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07df71887196cc7098ac9606a4a15b89d45219048c31280198ae75022bff9370\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:07:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bg45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://540ca207f31c9a8a9892f80dee153b3fb86f8de406f38360221512f424d8c884\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:07:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bg45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:07:00Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-qncw4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:16Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:16 crc kubenswrapper[4907]: I0313 14:07:16.544667 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5b4819028f34db81cc43a04e4854c30488ed876c63b8daff11105bf61546c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:16Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:16 crc kubenswrapper[4907]: I0313 14:07:16.561774 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:16Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:16 crc kubenswrapper[4907]: I0313 14:07:16.577374 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xtwf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64ebd338-0158-44dc-90ab-e7c285e87762\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://391b8bc1e975982b172175fcf4b4509756b018cbb99c12e48f4931ff472cdae2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6cwbq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xtwf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:16Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:16 crc kubenswrapper[4907]: I0313 14:07:16.590428 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:16Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:16 crc kubenswrapper[4907]: I0313 14:07:16.603912 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afb1d430459aafab2d6bc2507b4b4059529401e51fdf8e9848b0d7bb8651db68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:16Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:16 crc kubenswrapper[4907]: I0313 14:07:16.615850 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-745zx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58e11265-2fd9-4b9b-926c-3343bb84ee53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ce7ad32fd2bf2332f3ee41080115b083c34326b7dc31c950098dfba15f58422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mnz7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-745zx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:16Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:16 crc kubenswrapper[4907]: I0313 14:07:16.627927 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-tzmvc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f797b1bb-e769-49c4-84ce-93a6fad2df93\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tdzl9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tdzl9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:07:00Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-tzmvc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:16Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:16 crc kubenswrapper[4907]: I0313 14:07:16.640014 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"584d221c-be74-4a39-8173-6c8e3f55f686\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b782a795257057929fe616938d1483a9accbd7a93f0cba963f9791a3d4ec4749\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c826e7b9be5e3a0ecdaa6dfeaf2454535e542ff86acc532cc3c80acf4a95c1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4333bdf047392cb3f64e29e677fe8d2f21c7ebca1a7e7097167e70c6785b8d46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a4dea84b35a82ce16f0e0d38057dc994b26240cf45dcc9111ed0de573a62881\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a4dea84b35a82ce16f0e0d38057dc994b26240cf45dcc9111ed0de573a62881\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:16Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:16 crc kubenswrapper[4907]: I0313 14:07:16.652011 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:16Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:16 crc kubenswrapper[4907]: I0313 14:07:16.652448 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f797b1bb-e769-49c4-84ce-93a6fad2df93-metrics-certs\") pod \"network-metrics-daemon-tzmvc\" (UID: \"f797b1bb-e769-49c4-84ce-93a6fad2df93\") " pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:07:16 crc kubenswrapper[4907]: E0313 14:07:16.652524 4907 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 13 14:07:16 crc kubenswrapper[4907]: E0313 14:07:16.652625 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f797b1bb-e769-49c4-84ce-93a6fad2df93-metrics-certs podName:f797b1bb-e769-49c4-84ce-93a6fad2df93 nodeName:}" failed. No retries permitted until 2026-03-13 14:07:32.652597329 +0000 UTC m=+151.552385058 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/f797b1bb-e769-49c4-84ce-93a6fad2df93-metrics-certs") pod "network-metrics-daemon-tzmvc" (UID: "f797b1bb-e769-49c4-84ce-93a6fad2df93") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 13 14:07:16 crc kubenswrapper[4907]: I0313 14:07:16.666999 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-p8ptc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57e4e912-9a98-4419-87ce-f16e178ed921\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38f113418d62dab1728f0ba57cd1d289e3ec41621a338017e74c9996a8449b1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bdef68f401b5ba2c7cf61551eba416e74f295ab2281433a5115334d28094818\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bdef68f401b5ba2c7cf61551eba416e74f295ab2281433a5115334d28094818\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f8d1f40744d609fd964572ef5e323f0283d2582e98745602754165818f17847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f8d1f40744d609fd964572ef5e323f0283d2582e98745602754165818f17847\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c1ab2e148eaa177b98f5883a3f323d12f5b662bff8cb2d072a97be18ab8e120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c1ab2e148eaa177b98f5883a3f323d12f5b662bff8cb2d072a97be18ab8e120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-p8ptc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:16Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:16 crc kubenswrapper[4907]: I0313 14:07:16.678865 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3fb4213-144f-4b01-9d78-ff3315761bb3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba1140a2462713c2401b47a4cf1f231bc9d9ae7867a7d0b1b597214a1f7efe07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://770c882d9e9f9b2e9bd4f135e591952fa15b3898c42826aed8a75d953c9a875b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hm56j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:16Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:16 crc kubenswrapper[4907]: I0313 14:07:16.700709 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1006376d-4033-454e-8a23-2b67882bdd0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf6a3dbaa8dda61f461a906538824c8946d36dc07d4fb95bf4d8420ad0105600\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccef92e41501fab91189581d7e8875515c02d637426b329d2fb80a356c75b637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0ee0049aa4a6d49a6bcf7d347183d9cc8b85171f988662803a005e0cdd97395\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fff87d8ea572615eb1329164f9e536e6ee45cd00a9e71be0aa8a9b3ed2190e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c165053fd9933444b372ff40efe33c0e193a7f39e6f531a4f370e971eb41e6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fea41c11bebd2d9d622ed4b8bc50e03a106dadea6ac102cd60512ea2a1f8b8b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a99cabe30a7abc8da854bf86d79628b94a6ffc20c28f828b695080332b319db2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://987fce3ea49cd6a0ce7b37820e7cf26ee21b0721b0b5ffda8e9e0f45bb0d22e3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T14:06:59Z\\\",\\\"message\\\":\\\"r *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0313 14:06:59.164308 6921 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI0313 14:06:59.164611 6921 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0313 14:06:59.164770 6921 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0313 14:06:59.165234 6921 factory.go:656] Stopping watch factory\\\\nI0313 14:06:59.165255 6921 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0313 14:06:59.170923 6921 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI0313 14:06:59.170953 6921 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI0313 14:06:59.171016 6921 ovnkube.go:599] Stopped ovnkube\\\\nI0313 14:06:59.171061 6921 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0313 14:06:59.171151 6921 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:58Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a99cabe30a7abc8da854bf86d79628b94a6ffc20c28f828b695080332b319db2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T14:07:15Z\\\",\\\"message\\\":\\\"Sending *v1.Node event handler 7 for removal\\\\nI0313 14:07:15.619307 7162 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0313 14:07:15.619320 7162 handler.go:208] Removed *v1.Node event handler 2\\\\nI0313 14:07:15.619329 7162 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0313 14:07:15.619343 7162 handler.go:208] Removed *v1.Node event handler 7\\\\nI0313 14:07:15.619351 7162 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0313 14:07:15.619358 7162 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0313 14:07:15.619362 7162 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0313 14:07:15.619381 7162 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0313 14:07:15.619382 7162 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0313 14:07:15.619392 7162 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0313 14:07:15.619619 7162 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI0313 14:07:15.619713 7162 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI0313 14:07:15.619975 7162 factory.go:656] Stopping watch factory\\\\nI0313 14:07:15.620001 7162 ovnkube.go:599] Stopped ovnkube\\\\nI0313 14:07:15.620033 7162 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0313 14:07:15.620096 7162 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:07:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d60b5b774f8dd4cee16b9784f80d124640199a85331a2940772029be4884751\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qxlx5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:16Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:16 crc kubenswrapper[4907]: I0313 14:07:16.782063 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:07:16 crc kubenswrapper[4907]: I0313 14:07:16.782119 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:07:16 crc kubenswrapper[4907]: I0313 14:07:16.782084 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:07:16 crc kubenswrapper[4907]: I0313 14:07:16.782071 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:07:16 crc kubenswrapper[4907]: E0313 14:07:16.782203 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:07:16 crc kubenswrapper[4907]: E0313 14:07:16.782309 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:07:16 crc kubenswrapper[4907]: E0313 14:07:16.782407 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzmvc" podUID="f797b1bb-e769-49c4-84ce-93a6fad2df93" Mar 13 14:07:16 crc kubenswrapper[4907]: E0313 14:07:16.782501 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:07:16 crc kubenswrapper[4907]: E0313 14:07:16.890072 4907 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 13 14:07:17 crc kubenswrapper[4907]: I0313 14:07:17.447184 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:07:17 crc kubenswrapper[4907]: I0313 14:07:17.447227 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:07:17 crc kubenswrapper[4907]: I0313 14:07:17.447235 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:07:17 crc kubenswrapper[4907]: I0313 14:07:17.447252 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:07:17 crc kubenswrapper[4907]: I0313 14:07:17.447262 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:07:17Z","lastTransitionTime":"2026-03-13T14:07:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:07:17 crc kubenswrapper[4907]: E0313 14:07:17.458388 4907 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4b182bae-7f2d-4b54-934c-85a9a277d132\\\",\\\"systemUUID\\\":\\\"05bfbfa3-f04f-45ba-8233-a355fc774c24\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:17Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:17 crc kubenswrapper[4907]: I0313 14:07:17.461820 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:07:17 crc kubenswrapper[4907]: I0313 14:07:17.461913 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:07:17 crc kubenswrapper[4907]: I0313 14:07:17.461937 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:07:17 crc kubenswrapper[4907]: I0313 14:07:17.461960 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:07:17 crc kubenswrapper[4907]: I0313 14:07:17.461973 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:07:17Z","lastTransitionTime":"2026-03-13T14:07:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:07:17 crc kubenswrapper[4907]: I0313 14:07:17.471041 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qxlx5_1006376d-4033-454e-8a23-2b67882bdd0b/ovnkube-controller/2.log" Mar 13 14:07:17 crc kubenswrapper[4907]: E0313 14:07:17.474144 4907 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4b182bae-7f2d-4b54-934c-85a9a277d132\\\",\\\"systemUUID\\\":\\\"05bfbfa3-f04f-45ba-8233-a355fc774c24\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:17Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:17 crc kubenswrapper[4907]: I0313 14:07:17.475420 4907 scope.go:117] "RemoveContainer" containerID="a99cabe30a7abc8da854bf86d79628b94a6ffc20c28f828b695080332b319db2" Mar 13 14:07:17 crc kubenswrapper[4907]: E0313 14:07:17.475602 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-qxlx5_openshift-ovn-kubernetes(1006376d-4033-454e-8a23-2b67882bdd0b)\"" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" podUID="1006376d-4033-454e-8a23-2b67882bdd0b" Mar 13 14:07:17 crc kubenswrapper[4907]: I0313 14:07:17.477674 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:07:17 crc kubenswrapper[4907]: I0313 14:07:17.477708 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:07:17 crc kubenswrapper[4907]: I0313 14:07:17.477724 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:07:17 crc kubenswrapper[4907]: I0313 14:07:17.477741 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:07:17 crc kubenswrapper[4907]: I0313 14:07:17.477755 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:07:17Z","lastTransitionTime":"2026-03-13T14:07:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:07:17 crc kubenswrapper[4907]: I0313 14:07:17.488294 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5b4819028f34db81cc43a04e4854c30488ed876c63b8daff11105bf61546c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:17Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:17 crc kubenswrapper[4907]: E0313 14:07:17.488917 4907 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4b182bae-7f2d-4b54-934c-85a9a277d132\\\",\\\"systemUUID\\\":\\\"05bfbfa3-f04f-45ba-8233-a355fc774c24\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:17Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:17 crc kubenswrapper[4907]: I0313 14:07:17.492060 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:07:17 crc kubenswrapper[4907]: I0313 14:07:17.492092 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:07:17 crc kubenswrapper[4907]: I0313 14:07:17.492104 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:07:17 crc kubenswrapper[4907]: I0313 14:07:17.492121 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:07:17 crc kubenswrapper[4907]: I0313 14:07:17.492133 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:07:17Z","lastTransitionTime":"2026-03-13T14:07:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:07:17 crc kubenswrapper[4907]: I0313 14:07:17.503593 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:17Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:17 crc kubenswrapper[4907]: E0313 14:07:17.510373 4907 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4b182bae-7f2d-4b54-934c-85a9a277d132\\\",\\\"systemUUID\\\":\\\"05bfbfa3-f04f-45ba-8233-a355fc774c24\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:17Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:17 crc kubenswrapper[4907]: I0313 14:07:17.514583 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:07:17 crc kubenswrapper[4907]: I0313 14:07:17.514728 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:07:17 crc kubenswrapper[4907]: I0313 14:07:17.514745 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:07:17 crc kubenswrapper[4907]: I0313 14:07:17.514772 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:07:17 crc kubenswrapper[4907]: I0313 14:07:17.514792 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:07:17Z","lastTransitionTime":"2026-03-13T14:07:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:07:17 crc kubenswrapper[4907]: I0313 14:07:17.518824 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xtwf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64ebd338-0158-44dc-90ab-e7c285e87762\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://391b8bc1e975982b172175fcf4b4509756b018cbb99c12e48f4931ff472cdae2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6cwbq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xtwf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:17Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:17 crc kubenswrapper[4907]: E0313 14:07:17.525816 4907 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4b182bae-7f2d-4b54-934c-85a9a277d132\\\",\\\"systemUUID\\\":\\\"05bfbfa3-f04f-45ba-8233-a355fc774c24\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:17Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:17 crc kubenswrapper[4907]: E0313 14:07:17.526089 4907 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 13 14:07:17 crc kubenswrapper[4907]: I0313 14:07:17.528295 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-tzmvc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f797b1bb-e769-49c4-84ce-93a6fad2df93\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tdzl9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tdzl9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:07:00Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-tzmvc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:17Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:17 crc kubenswrapper[4907]: I0313 14:07:17.538843 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:17Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:17 crc kubenswrapper[4907]: I0313 14:07:17.550234 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afb1d430459aafab2d6bc2507b4b4059529401e51fdf8e9848b0d7bb8651db68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:17Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:17 crc kubenswrapper[4907]: I0313 14:07:17.560061 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-745zx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58e11265-2fd9-4b9b-926c-3343bb84ee53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ce7ad32fd2bf2332f3ee41080115b083c34326b7dc31c950098dfba15f58422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mnz7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-745zx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:17Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:17 crc kubenswrapper[4907]: I0313 14:07:17.571152 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3fb4213-144f-4b01-9d78-ff3315761bb3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba1140a2462713c2401b47a4cf1f231bc9d9ae7867a7d0b1b597214a1f7efe07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://770c882d9e9f9b2e9bd4f135e591952fa15b3898c42826aed8a75d953c9a875b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hm56j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:17Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:17 crc kubenswrapper[4907]: I0313 14:07:17.590845 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1006376d-4033-454e-8a23-2b67882bdd0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf6a3dbaa8dda61f461a906538824c8946d36dc07d4fb95bf4d8420ad0105600\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccef92e41501fab91189581d7e8875515c02d637426b329d2fb80a356c75b637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0ee0049aa4a6d49a6bcf7d347183d9cc8b85171f988662803a005e0cdd97395\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fff87d8ea572615eb1329164f9e536e6ee45cd00a9e71be0aa8a9b3ed2190e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c165053fd9933444b372ff40efe33c0e193a7f39e6f531a4f370e971eb41e6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fea41c11bebd2d9d622ed4b8bc50e03a106dadea6ac102cd60512ea2a1f8b8b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a99cabe30a7abc8da854bf86d79628b94a6ffc20c28f828b695080332b319db2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a99cabe30a7abc8da854bf86d79628b94a6ffc20c28f828b695080332b319db2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T14:07:15Z\\\",\\\"message\\\":\\\"Sending *v1.Node event handler 7 for removal\\\\nI0313 14:07:15.619307 7162 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0313 14:07:15.619320 7162 handler.go:208] Removed *v1.Node event handler 2\\\\nI0313 14:07:15.619329 7162 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0313 14:07:15.619343 7162 handler.go:208] Removed *v1.Node event handler 7\\\\nI0313 14:07:15.619351 7162 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0313 14:07:15.619358 7162 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0313 14:07:15.619362 7162 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0313 14:07:15.619381 7162 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0313 14:07:15.619382 7162 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0313 14:07:15.619392 7162 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0313 14:07:15.619619 7162 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI0313 14:07:15.619713 7162 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI0313 14:07:15.619975 7162 factory.go:656] Stopping watch factory\\\\nI0313 14:07:15.620001 7162 ovnkube.go:599] Stopped ovnkube\\\\nI0313 14:07:15.620033 7162 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0313 14:07:15.620096 7162 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:07:14Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-qxlx5_openshift-ovn-kubernetes(1006376d-4033-454e-8a23-2b67882bdd0b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d60b5b774f8dd4cee16b9784f80d124640199a85331a2940772029be4884751\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qxlx5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:17Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:17 crc kubenswrapper[4907]: I0313 14:07:17.600601 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"584d221c-be74-4a39-8173-6c8e3f55f686\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b782a795257057929fe616938d1483a9accbd7a93f0cba963f9791a3d4ec4749\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c826e7b9be5e3a0ecdaa6dfeaf2454535e542ff86acc532cc3c80acf4a95c1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4333bdf047392cb3f64e29e677fe8d2f21c7ebca1a7e7097167e70c6785b8d46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a4dea84b35a82ce16f0e0d38057dc994b26240cf45dcc9111ed0de573a62881\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a4dea84b35a82ce16f0e0d38057dc994b26240cf45dcc9111ed0de573a62881\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:17Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:17 crc kubenswrapper[4907]: I0313 14:07:17.614594 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:17Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:17 crc kubenswrapper[4907]: I0313 14:07:17.629602 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-p8ptc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57e4e912-9a98-4419-87ce-f16e178ed921\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38f113418d62dab1728f0ba57cd1d289e3ec41621a338017e74c9996a8449b1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bdef68f401b5ba2c7cf61551eba416e74f295ab2281433a5115334d28094818\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bdef68f401b5ba2c7cf61551eba416e74f295ab2281433a5115334d28094818\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f8d1f40744d609fd964572ef5e323f0283d2582e98745602754165818f17847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f8d1f40744d609fd964572ef5e323f0283d2582e98745602754165818f17847\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c1ab2e148eaa177b98f5883a3f323d12f5b662bff8cb2d072a97be18ab8e120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c1ab2e148eaa177b98f5883a3f323d12f5b662bff8cb2d072a97be18ab8e120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-p8ptc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:17Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:17 crc kubenswrapper[4907]: I0313 14:07:17.639552 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qncw4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1d7ee39-09d7-4c0e-942b-5e9d8c9f63be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07df71887196cc7098ac9606a4a15b89d45219048c31280198ae75022bff9370\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:07:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bg45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://540ca207f31c9a8a9892f80dee153b3fb86f8de406f38360221512f424d8c884\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:07:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bg45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:07:00Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-qncw4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:17Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:17 crc kubenswrapper[4907]: I0313 14:07:17.653584 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5d2b054-0ed8-419c-b731-f14d3598372f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4b503aa3b13268b3688d067f3e456c86afac1fda171aced0990899c3eb89b79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be40a1a33fa482eab363537afaa901127d560afbc1b5b287ef11e66391a5be4c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d1b539bbd918fc50428fc18cabcc4c8a43cc96f220b5692d67cf1c8dfbf8ed8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93d8c8c79be9d9ffb72de66fc2f3f9821936023071e95875582b45ccb319601f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T14:06:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0313 14:06:13.202871 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0313 14:06:13.203175 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0313 14:06:13.204427 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2471728650/tls.crt::/tmp/serving-cert-2471728650/tls.key\\\\\\\"\\\\nI0313 14:06:13.581160 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0313 14:06:13.584012 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0313 14:06:13.584036 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0313 14:06:13.584064 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0313 14:06:13.584072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0313 14:06:13.594504 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0313 14:06:13.594891 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0313 14:06:13.595323 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0313 14:06:13.595363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0313 14:06:13.595370 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0313 14:06:13.595376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0313 14:06:13.597226 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3f2b6756078925f8562b6053b3b83798efb72288f5fd18fff680a4ab0d33feb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:17Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:17 crc kubenswrapper[4907]: I0313 14:07:17.666800 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94b045cf2fed56457a2ca14f489bb69470198f159f16c207fe061d034dba1555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd28f7a5c35822582b70ed41645294aaede10e28ff59f3f3b5d36e7da8237b76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:17Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:17 crc kubenswrapper[4907]: I0313 14:07:17.676322 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l7zg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d105034b-f041-4fb5-8df6-6b1bb7681fda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21a207f917836785e320f4f79aca5d6449da2fdeec3ee0765f34d227d5617fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szmh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l7zg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:17Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:18 crc kubenswrapper[4907]: I0313 14:07:18.781662 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:07:18 crc kubenswrapper[4907]: I0313 14:07:18.781947 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:07:18 crc kubenswrapper[4907]: I0313 14:07:18.782062 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:07:18 crc kubenswrapper[4907]: I0313 14:07:18.782074 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:07:18 crc kubenswrapper[4907]: E0313 14:07:18.782059 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:07:18 crc kubenswrapper[4907]: E0313 14:07:18.782198 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:07:18 crc kubenswrapper[4907]: E0313 14:07:18.782308 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzmvc" podUID="f797b1bb-e769-49c4-84ce-93a6fad2df93" Mar 13 14:07:18 crc kubenswrapper[4907]: E0313 14:07:18.782380 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:07:20 crc kubenswrapper[4907]: I0313 14:07:20.782493 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:07:20 crc kubenswrapper[4907]: I0313 14:07:20.782537 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:07:20 crc kubenswrapper[4907]: I0313 14:07:20.782539 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:07:20 crc kubenswrapper[4907]: E0313 14:07:20.782683 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:07:20 crc kubenswrapper[4907]: I0313 14:07:20.782712 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:07:20 crc kubenswrapper[4907]: E0313 14:07:20.782925 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:07:20 crc kubenswrapper[4907]: E0313 14:07:20.783116 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzmvc" podUID="f797b1bb-e769-49c4-84ce-93a6fad2df93" Mar 13 14:07:20 crc kubenswrapper[4907]: E0313 14:07:20.783191 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:07:21 crc kubenswrapper[4907]: I0313 14:07:21.792335 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"584d221c-be74-4a39-8173-6c8e3f55f686\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b782a795257057929fe616938d1483a9accbd7a93f0cba963f9791a3d4ec4749\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c826e7b9be5e3a0ecdaa6dfeaf2454535e542ff86acc532cc3c80acf4a95c1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4333bdf047392cb3f64e29e677fe8d2f21c7ebca1a7e7097167e70c6785b8d46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a4dea84b35a82ce16f0e0d38057dc994b26240cf45dcc9111ed0de573a62881\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a4dea84b35a82ce16f0e0d38057dc994b26240cf45dcc9111ed0de573a62881\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:21Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:21 crc kubenswrapper[4907]: I0313 14:07:21.801831 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:21Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:21 crc kubenswrapper[4907]: I0313 14:07:21.813891 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-p8ptc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57e4e912-9a98-4419-87ce-f16e178ed921\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38f113418d62dab1728f0ba57cd1d289e3ec41621a338017e74c9996a8449b1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bdef68f401b5ba2c7cf61551eba416e74f295ab2281433a5115334d28094818\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bdef68f401b5ba2c7cf61551eba416e74f295ab2281433a5115334d28094818\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f8d1f40744d609fd964572ef5e323f0283d2582e98745602754165818f17847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f8d1f40744d609fd964572ef5e323f0283d2582e98745602754165818f17847\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c1ab2e148eaa177b98f5883a3f323d12f5b662bff8cb2d072a97be18ab8e120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c1ab2e148eaa177b98f5883a3f323d12f5b662bff8cb2d072a97be18ab8e120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-p8ptc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:21Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:21 crc kubenswrapper[4907]: I0313 14:07:21.824048 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3fb4213-144f-4b01-9d78-ff3315761bb3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba1140a2462713c2401b47a4cf1f231bc9d9ae7867a7d0b1b597214a1f7efe07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://770c882d9e9f9b2e9bd4f135e591952fa15b3898c42826aed8a75d953c9a875b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hm56j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:21Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:21 crc kubenswrapper[4907]: I0313 14:07:21.841851 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1006376d-4033-454e-8a23-2b67882bdd0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf6a3dbaa8dda61f461a906538824c8946d36dc07d4fb95bf4d8420ad0105600\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccef92e41501fab91189581d7e8875515c02d637426b329d2fb80a356c75b637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0ee0049aa4a6d49a6bcf7d347183d9cc8b85171f988662803a005e0cdd97395\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fff87d8ea572615eb1329164f9e536e6ee45cd00a9e71be0aa8a9b3ed2190e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c165053fd9933444b372ff40efe33c0e193a7f39e6f531a4f370e971eb41e6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fea41c11bebd2d9d622ed4b8bc50e03a106dadea6ac102cd60512ea2a1f8b8b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a99cabe30a7abc8da854bf86d79628b94a6ffc20c28f828b695080332b319db2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a99cabe30a7abc8da854bf86d79628b94a6ffc20c28f828b695080332b319db2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T14:07:15Z\\\",\\\"message\\\":\\\"Sending *v1.Node event handler 7 for removal\\\\nI0313 14:07:15.619307 7162 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0313 14:07:15.619320 7162 handler.go:208] Removed *v1.Node event handler 2\\\\nI0313 14:07:15.619329 7162 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0313 14:07:15.619343 7162 handler.go:208] Removed *v1.Node event handler 7\\\\nI0313 14:07:15.619351 7162 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0313 14:07:15.619358 7162 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0313 14:07:15.619362 7162 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0313 14:07:15.619381 7162 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0313 14:07:15.619382 7162 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0313 14:07:15.619392 7162 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0313 14:07:15.619619 7162 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI0313 14:07:15.619713 7162 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI0313 14:07:15.619975 7162 factory.go:656] Stopping watch factory\\\\nI0313 14:07:15.620001 7162 ovnkube.go:599] Stopped ovnkube\\\\nI0313 14:07:15.620033 7162 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0313 14:07:15.620096 7162 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:07:14Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-qxlx5_openshift-ovn-kubernetes(1006376d-4033-454e-8a23-2b67882bdd0b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d60b5b774f8dd4cee16b9784f80d124640199a85331a2940772029be4884751\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qxlx5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:21Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:21 crc kubenswrapper[4907]: I0313 14:07:21.856007 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5d2b054-0ed8-419c-b731-f14d3598372f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4b503aa3b13268b3688d067f3e456c86afac1fda171aced0990899c3eb89b79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be40a1a33fa482eab363537afaa901127d560afbc1b5b287ef11e66391a5be4c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d1b539bbd918fc50428fc18cabcc4c8a43cc96f220b5692d67cf1c8dfbf8ed8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93d8c8c79be9d9ffb72de66fc2f3f9821936023071e95875582b45ccb319601f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T14:06:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0313 14:06:13.202871 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0313 14:06:13.203175 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0313 14:06:13.204427 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2471728650/tls.crt::/tmp/serving-cert-2471728650/tls.key\\\\\\\"\\\\nI0313 14:06:13.581160 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0313 14:06:13.584012 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0313 14:06:13.584036 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0313 14:06:13.584064 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0313 14:06:13.584072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0313 14:06:13.594504 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0313 14:06:13.594891 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0313 14:06:13.595323 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0313 14:06:13.595363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0313 14:06:13.595370 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0313 14:06:13.595376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0313 14:06:13.597226 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3f2b6756078925f8562b6053b3b83798efb72288f5fd18fff680a4ab0d33feb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:21Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:21 crc kubenswrapper[4907]: I0313 14:07:21.869678 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94b045cf2fed56457a2ca14f489bb69470198f159f16c207fe061d034dba1555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd28f7a5c35822582b70ed41645294aaede10e28ff59f3f3b5d36e7da8237b76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:21Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:21 crc kubenswrapper[4907]: I0313 14:07:21.881308 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l7zg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d105034b-f041-4fb5-8df6-6b1bb7681fda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21a207f917836785e320f4f79aca5d6449da2fdeec3ee0765f34d227d5617fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szmh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l7zg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:21Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:21 crc kubenswrapper[4907]: E0313 14:07:21.890707 4907 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 13 14:07:21 crc kubenswrapper[4907]: I0313 14:07:21.896126 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qncw4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1d7ee39-09d7-4c0e-942b-5e9d8c9f63be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07df71887196cc7098ac9606a4a15b89d45219048c31280198ae75022bff9370\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:07:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bg45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://540ca207f31c9a8a9892f80dee153b3fb86f8de406f38360221512f424d8c884\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:07:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bg45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:07:00Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-qncw4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:21Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:21 crc kubenswrapper[4907]: I0313 14:07:21.910661 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5b4819028f34db81cc43a04e4854c30488ed876c63b8daff11105bf61546c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:21Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:21 crc kubenswrapper[4907]: I0313 14:07:21.924006 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:21Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:21 crc kubenswrapper[4907]: I0313 14:07:21.936392 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xtwf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64ebd338-0158-44dc-90ab-e7c285e87762\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://391b8bc1e975982b172175fcf4b4509756b018cbb99c12e48f4931ff472cdae2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6cwbq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xtwf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:21Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:21 crc kubenswrapper[4907]: I0313 14:07:21.951858 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:21Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:21 crc kubenswrapper[4907]: I0313 14:07:21.965671 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afb1d430459aafab2d6bc2507b4b4059529401e51fdf8e9848b0d7bb8651db68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:21Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:21 crc kubenswrapper[4907]: I0313 14:07:21.981405 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-745zx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58e11265-2fd9-4b9b-926c-3343bb84ee53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ce7ad32fd2bf2332f3ee41080115b083c34326b7dc31c950098dfba15f58422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mnz7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-745zx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:21Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:21 crc kubenswrapper[4907]: I0313 14:07:21.996426 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-tzmvc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f797b1bb-e769-49c4-84ce-93a6fad2df93\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tdzl9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tdzl9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:07:00Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-tzmvc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:21Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:22 crc kubenswrapper[4907]: I0313 14:07:22.781341 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:07:22 crc kubenswrapper[4907]: E0313 14:07:22.781480 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:07:22 crc kubenswrapper[4907]: I0313 14:07:22.781341 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:07:22 crc kubenswrapper[4907]: I0313 14:07:22.781360 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:07:22 crc kubenswrapper[4907]: E0313 14:07:22.781841 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzmvc" podUID="f797b1bb-e769-49c4-84ce-93a6fad2df93" Mar 13 14:07:22 crc kubenswrapper[4907]: E0313 14:07:22.781955 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:07:22 crc kubenswrapper[4907]: I0313 14:07:22.782062 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:07:22 crc kubenswrapper[4907]: E0313 14:07:22.782140 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:07:24 crc kubenswrapper[4907]: I0313 14:07:24.781355 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:07:24 crc kubenswrapper[4907]: I0313 14:07:24.781436 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:07:24 crc kubenswrapper[4907]: I0313 14:07:24.781458 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:07:24 crc kubenswrapper[4907]: I0313 14:07:24.781396 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:07:24 crc kubenswrapper[4907]: E0313 14:07:24.781585 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:07:24 crc kubenswrapper[4907]: E0313 14:07:24.781725 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:07:24 crc kubenswrapper[4907]: E0313 14:07:24.781859 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:07:24 crc kubenswrapper[4907]: E0313 14:07:24.782028 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzmvc" podUID="f797b1bb-e769-49c4-84ce-93a6fad2df93" Mar 13 14:07:26 crc kubenswrapper[4907]: I0313 14:07:26.782222 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:07:26 crc kubenswrapper[4907]: I0313 14:07:26.782266 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:07:26 crc kubenswrapper[4907]: E0313 14:07:26.782420 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzmvc" podUID="f797b1bb-e769-49c4-84ce-93a6fad2df93" Mar 13 14:07:26 crc kubenswrapper[4907]: I0313 14:07:26.782475 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:07:26 crc kubenswrapper[4907]: I0313 14:07:26.782582 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:07:26 crc kubenswrapper[4907]: E0313 14:07:26.782598 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:07:26 crc kubenswrapper[4907]: E0313 14:07:26.782746 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:07:26 crc kubenswrapper[4907]: E0313 14:07:26.782826 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:07:26 crc kubenswrapper[4907]: E0313 14:07:26.891536 4907 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 13 14:07:27 crc kubenswrapper[4907]: I0313 14:07:27.837877 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:07:27 crc kubenswrapper[4907]: I0313 14:07:27.838875 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:07:27 crc kubenswrapper[4907]: I0313 14:07:27.839290 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:07:27 crc kubenswrapper[4907]: I0313 14:07:27.839607 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:07:27 crc kubenswrapper[4907]: I0313 14:07:27.839932 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:07:27Z","lastTransitionTime":"2026-03-13T14:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:07:27 crc kubenswrapper[4907]: E0313 14:07:27.860066 4907 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4b182bae-7f2d-4b54-934c-85a9a277d132\\\",\\\"systemUUID\\\":\\\"05bfbfa3-f04f-45ba-8233-a355fc774c24\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:27Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:27 crc kubenswrapper[4907]: I0313 14:07:27.865802 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:07:27 crc kubenswrapper[4907]: I0313 14:07:27.865852 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:07:27 crc kubenswrapper[4907]: I0313 14:07:27.865863 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:07:27 crc kubenswrapper[4907]: I0313 14:07:27.865900 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:07:27 crc kubenswrapper[4907]: I0313 14:07:27.865913 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:07:27Z","lastTransitionTime":"2026-03-13T14:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:07:27 crc kubenswrapper[4907]: E0313 14:07:27.890088 4907 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4b182bae-7f2d-4b54-934c-85a9a277d132\\\",\\\"systemUUID\\\":\\\"05bfbfa3-f04f-45ba-8233-a355fc774c24\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:27Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:27 crc kubenswrapper[4907]: I0313 14:07:27.895662 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:07:27 crc kubenswrapper[4907]: I0313 14:07:27.895717 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:07:27 crc kubenswrapper[4907]: I0313 14:07:27.895737 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:07:27 crc kubenswrapper[4907]: I0313 14:07:27.895760 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:07:27 crc kubenswrapper[4907]: I0313 14:07:27.895778 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:07:27Z","lastTransitionTime":"2026-03-13T14:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:07:27 crc kubenswrapper[4907]: E0313 14:07:27.918045 4907 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4b182bae-7f2d-4b54-934c-85a9a277d132\\\",\\\"systemUUID\\\":\\\"05bfbfa3-f04f-45ba-8233-a355fc774c24\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:27Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:27 crc kubenswrapper[4907]: I0313 14:07:27.924171 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:07:27 crc kubenswrapper[4907]: I0313 14:07:27.924219 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:07:27 crc kubenswrapper[4907]: I0313 14:07:27.924237 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:07:27 crc kubenswrapper[4907]: I0313 14:07:27.924261 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:07:27 crc kubenswrapper[4907]: I0313 14:07:27.924277 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:07:27Z","lastTransitionTime":"2026-03-13T14:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:07:27 crc kubenswrapper[4907]: E0313 14:07:27.944493 4907 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4b182bae-7f2d-4b54-934c-85a9a277d132\\\",\\\"systemUUID\\\":\\\"05bfbfa3-f04f-45ba-8233-a355fc774c24\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:27Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:27 crc kubenswrapper[4907]: I0313 14:07:27.949448 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:07:27 crc kubenswrapper[4907]: I0313 14:07:27.949546 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:07:27 crc kubenswrapper[4907]: I0313 14:07:27.949563 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:07:27 crc kubenswrapper[4907]: I0313 14:07:27.949586 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:07:27 crc kubenswrapper[4907]: I0313 14:07:27.949603 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:07:27Z","lastTransitionTime":"2026-03-13T14:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:07:27 crc kubenswrapper[4907]: E0313 14:07:27.970052 4907 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4b182bae-7f2d-4b54-934c-85a9a277d132\\\",\\\"systemUUID\\\":\\\"05bfbfa3-f04f-45ba-8233-a355fc774c24\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:27Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:27 crc kubenswrapper[4907]: E0313 14:07:27.970580 4907 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 13 14:07:28 crc kubenswrapper[4907]: I0313 14:07:28.782203 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:07:28 crc kubenswrapper[4907]: I0313 14:07:28.782236 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:07:28 crc kubenswrapper[4907]: I0313 14:07:28.782261 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:07:28 crc kubenswrapper[4907]: E0313 14:07:28.782389 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:07:28 crc kubenswrapper[4907]: E0313 14:07:28.782633 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:07:28 crc kubenswrapper[4907]: I0313 14:07:28.782817 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:07:28 crc kubenswrapper[4907]: E0313 14:07:28.783273 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzmvc" podUID="f797b1bb-e769-49c4-84ce-93a6fad2df93" Mar 13 14:07:28 crc kubenswrapper[4907]: E0313 14:07:28.783332 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:07:28 crc kubenswrapper[4907]: I0313 14:07:28.783747 4907 scope.go:117] "RemoveContainer" containerID="a99cabe30a7abc8da854bf86d79628b94a6ffc20c28f828b695080332b319db2" Mar 13 14:07:28 crc kubenswrapper[4907]: E0313 14:07:28.784156 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-qxlx5_openshift-ovn-kubernetes(1006376d-4033-454e-8a23-2b67882bdd0b)\"" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" podUID="1006376d-4033-454e-8a23-2b67882bdd0b" Mar 13 14:07:29 crc kubenswrapper[4907]: I0313 14:07:29.798497 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Mar 13 14:07:30 crc kubenswrapper[4907]: I0313 14:07:30.781347 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:07:30 crc kubenswrapper[4907]: I0313 14:07:30.781383 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:07:30 crc kubenswrapper[4907]: I0313 14:07:30.781366 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:07:30 crc kubenswrapper[4907]: I0313 14:07:30.781971 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:07:30 crc kubenswrapper[4907]: E0313 14:07:30.782270 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:07:30 crc kubenswrapper[4907]: E0313 14:07:30.782612 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzmvc" podUID="f797b1bb-e769-49c4-84ce-93a6fad2df93" Mar 13 14:07:30 crc kubenswrapper[4907]: E0313 14:07:30.782784 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:07:30 crc kubenswrapper[4907]: E0313 14:07:30.782696 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:07:31 crc kubenswrapper[4907]: I0313 14:07:31.803310 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:31Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:31 crc kubenswrapper[4907]: I0313 14:07:31.823697 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xtwf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64ebd338-0158-44dc-90ab-e7c285e87762\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://391b8bc1e975982b172175fcf4b4509756b018cbb99c12e48f4931ff472cdae2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6cwbq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xtwf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:31Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:31 crc kubenswrapper[4907]: I0313 14:07:31.843604 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5b4819028f34db81cc43a04e4854c30488ed876c63b8daff11105bf61546c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:31Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:31 crc kubenswrapper[4907]: I0313 14:07:31.862806 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:31Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:31 crc kubenswrapper[4907]: I0313 14:07:31.880822 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afb1d430459aafab2d6bc2507b4b4059529401e51fdf8e9848b0d7bb8651db68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:31Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:31 crc kubenswrapper[4907]: E0313 14:07:31.892325 4907 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 13 14:07:31 crc kubenswrapper[4907]: I0313 14:07:31.901126 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-745zx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58e11265-2fd9-4b9b-926c-3343bb84ee53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ce7ad32fd2bf2332f3ee41080115b083c34326b7dc31c950098dfba15f58422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mnz7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-745zx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:31Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:31 crc kubenswrapper[4907]: I0313 14:07:31.920482 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-tzmvc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f797b1bb-e769-49c4-84ce-93a6fad2df93\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tdzl9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tdzl9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:07:00Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-tzmvc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:31Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:31 crc kubenswrapper[4907]: I0313 14:07:31.940170 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bed3bcea-e8c0-4b00-a985-e0c5b5bea100\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://402c1ef89085d840edb4313577f5b2fb967072cb3db00c34c98572f00699529c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f22becea3c68103ff0b2639c7ae8c051b76d19b407b0463645fc2e4a0682f23\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T14:05:32Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0313 14:05:03.905343 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0313 14:05:03.908030 1 observer_polling.go:159] Starting file observer\\\\nI0313 14:05:03.938230 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0313 14:05:03.942399 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0313 14:05:32.345768 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0313 14:05:32.345932 1 cmd.go:179] failed checking apiserver connectivity: Get \\\\\\\"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/openshift-kube-controller-manager/leases/cluster-policy-controller-lock\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:31Z is after 2026-02-23T05:33:13Z\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:03Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18eadc065bf4a307b7b443fcfe79b2b719938ee486b53321c0a9f20926a0a887\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9abfa5a9ad1a98a2907ead28390ee926e176646fe761b89f580101bd429cf747\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbc67477e0223db1e4b403be3709fe71fe6409570361e103eb730fdeaffe3e80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:31Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:31 crc kubenswrapper[4907]: I0313 14:07:31.954061 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"584d221c-be74-4a39-8173-6c8e3f55f686\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b782a795257057929fe616938d1483a9accbd7a93f0cba963f9791a3d4ec4749\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c826e7b9be5e3a0ecdaa6dfeaf2454535e542ff86acc532cc3c80acf4a95c1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4333bdf047392cb3f64e29e677fe8d2f21c7ebca1a7e7097167e70c6785b8d46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a4dea84b35a82ce16f0e0d38057dc994b26240cf45dcc9111ed0de573a62881\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a4dea84b35a82ce16f0e0d38057dc994b26240cf45dcc9111ed0de573a62881\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:31Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:31 crc kubenswrapper[4907]: I0313 14:07:31.965824 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:31Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:31 crc kubenswrapper[4907]: I0313 14:07:31.987853 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-p8ptc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57e4e912-9a98-4419-87ce-f16e178ed921\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38f113418d62dab1728f0ba57cd1d289e3ec41621a338017e74c9996a8449b1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bdef68f401b5ba2c7cf61551eba416e74f295ab2281433a5115334d28094818\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bdef68f401b5ba2c7cf61551eba416e74f295ab2281433a5115334d28094818\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f8d1f40744d609fd964572ef5e323f0283d2582e98745602754165818f17847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f8d1f40744d609fd964572ef5e323f0283d2582e98745602754165818f17847\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c1ab2e148eaa177b98f5883a3f323d12f5b662bff8cb2d072a97be18ab8e120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c1ab2e148eaa177b98f5883a3f323d12f5b662bff8cb2d072a97be18ab8e120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-p8ptc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:31Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:32 crc kubenswrapper[4907]: I0313 14:07:32.003051 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3fb4213-144f-4b01-9d78-ff3315761bb3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba1140a2462713c2401b47a4cf1f231bc9d9ae7867a7d0b1b597214a1f7efe07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://770c882d9e9f9b2e9bd4f135e591952fa15b3898c42826aed8a75d953c9a875b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hm56j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:32Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:32 crc kubenswrapper[4907]: I0313 14:07:32.022034 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1006376d-4033-454e-8a23-2b67882bdd0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf6a3dbaa8dda61f461a906538824c8946d36dc07d4fb95bf4d8420ad0105600\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccef92e41501fab91189581d7e8875515c02d637426b329d2fb80a356c75b637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0ee0049aa4a6d49a6bcf7d347183d9cc8b85171f988662803a005e0cdd97395\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fff87d8ea572615eb1329164f9e536e6ee45cd00a9e71be0aa8a9b3ed2190e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c165053fd9933444b372ff40efe33c0e193a7f39e6f531a4f370e971eb41e6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fea41c11bebd2d9d622ed4b8bc50e03a106dadea6ac102cd60512ea2a1f8b8b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a99cabe30a7abc8da854bf86d79628b94a6ffc20c28f828b695080332b319db2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a99cabe30a7abc8da854bf86d79628b94a6ffc20c28f828b695080332b319db2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T14:07:15Z\\\",\\\"message\\\":\\\"Sending *v1.Node event handler 7 for removal\\\\nI0313 14:07:15.619307 7162 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0313 14:07:15.619320 7162 handler.go:208] Removed *v1.Node event handler 2\\\\nI0313 14:07:15.619329 7162 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0313 14:07:15.619343 7162 handler.go:208] Removed *v1.Node event handler 7\\\\nI0313 14:07:15.619351 7162 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0313 14:07:15.619358 7162 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0313 14:07:15.619362 7162 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0313 14:07:15.619381 7162 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0313 14:07:15.619382 7162 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0313 14:07:15.619392 7162 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0313 14:07:15.619619 7162 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI0313 14:07:15.619713 7162 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI0313 14:07:15.619975 7162 factory.go:656] Stopping watch factory\\\\nI0313 14:07:15.620001 7162 ovnkube.go:599] Stopped ovnkube\\\\nI0313 14:07:15.620033 7162 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0313 14:07:15.620096 7162 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:07:14Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-qxlx5_openshift-ovn-kubernetes(1006376d-4033-454e-8a23-2b67882bdd0b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d60b5b774f8dd4cee16b9784f80d124640199a85331a2940772029be4884751\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qxlx5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:32Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:32 crc kubenswrapper[4907]: I0313 14:07:32.038022 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5d2b054-0ed8-419c-b731-f14d3598372f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4b503aa3b13268b3688d067f3e456c86afac1fda171aced0990899c3eb89b79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be40a1a33fa482eab363537afaa901127d560afbc1b5b287ef11e66391a5be4c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d1b539bbd918fc50428fc18cabcc4c8a43cc96f220b5692d67cf1c8dfbf8ed8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93d8c8c79be9d9ffb72de66fc2f3f9821936023071e95875582b45ccb319601f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T14:06:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0313 14:06:13.202871 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0313 14:06:13.203175 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0313 14:06:13.204427 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2471728650/tls.crt::/tmp/serving-cert-2471728650/tls.key\\\\\\\"\\\\nI0313 14:06:13.581160 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0313 14:06:13.584012 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0313 14:06:13.584036 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0313 14:06:13.584064 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0313 14:06:13.584072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0313 14:06:13.594504 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0313 14:06:13.594891 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0313 14:06:13.595323 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0313 14:06:13.595363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0313 14:06:13.595370 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0313 14:06:13.595376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0313 14:06:13.597226 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3f2b6756078925f8562b6053b3b83798efb72288f5fd18fff680a4ab0d33feb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:32Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:32 crc kubenswrapper[4907]: I0313 14:07:32.053173 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94b045cf2fed56457a2ca14f489bb69470198f159f16c207fe061d034dba1555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd28f7a5c35822582b70ed41645294aaede10e28ff59f3f3b5d36e7da8237b76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:32Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:32 crc kubenswrapper[4907]: I0313 14:07:32.068456 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l7zg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d105034b-f041-4fb5-8df6-6b1bb7681fda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21a207f917836785e320f4f79aca5d6449da2fdeec3ee0765f34d227d5617fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szmh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l7zg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:32Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:32 crc kubenswrapper[4907]: I0313 14:07:32.084093 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qncw4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1d7ee39-09d7-4c0e-942b-5e9d8c9f63be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07df71887196cc7098ac9606a4a15b89d45219048c31280198ae75022bff9370\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:07:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bg45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://540ca207f31c9a8a9892f80dee153b3fb86f8de406f38360221512f424d8c884\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:07:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bg45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:07:00Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-qncw4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:32Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:32 crc kubenswrapper[4907]: I0313 14:07:32.743875 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f797b1bb-e769-49c4-84ce-93a6fad2df93-metrics-certs\") pod \"network-metrics-daemon-tzmvc\" (UID: \"f797b1bb-e769-49c4-84ce-93a6fad2df93\") " pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:07:32 crc kubenswrapper[4907]: E0313 14:07:32.744037 4907 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 13 14:07:32 crc kubenswrapper[4907]: E0313 14:07:32.744088 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f797b1bb-e769-49c4-84ce-93a6fad2df93-metrics-certs podName:f797b1bb-e769-49c4-84ce-93a6fad2df93 nodeName:}" failed. No retries permitted until 2026-03-13 14:08:04.744075243 +0000 UTC m=+183.643862932 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/f797b1bb-e769-49c4-84ce-93a6fad2df93-metrics-certs") pod "network-metrics-daemon-tzmvc" (UID: "f797b1bb-e769-49c4-84ce-93a6fad2df93") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 13 14:07:32 crc kubenswrapper[4907]: I0313 14:07:32.781796 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:07:32 crc kubenswrapper[4907]: I0313 14:07:32.781827 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:07:32 crc kubenswrapper[4907]: I0313 14:07:32.781911 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:07:32 crc kubenswrapper[4907]: E0313 14:07:32.782005 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:07:32 crc kubenswrapper[4907]: I0313 14:07:32.782024 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:07:32 crc kubenswrapper[4907]: E0313 14:07:32.782118 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:07:32 crc kubenswrapper[4907]: E0313 14:07:32.782218 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:07:32 crc kubenswrapper[4907]: E0313 14:07:32.782302 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzmvc" podUID="f797b1bb-e769-49c4-84ce-93a6fad2df93" Mar 13 14:07:34 crc kubenswrapper[4907]: I0313 14:07:34.781351 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:07:34 crc kubenswrapper[4907]: I0313 14:07:34.781387 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:07:34 crc kubenswrapper[4907]: E0313 14:07:34.781875 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:07:34 crc kubenswrapper[4907]: I0313 14:07:34.781439 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:07:34 crc kubenswrapper[4907]: E0313 14:07:34.781985 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzmvc" podUID="f797b1bb-e769-49c4-84ce-93a6fad2df93" Mar 13 14:07:34 crc kubenswrapper[4907]: I0313 14:07:34.781470 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:07:34 crc kubenswrapper[4907]: E0313 14:07:34.782082 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:07:34 crc kubenswrapper[4907]: E0313 14:07:34.782311 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:07:35 crc kubenswrapper[4907]: I0313 14:07:35.545937 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-xtwf7_64ebd338-0158-44dc-90ab-e7c285e87762/kube-multus/0.log" Mar 13 14:07:35 crc kubenswrapper[4907]: I0313 14:07:35.546034 4907 generic.go:334] "Generic (PLEG): container finished" podID="64ebd338-0158-44dc-90ab-e7c285e87762" containerID="391b8bc1e975982b172175fcf4b4509756b018cbb99c12e48f4931ff472cdae2" exitCode=1 Mar 13 14:07:35 crc kubenswrapper[4907]: I0313 14:07:35.546101 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-xtwf7" event={"ID":"64ebd338-0158-44dc-90ab-e7c285e87762","Type":"ContainerDied","Data":"391b8bc1e975982b172175fcf4b4509756b018cbb99c12e48f4931ff472cdae2"} Mar 13 14:07:35 crc kubenswrapper[4907]: I0313 14:07:35.547028 4907 scope.go:117] "RemoveContainer" containerID="391b8bc1e975982b172175fcf4b4509756b018cbb99c12e48f4931ff472cdae2" Mar 13 14:07:35 crc kubenswrapper[4907]: I0313 14:07:35.567665 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"584d221c-be74-4a39-8173-6c8e3f55f686\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b782a795257057929fe616938d1483a9accbd7a93f0cba963f9791a3d4ec4749\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c826e7b9be5e3a0ecdaa6dfeaf2454535e542ff86acc532cc3c80acf4a95c1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4333bdf047392cb3f64e29e677fe8d2f21c7ebca1a7e7097167e70c6785b8d46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a4dea84b35a82ce16f0e0d38057dc994b26240cf45dcc9111ed0de573a62881\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a4dea84b35a82ce16f0e0d38057dc994b26240cf45dcc9111ed0de573a62881\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:35Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:35 crc kubenswrapper[4907]: I0313 14:07:35.587745 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:35Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:35 crc kubenswrapper[4907]: I0313 14:07:35.603142 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-p8ptc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57e4e912-9a98-4419-87ce-f16e178ed921\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38f113418d62dab1728f0ba57cd1d289e3ec41621a338017e74c9996a8449b1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bdef68f401b5ba2c7cf61551eba416e74f295ab2281433a5115334d28094818\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bdef68f401b5ba2c7cf61551eba416e74f295ab2281433a5115334d28094818\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f8d1f40744d609fd964572ef5e323f0283d2582e98745602754165818f17847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f8d1f40744d609fd964572ef5e323f0283d2582e98745602754165818f17847\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c1ab2e148eaa177b98f5883a3f323d12f5b662bff8cb2d072a97be18ab8e120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c1ab2e148eaa177b98f5883a3f323d12f5b662bff8cb2d072a97be18ab8e120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-p8ptc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:35Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:35 crc kubenswrapper[4907]: I0313 14:07:35.618972 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3fb4213-144f-4b01-9d78-ff3315761bb3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba1140a2462713c2401b47a4cf1f231bc9d9ae7867a7d0b1b597214a1f7efe07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://770c882d9e9f9b2e9bd4f135e591952fa15b3898c42826aed8a75d953c9a875b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hm56j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:35Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:35 crc kubenswrapper[4907]: I0313 14:07:35.641951 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1006376d-4033-454e-8a23-2b67882bdd0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf6a3dbaa8dda61f461a906538824c8946d36dc07d4fb95bf4d8420ad0105600\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccef92e41501fab91189581d7e8875515c02d637426b329d2fb80a356c75b637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0ee0049aa4a6d49a6bcf7d347183d9cc8b85171f988662803a005e0cdd97395\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fff87d8ea572615eb1329164f9e536e6ee45cd00a9e71be0aa8a9b3ed2190e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c165053fd9933444b372ff40efe33c0e193a7f39e6f531a4f370e971eb41e6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fea41c11bebd2d9d622ed4b8bc50e03a106dadea6ac102cd60512ea2a1f8b8b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a99cabe30a7abc8da854bf86d79628b94a6ffc20c28f828b695080332b319db2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a99cabe30a7abc8da854bf86d79628b94a6ffc20c28f828b695080332b319db2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T14:07:15Z\\\",\\\"message\\\":\\\"Sending *v1.Node event handler 7 for removal\\\\nI0313 14:07:15.619307 7162 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0313 14:07:15.619320 7162 handler.go:208] Removed *v1.Node event handler 2\\\\nI0313 14:07:15.619329 7162 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0313 14:07:15.619343 7162 handler.go:208] Removed *v1.Node event handler 7\\\\nI0313 14:07:15.619351 7162 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0313 14:07:15.619358 7162 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0313 14:07:15.619362 7162 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0313 14:07:15.619381 7162 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0313 14:07:15.619382 7162 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0313 14:07:15.619392 7162 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0313 14:07:15.619619 7162 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI0313 14:07:15.619713 7162 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI0313 14:07:15.619975 7162 factory.go:656] Stopping watch factory\\\\nI0313 14:07:15.620001 7162 ovnkube.go:599] Stopped ovnkube\\\\nI0313 14:07:15.620033 7162 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0313 14:07:15.620096 7162 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:07:14Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-qxlx5_openshift-ovn-kubernetes(1006376d-4033-454e-8a23-2b67882bdd0b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d60b5b774f8dd4cee16b9784f80d124640199a85331a2940772029be4884751\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qxlx5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:35Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:35 crc kubenswrapper[4907]: I0313 14:07:35.658422 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5d2b054-0ed8-419c-b731-f14d3598372f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4b503aa3b13268b3688d067f3e456c86afac1fda171aced0990899c3eb89b79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be40a1a33fa482eab363537afaa901127d560afbc1b5b287ef11e66391a5be4c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d1b539bbd918fc50428fc18cabcc4c8a43cc96f220b5692d67cf1c8dfbf8ed8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93d8c8c79be9d9ffb72de66fc2f3f9821936023071e95875582b45ccb319601f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T14:06:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0313 14:06:13.202871 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0313 14:06:13.203175 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0313 14:06:13.204427 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2471728650/tls.crt::/tmp/serving-cert-2471728650/tls.key\\\\\\\"\\\\nI0313 14:06:13.581160 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0313 14:06:13.584012 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0313 14:06:13.584036 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0313 14:06:13.584064 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0313 14:06:13.584072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0313 14:06:13.594504 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0313 14:06:13.594891 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0313 14:06:13.595323 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0313 14:06:13.595363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0313 14:06:13.595370 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0313 14:06:13.595376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0313 14:06:13.597226 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3f2b6756078925f8562b6053b3b83798efb72288f5fd18fff680a4ab0d33feb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:35Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:35 crc kubenswrapper[4907]: I0313 14:07:35.674439 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94b045cf2fed56457a2ca14f489bb69470198f159f16c207fe061d034dba1555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd28f7a5c35822582b70ed41645294aaede10e28ff59f3f3b5d36e7da8237b76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:35Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:35 crc kubenswrapper[4907]: I0313 14:07:35.690018 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l7zg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d105034b-f041-4fb5-8df6-6b1bb7681fda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21a207f917836785e320f4f79aca5d6449da2fdeec3ee0765f34d227d5617fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szmh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l7zg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:35Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:35 crc kubenswrapper[4907]: I0313 14:07:35.703952 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qncw4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1d7ee39-09d7-4c0e-942b-5e9d8c9f63be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07df71887196cc7098ac9606a4a15b89d45219048c31280198ae75022bff9370\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:07:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bg45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://540ca207f31c9a8a9892f80dee153b3fb86f8de406f38360221512f424d8c884\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:07:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bg45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:07:00Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-qncw4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:35Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:35 crc kubenswrapper[4907]: I0313 14:07:35.718418 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5b4819028f34db81cc43a04e4854c30488ed876c63b8daff11105bf61546c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:35Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:35 crc kubenswrapper[4907]: I0313 14:07:35.737440 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:35Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:35 crc kubenswrapper[4907]: I0313 14:07:35.758690 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xtwf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64ebd338-0158-44dc-90ab-e7c285e87762\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://391b8bc1e975982b172175fcf4b4509756b018cbb99c12e48f4931ff472cdae2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://391b8bc1e975982b172175fcf4b4509756b018cbb99c12e48f4931ff472cdae2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T14:07:34Z\\\",\\\"message\\\":\\\"2026-03-13T14:06:49+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_baa16ed5-a467-4fd5-8846-0c869be5e531\\\\n2026-03-13T14:06:49+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_baa16ed5-a467-4fd5-8846-0c869be5e531 to /host/opt/cni/bin/\\\\n2026-03-13T14:06:49Z [verbose] multus-daemon started\\\\n2026-03-13T14:06:49Z [verbose] Readiness Indicator file check\\\\n2026-03-13T14:07:34Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6cwbq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xtwf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:35Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:35 crc kubenswrapper[4907]: I0313 14:07:35.779046 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bed3bcea-e8c0-4b00-a985-e0c5b5bea100\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://402c1ef89085d840edb4313577f5b2fb967072cb3db00c34c98572f00699529c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f22becea3c68103ff0b2639c7ae8c051b76d19b407b0463645fc2e4a0682f23\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T14:05:32Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0313 14:05:03.905343 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0313 14:05:03.908030 1 observer_polling.go:159] Starting file observer\\\\nI0313 14:05:03.938230 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0313 14:05:03.942399 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0313 14:05:32.345768 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0313 14:05:32.345932 1 cmd.go:179] failed checking apiserver connectivity: Get \\\\\\\"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/openshift-kube-controller-manager/leases/cluster-policy-controller-lock\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:31Z is after 2026-02-23T05:33:13Z\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:03Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18eadc065bf4a307b7b443fcfe79b2b719938ee486b53321c0a9f20926a0a887\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9abfa5a9ad1a98a2907ead28390ee926e176646fe761b89f580101bd429cf747\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbc67477e0223db1e4b403be3709fe71fe6409570361e103eb730fdeaffe3e80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:35Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:35 crc kubenswrapper[4907]: I0313 14:07:35.795845 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:35Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:35 crc kubenswrapper[4907]: I0313 14:07:35.813251 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afb1d430459aafab2d6bc2507b4b4059529401e51fdf8e9848b0d7bb8651db68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:35Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:35 crc kubenswrapper[4907]: I0313 14:07:35.826201 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-745zx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58e11265-2fd9-4b9b-926c-3343bb84ee53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ce7ad32fd2bf2332f3ee41080115b083c34326b7dc31c950098dfba15f58422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mnz7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-745zx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:35Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:35 crc kubenswrapper[4907]: I0313 14:07:35.839401 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-tzmvc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f797b1bb-e769-49c4-84ce-93a6fad2df93\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tdzl9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tdzl9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:07:00Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-tzmvc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:35Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:36 crc kubenswrapper[4907]: I0313 14:07:36.557792 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-xtwf7_64ebd338-0158-44dc-90ab-e7c285e87762/kube-multus/0.log" Mar 13 14:07:36 crc kubenswrapper[4907]: I0313 14:07:36.557910 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-xtwf7" event={"ID":"64ebd338-0158-44dc-90ab-e7c285e87762","Type":"ContainerStarted","Data":"ef632bb17c8f529b35d0882c8d4ba3dc213b0308b46a4ade8426b9c08e64bb2b"} Mar 13 14:07:36 crc kubenswrapper[4907]: I0313 14:07:36.583674 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5b4819028f34db81cc43a04e4854c30488ed876c63b8daff11105bf61546c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:36Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:36 crc kubenswrapper[4907]: I0313 14:07:36.606989 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:36Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:36 crc kubenswrapper[4907]: I0313 14:07:36.628910 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xtwf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64ebd338-0158-44dc-90ab-e7c285e87762\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef632bb17c8f529b35d0882c8d4ba3dc213b0308b46a4ade8426b9c08e64bb2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://391b8bc1e975982b172175fcf4b4509756b018cbb99c12e48f4931ff472cdae2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T14:07:34Z\\\",\\\"message\\\":\\\"2026-03-13T14:06:49+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_baa16ed5-a467-4fd5-8846-0c869be5e531\\\\n2026-03-13T14:06:49+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_baa16ed5-a467-4fd5-8846-0c869be5e531 to /host/opt/cni/bin/\\\\n2026-03-13T14:06:49Z [verbose] multus-daemon started\\\\n2026-03-13T14:06:49Z [verbose] Readiness Indicator file check\\\\n2026-03-13T14:07:34Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:07:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6cwbq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xtwf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:36Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:36 crc kubenswrapper[4907]: I0313 14:07:36.648988 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bed3bcea-e8c0-4b00-a985-e0c5b5bea100\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://402c1ef89085d840edb4313577f5b2fb967072cb3db00c34c98572f00699529c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f22becea3c68103ff0b2639c7ae8c051b76d19b407b0463645fc2e4a0682f23\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T14:05:32Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0313 14:05:03.905343 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0313 14:05:03.908030 1 observer_polling.go:159] Starting file observer\\\\nI0313 14:05:03.938230 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0313 14:05:03.942399 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0313 14:05:32.345768 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0313 14:05:32.345932 1 cmd.go:179] failed checking apiserver connectivity: Get \\\\\\\"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/openshift-kube-controller-manager/leases/cluster-policy-controller-lock\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:31Z is after 2026-02-23T05:33:13Z\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:03Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18eadc065bf4a307b7b443fcfe79b2b719938ee486b53321c0a9f20926a0a887\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9abfa5a9ad1a98a2907ead28390ee926e176646fe761b89f580101bd429cf747\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbc67477e0223db1e4b403be3709fe71fe6409570361e103eb730fdeaffe3e80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:36Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:36 crc kubenswrapper[4907]: I0313 14:07:36.661920 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:36Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:36 crc kubenswrapper[4907]: I0313 14:07:36.675559 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afb1d430459aafab2d6bc2507b4b4059529401e51fdf8e9848b0d7bb8651db68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:36Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:36 crc kubenswrapper[4907]: I0313 14:07:36.688955 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-745zx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58e11265-2fd9-4b9b-926c-3343bb84ee53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ce7ad32fd2bf2332f3ee41080115b083c34326b7dc31c950098dfba15f58422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mnz7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-745zx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:36Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:36 crc kubenswrapper[4907]: I0313 14:07:36.697116 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:07:36 crc kubenswrapper[4907]: E0313 14:07:36.697661 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:08:40.697615688 +0000 UTC m=+219.597403377 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:07:36 crc kubenswrapper[4907]: I0313 14:07:36.709791 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-tzmvc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f797b1bb-e769-49c4-84ce-93a6fad2df93\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tdzl9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tdzl9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:07:00Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-tzmvc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:36Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:36 crc kubenswrapper[4907]: I0313 14:07:36.742488 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1006376d-4033-454e-8a23-2b67882bdd0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf6a3dbaa8dda61f461a906538824c8946d36dc07d4fb95bf4d8420ad0105600\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccef92e41501fab91189581d7e8875515c02d637426b329d2fb80a356c75b637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0ee0049aa4a6d49a6bcf7d347183d9cc8b85171f988662803a005e0cdd97395\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fff87d8ea572615eb1329164f9e536e6ee45cd00a9e71be0aa8a9b3ed2190e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c165053fd9933444b372ff40efe33c0e193a7f39e6f531a4f370e971eb41e6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fea41c11bebd2d9d622ed4b8bc50e03a106dadea6ac102cd60512ea2a1f8b8b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a99cabe30a7abc8da854bf86d79628b94a6ffc20c28f828b695080332b319db2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a99cabe30a7abc8da854bf86d79628b94a6ffc20c28f828b695080332b319db2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T14:07:15Z\\\",\\\"message\\\":\\\"Sending *v1.Node event handler 7 for removal\\\\nI0313 14:07:15.619307 7162 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0313 14:07:15.619320 7162 handler.go:208] Removed *v1.Node event handler 2\\\\nI0313 14:07:15.619329 7162 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0313 14:07:15.619343 7162 handler.go:208] Removed *v1.Node event handler 7\\\\nI0313 14:07:15.619351 7162 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0313 14:07:15.619358 7162 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0313 14:07:15.619362 7162 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0313 14:07:15.619381 7162 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0313 14:07:15.619382 7162 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0313 14:07:15.619392 7162 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0313 14:07:15.619619 7162 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI0313 14:07:15.619713 7162 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI0313 14:07:15.619975 7162 factory.go:656] Stopping watch factory\\\\nI0313 14:07:15.620001 7162 ovnkube.go:599] Stopped ovnkube\\\\nI0313 14:07:15.620033 7162 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0313 14:07:15.620096 7162 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:07:14Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-qxlx5_openshift-ovn-kubernetes(1006376d-4033-454e-8a23-2b67882bdd0b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d60b5b774f8dd4cee16b9784f80d124640199a85331a2940772029be4884751\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qxlx5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:36Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:36 crc kubenswrapper[4907]: I0313 14:07:36.764230 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"584d221c-be74-4a39-8173-6c8e3f55f686\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b782a795257057929fe616938d1483a9accbd7a93f0cba963f9791a3d4ec4749\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c826e7b9be5e3a0ecdaa6dfeaf2454535e542ff86acc532cc3c80acf4a95c1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4333bdf047392cb3f64e29e677fe8d2f21c7ebca1a7e7097167e70c6785b8d46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a4dea84b35a82ce16f0e0d38057dc994b26240cf45dcc9111ed0de573a62881\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a4dea84b35a82ce16f0e0d38057dc994b26240cf45dcc9111ed0de573a62881\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:36Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:36 crc kubenswrapper[4907]: I0313 14:07:36.781796 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:07:36 crc kubenswrapper[4907]: I0313 14:07:36.781819 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:07:36 crc kubenswrapper[4907]: I0313 14:07:36.781858 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:07:36 crc kubenswrapper[4907]: I0313 14:07:36.781916 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:07:36 crc kubenswrapper[4907]: E0313 14:07:36.782082 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:07:36 crc kubenswrapper[4907]: E0313 14:07:36.782277 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:07:36 crc kubenswrapper[4907]: E0313 14:07:36.782425 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:07:36 crc kubenswrapper[4907]: E0313 14:07:36.782540 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzmvc" podUID="f797b1bb-e769-49c4-84ce-93a6fad2df93" Mar 13 14:07:36 crc kubenswrapper[4907]: I0313 14:07:36.784210 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:36Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:36 crc kubenswrapper[4907]: I0313 14:07:36.798648 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:07:36 crc kubenswrapper[4907]: I0313 14:07:36.798702 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:07:36 crc kubenswrapper[4907]: I0313 14:07:36.798747 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:07:36 crc kubenswrapper[4907]: I0313 14:07:36.798782 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:07:36 crc kubenswrapper[4907]: E0313 14:07:36.798959 4907 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 13 14:07:36 crc kubenswrapper[4907]: E0313 14:07:36.798999 4907 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 13 14:07:36 crc kubenswrapper[4907]: E0313 14:07:36.799016 4907 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 13 14:07:36 crc kubenswrapper[4907]: E0313 14:07:36.799029 4907 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 13 14:07:36 crc kubenswrapper[4907]: E0313 14:07:36.799042 4907 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 13 14:07:36 crc kubenswrapper[4907]: E0313 14:07:36.799087 4907 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 14:07:36 crc kubenswrapper[4907]: E0313 14:07:36.799057 4907 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 13 14:07:36 crc kubenswrapper[4907]: E0313 14:07:36.799049 4907 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 14:07:36 crc kubenswrapper[4907]: E0313 14:07:36.799026 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-13 14:08:40.799005651 +0000 UTC m=+219.698793360 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 13 14:07:36 crc kubenswrapper[4907]: E0313 14:07:36.799375 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-13 14:08:40.799331735 +0000 UTC m=+219.699119474 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 14:07:36 crc kubenswrapper[4907]: E0313 14:07:36.799418 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-13 14:08:40.799403933 +0000 UTC m=+219.699191662 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 14:07:36 crc kubenswrapper[4907]: E0313 14:07:36.799465 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-13 14:08:40.799448567 +0000 UTC m=+219.699236296 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 13 14:07:36 crc kubenswrapper[4907]: I0313 14:07:36.811455 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-p8ptc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57e4e912-9a98-4419-87ce-f16e178ed921\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38f113418d62dab1728f0ba57cd1d289e3ec41621a338017e74c9996a8449b1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bdef68f401b5ba2c7cf61551eba416e74f295ab2281433a5115334d28094818\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bdef68f401b5ba2c7cf61551eba416e74f295ab2281433a5115334d28094818\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f8d1f40744d609fd964572ef5e323f0283d2582e98745602754165818f17847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f8d1f40744d609fd964572ef5e323f0283d2582e98745602754165818f17847\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c1ab2e148eaa177b98f5883a3f323d12f5b662bff8cb2d072a97be18ab8e120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c1ab2e148eaa177b98f5883a3f323d12f5b662bff8cb2d072a97be18ab8e120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-p8ptc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:36Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:36 crc kubenswrapper[4907]: I0313 14:07:36.829040 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3fb4213-144f-4b01-9d78-ff3315761bb3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba1140a2462713c2401b47a4cf1f231bc9d9ae7867a7d0b1b597214a1f7efe07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://770c882d9e9f9b2e9bd4f135e591952fa15b3898c42826aed8a75d953c9a875b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hm56j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:36Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:36 crc kubenswrapper[4907]: I0313 14:07:36.848995 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5d2b054-0ed8-419c-b731-f14d3598372f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4b503aa3b13268b3688d067f3e456c86afac1fda171aced0990899c3eb89b79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be40a1a33fa482eab363537afaa901127d560afbc1b5b287ef11e66391a5be4c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d1b539bbd918fc50428fc18cabcc4c8a43cc96f220b5692d67cf1c8dfbf8ed8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93d8c8c79be9d9ffb72de66fc2f3f9821936023071e95875582b45ccb319601f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T14:06:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0313 14:06:13.202871 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0313 14:06:13.203175 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0313 14:06:13.204427 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2471728650/tls.crt::/tmp/serving-cert-2471728650/tls.key\\\\\\\"\\\\nI0313 14:06:13.581160 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0313 14:06:13.584012 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0313 14:06:13.584036 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0313 14:06:13.584064 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0313 14:06:13.584072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0313 14:06:13.594504 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0313 14:06:13.594891 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0313 14:06:13.595323 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0313 14:06:13.595363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0313 14:06:13.595370 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0313 14:06:13.595376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0313 14:06:13.597226 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3f2b6756078925f8562b6053b3b83798efb72288f5fd18fff680a4ab0d33feb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:36Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:36 crc kubenswrapper[4907]: I0313 14:07:36.866643 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94b045cf2fed56457a2ca14f489bb69470198f159f16c207fe061d034dba1555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd28f7a5c35822582b70ed41645294aaede10e28ff59f3f3b5d36e7da8237b76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:36Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:36 crc kubenswrapper[4907]: I0313 14:07:36.883793 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l7zg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d105034b-f041-4fb5-8df6-6b1bb7681fda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21a207f917836785e320f4f79aca5d6449da2fdeec3ee0765f34d227d5617fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szmh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l7zg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:36Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:36 crc kubenswrapper[4907]: E0313 14:07:36.894703 4907 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 13 14:07:36 crc kubenswrapper[4907]: I0313 14:07:36.902248 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qncw4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1d7ee39-09d7-4c0e-942b-5e9d8c9f63be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07df71887196cc7098ac9606a4a15b89d45219048c31280198ae75022bff9370\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:07:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bg45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://540ca207f31c9a8a9892f80dee153b3fb86f8de406f38360221512f424d8c884\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:07:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bg45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:07:00Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-qncw4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:36Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:38 crc kubenswrapper[4907]: I0313 14:07:38.145725 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:07:38 crc kubenswrapper[4907]: I0313 14:07:38.145775 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:07:38 crc kubenswrapper[4907]: I0313 14:07:38.145793 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:07:38 crc kubenswrapper[4907]: I0313 14:07:38.145818 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:07:38 crc kubenswrapper[4907]: I0313 14:07:38.145835 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:07:38Z","lastTransitionTime":"2026-03-13T14:07:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:07:38 crc kubenswrapper[4907]: E0313 14:07:38.171276 4907 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4b182bae-7f2d-4b54-934c-85a9a277d132\\\",\\\"systemUUID\\\":\\\"05bfbfa3-f04f-45ba-8233-a355fc774c24\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:38Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:38 crc kubenswrapper[4907]: I0313 14:07:38.176838 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:07:38 crc kubenswrapper[4907]: I0313 14:07:38.176929 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:07:38 crc kubenswrapper[4907]: I0313 14:07:38.176948 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:07:38 crc kubenswrapper[4907]: I0313 14:07:38.176976 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:07:38 crc kubenswrapper[4907]: I0313 14:07:38.176993 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:07:38Z","lastTransitionTime":"2026-03-13T14:07:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:07:38 crc kubenswrapper[4907]: E0313 14:07:38.199763 4907 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4b182bae-7f2d-4b54-934c-85a9a277d132\\\",\\\"systemUUID\\\":\\\"05bfbfa3-f04f-45ba-8233-a355fc774c24\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:38Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:38 crc kubenswrapper[4907]: I0313 14:07:38.205427 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:07:38 crc kubenswrapper[4907]: I0313 14:07:38.205481 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:07:38 crc kubenswrapper[4907]: I0313 14:07:38.205500 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:07:38 crc kubenswrapper[4907]: I0313 14:07:38.205524 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:07:38 crc kubenswrapper[4907]: I0313 14:07:38.205543 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:07:38Z","lastTransitionTime":"2026-03-13T14:07:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:07:38 crc kubenswrapper[4907]: E0313 14:07:38.224732 4907 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4b182bae-7f2d-4b54-934c-85a9a277d132\\\",\\\"systemUUID\\\":\\\"05bfbfa3-f04f-45ba-8233-a355fc774c24\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:38Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:38 crc kubenswrapper[4907]: I0313 14:07:38.229348 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:07:38 crc kubenswrapper[4907]: I0313 14:07:38.229402 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:07:38 crc kubenswrapper[4907]: I0313 14:07:38.229416 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:07:38 crc kubenswrapper[4907]: I0313 14:07:38.229439 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:07:38 crc kubenswrapper[4907]: I0313 14:07:38.229453 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:07:38Z","lastTransitionTime":"2026-03-13T14:07:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:07:38 crc kubenswrapper[4907]: E0313 14:07:38.245055 4907 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4b182bae-7f2d-4b54-934c-85a9a277d132\\\",\\\"systemUUID\\\":\\\"05bfbfa3-f04f-45ba-8233-a355fc774c24\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:38Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:38 crc kubenswrapper[4907]: I0313 14:07:38.249376 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:07:38 crc kubenswrapper[4907]: I0313 14:07:38.249436 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:07:38 crc kubenswrapper[4907]: I0313 14:07:38.249453 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:07:38 crc kubenswrapper[4907]: I0313 14:07:38.249480 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:07:38 crc kubenswrapper[4907]: I0313 14:07:38.249498 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:07:38Z","lastTransitionTime":"2026-03-13T14:07:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:07:38 crc kubenswrapper[4907]: E0313 14:07:38.267639 4907 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4b182bae-7f2d-4b54-934c-85a9a277d132\\\",\\\"systemUUID\\\":\\\"05bfbfa3-f04f-45ba-8233-a355fc774c24\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:38Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:38 crc kubenswrapper[4907]: E0313 14:07:38.267850 4907 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 13 14:07:38 crc kubenswrapper[4907]: I0313 14:07:38.782199 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:07:38 crc kubenswrapper[4907]: I0313 14:07:38.782248 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:07:38 crc kubenswrapper[4907]: I0313 14:07:38.782274 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:07:38 crc kubenswrapper[4907]: I0313 14:07:38.782412 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:07:38 crc kubenswrapper[4907]: E0313 14:07:38.782516 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:07:38 crc kubenswrapper[4907]: E0313 14:07:38.782667 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:07:38 crc kubenswrapper[4907]: E0313 14:07:38.782774 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:07:38 crc kubenswrapper[4907]: E0313 14:07:38.782833 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzmvc" podUID="f797b1bb-e769-49c4-84ce-93a6fad2df93" Mar 13 14:07:38 crc kubenswrapper[4907]: I0313 14:07:38.796505 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Mar 13 14:07:40 crc kubenswrapper[4907]: I0313 14:07:40.782194 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:07:40 crc kubenswrapper[4907]: I0313 14:07:40.782275 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:07:40 crc kubenswrapper[4907]: E0313 14:07:40.782384 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:07:40 crc kubenswrapper[4907]: I0313 14:07:40.782227 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:07:40 crc kubenswrapper[4907]: E0313 14:07:40.782512 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:07:40 crc kubenswrapper[4907]: I0313 14:07:40.782557 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:07:40 crc kubenswrapper[4907]: E0313 14:07:40.782669 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzmvc" podUID="f797b1bb-e769-49c4-84ce-93a6fad2df93" Mar 13 14:07:40 crc kubenswrapper[4907]: E0313 14:07:40.782741 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:07:41 crc kubenswrapper[4907]: I0313 14:07:41.782525 4907 scope.go:117] "RemoveContainer" containerID="a99cabe30a7abc8da854bf86d79628b94a6ffc20c28f828b695080332b319db2" Mar 13 14:07:41 crc kubenswrapper[4907]: I0313 14:07:41.804866 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5b4819028f34db81cc43a04e4854c30488ed876c63b8daff11105bf61546c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:41Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:41 crc kubenswrapper[4907]: I0313 14:07:41.824039 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:41Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:41 crc kubenswrapper[4907]: I0313 14:07:41.841087 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xtwf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64ebd338-0158-44dc-90ab-e7c285e87762\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef632bb17c8f529b35d0882c8d4ba3dc213b0308b46a4ade8426b9c08e64bb2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://391b8bc1e975982b172175fcf4b4509756b018cbb99c12e48f4931ff472cdae2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T14:07:34Z\\\",\\\"message\\\":\\\"2026-03-13T14:06:49+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_baa16ed5-a467-4fd5-8846-0c869be5e531\\\\n2026-03-13T14:06:49+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_baa16ed5-a467-4fd5-8846-0c869be5e531 to /host/opt/cni/bin/\\\\n2026-03-13T14:06:49Z [verbose] multus-daemon started\\\\n2026-03-13T14:06:49Z [verbose] Readiness Indicator file check\\\\n2026-03-13T14:07:34Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:07:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6cwbq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xtwf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:41Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:41 crc kubenswrapper[4907]: I0313 14:07:41.856727 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bed3bcea-e8c0-4b00-a985-e0c5b5bea100\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://402c1ef89085d840edb4313577f5b2fb967072cb3db00c34c98572f00699529c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f22becea3c68103ff0b2639c7ae8c051b76d19b407b0463645fc2e4a0682f23\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T14:05:32Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0313 14:05:03.905343 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0313 14:05:03.908030 1 observer_polling.go:159] Starting file observer\\\\nI0313 14:05:03.938230 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0313 14:05:03.942399 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0313 14:05:32.345768 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0313 14:05:32.345932 1 cmd.go:179] failed checking apiserver connectivity: Get \\\\\\\"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/openshift-kube-controller-manager/leases/cluster-policy-controller-lock\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:31Z is after 2026-02-23T05:33:13Z\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:03Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18eadc065bf4a307b7b443fcfe79b2b719938ee486b53321c0a9f20926a0a887\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9abfa5a9ad1a98a2907ead28390ee926e176646fe761b89f580101bd429cf747\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbc67477e0223db1e4b403be3709fe71fe6409570361e103eb730fdeaffe3e80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:41Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:41 crc kubenswrapper[4907]: I0313 14:07:41.872448 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:41Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:41 crc kubenswrapper[4907]: I0313 14:07:41.886035 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afb1d430459aafab2d6bc2507b4b4059529401e51fdf8e9848b0d7bb8651db68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:41Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:41 crc kubenswrapper[4907]: E0313 14:07:41.895398 4907 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 13 14:07:41 crc kubenswrapper[4907]: I0313 14:07:41.903010 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-745zx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58e11265-2fd9-4b9b-926c-3343bb84ee53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ce7ad32fd2bf2332f3ee41080115b083c34326b7dc31c950098dfba15f58422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mnz7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-745zx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:41Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:41 crc kubenswrapper[4907]: I0313 14:07:41.918636 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-tzmvc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f797b1bb-e769-49c4-84ce-93a6fad2df93\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tdzl9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tdzl9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:07:00Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-tzmvc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:41Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:41 crc kubenswrapper[4907]: I0313 14:07:41.935334 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"584d221c-be74-4a39-8173-6c8e3f55f686\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b782a795257057929fe616938d1483a9accbd7a93f0cba963f9791a3d4ec4749\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c826e7b9be5e3a0ecdaa6dfeaf2454535e542ff86acc532cc3c80acf4a95c1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4333bdf047392cb3f64e29e677fe8d2f21c7ebca1a7e7097167e70c6785b8d46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a4dea84b35a82ce16f0e0d38057dc994b26240cf45dcc9111ed0de573a62881\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a4dea84b35a82ce16f0e0d38057dc994b26240cf45dcc9111ed0de573a62881\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:41Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:41 crc kubenswrapper[4907]: I0313 14:07:41.950925 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:41Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:41 crc kubenswrapper[4907]: I0313 14:07:41.968905 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-p8ptc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57e4e912-9a98-4419-87ce-f16e178ed921\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38f113418d62dab1728f0ba57cd1d289e3ec41621a338017e74c9996a8449b1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bdef68f401b5ba2c7cf61551eba416e74f295ab2281433a5115334d28094818\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bdef68f401b5ba2c7cf61551eba416e74f295ab2281433a5115334d28094818\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f8d1f40744d609fd964572ef5e323f0283d2582e98745602754165818f17847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f8d1f40744d609fd964572ef5e323f0283d2582e98745602754165818f17847\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c1ab2e148eaa177b98f5883a3f323d12f5b662bff8cb2d072a97be18ab8e120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c1ab2e148eaa177b98f5883a3f323d12f5b662bff8cb2d072a97be18ab8e120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-p8ptc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:41Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:41 crc kubenswrapper[4907]: I0313 14:07:41.981549 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3fb4213-144f-4b01-9d78-ff3315761bb3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba1140a2462713c2401b47a4cf1f231bc9d9ae7867a7d0b1b597214a1f7efe07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://770c882d9e9f9b2e9bd4f135e591952fa15b3898c42826aed8a75d953c9a875b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hm56j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:41Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:42 crc kubenswrapper[4907]: I0313 14:07:42.001655 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1006376d-4033-454e-8a23-2b67882bdd0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf6a3dbaa8dda61f461a906538824c8946d36dc07d4fb95bf4d8420ad0105600\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccef92e41501fab91189581d7e8875515c02d637426b329d2fb80a356c75b637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0ee0049aa4a6d49a6bcf7d347183d9cc8b85171f988662803a005e0cdd97395\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fff87d8ea572615eb1329164f9e536e6ee45cd00a9e71be0aa8a9b3ed2190e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c165053fd9933444b372ff40efe33c0e193a7f39e6f531a4f370e971eb41e6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fea41c11bebd2d9d622ed4b8bc50e03a106dadea6ac102cd60512ea2a1f8b8b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a99cabe30a7abc8da854bf86d79628b94a6ffc20c28f828b695080332b319db2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a99cabe30a7abc8da854bf86d79628b94a6ffc20c28f828b695080332b319db2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T14:07:15Z\\\",\\\"message\\\":\\\"Sending *v1.Node event handler 7 for removal\\\\nI0313 14:07:15.619307 7162 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0313 14:07:15.619320 7162 handler.go:208] Removed *v1.Node event handler 2\\\\nI0313 14:07:15.619329 7162 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0313 14:07:15.619343 7162 handler.go:208] Removed *v1.Node event handler 7\\\\nI0313 14:07:15.619351 7162 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0313 14:07:15.619358 7162 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0313 14:07:15.619362 7162 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0313 14:07:15.619381 7162 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0313 14:07:15.619382 7162 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0313 14:07:15.619392 7162 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0313 14:07:15.619619 7162 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI0313 14:07:15.619713 7162 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI0313 14:07:15.619975 7162 factory.go:656] Stopping watch factory\\\\nI0313 14:07:15.620001 7162 ovnkube.go:599] Stopped ovnkube\\\\nI0313 14:07:15.620033 7162 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0313 14:07:15.620096 7162 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:07:14Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-qxlx5_openshift-ovn-kubernetes(1006376d-4033-454e-8a23-2b67882bdd0b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d60b5b774f8dd4cee16b9784f80d124640199a85331a2940772029be4884751\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qxlx5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:41Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:42 crc kubenswrapper[4907]: I0313 14:07:42.023817 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31a594bd-13e7-4c68-ac7c-f481626088ec\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8141bbe3b2b50ad7d2aed26f813a951bc6db4c90bf8271cd23a125cc7ec2eb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb841f47d97b3e9495658ff7f6aa3f835905a9cd2ee9736774c3504d6c59a209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f04995f26a4c320fa30f2a97b1299284cbb2e23965bae16a6cca8d06eef630eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0568aef54a902fedcaaa71cabe0add0cbeb3a38743a7fe45584c5e315292afa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://171ea306a9ec518a88ef509adaab1a7c238abf7b5824a87d3425de7199cde520\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14a9ee19b96b2019de6ad9e5092e0004a4587a9f678fec417c4ac02e60695ac8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14a9ee19b96b2019de6ad9e5092e0004a4587a9f678fec417c4ac02e60695ac8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d46350802ae58ed4c464dc3f8d3f9dd7e091ef8091659e0711b9d2632e00708\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d46350802ae58ed4c464dc3f8d3f9dd7e091ef8091659e0711b9d2632e00708\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://65bd2db6ca4ffb26d9cb4bce44dc27a862353533bb0ec584bcc688b73fbec3eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65bd2db6ca4ffb26d9cb4bce44dc27a862353533bb0ec584bcc688b73fbec3eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:42Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:42 crc kubenswrapper[4907]: I0313 14:07:42.039817 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5d2b054-0ed8-419c-b731-f14d3598372f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4b503aa3b13268b3688d067f3e456c86afac1fda171aced0990899c3eb89b79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be40a1a33fa482eab363537afaa901127d560afbc1b5b287ef11e66391a5be4c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d1b539bbd918fc50428fc18cabcc4c8a43cc96f220b5692d67cf1c8dfbf8ed8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93d8c8c79be9d9ffb72de66fc2f3f9821936023071e95875582b45ccb319601f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T14:06:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0313 14:06:13.202871 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0313 14:06:13.203175 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0313 14:06:13.204427 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2471728650/tls.crt::/tmp/serving-cert-2471728650/tls.key\\\\\\\"\\\\nI0313 14:06:13.581160 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0313 14:06:13.584012 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0313 14:06:13.584036 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0313 14:06:13.584064 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0313 14:06:13.584072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0313 14:06:13.594504 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0313 14:06:13.594891 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0313 14:06:13.595323 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0313 14:06:13.595363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0313 14:06:13.595370 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0313 14:06:13.595376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0313 14:06:13.597226 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3f2b6756078925f8562b6053b3b83798efb72288f5fd18fff680a4ab0d33feb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:42Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:42 crc kubenswrapper[4907]: I0313 14:07:42.055220 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94b045cf2fed56457a2ca14f489bb69470198f159f16c207fe061d034dba1555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd28f7a5c35822582b70ed41645294aaede10e28ff59f3f3b5d36e7da8237b76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:42Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:42 crc kubenswrapper[4907]: I0313 14:07:42.065556 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l7zg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d105034b-f041-4fb5-8df6-6b1bb7681fda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21a207f917836785e320f4f79aca5d6449da2fdeec3ee0765f34d227d5617fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szmh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l7zg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:42Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:42 crc kubenswrapper[4907]: I0313 14:07:42.078097 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qncw4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1d7ee39-09d7-4c0e-942b-5e9d8c9f63be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07df71887196cc7098ac9606a4a15b89d45219048c31280198ae75022bff9370\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:07:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bg45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://540ca207f31c9a8a9892f80dee153b3fb86f8de406f38360221512f424d8c884\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:07:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bg45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:07:00Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-qncw4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:42Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:42 crc kubenswrapper[4907]: I0313 14:07:42.582990 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qxlx5_1006376d-4033-454e-8a23-2b67882bdd0b/ovnkube-controller/2.log" Mar 13 14:07:42 crc kubenswrapper[4907]: I0313 14:07:42.586419 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" event={"ID":"1006376d-4033-454e-8a23-2b67882bdd0b","Type":"ContainerStarted","Data":"9155434e2d8f0daa4504758fb72d6f498da107a64a92a0883ac687ed2af9caee"} Mar 13 14:07:42 crc kubenswrapper[4907]: I0313 14:07:42.586864 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:07:42 crc kubenswrapper[4907]: I0313 14:07:42.604895 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1006376d-4033-454e-8a23-2b67882bdd0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf6a3dbaa8dda61f461a906538824c8946d36dc07d4fb95bf4d8420ad0105600\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccef92e41501fab91189581d7e8875515c02d637426b329d2fb80a356c75b637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0ee0049aa4a6d49a6bcf7d347183d9cc8b85171f988662803a005e0cdd97395\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fff87d8ea572615eb1329164f9e536e6ee45cd00a9e71be0aa8a9b3ed2190e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c165053fd9933444b372ff40efe33c0e193a7f39e6f531a4f370e971eb41e6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fea41c11bebd2d9d622ed4b8bc50e03a106dadea6ac102cd60512ea2a1f8b8b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9155434e2d8f0daa4504758fb72d6f498da107a64a92a0883ac687ed2af9caee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a99cabe30a7abc8da854bf86d79628b94a6ffc20c28f828b695080332b319db2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T14:07:15Z\\\",\\\"message\\\":\\\"Sending *v1.Node event handler 7 for removal\\\\nI0313 14:07:15.619307 7162 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0313 14:07:15.619320 7162 handler.go:208] Removed *v1.Node event handler 2\\\\nI0313 14:07:15.619329 7162 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0313 14:07:15.619343 7162 handler.go:208] Removed *v1.Node event handler 7\\\\nI0313 14:07:15.619351 7162 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0313 14:07:15.619358 7162 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0313 14:07:15.619362 7162 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0313 14:07:15.619381 7162 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0313 14:07:15.619382 7162 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0313 14:07:15.619392 7162 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0313 14:07:15.619619 7162 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI0313 14:07:15.619713 7162 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI0313 14:07:15.619975 7162 factory.go:656] Stopping watch factory\\\\nI0313 14:07:15.620001 7162 ovnkube.go:599] Stopped ovnkube\\\\nI0313 14:07:15.620033 7162 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0313 14:07:15.620096 7162 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:07:14Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:07:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d60b5b774f8dd4cee16b9784f80d124640199a85331a2940772029be4884751\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qxlx5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:42Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:42 crc kubenswrapper[4907]: I0313 14:07:42.615338 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"584d221c-be74-4a39-8173-6c8e3f55f686\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b782a795257057929fe616938d1483a9accbd7a93f0cba963f9791a3d4ec4749\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c826e7b9be5e3a0ecdaa6dfeaf2454535e542ff86acc532cc3c80acf4a95c1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4333bdf047392cb3f64e29e677fe8d2f21c7ebca1a7e7097167e70c6785b8d46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a4dea84b35a82ce16f0e0d38057dc994b26240cf45dcc9111ed0de573a62881\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a4dea84b35a82ce16f0e0d38057dc994b26240cf45dcc9111ed0de573a62881\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:42Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:42 crc kubenswrapper[4907]: I0313 14:07:42.625216 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:42Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:42 crc kubenswrapper[4907]: I0313 14:07:42.639838 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-p8ptc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57e4e912-9a98-4419-87ce-f16e178ed921\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38f113418d62dab1728f0ba57cd1d289e3ec41621a338017e74c9996a8449b1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bdef68f401b5ba2c7cf61551eba416e74f295ab2281433a5115334d28094818\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bdef68f401b5ba2c7cf61551eba416e74f295ab2281433a5115334d28094818\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f8d1f40744d609fd964572ef5e323f0283d2582e98745602754165818f17847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f8d1f40744d609fd964572ef5e323f0283d2582e98745602754165818f17847\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c1ab2e148eaa177b98f5883a3f323d12f5b662bff8cb2d072a97be18ab8e120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c1ab2e148eaa177b98f5883a3f323d12f5b662bff8cb2d072a97be18ab8e120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-p8ptc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:42Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:42 crc kubenswrapper[4907]: I0313 14:07:42.649982 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3fb4213-144f-4b01-9d78-ff3315761bb3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba1140a2462713c2401b47a4cf1f231bc9d9ae7867a7d0b1b597214a1f7efe07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://770c882d9e9f9b2e9bd4f135e591952fa15b3898c42826aed8a75d953c9a875b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hm56j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:42Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:42 crc kubenswrapper[4907]: I0313 14:07:42.666139 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31a594bd-13e7-4c68-ac7c-f481626088ec\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8141bbe3b2b50ad7d2aed26f813a951bc6db4c90bf8271cd23a125cc7ec2eb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb841f47d97b3e9495658ff7f6aa3f835905a9cd2ee9736774c3504d6c59a209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f04995f26a4c320fa30f2a97b1299284cbb2e23965bae16a6cca8d06eef630eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0568aef54a902fedcaaa71cabe0add0cbeb3a38743a7fe45584c5e315292afa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://171ea306a9ec518a88ef509adaab1a7c238abf7b5824a87d3425de7199cde520\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14a9ee19b96b2019de6ad9e5092e0004a4587a9f678fec417c4ac02e60695ac8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14a9ee19b96b2019de6ad9e5092e0004a4587a9f678fec417c4ac02e60695ac8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d46350802ae58ed4c464dc3f8d3f9dd7e091ef8091659e0711b9d2632e00708\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d46350802ae58ed4c464dc3f8d3f9dd7e091ef8091659e0711b9d2632e00708\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://65bd2db6ca4ffb26d9cb4bce44dc27a862353533bb0ec584bcc688b73fbec3eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65bd2db6ca4ffb26d9cb4bce44dc27a862353533bb0ec584bcc688b73fbec3eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:42Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:42 crc kubenswrapper[4907]: I0313 14:07:42.682400 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5d2b054-0ed8-419c-b731-f14d3598372f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4b503aa3b13268b3688d067f3e456c86afac1fda171aced0990899c3eb89b79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be40a1a33fa482eab363537afaa901127d560afbc1b5b287ef11e66391a5be4c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d1b539bbd918fc50428fc18cabcc4c8a43cc96f220b5692d67cf1c8dfbf8ed8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93d8c8c79be9d9ffb72de66fc2f3f9821936023071e95875582b45ccb319601f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T14:06:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0313 14:06:13.202871 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0313 14:06:13.203175 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0313 14:06:13.204427 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2471728650/tls.crt::/tmp/serving-cert-2471728650/tls.key\\\\\\\"\\\\nI0313 14:06:13.581160 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0313 14:06:13.584012 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0313 14:06:13.584036 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0313 14:06:13.584064 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0313 14:06:13.584072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0313 14:06:13.594504 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0313 14:06:13.594891 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0313 14:06:13.595323 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0313 14:06:13.595363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0313 14:06:13.595370 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0313 14:06:13.595376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0313 14:06:13.597226 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3f2b6756078925f8562b6053b3b83798efb72288f5fd18fff680a4ab0d33feb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:42Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:42 crc kubenswrapper[4907]: I0313 14:07:42.692204 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94b045cf2fed56457a2ca14f489bb69470198f159f16c207fe061d034dba1555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd28f7a5c35822582b70ed41645294aaede10e28ff59f3f3b5d36e7da8237b76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:42Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:42 crc kubenswrapper[4907]: I0313 14:07:42.701874 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l7zg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d105034b-f041-4fb5-8df6-6b1bb7681fda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21a207f917836785e320f4f79aca5d6449da2fdeec3ee0765f34d227d5617fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szmh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l7zg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:42Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:42 crc kubenswrapper[4907]: I0313 14:07:42.712116 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qncw4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1d7ee39-09d7-4c0e-942b-5e9d8c9f63be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07df71887196cc7098ac9606a4a15b89d45219048c31280198ae75022bff9370\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:07:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bg45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://540ca207f31c9a8a9892f80dee153b3fb86f8de406f38360221512f424d8c884\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:07:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bg45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:07:00Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-qncw4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:42Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:42 crc kubenswrapper[4907]: I0313 14:07:42.722900 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5b4819028f34db81cc43a04e4854c30488ed876c63b8daff11105bf61546c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:42Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:42 crc kubenswrapper[4907]: I0313 14:07:42.733855 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:42Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:42 crc kubenswrapper[4907]: I0313 14:07:42.745799 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xtwf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64ebd338-0158-44dc-90ab-e7c285e87762\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef632bb17c8f529b35d0882c8d4ba3dc213b0308b46a4ade8426b9c08e64bb2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://391b8bc1e975982b172175fcf4b4509756b018cbb99c12e48f4931ff472cdae2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T14:07:34Z\\\",\\\"message\\\":\\\"2026-03-13T14:06:49+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_baa16ed5-a467-4fd5-8846-0c869be5e531\\\\n2026-03-13T14:06:49+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_baa16ed5-a467-4fd5-8846-0c869be5e531 to /host/opt/cni/bin/\\\\n2026-03-13T14:06:49Z [verbose] multus-daemon started\\\\n2026-03-13T14:06:49Z [verbose] Readiness Indicator file check\\\\n2026-03-13T14:07:34Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:07:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6cwbq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xtwf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:42Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:42 crc kubenswrapper[4907]: I0313 14:07:42.757538 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bed3bcea-e8c0-4b00-a985-e0c5b5bea100\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://402c1ef89085d840edb4313577f5b2fb967072cb3db00c34c98572f00699529c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f22becea3c68103ff0b2639c7ae8c051b76d19b407b0463645fc2e4a0682f23\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T14:05:32Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0313 14:05:03.905343 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0313 14:05:03.908030 1 observer_polling.go:159] Starting file observer\\\\nI0313 14:05:03.938230 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0313 14:05:03.942399 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0313 14:05:32.345768 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0313 14:05:32.345932 1 cmd.go:179] failed checking apiserver connectivity: Get \\\\\\\"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/openshift-kube-controller-manager/leases/cluster-policy-controller-lock\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:31Z is after 2026-02-23T05:33:13Z\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:03Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18eadc065bf4a307b7b443fcfe79b2b719938ee486b53321c0a9f20926a0a887\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9abfa5a9ad1a98a2907ead28390ee926e176646fe761b89f580101bd429cf747\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbc67477e0223db1e4b403be3709fe71fe6409570361e103eb730fdeaffe3e80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:42Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:42 crc kubenswrapper[4907]: I0313 14:07:42.769414 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:42Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:42 crc kubenswrapper[4907]: I0313 14:07:42.782377 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:07:42 crc kubenswrapper[4907]: I0313 14:07:42.782414 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:07:42 crc kubenswrapper[4907]: I0313 14:07:42.782457 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afb1d430459aafab2d6bc2507b4b4059529401e51fdf8e9848b0d7bb8651db68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:42Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:42 crc kubenswrapper[4907]: I0313 14:07:42.782562 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:07:42 crc kubenswrapper[4907]: E0313 14:07:42.782617 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:07:42 crc kubenswrapper[4907]: I0313 14:07:42.782632 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:07:42 crc kubenswrapper[4907]: E0313 14:07:42.782694 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:07:42 crc kubenswrapper[4907]: E0313 14:07:42.782967 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzmvc" podUID="f797b1bb-e769-49c4-84ce-93a6fad2df93" Mar 13 14:07:42 crc kubenswrapper[4907]: E0313 14:07:42.783129 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:07:42 crc kubenswrapper[4907]: I0313 14:07:42.795842 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-745zx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58e11265-2fd9-4b9b-926c-3343bb84ee53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ce7ad32fd2bf2332f3ee41080115b083c34326b7dc31c950098dfba15f58422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mnz7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-745zx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:42Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:42 crc kubenswrapper[4907]: I0313 14:07:42.808404 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-tzmvc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f797b1bb-e769-49c4-84ce-93a6fad2df93\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tdzl9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tdzl9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:07:00Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-tzmvc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:42Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:43 crc kubenswrapper[4907]: I0313 14:07:43.795955 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Mar 13 14:07:44 crc kubenswrapper[4907]: I0313 14:07:44.782032 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:07:44 crc kubenswrapper[4907]: I0313 14:07:44.782032 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:07:44 crc kubenswrapper[4907]: E0313 14:07:44.782171 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzmvc" podUID="f797b1bb-e769-49c4-84ce-93a6fad2df93" Mar 13 14:07:44 crc kubenswrapper[4907]: I0313 14:07:44.782055 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:07:44 crc kubenswrapper[4907]: E0313 14:07:44.782291 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:07:44 crc kubenswrapper[4907]: E0313 14:07:44.782314 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:07:44 crc kubenswrapper[4907]: I0313 14:07:44.782336 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:07:44 crc kubenswrapper[4907]: E0313 14:07:44.782399 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:07:46 crc kubenswrapper[4907]: I0313 14:07:46.781800 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:07:46 crc kubenswrapper[4907]: I0313 14:07:46.781842 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:07:46 crc kubenswrapper[4907]: I0313 14:07:46.781871 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:07:46 crc kubenswrapper[4907]: I0313 14:07:46.781848 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:07:46 crc kubenswrapper[4907]: E0313 14:07:46.781948 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:07:46 crc kubenswrapper[4907]: E0313 14:07:46.782026 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:07:46 crc kubenswrapper[4907]: E0313 14:07:46.782089 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzmvc" podUID="f797b1bb-e769-49c4-84ce-93a6fad2df93" Mar 13 14:07:46 crc kubenswrapper[4907]: E0313 14:07:46.782134 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:07:46 crc kubenswrapper[4907]: E0313 14:07:46.989943 4907 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 13 14:07:48 crc kubenswrapper[4907]: I0313 14:07:48.400792 4907 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" podUID="1006376d-4033-454e-8a23-2b67882bdd0b" containerName="ovnkube-controller" probeResult="failure" output="" Mar 13 14:07:48 crc kubenswrapper[4907]: I0313 14:07:48.543392 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:07:48 crc kubenswrapper[4907]: I0313 14:07:48.543451 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:07:48 crc kubenswrapper[4907]: I0313 14:07:48.543468 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:07:48 crc kubenswrapper[4907]: I0313 14:07:48.543490 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:07:48 crc kubenswrapper[4907]: I0313 14:07:48.543506 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:07:48Z","lastTransitionTime":"2026-03-13T14:07:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:07:48 crc kubenswrapper[4907]: E0313 14:07:48.556500 4907 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4b182bae-7f2d-4b54-934c-85a9a277d132\\\",\\\"systemUUID\\\":\\\"05bfbfa3-f04f-45ba-8233-a355fc774c24\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:48Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:48 crc kubenswrapper[4907]: I0313 14:07:48.559735 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:07:48 crc kubenswrapper[4907]: I0313 14:07:48.559776 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:07:48 crc kubenswrapper[4907]: I0313 14:07:48.559788 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:07:48 crc kubenswrapper[4907]: I0313 14:07:48.559805 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:07:48 crc kubenswrapper[4907]: I0313 14:07:48.559817 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:07:48Z","lastTransitionTime":"2026-03-13T14:07:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:07:48 crc kubenswrapper[4907]: E0313 14:07:48.577245 4907 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4b182bae-7f2d-4b54-934c-85a9a277d132\\\",\\\"systemUUID\\\":\\\"05bfbfa3-f04f-45ba-8233-a355fc774c24\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:48Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:48 crc kubenswrapper[4907]: I0313 14:07:48.581395 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:07:48 crc kubenswrapper[4907]: I0313 14:07:48.581473 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:07:48 crc kubenswrapper[4907]: I0313 14:07:48.581497 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:07:48 crc kubenswrapper[4907]: I0313 14:07:48.581529 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:07:48 crc kubenswrapper[4907]: I0313 14:07:48.581554 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:07:48Z","lastTransitionTime":"2026-03-13T14:07:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:07:48 crc kubenswrapper[4907]: E0313 14:07:48.596317 4907 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4b182bae-7f2d-4b54-934c-85a9a277d132\\\",\\\"systemUUID\\\":\\\"05bfbfa3-f04f-45ba-8233-a355fc774c24\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:48Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:48 crc kubenswrapper[4907]: I0313 14:07:48.600234 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:07:48 crc kubenswrapper[4907]: I0313 14:07:48.600273 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:07:48 crc kubenswrapper[4907]: I0313 14:07:48.600287 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:07:48 crc kubenswrapper[4907]: I0313 14:07:48.600307 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:07:48 crc kubenswrapper[4907]: I0313 14:07:48.600320 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:07:48Z","lastTransitionTime":"2026-03-13T14:07:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:07:48 crc kubenswrapper[4907]: E0313 14:07:48.618017 4907 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4b182bae-7f2d-4b54-934c-85a9a277d132\\\",\\\"systemUUID\\\":\\\"05bfbfa3-f04f-45ba-8233-a355fc774c24\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:48Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:48 crc kubenswrapper[4907]: I0313 14:07:48.621838 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:07:48 crc kubenswrapper[4907]: I0313 14:07:48.621873 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:07:48 crc kubenswrapper[4907]: I0313 14:07:48.621895 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:07:48 crc kubenswrapper[4907]: I0313 14:07:48.621908 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:07:48 crc kubenswrapper[4907]: I0313 14:07:48.621919 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:07:48Z","lastTransitionTime":"2026-03-13T14:07:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:07:48 crc kubenswrapper[4907]: E0313 14:07:48.634359 4907 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4b182bae-7f2d-4b54-934c-85a9a277d132\\\",\\\"systemUUID\\\":\\\"05bfbfa3-f04f-45ba-8233-a355fc774c24\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:48Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:48 crc kubenswrapper[4907]: E0313 14:07:48.634504 4907 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 13 14:07:48 crc kubenswrapper[4907]: I0313 14:07:48.781592 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:07:48 crc kubenswrapper[4907]: I0313 14:07:48.781650 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:07:48 crc kubenswrapper[4907]: I0313 14:07:48.781614 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:07:48 crc kubenswrapper[4907]: E0313 14:07:48.781741 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:07:48 crc kubenswrapper[4907]: I0313 14:07:48.781807 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:07:48 crc kubenswrapper[4907]: E0313 14:07:48.781959 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:07:48 crc kubenswrapper[4907]: E0313 14:07:48.781998 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzmvc" podUID="f797b1bb-e769-49c4-84ce-93a6fad2df93" Mar 13 14:07:48 crc kubenswrapper[4907]: E0313 14:07:48.782067 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:07:50 crc kubenswrapper[4907]: I0313 14:07:50.781499 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:07:50 crc kubenswrapper[4907]: I0313 14:07:50.781545 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:07:50 crc kubenswrapper[4907]: I0313 14:07:50.781589 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:07:50 crc kubenswrapper[4907]: I0313 14:07:50.781629 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:07:50 crc kubenswrapper[4907]: E0313 14:07:50.781623 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:07:50 crc kubenswrapper[4907]: E0313 14:07:50.781723 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:07:50 crc kubenswrapper[4907]: E0313 14:07:50.781864 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzmvc" podUID="f797b1bb-e769-49c4-84ce-93a6fad2df93" Mar 13 14:07:50 crc kubenswrapper[4907]: E0313 14:07:50.781922 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:07:51 crc kubenswrapper[4907]: I0313 14:07:51.803200 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5b4819028f34db81cc43a04e4854c30488ed876c63b8daff11105bf61546c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:51Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:51 crc kubenswrapper[4907]: I0313 14:07:51.824563 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:51Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:51 crc kubenswrapper[4907]: I0313 14:07:51.842691 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xtwf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64ebd338-0158-44dc-90ab-e7c285e87762\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef632bb17c8f529b35d0882c8d4ba3dc213b0308b46a4ade8426b9c08e64bb2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://391b8bc1e975982b172175fcf4b4509756b018cbb99c12e48f4931ff472cdae2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T14:07:34Z\\\",\\\"message\\\":\\\"2026-03-13T14:06:49+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_baa16ed5-a467-4fd5-8846-0c869be5e531\\\\n2026-03-13T14:06:49+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_baa16ed5-a467-4fd5-8846-0c869be5e531 to /host/opt/cni/bin/\\\\n2026-03-13T14:06:49Z [verbose] multus-daemon started\\\\n2026-03-13T14:06:49Z [verbose] Readiness Indicator file check\\\\n2026-03-13T14:07:34Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:07:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6cwbq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xtwf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:51Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:51 crc kubenswrapper[4907]: I0313 14:07:51.869066 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bed3bcea-e8c0-4b00-a985-e0c5b5bea100\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://402c1ef89085d840edb4313577f5b2fb967072cb3db00c34c98572f00699529c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f22becea3c68103ff0b2639c7ae8c051b76d19b407b0463645fc2e4a0682f23\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T14:05:32Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0313 14:05:03.905343 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0313 14:05:03.908030 1 observer_polling.go:159] Starting file observer\\\\nI0313 14:05:03.938230 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0313 14:05:03.942399 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0313 14:05:32.345768 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0313 14:05:32.345932 1 cmd.go:179] failed checking apiserver connectivity: Get \\\\\\\"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/openshift-kube-controller-manager/leases/cluster-policy-controller-lock\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:31Z is after 2026-02-23T05:33:13Z\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:03Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18eadc065bf4a307b7b443fcfe79b2b719938ee486b53321c0a9f20926a0a887\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9abfa5a9ad1a98a2907ead28390ee926e176646fe761b89f580101bd429cf747\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbc67477e0223db1e4b403be3709fe71fe6409570361e103eb730fdeaffe3e80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:51Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:51 crc kubenswrapper[4907]: I0313 14:07:51.921647 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:51Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:51 crc kubenswrapper[4907]: I0313 14:07:51.938813 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afb1d430459aafab2d6bc2507b4b4059529401e51fdf8e9848b0d7bb8651db68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:51Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:51 crc kubenswrapper[4907]: I0313 14:07:51.953438 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-745zx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58e11265-2fd9-4b9b-926c-3343bb84ee53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ce7ad32fd2bf2332f3ee41080115b083c34326b7dc31c950098dfba15f58422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mnz7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-745zx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:51Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:51 crc kubenswrapper[4907]: I0313 14:07:51.965784 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-tzmvc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f797b1bb-e769-49c4-84ce-93a6fad2df93\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tdzl9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tdzl9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:07:00Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-tzmvc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:51Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:51 crc kubenswrapper[4907]: I0313 14:07:51.975636 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a21d7e2-0481-4ea4-b129-0e2754315289\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35fd0db86ae409203bc57f755e489cd542ea68d19b5fcb9023794a77bcfa1a64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22af7218ec732d25874992d8015afd1d73b48cb6cf7d74e638c9db396cb9efe3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22af7218ec732d25874992d8015afd1d73b48cb6cf7d74e638c9db396cb9efe3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:51Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:51 crc kubenswrapper[4907]: I0313 14:07:51.987553 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"584d221c-be74-4a39-8173-6c8e3f55f686\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b782a795257057929fe616938d1483a9accbd7a93f0cba963f9791a3d4ec4749\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c826e7b9be5e3a0ecdaa6dfeaf2454535e542ff86acc532cc3c80acf4a95c1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4333bdf047392cb3f64e29e677fe8d2f21c7ebca1a7e7097167e70c6785b8d46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a4dea84b35a82ce16f0e0d38057dc994b26240cf45dcc9111ed0de573a62881\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a4dea84b35a82ce16f0e0d38057dc994b26240cf45dcc9111ed0de573a62881\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:51Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:51 crc kubenswrapper[4907]: E0313 14:07:51.990968 4907 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 13 14:07:52 crc kubenswrapper[4907]: I0313 14:07:52.001266 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:51Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:52 crc kubenswrapper[4907]: I0313 14:07:52.017471 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-p8ptc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57e4e912-9a98-4419-87ce-f16e178ed921\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38f113418d62dab1728f0ba57cd1d289e3ec41621a338017e74c9996a8449b1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bdef68f401b5ba2c7cf61551eba416e74f295ab2281433a5115334d28094818\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bdef68f401b5ba2c7cf61551eba416e74f295ab2281433a5115334d28094818\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f8d1f40744d609fd964572ef5e323f0283d2582e98745602754165818f17847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f8d1f40744d609fd964572ef5e323f0283d2582e98745602754165818f17847\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c1ab2e148eaa177b98f5883a3f323d12f5b662bff8cb2d072a97be18ab8e120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c1ab2e148eaa177b98f5883a3f323d12f5b662bff8cb2d072a97be18ab8e120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-p8ptc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:52Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:52 crc kubenswrapper[4907]: I0313 14:07:52.029071 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3fb4213-144f-4b01-9d78-ff3315761bb3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba1140a2462713c2401b47a4cf1f231bc9d9ae7867a7d0b1b597214a1f7efe07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://770c882d9e9f9b2e9bd4f135e591952fa15b3898c42826aed8a75d953c9a875b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hm56j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:52Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:52 crc kubenswrapper[4907]: I0313 14:07:52.055336 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1006376d-4033-454e-8a23-2b67882bdd0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf6a3dbaa8dda61f461a906538824c8946d36dc07d4fb95bf4d8420ad0105600\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccef92e41501fab91189581d7e8875515c02d637426b329d2fb80a356c75b637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0ee0049aa4a6d49a6bcf7d347183d9cc8b85171f988662803a005e0cdd97395\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fff87d8ea572615eb1329164f9e536e6ee45cd00a9e71be0aa8a9b3ed2190e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c165053fd9933444b372ff40efe33c0e193a7f39e6f531a4f370e971eb41e6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fea41c11bebd2d9d622ed4b8bc50e03a106dadea6ac102cd60512ea2a1f8b8b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9155434e2d8f0daa4504758fb72d6f498da107a64a92a0883ac687ed2af9caee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a99cabe30a7abc8da854bf86d79628b94a6ffc20c28f828b695080332b319db2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T14:07:15Z\\\",\\\"message\\\":\\\"Sending *v1.Node event handler 7 for removal\\\\nI0313 14:07:15.619307 7162 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0313 14:07:15.619320 7162 handler.go:208] Removed *v1.Node event handler 2\\\\nI0313 14:07:15.619329 7162 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0313 14:07:15.619343 7162 handler.go:208] Removed *v1.Node event handler 7\\\\nI0313 14:07:15.619351 7162 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0313 14:07:15.619358 7162 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0313 14:07:15.619362 7162 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0313 14:07:15.619381 7162 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0313 14:07:15.619382 7162 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0313 14:07:15.619392 7162 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0313 14:07:15.619619 7162 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI0313 14:07:15.619713 7162 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI0313 14:07:15.619975 7162 factory.go:656] Stopping watch factory\\\\nI0313 14:07:15.620001 7162 ovnkube.go:599] Stopped ovnkube\\\\nI0313 14:07:15.620033 7162 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0313 14:07:15.620096 7162 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:07:14Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:07:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d60b5b774f8dd4cee16b9784f80d124640199a85331a2940772029be4884751\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qxlx5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:52Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:52 crc kubenswrapper[4907]: I0313 14:07:52.086465 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31a594bd-13e7-4c68-ac7c-f481626088ec\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8141bbe3b2b50ad7d2aed26f813a951bc6db4c90bf8271cd23a125cc7ec2eb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb841f47d97b3e9495658ff7f6aa3f835905a9cd2ee9736774c3504d6c59a209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f04995f26a4c320fa30f2a97b1299284cbb2e23965bae16a6cca8d06eef630eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0568aef54a902fedcaaa71cabe0add0cbeb3a38743a7fe45584c5e315292afa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://171ea306a9ec518a88ef509adaab1a7c238abf7b5824a87d3425de7199cde520\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14a9ee19b96b2019de6ad9e5092e0004a4587a9f678fec417c4ac02e60695ac8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14a9ee19b96b2019de6ad9e5092e0004a4587a9f678fec417c4ac02e60695ac8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d46350802ae58ed4c464dc3f8d3f9dd7e091ef8091659e0711b9d2632e00708\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d46350802ae58ed4c464dc3f8d3f9dd7e091ef8091659e0711b9d2632e00708\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://65bd2db6ca4ffb26d9cb4bce44dc27a862353533bb0ec584bcc688b73fbec3eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65bd2db6ca4ffb26d9cb4bce44dc27a862353533bb0ec584bcc688b73fbec3eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:52Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:52 crc kubenswrapper[4907]: I0313 14:07:52.108641 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5d2b054-0ed8-419c-b731-f14d3598372f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4b503aa3b13268b3688d067f3e456c86afac1fda171aced0990899c3eb89b79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be40a1a33fa482eab363537afaa901127d560afbc1b5b287ef11e66391a5be4c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d1b539bbd918fc50428fc18cabcc4c8a43cc96f220b5692d67cf1c8dfbf8ed8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93d8c8c79be9d9ffb72de66fc2f3f9821936023071e95875582b45ccb319601f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T14:06:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0313 14:06:13.202871 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0313 14:06:13.203175 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0313 14:06:13.204427 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2471728650/tls.crt::/tmp/serving-cert-2471728650/tls.key\\\\\\\"\\\\nI0313 14:06:13.581160 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0313 14:06:13.584012 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0313 14:06:13.584036 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0313 14:06:13.584064 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0313 14:06:13.584072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0313 14:06:13.594504 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0313 14:06:13.594891 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0313 14:06:13.595323 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0313 14:06:13.595363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0313 14:06:13.595370 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0313 14:06:13.595376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0313 14:06:13.597226 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3f2b6756078925f8562b6053b3b83798efb72288f5fd18fff680a4ab0d33feb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:52Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:52 crc kubenswrapper[4907]: I0313 14:07:52.128425 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94b045cf2fed56457a2ca14f489bb69470198f159f16c207fe061d034dba1555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd28f7a5c35822582b70ed41645294aaede10e28ff59f3f3b5d36e7da8237b76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:52Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:52 crc kubenswrapper[4907]: I0313 14:07:52.144561 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l7zg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d105034b-f041-4fb5-8df6-6b1bb7681fda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21a207f917836785e320f4f79aca5d6449da2fdeec3ee0765f34d227d5617fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szmh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l7zg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:52Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:52 crc kubenswrapper[4907]: I0313 14:07:52.161034 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qncw4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1d7ee39-09d7-4c0e-942b-5e9d8c9f63be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07df71887196cc7098ac9606a4a15b89d45219048c31280198ae75022bff9370\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:07:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bg45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://540ca207f31c9a8a9892f80dee153b3fb86f8de406f38360221512f424d8c884\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:07:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bg45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:07:00Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-qncw4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:52Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:52 crc kubenswrapper[4907]: I0313 14:07:52.782062 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:07:52 crc kubenswrapper[4907]: I0313 14:07:52.782181 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:07:52 crc kubenswrapper[4907]: I0313 14:07:52.782144 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:07:52 crc kubenswrapper[4907]: I0313 14:07:52.782084 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:07:52 crc kubenswrapper[4907]: E0313 14:07:52.782372 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:07:52 crc kubenswrapper[4907]: E0313 14:07:52.782539 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:07:52 crc kubenswrapper[4907]: E0313 14:07:52.782873 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzmvc" podUID="f797b1bb-e769-49c4-84ce-93a6fad2df93" Mar 13 14:07:52 crc kubenswrapper[4907]: E0313 14:07:52.783065 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:07:54 crc kubenswrapper[4907]: I0313 14:07:54.781608 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:07:54 crc kubenswrapper[4907]: I0313 14:07:54.781673 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:07:54 crc kubenswrapper[4907]: E0313 14:07:54.781835 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:07:54 crc kubenswrapper[4907]: I0313 14:07:54.781996 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:07:54 crc kubenswrapper[4907]: I0313 14:07:54.782042 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:07:54 crc kubenswrapper[4907]: E0313 14:07:54.782205 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:07:54 crc kubenswrapper[4907]: E0313 14:07:54.782305 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzmvc" podUID="f797b1bb-e769-49c4-84ce-93a6fad2df93" Mar 13 14:07:54 crc kubenswrapper[4907]: E0313 14:07:54.782412 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:07:56 crc kubenswrapper[4907]: I0313 14:07:56.782039 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:07:56 crc kubenswrapper[4907]: E0313 14:07:56.782224 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:07:56 crc kubenswrapper[4907]: I0313 14:07:56.782323 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:07:56 crc kubenswrapper[4907]: E0313 14:07:56.782414 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzmvc" podUID="f797b1bb-e769-49c4-84ce-93a6fad2df93" Mar 13 14:07:56 crc kubenswrapper[4907]: I0313 14:07:56.782466 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:07:56 crc kubenswrapper[4907]: E0313 14:07:56.782531 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:07:56 crc kubenswrapper[4907]: I0313 14:07:56.782571 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:07:56 crc kubenswrapper[4907]: E0313 14:07:56.782649 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:07:56 crc kubenswrapper[4907]: E0313 14:07:56.992963 4907 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 13 14:07:58 crc kubenswrapper[4907]: I0313 14:07:58.782418 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:07:58 crc kubenswrapper[4907]: I0313 14:07:58.782540 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:07:58 crc kubenswrapper[4907]: E0313 14:07:58.782670 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:07:58 crc kubenswrapper[4907]: I0313 14:07:58.782934 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:07:58 crc kubenswrapper[4907]: I0313 14:07:58.783131 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:07:58 crc kubenswrapper[4907]: E0313 14:07:58.783299 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:07:58 crc kubenswrapper[4907]: E0313 14:07:58.783392 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzmvc" podUID="f797b1bb-e769-49c4-84ce-93a6fad2df93" Mar 13 14:07:58 crc kubenswrapper[4907]: E0313 14:07:58.783455 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:07:58 crc kubenswrapper[4907]: I0313 14:07:58.913548 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:07:58 crc kubenswrapper[4907]: I0313 14:07:58.913632 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:07:58 crc kubenswrapper[4907]: I0313 14:07:58.913658 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:07:58 crc kubenswrapper[4907]: I0313 14:07:58.913693 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:07:58 crc kubenswrapper[4907]: I0313 14:07:58.913717 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:07:58Z","lastTransitionTime":"2026-03-13T14:07:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:07:58 crc kubenswrapper[4907]: E0313 14:07:58.936196 4907 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4b182bae-7f2d-4b54-934c-85a9a277d132\\\",\\\"systemUUID\\\":\\\"05bfbfa3-f04f-45ba-8233-a355fc774c24\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:58Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:58 crc kubenswrapper[4907]: I0313 14:07:58.942223 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:07:58 crc kubenswrapper[4907]: I0313 14:07:58.942281 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:07:58 crc kubenswrapper[4907]: I0313 14:07:58.942300 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:07:58 crc kubenswrapper[4907]: I0313 14:07:58.942328 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:07:58 crc kubenswrapper[4907]: I0313 14:07:58.942347 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:07:58Z","lastTransitionTime":"2026-03-13T14:07:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:07:58 crc kubenswrapper[4907]: E0313 14:07:58.964148 4907 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4b182bae-7f2d-4b54-934c-85a9a277d132\\\",\\\"systemUUID\\\":\\\"05bfbfa3-f04f-45ba-8233-a355fc774c24\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:58Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:58 crc kubenswrapper[4907]: I0313 14:07:58.969912 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:07:58 crc kubenswrapper[4907]: I0313 14:07:58.969970 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:07:58 crc kubenswrapper[4907]: I0313 14:07:58.969991 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:07:58 crc kubenswrapper[4907]: I0313 14:07:58.970020 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:07:58 crc kubenswrapper[4907]: I0313 14:07:58.970040 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:07:58Z","lastTransitionTime":"2026-03-13T14:07:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:07:58 crc kubenswrapper[4907]: E0313 14:07:58.992728 4907 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4b182bae-7f2d-4b54-934c-85a9a277d132\\\",\\\"systemUUID\\\":\\\"05bfbfa3-f04f-45ba-8233-a355fc774c24\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:58Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:58 crc kubenswrapper[4907]: I0313 14:07:58.998966 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:07:58 crc kubenswrapper[4907]: I0313 14:07:58.999025 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:07:58 crc kubenswrapper[4907]: I0313 14:07:58.999041 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:07:58 crc kubenswrapper[4907]: I0313 14:07:58.999066 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:07:58 crc kubenswrapper[4907]: I0313 14:07:58.999087 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:07:58Z","lastTransitionTime":"2026-03-13T14:07:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:07:59 crc kubenswrapper[4907]: E0313 14:07:59.020542 4907 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4b182bae-7f2d-4b54-934c-85a9a277d132\\\",\\\"systemUUID\\\":\\\"05bfbfa3-f04f-45ba-8233-a355fc774c24\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:59Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:59 crc kubenswrapper[4907]: I0313 14:07:59.026297 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:07:59 crc kubenswrapper[4907]: I0313 14:07:59.026356 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:07:59 crc kubenswrapper[4907]: I0313 14:07:59.026372 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:07:59 crc kubenswrapper[4907]: I0313 14:07:59.026399 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:07:59 crc kubenswrapper[4907]: I0313 14:07:59.026418 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:07:59Z","lastTransitionTime":"2026-03-13T14:07:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:07:59 crc kubenswrapper[4907]: E0313 14:07:59.048007 4907 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4b182bae-7f2d-4b54-934c-85a9a277d132\\\",\\\"systemUUID\\\":\\\"05bfbfa3-f04f-45ba-8233-a355fc774c24\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:59Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:59 crc kubenswrapper[4907]: E0313 14:07:59.048300 4907 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 13 14:08:00 crc kubenswrapper[4907]: I0313 14:08:00.781781 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:08:00 crc kubenswrapper[4907]: I0313 14:08:00.781781 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:08:00 crc kubenswrapper[4907]: I0313 14:08:00.781828 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:08:00 crc kubenswrapper[4907]: I0313 14:08:00.781996 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:08:00 crc kubenswrapper[4907]: E0313 14:08:00.782034 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:08:00 crc kubenswrapper[4907]: E0313 14:08:00.782097 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzmvc" podUID="f797b1bb-e769-49c4-84ce-93a6fad2df93" Mar 13 14:08:00 crc kubenswrapper[4907]: E0313 14:08:00.782265 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:08:00 crc kubenswrapper[4907]: E0313 14:08:00.782461 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:08:01 crc kubenswrapper[4907]: I0313 14:08:01.805158 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afb1d430459aafab2d6bc2507b4b4059529401e51fdf8e9848b0d7bb8651db68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:08:01Z is after 2025-08-24T17:21:41Z" Mar 13 14:08:01 crc kubenswrapper[4907]: I0313 14:08:01.819116 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-745zx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58e11265-2fd9-4b9b-926c-3343bb84ee53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ce7ad32fd2bf2332f3ee41080115b083c34326b7dc31c950098dfba15f58422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mnz7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-745zx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:08:01Z is after 2025-08-24T17:21:41Z" Mar 13 14:08:01 crc kubenswrapper[4907]: I0313 14:08:01.831327 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-tzmvc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f797b1bb-e769-49c4-84ce-93a6fad2df93\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tdzl9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tdzl9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:07:00Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-tzmvc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:08:01Z is after 2025-08-24T17:21:41Z" Mar 13 14:08:01 crc kubenswrapper[4907]: I0313 14:08:01.843244 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bed3bcea-e8c0-4b00-a985-e0c5b5bea100\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://402c1ef89085d840edb4313577f5b2fb967072cb3db00c34c98572f00699529c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f22becea3c68103ff0b2639c7ae8c051b76d19b407b0463645fc2e4a0682f23\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T14:05:32Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0313 14:05:03.905343 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0313 14:05:03.908030 1 observer_polling.go:159] Starting file observer\\\\nI0313 14:05:03.938230 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0313 14:05:03.942399 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0313 14:05:32.345768 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0313 14:05:32.345932 1 cmd.go:179] failed checking apiserver connectivity: Get \\\\\\\"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/openshift-kube-controller-manager/leases/cluster-policy-controller-lock\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:31Z is after 2026-02-23T05:33:13Z\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:03Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18eadc065bf4a307b7b443fcfe79b2b719938ee486b53321c0a9f20926a0a887\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9abfa5a9ad1a98a2907ead28390ee926e176646fe761b89f580101bd429cf747\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbc67477e0223db1e4b403be3709fe71fe6409570361e103eb730fdeaffe3e80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:08:01Z is after 2025-08-24T17:21:41Z" Mar 13 14:08:01 crc kubenswrapper[4907]: I0313 14:08:01.854430 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:08:01Z is after 2025-08-24T17:21:41Z" Mar 13 14:08:01 crc kubenswrapper[4907]: I0313 14:08:01.868526 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:08:01Z is after 2025-08-24T17:21:41Z" Mar 13 14:08:01 crc kubenswrapper[4907]: I0313 14:08:01.885937 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-p8ptc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57e4e912-9a98-4419-87ce-f16e178ed921\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38f113418d62dab1728f0ba57cd1d289e3ec41621a338017e74c9996a8449b1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bdef68f401b5ba2c7cf61551eba416e74f295ab2281433a5115334d28094818\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bdef68f401b5ba2c7cf61551eba416e74f295ab2281433a5115334d28094818\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f8d1f40744d609fd964572ef5e323f0283d2582e98745602754165818f17847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f8d1f40744d609fd964572ef5e323f0283d2582e98745602754165818f17847\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c1ab2e148eaa177b98f5883a3f323d12f5b662bff8cb2d072a97be18ab8e120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c1ab2e148eaa177b98f5883a3f323d12f5b662bff8cb2d072a97be18ab8e120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-p8ptc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:08:01Z is after 2025-08-24T17:21:41Z" Mar 13 14:08:01 crc kubenswrapper[4907]: I0313 14:08:01.905691 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3fb4213-144f-4b01-9d78-ff3315761bb3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba1140a2462713c2401b47a4cf1f231bc9d9ae7867a7d0b1b597214a1f7efe07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://770c882d9e9f9b2e9bd4f135e591952fa15b3898c42826aed8a75d953c9a875b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hm56j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:08:01Z is after 2025-08-24T17:21:41Z" Mar 13 14:08:01 crc kubenswrapper[4907]: I0313 14:08:01.929268 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1006376d-4033-454e-8a23-2b67882bdd0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf6a3dbaa8dda61f461a906538824c8946d36dc07d4fb95bf4d8420ad0105600\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccef92e41501fab91189581d7e8875515c02d637426b329d2fb80a356c75b637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0ee0049aa4a6d49a6bcf7d347183d9cc8b85171f988662803a005e0cdd97395\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fff87d8ea572615eb1329164f9e536e6ee45cd00a9e71be0aa8a9b3ed2190e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c165053fd9933444b372ff40efe33c0e193a7f39e6f531a4f370e971eb41e6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fea41c11bebd2d9d622ed4b8bc50e03a106dadea6ac102cd60512ea2a1f8b8b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9155434e2d8f0daa4504758fb72d6f498da107a64a92a0883ac687ed2af9caee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a99cabe30a7abc8da854bf86d79628b94a6ffc20c28f828b695080332b319db2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T14:07:15Z\\\",\\\"message\\\":\\\"Sending *v1.Node event handler 7 for removal\\\\nI0313 14:07:15.619307 7162 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0313 14:07:15.619320 7162 handler.go:208] Removed *v1.Node event handler 2\\\\nI0313 14:07:15.619329 7162 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0313 14:07:15.619343 7162 handler.go:208] Removed *v1.Node event handler 7\\\\nI0313 14:07:15.619351 7162 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0313 14:07:15.619358 7162 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0313 14:07:15.619362 7162 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0313 14:07:15.619381 7162 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0313 14:07:15.619382 7162 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0313 14:07:15.619392 7162 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0313 14:07:15.619619 7162 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI0313 14:07:15.619713 7162 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI0313 14:07:15.619975 7162 factory.go:656] Stopping watch factory\\\\nI0313 14:07:15.620001 7162 ovnkube.go:599] Stopped ovnkube\\\\nI0313 14:07:15.620033 7162 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0313 14:07:15.620096 7162 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:07:14Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:07:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d60b5b774f8dd4cee16b9784f80d124640199a85331a2940772029be4884751\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qxlx5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:08:01Z is after 2025-08-24T17:21:41Z" Mar 13 14:08:01 crc kubenswrapper[4907]: I0313 14:08:01.942029 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a21d7e2-0481-4ea4-b129-0e2754315289\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35fd0db86ae409203bc57f755e489cd542ea68d19b5fcb9023794a77bcfa1a64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22af7218ec732d25874992d8015afd1d73b48cb6cf7d74e638c9db396cb9efe3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22af7218ec732d25874992d8015afd1d73b48cb6cf7d74e638c9db396cb9efe3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:08:01Z is after 2025-08-24T17:21:41Z" Mar 13 14:08:01 crc kubenswrapper[4907]: I0313 14:08:01.954961 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"584d221c-be74-4a39-8173-6c8e3f55f686\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b782a795257057929fe616938d1483a9accbd7a93f0cba963f9791a3d4ec4749\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c826e7b9be5e3a0ecdaa6dfeaf2454535e542ff86acc532cc3c80acf4a95c1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4333bdf047392cb3f64e29e677fe8d2f21c7ebca1a7e7097167e70c6785b8d46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a4dea84b35a82ce16f0e0d38057dc994b26240cf45dcc9111ed0de573a62881\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a4dea84b35a82ce16f0e0d38057dc994b26240cf45dcc9111ed0de573a62881\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:08:01Z is after 2025-08-24T17:21:41Z" Mar 13 14:08:01 crc kubenswrapper[4907]: I0313 14:08:01.970131 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94b045cf2fed56457a2ca14f489bb69470198f159f16c207fe061d034dba1555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd28f7a5c35822582b70ed41645294aaede10e28ff59f3f3b5d36e7da8237b76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:08:01Z is after 2025-08-24T17:21:41Z" Mar 13 14:08:01 crc kubenswrapper[4907]: I0313 14:08:01.984226 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l7zg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d105034b-f041-4fb5-8df6-6b1bb7681fda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21a207f917836785e320f4f79aca5d6449da2fdeec3ee0765f34d227d5617fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szmh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l7zg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:08:01Z is after 2025-08-24T17:21:41Z" Mar 13 14:08:01 crc kubenswrapper[4907]: E0313 14:08:01.994292 4907 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 13 14:08:02 crc kubenswrapper[4907]: I0313 14:08:02.002044 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qncw4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1d7ee39-09d7-4c0e-942b-5e9d8c9f63be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07df71887196cc7098ac9606a4a15b89d45219048c31280198ae75022bff9370\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:07:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bg45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://540ca207f31c9a8a9892f80dee153b3fb86f8de406f38360221512f424d8c884\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:07:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bg45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:07:00Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-qncw4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:08:01Z is after 2025-08-24T17:21:41Z" Mar 13 14:08:02 crc kubenswrapper[4907]: I0313 14:08:02.036083 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31a594bd-13e7-4c68-ac7c-f481626088ec\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8141bbe3b2b50ad7d2aed26f813a951bc6db4c90bf8271cd23a125cc7ec2eb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb841f47d97b3e9495658ff7f6aa3f835905a9cd2ee9736774c3504d6c59a209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f04995f26a4c320fa30f2a97b1299284cbb2e23965bae16a6cca8d06eef630eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0568aef54a902fedcaaa71cabe0add0cbeb3a38743a7fe45584c5e315292afa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://171ea306a9ec518a88ef509adaab1a7c238abf7b5824a87d3425de7199cde520\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14a9ee19b96b2019de6ad9e5092e0004a4587a9f678fec417c4ac02e60695ac8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14a9ee19b96b2019de6ad9e5092e0004a4587a9f678fec417c4ac02e60695ac8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d46350802ae58ed4c464dc3f8d3f9dd7e091ef8091659e0711b9d2632e00708\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d46350802ae58ed4c464dc3f8d3f9dd7e091ef8091659e0711b9d2632e00708\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://65bd2db6ca4ffb26d9cb4bce44dc27a862353533bb0ec584bcc688b73fbec3eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65bd2db6ca4ffb26d9cb4bce44dc27a862353533bb0ec584bcc688b73fbec3eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:08:02Z is after 2025-08-24T17:21:41Z" Mar 13 14:08:02 crc kubenswrapper[4907]: I0313 14:08:02.054124 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5d2b054-0ed8-419c-b731-f14d3598372f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4b503aa3b13268b3688d067f3e456c86afac1fda171aced0990899c3eb89b79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be40a1a33fa482eab363537afaa901127d560afbc1b5b287ef11e66391a5be4c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d1b539bbd918fc50428fc18cabcc4c8a43cc96f220b5692d67cf1c8dfbf8ed8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93d8c8c79be9d9ffb72de66fc2f3f9821936023071e95875582b45ccb319601f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T14:06:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0313 14:06:13.202871 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0313 14:06:13.203175 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0313 14:06:13.204427 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2471728650/tls.crt::/tmp/serving-cert-2471728650/tls.key\\\\\\\"\\\\nI0313 14:06:13.581160 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0313 14:06:13.584012 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0313 14:06:13.584036 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0313 14:06:13.584064 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0313 14:06:13.584072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0313 14:06:13.594504 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0313 14:06:13.594891 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0313 14:06:13.595323 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0313 14:06:13.595363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0313 14:06:13.595370 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0313 14:06:13.595376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0313 14:06:13.597226 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3f2b6756078925f8562b6053b3b83798efb72288f5fd18fff680a4ab0d33feb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:08:02Z is after 2025-08-24T17:21:41Z" Mar 13 14:08:02 crc kubenswrapper[4907]: I0313 14:08:02.075208 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xtwf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64ebd338-0158-44dc-90ab-e7c285e87762\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef632bb17c8f529b35d0882c8d4ba3dc213b0308b46a4ade8426b9c08e64bb2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://391b8bc1e975982b172175fcf4b4509756b018cbb99c12e48f4931ff472cdae2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T14:07:34Z\\\",\\\"message\\\":\\\"2026-03-13T14:06:49+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_baa16ed5-a467-4fd5-8846-0c869be5e531\\\\n2026-03-13T14:06:49+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_baa16ed5-a467-4fd5-8846-0c869be5e531 to /host/opt/cni/bin/\\\\n2026-03-13T14:06:49Z [verbose] multus-daemon started\\\\n2026-03-13T14:06:49Z [verbose] Readiness Indicator file check\\\\n2026-03-13T14:07:34Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:07:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6cwbq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xtwf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:08:02Z is after 2025-08-24T17:21:41Z" Mar 13 14:08:02 crc kubenswrapper[4907]: I0313 14:08:02.091690 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5b4819028f34db81cc43a04e4854c30488ed876c63b8daff11105bf61546c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:08:02Z is after 2025-08-24T17:21:41Z" Mar 13 14:08:02 crc kubenswrapper[4907]: I0313 14:08:02.111107 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:08:02Z is after 2025-08-24T17:21:41Z" Mar 13 14:08:02 crc kubenswrapper[4907]: I0313 14:08:02.781505 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:08:02 crc kubenswrapper[4907]: E0313 14:08:02.781706 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzmvc" podUID="f797b1bb-e769-49c4-84ce-93a6fad2df93" Mar 13 14:08:02 crc kubenswrapper[4907]: I0313 14:08:02.782066 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:08:02 crc kubenswrapper[4907]: I0313 14:08:02.782132 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:08:02 crc kubenswrapper[4907]: I0313 14:08:02.782172 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:08:02 crc kubenswrapper[4907]: E0313 14:08:02.782319 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:08:02 crc kubenswrapper[4907]: E0313 14:08:02.782495 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:08:02 crc kubenswrapper[4907]: E0313 14:08:02.782593 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:08:03 crc kubenswrapper[4907]: I0313 14:08:03.673350 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qxlx5_1006376d-4033-454e-8a23-2b67882bdd0b/ovnkube-controller/3.log" Mar 13 14:08:03 crc kubenswrapper[4907]: I0313 14:08:03.674474 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qxlx5_1006376d-4033-454e-8a23-2b67882bdd0b/ovnkube-controller/2.log" Mar 13 14:08:03 crc kubenswrapper[4907]: I0313 14:08:03.687504 4907 generic.go:334] "Generic (PLEG): container finished" podID="1006376d-4033-454e-8a23-2b67882bdd0b" containerID="9155434e2d8f0daa4504758fb72d6f498da107a64a92a0883ac687ed2af9caee" exitCode=1 Mar 13 14:08:03 crc kubenswrapper[4907]: I0313 14:08:03.687564 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" event={"ID":"1006376d-4033-454e-8a23-2b67882bdd0b","Type":"ContainerDied","Data":"9155434e2d8f0daa4504758fb72d6f498da107a64a92a0883ac687ed2af9caee"} Mar 13 14:08:03 crc kubenswrapper[4907]: I0313 14:08:03.687615 4907 scope.go:117] "RemoveContainer" containerID="a99cabe30a7abc8da854bf86d79628b94a6ffc20c28f828b695080332b319db2" Mar 13 14:08:03 crc kubenswrapper[4907]: I0313 14:08:03.688801 4907 scope.go:117] "RemoveContainer" containerID="9155434e2d8f0daa4504758fb72d6f498da107a64a92a0883ac687ed2af9caee" Mar 13 14:08:03 crc kubenswrapper[4907]: E0313 14:08:03.689216 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-qxlx5_openshift-ovn-kubernetes(1006376d-4033-454e-8a23-2b67882bdd0b)\"" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" podUID="1006376d-4033-454e-8a23-2b67882bdd0b" Mar 13 14:08:03 crc kubenswrapper[4907]: I0313 14:08:03.729115 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1006376d-4033-454e-8a23-2b67882bdd0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf6a3dbaa8dda61f461a906538824c8946d36dc07d4fb95bf4d8420ad0105600\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccef92e41501fab91189581d7e8875515c02d637426b329d2fb80a356c75b637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0ee0049aa4a6d49a6bcf7d347183d9cc8b85171f988662803a005e0cdd97395\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fff87d8ea572615eb1329164f9e536e6ee45cd00a9e71be0aa8a9b3ed2190e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c165053fd9933444b372ff40efe33c0e193a7f39e6f531a4f370e971eb41e6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fea41c11bebd2d9d622ed4b8bc50e03a106dadea6ac102cd60512ea2a1f8b8b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9155434e2d8f0daa4504758fb72d6f498da107a64a92a0883ac687ed2af9caee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a99cabe30a7abc8da854bf86d79628b94a6ffc20c28f828b695080332b319db2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T14:07:15Z\\\",\\\"message\\\":\\\"Sending *v1.Node event handler 7 for removal\\\\nI0313 14:07:15.619307 7162 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0313 14:07:15.619320 7162 handler.go:208] Removed *v1.Node event handler 2\\\\nI0313 14:07:15.619329 7162 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0313 14:07:15.619343 7162 handler.go:208] Removed *v1.Node event handler 7\\\\nI0313 14:07:15.619351 7162 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0313 14:07:15.619358 7162 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0313 14:07:15.619362 7162 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0313 14:07:15.619381 7162 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0313 14:07:15.619382 7162 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0313 14:07:15.619392 7162 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0313 14:07:15.619619 7162 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI0313 14:07:15.619713 7162 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI0313 14:07:15.619975 7162 factory.go:656] Stopping watch factory\\\\nI0313 14:07:15.620001 7162 ovnkube.go:599] Stopped ovnkube\\\\nI0313 14:07:15.620033 7162 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0313 14:07:15.620096 7162 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:07:14Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9155434e2d8f0daa4504758fb72d6f498da107a64a92a0883ac687ed2af9caee\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T14:08:02Z\\\",\\\"message\\\":\\\"dn-host-isolation-manager: error found while processing openshift-dns-operator/dns-operator-744455d44c-xkdjh: failed to check if pod openshift-dns-operator/dns-operator-744455d44c-xkdjh is in primary UDN: could not find OVN pod annotation in map[]\\\\nI0313 14:08:02.544936 7501 controller.go:257] Controller udn-host-isolation-manager: error found while processing openshift-kube-storage-version-migrator/migrator-59844c95c7-vfbsx: failed to check if pod openshift-kube-storage-version-migrator/migrator-59844c95c7-vfbsx is in primary UDN: could not find OVN pod annotation in map[openshift.io/required-scc:nonroot-v2 openshift.io/scc:nonroot-v2 seccomp.security.alpha.kubernetes.io/pod:runtime/default]\\\\nI0313 14:08:02.544957 7501 controller.go:257] Controller udn-host-isolation-manager: error found while processing openshift-multus/network-metrics-daemon-tzmvc: failed to check if pod openshift-multus/network-metrics-daemon-tzmvc is in primary UDN: could not find OVN pod annotation in map[cluster-autoscaler.kubernetes.io/enable-ds-eviction:false]\\\\nE0313 14:08:02.629503 7501 shared_informer.go:316] \\\\\\\"Unhandled Error\\\\\\\" err=\\\\\\\"unable to sync caches for ovn-lb-controller\\\\\\\" logger=\\\\\\\"UnhandledError\\\\\\\"\\\\nI0313 14:08:02.630816 7501 ovnkube.go:599] Stopped ovnkube\\\\nI0313 14:08:02.630949 7501 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:07:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d60b5b774f8dd4cee16b9784f80d124640199a85331a2940772029be4884751\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qxlx5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:08:03Z is after 2025-08-24T17:21:41Z" Mar 13 14:08:03 crc kubenswrapper[4907]: I0313 14:08:03.747752 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a21d7e2-0481-4ea4-b129-0e2754315289\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35fd0db86ae409203bc57f755e489cd542ea68d19b5fcb9023794a77bcfa1a64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22af7218ec732d25874992d8015afd1d73b48cb6cf7d74e638c9db396cb9efe3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22af7218ec732d25874992d8015afd1d73b48cb6cf7d74e638c9db396cb9efe3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:08:03Z is after 2025-08-24T17:21:41Z" Mar 13 14:08:03 crc kubenswrapper[4907]: I0313 14:08:03.764313 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"584d221c-be74-4a39-8173-6c8e3f55f686\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b782a795257057929fe616938d1483a9accbd7a93f0cba963f9791a3d4ec4749\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c826e7b9be5e3a0ecdaa6dfeaf2454535e542ff86acc532cc3c80acf4a95c1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4333bdf047392cb3f64e29e677fe8d2f21c7ebca1a7e7097167e70c6785b8d46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a4dea84b35a82ce16f0e0d38057dc994b26240cf45dcc9111ed0de573a62881\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a4dea84b35a82ce16f0e0d38057dc994b26240cf45dcc9111ed0de573a62881\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:08:03Z is after 2025-08-24T17:21:41Z" Mar 13 14:08:03 crc kubenswrapper[4907]: I0313 14:08:03.785779 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:08:03Z is after 2025-08-24T17:21:41Z" Mar 13 14:08:03 crc kubenswrapper[4907]: I0313 14:08:03.812618 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-p8ptc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57e4e912-9a98-4419-87ce-f16e178ed921\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38f113418d62dab1728f0ba57cd1d289e3ec41621a338017e74c9996a8449b1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bdef68f401b5ba2c7cf61551eba416e74f295ab2281433a5115334d28094818\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bdef68f401b5ba2c7cf61551eba416e74f295ab2281433a5115334d28094818\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f8d1f40744d609fd964572ef5e323f0283d2582e98745602754165818f17847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f8d1f40744d609fd964572ef5e323f0283d2582e98745602754165818f17847\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c1ab2e148eaa177b98f5883a3f323d12f5b662bff8cb2d072a97be18ab8e120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c1ab2e148eaa177b98f5883a3f323d12f5b662bff8cb2d072a97be18ab8e120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-p8ptc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:08:03Z is after 2025-08-24T17:21:41Z" Mar 13 14:08:03 crc kubenswrapper[4907]: I0313 14:08:03.831331 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3fb4213-144f-4b01-9d78-ff3315761bb3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba1140a2462713c2401b47a4cf1f231bc9d9ae7867a7d0b1b597214a1f7efe07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://770c882d9e9f9b2e9bd4f135e591952fa15b3898c42826aed8a75d953c9a875b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hm56j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:08:03Z is after 2025-08-24T17:21:41Z" Mar 13 14:08:03 crc kubenswrapper[4907]: I0313 14:08:03.863534 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31a594bd-13e7-4c68-ac7c-f481626088ec\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8141bbe3b2b50ad7d2aed26f813a951bc6db4c90bf8271cd23a125cc7ec2eb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb841f47d97b3e9495658ff7f6aa3f835905a9cd2ee9736774c3504d6c59a209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f04995f26a4c320fa30f2a97b1299284cbb2e23965bae16a6cca8d06eef630eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0568aef54a902fedcaaa71cabe0add0cbeb3a38743a7fe45584c5e315292afa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://171ea306a9ec518a88ef509adaab1a7c238abf7b5824a87d3425de7199cde520\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14a9ee19b96b2019de6ad9e5092e0004a4587a9f678fec417c4ac02e60695ac8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14a9ee19b96b2019de6ad9e5092e0004a4587a9f678fec417c4ac02e60695ac8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d46350802ae58ed4c464dc3f8d3f9dd7e091ef8091659e0711b9d2632e00708\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d46350802ae58ed4c464dc3f8d3f9dd7e091ef8091659e0711b9d2632e00708\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://65bd2db6ca4ffb26d9cb4bce44dc27a862353533bb0ec584bcc688b73fbec3eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65bd2db6ca4ffb26d9cb4bce44dc27a862353533bb0ec584bcc688b73fbec3eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:08:03Z is after 2025-08-24T17:21:41Z" Mar 13 14:08:03 crc kubenswrapper[4907]: I0313 14:08:03.882037 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5d2b054-0ed8-419c-b731-f14d3598372f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4b503aa3b13268b3688d067f3e456c86afac1fda171aced0990899c3eb89b79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be40a1a33fa482eab363537afaa901127d560afbc1b5b287ef11e66391a5be4c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d1b539bbd918fc50428fc18cabcc4c8a43cc96f220b5692d67cf1c8dfbf8ed8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93d8c8c79be9d9ffb72de66fc2f3f9821936023071e95875582b45ccb319601f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T14:06:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0313 14:06:13.202871 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0313 14:06:13.203175 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0313 14:06:13.204427 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2471728650/tls.crt::/tmp/serving-cert-2471728650/tls.key\\\\\\\"\\\\nI0313 14:06:13.581160 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0313 14:06:13.584012 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0313 14:06:13.584036 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0313 14:06:13.584064 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0313 14:06:13.584072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0313 14:06:13.594504 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0313 14:06:13.594891 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0313 14:06:13.595323 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0313 14:06:13.595363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0313 14:06:13.595370 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0313 14:06:13.595376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0313 14:06:13.597226 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3f2b6756078925f8562b6053b3b83798efb72288f5fd18fff680a4ab0d33feb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:08:03Z is after 2025-08-24T17:21:41Z" Mar 13 14:08:03 crc kubenswrapper[4907]: I0313 14:08:03.901813 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94b045cf2fed56457a2ca14f489bb69470198f159f16c207fe061d034dba1555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd28f7a5c35822582b70ed41645294aaede10e28ff59f3f3b5d36e7da8237b76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:08:03Z is after 2025-08-24T17:21:41Z" Mar 13 14:08:03 crc kubenswrapper[4907]: I0313 14:08:03.916073 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l7zg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d105034b-f041-4fb5-8df6-6b1bb7681fda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21a207f917836785e320f4f79aca5d6449da2fdeec3ee0765f34d227d5617fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szmh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l7zg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:08:03Z is after 2025-08-24T17:21:41Z" Mar 13 14:08:03 crc kubenswrapper[4907]: I0313 14:08:03.929797 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qncw4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1d7ee39-09d7-4c0e-942b-5e9d8c9f63be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07df71887196cc7098ac9606a4a15b89d45219048c31280198ae75022bff9370\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:07:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bg45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://540ca207f31c9a8a9892f80dee153b3fb86f8de406f38360221512f424d8c884\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:07:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bg45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:07:00Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-qncw4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:08:03Z is after 2025-08-24T17:21:41Z" Mar 13 14:08:03 crc kubenswrapper[4907]: I0313 14:08:03.946141 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5b4819028f34db81cc43a04e4854c30488ed876c63b8daff11105bf61546c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:08:03Z is after 2025-08-24T17:21:41Z" Mar 13 14:08:03 crc kubenswrapper[4907]: I0313 14:08:03.963022 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:08:03Z is after 2025-08-24T17:21:41Z" Mar 13 14:08:03 crc kubenswrapper[4907]: I0313 14:08:03.985403 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xtwf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64ebd338-0158-44dc-90ab-e7c285e87762\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef632bb17c8f529b35d0882c8d4ba3dc213b0308b46a4ade8426b9c08e64bb2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://391b8bc1e975982b172175fcf4b4509756b018cbb99c12e48f4931ff472cdae2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T14:07:34Z\\\",\\\"message\\\":\\\"2026-03-13T14:06:49+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_baa16ed5-a467-4fd5-8846-0c869be5e531\\\\n2026-03-13T14:06:49+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_baa16ed5-a467-4fd5-8846-0c869be5e531 to /host/opt/cni/bin/\\\\n2026-03-13T14:06:49Z [verbose] multus-daemon started\\\\n2026-03-13T14:06:49Z [verbose] Readiness Indicator file check\\\\n2026-03-13T14:07:34Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:07:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6cwbq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xtwf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:08:03Z is after 2025-08-24T17:21:41Z" Mar 13 14:08:04 crc kubenswrapper[4907]: I0313 14:08:04.002443 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bed3bcea-e8c0-4b00-a985-e0c5b5bea100\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://402c1ef89085d840edb4313577f5b2fb967072cb3db00c34c98572f00699529c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f22becea3c68103ff0b2639c7ae8c051b76d19b407b0463645fc2e4a0682f23\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T14:05:32Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0313 14:05:03.905343 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0313 14:05:03.908030 1 observer_polling.go:159] Starting file observer\\\\nI0313 14:05:03.938230 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0313 14:05:03.942399 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0313 14:05:32.345768 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0313 14:05:32.345932 1 cmd.go:179] failed checking apiserver connectivity: Get \\\\\\\"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/openshift-kube-controller-manager/leases/cluster-policy-controller-lock\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:31Z is after 2026-02-23T05:33:13Z\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:03Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18eadc065bf4a307b7b443fcfe79b2b719938ee486b53321c0a9f20926a0a887\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9abfa5a9ad1a98a2907ead28390ee926e176646fe761b89f580101bd429cf747\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbc67477e0223db1e4b403be3709fe71fe6409570361e103eb730fdeaffe3e80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:08:03Z is after 2025-08-24T17:21:41Z" Mar 13 14:08:04 crc kubenswrapper[4907]: I0313 14:08:04.021655 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:08:04Z is after 2025-08-24T17:21:41Z" Mar 13 14:08:04 crc kubenswrapper[4907]: I0313 14:08:04.035233 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afb1d430459aafab2d6bc2507b4b4059529401e51fdf8e9848b0d7bb8651db68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:08:04Z is after 2025-08-24T17:21:41Z" Mar 13 14:08:04 crc kubenswrapper[4907]: I0313 14:08:04.051281 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-745zx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58e11265-2fd9-4b9b-926c-3343bb84ee53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ce7ad32fd2bf2332f3ee41080115b083c34326b7dc31c950098dfba15f58422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mnz7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-745zx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:08:04Z is after 2025-08-24T17:21:41Z" Mar 13 14:08:04 crc kubenswrapper[4907]: I0313 14:08:04.069579 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-tzmvc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f797b1bb-e769-49c4-84ce-93a6fad2df93\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tdzl9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tdzl9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:07:00Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-tzmvc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:08:04Z is after 2025-08-24T17:21:41Z" Mar 13 14:08:04 crc kubenswrapper[4907]: I0313 14:08:04.694197 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qxlx5_1006376d-4033-454e-8a23-2b67882bdd0b/ovnkube-controller/3.log" Mar 13 14:08:04 crc kubenswrapper[4907]: I0313 14:08:04.781962 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:08:04 crc kubenswrapper[4907]: I0313 14:08:04.782007 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:08:04 crc kubenswrapper[4907]: I0313 14:08:04.782047 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:08:04 crc kubenswrapper[4907]: I0313 14:08:04.782067 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:08:04 crc kubenswrapper[4907]: E0313 14:08:04.782155 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:08:04 crc kubenswrapper[4907]: E0313 14:08:04.782263 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:08:04 crc kubenswrapper[4907]: E0313 14:08:04.782469 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzmvc" podUID="f797b1bb-e769-49c4-84ce-93a6fad2df93" Mar 13 14:08:04 crc kubenswrapper[4907]: E0313 14:08:04.782528 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:08:04 crc kubenswrapper[4907]: I0313 14:08:04.835674 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f797b1bb-e769-49c4-84ce-93a6fad2df93-metrics-certs\") pod \"network-metrics-daemon-tzmvc\" (UID: \"f797b1bb-e769-49c4-84ce-93a6fad2df93\") " pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:08:04 crc kubenswrapper[4907]: E0313 14:08:04.835918 4907 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 13 14:08:04 crc kubenswrapper[4907]: E0313 14:08:04.836061 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f797b1bb-e769-49c4-84ce-93a6fad2df93-metrics-certs podName:f797b1bb-e769-49c4-84ce-93a6fad2df93 nodeName:}" failed. No retries permitted until 2026-03-13 14:09:08.836021758 +0000 UTC m=+247.735809497 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/f797b1bb-e769-49c4-84ce-93a6fad2df93-metrics-certs") pod "network-metrics-daemon-tzmvc" (UID: "f797b1bb-e769-49c4-84ce-93a6fad2df93") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 13 14:08:06 crc kubenswrapper[4907]: I0313 14:08:06.781811 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:08:06 crc kubenswrapper[4907]: I0313 14:08:06.781811 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:08:06 crc kubenswrapper[4907]: E0313 14:08:06.782072 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:08:06 crc kubenswrapper[4907]: I0313 14:08:06.781864 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:08:06 crc kubenswrapper[4907]: E0313 14:08:06.782165 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:08:06 crc kubenswrapper[4907]: I0313 14:08:06.781844 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:08:06 crc kubenswrapper[4907]: E0313 14:08:06.782272 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:08:06 crc kubenswrapper[4907]: E0313 14:08:06.782508 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzmvc" podUID="f797b1bb-e769-49c4-84ce-93a6fad2df93" Mar 13 14:08:06 crc kubenswrapper[4907]: E0313 14:08:06.995989 4907 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 13 14:08:08 crc kubenswrapper[4907]: I0313 14:08:08.782057 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:08:08 crc kubenswrapper[4907]: I0313 14:08:08.782073 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:08:08 crc kubenswrapper[4907]: E0313 14:08:08.782232 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:08:08 crc kubenswrapper[4907]: E0313 14:08:08.782337 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzmvc" podUID="f797b1bb-e769-49c4-84ce-93a6fad2df93" Mar 13 14:08:08 crc kubenswrapper[4907]: I0313 14:08:08.782079 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:08:08 crc kubenswrapper[4907]: E0313 14:08:08.782522 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:08:08 crc kubenswrapper[4907]: I0313 14:08:08.783137 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:08:08 crc kubenswrapper[4907]: E0313 14:08:08.783299 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:08:09 crc kubenswrapper[4907]: I0313 14:08:09.298097 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:08:09 crc kubenswrapper[4907]: I0313 14:08:09.298160 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:08:09 crc kubenswrapper[4907]: I0313 14:08:09.298172 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:08:09 crc kubenswrapper[4907]: I0313 14:08:09.298196 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:08:09 crc kubenswrapper[4907]: I0313 14:08:09.298209 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:08:09Z","lastTransitionTime":"2026-03-13T14:08:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:08:09 crc kubenswrapper[4907]: E0313 14:08:09.319100 4907 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:08:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:08:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:08:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:08:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:08:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:08:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:08:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:08:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4b182bae-7f2d-4b54-934c-85a9a277d132\\\",\\\"systemUUID\\\":\\\"05bfbfa3-f04f-45ba-8233-a355fc774c24\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:08:09Z is after 2025-08-24T17:21:41Z" Mar 13 14:08:09 crc kubenswrapper[4907]: I0313 14:08:09.325982 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:08:09 crc kubenswrapper[4907]: I0313 14:08:09.326049 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:08:09 crc kubenswrapper[4907]: I0313 14:08:09.326069 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:08:09 crc kubenswrapper[4907]: I0313 14:08:09.326097 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:08:09 crc kubenswrapper[4907]: I0313 14:08:09.326117 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:08:09Z","lastTransitionTime":"2026-03-13T14:08:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:08:09 crc kubenswrapper[4907]: E0313 14:08:09.350203 4907 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:08:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:08:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:08:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:08:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:08:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:08:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:08:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:08:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4b182bae-7f2d-4b54-934c-85a9a277d132\\\",\\\"systemUUID\\\":\\\"05bfbfa3-f04f-45ba-8233-a355fc774c24\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:08:09Z is after 2025-08-24T17:21:41Z" Mar 13 14:08:09 crc kubenswrapper[4907]: I0313 14:08:09.355576 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:08:09 crc kubenswrapper[4907]: I0313 14:08:09.355655 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:08:09 crc kubenswrapper[4907]: I0313 14:08:09.355675 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:08:09 crc kubenswrapper[4907]: I0313 14:08:09.355707 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:08:09 crc kubenswrapper[4907]: I0313 14:08:09.355727 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:08:09Z","lastTransitionTime":"2026-03-13T14:08:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:08:09 crc kubenswrapper[4907]: E0313 14:08:09.375858 4907 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:08:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:08:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:08:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:08:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:08:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:08:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:08:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:08:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4b182bae-7f2d-4b54-934c-85a9a277d132\\\",\\\"systemUUID\\\":\\\"05bfbfa3-f04f-45ba-8233-a355fc774c24\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:08:09Z is after 2025-08-24T17:21:41Z" Mar 13 14:08:09 crc kubenswrapper[4907]: I0313 14:08:09.381674 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:08:09 crc kubenswrapper[4907]: I0313 14:08:09.381935 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:08:09 crc kubenswrapper[4907]: I0313 14:08:09.382109 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:08:09 crc kubenswrapper[4907]: I0313 14:08:09.382280 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:08:09 crc kubenswrapper[4907]: I0313 14:08:09.382453 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:08:09Z","lastTransitionTime":"2026-03-13T14:08:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:08:09 crc kubenswrapper[4907]: E0313 14:08:09.406198 4907 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:08:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:08:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:08:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:08:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:08:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:08:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:08:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:08:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4b182bae-7f2d-4b54-934c-85a9a277d132\\\",\\\"systemUUID\\\":\\\"05bfbfa3-f04f-45ba-8233-a355fc774c24\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:08:09Z is after 2025-08-24T17:21:41Z" Mar 13 14:08:09 crc kubenswrapper[4907]: I0313 14:08:09.412399 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:08:09 crc kubenswrapper[4907]: I0313 14:08:09.412503 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:08:09 crc kubenswrapper[4907]: I0313 14:08:09.412523 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:08:09 crc kubenswrapper[4907]: I0313 14:08:09.412558 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:08:09 crc kubenswrapper[4907]: I0313 14:08:09.412579 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:08:09Z","lastTransitionTime":"2026-03-13T14:08:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:08:09 crc kubenswrapper[4907]: E0313 14:08:09.435342 4907 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:08:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:08:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:08:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:08:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:08:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:08:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:08:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:08:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4b182bae-7f2d-4b54-934c-85a9a277d132\\\",\\\"systemUUID\\\":\\\"05bfbfa3-f04f-45ba-8233-a355fc774c24\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:08:09Z is after 2025-08-24T17:21:41Z" Mar 13 14:08:09 crc kubenswrapper[4907]: E0313 14:08:09.435570 4907 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 13 14:08:10 crc kubenswrapper[4907]: I0313 14:08:10.781550 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:08:10 crc kubenswrapper[4907]: I0313 14:08:10.781629 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:08:10 crc kubenswrapper[4907]: I0313 14:08:10.781731 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:08:10 crc kubenswrapper[4907]: I0313 14:08:10.781731 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:08:10 crc kubenswrapper[4907]: E0313 14:08:10.781986 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:08:10 crc kubenswrapper[4907]: E0313 14:08:10.782353 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:08:10 crc kubenswrapper[4907]: E0313 14:08:10.782441 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzmvc" podUID="f797b1bb-e769-49c4-84ce-93a6fad2df93" Mar 13 14:08:10 crc kubenswrapper[4907]: E0313 14:08:10.782611 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:08:11 crc kubenswrapper[4907]: I0313 14:08:11.803119 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5b4819028f34db81cc43a04e4854c30488ed876c63b8daff11105bf61546c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:08:11Z is after 2025-08-24T17:21:41Z" Mar 13 14:08:11 crc kubenswrapper[4907]: I0313 14:08:11.822643 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:08:11Z is after 2025-08-24T17:21:41Z" Mar 13 14:08:11 crc kubenswrapper[4907]: I0313 14:08:11.843527 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xtwf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64ebd338-0158-44dc-90ab-e7c285e87762\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef632bb17c8f529b35d0882c8d4ba3dc213b0308b46a4ade8426b9c08e64bb2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://391b8bc1e975982b172175fcf4b4509756b018cbb99c12e48f4931ff472cdae2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T14:07:34Z\\\",\\\"message\\\":\\\"2026-03-13T14:06:49+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_baa16ed5-a467-4fd5-8846-0c869be5e531\\\\n2026-03-13T14:06:49+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_baa16ed5-a467-4fd5-8846-0c869be5e531 to /host/opt/cni/bin/\\\\n2026-03-13T14:06:49Z [verbose] multus-daemon started\\\\n2026-03-13T14:06:49Z [verbose] Readiness Indicator file check\\\\n2026-03-13T14:07:34Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:07:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6cwbq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xtwf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:08:11Z is after 2025-08-24T17:21:41Z" Mar 13 14:08:11 crc kubenswrapper[4907]: I0313 14:08:11.864495 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bed3bcea-e8c0-4b00-a985-e0c5b5bea100\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://402c1ef89085d840edb4313577f5b2fb967072cb3db00c34c98572f00699529c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f22becea3c68103ff0b2639c7ae8c051b76d19b407b0463645fc2e4a0682f23\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T14:05:32Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0313 14:05:03.905343 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0313 14:05:03.908030 1 observer_polling.go:159] Starting file observer\\\\nI0313 14:05:03.938230 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0313 14:05:03.942399 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0313 14:05:32.345768 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0313 14:05:32.345932 1 cmd.go:179] failed checking apiserver connectivity: Get \\\\\\\"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/openshift-kube-controller-manager/leases/cluster-policy-controller-lock\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:31Z is after 2026-02-23T05:33:13Z\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:03Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18eadc065bf4a307b7b443fcfe79b2b719938ee486b53321c0a9f20926a0a887\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9abfa5a9ad1a98a2907ead28390ee926e176646fe761b89f580101bd429cf747\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbc67477e0223db1e4b403be3709fe71fe6409570361e103eb730fdeaffe3e80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:08:11Z is after 2025-08-24T17:21:41Z" Mar 13 14:08:11 crc kubenswrapper[4907]: I0313 14:08:11.882914 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:08:11Z is after 2025-08-24T17:21:41Z" Mar 13 14:08:11 crc kubenswrapper[4907]: I0313 14:08:11.900545 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afb1d430459aafab2d6bc2507b4b4059529401e51fdf8e9848b0d7bb8651db68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:08:11Z is after 2025-08-24T17:21:41Z" Mar 13 14:08:11 crc kubenswrapper[4907]: I0313 14:08:11.918515 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-745zx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58e11265-2fd9-4b9b-926c-3343bb84ee53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ce7ad32fd2bf2332f3ee41080115b083c34326b7dc31c950098dfba15f58422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mnz7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-745zx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:08:11Z is after 2025-08-24T17:21:41Z" Mar 13 14:08:11 crc kubenswrapper[4907]: I0313 14:08:11.935581 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-tzmvc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f797b1bb-e769-49c4-84ce-93a6fad2df93\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tdzl9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tdzl9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:07:00Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-tzmvc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:08:11Z is after 2025-08-24T17:21:41Z" Mar 13 14:08:11 crc kubenswrapper[4907]: I0313 14:08:11.960082 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1006376d-4033-454e-8a23-2b67882bdd0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf6a3dbaa8dda61f461a906538824c8946d36dc07d4fb95bf4d8420ad0105600\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccef92e41501fab91189581d7e8875515c02d637426b329d2fb80a356c75b637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0ee0049aa4a6d49a6bcf7d347183d9cc8b85171f988662803a005e0cdd97395\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fff87d8ea572615eb1329164f9e536e6ee45cd00a9e71be0aa8a9b3ed2190e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c165053fd9933444b372ff40efe33c0e193a7f39e6f531a4f370e971eb41e6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fea41c11bebd2d9d622ed4b8bc50e03a106dadea6ac102cd60512ea2a1f8b8b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9155434e2d8f0daa4504758fb72d6f498da107a64a92a0883ac687ed2af9caee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a99cabe30a7abc8da854bf86d79628b94a6ffc20c28f828b695080332b319db2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T14:07:15Z\\\",\\\"message\\\":\\\"Sending *v1.Node event handler 7 for removal\\\\nI0313 14:07:15.619307 7162 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0313 14:07:15.619320 7162 handler.go:208] Removed *v1.Node event handler 2\\\\nI0313 14:07:15.619329 7162 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0313 14:07:15.619343 7162 handler.go:208] Removed *v1.Node event handler 7\\\\nI0313 14:07:15.619351 7162 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0313 14:07:15.619358 7162 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0313 14:07:15.619362 7162 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0313 14:07:15.619381 7162 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0313 14:07:15.619382 7162 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0313 14:07:15.619392 7162 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0313 14:07:15.619619 7162 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI0313 14:07:15.619713 7162 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI0313 14:07:15.619975 7162 factory.go:656] Stopping watch factory\\\\nI0313 14:07:15.620001 7162 ovnkube.go:599] Stopped ovnkube\\\\nI0313 14:07:15.620033 7162 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0313 14:07:15.620096 7162 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:07:14Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9155434e2d8f0daa4504758fb72d6f498da107a64a92a0883ac687ed2af9caee\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T14:08:02Z\\\",\\\"message\\\":\\\"dn-host-isolation-manager: error found while processing openshift-dns-operator/dns-operator-744455d44c-xkdjh: failed to check if pod openshift-dns-operator/dns-operator-744455d44c-xkdjh is in primary UDN: could not find OVN pod annotation in map[]\\\\nI0313 14:08:02.544936 7501 controller.go:257] Controller udn-host-isolation-manager: error found while processing openshift-kube-storage-version-migrator/migrator-59844c95c7-vfbsx: failed to check if pod openshift-kube-storage-version-migrator/migrator-59844c95c7-vfbsx is in primary UDN: could not find OVN pod annotation in map[openshift.io/required-scc:nonroot-v2 openshift.io/scc:nonroot-v2 seccomp.security.alpha.kubernetes.io/pod:runtime/default]\\\\nI0313 14:08:02.544957 7501 controller.go:257] Controller udn-host-isolation-manager: error found while processing openshift-multus/network-metrics-daemon-tzmvc: failed to check if pod openshift-multus/network-metrics-daemon-tzmvc is in primary UDN: could not find OVN pod annotation in map[cluster-autoscaler.kubernetes.io/enable-ds-eviction:false]\\\\nE0313 14:08:02.629503 7501 shared_informer.go:316] \\\\\\\"Unhandled Error\\\\\\\" err=\\\\\\\"unable to sync caches for ovn-lb-controller\\\\\\\" logger=\\\\\\\"UnhandledError\\\\\\\"\\\\nI0313 14:08:02.630816 7501 ovnkube.go:599] Stopped ovnkube\\\\nI0313 14:08:02.630949 7501 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:07:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d60b5b774f8dd4cee16b9784f80d124640199a85331a2940772029be4884751\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qxlx5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:08:11Z is after 2025-08-24T17:21:41Z" Mar 13 14:08:11 crc kubenswrapper[4907]: I0313 14:08:11.976027 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a21d7e2-0481-4ea4-b129-0e2754315289\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35fd0db86ae409203bc57f755e489cd542ea68d19b5fcb9023794a77bcfa1a64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22af7218ec732d25874992d8015afd1d73b48cb6cf7d74e638c9db396cb9efe3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22af7218ec732d25874992d8015afd1d73b48cb6cf7d74e638c9db396cb9efe3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:08:11Z is after 2025-08-24T17:21:41Z" Mar 13 14:08:11 crc kubenswrapper[4907]: I0313 14:08:11.987954 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"584d221c-be74-4a39-8173-6c8e3f55f686\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b782a795257057929fe616938d1483a9accbd7a93f0cba963f9791a3d4ec4749\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c826e7b9be5e3a0ecdaa6dfeaf2454535e542ff86acc532cc3c80acf4a95c1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4333bdf047392cb3f64e29e677fe8d2f21c7ebca1a7e7097167e70c6785b8d46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a4dea84b35a82ce16f0e0d38057dc994b26240cf45dcc9111ed0de573a62881\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a4dea84b35a82ce16f0e0d38057dc994b26240cf45dcc9111ed0de573a62881\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:08:11Z is after 2025-08-24T17:21:41Z" Mar 13 14:08:11 crc kubenswrapper[4907]: E0313 14:08:11.997459 4907 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 13 14:08:12 crc kubenswrapper[4907]: I0313 14:08:12.000062 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:08:11Z is after 2025-08-24T17:21:41Z" Mar 13 14:08:12 crc kubenswrapper[4907]: I0313 14:08:12.016582 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-p8ptc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57e4e912-9a98-4419-87ce-f16e178ed921\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38f113418d62dab1728f0ba57cd1d289e3ec41621a338017e74c9996a8449b1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bdef68f401b5ba2c7cf61551eba416e74f295ab2281433a5115334d28094818\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bdef68f401b5ba2c7cf61551eba416e74f295ab2281433a5115334d28094818\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f8d1f40744d609fd964572ef5e323f0283d2582e98745602754165818f17847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f8d1f40744d609fd964572ef5e323f0283d2582e98745602754165818f17847\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c1ab2e148eaa177b98f5883a3f323d12f5b662bff8cb2d072a97be18ab8e120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c1ab2e148eaa177b98f5883a3f323d12f5b662bff8cb2d072a97be18ab8e120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-p8ptc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:08:12Z is after 2025-08-24T17:21:41Z" Mar 13 14:08:12 crc kubenswrapper[4907]: I0313 14:08:12.027402 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3fb4213-144f-4b01-9d78-ff3315761bb3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba1140a2462713c2401b47a4cf1f231bc9d9ae7867a7d0b1b597214a1f7efe07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://770c882d9e9f9b2e9bd4f135e591952fa15b3898c42826aed8a75d953c9a875b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hm56j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:08:12Z is after 2025-08-24T17:21:41Z" Mar 13 14:08:12 crc kubenswrapper[4907]: I0313 14:08:12.050426 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31a594bd-13e7-4c68-ac7c-f481626088ec\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8141bbe3b2b50ad7d2aed26f813a951bc6db4c90bf8271cd23a125cc7ec2eb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb841f47d97b3e9495658ff7f6aa3f835905a9cd2ee9736774c3504d6c59a209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f04995f26a4c320fa30f2a97b1299284cbb2e23965bae16a6cca8d06eef630eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0568aef54a902fedcaaa71cabe0add0cbeb3a38743a7fe45584c5e315292afa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://171ea306a9ec518a88ef509adaab1a7c238abf7b5824a87d3425de7199cde520\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14a9ee19b96b2019de6ad9e5092e0004a4587a9f678fec417c4ac02e60695ac8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14a9ee19b96b2019de6ad9e5092e0004a4587a9f678fec417c4ac02e60695ac8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d46350802ae58ed4c464dc3f8d3f9dd7e091ef8091659e0711b9d2632e00708\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d46350802ae58ed4c464dc3f8d3f9dd7e091ef8091659e0711b9d2632e00708\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://65bd2db6ca4ffb26d9cb4bce44dc27a862353533bb0ec584bcc688b73fbec3eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65bd2db6ca4ffb26d9cb4bce44dc27a862353533bb0ec584bcc688b73fbec3eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:08:12Z is after 2025-08-24T17:21:41Z" Mar 13 14:08:12 crc kubenswrapper[4907]: I0313 14:08:12.064794 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5d2b054-0ed8-419c-b731-f14d3598372f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4b503aa3b13268b3688d067f3e456c86afac1fda171aced0990899c3eb89b79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be40a1a33fa482eab363537afaa901127d560afbc1b5b287ef11e66391a5be4c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d1b539bbd918fc50428fc18cabcc4c8a43cc96f220b5692d67cf1c8dfbf8ed8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93d8c8c79be9d9ffb72de66fc2f3f9821936023071e95875582b45ccb319601f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T14:06:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0313 14:06:13.202871 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0313 14:06:13.203175 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0313 14:06:13.204427 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2471728650/tls.crt::/tmp/serving-cert-2471728650/tls.key\\\\\\\"\\\\nI0313 14:06:13.581160 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0313 14:06:13.584012 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0313 14:06:13.584036 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0313 14:06:13.584064 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0313 14:06:13.584072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0313 14:06:13.594504 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0313 14:06:13.594891 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0313 14:06:13.595323 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0313 14:06:13.595363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0313 14:06:13.595370 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0313 14:06:13.595376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0313 14:06:13.597226 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3f2b6756078925f8562b6053b3b83798efb72288f5fd18fff680a4ab0d33feb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:08:12Z is after 2025-08-24T17:21:41Z" Mar 13 14:08:12 crc kubenswrapper[4907]: I0313 14:08:12.078065 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94b045cf2fed56457a2ca14f489bb69470198f159f16c207fe061d034dba1555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd28f7a5c35822582b70ed41645294aaede10e28ff59f3f3b5d36e7da8237b76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:08:12Z is after 2025-08-24T17:21:41Z" Mar 13 14:08:12 crc kubenswrapper[4907]: I0313 14:08:12.089106 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l7zg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d105034b-f041-4fb5-8df6-6b1bb7681fda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21a207f917836785e320f4f79aca5d6449da2fdeec3ee0765f34d227d5617fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szmh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l7zg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:08:12Z is after 2025-08-24T17:21:41Z" Mar 13 14:08:12 crc kubenswrapper[4907]: I0313 14:08:12.103310 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qncw4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1d7ee39-09d7-4c0e-942b-5e9d8c9f63be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07df71887196cc7098ac9606a4a15b89d45219048c31280198ae75022bff9370\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:07:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bg45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://540ca207f31c9a8a9892f80dee153b3fb86f8de406f38360221512f424d8c884\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:07:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bg45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:07:00Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-qncw4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:08:12Z is after 2025-08-24T17:21:41Z" Mar 13 14:08:12 crc kubenswrapper[4907]: I0313 14:08:12.781860 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:08:12 crc kubenswrapper[4907]: I0313 14:08:12.781942 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:08:12 crc kubenswrapper[4907]: I0313 14:08:12.782069 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:08:12 crc kubenswrapper[4907]: I0313 14:08:12.782070 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:08:12 crc kubenswrapper[4907]: E0313 14:08:12.782256 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:08:12 crc kubenswrapper[4907]: E0313 14:08:12.782727 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzmvc" podUID="f797b1bb-e769-49c4-84ce-93a6fad2df93" Mar 13 14:08:12 crc kubenswrapper[4907]: E0313 14:08:12.782859 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:08:12 crc kubenswrapper[4907]: E0313 14:08:12.783066 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:08:14 crc kubenswrapper[4907]: I0313 14:08:14.782214 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:08:14 crc kubenswrapper[4907]: I0313 14:08:14.782247 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:08:14 crc kubenswrapper[4907]: I0313 14:08:14.782443 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:08:14 crc kubenswrapper[4907]: I0313 14:08:14.782479 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:08:14 crc kubenswrapper[4907]: E0313 14:08:14.783171 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:08:14 crc kubenswrapper[4907]: E0313 14:08:14.783207 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:08:14 crc kubenswrapper[4907]: E0313 14:08:14.783359 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzmvc" podUID="f797b1bb-e769-49c4-84ce-93a6fad2df93" Mar 13 14:08:14 crc kubenswrapper[4907]: E0313 14:08:14.783445 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:08:16 crc kubenswrapper[4907]: I0313 14:08:16.798789 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:08:16 crc kubenswrapper[4907]: E0313 14:08:16.799002 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:08:16 crc kubenswrapper[4907]: I0313 14:08:16.799283 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:08:16 crc kubenswrapper[4907]: E0313 14:08:16.799373 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:08:16 crc kubenswrapper[4907]: I0313 14:08:16.799434 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:08:16 crc kubenswrapper[4907]: E0313 14:08:16.799625 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:08:16 crc kubenswrapper[4907]: I0313 14:08:16.799687 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:08:16 crc kubenswrapper[4907]: E0313 14:08:16.799806 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzmvc" podUID="f797b1bb-e769-49c4-84ce-93a6fad2df93" Mar 13 14:08:16 crc kubenswrapper[4907]: E0313 14:08:16.998605 4907 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 13 14:08:18 crc kubenswrapper[4907]: I0313 14:08:18.781814 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:08:18 crc kubenswrapper[4907]: I0313 14:08:18.781847 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:08:18 crc kubenswrapper[4907]: E0313 14:08:18.781939 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:08:18 crc kubenswrapper[4907]: I0313 14:08:18.781974 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:08:18 crc kubenswrapper[4907]: E0313 14:08:18.782022 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzmvc" podUID="f797b1bb-e769-49c4-84ce-93a6fad2df93" Mar 13 14:08:18 crc kubenswrapper[4907]: I0313 14:08:18.782125 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:08:18 crc kubenswrapper[4907]: E0313 14:08:18.782302 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:08:18 crc kubenswrapper[4907]: E0313 14:08:18.782686 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:08:18 crc kubenswrapper[4907]: I0313 14:08:18.783108 4907 scope.go:117] "RemoveContainer" containerID="9155434e2d8f0daa4504758fb72d6f498da107a64a92a0883ac687ed2af9caee" Mar 13 14:08:18 crc kubenswrapper[4907]: E0313 14:08:18.783317 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-qxlx5_openshift-ovn-kubernetes(1006376d-4033-454e-8a23-2b67882bdd0b)\"" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" podUID="1006376d-4033-454e-8a23-2b67882bdd0b" Mar 13 14:08:18 crc kubenswrapper[4907]: I0313 14:08:18.816431 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=49.816414781 podStartE2EDuration="49.816414781s" podCreationTimestamp="2026-03-13 14:07:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:08:18.816289218 +0000 UTC m=+197.716076977" watchObservedRunningTime="2026-03-13 14:08:18.816414781 +0000 UTC m=+197.716202490" Mar 13 14:08:18 crc kubenswrapper[4907]: I0313 14:08:18.872142 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-745zx" podStartSLOduration=138.872107526 podStartE2EDuration="2m18.872107526s" podCreationTimestamp="2026-03-13 14:06:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:08:18.871831198 +0000 UTC m=+197.771618907" watchObservedRunningTime="2026-03-13 14:08:18.872107526 +0000 UTC m=+197.771895255" Mar 13 14:08:18 crc kubenswrapper[4907]: I0313 14:08:18.900807 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=35.900781476 podStartE2EDuration="35.900781476s" podCreationTimestamp="2026-03-13 14:07:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:08:18.899801189 +0000 UTC m=+197.799588918" watchObservedRunningTime="2026-03-13 14:08:18.900781476 +0000 UTC m=+197.800569175" Mar 13 14:08:18 crc kubenswrapper[4907]: I0313 14:08:18.917288 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=68.917269243 podStartE2EDuration="1m8.917269243s" podCreationTimestamp="2026-03-13 14:07:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:08:18.912419582 +0000 UTC m=+197.812207301" watchObservedRunningTime="2026-03-13 14:08:18.917269243 +0000 UTC m=+197.817056932" Mar 13 14:08:18 crc kubenswrapper[4907]: I0313 14:08:18.949958 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-p8ptc" podStartSLOduration=137.949931102 podStartE2EDuration="2m17.949931102s" podCreationTimestamp="2026-03-13 14:06:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:08:18.949808569 +0000 UTC m=+197.849596278" watchObservedRunningTime="2026-03-13 14:08:18.949931102 +0000 UTC m=+197.849718831" Mar 13 14:08:18 crc kubenswrapper[4907]: I0313 14:08:18.962758 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podStartSLOduration=137.96273609 podStartE2EDuration="2m17.96273609s" podCreationTimestamp="2026-03-13 14:06:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:08:18.962645018 +0000 UTC m=+197.862432727" watchObservedRunningTime="2026-03-13 14:08:18.96273609 +0000 UTC m=+197.862523779" Mar 13 14:08:19 crc kubenswrapper[4907]: I0313 14:08:19.019283 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=41.019261126 podStartE2EDuration="41.019261126s" podCreationTimestamp="2026-03-13 14:07:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:08:19.016848 +0000 UTC m=+197.916635709" watchObservedRunningTime="2026-03-13 14:08:19.019261126 +0000 UTC m=+197.919048835" Mar 13 14:08:19 crc kubenswrapper[4907]: I0313 14:08:19.038191 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=99.038163941 podStartE2EDuration="1m39.038163941s" podCreationTimestamp="2026-03-13 14:06:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:08:19.037753589 +0000 UTC m=+197.937541278" watchObservedRunningTime="2026-03-13 14:08:19.038163941 +0000 UTC m=+197.937951650" Mar 13 14:08:19 crc kubenswrapper[4907]: I0313 14:08:19.070574 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-l7zg7" podStartSLOduration=139.070554201 podStartE2EDuration="2m19.070554201s" podCreationTimestamp="2026-03-13 14:06:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:08:19.070445738 +0000 UTC m=+197.970233447" watchObservedRunningTime="2026-03-13 14:08:19.070554201 +0000 UTC m=+197.970341880" Mar 13 14:08:19 crc kubenswrapper[4907]: I0313 14:08:19.149243 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qncw4" podStartSLOduration=138.14922719 podStartE2EDuration="2m18.14922719s" podCreationTimestamp="2026-03-13 14:06:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:08:19.086528425 +0000 UTC m=+197.986316124" watchObservedRunningTime="2026-03-13 14:08:19.14922719 +0000 UTC m=+198.049014879" Mar 13 14:08:19 crc kubenswrapper[4907]: I0313 14:08:19.192027 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-xtwf7" podStartSLOduration=138.192000823 podStartE2EDuration="2m18.192000823s" podCreationTimestamp="2026-03-13 14:06:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:08:19.1907857 +0000 UTC m=+198.090573409" watchObservedRunningTime="2026-03-13 14:08:19.192000823 +0000 UTC m=+198.091788542" Mar 13 14:08:19 crc kubenswrapper[4907]: I0313 14:08:19.585135 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:08:19 crc kubenswrapper[4907]: I0313 14:08:19.585195 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:08:19 crc kubenswrapper[4907]: I0313 14:08:19.585211 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:08:19 crc kubenswrapper[4907]: I0313 14:08:19.585235 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:08:19 crc kubenswrapper[4907]: I0313 14:08:19.585251 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:08:19Z","lastTransitionTime":"2026-03-13T14:08:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:08:19 crc kubenswrapper[4907]: I0313 14:08:19.650939 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-h7pgw"] Mar 13 14:08:19 crc kubenswrapper[4907]: I0313 14:08:19.652078 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-h7pgw" Mar 13 14:08:19 crc kubenswrapper[4907]: I0313 14:08:19.655643 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Mar 13 14:08:19 crc kubenswrapper[4907]: I0313 14:08:19.656053 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Mar 13 14:08:19 crc kubenswrapper[4907]: I0313 14:08:19.655720 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Mar 13 14:08:19 crc kubenswrapper[4907]: I0313 14:08:19.655844 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Mar 13 14:08:19 crc kubenswrapper[4907]: I0313 14:08:19.696005 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2d0f4392-a2c3-4ca7-9b92-b7ea2c01579b-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-h7pgw\" (UID: \"2d0f4392-a2c3-4ca7-9b92-b7ea2c01579b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-h7pgw" Mar 13 14:08:19 crc kubenswrapper[4907]: I0313 14:08:19.696121 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/2d0f4392-a2c3-4ca7-9b92-b7ea2c01579b-service-ca\") pod \"cluster-version-operator-5c965bbfc6-h7pgw\" (UID: \"2d0f4392-a2c3-4ca7-9b92-b7ea2c01579b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-h7pgw" Mar 13 14:08:19 crc kubenswrapper[4907]: I0313 14:08:19.696236 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/2d0f4392-a2c3-4ca7-9b92-b7ea2c01579b-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-h7pgw\" (UID: \"2d0f4392-a2c3-4ca7-9b92-b7ea2c01579b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-h7pgw" Mar 13 14:08:19 crc kubenswrapper[4907]: I0313 14:08:19.696278 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2d0f4392-a2c3-4ca7-9b92-b7ea2c01579b-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-h7pgw\" (UID: \"2d0f4392-a2c3-4ca7-9b92-b7ea2c01579b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-h7pgw" Mar 13 14:08:19 crc kubenswrapper[4907]: I0313 14:08:19.696326 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/2d0f4392-a2c3-4ca7-9b92-b7ea2c01579b-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-h7pgw\" (UID: \"2d0f4392-a2c3-4ca7-9b92-b7ea2c01579b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-h7pgw" Mar 13 14:08:19 crc kubenswrapper[4907]: I0313 14:08:19.796940 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/2d0f4392-a2c3-4ca7-9b92-b7ea2c01579b-service-ca\") pod \"cluster-version-operator-5c965bbfc6-h7pgw\" (UID: \"2d0f4392-a2c3-4ca7-9b92-b7ea2c01579b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-h7pgw" Mar 13 14:08:19 crc kubenswrapper[4907]: I0313 14:08:19.797043 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/2d0f4392-a2c3-4ca7-9b92-b7ea2c01579b-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-h7pgw\" (UID: \"2d0f4392-a2c3-4ca7-9b92-b7ea2c01579b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-h7pgw" Mar 13 14:08:19 crc kubenswrapper[4907]: I0313 14:08:19.797073 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2d0f4392-a2c3-4ca7-9b92-b7ea2c01579b-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-h7pgw\" (UID: \"2d0f4392-a2c3-4ca7-9b92-b7ea2c01579b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-h7pgw" Mar 13 14:08:19 crc kubenswrapper[4907]: I0313 14:08:19.797111 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/2d0f4392-a2c3-4ca7-9b92-b7ea2c01579b-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-h7pgw\" (UID: \"2d0f4392-a2c3-4ca7-9b92-b7ea2c01579b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-h7pgw" Mar 13 14:08:19 crc kubenswrapper[4907]: I0313 14:08:19.797147 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2d0f4392-a2c3-4ca7-9b92-b7ea2c01579b-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-h7pgw\" (UID: \"2d0f4392-a2c3-4ca7-9b92-b7ea2c01579b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-h7pgw" Mar 13 14:08:19 crc kubenswrapper[4907]: I0313 14:08:19.798057 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/2d0f4392-a2c3-4ca7-9b92-b7ea2c01579b-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-h7pgw\" (UID: \"2d0f4392-a2c3-4ca7-9b92-b7ea2c01579b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-h7pgw" Mar 13 14:08:19 crc kubenswrapper[4907]: I0313 14:08:19.798097 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/2d0f4392-a2c3-4ca7-9b92-b7ea2c01579b-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-h7pgw\" (UID: \"2d0f4392-a2c3-4ca7-9b92-b7ea2c01579b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-h7pgw" Mar 13 14:08:19 crc kubenswrapper[4907]: I0313 14:08:19.798785 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/2d0f4392-a2c3-4ca7-9b92-b7ea2c01579b-service-ca\") pod \"cluster-version-operator-5c965bbfc6-h7pgw\" (UID: \"2d0f4392-a2c3-4ca7-9b92-b7ea2c01579b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-h7pgw" Mar 13 14:08:19 crc kubenswrapper[4907]: I0313 14:08:19.805746 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2d0f4392-a2c3-4ca7-9b92-b7ea2c01579b-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-h7pgw\" (UID: \"2d0f4392-a2c3-4ca7-9b92-b7ea2c01579b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-h7pgw" Mar 13 14:08:19 crc kubenswrapper[4907]: I0313 14:08:19.819517 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2d0f4392-a2c3-4ca7-9b92-b7ea2c01579b-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-h7pgw\" (UID: \"2d0f4392-a2c3-4ca7-9b92-b7ea2c01579b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-h7pgw" Mar 13 14:08:19 crc kubenswrapper[4907]: I0313 14:08:19.871477 4907 certificate_manager.go:356] kubernetes.io/kubelet-serving: Rotating certificates Mar 13 14:08:19 crc kubenswrapper[4907]: I0313 14:08:19.882676 4907 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Mar 13 14:08:19 crc kubenswrapper[4907]: I0313 14:08:19.977038 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-h7pgw" Mar 13 14:08:20 crc kubenswrapper[4907]: I0313 14:08:20.767155 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-h7pgw" event={"ID":"2d0f4392-a2c3-4ca7-9b92-b7ea2c01579b","Type":"ContainerStarted","Data":"4ed47fad7c3de74b8b9b0a421e06356ab4fa2cf4bcb3e89c6c7d4963e5c44def"} Mar 13 14:08:20 crc kubenswrapper[4907]: I0313 14:08:20.767658 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-h7pgw" event={"ID":"2d0f4392-a2c3-4ca7-9b92-b7ea2c01579b","Type":"ContainerStarted","Data":"dad1aa0752bc06328372e05af2c284010393fa1cd05fd1b800414d163c3d36b2"} Mar 13 14:08:20 crc kubenswrapper[4907]: I0313 14:08:20.781419 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:08:20 crc kubenswrapper[4907]: I0313 14:08:20.781455 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:08:20 crc kubenswrapper[4907]: E0313 14:08:20.781917 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:08:20 crc kubenswrapper[4907]: I0313 14:08:20.781553 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:08:20 crc kubenswrapper[4907]: I0313 14:08:20.781497 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:08:20 crc kubenswrapper[4907]: E0313 14:08:20.782308 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:08:20 crc kubenswrapper[4907]: E0313 14:08:20.782026 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzmvc" podUID="f797b1bb-e769-49c4-84ce-93a6fad2df93" Mar 13 14:08:20 crc kubenswrapper[4907]: E0313 14:08:20.782523 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:08:21 crc kubenswrapper[4907]: I0313 14:08:21.773112 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-xtwf7_64ebd338-0158-44dc-90ab-e7c285e87762/kube-multus/1.log" Mar 13 14:08:21 crc kubenswrapper[4907]: I0313 14:08:21.773595 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-xtwf7_64ebd338-0158-44dc-90ab-e7c285e87762/kube-multus/0.log" Mar 13 14:08:21 crc kubenswrapper[4907]: I0313 14:08:21.773623 4907 generic.go:334] "Generic (PLEG): container finished" podID="64ebd338-0158-44dc-90ab-e7c285e87762" containerID="ef632bb17c8f529b35d0882c8d4ba3dc213b0308b46a4ade8426b9c08e64bb2b" exitCode=1 Mar 13 14:08:21 crc kubenswrapper[4907]: I0313 14:08:21.773653 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-xtwf7" event={"ID":"64ebd338-0158-44dc-90ab-e7c285e87762","Type":"ContainerDied","Data":"ef632bb17c8f529b35d0882c8d4ba3dc213b0308b46a4ade8426b9c08e64bb2b"} Mar 13 14:08:21 crc kubenswrapper[4907]: I0313 14:08:21.773687 4907 scope.go:117] "RemoveContainer" containerID="391b8bc1e975982b172175fcf4b4509756b018cbb99c12e48f4931ff472cdae2" Mar 13 14:08:21 crc kubenswrapper[4907]: I0313 14:08:21.774086 4907 scope.go:117] "RemoveContainer" containerID="ef632bb17c8f529b35d0882c8d4ba3dc213b0308b46a4ade8426b9c08e64bb2b" Mar 13 14:08:21 crc kubenswrapper[4907]: E0313 14:08:21.774215 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-xtwf7_openshift-multus(64ebd338-0158-44dc-90ab-e7c285e87762)\"" pod="openshift-multus/multus-xtwf7" podUID="64ebd338-0158-44dc-90ab-e7c285e87762" Mar 13 14:08:21 crc kubenswrapper[4907]: I0313 14:08:21.811260 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-h7pgw" podStartSLOduration=141.811239521 podStartE2EDuration="2m21.811239521s" podCreationTimestamp="2026-03-13 14:06:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:08:20.78679233 +0000 UTC m=+199.686580029" watchObservedRunningTime="2026-03-13 14:08:21.811239521 +0000 UTC m=+200.711027220" Mar 13 14:08:22 crc kubenswrapper[4907]: E0313 14:08:22.000610 4907 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 13 14:08:22 crc kubenswrapper[4907]: I0313 14:08:22.780946 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-xtwf7_64ebd338-0158-44dc-90ab-e7c285e87762/kube-multus/1.log" Mar 13 14:08:22 crc kubenswrapper[4907]: I0313 14:08:22.782113 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:08:22 crc kubenswrapper[4907]: I0313 14:08:22.782193 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:08:22 crc kubenswrapper[4907]: I0313 14:08:22.782194 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:08:22 crc kubenswrapper[4907]: I0313 14:08:22.782327 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:08:22 crc kubenswrapper[4907]: E0313 14:08:22.782514 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzmvc" podUID="f797b1bb-e769-49c4-84ce-93a6fad2df93" Mar 13 14:08:22 crc kubenswrapper[4907]: E0313 14:08:22.782697 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:08:22 crc kubenswrapper[4907]: E0313 14:08:22.782854 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:08:22 crc kubenswrapper[4907]: E0313 14:08:22.783318 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:08:24 crc kubenswrapper[4907]: I0313 14:08:24.782242 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:08:24 crc kubenswrapper[4907]: I0313 14:08:24.782355 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:08:24 crc kubenswrapper[4907]: E0313 14:08:24.782392 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:08:24 crc kubenswrapper[4907]: I0313 14:08:24.782451 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:08:24 crc kubenswrapper[4907]: E0313 14:08:24.782570 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzmvc" podUID="f797b1bb-e769-49c4-84ce-93a6fad2df93" Mar 13 14:08:24 crc kubenswrapper[4907]: I0313 14:08:24.782601 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:08:24 crc kubenswrapper[4907]: E0313 14:08:24.782657 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:08:24 crc kubenswrapper[4907]: E0313 14:08:24.782716 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:08:26 crc kubenswrapper[4907]: I0313 14:08:26.781744 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:08:26 crc kubenswrapper[4907]: E0313 14:08:26.783303 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzmvc" podUID="f797b1bb-e769-49c4-84ce-93a6fad2df93" Mar 13 14:08:26 crc kubenswrapper[4907]: I0313 14:08:26.782435 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:08:26 crc kubenswrapper[4907]: I0313 14:08:26.782384 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:08:26 crc kubenswrapper[4907]: E0313 14:08:26.783501 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:08:26 crc kubenswrapper[4907]: I0313 14:08:26.782450 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:08:26 crc kubenswrapper[4907]: E0313 14:08:26.783801 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:08:26 crc kubenswrapper[4907]: E0313 14:08:26.783866 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:08:27 crc kubenswrapper[4907]: E0313 14:08:27.002723 4907 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 13 14:08:28 crc kubenswrapper[4907]: I0313 14:08:28.782141 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:08:28 crc kubenswrapper[4907]: E0313 14:08:28.782500 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:08:28 crc kubenswrapper[4907]: I0313 14:08:28.783055 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:08:28 crc kubenswrapper[4907]: I0313 14:08:28.783183 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:08:28 crc kubenswrapper[4907]: E0313 14:08:28.783212 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzmvc" podUID="f797b1bb-e769-49c4-84ce-93a6fad2df93" Mar 13 14:08:28 crc kubenswrapper[4907]: I0313 14:08:28.783383 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:08:28 crc kubenswrapper[4907]: E0313 14:08:28.783432 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:08:28 crc kubenswrapper[4907]: E0313 14:08:28.783364 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:08:30 crc kubenswrapper[4907]: I0313 14:08:30.782202 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:08:30 crc kubenswrapper[4907]: I0313 14:08:30.782287 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:08:30 crc kubenswrapper[4907]: I0313 14:08:30.782202 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:08:30 crc kubenswrapper[4907]: E0313 14:08:30.782378 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:08:30 crc kubenswrapper[4907]: E0313 14:08:30.782467 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzmvc" podUID="f797b1bb-e769-49c4-84ce-93a6fad2df93" Mar 13 14:08:30 crc kubenswrapper[4907]: I0313 14:08:30.782294 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:08:30 crc kubenswrapper[4907]: E0313 14:08:30.782587 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:08:30 crc kubenswrapper[4907]: E0313 14:08:30.782675 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:08:31 crc kubenswrapper[4907]: I0313 14:08:31.785094 4907 scope.go:117] "RemoveContainer" containerID="9155434e2d8f0daa4504758fb72d6f498da107a64a92a0883ac687ed2af9caee" Mar 13 14:08:31 crc kubenswrapper[4907]: E0313 14:08:31.785435 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-qxlx5_openshift-ovn-kubernetes(1006376d-4033-454e-8a23-2b67882bdd0b)\"" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" podUID="1006376d-4033-454e-8a23-2b67882bdd0b" Mar 13 14:08:32 crc kubenswrapper[4907]: E0313 14:08:32.004246 4907 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 13 14:08:32 crc kubenswrapper[4907]: I0313 14:08:32.781932 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:08:32 crc kubenswrapper[4907]: I0313 14:08:32.782091 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:08:32 crc kubenswrapper[4907]: I0313 14:08:32.782040 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:08:32 crc kubenswrapper[4907]: I0313 14:08:32.782014 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:08:32 crc kubenswrapper[4907]: E0313 14:08:32.782274 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:08:32 crc kubenswrapper[4907]: E0313 14:08:32.782453 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:08:32 crc kubenswrapper[4907]: E0313 14:08:32.782604 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzmvc" podUID="f797b1bb-e769-49c4-84ce-93a6fad2df93" Mar 13 14:08:32 crc kubenswrapper[4907]: E0313 14:08:32.782706 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:08:33 crc kubenswrapper[4907]: I0313 14:08:33.783733 4907 scope.go:117] "RemoveContainer" containerID="ef632bb17c8f529b35d0882c8d4ba3dc213b0308b46a4ade8426b9c08e64bb2b" Mar 13 14:08:34 crc kubenswrapper[4907]: I0313 14:08:34.782387 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:08:34 crc kubenswrapper[4907]: I0313 14:08:34.782388 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:08:34 crc kubenswrapper[4907]: E0313 14:08:34.782874 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:08:34 crc kubenswrapper[4907]: I0313 14:08:34.782541 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:08:34 crc kubenswrapper[4907]: I0313 14:08:34.782419 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:08:34 crc kubenswrapper[4907]: E0313 14:08:34.783216 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:08:34 crc kubenswrapper[4907]: E0313 14:08:34.783358 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:08:34 crc kubenswrapper[4907]: E0313 14:08:34.783522 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzmvc" podUID="f797b1bb-e769-49c4-84ce-93a6fad2df93" Mar 13 14:08:34 crc kubenswrapper[4907]: I0313 14:08:34.831774 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-xtwf7_64ebd338-0158-44dc-90ab-e7c285e87762/kube-multus/1.log" Mar 13 14:08:34 crc kubenswrapper[4907]: I0313 14:08:34.831865 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-xtwf7" event={"ID":"64ebd338-0158-44dc-90ab-e7c285e87762","Type":"ContainerStarted","Data":"8788f4b58c32d23e53448793bceb517c39726e72de35c52e959fc15e062ae173"} Mar 13 14:08:36 crc kubenswrapper[4907]: I0313 14:08:36.782359 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:08:36 crc kubenswrapper[4907]: I0313 14:08:36.782457 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:08:36 crc kubenswrapper[4907]: I0313 14:08:36.782481 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:08:36 crc kubenswrapper[4907]: I0313 14:08:36.782408 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:08:36 crc kubenswrapper[4907]: E0313 14:08:36.782637 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:08:36 crc kubenswrapper[4907]: E0313 14:08:36.782820 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:08:36 crc kubenswrapper[4907]: E0313 14:08:36.783132 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:08:36 crc kubenswrapper[4907]: E0313 14:08:36.783397 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzmvc" podUID="f797b1bb-e769-49c4-84ce-93a6fad2df93" Mar 13 14:08:37 crc kubenswrapper[4907]: E0313 14:08:37.005297 4907 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 13 14:08:38 crc kubenswrapper[4907]: I0313 14:08:38.781414 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:08:38 crc kubenswrapper[4907]: I0313 14:08:38.781443 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:08:38 crc kubenswrapper[4907]: E0313 14:08:38.781662 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:08:38 crc kubenswrapper[4907]: I0313 14:08:38.781475 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:08:38 crc kubenswrapper[4907]: I0313 14:08:38.781463 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:08:38 crc kubenswrapper[4907]: E0313 14:08:38.782032 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:08:38 crc kubenswrapper[4907]: E0313 14:08:38.782142 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:08:38 crc kubenswrapper[4907]: E0313 14:08:38.782315 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzmvc" podUID="f797b1bb-e769-49c4-84ce-93a6fad2df93" Mar 13 14:08:40 crc kubenswrapper[4907]: I0313 14:08:40.768335 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:08:40 crc kubenswrapper[4907]: E0313 14:08:40.768651 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:10:42.768599129 +0000 UTC m=+341.668386858 (durationBeforeRetry 2m2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:08:40 crc kubenswrapper[4907]: I0313 14:08:40.782204 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:08:40 crc kubenswrapper[4907]: I0313 14:08:40.782273 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:08:40 crc kubenswrapper[4907]: I0313 14:08:40.782377 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:08:40 crc kubenswrapper[4907]: I0313 14:08:40.782229 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:08:40 crc kubenswrapper[4907]: E0313 14:08:40.782420 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:08:40 crc kubenswrapper[4907]: E0313 14:08:40.782592 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzmvc" podUID="f797b1bb-e769-49c4-84ce-93a6fad2df93" Mar 13 14:08:40 crc kubenswrapper[4907]: E0313 14:08:40.782756 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:08:40 crc kubenswrapper[4907]: E0313 14:08:40.782987 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:08:40 crc kubenswrapper[4907]: I0313 14:08:40.870472 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:08:40 crc kubenswrapper[4907]: I0313 14:08:40.870553 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:08:40 crc kubenswrapper[4907]: I0313 14:08:40.870615 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:08:40 crc kubenswrapper[4907]: I0313 14:08:40.870674 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:08:40 crc kubenswrapper[4907]: E0313 14:08:40.870729 4907 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 13 14:08:40 crc kubenswrapper[4907]: E0313 14:08:40.870799 4907 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 13 14:08:40 crc kubenswrapper[4907]: E0313 14:08:40.870856 4907 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 13 14:08:40 crc kubenswrapper[4907]: E0313 14:08:40.870932 4907 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 13 14:08:40 crc kubenswrapper[4907]: E0313 14:08:40.870944 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-13 14:10:42.870831048 +0000 UTC m=+341.770618777 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 13 14:08:40 crc kubenswrapper[4907]: E0313 14:08:40.870973 4907 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 14:08:40 crc kubenswrapper[4907]: E0313 14:08:40.870993 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-13 14:10:42.870973872 +0000 UTC m=+341.770761601 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 13 14:08:40 crc kubenswrapper[4907]: E0313 14:08:40.871048 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-13 14:10:42.871028564 +0000 UTC m=+341.770816283 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 14:08:40 crc kubenswrapper[4907]: E0313 14:08:40.871171 4907 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 13 14:08:40 crc kubenswrapper[4907]: E0313 14:08:40.871244 4907 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 13 14:08:40 crc kubenswrapper[4907]: E0313 14:08:40.871273 4907 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 14:08:40 crc kubenswrapper[4907]: E0313 14:08:40.871403 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-13 14:10:42.871367573 +0000 UTC m=+341.771155302 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 14:08:42 crc kubenswrapper[4907]: E0313 14:08:42.007471 4907 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 13 14:08:42 crc kubenswrapper[4907]: I0313 14:08:42.781385 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:08:42 crc kubenswrapper[4907]: I0313 14:08:42.781436 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:08:42 crc kubenswrapper[4907]: I0313 14:08:42.781436 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:08:42 crc kubenswrapper[4907]: I0313 14:08:42.781371 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:08:42 crc kubenswrapper[4907]: E0313 14:08:42.781529 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:08:42 crc kubenswrapper[4907]: E0313 14:08:42.781757 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:08:42 crc kubenswrapper[4907]: E0313 14:08:42.782015 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:08:42 crc kubenswrapper[4907]: E0313 14:08:42.782314 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzmvc" podUID="f797b1bb-e769-49c4-84ce-93a6fad2df93" Mar 13 14:08:44 crc kubenswrapper[4907]: I0313 14:08:44.782376 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:08:44 crc kubenswrapper[4907]: I0313 14:08:44.782424 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:08:44 crc kubenswrapper[4907]: I0313 14:08:44.782540 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:08:44 crc kubenswrapper[4907]: E0313 14:08:44.784082 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzmvc" podUID="f797b1bb-e769-49c4-84ce-93a6fad2df93" Mar 13 14:08:44 crc kubenswrapper[4907]: I0313 14:08:44.782589 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:08:44 crc kubenswrapper[4907]: E0313 14:08:44.784194 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:08:44 crc kubenswrapper[4907]: E0313 14:08:44.784467 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:08:44 crc kubenswrapper[4907]: E0313 14:08:44.784671 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:08:45 crc kubenswrapper[4907]: I0313 14:08:45.783190 4907 scope.go:117] "RemoveContainer" containerID="9155434e2d8f0daa4504758fb72d6f498da107a64a92a0883ac687ed2af9caee" Mar 13 14:08:46 crc kubenswrapper[4907]: I0313 14:08:46.144936 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qxlx5_1006376d-4033-454e-8a23-2b67882bdd0b/ovnkube-controller/3.log" Mar 13 14:08:46 crc kubenswrapper[4907]: I0313 14:08:46.150805 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" event={"ID":"1006376d-4033-454e-8a23-2b67882bdd0b","Type":"ContainerStarted","Data":"a3d2a347b2f85682f2a54c30e63b7e9c88ffe402d6d71fa7ce6656786f93b886"} Mar 13 14:08:46 crc kubenswrapper[4907]: I0313 14:08:46.151592 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:08:46 crc kubenswrapper[4907]: I0313 14:08:46.191835 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" podStartSLOduration=165.191798508 podStartE2EDuration="2m45.191798508s" podCreationTimestamp="2026-03-13 14:06:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:08:46.188790897 +0000 UTC m=+225.088578586" watchObservedRunningTime="2026-03-13 14:08:46.191798508 +0000 UTC m=+225.091586237" Mar 13 14:08:46 crc kubenswrapper[4907]: I0313 14:08:46.781342 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:08:46 crc kubenswrapper[4907]: I0313 14:08:46.781428 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:08:46 crc kubenswrapper[4907]: E0313 14:08:46.781537 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzmvc" podUID="f797b1bb-e769-49c4-84ce-93a6fad2df93" Mar 13 14:08:46 crc kubenswrapper[4907]: I0313 14:08:46.781563 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:08:46 crc kubenswrapper[4907]: I0313 14:08:46.781342 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:08:46 crc kubenswrapper[4907]: E0313 14:08:46.781674 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:08:46 crc kubenswrapper[4907]: E0313 14:08:46.781782 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:08:46 crc kubenswrapper[4907]: E0313 14:08:46.782046 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:08:46 crc kubenswrapper[4907]: I0313 14:08:46.944104 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-tzmvc"] Mar 13 14:08:47 crc kubenswrapper[4907]: E0313 14:08:47.010207 4907 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 13 14:08:47 crc kubenswrapper[4907]: I0313 14:08:47.154751 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:08:47 crc kubenswrapper[4907]: E0313 14:08:47.154988 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzmvc" podUID="f797b1bb-e769-49c4-84ce-93a6fad2df93" Mar 13 14:08:48 crc kubenswrapper[4907]: I0313 14:08:48.041565 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 14:08:48 crc kubenswrapper[4907]: I0313 14:08:48.041676 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 14:08:48 crc kubenswrapper[4907]: I0313 14:08:48.782232 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:08:48 crc kubenswrapper[4907]: E0313 14:08:48.782934 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:08:48 crc kubenswrapper[4907]: I0313 14:08:48.782302 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:08:48 crc kubenswrapper[4907]: E0313 14:08:48.783078 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzmvc" podUID="f797b1bb-e769-49c4-84ce-93a6fad2df93" Mar 13 14:08:48 crc kubenswrapper[4907]: I0313 14:08:48.782401 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:08:48 crc kubenswrapper[4907]: E0313 14:08:48.783178 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:08:48 crc kubenswrapper[4907]: I0313 14:08:48.782241 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:08:48 crc kubenswrapper[4907]: E0313 14:08:48.783264 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:08:50 crc kubenswrapper[4907]: I0313 14:08:50.781519 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:08:50 crc kubenswrapper[4907]: I0313 14:08:50.781608 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:08:50 crc kubenswrapper[4907]: I0313 14:08:50.781664 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:08:50 crc kubenswrapper[4907]: I0313 14:08:50.781596 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:08:50 crc kubenswrapper[4907]: E0313 14:08:50.781773 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzmvc" podUID="f797b1bb-e769-49c4-84ce-93a6fad2df93" Mar 13 14:08:50 crc kubenswrapper[4907]: E0313 14:08:50.781991 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:08:50 crc kubenswrapper[4907]: E0313 14:08:50.782108 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:08:50 crc kubenswrapper[4907]: E0313 14:08:50.782260 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:08:52 crc kubenswrapper[4907]: I0313 14:08:52.782135 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:08:52 crc kubenswrapper[4907]: I0313 14:08:52.782169 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:08:52 crc kubenswrapper[4907]: I0313 14:08:52.782131 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:08:52 crc kubenswrapper[4907]: I0313 14:08:52.782432 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:08:52 crc kubenswrapper[4907]: I0313 14:08:52.787803 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Mar 13 14:08:52 crc kubenswrapper[4907]: I0313 14:08:52.788170 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Mar 13 14:08:52 crc kubenswrapper[4907]: I0313 14:08:52.788321 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Mar 13 14:08:52 crc kubenswrapper[4907]: I0313 14:08:52.788379 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Mar 13 14:08:52 crc kubenswrapper[4907]: I0313 14:08:52.788485 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Mar 13 14:08:52 crc kubenswrapper[4907]: I0313 14:08:52.788568 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.464995 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.512943 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-ksgr6"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.513688 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ksgr6" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.514634 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-rnh5t"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.515178 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-kwfhg"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.515586 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-rnh5t" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.515701 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-kwfhg" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.522547 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-4h9q7"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.523431 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4h9q7" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.525010 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.525803 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.526381 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.526612 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.526997 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.527222 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.527426 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.527691 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.528027 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.528239 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.528443 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.528655 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.528952 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.529219 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.529469 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.529676 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-r86r4"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.530645 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-r86r4" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.531170 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-prlhd"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.532351 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-prlhd" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.534249 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-52vhb"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.535201 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-52vhb" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.550016 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.550604 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.550990 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.551160 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.550709 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.551681 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.550784 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.552070 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.552208 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.552284 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-wjrms"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.552677 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.553052 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.553254 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-wjrms" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.553481 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-ng4kz"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.553820 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-ng4kz" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.553271 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.554211 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.569423 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.571598 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.574512 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.553392 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.553465 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.553572 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.553693 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.554009 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.593083 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.593866 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.594257 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.594267 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.594333 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.594437 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.594450 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.594612 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.596197 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.597099 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.597109 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5a9c889b-5954-4d3a-ae72-12000ee1cc25-trusted-ca-bundle\") pod \"apiserver-76f77b778f-r86r4\" (UID: \"5a9c889b-5954-4d3a-ae72-12000ee1cc25\") " pod="openshift-apiserver/apiserver-76f77b778f-r86r4" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.597160 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/814c6dec-6e96-4068-9aec-9e57fefca5db-config\") pod \"controller-manager-879f6c89f-kwfhg\" (UID: \"814c6dec-6e96-4068-9aec-9e57fefca5db\") " pod="openshift-controller-manager/controller-manager-879f6c89f-kwfhg" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.597197 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gn5jt\" (UniqueName: \"kubernetes.io/projected/814c6dec-6e96-4068-9aec-9e57fefca5db-kube-api-access-gn5jt\") pod \"controller-manager-879f6c89f-kwfhg\" (UID: \"814c6dec-6e96-4068-9aec-9e57fefca5db\") " pod="openshift-controller-manager/controller-manager-879f6c89f-kwfhg" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.597224 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/1be912b5-9d1d-48bd-a0ab-7f601f9eb9da-audit-policies\") pod \"apiserver-7bbb656c7d-ksgr6\" (UID: \"1be912b5-9d1d-48bd-a0ab-7f601f9eb9da\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ksgr6" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.597263 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/5a9c889b-5954-4d3a-ae72-12000ee1cc25-image-import-ca\") pod \"apiserver-76f77b778f-r86r4\" (UID: \"5a9c889b-5954-4d3a-ae72-12000ee1cc25\") " pod="openshift-apiserver/apiserver-76f77b778f-r86r4" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.597298 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/814c6dec-6e96-4068-9aec-9e57fefca5db-serving-cert\") pod \"controller-manager-879f6c89f-kwfhg\" (UID: \"814c6dec-6e96-4068-9aec-9e57fefca5db\") " pod="openshift-controller-manager/controller-manager-879f6c89f-kwfhg" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.597335 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/005dfe54-e402-4aa4-8b67-f7e7b685debd-machine-approver-tls\") pod \"machine-approver-56656f9798-52vhb\" (UID: \"005dfe54-e402-4aa4-8b67-f7e7b685debd\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-52vhb" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.597364 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a3baeb4f-988d-4d42-b469-a310ca127745-service-ca-bundle\") pod \"authentication-operator-69f744f599-rnh5t\" (UID: \"a3baeb4f-988d-4d42-b469-a310ca127745\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-rnh5t" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.597386 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1be912b5-9d1d-48bd-a0ab-7f601f9eb9da-serving-cert\") pod \"apiserver-7bbb656c7d-ksgr6\" (UID: \"1be912b5-9d1d-48bd-a0ab-7f601f9eb9da\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ksgr6" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.597408 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9bb1952d-0e55-4e45-81c0-69a8cddb827e-config\") pod \"route-controller-manager-6576b87f9c-4h9q7\" (UID: \"9bb1952d-0e55-4e45-81c0-69a8cddb827e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4h9q7" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.597433 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a3baeb4f-988d-4d42-b469-a310ca127745-config\") pod \"authentication-operator-69f744f599-rnh5t\" (UID: \"a3baeb4f-988d-4d42-b469-a310ca127745\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-rnh5t" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.597456 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/5a9c889b-5954-4d3a-ae72-12000ee1cc25-audit\") pod \"apiserver-76f77b778f-r86r4\" (UID: \"5a9c889b-5954-4d3a-ae72-12000ee1cc25\") " pod="openshift-apiserver/apiserver-76f77b778f-r86r4" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.597507 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-ng4kz\" (UID: \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\") " pod="openshift-authentication/oauth-openshift-558db77b4-ng4kz" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.597580 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-ng4kz\" (UID: \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\") " pod="openshift-authentication/oauth-openshift-558db77b4-ng4kz" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.597613 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/5a9c889b-5954-4d3a-ae72-12000ee1cc25-etcd-client\") pod \"apiserver-76f77b778f-r86r4\" (UID: \"5a9c889b-5954-4d3a-ae72-12000ee1cc25\") " pod="openshift-apiserver/apiserver-76f77b778f-r86r4" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.597638 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/1be912b5-9d1d-48bd-a0ab-7f601f9eb9da-audit-dir\") pod \"apiserver-7bbb656c7d-ksgr6\" (UID: \"1be912b5-9d1d-48bd-a0ab-7f601f9eb9da\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ksgr6" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.597662 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3661f6b2-1d5e-42f0-947f-a32b66258a1f-audit-dir\") pod \"oauth-openshift-558db77b4-ng4kz\" (UID: \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\") " pod="openshift-authentication/oauth-openshift-558db77b4-ng4kz" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.597672 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.597696 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7jwmk\" (UniqueName: \"kubernetes.io/projected/b75dbfe3-8887-4a0b-9541-a4e4000924cb-kube-api-access-7jwmk\") pod \"machine-api-operator-5694c8668f-prlhd\" (UID: \"b75dbfe3-8887-4a0b-9541-a4e4000924cb\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-prlhd" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.597725 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9bb1952d-0e55-4e45-81c0-69a8cddb827e-serving-cert\") pod \"route-controller-manager-6576b87f9c-4h9q7\" (UID: \"9bb1952d-0e55-4e45-81c0-69a8cddb827e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4h9q7" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.597750 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-ng4kz\" (UID: \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\") " pod="openshift-authentication/oauth-openshift-558db77b4-ng4kz" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.597786 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/814c6dec-6e96-4068-9aec-9e57fefca5db-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-kwfhg\" (UID: \"814c6dec-6e96-4068-9aec-9e57fefca5db\") " pod="openshift-controller-manager/controller-manager-879f6c89f-kwfhg" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.597815 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/005dfe54-e402-4aa4-8b67-f7e7b685debd-auth-proxy-config\") pod \"machine-approver-56656f9798-52vhb\" (UID: \"005dfe54-e402-4aa4-8b67-f7e7b685debd\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-52vhb" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.597835 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-ng4kz\" (UID: \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\") " pod="openshift-authentication/oauth-openshift-558db77b4-ng4kz" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.597855 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1be912b5-9d1d-48bd-a0ab-7f601f9eb9da-etcd-client\") pod \"apiserver-7bbb656c7d-ksgr6\" (UID: \"1be912b5-9d1d-48bd-a0ab-7f601f9eb9da\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ksgr6" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.597876 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-ng4kz\" (UID: \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\") " pod="openshift-authentication/oauth-openshift-558db77b4-ng4kz" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.597927 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1be912b5-9d1d-48bd-a0ab-7f601f9eb9da-encryption-config\") pod \"apiserver-7bbb656c7d-ksgr6\" (UID: \"1be912b5-9d1d-48bd-a0ab-7f601f9eb9da\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ksgr6" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.597949 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-57msq\" (UniqueName: \"kubernetes.io/projected/1be912b5-9d1d-48bd-a0ab-7f601f9eb9da-kube-api-access-57msq\") pod \"apiserver-7bbb656c7d-ksgr6\" (UID: \"1be912b5-9d1d-48bd-a0ab-7f601f9eb9da\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ksgr6" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.597969 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-ng4kz\" (UID: \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\") " pod="openshift-authentication/oauth-openshift-558db77b4-ng4kz" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.597990 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/814c6dec-6e96-4068-9aec-9e57fefca5db-client-ca\") pod \"controller-manager-879f6c89f-kwfhg\" (UID: \"814c6dec-6e96-4068-9aec-9e57fefca5db\") " pod="openshift-controller-manager/controller-manager-879f6c89f-kwfhg" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.598014 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a3baeb4f-988d-4d42-b469-a310ca127745-serving-cert\") pod \"authentication-operator-69f744f599-rnh5t\" (UID: \"a3baeb4f-988d-4d42-b469-a310ca127745\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-rnh5t" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.598034 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5a9c889b-5954-4d3a-ae72-12000ee1cc25-audit-dir\") pod \"apiserver-76f77b778f-r86r4\" (UID: \"5a9c889b-5954-4d3a-ae72-12000ee1cc25\") " pod="openshift-apiserver/apiserver-76f77b778f-r86r4" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.598054 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gqtv7\" (UniqueName: \"kubernetes.io/projected/9bb1952d-0e55-4e45-81c0-69a8cddb827e-kube-api-access-gqtv7\") pod \"route-controller-manager-6576b87f9c-4h9q7\" (UID: \"9bb1952d-0e55-4e45-81c0-69a8cddb827e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4h9q7" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.598078 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-289bg\" (UniqueName: \"kubernetes.io/projected/a3baeb4f-988d-4d42-b469-a310ca127745-kube-api-access-289bg\") pod \"authentication-operator-69f744f599-rnh5t\" (UID: \"a3baeb4f-988d-4d42-b469-a310ca127745\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-rnh5t" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.598099 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a3baeb4f-988d-4d42-b469-a310ca127745-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-rnh5t\" (UID: \"a3baeb4f-988d-4d42-b469-a310ca127745\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-rnh5t" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.598136 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/005dfe54-e402-4aa4-8b67-f7e7b685debd-config\") pod \"machine-approver-56656f9798-52vhb\" (UID: \"005dfe54-e402-4aa4-8b67-f7e7b685debd\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-52vhb" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.598173 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.598270 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.598606 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.599073 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.599185 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-wgpdd"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.599765 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-wgpdd" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.600015 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hfcsw"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.600341 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.600414 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hfcsw" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.601531 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-l9xpb"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.601947 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-l9xpb" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.602042 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-544df"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.602438 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-544df" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.604101 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.604113 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.604545 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.598159 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-ng4kz\" (UID: \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\") " pod="openshift-authentication/oauth-openshift-558db77b4-ng4kz" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.605643 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/71c234e9-5f3d-4b0c-94d4-80b4e33a6dda-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-wjrms\" (UID: \"71c234e9-5f3d-4b0c-94d4-80b4e33a6dda\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-wjrms" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.605680 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-ng4kz\" (UID: \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\") " pod="openshift-authentication/oauth-openshift-558db77b4-ng4kz" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.605704 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-ng4kz\" (UID: \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\") " pod="openshift-authentication/oauth-openshift-558db77b4-ng4kz" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.605739 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5a9c889b-5954-4d3a-ae72-12000ee1cc25-serving-cert\") pod \"apiserver-76f77b778f-r86r4\" (UID: \"5a9c889b-5954-4d3a-ae72-12000ee1cc25\") " pod="openshift-apiserver/apiserver-76f77b778f-r86r4" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.605763 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9qt98\" (UniqueName: \"kubernetes.io/projected/005dfe54-e402-4aa4-8b67-f7e7b685debd-kube-api-access-9qt98\") pod \"machine-approver-56656f9798-52vhb\" (UID: \"005dfe54-e402-4aa4-8b67-f7e7b685debd\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-52vhb" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.605785 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/5a9c889b-5954-4d3a-ae72-12000ee1cc25-encryption-config\") pod \"apiserver-76f77b778f-r86r4\" (UID: \"5a9c889b-5954-4d3a-ae72-12000ee1cc25\") " pod="openshift-apiserver/apiserver-76f77b778f-r86r4" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.605811 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b75dbfe3-8887-4a0b-9541-a4e4000924cb-config\") pod \"machine-api-operator-5694c8668f-prlhd\" (UID: \"b75dbfe3-8887-4a0b-9541-a4e4000924cb\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-prlhd" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.605845 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-ng4kz\" (UID: \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\") " pod="openshift-authentication/oauth-openshift-558db77b4-ng4kz" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.605901 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/5a9c889b-5954-4d3a-ae72-12000ee1cc25-etcd-serving-ca\") pod \"apiserver-76f77b778f-r86r4\" (UID: \"5a9c889b-5954-4d3a-ae72-12000ee1cc25\") " pod="openshift-apiserver/apiserver-76f77b778f-r86r4" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.605928 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mtqz9\" (UniqueName: \"kubernetes.io/projected/3661f6b2-1d5e-42f0-947f-a32b66258a1f-kube-api-access-mtqz9\") pod \"oauth-openshift-558db77b4-ng4kz\" (UID: \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\") " pod="openshift-authentication/oauth-openshift-558db77b4-ng4kz" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.605947 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hlg49\" (UniqueName: \"kubernetes.io/projected/5a9c889b-5954-4d3a-ae72-12000ee1cc25-kube-api-access-hlg49\") pod \"apiserver-76f77b778f-r86r4\" (UID: \"5a9c889b-5954-4d3a-ae72-12000ee1cc25\") " pod="openshift-apiserver/apiserver-76f77b778f-r86r4" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.605968 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/b75dbfe3-8887-4a0b-9541-a4e4000924cb-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-prlhd\" (UID: \"b75dbfe3-8887-4a0b-9541-a4e4000924cb\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-prlhd" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.605993 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5a9c889b-5954-4d3a-ae72-12000ee1cc25-config\") pod \"apiserver-76f77b778f-r86r4\" (UID: \"5a9c889b-5954-4d3a-ae72-12000ee1cc25\") " pod="openshift-apiserver/apiserver-76f77b778f-r86r4" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.606012 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/b75dbfe3-8887-4a0b-9541-a4e4000924cb-images\") pod \"machine-api-operator-5694c8668f-prlhd\" (UID: \"b75dbfe3-8887-4a0b-9541-a4e4000924cb\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-prlhd" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.606033 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/3661f6b2-1d5e-42f0-947f-a32b66258a1f-audit-policies\") pod \"oauth-openshift-558db77b4-ng4kz\" (UID: \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\") " pod="openshift-authentication/oauth-openshift-558db77b4-ng4kz" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.606078 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-ng4kz\" (UID: \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\") " pod="openshift-authentication/oauth-openshift-558db77b4-ng4kz" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.606104 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1be912b5-9d1d-48bd-a0ab-7f601f9eb9da-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-ksgr6\" (UID: \"1be912b5-9d1d-48bd-a0ab-7f601f9eb9da\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ksgr6" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.606124 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s47gg\" (UniqueName: \"kubernetes.io/projected/71c234e9-5f3d-4b0c-94d4-80b4e33a6dda-kube-api-access-s47gg\") pod \"cluster-samples-operator-665b6dd947-wjrms\" (UID: \"71c234e9-5f3d-4b0c-94d4-80b4e33a6dda\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-wjrms" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.606143 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1be912b5-9d1d-48bd-a0ab-7f601f9eb9da-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-ksgr6\" (UID: \"1be912b5-9d1d-48bd-a0ab-7f601f9eb9da\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ksgr6" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.606243 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9bb1952d-0e55-4e45-81c0-69a8cddb827e-client-ca\") pod \"route-controller-manager-6576b87f9c-4h9q7\" (UID: \"9bb1952d-0e55-4e45-81c0-69a8cddb827e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4h9q7" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.606296 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/5a9c889b-5954-4d3a-ae72-12000ee1cc25-node-pullsecrets\") pod \"apiserver-76f77b778f-r86r4\" (UID: \"5a9c889b-5954-4d3a-ae72-12000ee1cc25\") " pod="openshift-apiserver/apiserver-76f77b778f-r86r4" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.606696 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.609628 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-w76kw"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.610595 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-v9m24"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.609089 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.611132 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-xkdjh"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.611742 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-xkdjh" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.612163 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-w76kw" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.612503 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-v9m24" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.612505 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.615950 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-q9z8k"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.617788 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.620098 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-hwf6r"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.622542 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-hwf6r" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.621566 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.621737 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.621801 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.623656 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.623861 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.623749 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.623822 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.624327 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-f8mmz"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.624825 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-rnh5t"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.624998 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.625132 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-f8mmz" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.625264 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.625381 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.625539 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-56rqr"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.626106 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-56rqr" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.626595 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.628390 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-5szhc"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.629282 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5szhc" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.629387 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.632053 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.633309 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.634148 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.634317 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.634554 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.634689 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-zz77x"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.642732 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.643248 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.643658 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.646054 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.646820 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.647487 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.647628 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.648027 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-rmkfg"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.651232 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-rmkfg" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.651678 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-zz77x" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.651949 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.652664 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.652965 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.653209 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.653352 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.653634 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.653908 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.654377 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.687327 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.687492 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.687653 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.687922 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.689561 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-t68ql"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.689997 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.690416 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-hh9sk"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.690455 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.690772 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-ksgr6"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.690984 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-t68ql" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.691213 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.691739 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.692608 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.695271 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.696115 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-92gs7"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.696815 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-ksfb8"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.697379 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-ksfb8" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.697604 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-92gs7" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.698300 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.690874 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-hh9sk" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.705138 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.706118 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-vfbsx"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.706843 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/005dfe54-e402-4aa4-8b67-f7e7b685debd-auth-proxy-config\") pod \"machine-approver-56656f9798-52vhb\" (UID: \"005dfe54-e402-4aa4-8b67-f7e7b685debd\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-52vhb" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.706877 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-ng4kz\" (UID: \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\") " pod="openshift-authentication/oauth-openshift-558db77b4-ng4kz" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.706917 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-vfbsx" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.706944 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/4c8c0266-5452-4abb-92c5-c536df94dd41-console-serving-cert\") pod \"console-f9d7485db-l9xpb\" (UID: \"4c8c0266-5452-4abb-92c5-c536df94dd41\") " pod="openshift-console/console-f9d7485db-l9xpb" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.706967 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e9d44a1e-0eab-4bd6-833f-143436e43758-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-hh9sk\" (UID: \"e9d44a1e-0eab-4bd6-833f-143436e43758\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-hh9sk" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.706987 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/4c8c0266-5452-4abb-92c5-c536df94dd41-console-config\") pod \"console-f9d7485db-l9xpb\" (UID: \"4c8c0266-5452-4abb-92c5-c536df94dd41\") " pod="openshift-console/console-f9d7485db-l9xpb" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.707013 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1be912b5-9d1d-48bd-a0ab-7f601f9eb9da-etcd-client\") pod \"apiserver-7bbb656c7d-ksgr6\" (UID: \"1be912b5-9d1d-48bd-a0ab-7f601f9eb9da\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ksgr6" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.707032 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-ng4kz\" (UID: \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\") " pod="openshift-authentication/oauth-openshift-558db77b4-ng4kz" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.707052 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-ng4kz\" (UID: \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\") " pod="openshift-authentication/oauth-openshift-558db77b4-ng4kz" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.707078 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/814c6dec-6e96-4068-9aec-9e57fefca5db-client-ca\") pod \"controller-manager-879f6c89f-kwfhg\" (UID: \"814c6dec-6e96-4068-9aec-9e57fefca5db\") " pod="openshift-controller-manager/controller-manager-879f6c89f-kwfhg" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.707108 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a3baeb4f-988d-4d42-b469-a310ca127745-serving-cert\") pod \"authentication-operator-69f744f599-rnh5t\" (UID: \"a3baeb4f-988d-4d42-b469-a310ca127745\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-rnh5t" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.707129 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5a9c889b-5954-4d3a-ae72-12000ee1cc25-audit-dir\") pod \"apiserver-76f77b778f-r86r4\" (UID: \"5a9c889b-5954-4d3a-ae72-12000ee1cc25\") " pod="openshift-apiserver/apiserver-76f77b778f-r86r4" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.707150 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1be912b5-9d1d-48bd-a0ab-7f601f9eb9da-encryption-config\") pod \"apiserver-7bbb656c7d-ksgr6\" (UID: \"1be912b5-9d1d-48bd-a0ab-7f601f9eb9da\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ksgr6" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.707171 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-57msq\" (UniqueName: \"kubernetes.io/projected/1be912b5-9d1d-48bd-a0ab-7f601f9eb9da-kube-api-access-57msq\") pod \"apiserver-7bbb656c7d-ksgr6\" (UID: \"1be912b5-9d1d-48bd-a0ab-7f601f9eb9da\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ksgr6" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.707194 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/4c8c0266-5452-4abb-92c5-c536df94dd41-service-ca\") pod \"console-f9d7485db-l9xpb\" (UID: \"4c8c0266-5452-4abb-92c5-c536df94dd41\") " pod="openshift-console/console-f9d7485db-l9xpb" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.707217 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/963f6709-e807-40bb-aa9e-9bba24e82b71-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-v9m24\" (UID: \"963f6709-e807-40bb-aa9e-9bba24e82b71\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-v9m24" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.707238 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gqtv7\" (UniqueName: \"kubernetes.io/projected/9bb1952d-0e55-4e45-81c0-69a8cddb827e-kube-api-access-gqtv7\") pod \"route-controller-manager-6576b87f9c-4h9q7\" (UID: \"9bb1952d-0e55-4e45-81c0-69a8cddb827e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4h9q7" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.707259 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-289bg\" (UniqueName: \"kubernetes.io/projected/a3baeb4f-988d-4d42-b469-a310ca127745-kube-api-access-289bg\") pod \"authentication-operator-69f744f599-rnh5t\" (UID: \"a3baeb4f-988d-4d42-b469-a310ca127745\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-rnh5t" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.707255 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.707281 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ae320273-06e8-43c5-a64f-acd80ad16d1c-bound-sa-token\") pod \"ingress-operator-5b745b69d9-5szhc\" (UID: \"ae320273-06e8-43c5-a64f-acd80ad16d1c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5szhc" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.707305 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a3baeb4f-988d-4d42-b469-a310ca127745-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-rnh5t\" (UID: \"a3baeb4f-988d-4d42-b469-a310ca127745\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-rnh5t" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.707374 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/005dfe54-e402-4aa4-8b67-f7e7b685debd-config\") pod \"machine-approver-56656f9798-52vhb\" (UID: \"005dfe54-e402-4aa4-8b67-f7e7b685debd\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-52vhb" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.707399 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-ng4kz\" (UID: \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\") " pod="openshift-authentication/oauth-openshift-558db77b4-ng4kz" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.707423 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/ae320273-06e8-43c5-a64f-acd80ad16d1c-metrics-tls\") pod \"ingress-operator-5b745b69d9-5szhc\" (UID: \"ae320273-06e8-43c5-a64f-acd80ad16d1c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5szhc" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.707446 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-ng4kz\" (UID: \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\") " pod="openshift-authentication/oauth-openshift-558db77b4-ng4kz" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.707473 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5a9c889b-5954-4d3a-ae72-12000ee1cc25-serving-cert\") pod \"apiserver-76f77b778f-r86r4\" (UID: \"5a9c889b-5954-4d3a-ae72-12000ee1cc25\") " pod="openshift-apiserver/apiserver-76f77b778f-r86r4" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.707501 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9qt98\" (UniqueName: \"kubernetes.io/projected/005dfe54-e402-4aa4-8b67-f7e7b685debd-kube-api-access-9qt98\") pod \"machine-approver-56656f9798-52vhb\" (UID: \"005dfe54-e402-4aa4-8b67-f7e7b685debd\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-52vhb" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.707521 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/71c234e9-5f3d-4b0c-94d4-80b4e33a6dda-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-wjrms\" (UID: \"71c234e9-5f3d-4b0c-94d4-80b4e33a6dda\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-wjrms" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.707547 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-ng4kz\" (UID: \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\") " pod="openshift-authentication/oauth-openshift-558db77b4-ng4kz" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.707573 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/a8e9ea4a-43ad-4da0-99fa-c01eea712c8c-proxy-tls\") pod \"machine-config-controller-84d6567774-t68ql\" (UID: \"a8e9ea4a-43ad-4da0-99fa-c01eea712c8c\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-t68ql" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.707595 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a7b66665-221d-4334-8263-aeb3b20e1c92-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-hfcsw\" (UID: \"a7b66665-221d-4334-8263-aeb3b20e1c92\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hfcsw" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.707619 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rnmbh\" (UniqueName: \"kubernetes.io/projected/a7b66665-221d-4334-8263-aeb3b20e1c92-kube-api-access-rnmbh\") pod \"openshift-controller-manager-operator-756b6f6bc6-hfcsw\" (UID: \"a7b66665-221d-4334-8263-aeb3b20e1c92\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hfcsw" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.707675 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7h5xn\" (UniqueName: \"kubernetes.io/projected/963f6709-e807-40bb-aa9e-9bba24e82b71-kube-api-access-7h5xn\") pod \"cluster-image-registry-operator-dc59b4c8b-v9m24\" (UID: \"963f6709-e807-40bb-aa9e-9bba24e82b71\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-v9m24" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.707699 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-ng4kz\" (UID: \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\") " pod="openshift-authentication/oauth-openshift-558db77b4-ng4kz" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.707724 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/5a9c889b-5954-4d3a-ae72-12000ee1cc25-etcd-serving-ca\") pod \"apiserver-76f77b778f-r86r4\" (UID: \"5a9c889b-5954-4d3a-ae72-12000ee1cc25\") " pod="openshift-apiserver/apiserver-76f77b778f-r86r4" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.707747 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/5a9c889b-5954-4d3a-ae72-12000ee1cc25-encryption-config\") pod \"apiserver-76f77b778f-r86r4\" (UID: \"5a9c889b-5954-4d3a-ae72-12000ee1cc25\") " pod="openshift-apiserver/apiserver-76f77b778f-r86r4" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.707770 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b75dbfe3-8887-4a0b-9541-a4e4000924cb-config\") pod \"machine-api-operator-5694c8668f-prlhd\" (UID: \"b75dbfe3-8887-4a0b-9541-a4e4000924cb\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-prlhd" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.707786 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-djph7"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.708310 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-djph7" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.707790 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mtqz9\" (UniqueName: \"kubernetes.io/projected/3661f6b2-1d5e-42f0-947f-a32b66258a1f-kube-api-access-mtqz9\") pod \"oauth-openshift-558db77b4-ng4kz\" (UID: \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\") " pod="openshift-authentication/oauth-openshift-558db77b4-ng4kz" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.708518 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/4c8c0266-5452-4abb-92c5-c536df94dd41-console-oauth-config\") pod \"console-f9d7485db-l9xpb\" (UID: \"4c8c0266-5452-4abb-92c5-c536df94dd41\") " pod="openshift-console/console-f9d7485db-l9xpb" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.708557 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4c8c0266-5452-4abb-92c5-c536df94dd41-trusted-ca-bundle\") pod \"console-f9d7485db-l9xpb\" (UID: \"4c8c0266-5452-4abb-92c5-c536df94dd41\") " pod="openshift-console/console-f9d7485db-l9xpb" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.708599 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hlg49\" (UniqueName: \"kubernetes.io/projected/5a9c889b-5954-4d3a-ae72-12000ee1cc25-kube-api-access-hlg49\") pod \"apiserver-76f77b778f-r86r4\" (UID: \"5a9c889b-5954-4d3a-ae72-12000ee1cc25\") " pod="openshift-apiserver/apiserver-76f77b778f-r86r4" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.708627 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/b75dbfe3-8887-4a0b-9541-a4e4000924cb-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-prlhd\" (UID: \"b75dbfe3-8887-4a0b-9541-a4e4000924cb\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-prlhd" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.708655 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ae320273-06e8-43c5-a64f-acd80ad16d1c-trusted-ca\") pod \"ingress-operator-5b745b69d9-5szhc\" (UID: \"ae320273-06e8-43c5-a64f-acd80ad16d1c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5szhc" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.708679 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5a9c889b-5954-4d3a-ae72-12000ee1cc25-config\") pod \"apiserver-76f77b778f-r86r4\" (UID: \"5a9c889b-5954-4d3a-ae72-12000ee1cc25\") " pod="openshift-apiserver/apiserver-76f77b778f-r86r4" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.708680 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/005dfe54-e402-4aa4-8b67-f7e7b685debd-auth-proxy-config\") pod \"machine-approver-56656f9798-52vhb\" (UID: \"005dfe54-e402-4aa4-8b67-f7e7b685debd\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-52vhb" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.708702 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/b75dbfe3-8887-4a0b-9541-a4e4000924cb-images\") pod \"machine-api-operator-5694c8668f-prlhd\" (UID: \"b75dbfe3-8887-4a0b-9541-a4e4000924cb\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-prlhd" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.708727 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vxlx4\" (UniqueName: \"kubernetes.io/projected/4c8c0266-5452-4abb-92c5-c536df94dd41-kube-api-access-vxlx4\") pod \"console-f9d7485db-l9xpb\" (UID: \"4c8c0266-5452-4abb-92c5-c536df94dd41\") " pod="openshift-console/console-f9d7485db-l9xpb" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.708746 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/4c8c0266-5452-4abb-92c5-c536df94dd41-oauth-serving-cert\") pod \"console-f9d7485db-l9xpb\" (UID: \"4c8c0266-5452-4abb-92c5-c536df94dd41\") " pod="openshift-console/console-f9d7485db-l9xpb" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.708775 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1be912b5-9d1d-48bd-a0ab-7f601f9eb9da-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-ksgr6\" (UID: \"1be912b5-9d1d-48bd-a0ab-7f601f9eb9da\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ksgr6" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.708810 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s47gg\" (UniqueName: \"kubernetes.io/projected/71c234e9-5f3d-4b0c-94d4-80b4e33a6dda-kube-api-access-s47gg\") pod \"cluster-samples-operator-665b6dd947-wjrms\" (UID: \"71c234e9-5f3d-4b0c-94d4-80b4e33a6dda\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-wjrms" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.708838 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/3661f6b2-1d5e-42f0-947f-a32b66258a1f-audit-policies\") pod \"oauth-openshift-558db77b4-ng4kz\" (UID: \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\") " pod="openshift-authentication/oauth-openshift-558db77b4-ng4kz" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.708859 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-ng4kz\" (UID: \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\") " pod="openshift-authentication/oauth-openshift-558db77b4-ng4kz" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.708910 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9bb1952d-0e55-4e45-81c0-69a8cddb827e-client-ca\") pod \"route-controller-manager-6576b87f9c-4h9q7\" (UID: \"9bb1952d-0e55-4e45-81c0-69a8cddb827e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4h9q7" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.708930 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1be912b5-9d1d-48bd-a0ab-7f601f9eb9da-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-ksgr6\" (UID: \"1be912b5-9d1d-48bd-a0ab-7f601f9eb9da\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ksgr6" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.708949 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a7b66665-221d-4334-8263-aeb3b20e1c92-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-hfcsw\" (UID: \"a7b66665-221d-4334-8263-aeb3b20e1c92\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hfcsw" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.708967 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/5a9c889b-5954-4d3a-ae72-12000ee1cc25-node-pullsecrets\") pod \"apiserver-76f77b778f-r86r4\" (UID: \"5a9c889b-5954-4d3a-ae72-12000ee1cc25\") " pod="openshift-apiserver/apiserver-76f77b778f-r86r4" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.709058 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/814c6dec-6e96-4068-9aec-9e57fefca5db-config\") pod \"controller-manager-879f6c89f-kwfhg\" (UID: \"814c6dec-6e96-4068-9aec-9e57fefca5db\") " pod="openshift-controller-manager/controller-manager-879f6c89f-kwfhg" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.709842 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9bb1952d-0e55-4e45-81c0-69a8cddb827e-client-ca\") pod \"route-controller-manager-6576b87f9c-4h9q7\" (UID: \"9bb1952d-0e55-4e45-81c0-69a8cddb827e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4h9q7" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.709865 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/3661f6b2-1d5e-42f0-947f-a32b66258a1f-audit-policies\") pod \"oauth-openshift-558db77b4-ng4kz\" (UID: \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\") " pod="openshift-authentication/oauth-openshift-558db77b4-ng4kz" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.711415 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/814c6dec-6e96-4068-9aec-9e57fefca5db-client-ca\") pod \"controller-manager-879f6c89f-kwfhg\" (UID: \"814c6dec-6e96-4068-9aec-9e57fefca5db\") " pod="openshift-controller-manager/controller-manager-879f6c89f-kwfhg" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.711831 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-ng4kz\" (UID: \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\") " pod="openshift-authentication/oauth-openshift-558db77b4-ng4kz" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.711858 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a3baeb4f-988d-4d42-b469-a310ca127745-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-rnh5t\" (UID: \"a3baeb4f-988d-4d42-b469-a310ca127745\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-rnh5t" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.711961 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-wjrms"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.712390 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-ng4kz\" (UID: \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\") " pod="openshift-authentication/oauth-openshift-558db77b4-ng4kz" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.719819 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-kwfhg"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.720495 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/005dfe54-e402-4aa4-8b67-f7e7b685debd-config\") pod \"machine-approver-56656f9798-52vhb\" (UID: \"005dfe54-e402-4aa4-8b67-f7e7b685debd\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-52vhb" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.721260 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-5stnc"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.721362 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/5a9c889b-5954-4d3a-ae72-12000ee1cc25-etcd-serving-ca\") pod \"apiserver-76f77b778f-r86r4\" (UID: \"5a9c889b-5954-4d3a-ae72-12000ee1cc25\") " pod="openshift-apiserver/apiserver-76f77b778f-r86r4" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.721852 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-5stnc" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.730683 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pk65f"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.731967 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b75dbfe3-8887-4a0b-9541-a4e4000924cb-config\") pod \"machine-api-operator-5694c8668f-prlhd\" (UID: \"b75dbfe3-8887-4a0b-9541-a4e4000924cb\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-prlhd" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.731875 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a3baeb4f-988d-4d42-b469-a310ca127745-serving-cert\") pod \"authentication-operator-69f744f599-rnh5t\" (UID: \"a3baeb4f-988d-4d42-b469-a310ca127745\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-rnh5t" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.732214 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/5a9c889b-5954-4d3a-ae72-12000ee1cc25-encryption-config\") pod \"apiserver-76f77b778f-r86r4\" (UID: \"5a9c889b-5954-4d3a-ae72-12000ee1cc25\") " pod="openshift-apiserver/apiserver-76f77b778f-r86r4" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.732436 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1be912b5-9d1d-48bd-a0ab-7f601f9eb9da-etcd-client\") pod \"apiserver-7bbb656c7d-ksgr6\" (UID: \"1be912b5-9d1d-48bd-a0ab-7f601f9eb9da\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ksgr6" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.732494 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/5a9c889b-5954-4d3a-ae72-12000ee1cc25-node-pullsecrets\") pod \"apiserver-76f77b778f-r86r4\" (UID: \"5a9c889b-5954-4d3a-ae72-12000ee1cc25\") " pod="openshift-apiserver/apiserver-76f77b778f-r86r4" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.734483 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5a9c889b-5954-4d3a-ae72-12000ee1cc25-config\") pod \"apiserver-76f77b778f-r86r4\" (UID: \"5a9c889b-5954-4d3a-ae72-12000ee1cc25\") " pod="openshift-apiserver/apiserver-76f77b778f-r86r4" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.707446 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.735132 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/b75dbfe3-8887-4a0b-9541-a4e4000924cb-images\") pod \"machine-api-operator-5694c8668f-prlhd\" (UID: \"b75dbfe3-8887-4a0b-9541-a4e4000924cb\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-prlhd" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.736460 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/814c6dec-6e96-4068-9aec-9e57fefca5db-config\") pod \"controller-manager-879f6c89f-kwfhg\" (UID: \"814c6dec-6e96-4068-9aec-9e57fefca5db\") " pod="openshift-controller-manager/controller-manager-879f6c89f-kwfhg" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.737104 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1be912b5-9d1d-48bd-a0ab-7f601f9eb9da-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-ksgr6\" (UID: \"1be912b5-9d1d-48bd-a0ab-7f601f9eb9da\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ksgr6" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.707515 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5a9c889b-5954-4d3a-ae72-12000ee1cc25-audit-dir\") pod \"apiserver-76f77b778f-r86r4\" (UID: \"5a9c889b-5954-4d3a-ae72-12000ee1cc25\") " pod="openshift-apiserver/apiserver-76f77b778f-r86r4" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.737944 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.738278 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.738422 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5a9c889b-5954-4d3a-ae72-12000ee1cc25-trusted-ca-bundle\") pod \"apiserver-76f77b778f-r86r4\" (UID: \"5a9c889b-5954-4d3a-ae72-12000ee1cc25\") " pod="openshift-apiserver/apiserver-76f77b778f-r86r4" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.739014 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1be912b5-9d1d-48bd-a0ab-7f601f9eb9da-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-ksgr6\" (UID: \"1be912b5-9d1d-48bd-a0ab-7f601f9eb9da\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ksgr6" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.740002 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5a9c889b-5954-4d3a-ae72-12000ee1cc25-trusted-ca-bundle\") pod \"apiserver-76f77b778f-r86r4\" (UID: \"5a9c889b-5954-4d3a-ae72-12000ee1cc25\") " pod="openshift-apiserver/apiserver-76f77b778f-r86r4" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.740061 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gn5jt\" (UniqueName: \"kubernetes.io/projected/814c6dec-6e96-4068-9aec-9e57fefca5db-kube-api-access-gn5jt\") pod \"controller-manager-879f6c89f-kwfhg\" (UID: \"814c6dec-6e96-4068-9aec-9e57fefca5db\") " pod="openshift-controller-manager/controller-manager-879f6c89f-kwfhg" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.740122 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/1be912b5-9d1d-48bd-a0ab-7f601f9eb9da-audit-policies\") pod \"apiserver-7bbb656c7d-ksgr6\" (UID: \"1be912b5-9d1d-48bd-a0ab-7f601f9eb9da\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ksgr6" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.740170 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/963f6709-e807-40bb-aa9e-9bba24e82b71-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-v9m24\" (UID: \"963f6709-e807-40bb-aa9e-9bba24e82b71\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-v9m24" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.740639 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/1be912b5-9d1d-48bd-a0ab-7f601f9eb9da-audit-policies\") pod \"apiserver-7bbb656c7d-ksgr6\" (UID: \"1be912b5-9d1d-48bd-a0ab-7f601f9eb9da\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ksgr6" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.740701 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/5a9c889b-5954-4d3a-ae72-12000ee1cc25-image-import-ca\") pod \"apiserver-76f77b778f-r86r4\" (UID: \"5a9c889b-5954-4d3a-ae72-12000ee1cc25\") " pod="openshift-apiserver/apiserver-76f77b778f-r86r4" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.740727 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dslzm\" (UniqueName: \"kubernetes.io/projected/ae320273-06e8-43c5-a64f-acd80ad16d1c-kube-api-access-dslzm\") pod \"ingress-operator-5b745b69d9-5szhc\" (UID: \"ae320273-06e8-43c5-a64f-acd80ad16d1c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5szhc" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.741008 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/814c6dec-6e96-4068-9aec-9e57fefca5db-serving-cert\") pod \"controller-manager-879f6c89f-kwfhg\" (UID: \"814c6dec-6e96-4068-9aec-9e57fefca5db\") " pod="openshift-controller-manager/controller-manager-879f6c89f-kwfhg" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.741320 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-ng4kz\" (UID: \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\") " pod="openshift-authentication/oauth-openshift-558db77b4-ng4kz" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.741606 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e9d44a1e-0eab-4bd6-833f-143436e43758-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-hh9sk\" (UID: \"e9d44a1e-0eab-4bd6-833f-143436e43758\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-hh9sk" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.741665 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e9d44a1e-0eab-4bd6-833f-143436e43758-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-hh9sk\" (UID: \"e9d44a1e-0eab-4bd6-833f-143436e43758\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-hh9sk" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.741736 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/005dfe54-e402-4aa4-8b67-f7e7b685debd-machine-approver-tls\") pod \"machine-approver-56656f9798-52vhb\" (UID: \"005dfe54-e402-4aa4-8b67-f7e7b685debd\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-52vhb" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.742077 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a3baeb4f-988d-4d42-b469-a310ca127745-service-ca-bundle\") pod \"authentication-operator-69f744f599-rnh5t\" (UID: \"a3baeb4f-988d-4d42-b469-a310ca127745\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-rnh5t" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.742154 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1be912b5-9d1d-48bd-a0ab-7f601f9eb9da-serving-cert\") pod \"apiserver-7bbb656c7d-ksgr6\" (UID: \"1be912b5-9d1d-48bd-a0ab-7f601f9eb9da\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ksgr6" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.742419 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hkbqt"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.742722 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a3baeb4f-988d-4d42-b469-a310ca127745-service-ca-bundle\") pod \"authentication-operator-69f744f599-rnh5t\" (UID: \"a3baeb4f-988d-4d42-b469-a310ca127745\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-rnh5t" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.742839 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fwcwx"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.743227 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ccjsv"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.743605 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/b75dbfe3-8887-4a0b-9541-a4e4000924cb-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-prlhd\" (UID: \"b75dbfe3-8887-4a0b-9541-a4e4000924cb\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-prlhd" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.743868 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ccjsv" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.744181 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pk65f" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.744356 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hkbqt" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.744409 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/5a9c889b-5954-4d3a-ae72-12000ee1cc25-image-import-ca\") pod \"apiserver-76f77b778f-r86r4\" (UID: \"5a9c889b-5954-4d3a-ae72-12000ee1cc25\") " pod="openshift-apiserver/apiserver-76f77b778f-r86r4" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.744539 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fwcwx" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.744786 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.745071 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9bb1952d-0e55-4e45-81c0-69a8cddb827e-config\") pod \"route-controller-manager-6576b87f9c-4h9q7\" (UID: \"9bb1952d-0e55-4e45-81c0-69a8cddb827e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4h9q7" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.745131 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a3baeb4f-988d-4d42-b469-a310ca127745-config\") pod \"authentication-operator-69f744f599-rnh5t\" (UID: \"a3baeb4f-988d-4d42-b469-a310ca127745\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-rnh5t" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.745153 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/5a9c889b-5954-4d3a-ae72-12000ee1cc25-audit\") pod \"apiserver-76f77b778f-r86r4\" (UID: \"5a9c889b-5954-4d3a-ae72-12000ee1cc25\") " pod="openshift-apiserver/apiserver-76f77b778f-r86r4" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.745201 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-ng4kz\" (UID: \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\") " pod="openshift-authentication/oauth-openshift-558db77b4-ng4kz" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.745700 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/5a9c889b-5954-4d3a-ae72-12000ee1cc25-audit\") pod \"apiserver-76f77b778f-r86r4\" (UID: \"5a9c889b-5954-4d3a-ae72-12000ee1cc25\") " pod="openshift-apiserver/apiserver-76f77b778f-r86r4" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.746616 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9bb1952d-0e55-4e45-81c0-69a8cddb827e-config\") pod \"route-controller-manager-6576b87f9c-4h9q7\" (UID: \"9bb1952d-0e55-4e45-81c0-69a8cddb827e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4h9q7" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.746672 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-ng4kz\" (UID: \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\") " pod="openshift-authentication/oauth-openshift-558db77b4-ng4kz" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.746732 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/5a9c889b-5954-4d3a-ae72-12000ee1cc25-etcd-client\") pod \"apiserver-76f77b778f-r86r4\" (UID: \"5a9c889b-5954-4d3a-ae72-12000ee1cc25\") " pod="openshift-apiserver/apiserver-76f77b778f-r86r4" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.746790 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/1be912b5-9d1d-48bd-a0ab-7f601f9eb9da-audit-dir\") pod \"apiserver-7bbb656c7d-ksgr6\" (UID: \"1be912b5-9d1d-48bd-a0ab-7f601f9eb9da\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ksgr6" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.746816 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3661f6b2-1d5e-42f0-947f-a32b66258a1f-audit-dir\") pod \"oauth-openshift-558db77b4-ng4kz\" (UID: \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\") " pod="openshift-authentication/oauth-openshift-558db77b4-ng4kz" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.747054 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3661f6b2-1d5e-42f0-947f-a32b66258a1f-audit-dir\") pod \"oauth-openshift-558db77b4-ng4kz\" (UID: \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\") " pod="openshift-authentication/oauth-openshift-558db77b4-ng4kz" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.747119 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/1be912b5-9d1d-48bd-a0ab-7f601f9eb9da-audit-dir\") pod \"apiserver-7bbb656c7d-ksgr6\" (UID: \"1be912b5-9d1d-48bd-a0ab-7f601f9eb9da\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ksgr6" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.747176 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7jwmk\" (UniqueName: \"kubernetes.io/projected/b75dbfe3-8887-4a0b-9541-a4e4000924cb-kube-api-access-7jwmk\") pod \"machine-api-operator-5694c8668f-prlhd\" (UID: \"b75dbfe3-8887-4a0b-9541-a4e4000924cb\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-prlhd" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.747220 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9bb1952d-0e55-4e45-81c0-69a8cddb827e-serving-cert\") pod \"route-controller-manager-6576b87f9c-4h9q7\" (UID: \"9bb1952d-0e55-4e45-81c0-69a8cddb827e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4h9q7" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.747262 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-ng4kz\" (UID: \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\") " pod="openshift-authentication/oauth-openshift-558db77b4-ng4kz" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.747314 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jb574\" (UniqueName: \"kubernetes.io/projected/a8e9ea4a-43ad-4da0-99fa-c01eea712c8c-kube-api-access-jb574\") pod \"machine-config-controller-84d6567774-t68ql\" (UID: \"a8e9ea4a-43ad-4da0-99fa-c01eea712c8c\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-t68ql" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.747357 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/814c6dec-6e96-4068-9aec-9e57fefca5db-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-kwfhg\" (UID: \"814c6dec-6e96-4068-9aec-9e57fefca5db\") " pod="openshift-controller-manager/controller-manager-879f6c89f-kwfhg" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.747413 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/a8e9ea4a-43ad-4da0-99fa-c01eea712c8c-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-t68ql\" (UID: \"a8e9ea4a-43ad-4da0-99fa-c01eea712c8c\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-t68ql" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.747439 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/963f6709-e807-40bb-aa9e-9bba24e82b71-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-v9m24\" (UID: \"963f6709-e807-40bb-aa9e-9bba24e82b71\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-v9m24" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.749266 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/814c6dec-6e96-4068-9aec-9e57fefca5db-serving-cert\") pod \"controller-manager-879f6c89f-kwfhg\" (UID: \"814c6dec-6e96-4068-9aec-9e57fefca5db\") " pod="openshift-controller-manager/controller-manager-879f6c89f-kwfhg" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.750494 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/814c6dec-6e96-4068-9aec-9e57fefca5db-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-kwfhg\" (UID: \"814c6dec-6e96-4068-9aec-9e57fefca5db\") " pod="openshift-controller-manager/controller-manager-879f6c89f-kwfhg" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.758488 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-75gdt"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.760089 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556840-pgmwj"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.760114 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.760579 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556840-pgmwj" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.760822 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-75gdt" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.761909 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a3baeb4f-988d-4d42-b469-a310ca127745-config\") pod \"authentication-operator-69f744f599-rnh5t\" (UID: \"a3baeb4f-988d-4d42-b469-a310ca127745\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-rnh5t" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.761941 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-ng4kz\" (UID: \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\") " pod="openshift-authentication/oauth-openshift-558db77b4-ng4kz" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.776996 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/5a9c889b-5954-4d3a-ae72-12000ee1cc25-etcd-client\") pod \"apiserver-76f77b778f-r86r4\" (UID: \"5a9c889b-5954-4d3a-ae72-12000ee1cc25\") " pod="openshift-apiserver/apiserver-76f77b778f-r86r4" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.777134 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-ng4kz\" (UID: \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\") " pod="openshift-authentication/oauth-openshift-558db77b4-ng4kz" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.777502 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/005dfe54-e402-4aa4-8b67-f7e7b685debd-machine-approver-tls\") pod \"machine-approver-56656f9798-52vhb\" (UID: \"005dfe54-e402-4aa4-8b67-f7e7b685debd\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-52vhb" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.778224 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1be912b5-9d1d-48bd-a0ab-7f601f9eb9da-encryption-config\") pod \"apiserver-7bbb656c7d-ksgr6\" (UID: \"1be912b5-9d1d-48bd-a0ab-7f601f9eb9da\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ksgr6" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.778837 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-ng4kz\" (UID: \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\") " pod="openshift-authentication/oauth-openshift-558db77b4-ng4kz" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.779108 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-ng4kz\" (UID: \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\") " pod="openshift-authentication/oauth-openshift-558db77b4-ng4kz" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.779469 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5a9c889b-5954-4d3a-ae72-12000ee1cc25-serving-cert\") pod \"apiserver-76f77b778f-r86r4\" (UID: \"5a9c889b-5954-4d3a-ae72-12000ee1cc25\") " pod="openshift-apiserver/apiserver-76f77b778f-r86r4" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.779617 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.779939 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-ng4kz\" (UID: \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\") " pod="openshift-authentication/oauth-openshift-558db77b4-ng4kz" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.780266 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-ng4kz\" (UID: \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\") " pod="openshift-authentication/oauth-openshift-558db77b4-ng4kz" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.780665 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1be912b5-9d1d-48bd-a0ab-7f601f9eb9da-serving-cert\") pod \"apiserver-7bbb656c7d-ksgr6\" (UID: \"1be912b5-9d1d-48bd-a0ab-7f601f9eb9da\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ksgr6" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.781110 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9bb1952d-0e55-4e45-81c0-69a8cddb827e-serving-cert\") pod \"route-controller-manager-6576b87f9c-4h9q7\" (UID: \"9bb1952d-0e55-4e45-81c0-69a8cddb827e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4h9q7" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.781241 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556848-v8kv4"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.781355 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-ng4kz\" (UID: \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\") " pod="openshift-authentication/oauth-openshift-558db77b4-ng4kz" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.787477 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.788653 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556848-v8kv4" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.789248 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/71c234e9-5f3d-4b0c-94d4-80b4e33a6dda-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-wjrms\" (UID: \"71c234e9-5f3d-4b0c-94d4-80b4e33a6dda\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-wjrms" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.790621 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.790819 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-zwzc4"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.791770 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-zwzc4" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.794714 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xnr2b"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.796440 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xnr2b" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.797672 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-ng4kz\" (UID: \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\") " pod="openshift-authentication/oauth-openshift-558db77b4-ng4kz" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.799528 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.803452 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-w76kw"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.804817 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-4h9q7"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.805893 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hfcsw"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.810698 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-zz77x"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.810738 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-ng4kz"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.813220 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-g6q2r"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.814511 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-v9m24"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.814540 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-rmkfg"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.814687 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-g6q2r" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.815662 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-544df"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.816367 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.822567 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-hh9sk"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.822630 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-56rqr"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.824674 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-r86r4"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.827236 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hkbqt"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.829025 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-prlhd"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.830986 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-djph7"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.832469 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-l9xpb"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.835068 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.835222 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-xkdjh"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.836808 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-hwf6r"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.837993 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-92gs7"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.838951 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-252nq"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.839791 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-252nq" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.840027 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-sjwvn"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.841147 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-t68ql"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.841276 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-sjwvn" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.842320 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-ksfb8"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.844457 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-wgpdd"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.846798 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-75gdt"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.848112 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a7b66665-221d-4334-8263-aeb3b20e1c92-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-hfcsw\" (UID: \"a7b66665-221d-4334-8263-aeb3b20e1c92\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hfcsw" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.848155 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/963f6709-e807-40bb-aa9e-9bba24e82b71-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-v9m24\" (UID: \"963f6709-e807-40bb-aa9e-9bba24e82b71\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-v9m24" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.848176 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dslzm\" (UniqueName: \"kubernetes.io/projected/ae320273-06e8-43c5-a64f-acd80ad16d1c-kube-api-access-dslzm\") pod \"ingress-operator-5b745b69d9-5szhc\" (UID: \"ae320273-06e8-43c5-a64f-acd80ad16d1c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5szhc" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.848210 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e9d44a1e-0eab-4bd6-833f-143436e43758-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-hh9sk\" (UID: \"e9d44a1e-0eab-4bd6-833f-143436e43758\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-hh9sk" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.848226 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e9d44a1e-0eab-4bd6-833f-143436e43758-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-hh9sk\" (UID: \"e9d44a1e-0eab-4bd6-833f-143436e43758\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-hh9sk" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.848255 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jb574\" (UniqueName: \"kubernetes.io/projected/a8e9ea4a-43ad-4da0-99fa-c01eea712c8c-kube-api-access-jb574\") pod \"machine-config-controller-84d6567774-t68ql\" (UID: \"a8e9ea4a-43ad-4da0-99fa-c01eea712c8c\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-t68ql" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.848278 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/a8e9ea4a-43ad-4da0-99fa-c01eea712c8c-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-t68ql\" (UID: \"a8e9ea4a-43ad-4da0-99fa-c01eea712c8c\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-t68ql" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.848296 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/963f6709-e807-40bb-aa9e-9bba24e82b71-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-v9m24\" (UID: \"963f6709-e807-40bb-aa9e-9bba24e82b71\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-v9m24" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.848315 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/4c8c0266-5452-4abb-92c5-c536df94dd41-console-serving-cert\") pod \"console-f9d7485db-l9xpb\" (UID: \"4c8c0266-5452-4abb-92c5-c536df94dd41\") " pod="openshift-console/console-f9d7485db-l9xpb" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.848330 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e9d44a1e-0eab-4bd6-833f-143436e43758-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-hh9sk\" (UID: \"e9d44a1e-0eab-4bd6-833f-143436e43758\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-hh9sk" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.848345 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/4c8c0266-5452-4abb-92c5-c536df94dd41-console-config\") pod \"console-f9d7485db-l9xpb\" (UID: \"4c8c0266-5452-4abb-92c5-c536df94dd41\") " pod="openshift-console/console-f9d7485db-l9xpb" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.848366 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/4c8c0266-5452-4abb-92c5-c536df94dd41-service-ca\") pod \"console-f9d7485db-l9xpb\" (UID: \"4c8c0266-5452-4abb-92c5-c536df94dd41\") " pod="openshift-console/console-f9d7485db-l9xpb" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.848382 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/963f6709-e807-40bb-aa9e-9bba24e82b71-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-v9m24\" (UID: \"963f6709-e807-40bb-aa9e-9bba24e82b71\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-v9m24" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.848411 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ae320273-06e8-43c5-a64f-acd80ad16d1c-bound-sa-token\") pod \"ingress-operator-5b745b69d9-5szhc\" (UID: \"ae320273-06e8-43c5-a64f-acd80ad16d1c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5szhc" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.848434 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/ae320273-06e8-43c5-a64f-acd80ad16d1c-metrics-tls\") pod \"ingress-operator-5b745b69d9-5szhc\" (UID: \"ae320273-06e8-43c5-a64f-acd80ad16d1c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5szhc" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.848455 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/a8e9ea4a-43ad-4da0-99fa-c01eea712c8c-proxy-tls\") pod \"machine-config-controller-84d6567774-t68ql\" (UID: \"a8e9ea4a-43ad-4da0-99fa-c01eea712c8c\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-t68ql" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.848469 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a7b66665-221d-4334-8263-aeb3b20e1c92-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-hfcsw\" (UID: \"a7b66665-221d-4334-8263-aeb3b20e1c92\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hfcsw" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.848485 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rnmbh\" (UniqueName: \"kubernetes.io/projected/a7b66665-221d-4334-8263-aeb3b20e1c92-kube-api-access-rnmbh\") pod \"openshift-controller-manager-operator-756b6f6bc6-hfcsw\" (UID: \"a7b66665-221d-4334-8263-aeb3b20e1c92\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hfcsw" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.848502 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7h5xn\" (UniqueName: \"kubernetes.io/projected/963f6709-e807-40bb-aa9e-9bba24e82b71-kube-api-access-7h5xn\") pod \"cluster-image-registry-operator-dc59b4c8b-v9m24\" (UID: \"963f6709-e807-40bb-aa9e-9bba24e82b71\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-v9m24" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.848525 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/4c8c0266-5452-4abb-92c5-c536df94dd41-console-oauth-config\") pod \"console-f9d7485db-l9xpb\" (UID: \"4c8c0266-5452-4abb-92c5-c536df94dd41\") " pod="openshift-console/console-f9d7485db-l9xpb" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.848540 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4c8c0266-5452-4abb-92c5-c536df94dd41-trusted-ca-bundle\") pod \"console-f9d7485db-l9xpb\" (UID: \"4c8c0266-5452-4abb-92c5-c536df94dd41\") " pod="openshift-console/console-f9d7485db-l9xpb" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.848563 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ae320273-06e8-43c5-a64f-acd80ad16d1c-trusted-ca\") pod \"ingress-operator-5b745b69d9-5szhc\" (UID: \"ae320273-06e8-43c5-a64f-acd80ad16d1c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5szhc" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.848581 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vxlx4\" (UniqueName: \"kubernetes.io/projected/4c8c0266-5452-4abb-92c5-c536df94dd41-kube-api-access-vxlx4\") pod \"console-f9d7485db-l9xpb\" (UID: \"4c8c0266-5452-4abb-92c5-c536df94dd41\") " pod="openshift-console/console-f9d7485db-l9xpb" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.848597 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/4c8c0266-5452-4abb-92c5-c536df94dd41-oauth-serving-cert\") pod \"console-f9d7485db-l9xpb\" (UID: \"4c8c0266-5452-4abb-92c5-c536df94dd41\") " pod="openshift-console/console-f9d7485db-l9xpb" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.849066 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xnr2b"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.849463 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-8jqgq"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.849647 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a7b66665-221d-4334-8263-aeb3b20e1c92-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-hfcsw\" (UID: \"a7b66665-221d-4334-8263-aeb3b20e1c92\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hfcsw" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.850283 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-8jqgq" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.850465 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/4c8c0266-5452-4abb-92c5-c536df94dd41-oauth-serving-cert\") pod \"console-f9d7485db-l9xpb\" (UID: \"4c8c0266-5452-4abb-92c5-c536df94dd41\") " pod="openshift-console/console-f9d7485db-l9xpb" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.850792 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/963f6709-e807-40bb-aa9e-9bba24e82b71-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-v9m24\" (UID: \"963f6709-e807-40bb-aa9e-9bba24e82b71\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-v9m24" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.851349 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/4c8c0266-5452-4abb-92c5-c536df94dd41-console-config\") pod \"console-f9d7485db-l9xpb\" (UID: \"4c8c0266-5452-4abb-92c5-c536df94dd41\") " pod="openshift-console/console-f9d7485db-l9xpb" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.851450 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/4c8c0266-5452-4abb-92c5-c536df94dd41-service-ca\") pod \"console-f9d7485db-l9xpb\" (UID: \"4c8c0266-5452-4abb-92c5-c536df94dd41\") " pod="openshift-console/console-f9d7485db-l9xpb" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.851521 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/a8e9ea4a-43ad-4da0-99fa-c01eea712c8c-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-t68ql\" (UID: \"a8e9ea4a-43ad-4da0-99fa-c01eea712c8c\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-t68ql" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.851582 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pk65f"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.853070 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4c8c0266-5452-4abb-92c5-c536df94dd41-trusted-ca-bundle\") pod \"console-f9d7485db-l9xpb\" (UID: \"4c8c0266-5452-4abb-92c5-c536df94dd41\") " pod="openshift-console/console-f9d7485db-l9xpb" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.853116 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-5szhc"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.853512 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/963f6709-e807-40bb-aa9e-9bba24e82b71-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-v9m24\" (UID: \"963f6709-e807-40bb-aa9e-9bba24e82b71\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-v9m24" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.854245 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a7b66665-221d-4334-8263-aeb3b20e1c92-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-hfcsw\" (UID: \"a7b66665-221d-4334-8263-aeb3b20e1c92\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hfcsw" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.854519 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/4c8c0266-5452-4abb-92c5-c536df94dd41-console-serving-cert\") pod \"console-f9d7485db-l9xpb\" (UID: \"4c8c0266-5452-4abb-92c5-c536df94dd41\") " pod="openshift-console/console-f9d7485db-l9xpb" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.854668 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-252nq"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.855150 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.856159 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-q9z8k"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.857115 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ccjsv"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.857701 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/4c8c0266-5452-4abb-92c5-c536df94dd41-console-oauth-config\") pod \"console-f9d7485db-l9xpb\" (UID: \"4c8c0266-5452-4abb-92c5-c536df94dd41\") " pod="openshift-console/console-f9d7485db-l9xpb" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.858330 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fwcwx"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.860128 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-vfbsx"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.861331 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556840-pgmwj"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.862834 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-5stnc"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.864697 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-8jqgq"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.865975 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-sjwvn"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.867261 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-zwzc4"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.868526 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556848-v8kv4"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.875429 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.895915 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.915543 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.935423 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.955641 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.975549 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.996666 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Mar 13 14:09:01 crc kubenswrapper[4907]: I0313 14:09:01.016857 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Mar 13 14:09:01 crc kubenswrapper[4907]: I0313 14:09:01.035645 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Mar 13 14:09:01 crc kubenswrapper[4907]: I0313 14:09:01.056468 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Mar 13 14:09:01 crc kubenswrapper[4907]: I0313 14:09:01.076181 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Mar 13 14:09:01 crc kubenswrapper[4907]: I0313 14:09:01.096596 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Mar 13 14:09:01 crc kubenswrapper[4907]: I0313 14:09:01.115944 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Mar 13 14:09:01 crc kubenswrapper[4907]: I0313 14:09:01.135828 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Mar 13 14:09:01 crc kubenswrapper[4907]: I0313 14:09:01.143074 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/ae320273-06e8-43c5-a64f-acd80ad16d1c-metrics-tls\") pod \"ingress-operator-5b745b69d9-5szhc\" (UID: \"ae320273-06e8-43c5-a64f-acd80ad16d1c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5szhc" Mar 13 14:09:01 crc kubenswrapper[4907]: I0313 14:09:01.160792 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Mar 13 14:09:01 crc kubenswrapper[4907]: I0313 14:09:01.172818 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ae320273-06e8-43c5-a64f-acd80ad16d1c-trusted-ca\") pod \"ingress-operator-5b745b69d9-5szhc\" (UID: \"ae320273-06e8-43c5-a64f-acd80ad16d1c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5szhc" Mar 13 14:09:01 crc kubenswrapper[4907]: I0313 14:09:01.175533 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Mar 13 14:09:01 crc kubenswrapper[4907]: I0313 14:09:01.215926 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Mar 13 14:09:01 crc kubenswrapper[4907]: I0313 14:09:01.236151 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Mar 13 14:09:01 crc kubenswrapper[4907]: I0313 14:09:01.256841 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Mar 13 14:09:01 crc kubenswrapper[4907]: I0313 14:09:01.276221 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Mar 13 14:09:01 crc kubenswrapper[4907]: I0313 14:09:01.296707 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Mar 13 14:09:01 crc kubenswrapper[4907]: I0313 14:09:01.316145 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Mar 13 14:09:01 crc kubenswrapper[4907]: I0313 14:09:01.336733 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Mar 13 14:09:01 crc kubenswrapper[4907]: I0313 14:09:01.355564 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Mar 13 14:09:01 crc kubenswrapper[4907]: I0313 14:09:01.377095 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Mar 13 14:09:01 crc kubenswrapper[4907]: I0313 14:09:01.385435 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/a8e9ea4a-43ad-4da0-99fa-c01eea712c8c-proxy-tls\") pod \"machine-config-controller-84d6567774-t68ql\" (UID: \"a8e9ea4a-43ad-4da0-99fa-c01eea712c8c\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-t68ql" Mar 13 14:09:01 crc kubenswrapper[4907]: I0313 14:09:01.396311 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Mar 13 14:09:01 crc kubenswrapper[4907]: I0313 14:09:01.436708 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Mar 13 14:09:01 crc kubenswrapper[4907]: I0313 14:09:01.456110 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Mar 13 14:09:01 crc kubenswrapper[4907]: I0313 14:09:01.476261 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Mar 13 14:09:01 crc kubenswrapper[4907]: I0313 14:09:01.497051 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Mar 13 14:09:01 crc kubenswrapper[4907]: I0313 14:09:01.516031 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Mar 13 14:09:01 crc kubenswrapper[4907]: I0313 14:09:01.536141 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Mar 13 14:09:01 crc kubenswrapper[4907]: I0313 14:09:01.556680 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Mar 13 14:09:01 crc kubenswrapper[4907]: I0313 14:09:01.576914 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Mar 13 14:09:01 crc kubenswrapper[4907]: I0313 14:09:01.583219 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e9d44a1e-0eab-4bd6-833f-143436e43758-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-hh9sk\" (UID: \"e9d44a1e-0eab-4bd6-833f-143436e43758\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-hh9sk" Mar 13 14:09:01 crc kubenswrapper[4907]: I0313 14:09:01.596819 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Mar 13 14:09:01 crc kubenswrapper[4907]: I0313 14:09:01.600109 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e9d44a1e-0eab-4bd6-833f-143436e43758-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-hh9sk\" (UID: \"e9d44a1e-0eab-4bd6-833f-143436e43758\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-hh9sk" Mar 13 14:09:01 crc kubenswrapper[4907]: I0313 14:09:01.616375 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Mar 13 14:09:01 crc kubenswrapper[4907]: I0313 14:09:01.636366 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Mar 13 14:09:01 crc kubenswrapper[4907]: I0313 14:09:01.657948 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Mar 13 14:09:01 crc kubenswrapper[4907]: I0313 14:09:01.698372 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Mar 13 14:09:01 crc kubenswrapper[4907]: I0313 14:09:01.703525 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-289bg\" (UniqueName: \"kubernetes.io/projected/a3baeb4f-988d-4d42-b469-a310ca127745-kube-api-access-289bg\") pod \"authentication-operator-69f744f599-rnh5t\" (UID: \"a3baeb4f-988d-4d42-b469-a310ca127745\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-rnh5t" Mar 13 14:09:01 crc kubenswrapper[4907]: I0313 14:09:01.714623 4907 request.go:700] Waited for 1.005964719s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-api/secrets?fieldSelector=metadata.name%3Dcontrol-plane-machine-set-operator-dockercfg-k9rxt&limit=500&resourceVersion=0 Mar 13 14:09:01 crc kubenswrapper[4907]: I0313 14:09:01.717587 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Mar 13 14:09:01 crc kubenswrapper[4907]: I0313 14:09:01.796719 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Mar 13 14:09:01 crc kubenswrapper[4907]: I0313 14:09:01.816476 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Mar 13 14:09:01 crc kubenswrapper[4907]: I0313 14:09:01.836471 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Mar 13 14:09:01 crc kubenswrapper[4907]: I0313 14:09:01.862865 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Mar 13 14:09:01 crc kubenswrapper[4907]: I0313 14:09:01.876670 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Mar 13 14:09:01 crc kubenswrapper[4907]: I0313 14:09:01.975368 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Mar 13 14:09:01 crc kubenswrapper[4907]: I0313 14:09:01.996449 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Mar 13 14:09:02 crc kubenswrapper[4907]: I0313 14:09:02.016553 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Mar 13 14:09:02 crc kubenswrapper[4907]: I0313 14:09:02.035583 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Mar 13 14:09:02 crc kubenswrapper[4907]: I0313 14:09:02.056989 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Mar 13 14:09:02 crc kubenswrapper[4907]: I0313 14:09:02.076380 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Mar 13 14:09:02 crc kubenswrapper[4907]: I0313 14:09:02.096300 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Mar 13 14:09:02 crc kubenswrapper[4907]: I0313 14:09:02.116468 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Mar 13 14:09:02 crc kubenswrapper[4907]: I0313 14:09:02.156329 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 13 14:09:02 crc kubenswrapper[4907]: I0313 14:09:02.176566 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 13 14:09:02 crc kubenswrapper[4907]: I0313 14:09:02.196280 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Mar 13 14:09:02 crc kubenswrapper[4907]: I0313 14:09:02.216117 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Mar 13 14:09:02 crc kubenswrapper[4907]: I0313 14:09:02.235066 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Mar 13 14:09:02 crc kubenswrapper[4907]: I0313 14:09:02.255923 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Mar 13 14:09:02 crc kubenswrapper[4907]: I0313 14:09:02.276100 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Mar 13 14:09:02 crc kubenswrapper[4907]: I0313 14:09:02.295417 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 14:09:02 crc kubenswrapper[4907]: I0313 14:09:02.316322 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 14:09:02 crc kubenswrapper[4907]: I0313 14:09:02.335970 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Mar 13 14:09:02 crc kubenswrapper[4907]: I0313 14:09:02.356345 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Mar 13 14:09:02 crc kubenswrapper[4907]: I0313 14:09:02.375269 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Mar 13 14:09:02 crc kubenswrapper[4907]: I0313 14:09:02.395774 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Mar 13 14:09:02 crc kubenswrapper[4907]: I0313 14:09:02.416266 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Mar 13 14:09:02 crc kubenswrapper[4907]: I0313 14:09:02.435619 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Mar 13 14:09:02 crc kubenswrapper[4907]: I0313 14:09:02.456275 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Mar 13 14:09:02 crc kubenswrapper[4907]: I0313 14:09:02.476434 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Mar 13 14:09:02 crc kubenswrapper[4907]: I0313 14:09:02.496248 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Mar 13 14:09:02 crc kubenswrapper[4907]: I0313 14:09:02.516810 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Mar 13 14:09:02 crc kubenswrapper[4907]: I0313 14:09:02.536433 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Mar 13 14:09:02 crc kubenswrapper[4907]: I0313 14:09:02.556975 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Mar 13 14:09:02 crc kubenswrapper[4907]: I0313 14:09:02.577173 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Mar 13 14:09:02 crc kubenswrapper[4907]: I0313 14:09:02.595836 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Mar 13 14:09:02 crc kubenswrapper[4907]: I0313 14:09:02.617049 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Mar 13 14:09:02 crc kubenswrapper[4907]: I0313 14:09:02.635453 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Mar 13 14:09:02 crc kubenswrapper[4907]: I0313 14:09:02.656454 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Mar 13 14:09:02 crc kubenswrapper[4907]: I0313 14:09:02.676519 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Mar 13 14:09:02 crc kubenswrapper[4907]: I0313 14:09:02.696515 4907 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Mar 13 14:09:02 crc kubenswrapper[4907]: I0313 14:09:02.715651 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Mar 13 14:09:02 crc kubenswrapper[4907]: I0313 14:09:02.734432 4907 request.go:700] Waited for 1.88512071s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-image-registry/serviceaccounts/cluster-image-registry-operator/token Mar 13 14:09:02 crc kubenswrapper[4907]: E0313 14:09:02.751076 4907 projected.go:288] Couldn't get configMap openshift-oauth-apiserver/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Mar 13 14:09:02 crc kubenswrapper[4907]: I0313 14:09:02.757975 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/963f6709-e807-40bb-aa9e-9bba24e82b71-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-v9m24\" (UID: \"963f6709-e807-40bb-aa9e-9bba24e82b71\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-v9m24" Mar 13 14:09:02 crc kubenswrapper[4907]: E0313 14:09:02.766990 4907 projected.go:288] Couldn't get configMap openshift-authentication/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Mar 13 14:09:02 crc kubenswrapper[4907]: I0313 14:09:02.779571 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ae320273-06e8-43c5-a64f-acd80ad16d1c-bound-sa-token\") pod \"ingress-operator-5b745b69d9-5szhc\" (UID: \"ae320273-06e8-43c5-a64f-acd80ad16d1c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5szhc" Mar 13 14:09:02 crc kubenswrapper[4907]: E0313 14:09:02.789683 4907 projected.go:288] Couldn't get configMap openshift-apiserver/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Mar 13 14:09:02 crc kubenswrapper[4907]: I0313 14:09:02.811739 4907 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openshift-authentication-operator/authentication-operator-69f744f599-rnh5t" secret="" err="failed to sync secret cache: timed out waiting for the condition" Mar 13 14:09:02 crc kubenswrapper[4907]: I0313 14:09:02.811838 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-rnh5t" Mar 13 14:09:02 crc kubenswrapper[4907]: I0313 14:09:02.829974 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jb574\" (UniqueName: \"kubernetes.io/projected/a8e9ea4a-43ad-4da0-99fa-c01eea712c8c-kube-api-access-jb574\") pod \"machine-config-controller-84d6567774-t68ql\" (UID: \"a8e9ea4a-43ad-4da0-99fa-c01eea712c8c\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-t68ql" Mar 13 14:09:02 crc kubenswrapper[4907]: I0313 14:09:02.872640 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7h5xn\" (UniqueName: \"kubernetes.io/projected/963f6709-e807-40bb-aa9e-9bba24e82b71-kube-api-access-7h5xn\") pod \"cluster-image-registry-operator-dc59b4c8b-v9m24\" (UID: \"963f6709-e807-40bb-aa9e-9bba24e82b71\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-v9m24" Mar 13 14:09:02 crc kubenswrapper[4907]: I0313 14:09:02.897339 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Mar 13 14:09:02 crc kubenswrapper[4907]: E0313 14:09:02.911556 4907 projected.go:288] Couldn't get configMap openshift-cluster-samples-operator/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Mar 13 14:09:02 crc kubenswrapper[4907]: I0313 14:09:02.917747 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Mar 13 14:09:02 crc kubenswrapper[4907]: E0313 14:09:02.929139 4907 projected.go:288] Couldn't get configMap openshift-cluster-machine-approver/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Mar 13 14:09:02 crc kubenswrapper[4907]: I0313 14:09:02.937416 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Mar 13 14:09:02 crc kubenswrapper[4907]: E0313 14:09:02.947013 4907 projected.go:288] Couldn't get configMap openshift-route-controller-manager/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Mar 13 14:09:02 crc kubenswrapper[4907]: E0313 14:09:02.969283 4907 projected.go:288] Couldn't get configMap openshift-controller-manager/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Mar 13 14:09:02 crc kubenswrapper[4907]: I0313 14:09:02.996097 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.016627 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.039597 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.040127 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-rnh5t"] Mar 13 14:09:03 crc kubenswrapper[4907]: W0313 14:09:03.047451 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda3baeb4f_988d_4d42_b469_a310ca127745.slice/crio-c8e2efcf239dc9215a99f09f489a8e23320bb5126f2409f2a08ac75994d91c04 WatchSource:0}: Error finding container c8e2efcf239dc9215a99f09f489a8e23320bb5126f2409f2a08ac75994d91c04: Status 404 returned error can't find the container with id c8e2efcf239dc9215a99f09f489a8e23320bb5126f2409f2a08ac75994d91c04 Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.058331 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.075993 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.076439 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/59980734-2cf7-42dc-9f6f-a6de7e1c3665-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-56rqr\" (UID: \"59980734-2cf7-42dc-9f6f-a6de7e1c3665\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-56rqr" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.076499 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c26ccc9f-e682-4b94-9539-cb2607725f71-default-certificate\") pod \"router-default-5444994796-f8mmz\" (UID: \"c26ccc9f-e682-4b94-9539-cb2607725f71\") " pod="openshift-ingress/router-default-5444994796-f8mmz" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.076525 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/699da377-c61b-48f0-92e0-42bd247ee363-config\") pod \"etcd-operator-b45778765-hwf6r\" (UID: \"699da377-c61b-48f0-92e0-42bd247ee363\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hwf6r" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.076543 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/699da377-c61b-48f0-92e0-42bd247ee363-serving-cert\") pod \"etcd-operator-b45778765-hwf6r\" (UID: \"699da377-c61b-48f0-92e0-42bd247ee363\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hwf6r" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.076560 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/10e8bf08-339f-4948-9fe2-ab4092d9a942-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-zz77x\" (UID: \"10e8bf08-339f-4948-9fe2-ab4092d9a942\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-zz77x" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.076576 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0396c426-1d56-44e6-885a-1b97d69ad57a-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-rmkfg\" (UID: \"0396c426-1d56-44e6-885a-1b97d69ad57a\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-rmkfg" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.076651 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/59980734-2cf7-42dc-9f6f-a6de7e1c3665-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-56rqr\" (UID: \"59980734-2cf7-42dc-9f6f-a6de7e1c3665\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-56rqr" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.076705 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wk2bb\" (UniqueName: \"kubernetes.io/projected/ebd59055-333a-46c2-ba30-826296414a4c-kube-api-access-wk2bb\") pod \"dns-operator-744455d44c-xkdjh\" (UID: \"ebd59055-333a-46c2-ba30-826296414a4c\") " pod="openshift-dns-operator/dns-operator-744455d44c-xkdjh" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.076732 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/9c5d91bb-753e-4554-b7a1-6c85fe323875-registry-tls\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.076796 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/699da377-c61b-48f0-92e0-42bd247ee363-etcd-ca\") pod \"etcd-operator-b45778765-hwf6r\" (UID: \"699da377-c61b-48f0-92e0-42bd247ee363\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hwf6r" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.076845 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b75954c-b75d-4f17-82dc-4c0358fd8d0f-config\") pod \"console-operator-58897d9998-w76kw\" (UID: \"9b75954c-b75d-4f17-82dc-4c0358fd8d0f\") " pod="openshift-console-operator/console-operator-58897d9998-w76kw" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.077119 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9b75954c-b75d-4f17-82dc-4c0358fd8d0f-trusted-ca\") pod \"console-operator-58897d9998-w76kw\" (UID: \"9b75954c-b75d-4f17-82dc-4c0358fd8d0f\") " pod="openshift-console-operator/console-operator-58897d9998-w76kw" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.077144 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c26ccc9f-e682-4b94-9539-cb2607725f71-stats-auth\") pod \"router-default-5444994796-f8mmz\" (UID: \"c26ccc9f-e682-4b94-9539-cb2607725f71\") " pod="openshift-ingress/router-default-5444994796-f8mmz" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.077169 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/699da377-c61b-48f0-92e0-42bd247ee363-etcd-service-ca\") pod \"etcd-operator-b45778765-hwf6r\" (UID: \"699da377-c61b-48f0-92e0-42bd247ee363\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hwf6r" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.077218 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9b75954c-b75d-4f17-82dc-4c0358fd8d0f-serving-cert\") pod \"console-operator-58897d9998-w76kw\" (UID: \"9b75954c-b75d-4f17-82dc-4c0358fd8d0f\") " pod="openshift-console-operator/console-operator-58897d9998-w76kw" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.077246 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/0149e083-ed5b-4c7f-94e7-4f3f2ff8a557-available-featuregates\") pod \"openshift-config-operator-7777fb866f-wgpdd\" (UID: \"0149e083-ed5b-4c7f-94e7-4f3f2ff8a557\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-wgpdd" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.077267 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c26ccc9f-e682-4b94-9539-cb2607725f71-metrics-certs\") pod \"router-default-5444994796-f8mmz\" (UID: \"c26ccc9f-e682-4b94-9539-cb2607725f71\") " pod="openshift-ingress/router-default-5444994796-f8mmz" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.077365 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vcq96\" (UniqueName: \"kubernetes.io/projected/9b75954c-b75d-4f17-82dc-4c0358fd8d0f-kube-api-access-vcq96\") pod \"console-operator-58897d9998-w76kw\" (UID: \"9b75954c-b75d-4f17-82dc-4c0358fd8d0f\") " pod="openshift-console-operator/console-operator-58897d9998-w76kw" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.077448 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c26ccc9f-e682-4b94-9539-cb2607725f71-service-ca-bundle\") pod \"router-default-5444994796-f8mmz\" (UID: \"c26ccc9f-e682-4b94-9539-cb2607725f71\") " pod="openshift-ingress/router-default-5444994796-f8mmz" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.077547 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/9c5d91bb-753e-4554-b7a1-6c85fe323875-ca-trust-extracted\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.077592 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xwfx4\" (UniqueName: \"kubernetes.io/projected/699da377-c61b-48f0-92e0-42bd247ee363-kube-api-access-xwfx4\") pod \"etcd-operator-b45778765-hwf6r\" (UID: \"699da377-c61b-48f0-92e0-42bd247ee363\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hwf6r" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.077611 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5d6cj\" (UniqueName: \"kubernetes.io/projected/59980734-2cf7-42dc-9f6f-a6de7e1c3665-kube-api-access-5d6cj\") pod \"kube-storage-version-migrator-operator-b67b599dd-56rqr\" (UID: \"59980734-2cf7-42dc-9f6f-a6de7e1c3665\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-56rqr" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.077654 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/9c5d91bb-753e-4554-b7a1-6c85fe323875-registry-certificates\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.077771 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/9c5d91bb-753e-4554-b7a1-6c85fe323875-installation-pull-secrets\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.077822 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q6bl2\" (UniqueName: \"kubernetes.io/projected/c26ccc9f-e682-4b94-9539-cb2607725f71-kube-api-access-q6bl2\") pod \"router-default-5444994796-f8mmz\" (UID: \"c26ccc9f-e682-4b94-9539-cb2607725f71\") " pod="openshift-ingress/router-default-5444994796-f8mmz" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.077895 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9c5d91bb-753e-4554-b7a1-6c85fe323875-bound-sa-token\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.077914 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/ebd59055-333a-46c2-ba30-826296414a4c-metrics-tls\") pod \"dns-operator-744455d44c-xkdjh\" (UID: \"ebd59055-333a-46c2-ba30-826296414a4c\") " pod="openshift-dns-operator/dns-operator-744455d44c-xkdjh" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.077975 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0149e083-ed5b-4c7f-94e7-4f3f2ff8a557-serving-cert\") pod \"openshift-config-operator-7777fb866f-wgpdd\" (UID: \"0149e083-ed5b-4c7f-94e7-4f3f2ff8a557\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-wgpdd" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.078023 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/699da377-c61b-48f0-92e0-42bd247ee363-etcd-client\") pod \"etcd-operator-b45778765-hwf6r\" (UID: \"699da377-c61b-48f0-92e0-42bd247ee363\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hwf6r" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.078058 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8ctrn\" (UniqueName: \"kubernetes.io/projected/0149e083-ed5b-4c7f-94e7-4f3f2ff8a557-kube-api-access-8ctrn\") pod \"openshift-config-operator-7777fb866f-wgpdd\" (UID: \"0149e083-ed5b-4c7f-94e7-4f3f2ff8a557\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-wgpdd" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.078103 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j6nzw\" (UniqueName: \"kubernetes.io/projected/f8ea4f24-05b0-4661-a617-a4208f9a9188-kube-api-access-j6nzw\") pod \"downloads-7954f5f757-544df\" (UID: \"f8ea4f24-05b0-4661-a617-a4208f9a9188\") " pod="openshift-console/downloads-7954f5f757-544df" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.078137 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jb2r5\" (UniqueName: \"kubernetes.io/projected/9c5d91bb-753e-4554-b7a1-6c85fe323875-kube-api-access-jb2r5\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.078161 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/10e8bf08-339f-4948-9fe2-ab4092d9a942-config\") pod \"kube-controller-manager-operator-78b949d7b-zz77x\" (UID: \"10e8bf08-339f-4948-9fe2-ab4092d9a942\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-zz77x" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.078193 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/10e8bf08-339f-4948-9fe2-ab4092d9a942-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-zz77x\" (UID: \"10e8bf08-339f-4948-9fe2-ab4092d9a942\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-zz77x" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.078212 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0396c426-1d56-44e6-885a-1b97d69ad57a-config\") pod \"kube-apiserver-operator-766d6c64bb-rmkfg\" (UID: \"0396c426-1d56-44e6-885a-1b97d69ad57a\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-rmkfg" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.078239 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9c5d91bb-753e-4554-b7a1-6c85fe323875-trusted-ca\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.078318 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.078559 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0396c426-1d56-44e6-885a-1b97d69ad57a-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-rmkfg\" (UID: \"0396c426-1d56-44e6-885a-1b97d69ad57a\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-rmkfg" Mar 13 14:09:03 crc kubenswrapper[4907]: E0313 14:09:03.078689 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:03.578669328 +0000 UTC m=+242.478457167 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.096740 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.116276 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.135293 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.156463 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.176050 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.179283 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:03 crc kubenswrapper[4907]: E0313 14:09:03.179454 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:03.679417877 +0000 UTC m=+242.579205586 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.179509 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.179576 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/7e89f101-cb61-4117-841d-58b5cc2bdd80-tmpfs\") pod \"packageserver-d55dfcdfc-pk65f\" (UID: \"7e89f101-cb61-4117-841d-58b5cc2bdd80\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pk65f" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.179604 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/74b29e18-7d73-4ef5-9739-04137dd1b191-auth-proxy-config\") pod \"machine-config-operator-74547568cd-92gs7\" (UID: \"74b29e18-7d73-4ef5-9739-04137dd1b191\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-92gs7" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.179639 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c934a2b9-3585-4491-8962-df085c473d90-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-xnr2b\" (UID: \"c934a2b9-3585-4491-8962-df085c473d90\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xnr2b" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.179673 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0396c426-1d56-44e6-885a-1b97d69ad57a-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-rmkfg\" (UID: \"0396c426-1d56-44e6-885a-1b97d69ad57a\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-rmkfg" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.179820 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lfn75\" (UniqueName: \"kubernetes.io/projected/c934a2b9-3585-4491-8962-df085c473d90-kube-api-access-lfn75\") pod \"openshift-apiserver-operator-796bbdcf4f-xnr2b\" (UID: \"c934a2b9-3585-4491-8962-df085c473d90\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xnr2b" Mar 13 14:09:03 crc kubenswrapper[4907]: E0313 14:09:03.179872 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:03.679856608 +0000 UTC m=+242.579644317 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.180149 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/1954f507-7cb0-4942-a4c8-6586d370e717-signing-cabundle\") pod \"service-ca-9c57cc56f-zwzc4\" (UID: \"1954f507-7cb0-4942-a4c8-6586d370e717\") " pod="openshift-service-ca/service-ca-9c57cc56f-zwzc4" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.180177 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/b651a92c-e9f6-4a1a-8c49-b6500bb7c61c-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-ccjsv\" (UID: \"b651a92c-e9f6-4a1a-8c49-b6500bb7c61c\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ccjsv" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.180256 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bgp25\" (UniqueName: \"kubernetes.io/projected/4682bbe1-4257-4d04-a6d7-60c4655d6873-kube-api-access-bgp25\") pod \"olm-operator-6b444d44fb-hkbqt\" (UID: \"4682bbe1-4257-4d04-a6d7-60c4655d6873\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hkbqt" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.180346 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/59980734-2cf7-42dc-9f6f-a6de7e1c3665-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-56rqr\" (UID: \"59980734-2cf7-42dc-9f6f-a6de7e1c3665\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-56rqr" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.180438 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/699da377-c61b-48f0-92e0-42bd247ee363-config\") pod \"etcd-operator-b45778765-hwf6r\" (UID: \"699da377-c61b-48f0-92e0-42bd247ee363\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hwf6r" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.180462 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c26ccc9f-e682-4b94-9539-cb2607725f71-default-certificate\") pod \"router-default-5444994796-f8mmz\" (UID: \"c26ccc9f-e682-4b94-9539-cb2607725f71\") " pod="openshift-ingress/router-default-5444994796-f8mmz" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.180484 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hnbr7\" (UniqueName: \"kubernetes.io/projected/7810633c-9d54-4cbe-8861-b523f724ec9e-kube-api-access-hnbr7\") pod \"catalog-operator-68c6474976-fwcwx\" (UID: \"7810633c-9d54-4cbe-8861-b523f724ec9e\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fwcwx" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.180521 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/7810633c-9d54-4cbe-8861-b523f724ec9e-srv-cert\") pod \"catalog-operator-68c6474976-fwcwx\" (UID: \"7810633c-9d54-4cbe-8861-b523f724ec9e\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fwcwx" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.180626 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wq2kr\" (UniqueName: \"kubernetes.io/projected/1954f507-7cb0-4942-a4c8-6586d370e717-kube-api-access-wq2kr\") pod \"service-ca-9c57cc56f-zwzc4\" (UID: \"1954f507-7cb0-4942-a4c8-6586d370e717\") " pod="openshift-service-ca/service-ca-9c57cc56f-zwzc4" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.180751 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/699da377-c61b-48f0-92e0-42bd247ee363-serving-cert\") pod \"etcd-operator-b45778765-hwf6r\" (UID: \"699da377-c61b-48f0-92e0-42bd247ee363\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hwf6r" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.180940 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0290eb6f-0be7-4340-8ba8-ed6ea0662c33-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-5stnc\" (UID: \"0290eb6f-0be7-4340-8ba8-ed6ea0662c33\") " pod="openshift-marketplace/marketplace-operator-79b997595-5stnc" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.181265 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0396c426-1d56-44e6-885a-1b97d69ad57a-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-rmkfg\" (UID: \"0396c426-1d56-44e6-885a-1b97d69ad57a\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-rmkfg" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.181305 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/10e8bf08-339f-4948-9fe2-ab4092d9a942-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-zz77x\" (UID: \"10e8bf08-339f-4948-9fe2-ab4092d9a942\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-zz77x" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.181342 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/4a8597a4-c72f-4023-b2fb-1b1237411a36-socket-dir\") pod \"csi-hostpathplugin-sjwvn\" (UID: \"4a8597a4-c72f-4023-b2fb-1b1237411a36\") " pod="hostpath-provisioner/csi-hostpathplugin-sjwvn" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.181368 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dwlcl\" (UniqueName: \"kubernetes.io/projected/fa265490-c23b-4b05-94bc-78b4a636e065-kube-api-access-dwlcl\") pod \"dns-default-8jqgq\" (UID: \"fa265490-c23b-4b05-94bc-78b4a636e065\") " pod="openshift-dns/dns-default-8jqgq" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.181389 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/19edf37b-a1d3-4c7c-945e-5d0b8bb601bf-secret-volume\") pod \"collect-profiles-29556840-pgmwj\" (UID: \"19edf37b-a1d3-4c7c-945e-5d0b8bb601bf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556840-pgmwj" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.181409 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/7e89f101-cb61-4117-841d-58b5cc2bdd80-webhook-cert\") pod \"packageserver-d55dfcdfc-pk65f\" (UID: \"7e89f101-cb61-4117-841d-58b5cc2bdd80\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pk65f" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.181432 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tc8b2\" (UniqueName: \"kubernetes.io/projected/3a661298-271c-4c57-94ea-3d69a27d7ad5-kube-api-access-tc8b2\") pod \"control-plane-machine-set-operator-78cbb6b69f-djph7\" (UID: \"3a661298-271c-4c57-94ea-3d69a27d7ad5\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-djph7" Mar 13 14:09:03 crc kubenswrapper[4907]: E0313 14:09:03.181462 4907 projected.go:194] Error preparing data for projected volume kube-api-access-57msq for pod openshift-oauth-apiserver/apiserver-7bbb656c7d-ksgr6: failed to sync configmap cache: timed out waiting for the condition Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.181476 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/4a8597a4-c72f-4023-b2fb-1b1237411a36-csi-data-dir\") pod \"csi-hostpathplugin-sjwvn\" (UID: \"4a8597a4-c72f-4023-b2fb-1b1237411a36\") " pod="hostpath-provisioner/csi-hostpathplugin-sjwvn" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.181507 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/4682bbe1-4257-4d04-a6d7-60c4655d6873-srv-cert\") pod \"olm-operator-6b444d44fb-hkbqt\" (UID: \"4682bbe1-4257-4d04-a6d7-60c4655d6873\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hkbqt" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.181528 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/9c5d91bb-753e-4554-b7a1-6c85fe323875-registry-tls\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:03 crc kubenswrapper[4907]: E0313 14:09:03.181538 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/1be912b5-9d1d-48bd-a0ab-7f601f9eb9da-kube-api-access-57msq podName:1be912b5-9d1d-48bd-a0ab-7f601f9eb9da nodeName:}" failed. No retries permitted until 2026-03-13 14:09:03.681524853 +0000 UTC m=+242.581312562 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-57msq" (UniqueName: "kubernetes.io/projected/1be912b5-9d1d-48bd-a0ab-7f601f9eb9da-kube-api-access-57msq") pod "apiserver-7bbb656c7d-ksgr6" (UID: "1be912b5-9d1d-48bd-a0ab-7f601f9eb9da") : failed to sync configmap cache: timed out waiting for the condition Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.181663 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/59980734-2cf7-42dc-9f6f-a6de7e1c3665-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-56rqr\" (UID: \"59980734-2cf7-42dc-9f6f-a6de7e1c3665\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-56rqr" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.181708 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wk2bb\" (UniqueName: \"kubernetes.io/projected/ebd59055-333a-46c2-ba30-826296414a4c-kube-api-access-wk2bb\") pod \"dns-operator-744455d44c-xkdjh\" (UID: \"ebd59055-333a-46c2-ba30-826296414a4c\") " pod="openshift-dns-operator/dns-operator-744455d44c-xkdjh" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.181733 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/699da377-c61b-48f0-92e0-42bd247ee363-etcd-ca\") pod \"etcd-operator-b45778765-hwf6r\" (UID: \"699da377-c61b-48f0-92e0-42bd247ee363\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hwf6r" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.181767 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b75954c-b75d-4f17-82dc-4c0358fd8d0f-config\") pod \"console-operator-58897d9998-w76kw\" (UID: \"9b75954c-b75d-4f17-82dc-4c0358fd8d0f\") " pod="openshift-console-operator/console-operator-58897d9998-w76kw" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.181792 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9b75954c-b75d-4f17-82dc-4c0358fd8d0f-trusted-ca\") pod \"console-operator-58897d9998-w76kw\" (UID: \"9b75954c-b75d-4f17-82dc-4c0358fd8d0f\") " pod="openshift-console-operator/console-operator-58897d9998-w76kw" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.181834 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c26ccc9f-e682-4b94-9539-cb2607725f71-stats-auth\") pod \"router-default-5444994796-f8mmz\" (UID: \"c26ccc9f-e682-4b94-9539-cb2607725f71\") " pod="openshift-ingress/router-default-5444994796-f8mmz" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.181852 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qllsp\" (UniqueName: \"kubernetes.io/projected/4a8597a4-c72f-4023-b2fb-1b1237411a36-kube-api-access-qllsp\") pod \"csi-hostpathplugin-sjwvn\" (UID: \"4a8597a4-c72f-4023-b2fb-1b1237411a36\") " pod="hostpath-provisioner/csi-hostpathplugin-sjwvn" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.181889 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/699da377-c61b-48f0-92e0-42bd247ee363-etcd-service-ca\") pod \"etcd-operator-b45778765-hwf6r\" (UID: \"699da377-c61b-48f0-92e0-42bd247ee363\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hwf6r" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.181906 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ff0e893f-05cb-4751-9960-281c58000519-config\") pod \"service-ca-operator-777779d784-75gdt\" (UID: \"ff0e893f-05cb-4751-9960-281c58000519\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-75gdt" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.181935 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9b75954c-b75d-4f17-82dc-4c0358fd8d0f-serving-cert\") pod \"console-operator-58897d9998-w76kw\" (UID: \"9b75954c-b75d-4f17-82dc-4c0358fd8d0f\") " pod="openshift-console-operator/console-operator-58897d9998-w76kw" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.181953 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lx6xs\" (UniqueName: \"kubernetes.io/projected/8013cba2-6b3c-4548-af5c-6dd0057da0c7-kube-api-access-lx6xs\") pod \"multus-admission-controller-857f4d67dd-ksfb8\" (UID: \"8013cba2-6b3c-4548-af5c-6dd0057da0c7\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-ksfb8" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.181970 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/c77fd554-e957-4d49-b8cd-b4f7dcd4ba2d-node-bootstrap-token\") pod \"machine-config-server-g6q2r\" (UID: \"c77fd554-e957-4d49-b8cd-b4f7dcd4ba2d\") " pod="openshift-machine-config-operator/machine-config-server-g6q2r" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.182061 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/0149e083-ed5b-4c7f-94e7-4f3f2ff8a557-available-featuregates\") pod \"openshift-config-operator-7777fb866f-wgpdd\" (UID: \"0149e083-ed5b-4c7f-94e7-4f3f2ff8a557\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-wgpdd" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.182232 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c26ccc9f-e682-4b94-9539-cb2607725f71-metrics-certs\") pod \"router-default-5444994796-f8mmz\" (UID: \"c26ccc9f-e682-4b94-9539-cb2607725f71\") " pod="openshift-ingress/router-default-5444994796-f8mmz" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.182521 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/0149e083-ed5b-4c7f-94e7-4f3f2ff8a557-available-featuregates\") pod \"openshift-config-operator-7777fb866f-wgpdd\" (UID: \"0149e083-ed5b-4c7f-94e7-4f3f2ff8a557\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-wgpdd" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.183422 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zhdhq\" (UniqueName: \"kubernetes.io/projected/ff214b86-eb3d-46c6-b45a-18a9f63e742c-kube-api-access-zhdhq\") pod \"ingress-canary-252nq\" (UID: \"ff214b86-eb3d-46c6-b45a-18a9f63e742c\") " pod="openshift-ingress-canary/ingress-canary-252nq" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.183462 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vcq96\" (UniqueName: \"kubernetes.io/projected/9b75954c-b75d-4f17-82dc-4c0358fd8d0f-kube-api-access-vcq96\") pod \"console-operator-58897d9998-w76kw\" (UID: \"9b75954c-b75d-4f17-82dc-4c0358fd8d0f\") " pod="openshift-console-operator/console-operator-58897d9998-w76kw" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.183484 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/0290eb6f-0be7-4340-8ba8-ed6ea0662c33-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-5stnc\" (UID: \"0290eb6f-0be7-4340-8ba8-ed6ea0662c33\") " pod="openshift-marketplace/marketplace-operator-79b997595-5stnc" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.183503 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/4a8597a4-c72f-4023-b2fb-1b1237411a36-plugins-dir\") pod \"csi-hostpathplugin-sjwvn\" (UID: \"4a8597a4-c72f-4023-b2fb-1b1237411a36\") " pod="hostpath-provisioner/csi-hostpathplugin-sjwvn" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.183521 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/3a661298-271c-4c57-94ea-3d69a27d7ad5-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-djph7\" (UID: \"3a661298-271c-4c57-94ea-3d69a27d7ad5\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-djph7" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.184079 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/7e89f101-cb61-4117-841d-58b5cc2bdd80-apiservice-cert\") pod \"packageserver-d55dfcdfc-pk65f\" (UID: \"7e89f101-cb61-4117-841d-58b5cc2bdd80\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pk65f" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.184118 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/74b29e18-7d73-4ef5-9739-04137dd1b191-proxy-tls\") pod \"machine-config-operator-74547568cd-92gs7\" (UID: \"74b29e18-7d73-4ef5-9739-04137dd1b191\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-92gs7" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.184169 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c26ccc9f-e682-4b94-9539-cb2607725f71-service-ca-bundle\") pod \"router-default-5444994796-f8mmz\" (UID: \"c26ccc9f-e682-4b94-9539-cb2607725f71\") " pod="openshift-ingress/router-default-5444994796-f8mmz" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.185075 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gnmnd\" (UniqueName: \"kubernetes.io/projected/814322e3-6251-4ea3-a86b-7b89ca6ea728-kube-api-access-gnmnd\") pod \"migrator-59844c95c7-vfbsx\" (UID: \"814322e3-6251-4ea3-a86b-7b89ca6ea728\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-vfbsx" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.185246 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9pzkh\" (UniqueName: \"kubernetes.io/projected/7e89f101-cb61-4117-841d-58b5cc2bdd80-kube-api-access-9pzkh\") pod \"packageserver-d55dfcdfc-pk65f\" (UID: \"7e89f101-cb61-4117-841d-58b5cc2bdd80\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pk65f" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.185316 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/c77fd554-e957-4d49-b8cd-b4f7dcd4ba2d-certs\") pod \"machine-config-server-g6q2r\" (UID: \"c77fd554-e957-4d49-b8cd-b4f7dcd4ba2d\") " pod="openshift-machine-config-operator/machine-config-server-g6q2r" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.185492 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/9c5d91bb-753e-4554-b7a1-6c85fe323875-ca-trust-extracted\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.185538 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qppp6\" (UniqueName: \"kubernetes.io/projected/0290eb6f-0be7-4340-8ba8-ed6ea0662c33-kube-api-access-qppp6\") pod \"marketplace-operator-79b997595-5stnc\" (UID: \"0290eb6f-0be7-4340-8ba8-ed6ea0662c33\") " pod="openshift-marketplace/marketplace-operator-79b997595-5stnc" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.185588 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/19edf37b-a1d3-4c7c-945e-5d0b8bb601bf-config-volume\") pod \"collect-profiles-29556840-pgmwj\" (UID: \"19edf37b-a1d3-4c7c-945e-5d0b8bb601bf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556840-pgmwj" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.185636 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xwfx4\" (UniqueName: \"kubernetes.io/projected/699da377-c61b-48f0-92e0-42bd247ee363-kube-api-access-xwfx4\") pod \"etcd-operator-b45778765-hwf6r\" (UID: \"699da377-c61b-48f0-92e0-42bd247ee363\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hwf6r" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.185692 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5d6cj\" (UniqueName: \"kubernetes.io/projected/59980734-2cf7-42dc-9f6f-a6de7e1c3665-kube-api-access-5d6cj\") pod \"kube-storage-version-migrator-operator-b67b599dd-56rqr\" (UID: \"59980734-2cf7-42dc-9f6f-a6de7e1c3665\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-56rqr" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.185742 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/74b29e18-7d73-4ef5-9739-04137dd1b191-images\") pod \"machine-config-operator-74547568cd-92gs7\" (UID: \"74b29e18-7d73-4ef5-9739-04137dd1b191\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-92gs7" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.185845 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/9c5d91bb-753e-4554-b7a1-6c85fe323875-registry-certificates\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.185916 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ff214b86-eb3d-46c6-b45a-18a9f63e742c-cert\") pod \"ingress-canary-252nq\" (UID: \"ff214b86-eb3d-46c6-b45a-18a9f63e742c\") " pod="openshift-ingress-canary/ingress-canary-252nq" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.185993 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/9c5d91bb-753e-4554-b7a1-6c85fe323875-installation-pull-secrets\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.186028 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9f5cd\" (UniqueName: \"kubernetes.io/projected/a8867781-179d-47e4-9a55-330deb10d273-kube-api-access-9f5cd\") pod \"auto-csr-approver-29556848-v8kv4\" (UID: \"a8867781-179d-47e4-9a55-330deb10d273\") " pod="openshift-infra/auto-csr-approver-29556848-v8kv4" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.186069 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c934a2b9-3585-4491-8962-df085c473d90-config\") pod \"openshift-apiserver-operator-796bbdcf4f-xnr2b\" (UID: \"c934a2b9-3585-4491-8962-df085c473d90\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xnr2b" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.186142 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n5x9f\" (UniqueName: \"kubernetes.io/projected/ff0e893f-05cb-4751-9960-281c58000519-kube-api-access-n5x9f\") pod \"service-ca-operator-777779d784-75gdt\" (UID: \"ff0e893f-05cb-4751-9960-281c58000519\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-75gdt" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.186200 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q6bl2\" (UniqueName: \"kubernetes.io/projected/c26ccc9f-e682-4b94-9539-cb2607725f71-kube-api-access-q6bl2\") pod \"router-default-5444994796-f8mmz\" (UID: \"c26ccc9f-e682-4b94-9539-cb2607725f71\") " pod="openshift-ingress/router-default-5444994796-f8mmz" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.186238 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/4682bbe1-4257-4d04-a6d7-60c4655d6873-profile-collector-cert\") pod \"olm-operator-6b444d44fb-hkbqt\" (UID: \"4682bbe1-4257-4d04-a6d7-60c4655d6873\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hkbqt" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.186284 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/fa265490-c23b-4b05-94bc-78b4a636e065-metrics-tls\") pod \"dns-default-8jqgq\" (UID: \"fa265490-c23b-4b05-94bc-78b4a636e065\") " pod="openshift-dns/dns-default-8jqgq" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.186324 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/4a8597a4-c72f-4023-b2fb-1b1237411a36-mountpoint-dir\") pod \"csi-hostpathplugin-sjwvn\" (UID: \"4a8597a4-c72f-4023-b2fb-1b1237411a36\") " pod="hostpath-provisioner/csi-hostpathplugin-sjwvn" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.186369 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9qzhc\" (UniqueName: \"kubernetes.io/projected/74b29e18-7d73-4ef5-9739-04137dd1b191-kube-api-access-9qzhc\") pod \"machine-config-operator-74547568cd-92gs7\" (UID: \"74b29e18-7d73-4ef5-9739-04137dd1b191\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-92gs7" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.186368 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/9c5d91bb-753e-4554-b7a1-6c85fe323875-ca-trust-extracted\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.187200 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h9w5d\" (UniqueName: \"kubernetes.io/projected/19edf37b-a1d3-4c7c-945e-5d0b8bb601bf-kube-api-access-h9w5d\") pod \"collect-profiles-29556840-pgmwj\" (UID: \"19edf37b-a1d3-4c7c-945e-5d0b8bb601bf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556840-pgmwj" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.187247 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/fa265490-c23b-4b05-94bc-78b4a636e065-config-volume\") pod \"dns-default-8jqgq\" (UID: \"fa265490-c23b-4b05-94bc-78b4a636e065\") " pod="openshift-dns/dns-default-8jqgq" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.187276 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9c5d91bb-753e-4554-b7a1-6c85fe323875-bound-sa-token\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.187302 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/7810633c-9d54-4cbe-8861-b523f724ec9e-profile-collector-cert\") pod \"catalog-operator-68c6474976-fwcwx\" (UID: \"7810633c-9d54-4cbe-8861-b523f724ec9e\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fwcwx" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.187329 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/1954f507-7cb0-4942-a4c8-6586d370e717-signing-key\") pod \"service-ca-9c57cc56f-zwzc4\" (UID: \"1954f507-7cb0-4942-a4c8-6586d370e717\") " pod="openshift-service-ca/service-ca-9c57cc56f-zwzc4" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.187349 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/ebd59055-333a-46c2-ba30-826296414a4c-metrics-tls\") pod \"dns-operator-744455d44c-xkdjh\" (UID: \"ebd59055-333a-46c2-ba30-826296414a4c\") " pod="openshift-dns-operator/dns-operator-744455d44c-xkdjh" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.187373 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0149e083-ed5b-4c7f-94e7-4f3f2ff8a557-serving-cert\") pod \"openshift-config-operator-7777fb866f-wgpdd\" (UID: \"0149e083-ed5b-4c7f-94e7-4f3f2ff8a557\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-wgpdd" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.188340 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8ctrn\" (UniqueName: \"kubernetes.io/projected/0149e083-ed5b-4c7f-94e7-4f3f2ff8a557-kube-api-access-8ctrn\") pod \"openshift-config-operator-7777fb866f-wgpdd\" (UID: \"0149e083-ed5b-4c7f-94e7-4f3f2ff8a557\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-wgpdd" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.188393 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/699da377-c61b-48f0-92e0-42bd247ee363-etcd-client\") pod \"etcd-operator-b45778765-hwf6r\" (UID: \"699da377-c61b-48f0-92e0-42bd247ee363\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hwf6r" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.189031 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/4a8597a4-c72f-4023-b2fb-1b1237411a36-registration-dir\") pod \"csi-hostpathplugin-sjwvn\" (UID: \"4a8597a4-c72f-4023-b2fb-1b1237411a36\") " pod="hostpath-provisioner/csi-hostpathplugin-sjwvn" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.189100 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j6nzw\" (UniqueName: \"kubernetes.io/projected/f8ea4f24-05b0-4661-a617-a4208f9a9188-kube-api-access-j6nzw\") pod \"downloads-7954f5f757-544df\" (UID: \"f8ea4f24-05b0-4661-a617-a4208f9a9188\") " pod="openshift-console/downloads-7954f5f757-544df" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.189334 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jb2r5\" (UniqueName: \"kubernetes.io/projected/9c5d91bb-753e-4554-b7a1-6c85fe323875-kube-api-access-jb2r5\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.189493 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ff0e893f-05cb-4751-9960-281c58000519-serving-cert\") pod \"service-ca-operator-777779d784-75gdt\" (UID: \"ff0e893f-05cb-4751-9960-281c58000519\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-75gdt" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.189553 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/9c5d91bb-753e-4554-b7a1-6c85fe323875-registry-certificates\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.189837 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/10e8bf08-339f-4948-9fe2-ab4092d9a942-config\") pod \"kube-controller-manager-operator-78b949d7b-zz77x\" (UID: \"10e8bf08-339f-4948-9fe2-ab4092d9a942\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-zz77x" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.190086 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/10e8bf08-339f-4948-9fe2-ab4092d9a942-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-zz77x\" (UID: \"10e8bf08-339f-4948-9fe2-ab4092d9a942\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-zz77x" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.190126 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0396c426-1d56-44e6-885a-1b97d69ad57a-config\") pod \"kube-apiserver-operator-766d6c64bb-rmkfg\" (UID: \"0396c426-1d56-44e6-885a-1b97d69ad57a\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-rmkfg" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.190169 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/8013cba2-6b3c-4548-af5c-6dd0057da0c7-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-ksfb8\" (UID: \"8013cba2-6b3c-4548-af5c-6dd0057da0c7\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-ksfb8" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.190210 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dtbj8\" (UniqueName: \"kubernetes.io/projected/b651a92c-e9f6-4a1a-8c49-b6500bb7c61c-kube-api-access-dtbj8\") pod \"package-server-manager-789f6589d5-ccjsv\" (UID: \"b651a92c-e9f6-4a1a-8c49-b6500bb7c61c\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ccjsv" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.190513 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jgwz8\" (UniqueName: \"kubernetes.io/projected/c77fd554-e957-4d49-b8cd-b4f7dcd4ba2d-kube-api-access-jgwz8\") pod \"machine-config-server-g6q2r\" (UID: \"c77fd554-e957-4d49-b8cd-b4f7dcd4ba2d\") " pod="openshift-machine-config-operator/machine-config-server-g6q2r" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.190708 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9c5d91bb-753e-4554-b7a1-6c85fe323875-trusted-ca\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.192627 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9c5d91bb-753e-4554-b7a1-6c85fe323875-trusted-ca\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.196051 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Mar 13 14:09:03 crc kubenswrapper[4907]: E0313 14:09:03.198780 4907 projected.go:194] Error preparing data for projected volume kube-api-access-mtqz9 for pod openshift-authentication/oauth-openshift-558db77b4-ng4kz: failed to sync configmap cache: timed out waiting for the condition Mar 13 14:09:03 crc kubenswrapper[4907]: E0313 14:09:03.198843 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3661f6b2-1d5e-42f0-947f-a32b66258a1f-kube-api-access-mtqz9 podName:3661f6b2-1d5e-42f0-947f-a32b66258a1f nodeName:}" failed. No retries permitted until 2026-03-13 14:09:03.698826374 +0000 UTC m=+242.598614063 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-mtqz9" (UniqueName: "kubernetes.io/projected/3661f6b2-1d5e-42f0-947f-a32b66258a1f-kube-api-access-mtqz9") pod "oauth-openshift-558db77b4-ng4kz" (UID: "3661f6b2-1d5e-42f0-947f-a32b66258a1f") : failed to sync configmap cache: timed out waiting for the condition Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.216228 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Mar 13 14:09:03 crc kubenswrapper[4907]: E0313 14:09:03.219951 4907 projected.go:194] Error preparing data for projected volume kube-api-access-hlg49 for pod openshift-apiserver/apiserver-76f77b778f-r86r4: failed to sync configmap cache: timed out waiting for the condition Mar 13 14:09:03 crc kubenswrapper[4907]: E0313 14:09:03.220072 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/5a9c889b-5954-4d3a-ae72-12000ee1cc25-kube-api-access-hlg49 podName:5a9c889b-5954-4d3a-ae72-12000ee1cc25 nodeName:}" failed. No retries permitted until 2026-03-13 14:09:03.720043801 +0000 UTC m=+242.619831530 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-hlg49" (UniqueName: "kubernetes.io/projected/5a9c889b-5954-4d3a-ae72-12000ee1cc25-kube-api-access-hlg49") pod "apiserver-76f77b778f-r86r4" (UID: "5a9c889b-5954-4d3a-ae72-12000ee1cc25") : failed to sync configmap cache: timed out waiting for the condition Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.221822 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-rnh5t" event={"ID":"a3baeb4f-988d-4d42-b469-a310ca127745","Type":"ContainerStarted","Data":"c8e2efcf239dc9215a99f09f489a8e23320bb5126f2409f2a08ac75994d91c04"} Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.236205 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.256176 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.263961 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e9d44a1e-0eab-4bd6-833f-143436e43758-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-hh9sk\" (UID: \"e9d44a1e-0eab-4bd6-833f-143436e43758\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-hh9sk" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.278519 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.291906 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:03 crc kubenswrapper[4907]: E0313 14:09:03.292044 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:03.792022488 +0000 UTC m=+242.691810187 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.292150 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hnbr7\" (UniqueName: \"kubernetes.io/projected/7810633c-9d54-4cbe-8861-b523f724ec9e-kube-api-access-hnbr7\") pod \"catalog-operator-68c6474976-fwcwx\" (UID: \"7810633c-9d54-4cbe-8861-b523f724ec9e\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fwcwx" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.292209 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/7810633c-9d54-4cbe-8861-b523f724ec9e-srv-cert\") pod \"catalog-operator-68c6474976-fwcwx\" (UID: \"7810633c-9d54-4cbe-8861-b523f724ec9e\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fwcwx" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.292242 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0290eb6f-0be7-4340-8ba8-ed6ea0662c33-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-5stnc\" (UID: \"0290eb6f-0be7-4340-8ba8-ed6ea0662c33\") " pod="openshift-marketplace/marketplace-operator-79b997595-5stnc" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.292265 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wq2kr\" (UniqueName: \"kubernetes.io/projected/1954f507-7cb0-4942-a4c8-6586d370e717-kube-api-access-wq2kr\") pod \"service-ca-9c57cc56f-zwzc4\" (UID: \"1954f507-7cb0-4942-a4c8-6586d370e717\") " pod="openshift-service-ca/service-ca-9c57cc56f-zwzc4" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.292310 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/4a8597a4-c72f-4023-b2fb-1b1237411a36-socket-dir\") pod \"csi-hostpathplugin-sjwvn\" (UID: \"4a8597a4-c72f-4023-b2fb-1b1237411a36\") " pod="hostpath-provisioner/csi-hostpathplugin-sjwvn" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.292334 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/7e89f101-cb61-4117-841d-58b5cc2bdd80-webhook-cert\") pod \"packageserver-d55dfcdfc-pk65f\" (UID: \"7e89f101-cb61-4117-841d-58b5cc2bdd80\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pk65f" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.292357 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tc8b2\" (UniqueName: \"kubernetes.io/projected/3a661298-271c-4c57-94ea-3d69a27d7ad5-kube-api-access-tc8b2\") pod \"control-plane-machine-set-operator-78cbb6b69f-djph7\" (UID: \"3a661298-271c-4c57-94ea-3d69a27d7ad5\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-djph7" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.292379 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dwlcl\" (UniqueName: \"kubernetes.io/projected/fa265490-c23b-4b05-94bc-78b4a636e065-kube-api-access-dwlcl\") pod \"dns-default-8jqgq\" (UID: \"fa265490-c23b-4b05-94bc-78b4a636e065\") " pod="openshift-dns/dns-default-8jqgq" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.292400 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/19edf37b-a1d3-4c7c-945e-5d0b8bb601bf-secret-volume\") pod \"collect-profiles-29556840-pgmwj\" (UID: \"19edf37b-a1d3-4c7c-945e-5d0b8bb601bf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556840-pgmwj" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.292421 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/4682bbe1-4257-4d04-a6d7-60c4655d6873-srv-cert\") pod \"olm-operator-6b444d44fb-hkbqt\" (UID: \"4682bbe1-4257-4d04-a6d7-60c4655d6873\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hkbqt" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.292443 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/4a8597a4-c72f-4023-b2fb-1b1237411a36-csi-data-dir\") pod \"csi-hostpathplugin-sjwvn\" (UID: \"4a8597a4-c72f-4023-b2fb-1b1237411a36\") " pod="hostpath-provisioner/csi-hostpathplugin-sjwvn" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.292522 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qllsp\" (UniqueName: \"kubernetes.io/projected/4a8597a4-c72f-4023-b2fb-1b1237411a36-kube-api-access-qllsp\") pod \"csi-hostpathplugin-sjwvn\" (UID: \"4a8597a4-c72f-4023-b2fb-1b1237411a36\") " pod="hostpath-provisioner/csi-hostpathplugin-sjwvn" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.292552 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ff0e893f-05cb-4751-9960-281c58000519-config\") pod \"service-ca-operator-777779d784-75gdt\" (UID: \"ff0e893f-05cb-4751-9960-281c58000519\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-75gdt" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.292582 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lx6xs\" (UniqueName: \"kubernetes.io/projected/8013cba2-6b3c-4548-af5c-6dd0057da0c7-kube-api-access-lx6xs\") pod \"multus-admission-controller-857f4d67dd-ksfb8\" (UID: \"8013cba2-6b3c-4548-af5c-6dd0057da0c7\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-ksfb8" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.292607 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/c77fd554-e957-4d49-b8cd-b4f7dcd4ba2d-node-bootstrap-token\") pod \"machine-config-server-g6q2r\" (UID: \"c77fd554-e957-4d49-b8cd-b4f7dcd4ba2d\") " pod="openshift-machine-config-operator/machine-config-server-g6q2r" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.292639 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zhdhq\" (UniqueName: \"kubernetes.io/projected/ff214b86-eb3d-46c6-b45a-18a9f63e742c-kube-api-access-zhdhq\") pod \"ingress-canary-252nq\" (UID: \"ff214b86-eb3d-46c6-b45a-18a9f63e742c\") " pod="openshift-ingress-canary/ingress-canary-252nq" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.292688 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/0290eb6f-0be7-4340-8ba8-ed6ea0662c33-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-5stnc\" (UID: \"0290eb6f-0be7-4340-8ba8-ed6ea0662c33\") " pod="openshift-marketplace/marketplace-operator-79b997595-5stnc" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.292711 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/4a8597a4-c72f-4023-b2fb-1b1237411a36-plugins-dir\") pod \"csi-hostpathplugin-sjwvn\" (UID: \"4a8597a4-c72f-4023-b2fb-1b1237411a36\") " pod="hostpath-provisioner/csi-hostpathplugin-sjwvn" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.292735 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/3a661298-271c-4c57-94ea-3d69a27d7ad5-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-djph7\" (UID: \"3a661298-271c-4c57-94ea-3d69a27d7ad5\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-djph7" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.292760 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/74b29e18-7d73-4ef5-9739-04137dd1b191-proxy-tls\") pod \"machine-config-operator-74547568cd-92gs7\" (UID: \"74b29e18-7d73-4ef5-9739-04137dd1b191\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-92gs7" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.292784 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/7e89f101-cb61-4117-841d-58b5cc2bdd80-apiservice-cert\") pod \"packageserver-d55dfcdfc-pk65f\" (UID: \"7e89f101-cb61-4117-841d-58b5cc2bdd80\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pk65f" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.292819 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gnmnd\" (UniqueName: \"kubernetes.io/projected/814322e3-6251-4ea3-a86b-7b89ca6ea728-kube-api-access-gnmnd\") pod \"migrator-59844c95c7-vfbsx\" (UID: \"814322e3-6251-4ea3-a86b-7b89ca6ea728\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-vfbsx" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.292851 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9pzkh\" (UniqueName: \"kubernetes.io/projected/7e89f101-cb61-4117-841d-58b5cc2bdd80-kube-api-access-9pzkh\") pod \"packageserver-d55dfcdfc-pk65f\" (UID: \"7e89f101-cb61-4117-841d-58b5cc2bdd80\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pk65f" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.292929 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/c77fd554-e957-4d49-b8cd-b4f7dcd4ba2d-certs\") pod \"machine-config-server-g6q2r\" (UID: \"c77fd554-e957-4d49-b8cd-b4f7dcd4ba2d\") " pod="openshift-machine-config-operator/machine-config-server-g6q2r" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.292982 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qppp6\" (UniqueName: \"kubernetes.io/projected/0290eb6f-0be7-4340-8ba8-ed6ea0662c33-kube-api-access-qppp6\") pod \"marketplace-operator-79b997595-5stnc\" (UID: \"0290eb6f-0be7-4340-8ba8-ed6ea0662c33\") " pod="openshift-marketplace/marketplace-operator-79b997595-5stnc" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.293014 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/19edf37b-a1d3-4c7c-945e-5d0b8bb601bf-config-volume\") pod \"collect-profiles-29556840-pgmwj\" (UID: \"19edf37b-a1d3-4c7c-945e-5d0b8bb601bf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556840-pgmwj" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.293076 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/74b29e18-7d73-4ef5-9739-04137dd1b191-images\") pod \"machine-config-operator-74547568cd-92gs7\" (UID: \"74b29e18-7d73-4ef5-9739-04137dd1b191\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-92gs7" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.293100 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/4a8597a4-c72f-4023-b2fb-1b1237411a36-socket-dir\") pod \"csi-hostpathplugin-sjwvn\" (UID: \"4a8597a4-c72f-4023-b2fb-1b1237411a36\") " pod="hostpath-provisioner/csi-hostpathplugin-sjwvn" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.293117 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ff214b86-eb3d-46c6-b45a-18a9f63e742c-cert\") pod \"ingress-canary-252nq\" (UID: \"ff214b86-eb3d-46c6-b45a-18a9f63e742c\") " pod="openshift-ingress-canary/ingress-canary-252nq" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.293142 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c934a2b9-3585-4491-8962-df085c473d90-config\") pod \"openshift-apiserver-operator-796bbdcf4f-xnr2b\" (UID: \"c934a2b9-3585-4491-8962-df085c473d90\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xnr2b" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.293177 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9f5cd\" (UniqueName: \"kubernetes.io/projected/a8867781-179d-47e4-9a55-330deb10d273-kube-api-access-9f5cd\") pod \"auto-csr-approver-29556848-v8kv4\" (UID: \"a8867781-179d-47e4-9a55-330deb10d273\") " pod="openshift-infra/auto-csr-approver-29556848-v8kv4" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.293203 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n5x9f\" (UniqueName: \"kubernetes.io/projected/ff0e893f-05cb-4751-9960-281c58000519-kube-api-access-n5x9f\") pod \"service-ca-operator-777779d784-75gdt\" (UID: \"ff0e893f-05cb-4751-9960-281c58000519\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-75gdt" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.293235 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/4682bbe1-4257-4d04-a6d7-60c4655d6873-profile-collector-cert\") pod \"olm-operator-6b444d44fb-hkbqt\" (UID: \"4682bbe1-4257-4d04-a6d7-60c4655d6873\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hkbqt" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.293258 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/fa265490-c23b-4b05-94bc-78b4a636e065-metrics-tls\") pod \"dns-default-8jqgq\" (UID: \"fa265490-c23b-4b05-94bc-78b4a636e065\") " pod="openshift-dns/dns-default-8jqgq" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.293293 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/4a8597a4-c72f-4023-b2fb-1b1237411a36-mountpoint-dir\") pod \"csi-hostpathplugin-sjwvn\" (UID: \"4a8597a4-c72f-4023-b2fb-1b1237411a36\") " pod="hostpath-provisioner/csi-hostpathplugin-sjwvn" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.293333 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9qzhc\" (UniqueName: \"kubernetes.io/projected/74b29e18-7d73-4ef5-9739-04137dd1b191-kube-api-access-9qzhc\") pod \"machine-config-operator-74547568cd-92gs7\" (UID: \"74b29e18-7d73-4ef5-9739-04137dd1b191\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-92gs7" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.293381 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/fa265490-c23b-4b05-94bc-78b4a636e065-config-volume\") pod \"dns-default-8jqgq\" (UID: \"fa265490-c23b-4b05-94bc-78b4a636e065\") " pod="openshift-dns/dns-default-8jqgq" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.293414 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h9w5d\" (UniqueName: \"kubernetes.io/projected/19edf37b-a1d3-4c7c-945e-5d0b8bb601bf-kube-api-access-h9w5d\") pod \"collect-profiles-29556840-pgmwj\" (UID: \"19edf37b-a1d3-4c7c-945e-5d0b8bb601bf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556840-pgmwj" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.293456 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/7810633c-9d54-4cbe-8861-b523f724ec9e-profile-collector-cert\") pod \"catalog-operator-68c6474976-fwcwx\" (UID: \"7810633c-9d54-4cbe-8861-b523f724ec9e\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fwcwx" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.293493 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/1954f507-7cb0-4942-a4c8-6586d370e717-signing-key\") pod \"service-ca-9c57cc56f-zwzc4\" (UID: \"1954f507-7cb0-4942-a4c8-6586d370e717\") " pod="openshift-service-ca/service-ca-9c57cc56f-zwzc4" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.293562 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/4a8597a4-c72f-4023-b2fb-1b1237411a36-registration-dir\") pod \"csi-hostpathplugin-sjwvn\" (UID: \"4a8597a4-c72f-4023-b2fb-1b1237411a36\") " pod="hostpath-provisioner/csi-hostpathplugin-sjwvn" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.293634 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ff0e893f-05cb-4751-9960-281c58000519-serving-cert\") pod \"service-ca-operator-777779d784-75gdt\" (UID: \"ff0e893f-05cb-4751-9960-281c58000519\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-75gdt" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.293680 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/8013cba2-6b3c-4548-af5c-6dd0057da0c7-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-ksfb8\" (UID: \"8013cba2-6b3c-4548-af5c-6dd0057da0c7\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-ksfb8" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.293712 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dtbj8\" (UniqueName: \"kubernetes.io/projected/b651a92c-e9f6-4a1a-8c49-b6500bb7c61c-kube-api-access-dtbj8\") pod \"package-server-manager-789f6589d5-ccjsv\" (UID: \"b651a92c-e9f6-4a1a-8c49-b6500bb7c61c\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ccjsv" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.293719 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/4a8597a4-c72f-4023-b2fb-1b1237411a36-plugins-dir\") pod \"csi-hostpathplugin-sjwvn\" (UID: \"4a8597a4-c72f-4023-b2fb-1b1237411a36\") " pod="hostpath-provisioner/csi-hostpathplugin-sjwvn" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.293819 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jgwz8\" (UniqueName: \"kubernetes.io/projected/c77fd554-e957-4d49-b8cd-b4f7dcd4ba2d-kube-api-access-jgwz8\") pod \"machine-config-server-g6q2r\" (UID: \"c77fd554-e957-4d49-b8cd-b4f7dcd4ba2d\") " pod="openshift-machine-config-operator/machine-config-server-g6q2r" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.293860 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.293918 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/7e89f101-cb61-4117-841d-58b5cc2bdd80-tmpfs\") pod \"packageserver-d55dfcdfc-pk65f\" (UID: \"7e89f101-cb61-4117-841d-58b5cc2bdd80\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pk65f" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.293925 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/4a8597a4-c72f-4023-b2fb-1b1237411a36-csi-data-dir\") pod \"csi-hostpathplugin-sjwvn\" (UID: \"4a8597a4-c72f-4023-b2fb-1b1237411a36\") " pod="hostpath-provisioner/csi-hostpathplugin-sjwvn" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.293952 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/74b29e18-7d73-4ef5-9739-04137dd1b191-auth-proxy-config\") pod \"machine-config-operator-74547568cd-92gs7\" (UID: \"74b29e18-7d73-4ef5-9739-04137dd1b191\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-92gs7" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.293993 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c934a2b9-3585-4491-8962-df085c473d90-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-xnr2b\" (UID: \"c934a2b9-3585-4491-8962-df085c473d90\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xnr2b" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.294020 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lfn75\" (UniqueName: \"kubernetes.io/projected/c934a2b9-3585-4491-8962-df085c473d90-kube-api-access-lfn75\") pod \"openshift-apiserver-operator-796bbdcf4f-xnr2b\" (UID: \"c934a2b9-3585-4491-8962-df085c473d90\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xnr2b" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.294044 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/1954f507-7cb0-4942-a4c8-6586d370e717-signing-cabundle\") pod \"service-ca-9c57cc56f-zwzc4\" (UID: \"1954f507-7cb0-4942-a4c8-6586d370e717\") " pod="openshift-service-ca/service-ca-9c57cc56f-zwzc4" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.294067 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/b651a92c-e9f6-4a1a-8c49-b6500bb7c61c-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-ccjsv\" (UID: \"b651a92c-e9f6-4a1a-8c49-b6500bb7c61c\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ccjsv" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.294065 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ff0e893f-05cb-4751-9960-281c58000519-config\") pod \"service-ca-operator-777779d784-75gdt\" (UID: \"ff0e893f-05cb-4751-9960-281c58000519\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-75gdt" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.294093 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bgp25\" (UniqueName: \"kubernetes.io/projected/4682bbe1-4257-4d04-a6d7-60c4655d6873-kube-api-access-bgp25\") pod \"olm-operator-6b444d44fb-hkbqt\" (UID: \"4682bbe1-4257-4d04-a6d7-60c4655d6873\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hkbqt" Mar 13 14:09:03 crc kubenswrapper[4907]: E0313 14:09:03.294271 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:03.794259359 +0000 UTC m=+242.694047048 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.294353 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/19edf37b-a1d3-4c7c-945e-5d0b8bb601bf-config-volume\") pod \"collect-profiles-29556840-pgmwj\" (UID: \"19edf37b-a1d3-4c7c-945e-5d0b8bb601bf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556840-pgmwj" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.294628 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/7e89f101-cb61-4117-841d-58b5cc2bdd80-tmpfs\") pod \"packageserver-d55dfcdfc-pk65f\" (UID: \"7e89f101-cb61-4117-841d-58b5cc2bdd80\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pk65f" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.294833 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c934a2b9-3585-4491-8962-df085c473d90-config\") pod \"openshift-apiserver-operator-796bbdcf4f-xnr2b\" (UID: \"c934a2b9-3585-4491-8962-df085c473d90\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xnr2b" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.295320 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/74b29e18-7d73-4ef5-9739-04137dd1b191-auth-proxy-config\") pod \"machine-config-operator-74547568cd-92gs7\" (UID: \"74b29e18-7d73-4ef5-9739-04137dd1b191\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-92gs7" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.293724 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0290eb6f-0be7-4340-8ba8-ed6ea0662c33-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-5stnc\" (UID: \"0290eb6f-0be7-4340-8ba8-ed6ea0662c33\") " pod="openshift-marketplace/marketplace-operator-79b997595-5stnc" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.296104 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/c77fd554-e957-4d49-b8cd-b4f7dcd4ba2d-node-bootstrap-token\") pod \"machine-config-server-g6q2r\" (UID: \"c77fd554-e957-4d49-b8cd-b4f7dcd4ba2d\") " pod="openshift-machine-config-operator/machine-config-server-g6q2r" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.296806 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/1954f507-7cb0-4942-a4c8-6586d370e717-signing-cabundle\") pod \"service-ca-9c57cc56f-zwzc4\" (UID: \"1954f507-7cb0-4942-a4c8-6586d370e717\") " pod="openshift-service-ca/service-ca-9c57cc56f-zwzc4" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.296820 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/7e89f101-cb61-4117-841d-58b5cc2bdd80-apiservice-cert\") pod \"packageserver-d55dfcdfc-pk65f\" (UID: \"7e89f101-cb61-4117-841d-58b5cc2bdd80\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pk65f" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.297906 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/19edf37b-a1d3-4c7c-945e-5d0b8bb601bf-secret-volume\") pod \"collect-profiles-29556840-pgmwj\" (UID: \"19edf37b-a1d3-4c7c-945e-5d0b8bb601bf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556840-pgmwj" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.298286 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c934a2b9-3585-4491-8962-df085c473d90-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-xnr2b\" (UID: \"c934a2b9-3585-4491-8962-df085c473d90\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xnr2b" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.298938 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/7810633c-9d54-4cbe-8861-b523f724ec9e-profile-collector-cert\") pod \"catalog-operator-68c6474976-fwcwx\" (UID: \"7810633c-9d54-4cbe-8861-b523f724ec9e\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fwcwx" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.299392 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.299583 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/7810633c-9d54-4cbe-8861-b523f724ec9e-srv-cert\") pod \"catalog-operator-68c6474976-fwcwx\" (UID: \"7810633c-9d54-4cbe-8861-b523f724ec9e\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fwcwx" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.299683 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/4a8597a4-c72f-4023-b2fb-1b1237411a36-mountpoint-dir\") pod \"csi-hostpathplugin-sjwvn\" (UID: \"4a8597a4-c72f-4023-b2fb-1b1237411a36\") " pod="hostpath-provisioner/csi-hostpathplugin-sjwvn" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.299949 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/4a8597a4-c72f-4023-b2fb-1b1237411a36-registration-dir\") pod \"csi-hostpathplugin-sjwvn\" (UID: \"4a8597a4-c72f-4023-b2fb-1b1237411a36\") " pod="hostpath-provisioner/csi-hostpathplugin-sjwvn" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.300007 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/4682bbe1-4257-4d04-a6d7-60c4655d6873-profile-collector-cert\") pod \"olm-operator-6b444d44fb-hkbqt\" (UID: \"4682bbe1-4257-4d04-a6d7-60c4655d6873\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hkbqt" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.300068 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/fa265490-c23b-4b05-94bc-78b4a636e065-config-volume\") pod \"dns-default-8jqgq\" (UID: \"fa265490-c23b-4b05-94bc-78b4a636e065\") " pod="openshift-dns/dns-default-8jqgq" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.300382 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/4682bbe1-4257-4d04-a6d7-60c4655d6873-srv-cert\") pod \"olm-operator-6b444d44fb-hkbqt\" (UID: \"4682bbe1-4257-4d04-a6d7-60c4655d6873\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hkbqt" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.301067 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/0290eb6f-0be7-4340-8ba8-ed6ea0662c33-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-5stnc\" (UID: \"0290eb6f-0be7-4340-8ba8-ed6ea0662c33\") " pod="openshift-marketplace/marketplace-operator-79b997595-5stnc" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.301211 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/c77fd554-e957-4d49-b8cd-b4f7dcd4ba2d-certs\") pod \"machine-config-server-g6q2r\" (UID: \"c77fd554-e957-4d49-b8cd-b4f7dcd4ba2d\") " pod="openshift-machine-config-operator/machine-config-server-g6q2r" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.301285 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ff214b86-eb3d-46c6-b45a-18a9f63e742c-cert\") pod \"ingress-canary-252nq\" (UID: \"ff214b86-eb3d-46c6-b45a-18a9f63e742c\") " pod="openshift-ingress-canary/ingress-canary-252nq" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.301599 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/fa265490-c23b-4b05-94bc-78b4a636e065-metrics-tls\") pod \"dns-default-8jqgq\" (UID: \"fa265490-c23b-4b05-94bc-78b4a636e065\") " pod="openshift-dns/dns-default-8jqgq" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.301979 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/7e89f101-cb61-4117-841d-58b5cc2bdd80-webhook-cert\") pod \"packageserver-d55dfcdfc-pk65f\" (UID: \"7e89f101-cb61-4117-841d-58b5cc2bdd80\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pk65f" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.301994 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ff0e893f-05cb-4751-9960-281c58000519-serving-cert\") pod \"service-ca-operator-777779d784-75gdt\" (UID: \"ff0e893f-05cb-4751-9960-281c58000519\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-75gdt" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.302127 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/b651a92c-e9f6-4a1a-8c49-b6500bb7c61c-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-ccjsv\" (UID: \"b651a92c-e9f6-4a1a-8c49-b6500bb7c61c\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ccjsv" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.302316 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/1954f507-7cb0-4942-a4c8-6586d370e717-signing-key\") pod \"service-ca-9c57cc56f-zwzc4\" (UID: \"1954f507-7cb0-4942-a4c8-6586d370e717\") " pod="openshift-service-ca/service-ca-9c57cc56f-zwzc4" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.315745 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Mar 13 14:09:03 crc kubenswrapper[4907]: E0313 14:09:03.321949 4907 projected.go:194] Error preparing data for projected volume kube-api-access-s47gg for pod openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-wjrms: failed to sync configmap cache: timed out waiting for the condition Mar 13 14:09:03 crc kubenswrapper[4907]: E0313 14:09:03.322043 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/71c234e9-5f3d-4b0c-94d4-80b4e33a6dda-kube-api-access-s47gg podName:71c234e9-5f3d-4b0c-94d4-80b4e33a6dda nodeName:}" failed. No retries permitted until 2026-03-13 14:09:03.822020164 +0000 UTC m=+242.721807943 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s47gg" (UniqueName: "kubernetes.io/projected/71c234e9-5f3d-4b0c-94d4-80b4e33a6dda-kube-api-access-s47gg") pod "cluster-samples-operator-665b6dd947-wjrms" (UID: "71c234e9-5f3d-4b0c-94d4-80b4e33a6dda") : failed to sync configmap cache: timed out waiting for the condition Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.335690 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Mar 13 14:09:03 crc kubenswrapper[4907]: E0313 14:09:03.340107 4907 projected.go:194] Error preparing data for projected volume kube-api-access-9qt98 for pod openshift-cluster-machine-approver/machine-approver-56656f9798-52vhb: failed to sync configmap cache: timed out waiting for the condition Mar 13 14:09:03 crc kubenswrapper[4907]: E0313 14:09:03.340180 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/005dfe54-e402-4aa4-8b67-f7e7b685debd-kube-api-access-9qt98 podName:005dfe54-e402-4aa4-8b67-f7e7b685debd nodeName:}" failed. No retries permitted until 2026-03-13 14:09:03.840157947 +0000 UTC m=+242.739945706 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-9qt98" (UniqueName: "kubernetes.io/projected/005dfe54-e402-4aa4-8b67-f7e7b685debd-kube-api-access-9qt98") pod "machine-approver-56656f9798-52vhb" (UID: "005dfe54-e402-4aa4-8b67-f7e7b685debd") : failed to sync configmap cache: timed out waiting for the condition Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.355642 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 13 14:09:03 crc kubenswrapper[4907]: E0313 14:09:03.358440 4907 projected.go:194] Error preparing data for projected volume kube-api-access-gqtv7 for pod openshift-route-controller-manager/route-controller-manager-6576b87f9c-4h9q7: failed to sync configmap cache: timed out waiting for the condition Mar 13 14:09:03 crc kubenswrapper[4907]: E0313 14:09:03.358566 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9bb1952d-0e55-4e45-81c0-69a8cddb827e-kube-api-access-gqtv7 podName:9bb1952d-0e55-4e45-81c0-69a8cddb827e nodeName:}" failed. No retries permitted until 2026-03-13 14:09:03.858537926 +0000 UTC m=+242.758325655 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-gqtv7" (UniqueName: "kubernetes.io/projected/9bb1952d-0e55-4e45-81c0-69a8cddb827e-kube-api-access-gqtv7") pod "route-controller-manager-6576b87f9c-4h9q7" (UID: "9bb1952d-0e55-4e45-81c0-69a8cddb827e") : failed to sync configmap cache: timed out waiting for the condition Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.376606 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 13 14:09:03 crc kubenswrapper[4907]: E0313 14:09:03.379631 4907 projected.go:194] Error preparing data for projected volume kube-api-access-gn5jt for pod openshift-controller-manager/controller-manager-879f6c89f-kwfhg: failed to sync configmap cache: timed out waiting for the condition Mar 13 14:09:03 crc kubenswrapper[4907]: E0313 14:09:03.379741 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/814c6dec-6e96-4068-9aec-9e57fefca5db-kube-api-access-gn5jt podName:814c6dec-6e96-4068-9aec-9e57fefca5db nodeName:}" failed. No retries permitted until 2026-03-13 14:09:03.879710352 +0000 UTC m=+242.779498081 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-gn5jt" (UniqueName: "kubernetes.io/projected/814c6dec-6e96-4068-9aec-9e57fefca5db-kube-api-access-gn5jt") pod "controller-manager-879f6c89f-kwfhg" (UID: "814c6dec-6e96-4068-9aec-9e57fefca5db") : failed to sync configmap cache: timed out waiting for the condition Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.395539 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:03 crc kubenswrapper[4907]: E0313 14:09:03.395663 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:03.895648105 +0000 UTC m=+242.795435784 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.396191 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:03 crc kubenswrapper[4907]: E0313 14:09:03.396845 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:03.896827497 +0000 UTC m=+242.796615226 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.416362 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.417422 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-v9m24" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.435686 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.436606 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-t68ql" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.455743 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.463055 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7jwmk\" (UniqueName: \"kubernetes.io/projected/b75dbfe3-8887-4a0b-9541-a4e4000924cb-kube-api-access-7jwmk\") pod \"machine-api-operator-5694c8668f-prlhd\" (UID: \"b75dbfe3-8887-4a0b-9541-a4e4000924cb\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-prlhd" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.476456 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.486355 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0396c426-1d56-44e6-885a-1b97d69ad57a-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-rmkfg\" (UID: \"0396c426-1d56-44e6-885a-1b97d69ad57a\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-rmkfg" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.496206 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.498791 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:03 crc kubenswrapper[4907]: E0313 14:09:03.499015 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:03.998990264 +0000 UTC m=+242.898777973 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.499834 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:03 crc kubenswrapper[4907]: E0313 14:09:03.500375 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:04.000359592 +0000 UTC m=+242.900147311 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.504441 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/59980734-2cf7-42dc-9f6f-a6de7e1c3665-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-56rqr\" (UID: \"59980734-2cf7-42dc-9f6f-a6de7e1c3665\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-56rqr" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.516306 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.521420 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/699da377-c61b-48f0-92e0-42bd247ee363-config\") pod \"etcd-operator-b45778765-hwf6r\" (UID: \"699da377-c61b-48f0-92e0-42bd247ee363\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hwf6r" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.538608 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.546367 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c26ccc9f-e682-4b94-9539-cb2607725f71-default-certificate\") pod \"router-default-5444994796-f8mmz\" (UID: \"c26ccc9f-e682-4b94-9539-cb2607725f71\") " pod="openshift-ingress/router-default-5444994796-f8mmz" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.556567 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.564239 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/699da377-c61b-48f0-92e0-42bd247ee363-serving-cert\") pod \"etcd-operator-b45778765-hwf6r\" (UID: \"699da377-c61b-48f0-92e0-42bd247ee363\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hwf6r" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.578080 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.585144 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/9c5d91bb-753e-4554-b7a1-6c85fe323875-registry-tls\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.601513 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:03 crc kubenswrapper[4907]: E0313 14:09:03.601712 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:04.101677736 +0000 UTC m=+243.001465425 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.602182 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:03 crc kubenswrapper[4907]: E0313 14:09:03.602525 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:04.102512319 +0000 UTC m=+243.002300008 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.616082 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.625153 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/10e8bf08-339f-4948-9fe2-ab4092d9a942-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-zz77x\" (UID: \"10e8bf08-339f-4948-9fe2-ab4092d9a942\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-zz77x" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.655372 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.662558 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/699da377-c61b-48f0-92e0-42bd247ee363-etcd-service-ca\") pod \"etcd-operator-b45778765-hwf6r\" (UID: \"699da377-c61b-48f0-92e0-42bd247ee363\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hwf6r" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.675213 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.685546 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/59980734-2cf7-42dc-9f6f-a6de7e1c3665-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-56rqr\" (UID: \"59980734-2cf7-42dc-9f6f-a6de7e1c3665\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-56rqr" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.696386 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.702671 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:03 crc kubenswrapper[4907]: E0313 14:09:03.702807 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:04.202786655 +0000 UTC m=+243.102574344 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.702987 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-57msq\" (UniqueName: \"kubernetes.io/projected/1be912b5-9d1d-48bd-a0ab-7f601f9eb9da-kube-api-access-57msq\") pod \"apiserver-7bbb656c7d-ksgr6\" (UID: \"1be912b5-9d1d-48bd-a0ab-7f601f9eb9da\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ksgr6" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.703103 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mtqz9\" (UniqueName: \"kubernetes.io/projected/3661f6b2-1d5e-42f0-947f-a32b66258a1f-kube-api-access-mtqz9\") pod \"oauth-openshift-558db77b4-ng4kz\" (UID: \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\") " pod="openshift-authentication/oauth-openshift-558db77b4-ng4kz" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.703337 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/699da377-c61b-48f0-92e0-42bd247ee363-etcd-ca\") pod \"etcd-operator-b45778765-hwf6r\" (UID: \"699da377-c61b-48f0-92e0-42bd247ee363\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hwf6r" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.703394 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:03 crc kubenswrapper[4907]: E0313 14:09:03.703655 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:04.203638508 +0000 UTC m=+243.103426197 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.706260 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-57msq\" (UniqueName: \"kubernetes.io/projected/1be912b5-9d1d-48bd-a0ab-7f601f9eb9da-kube-api-access-57msq\") pod \"apiserver-7bbb656c7d-ksgr6\" (UID: \"1be912b5-9d1d-48bd-a0ab-7f601f9eb9da\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ksgr6" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.706696 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mtqz9\" (UniqueName: \"kubernetes.io/projected/3661f6b2-1d5e-42f0-947f-a32b66258a1f-kube-api-access-mtqz9\") pod \"oauth-openshift-558db77b4-ng4kz\" (UID: \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\") " pod="openshift-authentication/oauth-openshift-558db77b4-ng4kz" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.716214 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.728077 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9b75954c-b75d-4f17-82dc-4c0358fd8d0f-serving-cert\") pod \"console-operator-58897d9998-w76kw\" (UID: \"9b75954c-b75d-4f17-82dc-4c0358fd8d0f\") " pod="openshift-console-operator/console-operator-58897d9998-w76kw" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.743557 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.753944 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9b75954c-b75d-4f17-82dc-4c0358fd8d0f-trusted-ca\") pod \"console-operator-58897d9998-w76kw\" (UID: \"9b75954c-b75d-4f17-82dc-4c0358fd8d0f\") " pod="openshift-console-operator/console-operator-58897d9998-w76kw" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.755496 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.766195 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c26ccc9f-e682-4b94-9539-cb2607725f71-stats-auth\") pod \"router-default-5444994796-f8mmz\" (UID: \"c26ccc9f-e682-4b94-9539-cb2607725f71\") " pod="openshift-ingress/router-default-5444994796-f8mmz" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.776154 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.783067 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b75954c-b75d-4f17-82dc-4c0358fd8d0f-config\") pod \"console-operator-58897d9998-w76kw\" (UID: \"9b75954c-b75d-4f17-82dc-4c0358fd8d0f\") " pod="openshift-console-operator/console-operator-58897d9998-w76kw" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.796317 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.804972 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:03 crc kubenswrapper[4907]: E0313 14:09:03.805124 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:04.305087447 +0000 UTC m=+243.204875136 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.805375 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hlg49\" (UniqueName: \"kubernetes.io/projected/5a9c889b-5954-4d3a-ae72-12000ee1cc25-kube-api-access-hlg49\") pod \"apiserver-76f77b778f-r86r4\" (UID: \"5a9c889b-5954-4d3a-ae72-12000ee1cc25\") " pod="openshift-apiserver/apiserver-76f77b778f-r86r4" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.805683 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:03 crc kubenswrapper[4907]: E0313 14:09:03.806112 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:04.306095553 +0000 UTC m=+243.205883242 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.809875 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c26ccc9f-e682-4b94-9539-cb2607725f71-metrics-certs\") pod \"router-default-5444994796-f8mmz\" (UID: \"c26ccc9f-e682-4b94-9539-cb2607725f71\") " pod="openshift-ingress/router-default-5444994796-f8mmz" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.813261 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hlg49\" (UniqueName: \"kubernetes.io/projected/5a9c889b-5954-4d3a-ae72-12000ee1cc25-kube-api-access-hlg49\") pod \"apiserver-76f77b778f-r86r4\" (UID: \"5a9c889b-5954-4d3a-ae72-12000ee1cc25\") " pod="openshift-apiserver/apiserver-76f77b778f-r86r4" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.837307 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.838942 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-v9m24"] Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.839946 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-t68ql"] Mar 13 14:09:03 crc kubenswrapper[4907]: W0313 14:09:03.845239 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod963f6709_e807_40bb_aa9e_9bba24e82b71.slice/crio-cfdd9cd399e1567ff60aac5628b25525ccb0e61a814a25a3fc9df094f6cb8837 WatchSource:0}: Error finding container cfdd9cd399e1567ff60aac5628b25525ccb0e61a814a25a3fc9df094f6cb8837: Status 404 returned error can't find the container with id cfdd9cd399e1567ff60aac5628b25525ccb0e61a814a25a3fc9df094f6cb8837 Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.845664 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c26ccc9f-e682-4b94-9539-cb2607725f71-service-ca-bundle\") pod \"router-default-5444994796-f8mmz\" (UID: \"c26ccc9f-e682-4b94-9539-cb2607725f71\") " pod="openshift-ingress/router-default-5444994796-f8mmz" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.906366 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:03 crc kubenswrapper[4907]: E0313 14:09:03.906520 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:04.406503883 +0000 UTC m=+243.306291572 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.906655 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s47gg\" (UniqueName: \"kubernetes.io/projected/71c234e9-5f3d-4b0c-94d4-80b4e33a6dda-kube-api-access-s47gg\") pod \"cluster-samples-operator-665b6dd947-wjrms\" (UID: \"71c234e9-5f3d-4b0c-94d4-80b4e33a6dda\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-wjrms" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.907272 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gn5jt\" (UniqueName: \"kubernetes.io/projected/814c6dec-6e96-4068-9aec-9e57fefca5db-kube-api-access-gn5jt\") pod \"controller-manager-879f6c89f-kwfhg\" (UID: \"814c6dec-6e96-4068-9aec-9e57fefca5db\") " pod="openshift-controller-manager/controller-manager-879f6c89f-kwfhg" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.907591 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gqtv7\" (UniqueName: \"kubernetes.io/projected/9bb1952d-0e55-4e45-81c0-69a8cddb827e-kube-api-access-gqtv7\") pod \"route-controller-manager-6576b87f9c-4h9q7\" (UID: \"9bb1952d-0e55-4e45-81c0-69a8cddb827e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4h9q7" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.907958 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.908113 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9qt98\" (UniqueName: \"kubernetes.io/projected/005dfe54-e402-4aa4-8b67-f7e7b685debd-kube-api-access-9qt98\") pod \"machine-approver-56656f9798-52vhb\" (UID: \"005dfe54-e402-4aa4-8b67-f7e7b685debd\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-52vhb" Mar 13 14:09:03 crc kubenswrapper[4907]: E0313 14:09:03.908300 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:04.408282671 +0000 UTC m=+243.308070410 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.910561 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gqtv7\" (UniqueName: \"kubernetes.io/projected/9bb1952d-0e55-4e45-81c0-69a8cddb827e-kube-api-access-gqtv7\") pod \"route-controller-manager-6576b87f9c-4h9q7\" (UID: \"9bb1952d-0e55-4e45-81c0-69a8cddb827e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4h9q7" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.910620 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gn5jt\" (UniqueName: \"kubernetes.io/projected/814c6dec-6e96-4068-9aec-9e57fefca5db-kube-api-access-gn5jt\") pod \"controller-manager-879f6c89f-kwfhg\" (UID: \"814c6dec-6e96-4068-9aec-9e57fefca5db\") " pod="openshift-controller-manager/controller-manager-879f6c89f-kwfhg" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.910631 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s47gg\" (UniqueName: \"kubernetes.io/projected/71c234e9-5f3d-4b0c-94d4-80b4e33a6dda-kube-api-access-s47gg\") pod \"cluster-samples-operator-665b6dd947-wjrms\" (UID: \"71c234e9-5f3d-4b0c-94d4-80b4e33a6dda\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-wjrms" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.910862 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9c5d91bb-753e-4554-b7a1-6c85fe323875-bound-sa-token\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.912803 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9qt98\" (UniqueName: \"kubernetes.io/projected/005dfe54-e402-4aa4-8b67-f7e7b685debd-kube-api-access-9qt98\") pod \"machine-approver-56656f9798-52vhb\" (UID: \"005dfe54-e402-4aa4-8b67-f7e7b685debd\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-52vhb" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.936594 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.943634 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0149e083-ed5b-4c7f-94e7-4f3f2ff8a557-serving-cert\") pod \"openshift-config-operator-7777fb866f-wgpdd\" (UID: \"0149e083-ed5b-4c7f-94e7-4f3f2ff8a557\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-wgpdd" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.955362 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.964633 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/ebd59055-333a-46c2-ba30-826296414a4c-metrics-tls\") pod \"dns-operator-744455d44c-xkdjh\" (UID: \"ebd59055-333a-46c2-ba30-826296414a4c\") " pod="openshift-dns-operator/dns-operator-744455d44c-xkdjh" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.975047 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.982609 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/9c5d91bb-753e-4554-b7a1-6c85fe323875-installation-pull-secrets\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.009695 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:04 crc kubenswrapper[4907]: E0313 14:09:04.009834 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:04.509817452 +0000 UTC m=+243.409605141 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.010079 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:04 crc kubenswrapper[4907]: E0313 14:09:04.010413 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:04.510399358 +0000 UTC m=+243.410187057 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.015740 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.031445 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/699da377-c61b-48f0-92e0-42bd247ee363-etcd-client\") pod \"etcd-operator-b45778765-hwf6r\" (UID: \"699da377-c61b-48f0-92e0-42bd247ee363\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hwf6r" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.070016 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jb2r5\" (UniqueName: \"kubernetes.io/projected/9c5d91bb-753e-4554-b7a1-6c85fe323875-kube-api-access-jb2r5\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.095741 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.102566 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/10e8bf08-339f-4948-9fe2-ab4092d9a942-config\") pod \"kube-controller-manager-operator-78b949d7b-zz77x\" (UID: \"10e8bf08-339f-4948-9fe2-ab4092d9a942\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-zz77x" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.112231 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:04 crc kubenswrapper[4907]: E0313 14:09:04.112430 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:04.612391051 +0000 UTC m=+243.512178760 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.112844 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:04 crc kubenswrapper[4907]: E0313 14:09:04.113218 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:04.613201493 +0000 UTC m=+243.512989202 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.116528 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.121575 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0396c426-1d56-44e6-885a-1b97d69ad57a-config\") pod \"kube-apiserver-operator-766d6c64bb-rmkfg\" (UID: \"0396c426-1d56-44e6-885a-1b97d69ad57a\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-rmkfg" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.135937 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.146131 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dslzm\" (UniqueName: \"kubernetes.io/projected/ae320273-06e8-43c5-a64f-acd80ad16d1c-kube-api-access-dslzm\") pod \"ingress-operator-5b745b69d9-5szhc\" (UID: \"ae320273-06e8-43c5-a64f-acd80ad16d1c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5szhc" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.155965 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.167558 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rnmbh\" (UniqueName: \"kubernetes.io/projected/a7b66665-221d-4334-8263-aeb3b20e1c92-kube-api-access-rnmbh\") pod \"openshift-controller-manager-operator-756b6f6bc6-hfcsw\" (UID: \"a7b66665-221d-4334-8263-aeb3b20e1c92\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hfcsw" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.195326 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hnbr7\" (UniqueName: \"kubernetes.io/projected/7810633c-9d54-4cbe-8861-b523f724ec9e-kube-api-access-hnbr7\") pod \"catalog-operator-68c6474976-fwcwx\" (UID: \"7810633c-9d54-4cbe-8861-b523f724ec9e\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fwcwx" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.213259 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wq2kr\" (UniqueName: \"kubernetes.io/projected/1954f507-7cb0-4942-a4c8-6586d370e717-kube-api-access-wq2kr\") pod \"service-ca-9c57cc56f-zwzc4\" (UID: \"1954f507-7cb0-4942-a4c8-6586d370e717\") " pod="openshift-service-ca/service-ca-9c57cc56f-zwzc4" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.213871 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:04 crc kubenswrapper[4907]: E0313 14:09:04.214141 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:04.714111626 +0000 UTC m=+243.613899355 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.214648 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:04 crc kubenswrapper[4907]: E0313 14:09:04.215060 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:04.715042742 +0000 UTC m=+243.614830501 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.228484 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-t68ql" event={"ID":"a8e9ea4a-43ad-4da0-99fa-c01eea712c8c","Type":"ContainerStarted","Data":"f7123489a4da754753944b4f46a995a572d6483f103454fc1d99c5f31f23fc56"} Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.228531 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-t68ql" event={"ID":"a8e9ea4a-43ad-4da0-99fa-c01eea712c8c","Type":"ContainerStarted","Data":"0c7c592b44c27e210c582853669a9bdfbaaf66d54ff54f806a5a7d997b04aaff"} Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.228545 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-t68ql" event={"ID":"a8e9ea4a-43ad-4da0-99fa-c01eea712c8c","Type":"ContainerStarted","Data":"7becc3bfa86aaf701a6ac61da68318dbe069d09afac1a898ddc8d52a3fe71572"} Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.230308 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-rnh5t" event={"ID":"a3baeb4f-988d-4d42-b469-a310ca127745","Type":"ContainerStarted","Data":"093846580b2ef04d12bd65ee80e3c459927d2923b16915bfe3e4a7dabe9c4e6a"} Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.231985 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-v9m24" event={"ID":"963f6709-e807-40bb-aa9e-9bba24e82b71","Type":"ContainerStarted","Data":"227dc9775ef9c7db63ec63427d557314bc07d97ac6e9b074d22c6fedfee02153"} Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.232012 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-v9m24" event={"ID":"963f6709-e807-40bb-aa9e-9bba24e82b71","Type":"ContainerStarted","Data":"cfdd9cd399e1567ff60aac5628b25525ccb0e61a814a25a3fc9df094f6cb8837"} Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.238856 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tc8b2\" (UniqueName: \"kubernetes.io/projected/3a661298-271c-4c57-94ea-3d69a27d7ad5-kube-api-access-tc8b2\") pod \"control-plane-machine-set-operator-78cbb6b69f-djph7\" (UID: \"3a661298-271c-4c57-94ea-3d69a27d7ad5\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-djph7" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.251268 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9f5cd\" (UniqueName: \"kubernetes.io/projected/a8867781-179d-47e4-9a55-330deb10d273-kube-api-access-9f5cd\") pod \"auto-csr-approver-29556848-v8kv4\" (UID: \"a8867781-179d-47e4-9a55-330deb10d273\") " pod="openshift-infra/auto-csr-approver-29556848-v8kv4" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.274874 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9pzkh\" (UniqueName: \"kubernetes.io/projected/7e89f101-cb61-4117-841d-58b5cc2bdd80-kube-api-access-9pzkh\") pod \"packageserver-d55dfcdfc-pk65f\" (UID: \"7e89f101-cb61-4117-841d-58b5cc2bdd80\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pk65f" Mar 13 14:09:04 crc kubenswrapper[4907]: E0313 14:09:04.294133 4907 secret.go:188] Couldn't get secret openshift-machine-api/control-plane-machine-set-operator-tls: failed to sync secret cache: timed out waiting for the condition Mar 13 14:09:04 crc kubenswrapper[4907]: E0313 14:09:04.294245 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3a661298-271c-4c57-94ea-3d69a27d7ad5-control-plane-machine-set-operator-tls podName:3a661298-271c-4c57-94ea-3d69a27d7ad5 nodeName:}" failed. No retries permitted until 2026-03-13 14:09:04.794219354 +0000 UTC m=+243.694007113 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "control-plane-machine-set-operator-tls" (UniqueName: "kubernetes.io/secret/3a661298-271c-4c57-94ea-3d69a27d7ad5-control-plane-machine-set-operator-tls") pod "control-plane-machine-set-operator-78cbb6b69f-djph7" (UID: "3a661298-271c-4c57-94ea-3d69a27d7ad5") : failed to sync secret cache: timed out waiting for the condition Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.294163 4907 request.go:700] Waited for 1.000276454s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/serviceaccounts/marketplace-operator/token Mar 13 14:09:04 crc kubenswrapper[4907]: E0313 14:09:04.296471 4907 configmap.go:193] Couldn't get configMap openshift-machine-config-operator/machine-config-operator-images: failed to sync configmap cache: timed out waiting for the condition Mar 13 14:09:04 crc kubenswrapper[4907]: E0313 14:09:04.296529 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/74b29e18-7d73-4ef5-9739-04137dd1b191-images podName:74b29e18-7d73-4ef5-9739-04137dd1b191 nodeName:}" failed. No retries permitted until 2026-03-13 14:09:04.796517747 +0000 UTC m=+243.696305436 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "images" (UniqueName: "kubernetes.io/configmap/74b29e18-7d73-4ef5-9739-04137dd1b191-images") pod "machine-config-operator-74547568cd-92gs7" (UID: "74b29e18-7d73-4ef5-9739-04137dd1b191") : failed to sync configmap cache: timed out waiting for the condition Mar 13 14:09:04 crc kubenswrapper[4907]: E0313 14:09:04.300321 4907 secret.go:188] Couldn't get secret openshift-machine-config-operator/mco-proxy-tls: failed to sync secret cache: timed out waiting for the condition Mar 13 14:09:04 crc kubenswrapper[4907]: E0313 14:09:04.300353 4907 secret.go:188] Couldn't get secret openshift-multus/multus-admission-controller-secret: failed to sync secret cache: timed out waiting for the condition Mar 13 14:09:04 crc kubenswrapper[4907]: E0313 14:09:04.300399 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/74b29e18-7d73-4ef5-9739-04137dd1b191-proxy-tls podName:74b29e18-7d73-4ef5-9739-04137dd1b191 nodeName:}" failed. No retries permitted until 2026-03-13 14:09:04.800383081 +0000 UTC m=+243.700170770 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "proxy-tls" (UniqueName: "kubernetes.io/secret/74b29e18-7d73-4ef5-9739-04137dd1b191-proxy-tls") pod "machine-config-operator-74547568cd-92gs7" (UID: "74b29e18-7d73-4ef5-9739-04137dd1b191") : failed to sync secret cache: timed out waiting for the condition Mar 13 14:09:04 crc kubenswrapper[4907]: E0313 14:09:04.300419 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8013cba2-6b3c-4548-af5c-6dd0057da0c7-webhook-certs podName:8013cba2-6b3c-4548-af5c-6dd0057da0c7 nodeName:}" failed. No retries permitted until 2026-03-13 14:09:04.800412532 +0000 UTC m=+243.700200221 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/8013cba2-6b3c-4548-af5c-6dd0057da0c7-webhook-certs") pod "multus-admission-controller-857f4d67dd-ksfb8" (UID: "8013cba2-6b3c-4548-af5c-6dd0057da0c7") : failed to sync secret cache: timed out waiting for the condition Mar 13 14:09:04 crc kubenswrapper[4907]: E0313 14:09:04.301818 4907 projected.go:288] Couldn't get configMap openshift-console/openshift-service-ca.crt: failed to sync configmap cache: timed out waiting for the condition Mar 13 14:09:04 crc kubenswrapper[4907]: E0313 14:09:04.301861 4907 projected.go:194] Error preparing data for projected volume kube-api-access-vxlx4 for pod openshift-console/console-f9d7485db-l9xpb: failed to sync configmap cache: timed out waiting for the condition Mar 13 14:09:04 crc kubenswrapper[4907]: E0313 14:09:04.301968 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/4c8c0266-5452-4abb-92c5-c536df94dd41-kube-api-access-vxlx4 podName:4c8c0266-5452-4abb-92c5-c536df94dd41 nodeName:}" failed. No retries permitted until 2026-03-13 14:09:04.801939144 +0000 UTC m=+243.701726903 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-vxlx4" (UniqueName: "kubernetes.io/projected/4c8c0266-5452-4abb-92c5-c536df94dd41-kube-api-access-vxlx4") pod "console-f9d7485db-l9xpb" (UID: "4c8c0266-5452-4abb-92c5-c536df94dd41") : failed to sync configmap cache: timed out waiting for the condition Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.304066 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pk65f" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.314450 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qppp6\" (UniqueName: \"kubernetes.io/projected/0290eb6f-0be7-4340-8ba8-ed6ea0662c33-kube-api-access-qppp6\") pod \"marketplace-operator-79b997595-5stnc\" (UID: \"0290eb6f-0be7-4340-8ba8-ed6ea0662c33\") " pod="openshift-marketplace/marketplace-operator-79b997595-5stnc" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.315982 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:04 crc kubenswrapper[4907]: E0313 14:09:04.316295 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:04.816266614 +0000 UTC m=+243.716054343 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.317792 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:04 crc kubenswrapper[4907]: E0313 14:09:04.318394 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:04.818380231 +0000 UTC m=+243.718167920 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.331258 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n5x9f\" (UniqueName: \"kubernetes.io/projected/ff0e893f-05cb-4751-9960-281c58000519-kube-api-access-n5x9f\") pod \"service-ca-operator-777779d784-75gdt\" (UID: \"ff0e893f-05cb-4751-9960-281c58000519\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-75gdt" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.331631 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fwcwx" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.353644 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-75gdt" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.363664 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dwlcl\" (UniqueName: \"kubernetes.io/projected/fa265490-c23b-4b05-94bc-78b4a636e065-kube-api-access-dwlcl\") pod \"dns-default-8jqgq\" (UID: \"fa265490-c23b-4b05-94bc-78b4a636e065\") " pod="openshift-dns/dns-default-8jqgq" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.371196 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qllsp\" (UniqueName: \"kubernetes.io/projected/4a8597a4-c72f-4023-b2fb-1b1237411a36-kube-api-access-qllsp\") pod \"csi-hostpathplugin-sjwvn\" (UID: \"4a8597a4-c72f-4023-b2fb-1b1237411a36\") " pod="hostpath-provisioner/csi-hostpathplugin-sjwvn" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.376030 4907 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-hh9sk" secret="" err="failed to sync secret cache: timed out waiting for the condition" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.376087 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-hh9sk" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.379639 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556848-v8kv4" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.389990 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-zwzc4" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.397937 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zhdhq\" (UniqueName: \"kubernetes.io/projected/ff214b86-eb3d-46c6-b45a-18a9f63e742c-kube-api-access-zhdhq\") pod \"ingress-canary-252nq\" (UID: \"ff214b86-eb3d-46c6-b45a-18a9f63e742c\") " pod="openshift-ingress-canary/ingress-canary-252nq" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.417502 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.418596 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:04 crc kubenswrapper[4907]: E0313 14:09:04.418742 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:04.918720909 +0000 UTC m=+243.818508608 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.418968 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:04 crc kubenswrapper[4907]: E0313 14:09:04.419489 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:04.91947299 +0000 UTC m=+243.819260689 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.425271 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bgp25\" (UniqueName: \"kubernetes.io/projected/4682bbe1-4257-4d04-a6d7-60c4655d6873-kube-api-access-bgp25\") pod \"olm-operator-6b444d44fb-hkbqt\" (UID: \"4682bbe1-4257-4d04-a6d7-60c4655d6873\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hkbqt" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.430985 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-252nq" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.446775 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-sjwvn" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.461551 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jgwz8\" (UniqueName: \"kubernetes.io/projected/c77fd554-e957-4d49-b8cd-b4f7dcd4ba2d-kube-api-access-jgwz8\") pod \"machine-config-server-g6q2r\" (UID: \"c77fd554-e957-4d49-b8cd-b4f7dcd4ba2d\") " pod="openshift-machine-config-operator/machine-config-server-g6q2r" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.479710 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-8jqgq" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.481922 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.482152 4907 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openshift-machine-api/machine-api-operator-5694c8668f-prlhd" secret="" err="failed to sync secret cache: timed out waiting for the condition" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.482181 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-prlhd" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.484413 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lx6xs\" (UniqueName: \"kubernetes.io/projected/8013cba2-6b3c-4548-af5c-6dd0057da0c7-kube-api-access-lx6xs\") pod \"multus-admission-controller-857f4d67dd-ksfb8\" (UID: \"8013cba2-6b3c-4548-af5c-6dd0057da0c7\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-ksfb8" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.513765 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lfn75\" (UniqueName: \"kubernetes.io/projected/c934a2b9-3585-4491-8962-df085c473d90-kube-api-access-lfn75\") pod \"openshift-apiserver-operator-796bbdcf4f-xnr2b\" (UID: \"c934a2b9-3585-4491-8962-df085c473d90\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xnr2b" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.518591 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pk65f"] Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.521393 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:04 crc kubenswrapper[4907]: E0313 14:09:04.521550 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:05.021528063 +0000 UTC m=+243.921315752 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.522324 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:04 crc kubenswrapper[4907]: E0313 14:09:04.522560 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:05.022551962 +0000 UTC m=+243.922339651 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.535150 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9qzhc\" (UniqueName: \"kubernetes.io/projected/74b29e18-7d73-4ef5-9739-04137dd1b191-kube-api-access-9qzhc\") pod \"machine-config-operator-74547568cd-92gs7\" (UID: \"74b29e18-7d73-4ef5-9739-04137dd1b191\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-92gs7" Mar 13 14:09:04 crc kubenswrapper[4907]: W0313 14:09:04.535972 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7e89f101_cb61_4117_841d_58b5cc2bdd80.slice/crio-5b8f9375c0aaaf442ecdf173ca5f3a8f53bf0c84929853df0d2803ae5825f27a WatchSource:0}: Error finding container 5b8f9375c0aaaf442ecdf173ca5f3a8f53bf0c84929853df0d2803ae5825f27a: Status 404 returned error can't find the container with id 5b8f9375c0aaaf442ecdf173ca5f3a8f53bf0c84929853df0d2803ae5825f27a Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.559491 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h9w5d\" (UniqueName: \"kubernetes.io/projected/19edf37b-a1d3-4c7c-945e-5d0b8bb601bf-kube-api-access-h9w5d\") pod \"collect-profiles-29556840-pgmwj\" (UID: \"19edf37b-a1d3-4c7c-945e-5d0b8bb601bf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556840-pgmwj" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.575645 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.580028 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dtbj8\" (UniqueName: \"kubernetes.io/projected/b651a92c-e9f6-4a1a-8c49-b6500bb7c61c-kube-api-access-dtbj8\") pod \"package-server-manager-789f6589d5-ccjsv\" (UID: \"b651a92c-e9f6-4a1a-8c49-b6500bb7c61c\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ccjsv" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.588820 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-5stnc" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.598349 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ccjsv" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.601490 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Mar 13 14:09:04 crc kubenswrapper[4907]: E0313 14:09:04.607301 4907 projected.go:288] Couldn't get configMap openshift-kube-apiserver-operator/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Mar 13 14:09:04 crc kubenswrapper[4907]: E0313 14:09:04.607331 4907 projected.go:194] Error preparing data for projected volume kube-api-access for pod openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-rmkfg: failed to sync configmap cache: timed out waiting for the condition Mar 13 14:09:04 crc kubenswrapper[4907]: E0313 14:09:04.607384 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/0396c426-1d56-44e6-885a-1b97d69ad57a-kube-api-access podName:0396c426-1d56-44e6-885a-1b97d69ad57a nodeName:}" failed. No retries permitted until 2026-03-13 14:09:05.107366318 +0000 UTC m=+244.007154007 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access" (UniqueName: "kubernetes.io/projected/0396c426-1d56-44e6-885a-1b97d69ad57a-kube-api-access") pod "kube-apiserver-operator-766d6c64bb-rmkfg" (UID: "0396c426-1d56-44e6-885a-1b97d69ad57a") : failed to sync configmap cache: timed out waiting for the condition Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.615655 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fwcwx"] Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.616471 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.621983 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hkbqt" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.623494 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:04 crc kubenswrapper[4907]: E0313 14:09:04.623931 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:05.123912758 +0000 UTC m=+244.023700447 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.626730 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j6nzw\" (UniqueName: \"kubernetes.io/projected/f8ea4f24-05b0-4661-a617-a4208f9a9188-kube-api-access-j6nzw\") pod \"downloads-7954f5f757-544df\" (UID: \"f8ea4f24-05b0-4661-a617-a4208f9a9188\") " pod="openshift-console/downloads-7954f5f757-544df" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.640281 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.642191 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556840-pgmwj" Mar 13 14:09:04 crc kubenswrapper[4907]: E0313 14:09:04.647933 4907 projected.go:288] Couldn't get configMap openshift-dns-operator/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.677177 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.697701 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.699559 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xnr2b" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.715158 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-g6q2r" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.717286 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.735686 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:04 crc kubenswrapper[4907]: E0313 14:09:04.736291 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:05.236272822 +0000 UTC m=+244.136060511 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.738420 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.744574 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-ng4kz" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.796554 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.799554 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.799581 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.813114 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ksgr6" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.835310 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.836332 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.836516 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.836535 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/8013cba2-6b3c-4548-af5c-6dd0057da0c7-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-ksfb8\" (UID: \"8013cba2-6b3c-4548-af5c-6dd0057da0c7\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-ksfb8" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.836681 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/3a661298-271c-4c57-94ea-3d69a27d7ad5-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-djph7\" (UID: \"3a661298-271c-4c57-94ea-3d69a27d7ad5\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-djph7" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.836717 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/74b29e18-7d73-4ef5-9739-04137dd1b191-proxy-tls\") pod \"machine-config-operator-74547568cd-92gs7\" (UID: \"74b29e18-7d73-4ef5-9739-04137dd1b191\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-92gs7" Mar 13 14:09:04 crc kubenswrapper[4907]: E0313 14:09:04.836799 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:05.336760274 +0000 UTC m=+244.236547963 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.836931 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vxlx4\" (UniqueName: \"kubernetes.io/projected/4c8c0266-5452-4abb-92c5-c536df94dd41-kube-api-access-vxlx4\") pod \"console-f9d7485db-l9xpb\" (UID: \"4c8c0266-5452-4abb-92c5-c536df94dd41\") " pod="openshift-console/console-f9d7485db-l9xpb" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.836977 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/74b29e18-7d73-4ef5-9739-04137dd1b191-images\") pod \"machine-config-operator-74547568cd-92gs7\" (UID: \"74b29e18-7d73-4ef5-9739-04137dd1b191\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-92gs7" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.839306 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/74b29e18-7d73-4ef5-9739-04137dd1b191-images\") pod \"machine-config-operator-74547568cd-92gs7\" (UID: \"74b29e18-7d73-4ef5-9739-04137dd1b191\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-92gs7" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.848265 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vxlx4\" (UniqueName: \"kubernetes.io/projected/4c8c0266-5452-4abb-92c5-c536df94dd41-kube-api-access-vxlx4\") pod \"console-f9d7485db-l9xpb\" (UID: \"4c8c0266-5452-4abb-92c5-c536df94dd41\") " pod="openshift-console/console-f9d7485db-l9xpb" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.849072 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/74b29e18-7d73-4ef5-9739-04137dd1b191-proxy-tls\") pod \"machine-config-operator-74547568cd-92gs7\" (UID: \"74b29e18-7d73-4ef5-9739-04137dd1b191\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-92gs7" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.855147 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/8013cba2-6b3c-4548-af5c-6dd0057da0c7-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-ksfb8\" (UID: \"8013cba2-6b3c-4548-af5c-6dd0057da0c7\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-ksfb8" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.855303 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/3a661298-271c-4c57-94ea-3d69a27d7ad5-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-djph7\" (UID: \"3a661298-271c-4c57-94ea-3d69a27d7ad5\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-djph7" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.856506 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.865845 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-kwfhg" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.880692 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.888210 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-djph7" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.888553 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4h9q7" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.900110 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.915924 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.917637 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-r86r4" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.938039 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.938867 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:04 crc kubenswrapper[4907]: E0313 14:09:04.941562 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:05.441547652 +0000 UTC m=+244.341335341 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.952389 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/10e8bf08-339f-4948-9fe2-ab4092d9a942-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-zz77x\" (UID: \"10e8bf08-339f-4948-9fe2-ab4092d9a942\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-zz77x" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.956348 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.965370 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-52vhb" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.976280 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.978362 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-hh9sk"] Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.980698 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-75gdt"] Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.983974 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-wjrms" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.986320 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556848-v8kv4"] Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.995958 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.004183 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5szhc" Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.016311 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.035954 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.043015 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hfcsw" Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.043431 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:05 crc kubenswrapper[4907]: E0313 14:09:05.043898 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:05.543858254 +0000 UTC m=+244.443645933 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.046653 4907 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.065154 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Mar 13 14:09:05 crc kubenswrapper[4907]: E0313 14:09:05.068621 4907 projected.go:194] Error preparing data for projected volume kube-api-access-wk2bb for pod openshift-dns-operator/dns-operator-744455d44c-xkdjh: failed to sync configmap cache: timed out waiting for the condition Mar 13 14:09:05 crc kubenswrapper[4907]: E0313 14:09:05.068745 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/ebd59055-333a-46c2-ba30-826296414a4c-kube-api-access-wk2bb podName:ebd59055-333a-46c2-ba30-826296414a4c nodeName:}" failed. No retries permitted until 2026-03-13 14:09:05.568718 +0000 UTC m=+244.468505689 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-wk2bb" (UniqueName: "kubernetes.io/projected/ebd59055-333a-46c2-ba30-826296414a4c-kube-api-access-wk2bb") pod "dns-operator-744455d44c-xkdjh" (UID: "ebd59055-333a-46c2-ba30-826296414a4c") : failed to sync configmap cache: timed out waiting for the condition Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.071325 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-zwzc4"] Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.085652 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-prlhd"] Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.090372 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-252nq"] Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.096490 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.102191 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-sjwvn"] Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.104301 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-544df" Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.116136 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.121008 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vcq96\" (UniqueName: \"kubernetes.io/projected/9b75954c-b75d-4f17-82dc-4c0358fd8d0f-kube-api-access-vcq96\") pod \"console-operator-58897d9998-w76kw\" (UID: \"9b75954c-b75d-4f17-82dc-4c0358fd8d0f\") " pod="openshift-console-operator/console-operator-58897d9998-w76kw" Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.140939 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Mar 13 14:09:05 crc kubenswrapper[4907]: W0313 14:09:05.141240 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode9d44a1e_0eab_4bd6_833f_143436e43758.slice/crio-23e4a4539e1b648af539df572b036a3e5ceebfe6280418e6d7f8719a1fd54f93 WatchSource:0}: Error finding container 23e4a4539e1b648af539df572b036a3e5ceebfe6280418e6d7f8719a1fd54f93: Status 404 returned error can't find the container with id 23e4a4539e1b648af539df572b036a3e5ceebfe6280418e6d7f8719a1fd54f93 Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.147004 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q6bl2\" (UniqueName: \"kubernetes.io/projected/c26ccc9f-e682-4b94-9539-cb2607725f71-kube-api-access-q6bl2\") pod \"router-default-5444994796-f8mmz\" (UID: \"c26ccc9f-e682-4b94-9539-cb2607725f71\") " pod="openshift-ingress/router-default-5444994796-f8mmz" Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.147005 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.147169 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0396c426-1d56-44e6-885a-1b97d69ad57a-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-rmkfg\" (UID: \"0396c426-1d56-44e6-885a-1b97d69ad57a\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-rmkfg" Mar 13 14:09:05 crc kubenswrapper[4907]: E0313 14:09:05.147277 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:05.647264335 +0000 UTC m=+244.547052024 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.156521 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0396c426-1d56-44e6-885a-1b97d69ad57a-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-rmkfg\" (UID: \"0396c426-1d56-44e6-885a-1b97d69ad57a\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-rmkfg" Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.158266 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.173749 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xwfx4\" (UniqueName: \"kubernetes.io/projected/699da377-c61b-48f0-92e0-42bd247ee363-kube-api-access-xwfx4\") pod \"etcd-operator-b45778765-hwf6r\" (UID: \"699da377-c61b-48f0-92e0-42bd247ee363\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hwf6r" Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.176375 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.195154 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5d6cj\" (UniqueName: \"kubernetes.io/projected/59980734-2cf7-42dc-9f6f-a6de7e1c3665-kube-api-access-5d6cj\") pod \"kube-storage-version-migrator-operator-b67b599dd-56rqr\" (UID: \"59980734-2cf7-42dc-9f6f-a6de7e1c3665\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-56rqr" Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.201573 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.201760 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-92gs7" Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.216704 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.228759 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556840-pgmwj"] Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.229584 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hkbqt"] Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.231799 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8ctrn\" (UniqueName: \"kubernetes.io/projected/0149e083-ed5b-4c7f-94e7-4f3f2ff8a557-kube-api-access-8ctrn\") pod \"openshift-config-operator-7777fb866f-wgpdd\" (UID: \"0149e083-ed5b-4c7f-94e7-4f3f2ff8a557\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-wgpdd" Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.245076 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.253275 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:05 crc kubenswrapper[4907]: E0313 14:09:05.253647 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:05.753632197 +0000 UTC m=+244.653419886 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.260495 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gnmnd\" (UniqueName: \"kubernetes.io/projected/814322e3-6251-4ea3-a86b-7b89ca6ea728-kube-api-access-gnmnd\") pod \"migrator-59844c95c7-vfbsx\" (UID: \"814322e3-6251-4ea3-a86b-7b89ca6ea728\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-vfbsx" Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.266009 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.269529 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-l9xpb" Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.275559 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xnr2b"] Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.288720 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-8jqgq"] Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.288776 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-ksgr6"] Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.298346 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-ng4kz"] Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.301453 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.305760 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-52vhb" event={"ID":"005dfe54-e402-4aa4-8b67-f7e7b685debd","Type":"ContainerStarted","Data":"2640caee6d284f0bc297ee7eb18a0e186b53f36e618ca72302ac3834c8f7aeaa"} Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.307004 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556848-v8kv4" event={"ID":"a8867781-179d-47e4-9a55-330deb10d273","Type":"ContainerStarted","Data":"38620171197d496fb764e2284a8169d5eff52a1213907791553d488eaad39d9d"} Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.309290 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-zz77x" Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.316198 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.316508 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-ksfb8" Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.331284 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pk65f" event={"ID":"7e89f101-cb61-4117-841d-58b5cc2bdd80","Type":"ContainerStarted","Data":"ed7d771c7834263888b56c54ee8c8ea22b90724c2d98378a98feddf4aea2b65b"} Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.331329 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pk65f" event={"ID":"7e89f101-cb61-4117-841d-58b5cc2bdd80","Type":"ContainerStarted","Data":"5b8f9375c0aaaf442ecdf173ca5f3a8f53bf0c84929853df0d2803ae5825f27a"} Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.332434 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pk65f" Mar 13 14:09:05 crc kubenswrapper[4907]: W0313 14:09:05.333508 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod19edf37b_a1d3_4c7c_945e_5d0b8bb601bf.slice/crio-e8a5550c81aa8c4bb6491a89a3041cb9a6e5701c9a00e500620d96a20f452a36 WatchSource:0}: Error finding container e8a5550c81aa8c4bb6491a89a3041cb9a6e5701c9a00e500620d96a20f452a36: Status 404 returned error can't find the container with id e8a5550c81aa8c4bb6491a89a3041cb9a6e5701c9a00e500620d96a20f452a36 Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.337493 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-75gdt" event={"ID":"ff0e893f-05cb-4751-9960-281c58000519","Type":"ContainerStarted","Data":"4508a6e2419581abd5918c50470d548f610ef0c89853506353a8483b8ea3015d"} Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.340168 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-5stnc"] Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.341417 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ccjsv"] Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.343182 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-hh9sk" event={"ID":"e9d44a1e-0eab-4bd6-833f-143436e43758","Type":"ContainerStarted","Data":"23e4a4539e1b648af539df572b036a3e5ceebfe6280418e6d7f8719a1fd54f93"} Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.343898 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-sjwvn" event={"ID":"4a8597a4-c72f-4023-b2fb-1b1237411a36","Type":"ContainerStarted","Data":"a38dcdd81d9bc13e5a8dae158690d917c7091d0f83e3c30429ea70f671e8035f"} Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.352550 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fwcwx" event={"ID":"7810633c-9d54-4cbe-8861-b523f724ec9e","Type":"ContainerStarted","Data":"afd7d366e1cbef135eee0552758925fb36fb60a4c43a2c12c80b73605f7cb0fa"} Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.352701 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fwcwx" event={"ID":"7810633c-9d54-4cbe-8861-b523f724ec9e","Type":"ContainerStarted","Data":"6ef82b59f33daccea1490dc3bef81544cb46172092caf29fdc1c89470a456854"} Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.353123 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fwcwx" Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.354648 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:05 crc kubenswrapper[4907]: E0313 14:09:05.355560 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:05.855546818 +0000 UTC m=+244.755334507 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.364021 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.367604 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-wgpdd" Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.368789 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-g6q2r" event={"ID":"c77fd554-e957-4d49-b8cd-b4f7dcd4ba2d","Type":"ContainerStarted","Data":"f18d9e133ffdf3c70e7a18735c7450c5821e6f17755cb96b4dc3e0af638b395c"} Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.368820 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-g6q2r" event={"ID":"c77fd554-e957-4d49-b8cd-b4f7dcd4ba2d","Type":"ContainerStarted","Data":"9d2d79f538bec30c7e5525cc958a7eb716f40a8b39789138099658c3006687d4"} Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.376064 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-252nq" event={"ID":"ff214b86-eb3d-46c6-b45a-18a9f63e742c","Type":"ContainerStarted","Data":"eeeb42502a81f1b22e077bafdd88ba9c9f5f66a3268d2f62ba58eea461cb1789"} Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.385011 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-prlhd" event={"ID":"b75dbfe3-8887-4a0b-9541-a4e4000924cb","Type":"ContainerStarted","Data":"e7d019e94109b87308c200c5fd282b0abdb2116ded6cffd570ca3ed83675f3ad"} Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.393774 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-zwzc4" event={"ID":"1954f507-7cb0-4942-a4c8-6586d370e717","Type":"ContainerStarted","Data":"b2c639cb1f8fc378273d3d9722547069101a439330919f458fbd4e276e30511f"} Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.398701 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.410019 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-w76kw" Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.438572 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.446823 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-hwf6r" Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.450659 4907 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-pk65f container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.36:5443/healthz\": dial tcp 10.217.0.36:5443: connect: connection refused" start-of-body= Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.450746 4907 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pk65f" podUID="7e89f101-cb61-4117-841d-58b5cc2bdd80" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.36:5443/healthz\": dial tcp 10.217.0.36:5443: connect: connection refused" Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.455987 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:05 crc kubenswrapper[4907]: E0313 14:09:05.456431 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:05.95640496 +0000 UTC m=+244.856192649 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.459308 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.459408 4907 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-fwcwx container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.32:8443/healthz\": dial tcp 10.217.0.32:8443: connect: connection refused" start-of-body= Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.459471 4907 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fwcwx" podUID="7810633c-9d54-4cbe-8861-b523f724ec9e" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.32:8443/healthz\": dial tcp 10.217.0.32:8443: connect: connection refused" Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.460532 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-f8mmz" Mar 13 14:09:05 crc kubenswrapper[4907]: W0313 14:09:05.474585 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3661f6b2_1d5e_42f0_947f_a32b66258a1f.slice/crio-9696c0da9ec098e8a51fc14c33a99177653e98a7c7e1cea2ab3b0156346c1506 WatchSource:0}: Error finding container 9696c0da9ec098e8a51fc14c33a99177653e98a7c7e1cea2ab3b0156346c1506: Status 404 returned error can't find the container with id 9696c0da9ec098e8a51fc14c33a99177653e98a7c7e1cea2ab3b0156346c1506 Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.476943 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.484325 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-56rqr" Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.517907 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.522982 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-rmkfg" Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.536104 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.545724 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-vfbsx" Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.558267 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:05 crc kubenswrapper[4907]: E0313 14:09:05.558581 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:06.058570078 +0000 UTC m=+244.958357767 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.585295 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-kwfhg"] Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.597077 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-r86r4"] Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.597342 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hfcsw"] Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.663170 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.663326 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wk2bb\" (UniqueName: \"kubernetes.io/projected/ebd59055-333a-46c2-ba30-826296414a4c-kube-api-access-wk2bb\") pod \"dns-operator-744455d44c-xkdjh\" (UID: \"ebd59055-333a-46c2-ba30-826296414a4c\") " pod="openshift-dns-operator/dns-operator-744455d44c-xkdjh" Mar 13 14:09:05 crc kubenswrapper[4907]: E0313 14:09:05.664902 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:06.164857447 +0000 UTC m=+245.064645146 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.682802 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wk2bb\" (UniqueName: \"kubernetes.io/projected/ebd59055-333a-46c2-ba30-826296414a4c-kube-api-access-wk2bb\") pod \"dns-operator-744455d44c-xkdjh\" (UID: \"ebd59055-333a-46c2-ba30-826296414a4c\") " pod="openshift-dns-operator/dns-operator-744455d44c-xkdjh" Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.735359 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-5szhc"] Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.766600 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:05 crc kubenswrapper[4907]: E0313 14:09:05.766923 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:06.266911152 +0000 UTC m=+245.166698841 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:05 crc kubenswrapper[4907]: W0313 14:09:05.771567 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod814c6dec_6e96_4068_9aec_9e57fefca5db.slice/crio-0f1cf03420439a212080278b489e741330821d26fb8f1062a79437e4aeb8e4c5 WatchSource:0}: Error finding container 0f1cf03420439a212080278b489e741330821d26fb8f1062a79437e4aeb8e4c5: Status 404 returned error can't find the container with id 0f1cf03420439a212080278b489e741330821d26fb8f1062a79437e4aeb8e4c5 Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.805839 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-92gs7"] Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.810682 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-ksfb8"] Mar 13 14:09:05 crc kubenswrapper[4907]: W0313 14:09:05.835544 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda7b66665_221d_4334_8263_aeb3b20e1c92.slice/crio-bea2a55198cc235ae007e26c1440216d35071366920193590b872d42da99d43c WatchSource:0}: Error finding container bea2a55198cc235ae007e26c1440216d35071366920193590b872d42da99d43c: Status 404 returned error can't find the container with id bea2a55198cc235ae007e26c1440216d35071366920193590b872d42da99d43c Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.862549 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-djph7"] Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.871444 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-4h9q7"] Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.875491 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-wjrms"] Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.875751 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:05 crc kubenswrapper[4907]: E0313 14:09:05.876103 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:06.3760883 +0000 UTC m=+245.275875989 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.908616 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-544df"] Mar 13 14:09:05 crc kubenswrapper[4907]: W0313 14:09:05.917029 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5a9c889b_5954_4d3a_ae72_12000ee1cc25.slice/crio-23c50ef680bc0a0105f881d9b6d419e244388a4b1be6b0b744537678408be591 WatchSource:0}: Error finding container 23c50ef680bc0a0105f881d9b6d419e244388a4b1be6b0b744537678408be591: Status 404 returned error can't find the container with id 23c50ef680bc0a0105f881d9b6d419e244388a4b1be6b0b744537678408be591 Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.977145 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:05 crc kubenswrapper[4907]: E0313 14:09:05.977457 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:06.477442675 +0000 UTC m=+245.377230364 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.998420 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Mar 13 14:09:06 crc kubenswrapper[4907]: I0313 14:09:06.000781 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-xkdjh" Mar 13 14:09:06 crc kubenswrapper[4907]: I0313 14:09:06.001941 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-zz77x"] Mar 13 14:09:06 crc kubenswrapper[4907]: W0313 14:09:06.002925 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod74b29e18_7d73_4ef5_9739_04137dd1b191.slice/crio-28e1b4d5a08c50b259a144b90183e591003ab7707068b0fd9883bf559d1a9060 WatchSource:0}: Error finding container 28e1b4d5a08c50b259a144b90183e591003ab7707068b0fd9883bf559d1a9060: Status 404 returned error can't find the container with id 28e1b4d5a08c50b259a144b90183e591003ab7707068b0fd9883bf559d1a9060 Mar 13 14:09:06 crc kubenswrapper[4907]: W0313 14:09:06.003213 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8013cba2_6b3c_4548_af5c_6dd0057da0c7.slice/crio-bed4842855bdf9609aac5b814fc47449456580fd4be12ff6f7dd5cb79033bd15 WatchSource:0}: Error finding container bed4842855bdf9609aac5b814fc47449456580fd4be12ff6f7dd5cb79033bd15: Status 404 returned error can't find the container with id bed4842855bdf9609aac5b814fc47449456580fd4be12ff6f7dd5cb79033bd15 Mar 13 14:09:06 crc kubenswrapper[4907]: W0313 14:09:06.003730 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9bb1952d_0e55_4e45_81c0_69a8cddb827e.slice/crio-9aafc231113ab96a27a77794d70a165636d29d17e2ee245224f530f5fe6fe82a WatchSource:0}: Error finding container 9aafc231113ab96a27a77794d70a165636d29d17e2ee245224f530f5fe6fe82a: Status 404 returned error can't find the container with id 9aafc231113ab96a27a77794d70a165636d29d17e2ee245224f530f5fe6fe82a Mar 13 14:09:06 crc kubenswrapper[4907]: I0313 14:09:06.039361 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-l9xpb"] Mar 13 14:09:06 crc kubenswrapper[4907]: I0313 14:09:06.078444 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:06 crc kubenswrapper[4907]: E0313 14:09:06.079118 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:06.579074099 +0000 UTC m=+245.478861788 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:06 crc kubenswrapper[4907]: I0313 14:09:06.079336 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:06 crc kubenswrapper[4907]: E0313 14:09:06.079837 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:06.579826119 +0000 UTC m=+245.479613808 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:06 crc kubenswrapper[4907]: W0313 14:09:06.120354 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod10e8bf08_339f_4948_9fe2_ab4092d9a942.slice/crio-20ab56dcb33fadb51a23e57ce832fe3b3f6c0a4e913b6278419b0694bf504898 WatchSource:0}: Error finding container 20ab56dcb33fadb51a23e57ce832fe3b3f6c0a4e913b6278419b0694bf504898: Status 404 returned error can't find the container with id 20ab56dcb33fadb51a23e57ce832fe3b3f6c0a4e913b6278419b0694bf504898 Mar 13 14:09:06 crc kubenswrapper[4907]: I0313 14:09:06.158632 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-hwf6r"] Mar 13 14:09:06 crc kubenswrapper[4907]: I0313 14:09:06.180229 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:06 crc kubenswrapper[4907]: E0313 14:09:06.180762 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:06.680748072 +0000 UTC m=+245.580535761 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:06 crc kubenswrapper[4907]: I0313 14:09:06.263508 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-wgpdd"] Mar 13 14:09:06 crc kubenswrapper[4907]: I0313 14:09:06.281987 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:06 crc kubenswrapper[4907]: E0313 14:09:06.282304 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:06.782291213 +0000 UTC m=+245.682078902 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:06 crc kubenswrapper[4907]: I0313 14:09:06.327280 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fwcwx" podStartSLOduration=185.327261406 podStartE2EDuration="3m5.327261406s" podCreationTimestamp="2026-03-13 14:06:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:09:06.289031847 +0000 UTC m=+245.188819536" watchObservedRunningTime="2026-03-13 14:09:06.327261406 +0000 UTC m=+245.227049095" Mar 13 14:09:06 crc kubenswrapper[4907]: I0313 14:09:06.382897 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:06 crc kubenswrapper[4907]: E0313 14:09:06.383035 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:06.883018222 +0000 UTC m=+245.782805911 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:06 crc kubenswrapper[4907]: I0313 14:09:06.383080 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:06 crc kubenswrapper[4907]: E0313 14:09:06.383411 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:06.883402682 +0000 UTC m=+245.783190361 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:06 crc kubenswrapper[4907]: I0313 14:09:06.410184 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-hh9sk" event={"ID":"e9d44a1e-0eab-4bd6-833f-143436e43758","Type":"ContainerStarted","Data":"f2b59b770051ee0c497bccf354ff8d97a4c4e78381941599f2b81bcdb45aea51"} Mar 13 14:09:06 crc kubenswrapper[4907]: I0313 14:09:06.413796 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-ng4kz" event={"ID":"3661f6b2-1d5e-42f0-947f-a32b66258a1f","Type":"ContainerStarted","Data":"9696c0da9ec098e8a51fc14c33a99177653e98a7c7e1cea2ab3b0156346c1506"} Mar 13 14:09:06 crc kubenswrapper[4907]: I0313 14:09:06.415228 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-hwf6r" event={"ID":"699da377-c61b-48f0-92e0-42bd247ee363","Type":"ContainerStarted","Data":"98417ee0f0233faf0047977c4ec6593e0198d88401a88c6672ff9dc399e06656"} Mar 13 14:09:06 crc kubenswrapper[4907]: I0313 14:09:06.417441 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-r86r4" event={"ID":"5a9c889b-5954-4d3a-ae72-12000ee1cc25","Type":"ContainerStarted","Data":"23c50ef680bc0a0105f881d9b6d419e244388a4b1be6b0b744537678408be591"} Mar 13 14:09:06 crc kubenswrapper[4907]: I0313 14:09:06.418433 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-ksfb8" event={"ID":"8013cba2-6b3c-4548-af5c-6dd0057da0c7","Type":"ContainerStarted","Data":"bed4842855bdf9609aac5b814fc47449456580fd4be12ff6f7dd5cb79033bd15"} Mar 13 14:09:06 crc kubenswrapper[4907]: I0313 14:09:06.419957 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hfcsw" event={"ID":"a7b66665-221d-4334-8263-aeb3b20e1c92","Type":"ContainerStarted","Data":"bea2a55198cc235ae007e26c1440216d35071366920193590b872d42da99d43c"} Mar 13 14:09:06 crc kubenswrapper[4907]: I0313 14:09:06.421139 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-544df" event={"ID":"f8ea4f24-05b0-4661-a617-a4208f9a9188","Type":"ContainerStarted","Data":"76ec31cdef4dc0d683fc04866509997aaaa312dd52bb20f29c41405f80ba4c5c"} Mar 13 14:09:06 crc kubenswrapper[4907]: I0313 14:09:06.422272 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-5stnc" event={"ID":"0290eb6f-0be7-4340-8ba8-ed6ea0662c33","Type":"ContainerStarted","Data":"97e42fd7291025df6972a91466dcaf06ac63f386d1c5751baba2fa6aca2cf00a"} Mar 13 14:09:06 crc kubenswrapper[4907]: I0313 14:09:06.426111 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-zwzc4" event={"ID":"1954f507-7cb0-4942-a4c8-6586d370e717","Type":"ContainerStarted","Data":"292d35dc7c24966269168817ed4b7571bbf78be96c1a62a6688c9ddaa28634af"} Mar 13 14:09:06 crc kubenswrapper[4907]: I0313 14:09:06.433245 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-92gs7" event={"ID":"74b29e18-7d73-4ef5-9739-04137dd1b191","Type":"ContainerStarted","Data":"28e1b4d5a08c50b259a144b90183e591003ab7707068b0fd9883bf559d1a9060"} Mar 13 14:09:06 crc kubenswrapper[4907]: I0313 14:09:06.441593 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-zz77x" event={"ID":"10e8bf08-339f-4948-9fe2-ab4092d9a942","Type":"ContainerStarted","Data":"20ab56dcb33fadb51a23e57ce832fe3b3f6c0a4e913b6278419b0694bf504898"} Mar 13 14:09:06 crc kubenswrapper[4907]: I0313 14:09:06.447190 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-prlhd" event={"ID":"b75dbfe3-8887-4a0b-9541-a4e4000924cb","Type":"ContainerStarted","Data":"bfb81e78d3c0a64712627633cf3de282b68658dfd36fc8d0ea22fe07e87a7ca6"} Mar 13 14:09:06 crc kubenswrapper[4907]: I0313 14:09:06.448795 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xnr2b" event={"ID":"c934a2b9-3585-4491-8962-df085c473d90","Type":"ContainerStarted","Data":"e678dddce63a8e965db113d1ad5181b336626e09d4d5b053d64455d2374aa15f"} Mar 13 14:09:06 crc kubenswrapper[4907]: I0313 14:09:06.449548 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-l9xpb" event={"ID":"4c8c0266-5452-4abb-92c5-c536df94dd41","Type":"ContainerStarted","Data":"2ca75bf0ab6ab7b6d03a36b2b78748709fe1e70503c3e45b40703b8a20dadbca"} Mar 13 14:09:06 crc kubenswrapper[4907]: I0313 14:09:06.452329 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hkbqt" event={"ID":"4682bbe1-4257-4d04-a6d7-60c4655d6873","Type":"ContainerStarted","Data":"fe97ca6fd77d983d0419febba7976c1afd14125b4f9439aedf1df1df511bf7e0"} Mar 13 14:09:06 crc kubenswrapper[4907]: I0313 14:09:06.452360 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hkbqt" event={"ID":"4682bbe1-4257-4d04-a6d7-60c4655d6873","Type":"ContainerStarted","Data":"e3054c999bd2aae22709666dab86d8d7deb5c922a69e6d8a8238d836c3bea89e"} Mar 13 14:09:06 crc kubenswrapper[4907]: I0313 14:09:06.452907 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hkbqt" Mar 13 14:09:06 crc kubenswrapper[4907]: I0313 14:09:06.460055 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4h9q7" event={"ID":"9bb1952d-0e55-4e45-81c0-69a8cddb827e","Type":"ContainerStarted","Data":"9aafc231113ab96a27a77794d70a165636d29d17e2ee245224f530f5fe6fe82a"} Mar 13 14:09:06 crc kubenswrapper[4907]: I0313 14:09:06.467034 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-djph7" event={"ID":"3a661298-271c-4c57-94ea-3d69a27d7ad5","Type":"ContainerStarted","Data":"d38738b3846a8bae401e7e85068b5989891d2051495c94c078dd0dd338bbc4dd"} Mar 13 14:09:06 crc kubenswrapper[4907]: I0313 14:09:06.467192 4907 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-hkbqt container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.23:8443/healthz\": dial tcp 10.217.0.23:8443: connect: connection refused" start-of-body= Mar 13 14:09:06 crc kubenswrapper[4907]: I0313 14:09:06.467257 4907 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hkbqt" podUID="4682bbe1-4257-4d04-a6d7-60c4655d6873" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.23:8443/healthz\": dial tcp 10.217.0.23:8443: connect: connection refused" Mar 13 14:09:06 crc kubenswrapper[4907]: I0313 14:09:06.474478 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-8jqgq" event={"ID":"fa265490-c23b-4b05-94bc-78b4a636e065","Type":"ContainerStarted","Data":"7915bd087312e5a5b3f4cc701a590ae465651cc0364d1c824e4224d5547bc50b"} Mar 13 14:09:06 crc kubenswrapper[4907]: I0313 14:09:06.475482 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5szhc" event={"ID":"ae320273-06e8-43c5-a64f-acd80ad16d1c","Type":"ContainerStarted","Data":"cb025c3ea8ee29cc007c3e83fffde7d3aa1a71ee99015e66540499acd12ff294"} Mar 13 14:09:06 crc kubenswrapper[4907]: I0313 14:09:06.479294 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ksgr6" event={"ID":"1be912b5-9d1d-48bd-a0ab-7f601f9eb9da","Type":"ContainerStarted","Data":"917984672ee70e65cbe94b68435b329ada21f7abe45e3abae9a73f7190269e84"} Mar 13 14:09:06 crc kubenswrapper[4907]: I0313 14:09:06.481395 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-kwfhg" event={"ID":"814c6dec-6e96-4068-9aec-9e57fefca5db","Type":"ContainerStarted","Data":"0f1cf03420439a212080278b489e741330821d26fb8f1062a79437e4aeb8e4c5"} Mar 13 14:09:06 crc kubenswrapper[4907]: I0313 14:09:06.486041 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:06 crc kubenswrapper[4907]: I0313 14:09:06.487204 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-56rqr"] Mar 13 14:09:06 crc kubenswrapper[4907]: E0313 14:09:06.487523 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:06.987302776 +0000 UTC m=+245.887090465 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:06 crc kubenswrapper[4907]: W0313 14:09:06.509284 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0149e083_ed5b_4c7f_94e7_4f3f2ff8a557.slice/crio-b2a81aa0d935e7dfb8b83728724e4149f1365539869eda9b25d8e6618548a262 WatchSource:0}: Error finding container b2a81aa0d935e7dfb8b83728724e4149f1365539869eda9b25d8e6618548a262: Status 404 returned error can't find the container with id b2a81aa0d935e7dfb8b83728724e4149f1365539869eda9b25d8e6618548a262 Mar 13 14:09:06 crc kubenswrapper[4907]: I0313 14:09:06.517582 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556840-pgmwj" event={"ID":"19edf37b-a1d3-4c7c-945e-5d0b8bb601bf","Type":"ContainerStarted","Data":"d2284227e021b37d9e8d2812f8621bebf16efd8eeb5019b51cfa28773e7a5aa9"} Mar 13 14:09:06 crc kubenswrapper[4907]: I0313 14:09:06.517613 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556840-pgmwj" event={"ID":"19edf37b-a1d3-4c7c-945e-5d0b8bb601bf","Type":"ContainerStarted","Data":"e8a5550c81aa8c4bb6491a89a3041cb9a6e5701c9a00e500620d96a20f452a36"} Mar 13 14:09:06 crc kubenswrapper[4907]: I0313 14:09:06.529176 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-wjrms" event={"ID":"71c234e9-5f3d-4b0c-94d4-80b4e33a6dda","Type":"ContainerStarted","Data":"c587f565b1dce424d527ed3ea9bb069af779f1b90a7481e3460deb763887b3e9"} Mar 13 14:09:06 crc kubenswrapper[4907]: I0313 14:09:06.534809 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-52vhb" event={"ID":"005dfe54-e402-4aa4-8b67-f7e7b685debd","Type":"ContainerStarted","Data":"5995450a4779edf85718472df49cceacc642f80ac60d07e668e19501ff70c8b5"} Mar 13 14:09:06 crc kubenswrapper[4907]: I0313 14:09:06.540852 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-252nq" event={"ID":"ff214b86-eb3d-46c6-b45a-18a9f63e742c","Type":"ContainerStarted","Data":"51db44672a2c6c61b21d0063b559e91a222184620152fc66fe9d7966045175a8"} Mar 13 14:09:06 crc kubenswrapper[4907]: I0313 14:09:06.544490 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ccjsv" event={"ID":"b651a92c-e9f6-4a1a-8c49-b6500bb7c61c","Type":"ContainerStarted","Data":"7ef73955c4af33184c9094d1058ec7a76a57ccb504926b53434a49c80ea3b182"} Mar 13 14:09:06 crc kubenswrapper[4907]: I0313 14:09:06.556151 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-75gdt" event={"ID":"ff0e893f-05cb-4751-9960-281c58000519","Type":"ContainerStarted","Data":"8022ae37cda8bf634c08294061821922a69872cde924e7bb3e44901196573e79"} Mar 13 14:09:06 crc kubenswrapper[4907]: I0313 14:09:06.570620 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pk65f" podStartSLOduration=185.570600362 podStartE2EDuration="3m5.570600362s" podCreationTimestamp="2026-03-13 14:06:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:09:06.569252835 +0000 UTC m=+245.469040534" watchObservedRunningTime="2026-03-13 14:09:06.570600362 +0000 UTC m=+245.470388051" Mar 13 14:09:06 crc kubenswrapper[4907]: I0313 14:09:06.586917 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fwcwx" Mar 13 14:09:06 crc kubenswrapper[4907]: I0313 14:09:06.588685 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:06 crc kubenswrapper[4907]: E0313 14:09:06.591034 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:07.091021406 +0000 UTC m=+245.990809085 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:06 crc kubenswrapper[4907]: W0313 14:09:06.596330 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod59980734_2cf7_42dc_9f6f_a6de7e1c3665.slice/crio-1f36e2ae021694a6f49d8cea38eb878db4943e0747adab7e3fe2b00be56274fe WatchSource:0}: Error finding container 1f36e2ae021694a6f49d8cea38eb878db4943e0747adab7e3fe2b00be56274fe: Status 404 returned error can't find the container with id 1f36e2ae021694a6f49d8cea38eb878db4943e0747adab7e3fe2b00be56274fe Mar 13 14:09:06 crc kubenswrapper[4907]: I0313 14:09:06.655402 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-g6q2r" podStartSLOduration=6.655379216 podStartE2EDuration="6.655379216s" podCreationTimestamp="2026-03-13 14:09:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:09:06.607811643 +0000 UTC m=+245.507599332" watchObservedRunningTime="2026-03-13 14:09:06.655379216 +0000 UTC m=+245.555166905" Mar 13 14:09:06 crc kubenswrapper[4907]: I0313 14:09:06.686754 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-t68ql" podStartSLOduration=185.686734939 podStartE2EDuration="3m5.686734939s" podCreationTimestamp="2026-03-13 14:06:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:09:06.686192734 +0000 UTC m=+245.585980423" watchObservedRunningTime="2026-03-13 14:09:06.686734939 +0000 UTC m=+245.586522628" Mar 13 14:09:06 crc kubenswrapper[4907]: I0313 14:09:06.689792 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:06 crc kubenswrapper[4907]: E0313 14:09:06.690293 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:07.190272405 +0000 UTC m=+246.090060094 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:06 crc kubenswrapper[4907]: I0313 14:09:06.717402 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-rmkfg"] Mar 13 14:09:06 crc kubenswrapper[4907]: I0313 14:09:06.794728 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:06 crc kubenswrapper[4907]: E0313 14:09:06.795518 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:07.295505585 +0000 UTC m=+246.195293274 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:06 crc kubenswrapper[4907]: I0313 14:09:06.799044 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-xkdjh"] Mar 13 14:09:06 crc kubenswrapper[4907]: I0313 14:09:06.809799 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-w76kw"] Mar 13 14:09:06 crc kubenswrapper[4907]: I0313 14:09:06.860229 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-vfbsx"] Mar 13 14:09:06 crc kubenswrapper[4907]: I0313 14:09:06.896175 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:06 crc kubenswrapper[4907]: E0313 14:09:06.897069 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:07.397047896 +0000 UTC m=+246.296835585 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:06 crc kubenswrapper[4907]: W0313 14:09:06.961792 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9b75954c_b75d_4f17_82dc_4c0358fd8d0f.slice/crio-47724a9f48548516662c355e2e67e0051f4b8870e897b87bf044f8835576d09e WatchSource:0}: Error finding container 47724a9f48548516662c355e2e67e0051f4b8870e897b87bf044f8835576d09e: Status 404 returned error can't find the container with id 47724a9f48548516662c355e2e67e0051f4b8870e897b87bf044f8835576d09e Mar 13 14:09:06 crc kubenswrapper[4907]: W0313 14:09:06.963270 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podebd59055_333a_46c2_ba30_826296414a4c.slice/crio-be63aa8b685fc3b6de50ab1c5058dba0dd2fe771225810b4038045dc7a41c54c WatchSource:0}: Error finding container be63aa8b685fc3b6de50ab1c5058dba0dd2fe771225810b4038045dc7a41c54c: Status 404 returned error can't find the container with id be63aa8b685fc3b6de50ab1c5058dba0dd2fe771225810b4038045dc7a41c54c Mar 13 14:09:06 crc kubenswrapper[4907]: I0313 14:09:06.966418 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-v9m24" podStartSLOduration=185.966402382 podStartE2EDuration="3m5.966402382s" podCreationTimestamp="2026-03-13 14:06:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:09:06.927147755 +0000 UTC m=+245.826935434" watchObservedRunningTime="2026-03-13 14:09:06.966402382 +0000 UTC m=+245.866190071" Mar 13 14:09:06 crc kubenswrapper[4907]: I0313 14:09:06.998352 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:06 crc kubenswrapper[4907]: E0313 14:09:06.999270 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:07.499254075 +0000 UTC m=+246.399041764 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:07 crc kubenswrapper[4907]: I0313 14:09:07.099833 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:07 crc kubenswrapper[4907]: E0313 14:09:07.100022 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:07.599995964 +0000 UTC m=+246.499783673 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:07 crc kubenswrapper[4907]: I0313 14:09:07.100368 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:07 crc kubenswrapper[4907]: E0313 14:09:07.101534 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:07.601510315 +0000 UTC m=+246.501298004 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:07 crc kubenswrapper[4907]: I0313 14:09:07.137742 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-rnh5t" podStartSLOduration=187.13772344 podStartE2EDuration="3m7.13772344s" podCreationTimestamp="2026-03-13 14:06:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:09:07.130808112 +0000 UTC m=+246.030595801" watchObservedRunningTime="2026-03-13 14:09:07.13772344 +0000 UTC m=+246.037511119" Mar 13 14:09:07 crc kubenswrapper[4907]: I0313 14:09:07.161393 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pk65f" Mar 13 14:09:07 crc kubenswrapper[4907]: I0313 14:09:07.202018 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:07 crc kubenswrapper[4907]: E0313 14:09:07.202477 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:07.702456909 +0000 UTC m=+246.602244598 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:07 crc kubenswrapper[4907]: I0313 14:09:07.309510 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:07 crc kubenswrapper[4907]: E0313 14:09:07.310447 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:07.810431905 +0000 UTC m=+246.710219594 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:07 crc kubenswrapper[4907]: I0313 14:09:07.410742 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:07 crc kubenswrapper[4907]: E0313 14:09:07.411133 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:07.911117542 +0000 UTC m=+246.810905221 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:07 crc kubenswrapper[4907]: I0313 14:09:07.421442 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-252nq" podStartSLOduration=7.421422293 podStartE2EDuration="7.421422293s" podCreationTimestamp="2026-03-13 14:09:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:09:07.421013161 +0000 UTC m=+246.320800860" watchObservedRunningTime="2026-03-13 14:09:07.421422293 +0000 UTC m=+246.321209982" Mar 13 14:09:07 crc kubenswrapper[4907]: I0313 14:09:07.511841 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:07 crc kubenswrapper[4907]: E0313 14:09:07.512519 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:08.012506949 +0000 UTC m=+246.912294638 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:07 crc kubenswrapper[4907]: I0313 14:09:07.554448 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29556840-pgmwj" podStartSLOduration=186.554423459 podStartE2EDuration="3m6.554423459s" podCreationTimestamp="2026-03-13 14:06:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:09:07.553925655 +0000 UTC m=+246.453713344" watchObservedRunningTime="2026-03-13 14:09:07.554423459 +0000 UTC m=+246.454211148" Mar 13 14:09:07 crc kubenswrapper[4907]: I0313 14:09:07.593823 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-wgpdd" event={"ID":"0149e083-ed5b-4c7f-94e7-4f3f2ff8a557","Type":"ContainerStarted","Data":"40f0b1a25cd0d0a5030ec1dcae203661c52d45d8711eefed49ad300a5495f0ac"} Mar 13 14:09:07 crc kubenswrapper[4907]: I0313 14:09:07.593867 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-wgpdd" event={"ID":"0149e083-ed5b-4c7f-94e7-4f3f2ff8a557","Type":"ContainerStarted","Data":"b2a81aa0d935e7dfb8b83728724e4149f1365539869eda9b25d8e6618548a262"} Mar 13 14:09:07 crc kubenswrapper[4907]: I0313 14:09:07.615024 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:07 crc kubenswrapper[4907]: E0313 14:09:07.616155 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:08.116125976 +0000 UTC m=+247.015913665 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:07 crc kubenswrapper[4907]: I0313 14:09:07.629190 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-ksfb8" event={"ID":"8013cba2-6b3c-4548-af5c-6dd0057da0c7","Type":"ContainerStarted","Data":"fe367c6c7181382e906590402f45585b664a193ada8ba64c3ce22a4ffceb143a"} Mar 13 14:09:07 crc kubenswrapper[4907]: I0313 14:09:07.636371 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hfcsw" event={"ID":"a7b66665-221d-4334-8263-aeb3b20e1c92","Type":"ContainerStarted","Data":"e725ce3a4600879c951eaa2f5632cd7105e7bd88d5d27083b1e0b8ea856d9f8d"} Mar 13 14:09:07 crc kubenswrapper[4907]: I0313 14:09:07.664295 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-xkdjh" event={"ID":"ebd59055-333a-46c2-ba30-826296414a4c","Type":"ContainerStarted","Data":"be63aa8b685fc3b6de50ab1c5058dba0dd2fe771225810b4038045dc7a41c54c"} Mar 13 14:09:07 crc kubenswrapper[4907]: I0313 14:09:07.689150 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hkbqt" podStartSLOduration=186.689118 podStartE2EDuration="3m6.689118s" podCreationTimestamp="2026-03-13 14:06:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:09:07.686076588 +0000 UTC m=+246.585864277" watchObservedRunningTime="2026-03-13 14:09:07.689118 +0000 UTC m=+246.588905689" Mar 13 14:09:07 crc kubenswrapper[4907]: I0313 14:09:07.715031 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-zwzc4" podStartSLOduration=186.715014114 podStartE2EDuration="3m6.715014114s" podCreationTimestamp="2026-03-13 14:06:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:09:07.7133733 +0000 UTC m=+246.613160989" watchObservedRunningTime="2026-03-13 14:09:07.715014114 +0000 UTC m=+246.614801803" Mar 13 14:09:07 crc kubenswrapper[4907]: I0313 14:09:07.717602 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:07 crc kubenswrapper[4907]: I0313 14:09:07.718685 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-zz77x" event={"ID":"10e8bf08-339f-4948-9fe2-ab4092d9a942","Type":"ContainerStarted","Data":"cb1dbd959b8e29c65b2fa393f724e71e4139cca7b1c49f5b49d06e9a4ef4a6ee"} Mar 13 14:09:07 crc kubenswrapper[4907]: E0313 14:09:07.718747 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:08.218730435 +0000 UTC m=+247.118518124 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:07 crc kubenswrapper[4907]: I0313 14:09:07.736349 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ccjsv" event={"ID":"b651a92c-e9f6-4a1a-8c49-b6500bb7c61c","Type":"ContainerStarted","Data":"1060d8d803a2d78da1f8488a9dfc05a98c6bd3ae8f3e202cc535d226fa32fdcb"} Mar 13 14:09:07 crc kubenswrapper[4907]: I0313 14:09:07.736401 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ccjsv" event={"ID":"b651a92c-e9f6-4a1a-8c49-b6500bb7c61c","Type":"ContainerStarted","Data":"eee14898537e10df3b0f40a25557131ca5dc13864a11496c9fbd5305b6c03842"} Mar 13 14:09:07 crc kubenswrapper[4907]: I0313 14:09:07.737088 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ccjsv" Mar 13 14:09:07 crc kubenswrapper[4907]: I0313 14:09:07.748973 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-rmkfg" event={"ID":"0396c426-1d56-44e6-885a-1b97d69ad57a","Type":"ContainerStarted","Data":"a594462937e2e5a805dd6fe9e9b6c749f8349ed77ce762003ee5b78ec459bae8"} Mar 13 14:09:07 crc kubenswrapper[4907]: I0313 14:09:07.774240 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-8jqgq" event={"ID":"fa265490-c23b-4b05-94bc-78b4a636e065","Type":"ContainerStarted","Data":"59bec09f1e914543d24ba435710c730a198b7a8490d1fa1f391f63d54ca5e15b"} Mar 13 14:09:07 crc kubenswrapper[4907]: I0313 14:09:07.781843 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-kwfhg" event={"ID":"814c6dec-6e96-4068-9aec-9e57fefca5db","Type":"ContainerStarted","Data":"11d51b3d36d9283f41d21738ca7a244487cc24be4453cd47b19366b6bff8bcf1"} Mar 13 14:09:07 crc kubenswrapper[4907]: I0313 14:09:07.782512 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-kwfhg" Mar 13 14:09:07 crc kubenswrapper[4907]: I0313 14:09:07.785628 4907 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-kwfhg container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.7:8443/healthz\": dial tcp 10.217.0.7:8443: connect: connection refused" start-of-body= Mar 13 14:09:07 crc kubenswrapper[4907]: I0313 14:09:07.785789 4907 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-kwfhg" podUID="814c6dec-6e96-4068-9aec-9e57fefca5db" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.7:8443/healthz\": dial tcp 10.217.0.7:8443: connect: connection refused" Mar 13 14:09:07 crc kubenswrapper[4907]: I0313 14:09:07.798517 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-vfbsx" event={"ID":"814322e3-6251-4ea3-a86b-7b89ca6ea728","Type":"ContainerStarted","Data":"c0bccb4ffe867441f6cbe79f037b0841fedef0026ddfa347dd39ec3bd29ad30d"} Mar 13 14:09:07 crc kubenswrapper[4907]: I0313 14:09:07.800187 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-hh9sk" podStartSLOduration=186.800160369 podStartE2EDuration="3m6.800160369s" podCreationTimestamp="2026-03-13 14:06:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:09:07.792053198 +0000 UTC m=+246.691840877" watchObservedRunningTime="2026-03-13 14:09:07.800160369 +0000 UTC m=+246.699948058" Mar 13 14:09:07 crc kubenswrapper[4907]: I0313 14:09:07.803004 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-prlhd" event={"ID":"b75dbfe3-8887-4a0b-9541-a4e4000924cb","Type":"ContainerStarted","Data":"be13451dfb6a4d9101c5c56519e9742d9a74fed841a8fa3e62ab1471ed23edca"} Mar 13 14:09:07 crc kubenswrapper[4907]: I0313 14:09:07.805399 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xnr2b" event={"ID":"c934a2b9-3585-4491-8962-df085c473d90","Type":"ContainerStarted","Data":"78b55d953a014b3c151178d298a92988bb6e4a6dda66d0fb41b8a35f9cd910c8"} Mar 13 14:09:07 crc kubenswrapper[4907]: I0313 14:09:07.822910 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:07 crc kubenswrapper[4907]: E0313 14:09:07.824217 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:08.324191692 +0000 UTC m=+247.223979391 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:07 crc kubenswrapper[4907]: I0313 14:09:07.840000 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-ng4kz" event={"ID":"3661f6b2-1d5e-42f0-947f-a32b66258a1f","Type":"ContainerStarted","Data":"3d60ed9e1e0ff620e3a27d58d5fb879793f36dc7e7cd6201ecb90667f5bab7dc"} Mar 13 14:09:07 crc kubenswrapper[4907]: I0313 14:09:07.840820 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-ng4kz" Mar 13 14:09:07 crc kubenswrapper[4907]: I0313 14:09:07.842373 4907 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-ng4kz container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.11:6443/healthz\": dial tcp 10.217.0.11:6443: connect: connection refused" start-of-body= Mar 13 14:09:07 crc kubenswrapper[4907]: I0313 14:09:07.842429 4907 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-ng4kz" podUID="3661f6b2-1d5e-42f0-947f-a32b66258a1f" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.11:6443/healthz\": dial tcp 10.217.0.11:6443: connect: connection refused" Mar 13 14:09:07 crc kubenswrapper[4907]: I0313 14:09:07.848567 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-r86r4" event={"ID":"5a9c889b-5954-4d3a-ae72-12000ee1cc25","Type":"ContainerStarted","Data":"4651f82e390b3a5ac2c55407710eda6955584915c8001c9a1dc4182061851ee7"} Mar 13 14:09:07 crc kubenswrapper[4907]: I0313 14:09:07.860797 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-75gdt" podStartSLOduration=186.860774807 podStartE2EDuration="3m6.860774807s" podCreationTimestamp="2026-03-13 14:06:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:09:07.841145003 +0000 UTC m=+246.740932692" watchObservedRunningTime="2026-03-13 14:09:07.860774807 +0000 UTC m=+246.760562496" Mar 13 14:09:07 crc kubenswrapper[4907]: I0313 14:09:07.870711 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-kwfhg" podStartSLOduration=186.870686496 podStartE2EDuration="3m6.870686496s" podCreationTimestamp="2026-03-13 14:06:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:09:07.86971706 +0000 UTC m=+246.769504769" watchObservedRunningTime="2026-03-13 14:09:07.870686496 +0000 UTC m=+246.770474195" Mar 13 14:09:07 crc kubenswrapper[4907]: I0313 14:09:07.892911 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-56rqr" event={"ID":"59980734-2cf7-42dc-9f6f-a6de7e1c3665","Type":"ContainerStarted","Data":"1f36e2ae021694a6f49d8cea38eb878db4943e0747adab7e3fe2b00be56274fe"} Mar 13 14:09:07 crc kubenswrapper[4907]: I0313 14:09:07.898286 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-wjrms" event={"ID":"71c234e9-5f3d-4b0c-94d4-80b4e33a6dda","Type":"ContainerStarted","Data":"eb3bbb7e5376638446dc533d2a33fce47abb9f3e1fafc7162f893e7aa4800daf"} Mar 13 14:09:07 crc kubenswrapper[4907]: I0313 14:09:07.925763 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:07 crc kubenswrapper[4907]: E0313 14:09:07.926114 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:08.426100943 +0000 UTC m=+247.325888632 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:07 crc kubenswrapper[4907]: I0313 14:09:07.964284 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-52vhb" event={"ID":"005dfe54-e402-4aa4-8b67-f7e7b685debd","Type":"ContainerStarted","Data":"9537155f0efd7cadee4facaf87934d2aae807cb49c4f95882646566ae9e4d8a2"} Mar 13 14:09:07 crc kubenswrapper[4907]: I0313 14:09:07.978035 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-w76kw" event={"ID":"9b75954c-b75d-4f17-82dc-4c0358fd8d0f","Type":"ContainerStarted","Data":"47724a9f48548516662c355e2e67e0051f4b8870e897b87bf044f8835576d09e"} Mar 13 14:09:07 crc kubenswrapper[4907]: I0313 14:09:07.991474 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xnr2b" podStartSLOduration=187.991457 podStartE2EDuration="3m7.991457s" podCreationTimestamp="2026-03-13 14:06:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:09:07.961761672 +0000 UTC m=+246.861549361" watchObservedRunningTime="2026-03-13 14:09:07.991457 +0000 UTC m=+246.891244689" Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.013672 4907 generic.go:334] "Generic (PLEG): container finished" podID="1be912b5-9d1d-48bd-a0ab-7f601f9eb9da" containerID="9afd4edd5700b10026def68352336f5dc9ba3dd07805b6aa3edb298f50f16142" exitCode=0 Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.013750 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ksgr6" event={"ID":"1be912b5-9d1d-48bd-a0ab-7f601f9eb9da","Type":"ContainerDied","Data":"9afd4edd5700b10026def68352336f5dc9ba3dd07805b6aa3edb298f50f16142"} Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.027249 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:08 crc kubenswrapper[4907]: E0313 14:09:08.027370 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:08.527345605 +0000 UTC m=+247.427133294 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.027894 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.036199 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-zz77x" podStartSLOduration=187.036180516 podStartE2EDuration="3m7.036180516s" podCreationTimestamp="2026-03-13 14:06:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:09:07.995184391 +0000 UTC m=+246.894972100" watchObservedRunningTime="2026-03-13 14:09:08.036180516 +0000 UTC m=+246.935968205" Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.036708 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ccjsv" podStartSLOduration=187.03670085 podStartE2EDuration="3m7.03670085s" podCreationTimestamp="2026-03-13 14:06:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:09:08.032414223 +0000 UTC m=+246.932201912" watchObservedRunningTime="2026-03-13 14:09:08.03670085 +0000 UTC m=+246.936488529" Mar 13 14:09:08 crc kubenswrapper[4907]: E0313 14:09:08.047143 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:08.547124553 +0000 UTC m=+247.446912242 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.069778 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hfcsw" podStartSLOduration=187.069757539 podStartE2EDuration="3m7.069757539s" podCreationTimestamp="2026-03-13 14:06:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:09:08.068632808 +0000 UTC m=+246.968420497" watchObservedRunningTime="2026-03-13 14:09:08.069757539 +0000 UTC m=+246.969545228" Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.093454 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-92gs7" event={"ID":"74b29e18-7d73-4ef5-9739-04137dd1b191","Type":"ContainerStarted","Data":"9a979dcc7ab19700343b563444457c87eb78ce1bc410e31bf716d61b5b4c37b4"} Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.106564 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5szhc" event={"ID":"ae320273-06e8-43c5-a64f-acd80ad16d1c","Type":"ContainerStarted","Data":"2e80e8494d9dc6d61c40f9ceab4bc533f542e55642a745e2c701bf307a2db887"} Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.129867 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:08 crc kubenswrapper[4907]: E0313 14:09:08.135179 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:08.635147096 +0000 UTC m=+247.534934785 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.135663 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-5stnc" event={"ID":"0290eb6f-0be7-4340-8ba8-ed6ea0662c33","Type":"ContainerStarted","Data":"48afa92be95670a8e5bf5c9bca4dd6ed4ea5b0ffd5c088a0a66004cb2137ee6b"} Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.138206 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-5stnc" Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.147117 4907 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-5stnc container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.35:8080/healthz\": dial tcp 10.217.0.35:8080: connect: connection refused" start-of-body= Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.147184 4907 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-5stnc" podUID="0290eb6f-0be7-4340-8ba8-ed6ea0662c33" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.35:8080/healthz\": dial tcp 10.217.0.35:8080: connect: connection refused" Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.152326 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-544df" event={"ID":"f8ea4f24-05b0-4661-a617-a4208f9a9188","Type":"ContainerStarted","Data":"f48f2ca4ee981d838c5862139a85b9d2f101144d532b7ded20d21d6293a389b6"} Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.153350 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-544df" Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.162558 4907 patch_prober.go:28] interesting pod/downloads-7954f5f757-544df container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.162636 4907 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-544df" podUID="f8ea4f24-05b0-4661-a617-a4208f9a9188" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.167480 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-prlhd" podStartSLOduration=187.167457224 podStartE2EDuration="3m7.167457224s" podCreationTimestamp="2026-03-13 14:06:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:09:08.159376245 +0000 UTC m=+247.059163924" watchObservedRunningTime="2026-03-13 14:09:08.167457224 +0000 UTC m=+247.067244913" Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.168535 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-f8mmz" event={"ID":"c26ccc9f-e682-4b94-9539-cb2607725f71","Type":"ContainerStarted","Data":"4c9b269b13491d5272c98a727d99c52f5328c3fde0c6fcad1345bb2263323a7f"} Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.176314 4907 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-hkbqt container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.23:8443/healthz\": dial tcp 10.217.0.23:8443: connect: connection refused" start-of-body= Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.176375 4907 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hkbqt" podUID="4682bbe1-4257-4d04-a6d7-60c4655d6873" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.23:8443/healthz\": dial tcp 10.217.0.23:8443: connect: connection refused" Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.201415 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-ng4kz" podStartSLOduration=188.201389088 podStartE2EDuration="3m8.201389088s" podCreationTimestamp="2026-03-13 14:06:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:09:08.199108805 +0000 UTC m=+247.098896494" watchObservedRunningTime="2026-03-13 14:09:08.201389088 +0000 UTC m=+247.101176777" Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.242034 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.242394 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-djph7" podStartSLOduration=187.242372401 podStartE2EDuration="3m7.242372401s" podCreationTimestamp="2026-03-13 14:06:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:09:08.241466286 +0000 UTC m=+247.141253975" watchObservedRunningTime="2026-03-13 14:09:08.242372401 +0000 UTC m=+247.142160090" Mar 13 14:09:08 crc kubenswrapper[4907]: E0313 14:09:08.244031 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:08.744013376 +0000 UTC m=+247.643801065 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.257956 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-v6msv"] Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.259697 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-v6msv" Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.262976 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-v6msv"] Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.279727 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.338269 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-5stnc" podStartSLOduration=187.338232097 podStartE2EDuration="3m7.338232097s" podCreationTimestamp="2026-03-13 14:06:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:09:08.27249705 +0000 UTC m=+247.172284739" watchObservedRunningTime="2026-03-13 14:09:08.338232097 +0000 UTC m=+247.238019796" Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.343102 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.343251 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k6s9c\" (UniqueName: \"kubernetes.io/projected/84c2ada5-ceab-4327-802e-9ae459ac814d-kube-api-access-k6s9c\") pod \"community-operators-v6msv\" (UID: \"84c2ada5-ceab-4327-802e-9ae459ac814d\") " pod="openshift-marketplace/community-operators-v6msv" Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.343351 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/84c2ada5-ceab-4327-802e-9ae459ac814d-utilities\") pod \"community-operators-v6msv\" (UID: \"84c2ada5-ceab-4327-802e-9ae459ac814d\") " pod="openshift-marketplace/community-operators-v6msv" Mar 13 14:09:08 crc kubenswrapper[4907]: E0313 14:09:08.343386 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:08.843365447 +0000 UTC m=+247.743153146 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.343434 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/84c2ada5-ceab-4327-802e-9ae459ac814d-catalog-content\") pod \"community-operators-v6msv\" (UID: \"84c2ada5-ceab-4327-802e-9ae459ac814d\") " pod="openshift-marketplace/community-operators-v6msv" Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.383611 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-52vhb" podStartSLOduration=188.383587161 podStartE2EDuration="3m8.383587161s" podCreationTimestamp="2026-03-13 14:06:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:09:08.375735017 +0000 UTC m=+247.275522716" watchObservedRunningTime="2026-03-13 14:09:08.383587161 +0000 UTC m=+247.283374860" Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.422261 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-544df" podStartSLOduration=187.422237741 podStartE2EDuration="3m7.422237741s" podCreationTimestamp="2026-03-13 14:06:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:09:08.414823 +0000 UTC m=+247.314610689" watchObservedRunningTime="2026-03-13 14:09:08.422237741 +0000 UTC m=+247.322025430" Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.431151 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-42kpk"] Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.433215 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-42kpk" Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.447081 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/84c2ada5-ceab-4327-802e-9ae459ac814d-utilities\") pod \"community-operators-v6msv\" (UID: \"84c2ada5-ceab-4327-802e-9ae459ac814d\") " pod="openshift-marketplace/community-operators-v6msv" Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.448207 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/84c2ada5-ceab-4327-802e-9ae459ac814d-catalog-content\") pod \"community-operators-v6msv\" (UID: \"84c2ada5-ceab-4327-802e-9ae459ac814d\") " pod="openshift-marketplace/community-operators-v6msv" Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.451201 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k6s9c\" (UniqueName: \"kubernetes.io/projected/84c2ada5-ceab-4327-802e-9ae459ac814d-kube-api-access-k6s9c\") pod \"community-operators-v6msv\" (UID: \"84c2ada5-ceab-4327-802e-9ae459ac814d\") " pod="openshift-marketplace/community-operators-v6msv" Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.451445 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.451684 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-42kpk"] Mar 13 14:09:08 crc kubenswrapper[4907]: E0313 14:09:08.452930 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:08.952913835 +0000 UTC m=+247.852701524 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.453151 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/84c2ada5-ceab-4327-802e-9ae459ac814d-utilities\") pod \"community-operators-v6msv\" (UID: \"84c2ada5-ceab-4327-802e-9ae459ac814d\") " pod="openshift-marketplace/community-operators-v6msv" Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.453292 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/84c2ada5-ceab-4327-802e-9ae459ac814d-catalog-content\") pod \"community-operators-v6msv\" (UID: \"84c2ada5-ceab-4327-802e-9ae459ac814d\") " pod="openshift-marketplace/community-operators-v6msv" Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.458534 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.463904 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-f8mmz" Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.469421 4907 patch_prober.go:28] interesting pod/router-default-5444994796-f8mmz container/router namespace/openshift-ingress: Startup probe status=failure output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" start-of-body= Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.469485 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-f8mmz" podUID="c26ccc9f-e682-4b94-9539-cb2607725f71" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.528005 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k6s9c\" (UniqueName: \"kubernetes.io/projected/84c2ada5-ceab-4327-802e-9ae459ac814d-kube-api-access-k6s9c\") pod \"community-operators-v6msv\" (UID: \"84c2ada5-ceab-4327-802e-9ae459ac814d\") " pod="openshift-marketplace/community-operators-v6msv" Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.548555 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-f8mmz" podStartSLOduration=187.548537655 podStartE2EDuration="3m7.548537655s" podCreationTimestamp="2026-03-13 14:06:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:09:08.54689701 +0000 UTC m=+247.446684709" watchObservedRunningTime="2026-03-13 14:09:08.548537655 +0000 UTC m=+247.448325344" Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.553284 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.553788 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2x9qs\" (UniqueName: \"kubernetes.io/projected/8adb3c1c-bacd-4cca-9796-7ca96624e9f6-kube-api-access-2x9qs\") pod \"certified-operators-42kpk\" (UID: \"8adb3c1c-bacd-4cca-9796-7ca96624e9f6\") " pod="openshift-marketplace/certified-operators-42kpk" Mar 13 14:09:08 crc kubenswrapper[4907]: E0313 14:09:08.553927 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:09.053905521 +0000 UTC m=+247.953693210 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.554384 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8adb3c1c-bacd-4cca-9796-7ca96624e9f6-utilities\") pod \"certified-operators-42kpk\" (UID: \"8adb3c1c-bacd-4cca-9796-7ca96624e9f6\") " pod="openshift-marketplace/certified-operators-42kpk" Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.554523 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.554703 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8adb3c1c-bacd-4cca-9796-7ca96624e9f6-catalog-content\") pod \"certified-operators-42kpk\" (UID: \"8adb3c1c-bacd-4cca-9796-7ca96624e9f6\") " pod="openshift-marketplace/certified-operators-42kpk" Mar 13 14:09:08 crc kubenswrapper[4907]: E0313 14:09:08.555143 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:09.055131274 +0000 UTC m=+247.954918963 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.572204 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-56rqr" podStartSLOduration=187.572182867 podStartE2EDuration="3m7.572182867s" podCreationTimestamp="2026-03-13 14:06:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:09:08.57154052 +0000 UTC m=+247.471328209" watchObservedRunningTime="2026-03-13 14:09:08.572182867 +0000 UTC m=+247.471970556" Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.591240 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-v6msv" Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.655307 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.655591 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8adb3c1c-bacd-4cca-9796-7ca96624e9f6-catalog-content\") pod \"certified-operators-42kpk\" (UID: \"8adb3c1c-bacd-4cca-9796-7ca96624e9f6\") " pod="openshift-marketplace/certified-operators-42kpk" Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.655662 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2x9qs\" (UniqueName: \"kubernetes.io/projected/8adb3c1c-bacd-4cca-9796-7ca96624e9f6-kube-api-access-2x9qs\") pod \"certified-operators-42kpk\" (UID: \"8adb3c1c-bacd-4cca-9796-7ca96624e9f6\") " pod="openshift-marketplace/certified-operators-42kpk" Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.655693 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8adb3c1c-bacd-4cca-9796-7ca96624e9f6-utilities\") pod \"certified-operators-42kpk\" (UID: \"8adb3c1c-bacd-4cca-9796-7ca96624e9f6\") " pod="openshift-marketplace/certified-operators-42kpk" Mar 13 14:09:08 crc kubenswrapper[4907]: E0313 14:09:08.656324 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:09.156299445 +0000 UTC m=+248.056087134 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.658838 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8adb3c1c-bacd-4cca-9796-7ca96624e9f6-catalog-content\") pod \"certified-operators-42kpk\" (UID: \"8adb3c1c-bacd-4cca-9796-7ca96624e9f6\") " pod="openshift-marketplace/certified-operators-42kpk" Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.659913 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8adb3c1c-bacd-4cca-9796-7ca96624e9f6-utilities\") pod \"certified-operators-42kpk\" (UID: \"8adb3c1c-bacd-4cca-9796-7ca96624e9f6\") " pod="openshift-marketplace/certified-operators-42kpk" Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.710714 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-zbll9"] Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.712179 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zbll9" Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.712515 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2x9qs\" (UniqueName: \"kubernetes.io/projected/8adb3c1c-bacd-4cca-9796-7ca96624e9f6-kube-api-access-2x9qs\") pod \"certified-operators-42kpk\" (UID: \"8adb3c1c-bacd-4cca-9796-7ca96624e9f6\") " pod="openshift-marketplace/certified-operators-42kpk" Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.737385 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zbll9"] Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.757429 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6mhh2\" (UniqueName: \"kubernetes.io/projected/4b6ddc05-e51d-4648-84c6-6d1414398a4b-kube-api-access-6mhh2\") pod \"community-operators-zbll9\" (UID: \"4b6ddc05-e51d-4648-84c6-6d1414398a4b\") " pod="openshift-marketplace/community-operators-zbll9" Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.757492 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4b6ddc05-e51d-4648-84c6-6d1414398a4b-catalog-content\") pod \"community-operators-zbll9\" (UID: \"4b6ddc05-e51d-4648-84c6-6d1414398a4b\") " pod="openshift-marketplace/community-operators-zbll9" Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.757530 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.757557 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4b6ddc05-e51d-4648-84c6-6d1414398a4b-utilities\") pod \"community-operators-zbll9\" (UID: \"4b6ddc05-e51d-4648-84c6-6d1414398a4b\") " pod="openshift-marketplace/community-operators-zbll9" Mar 13 14:09:08 crc kubenswrapper[4907]: E0313 14:09:08.757867 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:09.257854556 +0000 UTC m=+248.157642245 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.770541 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-42kpk" Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.841302 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-gjdp7"] Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.852223 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gjdp7" Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.859870 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.860861 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6mhh2\" (UniqueName: \"kubernetes.io/projected/4b6ddc05-e51d-4648-84c6-6d1414398a4b-kube-api-access-6mhh2\") pod \"community-operators-zbll9\" (UID: \"4b6ddc05-e51d-4648-84c6-6d1414398a4b\") " pod="openshift-marketplace/community-operators-zbll9" Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.860950 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4b6ddc05-e51d-4648-84c6-6d1414398a4b-catalog-content\") pod \"community-operators-zbll9\" (UID: \"4b6ddc05-e51d-4648-84c6-6d1414398a4b\") " pod="openshift-marketplace/community-operators-zbll9" Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.861022 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f797b1bb-e769-49c4-84ce-93a6fad2df93-metrics-certs\") pod \"network-metrics-daemon-tzmvc\" (UID: \"f797b1bb-e769-49c4-84ce-93a6fad2df93\") " pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.861057 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4b6ddc05-e51d-4648-84c6-6d1414398a4b-utilities\") pod \"community-operators-zbll9\" (UID: \"4b6ddc05-e51d-4648-84c6-6d1414398a4b\") " pod="openshift-marketplace/community-operators-zbll9" Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.861802 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4b6ddc05-e51d-4648-84c6-6d1414398a4b-utilities\") pod \"community-operators-zbll9\" (UID: \"4b6ddc05-e51d-4648-84c6-6d1414398a4b\") " pod="openshift-marketplace/community-operators-zbll9" Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.867001 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4b6ddc05-e51d-4648-84c6-6d1414398a4b-catalog-content\") pod \"community-operators-zbll9\" (UID: \"4b6ddc05-e51d-4648-84c6-6d1414398a4b\") " pod="openshift-marketplace/community-operators-zbll9" Mar 13 14:09:08 crc kubenswrapper[4907]: E0313 14:09:08.867105 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:09.367051554 +0000 UTC m=+248.266839243 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.870548 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.871214 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-gjdp7"] Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.885209 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f797b1bb-e769-49c4-84ce-93a6fad2df93-metrics-certs\") pod \"network-metrics-daemon-tzmvc\" (UID: \"f797b1bb-e769-49c4-84ce-93a6fad2df93\") " pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.896087 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6mhh2\" (UniqueName: \"kubernetes.io/projected/4b6ddc05-e51d-4648-84c6-6d1414398a4b-kube-api-access-6mhh2\") pod \"community-operators-zbll9\" (UID: \"4b6ddc05-e51d-4648-84c6-6d1414398a4b\") " pod="openshift-marketplace/community-operators-zbll9" Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.966141 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6q2rc\" (UniqueName: \"kubernetes.io/projected/5ef16085-0f4d-435d-a559-20a33c237531-kube-api-access-6q2rc\") pod \"certified-operators-gjdp7\" (UID: \"5ef16085-0f4d-435d-a559-20a33c237531\") " pod="openshift-marketplace/certified-operators-gjdp7" Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.966751 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ef16085-0f4d-435d-a559-20a33c237531-utilities\") pod \"certified-operators-gjdp7\" (UID: \"5ef16085-0f4d-435d-a559-20a33c237531\") " pod="openshift-marketplace/certified-operators-gjdp7" Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.966988 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.967035 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ef16085-0f4d-435d-a559-20a33c237531-catalog-content\") pod \"certified-operators-gjdp7\" (UID: \"5ef16085-0f4d-435d-a559-20a33c237531\") " pod="openshift-marketplace/certified-operators-gjdp7" Mar 13 14:09:08 crc kubenswrapper[4907]: E0313 14:09:08.967844 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:09.467824994 +0000 UTC m=+248.367612683 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.059179 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zbll9" Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.062961 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.069130 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.069438 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ef16085-0f4d-435d-a559-20a33c237531-catalog-content\") pod \"certified-operators-gjdp7\" (UID: \"5ef16085-0f4d-435d-a559-20a33c237531\") " pod="openshift-marketplace/certified-operators-gjdp7" Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.069474 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6q2rc\" (UniqueName: \"kubernetes.io/projected/5ef16085-0f4d-435d-a559-20a33c237531-kube-api-access-6q2rc\") pod \"certified-operators-gjdp7\" (UID: \"5ef16085-0f4d-435d-a559-20a33c237531\") " pod="openshift-marketplace/certified-operators-gjdp7" Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.069519 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ef16085-0f4d-435d-a559-20a33c237531-utilities\") pod \"certified-operators-gjdp7\" (UID: \"5ef16085-0f4d-435d-a559-20a33c237531\") " pod="openshift-marketplace/certified-operators-gjdp7" Mar 13 14:09:09 crc kubenswrapper[4907]: E0313 14:09:09.070021 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:09.569988542 +0000 UTC m=+248.469776231 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.070109 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ef16085-0f4d-435d-a559-20a33c237531-utilities\") pod \"certified-operators-gjdp7\" (UID: \"5ef16085-0f4d-435d-a559-20a33c237531\") " pod="openshift-marketplace/certified-operators-gjdp7" Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.070330 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ef16085-0f4d-435d-a559-20a33c237531-catalog-content\") pod \"certified-operators-gjdp7\" (UID: \"5ef16085-0f4d-435d-a559-20a33c237531\") " pod="openshift-marketplace/certified-operators-gjdp7" Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.073339 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.086800 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-v6msv"] Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.102756 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6q2rc\" (UniqueName: \"kubernetes.io/projected/5ef16085-0f4d-435d-a559-20a33c237531-kube-api-access-6q2rc\") pod \"certified-operators-gjdp7\" (UID: \"5ef16085-0f4d-435d-a559-20a33c237531\") " pod="openshift-marketplace/certified-operators-gjdp7" Mar 13 14:09:09 crc kubenswrapper[4907]: W0313 14:09:09.157910 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod84c2ada5_ceab_4327_802e_9ae459ac814d.slice/crio-2b528cfb1a9dfb9a8a4cae32fcb0401d81595581c8b3e1b33eacedab76ae1cd0 WatchSource:0}: Error finding container 2b528cfb1a9dfb9a8a4cae32fcb0401d81595581c8b3e1b33eacedab76ae1cd0: Status 404 returned error can't find the container with id 2b528cfb1a9dfb9a8a4cae32fcb0401d81595581c8b3e1b33eacedab76ae1cd0 Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.171000 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:09 crc kubenswrapper[4907]: E0313 14:09:09.171553 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:09.671538592 +0000 UTC m=+248.571326281 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.205086 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4h9q7" event={"ID":"9bb1952d-0e55-4e45-81c0-69a8cddb827e","Type":"ContainerStarted","Data":"9435fd54ef37b6b3b6843aafb8ac4987dfdaac5f9146401bd984cd364cd498e2"} Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.205667 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4h9q7" Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.215077 4907 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-4h9q7 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.8:8443/healthz\": dial tcp 10.217.0.8:8443: connect: connection refused" start-of-body= Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.215151 4907 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4h9q7" podUID="9bb1952d-0e55-4e45-81c0-69a8cddb827e" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.8:8443/healthz\": dial tcp 10.217.0.8:8443: connect: connection refused" Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.216222 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-xkdjh" event={"ID":"ebd59055-333a-46c2-ba30-826296414a4c","Type":"ContainerStarted","Data":"96d1942009f1b8ce51bc3ce645a94bc1cff961168b69aaf647356e722a6a87aa"} Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.222353 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-sjwvn" event={"ID":"4a8597a4-c72f-4023-b2fb-1b1237411a36","Type":"ContainerStarted","Data":"e31cc3fa7fa433ae9637724792028e91568f4c8e8b4df50891f201ffb1364ddd"} Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.235209 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4h9q7" podStartSLOduration=188.235158972 podStartE2EDuration="3m8.235158972s" podCreationTimestamp="2026-03-13 14:06:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:09:09.233752894 +0000 UTC m=+248.133540583" watchObservedRunningTime="2026-03-13 14:09:09.235158972 +0000 UTC m=+248.134946661" Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.257118 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-42kpk"] Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.261047 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-wjrms" event={"ID":"71c234e9-5f3d-4b0c-94d4-80b4e33a6dda","Type":"ContainerStarted","Data":"bd285704b82eac1d7e6f05e8c5ba3360170af176524b5e7ebe3a2fe98c0557fb"} Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.272514 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:09 crc kubenswrapper[4907]: E0313 14:09:09.272846 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:09.772825726 +0000 UTC m=+248.672613415 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.281405 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-wjrms" podStartSLOduration=189.281382749 podStartE2EDuration="3m9.281382749s" podCreationTimestamp="2026-03-13 14:06:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:09:09.280453593 +0000 UTC m=+248.180241282" watchObservedRunningTime="2026-03-13 14:09:09.281382749 +0000 UTC m=+248.181170438" Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.281900 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-djph7" event={"ID":"3a661298-271c-4c57-94ea-3d69a27d7ad5","Type":"ContainerStarted","Data":"9a8f3e0d2c67c4386ad218f76fa5059a7db512cd7dfd04bfe2063252587ce4c2"} Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.315655 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gjdp7" Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.329321 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ksgr6" event={"ID":"1be912b5-9d1d-48bd-a0ab-7f601f9eb9da","Type":"ContainerStarted","Data":"a96675938619010b9f097dc54780a8799b01e25aee1e330d3f5306f2909342f1"} Mar 13 14:09:09 crc kubenswrapper[4907]: W0313 14:09:09.331375 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8adb3c1c_bacd_4cca_9796_7ca96624e9f6.slice/crio-04064aa9c533eaef085c99222c5a674a4c851c29b8b6bd8d0141f59c380540da WatchSource:0}: Error finding container 04064aa9c533eaef085c99222c5a674a4c851c29b8b6bd8d0141f59c380540da: Status 404 returned error can't find the container with id 04064aa9c533eaef085c99222c5a674a4c851c29b8b6bd8d0141f59c380540da Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.335822 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v6msv" event={"ID":"84c2ada5-ceab-4327-802e-9ae459ac814d","Type":"ContainerStarted","Data":"2b528cfb1a9dfb9a8a4cae32fcb0401d81595581c8b3e1b33eacedab76ae1cd0"} Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.353219 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-f8mmz" event={"ID":"c26ccc9f-e682-4b94-9539-cb2607725f71","Type":"ContainerStarted","Data":"787a19aa46c50e9190e7b8ab051f8b3388c5f04f9cff3b41bfd80250e281730a"} Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.375340 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-l9xpb" event={"ID":"4c8c0266-5452-4abb-92c5-c536df94dd41","Type":"ContainerStarted","Data":"27a56c8f2bc88df1fc0af00ca4f2cc9f8243a27d536703bf1460c381978493aa"} Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.376478 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:09 crc kubenswrapper[4907]: E0313 14:09:09.377764 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:09.877749039 +0000 UTC m=+248.777536728 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.395363 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ksgr6" podStartSLOduration=188.395336616 podStartE2EDuration="3m8.395336616s" podCreationTimestamp="2026-03-13 14:06:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:09:09.375024285 +0000 UTC m=+248.274811974" watchObservedRunningTime="2026-03-13 14:09:09.395336616 +0000 UTC m=+248.295124305" Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.430739 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-56rqr" event={"ID":"59980734-2cf7-42dc-9f6f-a6de7e1c3665","Type":"ContainerStarted","Data":"0a47e925b05361e00dba14ee9fb8ae5b234f948b23c9d549a757d4788ccadd08"} Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.470839 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-rmkfg" event={"ID":"0396c426-1d56-44e6-885a-1b97d69ad57a","Type":"ContainerStarted","Data":"b37cfd7cc95bf8691f1a46c915bb50b3bf1a28b831571a3218c322339ced6591"} Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.472021 4907 patch_prober.go:28] interesting pod/router-default-5444994796-f8mmz container/router namespace/openshift-ingress: Startup probe status=failure output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" start-of-body= Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.472090 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-f8mmz" podUID="c26ccc9f-e682-4b94-9539-cb2607725f71" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.478799 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:09 crc kubenswrapper[4907]: E0313 14:09:09.480680 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:09.980663796 +0000 UTC m=+248.880451485 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.503851 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-l9xpb" podStartSLOduration=188.503823896 podStartE2EDuration="3m8.503823896s" podCreationTimestamp="2026-03-13 14:06:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:09:09.396318573 +0000 UTC m=+248.296106262" watchObservedRunningTime="2026-03-13 14:09:09.503823896 +0000 UTC m=+248.403611585" Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.504224 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-rmkfg" podStartSLOduration=188.504217757 podStartE2EDuration="3m8.504217757s" podCreationTimestamp="2026-03-13 14:06:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:09:09.501628476 +0000 UTC m=+248.401416165" watchObservedRunningTime="2026-03-13 14:09:09.504217757 +0000 UTC m=+248.404005446" Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.542208 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-8jqgq" event={"ID":"fa265490-c23b-4b05-94bc-78b4a636e065","Type":"ContainerStarted","Data":"d87d12281854a7a4b9fc5df3cb864ebff9b2ac551f657b745dd2c8769835232b"} Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.542394 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-8jqgq" Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.560995 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-vfbsx" event={"ID":"814322e3-6251-4ea3-a86b-7b89ca6ea728","Type":"ContainerStarted","Data":"8e1acd7b778f9d3afc64ccd1ecf628a736572fda7bb48f3f9f0fce5bf77942f6"} Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.561039 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-vfbsx" event={"ID":"814322e3-6251-4ea3-a86b-7b89ca6ea728","Type":"ContainerStarted","Data":"a6498ada9850a819ec9f7e5a616b843cca7a2508a87755d09458256c9b442008"} Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.578516 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-hwf6r" event={"ID":"699da377-c61b-48f0-92e0-42bd247ee363","Type":"ContainerStarted","Data":"fa1cb4cbad0b730f0bb9a1a91d74909267b6ff61df39cbede52916f05f18ad8e"} Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.580429 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:09 crc kubenswrapper[4907]: E0313 14:09:09.582793 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:10.082780692 +0000 UTC m=+248.982568381 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.589912 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-w76kw" event={"ID":"9b75954c-b75d-4f17-82dc-4c0358fd8d0f","Type":"ContainerStarted","Data":"55c6769fb46a55fb50d7205a8ebc8698e80e046f65ffce5e932d9ab6bc93eae2"} Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.590303 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-w76kw" Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.619020 4907 patch_prober.go:28] interesting pod/console-operator-58897d9998-w76kw container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.13:8443/readyz\": dial tcp 10.217.0.13:8443: connect: connection refused" start-of-body= Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.619074 4907 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-w76kw" podUID="9b75954c-b75d-4f17-82dc-4c0358fd8d0f" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.13:8443/readyz\": dial tcp 10.217.0.13:8443: connect: connection refused" Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.633002 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-8jqgq" podStartSLOduration=9.632982837 podStartE2EDuration="9.632982837s" podCreationTimestamp="2026-03-13 14:09:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:09:09.62277182 +0000 UTC m=+248.522559509" watchObservedRunningTime="2026-03-13 14:09:09.632982837 +0000 UTC m=+248.532770526" Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.646228 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-ksfb8" event={"ID":"8013cba2-6b3c-4548-af5c-6dd0057da0c7","Type":"ContainerStarted","Data":"24c210dbfab7f5c7d92cdf58d42ef582dc91f8401eda4940a42300f5067b3f1b"} Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.682777 4907 generic.go:334] "Generic (PLEG): container finished" podID="5a9c889b-5954-4d3a-ae72-12000ee1cc25" containerID="4651f82e390b3a5ac2c55407710eda6955584915c8001c9a1dc4182061851ee7" exitCode=0 Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.682900 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-r86r4" event={"ID":"5a9c889b-5954-4d3a-ae72-12000ee1cc25","Type":"ContainerDied","Data":"4651f82e390b3a5ac2c55407710eda6955584915c8001c9a1dc4182061851ee7"} Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.682948 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-r86r4" event={"ID":"5a9c889b-5954-4d3a-ae72-12000ee1cc25","Type":"ContainerStarted","Data":"bcc0a7feb3950443a39009d894a278b8d6140c0e85b685de66a2bb1189e17496"} Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.685322 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:09 crc kubenswrapper[4907]: E0313 14:09:09.685587 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:10.185565197 +0000 UTC m=+249.085352936 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.685821 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:09 crc kubenswrapper[4907]: E0313 14:09:09.689602 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:10.189594346 +0000 UTC m=+249.089382035 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.715831 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-vfbsx" podStartSLOduration=188.715810129 podStartE2EDuration="3m8.715810129s" podCreationTimestamp="2026-03-13 14:06:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:09:09.661921774 +0000 UTC m=+248.561709463" watchObservedRunningTime="2026-03-13 14:09:09.715810129 +0000 UTC m=+248.615597818" Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.717480 4907 generic.go:334] "Generic (PLEG): container finished" podID="19edf37b-a1d3-4c7c-945e-5d0b8bb601bf" containerID="d2284227e021b37d9e8d2812f8621bebf16efd8eeb5019b51cfa28773e7a5aa9" exitCode=0 Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.717557 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556840-pgmwj" event={"ID":"19edf37b-a1d3-4c7c-945e-5d0b8bb601bf","Type":"ContainerDied","Data":"d2284227e021b37d9e8d2812f8621bebf16efd8eeb5019b51cfa28773e7a5aa9"} Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.739002 4907 generic.go:334] "Generic (PLEG): container finished" podID="0149e083-ed5b-4c7f-94e7-4f3f2ff8a557" containerID="40f0b1a25cd0d0a5030ec1dcae203661c52d45d8711eefed49ad300a5495f0ac" exitCode=0 Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.739772 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-wgpdd" event={"ID":"0149e083-ed5b-4c7f-94e7-4f3f2ff8a557","Type":"ContainerDied","Data":"40f0b1a25cd0d0a5030ec1dcae203661c52d45d8711eefed49ad300a5495f0ac"} Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.743732 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-92gs7" event={"ID":"74b29e18-7d73-4ef5-9739-04137dd1b191","Type":"ContainerStarted","Data":"ba1f38dc924a82f64021f6ef06f3ceb7e7b06c76b8f07b0d1bffc24eb269af27"} Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.781607 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5szhc" event={"ID":"ae320273-06e8-43c5-a64f-acd80ad16d1c","Type":"ContainerStarted","Data":"e7cd020657b593a0fe31337cf7b9bcad7216063124dfc53dafa29723e59f8c19"} Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.784045 4907 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-5stnc container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.35:8080/healthz\": dial tcp 10.217.0.35:8080: connect: connection refused" start-of-body= Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.784101 4907 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-5stnc" podUID="0290eb6f-0be7-4340-8ba8-ed6ea0662c33" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.35:8080/healthz\": dial tcp 10.217.0.35:8080: connect: connection refused" Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.785363 4907 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-kwfhg container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.7:8443/healthz\": dial tcp 10.217.0.7:8443: connect: connection refused" start-of-body= Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.785394 4907 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-kwfhg" podUID="814c6dec-6e96-4068-9aec-9e57fefca5db" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.7:8443/healthz\": dial tcp 10.217.0.7:8443: connect: connection refused" Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.785439 4907 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-ng4kz container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.11:6443/healthz\": dial tcp 10.217.0.11:6443: connect: connection refused" start-of-body= Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.785451 4907 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-ng4kz" podUID="3661f6b2-1d5e-42f0-947f-a32b66258a1f" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.11:6443/healthz\": dial tcp 10.217.0.11:6443: connect: connection refused" Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.785733 4907 patch_prober.go:28] interesting pod/downloads-7954f5f757-544df container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.785749 4907 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-544df" podUID="f8ea4f24-05b0-4661-a617-a4208f9a9188" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.785791 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-hwf6r" podStartSLOduration=188.785782031 podStartE2EDuration="3m8.785782031s" podCreationTimestamp="2026-03-13 14:06:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:09:09.718323038 +0000 UTC m=+248.618110737" watchObservedRunningTime="2026-03-13 14:09:09.785782031 +0000 UTC m=+248.685569720" Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.787182 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-tzmvc"] Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.787307 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:09 crc kubenswrapper[4907]: E0313 14:09:09.787510 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:10.287489738 +0000 UTC m=+249.187277427 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.787602 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:09 crc kubenswrapper[4907]: E0313 14:09:09.787846 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:10.287840428 +0000 UTC m=+249.187628117 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.816226 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ksgr6" Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.816258 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ksgr6" Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.840992 4907 patch_prober.go:28] interesting pod/apiserver-7bbb656c7d-ksgr6 container/oauth-apiserver namespace/openshift-oauth-apiserver: Startup probe status=failure output="Get \"https://10.217.0.6:8443/livez\": dial tcp 10.217.0.6:8443: connect: connection refused" start-of-body= Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.841041 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ksgr6" podUID="1be912b5-9d1d-48bd-a0ab-7f601f9eb9da" containerName="oauth-apiserver" probeResult="failure" output="Get \"https://10.217.0.6:8443/livez\": dial tcp 10.217.0.6:8443: connect: connection refused" Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.875725 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-w76kw" podStartSLOduration=188.875702666 podStartE2EDuration="3m8.875702666s" podCreationTimestamp="2026-03-13 14:06:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:09:09.809870116 +0000 UTC m=+248.709657805" watchObservedRunningTime="2026-03-13 14:09:09.875702666 +0000 UTC m=+248.775490355" Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.890153 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:09 crc kubenswrapper[4907]: E0313 14:09:09.905866 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:10.405819255 +0000 UTC m=+249.305606944 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.922422 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-92gs7" podStartSLOduration=188.922397646 podStartE2EDuration="3m8.922397646s" podCreationTimestamp="2026-03-13 14:06:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:09:09.922045877 +0000 UTC m=+248.821833566" watchObservedRunningTime="2026-03-13 14:09:09.922397646 +0000 UTC m=+248.822185335" Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.002488 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:10 crc kubenswrapper[4907]: E0313 14:09:10.003176 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:10.503155931 +0000 UTC m=+249.402943620 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.011621 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-ksfb8" podStartSLOduration=189.01157894 podStartE2EDuration="3m9.01157894s" podCreationTimestamp="2026-03-13 14:06:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:09:09.984546605 +0000 UTC m=+248.884334294" watchObservedRunningTime="2026-03-13 14:09:10.01157894 +0000 UTC m=+248.911366629" Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.012868 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5szhc" podStartSLOduration=189.012863375 podStartE2EDuration="3m9.012863375s" podCreationTimestamp="2026-03-13 14:06:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:09:10.003925222 +0000 UTC m=+248.903712911" watchObservedRunningTime="2026-03-13 14:09:10.012863375 +0000 UTC m=+248.912651064" Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.107682 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:10 crc kubenswrapper[4907]: E0313 14:09:10.108481 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:10.608459584 +0000 UTC m=+249.508247283 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.172194 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zbll9"] Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.185219 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-gjdp7"] Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.209268 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:10 crc kubenswrapper[4907]: E0313 14:09:10.209595 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:10.709584203 +0000 UTC m=+249.609371892 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:10 crc kubenswrapper[4907]: W0313 14:09:10.237143 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5ef16085_0f4d_435d_a559_20a33c237531.slice/crio-5380ac5cf0f032bd661fec9fad25b22b2b931ad5cc5954ae70fb9c5f8362990a WatchSource:0}: Error finding container 5380ac5cf0f032bd661fec9fad25b22b2b931ad5cc5954ae70fb9c5f8362990a: Status 404 returned error can't find the container with id 5380ac5cf0f032bd661fec9fad25b22b2b931ad5cc5954ae70fb9c5f8362990a Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.310951 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:10 crc kubenswrapper[4907]: E0313 14:09:10.311312 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:10.811295009 +0000 UTC m=+249.711082698 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.311423 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:10 crc kubenswrapper[4907]: E0313 14:09:10.311752 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:10.811743281 +0000 UTC m=+249.711530970 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.412252 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:10 crc kubenswrapper[4907]: E0313 14:09:10.412744 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:10.912725346 +0000 UTC m=+249.812513035 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.414785 4907 ???:1] "http: TLS handshake error from 192.168.126.11:55010: no serving certificate available for the kubelet" Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.429219 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-jk5nw"] Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.430501 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jk5nw" Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.433774 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.447618 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jk5nw"] Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.478095 4907 patch_prober.go:28] interesting pod/router-default-5444994796-f8mmz container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 13 14:09:10 crc kubenswrapper[4907]: [-]has-synced failed: reason withheld Mar 13 14:09:10 crc kubenswrapper[4907]: [+]process-running ok Mar 13 14:09:10 crc kubenswrapper[4907]: healthz check failed Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.478618 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-f8mmz" podUID="c26ccc9f-e682-4b94-9539-cb2607725f71" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.508137 4907 ???:1] "http: TLS handshake error from 192.168.126.11:55018: no serving certificate available for the kubelet" Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.513587 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.513684 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d342a6a2-43dd-4d15-8aff-f60f9a3383bf-catalog-content\") pod \"redhat-marketplace-jk5nw\" (UID: \"d342a6a2-43dd-4d15-8aff-f60f9a3383bf\") " pod="openshift-marketplace/redhat-marketplace-jk5nw" Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.513738 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-msd8t\" (UniqueName: \"kubernetes.io/projected/d342a6a2-43dd-4d15-8aff-f60f9a3383bf-kube-api-access-msd8t\") pod \"redhat-marketplace-jk5nw\" (UID: \"d342a6a2-43dd-4d15-8aff-f60f9a3383bf\") " pod="openshift-marketplace/redhat-marketplace-jk5nw" Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.513765 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d342a6a2-43dd-4d15-8aff-f60f9a3383bf-utilities\") pod \"redhat-marketplace-jk5nw\" (UID: \"d342a6a2-43dd-4d15-8aff-f60f9a3383bf\") " pod="openshift-marketplace/redhat-marketplace-jk5nw" Mar 13 14:09:10 crc kubenswrapper[4907]: E0313 14:09:10.514084 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:11.014072211 +0000 UTC m=+249.913859900 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.615117 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.615312 4907 ???:1] "http: TLS handshake error from 192.168.126.11:55024: no serving certificate available for the kubelet" Mar 13 14:09:10 crc kubenswrapper[4907]: E0313 14:09:10.615331 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:11.115298194 +0000 UTC m=+250.015085883 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.615442 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.615568 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d342a6a2-43dd-4d15-8aff-f60f9a3383bf-catalog-content\") pod \"redhat-marketplace-jk5nw\" (UID: \"d342a6a2-43dd-4d15-8aff-f60f9a3383bf\") " pod="openshift-marketplace/redhat-marketplace-jk5nw" Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.615620 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-msd8t\" (UniqueName: \"kubernetes.io/projected/d342a6a2-43dd-4d15-8aff-f60f9a3383bf-kube-api-access-msd8t\") pod \"redhat-marketplace-jk5nw\" (UID: \"d342a6a2-43dd-4d15-8aff-f60f9a3383bf\") " pod="openshift-marketplace/redhat-marketplace-jk5nw" Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.615678 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d342a6a2-43dd-4d15-8aff-f60f9a3383bf-utilities\") pod \"redhat-marketplace-jk5nw\" (UID: \"d342a6a2-43dd-4d15-8aff-f60f9a3383bf\") " pod="openshift-marketplace/redhat-marketplace-jk5nw" Mar 13 14:09:10 crc kubenswrapper[4907]: E0313 14:09:10.615718 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:11.115705885 +0000 UTC m=+250.015493574 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.616259 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d342a6a2-43dd-4d15-8aff-f60f9a3383bf-utilities\") pod \"redhat-marketplace-jk5nw\" (UID: \"d342a6a2-43dd-4d15-8aff-f60f9a3383bf\") " pod="openshift-marketplace/redhat-marketplace-jk5nw" Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.616281 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d342a6a2-43dd-4d15-8aff-f60f9a3383bf-catalog-content\") pod \"redhat-marketplace-jk5nw\" (UID: \"d342a6a2-43dd-4d15-8aff-f60f9a3383bf\") " pod="openshift-marketplace/redhat-marketplace-jk5nw" Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.643159 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-msd8t\" (UniqueName: \"kubernetes.io/projected/d342a6a2-43dd-4d15-8aff-f60f9a3383bf-kube-api-access-msd8t\") pod \"redhat-marketplace-jk5nw\" (UID: \"d342a6a2-43dd-4d15-8aff-f60f9a3383bf\") " pod="openshift-marketplace/redhat-marketplace-jk5nw" Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.701554 4907 ???:1] "http: TLS handshake error from 192.168.126.11:55040: no serving certificate available for the kubelet" Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.717300 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:10 crc kubenswrapper[4907]: E0313 14:09:10.717495 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:11.217446191 +0000 UTC m=+250.117233880 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.717673 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:10 crc kubenswrapper[4907]: E0313 14:09:10.718051 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:11.218041307 +0000 UTC m=+250.117828996 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.772588 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jk5nw" Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.814675 4907 ???:1] "http: TLS handshake error from 192.168.126.11:55050: no serving certificate available for the kubelet" Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.817736 4907 generic.go:334] "Generic (PLEG): container finished" podID="84c2ada5-ceab-4327-802e-9ae459ac814d" containerID="0bd7973d57fde27ab0e7adf532d86c90333b4abd96119d9c054a7fb89cbb6dd6" exitCode=0 Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.818370 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v6msv" event={"ID":"84c2ada5-ceab-4327-802e-9ae459ac814d","Type":"ContainerDied","Data":"0bd7973d57fde27ab0e7adf532d86c90333b4abd96119d9c054a7fb89cbb6dd6"} Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.818680 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:10 crc kubenswrapper[4907]: E0313 14:09:10.818975 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:11.31895735 +0000 UTC m=+250.218745039 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.829318 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-r86r4" event={"ID":"5a9c889b-5954-4d3a-ae72-12000ee1cc25","Type":"ContainerStarted","Data":"fbd6a73b2a9ffd348f53b7ffb74a46e3e4d4a0debddae3fc1d71bbfc0b464b7b"} Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.837144 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-p257z"] Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.841839 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-p257z" Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.846234 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-xkdjh" event={"ID":"ebd59055-333a-46c2-ba30-826296414a4c","Type":"ContainerStarted","Data":"008bbcbd9d87f286dcc70fceddeacf79d404fa4a3cc2853e4736cb3009c44561"} Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.864522 4907 generic.go:334] "Generic (PLEG): container finished" podID="4b6ddc05-e51d-4648-84c6-6d1414398a4b" containerID="a3e791567d7fec33aa07ec52ed2c7bf531de23472ee1120fe2b93a41d79ec2e7" exitCode=0 Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.864593 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zbll9" event={"ID":"4b6ddc05-e51d-4648-84c6-6d1414398a4b","Type":"ContainerDied","Data":"a3e791567d7fec33aa07ec52ed2c7bf531de23472ee1120fe2b93a41d79ec2e7"} Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.864619 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zbll9" event={"ID":"4b6ddc05-e51d-4648-84c6-6d1414398a4b","Type":"ContainerStarted","Data":"595859a9bf32c878bf56920b076fb65741978b6d9f83d6c4a352d65ef0a38504"} Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.872889 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-p257z"] Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.884553 4907 generic.go:334] "Generic (PLEG): container finished" podID="5ef16085-0f4d-435d-a559-20a33c237531" containerID="7e15a77e2ae6c9b87a9062d382cad40c9ca694c03f35d89b5511dd88c400a20c" exitCode=0 Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.884644 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gjdp7" event={"ID":"5ef16085-0f4d-435d-a559-20a33c237531","Type":"ContainerDied","Data":"7e15a77e2ae6c9b87a9062d382cad40c9ca694c03f35d89b5511dd88c400a20c"} Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.884672 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gjdp7" event={"ID":"5ef16085-0f4d-435d-a559-20a33c237531","Type":"ContainerStarted","Data":"5380ac5cf0f032bd661fec9fad25b22b2b931ad5cc5954ae70fb9c5f8362990a"} Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.899664 4907 generic.go:334] "Generic (PLEG): container finished" podID="8adb3c1c-bacd-4cca-9796-7ca96624e9f6" containerID="0d444f1a14d20e1e6a2efc71bfbe0eb77305492df11ed2559d59993943816ad1" exitCode=0 Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.899750 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-42kpk" event={"ID":"8adb3c1c-bacd-4cca-9796-7ca96624e9f6","Type":"ContainerDied","Data":"0d444f1a14d20e1e6a2efc71bfbe0eb77305492df11ed2559d59993943816ad1"} Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.899783 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-42kpk" event={"ID":"8adb3c1c-bacd-4cca-9796-7ca96624e9f6","Type":"ContainerStarted","Data":"04064aa9c533eaef085c99222c5a674a4c851c29b8b6bd8d0141f59c380540da"} Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.901922 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-tzmvc" event={"ID":"f797b1bb-e769-49c4-84ce-93a6fad2df93","Type":"ContainerStarted","Data":"559726274eea9d689eaef1bf7a9391dd2a474c1381281965a68d454600e7e5c0"} Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.901959 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-tzmvc" event={"ID":"f797b1bb-e769-49c4-84ce-93a6fad2df93","Type":"ContainerStarted","Data":"1a6c62d66ca1eb5eb6d74358b2a1c9d9d864fc4e20db8677f3c3e89606814482"} Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.913851 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-r86r4" podStartSLOduration=190.91383832 podStartE2EDuration="3m10.91383832s" podCreationTimestamp="2026-03-13 14:06:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:09:10.912277427 +0000 UTC m=+249.812065116" watchObservedRunningTime="2026-03-13 14:09:10.91383832 +0000 UTC m=+249.813626009" Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.920374 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.920471 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/775414ea-5b94-4ab6-8b7c-fe7c80671ced-utilities\") pod \"redhat-marketplace-p257z\" (UID: \"775414ea-5b94-4ab6-8b7c-fe7c80671ced\") " pod="openshift-marketplace/redhat-marketplace-p257z" Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.920535 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rjgtx\" (UniqueName: \"kubernetes.io/projected/775414ea-5b94-4ab6-8b7c-fe7c80671ced-kube-api-access-rjgtx\") pod \"redhat-marketplace-p257z\" (UID: \"775414ea-5b94-4ab6-8b7c-fe7c80671ced\") " pod="openshift-marketplace/redhat-marketplace-p257z" Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.920592 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/775414ea-5b94-4ab6-8b7c-fe7c80671ced-catalog-content\") pod \"redhat-marketplace-p257z\" (UID: \"775414ea-5b94-4ab6-8b7c-fe7c80671ced\") " pod="openshift-marketplace/redhat-marketplace-p257z" Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.922868 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-wgpdd" event={"ID":"0149e083-ed5b-4c7f-94e7-4f3f2ff8a557","Type":"ContainerStarted","Data":"9858b9ab2ce7880d483ef8f05b8275700629d4165e882f596ef9ef70ad0538c9"} Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.923983 4907 patch_prober.go:28] interesting pod/console-operator-58897d9998-w76kw container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.13:8443/readyz\": dial tcp 10.217.0.13:8443: connect: connection refused" start-of-body= Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.924048 4907 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-w76kw" podUID="9b75954c-b75d-4f17-82dc-4c0358fd8d0f" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.13:8443/readyz\": dial tcp 10.217.0.13:8443: connect: connection refused" Mar 13 14:09:10 crc kubenswrapper[4907]: E0313 14:09:10.925049 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:11.425023944 +0000 UTC m=+250.324811633 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.928125 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-wgpdd" Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.929126 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-ng4kz" Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.929562 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-kwfhg" Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.935897 4907 ???:1] "http: TLS handshake error from 192.168.126.11:55056: no serving certificate available for the kubelet" Mar 13 14:09:11 crc kubenswrapper[4907]: I0313 14:09:11.050061 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:11 crc kubenswrapper[4907]: I0313 14:09:11.050443 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rjgtx\" (UniqueName: \"kubernetes.io/projected/775414ea-5b94-4ab6-8b7c-fe7c80671ced-kube-api-access-rjgtx\") pod \"redhat-marketplace-p257z\" (UID: \"775414ea-5b94-4ab6-8b7c-fe7c80671ced\") " pod="openshift-marketplace/redhat-marketplace-p257z" Mar 13 14:09:11 crc kubenswrapper[4907]: I0313 14:09:11.050724 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/775414ea-5b94-4ab6-8b7c-fe7c80671ced-catalog-content\") pod \"redhat-marketplace-p257z\" (UID: \"775414ea-5b94-4ab6-8b7c-fe7c80671ced\") " pod="openshift-marketplace/redhat-marketplace-p257z" Mar 13 14:09:11 crc kubenswrapper[4907]: I0313 14:09:11.051264 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/775414ea-5b94-4ab6-8b7c-fe7c80671ced-utilities\") pod \"redhat-marketplace-p257z\" (UID: \"775414ea-5b94-4ab6-8b7c-fe7c80671ced\") " pod="openshift-marketplace/redhat-marketplace-p257z" Mar 13 14:09:11 crc kubenswrapper[4907]: E0313 14:09:11.057289 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:11.557244568 +0000 UTC m=+250.457032257 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:11 crc kubenswrapper[4907]: I0313 14:09:11.058003 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/775414ea-5b94-4ab6-8b7c-fe7c80671ced-utilities\") pod \"redhat-marketplace-p257z\" (UID: \"775414ea-5b94-4ab6-8b7c-fe7c80671ced\") " pod="openshift-marketplace/redhat-marketplace-p257z" Mar 13 14:09:11 crc kubenswrapper[4907]: I0313 14:09:11.060811 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/775414ea-5b94-4ab6-8b7c-fe7c80671ced-catalog-content\") pod \"redhat-marketplace-p257z\" (UID: \"775414ea-5b94-4ab6-8b7c-fe7c80671ced\") " pod="openshift-marketplace/redhat-marketplace-p257z" Mar 13 14:09:11 crc kubenswrapper[4907]: I0313 14:09:11.096956 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rjgtx\" (UniqueName: \"kubernetes.io/projected/775414ea-5b94-4ab6-8b7c-fe7c80671ced-kube-api-access-rjgtx\") pod \"redhat-marketplace-p257z\" (UID: \"775414ea-5b94-4ab6-8b7c-fe7c80671ced\") " pod="openshift-marketplace/redhat-marketplace-p257z" Mar 13 14:09:11 crc kubenswrapper[4907]: I0313 14:09:11.123268 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-xkdjh" podStartSLOduration=190.123250052 podStartE2EDuration="3m10.123250052s" podCreationTimestamp="2026-03-13 14:06:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:09:11.120388465 +0000 UTC m=+250.020176164" watchObservedRunningTime="2026-03-13 14:09:11.123250052 +0000 UTC m=+250.023037741" Mar 13 14:09:11 crc kubenswrapper[4907]: I0313 14:09:11.152314 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:11 crc kubenswrapper[4907]: E0313 14:09:11.153062 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:11.653050393 +0000 UTC m=+250.552838082 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:11 crc kubenswrapper[4907]: I0313 14:09:11.155181 4907 ???:1] "http: TLS handshake error from 192.168.126.11:55062: no serving certificate available for the kubelet" Mar 13 14:09:11 crc kubenswrapper[4907]: I0313 14:09:11.200566 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-p257z" Mar 13 14:09:11 crc kubenswrapper[4907]: I0313 14:09:11.254251 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:11 crc kubenswrapper[4907]: E0313 14:09:11.255012 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:11.754995335 +0000 UTC m=+250.654783024 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:11 crc kubenswrapper[4907]: I0313 14:09:11.302531 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4h9q7" Mar 13 14:09:11 crc kubenswrapper[4907]: I0313 14:09:11.335373 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-wgpdd" podStartSLOduration=191.335353549 podStartE2EDuration="3m11.335353549s" podCreationTimestamp="2026-03-13 14:06:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:09:11.301098538 +0000 UTC m=+250.200886227" watchObservedRunningTime="2026-03-13 14:09:11.335353549 +0000 UTC m=+250.235141238" Mar 13 14:09:11 crc kubenswrapper[4907]: I0313 14:09:11.355849 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:11 crc kubenswrapper[4907]: E0313 14:09:11.356206 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:11.856193836 +0000 UTC m=+250.755981525 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:11 crc kubenswrapper[4907]: I0313 14:09:11.443228 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-fmzmc"] Mar 13 14:09:11 crc kubenswrapper[4907]: I0313 14:09:11.444205 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fmzmc" Mar 13 14:09:11 crc kubenswrapper[4907]: I0313 14:09:11.458198 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:11 crc kubenswrapper[4907]: E0313 14:09:11.458659 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:11.958633601 +0000 UTC m=+250.858421290 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:11 crc kubenswrapper[4907]: I0313 14:09:11.458970 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Mar 13 14:09:11 crc kubenswrapper[4907]: I0313 14:09:11.474650 4907 patch_prober.go:28] interesting pod/router-default-5444994796-f8mmz container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 13 14:09:11 crc kubenswrapper[4907]: [-]has-synced failed: reason withheld Mar 13 14:09:11 crc kubenswrapper[4907]: [+]process-running ok Mar 13 14:09:11 crc kubenswrapper[4907]: healthz check failed Mar 13 14:09:11 crc kubenswrapper[4907]: I0313 14:09:11.474710 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-f8mmz" podUID="c26ccc9f-e682-4b94-9539-cb2607725f71" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 14:09:11 crc kubenswrapper[4907]: I0313 14:09:11.481991 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-fmzmc"] Mar 13 14:09:11 crc kubenswrapper[4907]: I0313 14:09:11.556991 4907 ???:1] "http: TLS handshake error from 192.168.126.11:55068: no serving certificate available for the kubelet" Mar 13 14:09:11 crc kubenswrapper[4907]: I0313 14:09:11.562107 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:11 crc kubenswrapper[4907]: I0313 14:09:11.562203 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ec3466c2-4954-459d-88d8-e8c6cbf309e8-catalog-content\") pod \"redhat-operators-fmzmc\" (UID: \"ec3466c2-4954-459d-88d8-e8c6cbf309e8\") " pod="openshift-marketplace/redhat-operators-fmzmc" Mar 13 14:09:11 crc kubenswrapper[4907]: I0313 14:09:11.562297 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-48ddb\" (UniqueName: \"kubernetes.io/projected/ec3466c2-4954-459d-88d8-e8c6cbf309e8-kube-api-access-48ddb\") pod \"redhat-operators-fmzmc\" (UID: \"ec3466c2-4954-459d-88d8-e8c6cbf309e8\") " pod="openshift-marketplace/redhat-operators-fmzmc" Mar 13 14:09:11 crc kubenswrapper[4907]: I0313 14:09:11.562355 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ec3466c2-4954-459d-88d8-e8c6cbf309e8-utilities\") pod \"redhat-operators-fmzmc\" (UID: \"ec3466c2-4954-459d-88d8-e8c6cbf309e8\") " pod="openshift-marketplace/redhat-operators-fmzmc" Mar 13 14:09:11 crc kubenswrapper[4907]: E0313 14:09:11.562840 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:12.062825023 +0000 UTC m=+250.962612702 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:11 crc kubenswrapper[4907]: I0313 14:09:11.617840 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556840-pgmwj" Mar 13 14:09:11 crc kubenswrapper[4907]: I0313 14:09:11.667933 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:11 crc kubenswrapper[4907]: I0313 14:09:11.668153 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ec3466c2-4954-459d-88d8-e8c6cbf309e8-utilities\") pod \"redhat-operators-fmzmc\" (UID: \"ec3466c2-4954-459d-88d8-e8c6cbf309e8\") " pod="openshift-marketplace/redhat-operators-fmzmc" Mar 13 14:09:11 crc kubenswrapper[4907]: I0313 14:09:11.668217 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ec3466c2-4954-459d-88d8-e8c6cbf309e8-catalog-content\") pod \"redhat-operators-fmzmc\" (UID: \"ec3466c2-4954-459d-88d8-e8c6cbf309e8\") " pod="openshift-marketplace/redhat-operators-fmzmc" Mar 13 14:09:11 crc kubenswrapper[4907]: I0313 14:09:11.668282 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-48ddb\" (UniqueName: \"kubernetes.io/projected/ec3466c2-4954-459d-88d8-e8c6cbf309e8-kube-api-access-48ddb\") pod \"redhat-operators-fmzmc\" (UID: \"ec3466c2-4954-459d-88d8-e8c6cbf309e8\") " pod="openshift-marketplace/redhat-operators-fmzmc" Mar 13 14:09:11 crc kubenswrapper[4907]: E0313 14:09:11.668682 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:12.168666841 +0000 UTC m=+251.068454530 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:11 crc kubenswrapper[4907]: I0313 14:09:11.669618 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ec3466c2-4954-459d-88d8-e8c6cbf309e8-utilities\") pod \"redhat-operators-fmzmc\" (UID: \"ec3466c2-4954-459d-88d8-e8c6cbf309e8\") " pod="openshift-marketplace/redhat-operators-fmzmc" Mar 13 14:09:11 crc kubenswrapper[4907]: I0313 14:09:11.669848 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ec3466c2-4954-459d-88d8-e8c6cbf309e8-catalog-content\") pod \"redhat-operators-fmzmc\" (UID: \"ec3466c2-4954-459d-88d8-e8c6cbf309e8\") " pod="openshift-marketplace/redhat-operators-fmzmc" Mar 13 14:09:11 crc kubenswrapper[4907]: I0313 14:09:11.726411 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-48ddb\" (UniqueName: \"kubernetes.io/projected/ec3466c2-4954-459d-88d8-e8c6cbf309e8-kube-api-access-48ddb\") pod \"redhat-operators-fmzmc\" (UID: \"ec3466c2-4954-459d-88d8-e8c6cbf309e8\") " pod="openshift-marketplace/redhat-operators-fmzmc" Mar 13 14:09:11 crc kubenswrapper[4907]: I0313 14:09:11.737343 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jk5nw"] Mar 13 14:09:11 crc kubenswrapper[4907]: I0313 14:09:11.770995 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h9w5d\" (UniqueName: \"kubernetes.io/projected/19edf37b-a1d3-4c7c-945e-5d0b8bb601bf-kube-api-access-h9w5d\") pod \"19edf37b-a1d3-4c7c-945e-5d0b8bb601bf\" (UID: \"19edf37b-a1d3-4c7c-945e-5d0b8bb601bf\") " Mar 13 14:09:11 crc kubenswrapper[4907]: I0313 14:09:11.771064 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/19edf37b-a1d3-4c7c-945e-5d0b8bb601bf-secret-volume\") pod \"19edf37b-a1d3-4c7c-945e-5d0b8bb601bf\" (UID: \"19edf37b-a1d3-4c7c-945e-5d0b8bb601bf\") " Mar 13 14:09:11 crc kubenswrapper[4907]: I0313 14:09:11.771392 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/19edf37b-a1d3-4c7c-945e-5d0b8bb601bf-config-volume\") pod \"19edf37b-a1d3-4c7c-945e-5d0b8bb601bf\" (UID: \"19edf37b-a1d3-4c7c-945e-5d0b8bb601bf\") " Mar 13 14:09:11 crc kubenswrapper[4907]: I0313 14:09:11.771598 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:11 crc kubenswrapper[4907]: E0313 14:09:11.772068 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:12.272052042 +0000 UTC m=+251.171839721 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:11 crc kubenswrapper[4907]: I0313 14:09:11.777313 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/19edf37b-a1d3-4c7c-945e-5d0b8bb601bf-config-volume" (OuterVolumeSpecName: "config-volume") pod "19edf37b-a1d3-4c7c-945e-5d0b8bb601bf" (UID: "19edf37b-a1d3-4c7c-945e-5d0b8bb601bf"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:09:11 crc kubenswrapper[4907]: I0313 14:09:11.782507 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19edf37b-a1d3-4c7c-945e-5d0b8bb601bf-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "19edf37b-a1d3-4c7c-945e-5d0b8bb601bf" (UID: "19edf37b-a1d3-4c7c-945e-5d0b8bb601bf"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:09:11 crc kubenswrapper[4907]: I0313 14:09:11.783925 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/19edf37b-a1d3-4c7c-945e-5d0b8bb601bf-kube-api-access-h9w5d" (OuterVolumeSpecName: "kube-api-access-h9w5d") pod "19edf37b-a1d3-4c7c-945e-5d0b8bb601bf" (UID: "19edf37b-a1d3-4c7c-945e-5d0b8bb601bf"). InnerVolumeSpecName "kube-api-access-h9w5d". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:09:11 crc kubenswrapper[4907]: I0313 14:09:11.792330 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fmzmc" Mar 13 14:09:11 crc kubenswrapper[4907]: I0313 14:09:11.806604 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-p257z"] Mar 13 14:09:11 crc kubenswrapper[4907]: I0313 14:09:11.876655 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:11 crc kubenswrapper[4907]: I0313 14:09:11.877158 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h9w5d\" (UniqueName: \"kubernetes.io/projected/19edf37b-a1d3-4c7c-945e-5d0b8bb601bf-kube-api-access-h9w5d\") on node \"crc\" DevicePath \"\"" Mar 13 14:09:11 crc kubenswrapper[4907]: I0313 14:09:11.877180 4907 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/19edf37b-a1d3-4c7c-945e-5d0b8bb601bf-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 13 14:09:11 crc kubenswrapper[4907]: I0313 14:09:11.877192 4907 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/19edf37b-a1d3-4c7c-945e-5d0b8bb601bf-config-volume\") on node \"crc\" DevicePath \"\"" Mar 13 14:09:11 crc kubenswrapper[4907]: E0313 14:09:11.877494 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:12.377472888 +0000 UTC m=+251.277260577 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:12 crc kubenswrapper[4907]: I0313 14:09:12.101797 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:12 crc kubenswrapper[4907]: E0313 14:09:12.102435 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:12.602413973 +0000 UTC m=+251.502201672 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:12 crc kubenswrapper[4907]: I0313 14:09:12.105691 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jk5nw" event={"ID":"d342a6a2-43dd-4d15-8aff-f60f9a3383bf","Type":"ContainerStarted","Data":"4fd25f47bfaf1894861177f77553de05bdcc490dd081b002f769378e51f3d5fb"} Mar 13 14:09:12 crc kubenswrapper[4907]: I0313 14:09:12.108449 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-vx5qd"] Mar 13 14:09:12 crc kubenswrapper[4907]: E0313 14:09:12.108790 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19edf37b-a1d3-4c7c-945e-5d0b8bb601bf" containerName="collect-profiles" Mar 13 14:09:12 crc kubenswrapper[4907]: I0313 14:09:12.108813 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="19edf37b-a1d3-4c7c-945e-5d0b8bb601bf" containerName="collect-profiles" Mar 13 14:09:12 crc kubenswrapper[4907]: I0313 14:09:12.108990 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="19edf37b-a1d3-4c7c-945e-5d0b8bb601bf" containerName="collect-profiles" Mar 13 14:09:12 crc kubenswrapper[4907]: I0313 14:09:12.109973 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vx5qd" Mar 13 14:09:12 crc kubenswrapper[4907]: I0313 14:09:12.122843 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vx5qd"] Mar 13 14:09:12 crc kubenswrapper[4907]: I0313 14:09:12.147708 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556840-pgmwj" event={"ID":"19edf37b-a1d3-4c7c-945e-5d0b8bb601bf","Type":"ContainerDied","Data":"e8a5550c81aa8c4bb6491a89a3041cb9a6e5701c9a00e500620d96a20f452a36"} Mar 13 14:09:12 crc kubenswrapper[4907]: I0313 14:09:12.147762 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e8a5550c81aa8c4bb6491a89a3041cb9a6e5701c9a00e500620d96a20f452a36" Mar 13 14:09:12 crc kubenswrapper[4907]: I0313 14:09:12.147844 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556840-pgmwj" Mar 13 14:09:12 crc kubenswrapper[4907]: I0313 14:09:12.150804 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-tzmvc" event={"ID":"f797b1bb-e769-49c4-84ce-93a6fad2df93","Type":"ContainerStarted","Data":"04f494188fd5ff2ecf6aeca648a652d5db8d54525a4fd3526e002308da68d7d6"} Mar 13 14:09:12 crc kubenswrapper[4907]: I0313 14:09:12.153155 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p257z" event={"ID":"775414ea-5b94-4ab6-8b7c-fe7c80671ced","Type":"ContainerStarted","Data":"f1572a3e7d8d63236fe0e9ac91306d55a1b117717a36d7026cf70c18e50c6571"} Mar 13 14:09:12 crc kubenswrapper[4907]: I0313 14:09:12.205049 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:12 crc kubenswrapper[4907]: I0313 14:09:12.205594 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mcf55\" (UniqueName: \"kubernetes.io/projected/8779a406-c08e-40db-b25d-8c45e784e94d-kube-api-access-mcf55\") pod \"redhat-operators-vx5qd\" (UID: \"8779a406-c08e-40db-b25d-8c45e784e94d\") " pod="openshift-marketplace/redhat-operators-vx5qd" Mar 13 14:09:12 crc kubenswrapper[4907]: I0313 14:09:12.205759 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8779a406-c08e-40db-b25d-8c45e784e94d-catalog-content\") pod \"redhat-operators-vx5qd\" (UID: \"8779a406-c08e-40db-b25d-8c45e784e94d\") " pod="openshift-marketplace/redhat-operators-vx5qd" Mar 13 14:09:12 crc kubenswrapper[4907]: I0313 14:09:12.206054 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8779a406-c08e-40db-b25d-8c45e784e94d-utilities\") pod \"redhat-operators-vx5qd\" (UID: \"8779a406-c08e-40db-b25d-8c45e784e94d\") " pod="openshift-marketplace/redhat-operators-vx5qd" Mar 13 14:09:12 crc kubenswrapper[4907]: E0313 14:09:12.209924 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:12.709871025 +0000 UTC m=+251.609658934 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:12 crc kubenswrapper[4907]: I0313 14:09:12.241981 4907 ???:1] "http: TLS handshake error from 192.168.126.11:55084: no serving certificate available for the kubelet" Mar 13 14:09:12 crc kubenswrapper[4907]: I0313 14:09:12.307130 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:12 crc kubenswrapper[4907]: I0313 14:09:12.307181 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8779a406-c08e-40db-b25d-8c45e784e94d-utilities\") pod \"redhat-operators-vx5qd\" (UID: \"8779a406-c08e-40db-b25d-8c45e784e94d\") " pod="openshift-marketplace/redhat-operators-vx5qd" Mar 13 14:09:12 crc kubenswrapper[4907]: I0313 14:09:12.307246 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mcf55\" (UniqueName: \"kubernetes.io/projected/8779a406-c08e-40db-b25d-8c45e784e94d-kube-api-access-mcf55\") pod \"redhat-operators-vx5qd\" (UID: \"8779a406-c08e-40db-b25d-8c45e784e94d\") " pod="openshift-marketplace/redhat-operators-vx5qd" Mar 13 14:09:12 crc kubenswrapper[4907]: I0313 14:09:12.307282 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8779a406-c08e-40db-b25d-8c45e784e94d-catalog-content\") pod \"redhat-operators-vx5qd\" (UID: \"8779a406-c08e-40db-b25d-8c45e784e94d\") " pod="openshift-marketplace/redhat-operators-vx5qd" Mar 13 14:09:12 crc kubenswrapper[4907]: E0313 14:09:12.307607 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:12.807587871 +0000 UTC m=+251.707375560 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:12 crc kubenswrapper[4907]: I0313 14:09:12.307915 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8779a406-c08e-40db-b25d-8c45e784e94d-catalog-content\") pod \"redhat-operators-vx5qd\" (UID: \"8779a406-c08e-40db-b25d-8c45e784e94d\") " pod="openshift-marketplace/redhat-operators-vx5qd" Mar 13 14:09:12 crc kubenswrapper[4907]: I0313 14:09:12.308086 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8779a406-c08e-40db-b25d-8c45e784e94d-utilities\") pod \"redhat-operators-vx5qd\" (UID: \"8779a406-c08e-40db-b25d-8c45e784e94d\") " pod="openshift-marketplace/redhat-operators-vx5qd" Mar 13 14:09:12 crc kubenswrapper[4907]: I0313 14:09:12.393257 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mcf55\" (UniqueName: \"kubernetes.io/projected/8779a406-c08e-40db-b25d-8c45e784e94d-kube-api-access-mcf55\") pod \"redhat-operators-vx5qd\" (UID: \"8779a406-c08e-40db-b25d-8c45e784e94d\") " pod="openshift-marketplace/redhat-operators-vx5qd" Mar 13 14:09:15 crc kubenswrapper[4907]: I0313 14:09:14.998741 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:15 crc kubenswrapper[4907]: E0313 14:09:14.999310 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:15.999289845 +0000 UTC m=+254.899077524 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:15 crc kubenswrapper[4907]: I0313 14:09:15.014419 4907 patch_prober.go:28] interesting pod/console-operator-58897d9998-w76kw container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.13:8443/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 14:09:15 crc kubenswrapper[4907]: I0313 14:09:15.014811 4907 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-w76kw" podUID="9b75954c-b75d-4f17-82dc-4c0358fd8d0f" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.13:8443/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 14:09:15 crc kubenswrapper[4907]: I0313 14:09:15.034706 4907 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-wgpdd container/openshift-config-operator namespace/openshift-config-operator: Liveness probe status=failure output="Get \"https://10.217.0.27:8443/healthz\": dial tcp 10.217.0.27:8443: connect: connection refused" start-of-body= Mar 13 14:09:15 crc kubenswrapper[4907]: I0313 14:09:15.034830 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-wgpdd" podUID="0149e083-ed5b-4c7f-94e7-4f3f2ff8a557" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.27:8443/healthz\": dial tcp 10.217.0.27:8443: connect: connection refused" Mar 13 14:09:15 crc kubenswrapper[4907]: I0313 14:09:15.034998 4907 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-wgpdd container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.27:8443/healthz\": dial tcp 10.217.0.27:8443: connect: connection refused" start-of-body= Mar 13 14:09:15 crc kubenswrapper[4907]: I0313 14:09:15.035021 4907 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-wgpdd" podUID="0149e083-ed5b-4c7f-94e7-4f3f2ff8a557" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.27:8443/healthz\": dial tcp 10.217.0.27:8443: connect: connection refused" Mar 13 14:09:15 crc kubenswrapper[4907]: I0313 14:09:15.036011 4907 patch_prober.go:28] interesting pod/router-default-5444994796-f8mmz container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 13 14:09:15 crc kubenswrapper[4907]: [-]has-synced failed: reason withheld Mar 13 14:09:15 crc kubenswrapper[4907]: [+]process-running ok Mar 13 14:09:15 crc kubenswrapper[4907]: healthz check failed Mar 13 14:09:15 crc kubenswrapper[4907]: I0313 14:09:15.036067 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-f8mmz" podUID="c26ccc9f-e682-4b94-9539-cb2607725f71" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 14:09:15 crc kubenswrapper[4907]: I0313 14:09:15.073196 4907 patch_prober.go:28] interesting pod/apiserver-76f77b778f-r86r4 container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="Get \"https://10.217.0.9:8443/livez\": dial tcp 10.217.0.9:8443: connect: connection refused" start-of-body= Mar 13 14:09:15 crc kubenswrapper[4907]: I0313 14:09:15.073280 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-r86r4" podUID="5a9c889b-5954-4d3a-ae72-12000ee1cc25" containerName="openshift-apiserver" probeResult="failure" output="Get \"https://10.217.0.9:8443/livez\": dial tcp 10.217.0.9:8443: connect: connection refused" Mar 13 14:09:15 crc kubenswrapper[4907]: I0313 14:09:15.093415 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-r86r4" Mar 13 14:09:15 crc kubenswrapper[4907]: I0313 14:09:15.093454 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-r86r4" Mar 13 14:09:15 crc kubenswrapper[4907]: I0313 14:09:15.093482 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hkbqt" Mar 13 14:09:15 crc kubenswrapper[4907]: I0313 14:09:15.097610 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-5stnc" Mar 13 14:09:15 crc kubenswrapper[4907]: I0313 14:09:15.100353 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-tzmvc" podStartSLOduration=194.100326813 podStartE2EDuration="3m14.100326813s" podCreationTimestamp="2026-03-13 14:06:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:09:12.173729723 +0000 UTC m=+251.073517412" watchObservedRunningTime="2026-03-13 14:09:15.100326813 +0000 UTC m=+254.000114502" Mar 13 14:09:15 crc kubenswrapper[4907]: I0313 14:09:15.101495 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:15 crc kubenswrapper[4907]: E0313 14:09:15.102050 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:15.60203051 +0000 UTC m=+254.501818199 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:15 crc kubenswrapper[4907]: I0313 14:09:15.107496 4907 patch_prober.go:28] interesting pod/downloads-7954f5f757-544df container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Mar 13 14:09:15 crc kubenswrapper[4907]: I0313 14:09:15.107564 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-544df" podUID="f8ea4f24-05b0-4661-a617-a4208f9a9188" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Mar 13 14:09:15 crc kubenswrapper[4907]: I0313 14:09:15.112583 4907 patch_prober.go:28] interesting pod/downloads-7954f5f757-544df container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Mar 13 14:09:15 crc kubenswrapper[4907]: I0313 14:09:15.112626 4907 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-544df" podUID="f8ea4f24-05b0-4661-a617-a4208f9a9188" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Mar 13 14:09:15 crc kubenswrapper[4907]: I0313 14:09:15.126467 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Mar 13 14:09:15 crc kubenswrapper[4907]: I0313 14:09:15.127571 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 13 14:09:15 crc kubenswrapper[4907]: I0313 14:09:15.168680 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Mar 13 14:09:15 crc kubenswrapper[4907]: I0313 14:09:15.171404 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Mar 13 14:09:15 crc kubenswrapper[4907]: I0313 14:09:15.179117 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Mar 13 14:09:15 crc kubenswrapper[4907]: I0313 14:09:15.204342 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/909a1142-83a3-43c2-9032-45f79a411200-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"909a1142-83a3-43c2-9032-45f79a411200\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 13 14:09:15 crc kubenswrapper[4907]: I0313 14:09:15.204491 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/909a1142-83a3-43c2-9032-45f79a411200-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"909a1142-83a3-43c2-9032-45f79a411200\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 13 14:09:15 crc kubenswrapper[4907]: I0313 14:09:15.204560 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:15 crc kubenswrapper[4907]: E0313 14:09:15.280941 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:15.780922636 +0000 UTC m=+254.680710325 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:15 crc kubenswrapper[4907]: I0313 14:09:15.283791 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-l9xpb" Mar 13 14:09:15 crc kubenswrapper[4907]: I0313 14:09:15.284024 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-l9xpb" Mar 13 14:09:15 crc kubenswrapper[4907]: I0313 14:09:15.300543 4907 ???:1] "http: TLS handshake error from 192.168.126.11:55094: no serving certificate available for the kubelet" Mar 13 14:09:15 crc kubenswrapper[4907]: E0313 14:09:15.309135 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:15.809086192 +0000 UTC m=+254.708873881 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:15 crc kubenswrapper[4907]: I0313 14:09:15.308671 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:15 crc kubenswrapper[4907]: I0313 14:09:15.332509 4907 patch_prober.go:28] interesting pod/console-f9d7485db-l9xpb container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.21:8443/health\": dial tcp 10.217.0.21:8443: connect: connection refused" start-of-body= Mar 13 14:09:15 crc kubenswrapper[4907]: I0313 14:09:15.334193 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-l9xpb" podUID="4c8c0266-5452-4abb-92c5-c536df94dd41" containerName="console" probeResult="failure" output="Get \"https://10.217.0.21:8443/health\": dial tcp 10.217.0.21:8443: connect: connection refused" Mar 13 14:09:15 crc kubenswrapper[4907]: I0313 14:09:15.362534 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/909a1142-83a3-43c2-9032-45f79a411200-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"909a1142-83a3-43c2-9032-45f79a411200\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 13 14:09:15 crc kubenswrapper[4907]: I0313 14:09:15.362982 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:15 crc kubenswrapper[4907]: I0313 14:09:15.363384 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/909a1142-83a3-43c2-9032-45f79a411200-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"909a1142-83a3-43c2-9032-45f79a411200\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 13 14:09:15 crc kubenswrapper[4907]: E0313 14:09:15.364466 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:15.864437797 +0000 UTC m=+254.764225486 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:15 crc kubenswrapper[4907]: I0313 14:09:15.419746 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/909a1142-83a3-43c2-9032-45f79a411200-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"909a1142-83a3-43c2-9032-45f79a411200\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 13 14:09:15 crc kubenswrapper[4907]: I0313 14:09:15.421568 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vx5qd" Mar 13 14:09:15 crc kubenswrapper[4907]: I0313 14:09:15.466418 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-f8mmz" Mar 13 14:09:15 crc kubenswrapper[4907]: I0313 14:09:15.470485 4907 patch_prober.go:28] interesting pod/router-default-5444994796-f8mmz container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 13 14:09:15 crc kubenswrapper[4907]: [-]has-synced failed: reason withheld Mar 13 14:09:15 crc kubenswrapper[4907]: [+]process-running ok Mar 13 14:09:15 crc kubenswrapper[4907]: healthz check failed Mar 13 14:09:15 crc kubenswrapper[4907]: I0313 14:09:15.470528 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-f8mmz" podUID="c26ccc9f-e682-4b94-9539-cb2607725f71" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 14:09:15 crc kubenswrapper[4907]: I0313 14:09:15.481623 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:15 crc kubenswrapper[4907]: E0313 14:09:15.485186 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:15.985147979 +0000 UTC m=+254.884935668 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:15 crc kubenswrapper[4907]: I0313 14:09:15.553779 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/909a1142-83a3-43c2-9032-45f79a411200-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"909a1142-83a3-43c2-9032-45f79a411200\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 13 14:09:15 crc kubenswrapper[4907]: I0313 14:09:15.574214 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-fmzmc"] Mar 13 14:09:15 crc kubenswrapper[4907]: I0313 14:09:15.581409 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-kwfhg"] Mar 13 14:09:15 crc kubenswrapper[4907]: I0313 14:09:15.581727 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-kwfhg" podUID="814c6dec-6e96-4068-9aec-9e57fefca5db" containerName="controller-manager" containerID="cri-o://11d51b3d36d9283f41d21738ca7a244487cc24be4453cd47b19366b6bff8bcf1" gracePeriod=30 Mar 13 14:09:15 crc kubenswrapper[4907]: I0313 14:09:15.583028 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:15 crc kubenswrapper[4907]: E0313 14:09:15.584115 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:16.084099178 +0000 UTC m=+254.983886867 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:15 crc kubenswrapper[4907]: W0313 14:09:15.596451 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podec3466c2_4954_459d_88d8_e8c6cbf309e8.slice/crio-33655dcb12d3167aaf6db2e72d760da63675955f0f25e1e18f759d3dc24a92c2 WatchSource:0}: Error finding container 33655dcb12d3167aaf6db2e72d760da63675955f0f25e1e18f759d3dc24a92c2: Status 404 returned error can't find the container with id 33655dcb12d3167aaf6db2e72d760da63675955f0f25e1e18f759d3dc24a92c2 Mar 13 14:09:15 crc kubenswrapper[4907]: I0313 14:09:15.655367 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-4h9q7"] Mar 13 14:09:15 crc kubenswrapper[4907]: I0313 14:09:15.655635 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4h9q7" podUID="9bb1952d-0e55-4e45-81c0-69a8cddb827e" containerName="route-controller-manager" containerID="cri-o://9435fd54ef37b6b3b6843aafb8ac4987dfdaac5f9146401bd984cd364cd498e2" gracePeriod=30 Mar 13 14:09:15 crc kubenswrapper[4907]: E0313 14:09:15.684999 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:16.18497153 +0000 UTC m=+255.084759219 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:15 crc kubenswrapper[4907]: I0313 14:09:15.685040 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:15 crc kubenswrapper[4907]: I0313 14:09:15.686037 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:15 crc kubenswrapper[4907]: E0313 14:09:15.686459 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:16.186449312 +0000 UTC m=+255.086236991 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:15 crc kubenswrapper[4907]: I0313 14:09:15.720639 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 13 14:09:15 crc kubenswrapper[4907]: I0313 14:09:15.793225 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:15 crc kubenswrapper[4907]: E0313 14:09:15.793722 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:16.293690396 +0000 UTC m=+255.193478085 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:15 crc kubenswrapper[4907]: I0313 14:09:15.895415 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:15 crc kubenswrapper[4907]: E0313 14:09:15.896114 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:16.396101751 +0000 UTC m=+255.295889440 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:15 crc kubenswrapper[4907]: I0313 14:09:15.943486 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ksgr6" Mar 13 14:09:15 crc kubenswrapper[4907]: I0313 14:09:15.954791 4907 patch_prober.go:28] interesting pod/apiserver-7bbb656c7d-ksgr6 container/oauth-apiserver namespace/openshift-oauth-apiserver: Readiness probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Mar 13 14:09:15 crc kubenswrapper[4907]: [+]log ok Mar 13 14:09:15 crc kubenswrapper[4907]: [+]etcd ok Mar 13 14:09:15 crc kubenswrapper[4907]: [+]etcd-readiness ok Mar 13 14:09:15 crc kubenswrapper[4907]: [+]poststarthook/start-apiserver-admission-initializer ok Mar 13 14:09:15 crc kubenswrapper[4907]: [-]informer-sync failed: reason withheld Mar 13 14:09:15 crc kubenswrapper[4907]: [+]poststarthook/generic-apiserver-start-informers ok Mar 13 14:09:15 crc kubenswrapper[4907]: [+]poststarthook/max-in-flight-filter ok Mar 13 14:09:15 crc kubenswrapper[4907]: [+]poststarthook/storage-object-count-tracker-hook ok Mar 13 14:09:15 crc kubenswrapper[4907]: [+]poststarthook/openshift.io-StartUserInformer ok Mar 13 14:09:15 crc kubenswrapper[4907]: [+]poststarthook/openshift.io-StartOAuthInformer ok Mar 13 14:09:15 crc kubenswrapper[4907]: [+]poststarthook/openshift.io-StartTokenTimeoutUpdater ok Mar 13 14:09:15 crc kubenswrapper[4907]: [+]shutdown ok Mar 13 14:09:15 crc kubenswrapper[4907]: readyz check failed Mar 13 14:09:15 crc kubenswrapper[4907]: I0313 14:09:15.954873 4907 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ksgr6" podUID="1be912b5-9d1d-48bd-a0ab-7f601f9eb9da" containerName="oauth-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.000451 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:16 crc kubenswrapper[4907]: E0313 14:09:16.000648 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:16.500616202 +0000 UTC m=+255.400403911 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.001083 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:16 crc kubenswrapper[4907]: E0313 14:09:16.001435 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:16.501423314 +0000 UTC m=+255.401210993 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.026570 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-w76kw" Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.028183 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vx5qd"] Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.101620 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:16 crc kubenswrapper[4907]: E0313 14:09:16.102614 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:16.602582345 +0000 UTC m=+255.502370034 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.124380 4907 generic.go:334] "Generic (PLEG): container finished" podID="9bb1952d-0e55-4e45-81c0-69a8cddb827e" containerID="9435fd54ef37b6b3b6843aafb8ac4987dfdaac5f9146401bd984cd364cd498e2" exitCode=0 Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.124468 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4h9q7" event={"ID":"9bb1952d-0e55-4e45-81c0-69a8cddb827e","Type":"ContainerDied","Data":"9435fd54ef37b6b3b6843aafb8ac4987dfdaac5f9146401bd984cd364cd498e2"} Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.135192 4907 generic.go:334] "Generic (PLEG): container finished" podID="814c6dec-6e96-4068-9aec-9e57fefca5db" containerID="11d51b3d36d9283f41d21738ca7a244487cc24be4453cd47b19366b6bff8bcf1" exitCode=0 Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.135307 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-kwfhg" event={"ID":"814c6dec-6e96-4068-9aec-9e57fefca5db","Type":"ContainerDied","Data":"11d51b3d36d9283f41d21738ca7a244487cc24be4453cd47b19366b6bff8bcf1"} Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.138941 4907 generic.go:334] "Generic (PLEG): container finished" podID="775414ea-5b94-4ab6-8b7c-fe7c80671ced" containerID="2f158ff96c417d8c3693d515dfe179799bf07e75ac8f792f12324fd7a18b5c50" exitCode=0 Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.138995 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p257z" event={"ID":"775414ea-5b94-4ab6-8b7c-fe7c80671ced","Type":"ContainerDied","Data":"2f158ff96c417d8c3693d515dfe179799bf07e75ac8f792f12324fd7a18b5c50"} Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.145461 4907 generic.go:334] "Generic (PLEG): container finished" podID="d342a6a2-43dd-4d15-8aff-f60f9a3383bf" containerID="167f39bf875dea575d6960ea51e3427665c28d2ee610e574bf41709836edc603" exitCode=0 Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.145729 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jk5nw" event={"ID":"d342a6a2-43dd-4d15-8aff-f60f9a3383bf","Type":"ContainerDied","Data":"167f39bf875dea575d6960ea51e3427665c28d2ee610e574bf41709836edc603"} Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.165761 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fmzmc" event={"ID":"ec3466c2-4954-459d-88d8-e8c6cbf309e8","Type":"ContainerStarted","Data":"5129da37c4baac30c1ed93e43558cb65af057e72717f5cffd7b1409299226e40"} Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.165834 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fmzmc" event={"ID":"ec3466c2-4954-459d-88d8-e8c6cbf309e8","Type":"ContainerStarted","Data":"33655dcb12d3167aaf6db2e72d760da63675955f0f25e1e18f759d3dc24a92c2"} Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.204320 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:16 crc kubenswrapper[4907]: E0313 14:09:16.204727 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:16.704707831 +0000 UTC m=+255.604495520 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.263432 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-kwfhg" Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.295402 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.319837 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:16 crc kubenswrapper[4907]: E0313 14:09:16.320082 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:16.820034416 +0000 UTC m=+255.719822115 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.321024 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:16 crc kubenswrapper[4907]: E0313 14:09:16.336405 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:16.836375771 +0000 UTC m=+255.736163460 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.338722 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Mar 13 14:09:16 crc kubenswrapper[4907]: E0313 14:09:16.351577 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="814c6dec-6e96-4068-9aec-9e57fefca5db" containerName="controller-manager" Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.351939 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="814c6dec-6e96-4068-9aec-9e57fefca5db" containerName="controller-manager" Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.352647 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="814c6dec-6e96-4068-9aec-9e57fefca5db" containerName="controller-manager" Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.353799 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.357176 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.360421 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.362673 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.422232 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/814c6dec-6e96-4068-9aec-9e57fefca5db-config\") pod \"814c6dec-6e96-4068-9aec-9e57fefca5db\" (UID: \"814c6dec-6e96-4068-9aec-9e57fefca5db\") " Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.422310 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/814c6dec-6e96-4068-9aec-9e57fefca5db-proxy-ca-bundles\") pod \"814c6dec-6e96-4068-9aec-9e57fefca5db\" (UID: \"814c6dec-6e96-4068-9aec-9e57fefca5db\") " Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.422345 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/814c6dec-6e96-4068-9aec-9e57fefca5db-client-ca\") pod \"814c6dec-6e96-4068-9aec-9e57fefca5db\" (UID: \"814c6dec-6e96-4068-9aec-9e57fefca5db\") " Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.422396 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/814c6dec-6e96-4068-9aec-9e57fefca5db-serving-cert\") pod \"814c6dec-6e96-4068-9aec-9e57fefca5db\" (UID: \"814c6dec-6e96-4068-9aec-9e57fefca5db\") " Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.422474 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gn5jt\" (UniqueName: \"kubernetes.io/projected/814c6dec-6e96-4068-9aec-9e57fefca5db-kube-api-access-gn5jt\") pod \"814c6dec-6e96-4068-9aec-9e57fefca5db\" (UID: \"814c6dec-6e96-4068-9aec-9e57fefca5db\") " Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.422580 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:16 crc kubenswrapper[4907]: E0313 14:09:16.422856 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:16.922840291 +0000 UTC m=+255.822627980 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.423900 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/814c6dec-6e96-4068-9aec-9e57fefca5db-config" (OuterVolumeSpecName: "config") pod "814c6dec-6e96-4068-9aec-9e57fefca5db" (UID: "814c6dec-6e96-4068-9aec-9e57fefca5db"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.424428 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/814c6dec-6e96-4068-9aec-9e57fefca5db-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "814c6dec-6e96-4068-9aec-9e57fefca5db" (UID: "814c6dec-6e96-4068-9aec-9e57fefca5db"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.424783 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/814c6dec-6e96-4068-9aec-9e57fefca5db-client-ca" (OuterVolumeSpecName: "client-ca") pod "814c6dec-6e96-4068-9aec-9e57fefca5db" (UID: "814c6dec-6e96-4068-9aec-9e57fefca5db"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.431409 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/814c6dec-6e96-4068-9aec-9e57fefca5db-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "814c6dec-6e96-4068-9aec-9e57fefca5db" (UID: "814c6dec-6e96-4068-9aec-9e57fefca5db"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.434738 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/814c6dec-6e96-4068-9aec-9e57fefca5db-kube-api-access-gn5jt" (OuterVolumeSpecName: "kube-api-access-gn5jt") pod "814c6dec-6e96-4068-9aec-9e57fefca5db" (UID: "814c6dec-6e96-4068-9aec-9e57fefca5db"). InnerVolumeSpecName "kube-api-access-gn5jt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.465164 4907 patch_prober.go:28] interesting pod/router-default-5444994796-f8mmz container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 13 14:09:16 crc kubenswrapper[4907]: [-]has-synced failed: reason withheld Mar 13 14:09:16 crc kubenswrapper[4907]: [+]process-running ok Mar 13 14:09:16 crc kubenswrapper[4907]: healthz check failed Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.465247 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-f8mmz" podUID="c26ccc9f-e682-4b94-9539-cb2607725f71" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.524350 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e0ad06d9-5751-43e2-bf40-b037e61934f5-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"e0ad06d9-5751-43e2-bf40-b037e61934f5\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.524447 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.524473 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e0ad06d9-5751-43e2-bf40-b037e61934f5-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"e0ad06d9-5751-43e2-bf40-b037e61934f5\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.524589 4907 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/814c6dec-6e96-4068-9aec-9e57fefca5db-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.524601 4907 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/814c6dec-6e96-4068-9aec-9e57fefca5db-client-ca\") on node \"crc\" DevicePath \"\"" Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.524610 4907 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/814c6dec-6e96-4068-9aec-9e57fefca5db-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.524619 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gn5jt\" (UniqueName: \"kubernetes.io/projected/814c6dec-6e96-4068-9aec-9e57fefca5db-kube-api-access-gn5jt\") on node \"crc\" DevicePath \"\"" Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.524631 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/814c6dec-6e96-4068-9aec-9e57fefca5db-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:09:16 crc kubenswrapper[4907]: E0313 14:09:16.525285 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:17.025266846 +0000 UTC m=+255.925054535 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.625636 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.626099 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e0ad06d9-5751-43e2-bf40-b037e61934f5-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"e0ad06d9-5751-43e2-bf40-b037e61934f5\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 13 14:09:16 crc kubenswrapper[4907]: E0313 14:09:16.626135 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:17.126097767 +0000 UTC m=+256.025885446 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.626372 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.626411 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e0ad06d9-5751-43e2-bf40-b037e61934f5-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"e0ad06d9-5751-43e2-bf40-b037e61934f5\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.626709 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e0ad06d9-5751-43e2-bf40-b037e61934f5-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"e0ad06d9-5751-43e2-bf40-b037e61934f5\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 13 14:09:16 crc kubenswrapper[4907]: E0313 14:09:16.626801 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:17.126788696 +0000 UTC m=+256.026576385 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.642777 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4h9q7" Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.653423 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e0ad06d9-5751-43e2-bf40-b037e61934f5-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"e0ad06d9-5751-43e2-bf40-b037e61934f5\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.700767 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.729576 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gqtv7\" (UniqueName: \"kubernetes.io/projected/9bb1952d-0e55-4e45-81c0-69a8cddb827e-kube-api-access-gqtv7\") pod \"9bb1952d-0e55-4e45-81c0-69a8cddb827e\" (UID: \"9bb1952d-0e55-4e45-81c0-69a8cddb827e\") " Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.729819 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.729852 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9bb1952d-0e55-4e45-81c0-69a8cddb827e-config\") pod \"9bb1952d-0e55-4e45-81c0-69a8cddb827e\" (UID: \"9bb1952d-0e55-4e45-81c0-69a8cddb827e\") " Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.729978 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9bb1952d-0e55-4e45-81c0-69a8cddb827e-serving-cert\") pod \"9bb1952d-0e55-4e45-81c0-69a8cddb827e\" (UID: \"9bb1952d-0e55-4e45-81c0-69a8cddb827e\") " Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.730073 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9bb1952d-0e55-4e45-81c0-69a8cddb827e-client-ca\") pod \"9bb1952d-0e55-4e45-81c0-69a8cddb827e\" (UID: \"9bb1952d-0e55-4e45-81c0-69a8cddb827e\") " Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.731564 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9bb1952d-0e55-4e45-81c0-69a8cddb827e-client-ca" (OuterVolumeSpecName: "client-ca") pod "9bb1952d-0e55-4e45-81c0-69a8cddb827e" (UID: "9bb1952d-0e55-4e45-81c0-69a8cddb827e"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.741236 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9bb1952d-0e55-4e45-81c0-69a8cddb827e-config" (OuterVolumeSpecName: "config") pod "9bb1952d-0e55-4e45-81c0-69a8cddb827e" (UID: "9bb1952d-0e55-4e45-81c0-69a8cddb827e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:09:16 crc kubenswrapper[4907]: E0313 14:09:16.741354 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:17.24133627 +0000 UTC m=+256.141123949 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.745048 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9bb1952d-0e55-4e45-81c0-69a8cddb827e-kube-api-access-gqtv7" (OuterVolumeSpecName: "kube-api-access-gqtv7") pod "9bb1952d-0e55-4e45-81c0-69a8cddb827e" (UID: "9bb1952d-0e55-4e45-81c0-69a8cddb827e"). InnerVolumeSpecName "kube-api-access-gqtv7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.755438 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9bb1952d-0e55-4e45-81c0-69a8cddb827e-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9bb1952d-0e55-4e45-81c0-69a8cddb827e" (UID: "9bb1952d-0e55-4e45-81c0-69a8cddb827e"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.835642 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.835914 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gqtv7\" (UniqueName: \"kubernetes.io/projected/9bb1952d-0e55-4e45-81c0-69a8cddb827e-kube-api-access-gqtv7\") on node \"crc\" DevicePath \"\"" Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.835929 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9bb1952d-0e55-4e45-81c0-69a8cddb827e-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.835940 4907 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9bb1952d-0e55-4e45-81c0-69a8cddb827e-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.835950 4907 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9bb1952d-0e55-4e45-81c0-69a8cddb827e-client-ca\") on node \"crc\" DevicePath \"\"" Mar 13 14:09:16 crc kubenswrapper[4907]: E0313 14:09:16.836199 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:17.336180279 +0000 UTC m=+256.235967968 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.937340 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:16 crc kubenswrapper[4907]: E0313 14:09:16.937483 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:17.437454362 +0000 UTC m=+256.337242051 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.937702 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:16 crc kubenswrapper[4907]: E0313 14:09:16.938066 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:17.438058518 +0000 UTC m=+256.337846197 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.017704 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-9c7f75cf9-5jvdt"] Mar 13 14:09:17 crc kubenswrapper[4907]: E0313 14:09:17.018056 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9bb1952d-0e55-4e45-81c0-69a8cddb827e" containerName="route-controller-manager" Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.018093 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="9bb1952d-0e55-4e45-81c0-69a8cddb827e" containerName="route-controller-manager" Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.018203 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="9bb1952d-0e55-4e45-81c0-69a8cddb827e" containerName="route-controller-manager" Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.018581 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-9c7f75cf9-5jvdt" Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.027072 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5f8c8c7848-sbqs7"] Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.027769 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5f8c8c7848-sbqs7" Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.030425 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-9c7f75cf9-5jvdt"] Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.045783 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5f8c8c7848-sbqs7"] Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.048599 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:17 crc kubenswrapper[4907]: E0313 14:09:17.059766 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:17.559727026 +0000 UTC m=+256.459514715 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.060709 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:17 crc kubenswrapper[4907]: E0313 14:09:17.061321 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:17.561307178 +0000 UTC m=+256.461095037 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.161793 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:17 crc kubenswrapper[4907]: E0313 14:09:17.162198 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:17.66213719 +0000 UTC m=+256.561924879 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.162400 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.162443 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e3ff9578-cc43-43df-a9f7-7480be932f63-serving-cert\") pod \"controller-manager-9c7f75cf9-5jvdt\" (UID: \"e3ff9578-cc43-43df-a9f7-7480be932f63\") " pod="openshift-controller-manager/controller-manager-9c7f75cf9-5jvdt" Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.162464 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e3ff9578-cc43-43df-a9f7-7480be932f63-proxy-ca-bundles\") pod \"controller-manager-9c7f75cf9-5jvdt\" (UID: \"e3ff9578-cc43-43df-a9f7-7480be932f63\") " pod="openshift-controller-manager/controller-manager-9c7f75cf9-5jvdt" Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.162500 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b17406bc-b62f-4633-a910-cbb294546ed8-serving-cert\") pod \"route-controller-manager-5f8c8c7848-sbqs7\" (UID: \"b17406bc-b62f-4633-a910-cbb294546ed8\") " pod="openshift-route-controller-manager/route-controller-manager-5f8c8c7848-sbqs7" Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.162531 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e3ff9578-cc43-43df-a9f7-7480be932f63-client-ca\") pod \"controller-manager-9c7f75cf9-5jvdt\" (UID: \"e3ff9578-cc43-43df-a9f7-7480be932f63\") " pod="openshift-controller-manager/controller-manager-9c7f75cf9-5jvdt" Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.162773 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q888d\" (UniqueName: \"kubernetes.io/projected/e3ff9578-cc43-43df-a9f7-7480be932f63-kube-api-access-q888d\") pod \"controller-manager-9c7f75cf9-5jvdt\" (UID: \"e3ff9578-cc43-43df-a9f7-7480be932f63\") " pod="openshift-controller-manager/controller-manager-9c7f75cf9-5jvdt" Mar 13 14:09:17 crc kubenswrapper[4907]: E0313 14:09:17.162866 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:17.662846959 +0000 UTC m=+256.562634648 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.162931 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pl7qj\" (UniqueName: \"kubernetes.io/projected/b17406bc-b62f-4633-a910-cbb294546ed8-kube-api-access-pl7qj\") pod \"route-controller-manager-5f8c8c7848-sbqs7\" (UID: \"b17406bc-b62f-4633-a910-cbb294546ed8\") " pod="openshift-route-controller-manager/route-controller-manager-5f8c8c7848-sbqs7" Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.162994 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b17406bc-b62f-4633-a910-cbb294546ed8-config\") pod \"route-controller-manager-5f8c8c7848-sbqs7\" (UID: \"b17406bc-b62f-4633-a910-cbb294546ed8\") " pod="openshift-route-controller-manager/route-controller-manager-5f8c8c7848-sbqs7" Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.163024 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e3ff9578-cc43-43df-a9f7-7480be932f63-config\") pod \"controller-manager-9c7f75cf9-5jvdt\" (UID: \"e3ff9578-cc43-43df-a9f7-7480be932f63\") " pod="openshift-controller-manager/controller-manager-9c7f75cf9-5jvdt" Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.163159 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b17406bc-b62f-4633-a910-cbb294546ed8-client-ca\") pod \"route-controller-manager-5f8c8c7848-sbqs7\" (UID: \"b17406bc-b62f-4633-a910-cbb294546ed8\") " pod="openshift-route-controller-manager/route-controller-manager-5f8c8c7848-sbqs7" Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.189747 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vx5qd" event={"ID":"8779a406-c08e-40db-b25d-8c45e784e94d","Type":"ContainerDied","Data":"8baf18c6d4683811fc8b3cb2c3dfe2ac02213d10d0c7cc70876c15379d595455"} Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.191929 4907 generic.go:334] "Generic (PLEG): container finished" podID="8779a406-c08e-40db-b25d-8c45e784e94d" containerID="8baf18c6d4683811fc8b3cb2c3dfe2ac02213d10d0c7cc70876c15379d595455" exitCode=0 Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.192064 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vx5qd" event={"ID":"8779a406-c08e-40db-b25d-8c45e784e94d","Type":"ContainerStarted","Data":"23bfb83692e34549e2b59eb14b5903640d0e07ca09276303909a34702eedee74"} Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.214991 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4h9q7" Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.215149 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4h9q7" event={"ID":"9bb1952d-0e55-4e45-81c0-69a8cddb827e","Type":"ContainerDied","Data":"9aafc231113ab96a27a77794d70a165636d29d17e2ee245224f530f5fe6fe82a"} Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.215212 4907 scope.go:117] "RemoveContainer" containerID="9435fd54ef37b6b3b6843aafb8ac4987dfdaac5f9146401bd984cd364cd498e2" Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.223412 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"909a1142-83a3-43c2-9032-45f79a411200","Type":"ContainerStarted","Data":"01d238bc9f78f8ddbe9c3e9ff416221f6b3b235da4624c67b2cbebfb3bfea1b2"} Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.228734 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-kwfhg" event={"ID":"814c6dec-6e96-4068-9aec-9e57fefca5db","Type":"ContainerDied","Data":"0f1cf03420439a212080278b489e741330821d26fb8f1062a79437e4aeb8e4c5"} Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.228838 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-kwfhg" Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.257665 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-sjwvn" event={"ID":"4a8597a4-c72f-4023-b2fb-1b1237411a36","Type":"ContainerStarted","Data":"597e9fba55f42bf26e073ef22245fee0726c592fd0c8e15ca812b30dbf8bba2e"} Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.264402 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.264910 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b17406bc-b62f-4633-a910-cbb294546ed8-serving-cert\") pod \"route-controller-manager-5f8c8c7848-sbqs7\" (UID: \"b17406bc-b62f-4633-a910-cbb294546ed8\") " pod="openshift-route-controller-manager/route-controller-manager-5f8c8c7848-sbqs7" Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.264959 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e3ff9578-cc43-43df-a9f7-7480be932f63-client-ca\") pod \"controller-manager-9c7f75cf9-5jvdt\" (UID: \"e3ff9578-cc43-43df-a9f7-7480be932f63\") " pod="openshift-controller-manager/controller-manager-9c7f75cf9-5jvdt" Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.265006 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q888d\" (UniqueName: \"kubernetes.io/projected/e3ff9578-cc43-43df-a9f7-7480be932f63-kube-api-access-q888d\") pod \"controller-manager-9c7f75cf9-5jvdt\" (UID: \"e3ff9578-cc43-43df-a9f7-7480be932f63\") " pod="openshift-controller-manager/controller-manager-9c7f75cf9-5jvdt" Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.265054 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pl7qj\" (UniqueName: \"kubernetes.io/projected/b17406bc-b62f-4633-a910-cbb294546ed8-kube-api-access-pl7qj\") pod \"route-controller-manager-5f8c8c7848-sbqs7\" (UID: \"b17406bc-b62f-4633-a910-cbb294546ed8\") " pod="openshift-route-controller-manager/route-controller-manager-5f8c8c7848-sbqs7" Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.265087 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b17406bc-b62f-4633-a910-cbb294546ed8-config\") pod \"route-controller-manager-5f8c8c7848-sbqs7\" (UID: \"b17406bc-b62f-4633-a910-cbb294546ed8\") " pod="openshift-route-controller-manager/route-controller-manager-5f8c8c7848-sbqs7" Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.265106 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e3ff9578-cc43-43df-a9f7-7480be932f63-config\") pod \"controller-manager-9c7f75cf9-5jvdt\" (UID: \"e3ff9578-cc43-43df-a9f7-7480be932f63\") " pod="openshift-controller-manager/controller-manager-9c7f75cf9-5jvdt" Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.265150 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b17406bc-b62f-4633-a910-cbb294546ed8-client-ca\") pod \"route-controller-manager-5f8c8c7848-sbqs7\" (UID: \"b17406bc-b62f-4633-a910-cbb294546ed8\") " pod="openshift-route-controller-manager/route-controller-manager-5f8c8c7848-sbqs7" Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.265232 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e3ff9578-cc43-43df-a9f7-7480be932f63-serving-cert\") pod \"controller-manager-9c7f75cf9-5jvdt\" (UID: \"e3ff9578-cc43-43df-a9f7-7480be932f63\") " pod="openshift-controller-manager/controller-manager-9c7f75cf9-5jvdt" Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.265256 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e3ff9578-cc43-43df-a9f7-7480be932f63-proxy-ca-bundles\") pod \"controller-manager-9c7f75cf9-5jvdt\" (UID: \"e3ff9578-cc43-43df-a9f7-7480be932f63\") " pod="openshift-controller-manager/controller-manager-9c7f75cf9-5jvdt" Mar 13 14:09:17 crc kubenswrapper[4907]: E0313 14:09:17.265808 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:17.765776147 +0000 UTC m=+256.665563836 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.267725 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e3ff9578-cc43-43df-a9f7-7480be932f63-config\") pod \"controller-manager-9c7f75cf9-5jvdt\" (UID: \"e3ff9578-cc43-43df-a9f7-7480be932f63\") " pod="openshift-controller-manager/controller-manager-9c7f75cf9-5jvdt" Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.267613 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e3ff9578-cc43-43df-a9f7-7480be932f63-proxy-ca-bundles\") pod \"controller-manager-9c7f75cf9-5jvdt\" (UID: \"e3ff9578-cc43-43df-a9f7-7480be932f63\") " pod="openshift-controller-manager/controller-manager-9c7f75cf9-5jvdt" Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.268230 4907 generic.go:334] "Generic (PLEG): container finished" podID="ec3466c2-4954-459d-88d8-e8c6cbf309e8" containerID="5129da37c4baac30c1ed93e43558cb65af057e72717f5cffd7b1409299226e40" exitCode=0 Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.268294 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fmzmc" event={"ID":"ec3466c2-4954-459d-88d8-e8c6cbf309e8","Type":"ContainerDied","Data":"5129da37c4baac30c1ed93e43558cb65af057e72717f5cffd7b1409299226e40"} Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.268421 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e3ff9578-cc43-43df-a9f7-7480be932f63-client-ca\") pod \"controller-manager-9c7f75cf9-5jvdt\" (UID: \"e3ff9578-cc43-43df-a9f7-7480be932f63\") " pod="openshift-controller-manager/controller-manager-9c7f75cf9-5jvdt" Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.269856 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b17406bc-b62f-4633-a910-cbb294546ed8-client-ca\") pod \"route-controller-manager-5f8c8c7848-sbqs7\" (UID: \"b17406bc-b62f-4633-a910-cbb294546ed8\") " pod="openshift-route-controller-manager/route-controller-manager-5f8c8c7848-sbqs7" Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.270441 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b17406bc-b62f-4633-a910-cbb294546ed8-config\") pod \"route-controller-manager-5f8c8c7848-sbqs7\" (UID: \"b17406bc-b62f-4633-a910-cbb294546ed8\") " pod="openshift-route-controller-manager/route-controller-manager-5f8c8c7848-sbqs7" Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.274737 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e3ff9578-cc43-43df-a9f7-7480be932f63-serving-cert\") pod \"controller-manager-9c7f75cf9-5jvdt\" (UID: \"e3ff9578-cc43-43df-a9f7-7480be932f63\") " pod="openshift-controller-manager/controller-manager-9c7f75cf9-5jvdt" Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.277023 4907 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.278813 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b17406bc-b62f-4633-a910-cbb294546ed8-serving-cert\") pod \"route-controller-manager-5f8c8c7848-sbqs7\" (UID: \"b17406bc-b62f-4633-a910-cbb294546ed8\") " pod="openshift-route-controller-manager/route-controller-manager-5f8c8c7848-sbqs7" Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.286435 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pl7qj\" (UniqueName: \"kubernetes.io/projected/b17406bc-b62f-4633-a910-cbb294546ed8-kube-api-access-pl7qj\") pod \"route-controller-manager-5f8c8c7848-sbqs7\" (UID: \"b17406bc-b62f-4633-a910-cbb294546ed8\") " pod="openshift-route-controller-manager/route-controller-manager-5f8c8c7848-sbqs7" Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.287112 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q888d\" (UniqueName: \"kubernetes.io/projected/e3ff9578-cc43-43df-a9f7-7480be932f63-kube-api-access-q888d\") pod \"controller-manager-9c7f75cf9-5jvdt\" (UID: \"e3ff9578-cc43-43df-a9f7-7480be932f63\") " pod="openshift-controller-manager/controller-manager-9c7f75cf9-5jvdt" Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.337927 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.348952 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-4h9q7"] Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.355567 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-4h9q7"] Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.363948 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-kwfhg"] Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.368618 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.368816 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-kwfhg"] Mar 13 14:09:17 crc kubenswrapper[4907]: E0313 14:09:17.369795 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:17.869773985 +0000 UTC m=+256.769561674 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.382464 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-wgpdd" Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.389686 4907 scope.go:117] "RemoveContainer" containerID="11d51b3d36d9283f41d21738ca7a244487cc24be4453cd47b19366b6bff8bcf1" Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.405960 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-9c7f75cf9-5jvdt" Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.435850 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5f8c8c7848-sbqs7" Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.469678 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:17 crc kubenswrapper[4907]: E0313 14:09:17.470182 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:17.970159604 +0000 UTC m=+256.869947293 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.472980 4907 patch_prober.go:28] interesting pod/router-default-5444994796-f8mmz container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 13 14:09:17 crc kubenswrapper[4907]: [-]has-synced failed: reason withheld Mar 13 14:09:17 crc kubenswrapper[4907]: [+]process-running ok Mar 13 14:09:17 crc kubenswrapper[4907]: healthz check failed Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.473032 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-f8mmz" podUID="c26ccc9f-e682-4b94-9539-cb2607725f71" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.571189 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:17 crc kubenswrapper[4907]: E0313 14:09:17.572899 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:18.072867147 +0000 UTC m=+256.972654836 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.678557 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:17 crc kubenswrapper[4907]: E0313 14:09:17.678791 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:18.178750895 +0000 UTC m=+257.078538584 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.679077 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:17 crc kubenswrapper[4907]: E0313 14:09:17.679470 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:18.179455734 +0000 UTC m=+257.079243413 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.784113 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:17 crc kubenswrapper[4907]: E0313 14:09:17.784473 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:18.284414748 +0000 UTC m=+257.184202437 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.784604 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:17 crc kubenswrapper[4907]: E0313 14:09:17.785080 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:18.285064475 +0000 UTC m=+257.184852164 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.835501 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="814c6dec-6e96-4068-9aec-9e57fefca5db" path="/var/lib/kubelet/pods/814c6dec-6e96-4068-9aec-9e57fefca5db/volumes" Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.836852 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9bb1952d-0e55-4e45-81c0-69a8cddb827e" path="/var/lib/kubelet/pods/9bb1952d-0e55-4e45-81c0-69a8cddb827e/volumes" Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.837326 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-9c7f75cf9-5jvdt"] Mar 13 14:09:17 crc kubenswrapper[4907]: W0313 14:09:17.848577 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode3ff9578_cc43_43df_a9f7_7480be932f63.slice/crio-65e068bf514e5d70b92a415eb6b665b20a27f8eb668a1c226c56693a6080a536 WatchSource:0}: Error finding container 65e068bf514e5d70b92a415eb6b665b20a27f8eb668a1c226c56693a6080a536: Status 404 returned error can't find the container with id 65e068bf514e5d70b92a415eb6b665b20a27f8eb668a1c226c56693a6080a536 Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.886720 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:17 crc kubenswrapper[4907]: E0313 14:09:17.886937 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:18.386896924 +0000 UTC m=+257.286684633 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.887343 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:17 crc kubenswrapper[4907]: E0313 14:09:17.887869 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:18.38786061 +0000 UTC m=+257.287648299 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.889048 4907 ???:1] "http: TLS handshake error from 192.168.126.11:55096: no serving certificate available for the kubelet" Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.989415 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:17 crc kubenswrapper[4907]: E0313 14:09:17.989536 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:18.489513593 +0000 UTC m=+257.389301282 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.990383 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:17 crc kubenswrapper[4907]: E0313 14:09:17.990953 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:18.490928832 +0000 UTC m=+257.390716721 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:18 crc kubenswrapper[4907]: I0313 14:09:18.020271 4907 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2026-03-13T14:09:17.277046235Z","Handler":null,"Name":""} Mar 13 14:09:18 crc kubenswrapper[4907]: I0313 14:09:18.041579 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 14:09:18 crc kubenswrapper[4907]: I0313 14:09:18.041640 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 14:09:18 crc kubenswrapper[4907]: I0313 14:09:18.041652 4907 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Mar 13 14:09:18 crc kubenswrapper[4907]: I0313 14:09:18.041718 4907 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Mar 13 14:09:18 crc kubenswrapper[4907]: I0313 14:09:18.091185 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:18 crc kubenswrapper[4907]: I0313 14:09:18.097704 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5f8c8c7848-sbqs7"] Mar 13 14:09:18 crc kubenswrapper[4907]: I0313 14:09:18.099529 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Mar 13 14:09:18 crc kubenswrapper[4907]: W0313 14:09:18.105108 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb17406bc_b62f_4633_a910_cbb294546ed8.slice/crio-f7b10d1097439391fe4ec620dc14b8023cbc5d2fa40a7a2598511514510245df WatchSource:0}: Error finding container f7b10d1097439391fe4ec620dc14b8023cbc5d2fa40a7a2598511514510245df: Status 404 returned error can't find the container with id f7b10d1097439391fe4ec620dc14b8023cbc5d2fa40a7a2598511514510245df Mar 13 14:09:18 crc kubenswrapper[4907]: I0313 14:09:18.192721 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:18 crc kubenswrapper[4907]: I0313 14:09:18.195645 4907 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 13 14:09:18 crc kubenswrapper[4907]: I0313 14:09:18.195689 4907 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:18 crc kubenswrapper[4907]: I0313 14:09:18.245199 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:18 crc kubenswrapper[4907]: I0313 14:09:18.297327 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Mar 13 14:09:18 crc kubenswrapper[4907]: I0313 14:09:18.306123 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:18 crc kubenswrapper[4907]: I0313 14:09:18.318915 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5f8c8c7848-sbqs7" event={"ID":"b17406bc-b62f-4633-a910-cbb294546ed8","Type":"ContainerStarted","Data":"f7b10d1097439391fe4ec620dc14b8023cbc5d2fa40a7a2598511514510245df"} Mar 13 14:09:18 crc kubenswrapper[4907]: I0313 14:09:18.351017 4907 generic.go:334] "Generic (PLEG): container finished" podID="909a1142-83a3-43c2-9032-45f79a411200" containerID="ae0944c8bd029d03d163fe172529c94d89626bc0ff58ec09f1ac2cc39d977fa6" exitCode=0 Mar 13 14:09:18 crc kubenswrapper[4907]: I0313 14:09:18.352012 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"909a1142-83a3-43c2-9032-45f79a411200","Type":"ContainerDied","Data":"ae0944c8bd029d03d163fe172529c94d89626bc0ff58ec09f1ac2cc39d977fa6"} Mar 13 14:09:18 crc kubenswrapper[4907]: I0313 14:09:18.386426 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-sjwvn" event={"ID":"4a8597a4-c72f-4023-b2fb-1b1237411a36","Type":"ContainerStarted","Data":"98989dbdfebe1005d6e35af2ea045605979542bf34f85e4631fcd7ebd963e87c"} Mar 13 14:09:18 crc kubenswrapper[4907]: I0313 14:09:18.395046 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-9c7f75cf9-5jvdt" event={"ID":"e3ff9578-cc43-43df-a9f7-7480be932f63","Type":"ContainerStarted","Data":"c215281d0e491b45da32a8e52ffe10fe7add03de5080c14587c1be4c05f1e89d"} Mar 13 14:09:18 crc kubenswrapper[4907]: I0313 14:09:18.395144 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-9c7f75cf9-5jvdt" Mar 13 14:09:18 crc kubenswrapper[4907]: I0313 14:09:18.395168 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-9c7f75cf9-5jvdt" event={"ID":"e3ff9578-cc43-43df-a9f7-7480be932f63","Type":"ContainerStarted","Data":"65e068bf514e5d70b92a415eb6b665b20a27f8eb668a1c226c56693a6080a536"} Mar 13 14:09:18 crc kubenswrapper[4907]: I0313 14:09:18.399114 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"e0ad06d9-5751-43e2-bf40-b037e61934f5","Type":"ContainerStarted","Data":"1e6d9369dcd3ed9950da83d6e5af4c20e4ff979f893db5aee4943245de2541e3"} Mar 13 14:09:18 crc kubenswrapper[4907]: I0313 14:09:18.399149 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"e0ad06d9-5751-43e2-bf40-b037e61934f5","Type":"ContainerStarted","Data":"e609619b64fb7b2f34b3160b59695ed7992caf3eb610ed3befd8b460e3b1a894"} Mar 13 14:09:18 crc kubenswrapper[4907]: I0313 14:09:18.401852 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-9c7f75cf9-5jvdt" Mar 13 14:09:18 crc kubenswrapper[4907]: I0313 14:09:18.417349 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-9c7f75cf9-5jvdt" podStartSLOduration=3.4173127340000002 podStartE2EDuration="3.417312734s" podCreationTimestamp="2026-03-13 14:09:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:09:18.415545387 +0000 UTC m=+257.315333076" watchObservedRunningTime="2026-03-13 14:09:18.417312734 +0000 UTC m=+257.317100423" Mar 13 14:09:18 crc kubenswrapper[4907]: I0313 14:09:18.463993 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=2.463965342 podStartE2EDuration="2.463965342s" podCreationTimestamp="2026-03-13 14:09:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:09:18.461196627 +0000 UTC m=+257.360984316" watchObservedRunningTime="2026-03-13 14:09:18.463965342 +0000 UTC m=+257.363753031" Mar 13 14:09:19 crc kubenswrapper[4907]: I0313 14:09:19.503047 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:09:19 crc kubenswrapper[4907]: I0313 14:09:19.526747 4907 patch_prober.go:28] interesting pod/router-default-5444994796-f8mmz container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 13 14:09:19 crc kubenswrapper[4907]: [-]has-synced failed: reason withheld Mar 13 14:09:19 crc kubenswrapper[4907]: [+]process-running ok Mar 13 14:09:19 crc kubenswrapper[4907]: healthz check failed Mar 13 14:09:19 crc kubenswrapper[4907]: I0313 14:09:19.526817 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-f8mmz" podUID="c26ccc9f-e682-4b94-9539-cb2607725f71" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 14:09:19 crc kubenswrapper[4907]: I0313 14:09:19.531843 4907 patch_prober.go:28] interesting pod/router-default-5444994796-f8mmz container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 13 14:09:19 crc kubenswrapper[4907]: [-]has-synced failed: reason withheld Mar 13 14:09:19 crc kubenswrapper[4907]: [+]process-running ok Mar 13 14:09:19 crc kubenswrapper[4907]: healthz check failed Mar 13 14:09:19 crc kubenswrapper[4907]: I0313 14:09:19.546325 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-f8mmz" podUID="c26ccc9f-e682-4b94-9539-cb2607725f71" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 14:09:19 crc kubenswrapper[4907]: I0313 14:09:19.547480 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-8jqgq" Mar 13 14:09:19 crc kubenswrapper[4907]: I0313 14:09:19.726100 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-sjwvn" event={"ID":"4a8597a4-c72f-4023-b2fb-1b1237411a36","Type":"ContainerStarted","Data":"c5bf646885b1b8dde36b6dc97c65514d19d56b34f146bae17445453273a1f619"} Mar 13 14:09:19 crc kubenswrapper[4907]: I0313 14:09:19.785605 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-sjwvn" podStartSLOduration=19.785579803 podStartE2EDuration="19.785579803s" podCreationTimestamp="2026-03-13 14:09:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:09:19.785036538 +0000 UTC m=+258.684824227" watchObservedRunningTime="2026-03-13 14:09:19.785579803 +0000 UTC m=+258.685367502" Mar 13 14:09:19 crc kubenswrapper[4907]: I0313 14:09:19.802436 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Mar 13 14:09:19 crc kubenswrapper[4907]: I0313 14:09:19.825282 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ksgr6" Mar 13 14:09:19 crc kubenswrapper[4907]: I0313 14:09:19.950169 4907 patch_prober.go:28] interesting pod/apiserver-76f77b778f-r86r4 container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Mar 13 14:09:19 crc kubenswrapper[4907]: [+]log ok Mar 13 14:09:19 crc kubenswrapper[4907]: [+]etcd ok Mar 13 14:09:19 crc kubenswrapper[4907]: [+]poststarthook/start-apiserver-admission-initializer ok Mar 13 14:09:19 crc kubenswrapper[4907]: [+]poststarthook/generic-apiserver-start-informers ok Mar 13 14:09:19 crc kubenswrapper[4907]: [+]poststarthook/max-in-flight-filter ok Mar 13 14:09:19 crc kubenswrapper[4907]: [+]poststarthook/storage-object-count-tracker-hook ok Mar 13 14:09:19 crc kubenswrapper[4907]: [+]poststarthook/image.openshift.io-apiserver-caches ok Mar 13 14:09:19 crc kubenswrapper[4907]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Mar 13 14:09:19 crc kubenswrapper[4907]: [+]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa ok Mar 13 14:09:19 crc kubenswrapper[4907]: [+]poststarthook/project.openshift.io-projectcache ok Mar 13 14:09:19 crc kubenswrapper[4907]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Mar 13 14:09:19 crc kubenswrapper[4907]: [+]poststarthook/openshift.io-startinformers ok Mar 13 14:09:19 crc kubenswrapper[4907]: [+]poststarthook/openshift.io-restmapperupdater ok Mar 13 14:09:19 crc kubenswrapper[4907]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Mar 13 14:09:19 crc kubenswrapper[4907]: livez check failed Mar 13 14:09:19 crc kubenswrapper[4907]: I0313 14:09:19.950238 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-r86r4" podUID="5a9c889b-5954-4d3a-ae72-12000ee1cc25" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 14:09:19 crc kubenswrapper[4907]: I0313 14:09:19.985186 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-q9z8k"] Mar 13 14:09:20 crc kubenswrapper[4907]: W0313 14:09:20.028059 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9c5d91bb_753e_4554_b7a1_6c85fe323875.slice/crio-b92738be37732b07d185fe67067541e81cf30ad45197ff888383d57285721ac7 WatchSource:0}: Error finding container b92738be37732b07d185fe67067541e81cf30ad45197ff888383d57285721ac7: Status 404 returned error can't find the container with id b92738be37732b07d185fe67067541e81cf30ad45197ff888383d57285721ac7 Mar 13 14:09:20 crc kubenswrapper[4907]: I0313 14:09:20.462618 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 13 14:09:20 crc kubenswrapper[4907]: I0313 14:09:20.476104 4907 patch_prober.go:28] interesting pod/router-default-5444994796-f8mmz container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 13 14:09:20 crc kubenswrapper[4907]: [-]has-synced failed: reason withheld Mar 13 14:09:20 crc kubenswrapper[4907]: [+]process-running ok Mar 13 14:09:20 crc kubenswrapper[4907]: healthz check failed Mar 13 14:09:20 crc kubenswrapper[4907]: I0313 14:09:20.476173 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-f8mmz" podUID="c26ccc9f-e682-4b94-9539-cb2607725f71" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 14:09:20 crc kubenswrapper[4907]: I0313 14:09:20.589102 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/909a1142-83a3-43c2-9032-45f79a411200-kubelet-dir\") pod \"909a1142-83a3-43c2-9032-45f79a411200\" (UID: \"909a1142-83a3-43c2-9032-45f79a411200\") " Mar 13 14:09:20 crc kubenswrapper[4907]: I0313 14:09:20.589531 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/909a1142-83a3-43c2-9032-45f79a411200-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "909a1142-83a3-43c2-9032-45f79a411200" (UID: "909a1142-83a3-43c2-9032-45f79a411200"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 14:09:20 crc kubenswrapper[4907]: I0313 14:09:20.589685 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/909a1142-83a3-43c2-9032-45f79a411200-kube-api-access\") pod \"909a1142-83a3-43c2-9032-45f79a411200\" (UID: \"909a1142-83a3-43c2-9032-45f79a411200\") " Mar 13 14:09:20 crc kubenswrapper[4907]: I0313 14:09:20.590081 4907 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/909a1142-83a3-43c2-9032-45f79a411200-kubelet-dir\") on node \"crc\" DevicePath \"\"" Mar 13 14:09:20 crc kubenswrapper[4907]: I0313 14:09:20.605407 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/909a1142-83a3-43c2-9032-45f79a411200-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "909a1142-83a3-43c2-9032-45f79a411200" (UID: "909a1142-83a3-43c2-9032-45f79a411200"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:09:20 crc kubenswrapper[4907]: I0313 14:09:20.696143 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/909a1142-83a3-43c2-9032-45f79a411200-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 13 14:09:20 crc kubenswrapper[4907]: I0313 14:09:20.730398 4907 ???:1] "http: TLS handshake error from 192.168.126.11:35328: no serving certificate available for the kubelet" Mar 13 14:09:20 crc kubenswrapper[4907]: I0313 14:09:20.848605 4907 generic.go:334] "Generic (PLEG): container finished" podID="e0ad06d9-5751-43e2-bf40-b037e61934f5" containerID="1e6d9369dcd3ed9950da83d6e5af4c20e4ff979f893db5aee4943245de2541e3" exitCode=0 Mar 13 14:09:20 crc kubenswrapper[4907]: I0313 14:09:20.848700 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"e0ad06d9-5751-43e2-bf40-b037e61934f5","Type":"ContainerDied","Data":"1e6d9369dcd3ed9950da83d6e5af4c20e4ff979f893db5aee4943245de2541e3"} Mar 13 14:09:20 crc kubenswrapper[4907]: I0313 14:09:20.866325 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" event={"ID":"9c5d91bb-753e-4554-b7a1-6c85fe323875","Type":"ContainerStarted","Data":"b92738be37732b07d185fe67067541e81cf30ad45197ff888383d57285721ac7"} Mar 13 14:09:20 crc kubenswrapper[4907]: I0313 14:09:20.883242 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5f8c8c7848-sbqs7" event={"ID":"b17406bc-b62f-4633-a910-cbb294546ed8","Type":"ContainerStarted","Data":"2fe15cd6ab7eff3941755a8f44ea7d5a0fd89e8b799f655d40ebc796fb88ee73"} Mar 13 14:09:20 crc kubenswrapper[4907]: I0313 14:09:20.884818 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-5f8c8c7848-sbqs7" Mar 13 14:09:20 crc kubenswrapper[4907]: I0313 14:09:20.901109 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-5f8c8c7848-sbqs7" Mar 13 14:09:20 crc kubenswrapper[4907]: I0313 14:09:20.923638 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"909a1142-83a3-43c2-9032-45f79a411200","Type":"ContainerDied","Data":"01d238bc9f78f8ddbe9c3e9ff416221f6b3b235da4624c67b2cbebfb3bfea1b2"} Mar 13 14:09:20 crc kubenswrapper[4907]: I0313 14:09:20.923715 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="01d238bc9f78f8ddbe9c3e9ff416221f6b3b235da4624c67b2cbebfb3bfea1b2" Mar 13 14:09:20 crc kubenswrapper[4907]: I0313 14:09:20.923813 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 13 14:09:20 crc kubenswrapper[4907]: I0313 14:09:20.965477 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-5f8c8c7848-sbqs7" podStartSLOduration=5.96545711 podStartE2EDuration="5.96545711s" podCreationTimestamp="2026-03-13 14:09:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:09:20.933563653 +0000 UTC m=+259.833351342" watchObservedRunningTime="2026-03-13 14:09:20.96545711 +0000 UTC m=+259.865244799" Mar 13 14:09:21 crc kubenswrapper[4907]: I0313 14:09:21.473766 4907 patch_prober.go:28] interesting pod/router-default-5444994796-f8mmz container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 13 14:09:21 crc kubenswrapper[4907]: [-]has-synced failed: reason withheld Mar 13 14:09:21 crc kubenswrapper[4907]: [+]process-running ok Mar 13 14:09:21 crc kubenswrapper[4907]: healthz check failed Mar 13 14:09:21 crc kubenswrapper[4907]: I0313 14:09:21.473896 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-f8mmz" podUID="c26ccc9f-e682-4b94-9539-cb2607725f71" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 14:09:21 crc kubenswrapper[4907]: I0313 14:09:21.972471 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" event={"ID":"9c5d91bb-753e-4554-b7a1-6c85fe323875","Type":"ContainerStarted","Data":"e2452e8d8fe737b72f44a8aba2c006a14d53bdfda4e4a01534d104020e7dd4aa"} Mar 13 14:09:21 crc kubenswrapper[4907]: I0313 14:09:21.973229 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:21 crc kubenswrapper[4907]: I0313 14:09:21.996389 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" podStartSLOduration=200.996362097 podStartE2EDuration="3m20.996362097s" podCreationTimestamp="2026-03-13 14:06:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:09:21.995030431 +0000 UTC m=+260.894818120" watchObservedRunningTime="2026-03-13 14:09:21.996362097 +0000 UTC m=+260.896149786" Mar 13 14:09:22 crc kubenswrapper[4907]: I0313 14:09:22.297954 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 13 14:09:22 crc kubenswrapper[4907]: I0313 14:09:22.354476 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e0ad06d9-5751-43e2-bf40-b037e61934f5-kubelet-dir\") pod \"e0ad06d9-5751-43e2-bf40-b037e61934f5\" (UID: \"e0ad06d9-5751-43e2-bf40-b037e61934f5\") " Mar 13 14:09:22 crc kubenswrapper[4907]: I0313 14:09:22.354645 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e0ad06d9-5751-43e2-bf40-b037e61934f5-kube-api-access\") pod \"e0ad06d9-5751-43e2-bf40-b037e61934f5\" (UID: \"e0ad06d9-5751-43e2-bf40-b037e61934f5\") " Mar 13 14:09:22 crc kubenswrapper[4907]: I0313 14:09:22.356042 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e0ad06d9-5751-43e2-bf40-b037e61934f5-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "e0ad06d9-5751-43e2-bf40-b037e61934f5" (UID: "e0ad06d9-5751-43e2-bf40-b037e61934f5"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 14:09:22 crc kubenswrapper[4907]: I0313 14:09:22.371257 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e0ad06d9-5751-43e2-bf40-b037e61934f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e0ad06d9-5751-43e2-bf40-b037e61934f5" (UID: "e0ad06d9-5751-43e2-bf40-b037e61934f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:09:22 crc kubenswrapper[4907]: I0313 14:09:22.456110 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e0ad06d9-5751-43e2-bf40-b037e61934f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 13 14:09:22 crc kubenswrapper[4907]: I0313 14:09:22.456149 4907 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e0ad06d9-5751-43e2-bf40-b037e61934f5-kubelet-dir\") on node \"crc\" DevicePath \"\"" Mar 13 14:09:22 crc kubenswrapper[4907]: I0313 14:09:22.464779 4907 patch_prober.go:28] interesting pod/router-default-5444994796-f8mmz container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 13 14:09:22 crc kubenswrapper[4907]: [-]has-synced failed: reason withheld Mar 13 14:09:22 crc kubenswrapper[4907]: [+]process-running ok Mar 13 14:09:22 crc kubenswrapper[4907]: healthz check failed Mar 13 14:09:22 crc kubenswrapper[4907]: I0313 14:09:22.464836 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-f8mmz" podUID="c26ccc9f-e682-4b94-9539-cb2607725f71" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 14:09:22 crc kubenswrapper[4907]: I0313 14:09:22.994945 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"e0ad06d9-5751-43e2-bf40-b037e61934f5","Type":"ContainerDied","Data":"e609619b64fb7b2f34b3160b59695ed7992caf3eb610ed3befd8b460e3b1a894"} Mar 13 14:09:22 crc kubenswrapper[4907]: I0313 14:09:22.995452 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e609619b64fb7b2f34b3160b59695ed7992caf3eb610ed3befd8b460e3b1a894" Mar 13 14:09:22 crc kubenswrapper[4907]: I0313 14:09:22.995080 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 13 14:09:23 crc kubenswrapper[4907]: I0313 14:09:23.040124 4907 ???:1] "http: TLS handshake error from 192.168.126.11:35338: no serving certificate available for the kubelet" Mar 13 14:09:23 crc kubenswrapper[4907]: I0313 14:09:23.464182 4907 patch_prober.go:28] interesting pod/router-default-5444994796-f8mmz container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 13 14:09:23 crc kubenswrapper[4907]: [-]has-synced failed: reason withheld Mar 13 14:09:23 crc kubenswrapper[4907]: [+]process-running ok Mar 13 14:09:23 crc kubenswrapper[4907]: healthz check failed Mar 13 14:09:23 crc kubenswrapper[4907]: I0313 14:09:23.464261 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-f8mmz" podUID="c26ccc9f-e682-4b94-9539-cb2607725f71" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 14:09:24 crc kubenswrapper[4907]: I0313 14:09:24.465556 4907 patch_prober.go:28] interesting pod/router-default-5444994796-f8mmz container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 13 14:09:24 crc kubenswrapper[4907]: [-]has-synced failed: reason withheld Mar 13 14:09:24 crc kubenswrapper[4907]: [+]process-running ok Mar 13 14:09:24 crc kubenswrapper[4907]: healthz check failed Mar 13 14:09:24 crc kubenswrapper[4907]: I0313 14:09:24.466032 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-f8mmz" podUID="c26ccc9f-e682-4b94-9539-cb2607725f71" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 14:09:24 crc kubenswrapper[4907]: I0313 14:09:24.924865 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-r86r4" Mar 13 14:09:24 crc kubenswrapper[4907]: I0313 14:09:24.931245 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-r86r4" Mar 13 14:09:25 crc kubenswrapper[4907]: I0313 14:09:25.108559 4907 patch_prober.go:28] interesting pod/downloads-7954f5f757-544df container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Mar 13 14:09:25 crc kubenswrapper[4907]: I0313 14:09:25.108624 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-544df" podUID="f8ea4f24-05b0-4661-a617-a4208f9a9188" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Mar 13 14:09:25 crc kubenswrapper[4907]: I0313 14:09:25.108559 4907 patch_prober.go:28] interesting pod/downloads-7954f5f757-544df container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Mar 13 14:09:25 crc kubenswrapper[4907]: I0313 14:09:25.109700 4907 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-544df" podUID="f8ea4f24-05b0-4661-a617-a4208f9a9188" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Mar 13 14:09:25 crc kubenswrapper[4907]: I0313 14:09:25.273171 4907 patch_prober.go:28] interesting pod/console-f9d7485db-l9xpb container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.21:8443/health\": dial tcp 10.217.0.21:8443: connect: connection refused" start-of-body= Mar 13 14:09:25 crc kubenswrapper[4907]: I0313 14:09:25.273255 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-l9xpb" podUID="4c8c0266-5452-4abb-92c5-c536df94dd41" containerName="console" probeResult="failure" output="Get \"https://10.217.0.21:8443/health\": dial tcp 10.217.0.21:8443: connect: connection refused" Mar 13 14:09:25 crc kubenswrapper[4907]: I0313 14:09:25.466679 4907 patch_prober.go:28] interesting pod/router-default-5444994796-f8mmz container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 13 14:09:25 crc kubenswrapper[4907]: [-]has-synced failed: reason withheld Mar 13 14:09:25 crc kubenswrapper[4907]: [+]process-running ok Mar 13 14:09:25 crc kubenswrapper[4907]: healthz check failed Mar 13 14:09:25 crc kubenswrapper[4907]: I0313 14:09:25.466782 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-f8mmz" podUID="c26ccc9f-e682-4b94-9539-cb2607725f71" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 14:09:26 crc kubenswrapper[4907]: I0313 14:09:26.463527 4907 patch_prober.go:28] interesting pod/router-default-5444994796-f8mmz container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 13 14:09:26 crc kubenswrapper[4907]: [+]has-synced ok Mar 13 14:09:26 crc kubenswrapper[4907]: [+]process-running ok Mar 13 14:09:26 crc kubenswrapper[4907]: healthz check failed Mar 13 14:09:26 crc kubenswrapper[4907]: I0313 14:09:26.463603 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-f8mmz" podUID="c26ccc9f-e682-4b94-9539-cb2607725f71" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 14:09:27 crc kubenswrapper[4907]: I0313 14:09:27.464395 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-f8mmz" Mar 13 14:09:27 crc kubenswrapper[4907]: I0313 14:09:27.467116 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-f8mmz" Mar 13 14:09:31 crc kubenswrapper[4907]: I0313 14:09:31.873329 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-9c7f75cf9-5jvdt"] Mar 13 14:09:31 crc kubenswrapper[4907]: I0313 14:09:31.874078 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-9c7f75cf9-5jvdt" podUID="e3ff9578-cc43-43df-a9f7-7480be932f63" containerName="controller-manager" containerID="cri-o://c215281d0e491b45da32a8e52ffe10fe7add03de5080c14587c1be4c05f1e89d" gracePeriod=30 Mar 13 14:09:31 crc kubenswrapper[4907]: I0313 14:09:31.887631 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5f8c8c7848-sbqs7"] Mar 13 14:09:31 crc kubenswrapper[4907]: I0313 14:09:31.887903 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-5f8c8c7848-sbqs7" podUID="b17406bc-b62f-4633-a910-cbb294546ed8" containerName="route-controller-manager" containerID="cri-o://2fe15cd6ab7eff3941755a8f44ea7d5a0fd89e8b799f655d40ebc796fb88ee73" gracePeriod=30 Mar 13 14:09:33 crc kubenswrapper[4907]: I0313 14:09:33.096066 4907 generic.go:334] "Generic (PLEG): container finished" podID="b17406bc-b62f-4633-a910-cbb294546ed8" containerID="2fe15cd6ab7eff3941755a8f44ea7d5a0fd89e8b799f655d40ebc796fb88ee73" exitCode=0 Mar 13 14:09:33 crc kubenswrapper[4907]: I0313 14:09:33.096152 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5f8c8c7848-sbqs7" event={"ID":"b17406bc-b62f-4633-a910-cbb294546ed8","Type":"ContainerDied","Data":"2fe15cd6ab7eff3941755a8f44ea7d5a0fd89e8b799f655d40ebc796fb88ee73"} Mar 13 14:09:33 crc kubenswrapper[4907]: I0313 14:09:33.099271 4907 generic.go:334] "Generic (PLEG): container finished" podID="e3ff9578-cc43-43df-a9f7-7480be932f63" containerID="c215281d0e491b45da32a8e52ffe10fe7add03de5080c14587c1be4c05f1e89d" exitCode=0 Mar 13 14:09:33 crc kubenswrapper[4907]: I0313 14:09:33.099328 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-9c7f75cf9-5jvdt" event={"ID":"e3ff9578-cc43-43df-a9f7-7480be932f63","Type":"ContainerDied","Data":"c215281d0e491b45da32a8e52ffe10fe7add03de5080c14587c1be4c05f1e89d"} Mar 13 14:09:35 crc kubenswrapper[4907]: I0313 14:09:35.123039 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-544df" Mar 13 14:09:35 crc kubenswrapper[4907]: I0313 14:09:35.291199 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-l9xpb" Mar 13 14:09:35 crc kubenswrapper[4907]: I0313 14:09:35.296059 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-l9xpb" Mar 13 14:09:37 crc kubenswrapper[4907]: I0313 14:09:37.320102 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5f8c8c7848-sbqs7" Mar 13 14:09:37 crc kubenswrapper[4907]: I0313 14:09:37.357909 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-65cb4586b7-dzz67"] Mar 13 14:09:37 crc kubenswrapper[4907]: E0313 14:09:37.358219 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0ad06d9-5751-43e2-bf40-b037e61934f5" containerName="pruner" Mar 13 14:09:37 crc kubenswrapper[4907]: I0313 14:09:37.358234 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0ad06d9-5751-43e2-bf40-b037e61934f5" containerName="pruner" Mar 13 14:09:37 crc kubenswrapper[4907]: E0313 14:09:37.358248 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b17406bc-b62f-4633-a910-cbb294546ed8" containerName="route-controller-manager" Mar 13 14:09:37 crc kubenswrapper[4907]: I0313 14:09:37.358254 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="b17406bc-b62f-4633-a910-cbb294546ed8" containerName="route-controller-manager" Mar 13 14:09:37 crc kubenswrapper[4907]: E0313 14:09:37.358273 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="909a1142-83a3-43c2-9032-45f79a411200" containerName="pruner" Mar 13 14:09:37 crc kubenswrapper[4907]: I0313 14:09:37.358281 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="909a1142-83a3-43c2-9032-45f79a411200" containerName="pruner" Mar 13 14:09:37 crc kubenswrapper[4907]: I0313 14:09:37.358428 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="b17406bc-b62f-4633-a910-cbb294546ed8" containerName="route-controller-manager" Mar 13 14:09:37 crc kubenswrapper[4907]: I0313 14:09:37.358452 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="e0ad06d9-5751-43e2-bf40-b037e61934f5" containerName="pruner" Mar 13 14:09:37 crc kubenswrapper[4907]: I0313 14:09:37.358465 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="909a1142-83a3-43c2-9032-45f79a411200" containerName="pruner" Mar 13 14:09:37 crc kubenswrapper[4907]: I0313 14:09:37.359066 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-65cb4586b7-dzz67" Mar 13 14:09:37 crc kubenswrapper[4907]: I0313 14:09:37.388198 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-65cb4586b7-dzz67"] Mar 13 14:09:37 crc kubenswrapper[4907]: I0313 14:09:37.469066 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b17406bc-b62f-4633-a910-cbb294546ed8-config\") pod \"b17406bc-b62f-4633-a910-cbb294546ed8\" (UID: \"b17406bc-b62f-4633-a910-cbb294546ed8\") " Mar 13 14:09:37 crc kubenswrapper[4907]: I0313 14:09:37.469345 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b17406bc-b62f-4633-a910-cbb294546ed8-client-ca\") pod \"b17406bc-b62f-4633-a910-cbb294546ed8\" (UID: \"b17406bc-b62f-4633-a910-cbb294546ed8\") " Mar 13 14:09:37 crc kubenswrapper[4907]: I0313 14:09:37.469391 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pl7qj\" (UniqueName: \"kubernetes.io/projected/b17406bc-b62f-4633-a910-cbb294546ed8-kube-api-access-pl7qj\") pod \"b17406bc-b62f-4633-a910-cbb294546ed8\" (UID: \"b17406bc-b62f-4633-a910-cbb294546ed8\") " Mar 13 14:09:37 crc kubenswrapper[4907]: I0313 14:09:37.469427 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b17406bc-b62f-4633-a910-cbb294546ed8-serving-cert\") pod \"b17406bc-b62f-4633-a910-cbb294546ed8\" (UID: \"b17406bc-b62f-4633-a910-cbb294546ed8\") " Mar 13 14:09:37 crc kubenswrapper[4907]: I0313 14:09:37.469818 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7b861e8a-0bd0-4b07-aa40-16b6310b19c1-config\") pod \"route-controller-manager-65cb4586b7-dzz67\" (UID: \"7b861e8a-0bd0-4b07-aa40-16b6310b19c1\") " pod="openshift-route-controller-manager/route-controller-manager-65cb4586b7-dzz67" Mar 13 14:09:37 crc kubenswrapper[4907]: I0313 14:09:37.469941 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fxxvw\" (UniqueName: \"kubernetes.io/projected/7b861e8a-0bd0-4b07-aa40-16b6310b19c1-kube-api-access-fxxvw\") pod \"route-controller-manager-65cb4586b7-dzz67\" (UID: \"7b861e8a-0bd0-4b07-aa40-16b6310b19c1\") " pod="openshift-route-controller-manager/route-controller-manager-65cb4586b7-dzz67" Mar 13 14:09:37 crc kubenswrapper[4907]: I0313 14:09:37.470015 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7b861e8a-0bd0-4b07-aa40-16b6310b19c1-client-ca\") pod \"route-controller-manager-65cb4586b7-dzz67\" (UID: \"7b861e8a-0bd0-4b07-aa40-16b6310b19c1\") " pod="openshift-route-controller-manager/route-controller-manager-65cb4586b7-dzz67" Mar 13 14:09:37 crc kubenswrapper[4907]: I0313 14:09:37.470022 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b17406bc-b62f-4633-a910-cbb294546ed8-config" (OuterVolumeSpecName: "config") pod "b17406bc-b62f-4633-a910-cbb294546ed8" (UID: "b17406bc-b62f-4633-a910-cbb294546ed8"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:09:37 crc kubenswrapper[4907]: I0313 14:09:37.470223 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7b861e8a-0bd0-4b07-aa40-16b6310b19c1-serving-cert\") pod \"route-controller-manager-65cb4586b7-dzz67\" (UID: \"7b861e8a-0bd0-4b07-aa40-16b6310b19c1\") " pod="openshift-route-controller-manager/route-controller-manager-65cb4586b7-dzz67" Mar 13 14:09:37 crc kubenswrapper[4907]: I0313 14:09:37.470385 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b17406bc-b62f-4633-a910-cbb294546ed8-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:09:37 crc kubenswrapper[4907]: I0313 14:09:37.471099 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b17406bc-b62f-4633-a910-cbb294546ed8-client-ca" (OuterVolumeSpecName: "client-ca") pod "b17406bc-b62f-4633-a910-cbb294546ed8" (UID: "b17406bc-b62f-4633-a910-cbb294546ed8"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:09:37 crc kubenswrapper[4907]: I0313 14:09:37.479160 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b17406bc-b62f-4633-a910-cbb294546ed8-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "b17406bc-b62f-4633-a910-cbb294546ed8" (UID: "b17406bc-b62f-4633-a910-cbb294546ed8"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:09:37 crc kubenswrapper[4907]: I0313 14:09:37.479201 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b17406bc-b62f-4633-a910-cbb294546ed8-kube-api-access-pl7qj" (OuterVolumeSpecName: "kube-api-access-pl7qj") pod "b17406bc-b62f-4633-a910-cbb294546ed8" (UID: "b17406bc-b62f-4633-a910-cbb294546ed8"). InnerVolumeSpecName "kube-api-access-pl7qj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:09:37 crc kubenswrapper[4907]: I0313 14:09:37.572070 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7b861e8a-0bd0-4b07-aa40-16b6310b19c1-serving-cert\") pod \"route-controller-manager-65cb4586b7-dzz67\" (UID: \"7b861e8a-0bd0-4b07-aa40-16b6310b19c1\") " pod="openshift-route-controller-manager/route-controller-manager-65cb4586b7-dzz67" Mar 13 14:09:37 crc kubenswrapper[4907]: I0313 14:09:37.572201 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7b861e8a-0bd0-4b07-aa40-16b6310b19c1-config\") pod \"route-controller-manager-65cb4586b7-dzz67\" (UID: \"7b861e8a-0bd0-4b07-aa40-16b6310b19c1\") " pod="openshift-route-controller-manager/route-controller-manager-65cb4586b7-dzz67" Mar 13 14:09:37 crc kubenswrapper[4907]: I0313 14:09:37.572233 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fxxvw\" (UniqueName: \"kubernetes.io/projected/7b861e8a-0bd0-4b07-aa40-16b6310b19c1-kube-api-access-fxxvw\") pod \"route-controller-manager-65cb4586b7-dzz67\" (UID: \"7b861e8a-0bd0-4b07-aa40-16b6310b19c1\") " pod="openshift-route-controller-manager/route-controller-manager-65cb4586b7-dzz67" Mar 13 14:09:37 crc kubenswrapper[4907]: I0313 14:09:37.572264 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7b861e8a-0bd0-4b07-aa40-16b6310b19c1-client-ca\") pod \"route-controller-manager-65cb4586b7-dzz67\" (UID: \"7b861e8a-0bd0-4b07-aa40-16b6310b19c1\") " pod="openshift-route-controller-manager/route-controller-manager-65cb4586b7-dzz67" Mar 13 14:09:37 crc kubenswrapper[4907]: I0313 14:09:37.572311 4907 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b17406bc-b62f-4633-a910-cbb294546ed8-client-ca\") on node \"crc\" DevicePath \"\"" Mar 13 14:09:37 crc kubenswrapper[4907]: I0313 14:09:37.572324 4907 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b17406bc-b62f-4633-a910-cbb294546ed8-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 14:09:37 crc kubenswrapper[4907]: I0313 14:09:37.572337 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pl7qj\" (UniqueName: \"kubernetes.io/projected/b17406bc-b62f-4633-a910-cbb294546ed8-kube-api-access-pl7qj\") on node \"crc\" DevicePath \"\"" Mar 13 14:09:37 crc kubenswrapper[4907]: I0313 14:09:37.573610 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7b861e8a-0bd0-4b07-aa40-16b6310b19c1-client-ca\") pod \"route-controller-manager-65cb4586b7-dzz67\" (UID: \"7b861e8a-0bd0-4b07-aa40-16b6310b19c1\") " pod="openshift-route-controller-manager/route-controller-manager-65cb4586b7-dzz67" Mar 13 14:09:37 crc kubenswrapper[4907]: I0313 14:09:37.574073 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7b861e8a-0bd0-4b07-aa40-16b6310b19c1-config\") pod \"route-controller-manager-65cb4586b7-dzz67\" (UID: \"7b861e8a-0bd0-4b07-aa40-16b6310b19c1\") " pod="openshift-route-controller-manager/route-controller-manager-65cb4586b7-dzz67" Mar 13 14:09:37 crc kubenswrapper[4907]: I0313 14:09:37.577466 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7b861e8a-0bd0-4b07-aa40-16b6310b19c1-serving-cert\") pod \"route-controller-manager-65cb4586b7-dzz67\" (UID: \"7b861e8a-0bd0-4b07-aa40-16b6310b19c1\") " pod="openshift-route-controller-manager/route-controller-manager-65cb4586b7-dzz67" Mar 13 14:09:37 crc kubenswrapper[4907]: I0313 14:09:37.590708 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fxxvw\" (UniqueName: \"kubernetes.io/projected/7b861e8a-0bd0-4b07-aa40-16b6310b19c1-kube-api-access-fxxvw\") pod \"route-controller-manager-65cb4586b7-dzz67\" (UID: \"7b861e8a-0bd0-4b07-aa40-16b6310b19c1\") " pod="openshift-route-controller-manager/route-controller-manager-65cb4586b7-dzz67" Mar 13 14:09:37 crc kubenswrapper[4907]: I0313 14:09:37.697932 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-65cb4586b7-dzz67" Mar 13 14:09:38 crc kubenswrapper[4907]: I0313 14:09:38.143534 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5f8c8c7848-sbqs7" event={"ID":"b17406bc-b62f-4633-a910-cbb294546ed8","Type":"ContainerDied","Data":"f7b10d1097439391fe4ec620dc14b8023cbc5d2fa40a7a2598511514510245df"} Mar 13 14:09:38 crc kubenswrapper[4907]: I0313 14:09:38.143607 4907 scope.go:117] "RemoveContainer" containerID="2fe15cd6ab7eff3941755a8f44ea7d5a0fd89e8b799f655d40ebc796fb88ee73" Mar 13 14:09:38 crc kubenswrapper[4907]: I0313 14:09:38.143651 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5f8c8c7848-sbqs7" Mar 13 14:09:38 crc kubenswrapper[4907]: I0313 14:09:38.168751 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5f8c8c7848-sbqs7"] Mar 13 14:09:38 crc kubenswrapper[4907]: I0313 14:09:38.172295 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5f8c8c7848-sbqs7"] Mar 13 14:09:38 crc kubenswrapper[4907]: I0313 14:09:38.316057 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:38 crc kubenswrapper[4907]: I0313 14:09:38.408009 4907 patch_prober.go:28] interesting pod/controller-manager-9c7f75cf9-5jvdt container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.55:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 14:09:38 crc kubenswrapper[4907]: I0313 14:09:38.408231 4907 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-9c7f75cf9-5jvdt" podUID="e3ff9578-cc43-43df-a9f7-7480be932f63" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.55:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 14:09:40 crc kubenswrapper[4907]: I0313 14:09:39.789526 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b17406bc-b62f-4633-a910-cbb294546ed8" path="/var/lib/kubelet/pods/b17406bc-b62f-4633-a910-cbb294546ed8/volumes" Mar 13 14:09:40 crc kubenswrapper[4907]: I0313 14:09:40.775903 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-9c7f75cf9-5jvdt" Mar 13 14:09:40 crc kubenswrapper[4907]: I0313 14:09:40.805115 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-6f9b8d4bb9-c6vxw"] Mar 13 14:09:40 crc kubenswrapper[4907]: E0313 14:09:40.805462 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3ff9578-cc43-43df-a9f7-7480be932f63" containerName="controller-manager" Mar 13 14:09:40 crc kubenswrapper[4907]: I0313 14:09:40.805479 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3ff9578-cc43-43df-a9f7-7480be932f63" containerName="controller-manager" Mar 13 14:09:40 crc kubenswrapper[4907]: I0313 14:09:40.805628 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3ff9578-cc43-43df-a9f7-7480be932f63" containerName="controller-manager" Mar 13 14:09:40 crc kubenswrapper[4907]: I0313 14:09:40.806211 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6f9b8d4bb9-c6vxw" Mar 13 14:09:40 crc kubenswrapper[4907]: I0313 14:09:40.829122 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6f9b8d4bb9-c6vxw"] Mar 13 14:09:40 crc kubenswrapper[4907]: I0313 14:09:40.920262 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e3ff9578-cc43-43df-a9f7-7480be932f63-proxy-ca-bundles\") pod \"e3ff9578-cc43-43df-a9f7-7480be932f63\" (UID: \"e3ff9578-cc43-43df-a9f7-7480be932f63\") " Mar 13 14:09:40 crc kubenswrapper[4907]: I0313 14:09:40.920399 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e3ff9578-cc43-43df-a9f7-7480be932f63-config\") pod \"e3ff9578-cc43-43df-a9f7-7480be932f63\" (UID: \"e3ff9578-cc43-43df-a9f7-7480be932f63\") " Mar 13 14:09:40 crc kubenswrapper[4907]: I0313 14:09:40.920485 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e3ff9578-cc43-43df-a9f7-7480be932f63-client-ca\") pod \"e3ff9578-cc43-43df-a9f7-7480be932f63\" (UID: \"e3ff9578-cc43-43df-a9f7-7480be932f63\") " Mar 13 14:09:40 crc kubenswrapper[4907]: I0313 14:09:40.920561 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e3ff9578-cc43-43df-a9f7-7480be932f63-serving-cert\") pod \"e3ff9578-cc43-43df-a9f7-7480be932f63\" (UID: \"e3ff9578-cc43-43df-a9f7-7480be932f63\") " Mar 13 14:09:40 crc kubenswrapper[4907]: I0313 14:09:40.920584 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q888d\" (UniqueName: \"kubernetes.io/projected/e3ff9578-cc43-43df-a9f7-7480be932f63-kube-api-access-q888d\") pod \"e3ff9578-cc43-43df-a9f7-7480be932f63\" (UID: \"e3ff9578-cc43-43df-a9f7-7480be932f63\") " Mar 13 14:09:40 crc kubenswrapper[4907]: I0313 14:09:40.920798 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/12427c08-13c7-4bad-9dc3-815d65a8860f-serving-cert\") pod \"controller-manager-6f9b8d4bb9-c6vxw\" (UID: \"12427c08-13c7-4bad-9dc3-815d65a8860f\") " pod="openshift-controller-manager/controller-manager-6f9b8d4bb9-c6vxw" Mar 13 14:09:40 crc kubenswrapper[4907]: I0313 14:09:40.920872 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/12427c08-13c7-4bad-9dc3-815d65a8860f-proxy-ca-bundles\") pod \"controller-manager-6f9b8d4bb9-c6vxw\" (UID: \"12427c08-13c7-4bad-9dc3-815d65a8860f\") " pod="openshift-controller-manager/controller-manager-6f9b8d4bb9-c6vxw" Mar 13 14:09:40 crc kubenswrapper[4907]: I0313 14:09:40.921112 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/12427c08-13c7-4bad-9dc3-815d65a8860f-config\") pod \"controller-manager-6f9b8d4bb9-c6vxw\" (UID: \"12427c08-13c7-4bad-9dc3-815d65a8860f\") " pod="openshift-controller-manager/controller-manager-6f9b8d4bb9-c6vxw" Mar 13 14:09:40 crc kubenswrapper[4907]: I0313 14:09:40.921166 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mbr8t\" (UniqueName: \"kubernetes.io/projected/12427c08-13c7-4bad-9dc3-815d65a8860f-kube-api-access-mbr8t\") pod \"controller-manager-6f9b8d4bb9-c6vxw\" (UID: \"12427c08-13c7-4bad-9dc3-815d65a8860f\") " pod="openshift-controller-manager/controller-manager-6f9b8d4bb9-c6vxw" Mar 13 14:09:40 crc kubenswrapper[4907]: I0313 14:09:40.921190 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e3ff9578-cc43-43df-a9f7-7480be932f63-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "e3ff9578-cc43-43df-a9f7-7480be932f63" (UID: "e3ff9578-cc43-43df-a9f7-7480be932f63"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:09:40 crc kubenswrapper[4907]: I0313 14:09:40.921269 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e3ff9578-cc43-43df-a9f7-7480be932f63-client-ca" (OuterVolumeSpecName: "client-ca") pod "e3ff9578-cc43-43df-a9f7-7480be932f63" (UID: "e3ff9578-cc43-43df-a9f7-7480be932f63"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:09:40 crc kubenswrapper[4907]: I0313 14:09:40.921401 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/12427c08-13c7-4bad-9dc3-815d65a8860f-client-ca\") pod \"controller-manager-6f9b8d4bb9-c6vxw\" (UID: \"12427c08-13c7-4bad-9dc3-815d65a8860f\") " pod="openshift-controller-manager/controller-manager-6f9b8d4bb9-c6vxw" Mar 13 14:09:40 crc kubenswrapper[4907]: I0313 14:09:40.921637 4907 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e3ff9578-cc43-43df-a9f7-7480be932f63-client-ca\") on node \"crc\" DevicePath \"\"" Mar 13 14:09:40 crc kubenswrapper[4907]: I0313 14:09:40.921656 4907 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e3ff9578-cc43-43df-a9f7-7480be932f63-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 13 14:09:40 crc kubenswrapper[4907]: I0313 14:09:40.922045 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e3ff9578-cc43-43df-a9f7-7480be932f63-config" (OuterVolumeSpecName: "config") pod "e3ff9578-cc43-43df-a9f7-7480be932f63" (UID: "e3ff9578-cc43-43df-a9f7-7480be932f63"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:09:40 crc kubenswrapper[4907]: I0313 14:09:40.940335 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3ff9578-cc43-43df-a9f7-7480be932f63-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e3ff9578-cc43-43df-a9f7-7480be932f63" (UID: "e3ff9578-cc43-43df-a9f7-7480be932f63"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:09:40 crc kubenswrapper[4907]: I0313 14:09:40.940491 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e3ff9578-cc43-43df-a9f7-7480be932f63-kube-api-access-q888d" (OuterVolumeSpecName: "kube-api-access-q888d") pod "e3ff9578-cc43-43df-a9f7-7480be932f63" (UID: "e3ff9578-cc43-43df-a9f7-7480be932f63"). InnerVolumeSpecName "kube-api-access-q888d". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:09:41 crc kubenswrapper[4907]: I0313 14:09:41.023197 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/12427c08-13c7-4bad-9dc3-815d65a8860f-serving-cert\") pod \"controller-manager-6f9b8d4bb9-c6vxw\" (UID: \"12427c08-13c7-4bad-9dc3-815d65a8860f\") " pod="openshift-controller-manager/controller-manager-6f9b8d4bb9-c6vxw" Mar 13 14:09:41 crc kubenswrapper[4907]: I0313 14:09:41.023267 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/12427c08-13c7-4bad-9dc3-815d65a8860f-proxy-ca-bundles\") pod \"controller-manager-6f9b8d4bb9-c6vxw\" (UID: \"12427c08-13c7-4bad-9dc3-815d65a8860f\") " pod="openshift-controller-manager/controller-manager-6f9b8d4bb9-c6vxw" Mar 13 14:09:41 crc kubenswrapper[4907]: I0313 14:09:41.023311 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/12427c08-13c7-4bad-9dc3-815d65a8860f-config\") pod \"controller-manager-6f9b8d4bb9-c6vxw\" (UID: \"12427c08-13c7-4bad-9dc3-815d65a8860f\") " pod="openshift-controller-manager/controller-manager-6f9b8d4bb9-c6vxw" Mar 13 14:09:41 crc kubenswrapper[4907]: I0313 14:09:41.023328 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mbr8t\" (UniqueName: \"kubernetes.io/projected/12427c08-13c7-4bad-9dc3-815d65a8860f-kube-api-access-mbr8t\") pod \"controller-manager-6f9b8d4bb9-c6vxw\" (UID: \"12427c08-13c7-4bad-9dc3-815d65a8860f\") " pod="openshift-controller-manager/controller-manager-6f9b8d4bb9-c6vxw" Mar 13 14:09:41 crc kubenswrapper[4907]: I0313 14:09:41.023556 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/12427c08-13c7-4bad-9dc3-815d65a8860f-client-ca\") pod \"controller-manager-6f9b8d4bb9-c6vxw\" (UID: \"12427c08-13c7-4bad-9dc3-815d65a8860f\") " pod="openshift-controller-manager/controller-manager-6f9b8d4bb9-c6vxw" Mar 13 14:09:41 crc kubenswrapper[4907]: I0313 14:09:41.023612 4907 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e3ff9578-cc43-43df-a9f7-7480be932f63-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 14:09:41 crc kubenswrapper[4907]: I0313 14:09:41.023623 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q888d\" (UniqueName: \"kubernetes.io/projected/e3ff9578-cc43-43df-a9f7-7480be932f63-kube-api-access-q888d\") on node \"crc\" DevicePath \"\"" Mar 13 14:09:41 crc kubenswrapper[4907]: I0313 14:09:41.023633 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e3ff9578-cc43-43df-a9f7-7480be932f63-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:09:41 crc kubenswrapper[4907]: I0313 14:09:41.024452 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/12427c08-13c7-4bad-9dc3-815d65a8860f-client-ca\") pod \"controller-manager-6f9b8d4bb9-c6vxw\" (UID: \"12427c08-13c7-4bad-9dc3-815d65a8860f\") " pod="openshift-controller-manager/controller-manager-6f9b8d4bb9-c6vxw" Mar 13 14:09:41 crc kubenswrapper[4907]: I0313 14:09:41.024855 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/12427c08-13c7-4bad-9dc3-815d65a8860f-config\") pod \"controller-manager-6f9b8d4bb9-c6vxw\" (UID: \"12427c08-13c7-4bad-9dc3-815d65a8860f\") " pod="openshift-controller-manager/controller-manager-6f9b8d4bb9-c6vxw" Mar 13 14:09:41 crc kubenswrapper[4907]: I0313 14:09:41.025772 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/12427c08-13c7-4bad-9dc3-815d65a8860f-proxy-ca-bundles\") pod \"controller-manager-6f9b8d4bb9-c6vxw\" (UID: \"12427c08-13c7-4bad-9dc3-815d65a8860f\") " pod="openshift-controller-manager/controller-manager-6f9b8d4bb9-c6vxw" Mar 13 14:09:41 crc kubenswrapper[4907]: I0313 14:09:41.028411 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/12427c08-13c7-4bad-9dc3-815d65a8860f-serving-cert\") pod \"controller-manager-6f9b8d4bb9-c6vxw\" (UID: \"12427c08-13c7-4bad-9dc3-815d65a8860f\") " pod="openshift-controller-manager/controller-manager-6f9b8d4bb9-c6vxw" Mar 13 14:09:41 crc kubenswrapper[4907]: I0313 14:09:41.040603 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mbr8t\" (UniqueName: \"kubernetes.io/projected/12427c08-13c7-4bad-9dc3-815d65a8860f-kube-api-access-mbr8t\") pod \"controller-manager-6f9b8d4bb9-c6vxw\" (UID: \"12427c08-13c7-4bad-9dc3-815d65a8860f\") " pod="openshift-controller-manager/controller-manager-6f9b8d4bb9-c6vxw" Mar 13 14:09:41 crc kubenswrapper[4907]: I0313 14:09:41.127831 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6f9b8d4bb9-c6vxw" Mar 13 14:09:41 crc kubenswrapper[4907]: I0313 14:09:41.165653 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-9c7f75cf9-5jvdt" event={"ID":"e3ff9578-cc43-43df-a9f7-7480be932f63","Type":"ContainerDied","Data":"65e068bf514e5d70b92a415eb6b665b20a27f8eb668a1c226c56693a6080a536"} Mar 13 14:09:41 crc kubenswrapper[4907]: I0313 14:09:41.165755 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-9c7f75cf9-5jvdt" Mar 13 14:09:41 crc kubenswrapper[4907]: I0313 14:09:41.215783 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-9c7f75cf9-5jvdt"] Mar 13 14:09:41 crc kubenswrapper[4907]: I0313 14:09:41.217220 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-9c7f75cf9-5jvdt"] Mar 13 14:09:41 crc kubenswrapper[4907]: I0313 14:09:41.794594 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e3ff9578-cc43-43df-a9f7-7480be932f63" path="/var/lib/kubelet/pods/e3ff9578-cc43-43df-a9f7-7480be932f63/volumes" Mar 13 14:09:43 crc kubenswrapper[4907]: I0313 14:09:43.542146 4907 ???:1] "http: TLS handshake error from 192.168.126.11:49470: no serving certificate available for the kubelet" Mar 13 14:09:44 crc kubenswrapper[4907]: I0313 14:09:44.606425 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ccjsv" Mar 13 14:09:45 crc kubenswrapper[4907]: E0313 14:09:45.975209 4907 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Mar 13 14:09:45 crc kubenswrapper[4907]: E0313 14:09:45.975379 4907 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6mhh2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-zbll9_openshift-marketplace(4b6ddc05-e51d-4648-84c6-6d1414398a4b): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 13 14:09:45 crc kubenswrapper[4907]: E0313 14:09:45.976850 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-zbll9" podUID="4b6ddc05-e51d-4648-84c6-6d1414398a4b" Mar 13 14:09:46 crc kubenswrapper[4907]: E0313 14:09:46.145346 4907 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/openshift4/ose-cli:latest" Mar 13 14:09:46 crc kubenswrapper[4907]: E0313 14:09:46.145539 4907 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 13 14:09:46 crc kubenswrapper[4907]: container &Container{Name:oc,Image:registry.redhat.io/openshift4/ose-cli:latest,Command:[/bin/bash -c oc get csr -o go-template='{{range .items}}{{if not .status}}{{.metadata.name}}{{"\n"}}{{end}}{{end}}' | xargs --no-run-if-empty oc adm certificate approve Mar 13 14:09:46 crc kubenswrapper[4907]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-9f5cd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod auto-csr-approver-29556848-v8kv4_openshift-infra(a8867781-179d-47e4-9a55-330deb10d273): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled Mar 13 14:09:46 crc kubenswrapper[4907]: > logger="UnhandledError" Mar 13 14:09:46 crc kubenswrapper[4907]: E0313 14:09:46.146769 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-infra/auto-csr-approver-29556848-v8kv4" podUID="a8867781-179d-47e4-9a55-330deb10d273" Mar 13 14:09:46 crc kubenswrapper[4907]: E0313 14:09:46.203599 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29556848-v8kv4" podUID="a8867781-179d-47e4-9a55-330deb10d273" Mar 13 14:09:47 crc kubenswrapper[4907]: E0313 14:09:47.694115 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-zbll9" podUID="4b6ddc05-e51d-4648-84c6-6d1414398a4b" Mar 13 14:09:47 crc kubenswrapper[4907]: E0313 14:09:47.784004 4907 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Mar 13 14:09:47 crc kubenswrapper[4907]: E0313 14:09:47.784519 4907 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rjgtx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-p257z_openshift-marketplace(775414ea-5b94-4ab6-8b7c-fe7c80671ced): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 13 14:09:47 crc kubenswrapper[4907]: E0313 14:09:47.786081 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-p257z" podUID="775414ea-5b94-4ab6-8b7c-fe7c80671ced" Mar 13 14:09:48 crc kubenswrapper[4907]: I0313 14:09:48.041674 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 14:09:48 crc kubenswrapper[4907]: I0313 14:09:48.041755 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 14:09:48 crc kubenswrapper[4907]: I0313 14:09:48.041811 4907 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" Mar 13 14:09:48 crc kubenswrapper[4907]: I0313 14:09:48.042423 4907 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"770c882d9e9f9b2e9bd4f135e591952fa15b3898c42826aed8a75d953c9a875b"} pod="openshift-machine-config-operator/machine-config-daemon-hm56j" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 14:09:48 crc kubenswrapper[4907]: I0313 14:09:48.042499 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" containerID="cri-o://770c882d9e9f9b2e9bd4f135e591952fa15b3898c42826aed8a75d953c9a875b" gracePeriod=600 Mar 13 14:09:48 crc kubenswrapper[4907]: I0313 14:09:48.211610 4907 generic.go:334] "Generic (PLEG): container finished" podID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerID="770c882d9e9f9b2e9bd4f135e591952fa15b3898c42826aed8a75d953c9a875b" exitCode=0 Mar 13 14:09:48 crc kubenswrapper[4907]: I0313 14:09:48.211691 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerDied","Data":"770c882d9e9f9b2e9bd4f135e591952fa15b3898c42826aed8a75d953c9a875b"} Mar 13 14:09:49 crc kubenswrapper[4907]: E0313 14:09:49.396548 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-p257z" podUID="775414ea-5b94-4ab6-8b7c-fe7c80671ced" Mar 13 14:09:49 crc kubenswrapper[4907]: E0313 14:09:49.583645 4907 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Mar 13 14:09:49 crc kubenswrapper[4907]: E0313 14:09:49.583833 4907 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-2x9qs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-42kpk_openshift-marketplace(8adb3c1c-bacd-4cca-9796-7ca96624e9f6): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 13 14:09:49 crc kubenswrapper[4907]: E0313 14:09:49.585091 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-42kpk" podUID="8adb3c1c-bacd-4cca-9796-7ca96624e9f6" Mar 13 14:09:49 crc kubenswrapper[4907]: E0313 14:09:49.719222 4907 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Mar 13 14:09:49 crc kubenswrapper[4907]: E0313 14:09:49.719646 4907 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6q2rc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-gjdp7_openshift-marketplace(5ef16085-0f4d-435d-a559-20a33c237531): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 13 14:09:49 crc kubenswrapper[4907]: E0313 14:09:49.720924 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-gjdp7" podUID="5ef16085-0f4d-435d-a559-20a33c237531" Mar 13 14:09:49 crc kubenswrapper[4907]: I0313 14:09:49.905175 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Mar 13 14:09:49 crc kubenswrapper[4907]: I0313 14:09:49.905913 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Mar 13 14:09:49 crc kubenswrapper[4907]: I0313 14:09:49.906018 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 13 14:09:49 crc kubenswrapper[4907]: I0313 14:09:49.908504 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Mar 13 14:09:49 crc kubenswrapper[4907]: I0313 14:09:49.909117 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Mar 13 14:09:50 crc kubenswrapper[4907]: I0313 14:09:50.000760 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a87e7070-cc91-43dd-94bd-63075734c3a8-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"a87e7070-cc91-43dd-94bd-63075734c3a8\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 13 14:09:50 crc kubenswrapper[4907]: I0313 14:09:50.001090 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a87e7070-cc91-43dd-94bd-63075734c3a8-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"a87e7070-cc91-43dd-94bd-63075734c3a8\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 13 14:09:50 crc kubenswrapper[4907]: I0313 14:09:50.102197 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a87e7070-cc91-43dd-94bd-63075734c3a8-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"a87e7070-cc91-43dd-94bd-63075734c3a8\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 13 14:09:50 crc kubenswrapper[4907]: I0313 14:09:50.102250 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a87e7070-cc91-43dd-94bd-63075734c3a8-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"a87e7070-cc91-43dd-94bd-63075734c3a8\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 13 14:09:50 crc kubenswrapper[4907]: I0313 14:09:50.102417 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a87e7070-cc91-43dd-94bd-63075734c3a8-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"a87e7070-cc91-43dd-94bd-63075734c3a8\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 13 14:09:50 crc kubenswrapper[4907]: I0313 14:09:50.140711 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a87e7070-cc91-43dd-94bd-63075734c3a8-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"a87e7070-cc91-43dd-94bd-63075734c3a8\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 13 14:09:50 crc kubenswrapper[4907]: I0313 14:09:50.226354 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 13 14:09:51 crc kubenswrapper[4907]: I0313 14:09:51.814663 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-6f9b8d4bb9-c6vxw"] Mar 13 14:09:51 crc kubenswrapper[4907]: I0313 14:09:51.908875 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-65cb4586b7-dzz67"] Mar 13 14:09:53 crc kubenswrapper[4907]: E0313 14:09:53.646417 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-gjdp7" podUID="5ef16085-0f4d-435d-a559-20a33c237531" Mar 13 14:09:53 crc kubenswrapper[4907]: E0313 14:09:53.646829 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-42kpk" podUID="8adb3c1c-bacd-4cca-9796-7ca96624e9f6" Mar 13 14:09:53 crc kubenswrapper[4907]: I0313 14:09:53.671061 4907 scope.go:117] "RemoveContainer" containerID="c215281d0e491b45da32a8e52ffe10fe7add03de5080c14587c1be4c05f1e89d" Mar 13 14:09:53 crc kubenswrapper[4907]: E0313 14:09:53.813385 4907 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Mar 13 14:09:53 crc kubenswrapper[4907]: E0313 14:09:53.813866 4907 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-48ddb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-fmzmc_openshift-marketplace(ec3466c2-4954-459d-88d8-e8c6cbf309e8): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 13 14:09:53 crc kubenswrapper[4907]: E0313 14:09:53.816171 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-fmzmc" podUID="ec3466c2-4954-459d-88d8-e8c6cbf309e8" Mar 13 14:09:53 crc kubenswrapper[4907]: E0313 14:09:53.869069 4907 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Mar 13 14:09:53 crc kubenswrapper[4907]: E0313 14:09:53.869482 4907 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-msd8t,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-jk5nw_openshift-marketplace(d342a6a2-43dd-4d15-8aff-f60f9a3383bf): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 13 14:09:53 crc kubenswrapper[4907]: E0313 14:09:53.869754 4907 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Mar 13 14:09:53 crc kubenswrapper[4907]: E0313 14:09:53.869981 4907 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-k6s9c,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-v6msv_openshift-marketplace(84c2ada5-ceab-4327-802e-9ae459ac814d): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 13 14:09:53 crc kubenswrapper[4907]: E0313 14:09:53.871701 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-jk5nw" podUID="d342a6a2-43dd-4d15-8aff-f60f9a3383bf" Mar 13 14:09:53 crc kubenswrapper[4907]: E0313 14:09:53.871736 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-v6msv" podUID="84c2ada5-ceab-4327-802e-9ae459ac814d" Mar 13 14:09:53 crc kubenswrapper[4907]: E0313 14:09:53.898348 4907 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Mar 13 14:09:53 crc kubenswrapper[4907]: E0313 14:09:53.898587 4907 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-mcf55,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-vx5qd_openshift-marketplace(8779a406-c08e-40db-b25d-8c45e784e94d): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 13 14:09:53 crc kubenswrapper[4907]: E0313 14:09:53.900482 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-vx5qd" podUID="8779a406-c08e-40db-b25d-8c45e784e94d" Mar 13 14:09:54 crc kubenswrapper[4907]: I0313 14:09:54.036315 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-65cb4586b7-dzz67"] Mar 13 14:09:54 crc kubenswrapper[4907]: I0313 14:09:54.093196 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Mar 13 14:09:54 crc kubenswrapper[4907]: I0313 14:09:54.093963 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Mar 13 14:09:54 crc kubenswrapper[4907]: I0313 14:09:54.105142 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Mar 13 14:09:54 crc kubenswrapper[4907]: I0313 14:09:54.154776 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/b66475a4-2e1f-4b18-b565-3cc70bb3720f-var-lock\") pod \"installer-9-crc\" (UID: \"b66475a4-2e1f-4b18-b565-3cc70bb3720f\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 13 14:09:54 crc kubenswrapper[4907]: I0313 14:09:54.155068 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b66475a4-2e1f-4b18-b565-3cc70bb3720f-kube-api-access\") pod \"installer-9-crc\" (UID: \"b66475a4-2e1f-4b18-b565-3cc70bb3720f\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 13 14:09:54 crc kubenswrapper[4907]: I0313 14:09:54.155113 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b66475a4-2e1f-4b18-b565-3cc70bb3720f-kubelet-dir\") pod \"installer-9-crc\" (UID: \"b66475a4-2e1f-4b18-b565-3cc70bb3720f\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 13 14:09:54 crc kubenswrapper[4907]: I0313 14:09:54.250109 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerStarted","Data":"93da0fb3f9dc1b1da4e8f0430569f23f1e76f6777c8f829bb761fb1ef946f9de"} Mar 13 14:09:54 crc kubenswrapper[4907]: I0313 14:09:54.256530 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/b66475a4-2e1f-4b18-b565-3cc70bb3720f-var-lock\") pod \"installer-9-crc\" (UID: \"b66475a4-2e1f-4b18-b565-3cc70bb3720f\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 13 14:09:54 crc kubenswrapper[4907]: I0313 14:09:54.256592 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b66475a4-2e1f-4b18-b565-3cc70bb3720f-kube-api-access\") pod \"installer-9-crc\" (UID: \"b66475a4-2e1f-4b18-b565-3cc70bb3720f\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 13 14:09:54 crc kubenswrapper[4907]: I0313 14:09:54.256630 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b66475a4-2e1f-4b18-b565-3cc70bb3720f-kubelet-dir\") pod \"installer-9-crc\" (UID: \"b66475a4-2e1f-4b18-b565-3cc70bb3720f\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 13 14:09:54 crc kubenswrapper[4907]: I0313 14:09:54.256691 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b66475a4-2e1f-4b18-b565-3cc70bb3720f-kubelet-dir\") pod \"installer-9-crc\" (UID: \"b66475a4-2e1f-4b18-b565-3cc70bb3720f\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 13 14:09:54 crc kubenswrapper[4907]: I0313 14:09:54.256722 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/b66475a4-2e1f-4b18-b565-3cc70bb3720f-var-lock\") pod \"installer-9-crc\" (UID: \"b66475a4-2e1f-4b18-b565-3cc70bb3720f\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 13 14:09:54 crc kubenswrapper[4907]: I0313 14:09:54.289871 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-65cb4586b7-dzz67" event={"ID":"7b861e8a-0bd0-4b07-aa40-16b6310b19c1","Type":"ContainerStarted","Data":"c886e79340573e19a4b1d551a8638eac839a10aac09e4c6b2631e0a6b5fec143"} Mar 13 14:09:54 crc kubenswrapper[4907]: I0313 14:09:54.292372 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b66475a4-2e1f-4b18-b565-3cc70bb3720f-kube-api-access\") pod \"installer-9-crc\" (UID: \"b66475a4-2e1f-4b18-b565-3cc70bb3720f\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 13 14:09:54 crc kubenswrapper[4907]: I0313 14:09:54.304393 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Mar 13 14:09:54 crc kubenswrapper[4907]: E0313 14:09:54.304714 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-jk5nw" podUID="d342a6a2-43dd-4d15-8aff-f60f9a3383bf" Mar 13 14:09:54 crc kubenswrapper[4907]: E0313 14:09:54.305454 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-vx5qd" podUID="8779a406-c08e-40db-b25d-8c45e784e94d" Mar 13 14:09:54 crc kubenswrapper[4907]: E0313 14:09:54.305559 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-v6msv" podUID="84c2ada5-ceab-4327-802e-9ae459ac814d" Mar 13 14:09:54 crc kubenswrapper[4907]: I0313 14:09:54.306437 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-6f9b8d4bb9-c6vxw"] Mar 13 14:09:54 crc kubenswrapper[4907]: E0313 14:09:54.310344 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-fmzmc" podUID="ec3466c2-4954-459d-88d8-e8c6cbf309e8" Mar 13 14:09:54 crc kubenswrapper[4907]: I0313 14:09:54.496674 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Mar 13 14:09:54 crc kubenswrapper[4907]: I0313 14:09:54.933170 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Mar 13 14:09:55 crc kubenswrapper[4907]: I0313 14:09:55.295722 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-65cb4586b7-dzz67" event={"ID":"7b861e8a-0bd0-4b07-aa40-16b6310b19c1","Type":"ContainerStarted","Data":"2807d645f250d96f0d23cefa64683a737d2ab00c01dfb33a7e730110d4ac31cc"} Mar 13 14:09:55 crc kubenswrapper[4907]: I0313 14:09:55.296573 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-65cb4586b7-dzz67" Mar 13 14:09:55 crc kubenswrapper[4907]: I0313 14:09:55.295853 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-65cb4586b7-dzz67" podUID="7b861e8a-0bd0-4b07-aa40-16b6310b19c1" containerName="route-controller-manager" containerID="cri-o://2807d645f250d96f0d23cefa64683a737d2ab00c01dfb33a7e730110d4ac31cc" gracePeriod=30 Mar 13 14:09:55 crc kubenswrapper[4907]: I0313 14:09:55.299113 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"b66475a4-2e1f-4b18-b565-3cc70bb3720f","Type":"ContainerStarted","Data":"a68e3ade1d545da2c68a88db1974b5d105f932a3f0f6695c49e0be2ffa9a3a68"} Mar 13 14:09:55 crc kubenswrapper[4907]: I0313 14:09:55.305505 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-65cb4586b7-dzz67" Mar 13 14:09:55 crc kubenswrapper[4907]: I0313 14:09:55.306047 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"a87e7070-cc91-43dd-94bd-63075734c3a8","Type":"ContainerStarted","Data":"5c0a54d32578b5bee8b628f6a41ee0c4081016228d46c69bfb1cadd7ed5ecaa2"} Mar 13 14:09:55 crc kubenswrapper[4907]: I0313 14:09:55.306080 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"a87e7070-cc91-43dd-94bd-63075734c3a8","Type":"ContainerStarted","Data":"38a1f0c2ada60deba963941149abcc4b4702a13faf071269ca0e030d1e00d6d1"} Mar 13 14:09:55 crc kubenswrapper[4907]: I0313 14:09:55.311924 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6f9b8d4bb9-c6vxw" event={"ID":"12427c08-13c7-4bad-9dc3-815d65a8860f","Type":"ContainerStarted","Data":"9d0c6bc13b0d5110984df2dac333ed4ee030d297ea1490e3744a9ec0ee9366ae"} Mar 13 14:09:55 crc kubenswrapper[4907]: I0313 14:09:55.311996 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6f9b8d4bb9-c6vxw" event={"ID":"12427c08-13c7-4bad-9dc3-815d65a8860f","Type":"ContainerStarted","Data":"d8255210963d6d02f148a7ef4aea4444e4de09330c2b6ebd0f599bf30f80400d"} Mar 13 14:09:55 crc kubenswrapper[4907]: I0313 14:09:55.312029 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-6f9b8d4bb9-c6vxw" Mar 13 14:09:55 crc kubenswrapper[4907]: I0313 14:09:55.312107 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-6f9b8d4bb9-c6vxw" podUID="12427c08-13c7-4bad-9dc3-815d65a8860f" containerName="controller-manager" containerID="cri-o://9d0c6bc13b0d5110984df2dac333ed4ee030d297ea1490e3744a9ec0ee9366ae" gracePeriod=30 Mar 13 14:09:55 crc kubenswrapper[4907]: I0313 14:09:55.320862 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-65cb4586b7-dzz67" podStartSLOduration=24.320842382 podStartE2EDuration="24.320842382s" podCreationTimestamp="2026-03-13 14:09:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:09:55.317435899 +0000 UTC m=+294.217223608" watchObservedRunningTime="2026-03-13 14:09:55.320842382 +0000 UTC m=+294.220630071" Mar 13 14:09:55 crc kubenswrapper[4907]: I0313 14:09:55.321683 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-6f9b8d4bb9-c6vxw" Mar 13 14:09:55 crc kubenswrapper[4907]: I0313 14:09:55.340837 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-6f9b8d4bb9-c6vxw" podStartSLOduration=24.340812353 podStartE2EDuration="24.340812353s" podCreationTimestamp="2026-03-13 14:09:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:09:55.334752789 +0000 UTC m=+294.234540478" watchObservedRunningTime="2026-03-13 14:09:55.340812353 +0000 UTC m=+294.240600052" Mar 13 14:09:55 crc kubenswrapper[4907]: I0313 14:09:55.347965 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-9-crc" podStartSLOduration=6.347944216 podStartE2EDuration="6.347944216s" podCreationTimestamp="2026-03-13 14:09:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:09:55.346045645 +0000 UTC m=+294.245833334" watchObservedRunningTime="2026-03-13 14:09:55.347944216 +0000 UTC m=+294.247731905" Mar 13 14:09:55 crc kubenswrapper[4907]: I0313 14:09:55.627852 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-65cb4586b7-dzz67" Mar 13 14:09:55 crc kubenswrapper[4907]: I0313 14:09:55.657853 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6f8c9bf8d5-trl8q"] Mar 13 14:09:55 crc kubenswrapper[4907]: E0313 14:09:55.658134 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b861e8a-0bd0-4b07-aa40-16b6310b19c1" containerName="route-controller-manager" Mar 13 14:09:55 crc kubenswrapper[4907]: I0313 14:09:55.658151 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b861e8a-0bd0-4b07-aa40-16b6310b19c1" containerName="route-controller-manager" Mar 13 14:09:55 crc kubenswrapper[4907]: I0313 14:09:55.658258 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b861e8a-0bd0-4b07-aa40-16b6310b19c1" containerName="route-controller-manager" Mar 13 14:09:55 crc kubenswrapper[4907]: I0313 14:09:55.658762 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6f8c9bf8d5-trl8q" Mar 13 14:09:55 crc kubenswrapper[4907]: I0313 14:09:55.677802 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7b861e8a-0bd0-4b07-aa40-16b6310b19c1-client-ca\") pod \"7b861e8a-0bd0-4b07-aa40-16b6310b19c1\" (UID: \"7b861e8a-0bd0-4b07-aa40-16b6310b19c1\") " Mar 13 14:09:55 crc kubenswrapper[4907]: I0313 14:09:55.678279 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7b861e8a-0bd0-4b07-aa40-16b6310b19c1-serving-cert\") pod \"7b861e8a-0bd0-4b07-aa40-16b6310b19c1\" (UID: \"7b861e8a-0bd0-4b07-aa40-16b6310b19c1\") " Mar 13 14:09:55 crc kubenswrapper[4907]: I0313 14:09:55.678335 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7b861e8a-0bd0-4b07-aa40-16b6310b19c1-config\") pod \"7b861e8a-0bd0-4b07-aa40-16b6310b19c1\" (UID: \"7b861e8a-0bd0-4b07-aa40-16b6310b19c1\") " Mar 13 14:09:55 crc kubenswrapper[4907]: I0313 14:09:55.678370 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fxxvw\" (UniqueName: \"kubernetes.io/projected/7b861e8a-0bd0-4b07-aa40-16b6310b19c1-kube-api-access-fxxvw\") pod \"7b861e8a-0bd0-4b07-aa40-16b6310b19c1\" (UID: \"7b861e8a-0bd0-4b07-aa40-16b6310b19c1\") " Mar 13 14:09:55 crc kubenswrapper[4907]: I0313 14:09:55.681861 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hhgz5\" (UniqueName: \"kubernetes.io/projected/7fb34c61-0980-40fc-9c6a-48af425786d0-kube-api-access-hhgz5\") pod \"route-controller-manager-6f8c9bf8d5-trl8q\" (UID: \"7fb34c61-0980-40fc-9c6a-48af425786d0\") " pod="openshift-route-controller-manager/route-controller-manager-6f8c9bf8d5-trl8q" Mar 13 14:09:55 crc kubenswrapper[4907]: I0313 14:09:55.682059 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7fb34c61-0980-40fc-9c6a-48af425786d0-config\") pod \"route-controller-manager-6f8c9bf8d5-trl8q\" (UID: \"7fb34c61-0980-40fc-9c6a-48af425786d0\") " pod="openshift-route-controller-manager/route-controller-manager-6f8c9bf8d5-trl8q" Mar 13 14:09:55 crc kubenswrapper[4907]: I0313 14:09:55.682109 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7fb34c61-0980-40fc-9c6a-48af425786d0-serving-cert\") pod \"route-controller-manager-6f8c9bf8d5-trl8q\" (UID: \"7fb34c61-0980-40fc-9c6a-48af425786d0\") " pod="openshift-route-controller-manager/route-controller-manager-6f8c9bf8d5-trl8q" Mar 13 14:09:55 crc kubenswrapper[4907]: I0313 14:09:55.682219 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7fb34c61-0980-40fc-9c6a-48af425786d0-client-ca\") pod \"route-controller-manager-6f8c9bf8d5-trl8q\" (UID: \"7fb34c61-0980-40fc-9c6a-48af425786d0\") " pod="openshift-route-controller-manager/route-controller-manager-6f8c9bf8d5-trl8q" Mar 13 14:09:55 crc kubenswrapper[4907]: I0313 14:09:55.683124 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7b861e8a-0bd0-4b07-aa40-16b6310b19c1-config" (OuterVolumeSpecName: "config") pod "7b861e8a-0bd0-4b07-aa40-16b6310b19c1" (UID: "7b861e8a-0bd0-4b07-aa40-16b6310b19c1"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:09:55 crc kubenswrapper[4907]: I0313 14:09:55.683772 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7b861e8a-0bd0-4b07-aa40-16b6310b19c1-client-ca" (OuterVolumeSpecName: "client-ca") pod "7b861e8a-0bd0-4b07-aa40-16b6310b19c1" (UID: "7b861e8a-0bd0-4b07-aa40-16b6310b19c1"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:09:55 crc kubenswrapper[4907]: I0313 14:09:55.689577 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b861e8a-0bd0-4b07-aa40-16b6310b19c1-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7b861e8a-0bd0-4b07-aa40-16b6310b19c1" (UID: "7b861e8a-0bd0-4b07-aa40-16b6310b19c1"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:09:55 crc kubenswrapper[4907]: I0313 14:09:55.689733 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7b861e8a-0bd0-4b07-aa40-16b6310b19c1-kube-api-access-fxxvw" (OuterVolumeSpecName: "kube-api-access-fxxvw") pod "7b861e8a-0bd0-4b07-aa40-16b6310b19c1" (UID: "7b861e8a-0bd0-4b07-aa40-16b6310b19c1"). InnerVolumeSpecName "kube-api-access-fxxvw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:09:55 crc kubenswrapper[4907]: I0313 14:09:55.712196 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6f8c9bf8d5-trl8q"] Mar 13 14:09:55 crc kubenswrapper[4907]: I0313 14:09:55.724597 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6f9b8d4bb9-c6vxw" Mar 13 14:09:55 crc kubenswrapper[4907]: I0313 14:09:55.782572 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/12427c08-13c7-4bad-9dc3-815d65a8860f-proxy-ca-bundles\") pod \"12427c08-13c7-4bad-9dc3-815d65a8860f\" (UID: \"12427c08-13c7-4bad-9dc3-815d65a8860f\") " Mar 13 14:09:55 crc kubenswrapper[4907]: I0313 14:09:55.782627 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/12427c08-13c7-4bad-9dc3-815d65a8860f-config\") pod \"12427c08-13c7-4bad-9dc3-815d65a8860f\" (UID: \"12427c08-13c7-4bad-9dc3-815d65a8860f\") " Mar 13 14:09:55 crc kubenswrapper[4907]: I0313 14:09:55.782667 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mbr8t\" (UniqueName: \"kubernetes.io/projected/12427c08-13c7-4bad-9dc3-815d65a8860f-kube-api-access-mbr8t\") pod \"12427c08-13c7-4bad-9dc3-815d65a8860f\" (UID: \"12427c08-13c7-4bad-9dc3-815d65a8860f\") " Mar 13 14:09:55 crc kubenswrapper[4907]: I0313 14:09:55.782709 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/12427c08-13c7-4bad-9dc3-815d65a8860f-client-ca\") pod \"12427c08-13c7-4bad-9dc3-815d65a8860f\" (UID: \"12427c08-13c7-4bad-9dc3-815d65a8860f\") " Mar 13 14:09:55 crc kubenswrapper[4907]: I0313 14:09:55.782731 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/12427c08-13c7-4bad-9dc3-815d65a8860f-serving-cert\") pod \"12427c08-13c7-4bad-9dc3-815d65a8860f\" (UID: \"12427c08-13c7-4bad-9dc3-815d65a8860f\") " Mar 13 14:09:55 crc kubenswrapper[4907]: I0313 14:09:55.782862 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7fb34c61-0980-40fc-9c6a-48af425786d0-config\") pod \"route-controller-manager-6f8c9bf8d5-trl8q\" (UID: \"7fb34c61-0980-40fc-9c6a-48af425786d0\") " pod="openshift-route-controller-manager/route-controller-manager-6f8c9bf8d5-trl8q" Mar 13 14:09:55 crc kubenswrapper[4907]: I0313 14:09:55.782909 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7fb34c61-0980-40fc-9c6a-48af425786d0-serving-cert\") pod \"route-controller-manager-6f8c9bf8d5-trl8q\" (UID: \"7fb34c61-0980-40fc-9c6a-48af425786d0\") " pod="openshift-route-controller-manager/route-controller-manager-6f8c9bf8d5-trl8q" Mar 13 14:09:55 crc kubenswrapper[4907]: I0313 14:09:55.782959 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7fb34c61-0980-40fc-9c6a-48af425786d0-client-ca\") pod \"route-controller-manager-6f8c9bf8d5-trl8q\" (UID: \"7fb34c61-0980-40fc-9c6a-48af425786d0\") " pod="openshift-route-controller-manager/route-controller-manager-6f8c9bf8d5-trl8q" Mar 13 14:09:55 crc kubenswrapper[4907]: I0313 14:09:55.783140 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hhgz5\" (UniqueName: \"kubernetes.io/projected/7fb34c61-0980-40fc-9c6a-48af425786d0-kube-api-access-hhgz5\") pod \"route-controller-manager-6f8c9bf8d5-trl8q\" (UID: \"7fb34c61-0980-40fc-9c6a-48af425786d0\") " pod="openshift-route-controller-manager/route-controller-manager-6f8c9bf8d5-trl8q" Mar 13 14:09:55 crc kubenswrapper[4907]: I0313 14:09:55.783364 4907 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7b861e8a-0bd0-4b07-aa40-16b6310b19c1-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 14:09:55 crc kubenswrapper[4907]: I0313 14:09:55.783385 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7b861e8a-0bd0-4b07-aa40-16b6310b19c1-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:09:55 crc kubenswrapper[4907]: I0313 14:09:55.783394 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fxxvw\" (UniqueName: \"kubernetes.io/projected/7b861e8a-0bd0-4b07-aa40-16b6310b19c1-kube-api-access-fxxvw\") on node \"crc\" DevicePath \"\"" Mar 13 14:09:55 crc kubenswrapper[4907]: I0313 14:09:55.783406 4907 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7b861e8a-0bd0-4b07-aa40-16b6310b19c1-client-ca\") on node \"crc\" DevicePath \"\"" Mar 13 14:09:55 crc kubenswrapper[4907]: I0313 14:09:55.783462 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/12427c08-13c7-4bad-9dc3-815d65a8860f-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "12427c08-13c7-4bad-9dc3-815d65a8860f" (UID: "12427c08-13c7-4bad-9dc3-815d65a8860f"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:09:55 crc kubenswrapper[4907]: I0313 14:09:55.783519 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/12427c08-13c7-4bad-9dc3-815d65a8860f-client-ca" (OuterVolumeSpecName: "client-ca") pod "12427c08-13c7-4bad-9dc3-815d65a8860f" (UID: "12427c08-13c7-4bad-9dc3-815d65a8860f"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:09:55 crc kubenswrapper[4907]: I0313 14:09:55.783793 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7fb34c61-0980-40fc-9c6a-48af425786d0-client-ca\") pod \"route-controller-manager-6f8c9bf8d5-trl8q\" (UID: \"7fb34c61-0980-40fc-9c6a-48af425786d0\") " pod="openshift-route-controller-manager/route-controller-manager-6f8c9bf8d5-trl8q" Mar 13 14:09:55 crc kubenswrapper[4907]: I0313 14:09:55.784174 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7fb34c61-0980-40fc-9c6a-48af425786d0-config\") pod \"route-controller-manager-6f8c9bf8d5-trl8q\" (UID: \"7fb34c61-0980-40fc-9c6a-48af425786d0\") " pod="openshift-route-controller-manager/route-controller-manager-6f8c9bf8d5-trl8q" Mar 13 14:09:55 crc kubenswrapper[4907]: I0313 14:09:55.784200 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/12427c08-13c7-4bad-9dc3-815d65a8860f-config" (OuterVolumeSpecName: "config") pod "12427c08-13c7-4bad-9dc3-815d65a8860f" (UID: "12427c08-13c7-4bad-9dc3-815d65a8860f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:09:55 crc kubenswrapper[4907]: I0313 14:09:55.787352 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7fb34c61-0980-40fc-9c6a-48af425786d0-serving-cert\") pod \"route-controller-manager-6f8c9bf8d5-trl8q\" (UID: \"7fb34c61-0980-40fc-9c6a-48af425786d0\") " pod="openshift-route-controller-manager/route-controller-manager-6f8c9bf8d5-trl8q" Mar 13 14:09:55 crc kubenswrapper[4907]: I0313 14:09:55.796418 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/12427c08-13c7-4bad-9dc3-815d65a8860f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "12427c08-13c7-4bad-9dc3-815d65a8860f" (UID: "12427c08-13c7-4bad-9dc3-815d65a8860f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:09:55 crc kubenswrapper[4907]: I0313 14:09:55.797107 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/12427c08-13c7-4bad-9dc3-815d65a8860f-kube-api-access-mbr8t" (OuterVolumeSpecName: "kube-api-access-mbr8t") pod "12427c08-13c7-4bad-9dc3-815d65a8860f" (UID: "12427c08-13c7-4bad-9dc3-815d65a8860f"). InnerVolumeSpecName "kube-api-access-mbr8t". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:09:55 crc kubenswrapper[4907]: I0313 14:09:55.804195 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hhgz5\" (UniqueName: \"kubernetes.io/projected/7fb34c61-0980-40fc-9c6a-48af425786d0-kube-api-access-hhgz5\") pod \"route-controller-manager-6f8c9bf8d5-trl8q\" (UID: \"7fb34c61-0980-40fc-9c6a-48af425786d0\") " pod="openshift-route-controller-manager/route-controller-manager-6f8c9bf8d5-trl8q" Mar 13 14:09:55 crc kubenswrapper[4907]: I0313 14:09:55.884983 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mbr8t\" (UniqueName: \"kubernetes.io/projected/12427c08-13c7-4bad-9dc3-815d65a8860f-kube-api-access-mbr8t\") on node \"crc\" DevicePath \"\"" Mar 13 14:09:55 crc kubenswrapper[4907]: I0313 14:09:55.885025 4907 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/12427c08-13c7-4bad-9dc3-815d65a8860f-client-ca\") on node \"crc\" DevicePath \"\"" Mar 13 14:09:55 crc kubenswrapper[4907]: I0313 14:09:55.885037 4907 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/12427c08-13c7-4bad-9dc3-815d65a8860f-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 14:09:55 crc kubenswrapper[4907]: I0313 14:09:55.885046 4907 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/12427c08-13c7-4bad-9dc3-815d65a8860f-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 13 14:09:55 crc kubenswrapper[4907]: I0313 14:09:55.885054 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/12427c08-13c7-4bad-9dc3-815d65a8860f-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:09:55 crc kubenswrapper[4907]: I0313 14:09:55.972904 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6f8c9bf8d5-trl8q" Mar 13 14:09:56 crc kubenswrapper[4907]: I0313 14:09:56.321446 4907 generic.go:334] "Generic (PLEG): container finished" podID="a87e7070-cc91-43dd-94bd-63075734c3a8" containerID="5c0a54d32578b5bee8b628f6a41ee0c4081016228d46c69bfb1cadd7ed5ecaa2" exitCode=0 Mar 13 14:09:56 crc kubenswrapper[4907]: I0313 14:09:56.321525 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"a87e7070-cc91-43dd-94bd-63075734c3a8","Type":"ContainerDied","Data":"5c0a54d32578b5bee8b628f6a41ee0c4081016228d46c69bfb1cadd7ed5ecaa2"} Mar 13 14:09:56 crc kubenswrapper[4907]: I0313 14:09:56.325030 4907 generic.go:334] "Generic (PLEG): container finished" podID="12427c08-13c7-4bad-9dc3-815d65a8860f" containerID="9d0c6bc13b0d5110984df2dac333ed4ee030d297ea1490e3744a9ec0ee9366ae" exitCode=0 Mar 13 14:09:56 crc kubenswrapper[4907]: I0313 14:09:56.325086 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6f9b8d4bb9-c6vxw" event={"ID":"12427c08-13c7-4bad-9dc3-815d65a8860f","Type":"ContainerDied","Data":"9d0c6bc13b0d5110984df2dac333ed4ee030d297ea1490e3744a9ec0ee9366ae"} Mar 13 14:09:56 crc kubenswrapper[4907]: I0313 14:09:56.325106 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6f9b8d4bb9-c6vxw" event={"ID":"12427c08-13c7-4bad-9dc3-815d65a8860f","Type":"ContainerDied","Data":"d8255210963d6d02f148a7ef4aea4444e4de09330c2b6ebd0f599bf30f80400d"} Mar 13 14:09:56 crc kubenswrapper[4907]: I0313 14:09:56.325123 4907 scope.go:117] "RemoveContainer" containerID="9d0c6bc13b0d5110984df2dac333ed4ee030d297ea1490e3744a9ec0ee9366ae" Mar 13 14:09:56 crc kubenswrapper[4907]: I0313 14:09:56.325117 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6f9b8d4bb9-c6vxw" Mar 13 14:09:56 crc kubenswrapper[4907]: I0313 14:09:56.328025 4907 generic.go:334] "Generic (PLEG): container finished" podID="7b861e8a-0bd0-4b07-aa40-16b6310b19c1" containerID="2807d645f250d96f0d23cefa64683a737d2ab00c01dfb33a7e730110d4ac31cc" exitCode=0 Mar 13 14:09:56 crc kubenswrapper[4907]: I0313 14:09:56.328155 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-65cb4586b7-dzz67" Mar 13 14:09:56 crc kubenswrapper[4907]: I0313 14:09:56.328209 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-65cb4586b7-dzz67" event={"ID":"7b861e8a-0bd0-4b07-aa40-16b6310b19c1","Type":"ContainerDied","Data":"2807d645f250d96f0d23cefa64683a737d2ab00c01dfb33a7e730110d4ac31cc"} Mar 13 14:09:56 crc kubenswrapper[4907]: I0313 14:09:56.328233 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-65cb4586b7-dzz67" event={"ID":"7b861e8a-0bd0-4b07-aa40-16b6310b19c1","Type":"ContainerDied","Data":"c886e79340573e19a4b1d551a8638eac839a10aac09e4c6b2631e0a6b5fec143"} Mar 13 14:09:56 crc kubenswrapper[4907]: I0313 14:09:56.331366 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"b66475a4-2e1f-4b18-b565-3cc70bb3720f","Type":"ContainerStarted","Data":"c818da15368a295079d572ed81796486cb0c9f878bd84d437fa33e0b3e3169f6"} Mar 13 14:09:56 crc kubenswrapper[4907]: I0313 14:09:56.357977 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-6f9b8d4bb9-c6vxw"] Mar 13 14:09:56 crc kubenswrapper[4907]: I0313 14:09:56.361571 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-6f9b8d4bb9-c6vxw"] Mar 13 14:09:56 crc kubenswrapper[4907]: I0313 14:09:56.365419 4907 scope.go:117] "RemoveContainer" containerID="9d0c6bc13b0d5110984df2dac333ed4ee030d297ea1490e3744a9ec0ee9366ae" Mar 13 14:09:56 crc kubenswrapper[4907]: E0313 14:09:56.365901 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9d0c6bc13b0d5110984df2dac333ed4ee030d297ea1490e3744a9ec0ee9366ae\": container with ID starting with 9d0c6bc13b0d5110984df2dac333ed4ee030d297ea1490e3744a9ec0ee9366ae not found: ID does not exist" containerID="9d0c6bc13b0d5110984df2dac333ed4ee030d297ea1490e3744a9ec0ee9366ae" Mar 13 14:09:56 crc kubenswrapper[4907]: I0313 14:09:56.365940 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9d0c6bc13b0d5110984df2dac333ed4ee030d297ea1490e3744a9ec0ee9366ae"} err="failed to get container status \"9d0c6bc13b0d5110984df2dac333ed4ee030d297ea1490e3744a9ec0ee9366ae\": rpc error: code = NotFound desc = could not find container \"9d0c6bc13b0d5110984df2dac333ed4ee030d297ea1490e3744a9ec0ee9366ae\": container with ID starting with 9d0c6bc13b0d5110984df2dac333ed4ee030d297ea1490e3744a9ec0ee9366ae not found: ID does not exist" Mar 13 14:09:56 crc kubenswrapper[4907]: I0313 14:09:56.365970 4907 scope.go:117] "RemoveContainer" containerID="2807d645f250d96f0d23cefa64683a737d2ab00c01dfb33a7e730110d4ac31cc" Mar 13 14:09:56 crc kubenswrapper[4907]: I0313 14:09:56.382851 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=2.38282815 podStartE2EDuration="2.38282815s" podCreationTimestamp="2026-03-13 14:09:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:09:56.374638169 +0000 UTC m=+295.274425868" watchObservedRunningTime="2026-03-13 14:09:56.38282815 +0000 UTC m=+295.282615839" Mar 13 14:09:56 crc kubenswrapper[4907]: I0313 14:09:56.386237 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-65cb4586b7-dzz67"] Mar 13 14:09:56 crc kubenswrapper[4907]: I0313 14:09:56.392773 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-65cb4586b7-dzz67"] Mar 13 14:09:56 crc kubenswrapper[4907]: I0313 14:09:56.398960 4907 scope.go:117] "RemoveContainer" containerID="2807d645f250d96f0d23cefa64683a737d2ab00c01dfb33a7e730110d4ac31cc" Mar 13 14:09:56 crc kubenswrapper[4907]: E0313 14:09:56.400165 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2807d645f250d96f0d23cefa64683a737d2ab00c01dfb33a7e730110d4ac31cc\": container with ID starting with 2807d645f250d96f0d23cefa64683a737d2ab00c01dfb33a7e730110d4ac31cc not found: ID does not exist" containerID="2807d645f250d96f0d23cefa64683a737d2ab00c01dfb33a7e730110d4ac31cc" Mar 13 14:09:56 crc kubenswrapper[4907]: I0313 14:09:56.400223 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2807d645f250d96f0d23cefa64683a737d2ab00c01dfb33a7e730110d4ac31cc"} err="failed to get container status \"2807d645f250d96f0d23cefa64683a737d2ab00c01dfb33a7e730110d4ac31cc\": rpc error: code = NotFound desc = could not find container \"2807d645f250d96f0d23cefa64683a737d2ab00c01dfb33a7e730110d4ac31cc\": container with ID starting with 2807d645f250d96f0d23cefa64683a737d2ab00c01dfb33a7e730110d4ac31cc not found: ID does not exist" Mar 13 14:09:56 crc kubenswrapper[4907]: I0313 14:09:56.420123 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6f8c9bf8d5-trl8q"] Mar 13 14:09:56 crc kubenswrapper[4907]: W0313 14:09:56.422855 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7fb34c61_0980_40fc_9c6a_48af425786d0.slice/crio-54f13a73875be71a32e36bb0c1839a08275e7ece063f4b0690961fb22cc8775b WatchSource:0}: Error finding container 54f13a73875be71a32e36bb0c1839a08275e7ece063f4b0690961fb22cc8775b: Status 404 returned error can't find the container with id 54f13a73875be71a32e36bb0c1839a08275e7ece063f4b0690961fb22cc8775b Mar 13 14:09:57 crc kubenswrapper[4907]: I0313 14:09:57.342437 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6f8c9bf8d5-trl8q" event={"ID":"7fb34c61-0980-40fc-9c6a-48af425786d0","Type":"ContainerStarted","Data":"6b3f156649bd3aebfecf7ea45641ce94397a3dcd4edc7dc31ce59e515d93132d"} Mar 13 14:09:57 crc kubenswrapper[4907]: I0313 14:09:57.342484 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6f8c9bf8d5-trl8q" event={"ID":"7fb34c61-0980-40fc-9c6a-48af425786d0","Type":"ContainerStarted","Data":"54f13a73875be71a32e36bb0c1839a08275e7ece063f4b0690961fb22cc8775b"} Mar 13 14:09:57 crc kubenswrapper[4907]: I0313 14:09:57.362004 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6f8c9bf8d5-trl8q" podStartSLOduration=6.361957104 podStartE2EDuration="6.361957104s" podCreationTimestamp="2026-03-13 14:09:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:09:57.356801764 +0000 UTC m=+296.256589453" watchObservedRunningTime="2026-03-13 14:09:57.361957104 +0000 UTC m=+296.261744793" Mar 13 14:09:57 crc kubenswrapper[4907]: I0313 14:09:57.584386 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 13 14:09:57 crc kubenswrapper[4907]: I0313 14:09:57.607411 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a87e7070-cc91-43dd-94bd-63075734c3a8-kube-api-access\") pod \"a87e7070-cc91-43dd-94bd-63075734c3a8\" (UID: \"a87e7070-cc91-43dd-94bd-63075734c3a8\") " Mar 13 14:09:57 crc kubenswrapper[4907]: I0313 14:09:57.607561 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a87e7070-cc91-43dd-94bd-63075734c3a8-kubelet-dir\") pod \"a87e7070-cc91-43dd-94bd-63075734c3a8\" (UID: \"a87e7070-cc91-43dd-94bd-63075734c3a8\") " Mar 13 14:09:57 crc kubenswrapper[4907]: I0313 14:09:57.607837 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a87e7070-cc91-43dd-94bd-63075734c3a8-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "a87e7070-cc91-43dd-94bd-63075734c3a8" (UID: "a87e7070-cc91-43dd-94bd-63075734c3a8"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 14:09:57 crc kubenswrapper[4907]: I0313 14:09:57.625262 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a87e7070-cc91-43dd-94bd-63075734c3a8-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "a87e7070-cc91-43dd-94bd-63075734c3a8" (UID: "a87e7070-cc91-43dd-94bd-63075734c3a8"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:09:57 crc kubenswrapper[4907]: I0313 14:09:57.708919 4907 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a87e7070-cc91-43dd-94bd-63075734c3a8-kubelet-dir\") on node \"crc\" DevicePath \"\"" Mar 13 14:09:57 crc kubenswrapper[4907]: I0313 14:09:57.708962 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a87e7070-cc91-43dd-94bd-63075734c3a8-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 13 14:09:57 crc kubenswrapper[4907]: I0313 14:09:57.791044 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="12427c08-13c7-4bad-9dc3-815d65a8860f" path="/var/lib/kubelet/pods/12427c08-13c7-4bad-9dc3-815d65a8860f/volumes" Mar 13 14:09:57 crc kubenswrapper[4907]: I0313 14:09:57.791939 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7b861e8a-0bd0-4b07-aa40-16b6310b19c1" path="/var/lib/kubelet/pods/7b861e8a-0bd0-4b07-aa40-16b6310b19c1/volumes" Mar 13 14:09:58 crc kubenswrapper[4907]: I0313 14:09:58.351629 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 13 14:09:58 crc kubenswrapper[4907]: I0313 14:09:58.351975 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"a87e7070-cc91-43dd-94bd-63075734c3a8","Type":"ContainerDied","Data":"38a1f0c2ada60deba963941149abcc4b4702a13faf071269ca0e030d1e00d6d1"} Mar 13 14:09:58 crc kubenswrapper[4907]: I0313 14:09:58.352004 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="38a1f0c2ada60deba963941149abcc4b4702a13faf071269ca0e030d1e00d6d1" Mar 13 14:09:58 crc kubenswrapper[4907]: I0313 14:09:58.352129 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6f8c9bf8d5-trl8q" Mar 13 14:09:58 crc kubenswrapper[4907]: I0313 14:09:58.357165 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6f8c9bf8d5-trl8q" Mar 13 14:09:58 crc kubenswrapper[4907]: I0313 14:09:58.536617 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-87cc55c9b-nlr2j"] Mar 13 14:09:58 crc kubenswrapper[4907]: E0313 14:09:58.536968 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a87e7070-cc91-43dd-94bd-63075734c3a8" containerName="pruner" Mar 13 14:09:58 crc kubenswrapper[4907]: I0313 14:09:58.536984 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="a87e7070-cc91-43dd-94bd-63075734c3a8" containerName="pruner" Mar 13 14:09:58 crc kubenswrapper[4907]: E0313 14:09:58.536999 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12427c08-13c7-4bad-9dc3-815d65a8860f" containerName="controller-manager" Mar 13 14:09:58 crc kubenswrapper[4907]: I0313 14:09:58.537005 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="12427c08-13c7-4bad-9dc3-815d65a8860f" containerName="controller-manager" Mar 13 14:09:58 crc kubenswrapper[4907]: I0313 14:09:58.537111 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="a87e7070-cc91-43dd-94bd-63075734c3a8" containerName="pruner" Mar 13 14:09:58 crc kubenswrapper[4907]: I0313 14:09:58.537127 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="12427c08-13c7-4bad-9dc3-815d65a8860f" containerName="controller-manager" Mar 13 14:09:58 crc kubenswrapper[4907]: I0313 14:09:58.537541 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-87cc55c9b-nlr2j" Mar 13 14:09:58 crc kubenswrapper[4907]: I0313 14:09:58.539720 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 13 14:09:58 crc kubenswrapper[4907]: I0313 14:09:58.540340 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 13 14:09:58 crc kubenswrapper[4907]: I0313 14:09:58.540612 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 13 14:09:58 crc kubenswrapper[4907]: I0313 14:09:58.540958 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 13 14:09:58 crc kubenswrapper[4907]: I0313 14:09:58.541031 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 13 14:09:58 crc kubenswrapper[4907]: I0313 14:09:58.541331 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 13 14:09:58 crc kubenswrapper[4907]: I0313 14:09:58.547340 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 13 14:09:58 crc kubenswrapper[4907]: I0313 14:09:58.550508 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-87cc55c9b-nlr2j"] Mar 13 14:09:58 crc kubenswrapper[4907]: I0313 14:09:58.621495 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5gj4x\" (UniqueName: \"kubernetes.io/projected/f341a2f4-0f47-47cb-b5e6-facbf98694b9-kube-api-access-5gj4x\") pod \"controller-manager-87cc55c9b-nlr2j\" (UID: \"f341a2f4-0f47-47cb-b5e6-facbf98694b9\") " pod="openshift-controller-manager/controller-manager-87cc55c9b-nlr2j" Mar 13 14:09:58 crc kubenswrapper[4907]: I0313 14:09:58.621631 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f341a2f4-0f47-47cb-b5e6-facbf98694b9-proxy-ca-bundles\") pod \"controller-manager-87cc55c9b-nlr2j\" (UID: \"f341a2f4-0f47-47cb-b5e6-facbf98694b9\") " pod="openshift-controller-manager/controller-manager-87cc55c9b-nlr2j" Mar 13 14:09:58 crc kubenswrapper[4907]: I0313 14:09:58.621662 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f341a2f4-0f47-47cb-b5e6-facbf98694b9-serving-cert\") pod \"controller-manager-87cc55c9b-nlr2j\" (UID: \"f341a2f4-0f47-47cb-b5e6-facbf98694b9\") " pod="openshift-controller-manager/controller-manager-87cc55c9b-nlr2j" Mar 13 14:09:58 crc kubenswrapper[4907]: I0313 14:09:58.621682 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f341a2f4-0f47-47cb-b5e6-facbf98694b9-client-ca\") pod \"controller-manager-87cc55c9b-nlr2j\" (UID: \"f341a2f4-0f47-47cb-b5e6-facbf98694b9\") " pod="openshift-controller-manager/controller-manager-87cc55c9b-nlr2j" Mar 13 14:09:58 crc kubenswrapper[4907]: I0313 14:09:58.621724 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f341a2f4-0f47-47cb-b5e6-facbf98694b9-config\") pod \"controller-manager-87cc55c9b-nlr2j\" (UID: \"f341a2f4-0f47-47cb-b5e6-facbf98694b9\") " pod="openshift-controller-manager/controller-manager-87cc55c9b-nlr2j" Mar 13 14:09:58 crc kubenswrapper[4907]: I0313 14:09:58.723106 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5gj4x\" (UniqueName: \"kubernetes.io/projected/f341a2f4-0f47-47cb-b5e6-facbf98694b9-kube-api-access-5gj4x\") pod \"controller-manager-87cc55c9b-nlr2j\" (UID: \"f341a2f4-0f47-47cb-b5e6-facbf98694b9\") " pod="openshift-controller-manager/controller-manager-87cc55c9b-nlr2j" Mar 13 14:09:58 crc kubenswrapper[4907]: I0313 14:09:58.723204 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f341a2f4-0f47-47cb-b5e6-facbf98694b9-proxy-ca-bundles\") pod \"controller-manager-87cc55c9b-nlr2j\" (UID: \"f341a2f4-0f47-47cb-b5e6-facbf98694b9\") " pod="openshift-controller-manager/controller-manager-87cc55c9b-nlr2j" Mar 13 14:09:58 crc kubenswrapper[4907]: I0313 14:09:58.723238 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f341a2f4-0f47-47cb-b5e6-facbf98694b9-serving-cert\") pod \"controller-manager-87cc55c9b-nlr2j\" (UID: \"f341a2f4-0f47-47cb-b5e6-facbf98694b9\") " pod="openshift-controller-manager/controller-manager-87cc55c9b-nlr2j" Mar 13 14:09:58 crc kubenswrapper[4907]: I0313 14:09:58.723260 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f341a2f4-0f47-47cb-b5e6-facbf98694b9-client-ca\") pod \"controller-manager-87cc55c9b-nlr2j\" (UID: \"f341a2f4-0f47-47cb-b5e6-facbf98694b9\") " pod="openshift-controller-manager/controller-manager-87cc55c9b-nlr2j" Mar 13 14:09:58 crc kubenswrapper[4907]: I0313 14:09:58.723317 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f341a2f4-0f47-47cb-b5e6-facbf98694b9-config\") pod \"controller-manager-87cc55c9b-nlr2j\" (UID: \"f341a2f4-0f47-47cb-b5e6-facbf98694b9\") " pod="openshift-controller-manager/controller-manager-87cc55c9b-nlr2j" Mar 13 14:09:58 crc kubenswrapper[4907]: I0313 14:09:58.724376 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f341a2f4-0f47-47cb-b5e6-facbf98694b9-client-ca\") pod \"controller-manager-87cc55c9b-nlr2j\" (UID: \"f341a2f4-0f47-47cb-b5e6-facbf98694b9\") " pod="openshift-controller-manager/controller-manager-87cc55c9b-nlr2j" Mar 13 14:09:58 crc kubenswrapper[4907]: I0313 14:09:58.724967 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f341a2f4-0f47-47cb-b5e6-facbf98694b9-config\") pod \"controller-manager-87cc55c9b-nlr2j\" (UID: \"f341a2f4-0f47-47cb-b5e6-facbf98694b9\") " pod="openshift-controller-manager/controller-manager-87cc55c9b-nlr2j" Mar 13 14:09:58 crc kubenswrapper[4907]: I0313 14:09:58.724964 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f341a2f4-0f47-47cb-b5e6-facbf98694b9-proxy-ca-bundles\") pod \"controller-manager-87cc55c9b-nlr2j\" (UID: \"f341a2f4-0f47-47cb-b5e6-facbf98694b9\") " pod="openshift-controller-manager/controller-manager-87cc55c9b-nlr2j" Mar 13 14:09:58 crc kubenswrapper[4907]: I0313 14:09:58.729761 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f341a2f4-0f47-47cb-b5e6-facbf98694b9-serving-cert\") pod \"controller-manager-87cc55c9b-nlr2j\" (UID: \"f341a2f4-0f47-47cb-b5e6-facbf98694b9\") " pod="openshift-controller-manager/controller-manager-87cc55c9b-nlr2j" Mar 13 14:09:58 crc kubenswrapper[4907]: I0313 14:09:58.741025 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5gj4x\" (UniqueName: \"kubernetes.io/projected/f341a2f4-0f47-47cb-b5e6-facbf98694b9-kube-api-access-5gj4x\") pod \"controller-manager-87cc55c9b-nlr2j\" (UID: \"f341a2f4-0f47-47cb-b5e6-facbf98694b9\") " pod="openshift-controller-manager/controller-manager-87cc55c9b-nlr2j" Mar 13 14:09:58 crc kubenswrapper[4907]: I0313 14:09:58.858973 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-87cc55c9b-nlr2j" Mar 13 14:09:59 crc kubenswrapper[4907]: I0313 14:09:59.277443 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-87cc55c9b-nlr2j"] Mar 13 14:09:59 crc kubenswrapper[4907]: W0313 14:09:59.286726 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf341a2f4_0f47_47cb_b5e6_facbf98694b9.slice/crio-7a696f22a4d1fcb996fa84f332f6a6f73364b4c819e37be9b8353e4327bf6378 WatchSource:0}: Error finding container 7a696f22a4d1fcb996fa84f332f6a6f73364b4c819e37be9b8353e4327bf6378: Status 404 returned error can't find the container with id 7a696f22a4d1fcb996fa84f332f6a6f73364b4c819e37be9b8353e4327bf6378 Mar 13 14:09:59 crc kubenswrapper[4907]: I0313 14:09:59.357794 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-87cc55c9b-nlr2j" event={"ID":"f341a2f4-0f47-47cb-b5e6-facbf98694b9","Type":"ContainerStarted","Data":"7a696f22a4d1fcb996fa84f332f6a6f73364b4c819e37be9b8353e4327bf6378"} Mar 13 14:10:00 crc kubenswrapper[4907]: I0313 14:10:00.141096 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556850-cl45g"] Mar 13 14:10:00 crc kubenswrapper[4907]: I0313 14:10:00.142298 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556850-cl45g" Mar 13 14:10:00 crc kubenswrapper[4907]: I0313 14:10:00.145633 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 14:10:00 crc kubenswrapper[4907]: I0313 14:10:00.146820 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556850-cl45g"] Mar 13 14:10:00 crc kubenswrapper[4907]: I0313 14:10:00.242612 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d8fjv\" (UniqueName: \"kubernetes.io/projected/10a1d486-2cf1-4078-8f0f-2dbc1ea97acc-kube-api-access-d8fjv\") pod \"auto-csr-approver-29556850-cl45g\" (UID: \"10a1d486-2cf1-4078-8f0f-2dbc1ea97acc\") " pod="openshift-infra/auto-csr-approver-29556850-cl45g" Mar 13 14:10:00 crc kubenswrapper[4907]: I0313 14:10:00.343814 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d8fjv\" (UniqueName: \"kubernetes.io/projected/10a1d486-2cf1-4078-8f0f-2dbc1ea97acc-kube-api-access-d8fjv\") pod \"auto-csr-approver-29556850-cl45g\" (UID: \"10a1d486-2cf1-4078-8f0f-2dbc1ea97acc\") " pod="openshift-infra/auto-csr-approver-29556850-cl45g" Mar 13 14:10:00 crc kubenswrapper[4907]: I0313 14:10:00.363955 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d8fjv\" (UniqueName: \"kubernetes.io/projected/10a1d486-2cf1-4078-8f0f-2dbc1ea97acc-kube-api-access-d8fjv\") pod \"auto-csr-approver-29556850-cl45g\" (UID: \"10a1d486-2cf1-4078-8f0f-2dbc1ea97acc\") " pod="openshift-infra/auto-csr-approver-29556850-cl45g" Mar 13 14:10:00 crc kubenswrapper[4907]: I0313 14:10:00.366080 4907 generic.go:334] "Generic (PLEG): container finished" podID="4b6ddc05-e51d-4648-84c6-6d1414398a4b" containerID="0a883aeefcfd6a2d4540c0a4e0e6bf02cdfb6c47d06b55f636280edfede41b98" exitCode=0 Mar 13 14:10:00 crc kubenswrapper[4907]: I0313 14:10:00.366169 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zbll9" event={"ID":"4b6ddc05-e51d-4648-84c6-6d1414398a4b","Type":"ContainerDied","Data":"0a883aeefcfd6a2d4540c0a4e0e6bf02cdfb6c47d06b55f636280edfede41b98"} Mar 13 14:10:00 crc kubenswrapper[4907]: I0313 14:10:00.370638 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-87cc55c9b-nlr2j" event={"ID":"f341a2f4-0f47-47cb-b5e6-facbf98694b9","Type":"ContainerStarted","Data":"695016fec526a7dbd7db71c59cc85a73528a2dc945096ce528d4e2858fa89f46"} Mar 13 14:10:00 crc kubenswrapper[4907]: I0313 14:10:00.370870 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-87cc55c9b-nlr2j" Mar 13 14:10:00 crc kubenswrapper[4907]: I0313 14:10:00.378930 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-87cc55c9b-nlr2j" Mar 13 14:10:00 crc kubenswrapper[4907]: I0313 14:10:00.417328 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-87cc55c9b-nlr2j" podStartSLOduration=9.417295381 podStartE2EDuration="9.417295381s" podCreationTimestamp="2026-03-13 14:09:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:10:00.414546617 +0000 UTC m=+299.314334306" watchObservedRunningTime="2026-03-13 14:10:00.417295381 +0000 UTC m=+299.317083080" Mar 13 14:10:00 crc kubenswrapper[4907]: I0313 14:10:00.457731 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556850-cl45g" Mar 13 14:10:00 crc kubenswrapper[4907]: I0313 14:10:00.631261 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556850-cl45g"] Mar 13 14:10:00 crc kubenswrapper[4907]: W0313 14:10:00.681774 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod10a1d486_2cf1_4078_8f0f_2dbc1ea97acc.slice/crio-91751e260bdf90455e0e2a20034e5c3e5c8d177cfcae4c3d592f866d7cff94d4 WatchSource:0}: Error finding container 91751e260bdf90455e0e2a20034e5c3e5c8d177cfcae4c3d592f866d7cff94d4: Status 404 returned error can't find the container with id 91751e260bdf90455e0e2a20034e5c3e5c8d177cfcae4c3d592f866d7cff94d4 Mar 13 14:10:01 crc kubenswrapper[4907]: I0313 14:10:01.241533 4907 csr.go:261] certificate signing request csr-lg4mf is approved, waiting to be issued Mar 13 14:10:01 crc kubenswrapper[4907]: I0313 14:10:01.248039 4907 csr.go:257] certificate signing request csr-lg4mf is issued Mar 13 14:10:01 crc kubenswrapper[4907]: I0313 14:10:01.376037 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556850-cl45g" event={"ID":"10a1d486-2cf1-4078-8f0f-2dbc1ea97acc","Type":"ContainerStarted","Data":"91751e260bdf90455e0e2a20034e5c3e5c8d177cfcae4c3d592f866d7cff94d4"} Mar 13 14:10:01 crc kubenswrapper[4907]: I0313 14:10:01.378206 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zbll9" event={"ID":"4b6ddc05-e51d-4648-84c6-6d1414398a4b","Type":"ContainerStarted","Data":"80864b09a701f09fdc5f3a4d57a08fa0f1943de0192a3eda8299785bc1d722e2"} Mar 13 14:10:01 crc kubenswrapper[4907]: I0313 14:10:01.380973 4907 generic.go:334] "Generic (PLEG): container finished" podID="a8867781-179d-47e4-9a55-330deb10d273" containerID="2e545187909a28cc5890020e1139c1d488e394bab920eab3166b317419887a60" exitCode=0 Mar 13 14:10:01 crc kubenswrapper[4907]: I0313 14:10:01.380992 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556848-v8kv4" event={"ID":"a8867781-179d-47e4-9a55-330deb10d273","Type":"ContainerDied","Data":"2e545187909a28cc5890020e1139c1d488e394bab920eab3166b317419887a60"} Mar 13 14:10:01 crc kubenswrapper[4907]: I0313 14:10:01.399091 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-zbll9" podStartSLOduration=3.493922748 podStartE2EDuration="53.399066386s" podCreationTimestamp="2026-03-13 14:09:08 +0000 UTC" firstStartedPulling="2026-03-13 14:09:10.86826136 +0000 UTC m=+249.768049049" lastFinishedPulling="2026-03-13 14:10:00.773404998 +0000 UTC m=+299.673192687" observedRunningTime="2026-03-13 14:10:01.39437653 +0000 UTC m=+300.294164239" watchObservedRunningTime="2026-03-13 14:10:01.399066386 +0000 UTC m=+300.298854075" Mar 13 14:10:02 crc kubenswrapper[4907]: I0313 14:10:02.249427 4907 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2027-02-24 05:54:36 +0000 UTC, rotation deadline is 2026-12-28 09:01:56.64747356 +0000 UTC Mar 13 14:10:02 crc kubenswrapper[4907]: I0313 14:10:02.249807 4907 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 6954h51m54.397670318s for next certificate rotation Mar 13 14:10:02 crc kubenswrapper[4907]: I0313 14:10:02.648820 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556848-v8kv4" Mar 13 14:10:02 crc kubenswrapper[4907]: I0313 14:10:02.772786 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9f5cd\" (UniqueName: \"kubernetes.io/projected/a8867781-179d-47e4-9a55-330deb10d273-kube-api-access-9f5cd\") pod \"a8867781-179d-47e4-9a55-330deb10d273\" (UID: \"a8867781-179d-47e4-9a55-330deb10d273\") " Mar 13 14:10:02 crc kubenswrapper[4907]: I0313 14:10:02.779579 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a8867781-179d-47e4-9a55-330deb10d273-kube-api-access-9f5cd" (OuterVolumeSpecName: "kube-api-access-9f5cd") pod "a8867781-179d-47e4-9a55-330deb10d273" (UID: "a8867781-179d-47e4-9a55-330deb10d273"). InnerVolumeSpecName "kube-api-access-9f5cd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:10:02 crc kubenswrapper[4907]: I0313 14:10:02.874429 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9f5cd\" (UniqueName: \"kubernetes.io/projected/a8867781-179d-47e4-9a55-330deb10d273-kube-api-access-9f5cd\") on node \"crc\" DevicePath \"\"" Mar 13 14:10:03 crc kubenswrapper[4907]: I0313 14:10:03.393720 4907 generic.go:334] "Generic (PLEG): container finished" podID="10a1d486-2cf1-4078-8f0f-2dbc1ea97acc" containerID="7118e101427b23d94455b072a1fe8c88f9a426be2d07a4a232a1f55839394ad6" exitCode=0 Mar 13 14:10:03 crc kubenswrapper[4907]: I0313 14:10:03.393805 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556850-cl45g" event={"ID":"10a1d486-2cf1-4078-8f0f-2dbc1ea97acc","Type":"ContainerDied","Data":"7118e101427b23d94455b072a1fe8c88f9a426be2d07a4a232a1f55839394ad6"} Mar 13 14:10:03 crc kubenswrapper[4907]: I0313 14:10:03.397024 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556848-v8kv4" event={"ID":"a8867781-179d-47e4-9a55-330deb10d273","Type":"ContainerDied","Data":"38620171197d496fb764e2284a8169d5eff52a1213907791553d488eaad39d9d"} Mar 13 14:10:03 crc kubenswrapper[4907]: I0313 14:10:03.397078 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="38620171197d496fb764e2284a8169d5eff52a1213907791553d488eaad39d9d" Mar 13 14:10:03 crc kubenswrapper[4907]: I0313 14:10:03.397082 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556848-v8kv4" Mar 13 14:10:04 crc kubenswrapper[4907]: I0313 14:10:04.689910 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556850-cl45g" Mar 13 14:10:04 crc kubenswrapper[4907]: I0313 14:10:04.799770 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d8fjv\" (UniqueName: \"kubernetes.io/projected/10a1d486-2cf1-4078-8f0f-2dbc1ea97acc-kube-api-access-d8fjv\") pod \"10a1d486-2cf1-4078-8f0f-2dbc1ea97acc\" (UID: \"10a1d486-2cf1-4078-8f0f-2dbc1ea97acc\") " Mar 13 14:10:04 crc kubenswrapper[4907]: I0313 14:10:04.808181 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/10a1d486-2cf1-4078-8f0f-2dbc1ea97acc-kube-api-access-d8fjv" (OuterVolumeSpecName: "kube-api-access-d8fjv") pod "10a1d486-2cf1-4078-8f0f-2dbc1ea97acc" (UID: "10a1d486-2cf1-4078-8f0f-2dbc1ea97acc"). InnerVolumeSpecName "kube-api-access-d8fjv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:10:04 crc kubenswrapper[4907]: I0313 14:10:04.901847 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d8fjv\" (UniqueName: \"kubernetes.io/projected/10a1d486-2cf1-4078-8f0f-2dbc1ea97acc-kube-api-access-d8fjv\") on node \"crc\" DevicePath \"\"" Mar 13 14:10:05 crc kubenswrapper[4907]: I0313 14:10:05.419285 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556850-cl45g" Mar 13 14:10:05 crc kubenswrapper[4907]: I0313 14:10:05.419308 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556850-cl45g" event={"ID":"10a1d486-2cf1-4078-8f0f-2dbc1ea97acc","Type":"ContainerDied","Data":"91751e260bdf90455e0e2a20034e5c3e5c8d177cfcae4c3d592f866d7cff94d4"} Mar 13 14:10:05 crc kubenswrapper[4907]: I0313 14:10:05.419360 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="91751e260bdf90455e0e2a20034e5c3e5c8d177cfcae4c3d592f866d7cff94d4" Mar 13 14:10:05 crc kubenswrapper[4907]: I0313 14:10:05.421717 4907 generic.go:334] "Generic (PLEG): container finished" podID="775414ea-5b94-4ab6-8b7c-fe7c80671ced" containerID="8fae8fbf21c4105d1f6b4e7581d3a799c097d7bce76e220e01e3d861c30432de" exitCode=0 Mar 13 14:10:05 crc kubenswrapper[4907]: I0313 14:10:05.421771 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p257z" event={"ID":"775414ea-5b94-4ab6-8b7c-fe7c80671ced","Type":"ContainerDied","Data":"8fae8fbf21c4105d1f6b4e7581d3a799c097d7bce76e220e01e3d861c30432de"} Mar 13 14:10:06 crc kubenswrapper[4907]: I0313 14:10:06.433016 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vx5qd" event={"ID":"8779a406-c08e-40db-b25d-8c45e784e94d","Type":"ContainerStarted","Data":"f50322f1c6601b0cca029ad907edd46e5e3069feeecd3b208e757340796cd8dd"} Mar 13 14:10:06 crc kubenswrapper[4907]: I0313 14:10:06.437153 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p257z" event={"ID":"775414ea-5b94-4ab6-8b7c-fe7c80671ced","Type":"ContainerStarted","Data":"b5828a5f740df8bf6f7ad67a6f9e5cd7c2cb33d99d741eccf259274586a3ee61"} Mar 13 14:10:06 crc kubenswrapper[4907]: I0313 14:10:06.440149 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fmzmc" event={"ID":"ec3466c2-4954-459d-88d8-e8c6cbf309e8","Type":"ContainerStarted","Data":"a06107e49962bc7b9a8f0ee9719c73d09f58fc4ae0da4f8a46e5df6c62712945"} Mar 13 14:10:07 crc kubenswrapper[4907]: I0313 14:10:07.447316 4907 generic.go:334] "Generic (PLEG): container finished" podID="ec3466c2-4954-459d-88d8-e8c6cbf309e8" containerID="a06107e49962bc7b9a8f0ee9719c73d09f58fc4ae0da4f8a46e5df6c62712945" exitCode=0 Mar 13 14:10:07 crc kubenswrapper[4907]: I0313 14:10:07.447481 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fmzmc" event={"ID":"ec3466c2-4954-459d-88d8-e8c6cbf309e8","Type":"ContainerDied","Data":"a06107e49962bc7b9a8f0ee9719c73d09f58fc4ae0da4f8a46e5df6c62712945"} Mar 13 14:10:07 crc kubenswrapper[4907]: I0313 14:10:07.449773 4907 generic.go:334] "Generic (PLEG): container finished" podID="8779a406-c08e-40db-b25d-8c45e784e94d" containerID="f50322f1c6601b0cca029ad907edd46e5e3069feeecd3b208e757340796cd8dd" exitCode=0 Mar 13 14:10:07 crc kubenswrapper[4907]: I0313 14:10:07.449877 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vx5qd" event={"ID":"8779a406-c08e-40db-b25d-8c45e784e94d","Type":"ContainerDied","Data":"f50322f1c6601b0cca029ad907edd46e5e3069feeecd3b208e757340796cd8dd"} Mar 13 14:10:07 crc kubenswrapper[4907]: I0313 14:10:07.525788 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-p257z" podStartSLOduration=7.454787329 podStartE2EDuration="57.525768593s" podCreationTimestamp="2026-03-13 14:09:10 +0000 UTC" firstStartedPulling="2026-03-13 14:09:16.143318192 +0000 UTC m=+255.043105881" lastFinishedPulling="2026-03-13 14:10:06.214299456 +0000 UTC m=+305.114087145" observedRunningTime="2026-03-13 14:10:07.520661924 +0000 UTC m=+306.420449613" watchObservedRunningTime="2026-03-13 14:10:07.525768593 +0000 UTC m=+306.425556282" Mar 13 14:10:09 crc kubenswrapper[4907]: I0313 14:10:09.060483 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-zbll9" Mar 13 14:10:09 crc kubenswrapper[4907]: I0313 14:10:09.060590 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-zbll9" Mar 13 14:10:09 crc kubenswrapper[4907]: I0313 14:10:09.485246 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-zbll9" Mar 13 14:10:10 crc kubenswrapper[4907]: I0313 14:10:10.524144 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-zbll9" Mar 13 14:10:11 crc kubenswrapper[4907]: I0313 14:10:11.200702 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-p257z" Mar 13 14:10:11 crc kubenswrapper[4907]: I0313 14:10:11.202057 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-p257z" Mar 13 14:10:11 crc kubenswrapper[4907]: I0313 14:10:11.254320 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-p257z" Mar 13 14:10:11 crc kubenswrapper[4907]: I0313 14:10:11.483518 4907 generic.go:334] "Generic (PLEG): container finished" podID="8adb3c1c-bacd-4cca-9796-7ca96624e9f6" containerID="059ebb72320a2144f692825d22d2d54a1e276a63cc41c87a83327001d7f8b539" exitCode=0 Mar 13 14:10:11 crc kubenswrapper[4907]: I0313 14:10:11.483555 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-42kpk" event={"ID":"8adb3c1c-bacd-4cca-9796-7ca96624e9f6","Type":"ContainerDied","Data":"059ebb72320a2144f692825d22d2d54a1e276a63cc41c87a83327001d7f8b539"} Mar 13 14:10:11 crc kubenswrapper[4907]: I0313 14:10:11.487141 4907 generic.go:334] "Generic (PLEG): container finished" podID="84c2ada5-ceab-4327-802e-9ae459ac814d" containerID="701b8cbaf2465addef85f7946e5e3a30ad0261ab612755a656d7086a119dc2d2" exitCode=0 Mar 13 14:10:11 crc kubenswrapper[4907]: I0313 14:10:11.487257 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v6msv" event={"ID":"84c2ada5-ceab-4327-802e-9ae459ac814d","Type":"ContainerDied","Data":"701b8cbaf2465addef85f7946e5e3a30ad0261ab612755a656d7086a119dc2d2"} Mar 13 14:10:11 crc kubenswrapper[4907]: I0313 14:10:11.540983 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-p257z" Mar 13 14:10:11 crc kubenswrapper[4907]: I0313 14:10:11.825313 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-87cc55c9b-nlr2j"] Mar 13 14:10:11 crc kubenswrapper[4907]: I0313 14:10:11.825590 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-87cc55c9b-nlr2j" podUID="f341a2f4-0f47-47cb-b5e6-facbf98694b9" containerName="controller-manager" containerID="cri-o://695016fec526a7dbd7db71c59cc85a73528a2dc945096ce528d4e2858fa89f46" gracePeriod=30 Mar 13 14:10:11 crc kubenswrapper[4907]: I0313 14:10:11.846477 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6f8c9bf8d5-trl8q"] Mar 13 14:10:11 crc kubenswrapper[4907]: I0313 14:10:11.846778 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6f8c9bf8d5-trl8q" podUID="7fb34c61-0980-40fc-9c6a-48af425786d0" containerName="route-controller-manager" containerID="cri-o://6b3f156649bd3aebfecf7ea45641ce94397a3dcd4edc7dc31ce59e515d93132d" gracePeriod=30 Mar 13 14:10:12 crc kubenswrapper[4907]: I0313 14:10:12.219839 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zbll9"] Mar 13 14:10:12 crc kubenswrapper[4907]: I0313 14:10:12.498472 4907 generic.go:334] "Generic (PLEG): container finished" podID="7fb34c61-0980-40fc-9c6a-48af425786d0" containerID="6b3f156649bd3aebfecf7ea45641ce94397a3dcd4edc7dc31ce59e515d93132d" exitCode=0 Mar 13 14:10:12 crc kubenswrapper[4907]: I0313 14:10:12.498566 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6f8c9bf8d5-trl8q" event={"ID":"7fb34c61-0980-40fc-9c6a-48af425786d0","Type":"ContainerDied","Data":"6b3f156649bd3aebfecf7ea45641ce94397a3dcd4edc7dc31ce59e515d93132d"} Mar 13 14:10:12 crc kubenswrapper[4907]: I0313 14:10:12.503216 4907 generic.go:334] "Generic (PLEG): container finished" podID="f341a2f4-0f47-47cb-b5e6-facbf98694b9" containerID="695016fec526a7dbd7db71c59cc85a73528a2dc945096ce528d4e2858fa89f46" exitCode=0 Mar 13 14:10:12 crc kubenswrapper[4907]: I0313 14:10:12.503300 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-87cc55c9b-nlr2j" event={"ID":"f341a2f4-0f47-47cb-b5e6-facbf98694b9","Type":"ContainerDied","Data":"695016fec526a7dbd7db71c59cc85a73528a2dc945096ce528d4e2858fa89f46"} Mar 13 14:10:12 crc kubenswrapper[4907]: I0313 14:10:12.504150 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-zbll9" podUID="4b6ddc05-e51d-4648-84c6-6d1414398a4b" containerName="registry-server" containerID="cri-o://80864b09a701f09fdc5f3a4d57a08fa0f1943de0192a3eda8299785bc1d722e2" gracePeriod=2 Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.073323 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6f8c9bf8d5-trl8q" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.106814 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-596c74884f-nxqmn"] Mar 13 14:10:13 crc kubenswrapper[4907]: E0313 14:10:13.107771 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fb34c61-0980-40fc-9c6a-48af425786d0" containerName="route-controller-manager" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.107840 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fb34c61-0980-40fc-9c6a-48af425786d0" containerName="route-controller-manager" Mar 13 14:10:13 crc kubenswrapper[4907]: E0313 14:10:13.107868 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10a1d486-2cf1-4078-8f0f-2dbc1ea97acc" containerName="oc" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.108010 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="10a1d486-2cf1-4078-8f0f-2dbc1ea97acc" containerName="oc" Mar 13 14:10:13 crc kubenswrapper[4907]: E0313 14:10:13.108028 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8867781-179d-47e4-9a55-330deb10d273" containerName="oc" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.108035 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8867781-179d-47e4-9a55-330deb10d273" containerName="oc" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.108153 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="10a1d486-2cf1-4078-8f0f-2dbc1ea97acc" containerName="oc" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.108189 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="a8867781-179d-47e4-9a55-330deb10d273" containerName="oc" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.108201 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="7fb34c61-0980-40fc-9c6a-48af425786d0" containerName="route-controller-manager" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.111549 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-596c74884f-nxqmn" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.114130 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-596c74884f-nxqmn"] Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.118244 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8c0d9c50-993a-4ade-9001-df91b492e893-client-ca\") pod \"route-controller-manager-596c74884f-nxqmn\" (UID: \"8c0d9c50-993a-4ade-9001-df91b492e893\") " pod="openshift-route-controller-manager/route-controller-manager-596c74884f-nxqmn" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.118344 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-smvjp\" (UniqueName: \"kubernetes.io/projected/8c0d9c50-993a-4ade-9001-df91b492e893-kube-api-access-smvjp\") pod \"route-controller-manager-596c74884f-nxqmn\" (UID: \"8c0d9c50-993a-4ade-9001-df91b492e893\") " pod="openshift-route-controller-manager/route-controller-manager-596c74884f-nxqmn" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.118373 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8c0d9c50-993a-4ade-9001-df91b492e893-serving-cert\") pod \"route-controller-manager-596c74884f-nxqmn\" (UID: \"8c0d9c50-993a-4ade-9001-df91b492e893\") " pod="openshift-route-controller-manager/route-controller-manager-596c74884f-nxqmn" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.118408 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8c0d9c50-993a-4ade-9001-df91b492e893-config\") pod \"route-controller-manager-596c74884f-nxqmn\" (UID: \"8c0d9c50-993a-4ade-9001-df91b492e893\") " pod="openshift-route-controller-manager/route-controller-manager-596c74884f-nxqmn" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.127369 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-87cc55c9b-nlr2j" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.219341 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7fb34c61-0980-40fc-9c6a-48af425786d0-config\") pod \"7fb34c61-0980-40fc-9c6a-48af425786d0\" (UID: \"7fb34c61-0980-40fc-9c6a-48af425786d0\") " Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.219385 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7fb34c61-0980-40fc-9c6a-48af425786d0-serving-cert\") pod \"7fb34c61-0980-40fc-9c6a-48af425786d0\" (UID: \"7fb34c61-0980-40fc-9c6a-48af425786d0\") " Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.219718 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f341a2f4-0f47-47cb-b5e6-facbf98694b9-config\") pod \"f341a2f4-0f47-47cb-b5e6-facbf98694b9\" (UID: \"f341a2f4-0f47-47cb-b5e6-facbf98694b9\") " Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.219743 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f341a2f4-0f47-47cb-b5e6-facbf98694b9-proxy-ca-bundles\") pod \"f341a2f4-0f47-47cb-b5e6-facbf98694b9\" (UID: \"f341a2f4-0f47-47cb-b5e6-facbf98694b9\") " Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.219762 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5gj4x\" (UniqueName: \"kubernetes.io/projected/f341a2f4-0f47-47cb-b5e6-facbf98694b9-kube-api-access-5gj4x\") pod \"f341a2f4-0f47-47cb-b5e6-facbf98694b9\" (UID: \"f341a2f4-0f47-47cb-b5e6-facbf98694b9\") " Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.219796 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f341a2f4-0f47-47cb-b5e6-facbf98694b9-serving-cert\") pod \"f341a2f4-0f47-47cb-b5e6-facbf98694b9\" (UID: \"f341a2f4-0f47-47cb-b5e6-facbf98694b9\") " Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.219815 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7fb34c61-0980-40fc-9c6a-48af425786d0-client-ca\") pod \"7fb34c61-0980-40fc-9c6a-48af425786d0\" (UID: \"7fb34c61-0980-40fc-9c6a-48af425786d0\") " Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.219840 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hhgz5\" (UniqueName: \"kubernetes.io/projected/7fb34c61-0980-40fc-9c6a-48af425786d0-kube-api-access-hhgz5\") pod \"7fb34c61-0980-40fc-9c6a-48af425786d0\" (UID: \"7fb34c61-0980-40fc-9c6a-48af425786d0\") " Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.219948 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8c0d9c50-993a-4ade-9001-df91b492e893-client-ca\") pod \"route-controller-manager-596c74884f-nxqmn\" (UID: \"8c0d9c50-993a-4ade-9001-df91b492e893\") " pod="openshift-route-controller-manager/route-controller-manager-596c74884f-nxqmn" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.219999 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-smvjp\" (UniqueName: \"kubernetes.io/projected/8c0d9c50-993a-4ade-9001-df91b492e893-kube-api-access-smvjp\") pod \"route-controller-manager-596c74884f-nxqmn\" (UID: \"8c0d9c50-993a-4ade-9001-df91b492e893\") " pod="openshift-route-controller-manager/route-controller-manager-596c74884f-nxqmn" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.220050 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8c0d9c50-993a-4ade-9001-df91b492e893-serving-cert\") pod \"route-controller-manager-596c74884f-nxqmn\" (UID: \"8c0d9c50-993a-4ade-9001-df91b492e893\") " pod="openshift-route-controller-manager/route-controller-manager-596c74884f-nxqmn" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.220071 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8c0d9c50-993a-4ade-9001-df91b492e893-config\") pod \"route-controller-manager-596c74884f-nxqmn\" (UID: \"8c0d9c50-993a-4ade-9001-df91b492e893\") " pod="openshift-route-controller-manager/route-controller-manager-596c74884f-nxqmn" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.220417 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7fb34c61-0980-40fc-9c6a-48af425786d0-config" (OuterVolumeSpecName: "config") pod "7fb34c61-0980-40fc-9c6a-48af425786d0" (UID: "7fb34c61-0980-40fc-9c6a-48af425786d0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.225020 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f341a2f4-0f47-47cb-b5e6-facbf98694b9-config" (OuterVolumeSpecName: "config") pod "f341a2f4-0f47-47cb-b5e6-facbf98694b9" (UID: "f341a2f4-0f47-47cb-b5e6-facbf98694b9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.225441 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f341a2f4-0f47-47cb-b5e6-facbf98694b9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "f341a2f4-0f47-47cb-b5e6-facbf98694b9" (UID: "f341a2f4-0f47-47cb-b5e6-facbf98694b9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.225462 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7fb34c61-0980-40fc-9c6a-48af425786d0-kube-api-access-hhgz5" (OuterVolumeSpecName: "kube-api-access-hhgz5") pod "7fb34c61-0980-40fc-9c6a-48af425786d0" (UID: "7fb34c61-0980-40fc-9c6a-48af425786d0"). InnerVolumeSpecName "kube-api-access-hhgz5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.226784 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7fb34c61-0980-40fc-9c6a-48af425786d0-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7fb34c61-0980-40fc-9c6a-48af425786d0" (UID: "7fb34c61-0980-40fc-9c6a-48af425786d0"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.228404 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f341a2f4-0f47-47cb-b5e6-facbf98694b9-kube-api-access-5gj4x" (OuterVolumeSpecName: "kube-api-access-5gj4x") pod "f341a2f4-0f47-47cb-b5e6-facbf98694b9" (UID: "f341a2f4-0f47-47cb-b5e6-facbf98694b9"). InnerVolumeSpecName "kube-api-access-5gj4x". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.229037 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f341a2f4-0f47-47cb-b5e6-facbf98694b9-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "f341a2f4-0f47-47cb-b5e6-facbf98694b9" (UID: "f341a2f4-0f47-47cb-b5e6-facbf98694b9"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.229661 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8c0d9c50-993a-4ade-9001-df91b492e893-client-ca\") pod \"route-controller-manager-596c74884f-nxqmn\" (UID: \"8c0d9c50-993a-4ade-9001-df91b492e893\") " pod="openshift-route-controller-manager/route-controller-manager-596c74884f-nxqmn" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.230467 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7fb34c61-0980-40fc-9c6a-48af425786d0-client-ca" (OuterVolumeSpecName: "client-ca") pod "7fb34c61-0980-40fc-9c6a-48af425786d0" (UID: "7fb34c61-0980-40fc-9c6a-48af425786d0"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.236662 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8c0d9c50-993a-4ade-9001-df91b492e893-serving-cert\") pod \"route-controller-manager-596c74884f-nxqmn\" (UID: \"8c0d9c50-993a-4ade-9001-df91b492e893\") " pod="openshift-route-controller-manager/route-controller-manager-596c74884f-nxqmn" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.236746 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8c0d9c50-993a-4ade-9001-df91b492e893-config\") pod \"route-controller-manager-596c74884f-nxqmn\" (UID: \"8c0d9c50-993a-4ade-9001-df91b492e893\") " pod="openshift-route-controller-manager/route-controller-manager-596c74884f-nxqmn" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.239753 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-smvjp\" (UniqueName: \"kubernetes.io/projected/8c0d9c50-993a-4ade-9001-df91b492e893-kube-api-access-smvjp\") pod \"route-controller-manager-596c74884f-nxqmn\" (UID: \"8c0d9c50-993a-4ade-9001-df91b492e893\") " pod="openshift-route-controller-manager/route-controller-manager-596c74884f-nxqmn" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.325803 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f341a2f4-0f47-47cb-b5e6-facbf98694b9-client-ca\") pod \"f341a2f4-0f47-47cb-b5e6-facbf98694b9\" (UID: \"f341a2f4-0f47-47cb-b5e6-facbf98694b9\") " Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.326767 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f341a2f4-0f47-47cb-b5e6-facbf98694b9-client-ca" (OuterVolumeSpecName: "client-ca") pod "f341a2f4-0f47-47cb-b5e6-facbf98694b9" (UID: "f341a2f4-0f47-47cb-b5e6-facbf98694b9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.326913 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hhgz5\" (UniqueName: \"kubernetes.io/projected/7fb34c61-0980-40fc-9c6a-48af425786d0-kube-api-access-hhgz5\") on node \"crc\" DevicePath \"\"" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.326940 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7fb34c61-0980-40fc-9c6a-48af425786d0-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.326951 4907 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7fb34c61-0980-40fc-9c6a-48af425786d0-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.326961 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f341a2f4-0f47-47cb-b5e6-facbf98694b9-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.326975 4907 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f341a2f4-0f47-47cb-b5e6-facbf98694b9-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.326985 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5gj4x\" (UniqueName: \"kubernetes.io/projected/f341a2f4-0f47-47cb-b5e6-facbf98694b9-kube-api-access-5gj4x\") on node \"crc\" DevicePath \"\"" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.326995 4907 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f341a2f4-0f47-47cb-b5e6-facbf98694b9-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.327004 4907 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7fb34c61-0980-40fc-9c6a-48af425786d0-client-ca\") on node \"crc\" DevicePath \"\"" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.389645 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zbll9" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.428748 4907 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f341a2f4-0f47-47cb-b5e6-facbf98694b9-client-ca\") on node \"crc\" DevicePath \"\"" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.440165 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-596c74884f-nxqmn" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.512586 4907 generic.go:334] "Generic (PLEG): container finished" podID="4b6ddc05-e51d-4648-84c6-6d1414398a4b" containerID="80864b09a701f09fdc5f3a4d57a08fa0f1943de0192a3eda8299785bc1d722e2" exitCode=0 Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.512641 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zbll9" event={"ID":"4b6ddc05-e51d-4648-84c6-6d1414398a4b","Type":"ContainerDied","Data":"80864b09a701f09fdc5f3a4d57a08fa0f1943de0192a3eda8299785bc1d722e2"} Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.512668 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zbll9" event={"ID":"4b6ddc05-e51d-4648-84c6-6d1414398a4b","Type":"ContainerDied","Data":"595859a9bf32c878bf56920b076fb65741978b6d9f83d6c4a352d65ef0a38504"} Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.512687 4907 scope.go:117] "RemoveContainer" containerID="80864b09a701f09fdc5f3a4d57a08fa0f1943de0192a3eda8299785bc1d722e2" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.512785 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zbll9" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.515114 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jk5nw" event={"ID":"d342a6a2-43dd-4d15-8aff-f60f9a3383bf","Type":"ContainerStarted","Data":"2ba2c849db65019cfc185d5e8fb816c1ea6e14d84b987e50fe979ce8ace8b5fb"} Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.518451 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fmzmc" event={"ID":"ec3466c2-4954-459d-88d8-e8c6cbf309e8","Type":"ContainerStarted","Data":"c60d7cfe210d5d163b6c716dca40da31b6a520923cab2dc2f6be072d3c2f2049"} Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.521743 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6f8c9bf8d5-trl8q" event={"ID":"7fb34c61-0980-40fc-9c6a-48af425786d0","Type":"ContainerDied","Data":"54f13a73875be71a32e36bb0c1839a08275e7ece063f4b0690961fb22cc8775b"} Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.521809 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6f8c9bf8d5-trl8q" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.527791 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vx5qd" event={"ID":"8779a406-c08e-40db-b25d-8c45e784e94d","Type":"ContainerStarted","Data":"8f15859e65d668d8b73b7ccd0f9c53101f5488796a1476b1c18893c5ad25807f"} Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.529462 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4b6ddc05-e51d-4648-84c6-6d1414398a4b-catalog-content\") pod \"4b6ddc05-e51d-4648-84c6-6d1414398a4b\" (UID: \"4b6ddc05-e51d-4648-84c6-6d1414398a4b\") " Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.529912 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6mhh2\" (UniqueName: \"kubernetes.io/projected/4b6ddc05-e51d-4648-84c6-6d1414398a4b-kube-api-access-6mhh2\") pod \"4b6ddc05-e51d-4648-84c6-6d1414398a4b\" (UID: \"4b6ddc05-e51d-4648-84c6-6d1414398a4b\") " Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.529960 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4b6ddc05-e51d-4648-84c6-6d1414398a4b-utilities\") pod \"4b6ddc05-e51d-4648-84c6-6d1414398a4b\" (UID: \"4b6ddc05-e51d-4648-84c6-6d1414398a4b\") " Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.530686 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4b6ddc05-e51d-4648-84c6-6d1414398a4b-utilities" (OuterVolumeSpecName: "utilities") pod "4b6ddc05-e51d-4648-84c6-6d1414398a4b" (UID: "4b6ddc05-e51d-4648-84c6-6d1414398a4b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.530924 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-87cc55c9b-nlr2j" event={"ID":"f341a2f4-0f47-47cb-b5e6-facbf98694b9","Type":"ContainerDied","Data":"7a696f22a4d1fcb996fa84f332f6a6f73364b4c819e37be9b8353e4327bf6378"} Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.530984 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-87cc55c9b-nlr2j" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.533318 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gjdp7" event={"ID":"5ef16085-0f4d-435d-a559-20a33c237531","Type":"ContainerStarted","Data":"031c710b0cf90428d9303b494400b216776d335d2ba2176b760375d7548140a7"} Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.534666 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4b6ddc05-e51d-4648-84c6-6d1414398a4b-kube-api-access-6mhh2" (OuterVolumeSpecName: "kube-api-access-6mhh2") pod "4b6ddc05-e51d-4648-84c6-6d1414398a4b" (UID: "4b6ddc05-e51d-4648-84c6-6d1414398a4b"). InnerVolumeSpecName "kube-api-access-6mhh2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.538061 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-42kpk" event={"ID":"8adb3c1c-bacd-4cca-9796-7ca96624e9f6","Type":"ContainerStarted","Data":"779541283cf16b48730f8ef9601537920e362fa4a7dbb572a3600beba4418986"} Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.548241 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v6msv" event={"ID":"84c2ada5-ceab-4327-802e-9ae459ac814d","Type":"ContainerStarted","Data":"6e642c222cc08b2da3f2c537efb7c366f76e02de754a7e9abbd7e89b0a44893f"} Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.599770 4907 scope.go:117] "RemoveContainer" containerID="0a883aeefcfd6a2d4540c0a4e0e6bf02cdfb6c47d06b55f636280edfede41b98" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.607712 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-fmzmc" podStartSLOduration=5.633796244 podStartE2EDuration="1m2.607688756s" podCreationTimestamp="2026-03-13 14:09:11 +0000 UTC" firstStartedPulling="2026-03-13 14:09:16.190183366 +0000 UTC m=+255.089971055" lastFinishedPulling="2026-03-13 14:10:13.164075878 +0000 UTC m=+312.063863567" observedRunningTime="2026-03-13 14:10:13.549790377 +0000 UTC m=+312.449578066" watchObservedRunningTime="2026-03-13 14:10:13.607688756 +0000 UTC m=+312.507476445" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.625957 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-v6msv" podStartSLOduration=3.28072038 podStartE2EDuration="1m5.62592687s" podCreationTimestamp="2026-03-13 14:09:08 +0000 UTC" firstStartedPulling="2026-03-13 14:09:10.830719429 +0000 UTC m=+249.730507118" lastFinishedPulling="2026-03-13 14:10:13.175925919 +0000 UTC m=+312.075713608" observedRunningTime="2026-03-13 14:10:13.607408987 +0000 UTC m=+312.507196676" watchObservedRunningTime="2026-03-13 14:10:13.62592687 +0000 UTC m=+312.525714549" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.631929 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6mhh2\" (UniqueName: \"kubernetes.io/projected/4b6ddc05-e51d-4648-84c6-6d1414398a4b-kube-api-access-6mhh2\") on node \"crc\" DevicePath \"\"" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.631953 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4b6ddc05-e51d-4648-84c6-6d1414398a4b-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.633055 4907 scope.go:117] "RemoveContainer" containerID="a3e791567d7fec33aa07ec52ed2c7bf531de23472ee1120fe2b93a41d79ec2e7" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.665678 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4b6ddc05-e51d-4648-84c6-6d1414398a4b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4b6ddc05-e51d-4648-84c6-6d1414398a4b" (UID: "4b6ddc05-e51d-4648-84c6-6d1414398a4b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.668096 4907 scope.go:117] "RemoveContainer" containerID="80864b09a701f09fdc5f3a4d57a08fa0f1943de0192a3eda8299785bc1d722e2" Mar 13 14:10:13 crc kubenswrapper[4907]: E0313 14:10:13.668621 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"80864b09a701f09fdc5f3a4d57a08fa0f1943de0192a3eda8299785bc1d722e2\": container with ID starting with 80864b09a701f09fdc5f3a4d57a08fa0f1943de0192a3eda8299785bc1d722e2 not found: ID does not exist" containerID="80864b09a701f09fdc5f3a4d57a08fa0f1943de0192a3eda8299785bc1d722e2" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.668664 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80864b09a701f09fdc5f3a4d57a08fa0f1943de0192a3eda8299785bc1d722e2"} err="failed to get container status \"80864b09a701f09fdc5f3a4d57a08fa0f1943de0192a3eda8299785bc1d722e2\": rpc error: code = NotFound desc = could not find container \"80864b09a701f09fdc5f3a4d57a08fa0f1943de0192a3eda8299785bc1d722e2\": container with ID starting with 80864b09a701f09fdc5f3a4d57a08fa0f1943de0192a3eda8299785bc1d722e2 not found: ID does not exist" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.668691 4907 scope.go:117] "RemoveContainer" containerID="0a883aeefcfd6a2d4540c0a4e0e6bf02cdfb6c47d06b55f636280edfede41b98" Mar 13 14:10:13 crc kubenswrapper[4907]: E0313 14:10:13.669067 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0a883aeefcfd6a2d4540c0a4e0e6bf02cdfb6c47d06b55f636280edfede41b98\": container with ID starting with 0a883aeefcfd6a2d4540c0a4e0e6bf02cdfb6c47d06b55f636280edfede41b98 not found: ID does not exist" containerID="0a883aeefcfd6a2d4540c0a4e0e6bf02cdfb6c47d06b55f636280edfede41b98" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.669099 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0a883aeefcfd6a2d4540c0a4e0e6bf02cdfb6c47d06b55f636280edfede41b98"} err="failed to get container status \"0a883aeefcfd6a2d4540c0a4e0e6bf02cdfb6c47d06b55f636280edfede41b98\": rpc error: code = NotFound desc = could not find container \"0a883aeefcfd6a2d4540c0a4e0e6bf02cdfb6c47d06b55f636280edfede41b98\": container with ID starting with 0a883aeefcfd6a2d4540c0a4e0e6bf02cdfb6c47d06b55f636280edfede41b98 not found: ID does not exist" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.669135 4907 scope.go:117] "RemoveContainer" containerID="a3e791567d7fec33aa07ec52ed2c7bf531de23472ee1120fe2b93a41d79ec2e7" Mar 13 14:10:13 crc kubenswrapper[4907]: E0313 14:10:13.669490 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a3e791567d7fec33aa07ec52ed2c7bf531de23472ee1120fe2b93a41d79ec2e7\": container with ID starting with a3e791567d7fec33aa07ec52ed2c7bf531de23472ee1120fe2b93a41d79ec2e7 not found: ID does not exist" containerID="a3e791567d7fec33aa07ec52ed2c7bf531de23472ee1120fe2b93a41d79ec2e7" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.669704 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a3e791567d7fec33aa07ec52ed2c7bf531de23472ee1120fe2b93a41d79ec2e7"} err="failed to get container status \"a3e791567d7fec33aa07ec52ed2c7bf531de23472ee1120fe2b93a41d79ec2e7\": rpc error: code = NotFound desc = could not find container \"a3e791567d7fec33aa07ec52ed2c7bf531de23472ee1120fe2b93a41d79ec2e7\": container with ID starting with a3e791567d7fec33aa07ec52ed2c7bf531de23472ee1120fe2b93a41d79ec2e7 not found: ID does not exist" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.669722 4907 scope.go:117] "RemoveContainer" containerID="6b3f156649bd3aebfecf7ea45641ce94397a3dcd4edc7dc31ce59e515d93132d" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.688301 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-42kpk" podStartSLOduration=3.300493233 podStartE2EDuration="1m5.688276728s" podCreationTimestamp="2026-03-13 14:09:08 +0000 UTC" firstStartedPulling="2026-03-13 14:09:10.908383601 +0000 UTC m=+249.808171280" lastFinishedPulling="2026-03-13 14:10:13.296167086 +0000 UTC m=+312.195954775" observedRunningTime="2026-03-13 14:10:13.673693654 +0000 UTC m=+312.573481363" watchObservedRunningTime="2026-03-13 14:10:13.688276728 +0000 UTC m=+312.588064417" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.709470 4907 scope.go:117] "RemoveContainer" containerID="695016fec526a7dbd7db71c59cc85a73528a2dc945096ce528d4e2858fa89f46" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.721911 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-vx5qd" podStartSLOduration=7.414619672 podStartE2EDuration="1m2.721866518s" podCreationTimestamp="2026-03-13 14:09:11 +0000 UTC" firstStartedPulling="2026-03-13 14:09:17.19376108 +0000 UTC m=+256.093548769" lastFinishedPulling="2026-03-13 14:10:12.501007926 +0000 UTC m=+311.400795615" observedRunningTime="2026-03-13 14:10:13.700758236 +0000 UTC m=+312.600545925" watchObservedRunningTime="2026-03-13 14:10:13.721866518 +0000 UTC m=+312.621654207" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.724001 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6f8c9bf8d5-trl8q"] Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.725688 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6f8c9bf8d5-trl8q"] Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.736554 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4b6ddc05-e51d-4648-84c6-6d1414398a4b-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.744735 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-87cc55c9b-nlr2j"] Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.757461 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-87cc55c9b-nlr2j"] Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.791504 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7fb34c61-0980-40fc-9c6a-48af425786d0" path="/var/lib/kubelet/pods/7fb34c61-0980-40fc-9c6a-48af425786d0/volumes" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.792019 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f341a2f4-0f47-47cb-b5e6-facbf98694b9" path="/var/lib/kubelet/pods/f341a2f4-0f47-47cb-b5e6-facbf98694b9/volumes" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.875536 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-596c74884f-nxqmn"] Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.923924 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zbll9"] Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.927652 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-zbll9"] Mar 13 14:10:14 crc kubenswrapper[4907]: I0313 14:10:14.017685 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-p257z"] Mar 13 14:10:14 crc kubenswrapper[4907]: I0313 14:10:14.554355 4907 generic.go:334] "Generic (PLEG): container finished" podID="d342a6a2-43dd-4d15-8aff-f60f9a3383bf" containerID="2ba2c849db65019cfc185d5e8fb816c1ea6e14d84b987e50fe979ce8ace8b5fb" exitCode=0 Mar 13 14:10:14 crc kubenswrapper[4907]: I0313 14:10:14.554602 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jk5nw" event={"ID":"d342a6a2-43dd-4d15-8aff-f60f9a3383bf","Type":"ContainerDied","Data":"2ba2c849db65019cfc185d5e8fb816c1ea6e14d84b987e50fe979ce8ace8b5fb"} Mar 13 14:10:14 crc kubenswrapper[4907]: I0313 14:10:14.558707 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-596c74884f-nxqmn" event={"ID":"8c0d9c50-993a-4ade-9001-df91b492e893","Type":"ContainerStarted","Data":"9f7688b312e9d673807fef17daaf30d9c92b2663246998695a6ef3925309315a"} Mar 13 14:10:14 crc kubenswrapper[4907]: I0313 14:10:14.558739 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-596c74884f-nxqmn" event={"ID":"8c0d9c50-993a-4ade-9001-df91b492e893","Type":"ContainerStarted","Data":"fe5db9115bda6db0c26db9bf40f10d6401ffba671bbe917f6eae2f4ff07fa2ca"} Mar 13 14:10:14 crc kubenswrapper[4907]: I0313 14:10:14.563586 4907 generic.go:334] "Generic (PLEG): container finished" podID="5ef16085-0f4d-435d-a559-20a33c237531" containerID="031c710b0cf90428d9303b494400b216776d335d2ba2176b760375d7548140a7" exitCode=0 Mar 13 14:10:14 crc kubenswrapper[4907]: I0313 14:10:14.563605 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gjdp7" event={"ID":"5ef16085-0f4d-435d-a559-20a33c237531","Type":"ContainerDied","Data":"031c710b0cf90428d9303b494400b216776d335d2ba2176b760375d7548140a7"} Mar 13 14:10:14 crc kubenswrapper[4907]: I0313 14:10:14.563835 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-p257z" podUID="775414ea-5b94-4ab6-8b7c-fe7c80671ced" containerName="registry-server" containerID="cri-o://b5828a5f740df8bf6f7ad67a6f9e5cd7c2cb33d99d741eccf259274586a3ee61" gracePeriod=2 Mar 13 14:10:15 crc kubenswrapper[4907]: I0313 14:10:15.422785 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-vx5qd" Mar 13 14:10:15 crc kubenswrapper[4907]: I0313 14:10:15.422837 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-vx5qd" Mar 13 14:10:15 crc kubenswrapper[4907]: I0313 14:10:15.554001 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-6797bc5cfb-75vkh"] Mar 13 14:10:15 crc kubenswrapper[4907]: E0313 14:10:15.554229 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b6ddc05-e51d-4648-84c6-6d1414398a4b" containerName="extract-content" Mar 13 14:10:15 crc kubenswrapper[4907]: I0313 14:10:15.554240 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b6ddc05-e51d-4648-84c6-6d1414398a4b" containerName="extract-content" Mar 13 14:10:15 crc kubenswrapper[4907]: E0313 14:10:15.554253 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f341a2f4-0f47-47cb-b5e6-facbf98694b9" containerName="controller-manager" Mar 13 14:10:15 crc kubenswrapper[4907]: I0313 14:10:15.554259 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="f341a2f4-0f47-47cb-b5e6-facbf98694b9" containerName="controller-manager" Mar 13 14:10:15 crc kubenswrapper[4907]: E0313 14:10:15.554273 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b6ddc05-e51d-4648-84c6-6d1414398a4b" containerName="registry-server" Mar 13 14:10:15 crc kubenswrapper[4907]: I0313 14:10:15.554279 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b6ddc05-e51d-4648-84c6-6d1414398a4b" containerName="registry-server" Mar 13 14:10:15 crc kubenswrapper[4907]: E0313 14:10:15.554287 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b6ddc05-e51d-4648-84c6-6d1414398a4b" containerName="extract-utilities" Mar 13 14:10:15 crc kubenswrapper[4907]: I0313 14:10:15.554292 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b6ddc05-e51d-4648-84c6-6d1414398a4b" containerName="extract-utilities" Mar 13 14:10:15 crc kubenswrapper[4907]: I0313 14:10:15.554416 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b6ddc05-e51d-4648-84c6-6d1414398a4b" containerName="registry-server" Mar 13 14:10:15 crc kubenswrapper[4907]: I0313 14:10:15.554425 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="f341a2f4-0f47-47cb-b5e6-facbf98694b9" containerName="controller-manager" Mar 13 14:10:15 crc kubenswrapper[4907]: I0313 14:10:15.555020 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6797bc5cfb-75vkh" Mar 13 14:10:15 crc kubenswrapper[4907]: I0313 14:10:15.557627 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 13 14:10:15 crc kubenswrapper[4907]: I0313 14:10:15.557772 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 13 14:10:15 crc kubenswrapper[4907]: I0313 14:10:15.557914 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 13 14:10:15 crc kubenswrapper[4907]: I0313 14:10:15.558075 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 13 14:10:15 crc kubenswrapper[4907]: I0313 14:10:15.558115 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 13 14:10:15 crc kubenswrapper[4907]: I0313 14:10:15.558118 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 13 14:10:15 crc kubenswrapper[4907]: I0313 14:10:15.566496 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 13 14:10:15 crc kubenswrapper[4907]: I0313 14:10:15.570821 4907 generic.go:334] "Generic (PLEG): container finished" podID="775414ea-5b94-4ab6-8b7c-fe7c80671ced" containerID="b5828a5f740df8bf6f7ad67a6f9e5cd7c2cb33d99d741eccf259274586a3ee61" exitCode=0 Mar 13 14:10:15 crc kubenswrapper[4907]: I0313 14:10:15.570919 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p257z" event={"ID":"775414ea-5b94-4ab6-8b7c-fe7c80671ced","Type":"ContainerDied","Data":"b5828a5f740df8bf6f7ad67a6f9e5cd7c2cb33d99d741eccf259274586a3ee61"} Mar 13 14:10:15 crc kubenswrapper[4907]: I0313 14:10:15.571379 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-596c74884f-nxqmn" Mar 13 14:10:15 crc kubenswrapper[4907]: I0313 14:10:15.576310 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-596c74884f-nxqmn" Mar 13 14:10:15 crc kubenswrapper[4907]: I0313 14:10:15.581059 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6797bc5cfb-75vkh"] Mar 13 14:10:15 crc kubenswrapper[4907]: I0313 14:10:15.661963 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/072a25a8-cae1-4572-846b-2d2feea5fb36-client-ca\") pod \"controller-manager-6797bc5cfb-75vkh\" (UID: \"072a25a8-cae1-4572-846b-2d2feea5fb36\") " pod="openshift-controller-manager/controller-manager-6797bc5cfb-75vkh" Mar 13 14:10:15 crc kubenswrapper[4907]: I0313 14:10:15.662537 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/072a25a8-cae1-4572-846b-2d2feea5fb36-serving-cert\") pod \"controller-manager-6797bc5cfb-75vkh\" (UID: \"072a25a8-cae1-4572-846b-2d2feea5fb36\") " pod="openshift-controller-manager/controller-manager-6797bc5cfb-75vkh" Mar 13 14:10:15 crc kubenswrapper[4907]: I0313 14:10:15.662992 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/072a25a8-cae1-4572-846b-2d2feea5fb36-proxy-ca-bundles\") pod \"controller-manager-6797bc5cfb-75vkh\" (UID: \"072a25a8-cae1-4572-846b-2d2feea5fb36\") " pod="openshift-controller-manager/controller-manager-6797bc5cfb-75vkh" Mar 13 14:10:15 crc kubenswrapper[4907]: I0313 14:10:15.663056 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/072a25a8-cae1-4572-846b-2d2feea5fb36-config\") pod \"controller-manager-6797bc5cfb-75vkh\" (UID: \"072a25a8-cae1-4572-846b-2d2feea5fb36\") " pod="openshift-controller-manager/controller-manager-6797bc5cfb-75vkh" Mar 13 14:10:15 crc kubenswrapper[4907]: I0313 14:10:15.663104 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-429xt\" (UniqueName: \"kubernetes.io/projected/072a25a8-cae1-4572-846b-2d2feea5fb36-kube-api-access-429xt\") pod \"controller-manager-6797bc5cfb-75vkh\" (UID: \"072a25a8-cae1-4572-846b-2d2feea5fb36\") " pod="openshift-controller-manager/controller-manager-6797bc5cfb-75vkh" Mar 13 14:10:15 crc kubenswrapper[4907]: I0313 14:10:15.764696 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/072a25a8-cae1-4572-846b-2d2feea5fb36-proxy-ca-bundles\") pod \"controller-manager-6797bc5cfb-75vkh\" (UID: \"072a25a8-cae1-4572-846b-2d2feea5fb36\") " pod="openshift-controller-manager/controller-manager-6797bc5cfb-75vkh" Mar 13 14:10:15 crc kubenswrapper[4907]: I0313 14:10:15.764753 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/072a25a8-cae1-4572-846b-2d2feea5fb36-config\") pod \"controller-manager-6797bc5cfb-75vkh\" (UID: \"072a25a8-cae1-4572-846b-2d2feea5fb36\") " pod="openshift-controller-manager/controller-manager-6797bc5cfb-75vkh" Mar 13 14:10:15 crc kubenswrapper[4907]: I0313 14:10:15.764788 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-429xt\" (UniqueName: \"kubernetes.io/projected/072a25a8-cae1-4572-846b-2d2feea5fb36-kube-api-access-429xt\") pod \"controller-manager-6797bc5cfb-75vkh\" (UID: \"072a25a8-cae1-4572-846b-2d2feea5fb36\") " pod="openshift-controller-manager/controller-manager-6797bc5cfb-75vkh" Mar 13 14:10:15 crc kubenswrapper[4907]: I0313 14:10:15.764833 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/072a25a8-cae1-4572-846b-2d2feea5fb36-client-ca\") pod \"controller-manager-6797bc5cfb-75vkh\" (UID: \"072a25a8-cae1-4572-846b-2d2feea5fb36\") " pod="openshift-controller-manager/controller-manager-6797bc5cfb-75vkh" Mar 13 14:10:15 crc kubenswrapper[4907]: I0313 14:10:15.764918 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/072a25a8-cae1-4572-846b-2d2feea5fb36-serving-cert\") pod \"controller-manager-6797bc5cfb-75vkh\" (UID: \"072a25a8-cae1-4572-846b-2d2feea5fb36\") " pod="openshift-controller-manager/controller-manager-6797bc5cfb-75vkh" Mar 13 14:10:15 crc kubenswrapper[4907]: I0313 14:10:15.766043 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/072a25a8-cae1-4572-846b-2d2feea5fb36-client-ca\") pod \"controller-manager-6797bc5cfb-75vkh\" (UID: \"072a25a8-cae1-4572-846b-2d2feea5fb36\") " pod="openshift-controller-manager/controller-manager-6797bc5cfb-75vkh" Mar 13 14:10:15 crc kubenswrapper[4907]: I0313 14:10:15.766385 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/072a25a8-cae1-4572-846b-2d2feea5fb36-config\") pod \"controller-manager-6797bc5cfb-75vkh\" (UID: \"072a25a8-cae1-4572-846b-2d2feea5fb36\") " pod="openshift-controller-manager/controller-manager-6797bc5cfb-75vkh" Mar 13 14:10:15 crc kubenswrapper[4907]: I0313 14:10:15.766449 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/072a25a8-cae1-4572-846b-2d2feea5fb36-proxy-ca-bundles\") pod \"controller-manager-6797bc5cfb-75vkh\" (UID: \"072a25a8-cae1-4572-846b-2d2feea5fb36\") " pod="openshift-controller-manager/controller-manager-6797bc5cfb-75vkh" Mar 13 14:10:15 crc kubenswrapper[4907]: I0313 14:10:15.770781 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/072a25a8-cae1-4572-846b-2d2feea5fb36-serving-cert\") pod \"controller-manager-6797bc5cfb-75vkh\" (UID: \"072a25a8-cae1-4572-846b-2d2feea5fb36\") " pod="openshift-controller-manager/controller-manager-6797bc5cfb-75vkh" Mar 13 14:10:15 crc kubenswrapper[4907]: I0313 14:10:15.781318 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-429xt\" (UniqueName: \"kubernetes.io/projected/072a25a8-cae1-4572-846b-2d2feea5fb36-kube-api-access-429xt\") pod \"controller-manager-6797bc5cfb-75vkh\" (UID: \"072a25a8-cae1-4572-846b-2d2feea5fb36\") " pod="openshift-controller-manager/controller-manager-6797bc5cfb-75vkh" Mar 13 14:10:15 crc kubenswrapper[4907]: I0313 14:10:15.793676 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4b6ddc05-e51d-4648-84c6-6d1414398a4b" path="/var/lib/kubelet/pods/4b6ddc05-e51d-4648-84c6-6d1414398a4b/volumes" Mar 13 14:10:15 crc kubenswrapper[4907]: I0313 14:10:15.871993 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6797bc5cfb-75vkh" Mar 13 14:10:16 crc kubenswrapper[4907]: I0313 14:10:16.079895 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-p257z" Mar 13 14:10:16 crc kubenswrapper[4907]: I0313 14:10:16.105373 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-596c74884f-nxqmn" podStartSLOduration=5.105350064 podStartE2EDuration="5.105350064s" podCreationTimestamp="2026-03-13 14:10:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:10:15.625709451 +0000 UTC m=+314.525497140" watchObservedRunningTime="2026-03-13 14:10:16.105350064 +0000 UTC m=+315.005137753" Mar 13 14:10:16 crc kubenswrapper[4907]: I0313 14:10:16.125452 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6797bc5cfb-75vkh"] Mar 13 14:10:16 crc kubenswrapper[4907]: W0313 14:10:16.133478 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod072a25a8_cae1_4572_846b_2d2feea5fb36.slice/crio-381bee78bac3d3aef3f83d5ea311981cbe79eee1108c280c12eebed6b19e7330 WatchSource:0}: Error finding container 381bee78bac3d3aef3f83d5ea311981cbe79eee1108c280c12eebed6b19e7330: Status 404 returned error can't find the container with id 381bee78bac3d3aef3f83d5ea311981cbe79eee1108c280c12eebed6b19e7330 Mar 13 14:10:16 crc kubenswrapper[4907]: I0313 14:10:16.273688 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/775414ea-5b94-4ab6-8b7c-fe7c80671ced-catalog-content\") pod \"775414ea-5b94-4ab6-8b7c-fe7c80671ced\" (UID: \"775414ea-5b94-4ab6-8b7c-fe7c80671ced\") " Mar 13 14:10:16 crc kubenswrapper[4907]: I0313 14:10:16.273740 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/775414ea-5b94-4ab6-8b7c-fe7c80671ced-utilities\") pod \"775414ea-5b94-4ab6-8b7c-fe7c80671ced\" (UID: \"775414ea-5b94-4ab6-8b7c-fe7c80671ced\") " Mar 13 14:10:16 crc kubenswrapper[4907]: I0313 14:10:16.273764 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rjgtx\" (UniqueName: \"kubernetes.io/projected/775414ea-5b94-4ab6-8b7c-fe7c80671ced-kube-api-access-rjgtx\") pod \"775414ea-5b94-4ab6-8b7c-fe7c80671ced\" (UID: \"775414ea-5b94-4ab6-8b7c-fe7c80671ced\") " Mar 13 14:10:16 crc kubenswrapper[4907]: I0313 14:10:16.277651 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/775414ea-5b94-4ab6-8b7c-fe7c80671ced-utilities" (OuterVolumeSpecName: "utilities") pod "775414ea-5b94-4ab6-8b7c-fe7c80671ced" (UID: "775414ea-5b94-4ab6-8b7c-fe7c80671ced"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:10:16 crc kubenswrapper[4907]: I0313 14:10:16.283063 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/775414ea-5b94-4ab6-8b7c-fe7c80671ced-kube-api-access-rjgtx" (OuterVolumeSpecName: "kube-api-access-rjgtx") pod "775414ea-5b94-4ab6-8b7c-fe7c80671ced" (UID: "775414ea-5b94-4ab6-8b7c-fe7c80671ced"). InnerVolumeSpecName "kube-api-access-rjgtx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:10:16 crc kubenswrapper[4907]: I0313 14:10:16.356456 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/775414ea-5b94-4ab6-8b7c-fe7c80671ced-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "775414ea-5b94-4ab6-8b7c-fe7c80671ced" (UID: "775414ea-5b94-4ab6-8b7c-fe7c80671ced"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:10:16 crc kubenswrapper[4907]: I0313 14:10:16.375259 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/775414ea-5b94-4ab6-8b7c-fe7c80671ced-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 14:10:16 crc kubenswrapper[4907]: I0313 14:10:16.375310 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/775414ea-5b94-4ab6-8b7c-fe7c80671ced-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 14:10:16 crc kubenswrapper[4907]: I0313 14:10:16.375320 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rjgtx\" (UniqueName: \"kubernetes.io/projected/775414ea-5b94-4ab6-8b7c-fe7c80671ced-kube-api-access-rjgtx\") on node \"crc\" DevicePath \"\"" Mar 13 14:10:16 crc kubenswrapper[4907]: I0313 14:10:16.474842 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-vx5qd" podUID="8779a406-c08e-40db-b25d-8c45e784e94d" containerName="registry-server" probeResult="failure" output=< Mar 13 14:10:16 crc kubenswrapper[4907]: timeout: failed to connect service ":50051" within 1s Mar 13 14:10:16 crc kubenswrapper[4907]: > Mar 13 14:10:16 crc kubenswrapper[4907]: I0313 14:10:16.578447 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gjdp7" event={"ID":"5ef16085-0f4d-435d-a559-20a33c237531","Type":"ContainerStarted","Data":"c851c33bb27b562ca9f87e1a35a7d7576757ef40b9f02ebed142608872748069"} Mar 13 14:10:16 crc kubenswrapper[4907]: I0313 14:10:16.580636 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p257z" event={"ID":"775414ea-5b94-4ab6-8b7c-fe7c80671ced","Type":"ContainerDied","Data":"f1572a3e7d8d63236fe0e9ac91306d55a1b117717a36d7026cf70c18e50c6571"} Mar 13 14:10:16 crc kubenswrapper[4907]: I0313 14:10:16.580702 4907 scope.go:117] "RemoveContainer" containerID="b5828a5f740df8bf6f7ad67a6f9e5cd7c2cb33d99d741eccf259274586a3ee61" Mar 13 14:10:16 crc kubenswrapper[4907]: I0313 14:10:16.580793 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-p257z" Mar 13 14:10:16 crc kubenswrapper[4907]: I0313 14:10:16.582607 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jk5nw" event={"ID":"d342a6a2-43dd-4d15-8aff-f60f9a3383bf","Type":"ContainerStarted","Data":"de346bdbb6259a737c213b9565b9465613c835283a744df647429b69400db2ec"} Mar 13 14:10:16 crc kubenswrapper[4907]: I0313 14:10:16.584700 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6797bc5cfb-75vkh" event={"ID":"072a25a8-cae1-4572-846b-2d2feea5fb36","Type":"ContainerStarted","Data":"3251605f371bf11c4b8c99a286c7e7a5947870c25059d47601917b31fe721247"} Mar 13 14:10:16 crc kubenswrapper[4907]: I0313 14:10:16.584742 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6797bc5cfb-75vkh" event={"ID":"072a25a8-cae1-4572-846b-2d2feea5fb36","Type":"ContainerStarted","Data":"381bee78bac3d3aef3f83d5ea311981cbe79eee1108c280c12eebed6b19e7330"} Mar 13 14:10:16 crc kubenswrapper[4907]: I0313 14:10:16.584923 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-6797bc5cfb-75vkh" Mar 13 14:10:16 crc kubenswrapper[4907]: I0313 14:10:16.592259 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-6797bc5cfb-75vkh" Mar 13 14:10:16 crc kubenswrapper[4907]: I0313 14:10:16.595072 4907 scope.go:117] "RemoveContainer" containerID="8fae8fbf21c4105d1f6b4e7581d3a799c097d7bce76e220e01e3d861c30432de" Mar 13 14:10:16 crc kubenswrapper[4907]: I0313 14:10:16.610641 4907 scope.go:117] "RemoveContainer" containerID="2f158ff96c417d8c3693d515dfe179799bf07e75ac8f792f12324fd7a18b5c50" Mar 13 14:10:16 crc kubenswrapper[4907]: I0313 14:10:16.626405 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-jk5nw" podStartSLOduration=6.742675659 podStartE2EDuration="1m6.626387639s" podCreationTimestamp="2026-03-13 14:09:10 +0000 UTC" firstStartedPulling="2026-03-13 14:09:16.152527152 +0000 UTC m=+255.052314841" lastFinishedPulling="2026-03-13 14:10:16.036239132 +0000 UTC m=+314.936026821" observedRunningTime="2026-03-13 14:10:16.622119882 +0000 UTC m=+315.521907571" watchObservedRunningTime="2026-03-13 14:10:16.626387639 +0000 UTC m=+315.526175318" Mar 13 14:10:16 crc kubenswrapper[4907]: I0313 14:10:16.651784 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-6797bc5cfb-75vkh" podStartSLOduration=5.651768796 podStartE2EDuration="5.651768796s" podCreationTimestamp="2026-03-13 14:10:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:10:16.650551573 +0000 UTC m=+315.550339262" watchObservedRunningTime="2026-03-13 14:10:16.651768796 +0000 UTC m=+315.551556485" Mar 13 14:10:16 crc kubenswrapper[4907]: I0313 14:10:16.669957 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-p257z"] Mar 13 14:10:16 crc kubenswrapper[4907]: I0313 14:10:16.673752 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-p257z"] Mar 13 14:10:17 crc kubenswrapper[4907]: I0313 14:10:17.616818 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-gjdp7" podStartSLOduration=4.339136938 podStartE2EDuration="1m9.616798898s" podCreationTimestamp="2026-03-13 14:09:08 +0000 UTC" firstStartedPulling="2026-03-13 14:09:10.889097687 +0000 UTC m=+249.788885376" lastFinishedPulling="2026-03-13 14:10:16.166759627 +0000 UTC m=+315.066547336" observedRunningTime="2026-03-13 14:10:17.615603075 +0000 UTC m=+316.515390784" watchObservedRunningTime="2026-03-13 14:10:17.616798898 +0000 UTC m=+316.516586587" Mar 13 14:10:17 crc kubenswrapper[4907]: I0313 14:10:17.788500 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="775414ea-5b94-4ab6-8b7c-fe7c80671ced" path="/var/lib/kubelet/pods/775414ea-5b94-4ab6-8b7c-fe7c80671ced/volumes" Mar 13 14:10:18 crc kubenswrapper[4907]: I0313 14:10:18.592966 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-v6msv" Mar 13 14:10:18 crc kubenswrapper[4907]: I0313 14:10:18.593051 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-v6msv" Mar 13 14:10:18 crc kubenswrapper[4907]: I0313 14:10:18.648443 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-v6msv" Mar 13 14:10:18 crc kubenswrapper[4907]: I0313 14:10:18.771726 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-42kpk" Mar 13 14:10:18 crc kubenswrapper[4907]: I0313 14:10:18.771784 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-42kpk" Mar 13 14:10:18 crc kubenswrapper[4907]: I0313 14:10:18.817510 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-42kpk" Mar 13 14:10:19 crc kubenswrapper[4907]: I0313 14:10:19.317204 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-gjdp7" Mar 13 14:10:19 crc kubenswrapper[4907]: I0313 14:10:19.317576 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-gjdp7" Mar 13 14:10:19 crc kubenswrapper[4907]: I0313 14:10:19.360308 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-gjdp7" Mar 13 14:10:19 crc kubenswrapper[4907]: I0313 14:10:19.662851 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-v6msv" Mar 13 14:10:19 crc kubenswrapper[4907]: I0313 14:10:19.685535 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-42kpk" Mar 13 14:10:20 crc kubenswrapper[4907]: I0313 14:10:20.773269 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-jk5nw" Mar 13 14:10:20 crc kubenswrapper[4907]: I0313 14:10:20.774009 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-jk5nw" Mar 13 14:10:20 crc kubenswrapper[4907]: I0313 14:10:20.830802 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-jk5nw" Mar 13 14:10:21 crc kubenswrapper[4907]: I0313 14:10:21.691251 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-jk5nw" Mar 13 14:10:21 crc kubenswrapper[4907]: I0313 14:10:21.793295 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-fmzmc" Mar 13 14:10:21 crc kubenswrapper[4907]: I0313 14:10:21.793800 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-fmzmc" Mar 13 14:10:21 crc kubenswrapper[4907]: I0313 14:10:21.842945 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-fmzmc" Mar 13 14:10:22 crc kubenswrapper[4907]: I0313 14:10:22.709797 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-fmzmc" Mar 13 14:10:25 crc kubenswrapper[4907]: I0313 14:10:25.502038 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-vx5qd" Mar 13 14:10:25 crc kubenswrapper[4907]: I0313 14:10:25.552761 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-vx5qd" Mar 13 14:10:28 crc kubenswrapper[4907]: I0313 14:10:28.422170 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vx5qd"] Mar 13 14:10:28 crc kubenswrapper[4907]: I0313 14:10:28.422450 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-vx5qd" podUID="8779a406-c08e-40db-b25d-8c45e784e94d" containerName="registry-server" containerID="cri-o://8f15859e65d668d8b73b7ccd0f9c53101f5488796a1476b1c18893c5ad25807f" gracePeriod=2 Mar 13 14:10:28 crc kubenswrapper[4907]: I0313 14:10:28.684875 4907 generic.go:334] "Generic (PLEG): container finished" podID="8779a406-c08e-40db-b25d-8c45e784e94d" containerID="8f15859e65d668d8b73b7ccd0f9c53101f5488796a1476b1c18893c5ad25807f" exitCode=0 Mar 13 14:10:28 crc kubenswrapper[4907]: I0313 14:10:28.684960 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vx5qd" event={"ID":"8779a406-c08e-40db-b25d-8c45e784e94d","Type":"ContainerDied","Data":"8f15859e65d668d8b73b7ccd0f9c53101f5488796a1476b1c18893c5ad25807f"} Mar 13 14:10:28 crc kubenswrapper[4907]: I0313 14:10:28.909864 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vx5qd" Mar 13 14:10:29 crc kubenswrapper[4907]: I0313 14:10:29.085657 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mcf55\" (UniqueName: \"kubernetes.io/projected/8779a406-c08e-40db-b25d-8c45e784e94d-kube-api-access-mcf55\") pod \"8779a406-c08e-40db-b25d-8c45e784e94d\" (UID: \"8779a406-c08e-40db-b25d-8c45e784e94d\") " Mar 13 14:10:29 crc kubenswrapper[4907]: I0313 14:10:29.085730 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8779a406-c08e-40db-b25d-8c45e784e94d-catalog-content\") pod \"8779a406-c08e-40db-b25d-8c45e784e94d\" (UID: \"8779a406-c08e-40db-b25d-8c45e784e94d\") " Mar 13 14:10:29 crc kubenswrapper[4907]: I0313 14:10:29.085813 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8779a406-c08e-40db-b25d-8c45e784e94d-utilities\") pod \"8779a406-c08e-40db-b25d-8c45e784e94d\" (UID: \"8779a406-c08e-40db-b25d-8c45e784e94d\") " Mar 13 14:10:29 crc kubenswrapper[4907]: I0313 14:10:29.087607 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8779a406-c08e-40db-b25d-8c45e784e94d-utilities" (OuterVolumeSpecName: "utilities") pod "8779a406-c08e-40db-b25d-8c45e784e94d" (UID: "8779a406-c08e-40db-b25d-8c45e784e94d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:10:29 crc kubenswrapper[4907]: I0313 14:10:29.100288 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8779a406-c08e-40db-b25d-8c45e784e94d-kube-api-access-mcf55" (OuterVolumeSpecName: "kube-api-access-mcf55") pod "8779a406-c08e-40db-b25d-8c45e784e94d" (UID: "8779a406-c08e-40db-b25d-8c45e784e94d"). InnerVolumeSpecName "kube-api-access-mcf55". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:10:29 crc kubenswrapper[4907]: I0313 14:10:29.187526 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8779a406-c08e-40db-b25d-8c45e784e94d-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 14:10:29 crc kubenswrapper[4907]: I0313 14:10:29.187574 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mcf55\" (UniqueName: \"kubernetes.io/projected/8779a406-c08e-40db-b25d-8c45e784e94d-kube-api-access-mcf55\") on node \"crc\" DevicePath \"\"" Mar 13 14:10:29 crc kubenswrapper[4907]: I0313 14:10:29.209046 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8779a406-c08e-40db-b25d-8c45e784e94d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8779a406-c08e-40db-b25d-8c45e784e94d" (UID: "8779a406-c08e-40db-b25d-8c45e784e94d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:10:29 crc kubenswrapper[4907]: I0313 14:10:29.289513 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8779a406-c08e-40db-b25d-8c45e784e94d-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 14:10:29 crc kubenswrapper[4907]: I0313 14:10:29.385557 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-gjdp7" Mar 13 14:10:29 crc kubenswrapper[4907]: I0313 14:10:29.698736 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vx5qd" event={"ID":"8779a406-c08e-40db-b25d-8c45e784e94d","Type":"ContainerDied","Data":"23bfb83692e34549e2b59eb14b5903640d0e07ca09276303909a34702eedee74"} Mar 13 14:10:29 crc kubenswrapper[4907]: I0313 14:10:29.698824 4907 scope.go:117] "RemoveContainer" containerID="8f15859e65d668d8b73b7ccd0f9c53101f5488796a1476b1c18893c5ad25807f" Mar 13 14:10:29 crc kubenswrapper[4907]: I0313 14:10:29.698873 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vx5qd" Mar 13 14:10:29 crc kubenswrapper[4907]: I0313 14:10:29.723039 4907 scope.go:117] "RemoveContainer" containerID="f50322f1c6601b0cca029ad907edd46e5e3069feeecd3b208e757340796cd8dd" Mar 13 14:10:29 crc kubenswrapper[4907]: I0313 14:10:29.751735 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vx5qd"] Mar 13 14:10:29 crc kubenswrapper[4907]: I0313 14:10:29.757106 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-vx5qd"] Mar 13 14:10:29 crc kubenswrapper[4907]: I0313 14:10:29.767116 4907 scope.go:117] "RemoveContainer" containerID="8baf18c6d4683811fc8b3cb2c3dfe2ac02213d10d0c7cc70876c15379d595455" Mar 13 14:10:29 crc kubenswrapper[4907]: I0313 14:10:29.798488 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8779a406-c08e-40db-b25d-8c45e784e94d" path="/var/lib/kubelet/pods/8779a406-c08e-40db-b25d-8c45e784e94d/volumes" Mar 13 14:10:31 crc kubenswrapper[4907]: I0313 14:10:31.883137 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-6797bc5cfb-75vkh"] Mar 13 14:10:31 crc kubenswrapper[4907]: I0313 14:10:31.883838 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-6797bc5cfb-75vkh" podUID="072a25a8-cae1-4572-846b-2d2feea5fb36" containerName="controller-manager" containerID="cri-o://3251605f371bf11c4b8c99a286c7e7a5947870c25059d47601917b31fe721247" gracePeriod=30 Mar 13 14:10:31 crc kubenswrapper[4907]: I0313 14:10:31.968350 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-596c74884f-nxqmn"] Mar 13 14:10:31 crc kubenswrapper[4907]: I0313 14:10:31.968653 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-596c74884f-nxqmn" podUID="8c0d9c50-993a-4ade-9001-df91b492e893" containerName="route-controller-manager" containerID="cri-o://9f7688b312e9d673807fef17daaf30d9c92b2663246998695a6ef3925309315a" gracePeriod=30 Mar 13 14:10:32 crc kubenswrapper[4907]: I0313 14:10:32.517205 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-ng4kz"] Mar 13 14:10:32 crc kubenswrapper[4907]: I0313 14:10:32.551737 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-596c74884f-nxqmn" Mar 13 14:10:32 crc kubenswrapper[4907]: I0313 14:10:32.556705 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6797bc5cfb-75vkh" Mar 13 14:10:32 crc kubenswrapper[4907]: I0313 14:10:32.655744 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-smvjp\" (UniqueName: \"kubernetes.io/projected/8c0d9c50-993a-4ade-9001-df91b492e893-kube-api-access-smvjp\") pod \"8c0d9c50-993a-4ade-9001-df91b492e893\" (UID: \"8c0d9c50-993a-4ade-9001-df91b492e893\") " Mar 13 14:10:32 crc kubenswrapper[4907]: I0313 14:10:32.655797 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/072a25a8-cae1-4572-846b-2d2feea5fb36-proxy-ca-bundles\") pod \"072a25a8-cae1-4572-846b-2d2feea5fb36\" (UID: \"072a25a8-cae1-4572-846b-2d2feea5fb36\") " Mar 13 14:10:32 crc kubenswrapper[4907]: I0313 14:10:32.655838 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/072a25a8-cae1-4572-846b-2d2feea5fb36-serving-cert\") pod \"072a25a8-cae1-4572-846b-2d2feea5fb36\" (UID: \"072a25a8-cae1-4572-846b-2d2feea5fb36\") " Mar 13 14:10:32 crc kubenswrapper[4907]: I0313 14:10:32.655872 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-429xt\" (UniqueName: \"kubernetes.io/projected/072a25a8-cae1-4572-846b-2d2feea5fb36-kube-api-access-429xt\") pod \"072a25a8-cae1-4572-846b-2d2feea5fb36\" (UID: \"072a25a8-cae1-4572-846b-2d2feea5fb36\") " Mar 13 14:10:32 crc kubenswrapper[4907]: I0313 14:10:32.655909 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8c0d9c50-993a-4ade-9001-df91b492e893-client-ca\") pod \"8c0d9c50-993a-4ade-9001-df91b492e893\" (UID: \"8c0d9c50-993a-4ade-9001-df91b492e893\") " Mar 13 14:10:32 crc kubenswrapper[4907]: I0313 14:10:32.655944 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8c0d9c50-993a-4ade-9001-df91b492e893-config\") pod \"8c0d9c50-993a-4ade-9001-df91b492e893\" (UID: \"8c0d9c50-993a-4ade-9001-df91b492e893\") " Mar 13 14:10:32 crc kubenswrapper[4907]: I0313 14:10:32.655982 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/072a25a8-cae1-4572-846b-2d2feea5fb36-config\") pod \"072a25a8-cae1-4572-846b-2d2feea5fb36\" (UID: \"072a25a8-cae1-4572-846b-2d2feea5fb36\") " Mar 13 14:10:32 crc kubenswrapper[4907]: I0313 14:10:32.656017 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/072a25a8-cae1-4572-846b-2d2feea5fb36-client-ca\") pod \"072a25a8-cae1-4572-846b-2d2feea5fb36\" (UID: \"072a25a8-cae1-4572-846b-2d2feea5fb36\") " Mar 13 14:10:32 crc kubenswrapper[4907]: I0313 14:10:32.656044 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8c0d9c50-993a-4ade-9001-df91b492e893-serving-cert\") pod \"8c0d9c50-993a-4ade-9001-df91b492e893\" (UID: \"8c0d9c50-993a-4ade-9001-df91b492e893\") " Mar 13 14:10:32 crc kubenswrapper[4907]: I0313 14:10:32.657618 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8c0d9c50-993a-4ade-9001-df91b492e893-config" (OuterVolumeSpecName: "config") pod "8c0d9c50-993a-4ade-9001-df91b492e893" (UID: "8c0d9c50-993a-4ade-9001-df91b492e893"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:10:32 crc kubenswrapper[4907]: I0313 14:10:32.657982 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8c0d9c50-993a-4ade-9001-df91b492e893-client-ca" (OuterVolumeSpecName: "client-ca") pod "8c0d9c50-993a-4ade-9001-df91b492e893" (UID: "8c0d9c50-993a-4ade-9001-df91b492e893"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:10:32 crc kubenswrapper[4907]: I0313 14:10:32.658140 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/072a25a8-cae1-4572-846b-2d2feea5fb36-config" (OuterVolumeSpecName: "config") pod "072a25a8-cae1-4572-846b-2d2feea5fb36" (UID: "072a25a8-cae1-4572-846b-2d2feea5fb36"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:10:32 crc kubenswrapper[4907]: I0313 14:10:32.658219 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/072a25a8-cae1-4572-846b-2d2feea5fb36-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "072a25a8-cae1-4572-846b-2d2feea5fb36" (UID: "072a25a8-cae1-4572-846b-2d2feea5fb36"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:10:32 crc kubenswrapper[4907]: I0313 14:10:32.658570 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/072a25a8-cae1-4572-846b-2d2feea5fb36-client-ca" (OuterVolumeSpecName: "client-ca") pod "072a25a8-cae1-4572-846b-2d2feea5fb36" (UID: "072a25a8-cae1-4572-846b-2d2feea5fb36"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:10:32 crc kubenswrapper[4907]: I0313 14:10:32.666766 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c0d9c50-993a-4ade-9001-df91b492e893-kube-api-access-smvjp" (OuterVolumeSpecName: "kube-api-access-smvjp") pod "8c0d9c50-993a-4ade-9001-df91b492e893" (UID: "8c0d9c50-993a-4ade-9001-df91b492e893"). InnerVolumeSpecName "kube-api-access-smvjp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:10:32 crc kubenswrapper[4907]: I0313 14:10:32.679211 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/072a25a8-cae1-4572-846b-2d2feea5fb36-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "072a25a8-cae1-4572-846b-2d2feea5fb36" (UID: "072a25a8-cae1-4572-846b-2d2feea5fb36"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:10:32 crc kubenswrapper[4907]: I0313 14:10:32.679320 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c0d9c50-993a-4ade-9001-df91b492e893-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8c0d9c50-993a-4ade-9001-df91b492e893" (UID: "8c0d9c50-993a-4ade-9001-df91b492e893"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:10:32 crc kubenswrapper[4907]: I0313 14:10:32.679390 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/072a25a8-cae1-4572-846b-2d2feea5fb36-kube-api-access-429xt" (OuterVolumeSpecName: "kube-api-access-429xt") pod "072a25a8-cae1-4572-846b-2d2feea5fb36" (UID: "072a25a8-cae1-4572-846b-2d2feea5fb36"). InnerVolumeSpecName "kube-api-access-429xt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:10:32 crc kubenswrapper[4907]: I0313 14:10:32.721579 4907 generic.go:334] "Generic (PLEG): container finished" podID="8c0d9c50-993a-4ade-9001-df91b492e893" containerID="9f7688b312e9d673807fef17daaf30d9c92b2663246998695a6ef3925309315a" exitCode=0 Mar 13 14:10:32 crc kubenswrapper[4907]: I0313 14:10:32.721661 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-596c74884f-nxqmn" event={"ID":"8c0d9c50-993a-4ade-9001-df91b492e893","Type":"ContainerDied","Data":"9f7688b312e9d673807fef17daaf30d9c92b2663246998695a6ef3925309315a"} Mar 13 14:10:32 crc kubenswrapper[4907]: I0313 14:10:32.721697 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-596c74884f-nxqmn" event={"ID":"8c0d9c50-993a-4ade-9001-df91b492e893","Type":"ContainerDied","Data":"fe5db9115bda6db0c26db9bf40f10d6401ffba671bbe917f6eae2f4ff07fa2ca"} Mar 13 14:10:32 crc kubenswrapper[4907]: I0313 14:10:32.721721 4907 scope.go:117] "RemoveContainer" containerID="9f7688b312e9d673807fef17daaf30d9c92b2663246998695a6ef3925309315a" Mar 13 14:10:32 crc kubenswrapper[4907]: I0313 14:10:32.721818 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-596c74884f-nxqmn" Mar 13 14:10:32 crc kubenswrapper[4907]: I0313 14:10:32.727960 4907 generic.go:334] "Generic (PLEG): container finished" podID="072a25a8-cae1-4572-846b-2d2feea5fb36" containerID="3251605f371bf11c4b8c99a286c7e7a5947870c25059d47601917b31fe721247" exitCode=0 Mar 13 14:10:32 crc kubenswrapper[4907]: I0313 14:10:32.727992 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6797bc5cfb-75vkh" event={"ID":"072a25a8-cae1-4572-846b-2d2feea5fb36","Type":"ContainerDied","Data":"3251605f371bf11c4b8c99a286c7e7a5947870c25059d47601917b31fe721247"} Mar 13 14:10:32 crc kubenswrapper[4907]: I0313 14:10:32.728029 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6797bc5cfb-75vkh" event={"ID":"072a25a8-cae1-4572-846b-2d2feea5fb36","Type":"ContainerDied","Data":"381bee78bac3d3aef3f83d5ea311981cbe79eee1108c280c12eebed6b19e7330"} Mar 13 14:10:32 crc kubenswrapper[4907]: I0313 14:10:32.728067 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6797bc5cfb-75vkh" Mar 13 14:10:32 crc kubenswrapper[4907]: I0313 14:10:32.757689 4907 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8c0d9c50-993a-4ade-9001-df91b492e893-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 14:10:32 crc kubenswrapper[4907]: I0313 14:10:32.757715 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-smvjp\" (UniqueName: \"kubernetes.io/projected/8c0d9c50-993a-4ade-9001-df91b492e893-kube-api-access-smvjp\") on node \"crc\" DevicePath \"\"" Mar 13 14:10:32 crc kubenswrapper[4907]: I0313 14:10:32.757725 4907 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/072a25a8-cae1-4572-846b-2d2feea5fb36-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 13 14:10:32 crc kubenswrapper[4907]: I0313 14:10:32.757733 4907 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/072a25a8-cae1-4572-846b-2d2feea5fb36-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 14:10:32 crc kubenswrapper[4907]: I0313 14:10:32.757744 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-429xt\" (UniqueName: \"kubernetes.io/projected/072a25a8-cae1-4572-846b-2d2feea5fb36-kube-api-access-429xt\") on node \"crc\" DevicePath \"\"" Mar 13 14:10:32 crc kubenswrapper[4907]: I0313 14:10:32.757753 4907 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8c0d9c50-993a-4ade-9001-df91b492e893-client-ca\") on node \"crc\" DevicePath \"\"" Mar 13 14:10:32 crc kubenswrapper[4907]: I0313 14:10:32.757761 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8c0d9c50-993a-4ade-9001-df91b492e893-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:10:32 crc kubenswrapper[4907]: I0313 14:10:32.757770 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/072a25a8-cae1-4572-846b-2d2feea5fb36-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:10:32 crc kubenswrapper[4907]: I0313 14:10:32.757777 4907 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/072a25a8-cae1-4572-846b-2d2feea5fb36-client-ca\") on node \"crc\" DevicePath \"\"" Mar 13 14:10:32 crc kubenswrapper[4907]: I0313 14:10:32.760520 4907 scope.go:117] "RemoveContainer" containerID="9f7688b312e9d673807fef17daaf30d9c92b2663246998695a6ef3925309315a" Mar 13 14:10:32 crc kubenswrapper[4907]: E0313 14:10:32.762434 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9f7688b312e9d673807fef17daaf30d9c92b2663246998695a6ef3925309315a\": container with ID starting with 9f7688b312e9d673807fef17daaf30d9c92b2663246998695a6ef3925309315a not found: ID does not exist" containerID="9f7688b312e9d673807fef17daaf30d9c92b2663246998695a6ef3925309315a" Mar 13 14:10:32 crc kubenswrapper[4907]: I0313 14:10:32.762476 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9f7688b312e9d673807fef17daaf30d9c92b2663246998695a6ef3925309315a"} err="failed to get container status \"9f7688b312e9d673807fef17daaf30d9c92b2663246998695a6ef3925309315a\": rpc error: code = NotFound desc = could not find container \"9f7688b312e9d673807fef17daaf30d9c92b2663246998695a6ef3925309315a\": container with ID starting with 9f7688b312e9d673807fef17daaf30d9c92b2663246998695a6ef3925309315a not found: ID does not exist" Mar 13 14:10:32 crc kubenswrapper[4907]: I0313 14:10:32.762505 4907 scope.go:117] "RemoveContainer" containerID="3251605f371bf11c4b8c99a286c7e7a5947870c25059d47601917b31fe721247" Mar 13 14:10:32 crc kubenswrapper[4907]: I0313 14:10:32.784400 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-6797bc5cfb-75vkh"] Mar 13 14:10:32 crc kubenswrapper[4907]: I0313 14:10:32.788136 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-6797bc5cfb-75vkh"] Mar 13 14:10:32 crc kubenswrapper[4907]: I0313 14:10:32.788227 4907 scope.go:117] "RemoveContainer" containerID="3251605f371bf11c4b8c99a286c7e7a5947870c25059d47601917b31fe721247" Mar 13 14:10:32 crc kubenswrapper[4907]: E0313 14:10:32.788630 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3251605f371bf11c4b8c99a286c7e7a5947870c25059d47601917b31fe721247\": container with ID starting with 3251605f371bf11c4b8c99a286c7e7a5947870c25059d47601917b31fe721247 not found: ID does not exist" containerID="3251605f371bf11c4b8c99a286c7e7a5947870c25059d47601917b31fe721247" Mar 13 14:10:32 crc kubenswrapper[4907]: I0313 14:10:32.788660 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3251605f371bf11c4b8c99a286c7e7a5947870c25059d47601917b31fe721247"} err="failed to get container status \"3251605f371bf11c4b8c99a286c7e7a5947870c25059d47601917b31fe721247\": rpc error: code = NotFound desc = could not find container \"3251605f371bf11c4b8c99a286c7e7a5947870c25059d47601917b31fe721247\": container with ID starting with 3251605f371bf11c4b8c99a286c7e7a5947870c25059d47601917b31fe721247 not found: ID does not exist" Mar 13 14:10:32 crc kubenswrapper[4907]: I0313 14:10:32.792947 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-596c74884f-nxqmn"] Mar 13 14:10:32 crc kubenswrapper[4907]: I0313 14:10:32.795829 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-596c74884f-nxqmn"] Mar 13 14:10:32 crc kubenswrapper[4907]: I0313 14:10:32.819782 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-gjdp7"] Mar 13 14:10:32 crc kubenswrapper[4907]: I0313 14:10:32.820013 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-gjdp7" podUID="5ef16085-0f4d-435d-a559-20a33c237531" containerName="registry-server" containerID="cri-o://c851c33bb27b562ca9f87e1a35a7d7576757ef40b9f02ebed142608872748069" gracePeriod=2 Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.034195 4907 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.034755 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://d4b503aa3b13268b3688d067f3e456c86afac1fda171aced0990899c3eb89b79" gracePeriod=15 Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.034823 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://93d8c8c79be9d9ffb72de66fc2f3f9821936023071e95875582b45ccb319601f" gracePeriod=15 Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.034971 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://e3f2b6756078925f8562b6053b3b83798efb72288f5fd18fff680a4ab0d33feb" gracePeriod=15 Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.035051 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://be40a1a33fa482eab363537afaa901127d560afbc1b5b287ef11e66391a5be4c" gracePeriod=15 Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.035143 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://1d1b539bbd918fc50428fc18cabcc4c8a43cc96f220b5692d67cf1c8dfbf8ed8" gracePeriod=15 Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.036070 4907 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Mar 13 14:10:33 crc kubenswrapper[4907]: E0313 14:10:33.036329 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.036347 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Mar 13 14:10:33 crc kubenswrapper[4907]: E0313 14:10:33.036356 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.036363 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 13 14:10:33 crc kubenswrapper[4907]: E0313 14:10:33.036373 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8779a406-c08e-40db-b25d-8c45e784e94d" containerName="extract-content" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.036379 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="8779a406-c08e-40db-b25d-8c45e784e94d" containerName="extract-content" Mar 13 14:10:33 crc kubenswrapper[4907]: E0313 14:10:33.036387 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="775414ea-5b94-4ab6-8b7c-fe7c80671ced" containerName="extract-content" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.036393 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="775414ea-5b94-4ab6-8b7c-fe7c80671ced" containerName="extract-content" Mar 13 14:10:33 crc kubenswrapper[4907]: E0313 14:10:33.036400 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.036405 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Mar 13 14:10:33 crc kubenswrapper[4907]: E0313 14:10:33.036414 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8779a406-c08e-40db-b25d-8c45e784e94d" containerName="extract-utilities" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.036419 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="8779a406-c08e-40db-b25d-8c45e784e94d" containerName="extract-utilities" Mar 13 14:10:33 crc kubenswrapper[4907]: E0313 14:10:33.036427 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="775414ea-5b94-4ab6-8b7c-fe7c80671ced" containerName="extract-utilities" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.036434 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="775414ea-5b94-4ab6-8b7c-fe7c80671ced" containerName="extract-utilities" Mar 13 14:10:33 crc kubenswrapper[4907]: E0313 14:10:33.036440 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="072a25a8-cae1-4572-846b-2d2feea5fb36" containerName="controller-manager" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.036446 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="072a25a8-cae1-4572-846b-2d2feea5fb36" containerName="controller-manager" Mar 13 14:10:33 crc kubenswrapper[4907]: E0313 14:10:33.036455 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="775414ea-5b94-4ab6-8b7c-fe7c80671ced" containerName="registry-server" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.036460 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="775414ea-5b94-4ab6-8b7c-fe7c80671ced" containerName="registry-server" Mar 13 14:10:33 crc kubenswrapper[4907]: E0313 14:10:33.036469 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8779a406-c08e-40db-b25d-8c45e784e94d" containerName="registry-server" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.036475 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="8779a406-c08e-40db-b25d-8c45e784e94d" containerName="registry-server" Mar 13 14:10:33 crc kubenswrapper[4907]: E0313 14:10:33.036483 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.036488 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Mar 13 14:10:33 crc kubenswrapper[4907]: E0313 14:10:33.036496 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c0d9c50-993a-4ade-9001-df91b492e893" containerName="route-controller-manager" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.036502 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c0d9c50-993a-4ade-9001-df91b492e893" containerName="route-controller-manager" Mar 13 14:10:33 crc kubenswrapper[4907]: E0313 14:10:33.036509 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.036514 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 13 14:10:33 crc kubenswrapper[4907]: E0313 14:10:33.036524 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.036531 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Mar 13 14:10:33 crc kubenswrapper[4907]: E0313 14:10:33.036542 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.036548 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 13 14:10:33 crc kubenswrapper[4907]: E0313 14:10:33.036560 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.036566 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 13 14:10:33 crc kubenswrapper[4907]: E0313 14:10:33.036574 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.036582 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.036669 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.036676 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c0d9c50-993a-4ade-9001-df91b492e893" containerName="route-controller-manager" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.036685 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.036694 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.036702 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.036708 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.036714 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="775414ea-5b94-4ab6-8b7c-fe7c80671ced" containerName="registry-server" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.036723 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="8779a406-c08e-40db-b25d-8c45e784e94d" containerName="registry-server" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.036730 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="072a25a8-cae1-4572-846b-2d2feea5fb36" containerName="controller-manager" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.036738 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.036744 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 13 14:10:33 crc kubenswrapper[4907]: E0313 14:10:33.036829 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.036835 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.036949 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.036959 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.038575 4907 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.039324 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.045711 4907 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="f4b27818a5e8e43d0dc095d08835c792" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" Mar 13 14:10:33 crc kubenswrapper[4907]: E0313 14:10:33.082829 4907 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.174:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.162933 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.163332 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.163364 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.163408 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.163580 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.163684 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.163766 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.164271 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.265247 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.265302 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.265376 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.265367 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.265452 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.265459 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.265398 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.265536 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.265588 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.265569 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.265698 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.265791 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.265842 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.265855 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.265998 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.266058 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.333044 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gjdp7" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.333847 4907 status_manager.go:851] "Failed to get status for pod" podUID="5ef16085-0f4d-435d-a559-20a33c237531" pod="openshift-marketplace/certified-operators-gjdp7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-gjdp7\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.383957 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 13 14:10:33 crc kubenswrapper[4907]: W0313 14:10:33.407762 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf85e55b1a89d02b0cb034b1ea31ed45a.slice/crio-a1023123e570e5f57c68d511b33b9e2a5b0ce18e77d1c18f79f77bbf5e2f961b WatchSource:0}: Error finding container a1023123e570e5f57c68d511b33b9e2a5b0ce18e77d1c18f79f77bbf5e2f961b: Status 404 returned error can't find the container with id a1023123e570e5f57c68d511b33b9e2a5b0ce18e77d1c18f79f77bbf5e2f961b Mar 13 14:10:33 crc kubenswrapper[4907]: E0313 14:10:33.412860 4907 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.174:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.189c6bf05f43c7db openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:10:33.411733467 +0000 UTC m=+332.311521186,LastTimestamp:2026-03-13 14:10:33.411733467 +0000 UTC m=+332.311521186,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.469445 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6q2rc\" (UniqueName: \"kubernetes.io/projected/5ef16085-0f4d-435d-a559-20a33c237531-kube-api-access-6q2rc\") pod \"5ef16085-0f4d-435d-a559-20a33c237531\" (UID: \"5ef16085-0f4d-435d-a559-20a33c237531\") " Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.469635 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ef16085-0f4d-435d-a559-20a33c237531-catalog-content\") pod \"5ef16085-0f4d-435d-a559-20a33c237531\" (UID: \"5ef16085-0f4d-435d-a559-20a33c237531\") " Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.469682 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ef16085-0f4d-435d-a559-20a33c237531-utilities\") pod \"5ef16085-0f4d-435d-a559-20a33c237531\" (UID: \"5ef16085-0f4d-435d-a559-20a33c237531\") " Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.472055 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5ef16085-0f4d-435d-a559-20a33c237531-utilities" (OuterVolumeSpecName: "utilities") pod "5ef16085-0f4d-435d-a559-20a33c237531" (UID: "5ef16085-0f4d-435d-a559-20a33c237531"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.475598 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ef16085-0f4d-435d-a559-20a33c237531-kube-api-access-6q2rc" (OuterVolumeSpecName: "kube-api-access-6q2rc") pod "5ef16085-0f4d-435d-a559-20a33c237531" (UID: "5ef16085-0f4d-435d-a559-20a33c237531"). InnerVolumeSpecName "kube-api-access-6q2rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.552358 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5ef16085-0f4d-435d-a559-20a33c237531-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5ef16085-0f4d-435d-a559-20a33c237531" (UID: "5ef16085-0f4d-435d-a559-20a33c237531"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.571281 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ef16085-0f4d-435d-a559-20a33c237531-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.571338 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ef16085-0f4d-435d-a559-20a33c237531-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.571361 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6q2rc\" (UniqueName: \"kubernetes.io/projected/5ef16085-0f4d-435d-a559-20a33c237531-kube-api-access-6q2rc\") on node \"crc\" DevicePath \"\"" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.750919 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.753681 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.755061 4907 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="93d8c8c79be9d9ffb72de66fc2f3f9821936023071e95875582b45ccb319601f" exitCode=0 Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.755102 4907 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="e3f2b6756078925f8562b6053b3b83798efb72288f5fd18fff680a4ab0d33feb" exitCode=0 Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.755125 4907 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="be40a1a33fa482eab363537afaa901127d560afbc1b5b287ef11e66391a5be4c" exitCode=0 Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.755140 4907 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="1d1b539bbd918fc50428fc18cabcc4c8a43cc96f220b5692d67cf1c8dfbf8ed8" exitCode=2 Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.755254 4907 scope.go:117] "RemoveContainer" containerID="1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.759771 4907 generic.go:334] "Generic (PLEG): container finished" podID="b66475a4-2e1f-4b18-b565-3cc70bb3720f" containerID="c818da15368a295079d572ed81796486cb0c9f878bd84d437fa33e0b3e3169f6" exitCode=0 Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.759837 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"b66475a4-2e1f-4b18-b565-3cc70bb3720f","Type":"ContainerDied","Data":"c818da15368a295079d572ed81796486cb0c9f878bd84d437fa33e0b3e3169f6"} Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.761443 4907 status_manager.go:851] "Failed to get status for pod" podUID="5ef16085-0f4d-435d-a559-20a33c237531" pod="openshift-marketplace/certified-operators-gjdp7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-gjdp7\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.761926 4907 status_manager.go:851] "Failed to get status for pod" podUID="b66475a4-2e1f-4b18-b565-3cc70bb3720f" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.778152 4907 generic.go:334] "Generic (PLEG): container finished" podID="5ef16085-0f4d-435d-a559-20a33c237531" containerID="c851c33bb27b562ca9f87e1a35a7d7576757ef40b9f02ebed142608872748069" exitCode=0 Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.778276 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gjdp7" event={"ID":"5ef16085-0f4d-435d-a559-20a33c237531","Type":"ContainerDied","Data":"c851c33bb27b562ca9f87e1a35a7d7576757ef40b9f02ebed142608872748069"} Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.778330 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gjdp7" event={"ID":"5ef16085-0f4d-435d-a559-20a33c237531","Type":"ContainerDied","Data":"5380ac5cf0f032bd661fec9fad25b22b2b931ad5cc5954ae70fb9c5f8362990a"} Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.778477 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gjdp7" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.782066 4907 status_manager.go:851] "Failed to get status for pod" podUID="b66475a4-2e1f-4b18-b565-3cc70bb3720f" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.783222 4907 status_manager.go:851] "Failed to get status for pod" podUID="5ef16085-0f4d-435d-a559-20a33c237531" pod="openshift-marketplace/certified-operators-gjdp7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-gjdp7\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.810455 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="072a25a8-cae1-4572-846b-2d2feea5fb36" path="/var/lib/kubelet/pods/072a25a8-cae1-4572-846b-2d2feea5fb36/volumes" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.811666 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8c0d9c50-993a-4ade-9001-df91b492e893" path="/var/lib/kubelet/pods/8c0d9c50-993a-4ade-9001-df91b492e893/volumes" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.812982 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"a1023123e570e5f57c68d511b33b9e2a5b0ce18e77d1c18f79f77bbf5e2f961b"} Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.813266 4907 scope.go:117] "RemoveContainer" containerID="c851c33bb27b562ca9f87e1a35a7d7576757ef40b9f02ebed142608872748069" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.814049 4907 status_manager.go:851] "Failed to get status for pod" podUID="b66475a4-2e1f-4b18-b565-3cc70bb3720f" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.819171 4907 status_manager.go:851] "Failed to get status for pod" podUID="5ef16085-0f4d-435d-a559-20a33c237531" pod="openshift-marketplace/certified-operators-gjdp7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-gjdp7\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.837862 4907 scope.go:117] "RemoveContainer" containerID="031c710b0cf90428d9303b494400b216776d335d2ba2176b760375d7548140a7" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.862398 4907 scope.go:117] "RemoveContainer" containerID="7e15a77e2ae6c9b87a9062d382cad40c9ca694c03f35d89b5511dd88c400a20c" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.881743 4907 scope.go:117] "RemoveContainer" containerID="c851c33bb27b562ca9f87e1a35a7d7576757ef40b9f02ebed142608872748069" Mar 13 14:10:33 crc kubenswrapper[4907]: E0313 14:10:33.882297 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c851c33bb27b562ca9f87e1a35a7d7576757ef40b9f02ebed142608872748069\": container with ID starting with c851c33bb27b562ca9f87e1a35a7d7576757ef40b9f02ebed142608872748069 not found: ID does not exist" containerID="c851c33bb27b562ca9f87e1a35a7d7576757ef40b9f02ebed142608872748069" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.882364 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c851c33bb27b562ca9f87e1a35a7d7576757ef40b9f02ebed142608872748069"} err="failed to get container status \"c851c33bb27b562ca9f87e1a35a7d7576757ef40b9f02ebed142608872748069\": rpc error: code = NotFound desc = could not find container \"c851c33bb27b562ca9f87e1a35a7d7576757ef40b9f02ebed142608872748069\": container with ID starting with c851c33bb27b562ca9f87e1a35a7d7576757ef40b9f02ebed142608872748069 not found: ID does not exist" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.882405 4907 scope.go:117] "RemoveContainer" containerID="031c710b0cf90428d9303b494400b216776d335d2ba2176b760375d7548140a7" Mar 13 14:10:33 crc kubenswrapper[4907]: E0313 14:10:33.882773 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"031c710b0cf90428d9303b494400b216776d335d2ba2176b760375d7548140a7\": container with ID starting with 031c710b0cf90428d9303b494400b216776d335d2ba2176b760375d7548140a7 not found: ID does not exist" containerID="031c710b0cf90428d9303b494400b216776d335d2ba2176b760375d7548140a7" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.882808 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"031c710b0cf90428d9303b494400b216776d335d2ba2176b760375d7548140a7"} err="failed to get container status \"031c710b0cf90428d9303b494400b216776d335d2ba2176b760375d7548140a7\": rpc error: code = NotFound desc = could not find container \"031c710b0cf90428d9303b494400b216776d335d2ba2176b760375d7548140a7\": container with ID starting with 031c710b0cf90428d9303b494400b216776d335d2ba2176b760375d7548140a7 not found: ID does not exist" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.882835 4907 scope.go:117] "RemoveContainer" containerID="7e15a77e2ae6c9b87a9062d382cad40c9ca694c03f35d89b5511dd88c400a20c" Mar 13 14:10:33 crc kubenswrapper[4907]: E0313 14:10:33.885052 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7e15a77e2ae6c9b87a9062d382cad40c9ca694c03f35d89b5511dd88c400a20c\": container with ID starting with 7e15a77e2ae6c9b87a9062d382cad40c9ca694c03f35d89b5511dd88c400a20c not found: ID does not exist" containerID="7e15a77e2ae6c9b87a9062d382cad40c9ca694c03f35d89b5511dd88c400a20c" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.885091 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7e15a77e2ae6c9b87a9062d382cad40c9ca694c03f35d89b5511dd88c400a20c"} err="failed to get container status \"7e15a77e2ae6c9b87a9062d382cad40c9ca694c03f35d89b5511dd88c400a20c\": rpc error: code = NotFound desc = could not find container \"7e15a77e2ae6c9b87a9062d382cad40c9ca694c03f35d89b5511dd88c400a20c\": container with ID starting with 7e15a77e2ae6c9b87a9062d382cad40c9ca694c03f35d89b5511dd88c400a20c not found: ID does not exist" Mar 13 14:10:34 crc kubenswrapper[4907]: E0313 14:10:34.004759 4907 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.174:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.189c6bf05f43c7db openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:10:33.411733467 +0000 UTC m=+332.311521186,LastTimestamp:2026-03-13 14:10:33.411733467 +0000 UTC m=+332.311521186,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:10:34 crc kubenswrapper[4907]: I0313 14:10:34.795083 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"d4f9add6c133d321c4423223d8d40952daa019ea189e4fd19cc3ba6806cde1d4"} Mar 13 14:10:34 crc kubenswrapper[4907]: E0313 14:10:34.796247 4907 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.174:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 13 14:10:34 crc kubenswrapper[4907]: I0313 14:10:34.796240 4907 status_manager.go:851] "Failed to get status for pod" podUID="b66475a4-2e1f-4b18-b565-3cc70bb3720f" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 13 14:10:34 crc kubenswrapper[4907]: I0313 14:10:34.797213 4907 status_manager.go:851] "Failed to get status for pod" podUID="5ef16085-0f4d-435d-a559-20a33c237531" pod="openshift-marketplace/certified-operators-gjdp7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-gjdp7\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 13 14:10:34 crc kubenswrapper[4907]: I0313 14:10:34.804172 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Mar 13 14:10:35 crc kubenswrapper[4907]: I0313 14:10:35.191036 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Mar 13 14:10:35 crc kubenswrapper[4907]: I0313 14:10:35.192061 4907 status_manager.go:851] "Failed to get status for pod" podUID="b66475a4-2e1f-4b18-b565-3cc70bb3720f" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 13 14:10:35 crc kubenswrapper[4907]: I0313 14:10:35.192280 4907 status_manager.go:851] "Failed to get status for pod" podUID="5ef16085-0f4d-435d-a559-20a33c237531" pod="openshift-marketplace/certified-operators-gjdp7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-gjdp7\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 13 14:10:35 crc kubenswrapper[4907]: I0313 14:10:35.222815 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b66475a4-2e1f-4b18-b565-3cc70bb3720f-kube-api-access\") pod \"b66475a4-2e1f-4b18-b565-3cc70bb3720f\" (UID: \"b66475a4-2e1f-4b18-b565-3cc70bb3720f\") " Mar 13 14:10:35 crc kubenswrapper[4907]: I0313 14:10:35.222975 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b66475a4-2e1f-4b18-b565-3cc70bb3720f-kubelet-dir\") pod \"b66475a4-2e1f-4b18-b565-3cc70bb3720f\" (UID: \"b66475a4-2e1f-4b18-b565-3cc70bb3720f\") " Mar 13 14:10:35 crc kubenswrapper[4907]: I0313 14:10:35.223039 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/b66475a4-2e1f-4b18-b565-3cc70bb3720f-var-lock\") pod \"b66475a4-2e1f-4b18-b565-3cc70bb3720f\" (UID: \"b66475a4-2e1f-4b18-b565-3cc70bb3720f\") " Mar 13 14:10:35 crc kubenswrapper[4907]: I0313 14:10:35.223404 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b66475a4-2e1f-4b18-b565-3cc70bb3720f-var-lock" (OuterVolumeSpecName: "var-lock") pod "b66475a4-2e1f-4b18-b565-3cc70bb3720f" (UID: "b66475a4-2e1f-4b18-b565-3cc70bb3720f"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 14:10:35 crc kubenswrapper[4907]: I0313 14:10:35.224521 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b66475a4-2e1f-4b18-b565-3cc70bb3720f-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "b66475a4-2e1f-4b18-b565-3cc70bb3720f" (UID: "b66475a4-2e1f-4b18-b565-3cc70bb3720f"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 14:10:35 crc kubenswrapper[4907]: I0313 14:10:35.230310 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b66475a4-2e1f-4b18-b565-3cc70bb3720f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "b66475a4-2e1f-4b18-b565-3cc70bb3720f" (UID: "b66475a4-2e1f-4b18-b565-3cc70bb3720f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:10:35 crc kubenswrapper[4907]: I0313 14:10:35.324745 4907 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b66475a4-2e1f-4b18-b565-3cc70bb3720f-kubelet-dir\") on node \"crc\" DevicePath \"\"" Mar 13 14:10:35 crc kubenswrapper[4907]: I0313 14:10:35.324950 4907 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/b66475a4-2e1f-4b18-b565-3cc70bb3720f-var-lock\") on node \"crc\" DevicePath \"\"" Mar 13 14:10:35 crc kubenswrapper[4907]: I0313 14:10:35.325058 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b66475a4-2e1f-4b18-b565-3cc70bb3720f-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 13 14:10:35 crc kubenswrapper[4907]: I0313 14:10:35.427904 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Mar 13 14:10:35 crc kubenswrapper[4907]: I0313 14:10:35.428831 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 14:10:35 crc kubenswrapper[4907]: I0313 14:10:35.429577 4907 status_manager.go:851] "Failed to get status for pod" podUID="b66475a4-2e1f-4b18-b565-3cc70bb3720f" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 13 14:10:35 crc kubenswrapper[4907]: I0313 14:10:35.429930 4907 status_manager.go:851] "Failed to get status for pod" podUID="5ef16085-0f4d-435d-a559-20a33c237531" pod="openshift-marketplace/certified-operators-gjdp7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-gjdp7\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 13 14:10:35 crc kubenswrapper[4907]: I0313 14:10:35.430923 4907 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 13 14:10:35 crc kubenswrapper[4907]: I0313 14:10:35.527520 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Mar 13 14:10:35 crc kubenswrapper[4907]: I0313 14:10:35.527718 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Mar 13 14:10:35 crc kubenswrapper[4907]: I0313 14:10:35.527780 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Mar 13 14:10:35 crc kubenswrapper[4907]: I0313 14:10:35.528091 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 14:10:35 crc kubenswrapper[4907]: I0313 14:10:35.528301 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 14:10:35 crc kubenswrapper[4907]: I0313 14:10:35.528555 4907 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Mar 13 14:10:35 crc kubenswrapper[4907]: I0313 14:10:35.528587 4907 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Mar 13 14:10:35 crc kubenswrapper[4907]: I0313 14:10:35.528335 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 14:10:35 crc kubenswrapper[4907]: I0313 14:10:35.629636 4907 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Mar 13 14:10:35 crc kubenswrapper[4907]: I0313 14:10:35.795991 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Mar 13 14:10:35 crc kubenswrapper[4907]: I0313 14:10:35.819172 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Mar 13 14:10:35 crc kubenswrapper[4907]: I0313 14:10:35.820608 4907 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="d4b503aa3b13268b3688d067f3e456c86afac1fda171aced0990899c3eb89b79" exitCode=0 Mar 13 14:10:35 crc kubenswrapper[4907]: I0313 14:10:35.820734 4907 scope.go:117] "RemoveContainer" containerID="93d8c8c79be9d9ffb72de66fc2f3f9821936023071e95875582b45ccb319601f" Mar 13 14:10:35 crc kubenswrapper[4907]: I0313 14:10:35.821456 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 14:10:35 crc kubenswrapper[4907]: I0313 14:10:35.822583 4907 status_manager.go:851] "Failed to get status for pod" podUID="b66475a4-2e1f-4b18-b565-3cc70bb3720f" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 13 14:10:35 crc kubenswrapper[4907]: I0313 14:10:35.823711 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"b66475a4-2e1f-4b18-b565-3cc70bb3720f","Type":"ContainerDied","Data":"a68e3ade1d545da2c68a88db1974b5d105f932a3f0f6695c49e0be2ffa9a3a68"} Mar 13 14:10:35 crc kubenswrapper[4907]: I0313 14:10:35.823755 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a68e3ade1d545da2c68a88db1974b5d105f932a3f0f6695c49e0be2ffa9a3a68" Mar 13 14:10:35 crc kubenswrapper[4907]: I0313 14:10:35.823727 4907 status_manager.go:851] "Failed to get status for pod" podUID="5ef16085-0f4d-435d-a559-20a33c237531" pod="openshift-marketplace/certified-operators-gjdp7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-gjdp7\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 13 14:10:35 crc kubenswrapper[4907]: I0313 14:10:35.824635 4907 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 13 14:10:35 crc kubenswrapper[4907]: I0313 14:10:35.826270 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Mar 13 14:10:35 crc kubenswrapper[4907]: E0313 14:10:35.826520 4907 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.174:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 13 14:10:35 crc kubenswrapper[4907]: I0313 14:10:35.843653 4907 status_manager.go:851] "Failed to get status for pod" podUID="b66475a4-2e1f-4b18-b565-3cc70bb3720f" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 13 14:10:35 crc kubenswrapper[4907]: I0313 14:10:35.844062 4907 status_manager.go:851] "Failed to get status for pod" podUID="5ef16085-0f4d-435d-a559-20a33c237531" pod="openshift-marketplace/certified-operators-gjdp7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-gjdp7\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 13 14:10:35 crc kubenswrapper[4907]: I0313 14:10:35.844583 4907 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 13 14:10:35 crc kubenswrapper[4907]: I0313 14:10:35.845598 4907 status_manager.go:851] "Failed to get status for pod" podUID="b66475a4-2e1f-4b18-b565-3cc70bb3720f" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 13 14:10:35 crc kubenswrapper[4907]: I0313 14:10:35.846431 4907 status_manager.go:851] "Failed to get status for pod" podUID="5ef16085-0f4d-435d-a559-20a33c237531" pod="openshift-marketplace/certified-operators-gjdp7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-gjdp7\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 13 14:10:35 crc kubenswrapper[4907]: I0313 14:10:35.847028 4907 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 13 14:10:35 crc kubenswrapper[4907]: I0313 14:10:35.854222 4907 scope.go:117] "RemoveContainer" containerID="e3f2b6756078925f8562b6053b3b83798efb72288f5fd18fff680a4ab0d33feb" Mar 13 14:10:35 crc kubenswrapper[4907]: I0313 14:10:35.878263 4907 scope.go:117] "RemoveContainer" containerID="be40a1a33fa482eab363537afaa901127d560afbc1b5b287ef11e66391a5be4c" Mar 13 14:10:35 crc kubenswrapper[4907]: I0313 14:10:35.898973 4907 scope.go:117] "RemoveContainer" containerID="1d1b539bbd918fc50428fc18cabcc4c8a43cc96f220b5692d67cf1c8dfbf8ed8" Mar 13 14:10:35 crc kubenswrapper[4907]: I0313 14:10:35.920814 4907 scope.go:117] "RemoveContainer" containerID="d4b503aa3b13268b3688d067f3e456c86afac1fda171aced0990899c3eb89b79" Mar 13 14:10:35 crc kubenswrapper[4907]: I0313 14:10:35.941726 4907 scope.go:117] "RemoveContainer" containerID="60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d" Mar 13 14:10:35 crc kubenswrapper[4907]: I0313 14:10:35.979594 4907 scope.go:117] "RemoveContainer" containerID="93d8c8c79be9d9ffb72de66fc2f3f9821936023071e95875582b45ccb319601f" Mar 13 14:10:35 crc kubenswrapper[4907]: E0313 14:10:35.980636 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"93d8c8c79be9d9ffb72de66fc2f3f9821936023071e95875582b45ccb319601f\": container with ID starting with 93d8c8c79be9d9ffb72de66fc2f3f9821936023071e95875582b45ccb319601f not found: ID does not exist" containerID="93d8c8c79be9d9ffb72de66fc2f3f9821936023071e95875582b45ccb319601f" Mar 13 14:10:35 crc kubenswrapper[4907]: I0313 14:10:35.981061 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"93d8c8c79be9d9ffb72de66fc2f3f9821936023071e95875582b45ccb319601f"} err="failed to get container status \"93d8c8c79be9d9ffb72de66fc2f3f9821936023071e95875582b45ccb319601f\": rpc error: code = NotFound desc = could not find container \"93d8c8c79be9d9ffb72de66fc2f3f9821936023071e95875582b45ccb319601f\": container with ID starting with 93d8c8c79be9d9ffb72de66fc2f3f9821936023071e95875582b45ccb319601f not found: ID does not exist" Mar 13 14:10:35 crc kubenswrapper[4907]: I0313 14:10:35.981365 4907 scope.go:117] "RemoveContainer" containerID="e3f2b6756078925f8562b6053b3b83798efb72288f5fd18fff680a4ab0d33feb" Mar 13 14:10:35 crc kubenswrapper[4907]: E0313 14:10:35.982355 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e3f2b6756078925f8562b6053b3b83798efb72288f5fd18fff680a4ab0d33feb\": container with ID starting with e3f2b6756078925f8562b6053b3b83798efb72288f5fd18fff680a4ab0d33feb not found: ID does not exist" containerID="e3f2b6756078925f8562b6053b3b83798efb72288f5fd18fff680a4ab0d33feb" Mar 13 14:10:35 crc kubenswrapper[4907]: I0313 14:10:35.982417 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e3f2b6756078925f8562b6053b3b83798efb72288f5fd18fff680a4ab0d33feb"} err="failed to get container status \"e3f2b6756078925f8562b6053b3b83798efb72288f5fd18fff680a4ab0d33feb\": rpc error: code = NotFound desc = could not find container \"e3f2b6756078925f8562b6053b3b83798efb72288f5fd18fff680a4ab0d33feb\": container with ID starting with e3f2b6756078925f8562b6053b3b83798efb72288f5fd18fff680a4ab0d33feb not found: ID does not exist" Mar 13 14:10:35 crc kubenswrapper[4907]: I0313 14:10:35.982455 4907 scope.go:117] "RemoveContainer" containerID="be40a1a33fa482eab363537afaa901127d560afbc1b5b287ef11e66391a5be4c" Mar 13 14:10:35 crc kubenswrapper[4907]: E0313 14:10:35.983395 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"be40a1a33fa482eab363537afaa901127d560afbc1b5b287ef11e66391a5be4c\": container with ID starting with be40a1a33fa482eab363537afaa901127d560afbc1b5b287ef11e66391a5be4c not found: ID does not exist" containerID="be40a1a33fa482eab363537afaa901127d560afbc1b5b287ef11e66391a5be4c" Mar 13 14:10:35 crc kubenswrapper[4907]: I0313 14:10:35.983441 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"be40a1a33fa482eab363537afaa901127d560afbc1b5b287ef11e66391a5be4c"} err="failed to get container status \"be40a1a33fa482eab363537afaa901127d560afbc1b5b287ef11e66391a5be4c\": rpc error: code = NotFound desc = could not find container \"be40a1a33fa482eab363537afaa901127d560afbc1b5b287ef11e66391a5be4c\": container with ID starting with be40a1a33fa482eab363537afaa901127d560afbc1b5b287ef11e66391a5be4c not found: ID does not exist" Mar 13 14:10:35 crc kubenswrapper[4907]: I0313 14:10:35.983471 4907 scope.go:117] "RemoveContainer" containerID="1d1b539bbd918fc50428fc18cabcc4c8a43cc96f220b5692d67cf1c8dfbf8ed8" Mar 13 14:10:35 crc kubenswrapper[4907]: E0313 14:10:35.983840 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1d1b539bbd918fc50428fc18cabcc4c8a43cc96f220b5692d67cf1c8dfbf8ed8\": container with ID starting with 1d1b539bbd918fc50428fc18cabcc4c8a43cc96f220b5692d67cf1c8dfbf8ed8 not found: ID does not exist" containerID="1d1b539bbd918fc50428fc18cabcc4c8a43cc96f220b5692d67cf1c8dfbf8ed8" Mar 13 14:10:35 crc kubenswrapper[4907]: I0313 14:10:35.983929 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d1b539bbd918fc50428fc18cabcc4c8a43cc96f220b5692d67cf1c8dfbf8ed8"} err="failed to get container status \"1d1b539bbd918fc50428fc18cabcc4c8a43cc96f220b5692d67cf1c8dfbf8ed8\": rpc error: code = NotFound desc = could not find container \"1d1b539bbd918fc50428fc18cabcc4c8a43cc96f220b5692d67cf1c8dfbf8ed8\": container with ID starting with 1d1b539bbd918fc50428fc18cabcc4c8a43cc96f220b5692d67cf1c8dfbf8ed8 not found: ID does not exist" Mar 13 14:10:35 crc kubenswrapper[4907]: I0313 14:10:35.983956 4907 scope.go:117] "RemoveContainer" containerID="d4b503aa3b13268b3688d067f3e456c86afac1fda171aced0990899c3eb89b79" Mar 13 14:10:35 crc kubenswrapper[4907]: E0313 14:10:35.984506 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d4b503aa3b13268b3688d067f3e456c86afac1fda171aced0990899c3eb89b79\": container with ID starting with d4b503aa3b13268b3688d067f3e456c86afac1fda171aced0990899c3eb89b79 not found: ID does not exist" containerID="d4b503aa3b13268b3688d067f3e456c86afac1fda171aced0990899c3eb89b79" Mar 13 14:10:35 crc kubenswrapper[4907]: I0313 14:10:35.984541 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d4b503aa3b13268b3688d067f3e456c86afac1fda171aced0990899c3eb89b79"} err="failed to get container status \"d4b503aa3b13268b3688d067f3e456c86afac1fda171aced0990899c3eb89b79\": rpc error: code = NotFound desc = could not find container \"d4b503aa3b13268b3688d067f3e456c86afac1fda171aced0990899c3eb89b79\": container with ID starting with d4b503aa3b13268b3688d067f3e456c86afac1fda171aced0990899c3eb89b79 not found: ID does not exist" Mar 13 14:10:35 crc kubenswrapper[4907]: I0313 14:10:35.984566 4907 scope.go:117] "RemoveContainer" containerID="60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d" Mar 13 14:10:35 crc kubenswrapper[4907]: E0313 14:10:35.984967 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\": container with ID starting with 60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d not found: ID does not exist" containerID="60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d" Mar 13 14:10:35 crc kubenswrapper[4907]: I0313 14:10:35.985014 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d"} err="failed to get container status \"60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\": rpc error: code = NotFound desc = could not find container \"60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\": container with ID starting with 60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d not found: ID does not exist" Mar 13 14:10:40 crc kubenswrapper[4907]: E0313 14:10:40.950321 4907 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 13 14:10:40 crc kubenswrapper[4907]: E0313 14:10:40.951422 4907 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 13 14:10:40 crc kubenswrapper[4907]: E0313 14:10:40.951825 4907 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 13 14:10:40 crc kubenswrapper[4907]: E0313 14:10:40.952318 4907 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 13 14:10:40 crc kubenswrapper[4907]: E0313 14:10:40.952744 4907 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 13 14:10:40 crc kubenswrapper[4907]: I0313 14:10:40.952785 4907 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Mar 13 14:10:40 crc kubenswrapper[4907]: E0313 14:10:40.953228 4907 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.174:6443: connect: connection refused" interval="200ms" Mar 13 14:10:41 crc kubenswrapper[4907]: E0313 14:10:41.154932 4907 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.174:6443: connect: connection refused" interval="400ms" Mar 13 14:10:41 crc kubenswrapper[4907]: E0313 14:10:41.556542 4907 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.174:6443: connect: connection refused" interval="800ms" Mar 13 14:10:41 crc kubenswrapper[4907]: I0313 14:10:41.787197 4907 status_manager.go:851] "Failed to get status for pod" podUID="b66475a4-2e1f-4b18-b565-3cc70bb3720f" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 13 14:10:41 crc kubenswrapper[4907]: I0313 14:10:41.787821 4907 status_manager.go:851] "Failed to get status for pod" podUID="5ef16085-0f4d-435d-a559-20a33c237531" pod="openshift-marketplace/certified-operators-gjdp7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-gjdp7\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 13 14:10:42 crc kubenswrapper[4907]: E0313 14:10:42.357008 4907 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.174:6443: connect: connection refused" interval="1.6s" Mar 13 14:10:42 crc kubenswrapper[4907]: E0313 14:10:42.780290 4907 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:10:42Z\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:10:42Z\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:10:42Z\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:10:42Z\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"}]}}\" for node \"crc\": Patch \"https://api-int.crc.testing:6443/api/v1/nodes/crc/status?timeout=10s\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 13 14:10:42 crc kubenswrapper[4907]: E0313 14:10:42.780651 4907 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 13 14:10:42 crc kubenswrapper[4907]: E0313 14:10:42.781137 4907 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 13 14:10:42 crc kubenswrapper[4907]: E0313 14:10:42.781398 4907 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 13 14:10:42 crc kubenswrapper[4907]: E0313 14:10:42.781975 4907 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 13 14:10:42 crc kubenswrapper[4907]: E0313 14:10:42.782041 4907 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 13 14:10:42 crc kubenswrapper[4907]: I0313 14:10:42.940052 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:10:42 crc kubenswrapper[4907]: I0313 14:10:42.940100 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:10:42 crc kubenswrapper[4907]: I0313 14:10:42.940171 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:10:42 crc kubenswrapper[4907]: I0313 14:10:42.940211 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:10:42 crc kubenswrapper[4907]: W0313 14:10:42.940656 4907 reflector.go:561] object-"openshift-network-diagnostics"/"kube-root-ca.crt": failed to list *v1.ConfigMap: Get "https://api-int.crc.testing:6443/api/v1/namespaces/openshift-network-diagnostics/configmaps?fieldSelector=metadata.name%3Dkube-root-ca.crt&resourceVersion=27253": dial tcp 38.102.83.174:6443: connect: connection refused Mar 13 14:10:42 crc kubenswrapper[4907]: E0313 14:10:42.940713 4907 reflector.go:158] "Unhandled Error" err="object-\"openshift-network-diagnostics\"/\"kube-root-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-network-diagnostics/configmaps?fieldSelector=metadata.name%3Dkube-root-ca.crt&resourceVersion=27253\": dial tcp 38.102.83.174:6443: connect: connection refused" logger="UnhandledError" Mar 13 14:10:42 crc kubenswrapper[4907]: W0313 14:10:42.940738 4907 reflector.go:561] object-"openshift-network-console"/"networking-console-plugin-cert": failed to list *v1.Secret: Get "https://api-int.crc.testing:6443/api/v1/namespaces/openshift-network-console/secrets?fieldSelector=metadata.name%3Dnetworking-console-plugin-cert&resourceVersion=27253": dial tcp 38.102.83.174:6443: connect: connection refused Mar 13 14:10:42 crc kubenswrapper[4907]: E0313 14:10:42.940870 4907 reflector.go:158] "Unhandled Error" err="object-\"openshift-network-console\"/\"networking-console-plugin-cert\": Failed to watch *v1.Secret: failed to list *v1.Secret: Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-network-console/secrets?fieldSelector=metadata.name%3Dnetworking-console-plugin-cert&resourceVersion=27253\": dial tcp 38.102.83.174:6443: connect: connection refused" logger="UnhandledError" Mar 13 14:10:42 crc kubenswrapper[4907]: W0313 14:10:42.941145 4907 reflector.go:561] object-"openshift-network-console"/"networking-console-plugin": failed to list *v1.ConfigMap: Get "https://api-int.crc.testing:6443/api/v1/namespaces/openshift-network-console/configmaps?fieldSelector=metadata.name%3Dnetworking-console-plugin&resourceVersion=27253": dial tcp 38.102.83.174:6443: connect: connection refused Mar 13 14:10:42 crc kubenswrapper[4907]: E0313 14:10:42.941259 4907 reflector.go:158] "Unhandled Error" err="object-\"openshift-network-console\"/\"networking-console-plugin\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-network-console/configmaps?fieldSelector=metadata.name%3Dnetworking-console-plugin&resourceVersion=27253\": dial tcp 38.102.83.174:6443: connect: connection refused" logger="UnhandledError" Mar 13 14:10:43 crc kubenswrapper[4907]: I0313 14:10:43.781742 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 14:10:43 crc kubenswrapper[4907]: I0313 14:10:43.783008 4907 status_manager.go:851] "Failed to get status for pod" podUID="b66475a4-2e1f-4b18-b565-3cc70bb3720f" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 13 14:10:43 crc kubenswrapper[4907]: I0313 14:10:43.783418 4907 status_manager.go:851] "Failed to get status for pod" podUID="5ef16085-0f4d-435d-a559-20a33c237531" pod="openshift-marketplace/certified-operators-gjdp7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-gjdp7\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 13 14:10:43 crc kubenswrapper[4907]: I0313 14:10:43.807087 4907 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="c5d2b054-0ed8-419c-b731-f14d3598372f" Mar 13 14:10:43 crc kubenswrapper[4907]: I0313 14:10:43.807123 4907 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="c5d2b054-0ed8-419c-b731-f14d3598372f" Mar 13 14:10:43 crc kubenswrapper[4907]: E0313 14:10:43.807360 4907 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 14:10:43 crc kubenswrapper[4907]: I0313 14:10:43.807920 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 14:10:43 crc kubenswrapper[4907]: I0313 14:10:43.863594 4907 pod_container_manager_linux.go:210] "Failed to delete cgroup paths" cgroupName=["kubepods","burstable","pod4b6ddc05-e51d-4648-84c6-6d1414398a4b"] err="unable to destroy cgroup paths for cgroup [kubepods burstable pod4b6ddc05-e51d-4648-84c6-6d1414398a4b] : Timed out while waiting for systemd to remove kubepods-burstable-pod4b6ddc05_e51d_4648_84c6_6d1414398a4b.slice" Mar 13 14:10:43 crc kubenswrapper[4907]: I0313 14:10:43.885235 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"6b2431eedeb3a3f40d4173ed9069eeef6a78c892216287534c57a19ae6b0eec9"} Mar 13 14:10:43 crc kubenswrapper[4907]: E0313 14:10:43.940679 4907 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: failed to sync secret cache: timed out waiting for the condition Mar 13 14:10:43 crc kubenswrapper[4907]: E0313 14:10:43.941176 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-13 14:12:45.941147757 +0000 UTC m=+464.840935456 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : failed to sync secret cache: timed out waiting for the condition Mar 13 14:10:43 crc kubenswrapper[4907]: E0313 14:10:43.940714 4907 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Mar 13 14:10:43 crc kubenswrapper[4907]: E0313 14:10:43.940730 4907 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Mar 13 14:10:43 crc kubenswrapper[4907]: E0313 14:10:43.940796 4907 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: failed to sync configmap cache: timed out waiting for the condition Mar 13 14:10:43 crc kubenswrapper[4907]: E0313 14:10:43.941341 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-13 14:12:45.941324112 +0000 UTC m=+464.841111801 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : failed to sync configmap cache: timed out waiting for the condition Mar 13 14:10:43 crc kubenswrapper[4907]: W0313 14:10:43.941703 4907 reflector.go:561] object-"openshift-network-diagnostics"/"openshift-service-ca.crt": failed to list *v1.ConfigMap: Get "https://api-int.crc.testing:6443/api/v1/namespaces/openshift-network-diagnostics/configmaps?fieldSelector=metadata.name%3Dopenshift-service-ca.crt&resourceVersion=27253": dial tcp 38.102.83.174:6443: connect: connection refused Mar 13 14:10:43 crc kubenswrapper[4907]: E0313 14:10:43.941797 4907 reflector.go:158] "Unhandled Error" err="object-\"openshift-network-diagnostics\"/\"openshift-service-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-network-diagnostics/configmaps?fieldSelector=metadata.name%3Dopenshift-service-ca.crt&resourceVersion=27253\": dial tcp 38.102.83.174:6443: connect: connection refused" logger="UnhandledError" Mar 13 14:10:43 crc kubenswrapper[4907]: E0313 14:10:43.957766 4907 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.174:6443: connect: connection refused" interval="3.2s" Mar 13 14:10:44 crc kubenswrapper[4907]: E0313 14:10:44.005611 4907 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.174:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.189c6bf05f43c7db openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:10:33.411733467 +0000 UTC m=+332.311521186,LastTimestamp:2026-03-13 14:10:33.411733467 +0000 UTC m=+332.311521186,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:10:44 crc kubenswrapper[4907]: E0313 14:10:44.844673 4907 desired_state_of_world_populator.go:312] "Error processing volume" err="error processing PVC openshift-image-registry/crc-image-registry-storage: failed to fetch PVC from API server: Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-image-registry/persistentvolumeclaims/crc-image-registry-storage\": dial tcp 38.102.83.174:6443: connect: connection refused" pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" volumeName="registry-storage" Mar 13 14:10:44 crc kubenswrapper[4907]: I0313 14:10:44.896281 4907 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="ef2c70913dd6f3338ee16f76a9023591596215ccf7134159f4cb7341cffbf0c1" exitCode=0 Mar 13 14:10:44 crc kubenswrapper[4907]: I0313 14:10:44.896387 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"ef2c70913dd6f3338ee16f76a9023591596215ccf7134159f4cb7341cffbf0c1"} Mar 13 14:10:44 crc kubenswrapper[4907]: I0313 14:10:44.897002 4907 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="c5d2b054-0ed8-419c-b731-f14d3598372f" Mar 13 14:10:44 crc kubenswrapper[4907]: I0313 14:10:44.897056 4907 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="c5d2b054-0ed8-419c-b731-f14d3598372f" Mar 13 14:10:44 crc kubenswrapper[4907]: E0313 14:10:44.898157 4907 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 14:10:44 crc kubenswrapper[4907]: I0313 14:10:44.898156 4907 status_manager.go:851] "Failed to get status for pod" podUID="b66475a4-2e1f-4b18-b565-3cc70bb3720f" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 13 14:10:44 crc kubenswrapper[4907]: I0313 14:10:44.899115 4907 status_manager.go:851] "Failed to get status for pod" podUID="5ef16085-0f4d-435d-a559-20a33c237531" pod="openshift-marketplace/certified-operators-gjdp7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-gjdp7\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 13 14:10:44 crc kubenswrapper[4907]: E0313 14:10:44.942366 4907 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: failed to sync configmap cache: timed out waiting for the condition Mar 13 14:10:44 crc kubenswrapper[4907]: E0313 14:10:44.942418 4907 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: failed to sync configmap cache: timed out waiting for the condition Mar 13 14:10:44 crc kubenswrapper[4907]: E0313 14:10:44.942445 4907 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: failed to sync configmap cache: timed out waiting for the condition Mar 13 14:10:44 crc kubenswrapper[4907]: E0313 14:10:44.942490 4907 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: failed to sync configmap cache: timed out waiting for the condition Mar 13 14:10:44 crc kubenswrapper[4907]: E0313 14:10:44.942529 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-13 14:12:46.942494918 +0000 UTC m=+465.842282637 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : failed to sync configmap cache: timed out waiting for the condition Mar 13 14:10:44 crc kubenswrapper[4907]: E0313 14:10:44.942591 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-13 14:12:46.94255852 +0000 UTC m=+465.842346249 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : failed to sync configmap cache: timed out waiting for the condition Mar 13 14:10:45 crc kubenswrapper[4907]: W0313 14:10:45.060966 4907 reflector.go:561] object-"openshift-network-console"/"networking-console-plugin-cert": failed to list *v1.Secret: Get "https://api-int.crc.testing:6443/api/v1/namespaces/openshift-network-console/secrets?fieldSelector=metadata.name%3Dnetworking-console-plugin-cert&resourceVersion=27253": dial tcp 38.102.83.174:6443: connect: connection refused Mar 13 14:10:45 crc kubenswrapper[4907]: E0313 14:10:45.061140 4907 reflector.go:158] "Unhandled Error" err="object-\"openshift-network-console\"/\"networking-console-plugin-cert\": Failed to watch *v1.Secret: failed to list *v1.Secret: Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-network-console/secrets?fieldSelector=metadata.name%3Dnetworking-console-plugin-cert&resourceVersion=27253\": dial tcp 38.102.83.174:6443: connect: connection refused" logger="UnhandledError" Mar 13 14:10:45 crc kubenswrapper[4907]: I0313 14:10:45.340573 4907 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Readiness probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Mar 13 14:10:45 crc kubenswrapper[4907]: I0313 14:10:45.341110 4907 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Mar 13 14:10:45 crc kubenswrapper[4907]: I0313 14:10:45.911198 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"b74496d1c4984a1e4f4d9b05b79122c2e6c035b15b2f018289763da1eeeeb60c"} Mar 13 14:10:45 crc kubenswrapper[4907]: I0313 14:10:45.911259 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"05a5708731085b175d3a5f8b7c9c7da805199496debf2089469cf2e6f29cd42c"} Mar 13 14:10:45 crc kubenswrapper[4907]: I0313 14:10:45.917146 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/0.log" Mar 13 14:10:45 crc kubenswrapper[4907]: I0313 14:10:45.919074 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Mar 13 14:10:45 crc kubenswrapper[4907]: I0313 14:10:45.919144 4907 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="18eadc065bf4a307b7b443fcfe79b2b719938ee486b53321c0a9f20926a0a887" exitCode=1 Mar 13 14:10:45 crc kubenswrapper[4907]: I0313 14:10:45.919189 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"18eadc065bf4a307b7b443fcfe79b2b719938ee486b53321c0a9f20926a0a887"} Mar 13 14:10:45 crc kubenswrapper[4907]: I0313 14:10:45.919674 4907 scope.go:117] "RemoveContainer" containerID="18eadc065bf4a307b7b443fcfe79b2b719938ee486b53321c0a9f20926a0a887" Mar 13 14:10:46 crc kubenswrapper[4907]: I0313 14:10:46.927560 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/0.log" Mar 13 14:10:46 crc kubenswrapper[4907]: I0313 14:10:46.929516 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Mar 13 14:10:46 crc kubenswrapper[4907]: I0313 14:10:46.929620 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"ce3d04da2d6730474f09c50e099717f844702dbedb6d0bcf6e1beaf09bfaa48e"} Mar 13 14:10:46 crc kubenswrapper[4907]: I0313 14:10:46.932540 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"abae9b949d1bc83d19f3001605ff34418df6c5d957b611db5e92a058405e1b89"} Mar 13 14:10:46 crc kubenswrapper[4907]: I0313 14:10:46.932569 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"065c5409509743840539ff7778a7053b2a294210fadb97a6f1663e20bd78a497"} Mar 13 14:10:46 crc kubenswrapper[4907]: I0313 14:10:46.932581 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"08143c76ef7b653c5bfee9a2778f3a8be16e2709753dcfd74f0f43dc83614561"} Mar 13 14:10:46 crc kubenswrapper[4907]: I0313 14:10:46.932816 4907 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="c5d2b054-0ed8-419c-b731-f14d3598372f" Mar 13 14:10:46 crc kubenswrapper[4907]: I0313 14:10:46.932835 4907 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="c5d2b054-0ed8-419c-b731-f14d3598372f" Mar 13 14:10:46 crc kubenswrapper[4907]: I0313 14:10:46.933045 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 14:10:48 crc kubenswrapper[4907]: I0313 14:10:48.809031 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 14:10:48 crc kubenswrapper[4907]: I0313 14:10:48.809124 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 14:10:48 crc kubenswrapper[4907]: I0313 14:10:48.818814 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 14:10:50 crc kubenswrapper[4907]: I0313 14:10:50.075421 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 13 14:10:50 crc kubenswrapper[4907]: I0313 14:10:50.083485 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 13 14:10:50 crc kubenswrapper[4907]: I0313 14:10:50.961793 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 13 14:10:51 crc kubenswrapper[4907]: I0313 14:10:51.355387 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Mar 13 14:10:51 crc kubenswrapper[4907]: I0313 14:10:51.850594 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Mar 13 14:10:51 crc kubenswrapper[4907]: I0313 14:10:51.851431 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Mar 13 14:10:51 crc kubenswrapper[4907]: I0313 14:10:51.853769 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Mar 13 14:10:51 crc kubenswrapper[4907]: I0313 14:10:51.943187 4907 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 14:10:51 crc kubenswrapper[4907]: I0313 14:10:51.968519 4907 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="c5d2b054-0ed8-419c-b731-f14d3598372f" Mar 13 14:10:51 crc kubenswrapper[4907]: I0313 14:10:51.968569 4907 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="c5d2b054-0ed8-419c-b731-f14d3598372f" Mar 13 14:10:51 crc kubenswrapper[4907]: I0313 14:10:51.973489 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 14:10:52 crc kubenswrapper[4907]: I0313 14:10:52.008128 4907 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="86097500-e6a8-4abd-8b9d-7198222956b2" Mar 13 14:10:52 crc kubenswrapper[4907]: I0313 14:10:52.976441 4907 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="c5d2b054-0ed8-419c-b731-f14d3598372f" Mar 13 14:10:52 crc kubenswrapper[4907]: I0313 14:10:52.977963 4907 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="c5d2b054-0ed8-419c-b731-f14d3598372f" Mar 13 14:10:52 crc kubenswrapper[4907]: I0313 14:10:52.983117 4907 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="86097500-e6a8-4abd-8b9d-7198222956b2" Mar 13 14:10:55 crc kubenswrapper[4907]: E0313 14:10:55.811917 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[kube-api-access-s2dwl], unattached volumes=[], failed to process volumes=[]: context deadline exceeded" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:10:55 crc kubenswrapper[4907]: E0313 14:10:55.832248 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[networking-console-plugin-cert nginx-conf], unattached volumes=[], failed to process volumes=[]: context deadline exceeded" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:10:55 crc kubenswrapper[4907]: E0313 14:10:55.845709 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[kube-api-access-cqllr], unattached volumes=[], failed to process volumes=[]: context deadline exceeded" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:10:57 crc kubenswrapper[4907]: I0313 14:10:57.561272 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-ng4kz" podUID="3661f6b2-1d5e-42f0-947f-a32b66258a1f" containerName="oauth-openshift" containerID="cri-o://3d60ed9e1e0ff620e3a27d58d5fb879793f36dc7e7cd6201ecb90667f5bab7dc" gracePeriod=15 Mar 13 14:10:58 crc kubenswrapper[4907]: I0313 14:10:58.009807 4907 generic.go:334] "Generic (PLEG): container finished" podID="3661f6b2-1d5e-42f0-947f-a32b66258a1f" containerID="3d60ed9e1e0ff620e3a27d58d5fb879793f36dc7e7cd6201ecb90667f5bab7dc" exitCode=0 Mar 13 14:10:58 crc kubenswrapper[4907]: I0313 14:10:58.009907 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-ng4kz" event={"ID":"3661f6b2-1d5e-42f0-947f-a32b66258a1f","Type":"ContainerDied","Data":"3d60ed9e1e0ff620e3a27d58d5fb879793f36dc7e7cd6201ecb90667f5bab7dc"} Mar 13 14:10:58 crc kubenswrapper[4907]: I0313 14:10:58.101194 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-ng4kz" Mar 13 14:10:58 crc kubenswrapper[4907]: I0313 14:10:58.270865 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-user-template-error\") pod \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\" (UID: \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\") " Mar 13 14:10:58 crc kubenswrapper[4907]: I0313 14:10:58.270945 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-system-session\") pod \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\" (UID: \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\") " Mar 13 14:10:58 crc kubenswrapper[4907]: I0313 14:10:58.271182 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-user-template-provider-selection\") pod \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\" (UID: \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\") " Mar 13 14:10:58 crc kubenswrapper[4907]: I0313 14:10:58.271234 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-system-service-ca\") pod \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\" (UID: \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\") " Mar 13 14:10:58 crc kubenswrapper[4907]: I0313 14:10:58.271292 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-system-serving-cert\") pod \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\" (UID: \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\") " Mar 13 14:10:58 crc kubenswrapper[4907]: I0313 14:10:58.271326 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-system-trusted-ca-bundle\") pod \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\" (UID: \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\") " Mar 13 14:10:58 crc kubenswrapper[4907]: I0313 14:10:58.271355 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mtqz9\" (UniqueName: \"kubernetes.io/projected/3661f6b2-1d5e-42f0-947f-a32b66258a1f-kube-api-access-mtqz9\") pod \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\" (UID: \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\") " Mar 13 14:10:58 crc kubenswrapper[4907]: I0313 14:10:58.271388 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-system-router-certs\") pod \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\" (UID: \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\") " Mar 13 14:10:58 crc kubenswrapper[4907]: I0313 14:10:58.271440 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/3661f6b2-1d5e-42f0-947f-a32b66258a1f-audit-policies\") pod \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\" (UID: \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\") " Mar 13 14:10:58 crc kubenswrapper[4907]: I0313 14:10:58.271463 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-system-cliconfig\") pod \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\" (UID: \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\") " Mar 13 14:10:58 crc kubenswrapper[4907]: I0313 14:10:58.271481 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-system-ocp-branding-template\") pod \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\" (UID: \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\") " Mar 13 14:10:58 crc kubenswrapper[4907]: I0313 14:10:58.271508 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3661f6b2-1d5e-42f0-947f-a32b66258a1f-audit-dir\") pod \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\" (UID: \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\") " Mar 13 14:10:58 crc kubenswrapper[4907]: I0313 14:10:58.271541 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-user-template-login\") pod \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\" (UID: \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\") " Mar 13 14:10:58 crc kubenswrapper[4907]: I0313 14:10:58.271578 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-user-idp-0-file-data\") pod \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\" (UID: \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\") " Mar 13 14:10:58 crc kubenswrapper[4907]: I0313 14:10:58.273619 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "3661f6b2-1d5e-42f0-947f-a32b66258a1f" (UID: "3661f6b2-1d5e-42f0-947f-a32b66258a1f"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:10:58 crc kubenswrapper[4907]: I0313 14:10:58.274193 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "3661f6b2-1d5e-42f0-947f-a32b66258a1f" (UID: "3661f6b2-1d5e-42f0-947f-a32b66258a1f"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:10:58 crc kubenswrapper[4907]: I0313 14:10:58.275965 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3661f6b2-1d5e-42f0-947f-a32b66258a1f-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "3661f6b2-1d5e-42f0-947f-a32b66258a1f" (UID: "3661f6b2-1d5e-42f0-947f-a32b66258a1f"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 14:10:58 crc kubenswrapper[4907]: I0313 14:10:58.277452 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3661f6b2-1d5e-42f0-947f-a32b66258a1f-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "3661f6b2-1d5e-42f0-947f-a32b66258a1f" (UID: "3661f6b2-1d5e-42f0-947f-a32b66258a1f"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:10:58 crc kubenswrapper[4907]: I0313 14:10:58.279292 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "3661f6b2-1d5e-42f0-947f-a32b66258a1f" (UID: "3661f6b2-1d5e-42f0-947f-a32b66258a1f"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:10:58 crc kubenswrapper[4907]: I0313 14:10:58.290020 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "3661f6b2-1d5e-42f0-947f-a32b66258a1f" (UID: "3661f6b2-1d5e-42f0-947f-a32b66258a1f"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:10:58 crc kubenswrapper[4907]: I0313 14:10:58.291990 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3661f6b2-1d5e-42f0-947f-a32b66258a1f-kube-api-access-mtqz9" (OuterVolumeSpecName: "kube-api-access-mtqz9") pod "3661f6b2-1d5e-42f0-947f-a32b66258a1f" (UID: "3661f6b2-1d5e-42f0-947f-a32b66258a1f"). InnerVolumeSpecName "kube-api-access-mtqz9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:10:58 crc kubenswrapper[4907]: I0313 14:10:58.292086 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "3661f6b2-1d5e-42f0-947f-a32b66258a1f" (UID: "3661f6b2-1d5e-42f0-947f-a32b66258a1f"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:10:58 crc kubenswrapper[4907]: I0313 14:10:58.292384 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "3661f6b2-1d5e-42f0-947f-a32b66258a1f" (UID: "3661f6b2-1d5e-42f0-947f-a32b66258a1f"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:10:58 crc kubenswrapper[4907]: I0313 14:10:58.292951 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "3661f6b2-1d5e-42f0-947f-a32b66258a1f" (UID: "3661f6b2-1d5e-42f0-947f-a32b66258a1f"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:10:58 crc kubenswrapper[4907]: I0313 14:10:58.293568 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "3661f6b2-1d5e-42f0-947f-a32b66258a1f" (UID: "3661f6b2-1d5e-42f0-947f-a32b66258a1f"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:10:58 crc kubenswrapper[4907]: I0313 14:10:58.294397 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "3661f6b2-1d5e-42f0-947f-a32b66258a1f" (UID: "3661f6b2-1d5e-42f0-947f-a32b66258a1f"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:10:58 crc kubenswrapper[4907]: I0313 14:10:58.298607 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "3661f6b2-1d5e-42f0-947f-a32b66258a1f" (UID: "3661f6b2-1d5e-42f0-947f-a32b66258a1f"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:10:58 crc kubenswrapper[4907]: I0313 14:10:58.299037 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "3661f6b2-1d5e-42f0-947f-a32b66258a1f" (UID: "3661f6b2-1d5e-42f0-947f-a32b66258a1f"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:10:58 crc kubenswrapper[4907]: I0313 14:10:58.379386 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mtqz9\" (UniqueName: \"kubernetes.io/projected/3661f6b2-1d5e-42f0-947f-a32b66258a1f-kube-api-access-mtqz9\") on node \"crc\" DevicePath \"\"" Mar 13 14:10:58 crc kubenswrapper[4907]: I0313 14:10:58.379445 4907 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Mar 13 14:10:58 crc kubenswrapper[4907]: I0313 14:10:58.379473 4907 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/3661f6b2-1d5e-42f0-947f-a32b66258a1f-audit-policies\") on node \"crc\" DevicePath \"\"" Mar 13 14:10:58 crc kubenswrapper[4907]: I0313 14:10:58.379498 4907 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Mar 13 14:10:58 crc kubenswrapper[4907]: I0313 14:10:58.379520 4907 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3661f6b2-1d5e-42f0-947f-a32b66258a1f-audit-dir\") on node \"crc\" DevicePath \"\"" Mar 13 14:10:58 crc kubenswrapper[4907]: I0313 14:10:58.379540 4907 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Mar 13 14:10:58 crc kubenswrapper[4907]: I0313 14:10:58.379561 4907 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Mar 13 14:10:58 crc kubenswrapper[4907]: I0313 14:10:58.379583 4907 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Mar 13 14:10:58 crc kubenswrapper[4907]: I0313 14:10:58.379604 4907 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Mar 13 14:10:58 crc kubenswrapper[4907]: I0313 14:10:58.379626 4907 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Mar 13 14:10:58 crc kubenswrapper[4907]: I0313 14:10:58.379647 4907 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Mar 13 14:10:58 crc kubenswrapper[4907]: I0313 14:10:58.379669 4907 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Mar 13 14:10:58 crc kubenswrapper[4907]: I0313 14:10:58.379688 4907 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 14:10:58 crc kubenswrapper[4907]: I0313 14:10:58.379709 4907 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:10:59 crc kubenswrapper[4907]: I0313 14:10:59.020351 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-ng4kz" event={"ID":"3661f6b2-1d5e-42f0-947f-a32b66258a1f","Type":"ContainerDied","Data":"9696c0da9ec098e8a51fc14c33a99177653e98a7c7e1cea2ab3b0156346c1506"} Mar 13 14:10:59 crc kubenswrapper[4907]: I0313 14:10:59.020467 4907 scope.go:117] "RemoveContainer" containerID="3d60ed9e1e0ff620e3a27d58d5fb879793f36dc7e7cd6201ecb90667f5bab7dc" Mar 13 14:10:59 crc kubenswrapper[4907]: I0313 14:10:59.020474 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-ng4kz" Mar 13 14:11:01 crc kubenswrapper[4907]: I0313 14:11:01.614299 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Mar 13 14:11:02 crc kubenswrapper[4907]: I0313 14:11:02.092526 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Mar 13 14:11:02 crc kubenswrapper[4907]: I0313 14:11:02.590477 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Mar 13 14:11:02 crc kubenswrapper[4907]: I0313 14:11:02.642844 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Mar 13 14:11:02 crc kubenswrapper[4907]: I0313 14:11:02.884269 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Mar 13 14:11:03 crc kubenswrapper[4907]: I0313 14:11:03.257446 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Mar 13 14:11:03 crc kubenswrapper[4907]: I0313 14:11:03.351096 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Mar 13 14:11:03 crc kubenswrapper[4907]: I0313 14:11:03.373265 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Mar 13 14:11:03 crc kubenswrapper[4907]: I0313 14:11:03.549420 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Mar 13 14:11:03 crc kubenswrapper[4907]: I0313 14:11:03.742632 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Mar 13 14:11:04 crc kubenswrapper[4907]: I0313 14:11:04.687076 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Mar 13 14:11:04 crc kubenswrapper[4907]: I0313 14:11:04.940681 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Mar 13 14:11:05 crc kubenswrapper[4907]: I0313 14:11:05.288662 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Mar 13 14:11:05 crc kubenswrapper[4907]: I0313 14:11:05.343630 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 13 14:11:05 crc kubenswrapper[4907]: I0313 14:11:05.395263 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Mar 13 14:11:05 crc kubenswrapper[4907]: I0313 14:11:05.729534 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Mar 13 14:11:05 crc kubenswrapper[4907]: I0313 14:11:05.732996 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Mar 13 14:11:05 crc kubenswrapper[4907]: I0313 14:11:05.893811 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Mar 13 14:11:05 crc kubenswrapper[4907]: I0313 14:11:05.957420 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Mar 13 14:11:06 crc kubenswrapper[4907]: I0313 14:11:06.253602 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Mar 13 14:11:06 crc kubenswrapper[4907]: I0313 14:11:06.258147 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Mar 13 14:11:06 crc kubenswrapper[4907]: I0313 14:11:06.303586 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Mar 13 14:11:06 crc kubenswrapper[4907]: I0313 14:11:06.394396 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Mar 13 14:11:06 crc kubenswrapper[4907]: I0313 14:11:06.535331 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Mar 13 14:11:06 crc kubenswrapper[4907]: I0313 14:11:06.636113 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Mar 13 14:11:06 crc kubenswrapper[4907]: I0313 14:11:06.713698 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Mar 13 14:11:06 crc kubenswrapper[4907]: I0313 14:11:06.747463 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Mar 13 14:11:06 crc kubenswrapper[4907]: I0313 14:11:06.860765 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Mar 13 14:11:07 crc kubenswrapper[4907]: I0313 14:11:07.138016 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Mar 13 14:11:07 crc kubenswrapper[4907]: I0313 14:11:07.157136 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Mar 13 14:11:07 crc kubenswrapper[4907]: I0313 14:11:07.161128 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Mar 13 14:11:07 crc kubenswrapper[4907]: I0313 14:11:07.445471 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Mar 13 14:11:07 crc kubenswrapper[4907]: I0313 14:11:07.521372 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Mar 13 14:11:07 crc kubenswrapper[4907]: I0313 14:11:07.528676 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Mar 13 14:11:07 crc kubenswrapper[4907]: I0313 14:11:07.537550 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Mar 13 14:11:07 crc kubenswrapper[4907]: I0313 14:11:07.594021 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Mar 13 14:11:07 crc kubenswrapper[4907]: I0313 14:11:07.640966 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Mar 13 14:11:07 crc kubenswrapper[4907]: I0313 14:11:07.676989 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Mar 13 14:11:07 crc kubenswrapper[4907]: I0313 14:11:07.806512 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Mar 13 14:11:07 crc kubenswrapper[4907]: I0313 14:11:07.910567 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Mar 13 14:11:08 crc kubenswrapper[4907]: I0313 14:11:08.076807 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Mar 13 14:11:08 crc kubenswrapper[4907]: I0313 14:11:08.137640 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Mar 13 14:11:08 crc kubenswrapper[4907]: I0313 14:11:08.245335 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Mar 13 14:11:08 crc kubenswrapper[4907]: I0313 14:11:08.254180 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Mar 13 14:11:08 crc kubenswrapper[4907]: I0313 14:11:08.259957 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Mar 13 14:11:08 crc kubenswrapper[4907]: I0313 14:11:08.303719 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Mar 13 14:11:08 crc kubenswrapper[4907]: I0313 14:11:08.521811 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Mar 13 14:11:08 crc kubenswrapper[4907]: I0313 14:11:08.700092 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Mar 13 14:11:08 crc kubenswrapper[4907]: I0313 14:11:08.781342 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:11:08 crc kubenswrapper[4907]: I0313 14:11:08.855603 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.015147 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.039513 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.051768 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.067228 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.071065 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.080125 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.098358 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.098517 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.118130 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.217564 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.315049 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.338061 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.349179 4907 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.351720 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.354406 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-gjdp7","openshift-authentication/oauth-openshift-558db77b4-ng4kz","openshift-kube-apiserver/kube-apiserver-crc"] Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.354467 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5c879c7f8d-2mxxc","openshift-kube-apiserver/kube-apiserver-crc","openshift-controller-manager/controller-manager-5f4c4b8dd4-9sl4b","openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp"] Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.355003 4907 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="c5d2b054-0ed8-419c-b731-f14d3598372f" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.355221 4907 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="c5d2b054-0ed8-419c-b731-f14d3598372f" Mar 13 14:11:09 crc kubenswrapper[4907]: E0313 14:11:09.355447 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b66475a4-2e1f-4b18-b565-3cc70bb3720f" containerName="installer" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.355476 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="b66475a4-2e1f-4b18-b565-3cc70bb3720f" containerName="installer" Mar 13 14:11:09 crc kubenswrapper[4907]: E0313 14:11:09.355495 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3661f6b2-1d5e-42f0-947f-a32b66258a1f" containerName="oauth-openshift" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.355502 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="3661f6b2-1d5e-42f0-947f-a32b66258a1f" containerName="oauth-openshift" Mar 13 14:11:09 crc kubenswrapper[4907]: E0313 14:11:09.355521 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ef16085-0f4d-435d-a559-20a33c237531" containerName="extract-content" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.355528 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ef16085-0f4d-435d-a559-20a33c237531" containerName="extract-content" Mar 13 14:11:09 crc kubenswrapper[4907]: E0313 14:11:09.355541 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ef16085-0f4d-435d-a559-20a33c237531" containerName="registry-server" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.355548 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ef16085-0f4d-435d-a559-20a33c237531" containerName="registry-server" Mar 13 14:11:09 crc kubenswrapper[4907]: E0313 14:11:09.355561 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ef16085-0f4d-435d-a559-20a33c237531" containerName="extract-utilities" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.355569 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ef16085-0f4d-435d-a559-20a33c237531" containerName="extract-utilities" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.355738 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="b66475a4-2e1f-4b18-b565-3cc70bb3720f" containerName="installer" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.355749 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="3661f6b2-1d5e-42f0-947f-a32b66258a1f" containerName="oauth-openshift" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.355761 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ef16085-0f4d-435d-a559-20a33c237531" containerName="registry-server" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.357900 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5f4c4b8dd4-9sl4b" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.358617 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5c879c7f8d-2mxxc" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.360092 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.360974 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.361236 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.361684 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.365122 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.365302 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.365378 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.365408 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.365584 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.365742 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.365873 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.365993 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.366148 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.366187 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.366277 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.366320 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.366370 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.366296 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.366397 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.366434 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.366581 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.366684 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.366709 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.366745 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.366963 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.367231 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.367427 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.381271 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.381327 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.383378 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.389414 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.391129 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.443101 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=18.443083684 podStartE2EDuration="18.443083684s" podCreationTimestamp="2026-03-13 14:10:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:11:09.440236688 +0000 UTC m=+368.340024377" watchObservedRunningTime="2026-03-13 14:11:09.443083684 +0000 UTC m=+368.342871373" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.445419 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.528253 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.538352 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.540830 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/03ca839d-b5ba-4b55-87b3-48abca197163-config\") pod \"controller-manager-5f4c4b8dd4-9sl4b\" (UID: \"03ca839d-b5ba-4b55-87b3-48abca197163\") " pod="openshift-controller-manager/controller-manager-5f4c4b8dd4-9sl4b" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.541055 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b98r7\" (UniqueName: \"kubernetes.io/projected/a704bdfc-46ff-4a8a-8d31-21a140b49c99-kube-api-access-b98r7\") pod \"route-controller-manager-5c879c7f8d-2mxxc\" (UID: \"a704bdfc-46ff-4a8a-8d31-21a140b49c99\") " pod="openshift-route-controller-manager/route-controller-manager-5c879c7f8d-2mxxc" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.541257 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/8746df2e-524e-42a7-9ca5-c5300c0a1a09-v4-0-config-system-service-ca\") pod \"oauth-openshift-7c65b78cb7-ld7fp\" (UID: \"8746df2e-524e-42a7-9ca5-c5300c0a1a09\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.541378 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/8746df2e-524e-42a7-9ca5-c5300c0a1a09-v4-0-config-user-template-error\") pod \"oauth-openshift-7c65b78cb7-ld7fp\" (UID: \"8746df2e-524e-42a7-9ca5-c5300c0a1a09\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.541481 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/8746df2e-524e-42a7-9ca5-c5300c0a1a09-v4-0-config-system-router-certs\") pod \"oauth-openshift-7c65b78cb7-ld7fp\" (UID: \"8746df2e-524e-42a7-9ca5-c5300c0a1a09\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.541633 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/8746df2e-524e-42a7-9ca5-c5300c0a1a09-v4-0-config-user-template-login\") pod \"oauth-openshift-7c65b78cb7-ld7fp\" (UID: \"8746df2e-524e-42a7-9ca5-c5300c0a1a09\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.541741 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lwxs7\" (UniqueName: \"kubernetes.io/projected/03ca839d-b5ba-4b55-87b3-48abca197163-kube-api-access-lwxs7\") pod \"controller-manager-5f4c4b8dd4-9sl4b\" (UID: \"03ca839d-b5ba-4b55-87b3-48abca197163\") " pod="openshift-controller-manager/controller-manager-5f4c4b8dd4-9sl4b" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.541851 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/8746df2e-524e-42a7-9ca5-c5300c0a1a09-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-7c65b78cb7-ld7fp\" (UID: \"8746df2e-524e-42a7-9ca5-c5300c0a1a09\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.541988 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/03ca839d-b5ba-4b55-87b3-48abca197163-serving-cert\") pod \"controller-manager-5f4c4b8dd4-9sl4b\" (UID: \"03ca839d-b5ba-4b55-87b3-48abca197163\") " pod="openshift-controller-manager/controller-manager-5f4c4b8dd4-9sl4b" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.542109 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/8746df2e-524e-42a7-9ca5-c5300c0a1a09-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-7c65b78cb7-ld7fp\" (UID: \"8746df2e-524e-42a7-9ca5-c5300c0a1a09\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.542215 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/03ca839d-b5ba-4b55-87b3-48abca197163-proxy-ca-bundles\") pod \"controller-manager-5f4c4b8dd4-9sl4b\" (UID: \"03ca839d-b5ba-4b55-87b3-48abca197163\") " pod="openshift-controller-manager/controller-manager-5f4c4b8dd4-9sl4b" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.542320 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/8746df2e-524e-42a7-9ca5-c5300c0a1a09-v4-0-config-system-session\") pod \"oauth-openshift-7c65b78cb7-ld7fp\" (UID: \"8746df2e-524e-42a7-9ca5-c5300c0a1a09\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.542418 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8746df2e-524e-42a7-9ca5-c5300c0a1a09-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-7c65b78cb7-ld7fp\" (UID: \"8746df2e-524e-42a7-9ca5-c5300c0a1a09\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.542525 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/8746df2e-524e-42a7-9ca5-c5300c0a1a09-v4-0-config-system-cliconfig\") pod \"oauth-openshift-7c65b78cb7-ld7fp\" (UID: \"8746df2e-524e-42a7-9ca5-c5300c0a1a09\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.542625 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a704bdfc-46ff-4a8a-8d31-21a140b49c99-config\") pod \"route-controller-manager-5c879c7f8d-2mxxc\" (UID: \"a704bdfc-46ff-4a8a-8d31-21a140b49c99\") " pod="openshift-route-controller-manager/route-controller-manager-5c879c7f8d-2mxxc" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.542726 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/03ca839d-b5ba-4b55-87b3-48abca197163-client-ca\") pod \"controller-manager-5f4c4b8dd4-9sl4b\" (UID: \"03ca839d-b5ba-4b55-87b3-48abca197163\") " pod="openshift-controller-manager/controller-manager-5f4c4b8dd4-9sl4b" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.542841 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a704bdfc-46ff-4a8a-8d31-21a140b49c99-serving-cert\") pod \"route-controller-manager-5c879c7f8d-2mxxc\" (UID: \"a704bdfc-46ff-4a8a-8d31-21a140b49c99\") " pod="openshift-route-controller-manager/route-controller-manager-5c879c7f8d-2mxxc" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.542960 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/8746df2e-524e-42a7-9ca5-c5300c0a1a09-audit-policies\") pod \"oauth-openshift-7c65b78cb7-ld7fp\" (UID: \"8746df2e-524e-42a7-9ca5-c5300c0a1a09\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.543107 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8746df2e-524e-42a7-9ca5-c5300c0a1a09-audit-dir\") pod \"oauth-openshift-7c65b78cb7-ld7fp\" (UID: \"8746df2e-524e-42a7-9ca5-c5300c0a1a09\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.543202 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/8746df2e-524e-42a7-9ca5-c5300c0a1a09-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-7c65b78cb7-ld7fp\" (UID: \"8746df2e-524e-42a7-9ca5-c5300c0a1a09\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.543240 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/8746df2e-524e-42a7-9ca5-c5300c0a1a09-v4-0-config-system-serving-cert\") pod \"oauth-openshift-7c65b78cb7-ld7fp\" (UID: \"8746df2e-524e-42a7-9ca5-c5300c0a1a09\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.543265 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wml9p\" (UniqueName: \"kubernetes.io/projected/8746df2e-524e-42a7-9ca5-c5300c0a1a09-kube-api-access-wml9p\") pod \"oauth-openshift-7c65b78cb7-ld7fp\" (UID: \"8746df2e-524e-42a7-9ca5-c5300c0a1a09\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.543291 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a704bdfc-46ff-4a8a-8d31-21a140b49c99-client-ca\") pod \"route-controller-manager-5c879c7f8d-2mxxc\" (UID: \"a704bdfc-46ff-4a8a-8d31-21a140b49c99\") " pod="openshift-route-controller-manager/route-controller-manager-5c879c7f8d-2mxxc" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.644077 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/8746df2e-524e-42a7-9ca5-c5300c0a1a09-v4-0-config-user-template-login\") pod \"oauth-openshift-7c65b78cb7-ld7fp\" (UID: \"8746df2e-524e-42a7-9ca5-c5300c0a1a09\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.644116 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lwxs7\" (UniqueName: \"kubernetes.io/projected/03ca839d-b5ba-4b55-87b3-48abca197163-kube-api-access-lwxs7\") pod \"controller-manager-5f4c4b8dd4-9sl4b\" (UID: \"03ca839d-b5ba-4b55-87b3-48abca197163\") " pod="openshift-controller-manager/controller-manager-5f4c4b8dd4-9sl4b" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.644139 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/8746df2e-524e-42a7-9ca5-c5300c0a1a09-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-7c65b78cb7-ld7fp\" (UID: \"8746df2e-524e-42a7-9ca5-c5300c0a1a09\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.644172 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/03ca839d-b5ba-4b55-87b3-48abca197163-serving-cert\") pod \"controller-manager-5f4c4b8dd4-9sl4b\" (UID: \"03ca839d-b5ba-4b55-87b3-48abca197163\") " pod="openshift-controller-manager/controller-manager-5f4c4b8dd4-9sl4b" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.644191 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/8746df2e-524e-42a7-9ca5-c5300c0a1a09-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-7c65b78cb7-ld7fp\" (UID: \"8746df2e-524e-42a7-9ca5-c5300c0a1a09\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.644209 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/03ca839d-b5ba-4b55-87b3-48abca197163-proxy-ca-bundles\") pod \"controller-manager-5f4c4b8dd4-9sl4b\" (UID: \"03ca839d-b5ba-4b55-87b3-48abca197163\") " pod="openshift-controller-manager/controller-manager-5f4c4b8dd4-9sl4b" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.644229 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/8746df2e-524e-42a7-9ca5-c5300c0a1a09-v4-0-config-system-session\") pod \"oauth-openshift-7c65b78cb7-ld7fp\" (UID: \"8746df2e-524e-42a7-9ca5-c5300c0a1a09\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.644248 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8746df2e-524e-42a7-9ca5-c5300c0a1a09-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-7c65b78cb7-ld7fp\" (UID: \"8746df2e-524e-42a7-9ca5-c5300c0a1a09\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.644266 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/8746df2e-524e-42a7-9ca5-c5300c0a1a09-v4-0-config-system-cliconfig\") pod \"oauth-openshift-7c65b78cb7-ld7fp\" (UID: \"8746df2e-524e-42a7-9ca5-c5300c0a1a09\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.644287 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a704bdfc-46ff-4a8a-8d31-21a140b49c99-config\") pod \"route-controller-manager-5c879c7f8d-2mxxc\" (UID: \"a704bdfc-46ff-4a8a-8d31-21a140b49c99\") " pod="openshift-route-controller-manager/route-controller-manager-5c879c7f8d-2mxxc" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.644302 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/03ca839d-b5ba-4b55-87b3-48abca197163-client-ca\") pod \"controller-manager-5f4c4b8dd4-9sl4b\" (UID: \"03ca839d-b5ba-4b55-87b3-48abca197163\") " pod="openshift-controller-manager/controller-manager-5f4c4b8dd4-9sl4b" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.644331 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a704bdfc-46ff-4a8a-8d31-21a140b49c99-serving-cert\") pod \"route-controller-manager-5c879c7f8d-2mxxc\" (UID: \"a704bdfc-46ff-4a8a-8d31-21a140b49c99\") " pod="openshift-route-controller-manager/route-controller-manager-5c879c7f8d-2mxxc" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.644359 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/8746df2e-524e-42a7-9ca5-c5300c0a1a09-audit-policies\") pod \"oauth-openshift-7c65b78cb7-ld7fp\" (UID: \"8746df2e-524e-42a7-9ca5-c5300c0a1a09\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.644376 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8746df2e-524e-42a7-9ca5-c5300c0a1a09-audit-dir\") pod \"oauth-openshift-7c65b78cb7-ld7fp\" (UID: \"8746df2e-524e-42a7-9ca5-c5300c0a1a09\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.644394 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/8746df2e-524e-42a7-9ca5-c5300c0a1a09-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-7c65b78cb7-ld7fp\" (UID: \"8746df2e-524e-42a7-9ca5-c5300c0a1a09\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.644411 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/8746df2e-524e-42a7-9ca5-c5300c0a1a09-v4-0-config-system-serving-cert\") pod \"oauth-openshift-7c65b78cb7-ld7fp\" (UID: \"8746df2e-524e-42a7-9ca5-c5300c0a1a09\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.644427 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wml9p\" (UniqueName: \"kubernetes.io/projected/8746df2e-524e-42a7-9ca5-c5300c0a1a09-kube-api-access-wml9p\") pod \"oauth-openshift-7c65b78cb7-ld7fp\" (UID: \"8746df2e-524e-42a7-9ca5-c5300c0a1a09\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.644445 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a704bdfc-46ff-4a8a-8d31-21a140b49c99-client-ca\") pod \"route-controller-manager-5c879c7f8d-2mxxc\" (UID: \"a704bdfc-46ff-4a8a-8d31-21a140b49c99\") " pod="openshift-route-controller-manager/route-controller-manager-5c879c7f8d-2mxxc" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.644476 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/03ca839d-b5ba-4b55-87b3-48abca197163-config\") pod \"controller-manager-5f4c4b8dd4-9sl4b\" (UID: \"03ca839d-b5ba-4b55-87b3-48abca197163\") " pod="openshift-controller-manager/controller-manager-5f4c4b8dd4-9sl4b" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.644493 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b98r7\" (UniqueName: \"kubernetes.io/projected/a704bdfc-46ff-4a8a-8d31-21a140b49c99-kube-api-access-b98r7\") pod \"route-controller-manager-5c879c7f8d-2mxxc\" (UID: \"a704bdfc-46ff-4a8a-8d31-21a140b49c99\") " pod="openshift-route-controller-manager/route-controller-manager-5c879c7f8d-2mxxc" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.644510 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/8746df2e-524e-42a7-9ca5-c5300c0a1a09-v4-0-config-system-service-ca\") pod \"oauth-openshift-7c65b78cb7-ld7fp\" (UID: \"8746df2e-524e-42a7-9ca5-c5300c0a1a09\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.644529 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/8746df2e-524e-42a7-9ca5-c5300c0a1a09-v4-0-config-user-template-error\") pod \"oauth-openshift-7c65b78cb7-ld7fp\" (UID: \"8746df2e-524e-42a7-9ca5-c5300c0a1a09\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.644550 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/8746df2e-524e-42a7-9ca5-c5300c0a1a09-v4-0-config-system-router-certs\") pod \"oauth-openshift-7c65b78cb7-ld7fp\" (UID: \"8746df2e-524e-42a7-9ca5-c5300c0a1a09\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.645374 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/8746df2e-524e-42a7-9ca5-c5300c0a1a09-audit-policies\") pod \"oauth-openshift-7c65b78cb7-ld7fp\" (UID: \"8746df2e-524e-42a7-9ca5-c5300c0a1a09\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.645522 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/03ca839d-b5ba-4b55-87b3-48abca197163-client-ca\") pod \"controller-manager-5f4c4b8dd4-9sl4b\" (UID: \"03ca839d-b5ba-4b55-87b3-48abca197163\") " pod="openshift-controller-manager/controller-manager-5f4c4b8dd4-9sl4b" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.645837 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a704bdfc-46ff-4a8a-8d31-21a140b49c99-config\") pod \"route-controller-manager-5c879c7f8d-2mxxc\" (UID: \"a704bdfc-46ff-4a8a-8d31-21a140b49c99\") " pod="openshift-route-controller-manager/route-controller-manager-5c879c7f8d-2mxxc" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.646480 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8746df2e-524e-42a7-9ca5-c5300c0a1a09-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-7c65b78cb7-ld7fp\" (UID: \"8746df2e-524e-42a7-9ca5-c5300c0a1a09\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.646609 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/8746df2e-524e-42a7-9ca5-c5300c0a1a09-v4-0-config-system-service-ca\") pod \"oauth-openshift-7c65b78cb7-ld7fp\" (UID: \"8746df2e-524e-42a7-9ca5-c5300c0a1a09\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.647425 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a704bdfc-46ff-4a8a-8d31-21a140b49c99-client-ca\") pod \"route-controller-manager-5c879c7f8d-2mxxc\" (UID: \"a704bdfc-46ff-4a8a-8d31-21a140b49c99\") " pod="openshift-route-controller-manager/route-controller-manager-5c879c7f8d-2mxxc" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.647830 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8746df2e-524e-42a7-9ca5-c5300c0a1a09-audit-dir\") pod \"oauth-openshift-7c65b78cb7-ld7fp\" (UID: \"8746df2e-524e-42a7-9ca5-c5300c0a1a09\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.647996 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/8746df2e-524e-42a7-9ca5-c5300c0a1a09-v4-0-config-system-cliconfig\") pod \"oauth-openshift-7c65b78cb7-ld7fp\" (UID: \"8746df2e-524e-42a7-9ca5-c5300c0a1a09\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.650026 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/03ca839d-b5ba-4b55-87b3-48abca197163-config\") pod \"controller-manager-5f4c4b8dd4-9sl4b\" (UID: \"03ca839d-b5ba-4b55-87b3-48abca197163\") " pod="openshift-controller-manager/controller-manager-5f4c4b8dd4-9sl4b" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.651208 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/03ca839d-b5ba-4b55-87b3-48abca197163-proxy-ca-bundles\") pod \"controller-manager-5f4c4b8dd4-9sl4b\" (UID: \"03ca839d-b5ba-4b55-87b3-48abca197163\") " pod="openshift-controller-manager/controller-manager-5f4c4b8dd4-9sl4b" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.651649 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/03ca839d-b5ba-4b55-87b3-48abca197163-serving-cert\") pod \"controller-manager-5f4c4b8dd4-9sl4b\" (UID: \"03ca839d-b5ba-4b55-87b3-48abca197163\") " pod="openshift-controller-manager/controller-manager-5f4c4b8dd4-9sl4b" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.652986 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/8746df2e-524e-42a7-9ca5-c5300c0a1a09-v4-0-config-user-template-error\") pod \"oauth-openshift-7c65b78cb7-ld7fp\" (UID: \"8746df2e-524e-42a7-9ca5-c5300c0a1a09\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.652991 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/8746df2e-524e-42a7-9ca5-c5300c0a1a09-v4-0-config-system-serving-cert\") pod \"oauth-openshift-7c65b78cb7-ld7fp\" (UID: \"8746df2e-524e-42a7-9ca5-c5300c0a1a09\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.654530 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.654704 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/8746df2e-524e-42a7-9ca5-c5300c0a1a09-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-7c65b78cb7-ld7fp\" (UID: \"8746df2e-524e-42a7-9ca5-c5300c0a1a09\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.656660 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/8746df2e-524e-42a7-9ca5-c5300c0a1a09-v4-0-config-user-template-login\") pod \"oauth-openshift-7c65b78cb7-ld7fp\" (UID: \"8746df2e-524e-42a7-9ca5-c5300c0a1a09\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.658032 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/8746df2e-524e-42a7-9ca5-c5300c0a1a09-v4-0-config-system-router-certs\") pod \"oauth-openshift-7c65b78cb7-ld7fp\" (UID: \"8746df2e-524e-42a7-9ca5-c5300c0a1a09\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.659652 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/8746df2e-524e-42a7-9ca5-c5300c0a1a09-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-7c65b78cb7-ld7fp\" (UID: \"8746df2e-524e-42a7-9ca5-c5300c0a1a09\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.662113 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/8746df2e-524e-42a7-9ca5-c5300c0a1a09-v4-0-config-system-session\") pod \"oauth-openshift-7c65b78cb7-ld7fp\" (UID: \"8746df2e-524e-42a7-9ca5-c5300c0a1a09\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.662355 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.663423 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a704bdfc-46ff-4a8a-8d31-21a140b49c99-serving-cert\") pod \"route-controller-manager-5c879c7f8d-2mxxc\" (UID: \"a704bdfc-46ff-4a8a-8d31-21a140b49c99\") " pod="openshift-route-controller-manager/route-controller-manager-5c879c7f8d-2mxxc" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.664280 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lwxs7\" (UniqueName: \"kubernetes.io/projected/03ca839d-b5ba-4b55-87b3-48abca197163-kube-api-access-lwxs7\") pod \"controller-manager-5f4c4b8dd4-9sl4b\" (UID: \"03ca839d-b5ba-4b55-87b3-48abca197163\") " pod="openshift-controller-manager/controller-manager-5f4c4b8dd4-9sl4b" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.667997 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wml9p\" (UniqueName: \"kubernetes.io/projected/8746df2e-524e-42a7-9ca5-c5300c0a1a09-kube-api-access-wml9p\") pod \"oauth-openshift-7c65b78cb7-ld7fp\" (UID: \"8746df2e-524e-42a7-9ca5-c5300c0a1a09\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.668086 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b98r7\" (UniqueName: \"kubernetes.io/projected/a704bdfc-46ff-4a8a-8d31-21a140b49c99-kube-api-access-b98r7\") pod \"route-controller-manager-5c879c7f8d-2mxxc\" (UID: \"a704bdfc-46ff-4a8a-8d31-21a140b49c99\") " pod="openshift-route-controller-manager/route-controller-manager-5c879c7f8d-2mxxc" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.668465 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/8746df2e-524e-42a7-9ca5-c5300c0a1a09-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-7c65b78cb7-ld7fp\" (UID: \"8746df2e-524e-42a7-9ca5-c5300c0a1a09\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.686129 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5f4c4b8dd4-9sl4b" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.692535 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5c879c7f8d-2mxxc" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.703238 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.708490 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.748907 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.782247 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.790793 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3661f6b2-1d5e-42f0-947f-a32b66258a1f" path="/var/lib/kubelet/pods/3661f6b2-1d5e-42f0-947f-a32b66258a1f/volumes" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.791634 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5ef16085-0f4d-435d-a559-20a33c237531" path="/var/lib/kubelet/pods/5ef16085-0f4d-435d-a559-20a33c237531/volumes" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.831220 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.867158 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.884037 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.931865 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Mar 13 14:11:10 crc kubenswrapper[4907]: I0313 14:11:10.034709 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Mar 13 14:11:10 crc kubenswrapper[4907]: I0313 14:11:10.045492 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Mar 13 14:11:10 crc kubenswrapper[4907]: I0313 14:11:10.119710 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Mar 13 14:11:10 crc kubenswrapper[4907]: I0313 14:11:10.216638 4907 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Mar 13 14:11:10 crc kubenswrapper[4907]: I0313 14:11:10.247723 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Mar 13 14:11:10 crc kubenswrapper[4907]: I0313 14:11:10.252099 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Mar 13 14:11:10 crc kubenswrapper[4907]: I0313 14:11:10.254390 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Mar 13 14:11:10 crc kubenswrapper[4907]: I0313 14:11:10.363764 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Mar 13 14:11:10 crc kubenswrapper[4907]: I0313 14:11:10.381855 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Mar 13 14:11:10 crc kubenswrapper[4907]: I0313 14:11:10.407209 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Mar 13 14:11:10 crc kubenswrapper[4907]: I0313 14:11:10.468650 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Mar 13 14:11:10 crc kubenswrapper[4907]: I0313 14:11:10.475305 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Mar 13 14:11:10 crc kubenswrapper[4907]: I0313 14:11:10.570621 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Mar 13 14:11:10 crc kubenswrapper[4907]: I0313 14:11:10.672586 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Mar 13 14:11:10 crc kubenswrapper[4907]: I0313 14:11:10.676418 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Mar 13 14:11:10 crc kubenswrapper[4907]: I0313 14:11:10.740603 4907 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Mar 13 14:11:10 crc kubenswrapper[4907]: I0313 14:11:10.781788 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:11:10 crc kubenswrapper[4907]: I0313 14:11:10.948685 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Mar 13 14:11:11 crc kubenswrapper[4907]: I0313 14:11:11.125603 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Mar 13 14:11:11 crc kubenswrapper[4907]: I0313 14:11:11.185056 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Mar 13 14:11:11 crc kubenswrapper[4907]: I0313 14:11:11.291594 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Mar 13 14:11:11 crc kubenswrapper[4907]: I0313 14:11:11.413974 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Mar 13 14:11:11 crc kubenswrapper[4907]: I0313 14:11:11.559392 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Mar 13 14:11:11 crc kubenswrapper[4907]: I0313 14:11:11.584267 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Mar 13 14:11:11 crc kubenswrapper[4907]: I0313 14:11:11.681723 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Mar 13 14:11:11 crc kubenswrapper[4907]: I0313 14:11:11.749628 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Mar 13 14:11:11 crc kubenswrapper[4907]: I0313 14:11:11.759330 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Mar 13 14:11:11 crc kubenswrapper[4907]: I0313 14:11:11.997266 4907 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Mar 13 14:11:11 crc kubenswrapper[4907]: I0313 14:11:11.997785 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Mar 13 14:11:12 crc kubenswrapper[4907]: I0313 14:11:12.036063 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Mar 13 14:11:12 crc kubenswrapper[4907]: I0313 14:11:12.061494 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Mar 13 14:11:12 crc kubenswrapper[4907]: I0313 14:11:12.196125 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Mar 13 14:11:12 crc kubenswrapper[4907]: I0313 14:11:12.243853 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Mar 13 14:11:12 crc kubenswrapper[4907]: I0313 14:11:12.288125 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Mar 13 14:11:12 crc kubenswrapper[4907]: I0313 14:11:12.353871 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Mar 13 14:11:12 crc kubenswrapper[4907]: I0313 14:11:12.427532 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Mar 13 14:11:12 crc kubenswrapper[4907]: I0313 14:11:12.559401 4907 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Mar 13 14:11:12 crc kubenswrapper[4907]: I0313 14:11:12.587241 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Mar 13 14:11:12 crc kubenswrapper[4907]: I0313 14:11:12.630182 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Mar 13 14:11:12 crc kubenswrapper[4907]: I0313 14:11:12.863034 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Mar 13 14:11:12 crc kubenswrapper[4907]: I0313 14:11:12.919249 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Mar 13 14:11:12 crc kubenswrapper[4907]: E0313 14:11:12.943467 4907 log.go:32] "RunPodSandbox from runtime service failed" err=< Mar 13 14:11:12 crc kubenswrapper[4907]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_route-controller-manager-5c879c7f8d-2mxxc_openshift-route-controller-manager_a704bdfc-46ff-4a8a-8d31-21a140b49c99_0(f75e30dd6fbe4d97da9cd85143325c74ecdd86dbb3207f61697dca55b64c1621): error adding pod openshift-route-controller-manager_route-controller-manager-5c879c7f8d-2mxxc to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"f75e30dd6fbe4d97da9cd85143325c74ecdd86dbb3207f61697dca55b64c1621" Netns:"/var/run/netns/208f84ca-664c-4bf4-addd-1a01472d31bb" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-route-controller-manager;K8S_POD_NAME=route-controller-manager-5c879c7f8d-2mxxc;K8S_POD_INFRA_CONTAINER_ID=f75e30dd6fbe4d97da9cd85143325c74ecdd86dbb3207f61697dca55b64c1621;K8S_POD_UID=a704bdfc-46ff-4a8a-8d31-21a140b49c99" Path:"" ERRORED: error configuring pod [openshift-route-controller-manager/route-controller-manager-5c879c7f8d-2mxxc] networking: Multus: [openshift-route-controller-manager/route-controller-manager-5c879c7f8d-2mxxc/a704bdfc-46ff-4a8a-8d31-21a140b49c99]: error setting the networks status, pod was already deleted: SetPodNetworkStatusAnnotation: failed to query the pod route-controller-manager-5c879c7f8d-2mxxc in out of cluster comm: pod "route-controller-manager-5c879c7f8d-2mxxc" not found Mar 13 14:11:12 crc kubenswrapper[4907]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Mar 13 14:11:12 crc kubenswrapper[4907]: > Mar 13 14:11:12 crc kubenswrapper[4907]: E0313 14:11:12.943596 4907 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err=< Mar 13 14:11:12 crc kubenswrapper[4907]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_route-controller-manager-5c879c7f8d-2mxxc_openshift-route-controller-manager_a704bdfc-46ff-4a8a-8d31-21a140b49c99_0(f75e30dd6fbe4d97da9cd85143325c74ecdd86dbb3207f61697dca55b64c1621): error adding pod openshift-route-controller-manager_route-controller-manager-5c879c7f8d-2mxxc to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"f75e30dd6fbe4d97da9cd85143325c74ecdd86dbb3207f61697dca55b64c1621" Netns:"/var/run/netns/208f84ca-664c-4bf4-addd-1a01472d31bb" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-route-controller-manager;K8S_POD_NAME=route-controller-manager-5c879c7f8d-2mxxc;K8S_POD_INFRA_CONTAINER_ID=f75e30dd6fbe4d97da9cd85143325c74ecdd86dbb3207f61697dca55b64c1621;K8S_POD_UID=a704bdfc-46ff-4a8a-8d31-21a140b49c99" Path:"" ERRORED: error configuring pod [openshift-route-controller-manager/route-controller-manager-5c879c7f8d-2mxxc] networking: Multus: [openshift-route-controller-manager/route-controller-manager-5c879c7f8d-2mxxc/a704bdfc-46ff-4a8a-8d31-21a140b49c99]: error setting the networks status, pod was already deleted: SetPodNetworkStatusAnnotation: failed to query the pod route-controller-manager-5c879c7f8d-2mxxc in out of cluster comm: pod "route-controller-manager-5c879c7f8d-2mxxc" not found Mar 13 14:11:12 crc kubenswrapper[4907]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Mar 13 14:11:12 crc kubenswrapper[4907]: > pod="openshift-route-controller-manager/route-controller-manager-5c879c7f8d-2mxxc" Mar 13 14:11:12 crc kubenswrapper[4907]: E0313 14:11:12.943640 4907 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err=< Mar 13 14:11:12 crc kubenswrapper[4907]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_route-controller-manager-5c879c7f8d-2mxxc_openshift-route-controller-manager_a704bdfc-46ff-4a8a-8d31-21a140b49c99_0(f75e30dd6fbe4d97da9cd85143325c74ecdd86dbb3207f61697dca55b64c1621): error adding pod openshift-route-controller-manager_route-controller-manager-5c879c7f8d-2mxxc to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"f75e30dd6fbe4d97da9cd85143325c74ecdd86dbb3207f61697dca55b64c1621" Netns:"/var/run/netns/208f84ca-664c-4bf4-addd-1a01472d31bb" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-route-controller-manager;K8S_POD_NAME=route-controller-manager-5c879c7f8d-2mxxc;K8S_POD_INFRA_CONTAINER_ID=f75e30dd6fbe4d97da9cd85143325c74ecdd86dbb3207f61697dca55b64c1621;K8S_POD_UID=a704bdfc-46ff-4a8a-8d31-21a140b49c99" Path:"" ERRORED: error configuring pod [openshift-route-controller-manager/route-controller-manager-5c879c7f8d-2mxxc] networking: Multus: [openshift-route-controller-manager/route-controller-manager-5c879c7f8d-2mxxc/a704bdfc-46ff-4a8a-8d31-21a140b49c99]: error setting the networks status, pod was already deleted: SetPodNetworkStatusAnnotation: failed to query the pod route-controller-manager-5c879c7f8d-2mxxc in out of cluster comm: pod "route-controller-manager-5c879c7f8d-2mxxc" not found Mar 13 14:11:12 crc kubenswrapper[4907]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Mar 13 14:11:12 crc kubenswrapper[4907]: > pod="openshift-route-controller-manager/route-controller-manager-5c879c7f8d-2mxxc" Mar 13 14:11:12 crc kubenswrapper[4907]: E0313 14:11:12.944540 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"route-controller-manager-5c879c7f8d-2mxxc_openshift-route-controller-manager(a704bdfc-46ff-4a8a-8d31-21a140b49c99)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"route-controller-manager-5c879c7f8d-2mxxc_openshift-route-controller-manager(a704bdfc-46ff-4a8a-8d31-21a140b49c99)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_route-controller-manager-5c879c7f8d-2mxxc_openshift-route-controller-manager_a704bdfc-46ff-4a8a-8d31-21a140b49c99_0(f75e30dd6fbe4d97da9cd85143325c74ecdd86dbb3207f61697dca55b64c1621): error adding pod openshift-route-controller-manager_route-controller-manager-5c879c7f8d-2mxxc to CNI network \\\"multus-cni-network\\\": plugin type=\\\"multus-shim\\\" name=\\\"multus-cni-network\\\" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:\\\"f75e30dd6fbe4d97da9cd85143325c74ecdd86dbb3207f61697dca55b64c1621\\\" Netns:\\\"/var/run/netns/208f84ca-664c-4bf4-addd-1a01472d31bb\\\" IfName:\\\"eth0\\\" Args:\\\"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-route-controller-manager;K8S_POD_NAME=route-controller-manager-5c879c7f8d-2mxxc;K8S_POD_INFRA_CONTAINER_ID=f75e30dd6fbe4d97da9cd85143325c74ecdd86dbb3207f61697dca55b64c1621;K8S_POD_UID=a704bdfc-46ff-4a8a-8d31-21a140b49c99\\\" Path:\\\"\\\" ERRORED: error configuring pod [openshift-route-controller-manager/route-controller-manager-5c879c7f8d-2mxxc] networking: Multus: [openshift-route-controller-manager/route-controller-manager-5c879c7f8d-2mxxc/a704bdfc-46ff-4a8a-8d31-21a140b49c99]: error setting the networks status, pod was already deleted: SetPodNetworkStatusAnnotation: failed to query the pod route-controller-manager-5c879c7f8d-2mxxc in out of cluster comm: pod \\\"route-controller-manager-5c879c7f8d-2mxxc\\\" not found\\n': StdinData: {\\\"binDir\\\":\\\"/var/lib/cni/bin\\\",\\\"clusterNetwork\\\":\\\"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf\\\",\\\"cniVersion\\\":\\\"0.3.1\\\",\\\"daemonSocketDir\\\":\\\"/run/multus/socket\\\",\\\"globalNamespaces\\\":\\\"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv\\\",\\\"logLevel\\\":\\\"verbose\\\",\\\"logToStderr\\\":true,\\\"name\\\":\\\"multus-cni-network\\\",\\\"namespaceIsolation\\\":true,\\\"type\\\":\\\"multus-shim\\\"}\"" pod="openshift-route-controller-manager/route-controller-manager-5c879c7f8d-2mxxc" podUID="a704bdfc-46ff-4a8a-8d31-21a140b49c99" Mar 13 14:11:12 crc kubenswrapper[4907]: I0313 14:11:12.967857 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Mar 13 14:11:13 crc kubenswrapper[4907]: E0313 14:11:13.000229 4907 log.go:32] "RunPodSandbox from runtime service failed" err=< Mar 13 14:11:13 crc kubenswrapper[4907]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_oauth-openshift-7c65b78cb7-ld7fp_openshift-authentication_8746df2e-524e-42a7-9ca5-c5300c0a1a09_0(c45269bc024ec530d94f1f1c216da78b5947c8922810d8bce846a0c2e15b9b0e): error adding pod openshift-authentication_oauth-openshift-7c65b78cb7-ld7fp to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"c45269bc024ec530d94f1f1c216da78b5947c8922810d8bce846a0c2e15b9b0e" Netns:"/var/run/netns/1829481a-098a-4a48-b8c3-231563e389b5" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-authentication;K8S_POD_NAME=oauth-openshift-7c65b78cb7-ld7fp;K8S_POD_INFRA_CONTAINER_ID=c45269bc024ec530d94f1f1c216da78b5947c8922810d8bce846a0c2e15b9b0e;K8S_POD_UID=8746df2e-524e-42a7-9ca5-c5300c0a1a09" Path:"" ERRORED: error configuring pod [openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp] networking: Multus: [openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp/8746df2e-524e-42a7-9ca5-c5300c0a1a09]: error setting the networks status, pod was already deleted: SetPodNetworkStatusAnnotation: failed to query the pod oauth-openshift-7c65b78cb7-ld7fp in out of cluster comm: pod "oauth-openshift-7c65b78cb7-ld7fp" not found Mar 13 14:11:13 crc kubenswrapper[4907]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Mar 13 14:11:13 crc kubenswrapper[4907]: > Mar 13 14:11:13 crc kubenswrapper[4907]: E0313 14:11:13.000300 4907 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err=< Mar 13 14:11:13 crc kubenswrapper[4907]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_oauth-openshift-7c65b78cb7-ld7fp_openshift-authentication_8746df2e-524e-42a7-9ca5-c5300c0a1a09_0(c45269bc024ec530d94f1f1c216da78b5947c8922810d8bce846a0c2e15b9b0e): error adding pod openshift-authentication_oauth-openshift-7c65b78cb7-ld7fp to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"c45269bc024ec530d94f1f1c216da78b5947c8922810d8bce846a0c2e15b9b0e" Netns:"/var/run/netns/1829481a-098a-4a48-b8c3-231563e389b5" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-authentication;K8S_POD_NAME=oauth-openshift-7c65b78cb7-ld7fp;K8S_POD_INFRA_CONTAINER_ID=c45269bc024ec530d94f1f1c216da78b5947c8922810d8bce846a0c2e15b9b0e;K8S_POD_UID=8746df2e-524e-42a7-9ca5-c5300c0a1a09" Path:"" ERRORED: error configuring pod [openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp] networking: Multus: [openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp/8746df2e-524e-42a7-9ca5-c5300c0a1a09]: error setting the networks status, pod was already deleted: SetPodNetworkStatusAnnotation: failed to query the pod oauth-openshift-7c65b78cb7-ld7fp in out of cluster comm: pod "oauth-openshift-7c65b78cb7-ld7fp" not found Mar 13 14:11:13 crc kubenswrapper[4907]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Mar 13 14:11:13 crc kubenswrapper[4907]: > pod="openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp" Mar 13 14:11:13 crc kubenswrapper[4907]: E0313 14:11:13.000331 4907 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err=< Mar 13 14:11:13 crc kubenswrapper[4907]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_oauth-openshift-7c65b78cb7-ld7fp_openshift-authentication_8746df2e-524e-42a7-9ca5-c5300c0a1a09_0(c45269bc024ec530d94f1f1c216da78b5947c8922810d8bce846a0c2e15b9b0e): error adding pod openshift-authentication_oauth-openshift-7c65b78cb7-ld7fp to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"c45269bc024ec530d94f1f1c216da78b5947c8922810d8bce846a0c2e15b9b0e" Netns:"/var/run/netns/1829481a-098a-4a48-b8c3-231563e389b5" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-authentication;K8S_POD_NAME=oauth-openshift-7c65b78cb7-ld7fp;K8S_POD_INFRA_CONTAINER_ID=c45269bc024ec530d94f1f1c216da78b5947c8922810d8bce846a0c2e15b9b0e;K8S_POD_UID=8746df2e-524e-42a7-9ca5-c5300c0a1a09" Path:"" ERRORED: error configuring pod [openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp] networking: Multus: [openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp/8746df2e-524e-42a7-9ca5-c5300c0a1a09]: error setting the networks status, pod was already deleted: SetPodNetworkStatusAnnotation: failed to query the pod oauth-openshift-7c65b78cb7-ld7fp in out of cluster comm: pod "oauth-openshift-7c65b78cb7-ld7fp" not found Mar 13 14:11:13 crc kubenswrapper[4907]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Mar 13 14:11:13 crc kubenswrapper[4907]: > pod="openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp" Mar 13 14:11:13 crc kubenswrapper[4907]: E0313 14:11:13.000396 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"oauth-openshift-7c65b78cb7-ld7fp_openshift-authentication(8746df2e-524e-42a7-9ca5-c5300c0a1a09)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"oauth-openshift-7c65b78cb7-ld7fp_openshift-authentication(8746df2e-524e-42a7-9ca5-c5300c0a1a09)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_oauth-openshift-7c65b78cb7-ld7fp_openshift-authentication_8746df2e-524e-42a7-9ca5-c5300c0a1a09_0(c45269bc024ec530d94f1f1c216da78b5947c8922810d8bce846a0c2e15b9b0e): error adding pod openshift-authentication_oauth-openshift-7c65b78cb7-ld7fp to CNI network \\\"multus-cni-network\\\": plugin type=\\\"multus-shim\\\" name=\\\"multus-cni-network\\\" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:\\\"c45269bc024ec530d94f1f1c216da78b5947c8922810d8bce846a0c2e15b9b0e\\\" Netns:\\\"/var/run/netns/1829481a-098a-4a48-b8c3-231563e389b5\\\" IfName:\\\"eth0\\\" Args:\\\"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-authentication;K8S_POD_NAME=oauth-openshift-7c65b78cb7-ld7fp;K8S_POD_INFRA_CONTAINER_ID=c45269bc024ec530d94f1f1c216da78b5947c8922810d8bce846a0c2e15b9b0e;K8S_POD_UID=8746df2e-524e-42a7-9ca5-c5300c0a1a09\\\" Path:\\\"\\\" ERRORED: error configuring pod [openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp] networking: Multus: [openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp/8746df2e-524e-42a7-9ca5-c5300c0a1a09]: error setting the networks status, pod was already deleted: SetPodNetworkStatusAnnotation: failed to query the pod oauth-openshift-7c65b78cb7-ld7fp in out of cluster comm: pod \\\"oauth-openshift-7c65b78cb7-ld7fp\\\" not found\\n': StdinData: {\\\"binDir\\\":\\\"/var/lib/cni/bin\\\",\\\"clusterNetwork\\\":\\\"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf\\\",\\\"cniVersion\\\":\\\"0.3.1\\\",\\\"daemonSocketDir\\\":\\\"/run/multus/socket\\\",\\\"globalNamespaces\\\":\\\"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv\\\",\\\"logLevel\\\":\\\"verbose\\\",\\\"logToStderr\\\":true,\\\"name\\\":\\\"multus-cni-network\\\",\\\"namespaceIsolation\\\":true,\\\"type\\\":\\\"multus-shim\\\"}\"" pod="openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp" podUID="8746df2e-524e-42a7-9ca5-c5300c0a1a09" Mar 13 14:11:13 crc kubenswrapper[4907]: I0313 14:11:13.003558 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Mar 13 14:11:13 crc kubenswrapper[4907]: E0313 14:11:13.038713 4907 log.go:32] "RunPodSandbox from runtime service failed" err=< Mar 13 14:11:13 crc kubenswrapper[4907]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_controller-manager-5f4c4b8dd4-9sl4b_openshift-controller-manager_03ca839d-b5ba-4b55-87b3-48abca197163_0(5c1ba8ff43ed33e226001737ee4130c9659d7444a294f4dcc3db6844b59f44df): error adding pod openshift-controller-manager_controller-manager-5f4c4b8dd4-9sl4b to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"5c1ba8ff43ed33e226001737ee4130c9659d7444a294f4dcc3db6844b59f44df" Netns:"/var/run/netns/5a26f034-0cf7-47c9-babd-1832d27300bb" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-controller-manager;K8S_POD_NAME=controller-manager-5f4c4b8dd4-9sl4b;K8S_POD_INFRA_CONTAINER_ID=5c1ba8ff43ed33e226001737ee4130c9659d7444a294f4dcc3db6844b59f44df;K8S_POD_UID=03ca839d-b5ba-4b55-87b3-48abca197163" Path:"" ERRORED: error configuring pod [openshift-controller-manager/controller-manager-5f4c4b8dd4-9sl4b] networking: Multus: [openshift-controller-manager/controller-manager-5f4c4b8dd4-9sl4b/03ca839d-b5ba-4b55-87b3-48abca197163]: error setting the networks status, pod was already deleted: SetPodNetworkStatusAnnotation: failed to query the pod controller-manager-5f4c4b8dd4-9sl4b in out of cluster comm: pod "controller-manager-5f4c4b8dd4-9sl4b" not found Mar 13 14:11:13 crc kubenswrapper[4907]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Mar 13 14:11:13 crc kubenswrapper[4907]: > Mar 13 14:11:13 crc kubenswrapper[4907]: E0313 14:11:13.038837 4907 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err=< Mar 13 14:11:13 crc kubenswrapper[4907]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_controller-manager-5f4c4b8dd4-9sl4b_openshift-controller-manager_03ca839d-b5ba-4b55-87b3-48abca197163_0(5c1ba8ff43ed33e226001737ee4130c9659d7444a294f4dcc3db6844b59f44df): error adding pod openshift-controller-manager_controller-manager-5f4c4b8dd4-9sl4b to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"5c1ba8ff43ed33e226001737ee4130c9659d7444a294f4dcc3db6844b59f44df" Netns:"/var/run/netns/5a26f034-0cf7-47c9-babd-1832d27300bb" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-controller-manager;K8S_POD_NAME=controller-manager-5f4c4b8dd4-9sl4b;K8S_POD_INFRA_CONTAINER_ID=5c1ba8ff43ed33e226001737ee4130c9659d7444a294f4dcc3db6844b59f44df;K8S_POD_UID=03ca839d-b5ba-4b55-87b3-48abca197163" Path:"" ERRORED: error configuring pod [openshift-controller-manager/controller-manager-5f4c4b8dd4-9sl4b] networking: Multus: [openshift-controller-manager/controller-manager-5f4c4b8dd4-9sl4b/03ca839d-b5ba-4b55-87b3-48abca197163]: error setting the networks status, pod was already deleted: SetPodNetworkStatusAnnotation: failed to query the pod controller-manager-5f4c4b8dd4-9sl4b in out of cluster comm: pod "controller-manager-5f4c4b8dd4-9sl4b" not found Mar 13 14:11:13 crc kubenswrapper[4907]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Mar 13 14:11:13 crc kubenswrapper[4907]: > pod="openshift-controller-manager/controller-manager-5f4c4b8dd4-9sl4b" Mar 13 14:11:13 crc kubenswrapper[4907]: E0313 14:11:13.038869 4907 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err=< Mar 13 14:11:13 crc kubenswrapper[4907]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_controller-manager-5f4c4b8dd4-9sl4b_openshift-controller-manager_03ca839d-b5ba-4b55-87b3-48abca197163_0(5c1ba8ff43ed33e226001737ee4130c9659d7444a294f4dcc3db6844b59f44df): error adding pod openshift-controller-manager_controller-manager-5f4c4b8dd4-9sl4b to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"5c1ba8ff43ed33e226001737ee4130c9659d7444a294f4dcc3db6844b59f44df" Netns:"/var/run/netns/5a26f034-0cf7-47c9-babd-1832d27300bb" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-controller-manager;K8S_POD_NAME=controller-manager-5f4c4b8dd4-9sl4b;K8S_POD_INFRA_CONTAINER_ID=5c1ba8ff43ed33e226001737ee4130c9659d7444a294f4dcc3db6844b59f44df;K8S_POD_UID=03ca839d-b5ba-4b55-87b3-48abca197163" Path:"" ERRORED: error configuring pod [openshift-controller-manager/controller-manager-5f4c4b8dd4-9sl4b] networking: Multus: [openshift-controller-manager/controller-manager-5f4c4b8dd4-9sl4b/03ca839d-b5ba-4b55-87b3-48abca197163]: error setting the networks status, pod was already deleted: SetPodNetworkStatusAnnotation: failed to query the pod controller-manager-5f4c4b8dd4-9sl4b in out of cluster comm: pod "controller-manager-5f4c4b8dd4-9sl4b" not found Mar 13 14:11:13 crc kubenswrapper[4907]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Mar 13 14:11:13 crc kubenswrapper[4907]: > pod="openshift-controller-manager/controller-manager-5f4c4b8dd4-9sl4b" Mar 13 14:11:13 crc kubenswrapper[4907]: E0313 14:11:13.038987 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"controller-manager-5f4c4b8dd4-9sl4b_openshift-controller-manager(03ca839d-b5ba-4b55-87b3-48abca197163)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"controller-manager-5f4c4b8dd4-9sl4b_openshift-controller-manager(03ca839d-b5ba-4b55-87b3-48abca197163)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_controller-manager-5f4c4b8dd4-9sl4b_openshift-controller-manager_03ca839d-b5ba-4b55-87b3-48abca197163_0(5c1ba8ff43ed33e226001737ee4130c9659d7444a294f4dcc3db6844b59f44df): error adding pod openshift-controller-manager_controller-manager-5f4c4b8dd4-9sl4b to CNI network \\\"multus-cni-network\\\": plugin type=\\\"multus-shim\\\" name=\\\"multus-cni-network\\\" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:\\\"5c1ba8ff43ed33e226001737ee4130c9659d7444a294f4dcc3db6844b59f44df\\\" Netns:\\\"/var/run/netns/5a26f034-0cf7-47c9-babd-1832d27300bb\\\" IfName:\\\"eth0\\\" Args:\\\"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-controller-manager;K8S_POD_NAME=controller-manager-5f4c4b8dd4-9sl4b;K8S_POD_INFRA_CONTAINER_ID=5c1ba8ff43ed33e226001737ee4130c9659d7444a294f4dcc3db6844b59f44df;K8S_POD_UID=03ca839d-b5ba-4b55-87b3-48abca197163\\\" Path:\\\"\\\" ERRORED: error configuring pod [openshift-controller-manager/controller-manager-5f4c4b8dd4-9sl4b] networking: Multus: [openshift-controller-manager/controller-manager-5f4c4b8dd4-9sl4b/03ca839d-b5ba-4b55-87b3-48abca197163]: error setting the networks status, pod was already deleted: SetPodNetworkStatusAnnotation: failed to query the pod controller-manager-5f4c4b8dd4-9sl4b in out of cluster comm: pod \\\"controller-manager-5f4c4b8dd4-9sl4b\\\" not found\\n': StdinData: {\\\"binDir\\\":\\\"/var/lib/cni/bin\\\",\\\"clusterNetwork\\\":\\\"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf\\\",\\\"cniVersion\\\":\\\"0.3.1\\\",\\\"daemonSocketDir\\\":\\\"/run/multus/socket\\\",\\\"globalNamespaces\\\":\\\"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv\\\",\\\"logLevel\\\":\\\"verbose\\\",\\\"logToStderr\\\":true,\\\"name\\\":\\\"multus-cni-network\\\",\\\"namespaceIsolation\\\":true,\\\"type\\\":\\\"multus-shim\\\"}\"" pod="openshift-controller-manager/controller-manager-5f4c4b8dd4-9sl4b" podUID="03ca839d-b5ba-4b55-87b3-48abca197163" Mar 13 14:11:13 crc kubenswrapper[4907]: I0313 14:11:13.040402 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Mar 13 14:11:13 crc kubenswrapper[4907]: I0313 14:11:13.090017 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Mar 13 14:11:13 crc kubenswrapper[4907]: I0313 14:11:13.134429 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Mar 13 14:11:13 crc kubenswrapper[4907]: I0313 14:11:13.164782 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Mar 13 14:11:13 crc kubenswrapper[4907]: I0313 14:11:13.197923 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Mar 13 14:11:13 crc kubenswrapper[4907]: I0313 14:11:13.227044 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Mar 13 14:11:13 crc kubenswrapper[4907]: I0313 14:11:13.293909 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Mar 13 14:11:13 crc kubenswrapper[4907]: I0313 14:11:13.506297 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Mar 13 14:11:13 crc kubenswrapper[4907]: I0313 14:11:13.525749 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Mar 13 14:11:13 crc kubenswrapper[4907]: I0313 14:11:13.584660 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Mar 13 14:11:13 crc kubenswrapper[4907]: I0313 14:11:13.621929 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Mar 13 14:11:13 crc kubenswrapper[4907]: I0313 14:11:13.632322 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Mar 13 14:11:13 crc kubenswrapper[4907]: I0313 14:11:13.745237 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Mar 13 14:11:13 crc kubenswrapper[4907]: I0313 14:11:13.859126 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Mar 13 14:11:13 crc kubenswrapper[4907]: I0313 14:11:13.872872 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Mar 13 14:11:13 crc kubenswrapper[4907]: I0313 14:11:13.977113 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Mar 13 14:11:13 crc kubenswrapper[4907]: I0313 14:11:13.993914 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Mar 13 14:11:14 crc kubenswrapper[4907]: I0313 14:11:14.021235 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Mar 13 14:11:14 crc kubenswrapper[4907]: I0313 14:11:14.049114 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Mar 13 14:11:14 crc kubenswrapper[4907]: I0313 14:11:14.233506 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Mar 13 14:11:14 crc kubenswrapper[4907]: I0313 14:11:14.260650 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Mar 13 14:11:14 crc kubenswrapper[4907]: I0313 14:11:14.294789 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Mar 13 14:11:14 crc kubenswrapper[4907]: I0313 14:11:14.296669 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Mar 13 14:11:14 crc kubenswrapper[4907]: I0313 14:11:14.413356 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Mar 13 14:11:14 crc kubenswrapper[4907]: I0313 14:11:14.415514 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Mar 13 14:11:14 crc kubenswrapper[4907]: I0313 14:11:14.461650 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Mar 13 14:11:14 crc kubenswrapper[4907]: I0313 14:11:14.490782 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Mar 13 14:11:14 crc kubenswrapper[4907]: I0313 14:11:14.598728 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Mar 13 14:11:14 crc kubenswrapper[4907]: I0313 14:11:14.618280 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Mar 13 14:11:14 crc kubenswrapper[4907]: I0313 14:11:14.657615 4907 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Mar 13 14:11:14 crc kubenswrapper[4907]: I0313 14:11:14.657940 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://d4f9add6c133d321c4423223d8d40952daa019ea189e4fd19cc3ba6806cde1d4" gracePeriod=5 Mar 13 14:11:14 crc kubenswrapper[4907]: I0313 14:11:14.700326 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Mar 13 14:11:14 crc kubenswrapper[4907]: I0313 14:11:14.746459 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Mar 13 14:11:14 crc kubenswrapper[4907]: I0313 14:11:14.761322 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Mar 13 14:11:14 crc kubenswrapper[4907]: I0313 14:11:14.846540 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Mar 13 14:11:14 crc kubenswrapper[4907]: I0313 14:11:14.863606 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Mar 13 14:11:14 crc kubenswrapper[4907]: I0313 14:11:14.870912 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Mar 13 14:11:14 crc kubenswrapper[4907]: I0313 14:11:14.881372 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Mar 13 14:11:14 crc kubenswrapper[4907]: I0313 14:11:14.910621 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Mar 13 14:11:14 crc kubenswrapper[4907]: I0313 14:11:14.942074 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Mar 13 14:11:14 crc kubenswrapper[4907]: I0313 14:11:14.972004 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Mar 13 14:11:15 crc kubenswrapper[4907]: I0313 14:11:15.080213 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Mar 13 14:11:15 crc kubenswrapper[4907]: I0313 14:11:15.111398 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Mar 13 14:11:15 crc kubenswrapper[4907]: I0313 14:11:15.156561 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Mar 13 14:11:15 crc kubenswrapper[4907]: I0313 14:11:15.224266 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Mar 13 14:11:15 crc kubenswrapper[4907]: I0313 14:11:15.265269 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Mar 13 14:11:15 crc kubenswrapper[4907]: I0313 14:11:15.283335 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Mar 13 14:11:15 crc kubenswrapper[4907]: I0313 14:11:15.387635 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Mar 13 14:11:15 crc kubenswrapper[4907]: I0313 14:11:15.387849 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Mar 13 14:11:15 crc kubenswrapper[4907]: I0313 14:11:15.406396 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Mar 13 14:11:15 crc kubenswrapper[4907]: I0313 14:11:15.479217 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Mar 13 14:11:15 crc kubenswrapper[4907]: I0313 14:11:15.479630 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Mar 13 14:11:15 crc kubenswrapper[4907]: I0313 14:11:15.480957 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Mar 13 14:11:15 crc kubenswrapper[4907]: I0313 14:11:15.766632 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Mar 13 14:11:15 crc kubenswrapper[4907]: I0313 14:11:15.773754 4907 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Mar 13 14:11:15 crc kubenswrapper[4907]: I0313 14:11:15.787283 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Mar 13 14:11:15 crc kubenswrapper[4907]: I0313 14:11:15.825081 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Mar 13 14:11:15 crc kubenswrapper[4907]: I0313 14:11:15.898787 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Mar 13 14:11:15 crc kubenswrapper[4907]: I0313 14:11:15.926795 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Mar 13 14:11:15 crc kubenswrapper[4907]: I0313 14:11:15.927494 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Mar 13 14:11:15 crc kubenswrapper[4907]: I0313 14:11:15.931188 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Mar 13 14:11:15 crc kubenswrapper[4907]: I0313 14:11:15.971633 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Mar 13 14:11:16 crc kubenswrapper[4907]: I0313 14:11:16.010562 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Mar 13 14:11:16 crc kubenswrapper[4907]: I0313 14:11:16.031998 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Mar 13 14:11:16 crc kubenswrapper[4907]: I0313 14:11:16.130230 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Mar 13 14:11:16 crc kubenswrapper[4907]: I0313 14:11:16.248607 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Mar 13 14:11:16 crc kubenswrapper[4907]: I0313 14:11:16.290010 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Mar 13 14:11:16 crc kubenswrapper[4907]: I0313 14:11:16.324510 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Mar 13 14:11:16 crc kubenswrapper[4907]: I0313 14:11:16.393523 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Mar 13 14:11:16 crc kubenswrapper[4907]: I0313 14:11:16.459815 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Mar 13 14:11:16 crc kubenswrapper[4907]: I0313 14:11:16.593924 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Mar 13 14:11:16 crc kubenswrapper[4907]: I0313 14:11:16.650282 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Mar 13 14:11:16 crc kubenswrapper[4907]: I0313 14:11:16.656244 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Mar 13 14:11:16 crc kubenswrapper[4907]: I0313 14:11:16.882919 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Mar 13 14:11:16 crc kubenswrapper[4907]: I0313 14:11:16.895967 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Mar 13 14:11:17 crc kubenswrapper[4907]: I0313 14:11:17.024061 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Mar 13 14:11:17 crc kubenswrapper[4907]: I0313 14:11:17.039914 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Mar 13 14:11:17 crc kubenswrapper[4907]: I0313 14:11:17.293662 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Mar 13 14:11:17 crc kubenswrapper[4907]: I0313 14:11:17.313386 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Mar 13 14:11:17 crc kubenswrapper[4907]: I0313 14:11:17.315191 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Mar 13 14:11:17 crc kubenswrapper[4907]: I0313 14:11:17.390267 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Mar 13 14:11:17 crc kubenswrapper[4907]: I0313 14:11:17.472966 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Mar 13 14:11:17 crc kubenswrapper[4907]: I0313 14:11:17.649465 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Mar 13 14:11:17 crc kubenswrapper[4907]: I0313 14:11:17.747068 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Mar 13 14:11:17 crc kubenswrapper[4907]: I0313 14:11:17.923872 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Mar 13 14:11:18 crc kubenswrapper[4907]: I0313 14:11:18.034281 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Mar 13 14:11:18 crc kubenswrapper[4907]: I0313 14:11:18.105934 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Mar 13 14:11:18 crc kubenswrapper[4907]: I0313 14:11:18.167740 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Mar 13 14:11:18 crc kubenswrapper[4907]: I0313 14:11:18.240889 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Mar 13 14:11:18 crc kubenswrapper[4907]: I0313 14:11:18.409456 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Mar 13 14:11:18 crc kubenswrapper[4907]: I0313 14:11:18.414809 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Mar 13 14:11:18 crc kubenswrapper[4907]: I0313 14:11:18.469139 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Mar 13 14:11:18 crc kubenswrapper[4907]: I0313 14:11:18.499685 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Mar 13 14:11:18 crc kubenswrapper[4907]: I0313 14:11:18.741176 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Mar 13 14:11:19 crc kubenswrapper[4907]: I0313 14:11:19.059500 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Mar 13 14:11:19 crc kubenswrapper[4907]: I0313 14:11:19.075649 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Mar 13 14:11:19 crc kubenswrapper[4907]: I0313 14:11:19.256190 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Mar 13 14:11:19 crc kubenswrapper[4907]: I0313 14:11:19.394705 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Mar 13 14:11:19 crc kubenswrapper[4907]: I0313 14:11:19.458135 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Mar 13 14:11:19 crc kubenswrapper[4907]: I0313 14:11:19.603164 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Mar 13 14:11:19 crc kubenswrapper[4907]: I0313 14:11:19.897333 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Mar 13 14:11:20 crc kubenswrapper[4907]: I0313 14:11:20.050799 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Mar 13 14:11:20 crc kubenswrapper[4907]: I0313 14:11:20.154625 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Mar 13 14:11:20 crc kubenswrapper[4907]: I0313 14:11:20.154695 4907 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="d4f9add6c133d321c4423223d8d40952daa019ea189e4fd19cc3ba6806cde1d4" exitCode=137 Mar 13 14:11:20 crc kubenswrapper[4907]: I0313 14:11:20.265244 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Mar 13 14:11:20 crc kubenswrapper[4907]: I0313 14:11:20.265353 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 13 14:11:20 crc kubenswrapper[4907]: I0313 14:11:20.310571 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 13 14:11:20 crc kubenswrapper[4907]: I0313 14:11:20.310629 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 13 14:11:20 crc kubenswrapper[4907]: I0313 14:11:20.310662 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 13 14:11:20 crc kubenswrapper[4907]: I0313 14:11:20.310686 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 13 14:11:20 crc kubenswrapper[4907]: I0313 14:11:20.310757 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 14:11:20 crc kubenswrapper[4907]: I0313 14:11:20.310823 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 14:11:20 crc kubenswrapper[4907]: I0313 14:11:20.310896 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 14:11:20 crc kubenswrapper[4907]: I0313 14:11:20.310963 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 14:11:20 crc kubenswrapper[4907]: I0313 14:11:20.311177 4907 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Mar 13 14:11:20 crc kubenswrapper[4907]: I0313 14:11:20.311204 4907 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Mar 13 14:11:20 crc kubenswrapper[4907]: I0313 14:11:20.311226 4907 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Mar 13 14:11:20 crc kubenswrapper[4907]: I0313 14:11:20.311244 4907 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Mar 13 14:11:20 crc kubenswrapper[4907]: I0313 14:11:20.413084 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 13 14:11:20 crc kubenswrapper[4907]: I0313 14:11:20.428998 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 14:11:20 crc kubenswrapper[4907]: I0313 14:11:20.515337 4907 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Mar 13 14:11:21 crc kubenswrapper[4907]: I0313 14:11:21.101700 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Mar 13 14:11:21 crc kubenswrapper[4907]: I0313 14:11:21.166163 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Mar 13 14:11:21 crc kubenswrapper[4907]: I0313 14:11:21.166251 4907 scope.go:117] "RemoveContainer" containerID="d4f9add6c133d321c4423223d8d40952daa019ea189e4fd19cc3ba6806cde1d4" Mar 13 14:11:21 crc kubenswrapper[4907]: I0313 14:11:21.166409 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 13 14:11:21 crc kubenswrapper[4907]: I0313 14:11:21.408731 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Mar 13 14:11:21 crc kubenswrapper[4907]: I0313 14:11:21.672861 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Mar 13 14:11:21 crc kubenswrapper[4907]: I0313 14:11:21.793240 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Mar 13 14:11:23 crc kubenswrapper[4907]: I0313 14:11:23.781936 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp" Mar 13 14:11:23 crc kubenswrapper[4907]: I0313 14:11:23.782546 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp" Mar 13 14:11:24 crc kubenswrapper[4907]: I0313 14:11:24.326606 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp"] Mar 13 14:11:25 crc kubenswrapper[4907]: I0313 14:11:25.196991 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp" event={"ID":"8746df2e-524e-42a7-9ca5-c5300c0a1a09","Type":"ContainerStarted","Data":"27be89dc86edaee6c3a0b2c99e08d3ec32bb36d3d79ee3a0893758666c47e531"} Mar 13 14:11:25 crc kubenswrapper[4907]: I0313 14:11:25.197542 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp" Mar 13 14:11:25 crc kubenswrapper[4907]: I0313 14:11:25.197565 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp" event={"ID":"8746df2e-524e-42a7-9ca5-c5300c0a1a09","Type":"ContainerStarted","Data":"e4b37a2df1824ba52cb67284261294565312dc7c7f4b756f55bce06a3ca979d2"} Mar 13 14:11:25 crc kubenswrapper[4907]: I0313 14:11:25.208116 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp" Mar 13 14:11:25 crc kubenswrapper[4907]: I0313 14:11:25.233091 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp" podStartSLOduration=53.233048975 podStartE2EDuration="53.233048975s" podCreationTimestamp="2026-03-13 14:10:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:11:25.224326203 +0000 UTC m=+384.124113912" watchObservedRunningTime="2026-03-13 14:11:25.233048975 +0000 UTC m=+384.132836704" Mar 13 14:11:26 crc kubenswrapper[4907]: I0313 14:11:26.782330 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5c879c7f8d-2mxxc" Mar 13 14:11:26 crc kubenswrapper[4907]: I0313 14:11:26.783038 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5c879c7f8d-2mxxc" Mar 13 14:11:27 crc kubenswrapper[4907]: I0313 14:11:27.029467 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5c879c7f8d-2mxxc"] Mar 13 14:11:27 crc kubenswrapper[4907]: I0313 14:11:27.217200 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5c879c7f8d-2mxxc" event={"ID":"a704bdfc-46ff-4a8a-8d31-21a140b49c99","Type":"ContainerStarted","Data":"44caeeeda87dc10eb9a28d60f555da1b49720cc085c4d59ad8abb1edafade068"} Mar 13 14:11:27 crc kubenswrapper[4907]: I0313 14:11:27.218029 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5c879c7f8d-2mxxc" event={"ID":"a704bdfc-46ff-4a8a-8d31-21a140b49c99","Type":"ContainerStarted","Data":"daed4ef5759632eb1579bb1e0cd602bb20107fbcc89063bcb1d061a61c877b80"} Mar 13 14:11:27 crc kubenswrapper[4907]: I0313 14:11:27.250600 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-5c879c7f8d-2mxxc" podStartSLOduration=55.250561214 podStartE2EDuration="55.250561214s" podCreationTimestamp="2026-03-13 14:10:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:11:27.248256184 +0000 UTC m=+386.148043903" watchObservedRunningTime="2026-03-13 14:11:27.250561214 +0000 UTC m=+386.150348933" Mar 13 14:11:27 crc kubenswrapper[4907]: I0313 14:11:27.782043 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5f4c4b8dd4-9sl4b" Mar 13 14:11:27 crc kubenswrapper[4907]: I0313 14:11:27.782726 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5f4c4b8dd4-9sl4b" Mar 13 14:11:28 crc kubenswrapper[4907]: I0313 14:11:28.225162 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-5c879c7f8d-2mxxc" Mar 13 14:11:28 crc kubenswrapper[4907]: I0313 14:11:28.233055 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-5c879c7f8d-2mxxc" Mar 13 14:11:28 crc kubenswrapper[4907]: I0313 14:11:28.331248 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5f4c4b8dd4-9sl4b"] Mar 13 14:11:28 crc kubenswrapper[4907]: W0313 14:11:28.339508 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod03ca839d_b5ba_4b55_87b3_48abca197163.slice/crio-f2f5e884109c6f169297c52fed01016438e86d1f9eba43973bb35837fea367d5 WatchSource:0}: Error finding container f2f5e884109c6f169297c52fed01016438e86d1f9eba43973bb35837fea367d5: Status 404 returned error can't find the container with id f2f5e884109c6f169297c52fed01016438e86d1f9eba43973bb35837fea367d5 Mar 13 14:11:29 crc kubenswrapper[4907]: I0313 14:11:29.234288 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5f4c4b8dd4-9sl4b" event={"ID":"03ca839d-b5ba-4b55-87b3-48abca197163","Type":"ContainerStarted","Data":"1b8fbae953b8242b87d46728d1b3fbed97457d9be26f86a62170e7caaa324132"} Mar 13 14:11:29 crc kubenswrapper[4907]: I0313 14:11:29.234808 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5f4c4b8dd4-9sl4b" event={"ID":"03ca839d-b5ba-4b55-87b3-48abca197163","Type":"ContainerStarted","Data":"f2f5e884109c6f169297c52fed01016438e86d1f9eba43973bb35837fea367d5"} Mar 13 14:11:29 crc kubenswrapper[4907]: I0313 14:11:29.258312 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-5f4c4b8dd4-9sl4b" podStartSLOduration=58.258284449 podStartE2EDuration="58.258284449s" podCreationTimestamp="2026-03-13 14:10:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:11:29.257460393 +0000 UTC m=+388.157248142" watchObservedRunningTime="2026-03-13 14:11:29.258284449 +0000 UTC m=+388.158072158" Mar 13 14:11:29 crc kubenswrapper[4907]: I0313 14:11:29.687520 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-5f4c4b8dd4-9sl4b" Mar 13 14:11:29 crc kubenswrapper[4907]: I0313 14:11:29.695760 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-5f4c4b8dd4-9sl4b" Mar 13 14:11:40 crc kubenswrapper[4907]: I0313 14:11:40.331411 4907 generic.go:334] "Generic (PLEG): container finished" podID="0290eb6f-0be7-4340-8ba8-ed6ea0662c33" containerID="48afa92be95670a8e5bf5c9bca4dd6ed4ea5b0ffd5c088a0a66004cb2137ee6b" exitCode=0 Mar 13 14:11:40 crc kubenswrapper[4907]: I0313 14:11:40.331488 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-5stnc" event={"ID":"0290eb6f-0be7-4340-8ba8-ed6ea0662c33","Type":"ContainerDied","Data":"48afa92be95670a8e5bf5c9bca4dd6ed4ea5b0ffd5c088a0a66004cb2137ee6b"} Mar 13 14:11:40 crc kubenswrapper[4907]: I0313 14:11:40.333506 4907 scope.go:117] "RemoveContainer" containerID="48afa92be95670a8e5bf5c9bca4dd6ed4ea5b0ffd5c088a0a66004cb2137ee6b" Mar 13 14:11:41 crc kubenswrapper[4907]: I0313 14:11:41.345495 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-5stnc" event={"ID":"0290eb6f-0be7-4340-8ba8-ed6ea0662c33","Type":"ContainerStarted","Data":"a4c4a9be5faa2f17ce6e81edeed8ac29998640aca39360db9beeef84d8b0752a"} Mar 13 14:11:41 crc kubenswrapper[4907]: I0313 14:11:41.346358 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-5stnc" Mar 13 14:11:41 crc kubenswrapper[4907]: I0313 14:11:41.351963 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-5stnc" Mar 13 14:12:00 crc kubenswrapper[4907]: I0313 14:12:00.186222 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556852-bg4rn"] Mar 13 14:12:00 crc kubenswrapper[4907]: E0313 14:12:00.187222 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Mar 13 14:12:00 crc kubenswrapper[4907]: I0313 14:12:00.187238 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Mar 13 14:12:00 crc kubenswrapper[4907]: I0313 14:12:00.187374 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Mar 13 14:12:00 crc kubenswrapper[4907]: I0313 14:12:00.187823 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556852-bg4rn" Mar 13 14:12:00 crc kubenswrapper[4907]: I0313 14:12:00.191143 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 14:12:00 crc kubenswrapper[4907]: I0313 14:12:00.191191 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 14:12:00 crc kubenswrapper[4907]: I0313 14:12:00.193534 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 14:12:00 crc kubenswrapper[4907]: I0313 14:12:00.207965 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556852-bg4rn"] Mar 13 14:12:00 crc kubenswrapper[4907]: I0313 14:12:00.290345 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fxcqc\" (UniqueName: \"kubernetes.io/projected/5cd3c3cd-9c71-4d90-a5cb-800de65b86de-kube-api-access-fxcqc\") pod \"auto-csr-approver-29556852-bg4rn\" (UID: \"5cd3c3cd-9c71-4d90-a5cb-800de65b86de\") " pod="openshift-infra/auto-csr-approver-29556852-bg4rn" Mar 13 14:12:00 crc kubenswrapper[4907]: I0313 14:12:00.391090 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fxcqc\" (UniqueName: \"kubernetes.io/projected/5cd3c3cd-9c71-4d90-a5cb-800de65b86de-kube-api-access-fxcqc\") pod \"auto-csr-approver-29556852-bg4rn\" (UID: \"5cd3c3cd-9c71-4d90-a5cb-800de65b86de\") " pod="openshift-infra/auto-csr-approver-29556852-bg4rn" Mar 13 14:12:00 crc kubenswrapper[4907]: I0313 14:12:00.413764 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fxcqc\" (UniqueName: \"kubernetes.io/projected/5cd3c3cd-9c71-4d90-a5cb-800de65b86de-kube-api-access-fxcqc\") pod \"auto-csr-approver-29556852-bg4rn\" (UID: \"5cd3c3cd-9c71-4d90-a5cb-800de65b86de\") " pod="openshift-infra/auto-csr-approver-29556852-bg4rn" Mar 13 14:12:00 crc kubenswrapper[4907]: I0313 14:12:00.508625 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556852-bg4rn" Mar 13 14:12:01 crc kubenswrapper[4907]: I0313 14:12:01.012543 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556852-bg4rn"] Mar 13 14:12:01 crc kubenswrapper[4907]: I0313 14:12:01.469120 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556852-bg4rn" event={"ID":"5cd3c3cd-9c71-4d90-a5cb-800de65b86de","Type":"ContainerStarted","Data":"4d8bca074dafc377113ee92ef77c5daac837754efb134951fc4f369ffa3d81f3"} Mar 13 14:12:03 crc kubenswrapper[4907]: I0313 14:12:03.486142 4907 generic.go:334] "Generic (PLEG): container finished" podID="5cd3c3cd-9c71-4d90-a5cb-800de65b86de" containerID="cbcf7b779b5c51c01226dc524be3c2eb43327ef749d0c0644c47a0b6f9140e3e" exitCode=0 Mar 13 14:12:03 crc kubenswrapper[4907]: I0313 14:12:03.486227 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556852-bg4rn" event={"ID":"5cd3c3cd-9c71-4d90-a5cb-800de65b86de","Type":"ContainerDied","Data":"cbcf7b779b5c51c01226dc524be3c2eb43327ef749d0c0644c47a0b6f9140e3e"} Mar 13 14:12:04 crc kubenswrapper[4907]: I0313 14:12:04.888539 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556852-bg4rn" Mar 13 14:12:04 crc kubenswrapper[4907]: I0313 14:12:04.970302 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fxcqc\" (UniqueName: \"kubernetes.io/projected/5cd3c3cd-9c71-4d90-a5cb-800de65b86de-kube-api-access-fxcqc\") pod \"5cd3c3cd-9c71-4d90-a5cb-800de65b86de\" (UID: \"5cd3c3cd-9c71-4d90-a5cb-800de65b86de\") " Mar 13 14:12:04 crc kubenswrapper[4907]: I0313 14:12:04.977847 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5cd3c3cd-9c71-4d90-a5cb-800de65b86de-kube-api-access-fxcqc" (OuterVolumeSpecName: "kube-api-access-fxcqc") pod "5cd3c3cd-9c71-4d90-a5cb-800de65b86de" (UID: "5cd3c3cd-9c71-4d90-a5cb-800de65b86de"). InnerVolumeSpecName "kube-api-access-fxcqc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:12:05 crc kubenswrapper[4907]: I0313 14:12:05.071742 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fxcqc\" (UniqueName: \"kubernetes.io/projected/5cd3c3cd-9c71-4d90-a5cb-800de65b86de-kube-api-access-fxcqc\") on node \"crc\" DevicePath \"\"" Mar 13 14:12:05 crc kubenswrapper[4907]: I0313 14:12:05.501909 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556852-bg4rn" event={"ID":"5cd3c3cd-9c71-4d90-a5cb-800de65b86de","Type":"ContainerDied","Data":"4d8bca074dafc377113ee92ef77c5daac837754efb134951fc4f369ffa3d81f3"} Mar 13 14:12:05 crc kubenswrapper[4907]: I0313 14:12:05.502363 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4d8bca074dafc377113ee92ef77c5daac837754efb134951fc4f369ffa3d81f3" Mar 13 14:12:05 crc kubenswrapper[4907]: I0313 14:12:05.501987 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556852-bg4rn" Mar 13 14:12:17 crc kubenswrapper[4907]: I0313 14:12:17.881213 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-m64n6"] Mar 13 14:12:17 crc kubenswrapper[4907]: E0313 14:12:17.882490 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cd3c3cd-9c71-4d90-a5cb-800de65b86de" containerName="oc" Mar 13 14:12:17 crc kubenswrapper[4907]: I0313 14:12:17.882518 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cd3c3cd-9c71-4d90-a5cb-800de65b86de" containerName="oc" Mar 13 14:12:17 crc kubenswrapper[4907]: I0313 14:12:17.882765 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="5cd3c3cd-9c71-4d90-a5cb-800de65b86de" containerName="oc" Mar 13 14:12:17 crc kubenswrapper[4907]: I0313 14:12:17.883552 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-m64n6" Mar 13 14:12:17 crc kubenswrapper[4907]: I0313 14:12:17.894192 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-m64n6"] Mar 13 14:12:18 crc kubenswrapper[4907]: I0313 14:12:18.004992 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-m64n6\" (UID: \"e905031a-4656-47cf-8f91-92feb42bd418\") " pod="openshift-image-registry/image-registry-66df7c8f76-m64n6" Mar 13 14:12:18 crc kubenswrapper[4907]: I0313 14:12:18.005067 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/e905031a-4656-47cf-8f91-92feb42bd418-registry-certificates\") pod \"image-registry-66df7c8f76-m64n6\" (UID: \"e905031a-4656-47cf-8f91-92feb42bd418\") " pod="openshift-image-registry/image-registry-66df7c8f76-m64n6" Mar 13 14:12:18 crc kubenswrapper[4907]: I0313 14:12:18.005139 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/e905031a-4656-47cf-8f91-92feb42bd418-installation-pull-secrets\") pod \"image-registry-66df7c8f76-m64n6\" (UID: \"e905031a-4656-47cf-8f91-92feb42bd418\") " pod="openshift-image-registry/image-registry-66df7c8f76-m64n6" Mar 13 14:12:18 crc kubenswrapper[4907]: I0313 14:12:18.005174 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z7mcj\" (UniqueName: \"kubernetes.io/projected/e905031a-4656-47cf-8f91-92feb42bd418-kube-api-access-z7mcj\") pod \"image-registry-66df7c8f76-m64n6\" (UID: \"e905031a-4656-47cf-8f91-92feb42bd418\") " pod="openshift-image-registry/image-registry-66df7c8f76-m64n6" Mar 13 14:12:18 crc kubenswrapper[4907]: I0313 14:12:18.005202 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/e905031a-4656-47cf-8f91-92feb42bd418-ca-trust-extracted\") pod \"image-registry-66df7c8f76-m64n6\" (UID: \"e905031a-4656-47cf-8f91-92feb42bd418\") " pod="openshift-image-registry/image-registry-66df7c8f76-m64n6" Mar 13 14:12:18 crc kubenswrapper[4907]: I0313 14:12:18.005242 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e905031a-4656-47cf-8f91-92feb42bd418-bound-sa-token\") pod \"image-registry-66df7c8f76-m64n6\" (UID: \"e905031a-4656-47cf-8f91-92feb42bd418\") " pod="openshift-image-registry/image-registry-66df7c8f76-m64n6" Mar 13 14:12:18 crc kubenswrapper[4907]: I0313 14:12:18.005274 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/e905031a-4656-47cf-8f91-92feb42bd418-registry-tls\") pod \"image-registry-66df7c8f76-m64n6\" (UID: \"e905031a-4656-47cf-8f91-92feb42bd418\") " pod="openshift-image-registry/image-registry-66df7c8f76-m64n6" Mar 13 14:12:18 crc kubenswrapper[4907]: I0313 14:12:18.005454 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e905031a-4656-47cf-8f91-92feb42bd418-trusted-ca\") pod \"image-registry-66df7c8f76-m64n6\" (UID: \"e905031a-4656-47cf-8f91-92feb42bd418\") " pod="openshift-image-registry/image-registry-66df7c8f76-m64n6" Mar 13 14:12:18 crc kubenswrapper[4907]: I0313 14:12:18.041699 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 14:12:18 crc kubenswrapper[4907]: I0313 14:12:18.041814 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 14:12:18 crc kubenswrapper[4907]: I0313 14:12:18.047133 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-m64n6\" (UID: \"e905031a-4656-47cf-8f91-92feb42bd418\") " pod="openshift-image-registry/image-registry-66df7c8f76-m64n6" Mar 13 14:12:18 crc kubenswrapper[4907]: I0313 14:12:18.106511 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/e905031a-4656-47cf-8f91-92feb42bd418-registry-certificates\") pod \"image-registry-66df7c8f76-m64n6\" (UID: \"e905031a-4656-47cf-8f91-92feb42bd418\") " pod="openshift-image-registry/image-registry-66df7c8f76-m64n6" Mar 13 14:12:18 crc kubenswrapper[4907]: I0313 14:12:18.107223 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/e905031a-4656-47cf-8f91-92feb42bd418-installation-pull-secrets\") pod \"image-registry-66df7c8f76-m64n6\" (UID: \"e905031a-4656-47cf-8f91-92feb42bd418\") " pod="openshift-image-registry/image-registry-66df7c8f76-m64n6" Mar 13 14:12:18 crc kubenswrapper[4907]: I0313 14:12:18.107436 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z7mcj\" (UniqueName: \"kubernetes.io/projected/e905031a-4656-47cf-8f91-92feb42bd418-kube-api-access-z7mcj\") pod \"image-registry-66df7c8f76-m64n6\" (UID: \"e905031a-4656-47cf-8f91-92feb42bd418\") " pod="openshift-image-registry/image-registry-66df7c8f76-m64n6" Mar 13 14:12:18 crc kubenswrapper[4907]: I0313 14:12:18.107666 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/e905031a-4656-47cf-8f91-92feb42bd418-ca-trust-extracted\") pod \"image-registry-66df7c8f76-m64n6\" (UID: \"e905031a-4656-47cf-8f91-92feb42bd418\") " pod="openshift-image-registry/image-registry-66df7c8f76-m64n6" Mar 13 14:12:18 crc kubenswrapper[4907]: I0313 14:12:18.107910 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e905031a-4656-47cf-8f91-92feb42bd418-bound-sa-token\") pod \"image-registry-66df7c8f76-m64n6\" (UID: \"e905031a-4656-47cf-8f91-92feb42bd418\") " pod="openshift-image-registry/image-registry-66df7c8f76-m64n6" Mar 13 14:12:18 crc kubenswrapper[4907]: I0313 14:12:18.108040 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/e905031a-4656-47cf-8f91-92feb42bd418-registry-tls\") pod \"image-registry-66df7c8f76-m64n6\" (UID: \"e905031a-4656-47cf-8f91-92feb42bd418\") " pod="openshift-image-registry/image-registry-66df7c8f76-m64n6" Mar 13 14:12:18 crc kubenswrapper[4907]: I0313 14:12:18.108333 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e905031a-4656-47cf-8f91-92feb42bd418-trusted-ca\") pod \"image-registry-66df7c8f76-m64n6\" (UID: \"e905031a-4656-47cf-8f91-92feb42bd418\") " pod="openshift-image-registry/image-registry-66df7c8f76-m64n6" Mar 13 14:12:18 crc kubenswrapper[4907]: I0313 14:12:18.110011 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/e905031a-4656-47cf-8f91-92feb42bd418-registry-certificates\") pod \"image-registry-66df7c8f76-m64n6\" (UID: \"e905031a-4656-47cf-8f91-92feb42bd418\") " pod="openshift-image-registry/image-registry-66df7c8f76-m64n6" Mar 13 14:12:18 crc kubenswrapper[4907]: I0313 14:12:18.110937 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/e905031a-4656-47cf-8f91-92feb42bd418-ca-trust-extracted\") pod \"image-registry-66df7c8f76-m64n6\" (UID: \"e905031a-4656-47cf-8f91-92feb42bd418\") " pod="openshift-image-registry/image-registry-66df7c8f76-m64n6" Mar 13 14:12:18 crc kubenswrapper[4907]: I0313 14:12:18.113911 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e905031a-4656-47cf-8f91-92feb42bd418-trusted-ca\") pod \"image-registry-66df7c8f76-m64n6\" (UID: \"e905031a-4656-47cf-8f91-92feb42bd418\") " pod="openshift-image-registry/image-registry-66df7c8f76-m64n6" Mar 13 14:12:18 crc kubenswrapper[4907]: I0313 14:12:18.115767 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/e905031a-4656-47cf-8f91-92feb42bd418-installation-pull-secrets\") pod \"image-registry-66df7c8f76-m64n6\" (UID: \"e905031a-4656-47cf-8f91-92feb42bd418\") " pod="openshift-image-registry/image-registry-66df7c8f76-m64n6" Mar 13 14:12:18 crc kubenswrapper[4907]: I0313 14:12:18.115979 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/e905031a-4656-47cf-8f91-92feb42bd418-registry-tls\") pod \"image-registry-66df7c8f76-m64n6\" (UID: \"e905031a-4656-47cf-8f91-92feb42bd418\") " pod="openshift-image-registry/image-registry-66df7c8f76-m64n6" Mar 13 14:12:18 crc kubenswrapper[4907]: I0313 14:12:18.127571 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e905031a-4656-47cf-8f91-92feb42bd418-bound-sa-token\") pod \"image-registry-66df7c8f76-m64n6\" (UID: \"e905031a-4656-47cf-8f91-92feb42bd418\") " pod="openshift-image-registry/image-registry-66df7c8f76-m64n6" Mar 13 14:12:18 crc kubenswrapper[4907]: I0313 14:12:18.130837 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z7mcj\" (UniqueName: \"kubernetes.io/projected/e905031a-4656-47cf-8f91-92feb42bd418-kube-api-access-z7mcj\") pod \"image-registry-66df7c8f76-m64n6\" (UID: \"e905031a-4656-47cf-8f91-92feb42bd418\") " pod="openshift-image-registry/image-registry-66df7c8f76-m64n6" Mar 13 14:12:18 crc kubenswrapper[4907]: I0313 14:12:18.210567 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-m64n6" Mar 13 14:12:18 crc kubenswrapper[4907]: I0313 14:12:18.751363 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-m64n6"] Mar 13 14:12:18 crc kubenswrapper[4907]: W0313 14:12:18.765324 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode905031a_4656_47cf_8f91_92feb42bd418.slice/crio-9e3ef090e9727e8123ae7bc6389b4e34aa9b15f035fbad5f010800f5620b937c WatchSource:0}: Error finding container 9e3ef090e9727e8123ae7bc6389b4e34aa9b15f035fbad5f010800f5620b937c: Status 404 returned error can't find the container with id 9e3ef090e9727e8123ae7bc6389b4e34aa9b15f035fbad5f010800f5620b937c Mar 13 14:12:19 crc kubenswrapper[4907]: I0313 14:12:19.606420 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-m64n6" event={"ID":"e905031a-4656-47cf-8f91-92feb42bd418","Type":"ContainerStarted","Data":"d2b2e692b861662142157ab27f24d6a4e1fd8452f651b0b9aff7165302235cf1"} Mar 13 14:12:19 crc kubenswrapper[4907]: I0313 14:12:19.606911 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-m64n6" event={"ID":"e905031a-4656-47cf-8f91-92feb42bd418","Type":"ContainerStarted","Data":"9e3ef090e9727e8123ae7bc6389b4e34aa9b15f035fbad5f010800f5620b937c"} Mar 13 14:12:19 crc kubenswrapper[4907]: I0313 14:12:19.607067 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-m64n6" Mar 13 14:12:19 crc kubenswrapper[4907]: I0313 14:12:19.637731 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-m64n6" podStartSLOduration=2.637699189 podStartE2EDuration="2.637699189s" podCreationTimestamp="2026-03-13 14:12:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:12:19.632326635 +0000 UTC m=+438.532114374" watchObservedRunningTime="2026-03-13 14:12:19.637699189 +0000 UTC m=+438.537486888" Mar 13 14:12:38 crc kubenswrapper[4907]: I0313 14:12:38.219324 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-m64n6" Mar 13 14:12:38 crc kubenswrapper[4907]: I0313 14:12:38.299414 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-q9z8k"] Mar 13 14:12:45 crc kubenswrapper[4907]: I0313 14:12:45.964003 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:12:45 crc kubenswrapper[4907]: I0313 14:12:45.964437 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:12:45 crc kubenswrapper[4907]: I0313 14:12:45.965381 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:12:45 crc kubenswrapper[4907]: I0313 14:12:45.972530 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:12:46 crc kubenswrapper[4907]: I0313 14:12:46.083692 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:12:46 crc kubenswrapper[4907]: W0313 14:12:46.552585 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fe485a1_e14f_4c09_b5b9_f252bc42b7e8.slice/crio-5d3e13e6ce85bd005c8351c1e2a74195a6cc191dcea89de87006f40de5e7faf1 WatchSource:0}: Error finding container 5d3e13e6ce85bd005c8351c1e2a74195a6cc191dcea89de87006f40de5e7faf1: Status 404 returned error can't find the container with id 5d3e13e6ce85bd005c8351c1e2a74195a6cc191dcea89de87006f40de5e7faf1 Mar 13 14:12:46 crc kubenswrapper[4907]: I0313 14:12:46.787905 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"eef0b64c634c7fd2961d0bde201f0af44757923cc03467a3eafaed052eaee559"} Mar 13 14:12:46 crc kubenswrapper[4907]: I0313 14:12:46.788184 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"5d3e13e6ce85bd005c8351c1e2a74195a6cc191dcea89de87006f40de5e7faf1"} Mar 13 14:12:46 crc kubenswrapper[4907]: I0313 14:12:46.976461 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:12:46 crc kubenswrapper[4907]: I0313 14:12:46.976536 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:12:46 crc kubenswrapper[4907]: I0313 14:12:46.983709 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:12:46 crc kubenswrapper[4907]: I0313 14:12:46.983724 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:12:47 crc kubenswrapper[4907]: I0313 14:12:47.083012 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:12:47 crc kubenswrapper[4907]: I0313 14:12:47.182966 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:12:47 crc kubenswrapper[4907]: W0313 14:12:47.622922 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d751cbb_f2e2_430d_9754_c882a5e924a5.slice/crio-b589f6199d516fdd4b4f70c3b9a2d237feaf198d2736eb5b0f1edaee3f88ecbc WatchSource:0}: Error finding container b589f6199d516fdd4b4f70c3b9a2d237feaf198d2736eb5b0f1edaee3f88ecbc: Status 404 returned error can't find the container with id b589f6199d516fdd4b4f70c3b9a2d237feaf198d2736eb5b0f1edaee3f88ecbc Mar 13 14:12:47 crc kubenswrapper[4907]: I0313 14:12:47.794805 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"2c35eeee99bdfd18cc00cf2324a49b157be30fef45a8552698e3dd13c84db49b"} Mar 13 14:12:47 crc kubenswrapper[4907]: I0313 14:12:47.794870 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"b589f6199d516fdd4b4f70c3b9a2d237feaf198d2736eb5b0f1edaee3f88ecbc"} Mar 13 14:12:47 crc kubenswrapper[4907]: I0313 14:12:47.800975 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"7f17450bbf9f2a07a348d0fc0308410472e1ad570e624a35de5fcbe6939b6ba2"} Mar 13 14:12:47 crc kubenswrapper[4907]: I0313 14:12:47.801037 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"9ba1178efe0855315d30bc9b7bd8d2497870c71f7a0bf4c277765b33fd0a3dd8"} Mar 13 14:12:47 crc kubenswrapper[4907]: I0313 14:12:47.801476 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:12:48 crc kubenswrapper[4907]: I0313 14:12:48.041291 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 14:12:48 crc kubenswrapper[4907]: I0313 14:12:48.041396 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 14:12:53 crc kubenswrapper[4907]: I0313 14:12:53.526725 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-42kpk"] Mar 13 14:12:53 crc kubenswrapper[4907]: I0313 14:12:53.527750 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-42kpk" podUID="8adb3c1c-bacd-4cca-9796-7ca96624e9f6" containerName="registry-server" containerID="cri-o://779541283cf16b48730f8ef9601537920e362fa4a7dbb572a3600beba4418986" gracePeriod=30 Mar 13 14:12:53 crc kubenswrapper[4907]: I0313 14:12:53.549658 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-v6msv"] Mar 13 14:12:53 crc kubenswrapper[4907]: I0313 14:12:53.550257 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-v6msv" podUID="84c2ada5-ceab-4327-802e-9ae459ac814d" containerName="registry-server" containerID="cri-o://6e642c222cc08b2da3f2c537efb7c366f76e02de754a7e9abbd7e89b0a44893f" gracePeriod=30 Mar 13 14:12:53 crc kubenswrapper[4907]: I0313 14:12:53.556185 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-5stnc"] Mar 13 14:12:53 crc kubenswrapper[4907]: I0313 14:12:53.556581 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-5stnc" podUID="0290eb6f-0be7-4340-8ba8-ed6ea0662c33" containerName="marketplace-operator" containerID="cri-o://a4c4a9be5faa2f17ce6e81edeed8ac29998640aca39360db9beeef84d8b0752a" gracePeriod=30 Mar 13 14:12:53 crc kubenswrapper[4907]: I0313 14:12:53.566057 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-jk5nw"] Mar 13 14:12:53 crc kubenswrapper[4907]: I0313 14:12:53.566326 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-jk5nw" podUID="d342a6a2-43dd-4d15-8aff-f60f9a3383bf" containerName="registry-server" containerID="cri-o://de346bdbb6259a737c213b9565b9465613c835283a744df647429b69400db2ec" gracePeriod=30 Mar 13 14:12:53 crc kubenswrapper[4907]: I0313 14:12:53.583613 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-h7vlt"] Mar 13 14:12:53 crc kubenswrapper[4907]: I0313 14:12:53.586422 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-h7vlt" Mar 13 14:12:53 crc kubenswrapper[4907]: I0313 14:12:53.595353 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-fmzmc"] Mar 13 14:12:53 crc kubenswrapper[4907]: I0313 14:12:53.595918 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-fmzmc" podUID="ec3466c2-4954-459d-88d8-e8c6cbf309e8" containerName="registry-server" containerID="cri-o://c60d7cfe210d5d163b6c716dca40da31b6a520923cab2dc2f6be072d3c2f2049" gracePeriod=30 Mar 13 14:12:53 crc kubenswrapper[4907]: I0313 14:12:53.600610 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-h7vlt"] Mar 13 14:12:53 crc kubenswrapper[4907]: I0313 14:12:53.685082 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/05325ebe-cad6-417a-b1fb-af1efdf24b76-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-h7vlt\" (UID: \"05325ebe-cad6-417a-b1fb-af1efdf24b76\") " pod="openshift-marketplace/marketplace-operator-79b997595-h7vlt" Mar 13 14:12:53 crc kubenswrapper[4907]: I0313 14:12:53.685133 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/05325ebe-cad6-417a-b1fb-af1efdf24b76-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-h7vlt\" (UID: \"05325ebe-cad6-417a-b1fb-af1efdf24b76\") " pod="openshift-marketplace/marketplace-operator-79b997595-h7vlt" Mar 13 14:12:53 crc kubenswrapper[4907]: I0313 14:12:53.685163 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-266nc\" (UniqueName: \"kubernetes.io/projected/05325ebe-cad6-417a-b1fb-af1efdf24b76-kube-api-access-266nc\") pod \"marketplace-operator-79b997595-h7vlt\" (UID: \"05325ebe-cad6-417a-b1fb-af1efdf24b76\") " pod="openshift-marketplace/marketplace-operator-79b997595-h7vlt" Mar 13 14:12:53 crc kubenswrapper[4907]: I0313 14:12:53.786019 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/05325ebe-cad6-417a-b1fb-af1efdf24b76-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-h7vlt\" (UID: \"05325ebe-cad6-417a-b1fb-af1efdf24b76\") " pod="openshift-marketplace/marketplace-operator-79b997595-h7vlt" Mar 13 14:12:53 crc kubenswrapper[4907]: I0313 14:12:53.786115 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/05325ebe-cad6-417a-b1fb-af1efdf24b76-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-h7vlt\" (UID: \"05325ebe-cad6-417a-b1fb-af1efdf24b76\") " pod="openshift-marketplace/marketplace-operator-79b997595-h7vlt" Mar 13 14:12:53 crc kubenswrapper[4907]: I0313 14:12:53.786145 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-266nc\" (UniqueName: \"kubernetes.io/projected/05325ebe-cad6-417a-b1fb-af1efdf24b76-kube-api-access-266nc\") pod \"marketplace-operator-79b997595-h7vlt\" (UID: \"05325ebe-cad6-417a-b1fb-af1efdf24b76\") " pod="openshift-marketplace/marketplace-operator-79b997595-h7vlt" Mar 13 14:12:53 crc kubenswrapper[4907]: I0313 14:12:53.788340 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/05325ebe-cad6-417a-b1fb-af1efdf24b76-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-h7vlt\" (UID: \"05325ebe-cad6-417a-b1fb-af1efdf24b76\") " pod="openshift-marketplace/marketplace-operator-79b997595-h7vlt" Mar 13 14:12:53 crc kubenswrapper[4907]: I0313 14:12:53.799938 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/05325ebe-cad6-417a-b1fb-af1efdf24b76-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-h7vlt\" (UID: \"05325ebe-cad6-417a-b1fb-af1efdf24b76\") " pod="openshift-marketplace/marketplace-operator-79b997595-h7vlt" Mar 13 14:12:53 crc kubenswrapper[4907]: I0313 14:12:53.820581 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-266nc\" (UniqueName: \"kubernetes.io/projected/05325ebe-cad6-417a-b1fb-af1efdf24b76-kube-api-access-266nc\") pod \"marketplace-operator-79b997595-h7vlt\" (UID: \"05325ebe-cad6-417a-b1fb-af1efdf24b76\") " pod="openshift-marketplace/marketplace-operator-79b997595-h7vlt" Mar 13 14:12:53 crc kubenswrapper[4907]: I0313 14:12:53.859117 4907 generic.go:334] "Generic (PLEG): container finished" podID="8adb3c1c-bacd-4cca-9796-7ca96624e9f6" containerID="779541283cf16b48730f8ef9601537920e362fa4a7dbb572a3600beba4418986" exitCode=0 Mar 13 14:12:53 crc kubenswrapper[4907]: I0313 14:12:53.859186 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-42kpk" event={"ID":"8adb3c1c-bacd-4cca-9796-7ca96624e9f6","Type":"ContainerDied","Data":"779541283cf16b48730f8ef9601537920e362fa4a7dbb572a3600beba4418986"} Mar 13 14:12:53 crc kubenswrapper[4907]: I0313 14:12:53.863895 4907 generic.go:334] "Generic (PLEG): container finished" podID="0290eb6f-0be7-4340-8ba8-ed6ea0662c33" containerID="a4c4a9be5faa2f17ce6e81edeed8ac29998640aca39360db9beeef84d8b0752a" exitCode=0 Mar 13 14:12:53 crc kubenswrapper[4907]: I0313 14:12:53.864046 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-5stnc" event={"ID":"0290eb6f-0be7-4340-8ba8-ed6ea0662c33","Type":"ContainerDied","Data":"a4c4a9be5faa2f17ce6e81edeed8ac29998640aca39360db9beeef84d8b0752a"} Mar 13 14:12:53 crc kubenswrapper[4907]: I0313 14:12:53.864119 4907 scope.go:117] "RemoveContainer" containerID="48afa92be95670a8e5bf5c9bca4dd6ed4ea5b0ffd5c088a0a66004cb2137ee6b" Mar 13 14:12:53 crc kubenswrapper[4907]: I0313 14:12:53.867366 4907 generic.go:334] "Generic (PLEG): container finished" podID="d342a6a2-43dd-4d15-8aff-f60f9a3383bf" containerID="de346bdbb6259a737c213b9565b9465613c835283a744df647429b69400db2ec" exitCode=0 Mar 13 14:12:53 crc kubenswrapper[4907]: I0313 14:12:53.867415 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jk5nw" event={"ID":"d342a6a2-43dd-4d15-8aff-f60f9a3383bf","Type":"ContainerDied","Data":"de346bdbb6259a737c213b9565b9465613c835283a744df647429b69400db2ec"} Mar 13 14:12:53 crc kubenswrapper[4907]: I0313 14:12:53.870986 4907 generic.go:334] "Generic (PLEG): container finished" podID="ec3466c2-4954-459d-88d8-e8c6cbf309e8" containerID="c60d7cfe210d5d163b6c716dca40da31b6a520923cab2dc2f6be072d3c2f2049" exitCode=0 Mar 13 14:12:53 crc kubenswrapper[4907]: I0313 14:12:53.871133 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fmzmc" event={"ID":"ec3466c2-4954-459d-88d8-e8c6cbf309e8","Type":"ContainerDied","Data":"c60d7cfe210d5d163b6c716dca40da31b6a520923cab2dc2f6be072d3c2f2049"} Mar 13 14:12:53 crc kubenswrapper[4907]: I0313 14:12:53.873207 4907 generic.go:334] "Generic (PLEG): container finished" podID="84c2ada5-ceab-4327-802e-9ae459ac814d" containerID="6e642c222cc08b2da3f2c537efb7c366f76e02de754a7e9abbd7e89b0a44893f" exitCode=0 Mar 13 14:12:53 crc kubenswrapper[4907]: I0313 14:12:53.873237 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v6msv" event={"ID":"84c2ada5-ceab-4327-802e-9ae459ac814d","Type":"ContainerDied","Data":"6e642c222cc08b2da3f2c537efb7c366f76e02de754a7e9abbd7e89b0a44893f"} Mar 13 14:12:53 crc kubenswrapper[4907]: I0313 14:12:53.924777 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-h7vlt" Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.036679 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-v6msv" Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.090620 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k6s9c\" (UniqueName: \"kubernetes.io/projected/84c2ada5-ceab-4327-802e-9ae459ac814d-kube-api-access-k6s9c\") pod \"84c2ada5-ceab-4327-802e-9ae459ac814d\" (UID: \"84c2ada5-ceab-4327-802e-9ae459ac814d\") " Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.090672 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/84c2ada5-ceab-4327-802e-9ae459ac814d-utilities\") pod \"84c2ada5-ceab-4327-802e-9ae459ac814d\" (UID: \"84c2ada5-ceab-4327-802e-9ae459ac814d\") " Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.090750 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/84c2ada5-ceab-4327-802e-9ae459ac814d-catalog-content\") pod \"84c2ada5-ceab-4327-802e-9ae459ac814d\" (UID: \"84c2ada5-ceab-4327-802e-9ae459ac814d\") " Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.092006 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/84c2ada5-ceab-4327-802e-9ae459ac814d-utilities" (OuterVolumeSpecName: "utilities") pod "84c2ada5-ceab-4327-802e-9ae459ac814d" (UID: "84c2ada5-ceab-4327-802e-9ae459ac814d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.098443 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/84c2ada5-ceab-4327-802e-9ae459ac814d-kube-api-access-k6s9c" (OuterVolumeSpecName: "kube-api-access-k6s9c") pod "84c2ada5-ceab-4327-802e-9ae459ac814d" (UID: "84c2ada5-ceab-4327-802e-9ae459ac814d"). InnerVolumeSpecName "kube-api-access-k6s9c". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.108834 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jk5nw" Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.116833 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-5stnc" Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.117558 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fmzmc" Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.158645 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/84c2ada5-ceab-4327-802e-9ae459ac814d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "84c2ada5-ceab-4327-802e-9ae459ac814d" (UID: "84c2ada5-ceab-4327-802e-9ae459ac814d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.164640 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-42kpk" Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.192151 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/0290eb6f-0be7-4340-8ba8-ed6ea0662c33-marketplace-operator-metrics\") pod \"0290eb6f-0be7-4340-8ba8-ed6ea0662c33\" (UID: \"0290eb6f-0be7-4340-8ba8-ed6ea0662c33\") " Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.192208 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qppp6\" (UniqueName: \"kubernetes.io/projected/0290eb6f-0be7-4340-8ba8-ed6ea0662c33-kube-api-access-qppp6\") pod \"0290eb6f-0be7-4340-8ba8-ed6ea0662c33\" (UID: \"0290eb6f-0be7-4340-8ba8-ed6ea0662c33\") " Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.192327 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ec3466c2-4954-459d-88d8-e8c6cbf309e8-utilities\") pod \"ec3466c2-4954-459d-88d8-e8c6cbf309e8\" (UID: \"ec3466c2-4954-459d-88d8-e8c6cbf309e8\") " Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.192374 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-48ddb\" (UniqueName: \"kubernetes.io/projected/ec3466c2-4954-459d-88d8-e8c6cbf309e8-kube-api-access-48ddb\") pod \"ec3466c2-4954-459d-88d8-e8c6cbf309e8\" (UID: \"ec3466c2-4954-459d-88d8-e8c6cbf309e8\") " Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.192399 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8adb3c1c-bacd-4cca-9796-7ca96624e9f6-utilities\") pod \"8adb3c1c-bacd-4cca-9796-7ca96624e9f6\" (UID: \"8adb3c1c-bacd-4cca-9796-7ca96624e9f6\") " Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.192424 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d342a6a2-43dd-4d15-8aff-f60f9a3383bf-catalog-content\") pod \"d342a6a2-43dd-4d15-8aff-f60f9a3383bf\" (UID: \"d342a6a2-43dd-4d15-8aff-f60f9a3383bf\") " Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.192442 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d342a6a2-43dd-4d15-8aff-f60f9a3383bf-utilities\") pod \"d342a6a2-43dd-4d15-8aff-f60f9a3383bf\" (UID: \"d342a6a2-43dd-4d15-8aff-f60f9a3383bf\") " Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.192463 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8adb3c1c-bacd-4cca-9796-7ca96624e9f6-catalog-content\") pod \"8adb3c1c-bacd-4cca-9796-7ca96624e9f6\" (UID: \"8adb3c1c-bacd-4cca-9796-7ca96624e9f6\") " Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.192482 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2x9qs\" (UniqueName: \"kubernetes.io/projected/8adb3c1c-bacd-4cca-9796-7ca96624e9f6-kube-api-access-2x9qs\") pod \"8adb3c1c-bacd-4cca-9796-7ca96624e9f6\" (UID: \"8adb3c1c-bacd-4cca-9796-7ca96624e9f6\") " Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.192498 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-msd8t\" (UniqueName: \"kubernetes.io/projected/d342a6a2-43dd-4d15-8aff-f60f9a3383bf-kube-api-access-msd8t\") pod \"d342a6a2-43dd-4d15-8aff-f60f9a3383bf\" (UID: \"d342a6a2-43dd-4d15-8aff-f60f9a3383bf\") " Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.192520 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0290eb6f-0be7-4340-8ba8-ed6ea0662c33-marketplace-trusted-ca\") pod \"0290eb6f-0be7-4340-8ba8-ed6ea0662c33\" (UID: \"0290eb6f-0be7-4340-8ba8-ed6ea0662c33\") " Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.192538 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ec3466c2-4954-459d-88d8-e8c6cbf309e8-catalog-content\") pod \"ec3466c2-4954-459d-88d8-e8c6cbf309e8\" (UID: \"ec3466c2-4954-459d-88d8-e8c6cbf309e8\") " Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.192802 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/84c2ada5-ceab-4327-802e-9ae459ac814d-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.192815 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/84c2ada5-ceab-4327-802e-9ae459ac814d-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.192826 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k6s9c\" (UniqueName: \"kubernetes.io/projected/84c2ada5-ceab-4327-802e-9ae459ac814d-kube-api-access-k6s9c\") on node \"crc\" DevicePath \"\"" Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.196044 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d342a6a2-43dd-4d15-8aff-f60f9a3383bf-utilities" (OuterVolumeSpecName: "utilities") pod "d342a6a2-43dd-4d15-8aff-f60f9a3383bf" (UID: "d342a6a2-43dd-4d15-8aff-f60f9a3383bf"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.196436 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0290eb6f-0be7-4340-8ba8-ed6ea0662c33-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "0290eb6f-0be7-4340-8ba8-ed6ea0662c33" (UID: "0290eb6f-0be7-4340-8ba8-ed6ea0662c33"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.197137 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ec3466c2-4954-459d-88d8-e8c6cbf309e8-utilities" (OuterVolumeSpecName: "utilities") pod "ec3466c2-4954-459d-88d8-e8c6cbf309e8" (UID: "ec3466c2-4954-459d-88d8-e8c6cbf309e8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.197363 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8adb3c1c-bacd-4cca-9796-7ca96624e9f6-utilities" (OuterVolumeSpecName: "utilities") pod "8adb3c1c-bacd-4cca-9796-7ca96624e9f6" (UID: "8adb3c1c-bacd-4cca-9796-7ca96624e9f6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.200800 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ec3466c2-4954-459d-88d8-e8c6cbf309e8-kube-api-access-48ddb" (OuterVolumeSpecName: "kube-api-access-48ddb") pod "ec3466c2-4954-459d-88d8-e8c6cbf309e8" (UID: "ec3466c2-4954-459d-88d8-e8c6cbf309e8"). InnerVolumeSpecName "kube-api-access-48ddb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.201333 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0290eb6f-0be7-4340-8ba8-ed6ea0662c33-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "0290eb6f-0be7-4340-8ba8-ed6ea0662c33" (UID: "0290eb6f-0be7-4340-8ba8-ed6ea0662c33"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.202024 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0290eb6f-0be7-4340-8ba8-ed6ea0662c33-kube-api-access-qppp6" (OuterVolumeSpecName: "kube-api-access-qppp6") pod "0290eb6f-0be7-4340-8ba8-ed6ea0662c33" (UID: "0290eb6f-0be7-4340-8ba8-ed6ea0662c33"). InnerVolumeSpecName "kube-api-access-qppp6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.202570 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d342a6a2-43dd-4d15-8aff-f60f9a3383bf-kube-api-access-msd8t" (OuterVolumeSpecName: "kube-api-access-msd8t") pod "d342a6a2-43dd-4d15-8aff-f60f9a3383bf" (UID: "d342a6a2-43dd-4d15-8aff-f60f9a3383bf"). InnerVolumeSpecName "kube-api-access-msd8t". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.203782 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8adb3c1c-bacd-4cca-9796-7ca96624e9f6-kube-api-access-2x9qs" (OuterVolumeSpecName: "kube-api-access-2x9qs") pod "8adb3c1c-bacd-4cca-9796-7ca96624e9f6" (UID: "8adb3c1c-bacd-4cca-9796-7ca96624e9f6"). InnerVolumeSpecName "kube-api-access-2x9qs". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.227498 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d342a6a2-43dd-4d15-8aff-f60f9a3383bf-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d342a6a2-43dd-4d15-8aff-f60f9a3383bf" (UID: "d342a6a2-43dd-4d15-8aff-f60f9a3383bf"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.262651 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8adb3c1c-bacd-4cca-9796-7ca96624e9f6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8adb3c1c-bacd-4cca-9796-7ca96624e9f6" (UID: "8adb3c1c-bacd-4cca-9796-7ca96624e9f6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.293542 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ec3466c2-4954-459d-88d8-e8c6cbf309e8-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.293566 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-48ddb\" (UniqueName: \"kubernetes.io/projected/ec3466c2-4954-459d-88d8-e8c6cbf309e8-kube-api-access-48ddb\") on node \"crc\" DevicePath \"\"" Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.293579 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8adb3c1c-bacd-4cca-9796-7ca96624e9f6-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.293587 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d342a6a2-43dd-4d15-8aff-f60f9a3383bf-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.293595 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d342a6a2-43dd-4d15-8aff-f60f9a3383bf-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.293603 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8adb3c1c-bacd-4cca-9796-7ca96624e9f6-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.293611 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2x9qs\" (UniqueName: \"kubernetes.io/projected/8adb3c1c-bacd-4cca-9796-7ca96624e9f6-kube-api-access-2x9qs\") on node \"crc\" DevicePath \"\"" Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.293619 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-msd8t\" (UniqueName: \"kubernetes.io/projected/d342a6a2-43dd-4d15-8aff-f60f9a3383bf-kube-api-access-msd8t\") on node \"crc\" DevicePath \"\"" Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.293628 4907 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0290eb6f-0be7-4340-8ba8-ed6ea0662c33-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.293636 4907 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/0290eb6f-0be7-4340-8ba8-ed6ea0662c33-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.293648 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qppp6\" (UniqueName: \"kubernetes.io/projected/0290eb6f-0be7-4340-8ba8-ed6ea0662c33-kube-api-access-qppp6\") on node \"crc\" DevicePath \"\"" Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.317020 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ec3466c2-4954-459d-88d8-e8c6cbf309e8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ec3466c2-4954-459d-88d8-e8c6cbf309e8" (UID: "ec3466c2-4954-459d-88d8-e8c6cbf309e8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.395672 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ec3466c2-4954-459d-88d8-e8c6cbf309e8-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.553254 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-h7vlt"] Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.881777 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-42kpk" event={"ID":"8adb3c1c-bacd-4cca-9796-7ca96624e9f6","Type":"ContainerDied","Data":"04064aa9c533eaef085c99222c5a674a4c851c29b8b6bd8d0141f59c380540da"} Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.881826 4907 scope.go:117] "RemoveContainer" containerID="779541283cf16b48730f8ef9601537920e362fa4a7dbb572a3600beba4418986" Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.881952 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-42kpk" Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.890495 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-5stnc" event={"ID":"0290eb6f-0be7-4340-8ba8-ed6ea0662c33","Type":"ContainerDied","Data":"97e42fd7291025df6972a91466dcaf06ac63f386d1c5751baba2fa6aca2cf00a"} Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.890592 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-5stnc" Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.897758 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jk5nw" event={"ID":"d342a6a2-43dd-4d15-8aff-f60f9a3383bf","Type":"ContainerDied","Data":"4fd25f47bfaf1894861177f77553de05bdcc490dd081b002f769378e51f3d5fb"} Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.897799 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jk5nw" Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.901532 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fmzmc" event={"ID":"ec3466c2-4954-459d-88d8-e8c6cbf309e8","Type":"ContainerDied","Data":"33655dcb12d3167aaf6db2e72d760da63675955f0f25e1e18f759d3dc24a92c2"} Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.901599 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fmzmc" Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.909957 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v6msv" event={"ID":"84c2ada5-ceab-4327-802e-9ae459ac814d","Type":"ContainerDied","Data":"2b528cfb1a9dfb9a8a4cae32fcb0401d81595581c8b3e1b33eacedab76ae1cd0"} Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.909982 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-v6msv" Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.911372 4907 scope.go:117] "RemoveContainer" containerID="059ebb72320a2144f692825d22d2d54a1e276a63cc41c87a83327001d7f8b539" Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.914698 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-h7vlt" event={"ID":"05325ebe-cad6-417a-b1fb-af1efdf24b76","Type":"ContainerStarted","Data":"0a42b4b4dddcb9d47eabfa57af67fd59f71354ad57a5dd141a5e496ca7bcc386"} Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.915023 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-h7vlt" event={"ID":"05325ebe-cad6-417a-b1fb-af1efdf24b76","Type":"ContainerStarted","Data":"6c4482ad907f1cc1aba645f7a3ee94d3b6cf5739b1adec2a4a8c45af04f764e2"} Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.915903 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-h7vlt" Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.917375 4907 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-h7vlt container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.72:8080/healthz\": dial tcp 10.217.0.72:8080: connect: connection refused" start-of-body= Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.917411 4907 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-h7vlt" podUID="05325ebe-cad6-417a-b1fb-af1efdf24b76" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.72:8080/healthz\": dial tcp 10.217.0.72:8080: connect: connection refused" Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.926965 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-42kpk"] Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.932732 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-42kpk"] Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.951816 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-5stnc"] Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.955338 4907 scope.go:117] "RemoveContainer" containerID="0d444f1a14d20e1e6a2efc71bfbe0eb77305492df11ed2559d59993943816ad1" Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.961159 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-5stnc"] Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.972036 4907 scope.go:117] "RemoveContainer" containerID="a4c4a9be5faa2f17ce6e81edeed8ac29998640aca39360db9beeef84d8b0752a" Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.975953 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-h7vlt" podStartSLOduration=1.975937745 podStartE2EDuration="1.975937745s" podCreationTimestamp="2026-03-13 14:12:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:12:54.975036741 +0000 UTC m=+473.874824430" watchObservedRunningTime="2026-03-13 14:12:54.975937745 +0000 UTC m=+473.875725444" Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.997167 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-jk5nw"] Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.998757 4907 scope.go:117] "RemoveContainer" containerID="de346bdbb6259a737c213b9565b9465613c835283a744df647429b69400db2ec" Mar 13 14:12:55 crc kubenswrapper[4907]: I0313 14:12:55.002068 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-jk5nw"] Mar 13 14:12:55 crc kubenswrapper[4907]: I0313 14:12:55.012740 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-v6msv"] Mar 13 14:12:55 crc kubenswrapper[4907]: I0313 14:12:55.014052 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-v6msv"] Mar 13 14:12:55 crc kubenswrapper[4907]: I0313 14:12:55.020523 4907 scope.go:117] "RemoveContainer" containerID="2ba2c849db65019cfc185d5e8fb816c1ea6e14d84b987e50fe979ce8ace8b5fb" Mar 13 14:12:55 crc kubenswrapper[4907]: I0313 14:12:55.026479 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-fmzmc"] Mar 13 14:12:55 crc kubenswrapper[4907]: I0313 14:12:55.029638 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-fmzmc"] Mar 13 14:12:55 crc kubenswrapper[4907]: I0313 14:12:55.034234 4907 scope.go:117] "RemoveContainer" containerID="167f39bf875dea575d6960ea51e3427665c28d2ee610e574bf41709836edc603" Mar 13 14:12:55 crc kubenswrapper[4907]: I0313 14:12:55.055229 4907 scope.go:117] "RemoveContainer" containerID="c60d7cfe210d5d163b6c716dca40da31b6a520923cab2dc2f6be072d3c2f2049" Mar 13 14:12:55 crc kubenswrapper[4907]: I0313 14:12:55.087252 4907 scope.go:117] "RemoveContainer" containerID="a06107e49962bc7b9a8f0ee9719c73d09f58fc4ae0da4f8a46e5df6c62712945" Mar 13 14:12:55 crc kubenswrapper[4907]: I0313 14:12:55.112237 4907 scope.go:117] "RemoveContainer" containerID="5129da37c4baac30c1ed93e43558cb65af057e72717f5cffd7b1409299226e40" Mar 13 14:12:55 crc kubenswrapper[4907]: I0313 14:12:55.135235 4907 scope.go:117] "RemoveContainer" containerID="6e642c222cc08b2da3f2c537efb7c366f76e02de754a7e9abbd7e89b0a44893f" Mar 13 14:12:55 crc kubenswrapper[4907]: I0313 14:12:55.160048 4907 scope.go:117] "RemoveContainer" containerID="701b8cbaf2465addef85f7946e5e3a30ad0261ab612755a656d7086a119dc2d2" Mar 13 14:12:55 crc kubenswrapper[4907]: I0313 14:12:55.177895 4907 scope.go:117] "RemoveContainer" containerID="0bd7973d57fde27ab0e7adf532d86c90333b4abd96119d9c054a7fb89cbb6dd6" Mar 13 14:12:55 crc kubenswrapper[4907]: I0313 14:12:55.790911 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0290eb6f-0be7-4340-8ba8-ed6ea0662c33" path="/var/lib/kubelet/pods/0290eb6f-0be7-4340-8ba8-ed6ea0662c33/volumes" Mar 13 14:12:55 crc kubenswrapper[4907]: I0313 14:12:55.791571 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="84c2ada5-ceab-4327-802e-9ae459ac814d" path="/var/lib/kubelet/pods/84c2ada5-ceab-4327-802e-9ae459ac814d/volumes" Mar 13 14:12:55 crc kubenswrapper[4907]: I0313 14:12:55.792405 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8adb3c1c-bacd-4cca-9796-7ca96624e9f6" path="/var/lib/kubelet/pods/8adb3c1c-bacd-4cca-9796-7ca96624e9f6/volumes" Mar 13 14:12:55 crc kubenswrapper[4907]: I0313 14:12:55.793753 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d342a6a2-43dd-4d15-8aff-f60f9a3383bf" path="/var/lib/kubelet/pods/d342a6a2-43dd-4d15-8aff-f60f9a3383bf/volumes" Mar 13 14:12:55 crc kubenswrapper[4907]: I0313 14:12:55.794531 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ec3466c2-4954-459d-88d8-e8c6cbf309e8" path="/var/lib/kubelet/pods/ec3466c2-4954-459d-88d8-e8c6cbf309e8/volumes" Mar 13 14:12:55 crc kubenswrapper[4907]: I0313 14:12:55.934652 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-h7vlt" Mar 13 14:12:56 crc kubenswrapper[4907]: I0313 14:12:56.548978 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-8dzhn"] Mar 13 14:12:56 crc kubenswrapper[4907]: E0313 14:12:56.549176 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0290eb6f-0be7-4340-8ba8-ed6ea0662c33" containerName="marketplace-operator" Mar 13 14:12:56 crc kubenswrapper[4907]: I0313 14:12:56.549189 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="0290eb6f-0be7-4340-8ba8-ed6ea0662c33" containerName="marketplace-operator" Mar 13 14:12:56 crc kubenswrapper[4907]: E0313 14:12:56.549200 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d342a6a2-43dd-4d15-8aff-f60f9a3383bf" containerName="registry-server" Mar 13 14:12:56 crc kubenswrapper[4907]: I0313 14:12:56.549206 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="d342a6a2-43dd-4d15-8aff-f60f9a3383bf" containerName="registry-server" Mar 13 14:12:56 crc kubenswrapper[4907]: E0313 14:12:56.549215 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84c2ada5-ceab-4327-802e-9ae459ac814d" containerName="extract-content" Mar 13 14:12:56 crc kubenswrapper[4907]: I0313 14:12:56.549220 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="84c2ada5-ceab-4327-802e-9ae459ac814d" containerName="extract-content" Mar 13 14:12:56 crc kubenswrapper[4907]: E0313 14:12:56.549229 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8adb3c1c-bacd-4cca-9796-7ca96624e9f6" containerName="registry-server" Mar 13 14:12:56 crc kubenswrapper[4907]: I0313 14:12:56.549235 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="8adb3c1c-bacd-4cca-9796-7ca96624e9f6" containerName="registry-server" Mar 13 14:12:56 crc kubenswrapper[4907]: E0313 14:12:56.549243 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec3466c2-4954-459d-88d8-e8c6cbf309e8" containerName="extract-content" Mar 13 14:12:56 crc kubenswrapper[4907]: I0313 14:12:56.549248 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec3466c2-4954-459d-88d8-e8c6cbf309e8" containerName="extract-content" Mar 13 14:12:56 crc kubenswrapper[4907]: E0313 14:12:56.549256 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0290eb6f-0be7-4340-8ba8-ed6ea0662c33" containerName="marketplace-operator" Mar 13 14:12:56 crc kubenswrapper[4907]: I0313 14:12:56.549262 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="0290eb6f-0be7-4340-8ba8-ed6ea0662c33" containerName="marketplace-operator" Mar 13 14:12:56 crc kubenswrapper[4907]: E0313 14:12:56.549271 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84c2ada5-ceab-4327-802e-9ae459ac814d" containerName="extract-utilities" Mar 13 14:12:56 crc kubenswrapper[4907]: I0313 14:12:56.549277 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="84c2ada5-ceab-4327-802e-9ae459ac814d" containerName="extract-utilities" Mar 13 14:12:56 crc kubenswrapper[4907]: E0313 14:12:56.549288 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8adb3c1c-bacd-4cca-9796-7ca96624e9f6" containerName="extract-content" Mar 13 14:12:56 crc kubenswrapper[4907]: I0313 14:12:56.549294 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="8adb3c1c-bacd-4cca-9796-7ca96624e9f6" containerName="extract-content" Mar 13 14:12:56 crc kubenswrapper[4907]: E0313 14:12:56.549316 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d342a6a2-43dd-4d15-8aff-f60f9a3383bf" containerName="extract-content" Mar 13 14:12:56 crc kubenswrapper[4907]: I0313 14:12:56.549322 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="d342a6a2-43dd-4d15-8aff-f60f9a3383bf" containerName="extract-content" Mar 13 14:12:56 crc kubenswrapper[4907]: E0313 14:12:56.549329 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec3466c2-4954-459d-88d8-e8c6cbf309e8" containerName="registry-server" Mar 13 14:12:56 crc kubenswrapper[4907]: I0313 14:12:56.549335 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec3466c2-4954-459d-88d8-e8c6cbf309e8" containerName="registry-server" Mar 13 14:12:56 crc kubenswrapper[4907]: E0313 14:12:56.549342 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d342a6a2-43dd-4d15-8aff-f60f9a3383bf" containerName="extract-utilities" Mar 13 14:12:56 crc kubenswrapper[4907]: I0313 14:12:56.549348 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="d342a6a2-43dd-4d15-8aff-f60f9a3383bf" containerName="extract-utilities" Mar 13 14:12:56 crc kubenswrapper[4907]: E0313 14:12:56.549358 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec3466c2-4954-459d-88d8-e8c6cbf309e8" containerName="extract-utilities" Mar 13 14:12:56 crc kubenswrapper[4907]: I0313 14:12:56.549363 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec3466c2-4954-459d-88d8-e8c6cbf309e8" containerName="extract-utilities" Mar 13 14:12:56 crc kubenswrapper[4907]: E0313 14:12:56.549372 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84c2ada5-ceab-4327-802e-9ae459ac814d" containerName="registry-server" Mar 13 14:12:56 crc kubenswrapper[4907]: I0313 14:12:56.549378 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="84c2ada5-ceab-4327-802e-9ae459ac814d" containerName="registry-server" Mar 13 14:12:56 crc kubenswrapper[4907]: E0313 14:12:56.549385 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8adb3c1c-bacd-4cca-9796-7ca96624e9f6" containerName="extract-utilities" Mar 13 14:12:56 crc kubenswrapper[4907]: I0313 14:12:56.549391 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="8adb3c1c-bacd-4cca-9796-7ca96624e9f6" containerName="extract-utilities" Mar 13 14:12:56 crc kubenswrapper[4907]: I0313 14:12:56.549469 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="84c2ada5-ceab-4327-802e-9ae459ac814d" containerName="registry-server" Mar 13 14:12:56 crc kubenswrapper[4907]: I0313 14:12:56.549479 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="d342a6a2-43dd-4d15-8aff-f60f9a3383bf" containerName="registry-server" Mar 13 14:12:56 crc kubenswrapper[4907]: I0313 14:12:56.549493 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="ec3466c2-4954-459d-88d8-e8c6cbf309e8" containerName="registry-server" Mar 13 14:12:56 crc kubenswrapper[4907]: I0313 14:12:56.549501 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="8adb3c1c-bacd-4cca-9796-7ca96624e9f6" containerName="registry-server" Mar 13 14:12:56 crc kubenswrapper[4907]: I0313 14:12:56.549510 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="0290eb6f-0be7-4340-8ba8-ed6ea0662c33" containerName="marketplace-operator" Mar 13 14:12:56 crc kubenswrapper[4907]: I0313 14:12:56.549517 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="0290eb6f-0be7-4340-8ba8-ed6ea0662c33" containerName="marketplace-operator" Mar 13 14:12:56 crc kubenswrapper[4907]: I0313 14:12:56.550234 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8dzhn" Mar 13 14:12:56 crc kubenswrapper[4907]: I0313 14:12:56.554124 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Mar 13 14:12:56 crc kubenswrapper[4907]: I0313 14:12:56.570054 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8dzhn"] Mar 13 14:12:56 crc kubenswrapper[4907]: I0313 14:12:56.726275 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6l4kd\" (UniqueName: \"kubernetes.io/projected/a4b992e7-af9d-44a5-91ea-4942ad99bce2-kube-api-access-6l4kd\") pod \"redhat-marketplace-8dzhn\" (UID: \"a4b992e7-af9d-44a5-91ea-4942ad99bce2\") " pod="openshift-marketplace/redhat-marketplace-8dzhn" Mar 13 14:12:56 crc kubenswrapper[4907]: I0313 14:12:56.726437 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4b992e7-af9d-44a5-91ea-4942ad99bce2-utilities\") pod \"redhat-marketplace-8dzhn\" (UID: \"a4b992e7-af9d-44a5-91ea-4942ad99bce2\") " pod="openshift-marketplace/redhat-marketplace-8dzhn" Mar 13 14:12:56 crc kubenswrapper[4907]: I0313 14:12:56.726477 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4b992e7-af9d-44a5-91ea-4942ad99bce2-catalog-content\") pod \"redhat-marketplace-8dzhn\" (UID: \"a4b992e7-af9d-44a5-91ea-4942ad99bce2\") " pod="openshift-marketplace/redhat-marketplace-8dzhn" Mar 13 14:12:56 crc kubenswrapper[4907]: I0313 14:12:56.829100 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4b992e7-af9d-44a5-91ea-4942ad99bce2-utilities\") pod \"redhat-marketplace-8dzhn\" (UID: \"a4b992e7-af9d-44a5-91ea-4942ad99bce2\") " pod="openshift-marketplace/redhat-marketplace-8dzhn" Mar 13 14:12:56 crc kubenswrapper[4907]: I0313 14:12:56.829170 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4b992e7-af9d-44a5-91ea-4942ad99bce2-catalog-content\") pod \"redhat-marketplace-8dzhn\" (UID: \"a4b992e7-af9d-44a5-91ea-4942ad99bce2\") " pod="openshift-marketplace/redhat-marketplace-8dzhn" Mar 13 14:12:56 crc kubenswrapper[4907]: I0313 14:12:56.829250 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6l4kd\" (UniqueName: \"kubernetes.io/projected/a4b992e7-af9d-44a5-91ea-4942ad99bce2-kube-api-access-6l4kd\") pod \"redhat-marketplace-8dzhn\" (UID: \"a4b992e7-af9d-44a5-91ea-4942ad99bce2\") " pod="openshift-marketplace/redhat-marketplace-8dzhn" Mar 13 14:12:56 crc kubenswrapper[4907]: I0313 14:12:56.829634 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4b992e7-af9d-44a5-91ea-4942ad99bce2-utilities\") pod \"redhat-marketplace-8dzhn\" (UID: \"a4b992e7-af9d-44a5-91ea-4942ad99bce2\") " pod="openshift-marketplace/redhat-marketplace-8dzhn" Mar 13 14:12:56 crc kubenswrapper[4907]: I0313 14:12:56.829874 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4b992e7-af9d-44a5-91ea-4942ad99bce2-catalog-content\") pod \"redhat-marketplace-8dzhn\" (UID: \"a4b992e7-af9d-44a5-91ea-4942ad99bce2\") " pod="openshift-marketplace/redhat-marketplace-8dzhn" Mar 13 14:12:56 crc kubenswrapper[4907]: I0313 14:12:56.851420 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6l4kd\" (UniqueName: \"kubernetes.io/projected/a4b992e7-af9d-44a5-91ea-4942ad99bce2-kube-api-access-6l4kd\") pod \"redhat-marketplace-8dzhn\" (UID: \"a4b992e7-af9d-44a5-91ea-4942ad99bce2\") " pod="openshift-marketplace/redhat-marketplace-8dzhn" Mar 13 14:12:56 crc kubenswrapper[4907]: I0313 14:12:56.884958 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8dzhn" Mar 13 14:12:57 crc kubenswrapper[4907]: I0313 14:12:57.145045 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-z7sd2"] Mar 13 14:12:57 crc kubenswrapper[4907]: I0313 14:12:57.146834 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-z7sd2" Mar 13 14:12:57 crc kubenswrapper[4907]: I0313 14:12:57.164228 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Mar 13 14:12:57 crc kubenswrapper[4907]: I0313 14:12:57.166839 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-z7sd2"] Mar 13 14:12:57 crc kubenswrapper[4907]: I0313 14:12:57.335968 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9d354400-9ba4-45b0-a866-09da4921210a-utilities\") pod \"redhat-operators-z7sd2\" (UID: \"9d354400-9ba4-45b0-a866-09da4921210a\") " pod="openshift-marketplace/redhat-operators-z7sd2" Mar 13 14:12:57 crc kubenswrapper[4907]: I0313 14:12:57.336066 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z568x\" (UniqueName: \"kubernetes.io/projected/9d354400-9ba4-45b0-a866-09da4921210a-kube-api-access-z568x\") pod \"redhat-operators-z7sd2\" (UID: \"9d354400-9ba4-45b0-a866-09da4921210a\") " pod="openshift-marketplace/redhat-operators-z7sd2" Mar 13 14:12:57 crc kubenswrapper[4907]: I0313 14:12:57.336112 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9d354400-9ba4-45b0-a866-09da4921210a-catalog-content\") pod \"redhat-operators-z7sd2\" (UID: \"9d354400-9ba4-45b0-a866-09da4921210a\") " pod="openshift-marketplace/redhat-operators-z7sd2" Mar 13 14:12:57 crc kubenswrapper[4907]: I0313 14:12:57.355406 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8dzhn"] Mar 13 14:12:57 crc kubenswrapper[4907]: W0313 14:12:57.360648 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda4b992e7_af9d_44a5_91ea_4942ad99bce2.slice/crio-d3f829163785e2f0fbd5b9a1c153ca084ff17e459a2df81a1f77d5c94887b538 WatchSource:0}: Error finding container d3f829163785e2f0fbd5b9a1c153ca084ff17e459a2df81a1f77d5c94887b538: Status 404 returned error can't find the container with id d3f829163785e2f0fbd5b9a1c153ca084ff17e459a2df81a1f77d5c94887b538 Mar 13 14:12:57 crc kubenswrapper[4907]: I0313 14:12:57.437597 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9d354400-9ba4-45b0-a866-09da4921210a-utilities\") pod \"redhat-operators-z7sd2\" (UID: \"9d354400-9ba4-45b0-a866-09da4921210a\") " pod="openshift-marketplace/redhat-operators-z7sd2" Mar 13 14:12:57 crc kubenswrapper[4907]: I0313 14:12:57.437666 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z568x\" (UniqueName: \"kubernetes.io/projected/9d354400-9ba4-45b0-a866-09da4921210a-kube-api-access-z568x\") pod \"redhat-operators-z7sd2\" (UID: \"9d354400-9ba4-45b0-a866-09da4921210a\") " pod="openshift-marketplace/redhat-operators-z7sd2" Mar 13 14:12:57 crc kubenswrapper[4907]: I0313 14:12:57.437697 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9d354400-9ba4-45b0-a866-09da4921210a-catalog-content\") pod \"redhat-operators-z7sd2\" (UID: \"9d354400-9ba4-45b0-a866-09da4921210a\") " pod="openshift-marketplace/redhat-operators-z7sd2" Mar 13 14:12:57 crc kubenswrapper[4907]: I0313 14:12:57.438302 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9d354400-9ba4-45b0-a866-09da4921210a-catalog-content\") pod \"redhat-operators-z7sd2\" (UID: \"9d354400-9ba4-45b0-a866-09da4921210a\") " pod="openshift-marketplace/redhat-operators-z7sd2" Mar 13 14:12:57 crc kubenswrapper[4907]: I0313 14:12:57.438302 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9d354400-9ba4-45b0-a866-09da4921210a-utilities\") pod \"redhat-operators-z7sd2\" (UID: \"9d354400-9ba4-45b0-a866-09da4921210a\") " pod="openshift-marketplace/redhat-operators-z7sd2" Mar 13 14:12:57 crc kubenswrapper[4907]: I0313 14:12:57.467629 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z568x\" (UniqueName: \"kubernetes.io/projected/9d354400-9ba4-45b0-a866-09da4921210a-kube-api-access-z568x\") pod \"redhat-operators-z7sd2\" (UID: \"9d354400-9ba4-45b0-a866-09da4921210a\") " pod="openshift-marketplace/redhat-operators-z7sd2" Mar 13 14:12:57 crc kubenswrapper[4907]: I0313 14:12:57.490985 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-z7sd2" Mar 13 14:12:57 crc kubenswrapper[4907]: I0313 14:12:57.959458 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-z7sd2"] Mar 13 14:12:57 crc kubenswrapper[4907]: I0313 14:12:57.968235 4907 generic.go:334] "Generic (PLEG): container finished" podID="a4b992e7-af9d-44a5-91ea-4942ad99bce2" containerID="8ff120885864b840470bc227f13746fd7a689b3d184728b2e61e8f3a9ba2cdee" exitCode=0 Mar 13 14:12:57 crc kubenswrapper[4907]: I0313 14:12:57.968327 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8dzhn" event={"ID":"a4b992e7-af9d-44a5-91ea-4942ad99bce2","Type":"ContainerDied","Data":"8ff120885864b840470bc227f13746fd7a689b3d184728b2e61e8f3a9ba2cdee"} Mar 13 14:12:57 crc kubenswrapper[4907]: I0313 14:12:57.968383 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8dzhn" event={"ID":"a4b992e7-af9d-44a5-91ea-4942ad99bce2","Type":"ContainerStarted","Data":"d3f829163785e2f0fbd5b9a1c153ca084ff17e459a2df81a1f77d5c94887b538"} Mar 13 14:12:58 crc kubenswrapper[4907]: I0313 14:12:58.953916 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-bkqkq"] Mar 13 14:12:58 crc kubenswrapper[4907]: I0313 14:12:58.958251 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bkqkq" Mar 13 14:12:58 crc kubenswrapper[4907]: I0313 14:12:58.962381 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Mar 13 14:12:58 crc kubenswrapper[4907]: I0313 14:12:58.963643 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-bkqkq"] Mar 13 14:12:58 crc kubenswrapper[4907]: I0313 14:12:58.985555 4907 generic.go:334] "Generic (PLEG): container finished" podID="9d354400-9ba4-45b0-a866-09da4921210a" containerID="d6c37d7b57c4e69c9a3f6b656d4baa52460de34948395e9868d67ec94be94892" exitCode=0 Mar 13 14:12:58 crc kubenswrapper[4907]: I0313 14:12:58.985623 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z7sd2" event={"ID":"9d354400-9ba4-45b0-a866-09da4921210a","Type":"ContainerDied","Data":"d6c37d7b57c4e69c9a3f6b656d4baa52460de34948395e9868d67ec94be94892"} Mar 13 14:12:58 crc kubenswrapper[4907]: I0313 14:12:58.985648 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z7sd2" event={"ID":"9d354400-9ba4-45b0-a866-09da4921210a","Type":"ContainerStarted","Data":"47ed7182a9338cb13165cfe12b9a8fb553637da7ef787dcade4cffd32ec34556"} Mar 13 14:12:58 crc kubenswrapper[4907]: I0313 14:12:58.989988 4907 generic.go:334] "Generic (PLEG): container finished" podID="a4b992e7-af9d-44a5-91ea-4942ad99bce2" containerID="547879c2a6adc0d798d067b0e8539fbf29b569018429785d08c1af1bcb78f58c" exitCode=0 Mar 13 14:12:58 crc kubenswrapper[4907]: I0313 14:12:58.990045 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8dzhn" event={"ID":"a4b992e7-af9d-44a5-91ea-4942ad99bce2","Type":"ContainerDied","Data":"547879c2a6adc0d798d067b0e8539fbf29b569018429785d08c1af1bcb78f58c"} Mar 13 14:12:59 crc kubenswrapper[4907]: I0313 14:12:59.059184 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f55641f-96d1-4abe-b29e-35c03423c86f-catalog-content\") pod \"community-operators-bkqkq\" (UID: \"8f55641f-96d1-4abe-b29e-35c03423c86f\") " pod="openshift-marketplace/community-operators-bkqkq" Mar 13 14:12:59 crc kubenswrapper[4907]: I0313 14:12:59.059251 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-blsqh\" (UniqueName: \"kubernetes.io/projected/8f55641f-96d1-4abe-b29e-35c03423c86f-kube-api-access-blsqh\") pod \"community-operators-bkqkq\" (UID: \"8f55641f-96d1-4abe-b29e-35c03423c86f\") " pod="openshift-marketplace/community-operators-bkqkq" Mar 13 14:12:59 crc kubenswrapper[4907]: I0313 14:12:59.059277 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f55641f-96d1-4abe-b29e-35c03423c86f-utilities\") pod \"community-operators-bkqkq\" (UID: \"8f55641f-96d1-4abe-b29e-35c03423c86f\") " pod="openshift-marketplace/community-operators-bkqkq" Mar 13 14:12:59 crc kubenswrapper[4907]: I0313 14:12:59.160967 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f55641f-96d1-4abe-b29e-35c03423c86f-catalog-content\") pod \"community-operators-bkqkq\" (UID: \"8f55641f-96d1-4abe-b29e-35c03423c86f\") " pod="openshift-marketplace/community-operators-bkqkq" Mar 13 14:12:59 crc kubenswrapper[4907]: I0313 14:12:59.161091 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-blsqh\" (UniqueName: \"kubernetes.io/projected/8f55641f-96d1-4abe-b29e-35c03423c86f-kube-api-access-blsqh\") pod \"community-operators-bkqkq\" (UID: \"8f55641f-96d1-4abe-b29e-35c03423c86f\") " pod="openshift-marketplace/community-operators-bkqkq" Mar 13 14:12:59 crc kubenswrapper[4907]: I0313 14:12:59.161128 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f55641f-96d1-4abe-b29e-35c03423c86f-utilities\") pod \"community-operators-bkqkq\" (UID: \"8f55641f-96d1-4abe-b29e-35c03423c86f\") " pod="openshift-marketplace/community-operators-bkqkq" Mar 13 14:12:59 crc kubenswrapper[4907]: I0313 14:12:59.161758 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f55641f-96d1-4abe-b29e-35c03423c86f-catalog-content\") pod \"community-operators-bkqkq\" (UID: \"8f55641f-96d1-4abe-b29e-35c03423c86f\") " pod="openshift-marketplace/community-operators-bkqkq" Mar 13 14:12:59 crc kubenswrapper[4907]: I0313 14:12:59.161843 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f55641f-96d1-4abe-b29e-35c03423c86f-utilities\") pod \"community-operators-bkqkq\" (UID: \"8f55641f-96d1-4abe-b29e-35c03423c86f\") " pod="openshift-marketplace/community-operators-bkqkq" Mar 13 14:12:59 crc kubenswrapper[4907]: I0313 14:12:59.183328 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-blsqh\" (UniqueName: \"kubernetes.io/projected/8f55641f-96d1-4abe-b29e-35c03423c86f-kube-api-access-blsqh\") pod \"community-operators-bkqkq\" (UID: \"8f55641f-96d1-4abe-b29e-35c03423c86f\") " pod="openshift-marketplace/community-operators-bkqkq" Mar 13 14:12:59 crc kubenswrapper[4907]: I0313 14:12:59.284115 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bkqkq" Mar 13 14:12:59 crc kubenswrapper[4907]: I0313 14:12:59.489107 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-bkqkq"] Mar 13 14:12:59 crc kubenswrapper[4907]: W0313 14:12:59.499098 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8f55641f_96d1_4abe_b29e_35c03423c86f.slice/crio-51741992329c7cbf452d56e06ea4b68fc6d3bd2347af300a47210ccc94035b08 WatchSource:0}: Error finding container 51741992329c7cbf452d56e06ea4b68fc6d3bd2347af300a47210ccc94035b08: Status 404 returned error can't find the container with id 51741992329c7cbf452d56e06ea4b68fc6d3bd2347af300a47210ccc94035b08 Mar 13 14:12:59 crc kubenswrapper[4907]: I0313 14:12:59.548596 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-slm6d"] Mar 13 14:12:59 crc kubenswrapper[4907]: I0313 14:12:59.552651 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-slm6d" Mar 13 14:12:59 crc kubenswrapper[4907]: I0313 14:12:59.554817 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Mar 13 14:12:59 crc kubenswrapper[4907]: I0313 14:12:59.556660 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-slm6d"] Mar 13 14:12:59 crc kubenswrapper[4907]: I0313 14:12:59.670659 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa7e1b0d-049f-4775-bd6b-d7430126fbd3-utilities\") pod \"certified-operators-slm6d\" (UID: \"aa7e1b0d-049f-4775-bd6b-d7430126fbd3\") " pod="openshift-marketplace/certified-operators-slm6d" Mar 13 14:12:59 crc kubenswrapper[4907]: I0313 14:12:59.670705 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4w459\" (UniqueName: \"kubernetes.io/projected/aa7e1b0d-049f-4775-bd6b-d7430126fbd3-kube-api-access-4w459\") pod \"certified-operators-slm6d\" (UID: \"aa7e1b0d-049f-4775-bd6b-d7430126fbd3\") " pod="openshift-marketplace/certified-operators-slm6d" Mar 13 14:12:59 crc kubenswrapper[4907]: I0313 14:12:59.670731 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa7e1b0d-049f-4775-bd6b-d7430126fbd3-catalog-content\") pod \"certified-operators-slm6d\" (UID: \"aa7e1b0d-049f-4775-bd6b-d7430126fbd3\") " pod="openshift-marketplace/certified-operators-slm6d" Mar 13 14:12:59 crc kubenswrapper[4907]: I0313 14:12:59.772449 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa7e1b0d-049f-4775-bd6b-d7430126fbd3-utilities\") pod \"certified-operators-slm6d\" (UID: \"aa7e1b0d-049f-4775-bd6b-d7430126fbd3\") " pod="openshift-marketplace/certified-operators-slm6d" Mar 13 14:12:59 crc kubenswrapper[4907]: I0313 14:12:59.772671 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4w459\" (UniqueName: \"kubernetes.io/projected/aa7e1b0d-049f-4775-bd6b-d7430126fbd3-kube-api-access-4w459\") pod \"certified-operators-slm6d\" (UID: \"aa7e1b0d-049f-4775-bd6b-d7430126fbd3\") " pod="openshift-marketplace/certified-operators-slm6d" Mar 13 14:12:59 crc kubenswrapper[4907]: I0313 14:12:59.772753 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa7e1b0d-049f-4775-bd6b-d7430126fbd3-catalog-content\") pod \"certified-operators-slm6d\" (UID: \"aa7e1b0d-049f-4775-bd6b-d7430126fbd3\") " pod="openshift-marketplace/certified-operators-slm6d" Mar 13 14:12:59 crc kubenswrapper[4907]: I0313 14:12:59.773035 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa7e1b0d-049f-4775-bd6b-d7430126fbd3-utilities\") pod \"certified-operators-slm6d\" (UID: \"aa7e1b0d-049f-4775-bd6b-d7430126fbd3\") " pod="openshift-marketplace/certified-operators-slm6d" Mar 13 14:12:59 crc kubenswrapper[4907]: I0313 14:12:59.773305 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa7e1b0d-049f-4775-bd6b-d7430126fbd3-catalog-content\") pod \"certified-operators-slm6d\" (UID: \"aa7e1b0d-049f-4775-bd6b-d7430126fbd3\") " pod="openshift-marketplace/certified-operators-slm6d" Mar 13 14:12:59 crc kubenswrapper[4907]: I0313 14:12:59.790616 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4w459\" (UniqueName: \"kubernetes.io/projected/aa7e1b0d-049f-4775-bd6b-d7430126fbd3-kube-api-access-4w459\") pod \"certified-operators-slm6d\" (UID: \"aa7e1b0d-049f-4775-bd6b-d7430126fbd3\") " pod="openshift-marketplace/certified-operators-slm6d" Mar 13 14:12:59 crc kubenswrapper[4907]: I0313 14:12:59.977993 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-slm6d" Mar 13 14:12:59 crc kubenswrapper[4907]: I0313 14:12:59.997011 4907 generic.go:334] "Generic (PLEG): container finished" podID="8f55641f-96d1-4abe-b29e-35c03423c86f" containerID="4a29b590b6d8237b3616d05e54dcdd32085b2501383cd2a01ceaf05da4c45946" exitCode=0 Mar 13 14:12:59 crc kubenswrapper[4907]: I0313 14:12:59.997087 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bkqkq" event={"ID":"8f55641f-96d1-4abe-b29e-35c03423c86f","Type":"ContainerDied","Data":"4a29b590b6d8237b3616d05e54dcdd32085b2501383cd2a01ceaf05da4c45946"} Mar 13 14:12:59 crc kubenswrapper[4907]: I0313 14:12:59.997138 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bkqkq" event={"ID":"8f55641f-96d1-4abe-b29e-35c03423c86f","Type":"ContainerStarted","Data":"51741992329c7cbf452d56e06ea4b68fc6d3bd2347af300a47210ccc94035b08"} Mar 13 14:13:00 crc kubenswrapper[4907]: I0313 14:13:00.000065 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8dzhn" event={"ID":"a4b992e7-af9d-44a5-91ea-4942ad99bce2","Type":"ContainerStarted","Data":"61dc0c95d112b04f5718295d9fa9342b536564f7e473174dfc22f62bfc51b010"} Mar 13 14:13:00 crc kubenswrapper[4907]: I0313 14:13:00.004397 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z7sd2" event={"ID":"9d354400-9ba4-45b0-a866-09da4921210a","Type":"ContainerStarted","Data":"dab525c88cc920e8d4d1cece348d2b1fdef07a1d9676ecd45d6453cc3380f24a"} Mar 13 14:13:00 crc kubenswrapper[4907]: I0313 14:13:00.047153 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-8dzhn" podStartSLOduration=2.533633002 podStartE2EDuration="4.047134278s" podCreationTimestamp="2026-03-13 14:12:56 +0000 UTC" firstStartedPulling="2026-03-13 14:12:57.980303097 +0000 UTC m=+476.880090786" lastFinishedPulling="2026-03-13 14:12:59.493804363 +0000 UTC m=+478.393592062" observedRunningTime="2026-03-13 14:13:00.046425709 +0000 UTC m=+478.946213398" watchObservedRunningTime="2026-03-13 14:13:00.047134278 +0000 UTC m=+478.946921967" Mar 13 14:13:00 crc kubenswrapper[4907]: I0313 14:13:00.412655 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-slm6d"] Mar 13 14:13:01 crc kubenswrapper[4907]: I0313 14:13:01.009115 4907 generic.go:334] "Generic (PLEG): container finished" podID="aa7e1b0d-049f-4775-bd6b-d7430126fbd3" containerID="94cfc15a8b67ee390217c4c4f7a6e9d139f4f9130b30d47ce32fe764058f9702" exitCode=0 Mar 13 14:13:01 crc kubenswrapper[4907]: I0313 14:13:01.009209 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-slm6d" event={"ID":"aa7e1b0d-049f-4775-bd6b-d7430126fbd3","Type":"ContainerDied","Data":"94cfc15a8b67ee390217c4c4f7a6e9d139f4f9130b30d47ce32fe764058f9702"} Mar 13 14:13:01 crc kubenswrapper[4907]: I0313 14:13:01.009507 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-slm6d" event={"ID":"aa7e1b0d-049f-4775-bd6b-d7430126fbd3","Type":"ContainerStarted","Data":"bb680d736d304ce74d4c50a4ab2ec6d4e3e931217becaf9f28120e12c315dcc1"} Mar 13 14:13:01 crc kubenswrapper[4907]: I0313 14:13:01.011320 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bkqkq" event={"ID":"8f55641f-96d1-4abe-b29e-35c03423c86f","Type":"ContainerStarted","Data":"7fb98f669ca95b8966acee50a69bfe3ca16749a7b12337396bf39c2c672c9654"} Mar 13 14:13:01 crc kubenswrapper[4907]: I0313 14:13:01.013214 4907 generic.go:334] "Generic (PLEG): container finished" podID="9d354400-9ba4-45b0-a866-09da4921210a" containerID="dab525c88cc920e8d4d1cece348d2b1fdef07a1d9676ecd45d6453cc3380f24a" exitCode=0 Mar 13 14:13:01 crc kubenswrapper[4907]: I0313 14:13:01.014115 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z7sd2" event={"ID":"9d354400-9ba4-45b0-a866-09da4921210a","Type":"ContainerDied","Data":"dab525c88cc920e8d4d1cece348d2b1fdef07a1d9676ecd45d6453cc3380f24a"} Mar 13 14:13:02 crc kubenswrapper[4907]: I0313 14:13:02.041919 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z7sd2" event={"ID":"9d354400-9ba4-45b0-a866-09da4921210a","Type":"ContainerStarted","Data":"5f9ae23bba94014c66ed4e0accae3b53de94f902d021871cc264533c7fb6c724"} Mar 13 14:13:02 crc kubenswrapper[4907]: I0313 14:13:02.049060 4907 generic.go:334] "Generic (PLEG): container finished" podID="8f55641f-96d1-4abe-b29e-35c03423c86f" containerID="7fb98f669ca95b8966acee50a69bfe3ca16749a7b12337396bf39c2c672c9654" exitCode=0 Mar 13 14:13:02 crc kubenswrapper[4907]: I0313 14:13:02.049132 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bkqkq" event={"ID":"8f55641f-96d1-4abe-b29e-35c03423c86f","Type":"ContainerDied","Data":"7fb98f669ca95b8966acee50a69bfe3ca16749a7b12337396bf39c2c672c9654"} Mar 13 14:13:02 crc kubenswrapper[4907]: I0313 14:13:02.069187 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-z7sd2" podStartSLOduration=2.58717776 podStartE2EDuration="5.06916975s" podCreationTimestamp="2026-03-13 14:12:57 +0000 UTC" firstStartedPulling="2026-03-13 14:12:58.988480678 +0000 UTC m=+477.888268367" lastFinishedPulling="2026-03-13 14:13:01.470472658 +0000 UTC m=+480.370260357" observedRunningTime="2026-03-13 14:13:02.069068076 +0000 UTC m=+480.968855795" watchObservedRunningTime="2026-03-13 14:13:02.06916975 +0000 UTC m=+480.968957449" Mar 13 14:13:03 crc kubenswrapper[4907]: I0313 14:13:03.057726 4907 generic.go:334] "Generic (PLEG): container finished" podID="aa7e1b0d-049f-4775-bd6b-d7430126fbd3" containerID="bb225b43511b66781eff07b7c196782d39311083eeb0149c3f9bb46c2bce2771" exitCode=0 Mar 13 14:13:03 crc kubenswrapper[4907]: I0313 14:13:03.057786 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-slm6d" event={"ID":"aa7e1b0d-049f-4775-bd6b-d7430126fbd3","Type":"ContainerDied","Data":"bb225b43511b66781eff07b7c196782d39311083eeb0149c3f9bb46c2bce2771"} Mar 13 14:13:03 crc kubenswrapper[4907]: I0313 14:13:03.061065 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bkqkq" event={"ID":"8f55641f-96d1-4abe-b29e-35c03423c86f","Type":"ContainerStarted","Data":"dd30544066f21dab70ce804b909280851ba1f1bd4d4cb32479861077957a9178"} Mar 13 14:13:03 crc kubenswrapper[4907]: I0313 14:13:03.092316 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-bkqkq" podStartSLOduration=2.612751367 podStartE2EDuration="5.092297588s" podCreationTimestamp="2026-03-13 14:12:58 +0000 UTC" firstStartedPulling="2026-03-13 14:12:59.998771559 +0000 UTC m=+478.898559248" lastFinishedPulling="2026-03-13 14:13:02.47831775 +0000 UTC m=+481.378105469" observedRunningTime="2026-03-13 14:13:03.090461857 +0000 UTC m=+481.990249546" watchObservedRunningTime="2026-03-13 14:13:03.092297588 +0000 UTC m=+481.992085277" Mar 13 14:13:03 crc kubenswrapper[4907]: I0313 14:13:03.371566 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" podUID="9c5d91bb-753e-4554-b7a1-6c85fe323875" containerName="registry" containerID="cri-o://e2452e8d8fe737b72f44a8aba2c006a14d53bdfda4e4a01534d104020e7dd4aa" gracePeriod=30 Mar 13 14:13:03 crc kubenswrapper[4907]: I0313 14:13:03.789507 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:13:03 crc kubenswrapper[4907]: I0313 14:13:03.927203 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"9c5d91bb-753e-4554-b7a1-6c85fe323875\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " Mar 13 14:13:03 crc kubenswrapper[4907]: I0313 14:13:03.927288 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jb2r5\" (UniqueName: \"kubernetes.io/projected/9c5d91bb-753e-4554-b7a1-6c85fe323875-kube-api-access-jb2r5\") pod \"9c5d91bb-753e-4554-b7a1-6c85fe323875\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " Mar 13 14:13:03 crc kubenswrapper[4907]: I0313 14:13:03.927314 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9c5d91bb-753e-4554-b7a1-6c85fe323875-trusted-ca\") pod \"9c5d91bb-753e-4554-b7a1-6c85fe323875\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " Mar 13 14:13:03 crc kubenswrapper[4907]: I0313 14:13:03.927355 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9c5d91bb-753e-4554-b7a1-6c85fe323875-bound-sa-token\") pod \"9c5d91bb-753e-4554-b7a1-6c85fe323875\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " Mar 13 14:13:03 crc kubenswrapper[4907]: I0313 14:13:03.927392 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/9c5d91bb-753e-4554-b7a1-6c85fe323875-registry-certificates\") pod \"9c5d91bb-753e-4554-b7a1-6c85fe323875\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " Mar 13 14:13:03 crc kubenswrapper[4907]: I0313 14:13:03.927423 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/9c5d91bb-753e-4554-b7a1-6c85fe323875-ca-trust-extracted\") pod \"9c5d91bb-753e-4554-b7a1-6c85fe323875\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " Mar 13 14:13:03 crc kubenswrapper[4907]: I0313 14:13:03.927444 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/9c5d91bb-753e-4554-b7a1-6c85fe323875-installation-pull-secrets\") pod \"9c5d91bb-753e-4554-b7a1-6c85fe323875\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " Mar 13 14:13:03 crc kubenswrapper[4907]: I0313 14:13:03.927490 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/9c5d91bb-753e-4554-b7a1-6c85fe323875-registry-tls\") pod \"9c5d91bb-753e-4554-b7a1-6c85fe323875\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " Mar 13 14:13:03 crc kubenswrapper[4907]: I0313 14:13:03.928463 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9c5d91bb-753e-4554-b7a1-6c85fe323875-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9c5d91bb-753e-4554-b7a1-6c85fe323875" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:13:03 crc kubenswrapper[4907]: I0313 14:13:03.928583 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9c5d91bb-753e-4554-b7a1-6c85fe323875-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "9c5d91bb-753e-4554-b7a1-6c85fe323875" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:13:03 crc kubenswrapper[4907]: I0313 14:13:03.932992 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c5d91bb-753e-4554-b7a1-6c85fe323875-kube-api-access-jb2r5" (OuterVolumeSpecName: "kube-api-access-jb2r5") pod "9c5d91bb-753e-4554-b7a1-6c85fe323875" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875"). InnerVolumeSpecName "kube-api-access-jb2r5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:13:03 crc kubenswrapper[4907]: I0313 14:13:03.933337 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c5d91bb-753e-4554-b7a1-6c85fe323875-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "9c5d91bb-753e-4554-b7a1-6c85fe323875" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:13:03 crc kubenswrapper[4907]: I0313 14:13:03.933668 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c5d91bb-753e-4554-b7a1-6c85fe323875-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "9c5d91bb-753e-4554-b7a1-6c85fe323875" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:13:03 crc kubenswrapper[4907]: I0313 14:13:03.937716 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c5d91bb-753e-4554-b7a1-6c85fe323875-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "9c5d91bb-753e-4554-b7a1-6c85fe323875" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:13:03 crc kubenswrapper[4907]: I0313 14:13:03.949968 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9c5d91bb-753e-4554-b7a1-6c85fe323875-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "9c5d91bb-753e-4554-b7a1-6c85fe323875" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:13:03 crc kubenswrapper[4907]: I0313 14:13:03.954268 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "9c5d91bb-753e-4554-b7a1-6c85fe323875" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Mar 13 14:13:04 crc kubenswrapper[4907]: I0313 14:13:04.029028 4907 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/9c5d91bb-753e-4554-b7a1-6c85fe323875-registry-certificates\") on node \"crc\" DevicePath \"\"" Mar 13 14:13:04 crc kubenswrapper[4907]: I0313 14:13:04.029068 4907 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/9c5d91bb-753e-4554-b7a1-6c85fe323875-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Mar 13 14:13:04 crc kubenswrapper[4907]: I0313 14:13:04.029079 4907 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/9c5d91bb-753e-4554-b7a1-6c85fe323875-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Mar 13 14:13:04 crc kubenswrapper[4907]: I0313 14:13:04.029091 4907 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/9c5d91bb-753e-4554-b7a1-6c85fe323875-registry-tls\") on node \"crc\" DevicePath \"\"" Mar 13 14:13:04 crc kubenswrapper[4907]: I0313 14:13:04.029099 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jb2r5\" (UniqueName: \"kubernetes.io/projected/9c5d91bb-753e-4554-b7a1-6c85fe323875-kube-api-access-jb2r5\") on node \"crc\" DevicePath \"\"" Mar 13 14:13:04 crc kubenswrapper[4907]: I0313 14:13:04.029107 4907 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9c5d91bb-753e-4554-b7a1-6c85fe323875-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 13 14:13:04 crc kubenswrapper[4907]: I0313 14:13:04.029119 4907 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9c5d91bb-753e-4554-b7a1-6c85fe323875-bound-sa-token\") on node \"crc\" DevicePath \"\"" Mar 13 14:13:04 crc kubenswrapper[4907]: I0313 14:13:04.066930 4907 generic.go:334] "Generic (PLEG): container finished" podID="9c5d91bb-753e-4554-b7a1-6c85fe323875" containerID="e2452e8d8fe737b72f44a8aba2c006a14d53bdfda4e4a01534d104020e7dd4aa" exitCode=0 Mar 13 14:13:04 crc kubenswrapper[4907]: I0313 14:13:04.066982 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" event={"ID":"9c5d91bb-753e-4554-b7a1-6c85fe323875","Type":"ContainerDied","Data":"e2452e8d8fe737b72f44a8aba2c006a14d53bdfda4e4a01534d104020e7dd4aa"} Mar 13 14:13:04 crc kubenswrapper[4907]: I0313 14:13:04.067009 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" event={"ID":"9c5d91bb-753e-4554-b7a1-6c85fe323875","Type":"ContainerDied","Data":"b92738be37732b07d185fe67067541e81cf30ad45197ff888383d57285721ac7"} Mar 13 14:13:04 crc kubenswrapper[4907]: I0313 14:13:04.066967 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:13:04 crc kubenswrapper[4907]: I0313 14:13:04.067023 4907 scope.go:117] "RemoveContainer" containerID="e2452e8d8fe737b72f44a8aba2c006a14d53bdfda4e4a01534d104020e7dd4aa" Mar 13 14:13:04 crc kubenswrapper[4907]: I0313 14:13:04.069336 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-slm6d" event={"ID":"aa7e1b0d-049f-4775-bd6b-d7430126fbd3","Type":"ContainerStarted","Data":"39cddf09ec86749cd87a8e8442276ad94033823f9427afcca89f9cc001c858da"} Mar 13 14:13:04 crc kubenswrapper[4907]: I0313 14:13:04.086511 4907 scope.go:117] "RemoveContainer" containerID="e2452e8d8fe737b72f44a8aba2c006a14d53bdfda4e4a01534d104020e7dd4aa" Mar 13 14:13:04 crc kubenswrapper[4907]: E0313 14:13:04.086917 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e2452e8d8fe737b72f44a8aba2c006a14d53bdfda4e4a01534d104020e7dd4aa\": container with ID starting with e2452e8d8fe737b72f44a8aba2c006a14d53bdfda4e4a01534d104020e7dd4aa not found: ID does not exist" containerID="e2452e8d8fe737b72f44a8aba2c006a14d53bdfda4e4a01534d104020e7dd4aa" Mar 13 14:13:04 crc kubenswrapper[4907]: I0313 14:13:04.086952 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e2452e8d8fe737b72f44a8aba2c006a14d53bdfda4e4a01534d104020e7dd4aa"} err="failed to get container status \"e2452e8d8fe737b72f44a8aba2c006a14d53bdfda4e4a01534d104020e7dd4aa\": rpc error: code = NotFound desc = could not find container \"e2452e8d8fe737b72f44a8aba2c006a14d53bdfda4e4a01534d104020e7dd4aa\": container with ID starting with e2452e8d8fe737b72f44a8aba2c006a14d53bdfda4e4a01534d104020e7dd4aa not found: ID does not exist" Mar 13 14:13:04 crc kubenswrapper[4907]: I0313 14:13:04.098771 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-slm6d" podStartSLOduration=2.622345007 podStartE2EDuration="5.098752891s" podCreationTimestamp="2026-03-13 14:12:59 +0000 UTC" firstStartedPulling="2026-03-13 14:13:01.013287096 +0000 UTC m=+479.913074825" lastFinishedPulling="2026-03-13 14:13:03.48969502 +0000 UTC m=+482.389482709" observedRunningTime="2026-03-13 14:13:04.098407802 +0000 UTC m=+482.998195501" watchObservedRunningTime="2026-03-13 14:13:04.098752891 +0000 UTC m=+482.998540600" Mar 13 14:13:04 crc kubenswrapper[4907]: I0313 14:13:04.124030 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-q9z8k"] Mar 13 14:13:04 crc kubenswrapper[4907]: I0313 14:13:04.130540 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-q9z8k"] Mar 13 14:13:05 crc kubenswrapper[4907]: I0313 14:13:05.800434 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9c5d91bb-753e-4554-b7a1-6c85fe323875" path="/var/lib/kubelet/pods/9c5d91bb-753e-4554-b7a1-6c85fe323875/volumes" Mar 13 14:13:06 crc kubenswrapper[4907]: I0313 14:13:06.886115 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-8dzhn" Mar 13 14:13:06 crc kubenswrapper[4907]: I0313 14:13:06.886699 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-8dzhn" Mar 13 14:13:06 crc kubenswrapper[4907]: I0313 14:13:06.941318 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-8dzhn" Mar 13 14:13:07 crc kubenswrapper[4907]: I0313 14:13:07.154693 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-8dzhn" Mar 13 14:13:07 crc kubenswrapper[4907]: I0313 14:13:07.491186 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-z7sd2" Mar 13 14:13:07 crc kubenswrapper[4907]: I0313 14:13:07.491238 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-z7sd2" Mar 13 14:13:08 crc kubenswrapper[4907]: I0313 14:13:08.534214 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-z7sd2" podUID="9d354400-9ba4-45b0-a866-09da4921210a" containerName="registry-server" probeResult="failure" output=< Mar 13 14:13:08 crc kubenswrapper[4907]: timeout: failed to connect service ":50051" within 1s Mar 13 14:13:08 crc kubenswrapper[4907]: > Mar 13 14:13:09 crc kubenswrapper[4907]: I0313 14:13:09.285308 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-bkqkq" Mar 13 14:13:09 crc kubenswrapper[4907]: I0313 14:13:09.285394 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-bkqkq" Mar 13 14:13:09 crc kubenswrapper[4907]: I0313 14:13:09.322967 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-bkqkq" Mar 13 14:13:09 crc kubenswrapper[4907]: I0313 14:13:09.980140 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-slm6d" Mar 13 14:13:09 crc kubenswrapper[4907]: I0313 14:13:09.980197 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-slm6d" Mar 13 14:13:10 crc kubenswrapper[4907]: I0313 14:13:10.014259 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-slm6d" Mar 13 14:13:10 crc kubenswrapper[4907]: I0313 14:13:10.144606 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-bkqkq" Mar 13 14:13:10 crc kubenswrapper[4907]: I0313 14:13:10.156766 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-slm6d" Mar 13 14:13:17 crc kubenswrapper[4907]: I0313 14:13:17.101089 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:13:17 crc kubenswrapper[4907]: I0313 14:13:17.532487 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-z7sd2" Mar 13 14:13:17 crc kubenswrapper[4907]: I0313 14:13:17.602617 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-z7sd2" Mar 13 14:13:18 crc kubenswrapper[4907]: I0313 14:13:18.041931 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 14:13:18 crc kubenswrapper[4907]: I0313 14:13:18.041984 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 14:13:18 crc kubenswrapper[4907]: I0313 14:13:18.042033 4907 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" Mar 13 14:13:18 crc kubenswrapper[4907]: I0313 14:13:18.042614 4907 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"93da0fb3f9dc1b1da4e8f0430569f23f1e76f6777c8f829bb761fb1ef946f9de"} pod="openshift-machine-config-operator/machine-config-daemon-hm56j" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 14:13:18 crc kubenswrapper[4907]: I0313 14:13:18.042656 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" containerID="cri-o://93da0fb3f9dc1b1da4e8f0430569f23f1e76f6777c8f829bb761fb1ef946f9de" gracePeriod=600 Mar 13 14:13:18 crc kubenswrapper[4907]: I0313 14:13:18.180394 4907 generic.go:334] "Generic (PLEG): container finished" podID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerID="93da0fb3f9dc1b1da4e8f0430569f23f1e76f6777c8f829bb761fb1ef946f9de" exitCode=0 Mar 13 14:13:18 crc kubenswrapper[4907]: I0313 14:13:18.180502 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerDied","Data":"93da0fb3f9dc1b1da4e8f0430569f23f1e76f6777c8f829bb761fb1ef946f9de"} Mar 13 14:13:18 crc kubenswrapper[4907]: I0313 14:13:18.180577 4907 scope.go:117] "RemoveContainer" containerID="770c882d9e9f9b2e9bd4f135e591952fa15b3898c42826aed8a75d953c9a875b" Mar 13 14:13:19 crc kubenswrapper[4907]: I0313 14:13:19.186924 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerStarted","Data":"c6951e52b25c8a7e922f875d35cd18c0cb0419427475b81395b9907dcedc0195"} Mar 13 14:14:00 crc kubenswrapper[4907]: I0313 14:14:00.165501 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556854-l7l9t"] Mar 13 14:14:00 crc kubenswrapper[4907]: E0313 14:14:00.168340 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c5d91bb-753e-4554-b7a1-6c85fe323875" containerName="registry" Mar 13 14:14:00 crc kubenswrapper[4907]: I0313 14:14:00.168527 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c5d91bb-753e-4554-b7a1-6c85fe323875" containerName="registry" Mar 13 14:14:00 crc kubenswrapper[4907]: I0313 14:14:00.168986 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c5d91bb-753e-4554-b7a1-6c85fe323875" containerName="registry" Mar 13 14:14:00 crc kubenswrapper[4907]: I0313 14:14:00.169941 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556854-l7l9t" Mar 13 14:14:00 crc kubenswrapper[4907]: I0313 14:14:00.172794 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 14:14:00 crc kubenswrapper[4907]: I0313 14:14:00.173867 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556854-l7l9t"] Mar 13 14:14:00 crc kubenswrapper[4907]: I0313 14:14:00.173962 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 14:14:00 crc kubenswrapper[4907]: I0313 14:14:00.178970 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 14:14:00 crc kubenswrapper[4907]: I0313 14:14:00.241475 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nkz6n\" (UniqueName: \"kubernetes.io/projected/82df106b-e5b8-4798-942a-999c2ac7a4be-kube-api-access-nkz6n\") pod \"auto-csr-approver-29556854-l7l9t\" (UID: \"82df106b-e5b8-4798-942a-999c2ac7a4be\") " pod="openshift-infra/auto-csr-approver-29556854-l7l9t" Mar 13 14:14:00 crc kubenswrapper[4907]: I0313 14:14:00.343657 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nkz6n\" (UniqueName: \"kubernetes.io/projected/82df106b-e5b8-4798-942a-999c2ac7a4be-kube-api-access-nkz6n\") pod \"auto-csr-approver-29556854-l7l9t\" (UID: \"82df106b-e5b8-4798-942a-999c2ac7a4be\") " pod="openshift-infra/auto-csr-approver-29556854-l7l9t" Mar 13 14:14:00 crc kubenswrapper[4907]: I0313 14:14:00.384493 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nkz6n\" (UniqueName: \"kubernetes.io/projected/82df106b-e5b8-4798-942a-999c2ac7a4be-kube-api-access-nkz6n\") pod \"auto-csr-approver-29556854-l7l9t\" (UID: \"82df106b-e5b8-4798-942a-999c2ac7a4be\") " pod="openshift-infra/auto-csr-approver-29556854-l7l9t" Mar 13 14:14:00 crc kubenswrapper[4907]: I0313 14:14:00.503847 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556854-l7l9t" Mar 13 14:14:00 crc kubenswrapper[4907]: I0313 14:14:00.993671 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556854-l7l9t"] Mar 13 14:14:01 crc kubenswrapper[4907]: I0313 14:14:01.460228 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556854-l7l9t" event={"ID":"82df106b-e5b8-4798-942a-999c2ac7a4be","Type":"ContainerStarted","Data":"967dbb55436e1bea2443bf62f90aa9ec79cd3b235fe5f9992286f8657a1468eb"} Mar 13 14:14:02 crc kubenswrapper[4907]: I0313 14:14:02.469423 4907 generic.go:334] "Generic (PLEG): container finished" podID="82df106b-e5b8-4798-942a-999c2ac7a4be" containerID="2fab8f0b7470e3acd4d60ef3af4057ac0e1b8ec4f3de833c7e05ac6623d0bf47" exitCode=0 Mar 13 14:14:02 crc kubenswrapper[4907]: I0313 14:14:02.469782 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556854-l7l9t" event={"ID":"82df106b-e5b8-4798-942a-999c2ac7a4be","Type":"ContainerDied","Data":"2fab8f0b7470e3acd4d60ef3af4057ac0e1b8ec4f3de833c7e05ac6623d0bf47"} Mar 13 14:14:03 crc kubenswrapper[4907]: I0313 14:14:03.746798 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556854-l7l9t" Mar 13 14:14:03 crc kubenswrapper[4907]: I0313 14:14:03.892963 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nkz6n\" (UniqueName: \"kubernetes.io/projected/82df106b-e5b8-4798-942a-999c2ac7a4be-kube-api-access-nkz6n\") pod \"82df106b-e5b8-4798-942a-999c2ac7a4be\" (UID: \"82df106b-e5b8-4798-942a-999c2ac7a4be\") " Mar 13 14:14:03 crc kubenswrapper[4907]: I0313 14:14:03.900942 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/82df106b-e5b8-4798-942a-999c2ac7a4be-kube-api-access-nkz6n" (OuterVolumeSpecName: "kube-api-access-nkz6n") pod "82df106b-e5b8-4798-942a-999c2ac7a4be" (UID: "82df106b-e5b8-4798-942a-999c2ac7a4be"). InnerVolumeSpecName "kube-api-access-nkz6n". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:14:03 crc kubenswrapper[4907]: I0313 14:14:03.995586 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nkz6n\" (UniqueName: \"kubernetes.io/projected/82df106b-e5b8-4798-942a-999c2ac7a4be-kube-api-access-nkz6n\") on node \"crc\" DevicePath \"\"" Mar 13 14:14:04 crc kubenswrapper[4907]: I0313 14:14:04.489076 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556854-l7l9t" event={"ID":"82df106b-e5b8-4798-942a-999c2ac7a4be","Type":"ContainerDied","Data":"967dbb55436e1bea2443bf62f90aa9ec79cd3b235fe5f9992286f8657a1468eb"} Mar 13 14:14:04 crc kubenswrapper[4907]: I0313 14:14:04.489158 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="967dbb55436e1bea2443bf62f90aa9ec79cd3b235fe5f9992286f8657a1468eb" Mar 13 14:14:04 crc kubenswrapper[4907]: I0313 14:14:04.489397 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556854-l7l9t" Mar 13 14:14:04 crc kubenswrapper[4907]: I0313 14:14:04.844963 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556848-v8kv4"] Mar 13 14:14:04 crc kubenswrapper[4907]: I0313 14:14:04.852307 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556848-v8kv4"] Mar 13 14:14:05 crc kubenswrapper[4907]: I0313 14:14:05.797935 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a8867781-179d-47e4-9a55-330deb10d273" path="/var/lib/kubelet/pods/a8867781-179d-47e4-9a55-330deb10d273/volumes" Mar 13 14:15:00 crc kubenswrapper[4907]: I0313 14:15:00.155639 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556855-t8w6z"] Mar 13 14:15:00 crc kubenswrapper[4907]: E0313 14:15:00.156980 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82df106b-e5b8-4798-942a-999c2ac7a4be" containerName="oc" Mar 13 14:15:00 crc kubenswrapper[4907]: I0313 14:15:00.157341 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="82df106b-e5b8-4798-942a-999c2ac7a4be" containerName="oc" Mar 13 14:15:00 crc kubenswrapper[4907]: I0313 14:15:00.157616 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="82df106b-e5b8-4798-942a-999c2ac7a4be" containerName="oc" Mar 13 14:15:00 crc kubenswrapper[4907]: I0313 14:15:00.158325 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556855-t8w6z" Mar 13 14:15:00 crc kubenswrapper[4907]: I0313 14:15:00.164039 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 13 14:15:00 crc kubenswrapper[4907]: I0313 14:15:00.164864 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 13 14:15:00 crc kubenswrapper[4907]: I0313 14:15:00.182493 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556855-t8w6z"] Mar 13 14:15:00 crc kubenswrapper[4907]: I0313 14:15:00.320778 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/80282185-9b05-4101-b906-e069f57e9bae-config-volume\") pod \"collect-profiles-29556855-t8w6z\" (UID: \"80282185-9b05-4101-b906-e069f57e9bae\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556855-t8w6z" Mar 13 14:15:00 crc kubenswrapper[4907]: I0313 14:15:00.320860 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hjwzv\" (UniqueName: \"kubernetes.io/projected/80282185-9b05-4101-b906-e069f57e9bae-kube-api-access-hjwzv\") pod \"collect-profiles-29556855-t8w6z\" (UID: \"80282185-9b05-4101-b906-e069f57e9bae\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556855-t8w6z" Mar 13 14:15:00 crc kubenswrapper[4907]: I0313 14:15:00.320972 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/80282185-9b05-4101-b906-e069f57e9bae-secret-volume\") pod \"collect-profiles-29556855-t8w6z\" (UID: \"80282185-9b05-4101-b906-e069f57e9bae\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556855-t8w6z" Mar 13 14:15:00 crc kubenswrapper[4907]: I0313 14:15:00.422263 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/80282185-9b05-4101-b906-e069f57e9bae-secret-volume\") pod \"collect-profiles-29556855-t8w6z\" (UID: \"80282185-9b05-4101-b906-e069f57e9bae\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556855-t8w6z" Mar 13 14:15:00 crc kubenswrapper[4907]: I0313 14:15:00.422772 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/80282185-9b05-4101-b906-e069f57e9bae-config-volume\") pod \"collect-profiles-29556855-t8w6z\" (UID: \"80282185-9b05-4101-b906-e069f57e9bae\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556855-t8w6z" Mar 13 14:15:00 crc kubenswrapper[4907]: I0313 14:15:00.423091 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hjwzv\" (UniqueName: \"kubernetes.io/projected/80282185-9b05-4101-b906-e069f57e9bae-kube-api-access-hjwzv\") pod \"collect-profiles-29556855-t8w6z\" (UID: \"80282185-9b05-4101-b906-e069f57e9bae\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556855-t8w6z" Mar 13 14:15:00 crc kubenswrapper[4907]: I0313 14:15:00.424470 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/80282185-9b05-4101-b906-e069f57e9bae-config-volume\") pod \"collect-profiles-29556855-t8w6z\" (UID: \"80282185-9b05-4101-b906-e069f57e9bae\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556855-t8w6z" Mar 13 14:15:00 crc kubenswrapper[4907]: I0313 14:15:00.429705 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/80282185-9b05-4101-b906-e069f57e9bae-secret-volume\") pod \"collect-profiles-29556855-t8w6z\" (UID: \"80282185-9b05-4101-b906-e069f57e9bae\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556855-t8w6z" Mar 13 14:15:00 crc kubenswrapper[4907]: I0313 14:15:00.443430 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hjwzv\" (UniqueName: \"kubernetes.io/projected/80282185-9b05-4101-b906-e069f57e9bae-kube-api-access-hjwzv\") pod \"collect-profiles-29556855-t8w6z\" (UID: \"80282185-9b05-4101-b906-e069f57e9bae\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556855-t8w6z" Mar 13 14:15:00 crc kubenswrapper[4907]: I0313 14:15:00.530484 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556855-t8w6z" Mar 13 14:15:00 crc kubenswrapper[4907]: I0313 14:15:00.748542 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556855-t8w6z"] Mar 13 14:15:00 crc kubenswrapper[4907]: I0313 14:15:00.872469 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556855-t8w6z" event={"ID":"80282185-9b05-4101-b906-e069f57e9bae","Type":"ContainerStarted","Data":"1113c32af3f3a8484a7765dfa4c4a4904cf0b0cad6e1387975333f03729886ff"} Mar 13 14:15:00 crc kubenswrapper[4907]: I0313 14:15:00.872507 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556855-t8w6z" event={"ID":"80282185-9b05-4101-b906-e069f57e9bae","Type":"ContainerStarted","Data":"e783ab47360b0f678f98f9fecaaf31bb829cbc86f6e08ca80e77430e665ca4a4"} Mar 13 14:15:00 crc kubenswrapper[4907]: I0313 14:15:00.889761 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29556855-t8w6z" podStartSLOduration=0.889742696 podStartE2EDuration="889.742696ms" podCreationTimestamp="2026-03-13 14:15:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:15:00.88665707 +0000 UTC m=+599.786444759" watchObservedRunningTime="2026-03-13 14:15:00.889742696 +0000 UTC m=+599.789530395" Mar 13 14:15:01 crc kubenswrapper[4907]: I0313 14:15:01.880719 4907 generic.go:334] "Generic (PLEG): container finished" podID="80282185-9b05-4101-b906-e069f57e9bae" containerID="1113c32af3f3a8484a7765dfa4c4a4904cf0b0cad6e1387975333f03729886ff" exitCode=0 Mar 13 14:15:01 crc kubenswrapper[4907]: I0313 14:15:01.880764 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556855-t8w6z" event={"ID":"80282185-9b05-4101-b906-e069f57e9bae","Type":"ContainerDied","Data":"1113c32af3f3a8484a7765dfa4c4a4904cf0b0cad6e1387975333f03729886ff"} Mar 13 14:15:03 crc kubenswrapper[4907]: I0313 14:15:03.190584 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556855-t8w6z" Mar 13 14:15:03 crc kubenswrapper[4907]: I0313 14:15:03.359425 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/80282185-9b05-4101-b906-e069f57e9bae-config-volume\") pod \"80282185-9b05-4101-b906-e069f57e9bae\" (UID: \"80282185-9b05-4101-b906-e069f57e9bae\") " Mar 13 14:15:03 crc kubenswrapper[4907]: I0313 14:15:03.359634 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hjwzv\" (UniqueName: \"kubernetes.io/projected/80282185-9b05-4101-b906-e069f57e9bae-kube-api-access-hjwzv\") pod \"80282185-9b05-4101-b906-e069f57e9bae\" (UID: \"80282185-9b05-4101-b906-e069f57e9bae\") " Mar 13 14:15:03 crc kubenswrapper[4907]: I0313 14:15:03.359724 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/80282185-9b05-4101-b906-e069f57e9bae-secret-volume\") pod \"80282185-9b05-4101-b906-e069f57e9bae\" (UID: \"80282185-9b05-4101-b906-e069f57e9bae\") " Mar 13 14:15:03 crc kubenswrapper[4907]: I0313 14:15:03.360698 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/80282185-9b05-4101-b906-e069f57e9bae-config-volume" (OuterVolumeSpecName: "config-volume") pod "80282185-9b05-4101-b906-e069f57e9bae" (UID: "80282185-9b05-4101-b906-e069f57e9bae"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:15:03 crc kubenswrapper[4907]: I0313 14:15:03.367751 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/80282185-9b05-4101-b906-e069f57e9bae-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "80282185-9b05-4101-b906-e069f57e9bae" (UID: "80282185-9b05-4101-b906-e069f57e9bae"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:15:03 crc kubenswrapper[4907]: I0313 14:15:03.368790 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/80282185-9b05-4101-b906-e069f57e9bae-kube-api-access-hjwzv" (OuterVolumeSpecName: "kube-api-access-hjwzv") pod "80282185-9b05-4101-b906-e069f57e9bae" (UID: "80282185-9b05-4101-b906-e069f57e9bae"). InnerVolumeSpecName "kube-api-access-hjwzv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:15:03 crc kubenswrapper[4907]: I0313 14:15:03.461412 4907 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/80282185-9b05-4101-b906-e069f57e9bae-config-volume\") on node \"crc\" DevicePath \"\"" Mar 13 14:15:03 crc kubenswrapper[4907]: I0313 14:15:03.461464 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hjwzv\" (UniqueName: \"kubernetes.io/projected/80282185-9b05-4101-b906-e069f57e9bae-kube-api-access-hjwzv\") on node \"crc\" DevicePath \"\"" Mar 13 14:15:03 crc kubenswrapper[4907]: I0313 14:15:03.461488 4907 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/80282185-9b05-4101-b906-e069f57e9bae-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 13 14:15:03 crc kubenswrapper[4907]: I0313 14:15:03.896387 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556855-t8w6z" event={"ID":"80282185-9b05-4101-b906-e069f57e9bae","Type":"ContainerDied","Data":"e783ab47360b0f678f98f9fecaaf31bb829cbc86f6e08ca80e77430e665ca4a4"} Mar 13 14:15:03 crc kubenswrapper[4907]: I0313 14:15:03.896439 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e783ab47360b0f678f98f9fecaaf31bb829cbc86f6e08ca80e77430e665ca4a4" Mar 13 14:15:03 crc kubenswrapper[4907]: I0313 14:15:03.896456 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556855-t8w6z" Mar 13 14:15:18 crc kubenswrapper[4907]: I0313 14:15:18.042024 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 14:15:18 crc kubenswrapper[4907]: I0313 14:15:18.042645 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 14:15:48 crc kubenswrapper[4907]: I0313 14:15:48.041629 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 14:15:48 crc kubenswrapper[4907]: I0313 14:15:48.042319 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 14:16:00 crc kubenswrapper[4907]: I0313 14:16:00.139582 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556856-8grxd"] Mar 13 14:16:00 crc kubenswrapper[4907]: E0313 14:16:00.140192 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80282185-9b05-4101-b906-e069f57e9bae" containerName="collect-profiles" Mar 13 14:16:00 crc kubenswrapper[4907]: I0313 14:16:00.140203 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="80282185-9b05-4101-b906-e069f57e9bae" containerName="collect-profiles" Mar 13 14:16:00 crc kubenswrapper[4907]: I0313 14:16:00.140303 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="80282185-9b05-4101-b906-e069f57e9bae" containerName="collect-profiles" Mar 13 14:16:00 crc kubenswrapper[4907]: I0313 14:16:00.140605 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556856-8grxd" Mar 13 14:16:00 crc kubenswrapper[4907]: I0313 14:16:00.143110 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 14:16:00 crc kubenswrapper[4907]: I0313 14:16:00.144145 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 14:16:00 crc kubenswrapper[4907]: I0313 14:16:00.148488 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 14:16:00 crc kubenswrapper[4907]: I0313 14:16:00.162845 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556856-8grxd"] Mar 13 14:16:00 crc kubenswrapper[4907]: I0313 14:16:00.282705 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6sp5k\" (UniqueName: \"kubernetes.io/projected/1604551e-a53d-483c-a03e-3c2159768001-kube-api-access-6sp5k\") pod \"auto-csr-approver-29556856-8grxd\" (UID: \"1604551e-a53d-483c-a03e-3c2159768001\") " pod="openshift-infra/auto-csr-approver-29556856-8grxd" Mar 13 14:16:00 crc kubenswrapper[4907]: I0313 14:16:00.384694 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6sp5k\" (UniqueName: \"kubernetes.io/projected/1604551e-a53d-483c-a03e-3c2159768001-kube-api-access-6sp5k\") pod \"auto-csr-approver-29556856-8grxd\" (UID: \"1604551e-a53d-483c-a03e-3c2159768001\") " pod="openshift-infra/auto-csr-approver-29556856-8grxd" Mar 13 14:16:00 crc kubenswrapper[4907]: I0313 14:16:00.419509 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6sp5k\" (UniqueName: \"kubernetes.io/projected/1604551e-a53d-483c-a03e-3c2159768001-kube-api-access-6sp5k\") pod \"auto-csr-approver-29556856-8grxd\" (UID: \"1604551e-a53d-483c-a03e-3c2159768001\") " pod="openshift-infra/auto-csr-approver-29556856-8grxd" Mar 13 14:16:00 crc kubenswrapper[4907]: I0313 14:16:00.459167 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556856-8grxd" Mar 13 14:16:00 crc kubenswrapper[4907]: I0313 14:16:00.709862 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556856-8grxd"] Mar 13 14:16:00 crc kubenswrapper[4907]: I0313 14:16:00.719141 4907 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 13 14:16:01 crc kubenswrapper[4907]: I0313 14:16:01.434686 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556856-8grxd" event={"ID":"1604551e-a53d-483c-a03e-3c2159768001","Type":"ContainerStarted","Data":"c5e66cf058a4411962cd24af99e75819cdb33969ea8b7d7705c9750d7b4c56eb"} Mar 13 14:16:02 crc kubenswrapper[4907]: I0313 14:16:02.200814 4907 scope.go:117] "RemoveContainer" containerID="2e545187909a28cc5890020e1139c1d488e394bab920eab3166b317419887a60" Mar 13 14:16:02 crc kubenswrapper[4907]: I0313 14:16:02.444957 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556856-8grxd" event={"ID":"1604551e-a53d-483c-a03e-3c2159768001","Type":"ContainerStarted","Data":"ee5ef069f4132418de863818eaddffd619fd31484fffe36ce7827c7d59840bbd"} Mar 13 14:16:02 crc kubenswrapper[4907]: I0313 14:16:02.464091 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29556856-8grxd" podStartSLOduration=1.092545248 podStartE2EDuration="2.464065981s" podCreationTimestamp="2026-03-13 14:16:00 +0000 UTC" firstStartedPulling="2026-03-13 14:16:00.718784855 +0000 UTC m=+659.618572554" lastFinishedPulling="2026-03-13 14:16:02.090305598 +0000 UTC m=+660.990093287" observedRunningTime="2026-03-13 14:16:02.45975387 +0000 UTC m=+661.359541559" watchObservedRunningTime="2026-03-13 14:16:02.464065981 +0000 UTC m=+661.363853670" Mar 13 14:16:03 crc kubenswrapper[4907]: I0313 14:16:03.458024 4907 generic.go:334] "Generic (PLEG): container finished" podID="1604551e-a53d-483c-a03e-3c2159768001" containerID="ee5ef069f4132418de863818eaddffd619fd31484fffe36ce7827c7d59840bbd" exitCode=0 Mar 13 14:16:03 crc kubenswrapper[4907]: I0313 14:16:03.458143 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556856-8grxd" event={"ID":"1604551e-a53d-483c-a03e-3c2159768001","Type":"ContainerDied","Data":"ee5ef069f4132418de863818eaddffd619fd31484fffe36ce7827c7d59840bbd"} Mar 13 14:16:04 crc kubenswrapper[4907]: I0313 14:16:04.809461 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556856-8grxd" Mar 13 14:16:04 crc kubenswrapper[4907]: I0313 14:16:04.973700 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6sp5k\" (UniqueName: \"kubernetes.io/projected/1604551e-a53d-483c-a03e-3c2159768001-kube-api-access-6sp5k\") pod \"1604551e-a53d-483c-a03e-3c2159768001\" (UID: \"1604551e-a53d-483c-a03e-3c2159768001\") " Mar 13 14:16:04 crc kubenswrapper[4907]: I0313 14:16:04.981376 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1604551e-a53d-483c-a03e-3c2159768001-kube-api-access-6sp5k" (OuterVolumeSpecName: "kube-api-access-6sp5k") pod "1604551e-a53d-483c-a03e-3c2159768001" (UID: "1604551e-a53d-483c-a03e-3c2159768001"). InnerVolumeSpecName "kube-api-access-6sp5k". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:16:05 crc kubenswrapper[4907]: I0313 14:16:05.075632 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6sp5k\" (UniqueName: \"kubernetes.io/projected/1604551e-a53d-483c-a03e-3c2159768001-kube-api-access-6sp5k\") on node \"crc\" DevicePath \"\"" Mar 13 14:16:05 crc kubenswrapper[4907]: I0313 14:16:05.473285 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556856-8grxd" event={"ID":"1604551e-a53d-483c-a03e-3c2159768001","Type":"ContainerDied","Data":"c5e66cf058a4411962cd24af99e75819cdb33969ea8b7d7705c9750d7b4c56eb"} Mar 13 14:16:05 crc kubenswrapper[4907]: I0313 14:16:05.473374 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c5e66cf058a4411962cd24af99e75819cdb33969ea8b7d7705c9750d7b4c56eb" Mar 13 14:16:05 crc kubenswrapper[4907]: I0313 14:16:05.473708 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556856-8grxd" Mar 13 14:16:05 crc kubenswrapper[4907]: I0313 14:16:05.880154 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556850-cl45g"] Mar 13 14:16:05 crc kubenswrapper[4907]: I0313 14:16:05.885290 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556850-cl45g"] Mar 13 14:16:07 crc kubenswrapper[4907]: I0313 14:16:07.793628 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="10a1d486-2cf1-4078-8f0f-2dbc1ea97acc" path="/var/lib/kubelet/pods/10a1d486-2cf1-4078-8f0f-2dbc1ea97acc/volumes" Mar 13 14:16:18 crc kubenswrapper[4907]: I0313 14:16:18.041526 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 14:16:18 crc kubenswrapper[4907]: I0313 14:16:18.042686 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 14:16:18 crc kubenswrapper[4907]: I0313 14:16:18.042778 4907 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" Mar 13 14:16:18 crc kubenswrapper[4907]: I0313 14:16:18.043834 4907 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c6951e52b25c8a7e922f875d35cd18c0cb0419427475b81395b9907dcedc0195"} pod="openshift-machine-config-operator/machine-config-daemon-hm56j" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 14:16:18 crc kubenswrapper[4907]: I0313 14:16:18.043965 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" containerID="cri-o://c6951e52b25c8a7e922f875d35cd18c0cb0419427475b81395b9907dcedc0195" gracePeriod=600 Mar 13 14:16:18 crc kubenswrapper[4907]: I0313 14:16:18.567631 4907 generic.go:334] "Generic (PLEG): container finished" podID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerID="c6951e52b25c8a7e922f875d35cd18c0cb0419427475b81395b9907dcedc0195" exitCode=0 Mar 13 14:16:18 crc kubenswrapper[4907]: I0313 14:16:18.567722 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerDied","Data":"c6951e52b25c8a7e922f875d35cd18c0cb0419427475b81395b9907dcedc0195"} Mar 13 14:16:18 crc kubenswrapper[4907]: I0313 14:16:18.568407 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerStarted","Data":"da83df5fde38cf429de6b21ade144e7c74078db3c329650c84ad8d669e1beb0e"} Mar 13 14:16:18 crc kubenswrapper[4907]: I0313 14:16:18.568477 4907 scope.go:117] "RemoveContainer" containerID="93da0fb3f9dc1b1da4e8f0430569f23f1e76f6777c8f829bb761fb1ef946f9de" Mar 13 14:17:02 crc kubenswrapper[4907]: I0313 14:17:02.258144 4907 scope.go:117] "RemoveContainer" containerID="7118e101427b23d94455b072a1fe8c88f9a426be2d07a4a232a1f55839394ad6" Mar 13 14:18:00 crc kubenswrapper[4907]: I0313 14:18:00.150157 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556858-bk6rj"] Mar 13 14:18:00 crc kubenswrapper[4907]: E0313 14:18:00.151087 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1604551e-a53d-483c-a03e-3c2159768001" containerName="oc" Mar 13 14:18:00 crc kubenswrapper[4907]: I0313 14:18:00.151108 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="1604551e-a53d-483c-a03e-3c2159768001" containerName="oc" Mar 13 14:18:00 crc kubenswrapper[4907]: I0313 14:18:00.151277 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="1604551e-a53d-483c-a03e-3c2159768001" containerName="oc" Mar 13 14:18:00 crc kubenswrapper[4907]: I0313 14:18:00.151939 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556858-bk6rj" Mar 13 14:18:00 crc kubenswrapper[4907]: I0313 14:18:00.154556 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 14:18:00 crc kubenswrapper[4907]: I0313 14:18:00.155044 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 14:18:00 crc kubenswrapper[4907]: I0313 14:18:00.155349 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 14:18:00 crc kubenswrapper[4907]: I0313 14:18:00.170538 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556858-bk6rj"] Mar 13 14:18:00 crc kubenswrapper[4907]: I0313 14:18:00.249087 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dkhtw\" (UniqueName: \"kubernetes.io/projected/03fa7dad-f13d-4e5c-8b9c-60b2ef2d272a-kube-api-access-dkhtw\") pod \"auto-csr-approver-29556858-bk6rj\" (UID: \"03fa7dad-f13d-4e5c-8b9c-60b2ef2d272a\") " pod="openshift-infra/auto-csr-approver-29556858-bk6rj" Mar 13 14:18:00 crc kubenswrapper[4907]: I0313 14:18:00.351293 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dkhtw\" (UniqueName: \"kubernetes.io/projected/03fa7dad-f13d-4e5c-8b9c-60b2ef2d272a-kube-api-access-dkhtw\") pod \"auto-csr-approver-29556858-bk6rj\" (UID: \"03fa7dad-f13d-4e5c-8b9c-60b2ef2d272a\") " pod="openshift-infra/auto-csr-approver-29556858-bk6rj" Mar 13 14:18:00 crc kubenswrapper[4907]: I0313 14:18:00.380393 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dkhtw\" (UniqueName: \"kubernetes.io/projected/03fa7dad-f13d-4e5c-8b9c-60b2ef2d272a-kube-api-access-dkhtw\") pod \"auto-csr-approver-29556858-bk6rj\" (UID: \"03fa7dad-f13d-4e5c-8b9c-60b2ef2d272a\") " pod="openshift-infra/auto-csr-approver-29556858-bk6rj" Mar 13 14:18:00 crc kubenswrapper[4907]: I0313 14:18:00.482441 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556858-bk6rj" Mar 13 14:18:00 crc kubenswrapper[4907]: I0313 14:18:00.767073 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556858-bk6rj"] Mar 13 14:18:01 crc kubenswrapper[4907]: I0313 14:18:01.452773 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556858-bk6rj" event={"ID":"03fa7dad-f13d-4e5c-8b9c-60b2ef2d272a","Type":"ContainerStarted","Data":"6cee4ab91f9e43ac6ed054dabe1750fcb13c6fcca497178aea30982262b12a53"} Mar 13 14:18:02 crc kubenswrapper[4907]: I0313 14:18:02.463580 4907 generic.go:334] "Generic (PLEG): container finished" podID="03fa7dad-f13d-4e5c-8b9c-60b2ef2d272a" containerID="7377ca75e3097f224f1949f7536a8f7a1c19b44224ea656c9d4dd6e3d82e1652" exitCode=0 Mar 13 14:18:02 crc kubenswrapper[4907]: I0313 14:18:02.463667 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556858-bk6rj" event={"ID":"03fa7dad-f13d-4e5c-8b9c-60b2ef2d272a","Type":"ContainerDied","Data":"7377ca75e3097f224f1949f7536a8f7a1c19b44224ea656c9d4dd6e3d82e1652"} Mar 13 14:18:03 crc kubenswrapper[4907]: I0313 14:18:03.795878 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556858-bk6rj" Mar 13 14:18:03 crc kubenswrapper[4907]: I0313 14:18:03.903398 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dkhtw\" (UniqueName: \"kubernetes.io/projected/03fa7dad-f13d-4e5c-8b9c-60b2ef2d272a-kube-api-access-dkhtw\") pod \"03fa7dad-f13d-4e5c-8b9c-60b2ef2d272a\" (UID: \"03fa7dad-f13d-4e5c-8b9c-60b2ef2d272a\") " Mar 13 14:18:03 crc kubenswrapper[4907]: I0313 14:18:03.913676 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/03fa7dad-f13d-4e5c-8b9c-60b2ef2d272a-kube-api-access-dkhtw" (OuterVolumeSpecName: "kube-api-access-dkhtw") pod "03fa7dad-f13d-4e5c-8b9c-60b2ef2d272a" (UID: "03fa7dad-f13d-4e5c-8b9c-60b2ef2d272a"). InnerVolumeSpecName "kube-api-access-dkhtw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:18:04 crc kubenswrapper[4907]: I0313 14:18:04.005235 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dkhtw\" (UniqueName: \"kubernetes.io/projected/03fa7dad-f13d-4e5c-8b9c-60b2ef2d272a-kube-api-access-dkhtw\") on node \"crc\" DevicePath \"\"" Mar 13 14:18:04 crc kubenswrapper[4907]: I0313 14:18:04.478719 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556858-bk6rj" event={"ID":"03fa7dad-f13d-4e5c-8b9c-60b2ef2d272a","Type":"ContainerDied","Data":"6cee4ab91f9e43ac6ed054dabe1750fcb13c6fcca497178aea30982262b12a53"} Mar 13 14:18:04 crc kubenswrapper[4907]: I0313 14:18:04.479379 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6cee4ab91f9e43ac6ed054dabe1750fcb13c6fcca497178aea30982262b12a53" Mar 13 14:18:04 crc kubenswrapper[4907]: I0313 14:18:04.478801 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556858-bk6rj" Mar 13 14:18:04 crc kubenswrapper[4907]: I0313 14:18:04.877386 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556852-bg4rn"] Mar 13 14:18:04 crc kubenswrapper[4907]: I0313 14:18:04.884510 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556852-bg4rn"] Mar 13 14:18:05 crc kubenswrapper[4907]: I0313 14:18:05.795203 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5cd3c3cd-9c71-4d90-a5cb-800de65b86de" path="/var/lib/kubelet/pods/5cd3c3cd-9c71-4d90-a5cb-800de65b86de/volumes" Mar 13 14:18:18 crc kubenswrapper[4907]: I0313 14:18:18.323065 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 14:18:18 crc kubenswrapper[4907]: I0313 14:18:18.324180 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 14:18:48 crc kubenswrapper[4907]: I0313 14:18:48.041521 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 14:18:48 crc kubenswrapper[4907]: I0313 14:18:48.042180 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 14:19:02 crc kubenswrapper[4907]: I0313 14:19:02.343096 4907 scope.go:117] "RemoveContainer" containerID="cbcf7b779b5c51c01226dc524be3c2eb43327ef749d0c0644c47a0b6f9140e3e" Mar 13 14:19:18 crc kubenswrapper[4907]: I0313 14:19:18.042105 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 14:19:18 crc kubenswrapper[4907]: I0313 14:19:18.042753 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 14:19:18 crc kubenswrapper[4907]: I0313 14:19:18.042815 4907 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" Mar 13 14:19:18 crc kubenswrapper[4907]: I0313 14:19:18.043631 4907 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"da83df5fde38cf429de6b21ade144e7c74078db3c329650c84ad8d669e1beb0e"} pod="openshift-machine-config-operator/machine-config-daemon-hm56j" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 14:19:18 crc kubenswrapper[4907]: I0313 14:19:18.043732 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" containerID="cri-o://da83df5fde38cf429de6b21ade144e7c74078db3c329650c84ad8d669e1beb0e" gracePeriod=600 Mar 13 14:19:18 crc kubenswrapper[4907]: I0313 14:19:18.727012 4907 generic.go:334] "Generic (PLEG): container finished" podID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerID="da83df5fde38cf429de6b21ade144e7c74078db3c329650c84ad8d669e1beb0e" exitCode=0 Mar 13 14:19:18 crc kubenswrapper[4907]: I0313 14:19:18.727086 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerDied","Data":"da83df5fde38cf429de6b21ade144e7c74078db3c329650c84ad8d669e1beb0e"} Mar 13 14:19:18 crc kubenswrapper[4907]: I0313 14:19:18.727671 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerStarted","Data":"abe5ce1bc0297c29b962bd8d44bae1a474036514922a5df8e114b5dafeb79cce"} Mar 13 14:19:18 crc kubenswrapper[4907]: I0313 14:19:18.727695 4907 scope.go:117] "RemoveContainer" containerID="c6951e52b25c8a7e922f875d35cd18c0cb0419427475b81395b9907dcedc0195" Mar 13 14:19:20 crc kubenswrapper[4907]: I0313 14:19:20.879383 4907 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.269510 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-qxlx5"] Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.270371 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" podUID="1006376d-4033-454e-8a23-2b67882bdd0b" containerName="ovn-controller" containerID="cri-o://fea41c11bebd2d9d622ed4b8bc50e03a106dadea6ac102cd60512ea2a1f8b8b0" gracePeriod=30 Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.270731 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" podUID="1006376d-4033-454e-8a23-2b67882bdd0b" containerName="sbdb" containerID="cri-o://4d60b5b774f8dd4cee16b9784f80d124640199a85331a2940772029be4884751" gracePeriod=30 Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.270767 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" podUID="1006376d-4033-454e-8a23-2b67882bdd0b" containerName="nbdb" containerID="cri-o://a0ee0049aa4a6d49a6bcf7d347183d9cc8b85171f988662803a005e0cdd97395" gracePeriod=30 Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.270840 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" podUID="1006376d-4033-454e-8a23-2b67882bdd0b" containerName="northd" containerID="cri-o://0fff87d8ea572615eb1329164f9e536e6ee45cd00a9e71be0aa8a9b3ed2190e2" gracePeriod=30 Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.270919 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" podUID="1006376d-4033-454e-8a23-2b67882bdd0b" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://ccef92e41501fab91189581d7e8875515c02d637426b329d2fb80a356c75b637" gracePeriod=30 Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.270957 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" podUID="1006376d-4033-454e-8a23-2b67882bdd0b" containerName="kube-rbac-proxy-node" containerID="cri-o://bf6a3dbaa8dda61f461a906538824c8946d36dc07d4fb95bf4d8420ad0105600" gracePeriod=30 Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.271018 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" podUID="1006376d-4033-454e-8a23-2b67882bdd0b" containerName="ovn-acl-logging" containerID="cri-o://3c165053fd9933444b372ff40efe33c0e193a7f39e6f531a4f370e971eb41e6d" gracePeriod=30 Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.314645 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" podUID="1006376d-4033-454e-8a23-2b67882bdd0b" containerName="ovnkube-controller" containerID="cri-o://a3d2a347b2f85682f2a54c30e63b7e9c88ffe402d6d71fa7ce6656786f93b886" gracePeriod=30 Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.611833 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qxlx5_1006376d-4033-454e-8a23-2b67882bdd0b/ovnkube-controller/3.log" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.615351 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qxlx5_1006376d-4033-454e-8a23-2b67882bdd0b/ovn-acl-logging/0.log" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.616016 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qxlx5_1006376d-4033-454e-8a23-2b67882bdd0b/ovn-controller/0.log" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.616453 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.678091 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-8pphf"] Mar 13 14:19:42 crc kubenswrapper[4907]: E0313 14:19:42.678466 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1006376d-4033-454e-8a23-2b67882bdd0b" containerName="ovnkube-controller" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.678493 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="1006376d-4033-454e-8a23-2b67882bdd0b" containerName="ovnkube-controller" Mar 13 14:19:42 crc kubenswrapper[4907]: E0313 14:19:42.678531 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1006376d-4033-454e-8a23-2b67882bdd0b" containerName="nbdb" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.678542 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="1006376d-4033-454e-8a23-2b67882bdd0b" containerName="nbdb" Mar 13 14:19:42 crc kubenswrapper[4907]: E0313 14:19:42.678555 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1006376d-4033-454e-8a23-2b67882bdd0b" containerName="sbdb" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.678563 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="1006376d-4033-454e-8a23-2b67882bdd0b" containerName="sbdb" Mar 13 14:19:42 crc kubenswrapper[4907]: E0313 14:19:42.678576 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1006376d-4033-454e-8a23-2b67882bdd0b" containerName="ovn-acl-logging" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.678584 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="1006376d-4033-454e-8a23-2b67882bdd0b" containerName="ovn-acl-logging" Mar 13 14:19:42 crc kubenswrapper[4907]: E0313 14:19:42.678622 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1006376d-4033-454e-8a23-2b67882bdd0b" containerName="kubecfg-setup" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.678633 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="1006376d-4033-454e-8a23-2b67882bdd0b" containerName="kubecfg-setup" Mar 13 14:19:42 crc kubenswrapper[4907]: E0313 14:19:42.678646 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1006376d-4033-454e-8a23-2b67882bdd0b" containerName="ovnkube-controller" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.678655 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="1006376d-4033-454e-8a23-2b67882bdd0b" containerName="ovnkube-controller" Mar 13 14:19:42 crc kubenswrapper[4907]: E0313 14:19:42.678664 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1006376d-4033-454e-8a23-2b67882bdd0b" containerName="ovnkube-controller" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.678694 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="1006376d-4033-454e-8a23-2b67882bdd0b" containerName="ovnkube-controller" Mar 13 14:19:42 crc kubenswrapper[4907]: E0313 14:19:42.678710 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03fa7dad-f13d-4e5c-8b9c-60b2ef2d272a" containerName="oc" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.678721 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="03fa7dad-f13d-4e5c-8b9c-60b2ef2d272a" containerName="oc" Mar 13 14:19:42 crc kubenswrapper[4907]: E0313 14:19:42.678733 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1006376d-4033-454e-8a23-2b67882bdd0b" containerName="northd" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.678740 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="1006376d-4033-454e-8a23-2b67882bdd0b" containerName="northd" Mar 13 14:19:42 crc kubenswrapper[4907]: E0313 14:19:42.678814 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1006376d-4033-454e-8a23-2b67882bdd0b" containerName="ovn-controller" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.678827 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="1006376d-4033-454e-8a23-2b67882bdd0b" containerName="ovn-controller" Mar 13 14:19:42 crc kubenswrapper[4907]: E0313 14:19:42.678906 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1006376d-4033-454e-8a23-2b67882bdd0b" containerName="kube-rbac-proxy-node" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.678920 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="1006376d-4033-454e-8a23-2b67882bdd0b" containerName="kube-rbac-proxy-node" Mar 13 14:19:42 crc kubenswrapper[4907]: E0313 14:19:42.678931 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1006376d-4033-454e-8a23-2b67882bdd0b" containerName="kube-rbac-proxy-ovn-metrics" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.678939 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="1006376d-4033-454e-8a23-2b67882bdd0b" containerName="kube-rbac-proxy-ovn-metrics" Mar 13 14:19:42 crc kubenswrapper[4907]: E0313 14:19:42.678949 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1006376d-4033-454e-8a23-2b67882bdd0b" containerName="ovnkube-controller" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.678958 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="1006376d-4033-454e-8a23-2b67882bdd0b" containerName="ovnkube-controller" Mar 13 14:19:42 crc kubenswrapper[4907]: E0313 14:19:42.678970 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1006376d-4033-454e-8a23-2b67882bdd0b" containerName="ovnkube-controller" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.678979 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="1006376d-4033-454e-8a23-2b67882bdd0b" containerName="ovnkube-controller" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.679111 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="1006376d-4033-454e-8a23-2b67882bdd0b" containerName="kube-rbac-proxy-ovn-metrics" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.679124 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="1006376d-4033-454e-8a23-2b67882bdd0b" containerName="ovnkube-controller" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.679136 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="03fa7dad-f13d-4e5c-8b9c-60b2ef2d272a" containerName="oc" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.679148 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="1006376d-4033-454e-8a23-2b67882bdd0b" containerName="kube-rbac-proxy-node" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.679157 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="1006376d-4033-454e-8a23-2b67882bdd0b" containerName="northd" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.679167 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="1006376d-4033-454e-8a23-2b67882bdd0b" containerName="ovnkube-controller" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.679178 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="1006376d-4033-454e-8a23-2b67882bdd0b" containerName="nbdb" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.679187 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="1006376d-4033-454e-8a23-2b67882bdd0b" containerName="ovnkube-controller" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.679196 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="1006376d-4033-454e-8a23-2b67882bdd0b" containerName="ovnkube-controller" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.679206 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="1006376d-4033-454e-8a23-2b67882bdd0b" containerName="sbdb" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.679217 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="1006376d-4033-454e-8a23-2b67882bdd0b" containerName="ovnkube-controller" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.679227 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="1006376d-4033-454e-8a23-2b67882bdd0b" containerName="ovn-controller" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.679237 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="1006376d-4033-454e-8a23-2b67882bdd0b" containerName="ovn-acl-logging" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.682114 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.760142 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-host-kubelet\") pod \"1006376d-4033-454e-8a23-2b67882bdd0b\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.760238 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/1006376d-4033-454e-8a23-2b67882bdd0b-ovn-node-metrics-cert\") pod \"1006376d-4033-454e-8a23-2b67882bdd0b\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.760273 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-host-slash\") pod \"1006376d-4033-454e-8a23-2b67882bdd0b\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.760252 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "1006376d-4033-454e-8a23-2b67882bdd0b" (UID: "1006376d-4033-454e-8a23-2b67882bdd0b"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.760295 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-var-lib-openvswitch\") pod \"1006376d-4033-454e-8a23-2b67882bdd0b\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.760325 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-systemd-units\") pod \"1006376d-4033-454e-8a23-2b67882bdd0b\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.760367 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-run-systemd\") pod \"1006376d-4033-454e-8a23-2b67882bdd0b\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.760398 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-host-slash" (OuterVolumeSpecName: "host-slash") pod "1006376d-4033-454e-8a23-2b67882bdd0b" (UID: "1006376d-4033-454e-8a23-2b67882bdd0b"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.760408 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "1006376d-4033-454e-8a23-2b67882bdd0b" (UID: "1006376d-4033-454e-8a23-2b67882bdd0b"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.760416 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/1006376d-4033-454e-8a23-2b67882bdd0b-ovnkube-script-lib\") pod \"1006376d-4033-454e-8a23-2b67882bdd0b\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.760527 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-host-cni-netd\") pod \"1006376d-4033-454e-8a23-2b67882bdd0b\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.760556 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-run-openvswitch\") pod \"1006376d-4033-454e-8a23-2b67882bdd0b\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.760612 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-run-ovn\") pod \"1006376d-4033-454e-8a23-2b67882bdd0b\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.760665 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-log-socket\") pod \"1006376d-4033-454e-8a23-2b67882bdd0b\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.760694 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/1006376d-4033-454e-8a23-2b67882bdd0b-ovnkube-config\") pod \"1006376d-4033-454e-8a23-2b67882bdd0b\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.760731 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-host-run-netns\") pod \"1006376d-4033-454e-8a23-2b67882bdd0b\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.760770 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-etc-openvswitch\") pod \"1006376d-4033-454e-8a23-2b67882bdd0b\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.760799 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-host-run-ovn-kubernetes\") pod \"1006376d-4033-454e-8a23-2b67882bdd0b\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.760823 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-host-cni-bin\") pod \"1006376d-4033-454e-8a23-2b67882bdd0b\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.760852 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-node-log\") pod \"1006376d-4033-454e-8a23-2b67882bdd0b\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.760908 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tm9lw\" (UniqueName: \"kubernetes.io/projected/1006376d-4033-454e-8a23-2b67882bdd0b-kube-api-access-tm9lw\") pod \"1006376d-4033-454e-8a23-2b67882bdd0b\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.760929 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-host-var-lib-cni-networks-ovn-kubernetes\") pod \"1006376d-4033-454e-8a23-2b67882bdd0b\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.760960 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/1006376d-4033-454e-8a23-2b67882bdd0b-env-overrides\") pod \"1006376d-4033-454e-8a23-2b67882bdd0b\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.761067 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "1006376d-4033-454e-8a23-2b67882bdd0b" (UID: "1006376d-4033-454e-8a23-2b67882bdd0b"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.761100 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "1006376d-4033-454e-8a23-2b67882bdd0b" (UID: "1006376d-4033-454e-8a23-2b67882bdd0b"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.761103 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "1006376d-4033-454e-8a23-2b67882bdd0b" (UID: "1006376d-4033-454e-8a23-2b67882bdd0b"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.761119 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "1006376d-4033-454e-8a23-2b67882bdd0b" (UID: "1006376d-4033-454e-8a23-2b67882bdd0b"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.761139 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-node-log" (OuterVolumeSpecName: "node-log") pod "1006376d-4033-454e-8a23-2b67882bdd0b" (UID: "1006376d-4033-454e-8a23-2b67882bdd0b"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.761160 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-log-socket" (OuterVolumeSpecName: "log-socket") pod "1006376d-4033-454e-8a23-2b67882bdd0b" (UID: "1006376d-4033-454e-8a23-2b67882bdd0b"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.761179 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "1006376d-4033-454e-8a23-2b67882bdd0b" (UID: "1006376d-4033-454e-8a23-2b67882bdd0b"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.761168 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "1006376d-4033-454e-8a23-2b67882bdd0b" (UID: "1006376d-4033-454e-8a23-2b67882bdd0b"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.761221 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1006376d-4033-454e-8a23-2b67882bdd0b-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "1006376d-4033-454e-8a23-2b67882bdd0b" (UID: "1006376d-4033-454e-8a23-2b67882bdd0b"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.761090 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "1006376d-4033-454e-8a23-2b67882bdd0b" (UID: "1006376d-4033-454e-8a23-2b67882bdd0b"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.761245 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "1006376d-4033-454e-8a23-2b67882bdd0b" (UID: "1006376d-4033-454e-8a23-2b67882bdd0b"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.761200 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "1006376d-4033-454e-8a23-2b67882bdd0b" (UID: "1006376d-4033-454e-8a23-2b67882bdd0b"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.761524 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1006376d-4033-454e-8a23-2b67882bdd0b-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "1006376d-4033-454e-8a23-2b67882bdd0b" (UID: "1006376d-4033-454e-8a23-2b67882bdd0b"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.761536 4907 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-host-cni-netd\") on node \"crc\" DevicePath \"\"" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.761548 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1006376d-4033-454e-8a23-2b67882bdd0b-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "1006376d-4033-454e-8a23-2b67882bdd0b" (UID: "1006376d-4033-454e-8a23-2b67882bdd0b"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.761567 4907 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-run-openvswitch\") on node \"crc\" DevicePath \"\"" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.761646 4907 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-run-ovn\") on node \"crc\" DevicePath \"\"" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.761667 4907 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-log-socket\") on node \"crc\" DevicePath \"\"" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.761688 4907 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-host-run-netns\") on node \"crc\" DevicePath \"\"" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.761707 4907 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.761729 4907 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.761752 4907 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-host-cni-bin\") on node \"crc\" DevicePath \"\"" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.761770 4907 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-node-log\") on node \"crc\" DevicePath \"\"" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.761789 4907 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.761809 4907 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-host-kubelet\") on node \"crc\" DevicePath \"\"" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.761830 4907 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-host-slash\") on node \"crc\" DevicePath \"\"" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.761849 4907 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.761867 4907 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-systemd-units\") on node \"crc\" DevicePath \"\"" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.761921 4907 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/1006376d-4033-454e-8a23-2b67882bdd0b-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.766090 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1006376d-4033-454e-8a23-2b67882bdd0b-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "1006376d-4033-454e-8a23-2b67882bdd0b" (UID: "1006376d-4033-454e-8a23-2b67882bdd0b"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.766820 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1006376d-4033-454e-8a23-2b67882bdd0b-kube-api-access-tm9lw" (OuterVolumeSpecName: "kube-api-access-tm9lw") pod "1006376d-4033-454e-8a23-2b67882bdd0b" (UID: "1006376d-4033-454e-8a23-2b67882bdd0b"). InnerVolumeSpecName "kube-api-access-tm9lw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.773326 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "1006376d-4033-454e-8a23-2b67882bdd0b" (UID: "1006376d-4033-454e-8a23-2b67882bdd0b"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.863756 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/42f6fac8-4aa0-4cd7-a1a1-c80ced66c462-host-run-ovn-kubernetes\") pod \"ovnkube-node-8pphf\" (UID: \"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462\") " pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.863846 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jbxsh\" (UniqueName: \"kubernetes.io/projected/42f6fac8-4aa0-4cd7-a1a1-c80ced66c462-kube-api-access-jbxsh\") pod \"ovnkube-node-8pphf\" (UID: \"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462\") " pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.864647 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/42f6fac8-4aa0-4cd7-a1a1-c80ced66c462-host-run-netns\") pod \"ovnkube-node-8pphf\" (UID: \"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462\") " pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.864688 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/42f6fac8-4aa0-4cd7-a1a1-c80ced66c462-host-cni-netd\") pod \"ovnkube-node-8pphf\" (UID: \"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462\") " pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.864720 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/42f6fac8-4aa0-4cd7-a1a1-c80ced66c462-host-kubelet\") pod \"ovnkube-node-8pphf\" (UID: \"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462\") " pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.864876 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/42f6fac8-4aa0-4cd7-a1a1-c80ced66c462-env-overrides\") pod \"ovnkube-node-8pphf\" (UID: \"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462\") " pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.864993 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/42f6fac8-4aa0-4cd7-a1a1-c80ced66c462-log-socket\") pod \"ovnkube-node-8pphf\" (UID: \"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462\") " pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.865022 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/42f6fac8-4aa0-4cd7-a1a1-c80ced66c462-ovnkube-script-lib\") pod \"ovnkube-node-8pphf\" (UID: \"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462\") " pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.865096 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/42f6fac8-4aa0-4cd7-a1a1-c80ced66c462-run-openvswitch\") pod \"ovnkube-node-8pphf\" (UID: \"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462\") " pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.865169 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/42f6fac8-4aa0-4cd7-a1a1-c80ced66c462-ovnkube-config\") pod \"ovnkube-node-8pphf\" (UID: \"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462\") " pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.865214 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/42f6fac8-4aa0-4cd7-a1a1-c80ced66c462-var-lib-openvswitch\") pod \"ovnkube-node-8pphf\" (UID: \"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462\") " pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.865268 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/42f6fac8-4aa0-4cd7-a1a1-c80ced66c462-etc-openvswitch\") pod \"ovnkube-node-8pphf\" (UID: \"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462\") " pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.865294 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/42f6fac8-4aa0-4cd7-a1a1-c80ced66c462-node-log\") pod \"ovnkube-node-8pphf\" (UID: \"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462\") " pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.865352 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/42f6fac8-4aa0-4cd7-a1a1-c80ced66c462-run-ovn\") pod \"ovnkube-node-8pphf\" (UID: \"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462\") " pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.865408 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/42f6fac8-4aa0-4cd7-a1a1-c80ced66c462-run-systemd\") pod \"ovnkube-node-8pphf\" (UID: \"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462\") " pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.865439 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/42f6fac8-4aa0-4cd7-a1a1-c80ced66c462-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-8pphf\" (UID: \"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462\") " pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.865465 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/42f6fac8-4aa0-4cd7-a1a1-c80ced66c462-host-cni-bin\") pod \"ovnkube-node-8pphf\" (UID: \"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462\") " pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.865514 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/42f6fac8-4aa0-4cd7-a1a1-c80ced66c462-systemd-units\") pod \"ovnkube-node-8pphf\" (UID: \"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462\") " pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.865529 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/42f6fac8-4aa0-4cd7-a1a1-c80ced66c462-host-slash\") pod \"ovnkube-node-8pphf\" (UID: \"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462\") " pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.865566 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/42f6fac8-4aa0-4cd7-a1a1-c80ced66c462-ovn-node-metrics-cert\") pod \"ovnkube-node-8pphf\" (UID: \"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462\") " pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.865642 4907 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/1006376d-4033-454e-8a23-2b67882bdd0b-ovnkube-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.865679 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tm9lw\" (UniqueName: \"kubernetes.io/projected/1006376d-4033-454e-8a23-2b67882bdd0b-kube-api-access-tm9lw\") on node \"crc\" DevicePath \"\"" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.865689 4907 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/1006376d-4033-454e-8a23-2b67882bdd0b-env-overrides\") on node \"crc\" DevicePath \"\"" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.865698 4907 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/1006376d-4033-454e-8a23-2b67882bdd0b-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.865708 4907 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-run-systemd\") on node \"crc\" DevicePath \"\"" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.899110 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qxlx5_1006376d-4033-454e-8a23-2b67882bdd0b/ovnkube-controller/3.log" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.901734 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qxlx5_1006376d-4033-454e-8a23-2b67882bdd0b/ovn-acl-logging/0.log" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.902264 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qxlx5_1006376d-4033-454e-8a23-2b67882bdd0b/ovn-controller/0.log" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.902754 4907 generic.go:334] "Generic (PLEG): container finished" podID="1006376d-4033-454e-8a23-2b67882bdd0b" containerID="a3d2a347b2f85682f2a54c30e63b7e9c88ffe402d6d71fa7ce6656786f93b886" exitCode=0 Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.902781 4907 generic.go:334] "Generic (PLEG): container finished" podID="1006376d-4033-454e-8a23-2b67882bdd0b" containerID="4d60b5b774f8dd4cee16b9784f80d124640199a85331a2940772029be4884751" exitCode=0 Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.902789 4907 generic.go:334] "Generic (PLEG): container finished" podID="1006376d-4033-454e-8a23-2b67882bdd0b" containerID="a0ee0049aa4a6d49a6bcf7d347183d9cc8b85171f988662803a005e0cdd97395" exitCode=0 Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.902798 4907 generic.go:334] "Generic (PLEG): container finished" podID="1006376d-4033-454e-8a23-2b67882bdd0b" containerID="0fff87d8ea572615eb1329164f9e536e6ee45cd00a9e71be0aa8a9b3ed2190e2" exitCode=0 Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.902805 4907 generic.go:334] "Generic (PLEG): container finished" podID="1006376d-4033-454e-8a23-2b67882bdd0b" containerID="ccef92e41501fab91189581d7e8875515c02d637426b329d2fb80a356c75b637" exitCode=0 Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.902811 4907 generic.go:334] "Generic (PLEG): container finished" podID="1006376d-4033-454e-8a23-2b67882bdd0b" containerID="bf6a3dbaa8dda61f461a906538824c8946d36dc07d4fb95bf4d8420ad0105600" exitCode=0 Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.902818 4907 generic.go:334] "Generic (PLEG): container finished" podID="1006376d-4033-454e-8a23-2b67882bdd0b" containerID="3c165053fd9933444b372ff40efe33c0e193a7f39e6f531a4f370e971eb41e6d" exitCode=143 Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.902826 4907 generic.go:334] "Generic (PLEG): container finished" podID="1006376d-4033-454e-8a23-2b67882bdd0b" containerID="fea41c11bebd2d9d622ed4b8bc50e03a106dadea6ac102cd60512ea2a1f8b8b0" exitCode=143 Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.902832 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" event={"ID":"1006376d-4033-454e-8a23-2b67882bdd0b","Type":"ContainerDied","Data":"a3d2a347b2f85682f2a54c30e63b7e9c88ffe402d6d71fa7ce6656786f93b886"} Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.902856 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.902873 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" event={"ID":"1006376d-4033-454e-8a23-2b67882bdd0b","Type":"ContainerDied","Data":"4d60b5b774f8dd4cee16b9784f80d124640199a85331a2940772029be4884751"} Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.902897 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" event={"ID":"1006376d-4033-454e-8a23-2b67882bdd0b","Type":"ContainerDied","Data":"a0ee0049aa4a6d49a6bcf7d347183d9cc8b85171f988662803a005e0cdd97395"} Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.902909 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" event={"ID":"1006376d-4033-454e-8a23-2b67882bdd0b","Type":"ContainerDied","Data":"0fff87d8ea572615eb1329164f9e536e6ee45cd00a9e71be0aa8a9b3ed2190e2"} Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.902924 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" event={"ID":"1006376d-4033-454e-8a23-2b67882bdd0b","Type":"ContainerDied","Data":"ccef92e41501fab91189581d7e8875515c02d637426b329d2fb80a356c75b637"} Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.902934 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" event={"ID":"1006376d-4033-454e-8a23-2b67882bdd0b","Type":"ContainerDied","Data":"bf6a3dbaa8dda61f461a906538824c8946d36dc07d4fb95bf4d8420ad0105600"} Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.902945 4907 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9155434e2d8f0daa4504758fb72d6f498da107a64a92a0883ac687ed2af9caee"} Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.902954 4907 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4d60b5b774f8dd4cee16b9784f80d124640199a85331a2940772029be4884751"} Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.902959 4907 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a0ee0049aa4a6d49a6bcf7d347183d9cc8b85171f988662803a005e0cdd97395"} Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.902964 4907 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0fff87d8ea572615eb1329164f9e536e6ee45cd00a9e71be0aa8a9b3ed2190e2"} Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.902970 4907 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ccef92e41501fab91189581d7e8875515c02d637426b329d2fb80a356c75b637"} Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.902975 4907 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"bf6a3dbaa8dda61f461a906538824c8946d36dc07d4fb95bf4d8420ad0105600"} Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.902980 4907 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3c165053fd9933444b372ff40efe33c0e193a7f39e6f531a4f370e971eb41e6d"} Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.902986 4907 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"fea41c11bebd2d9d622ed4b8bc50e03a106dadea6ac102cd60512ea2a1f8b8b0"} Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.902991 4907 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279"} Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.902998 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" event={"ID":"1006376d-4033-454e-8a23-2b67882bdd0b","Type":"ContainerDied","Data":"3c165053fd9933444b372ff40efe33c0e193a7f39e6f531a4f370e971eb41e6d"} Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.903004 4907 scope.go:117] "RemoveContainer" containerID="a3d2a347b2f85682f2a54c30e63b7e9c88ffe402d6d71fa7ce6656786f93b886" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.903009 4907 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a3d2a347b2f85682f2a54c30e63b7e9c88ffe402d6d71fa7ce6656786f93b886"} Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.903105 4907 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9155434e2d8f0daa4504758fb72d6f498da107a64a92a0883ac687ed2af9caee"} Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.903118 4907 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4d60b5b774f8dd4cee16b9784f80d124640199a85331a2940772029be4884751"} Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.903125 4907 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a0ee0049aa4a6d49a6bcf7d347183d9cc8b85171f988662803a005e0cdd97395"} Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.903131 4907 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0fff87d8ea572615eb1329164f9e536e6ee45cd00a9e71be0aa8a9b3ed2190e2"} Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.903136 4907 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ccef92e41501fab91189581d7e8875515c02d637426b329d2fb80a356c75b637"} Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.903141 4907 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"bf6a3dbaa8dda61f461a906538824c8946d36dc07d4fb95bf4d8420ad0105600"} Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.903147 4907 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3c165053fd9933444b372ff40efe33c0e193a7f39e6f531a4f370e971eb41e6d"} Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.903153 4907 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"fea41c11bebd2d9d622ed4b8bc50e03a106dadea6ac102cd60512ea2a1f8b8b0"} Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.903158 4907 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279"} Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.903176 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" event={"ID":"1006376d-4033-454e-8a23-2b67882bdd0b","Type":"ContainerDied","Data":"fea41c11bebd2d9d622ed4b8bc50e03a106dadea6ac102cd60512ea2a1f8b8b0"} Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.903196 4907 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a3d2a347b2f85682f2a54c30e63b7e9c88ffe402d6d71fa7ce6656786f93b886"} Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.903202 4907 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9155434e2d8f0daa4504758fb72d6f498da107a64a92a0883ac687ed2af9caee"} Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.903209 4907 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4d60b5b774f8dd4cee16b9784f80d124640199a85331a2940772029be4884751"} Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.903214 4907 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a0ee0049aa4a6d49a6bcf7d347183d9cc8b85171f988662803a005e0cdd97395"} Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.903220 4907 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0fff87d8ea572615eb1329164f9e536e6ee45cd00a9e71be0aa8a9b3ed2190e2"} Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.903225 4907 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ccef92e41501fab91189581d7e8875515c02d637426b329d2fb80a356c75b637"} Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.903230 4907 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"bf6a3dbaa8dda61f461a906538824c8946d36dc07d4fb95bf4d8420ad0105600"} Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.903235 4907 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3c165053fd9933444b372ff40efe33c0e193a7f39e6f531a4f370e971eb41e6d"} Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.903240 4907 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"fea41c11bebd2d9d622ed4b8bc50e03a106dadea6ac102cd60512ea2a1f8b8b0"} Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.903245 4907 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279"} Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.903253 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" event={"ID":"1006376d-4033-454e-8a23-2b67882bdd0b","Type":"ContainerDied","Data":"4a8bfe852be3826bb96e2a956b4a0cd7eaff7036e060420d181a2514210039c0"} Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.903261 4907 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a3d2a347b2f85682f2a54c30e63b7e9c88ffe402d6d71fa7ce6656786f93b886"} Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.903269 4907 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9155434e2d8f0daa4504758fb72d6f498da107a64a92a0883ac687ed2af9caee"} Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.903274 4907 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4d60b5b774f8dd4cee16b9784f80d124640199a85331a2940772029be4884751"} Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.903280 4907 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a0ee0049aa4a6d49a6bcf7d347183d9cc8b85171f988662803a005e0cdd97395"} Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.903286 4907 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0fff87d8ea572615eb1329164f9e536e6ee45cd00a9e71be0aa8a9b3ed2190e2"} Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.903291 4907 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ccef92e41501fab91189581d7e8875515c02d637426b329d2fb80a356c75b637"} Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.903296 4907 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"bf6a3dbaa8dda61f461a906538824c8946d36dc07d4fb95bf4d8420ad0105600"} Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.903302 4907 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3c165053fd9933444b372ff40efe33c0e193a7f39e6f531a4f370e971eb41e6d"} Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.903308 4907 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"fea41c11bebd2d9d622ed4b8bc50e03a106dadea6ac102cd60512ea2a1f8b8b0"} Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.903314 4907 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279"} Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.904726 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-xtwf7_64ebd338-0158-44dc-90ab-e7c285e87762/kube-multus/2.log" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.905479 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-xtwf7_64ebd338-0158-44dc-90ab-e7c285e87762/kube-multus/1.log" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.905506 4907 generic.go:334] "Generic (PLEG): container finished" podID="64ebd338-0158-44dc-90ab-e7c285e87762" containerID="8788f4b58c32d23e53448793bceb517c39726e72de35c52e959fc15e062ae173" exitCode=2 Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.905526 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-xtwf7" event={"ID":"64ebd338-0158-44dc-90ab-e7c285e87762","Type":"ContainerDied","Data":"8788f4b58c32d23e53448793bceb517c39726e72de35c52e959fc15e062ae173"} Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.905546 4907 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ef632bb17c8f529b35d0882c8d4ba3dc213b0308b46a4ade8426b9c08e64bb2b"} Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.906030 4907 scope.go:117] "RemoveContainer" containerID="8788f4b58c32d23e53448793bceb517c39726e72de35c52e959fc15e062ae173" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.935473 4907 scope.go:117] "RemoveContainer" containerID="9155434e2d8f0daa4504758fb72d6f498da107a64a92a0883ac687ed2af9caee" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.957239 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-qxlx5"] Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.962441 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-qxlx5"] Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.967207 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/42f6fac8-4aa0-4cd7-a1a1-c80ced66c462-node-log\") pod \"ovnkube-node-8pphf\" (UID: \"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462\") " pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.967258 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/42f6fac8-4aa0-4cd7-a1a1-c80ced66c462-run-ovn\") pod \"ovnkube-node-8pphf\" (UID: \"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462\") " pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.967296 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/42f6fac8-4aa0-4cd7-a1a1-c80ced66c462-run-systemd\") pod \"ovnkube-node-8pphf\" (UID: \"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462\") " pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.967338 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/42f6fac8-4aa0-4cd7-a1a1-c80ced66c462-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-8pphf\" (UID: \"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462\") " pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.967360 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/42f6fac8-4aa0-4cd7-a1a1-c80ced66c462-node-log\") pod \"ovnkube-node-8pphf\" (UID: \"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462\") " pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.967389 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/42f6fac8-4aa0-4cd7-a1a1-c80ced66c462-run-ovn\") pod \"ovnkube-node-8pphf\" (UID: \"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462\") " pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.967375 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/42f6fac8-4aa0-4cd7-a1a1-c80ced66c462-host-cni-bin\") pod \"ovnkube-node-8pphf\" (UID: \"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462\") " pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.967432 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/42f6fac8-4aa0-4cd7-a1a1-c80ced66c462-host-cni-bin\") pod \"ovnkube-node-8pphf\" (UID: \"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462\") " pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.967439 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/42f6fac8-4aa0-4cd7-a1a1-c80ced66c462-run-systemd\") pod \"ovnkube-node-8pphf\" (UID: \"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462\") " pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.967792 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/42f6fac8-4aa0-4cd7-a1a1-c80ced66c462-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-8pphf\" (UID: \"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462\") " pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.967898 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/42f6fac8-4aa0-4cd7-a1a1-c80ced66c462-systemd-units\") pod \"ovnkube-node-8pphf\" (UID: \"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462\") " pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.967922 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/42f6fac8-4aa0-4cd7-a1a1-c80ced66c462-host-slash\") pod \"ovnkube-node-8pphf\" (UID: \"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462\") " pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.967959 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/42f6fac8-4aa0-4cd7-a1a1-c80ced66c462-ovn-node-metrics-cert\") pod \"ovnkube-node-8pphf\" (UID: \"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462\") " pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.967982 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/42f6fac8-4aa0-4cd7-a1a1-c80ced66c462-host-run-ovn-kubernetes\") pod \"ovnkube-node-8pphf\" (UID: \"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462\") " pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.968003 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jbxsh\" (UniqueName: \"kubernetes.io/projected/42f6fac8-4aa0-4cd7-a1a1-c80ced66c462-kube-api-access-jbxsh\") pod \"ovnkube-node-8pphf\" (UID: \"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462\") " pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.968041 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/42f6fac8-4aa0-4cd7-a1a1-c80ced66c462-host-run-netns\") pod \"ovnkube-node-8pphf\" (UID: \"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462\") " pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.968056 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/42f6fac8-4aa0-4cd7-a1a1-c80ced66c462-host-cni-netd\") pod \"ovnkube-node-8pphf\" (UID: \"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462\") " pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.968073 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/42f6fac8-4aa0-4cd7-a1a1-c80ced66c462-host-kubelet\") pod \"ovnkube-node-8pphf\" (UID: \"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462\") " pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.968092 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/42f6fac8-4aa0-4cd7-a1a1-c80ced66c462-env-overrides\") pod \"ovnkube-node-8pphf\" (UID: \"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462\") " pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.968113 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/42f6fac8-4aa0-4cd7-a1a1-c80ced66c462-log-socket\") pod \"ovnkube-node-8pphf\" (UID: \"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462\") " pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.968128 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/42f6fac8-4aa0-4cd7-a1a1-c80ced66c462-ovnkube-script-lib\") pod \"ovnkube-node-8pphf\" (UID: \"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462\") " pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.968160 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/42f6fac8-4aa0-4cd7-a1a1-c80ced66c462-run-openvswitch\") pod \"ovnkube-node-8pphf\" (UID: \"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462\") " pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.968181 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/42f6fac8-4aa0-4cd7-a1a1-c80ced66c462-ovnkube-config\") pod \"ovnkube-node-8pphf\" (UID: \"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462\") " pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.968231 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/42f6fac8-4aa0-4cd7-a1a1-c80ced66c462-var-lib-openvswitch\") pod \"ovnkube-node-8pphf\" (UID: \"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462\") " pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.968263 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/42f6fac8-4aa0-4cd7-a1a1-c80ced66c462-etc-openvswitch\") pod \"ovnkube-node-8pphf\" (UID: \"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462\") " pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.968346 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/42f6fac8-4aa0-4cd7-a1a1-c80ced66c462-etc-openvswitch\") pod \"ovnkube-node-8pphf\" (UID: \"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462\") " pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.968370 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/42f6fac8-4aa0-4cd7-a1a1-c80ced66c462-systemd-units\") pod \"ovnkube-node-8pphf\" (UID: \"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462\") " pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.968393 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/42f6fac8-4aa0-4cd7-a1a1-c80ced66c462-host-slash\") pod \"ovnkube-node-8pphf\" (UID: \"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462\") " pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.968859 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/42f6fac8-4aa0-4cd7-a1a1-c80ced66c462-run-openvswitch\") pod \"ovnkube-node-8pphf\" (UID: \"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462\") " pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.968874 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/42f6fac8-4aa0-4cd7-a1a1-c80ced66c462-host-run-ovn-kubernetes\") pod \"ovnkube-node-8pphf\" (UID: \"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462\") " pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.968913 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/42f6fac8-4aa0-4cd7-a1a1-c80ced66c462-host-cni-netd\") pod \"ovnkube-node-8pphf\" (UID: \"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462\") " pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.968932 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/42f6fac8-4aa0-4cd7-a1a1-c80ced66c462-host-run-netns\") pod \"ovnkube-node-8pphf\" (UID: \"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462\") " pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.968965 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/42f6fac8-4aa0-4cd7-a1a1-c80ced66c462-log-socket\") pod \"ovnkube-node-8pphf\" (UID: \"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462\") " pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.969148 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/42f6fac8-4aa0-4cd7-a1a1-c80ced66c462-host-kubelet\") pod \"ovnkube-node-8pphf\" (UID: \"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462\") " pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.969239 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/42f6fac8-4aa0-4cd7-a1a1-c80ced66c462-var-lib-openvswitch\") pod \"ovnkube-node-8pphf\" (UID: \"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462\") " pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.969562 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/42f6fac8-4aa0-4cd7-a1a1-c80ced66c462-ovnkube-script-lib\") pod \"ovnkube-node-8pphf\" (UID: \"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462\") " pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.970002 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/42f6fac8-4aa0-4cd7-a1a1-c80ced66c462-ovnkube-config\") pod \"ovnkube-node-8pphf\" (UID: \"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462\") " pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.974353 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/42f6fac8-4aa0-4cd7-a1a1-c80ced66c462-env-overrides\") pod \"ovnkube-node-8pphf\" (UID: \"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462\") " pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.976157 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/42f6fac8-4aa0-4cd7-a1a1-c80ced66c462-ovn-node-metrics-cert\") pod \"ovnkube-node-8pphf\" (UID: \"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462\") " pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.980996 4907 scope.go:117] "RemoveContainer" containerID="4d60b5b774f8dd4cee16b9784f80d124640199a85331a2940772029be4884751" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.993645 4907 scope.go:117] "RemoveContainer" containerID="a0ee0049aa4a6d49a6bcf7d347183d9cc8b85171f988662803a005e0cdd97395" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.994280 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jbxsh\" (UniqueName: \"kubernetes.io/projected/42f6fac8-4aa0-4cd7-a1a1-c80ced66c462-kube-api-access-jbxsh\") pod \"ovnkube-node-8pphf\" (UID: \"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462\") " pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.006304 4907 scope.go:117] "RemoveContainer" containerID="0fff87d8ea572615eb1329164f9e536e6ee45cd00a9e71be0aa8a9b3ed2190e2" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.019797 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.021586 4907 scope.go:117] "RemoveContainer" containerID="ccef92e41501fab91189581d7e8875515c02d637426b329d2fb80a356c75b637" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.046185 4907 scope.go:117] "RemoveContainer" containerID="bf6a3dbaa8dda61f461a906538824c8946d36dc07d4fb95bf4d8420ad0105600" Mar 13 14:19:43 crc kubenswrapper[4907]: W0313 14:19:43.050469 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod42f6fac8_4aa0_4cd7_a1a1_c80ced66c462.slice/crio-b113dbc4695af8ad071eea34316d474e13189383fd3ee685adb103d09acf57d4 WatchSource:0}: Error finding container b113dbc4695af8ad071eea34316d474e13189383fd3ee685adb103d09acf57d4: Status 404 returned error can't find the container with id b113dbc4695af8ad071eea34316d474e13189383fd3ee685adb103d09acf57d4 Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.060761 4907 scope.go:117] "RemoveContainer" containerID="3c165053fd9933444b372ff40efe33c0e193a7f39e6f531a4f370e971eb41e6d" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.073958 4907 scope.go:117] "RemoveContainer" containerID="fea41c11bebd2d9d622ed4b8bc50e03a106dadea6ac102cd60512ea2a1f8b8b0" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.105059 4907 scope.go:117] "RemoveContainer" containerID="858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.131285 4907 scope.go:117] "RemoveContainer" containerID="a3d2a347b2f85682f2a54c30e63b7e9c88ffe402d6d71fa7ce6656786f93b886" Mar 13 14:19:43 crc kubenswrapper[4907]: E0313 14:19:43.131806 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a3d2a347b2f85682f2a54c30e63b7e9c88ffe402d6d71fa7ce6656786f93b886\": container with ID starting with a3d2a347b2f85682f2a54c30e63b7e9c88ffe402d6d71fa7ce6656786f93b886 not found: ID does not exist" containerID="a3d2a347b2f85682f2a54c30e63b7e9c88ffe402d6d71fa7ce6656786f93b886" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.131845 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a3d2a347b2f85682f2a54c30e63b7e9c88ffe402d6d71fa7ce6656786f93b886"} err="failed to get container status \"a3d2a347b2f85682f2a54c30e63b7e9c88ffe402d6d71fa7ce6656786f93b886\": rpc error: code = NotFound desc = could not find container \"a3d2a347b2f85682f2a54c30e63b7e9c88ffe402d6d71fa7ce6656786f93b886\": container with ID starting with a3d2a347b2f85682f2a54c30e63b7e9c88ffe402d6d71fa7ce6656786f93b886 not found: ID does not exist" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.131873 4907 scope.go:117] "RemoveContainer" containerID="9155434e2d8f0daa4504758fb72d6f498da107a64a92a0883ac687ed2af9caee" Mar 13 14:19:43 crc kubenswrapper[4907]: E0313 14:19:43.132513 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9155434e2d8f0daa4504758fb72d6f498da107a64a92a0883ac687ed2af9caee\": container with ID starting with 9155434e2d8f0daa4504758fb72d6f498da107a64a92a0883ac687ed2af9caee not found: ID does not exist" containerID="9155434e2d8f0daa4504758fb72d6f498da107a64a92a0883ac687ed2af9caee" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.132540 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9155434e2d8f0daa4504758fb72d6f498da107a64a92a0883ac687ed2af9caee"} err="failed to get container status \"9155434e2d8f0daa4504758fb72d6f498da107a64a92a0883ac687ed2af9caee\": rpc error: code = NotFound desc = could not find container \"9155434e2d8f0daa4504758fb72d6f498da107a64a92a0883ac687ed2af9caee\": container with ID starting with 9155434e2d8f0daa4504758fb72d6f498da107a64a92a0883ac687ed2af9caee not found: ID does not exist" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.132561 4907 scope.go:117] "RemoveContainer" containerID="4d60b5b774f8dd4cee16b9784f80d124640199a85331a2940772029be4884751" Mar 13 14:19:43 crc kubenswrapper[4907]: E0313 14:19:43.132919 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4d60b5b774f8dd4cee16b9784f80d124640199a85331a2940772029be4884751\": container with ID starting with 4d60b5b774f8dd4cee16b9784f80d124640199a85331a2940772029be4884751 not found: ID does not exist" containerID="4d60b5b774f8dd4cee16b9784f80d124640199a85331a2940772029be4884751" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.132948 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4d60b5b774f8dd4cee16b9784f80d124640199a85331a2940772029be4884751"} err="failed to get container status \"4d60b5b774f8dd4cee16b9784f80d124640199a85331a2940772029be4884751\": rpc error: code = NotFound desc = could not find container \"4d60b5b774f8dd4cee16b9784f80d124640199a85331a2940772029be4884751\": container with ID starting with 4d60b5b774f8dd4cee16b9784f80d124640199a85331a2940772029be4884751 not found: ID does not exist" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.132972 4907 scope.go:117] "RemoveContainer" containerID="a0ee0049aa4a6d49a6bcf7d347183d9cc8b85171f988662803a005e0cdd97395" Mar 13 14:19:43 crc kubenswrapper[4907]: E0313 14:19:43.133270 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a0ee0049aa4a6d49a6bcf7d347183d9cc8b85171f988662803a005e0cdd97395\": container with ID starting with a0ee0049aa4a6d49a6bcf7d347183d9cc8b85171f988662803a005e0cdd97395 not found: ID does not exist" containerID="a0ee0049aa4a6d49a6bcf7d347183d9cc8b85171f988662803a005e0cdd97395" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.133298 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a0ee0049aa4a6d49a6bcf7d347183d9cc8b85171f988662803a005e0cdd97395"} err="failed to get container status \"a0ee0049aa4a6d49a6bcf7d347183d9cc8b85171f988662803a005e0cdd97395\": rpc error: code = NotFound desc = could not find container \"a0ee0049aa4a6d49a6bcf7d347183d9cc8b85171f988662803a005e0cdd97395\": container with ID starting with a0ee0049aa4a6d49a6bcf7d347183d9cc8b85171f988662803a005e0cdd97395 not found: ID does not exist" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.133315 4907 scope.go:117] "RemoveContainer" containerID="0fff87d8ea572615eb1329164f9e536e6ee45cd00a9e71be0aa8a9b3ed2190e2" Mar 13 14:19:43 crc kubenswrapper[4907]: E0313 14:19:43.133519 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0fff87d8ea572615eb1329164f9e536e6ee45cd00a9e71be0aa8a9b3ed2190e2\": container with ID starting with 0fff87d8ea572615eb1329164f9e536e6ee45cd00a9e71be0aa8a9b3ed2190e2 not found: ID does not exist" containerID="0fff87d8ea572615eb1329164f9e536e6ee45cd00a9e71be0aa8a9b3ed2190e2" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.133547 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0fff87d8ea572615eb1329164f9e536e6ee45cd00a9e71be0aa8a9b3ed2190e2"} err="failed to get container status \"0fff87d8ea572615eb1329164f9e536e6ee45cd00a9e71be0aa8a9b3ed2190e2\": rpc error: code = NotFound desc = could not find container \"0fff87d8ea572615eb1329164f9e536e6ee45cd00a9e71be0aa8a9b3ed2190e2\": container with ID starting with 0fff87d8ea572615eb1329164f9e536e6ee45cd00a9e71be0aa8a9b3ed2190e2 not found: ID does not exist" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.133565 4907 scope.go:117] "RemoveContainer" containerID="ccef92e41501fab91189581d7e8875515c02d637426b329d2fb80a356c75b637" Mar 13 14:19:43 crc kubenswrapper[4907]: E0313 14:19:43.133897 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ccef92e41501fab91189581d7e8875515c02d637426b329d2fb80a356c75b637\": container with ID starting with ccef92e41501fab91189581d7e8875515c02d637426b329d2fb80a356c75b637 not found: ID does not exist" containerID="ccef92e41501fab91189581d7e8875515c02d637426b329d2fb80a356c75b637" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.133926 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ccef92e41501fab91189581d7e8875515c02d637426b329d2fb80a356c75b637"} err="failed to get container status \"ccef92e41501fab91189581d7e8875515c02d637426b329d2fb80a356c75b637\": rpc error: code = NotFound desc = could not find container \"ccef92e41501fab91189581d7e8875515c02d637426b329d2fb80a356c75b637\": container with ID starting with ccef92e41501fab91189581d7e8875515c02d637426b329d2fb80a356c75b637 not found: ID does not exist" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.133945 4907 scope.go:117] "RemoveContainer" containerID="bf6a3dbaa8dda61f461a906538824c8946d36dc07d4fb95bf4d8420ad0105600" Mar 13 14:19:43 crc kubenswrapper[4907]: E0313 14:19:43.134394 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bf6a3dbaa8dda61f461a906538824c8946d36dc07d4fb95bf4d8420ad0105600\": container with ID starting with bf6a3dbaa8dda61f461a906538824c8946d36dc07d4fb95bf4d8420ad0105600 not found: ID does not exist" containerID="bf6a3dbaa8dda61f461a906538824c8946d36dc07d4fb95bf4d8420ad0105600" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.134421 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bf6a3dbaa8dda61f461a906538824c8946d36dc07d4fb95bf4d8420ad0105600"} err="failed to get container status \"bf6a3dbaa8dda61f461a906538824c8946d36dc07d4fb95bf4d8420ad0105600\": rpc error: code = NotFound desc = could not find container \"bf6a3dbaa8dda61f461a906538824c8946d36dc07d4fb95bf4d8420ad0105600\": container with ID starting with bf6a3dbaa8dda61f461a906538824c8946d36dc07d4fb95bf4d8420ad0105600 not found: ID does not exist" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.134438 4907 scope.go:117] "RemoveContainer" containerID="3c165053fd9933444b372ff40efe33c0e193a7f39e6f531a4f370e971eb41e6d" Mar 13 14:19:43 crc kubenswrapper[4907]: E0313 14:19:43.134723 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3c165053fd9933444b372ff40efe33c0e193a7f39e6f531a4f370e971eb41e6d\": container with ID starting with 3c165053fd9933444b372ff40efe33c0e193a7f39e6f531a4f370e971eb41e6d not found: ID does not exist" containerID="3c165053fd9933444b372ff40efe33c0e193a7f39e6f531a4f370e971eb41e6d" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.134753 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3c165053fd9933444b372ff40efe33c0e193a7f39e6f531a4f370e971eb41e6d"} err="failed to get container status \"3c165053fd9933444b372ff40efe33c0e193a7f39e6f531a4f370e971eb41e6d\": rpc error: code = NotFound desc = could not find container \"3c165053fd9933444b372ff40efe33c0e193a7f39e6f531a4f370e971eb41e6d\": container with ID starting with 3c165053fd9933444b372ff40efe33c0e193a7f39e6f531a4f370e971eb41e6d not found: ID does not exist" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.134770 4907 scope.go:117] "RemoveContainer" containerID="fea41c11bebd2d9d622ed4b8bc50e03a106dadea6ac102cd60512ea2a1f8b8b0" Mar 13 14:19:43 crc kubenswrapper[4907]: E0313 14:19:43.135052 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fea41c11bebd2d9d622ed4b8bc50e03a106dadea6ac102cd60512ea2a1f8b8b0\": container with ID starting with fea41c11bebd2d9d622ed4b8bc50e03a106dadea6ac102cd60512ea2a1f8b8b0 not found: ID does not exist" containerID="fea41c11bebd2d9d622ed4b8bc50e03a106dadea6ac102cd60512ea2a1f8b8b0" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.135084 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fea41c11bebd2d9d622ed4b8bc50e03a106dadea6ac102cd60512ea2a1f8b8b0"} err="failed to get container status \"fea41c11bebd2d9d622ed4b8bc50e03a106dadea6ac102cd60512ea2a1f8b8b0\": rpc error: code = NotFound desc = could not find container \"fea41c11bebd2d9d622ed4b8bc50e03a106dadea6ac102cd60512ea2a1f8b8b0\": container with ID starting with fea41c11bebd2d9d622ed4b8bc50e03a106dadea6ac102cd60512ea2a1f8b8b0 not found: ID does not exist" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.135101 4907 scope.go:117] "RemoveContainer" containerID="858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279" Mar 13 14:19:43 crc kubenswrapper[4907]: E0313 14:19:43.135325 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\": container with ID starting with 858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279 not found: ID does not exist" containerID="858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.135368 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279"} err="failed to get container status \"858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\": rpc error: code = NotFound desc = could not find container \"858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\": container with ID starting with 858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279 not found: ID does not exist" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.135383 4907 scope.go:117] "RemoveContainer" containerID="a3d2a347b2f85682f2a54c30e63b7e9c88ffe402d6d71fa7ce6656786f93b886" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.135665 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a3d2a347b2f85682f2a54c30e63b7e9c88ffe402d6d71fa7ce6656786f93b886"} err="failed to get container status \"a3d2a347b2f85682f2a54c30e63b7e9c88ffe402d6d71fa7ce6656786f93b886\": rpc error: code = NotFound desc = could not find container \"a3d2a347b2f85682f2a54c30e63b7e9c88ffe402d6d71fa7ce6656786f93b886\": container with ID starting with a3d2a347b2f85682f2a54c30e63b7e9c88ffe402d6d71fa7ce6656786f93b886 not found: ID does not exist" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.135690 4907 scope.go:117] "RemoveContainer" containerID="9155434e2d8f0daa4504758fb72d6f498da107a64a92a0883ac687ed2af9caee" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.135931 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9155434e2d8f0daa4504758fb72d6f498da107a64a92a0883ac687ed2af9caee"} err="failed to get container status \"9155434e2d8f0daa4504758fb72d6f498da107a64a92a0883ac687ed2af9caee\": rpc error: code = NotFound desc = could not find container \"9155434e2d8f0daa4504758fb72d6f498da107a64a92a0883ac687ed2af9caee\": container with ID starting with 9155434e2d8f0daa4504758fb72d6f498da107a64a92a0883ac687ed2af9caee not found: ID does not exist" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.135953 4907 scope.go:117] "RemoveContainer" containerID="4d60b5b774f8dd4cee16b9784f80d124640199a85331a2940772029be4884751" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.136160 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4d60b5b774f8dd4cee16b9784f80d124640199a85331a2940772029be4884751"} err="failed to get container status \"4d60b5b774f8dd4cee16b9784f80d124640199a85331a2940772029be4884751\": rpc error: code = NotFound desc = could not find container \"4d60b5b774f8dd4cee16b9784f80d124640199a85331a2940772029be4884751\": container with ID starting with 4d60b5b774f8dd4cee16b9784f80d124640199a85331a2940772029be4884751 not found: ID does not exist" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.136189 4907 scope.go:117] "RemoveContainer" containerID="a0ee0049aa4a6d49a6bcf7d347183d9cc8b85171f988662803a005e0cdd97395" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.136406 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a0ee0049aa4a6d49a6bcf7d347183d9cc8b85171f988662803a005e0cdd97395"} err="failed to get container status \"a0ee0049aa4a6d49a6bcf7d347183d9cc8b85171f988662803a005e0cdd97395\": rpc error: code = NotFound desc = could not find container \"a0ee0049aa4a6d49a6bcf7d347183d9cc8b85171f988662803a005e0cdd97395\": container with ID starting with a0ee0049aa4a6d49a6bcf7d347183d9cc8b85171f988662803a005e0cdd97395 not found: ID does not exist" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.136427 4907 scope.go:117] "RemoveContainer" containerID="0fff87d8ea572615eb1329164f9e536e6ee45cd00a9e71be0aa8a9b3ed2190e2" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.136663 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0fff87d8ea572615eb1329164f9e536e6ee45cd00a9e71be0aa8a9b3ed2190e2"} err="failed to get container status \"0fff87d8ea572615eb1329164f9e536e6ee45cd00a9e71be0aa8a9b3ed2190e2\": rpc error: code = NotFound desc = could not find container \"0fff87d8ea572615eb1329164f9e536e6ee45cd00a9e71be0aa8a9b3ed2190e2\": container with ID starting with 0fff87d8ea572615eb1329164f9e536e6ee45cd00a9e71be0aa8a9b3ed2190e2 not found: ID does not exist" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.136685 4907 scope.go:117] "RemoveContainer" containerID="ccef92e41501fab91189581d7e8875515c02d637426b329d2fb80a356c75b637" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.136863 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ccef92e41501fab91189581d7e8875515c02d637426b329d2fb80a356c75b637"} err="failed to get container status \"ccef92e41501fab91189581d7e8875515c02d637426b329d2fb80a356c75b637\": rpc error: code = NotFound desc = could not find container \"ccef92e41501fab91189581d7e8875515c02d637426b329d2fb80a356c75b637\": container with ID starting with ccef92e41501fab91189581d7e8875515c02d637426b329d2fb80a356c75b637 not found: ID does not exist" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.136902 4907 scope.go:117] "RemoveContainer" containerID="bf6a3dbaa8dda61f461a906538824c8946d36dc07d4fb95bf4d8420ad0105600" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.137122 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bf6a3dbaa8dda61f461a906538824c8946d36dc07d4fb95bf4d8420ad0105600"} err="failed to get container status \"bf6a3dbaa8dda61f461a906538824c8946d36dc07d4fb95bf4d8420ad0105600\": rpc error: code = NotFound desc = could not find container \"bf6a3dbaa8dda61f461a906538824c8946d36dc07d4fb95bf4d8420ad0105600\": container with ID starting with bf6a3dbaa8dda61f461a906538824c8946d36dc07d4fb95bf4d8420ad0105600 not found: ID does not exist" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.137143 4907 scope.go:117] "RemoveContainer" containerID="3c165053fd9933444b372ff40efe33c0e193a7f39e6f531a4f370e971eb41e6d" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.137394 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3c165053fd9933444b372ff40efe33c0e193a7f39e6f531a4f370e971eb41e6d"} err="failed to get container status \"3c165053fd9933444b372ff40efe33c0e193a7f39e6f531a4f370e971eb41e6d\": rpc error: code = NotFound desc = could not find container \"3c165053fd9933444b372ff40efe33c0e193a7f39e6f531a4f370e971eb41e6d\": container with ID starting with 3c165053fd9933444b372ff40efe33c0e193a7f39e6f531a4f370e971eb41e6d not found: ID does not exist" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.137477 4907 scope.go:117] "RemoveContainer" containerID="fea41c11bebd2d9d622ed4b8bc50e03a106dadea6ac102cd60512ea2a1f8b8b0" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.137691 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fea41c11bebd2d9d622ed4b8bc50e03a106dadea6ac102cd60512ea2a1f8b8b0"} err="failed to get container status \"fea41c11bebd2d9d622ed4b8bc50e03a106dadea6ac102cd60512ea2a1f8b8b0\": rpc error: code = NotFound desc = could not find container \"fea41c11bebd2d9d622ed4b8bc50e03a106dadea6ac102cd60512ea2a1f8b8b0\": container with ID starting with fea41c11bebd2d9d622ed4b8bc50e03a106dadea6ac102cd60512ea2a1f8b8b0 not found: ID does not exist" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.137729 4907 scope.go:117] "RemoveContainer" containerID="858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.137976 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279"} err="failed to get container status \"858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\": rpc error: code = NotFound desc = could not find container \"858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\": container with ID starting with 858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279 not found: ID does not exist" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.138000 4907 scope.go:117] "RemoveContainer" containerID="a3d2a347b2f85682f2a54c30e63b7e9c88ffe402d6d71fa7ce6656786f93b886" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.138183 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a3d2a347b2f85682f2a54c30e63b7e9c88ffe402d6d71fa7ce6656786f93b886"} err="failed to get container status \"a3d2a347b2f85682f2a54c30e63b7e9c88ffe402d6d71fa7ce6656786f93b886\": rpc error: code = NotFound desc = could not find container \"a3d2a347b2f85682f2a54c30e63b7e9c88ffe402d6d71fa7ce6656786f93b886\": container with ID starting with a3d2a347b2f85682f2a54c30e63b7e9c88ffe402d6d71fa7ce6656786f93b886 not found: ID does not exist" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.138207 4907 scope.go:117] "RemoveContainer" containerID="9155434e2d8f0daa4504758fb72d6f498da107a64a92a0883ac687ed2af9caee" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.138398 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9155434e2d8f0daa4504758fb72d6f498da107a64a92a0883ac687ed2af9caee"} err="failed to get container status \"9155434e2d8f0daa4504758fb72d6f498da107a64a92a0883ac687ed2af9caee\": rpc error: code = NotFound desc = could not find container \"9155434e2d8f0daa4504758fb72d6f498da107a64a92a0883ac687ed2af9caee\": container with ID starting with 9155434e2d8f0daa4504758fb72d6f498da107a64a92a0883ac687ed2af9caee not found: ID does not exist" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.138419 4907 scope.go:117] "RemoveContainer" containerID="4d60b5b774f8dd4cee16b9784f80d124640199a85331a2940772029be4884751" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.138591 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4d60b5b774f8dd4cee16b9784f80d124640199a85331a2940772029be4884751"} err="failed to get container status \"4d60b5b774f8dd4cee16b9784f80d124640199a85331a2940772029be4884751\": rpc error: code = NotFound desc = could not find container \"4d60b5b774f8dd4cee16b9784f80d124640199a85331a2940772029be4884751\": container with ID starting with 4d60b5b774f8dd4cee16b9784f80d124640199a85331a2940772029be4884751 not found: ID does not exist" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.138611 4907 scope.go:117] "RemoveContainer" containerID="a0ee0049aa4a6d49a6bcf7d347183d9cc8b85171f988662803a005e0cdd97395" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.138859 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a0ee0049aa4a6d49a6bcf7d347183d9cc8b85171f988662803a005e0cdd97395"} err="failed to get container status \"a0ee0049aa4a6d49a6bcf7d347183d9cc8b85171f988662803a005e0cdd97395\": rpc error: code = NotFound desc = could not find container \"a0ee0049aa4a6d49a6bcf7d347183d9cc8b85171f988662803a005e0cdd97395\": container with ID starting with a0ee0049aa4a6d49a6bcf7d347183d9cc8b85171f988662803a005e0cdd97395 not found: ID does not exist" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.138897 4907 scope.go:117] "RemoveContainer" containerID="0fff87d8ea572615eb1329164f9e536e6ee45cd00a9e71be0aa8a9b3ed2190e2" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.139151 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0fff87d8ea572615eb1329164f9e536e6ee45cd00a9e71be0aa8a9b3ed2190e2"} err="failed to get container status \"0fff87d8ea572615eb1329164f9e536e6ee45cd00a9e71be0aa8a9b3ed2190e2\": rpc error: code = NotFound desc = could not find container \"0fff87d8ea572615eb1329164f9e536e6ee45cd00a9e71be0aa8a9b3ed2190e2\": container with ID starting with 0fff87d8ea572615eb1329164f9e536e6ee45cd00a9e71be0aa8a9b3ed2190e2 not found: ID does not exist" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.139172 4907 scope.go:117] "RemoveContainer" containerID="ccef92e41501fab91189581d7e8875515c02d637426b329d2fb80a356c75b637" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.139373 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ccef92e41501fab91189581d7e8875515c02d637426b329d2fb80a356c75b637"} err="failed to get container status \"ccef92e41501fab91189581d7e8875515c02d637426b329d2fb80a356c75b637\": rpc error: code = NotFound desc = could not find container \"ccef92e41501fab91189581d7e8875515c02d637426b329d2fb80a356c75b637\": container with ID starting with ccef92e41501fab91189581d7e8875515c02d637426b329d2fb80a356c75b637 not found: ID does not exist" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.139394 4907 scope.go:117] "RemoveContainer" containerID="bf6a3dbaa8dda61f461a906538824c8946d36dc07d4fb95bf4d8420ad0105600" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.139574 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bf6a3dbaa8dda61f461a906538824c8946d36dc07d4fb95bf4d8420ad0105600"} err="failed to get container status \"bf6a3dbaa8dda61f461a906538824c8946d36dc07d4fb95bf4d8420ad0105600\": rpc error: code = NotFound desc = could not find container \"bf6a3dbaa8dda61f461a906538824c8946d36dc07d4fb95bf4d8420ad0105600\": container with ID starting with bf6a3dbaa8dda61f461a906538824c8946d36dc07d4fb95bf4d8420ad0105600 not found: ID does not exist" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.139595 4907 scope.go:117] "RemoveContainer" containerID="3c165053fd9933444b372ff40efe33c0e193a7f39e6f531a4f370e971eb41e6d" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.139778 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3c165053fd9933444b372ff40efe33c0e193a7f39e6f531a4f370e971eb41e6d"} err="failed to get container status \"3c165053fd9933444b372ff40efe33c0e193a7f39e6f531a4f370e971eb41e6d\": rpc error: code = NotFound desc = could not find container \"3c165053fd9933444b372ff40efe33c0e193a7f39e6f531a4f370e971eb41e6d\": container with ID starting with 3c165053fd9933444b372ff40efe33c0e193a7f39e6f531a4f370e971eb41e6d not found: ID does not exist" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.139798 4907 scope.go:117] "RemoveContainer" containerID="fea41c11bebd2d9d622ed4b8bc50e03a106dadea6ac102cd60512ea2a1f8b8b0" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.140019 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fea41c11bebd2d9d622ed4b8bc50e03a106dadea6ac102cd60512ea2a1f8b8b0"} err="failed to get container status \"fea41c11bebd2d9d622ed4b8bc50e03a106dadea6ac102cd60512ea2a1f8b8b0\": rpc error: code = NotFound desc = could not find container \"fea41c11bebd2d9d622ed4b8bc50e03a106dadea6ac102cd60512ea2a1f8b8b0\": container with ID starting with fea41c11bebd2d9d622ed4b8bc50e03a106dadea6ac102cd60512ea2a1f8b8b0 not found: ID does not exist" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.140040 4907 scope.go:117] "RemoveContainer" containerID="858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.140230 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279"} err="failed to get container status \"858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\": rpc error: code = NotFound desc = could not find container \"858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\": container with ID starting with 858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279 not found: ID does not exist" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.140251 4907 scope.go:117] "RemoveContainer" containerID="a3d2a347b2f85682f2a54c30e63b7e9c88ffe402d6d71fa7ce6656786f93b886" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.140439 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a3d2a347b2f85682f2a54c30e63b7e9c88ffe402d6d71fa7ce6656786f93b886"} err="failed to get container status \"a3d2a347b2f85682f2a54c30e63b7e9c88ffe402d6d71fa7ce6656786f93b886\": rpc error: code = NotFound desc = could not find container \"a3d2a347b2f85682f2a54c30e63b7e9c88ffe402d6d71fa7ce6656786f93b886\": container with ID starting with a3d2a347b2f85682f2a54c30e63b7e9c88ffe402d6d71fa7ce6656786f93b886 not found: ID does not exist" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.140463 4907 scope.go:117] "RemoveContainer" containerID="9155434e2d8f0daa4504758fb72d6f498da107a64a92a0883ac687ed2af9caee" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.140653 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9155434e2d8f0daa4504758fb72d6f498da107a64a92a0883ac687ed2af9caee"} err="failed to get container status \"9155434e2d8f0daa4504758fb72d6f498da107a64a92a0883ac687ed2af9caee\": rpc error: code = NotFound desc = could not find container \"9155434e2d8f0daa4504758fb72d6f498da107a64a92a0883ac687ed2af9caee\": container with ID starting with 9155434e2d8f0daa4504758fb72d6f498da107a64a92a0883ac687ed2af9caee not found: ID does not exist" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.140675 4907 scope.go:117] "RemoveContainer" containerID="4d60b5b774f8dd4cee16b9784f80d124640199a85331a2940772029be4884751" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.141348 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4d60b5b774f8dd4cee16b9784f80d124640199a85331a2940772029be4884751"} err="failed to get container status \"4d60b5b774f8dd4cee16b9784f80d124640199a85331a2940772029be4884751\": rpc error: code = NotFound desc = could not find container \"4d60b5b774f8dd4cee16b9784f80d124640199a85331a2940772029be4884751\": container with ID starting with 4d60b5b774f8dd4cee16b9784f80d124640199a85331a2940772029be4884751 not found: ID does not exist" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.141369 4907 scope.go:117] "RemoveContainer" containerID="a0ee0049aa4a6d49a6bcf7d347183d9cc8b85171f988662803a005e0cdd97395" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.141568 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a0ee0049aa4a6d49a6bcf7d347183d9cc8b85171f988662803a005e0cdd97395"} err="failed to get container status \"a0ee0049aa4a6d49a6bcf7d347183d9cc8b85171f988662803a005e0cdd97395\": rpc error: code = NotFound desc = could not find container \"a0ee0049aa4a6d49a6bcf7d347183d9cc8b85171f988662803a005e0cdd97395\": container with ID starting with a0ee0049aa4a6d49a6bcf7d347183d9cc8b85171f988662803a005e0cdd97395 not found: ID does not exist" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.141591 4907 scope.go:117] "RemoveContainer" containerID="0fff87d8ea572615eb1329164f9e536e6ee45cd00a9e71be0aa8a9b3ed2190e2" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.141816 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0fff87d8ea572615eb1329164f9e536e6ee45cd00a9e71be0aa8a9b3ed2190e2"} err="failed to get container status \"0fff87d8ea572615eb1329164f9e536e6ee45cd00a9e71be0aa8a9b3ed2190e2\": rpc error: code = NotFound desc = could not find container \"0fff87d8ea572615eb1329164f9e536e6ee45cd00a9e71be0aa8a9b3ed2190e2\": container with ID starting with 0fff87d8ea572615eb1329164f9e536e6ee45cd00a9e71be0aa8a9b3ed2190e2 not found: ID does not exist" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.141838 4907 scope.go:117] "RemoveContainer" containerID="ccef92e41501fab91189581d7e8875515c02d637426b329d2fb80a356c75b637" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.142117 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ccef92e41501fab91189581d7e8875515c02d637426b329d2fb80a356c75b637"} err="failed to get container status \"ccef92e41501fab91189581d7e8875515c02d637426b329d2fb80a356c75b637\": rpc error: code = NotFound desc = could not find container \"ccef92e41501fab91189581d7e8875515c02d637426b329d2fb80a356c75b637\": container with ID starting with ccef92e41501fab91189581d7e8875515c02d637426b329d2fb80a356c75b637 not found: ID does not exist" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.142138 4907 scope.go:117] "RemoveContainer" containerID="bf6a3dbaa8dda61f461a906538824c8946d36dc07d4fb95bf4d8420ad0105600" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.142350 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bf6a3dbaa8dda61f461a906538824c8946d36dc07d4fb95bf4d8420ad0105600"} err="failed to get container status \"bf6a3dbaa8dda61f461a906538824c8946d36dc07d4fb95bf4d8420ad0105600\": rpc error: code = NotFound desc = could not find container \"bf6a3dbaa8dda61f461a906538824c8946d36dc07d4fb95bf4d8420ad0105600\": container with ID starting with bf6a3dbaa8dda61f461a906538824c8946d36dc07d4fb95bf4d8420ad0105600 not found: ID does not exist" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.142371 4907 scope.go:117] "RemoveContainer" containerID="3c165053fd9933444b372ff40efe33c0e193a7f39e6f531a4f370e971eb41e6d" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.142561 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3c165053fd9933444b372ff40efe33c0e193a7f39e6f531a4f370e971eb41e6d"} err="failed to get container status \"3c165053fd9933444b372ff40efe33c0e193a7f39e6f531a4f370e971eb41e6d\": rpc error: code = NotFound desc = could not find container \"3c165053fd9933444b372ff40efe33c0e193a7f39e6f531a4f370e971eb41e6d\": container with ID starting with 3c165053fd9933444b372ff40efe33c0e193a7f39e6f531a4f370e971eb41e6d not found: ID does not exist" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.142582 4907 scope.go:117] "RemoveContainer" containerID="fea41c11bebd2d9d622ed4b8bc50e03a106dadea6ac102cd60512ea2a1f8b8b0" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.142765 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fea41c11bebd2d9d622ed4b8bc50e03a106dadea6ac102cd60512ea2a1f8b8b0"} err="failed to get container status \"fea41c11bebd2d9d622ed4b8bc50e03a106dadea6ac102cd60512ea2a1f8b8b0\": rpc error: code = NotFound desc = could not find container \"fea41c11bebd2d9d622ed4b8bc50e03a106dadea6ac102cd60512ea2a1f8b8b0\": container with ID starting with fea41c11bebd2d9d622ed4b8bc50e03a106dadea6ac102cd60512ea2a1f8b8b0 not found: ID does not exist" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.142787 4907 scope.go:117] "RemoveContainer" containerID="858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.143046 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279"} err="failed to get container status \"858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\": rpc error: code = NotFound desc = could not find container \"858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\": container with ID starting with 858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279 not found: ID does not exist" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.143069 4907 scope.go:117] "RemoveContainer" containerID="a3d2a347b2f85682f2a54c30e63b7e9c88ffe402d6d71fa7ce6656786f93b886" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.143262 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a3d2a347b2f85682f2a54c30e63b7e9c88ffe402d6d71fa7ce6656786f93b886"} err="failed to get container status \"a3d2a347b2f85682f2a54c30e63b7e9c88ffe402d6d71fa7ce6656786f93b886\": rpc error: code = NotFound desc = could not find container \"a3d2a347b2f85682f2a54c30e63b7e9c88ffe402d6d71fa7ce6656786f93b886\": container with ID starting with a3d2a347b2f85682f2a54c30e63b7e9c88ffe402d6d71fa7ce6656786f93b886 not found: ID does not exist" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.791831 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1006376d-4033-454e-8a23-2b67882bdd0b" path="/var/lib/kubelet/pods/1006376d-4033-454e-8a23-2b67882bdd0b/volumes" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.915126 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-xtwf7_64ebd338-0158-44dc-90ab-e7c285e87762/kube-multus/2.log" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.915906 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-xtwf7_64ebd338-0158-44dc-90ab-e7c285e87762/kube-multus/1.log" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.916107 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-xtwf7" event={"ID":"64ebd338-0158-44dc-90ab-e7c285e87762","Type":"ContainerStarted","Data":"569f8a799f7f20725d989134c554036cb2d5eb61b9ee2129f2d173a5c1c3e1a9"} Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.921126 4907 generic.go:334] "Generic (PLEG): container finished" podID="42f6fac8-4aa0-4cd7-a1a1-c80ced66c462" containerID="d106a54791454eb6427df73c45bd36a1109922c53a635e0a8d7cf85779280458" exitCode=0 Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.921212 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" event={"ID":"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462","Type":"ContainerDied","Data":"d106a54791454eb6427df73c45bd36a1109922c53a635e0a8d7cf85779280458"} Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.921256 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" event={"ID":"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462","Type":"ContainerStarted","Data":"b113dbc4695af8ad071eea34316d474e13189383fd3ee685adb103d09acf57d4"} Mar 13 14:19:44 crc kubenswrapper[4907]: I0313 14:19:44.933911 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" event={"ID":"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462","Type":"ContainerStarted","Data":"578de0ccce050e57e4f8e80bbe09f048bb5cd0dae436f2ba29311e78a016cc11"} Mar 13 14:19:44 crc kubenswrapper[4907]: I0313 14:19:44.935079 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" event={"ID":"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462","Type":"ContainerStarted","Data":"25b13197844fea9f278eff609105bf3c042456ddb283a72136938e344204cf4f"} Mar 13 14:19:44 crc kubenswrapper[4907]: I0313 14:19:44.935186 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" event={"ID":"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462","Type":"ContainerStarted","Data":"0d0df32e73dca5d01127944f2d46839aeee380a1b2ea551c278b048e37bffe05"} Mar 13 14:19:44 crc kubenswrapper[4907]: I0313 14:19:44.935260 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" event={"ID":"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462","Type":"ContainerStarted","Data":"c23967456e62fa626525f0f5f03be5f1b03daa0c6c67d13556df25e9cf70b1ed"} Mar 13 14:19:45 crc kubenswrapper[4907]: I0313 14:19:45.945541 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" event={"ID":"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462","Type":"ContainerStarted","Data":"28cd9af2d20dd6887662d65303cd8fba04b00f95081ebede4b2cbadeabf91813"} Mar 13 14:19:45 crc kubenswrapper[4907]: I0313 14:19:45.945917 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" event={"ID":"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462","Type":"ContainerStarted","Data":"9af887e10ec13a531557ac2b8baedf3614b35ac1b0b62a832b549ded6ce7fd7f"} Mar 13 14:19:47 crc kubenswrapper[4907]: I0313 14:19:47.963302 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" event={"ID":"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462","Type":"ContainerStarted","Data":"347baab8da25f1e8a13e9500f5636a4f8b732c7dd039e0d6df7ce84b3cf22f01"} Mar 13 14:19:48 crc kubenswrapper[4907]: I0313 14:19:48.732144 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["crc-storage/crc-storage-crc-wds2z"] Mar 13 14:19:48 crc kubenswrapper[4907]: I0313 14:19:48.733177 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-wds2z" Mar 13 14:19:48 crc kubenswrapper[4907]: I0313 14:19:48.735227 4907 reflector.go:368] Caches populated for *v1.Secret from object-"crc-storage"/"crc-storage-dockercfg-2d28j" Mar 13 14:19:48 crc kubenswrapper[4907]: I0313 14:19:48.735339 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"kube-root-ca.crt" Mar 13 14:19:48 crc kubenswrapper[4907]: I0313 14:19:48.735458 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"openshift-service-ca.crt" Mar 13 14:19:48 crc kubenswrapper[4907]: I0313 14:19:48.735913 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"crc-storage" Mar 13 14:19:48 crc kubenswrapper[4907]: I0313 14:19:48.792226 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/b501aea0-66d1-4655-b67f-c6a3bacb7cdb-node-mnt\") pod \"crc-storage-crc-wds2z\" (UID: \"b501aea0-66d1-4655-b67f-c6a3bacb7cdb\") " pod="crc-storage/crc-storage-crc-wds2z" Mar 13 14:19:48 crc kubenswrapper[4907]: I0313 14:19:48.792415 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fdjxm\" (UniqueName: \"kubernetes.io/projected/b501aea0-66d1-4655-b67f-c6a3bacb7cdb-kube-api-access-fdjxm\") pod \"crc-storage-crc-wds2z\" (UID: \"b501aea0-66d1-4655-b67f-c6a3bacb7cdb\") " pod="crc-storage/crc-storage-crc-wds2z" Mar 13 14:19:48 crc kubenswrapper[4907]: I0313 14:19:48.792654 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/b501aea0-66d1-4655-b67f-c6a3bacb7cdb-crc-storage\") pod \"crc-storage-crc-wds2z\" (UID: \"b501aea0-66d1-4655-b67f-c6a3bacb7cdb\") " pod="crc-storage/crc-storage-crc-wds2z" Mar 13 14:19:48 crc kubenswrapper[4907]: I0313 14:19:48.893962 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/b501aea0-66d1-4655-b67f-c6a3bacb7cdb-crc-storage\") pod \"crc-storage-crc-wds2z\" (UID: \"b501aea0-66d1-4655-b67f-c6a3bacb7cdb\") " pod="crc-storage/crc-storage-crc-wds2z" Mar 13 14:19:48 crc kubenswrapper[4907]: I0313 14:19:48.894041 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/b501aea0-66d1-4655-b67f-c6a3bacb7cdb-node-mnt\") pod \"crc-storage-crc-wds2z\" (UID: \"b501aea0-66d1-4655-b67f-c6a3bacb7cdb\") " pod="crc-storage/crc-storage-crc-wds2z" Mar 13 14:19:48 crc kubenswrapper[4907]: I0313 14:19:48.894131 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fdjxm\" (UniqueName: \"kubernetes.io/projected/b501aea0-66d1-4655-b67f-c6a3bacb7cdb-kube-api-access-fdjxm\") pod \"crc-storage-crc-wds2z\" (UID: \"b501aea0-66d1-4655-b67f-c6a3bacb7cdb\") " pod="crc-storage/crc-storage-crc-wds2z" Mar 13 14:19:48 crc kubenswrapper[4907]: I0313 14:19:48.894549 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/b501aea0-66d1-4655-b67f-c6a3bacb7cdb-node-mnt\") pod \"crc-storage-crc-wds2z\" (UID: \"b501aea0-66d1-4655-b67f-c6a3bacb7cdb\") " pod="crc-storage/crc-storage-crc-wds2z" Mar 13 14:19:48 crc kubenswrapper[4907]: I0313 14:19:48.895525 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/b501aea0-66d1-4655-b67f-c6a3bacb7cdb-crc-storage\") pod \"crc-storage-crc-wds2z\" (UID: \"b501aea0-66d1-4655-b67f-c6a3bacb7cdb\") " pod="crc-storage/crc-storage-crc-wds2z" Mar 13 14:19:48 crc kubenswrapper[4907]: I0313 14:19:48.924010 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fdjxm\" (UniqueName: \"kubernetes.io/projected/b501aea0-66d1-4655-b67f-c6a3bacb7cdb-kube-api-access-fdjxm\") pod \"crc-storage-crc-wds2z\" (UID: \"b501aea0-66d1-4655-b67f-c6a3bacb7cdb\") " pod="crc-storage/crc-storage-crc-wds2z" Mar 13 14:19:49 crc kubenswrapper[4907]: I0313 14:19:49.059175 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-wds2z" Mar 13 14:19:49 crc kubenswrapper[4907]: E0313 14:19:49.111605 4907 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-wds2z_crc-storage_b501aea0-66d1-4655-b67f-c6a3bacb7cdb_0(664f7b52eff02406755a65652f95b07d39e80c06474459a884c4a4c3a60ed0fe): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 13 14:19:49 crc kubenswrapper[4907]: E0313 14:19:49.111767 4907 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-wds2z_crc-storage_b501aea0-66d1-4655-b67f-c6a3bacb7cdb_0(664f7b52eff02406755a65652f95b07d39e80c06474459a884c4a4c3a60ed0fe): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-wds2z" Mar 13 14:19:49 crc kubenswrapper[4907]: E0313 14:19:49.111812 4907 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-wds2z_crc-storage_b501aea0-66d1-4655-b67f-c6a3bacb7cdb_0(664f7b52eff02406755a65652f95b07d39e80c06474459a884c4a4c3a60ed0fe): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-wds2z" Mar 13 14:19:49 crc kubenswrapper[4907]: E0313 14:19:49.111920 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"crc-storage-crc-wds2z_crc-storage(b501aea0-66d1-4655-b67f-c6a3bacb7cdb)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"crc-storage-crc-wds2z_crc-storage(b501aea0-66d1-4655-b67f-c6a3bacb7cdb)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-wds2z_crc-storage_b501aea0-66d1-4655-b67f-c6a3bacb7cdb_0(664f7b52eff02406755a65652f95b07d39e80c06474459a884c4a4c3a60ed0fe): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="crc-storage/crc-storage-crc-wds2z" podUID="b501aea0-66d1-4655-b67f-c6a3bacb7cdb" Mar 13 14:19:50 crc kubenswrapper[4907]: I0313 14:19:50.734059 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-wds2z"] Mar 13 14:19:50 crc kubenswrapper[4907]: I0313 14:19:50.735383 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-wds2z" Mar 13 14:19:50 crc kubenswrapper[4907]: I0313 14:19:50.736303 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-wds2z" Mar 13 14:19:50 crc kubenswrapper[4907]: E0313 14:19:50.780165 4907 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-wds2z_crc-storage_b501aea0-66d1-4655-b67f-c6a3bacb7cdb_0(fa9f25ad6b8187a7a88d3dd386755b7b0b93bda2b9cade1fbc1357b71fa54a4b): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 13 14:19:50 crc kubenswrapper[4907]: E0313 14:19:50.780261 4907 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-wds2z_crc-storage_b501aea0-66d1-4655-b67f-c6a3bacb7cdb_0(fa9f25ad6b8187a7a88d3dd386755b7b0b93bda2b9cade1fbc1357b71fa54a4b): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-wds2z" Mar 13 14:19:50 crc kubenswrapper[4907]: E0313 14:19:50.780295 4907 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-wds2z_crc-storage_b501aea0-66d1-4655-b67f-c6a3bacb7cdb_0(fa9f25ad6b8187a7a88d3dd386755b7b0b93bda2b9cade1fbc1357b71fa54a4b): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-wds2z" Mar 13 14:19:50 crc kubenswrapper[4907]: E0313 14:19:50.780369 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"crc-storage-crc-wds2z_crc-storage(b501aea0-66d1-4655-b67f-c6a3bacb7cdb)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"crc-storage-crc-wds2z_crc-storage(b501aea0-66d1-4655-b67f-c6a3bacb7cdb)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-wds2z_crc-storage_b501aea0-66d1-4655-b67f-c6a3bacb7cdb_0(fa9f25ad6b8187a7a88d3dd386755b7b0b93bda2b9cade1fbc1357b71fa54a4b): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="crc-storage/crc-storage-crc-wds2z" podUID="b501aea0-66d1-4655-b67f-c6a3bacb7cdb" Mar 13 14:19:50 crc kubenswrapper[4907]: I0313 14:19:50.987326 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" event={"ID":"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462","Type":"ContainerStarted","Data":"a0f5a9d3eb46fe980c3893c4201a77f3a1ab6fe006f338a8077b3da098ea6914"} Mar 13 14:19:50 crc kubenswrapper[4907]: I0313 14:19:50.987872 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:50 crc kubenswrapper[4907]: I0313 14:19:50.988091 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:50 crc kubenswrapper[4907]: I0313 14:19:50.988220 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:51 crc kubenswrapper[4907]: I0313 14:19:51.040228 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" podStartSLOduration=9.040197703 podStartE2EDuration="9.040197703s" podCreationTimestamp="2026-03-13 14:19:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:19:51.027082755 +0000 UTC m=+889.926870484" watchObservedRunningTime="2026-03-13 14:19:51.040197703 +0000 UTC m=+889.939985432" Mar 13 14:19:51 crc kubenswrapper[4907]: I0313 14:19:51.049633 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:51 crc kubenswrapper[4907]: I0313 14:19:51.056619 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:20:00 crc kubenswrapper[4907]: I0313 14:20:00.148531 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556860-ljdj7"] Mar 13 14:20:00 crc kubenswrapper[4907]: I0313 14:20:00.152439 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556860-ljdj7" Mar 13 14:20:00 crc kubenswrapper[4907]: I0313 14:20:00.155249 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 14:20:00 crc kubenswrapper[4907]: I0313 14:20:00.157244 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 14:20:00 crc kubenswrapper[4907]: I0313 14:20:00.157403 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 14:20:00 crc kubenswrapper[4907]: I0313 14:20:00.161833 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556860-ljdj7"] Mar 13 14:20:00 crc kubenswrapper[4907]: I0313 14:20:00.249612 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j2h2l\" (UniqueName: \"kubernetes.io/projected/257ee294-461c-4344-9dcb-b44f1497f8bf-kube-api-access-j2h2l\") pod \"auto-csr-approver-29556860-ljdj7\" (UID: \"257ee294-461c-4344-9dcb-b44f1497f8bf\") " pod="openshift-infra/auto-csr-approver-29556860-ljdj7" Mar 13 14:20:00 crc kubenswrapper[4907]: I0313 14:20:00.351174 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j2h2l\" (UniqueName: \"kubernetes.io/projected/257ee294-461c-4344-9dcb-b44f1497f8bf-kube-api-access-j2h2l\") pod \"auto-csr-approver-29556860-ljdj7\" (UID: \"257ee294-461c-4344-9dcb-b44f1497f8bf\") " pod="openshift-infra/auto-csr-approver-29556860-ljdj7" Mar 13 14:20:00 crc kubenswrapper[4907]: I0313 14:20:00.373476 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j2h2l\" (UniqueName: \"kubernetes.io/projected/257ee294-461c-4344-9dcb-b44f1497f8bf-kube-api-access-j2h2l\") pod \"auto-csr-approver-29556860-ljdj7\" (UID: \"257ee294-461c-4344-9dcb-b44f1497f8bf\") " pod="openshift-infra/auto-csr-approver-29556860-ljdj7" Mar 13 14:20:00 crc kubenswrapper[4907]: I0313 14:20:00.484078 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556860-ljdj7" Mar 13 14:20:00 crc kubenswrapper[4907]: I0313 14:20:00.752183 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556860-ljdj7"] Mar 13 14:20:00 crc kubenswrapper[4907]: W0313 14:20:00.764092 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod257ee294_461c_4344_9dcb_b44f1497f8bf.slice/crio-4054e86db3967d65bfa7504ca3ded70759d07d9bcb2b6c5f7340940d6a6f850f WatchSource:0}: Error finding container 4054e86db3967d65bfa7504ca3ded70759d07d9bcb2b6c5f7340940d6a6f850f: Status 404 returned error can't find the container with id 4054e86db3967d65bfa7504ca3ded70759d07d9bcb2b6c5f7340940d6a6f850f Mar 13 14:20:01 crc kubenswrapper[4907]: I0313 14:20:01.062010 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556860-ljdj7" event={"ID":"257ee294-461c-4344-9dcb-b44f1497f8bf","Type":"ContainerStarted","Data":"4054e86db3967d65bfa7504ca3ded70759d07d9bcb2b6c5f7340940d6a6f850f"} Mar 13 14:20:02 crc kubenswrapper[4907]: I0313 14:20:02.072055 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556860-ljdj7" event={"ID":"257ee294-461c-4344-9dcb-b44f1497f8bf","Type":"ContainerStarted","Data":"dd6efe3aa3a79d161b96b8d84e7fa19f50f23cdeadb3268906a3667faac57368"} Mar 13 14:20:02 crc kubenswrapper[4907]: I0313 14:20:02.426114 4907 scope.go:117] "RemoveContainer" containerID="ef632bb17c8f529b35d0882c8d4ba3dc213b0308b46a4ade8426b9c08e64bb2b" Mar 13 14:20:02 crc kubenswrapper[4907]: I0313 14:20:02.781698 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-wds2z" Mar 13 14:20:02 crc kubenswrapper[4907]: I0313 14:20:02.782517 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-wds2z" Mar 13 14:20:03 crc kubenswrapper[4907]: I0313 14:20:03.094423 4907 generic.go:334] "Generic (PLEG): container finished" podID="257ee294-461c-4344-9dcb-b44f1497f8bf" containerID="dd6efe3aa3a79d161b96b8d84e7fa19f50f23cdeadb3268906a3667faac57368" exitCode=0 Mar 13 14:20:03 crc kubenswrapper[4907]: I0313 14:20:03.094545 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556860-ljdj7" event={"ID":"257ee294-461c-4344-9dcb-b44f1497f8bf","Type":"ContainerDied","Data":"dd6efe3aa3a79d161b96b8d84e7fa19f50f23cdeadb3268906a3667faac57368"} Mar 13 14:20:03 crc kubenswrapper[4907]: I0313 14:20:03.101126 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-xtwf7_64ebd338-0158-44dc-90ab-e7c285e87762/kube-multus/2.log" Mar 13 14:20:03 crc kubenswrapper[4907]: I0313 14:20:03.239182 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-wds2z"] Mar 13 14:20:03 crc kubenswrapper[4907]: W0313 14:20:03.245973 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb501aea0_66d1_4655_b67f_c6a3bacb7cdb.slice/crio-5e8decbb8a440a87fe3c9efeb0e3bba54f69864a8dedb5e80d99fcf912b2628a WatchSource:0}: Error finding container 5e8decbb8a440a87fe3c9efeb0e3bba54f69864a8dedb5e80d99fcf912b2628a: Status 404 returned error can't find the container with id 5e8decbb8a440a87fe3c9efeb0e3bba54f69864a8dedb5e80d99fcf912b2628a Mar 13 14:20:04 crc kubenswrapper[4907]: I0313 14:20:04.111540 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-wds2z" event={"ID":"b501aea0-66d1-4655-b67f-c6a3bacb7cdb","Type":"ContainerStarted","Data":"5e8decbb8a440a87fe3c9efeb0e3bba54f69864a8dedb5e80d99fcf912b2628a"} Mar 13 14:20:04 crc kubenswrapper[4907]: I0313 14:20:04.556988 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556860-ljdj7" Mar 13 14:20:04 crc kubenswrapper[4907]: I0313 14:20:04.609923 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j2h2l\" (UniqueName: \"kubernetes.io/projected/257ee294-461c-4344-9dcb-b44f1497f8bf-kube-api-access-j2h2l\") pod \"257ee294-461c-4344-9dcb-b44f1497f8bf\" (UID: \"257ee294-461c-4344-9dcb-b44f1497f8bf\") " Mar 13 14:20:04 crc kubenswrapper[4907]: I0313 14:20:04.616521 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/257ee294-461c-4344-9dcb-b44f1497f8bf-kube-api-access-j2h2l" (OuterVolumeSpecName: "kube-api-access-j2h2l") pod "257ee294-461c-4344-9dcb-b44f1497f8bf" (UID: "257ee294-461c-4344-9dcb-b44f1497f8bf"). InnerVolumeSpecName "kube-api-access-j2h2l". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:20:04 crc kubenswrapper[4907]: I0313 14:20:04.711519 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j2h2l\" (UniqueName: \"kubernetes.io/projected/257ee294-461c-4344-9dcb-b44f1497f8bf-kube-api-access-j2h2l\") on node \"crc\" DevicePath \"\"" Mar 13 14:20:04 crc kubenswrapper[4907]: I0313 14:20:04.872805 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556854-l7l9t"] Mar 13 14:20:04 crc kubenswrapper[4907]: I0313 14:20:04.879072 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556854-l7l9t"] Mar 13 14:20:05 crc kubenswrapper[4907]: I0313 14:20:05.118357 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556860-ljdj7" event={"ID":"257ee294-461c-4344-9dcb-b44f1497f8bf","Type":"ContainerDied","Data":"4054e86db3967d65bfa7504ca3ded70759d07d9bcb2b6c5f7340940d6a6f850f"} Mar 13 14:20:05 crc kubenswrapper[4907]: I0313 14:20:05.118394 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4054e86db3967d65bfa7504ca3ded70759d07d9bcb2b6c5f7340940d6a6f850f" Mar 13 14:20:05 crc kubenswrapper[4907]: I0313 14:20:05.118495 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556860-ljdj7" Mar 13 14:20:05 crc kubenswrapper[4907]: I0313 14:20:05.120003 4907 generic.go:334] "Generic (PLEG): container finished" podID="b501aea0-66d1-4655-b67f-c6a3bacb7cdb" containerID="311a031e1ff4ef3b74d10161b82c176889b7b82821fc88f2a4edeca5e5657ee5" exitCode=0 Mar 13 14:20:05 crc kubenswrapper[4907]: I0313 14:20:05.120050 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-wds2z" event={"ID":"b501aea0-66d1-4655-b67f-c6a3bacb7cdb","Type":"ContainerDied","Data":"311a031e1ff4ef3b74d10161b82c176889b7b82821fc88f2a4edeca5e5657ee5"} Mar 13 14:20:05 crc kubenswrapper[4907]: E0313 14:20:05.178254 4907 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod257ee294_461c_4344_9dcb_b44f1497f8bf.slice/crio-4054e86db3967d65bfa7504ca3ded70759d07d9bcb2b6c5f7340940d6a6f850f\": RecentStats: unable to find data in memory cache]" Mar 13 14:20:05 crc kubenswrapper[4907]: I0313 14:20:05.794257 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="82df106b-e5b8-4798-942a-999c2ac7a4be" path="/var/lib/kubelet/pods/82df106b-e5b8-4798-942a-999c2ac7a4be/volumes" Mar 13 14:20:06 crc kubenswrapper[4907]: I0313 14:20:06.408726 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-wds2z" Mar 13 14:20:06 crc kubenswrapper[4907]: I0313 14:20:06.432839 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/b501aea0-66d1-4655-b67f-c6a3bacb7cdb-crc-storage\") pod \"b501aea0-66d1-4655-b67f-c6a3bacb7cdb\" (UID: \"b501aea0-66d1-4655-b67f-c6a3bacb7cdb\") " Mar 13 14:20:06 crc kubenswrapper[4907]: I0313 14:20:06.432933 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fdjxm\" (UniqueName: \"kubernetes.io/projected/b501aea0-66d1-4655-b67f-c6a3bacb7cdb-kube-api-access-fdjxm\") pod \"b501aea0-66d1-4655-b67f-c6a3bacb7cdb\" (UID: \"b501aea0-66d1-4655-b67f-c6a3bacb7cdb\") " Mar 13 14:20:06 crc kubenswrapper[4907]: I0313 14:20:06.432982 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/b501aea0-66d1-4655-b67f-c6a3bacb7cdb-node-mnt\") pod \"b501aea0-66d1-4655-b67f-c6a3bacb7cdb\" (UID: \"b501aea0-66d1-4655-b67f-c6a3bacb7cdb\") " Mar 13 14:20:06 crc kubenswrapper[4907]: I0313 14:20:06.433257 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b501aea0-66d1-4655-b67f-c6a3bacb7cdb-node-mnt" (OuterVolumeSpecName: "node-mnt") pod "b501aea0-66d1-4655-b67f-c6a3bacb7cdb" (UID: "b501aea0-66d1-4655-b67f-c6a3bacb7cdb"). InnerVolumeSpecName "node-mnt". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 14:20:06 crc kubenswrapper[4907]: I0313 14:20:06.439109 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b501aea0-66d1-4655-b67f-c6a3bacb7cdb-kube-api-access-fdjxm" (OuterVolumeSpecName: "kube-api-access-fdjxm") pod "b501aea0-66d1-4655-b67f-c6a3bacb7cdb" (UID: "b501aea0-66d1-4655-b67f-c6a3bacb7cdb"). InnerVolumeSpecName "kube-api-access-fdjxm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:20:06 crc kubenswrapper[4907]: I0313 14:20:06.458774 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b501aea0-66d1-4655-b67f-c6a3bacb7cdb-crc-storage" (OuterVolumeSpecName: "crc-storage") pod "b501aea0-66d1-4655-b67f-c6a3bacb7cdb" (UID: "b501aea0-66d1-4655-b67f-c6a3bacb7cdb"). InnerVolumeSpecName "crc-storage". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:20:06 crc kubenswrapper[4907]: I0313 14:20:06.535080 4907 reconciler_common.go:293] "Volume detached for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/b501aea0-66d1-4655-b67f-c6a3bacb7cdb-node-mnt\") on node \"crc\" DevicePath \"\"" Mar 13 14:20:06 crc kubenswrapper[4907]: I0313 14:20:06.535143 4907 reconciler_common.go:293] "Volume detached for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/b501aea0-66d1-4655-b67f-c6a3bacb7cdb-crc-storage\") on node \"crc\" DevicePath \"\"" Mar 13 14:20:06 crc kubenswrapper[4907]: I0313 14:20:06.535163 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fdjxm\" (UniqueName: \"kubernetes.io/projected/b501aea0-66d1-4655-b67f-c6a3bacb7cdb-kube-api-access-fdjxm\") on node \"crc\" DevicePath \"\"" Mar 13 14:20:07 crc kubenswrapper[4907]: I0313 14:20:07.137441 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-wds2z" event={"ID":"b501aea0-66d1-4655-b67f-c6a3bacb7cdb","Type":"ContainerDied","Data":"5e8decbb8a440a87fe3c9efeb0e3bba54f69864a8dedb5e80d99fcf912b2628a"} Mar 13 14:20:07 crc kubenswrapper[4907]: I0313 14:20:07.137863 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5e8decbb8a440a87fe3c9efeb0e3bba54f69864a8dedb5e80d99fcf912b2628a" Mar 13 14:20:07 crc kubenswrapper[4907]: I0313 14:20:07.137552 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-wds2z" Mar 13 14:20:13 crc kubenswrapper[4907]: I0313 14:20:13.048651 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:20:14 crc kubenswrapper[4907]: I0313 14:20:14.750858 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874cgs5s"] Mar 13 14:20:14 crc kubenswrapper[4907]: E0313 14:20:14.751616 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b501aea0-66d1-4655-b67f-c6a3bacb7cdb" containerName="storage" Mar 13 14:20:14 crc kubenswrapper[4907]: I0313 14:20:14.751643 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="b501aea0-66d1-4655-b67f-c6a3bacb7cdb" containerName="storage" Mar 13 14:20:14 crc kubenswrapper[4907]: E0313 14:20:14.751665 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="257ee294-461c-4344-9dcb-b44f1497f8bf" containerName="oc" Mar 13 14:20:14 crc kubenswrapper[4907]: I0313 14:20:14.751679 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="257ee294-461c-4344-9dcb-b44f1497f8bf" containerName="oc" Mar 13 14:20:14 crc kubenswrapper[4907]: I0313 14:20:14.751951 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="257ee294-461c-4344-9dcb-b44f1497f8bf" containerName="oc" Mar 13 14:20:14 crc kubenswrapper[4907]: I0313 14:20:14.751997 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="b501aea0-66d1-4655-b67f-c6a3bacb7cdb" containerName="storage" Mar 13 14:20:14 crc kubenswrapper[4907]: I0313 14:20:14.753477 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874cgs5s" Mar 13 14:20:14 crc kubenswrapper[4907]: I0313 14:20:14.755778 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Mar 13 14:20:14 crc kubenswrapper[4907]: I0313 14:20:14.763645 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874cgs5s"] Mar 13 14:20:14 crc kubenswrapper[4907]: I0313 14:20:14.854847 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a387235c-ef5d-47cc-8dff-8af5e8b5690f-bundle\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874cgs5s\" (UID: \"a387235c-ef5d-47cc-8dff-8af5e8b5690f\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874cgs5s" Mar 13 14:20:14 crc kubenswrapper[4907]: I0313 14:20:14.855113 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qrhj4\" (UniqueName: \"kubernetes.io/projected/a387235c-ef5d-47cc-8dff-8af5e8b5690f-kube-api-access-qrhj4\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874cgs5s\" (UID: \"a387235c-ef5d-47cc-8dff-8af5e8b5690f\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874cgs5s" Mar 13 14:20:14 crc kubenswrapper[4907]: I0313 14:20:14.855335 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a387235c-ef5d-47cc-8dff-8af5e8b5690f-util\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874cgs5s\" (UID: \"a387235c-ef5d-47cc-8dff-8af5e8b5690f\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874cgs5s" Mar 13 14:20:14 crc kubenswrapper[4907]: I0313 14:20:14.957164 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a387235c-ef5d-47cc-8dff-8af5e8b5690f-util\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874cgs5s\" (UID: \"a387235c-ef5d-47cc-8dff-8af5e8b5690f\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874cgs5s" Mar 13 14:20:14 crc kubenswrapper[4907]: I0313 14:20:14.957403 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a387235c-ef5d-47cc-8dff-8af5e8b5690f-bundle\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874cgs5s\" (UID: \"a387235c-ef5d-47cc-8dff-8af5e8b5690f\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874cgs5s" Mar 13 14:20:14 crc kubenswrapper[4907]: I0313 14:20:14.957510 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qrhj4\" (UniqueName: \"kubernetes.io/projected/a387235c-ef5d-47cc-8dff-8af5e8b5690f-kube-api-access-qrhj4\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874cgs5s\" (UID: \"a387235c-ef5d-47cc-8dff-8af5e8b5690f\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874cgs5s" Mar 13 14:20:14 crc kubenswrapper[4907]: I0313 14:20:14.958034 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a387235c-ef5d-47cc-8dff-8af5e8b5690f-util\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874cgs5s\" (UID: \"a387235c-ef5d-47cc-8dff-8af5e8b5690f\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874cgs5s" Mar 13 14:20:14 crc kubenswrapper[4907]: I0313 14:20:14.958044 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a387235c-ef5d-47cc-8dff-8af5e8b5690f-bundle\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874cgs5s\" (UID: \"a387235c-ef5d-47cc-8dff-8af5e8b5690f\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874cgs5s" Mar 13 14:20:14 crc kubenswrapper[4907]: I0313 14:20:14.979299 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qrhj4\" (UniqueName: \"kubernetes.io/projected/a387235c-ef5d-47cc-8dff-8af5e8b5690f-kube-api-access-qrhj4\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874cgs5s\" (UID: \"a387235c-ef5d-47cc-8dff-8af5e8b5690f\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874cgs5s" Mar 13 14:20:15 crc kubenswrapper[4907]: I0313 14:20:15.088036 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874cgs5s" Mar 13 14:20:15 crc kubenswrapper[4907]: I0313 14:20:15.504733 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874cgs5s"] Mar 13 14:20:16 crc kubenswrapper[4907]: I0313 14:20:16.191675 4907 generic.go:334] "Generic (PLEG): container finished" podID="a387235c-ef5d-47cc-8dff-8af5e8b5690f" containerID="8681768c7f655c9c29e81df6bafc91d92b79652dac2656bc6393664d83a817a3" exitCode=0 Mar 13 14:20:16 crc kubenswrapper[4907]: I0313 14:20:16.191759 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874cgs5s" event={"ID":"a387235c-ef5d-47cc-8dff-8af5e8b5690f","Type":"ContainerDied","Data":"8681768c7f655c9c29e81df6bafc91d92b79652dac2656bc6393664d83a817a3"} Mar 13 14:20:16 crc kubenswrapper[4907]: I0313 14:20:16.191961 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874cgs5s" event={"ID":"a387235c-ef5d-47cc-8dff-8af5e8b5690f","Type":"ContainerStarted","Data":"9144c72293eb14856491b3ee7a3a07c84f6294a348c2752001d7e79b3abf4e79"} Mar 13 14:20:17 crc kubenswrapper[4907]: I0313 14:20:17.094393 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-sftcn"] Mar 13 14:20:17 crc kubenswrapper[4907]: I0313 14:20:17.095734 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sftcn" Mar 13 14:20:17 crc kubenswrapper[4907]: I0313 14:20:17.109079 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-sftcn"] Mar 13 14:20:17 crc kubenswrapper[4907]: I0313 14:20:17.289783 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zp4z9\" (UniqueName: \"kubernetes.io/projected/1885cbd6-ea2e-4d89-b9be-5d8ab9d8d959-kube-api-access-zp4z9\") pod \"redhat-operators-sftcn\" (UID: \"1885cbd6-ea2e-4d89-b9be-5d8ab9d8d959\") " pod="openshift-marketplace/redhat-operators-sftcn" Mar 13 14:20:17 crc kubenswrapper[4907]: I0313 14:20:17.290052 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1885cbd6-ea2e-4d89-b9be-5d8ab9d8d959-utilities\") pod \"redhat-operators-sftcn\" (UID: \"1885cbd6-ea2e-4d89-b9be-5d8ab9d8d959\") " pod="openshift-marketplace/redhat-operators-sftcn" Mar 13 14:20:17 crc kubenswrapper[4907]: I0313 14:20:17.290075 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1885cbd6-ea2e-4d89-b9be-5d8ab9d8d959-catalog-content\") pod \"redhat-operators-sftcn\" (UID: \"1885cbd6-ea2e-4d89-b9be-5d8ab9d8d959\") " pod="openshift-marketplace/redhat-operators-sftcn" Mar 13 14:20:17 crc kubenswrapper[4907]: I0313 14:20:17.391534 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1885cbd6-ea2e-4d89-b9be-5d8ab9d8d959-utilities\") pod \"redhat-operators-sftcn\" (UID: \"1885cbd6-ea2e-4d89-b9be-5d8ab9d8d959\") " pod="openshift-marketplace/redhat-operators-sftcn" Mar 13 14:20:17 crc kubenswrapper[4907]: I0313 14:20:17.391590 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1885cbd6-ea2e-4d89-b9be-5d8ab9d8d959-catalog-content\") pod \"redhat-operators-sftcn\" (UID: \"1885cbd6-ea2e-4d89-b9be-5d8ab9d8d959\") " pod="openshift-marketplace/redhat-operators-sftcn" Mar 13 14:20:17 crc kubenswrapper[4907]: I0313 14:20:17.391673 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zp4z9\" (UniqueName: \"kubernetes.io/projected/1885cbd6-ea2e-4d89-b9be-5d8ab9d8d959-kube-api-access-zp4z9\") pod \"redhat-operators-sftcn\" (UID: \"1885cbd6-ea2e-4d89-b9be-5d8ab9d8d959\") " pod="openshift-marketplace/redhat-operators-sftcn" Mar 13 14:20:17 crc kubenswrapper[4907]: I0313 14:20:17.392175 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1885cbd6-ea2e-4d89-b9be-5d8ab9d8d959-utilities\") pod \"redhat-operators-sftcn\" (UID: \"1885cbd6-ea2e-4d89-b9be-5d8ab9d8d959\") " pod="openshift-marketplace/redhat-operators-sftcn" Mar 13 14:20:17 crc kubenswrapper[4907]: I0313 14:20:17.392213 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1885cbd6-ea2e-4d89-b9be-5d8ab9d8d959-catalog-content\") pod \"redhat-operators-sftcn\" (UID: \"1885cbd6-ea2e-4d89-b9be-5d8ab9d8d959\") " pod="openshift-marketplace/redhat-operators-sftcn" Mar 13 14:20:17 crc kubenswrapper[4907]: I0313 14:20:17.410452 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zp4z9\" (UniqueName: \"kubernetes.io/projected/1885cbd6-ea2e-4d89-b9be-5d8ab9d8d959-kube-api-access-zp4z9\") pod \"redhat-operators-sftcn\" (UID: \"1885cbd6-ea2e-4d89-b9be-5d8ab9d8d959\") " pod="openshift-marketplace/redhat-operators-sftcn" Mar 13 14:20:17 crc kubenswrapper[4907]: I0313 14:20:17.433072 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sftcn" Mar 13 14:20:17 crc kubenswrapper[4907]: I0313 14:20:17.658068 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-sftcn"] Mar 13 14:20:17 crc kubenswrapper[4907]: W0313 14:20:17.666041 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1885cbd6_ea2e_4d89_b9be_5d8ab9d8d959.slice/crio-32ac1d4a8e7dc17b286bf397735fbaec81ae96dd50b7a9ab89b1fd768ac31fd8 WatchSource:0}: Error finding container 32ac1d4a8e7dc17b286bf397735fbaec81ae96dd50b7a9ab89b1fd768ac31fd8: Status 404 returned error can't find the container with id 32ac1d4a8e7dc17b286bf397735fbaec81ae96dd50b7a9ab89b1fd768ac31fd8 Mar 13 14:20:18 crc kubenswrapper[4907]: I0313 14:20:18.204510 4907 generic.go:334] "Generic (PLEG): container finished" podID="1885cbd6-ea2e-4d89-b9be-5d8ab9d8d959" containerID="c6fc8c78c0d2400b40d785366f4fdd78a06e3d1fb3c60436c40f635cf1fa15ee" exitCode=0 Mar 13 14:20:18 crc kubenswrapper[4907]: I0313 14:20:18.204624 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sftcn" event={"ID":"1885cbd6-ea2e-4d89-b9be-5d8ab9d8d959","Type":"ContainerDied","Data":"c6fc8c78c0d2400b40d785366f4fdd78a06e3d1fb3c60436c40f635cf1fa15ee"} Mar 13 14:20:18 crc kubenswrapper[4907]: I0313 14:20:18.204969 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sftcn" event={"ID":"1885cbd6-ea2e-4d89-b9be-5d8ab9d8d959","Type":"ContainerStarted","Data":"32ac1d4a8e7dc17b286bf397735fbaec81ae96dd50b7a9ab89b1fd768ac31fd8"} Mar 13 14:20:18 crc kubenswrapper[4907]: I0313 14:20:18.206925 4907 generic.go:334] "Generic (PLEG): container finished" podID="a387235c-ef5d-47cc-8dff-8af5e8b5690f" containerID="b1907d331967d1a964d1cfc7f63d248b5cfe9d2b0d45e04047cb097b8077faf3" exitCode=0 Mar 13 14:20:18 crc kubenswrapper[4907]: I0313 14:20:18.206948 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874cgs5s" event={"ID":"a387235c-ef5d-47cc-8dff-8af5e8b5690f","Type":"ContainerDied","Data":"b1907d331967d1a964d1cfc7f63d248b5cfe9d2b0d45e04047cb097b8077faf3"} Mar 13 14:20:19 crc kubenswrapper[4907]: I0313 14:20:19.217560 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sftcn" event={"ID":"1885cbd6-ea2e-4d89-b9be-5d8ab9d8d959","Type":"ContainerStarted","Data":"df47c4a79801f3d5add4de851978c2a6b00163c4440a6a765aebda87af0252bc"} Mar 13 14:20:19 crc kubenswrapper[4907]: I0313 14:20:19.222458 4907 generic.go:334] "Generic (PLEG): container finished" podID="a387235c-ef5d-47cc-8dff-8af5e8b5690f" containerID="2e38829b5a9579bee5bd078fe0fab962d0e0f2f365a6b75024fe01a2cd8d9e8a" exitCode=0 Mar 13 14:20:19 crc kubenswrapper[4907]: I0313 14:20:19.222584 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874cgs5s" event={"ID":"a387235c-ef5d-47cc-8dff-8af5e8b5690f","Type":"ContainerDied","Data":"2e38829b5a9579bee5bd078fe0fab962d0e0f2f365a6b75024fe01a2cd8d9e8a"} Mar 13 14:20:20 crc kubenswrapper[4907]: I0313 14:20:20.231556 4907 generic.go:334] "Generic (PLEG): container finished" podID="1885cbd6-ea2e-4d89-b9be-5d8ab9d8d959" containerID="df47c4a79801f3d5add4de851978c2a6b00163c4440a6a765aebda87af0252bc" exitCode=0 Mar 13 14:20:20 crc kubenswrapper[4907]: I0313 14:20:20.231663 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sftcn" event={"ID":"1885cbd6-ea2e-4d89-b9be-5d8ab9d8d959","Type":"ContainerDied","Data":"df47c4a79801f3d5add4de851978c2a6b00163c4440a6a765aebda87af0252bc"} Mar 13 14:20:20 crc kubenswrapper[4907]: I0313 14:20:20.629352 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874cgs5s" Mar 13 14:20:20 crc kubenswrapper[4907]: I0313 14:20:20.832478 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qrhj4\" (UniqueName: \"kubernetes.io/projected/a387235c-ef5d-47cc-8dff-8af5e8b5690f-kube-api-access-qrhj4\") pod \"a387235c-ef5d-47cc-8dff-8af5e8b5690f\" (UID: \"a387235c-ef5d-47cc-8dff-8af5e8b5690f\") " Mar 13 14:20:20 crc kubenswrapper[4907]: I0313 14:20:20.832711 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a387235c-ef5d-47cc-8dff-8af5e8b5690f-bundle\") pod \"a387235c-ef5d-47cc-8dff-8af5e8b5690f\" (UID: \"a387235c-ef5d-47cc-8dff-8af5e8b5690f\") " Mar 13 14:20:20 crc kubenswrapper[4907]: I0313 14:20:20.832803 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a387235c-ef5d-47cc-8dff-8af5e8b5690f-util\") pod \"a387235c-ef5d-47cc-8dff-8af5e8b5690f\" (UID: \"a387235c-ef5d-47cc-8dff-8af5e8b5690f\") " Mar 13 14:20:20 crc kubenswrapper[4907]: I0313 14:20:20.833569 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a387235c-ef5d-47cc-8dff-8af5e8b5690f-bundle" (OuterVolumeSpecName: "bundle") pod "a387235c-ef5d-47cc-8dff-8af5e8b5690f" (UID: "a387235c-ef5d-47cc-8dff-8af5e8b5690f"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:20:20 crc kubenswrapper[4907]: I0313 14:20:20.839641 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a387235c-ef5d-47cc-8dff-8af5e8b5690f-kube-api-access-qrhj4" (OuterVolumeSpecName: "kube-api-access-qrhj4") pod "a387235c-ef5d-47cc-8dff-8af5e8b5690f" (UID: "a387235c-ef5d-47cc-8dff-8af5e8b5690f"). InnerVolumeSpecName "kube-api-access-qrhj4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:20:20 crc kubenswrapper[4907]: I0313 14:20:20.852535 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a387235c-ef5d-47cc-8dff-8af5e8b5690f-util" (OuterVolumeSpecName: "util") pod "a387235c-ef5d-47cc-8dff-8af5e8b5690f" (UID: "a387235c-ef5d-47cc-8dff-8af5e8b5690f"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:20:20 crc kubenswrapper[4907]: I0313 14:20:20.934501 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qrhj4\" (UniqueName: \"kubernetes.io/projected/a387235c-ef5d-47cc-8dff-8af5e8b5690f-kube-api-access-qrhj4\") on node \"crc\" DevicePath \"\"" Mar 13 14:20:20 crc kubenswrapper[4907]: I0313 14:20:20.934537 4907 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a387235c-ef5d-47cc-8dff-8af5e8b5690f-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:20:20 crc kubenswrapper[4907]: I0313 14:20:20.934550 4907 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a387235c-ef5d-47cc-8dff-8af5e8b5690f-util\") on node \"crc\" DevicePath \"\"" Mar 13 14:20:21 crc kubenswrapper[4907]: I0313 14:20:21.239661 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sftcn" event={"ID":"1885cbd6-ea2e-4d89-b9be-5d8ab9d8d959","Type":"ContainerStarted","Data":"a9ac0599c640beb63d63a6aae7a564ac63e2bcc0d12ec69b4aa940c25b9e8356"} Mar 13 14:20:21 crc kubenswrapper[4907]: I0313 14:20:21.243545 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874cgs5s" event={"ID":"a387235c-ef5d-47cc-8dff-8af5e8b5690f","Type":"ContainerDied","Data":"9144c72293eb14856491b3ee7a3a07c84f6294a348c2752001d7e79b3abf4e79"} Mar 13 14:20:21 crc kubenswrapper[4907]: I0313 14:20:21.243597 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9144c72293eb14856491b3ee7a3a07c84f6294a348c2752001d7e79b3abf4e79" Mar 13 14:20:21 crc kubenswrapper[4907]: I0313 14:20:21.243609 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874cgs5s" Mar 13 14:20:21 crc kubenswrapper[4907]: I0313 14:20:21.676047 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-sftcn" podStartSLOduration=2.083378881 podStartE2EDuration="4.676028288s" podCreationTimestamp="2026-03-13 14:20:17 +0000 UTC" firstStartedPulling="2026-03-13 14:20:18.20620721 +0000 UTC m=+917.105994899" lastFinishedPulling="2026-03-13 14:20:20.798856607 +0000 UTC m=+919.698644306" observedRunningTime="2026-03-13 14:20:21.258513739 +0000 UTC m=+920.158301448" watchObservedRunningTime="2026-03-13 14:20:21.676028288 +0000 UTC m=+920.575815977" Mar 13 14:20:24 crc kubenswrapper[4907]: I0313 14:20:24.994393 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-796d4cfff4-jrdzh"] Mar 13 14:20:24 crc kubenswrapper[4907]: E0313 14:20:24.995267 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a387235c-ef5d-47cc-8dff-8af5e8b5690f" containerName="extract" Mar 13 14:20:24 crc kubenswrapper[4907]: I0313 14:20:24.995289 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="a387235c-ef5d-47cc-8dff-8af5e8b5690f" containerName="extract" Mar 13 14:20:24 crc kubenswrapper[4907]: E0313 14:20:24.995316 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a387235c-ef5d-47cc-8dff-8af5e8b5690f" containerName="util" Mar 13 14:20:24 crc kubenswrapper[4907]: I0313 14:20:24.995324 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="a387235c-ef5d-47cc-8dff-8af5e8b5690f" containerName="util" Mar 13 14:20:24 crc kubenswrapper[4907]: E0313 14:20:24.995344 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a387235c-ef5d-47cc-8dff-8af5e8b5690f" containerName="pull" Mar 13 14:20:24 crc kubenswrapper[4907]: I0313 14:20:24.995353 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="a387235c-ef5d-47cc-8dff-8af5e8b5690f" containerName="pull" Mar 13 14:20:24 crc kubenswrapper[4907]: I0313 14:20:24.995476 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="a387235c-ef5d-47cc-8dff-8af5e8b5690f" containerName="extract" Mar 13 14:20:24 crc kubenswrapper[4907]: I0313 14:20:24.996060 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-796d4cfff4-jrdzh" Mar 13 14:20:24 crc kubenswrapper[4907]: I0313 14:20:24.997963 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Mar 13 14:20:24 crc kubenswrapper[4907]: I0313 14:20:24.998332 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-4xlqd" Mar 13 14:20:24 crc kubenswrapper[4907]: I0313 14:20:24.998382 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Mar 13 14:20:25 crc kubenswrapper[4907]: I0313 14:20:25.031417 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-796d4cfff4-jrdzh"] Mar 13 14:20:25 crc kubenswrapper[4907]: I0313 14:20:25.188192 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vjjcp\" (UniqueName: \"kubernetes.io/projected/443989a2-3471-4c69-a460-c441d90064fe-kube-api-access-vjjcp\") pod \"nmstate-operator-796d4cfff4-jrdzh\" (UID: \"443989a2-3471-4c69-a460-c441d90064fe\") " pod="openshift-nmstate/nmstate-operator-796d4cfff4-jrdzh" Mar 13 14:20:25 crc kubenswrapper[4907]: I0313 14:20:25.289082 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vjjcp\" (UniqueName: \"kubernetes.io/projected/443989a2-3471-4c69-a460-c441d90064fe-kube-api-access-vjjcp\") pod \"nmstate-operator-796d4cfff4-jrdzh\" (UID: \"443989a2-3471-4c69-a460-c441d90064fe\") " pod="openshift-nmstate/nmstate-operator-796d4cfff4-jrdzh" Mar 13 14:20:25 crc kubenswrapper[4907]: I0313 14:20:25.308579 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vjjcp\" (UniqueName: \"kubernetes.io/projected/443989a2-3471-4c69-a460-c441d90064fe-kube-api-access-vjjcp\") pod \"nmstate-operator-796d4cfff4-jrdzh\" (UID: \"443989a2-3471-4c69-a460-c441d90064fe\") " pod="openshift-nmstate/nmstate-operator-796d4cfff4-jrdzh" Mar 13 14:20:25 crc kubenswrapper[4907]: I0313 14:20:25.366593 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-796d4cfff4-jrdzh" Mar 13 14:20:25 crc kubenswrapper[4907]: I0313 14:20:25.602756 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-796d4cfff4-jrdzh"] Mar 13 14:20:26 crc kubenswrapper[4907]: I0313 14:20:26.269896 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-796d4cfff4-jrdzh" event={"ID":"443989a2-3471-4c69-a460-c441d90064fe","Type":"ContainerStarted","Data":"8a754bbd1938636e6c7ebb1e6c1879692f0daf6d651e0cbf15f733a870f4a158"} Mar 13 14:20:27 crc kubenswrapper[4907]: I0313 14:20:27.434160 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-sftcn" Mar 13 14:20:27 crc kubenswrapper[4907]: I0313 14:20:27.434210 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-sftcn" Mar 13 14:20:28 crc kubenswrapper[4907]: I0313 14:20:28.484251 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-sftcn" podUID="1885cbd6-ea2e-4d89-b9be-5d8ab9d8d959" containerName="registry-server" probeResult="failure" output=< Mar 13 14:20:28 crc kubenswrapper[4907]: timeout: failed to connect service ":50051" within 1s Mar 13 14:20:28 crc kubenswrapper[4907]: > Mar 13 14:20:30 crc kubenswrapper[4907]: I0313 14:20:30.297659 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-796d4cfff4-jrdzh" event={"ID":"443989a2-3471-4c69-a460-c441d90064fe","Type":"ContainerStarted","Data":"24f35f39833c1827d4c2274e72061a384384a16f92297d22076b13413798882e"} Mar 13 14:20:30 crc kubenswrapper[4907]: I0313 14:20:30.316422 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-796d4cfff4-jrdzh" podStartSLOduration=2.273083783 podStartE2EDuration="6.316399174s" podCreationTimestamp="2026-03-13 14:20:24 +0000 UTC" firstStartedPulling="2026-03-13 14:20:25.62311446 +0000 UTC m=+924.522902149" lastFinishedPulling="2026-03-13 14:20:29.666429851 +0000 UTC m=+928.566217540" observedRunningTime="2026-03-13 14:20:30.313464662 +0000 UTC m=+929.213252351" watchObservedRunningTime="2026-03-13 14:20:30.316399174 +0000 UTC m=+929.216186863" Mar 13 14:20:34 crc kubenswrapper[4907]: I0313 14:20:34.850873 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-9b8c8685d-fnz7h"] Mar 13 14:20:34 crc kubenswrapper[4907]: I0313 14:20:34.852317 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-9b8c8685d-fnz7h" Mar 13 14:20:34 crc kubenswrapper[4907]: I0313 14:20:34.854199 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-nrmnv" Mar 13 14:20:34 crc kubenswrapper[4907]: I0313 14:20:34.866726 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-5f558f5558-cqjnn"] Mar 13 14:20:34 crc kubenswrapper[4907]: I0313 14:20:34.867604 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f558f5558-cqjnn" Mar 13 14:20:34 crc kubenswrapper[4907]: I0313 14:20:34.874548 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Mar 13 14:20:34 crc kubenswrapper[4907]: I0313 14:20:34.904265 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-9b8c8685d-fnz7h"] Mar 13 14:20:34 crc kubenswrapper[4907]: I0313 14:20:34.913075 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f558f5558-cqjnn"] Mar 13 14:20:34 crc kubenswrapper[4907]: I0313 14:20:34.919995 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-4drs7"] Mar 13 14:20:34 crc kubenswrapper[4907]: I0313 14:20:34.920764 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-4drs7" Mar 13 14:20:34 crc kubenswrapper[4907]: I0313 14:20:34.979257 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-86f58fcf4-wrb6h"] Mar 13 14:20:34 crc kubenswrapper[4907]: I0313 14:20:34.980089 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-wrb6h" Mar 13 14:20:34 crc kubenswrapper[4907]: I0313 14:20:34.984429 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-86f58fcf4-wrb6h"] Mar 13 14:20:34 crc kubenswrapper[4907]: I0313 14:20:34.985440 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Mar 13 14:20:34 crc kubenswrapper[4907]: I0313 14:20:34.985649 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Mar 13 14:20:34 crc kubenswrapper[4907]: I0313 14:20:34.985830 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-vxrgq" Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.017409 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qplkl\" (UniqueName: \"kubernetes.io/projected/57da19c3-e374-4e40-911a-459c1d9da8f7-kube-api-access-qplkl\") pod \"nmstate-webhook-5f558f5558-cqjnn\" (UID: \"57da19c3-e374-4e40-911a-459c1d9da8f7\") " pod="openshift-nmstate/nmstate-webhook-5f558f5558-cqjnn" Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.017460 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6894k\" (UniqueName: \"kubernetes.io/projected/5e672ba1-22be-4fb4-b11c-3aa6a2b69c81-kube-api-access-6894k\") pod \"nmstate-metrics-9b8c8685d-fnz7h\" (UID: \"5e672ba1-22be-4fb4-b11c-3aa6a2b69c81\") " pod="openshift-nmstate/nmstate-metrics-9b8c8685d-fnz7h" Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.017737 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/57da19c3-e374-4e40-911a-459c1d9da8f7-tls-key-pair\") pod \"nmstate-webhook-5f558f5558-cqjnn\" (UID: \"57da19c3-e374-4e40-911a-459c1d9da8f7\") " pod="openshift-nmstate/nmstate-webhook-5f558f5558-cqjnn" Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.119458 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qplkl\" (UniqueName: \"kubernetes.io/projected/57da19c3-e374-4e40-911a-459c1d9da8f7-kube-api-access-qplkl\") pod \"nmstate-webhook-5f558f5558-cqjnn\" (UID: \"57da19c3-e374-4e40-911a-459c1d9da8f7\") " pod="openshift-nmstate/nmstate-webhook-5f558f5558-cqjnn" Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.119594 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zs2zt\" (UniqueName: \"kubernetes.io/projected/46a0badb-4ddd-4f72-a6ae-c415a5485062-kube-api-access-zs2zt\") pod \"nmstate-console-plugin-86f58fcf4-wrb6h\" (UID: \"46a0badb-4ddd-4f72-a6ae-c415a5485062\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-wrb6h" Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.119654 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6894k\" (UniqueName: \"kubernetes.io/projected/5e672ba1-22be-4fb4-b11c-3aa6a2b69c81-kube-api-access-6894k\") pod \"nmstate-metrics-9b8c8685d-fnz7h\" (UID: \"5e672ba1-22be-4fb4-b11c-3aa6a2b69c81\") " pod="openshift-nmstate/nmstate-metrics-9b8c8685d-fnz7h" Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.119727 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/f945fc96-0cb3-469c-ac53-8a37e0b01f51-nmstate-lock\") pod \"nmstate-handler-4drs7\" (UID: \"f945fc96-0cb3-469c-ac53-8a37e0b01f51\") " pod="openshift-nmstate/nmstate-handler-4drs7" Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.119828 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/46a0badb-4ddd-4f72-a6ae-c415a5485062-nginx-conf\") pod \"nmstate-console-plugin-86f58fcf4-wrb6h\" (UID: \"46a0badb-4ddd-4f72-a6ae-c415a5485062\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-wrb6h" Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.119910 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4m246\" (UniqueName: \"kubernetes.io/projected/f945fc96-0cb3-469c-ac53-8a37e0b01f51-kube-api-access-4m246\") pod \"nmstate-handler-4drs7\" (UID: \"f945fc96-0cb3-469c-ac53-8a37e0b01f51\") " pod="openshift-nmstate/nmstate-handler-4drs7" Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.119968 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/57da19c3-e374-4e40-911a-459c1d9da8f7-tls-key-pair\") pod \"nmstate-webhook-5f558f5558-cqjnn\" (UID: \"57da19c3-e374-4e40-911a-459c1d9da8f7\") " pod="openshift-nmstate/nmstate-webhook-5f558f5558-cqjnn" Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.120021 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/f945fc96-0cb3-469c-ac53-8a37e0b01f51-ovs-socket\") pod \"nmstate-handler-4drs7\" (UID: \"f945fc96-0cb3-469c-ac53-8a37e0b01f51\") " pod="openshift-nmstate/nmstate-handler-4drs7" Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.120141 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/f945fc96-0cb3-469c-ac53-8a37e0b01f51-dbus-socket\") pod \"nmstate-handler-4drs7\" (UID: \"f945fc96-0cb3-469c-ac53-8a37e0b01f51\") " pod="openshift-nmstate/nmstate-handler-4drs7" Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.120184 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/46a0badb-4ddd-4f72-a6ae-c415a5485062-plugin-serving-cert\") pod \"nmstate-console-plugin-86f58fcf4-wrb6h\" (UID: \"46a0badb-4ddd-4f72-a6ae-c415a5485062\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-wrb6h" Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.137904 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/57da19c3-e374-4e40-911a-459c1d9da8f7-tls-key-pair\") pod \"nmstate-webhook-5f558f5558-cqjnn\" (UID: \"57da19c3-e374-4e40-911a-459c1d9da8f7\") " pod="openshift-nmstate/nmstate-webhook-5f558f5558-cqjnn" Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.144161 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qplkl\" (UniqueName: \"kubernetes.io/projected/57da19c3-e374-4e40-911a-459c1d9da8f7-kube-api-access-qplkl\") pod \"nmstate-webhook-5f558f5558-cqjnn\" (UID: \"57da19c3-e374-4e40-911a-459c1d9da8f7\") " pod="openshift-nmstate/nmstate-webhook-5f558f5558-cqjnn" Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.147628 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6894k\" (UniqueName: \"kubernetes.io/projected/5e672ba1-22be-4fb4-b11c-3aa6a2b69c81-kube-api-access-6894k\") pod \"nmstate-metrics-9b8c8685d-fnz7h\" (UID: \"5e672ba1-22be-4fb4-b11c-3aa6a2b69c81\") " pod="openshift-nmstate/nmstate-metrics-9b8c8685d-fnz7h" Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.171383 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-9b8c8685d-fnz7h" Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.199350 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f558f5558-cqjnn" Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.220117 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-65d688f545-w8jgd"] Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.220902 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-65d688f545-w8jgd" Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.221100 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zs2zt\" (UniqueName: \"kubernetes.io/projected/46a0badb-4ddd-4f72-a6ae-c415a5485062-kube-api-access-zs2zt\") pod \"nmstate-console-plugin-86f58fcf4-wrb6h\" (UID: \"46a0badb-4ddd-4f72-a6ae-c415a5485062\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-wrb6h" Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.221160 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/f945fc96-0cb3-469c-ac53-8a37e0b01f51-nmstate-lock\") pod \"nmstate-handler-4drs7\" (UID: \"f945fc96-0cb3-469c-ac53-8a37e0b01f51\") " pod="openshift-nmstate/nmstate-handler-4drs7" Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.221195 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/46a0badb-4ddd-4f72-a6ae-c415a5485062-nginx-conf\") pod \"nmstate-console-plugin-86f58fcf4-wrb6h\" (UID: \"46a0badb-4ddd-4f72-a6ae-c415a5485062\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-wrb6h" Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.221217 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4m246\" (UniqueName: \"kubernetes.io/projected/f945fc96-0cb3-469c-ac53-8a37e0b01f51-kube-api-access-4m246\") pod \"nmstate-handler-4drs7\" (UID: \"f945fc96-0cb3-469c-ac53-8a37e0b01f51\") " pod="openshift-nmstate/nmstate-handler-4drs7" Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.221267 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/f945fc96-0cb3-469c-ac53-8a37e0b01f51-ovs-socket\") pod \"nmstate-handler-4drs7\" (UID: \"f945fc96-0cb3-469c-ac53-8a37e0b01f51\") " pod="openshift-nmstate/nmstate-handler-4drs7" Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.221342 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/f945fc96-0cb3-469c-ac53-8a37e0b01f51-nmstate-lock\") pod \"nmstate-handler-4drs7\" (UID: \"f945fc96-0cb3-469c-ac53-8a37e0b01f51\") " pod="openshift-nmstate/nmstate-handler-4drs7" Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.221849 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/f945fc96-0cb3-469c-ac53-8a37e0b01f51-ovs-socket\") pod \"nmstate-handler-4drs7\" (UID: \"f945fc96-0cb3-469c-ac53-8a37e0b01f51\") " pod="openshift-nmstate/nmstate-handler-4drs7" Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.222705 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/46a0badb-4ddd-4f72-a6ae-c415a5485062-nginx-conf\") pod \"nmstate-console-plugin-86f58fcf4-wrb6h\" (UID: \"46a0badb-4ddd-4f72-a6ae-c415a5485062\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-wrb6h" Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.222765 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/f945fc96-0cb3-469c-ac53-8a37e0b01f51-dbus-socket\") pod \"nmstate-handler-4drs7\" (UID: \"f945fc96-0cb3-469c-ac53-8a37e0b01f51\") " pod="openshift-nmstate/nmstate-handler-4drs7" Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.222792 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/46a0badb-4ddd-4f72-a6ae-c415a5485062-plugin-serving-cert\") pod \"nmstate-console-plugin-86f58fcf4-wrb6h\" (UID: \"46a0badb-4ddd-4f72-a6ae-c415a5485062\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-wrb6h" Mar 13 14:20:35 crc kubenswrapper[4907]: E0313 14:20:35.222963 4907 secret.go:188] Couldn't get secret openshift-nmstate/plugin-serving-cert: secret "plugin-serving-cert" not found Mar 13 14:20:35 crc kubenswrapper[4907]: E0313 14:20:35.223029 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/46a0badb-4ddd-4f72-a6ae-c415a5485062-plugin-serving-cert podName:46a0badb-4ddd-4f72-a6ae-c415a5485062 nodeName:}" failed. No retries permitted until 2026-03-13 14:20:35.723010738 +0000 UTC m=+934.622798427 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "plugin-serving-cert" (UniqueName: "kubernetes.io/secret/46a0badb-4ddd-4f72-a6ae-c415a5485062-plugin-serving-cert") pod "nmstate-console-plugin-86f58fcf4-wrb6h" (UID: "46a0badb-4ddd-4f72-a6ae-c415a5485062") : secret "plugin-serving-cert" not found Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.223096 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/f945fc96-0cb3-469c-ac53-8a37e0b01f51-dbus-socket\") pod \"nmstate-handler-4drs7\" (UID: \"f945fc96-0cb3-469c-ac53-8a37e0b01f51\") " pod="openshift-nmstate/nmstate-handler-4drs7" Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.238324 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-65d688f545-w8jgd"] Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.248091 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zs2zt\" (UniqueName: \"kubernetes.io/projected/46a0badb-4ddd-4f72-a6ae-c415a5485062-kube-api-access-zs2zt\") pod \"nmstate-console-plugin-86f58fcf4-wrb6h\" (UID: \"46a0badb-4ddd-4f72-a6ae-c415a5485062\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-wrb6h" Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.248093 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4m246\" (UniqueName: \"kubernetes.io/projected/f945fc96-0cb3-469c-ac53-8a37e0b01f51-kube-api-access-4m246\") pod \"nmstate-handler-4drs7\" (UID: \"f945fc96-0cb3-469c-ac53-8a37e0b01f51\") " pod="openshift-nmstate/nmstate-handler-4drs7" Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.325247 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/25a755e0-d2c0-414f-8a68-aa0c9c28079c-console-oauth-config\") pod \"console-65d688f545-w8jgd\" (UID: \"25a755e0-d2c0-414f-8a68-aa0c9c28079c\") " pod="openshift-console/console-65d688f545-w8jgd" Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.326185 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/25a755e0-d2c0-414f-8a68-aa0c9c28079c-oauth-serving-cert\") pod \"console-65d688f545-w8jgd\" (UID: \"25a755e0-d2c0-414f-8a68-aa0c9c28079c\") " pod="openshift-console/console-65d688f545-w8jgd" Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.326231 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/25a755e0-d2c0-414f-8a68-aa0c9c28079c-trusted-ca-bundle\") pod \"console-65d688f545-w8jgd\" (UID: \"25a755e0-d2c0-414f-8a68-aa0c9c28079c\") " pod="openshift-console/console-65d688f545-w8jgd" Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.326308 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wv7v8\" (UniqueName: \"kubernetes.io/projected/25a755e0-d2c0-414f-8a68-aa0c9c28079c-kube-api-access-wv7v8\") pod \"console-65d688f545-w8jgd\" (UID: \"25a755e0-d2c0-414f-8a68-aa0c9c28079c\") " pod="openshift-console/console-65d688f545-w8jgd" Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.326335 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/25a755e0-d2c0-414f-8a68-aa0c9c28079c-service-ca\") pod \"console-65d688f545-w8jgd\" (UID: \"25a755e0-d2c0-414f-8a68-aa0c9c28079c\") " pod="openshift-console/console-65d688f545-w8jgd" Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.326374 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/25a755e0-d2c0-414f-8a68-aa0c9c28079c-console-serving-cert\") pod \"console-65d688f545-w8jgd\" (UID: \"25a755e0-d2c0-414f-8a68-aa0c9c28079c\") " pod="openshift-console/console-65d688f545-w8jgd" Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.326392 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/25a755e0-d2c0-414f-8a68-aa0c9c28079c-console-config\") pod \"console-65d688f545-w8jgd\" (UID: \"25a755e0-d2c0-414f-8a68-aa0c9c28079c\") " pod="openshift-console/console-65d688f545-w8jgd" Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.427615 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wv7v8\" (UniqueName: \"kubernetes.io/projected/25a755e0-d2c0-414f-8a68-aa0c9c28079c-kube-api-access-wv7v8\") pod \"console-65d688f545-w8jgd\" (UID: \"25a755e0-d2c0-414f-8a68-aa0c9c28079c\") " pod="openshift-console/console-65d688f545-w8jgd" Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.427654 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/25a755e0-d2c0-414f-8a68-aa0c9c28079c-service-ca\") pod \"console-65d688f545-w8jgd\" (UID: \"25a755e0-d2c0-414f-8a68-aa0c9c28079c\") " pod="openshift-console/console-65d688f545-w8jgd" Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.427688 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/25a755e0-d2c0-414f-8a68-aa0c9c28079c-console-serving-cert\") pod \"console-65d688f545-w8jgd\" (UID: \"25a755e0-d2c0-414f-8a68-aa0c9c28079c\") " pod="openshift-console/console-65d688f545-w8jgd" Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.427704 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/25a755e0-d2c0-414f-8a68-aa0c9c28079c-console-config\") pod \"console-65d688f545-w8jgd\" (UID: \"25a755e0-d2c0-414f-8a68-aa0c9c28079c\") " pod="openshift-console/console-65d688f545-w8jgd" Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.427743 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/25a755e0-d2c0-414f-8a68-aa0c9c28079c-console-oauth-config\") pod \"console-65d688f545-w8jgd\" (UID: \"25a755e0-d2c0-414f-8a68-aa0c9c28079c\") " pod="openshift-console/console-65d688f545-w8jgd" Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.427766 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/25a755e0-d2c0-414f-8a68-aa0c9c28079c-oauth-serving-cert\") pod \"console-65d688f545-w8jgd\" (UID: \"25a755e0-d2c0-414f-8a68-aa0c9c28079c\") " pod="openshift-console/console-65d688f545-w8jgd" Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.427783 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/25a755e0-d2c0-414f-8a68-aa0c9c28079c-trusted-ca-bundle\") pod \"console-65d688f545-w8jgd\" (UID: \"25a755e0-d2c0-414f-8a68-aa0c9c28079c\") " pod="openshift-console/console-65d688f545-w8jgd" Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.428778 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/25a755e0-d2c0-414f-8a68-aa0c9c28079c-trusted-ca-bundle\") pod \"console-65d688f545-w8jgd\" (UID: \"25a755e0-d2c0-414f-8a68-aa0c9c28079c\") " pod="openshift-console/console-65d688f545-w8jgd" Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.428894 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/25a755e0-d2c0-414f-8a68-aa0c9c28079c-oauth-serving-cert\") pod \"console-65d688f545-w8jgd\" (UID: \"25a755e0-d2c0-414f-8a68-aa0c9c28079c\") " pod="openshift-console/console-65d688f545-w8jgd" Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.430223 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/25a755e0-d2c0-414f-8a68-aa0c9c28079c-console-config\") pod \"console-65d688f545-w8jgd\" (UID: \"25a755e0-d2c0-414f-8a68-aa0c9c28079c\") " pod="openshift-console/console-65d688f545-w8jgd" Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.431515 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/25a755e0-d2c0-414f-8a68-aa0c9c28079c-service-ca\") pod \"console-65d688f545-w8jgd\" (UID: \"25a755e0-d2c0-414f-8a68-aa0c9c28079c\") " pod="openshift-console/console-65d688f545-w8jgd" Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.433807 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/25a755e0-d2c0-414f-8a68-aa0c9c28079c-console-oauth-config\") pod \"console-65d688f545-w8jgd\" (UID: \"25a755e0-d2c0-414f-8a68-aa0c9c28079c\") " pod="openshift-console/console-65d688f545-w8jgd" Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.434526 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/25a755e0-d2c0-414f-8a68-aa0c9c28079c-console-serving-cert\") pod \"console-65d688f545-w8jgd\" (UID: \"25a755e0-d2c0-414f-8a68-aa0c9c28079c\") " pod="openshift-console/console-65d688f545-w8jgd" Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.444222 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wv7v8\" (UniqueName: \"kubernetes.io/projected/25a755e0-d2c0-414f-8a68-aa0c9c28079c-kube-api-access-wv7v8\") pod \"console-65d688f545-w8jgd\" (UID: \"25a755e0-d2c0-414f-8a68-aa0c9c28079c\") " pod="openshift-console/console-65d688f545-w8jgd" Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.456008 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f558f5558-cqjnn"] Mar 13 14:20:35 crc kubenswrapper[4907]: W0313 14:20:35.457796 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod57da19c3_e374_4e40_911a_459c1d9da8f7.slice/crio-9046362e8ce8a6b9344f48e54f7ce1f08c91443317b3ba413fbbc123ec101bc6 WatchSource:0}: Error finding container 9046362e8ce8a6b9344f48e54f7ce1f08c91443317b3ba413fbbc123ec101bc6: Status 404 returned error can't find the container with id 9046362e8ce8a6b9344f48e54f7ce1f08c91443317b3ba413fbbc123ec101bc6 Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.537595 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-4drs7" Mar 13 14:20:35 crc kubenswrapper[4907]: W0313 14:20:35.564512 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf945fc96_0cb3_469c_ac53_8a37e0b01f51.slice/crio-2dff306914b4121425b0b92bc2102810f2dd230dbcd3319f0293820db567826b WatchSource:0}: Error finding container 2dff306914b4121425b0b92bc2102810f2dd230dbcd3319f0293820db567826b: Status 404 returned error can't find the container with id 2dff306914b4121425b0b92bc2102810f2dd230dbcd3319f0293820db567826b Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.601033 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-65d688f545-w8jgd" Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.630688 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-9b8c8685d-fnz7h"] Mar 13 14:20:35 crc kubenswrapper[4907]: W0313 14:20:35.636292 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5e672ba1_22be_4fb4_b11c_3aa6a2b69c81.slice/crio-d63fd92a574dd16432e8cc6b1570153bb6e2b7cb09e0e9c81696e31884132f2e WatchSource:0}: Error finding container d63fd92a574dd16432e8cc6b1570153bb6e2b7cb09e0e9c81696e31884132f2e: Status 404 returned error can't find the container with id d63fd92a574dd16432e8cc6b1570153bb6e2b7cb09e0e9c81696e31884132f2e Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.730653 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/46a0badb-4ddd-4f72-a6ae-c415a5485062-plugin-serving-cert\") pod \"nmstate-console-plugin-86f58fcf4-wrb6h\" (UID: \"46a0badb-4ddd-4f72-a6ae-c415a5485062\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-wrb6h" Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.735626 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/46a0badb-4ddd-4f72-a6ae-c415a5485062-plugin-serving-cert\") pod \"nmstate-console-plugin-86f58fcf4-wrb6h\" (UID: \"46a0badb-4ddd-4f72-a6ae-c415a5485062\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-wrb6h" Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.800963 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-65d688f545-w8jgd"] Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.897502 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-wrb6h" Mar 13 14:20:36 crc kubenswrapper[4907]: I0313 14:20:36.106841 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-86f58fcf4-wrb6h"] Mar 13 14:20:36 crc kubenswrapper[4907]: W0313 14:20:36.111846 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod46a0badb_4ddd_4f72_a6ae_c415a5485062.slice/crio-2b51f1fbe6aaca4e97833e4f8fca8a3a1e0b6f3117495a6860581c66c8562ff7 WatchSource:0}: Error finding container 2b51f1fbe6aaca4e97833e4f8fca8a3a1e0b6f3117495a6860581c66c8562ff7: Status 404 returned error can't find the container with id 2b51f1fbe6aaca4e97833e4f8fca8a3a1e0b6f3117495a6860581c66c8562ff7 Mar 13 14:20:36 crc kubenswrapper[4907]: I0313 14:20:36.362734 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f558f5558-cqjnn" event={"ID":"57da19c3-e374-4e40-911a-459c1d9da8f7","Type":"ContainerStarted","Data":"9046362e8ce8a6b9344f48e54f7ce1f08c91443317b3ba413fbbc123ec101bc6"} Mar 13 14:20:36 crc kubenswrapper[4907]: I0313 14:20:36.364200 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-wrb6h" event={"ID":"46a0badb-4ddd-4f72-a6ae-c415a5485062","Type":"ContainerStarted","Data":"2b51f1fbe6aaca4e97833e4f8fca8a3a1e0b6f3117495a6860581c66c8562ff7"} Mar 13 14:20:36 crc kubenswrapper[4907]: I0313 14:20:36.365954 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-9b8c8685d-fnz7h" event={"ID":"5e672ba1-22be-4fb4-b11c-3aa6a2b69c81","Type":"ContainerStarted","Data":"d63fd92a574dd16432e8cc6b1570153bb6e2b7cb09e0e9c81696e31884132f2e"} Mar 13 14:20:36 crc kubenswrapper[4907]: I0313 14:20:36.368129 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-65d688f545-w8jgd" event={"ID":"25a755e0-d2c0-414f-8a68-aa0c9c28079c","Type":"ContainerStarted","Data":"6a7d64edf00f3d3262dc2c6224febcf7d60e0f32623e61efc8de67a598cf7646"} Mar 13 14:20:36 crc kubenswrapper[4907]: I0313 14:20:36.368161 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-65d688f545-w8jgd" event={"ID":"25a755e0-d2c0-414f-8a68-aa0c9c28079c","Type":"ContainerStarted","Data":"1af2fc91e62a0ba1fbf071d2fffc6640ff0e146852d1684462d2d89243a2698a"} Mar 13 14:20:36 crc kubenswrapper[4907]: I0313 14:20:36.369486 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-4drs7" event={"ID":"f945fc96-0cb3-469c-ac53-8a37e0b01f51","Type":"ContainerStarted","Data":"2dff306914b4121425b0b92bc2102810f2dd230dbcd3319f0293820db567826b"} Mar 13 14:20:36 crc kubenswrapper[4907]: I0313 14:20:36.389721 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-65d688f545-w8jgd" podStartSLOduration=1.389702041 podStartE2EDuration="1.389702041s" podCreationTimestamp="2026-03-13 14:20:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:20:36.386376939 +0000 UTC m=+935.286164618" watchObservedRunningTime="2026-03-13 14:20:36.389702041 +0000 UTC m=+935.289489750" Mar 13 14:20:37 crc kubenswrapper[4907]: I0313 14:20:37.474212 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-sftcn" Mar 13 14:20:37 crc kubenswrapper[4907]: I0313 14:20:37.523290 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-sftcn" Mar 13 14:20:37 crc kubenswrapper[4907]: I0313 14:20:37.703767 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-sftcn"] Mar 13 14:20:38 crc kubenswrapper[4907]: I0313 14:20:38.383455 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-4drs7" event={"ID":"f945fc96-0cb3-469c-ac53-8a37e0b01f51","Type":"ContainerStarted","Data":"b63fe2e822839caaa6054f85481e03baef52d46da7bbf472cfdd35feb6cc2ae7"} Mar 13 14:20:38 crc kubenswrapper[4907]: I0313 14:20:38.383791 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-4drs7" Mar 13 14:20:38 crc kubenswrapper[4907]: I0313 14:20:38.389615 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f558f5558-cqjnn" event={"ID":"57da19c3-e374-4e40-911a-459c1d9da8f7","Type":"ContainerStarted","Data":"5b7751fd2f4a05db76d4f555271c5c0b9ab1b77200a9a6a2dbf746f37e7d52a9"} Mar 13 14:20:38 crc kubenswrapper[4907]: I0313 14:20:38.390236 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-5f558f5558-cqjnn" Mar 13 14:20:38 crc kubenswrapper[4907]: I0313 14:20:38.391695 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-9b8c8685d-fnz7h" event={"ID":"5e672ba1-22be-4fb4-b11c-3aa6a2b69c81","Type":"ContainerStarted","Data":"0733092feefb0ff9118e05d8d0be492c02cc5f9c3b863e4cb7897ca1e6c58c29"} Mar 13 14:20:38 crc kubenswrapper[4907]: I0313 14:20:38.402409 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-4drs7" podStartSLOduration=2.022649089 podStartE2EDuration="4.402384158s" podCreationTimestamp="2026-03-13 14:20:34 +0000 UTC" firstStartedPulling="2026-03-13 14:20:35.567848402 +0000 UTC m=+934.467636121" lastFinishedPulling="2026-03-13 14:20:37.947583461 +0000 UTC m=+936.847371190" observedRunningTime="2026-03-13 14:20:38.39745281 +0000 UTC m=+937.297240499" watchObservedRunningTime="2026-03-13 14:20:38.402384158 +0000 UTC m=+937.302171867" Mar 13 14:20:38 crc kubenswrapper[4907]: I0313 14:20:38.419493 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-5f558f5558-cqjnn" podStartSLOduration=1.907904863 podStartE2EDuration="4.419473883s" podCreationTimestamp="2026-03-13 14:20:34 +0000 UTC" firstStartedPulling="2026-03-13 14:20:35.459468744 +0000 UTC m=+934.359256433" lastFinishedPulling="2026-03-13 14:20:37.971037764 +0000 UTC m=+936.870825453" observedRunningTime="2026-03-13 14:20:38.41324132 +0000 UTC m=+937.313029009" watchObservedRunningTime="2026-03-13 14:20:38.419473883 +0000 UTC m=+937.319261572" Mar 13 14:20:39 crc kubenswrapper[4907]: I0313 14:20:39.409578 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-wrb6h" event={"ID":"46a0badb-4ddd-4f72-a6ae-c415a5485062","Type":"ContainerStarted","Data":"0c762005459774c9f2cf4e2c09c71c5c6d8a5001c6e4b196411b93ad10ff0ce7"} Mar 13 14:20:39 crc kubenswrapper[4907]: I0313 14:20:39.410233 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-sftcn" podUID="1885cbd6-ea2e-4d89-b9be-5d8ab9d8d959" containerName="registry-server" containerID="cri-o://a9ac0599c640beb63d63a6aae7a564ac63e2bcc0d12ec69b4aa940c25b9e8356" gracePeriod=2 Mar 13 14:20:39 crc kubenswrapper[4907]: I0313 14:20:39.428247 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-wrb6h" podStartSLOduration=2.715792122 podStartE2EDuration="5.428202737s" podCreationTimestamp="2026-03-13 14:20:34 +0000 UTC" firstStartedPulling="2026-03-13 14:20:36.113839328 +0000 UTC m=+935.013627017" lastFinishedPulling="2026-03-13 14:20:38.826249933 +0000 UTC m=+937.726037632" observedRunningTime="2026-03-13 14:20:39.423488236 +0000 UTC m=+938.323275925" watchObservedRunningTime="2026-03-13 14:20:39.428202737 +0000 UTC m=+938.327990426" Mar 13 14:20:39 crc kubenswrapper[4907]: I0313 14:20:39.745810 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sftcn" Mar 13 14:20:39 crc kubenswrapper[4907]: I0313 14:20:39.788802 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zp4z9\" (UniqueName: \"kubernetes.io/projected/1885cbd6-ea2e-4d89-b9be-5d8ab9d8d959-kube-api-access-zp4z9\") pod \"1885cbd6-ea2e-4d89-b9be-5d8ab9d8d959\" (UID: \"1885cbd6-ea2e-4d89-b9be-5d8ab9d8d959\") " Mar 13 14:20:39 crc kubenswrapper[4907]: I0313 14:20:39.793928 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1885cbd6-ea2e-4d89-b9be-5d8ab9d8d959-kube-api-access-zp4z9" (OuterVolumeSpecName: "kube-api-access-zp4z9") pod "1885cbd6-ea2e-4d89-b9be-5d8ab9d8d959" (UID: "1885cbd6-ea2e-4d89-b9be-5d8ab9d8d959"). InnerVolumeSpecName "kube-api-access-zp4z9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:20:39 crc kubenswrapper[4907]: I0313 14:20:39.889683 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1885cbd6-ea2e-4d89-b9be-5d8ab9d8d959-catalog-content\") pod \"1885cbd6-ea2e-4d89-b9be-5d8ab9d8d959\" (UID: \"1885cbd6-ea2e-4d89-b9be-5d8ab9d8d959\") " Mar 13 14:20:39 crc kubenswrapper[4907]: I0313 14:20:39.889741 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1885cbd6-ea2e-4d89-b9be-5d8ab9d8d959-utilities\") pod \"1885cbd6-ea2e-4d89-b9be-5d8ab9d8d959\" (UID: \"1885cbd6-ea2e-4d89-b9be-5d8ab9d8d959\") " Mar 13 14:20:39 crc kubenswrapper[4907]: I0313 14:20:39.889956 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zp4z9\" (UniqueName: \"kubernetes.io/projected/1885cbd6-ea2e-4d89-b9be-5d8ab9d8d959-kube-api-access-zp4z9\") on node \"crc\" DevicePath \"\"" Mar 13 14:20:39 crc kubenswrapper[4907]: I0313 14:20:39.890642 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1885cbd6-ea2e-4d89-b9be-5d8ab9d8d959-utilities" (OuterVolumeSpecName: "utilities") pod "1885cbd6-ea2e-4d89-b9be-5d8ab9d8d959" (UID: "1885cbd6-ea2e-4d89-b9be-5d8ab9d8d959"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:20:39 crc kubenswrapper[4907]: I0313 14:20:39.991258 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1885cbd6-ea2e-4d89-b9be-5d8ab9d8d959-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 14:20:40 crc kubenswrapper[4907]: I0313 14:20:40.018820 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1885cbd6-ea2e-4d89-b9be-5d8ab9d8d959-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1885cbd6-ea2e-4d89-b9be-5d8ab9d8d959" (UID: "1885cbd6-ea2e-4d89-b9be-5d8ab9d8d959"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:20:40 crc kubenswrapper[4907]: I0313 14:20:40.092199 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1885cbd6-ea2e-4d89-b9be-5d8ab9d8d959-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 14:20:40 crc kubenswrapper[4907]: I0313 14:20:40.417434 4907 generic.go:334] "Generic (PLEG): container finished" podID="1885cbd6-ea2e-4d89-b9be-5d8ab9d8d959" containerID="a9ac0599c640beb63d63a6aae7a564ac63e2bcc0d12ec69b4aa940c25b9e8356" exitCode=0 Mar 13 14:20:40 crc kubenswrapper[4907]: I0313 14:20:40.418562 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sftcn" Mar 13 14:20:40 crc kubenswrapper[4907]: I0313 14:20:40.429952 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sftcn" event={"ID":"1885cbd6-ea2e-4d89-b9be-5d8ab9d8d959","Type":"ContainerDied","Data":"a9ac0599c640beb63d63a6aae7a564ac63e2bcc0d12ec69b4aa940c25b9e8356"} Mar 13 14:20:40 crc kubenswrapper[4907]: I0313 14:20:40.430018 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sftcn" event={"ID":"1885cbd6-ea2e-4d89-b9be-5d8ab9d8d959","Type":"ContainerDied","Data":"32ac1d4a8e7dc17b286bf397735fbaec81ae96dd50b7a9ab89b1fd768ac31fd8"} Mar 13 14:20:40 crc kubenswrapper[4907]: I0313 14:20:40.430048 4907 scope.go:117] "RemoveContainer" containerID="a9ac0599c640beb63d63a6aae7a564ac63e2bcc0d12ec69b4aa940c25b9e8356" Mar 13 14:20:40 crc kubenswrapper[4907]: I0313 14:20:40.457928 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-sftcn"] Mar 13 14:20:40 crc kubenswrapper[4907]: I0313 14:20:40.465251 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-sftcn"] Mar 13 14:20:40 crc kubenswrapper[4907]: I0313 14:20:40.467760 4907 scope.go:117] "RemoveContainer" containerID="df47c4a79801f3d5add4de851978c2a6b00163c4440a6a765aebda87af0252bc" Mar 13 14:20:40 crc kubenswrapper[4907]: I0313 14:20:40.516771 4907 scope.go:117] "RemoveContainer" containerID="c6fc8c78c0d2400b40d785366f4fdd78a06e3d1fb3c60436c40f635cf1fa15ee" Mar 13 14:20:40 crc kubenswrapper[4907]: I0313 14:20:40.530903 4907 scope.go:117] "RemoveContainer" containerID="a9ac0599c640beb63d63a6aae7a564ac63e2bcc0d12ec69b4aa940c25b9e8356" Mar 13 14:20:40 crc kubenswrapper[4907]: E0313 14:20:40.531259 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a9ac0599c640beb63d63a6aae7a564ac63e2bcc0d12ec69b4aa940c25b9e8356\": container with ID starting with a9ac0599c640beb63d63a6aae7a564ac63e2bcc0d12ec69b4aa940c25b9e8356 not found: ID does not exist" containerID="a9ac0599c640beb63d63a6aae7a564ac63e2bcc0d12ec69b4aa940c25b9e8356" Mar 13 14:20:40 crc kubenswrapper[4907]: I0313 14:20:40.531305 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a9ac0599c640beb63d63a6aae7a564ac63e2bcc0d12ec69b4aa940c25b9e8356"} err="failed to get container status \"a9ac0599c640beb63d63a6aae7a564ac63e2bcc0d12ec69b4aa940c25b9e8356\": rpc error: code = NotFound desc = could not find container \"a9ac0599c640beb63d63a6aae7a564ac63e2bcc0d12ec69b4aa940c25b9e8356\": container with ID starting with a9ac0599c640beb63d63a6aae7a564ac63e2bcc0d12ec69b4aa940c25b9e8356 not found: ID does not exist" Mar 13 14:20:40 crc kubenswrapper[4907]: I0313 14:20:40.531335 4907 scope.go:117] "RemoveContainer" containerID="df47c4a79801f3d5add4de851978c2a6b00163c4440a6a765aebda87af0252bc" Mar 13 14:20:40 crc kubenswrapper[4907]: E0313 14:20:40.531731 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"df47c4a79801f3d5add4de851978c2a6b00163c4440a6a765aebda87af0252bc\": container with ID starting with df47c4a79801f3d5add4de851978c2a6b00163c4440a6a765aebda87af0252bc not found: ID does not exist" containerID="df47c4a79801f3d5add4de851978c2a6b00163c4440a6a765aebda87af0252bc" Mar 13 14:20:40 crc kubenswrapper[4907]: I0313 14:20:40.531762 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"df47c4a79801f3d5add4de851978c2a6b00163c4440a6a765aebda87af0252bc"} err="failed to get container status \"df47c4a79801f3d5add4de851978c2a6b00163c4440a6a765aebda87af0252bc\": rpc error: code = NotFound desc = could not find container \"df47c4a79801f3d5add4de851978c2a6b00163c4440a6a765aebda87af0252bc\": container with ID starting with df47c4a79801f3d5add4de851978c2a6b00163c4440a6a765aebda87af0252bc not found: ID does not exist" Mar 13 14:20:40 crc kubenswrapper[4907]: I0313 14:20:40.532045 4907 scope.go:117] "RemoveContainer" containerID="c6fc8c78c0d2400b40d785366f4fdd78a06e3d1fb3c60436c40f635cf1fa15ee" Mar 13 14:20:40 crc kubenswrapper[4907]: E0313 14:20:40.532383 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c6fc8c78c0d2400b40d785366f4fdd78a06e3d1fb3c60436c40f635cf1fa15ee\": container with ID starting with c6fc8c78c0d2400b40d785366f4fdd78a06e3d1fb3c60436c40f635cf1fa15ee not found: ID does not exist" containerID="c6fc8c78c0d2400b40d785366f4fdd78a06e3d1fb3c60436c40f635cf1fa15ee" Mar 13 14:20:40 crc kubenswrapper[4907]: I0313 14:20:40.532414 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c6fc8c78c0d2400b40d785366f4fdd78a06e3d1fb3c60436c40f635cf1fa15ee"} err="failed to get container status \"c6fc8c78c0d2400b40d785366f4fdd78a06e3d1fb3c60436c40f635cf1fa15ee\": rpc error: code = NotFound desc = could not find container \"c6fc8c78c0d2400b40d785366f4fdd78a06e3d1fb3c60436c40f635cf1fa15ee\": container with ID starting with c6fc8c78c0d2400b40d785366f4fdd78a06e3d1fb3c60436c40f635cf1fa15ee not found: ID does not exist" Mar 13 14:20:41 crc kubenswrapper[4907]: I0313 14:20:41.431161 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-9b8c8685d-fnz7h" event={"ID":"5e672ba1-22be-4fb4-b11c-3aa6a2b69c81","Type":"ContainerStarted","Data":"f40113f178ce01366186258c102e3799aab0ed2e3849d2235a8b7cb62f5bc4a0"} Mar 13 14:20:41 crc kubenswrapper[4907]: I0313 14:20:41.461993 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-9b8c8685d-fnz7h" podStartSLOduration=2.570348044 podStartE2EDuration="7.461965831s" podCreationTimestamp="2026-03-13 14:20:34 +0000 UTC" firstStartedPulling="2026-03-13 14:20:35.641080592 +0000 UTC m=+934.540868281" lastFinishedPulling="2026-03-13 14:20:40.532698379 +0000 UTC m=+939.432486068" observedRunningTime="2026-03-13 14:20:41.456715044 +0000 UTC m=+940.356502743" watchObservedRunningTime="2026-03-13 14:20:41.461965831 +0000 UTC m=+940.361753570" Mar 13 14:20:41 crc kubenswrapper[4907]: I0313 14:20:41.797366 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1885cbd6-ea2e-4d89-b9be-5d8ab9d8d959" path="/var/lib/kubelet/pods/1885cbd6-ea2e-4d89-b9be-5d8ab9d8d959/volumes" Mar 13 14:20:45 crc kubenswrapper[4907]: I0313 14:20:45.562532 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-4drs7" Mar 13 14:20:45 crc kubenswrapper[4907]: I0313 14:20:45.602238 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-65d688f545-w8jgd" Mar 13 14:20:45 crc kubenswrapper[4907]: I0313 14:20:45.602301 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-65d688f545-w8jgd" Mar 13 14:20:45 crc kubenswrapper[4907]: I0313 14:20:45.606472 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-65d688f545-w8jgd" Mar 13 14:20:46 crc kubenswrapper[4907]: I0313 14:20:46.475608 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-65d688f545-w8jgd" Mar 13 14:20:46 crc kubenswrapper[4907]: I0313 14:20:46.517973 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-l9xpb"] Mar 13 14:20:53 crc kubenswrapper[4907]: I0313 14:20:53.008028 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-pqj8w"] Mar 13 14:20:53 crc kubenswrapper[4907]: E0313 14:20:53.008680 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1885cbd6-ea2e-4d89-b9be-5d8ab9d8d959" containerName="extract-utilities" Mar 13 14:20:53 crc kubenswrapper[4907]: I0313 14:20:53.008703 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="1885cbd6-ea2e-4d89-b9be-5d8ab9d8d959" containerName="extract-utilities" Mar 13 14:20:53 crc kubenswrapper[4907]: E0313 14:20:53.008724 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1885cbd6-ea2e-4d89-b9be-5d8ab9d8d959" containerName="registry-server" Mar 13 14:20:53 crc kubenswrapper[4907]: I0313 14:20:53.008735 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="1885cbd6-ea2e-4d89-b9be-5d8ab9d8d959" containerName="registry-server" Mar 13 14:20:53 crc kubenswrapper[4907]: E0313 14:20:53.008750 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1885cbd6-ea2e-4d89-b9be-5d8ab9d8d959" containerName="extract-content" Mar 13 14:20:53 crc kubenswrapper[4907]: I0313 14:20:53.008762 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="1885cbd6-ea2e-4d89-b9be-5d8ab9d8d959" containerName="extract-content" Mar 13 14:20:53 crc kubenswrapper[4907]: I0313 14:20:53.008958 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="1885cbd6-ea2e-4d89-b9be-5d8ab9d8d959" containerName="registry-server" Mar 13 14:20:53 crc kubenswrapper[4907]: I0313 14:20:53.010077 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pqj8w" Mar 13 14:20:53 crc kubenswrapper[4907]: I0313 14:20:53.025537 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-pqj8w"] Mar 13 14:20:53 crc kubenswrapper[4907]: I0313 14:20:53.164451 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb19cbd4-fd73-485d-9b1e-a9ed01fa73ac-utilities\") pod \"certified-operators-pqj8w\" (UID: \"cb19cbd4-fd73-485d-9b1e-a9ed01fa73ac\") " pod="openshift-marketplace/certified-operators-pqj8w" Mar 13 14:20:53 crc kubenswrapper[4907]: I0313 14:20:53.164525 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ch7ft\" (UniqueName: \"kubernetes.io/projected/cb19cbd4-fd73-485d-9b1e-a9ed01fa73ac-kube-api-access-ch7ft\") pod \"certified-operators-pqj8w\" (UID: \"cb19cbd4-fd73-485d-9b1e-a9ed01fa73ac\") " pod="openshift-marketplace/certified-operators-pqj8w" Mar 13 14:20:53 crc kubenswrapper[4907]: I0313 14:20:53.164573 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb19cbd4-fd73-485d-9b1e-a9ed01fa73ac-catalog-content\") pod \"certified-operators-pqj8w\" (UID: \"cb19cbd4-fd73-485d-9b1e-a9ed01fa73ac\") " pod="openshift-marketplace/certified-operators-pqj8w" Mar 13 14:20:53 crc kubenswrapper[4907]: I0313 14:20:53.265493 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb19cbd4-fd73-485d-9b1e-a9ed01fa73ac-utilities\") pod \"certified-operators-pqj8w\" (UID: \"cb19cbd4-fd73-485d-9b1e-a9ed01fa73ac\") " pod="openshift-marketplace/certified-operators-pqj8w" Mar 13 14:20:53 crc kubenswrapper[4907]: I0313 14:20:53.265566 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ch7ft\" (UniqueName: \"kubernetes.io/projected/cb19cbd4-fd73-485d-9b1e-a9ed01fa73ac-kube-api-access-ch7ft\") pod \"certified-operators-pqj8w\" (UID: \"cb19cbd4-fd73-485d-9b1e-a9ed01fa73ac\") " pod="openshift-marketplace/certified-operators-pqj8w" Mar 13 14:20:53 crc kubenswrapper[4907]: I0313 14:20:53.265612 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb19cbd4-fd73-485d-9b1e-a9ed01fa73ac-catalog-content\") pod \"certified-operators-pqj8w\" (UID: \"cb19cbd4-fd73-485d-9b1e-a9ed01fa73ac\") " pod="openshift-marketplace/certified-operators-pqj8w" Mar 13 14:20:53 crc kubenswrapper[4907]: I0313 14:20:53.266110 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb19cbd4-fd73-485d-9b1e-a9ed01fa73ac-catalog-content\") pod \"certified-operators-pqj8w\" (UID: \"cb19cbd4-fd73-485d-9b1e-a9ed01fa73ac\") " pod="openshift-marketplace/certified-operators-pqj8w" Mar 13 14:20:53 crc kubenswrapper[4907]: I0313 14:20:53.266332 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb19cbd4-fd73-485d-9b1e-a9ed01fa73ac-utilities\") pod \"certified-operators-pqj8w\" (UID: \"cb19cbd4-fd73-485d-9b1e-a9ed01fa73ac\") " pod="openshift-marketplace/certified-operators-pqj8w" Mar 13 14:20:53 crc kubenswrapper[4907]: I0313 14:20:53.289011 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ch7ft\" (UniqueName: \"kubernetes.io/projected/cb19cbd4-fd73-485d-9b1e-a9ed01fa73ac-kube-api-access-ch7ft\") pod \"certified-operators-pqj8w\" (UID: \"cb19cbd4-fd73-485d-9b1e-a9ed01fa73ac\") " pod="openshift-marketplace/certified-operators-pqj8w" Mar 13 14:20:53 crc kubenswrapper[4907]: I0313 14:20:53.348497 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pqj8w" Mar 13 14:20:53 crc kubenswrapper[4907]: I0313 14:20:53.780526 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-pqj8w"] Mar 13 14:20:53 crc kubenswrapper[4907]: W0313 14:20:53.789681 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcb19cbd4_fd73_485d_9b1e_a9ed01fa73ac.slice/crio-a30ed723dfd93724481722c622f269a4f11046d220f97345500ef389c7268947 WatchSource:0}: Error finding container a30ed723dfd93724481722c622f269a4f11046d220f97345500ef389c7268947: Status 404 returned error can't find the container with id a30ed723dfd93724481722c622f269a4f11046d220f97345500ef389c7268947 Mar 13 14:20:54 crc kubenswrapper[4907]: I0313 14:20:54.531477 4907 generic.go:334] "Generic (PLEG): container finished" podID="cb19cbd4-fd73-485d-9b1e-a9ed01fa73ac" containerID="270f522a63d6b7db266e9f432dbe7c6050c412adf666d82c89c463a2eae54bb1" exitCode=0 Mar 13 14:20:54 crc kubenswrapper[4907]: I0313 14:20:54.531956 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pqj8w" event={"ID":"cb19cbd4-fd73-485d-9b1e-a9ed01fa73ac","Type":"ContainerDied","Data":"270f522a63d6b7db266e9f432dbe7c6050c412adf666d82c89c463a2eae54bb1"} Mar 13 14:20:54 crc kubenswrapper[4907]: I0313 14:20:54.532012 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pqj8w" event={"ID":"cb19cbd4-fd73-485d-9b1e-a9ed01fa73ac","Type":"ContainerStarted","Data":"a30ed723dfd93724481722c622f269a4f11046d220f97345500ef389c7268947"} Mar 13 14:20:55 crc kubenswrapper[4907]: I0313 14:20:55.205957 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-5f558f5558-cqjnn" Mar 13 14:20:55 crc kubenswrapper[4907]: I0313 14:20:55.545584 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pqj8w" event={"ID":"cb19cbd4-fd73-485d-9b1e-a9ed01fa73ac","Type":"ContainerStarted","Data":"a8fc5a1c00d8ab234f0ce968e9fc0a3754194014ed118abc584a840f27aef640"} Mar 13 14:20:56 crc kubenswrapper[4907]: I0313 14:20:56.554465 4907 generic.go:334] "Generic (PLEG): container finished" podID="cb19cbd4-fd73-485d-9b1e-a9ed01fa73ac" containerID="a8fc5a1c00d8ab234f0ce968e9fc0a3754194014ed118abc584a840f27aef640" exitCode=0 Mar 13 14:20:56 crc kubenswrapper[4907]: I0313 14:20:56.554513 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pqj8w" event={"ID":"cb19cbd4-fd73-485d-9b1e-a9ed01fa73ac","Type":"ContainerDied","Data":"a8fc5a1c00d8ab234f0ce968e9fc0a3754194014ed118abc584a840f27aef640"} Mar 13 14:20:57 crc kubenswrapper[4907]: I0313 14:20:57.563072 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pqj8w" event={"ID":"cb19cbd4-fd73-485d-9b1e-a9ed01fa73ac","Type":"ContainerStarted","Data":"dac05b5102dfe562f7f72b8cbc866e0b01dabd3fb6d1bf5ab854368d949c6778"} Mar 13 14:20:57 crc kubenswrapper[4907]: I0313 14:20:57.580140 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-pqj8w" podStartSLOduration=3.114961532 podStartE2EDuration="5.58011902s" podCreationTimestamp="2026-03-13 14:20:52 +0000 UTC" firstStartedPulling="2026-03-13 14:20:54.536493921 +0000 UTC m=+953.436281610" lastFinishedPulling="2026-03-13 14:20:57.001651409 +0000 UTC m=+955.901439098" observedRunningTime="2026-03-13 14:20:57.578336911 +0000 UTC m=+956.478124600" watchObservedRunningTime="2026-03-13 14:20:57.58011902 +0000 UTC m=+956.479906719" Mar 13 14:21:02 crc kubenswrapper[4907]: I0313 14:21:02.489097 4907 scope.go:117] "RemoveContainer" containerID="2fab8f0b7470e3acd4d60ef3af4057ac0e1b8ec4f3de833c7e05ac6623d0bf47" Mar 13 14:21:03 crc kubenswrapper[4907]: I0313 14:21:03.350440 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-pqj8w" Mar 13 14:21:03 crc kubenswrapper[4907]: I0313 14:21:03.350862 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-pqj8w" Mar 13 14:21:03 crc kubenswrapper[4907]: I0313 14:21:03.412685 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-pqj8w" Mar 13 14:21:03 crc kubenswrapper[4907]: I0313 14:21:03.642804 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-pqj8w" Mar 13 14:21:03 crc kubenswrapper[4907]: I0313 14:21:03.685065 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-pqj8w"] Mar 13 14:21:05 crc kubenswrapper[4907]: I0313 14:21:05.614410 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-pqj8w" podUID="cb19cbd4-fd73-485d-9b1e-a9ed01fa73ac" containerName="registry-server" containerID="cri-o://dac05b5102dfe562f7f72b8cbc866e0b01dabd3fb6d1bf5ab854368d949c6778" gracePeriod=2 Mar 13 14:21:06 crc kubenswrapper[4907]: I0313 14:21:06.463489 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pqj8w" Mar 13 14:21:06 crc kubenswrapper[4907]: I0313 14:21:06.584056 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb19cbd4-fd73-485d-9b1e-a9ed01fa73ac-utilities\") pod \"cb19cbd4-fd73-485d-9b1e-a9ed01fa73ac\" (UID: \"cb19cbd4-fd73-485d-9b1e-a9ed01fa73ac\") " Mar 13 14:21:06 crc kubenswrapper[4907]: I0313 14:21:06.584115 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ch7ft\" (UniqueName: \"kubernetes.io/projected/cb19cbd4-fd73-485d-9b1e-a9ed01fa73ac-kube-api-access-ch7ft\") pod \"cb19cbd4-fd73-485d-9b1e-a9ed01fa73ac\" (UID: \"cb19cbd4-fd73-485d-9b1e-a9ed01fa73ac\") " Mar 13 14:21:06 crc kubenswrapper[4907]: I0313 14:21:06.584159 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb19cbd4-fd73-485d-9b1e-a9ed01fa73ac-catalog-content\") pod \"cb19cbd4-fd73-485d-9b1e-a9ed01fa73ac\" (UID: \"cb19cbd4-fd73-485d-9b1e-a9ed01fa73ac\") " Mar 13 14:21:06 crc kubenswrapper[4907]: I0313 14:21:06.585206 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cb19cbd4-fd73-485d-9b1e-a9ed01fa73ac-utilities" (OuterVolumeSpecName: "utilities") pod "cb19cbd4-fd73-485d-9b1e-a9ed01fa73ac" (UID: "cb19cbd4-fd73-485d-9b1e-a9ed01fa73ac"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:21:06 crc kubenswrapper[4907]: I0313 14:21:06.593976 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cb19cbd4-fd73-485d-9b1e-a9ed01fa73ac-kube-api-access-ch7ft" (OuterVolumeSpecName: "kube-api-access-ch7ft") pod "cb19cbd4-fd73-485d-9b1e-a9ed01fa73ac" (UID: "cb19cbd4-fd73-485d-9b1e-a9ed01fa73ac"). InnerVolumeSpecName "kube-api-access-ch7ft". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:21:06 crc kubenswrapper[4907]: I0313 14:21:06.626251 4907 generic.go:334] "Generic (PLEG): container finished" podID="cb19cbd4-fd73-485d-9b1e-a9ed01fa73ac" containerID="dac05b5102dfe562f7f72b8cbc866e0b01dabd3fb6d1bf5ab854368d949c6778" exitCode=0 Mar 13 14:21:06 crc kubenswrapper[4907]: I0313 14:21:06.626315 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pqj8w" Mar 13 14:21:06 crc kubenswrapper[4907]: I0313 14:21:06.626323 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pqj8w" event={"ID":"cb19cbd4-fd73-485d-9b1e-a9ed01fa73ac","Type":"ContainerDied","Data":"dac05b5102dfe562f7f72b8cbc866e0b01dabd3fb6d1bf5ab854368d949c6778"} Mar 13 14:21:06 crc kubenswrapper[4907]: I0313 14:21:06.626641 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pqj8w" event={"ID":"cb19cbd4-fd73-485d-9b1e-a9ed01fa73ac","Type":"ContainerDied","Data":"a30ed723dfd93724481722c622f269a4f11046d220f97345500ef389c7268947"} Mar 13 14:21:06 crc kubenswrapper[4907]: I0313 14:21:06.626673 4907 scope.go:117] "RemoveContainer" containerID="dac05b5102dfe562f7f72b8cbc866e0b01dabd3fb6d1bf5ab854368d949c6778" Mar 13 14:21:06 crc kubenswrapper[4907]: I0313 14:21:06.646246 4907 scope.go:117] "RemoveContainer" containerID="a8fc5a1c00d8ab234f0ce968e9fc0a3754194014ed118abc584a840f27aef640" Mar 13 14:21:06 crc kubenswrapper[4907]: I0313 14:21:06.662916 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cb19cbd4-fd73-485d-9b1e-a9ed01fa73ac-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cb19cbd4-fd73-485d-9b1e-a9ed01fa73ac" (UID: "cb19cbd4-fd73-485d-9b1e-a9ed01fa73ac"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:21:06 crc kubenswrapper[4907]: I0313 14:21:06.664848 4907 scope.go:117] "RemoveContainer" containerID="270f522a63d6b7db266e9f432dbe7c6050c412adf666d82c89c463a2eae54bb1" Mar 13 14:21:06 crc kubenswrapper[4907]: I0313 14:21:06.686560 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb19cbd4-fd73-485d-9b1e-a9ed01fa73ac-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 14:21:06 crc kubenswrapper[4907]: I0313 14:21:06.686585 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ch7ft\" (UniqueName: \"kubernetes.io/projected/cb19cbd4-fd73-485d-9b1e-a9ed01fa73ac-kube-api-access-ch7ft\") on node \"crc\" DevicePath \"\"" Mar 13 14:21:06 crc kubenswrapper[4907]: I0313 14:21:06.686595 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb19cbd4-fd73-485d-9b1e-a9ed01fa73ac-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 14:21:06 crc kubenswrapper[4907]: I0313 14:21:06.696682 4907 scope.go:117] "RemoveContainer" containerID="dac05b5102dfe562f7f72b8cbc866e0b01dabd3fb6d1bf5ab854368d949c6778" Mar 13 14:21:06 crc kubenswrapper[4907]: E0313 14:21:06.697114 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dac05b5102dfe562f7f72b8cbc866e0b01dabd3fb6d1bf5ab854368d949c6778\": container with ID starting with dac05b5102dfe562f7f72b8cbc866e0b01dabd3fb6d1bf5ab854368d949c6778 not found: ID does not exist" containerID="dac05b5102dfe562f7f72b8cbc866e0b01dabd3fb6d1bf5ab854368d949c6778" Mar 13 14:21:06 crc kubenswrapper[4907]: I0313 14:21:06.697205 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dac05b5102dfe562f7f72b8cbc866e0b01dabd3fb6d1bf5ab854368d949c6778"} err="failed to get container status \"dac05b5102dfe562f7f72b8cbc866e0b01dabd3fb6d1bf5ab854368d949c6778\": rpc error: code = NotFound desc = could not find container \"dac05b5102dfe562f7f72b8cbc866e0b01dabd3fb6d1bf5ab854368d949c6778\": container with ID starting with dac05b5102dfe562f7f72b8cbc866e0b01dabd3fb6d1bf5ab854368d949c6778 not found: ID does not exist" Mar 13 14:21:06 crc kubenswrapper[4907]: I0313 14:21:06.697239 4907 scope.go:117] "RemoveContainer" containerID="a8fc5a1c00d8ab234f0ce968e9fc0a3754194014ed118abc584a840f27aef640" Mar 13 14:21:06 crc kubenswrapper[4907]: E0313 14:21:06.697656 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a8fc5a1c00d8ab234f0ce968e9fc0a3754194014ed118abc584a840f27aef640\": container with ID starting with a8fc5a1c00d8ab234f0ce968e9fc0a3754194014ed118abc584a840f27aef640 not found: ID does not exist" containerID="a8fc5a1c00d8ab234f0ce968e9fc0a3754194014ed118abc584a840f27aef640" Mar 13 14:21:06 crc kubenswrapper[4907]: I0313 14:21:06.697694 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a8fc5a1c00d8ab234f0ce968e9fc0a3754194014ed118abc584a840f27aef640"} err="failed to get container status \"a8fc5a1c00d8ab234f0ce968e9fc0a3754194014ed118abc584a840f27aef640\": rpc error: code = NotFound desc = could not find container \"a8fc5a1c00d8ab234f0ce968e9fc0a3754194014ed118abc584a840f27aef640\": container with ID starting with a8fc5a1c00d8ab234f0ce968e9fc0a3754194014ed118abc584a840f27aef640 not found: ID does not exist" Mar 13 14:21:06 crc kubenswrapper[4907]: I0313 14:21:06.697724 4907 scope.go:117] "RemoveContainer" containerID="270f522a63d6b7db266e9f432dbe7c6050c412adf666d82c89c463a2eae54bb1" Mar 13 14:21:06 crc kubenswrapper[4907]: E0313 14:21:06.698093 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"270f522a63d6b7db266e9f432dbe7c6050c412adf666d82c89c463a2eae54bb1\": container with ID starting with 270f522a63d6b7db266e9f432dbe7c6050c412adf666d82c89c463a2eae54bb1 not found: ID does not exist" containerID="270f522a63d6b7db266e9f432dbe7c6050c412adf666d82c89c463a2eae54bb1" Mar 13 14:21:06 crc kubenswrapper[4907]: I0313 14:21:06.698114 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"270f522a63d6b7db266e9f432dbe7c6050c412adf666d82c89c463a2eae54bb1"} err="failed to get container status \"270f522a63d6b7db266e9f432dbe7c6050c412adf666d82c89c463a2eae54bb1\": rpc error: code = NotFound desc = could not find container \"270f522a63d6b7db266e9f432dbe7c6050c412adf666d82c89c463a2eae54bb1\": container with ID starting with 270f522a63d6b7db266e9f432dbe7c6050c412adf666d82c89c463a2eae54bb1 not found: ID does not exist" Mar 13 14:21:06 crc kubenswrapper[4907]: I0313 14:21:06.965625 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-pqj8w"] Mar 13 14:21:06 crc kubenswrapper[4907]: I0313 14:21:06.972228 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-pqj8w"] Mar 13 14:21:07 crc kubenswrapper[4907]: I0313 14:21:07.790805 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cb19cbd4-fd73-485d-9b1e-a9ed01fa73ac" path="/var/lib/kubelet/pods/cb19cbd4-fd73-485d-9b1e-a9ed01fa73ac/volumes" Mar 13 14:21:10 crc kubenswrapper[4907]: I0313 14:21:10.512674 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1qkjpl"] Mar 13 14:21:10 crc kubenswrapper[4907]: E0313 14:21:10.513228 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb19cbd4-fd73-485d-9b1e-a9ed01fa73ac" containerName="extract-utilities" Mar 13 14:21:10 crc kubenswrapper[4907]: I0313 14:21:10.513243 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb19cbd4-fd73-485d-9b1e-a9ed01fa73ac" containerName="extract-utilities" Mar 13 14:21:10 crc kubenswrapper[4907]: E0313 14:21:10.513266 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb19cbd4-fd73-485d-9b1e-a9ed01fa73ac" containerName="extract-content" Mar 13 14:21:10 crc kubenswrapper[4907]: I0313 14:21:10.513274 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb19cbd4-fd73-485d-9b1e-a9ed01fa73ac" containerName="extract-content" Mar 13 14:21:10 crc kubenswrapper[4907]: E0313 14:21:10.513289 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb19cbd4-fd73-485d-9b1e-a9ed01fa73ac" containerName="registry-server" Mar 13 14:21:10 crc kubenswrapper[4907]: I0313 14:21:10.513300 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb19cbd4-fd73-485d-9b1e-a9ed01fa73ac" containerName="registry-server" Mar 13 14:21:10 crc kubenswrapper[4907]: I0313 14:21:10.513427 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="cb19cbd4-fd73-485d-9b1e-a9ed01fa73ac" containerName="registry-server" Mar 13 14:21:10 crc kubenswrapper[4907]: I0313 14:21:10.514384 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1qkjpl" Mar 13 14:21:10 crc kubenswrapper[4907]: I0313 14:21:10.516964 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Mar 13 14:21:10 crc kubenswrapper[4907]: I0313 14:21:10.523040 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1qkjpl"] Mar 13 14:21:10 crc kubenswrapper[4907]: I0313 14:21:10.534740 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4sr9w\" (UniqueName: \"kubernetes.io/projected/385226e5-9bb7-4f71-888c-fdd2c3a59b6e-kube-api-access-4sr9w\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1qkjpl\" (UID: \"385226e5-9bb7-4f71-888c-fdd2c3a59b6e\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1qkjpl" Mar 13 14:21:10 crc kubenswrapper[4907]: I0313 14:21:10.534841 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/385226e5-9bb7-4f71-888c-fdd2c3a59b6e-bundle\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1qkjpl\" (UID: \"385226e5-9bb7-4f71-888c-fdd2c3a59b6e\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1qkjpl" Mar 13 14:21:10 crc kubenswrapper[4907]: I0313 14:21:10.534974 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/385226e5-9bb7-4f71-888c-fdd2c3a59b6e-util\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1qkjpl\" (UID: \"385226e5-9bb7-4f71-888c-fdd2c3a59b6e\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1qkjpl" Mar 13 14:21:10 crc kubenswrapper[4907]: I0313 14:21:10.650613 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4sr9w\" (UniqueName: \"kubernetes.io/projected/385226e5-9bb7-4f71-888c-fdd2c3a59b6e-kube-api-access-4sr9w\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1qkjpl\" (UID: \"385226e5-9bb7-4f71-888c-fdd2c3a59b6e\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1qkjpl" Mar 13 14:21:10 crc kubenswrapper[4907]: I0313 14:21:10.651221 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/385226e5-9bb7-4f71-888c-fdd2c3a59b6e-bundle\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1qkjpl\" (UID: \"385226e5-9bb7-4f71-888c-fdd2c3a59b6e\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1qkjpl" Mar 13 14:21:10 crc kubenswrapper[4907]: I0313 14:21:10.651358 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/385226e5-9bb7-4f71-888c-fdd2c3a59b6e-util\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1qkjpl\" (UID: \"385226e5-9bb7-4f71-888c-fdd2c3a59b6e\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1qkjpl" Mar 13 14:21:10 crc kubenswrapper[4907]: I0313 14:21:10.651527 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/385226e5-9bb7-4f71-888c-fdd2c3a59b6e-bundle\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1qkjpl\" (UID: \"385226e5-9bb7-4f71-888c-fdd2c3a59b6e\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1qkjpl" Mar 13 14:21:10 crc kubenswrapper[4907]: I0313 14:21:10.651724 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/385226e5-9bb7-4f71-888c-fdd2c3a59b6e-util\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1qkjpl\" (UID: \"385226e5-9bb7-4f71-888c-fdd2c3a59b6e\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1qkjpl" Mar 13 14:21:10 crc kubenswrapper[4907]: I0313 14:21:10.670407 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4sr9w\" (UniqueName: \"kubernetes.io/projected/385226e5-9bb7-4f71-888c-fdd2c3a59b6e-kube-api-access-4sr9w\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1qkjpl\" (UID: \"385226e5-9bb7-4f71-888c-fdd2c3a59b6e\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1qkjpl" Mar 13 14:21:10 crc kubenswrapper[4907]: I0313 14:21:10.843109 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1qkjpl" Mar 13 14:21:11 crc kubenswrapper[4907]: I0313 14:21:11.305041 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1qkjpl"] Mar 13 14:21:11 crc kubenswrapper[4907]: I0313 14:21:11.579746 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-l9xpb" podUID="4c8c0266-5452-4abb-92c5-c536df94dd41" containerName="console" containerID="cri-o://27a56c8f2bc88df1fc0af00ca4f2cc9f8243a27d536703bf1460c381978493aa" gracePeriod=15 Mar 13 14:21:11 crc kubenswrapper[4907]: I0313 14:21:11.660620 4907 generic.go:334] "Generic (PLEG): container finished" podID="385226e5-9bb7-4f71-888c-fdd2c3a59b6e" containerID="30a631b3f50b179ac1b95bb880442cbed61c80cb1509ef7f12866056fb40c103" exitCode=0 Mar 13 14:21:11 crc kubenswrapper[4907]: I0313 14:21:11.660686 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1qkjpl" event={"ID":"385226e5-9bb7-4f71-888c-fdd2c3a59b6e","Type":"ContainerDied","Data":"30a631b3f50b179ac1b95bb880442cbed61c80cb1509ef7f12866056fb40c103"} Mar 13 14:21:11 crc kubenswrapper[4907]: I0313 14:21:11.660717 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1qkjpl" event={"ID":"385226e5-9bb7-4f71-888c-fdd2c3a59b6e","Type":"ContainerStarted","Data":"ee977785926063e5f131d49c4b9de5e11f45ee3e668dd114d5625a14c04dadc5"} Mar 13 14:21:11 crc kubenswrapper[4907]: I0313 14:21:11.667220 4907 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 13 14:21:12 crc kubenswrapper[4907]: I0313 14:21:12.018199 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-l9xpb_4c8c0266-5452-4abb-92c5-c536df94dd41/console/0.log" Mar 13 14:21:12 crc kubenswrapper[4907]: I0313 14:21:12.018271 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-l9xpb" Mar 13 14:21:12 crc kubenswrapper[4907]: I0313 14:21:12.172061 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4c8c0266-5452-4abb-92c5-c536df94dd41-trusted-ca-bundle\") pod \"4c8c0266-5452-4abb-92c5-c536df94dd41\" (UID: \"4c8c0266-5452-4abb-92c5-c536df94dd41\") " Mar 13 14:21:12 crc kubenswrapper[4907]: I0313 14:21:12.172129 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/4c8c0266-5452-4abb-92c5-c536df94dd41-console-serving-cert\") pod \"4c8c0266-5452-4abb-92c5-c536df94dd41\" (UID: \"4c8c0266-5452-4abb-92c5-c536df94dd41\") " Mar 13 14:21:12 crc kubenswrapper[4907]: I0313 14:21:12.172168 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/4c8c0266-5452-4abb-92c5-c536df94dd41-service-ca\") pod \"4c8c0266-5452-4abb-92c5-c536df94dd41\" (UID: \"4c8c0266-5452-4abb-92c5-c536df94dd41\") " Mar 13 14:21:12 crc kubenswrapper[4907]: I0313 14:21:12.172186 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/4c8c0266-5452-4abb-92c5-c536df94dd41-console-oauth-config\") pod \"4c8c0266-5452-4abb-92c5-c536df94dd41\" (UID: \"4c8c0266-5452-4abb-92c5-c536df94dd41\") " Mar 13 14:21:12 crc kubenswrapper[4907]: I0313 14:21:12.172211 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vxlx4\" (UniqueName: \"kubernetes.io/projected/4c8c0266-5452-4abb-92c5-c536df94dd41-kube-api-access-vxlx4\") pod \"4c8c0266-5452-4abb-92c5-c536df94dd41\" (UID: \"4c8c0266-5452-4abb-92c5-c536df94dd41\") " Mar 13 14:21:12 crc kubenswrapper[4907]: I0313 14:21:12.172249 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/4c8c0266-5452-4abb-92c5-c536df94dd41-console-config\") pod \"4c8c0266-5452-4abb-92c5-c536df94dd41\" (UID: \"4c8c0266-5452-4abb-92c5-c536df94dd41\") " Mar 13 14:21:12 crc kubenswrapper[4907]: I0313 14:21:12.172267 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/4c8c0266-5452-4abb-92c5-c536df94dd41-oauth-serving-cert\") pod \"4c8c0266-5452-4abb-92c5-c536df94dd41\" (UID: \"4c8c0266-5452-4abb-92c5-c536df94dd41\") " Mar 13 14:21:12 crc kubenswrapper[4907]: I0313 14:21:12.172870 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4c8c0266-5452-4abb-92c5-c536df94dd41-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "4c8c0266-5452-4abb-92c5-c536df94dd41" (UID: "4c8c0266-5452-4abb-92c5-c536df94dd41"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:21:12 crc kubenswrapper[4907]: I0313 14:21:12.172905 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4c8c0266-5452-4abb-92c5-c536df94dd41-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "4c8c0266-5452-4abb-92c5-c536df94dd41" (UID: "4c8c0266-5452-4abb-92c5-c536df94dd41"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:21:12 crc kubenswrapper[4907]: I0313 14:21:12.173342 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4c8c0266-5452-4abb-92c5-c536df94dd41-console-config" (OuterVolumeSpecName: "console-config") pod "4c8c0266-5452-4abb-92c5-c536df94dd41" (UID: "4c8c0266-5452-4abb-92c5-c536df94dd41"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:21:12 crc kubenswrapper[4907]: I0313 14:21:12.173501 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4c8c0266-5452-4abb-92c5-c536df94dd41-service-ca" (OuterVolumeSpecName: "service-ca") pod "4c8c0266-5452-4abb-92c5-c536df94dd41" (UID: "4c8c0266-5452-4abb-92c5-c536df94dd41"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:21:12 crc kubenswrapper[4907]: I0313 14:21:12.179372 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c8c0266-5452-4abb-92c5-c536df94dd41-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "4c8c0266-5452-4abb-92c5-c536df94dd41" (UID: "4c8c0266-5452-4abb-92c5-c536df94dd41"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:21:12 crc kubenswrapper[4907]: I0313 14:21:12.186581 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c8c0266-5452-4abb-92c5-c536df94dd41-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "4c8c0266-5452-4abb-92c5-c536df94dd41" (UID: "4c8c0266-5452-4abb-92c5-c536df94dd41"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:21:12 crc kubenswrapper[4907]: I0313 14:21:12.191232 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c8c0266-5452-4abb-92c5-c536df94dd41-kube-api-access-vxlx4" (OuterVolumeSpecName: "kube-api-access-vxlx4") pod "4c8c0266-5452-4abb-92c5-c536df94dd41" (UID: "4c8c0266-5452-4abb-92c5-c536df94dd41"). InnerVolumeSpecName "kube-api-access-vxlx4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:21:12 crc kubenswrapper[4907]: I0313 14:21:12.274401 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vxlx4\" (UniqueName: \"kubernetes.io/projected/4c8c0266-5452-4abb-92c5-c536df94dd41-kube-api-access-vxlx4\") on node \"crc\" DevicePath \"\"" Mar 13 14:21:12 crc kubenswrapper[4907]: I0313 14:21:12.274622 4907 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/4c8c0266-5452-4abb-92c5-c536df94dd41-console-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:21:12 crc kubenswrapper[4907]: I0313 14:21:12.274695 4907 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/4c8c0266-5452-4abb-92c5-c536df94dd41-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 14:21:12 crc kubenswrapper[4907]: I0313 14:21:12.274754 4907 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4c8c0266-5452-4abb-92c5-c536df94dd41-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:21:12 crc kubenswrapper[4907]: I0313 14:21:12.274806 4907 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/4c8c0266-5452-4abb-92c5-c536df94dd41-console-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 14:21:12 crc kubenswrapper[4907]: I0313 14:21:12.274858 4907 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/4c8c0266-5452-4abb-92c5-c536df94dd41-service-ca\") on node \"crc\" DevicePath \"\"" Mar 13 14:21:12 crc kubenswrapper[4907]: I0313 14:21:12.274974 4907 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/4c8c0266-5452-4abb-92c5-c536df94dd41-console-oauth-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:21:12 crc kubenswrapper[4907]: I0313 14:21:12.669478 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-l9xpb_4c8c0266-5452-4abb-92c5-c536df94dd41/console/0.log" Mar 13 14:21:12 crc kubenswrapper[4907]: I0313 14:21:12.669832 4907 generic.go:334] "Generic (PLEG): container finished" podID="4c8c0266-5452-4abb-92c5-c536df94dd41" containerID="27a56c8f2bc88df1fc0af00ca4f2cc9f8243a27d536703bf1460c381978493aa" exitCode=2 Mar 13 14:21:12 crc kubenswrapper[4907]: I0313 14:21:12.669874 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-l9xpb" event={"ID":"4c8c0266-5452-4abb-92c5-c536df94dd41","Type":"ContainerDied","Data":"27a56c8f2bc88df1fc0af00ca4f2cc9f8243a27d536703bf1460c381978493aa"} Mar 13 14:21:12 crc kubenswrapper[4907]: I0313 14:21:12.669953 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-l9xpb" event={"ID":"4c8c0266-5452-4abb-92c5-c536df94dd41","Type":"ContainerDied","Data":"2ca75bf0ab6ab7b6d03a36b2b78748709fe1e70503c3e45b40703b8a20dadbca"} Mar 13 14:21:12 crc kubenswrapper[4907]: I0313 14:21:12.669980 4907 scope.go:117] "RemoveContainer" containerID="27a56c8f2bc88df1fc0af00ca4f2cc9f8243a27d536703bf1460c381978493aa" Mar 13 14:21:12 crc kubenswrapper[4907]: I0313 14:21:12.670130 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-l9xpb" Mar 13 14:21:12 crc kubenswrapper[4907]: I0313 14:21:12.702218 4907 scope.go:117] "RemoveContainer" containerID="27a56c8f2bc88df1fc0af00ca4f2cc9f8243a27d536703bf1460c381978493aa" Mar 13 14:21:12 crc kubenswrapper[4907]: E0313 14:21:12.702814 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"27a56c8f2bc88df1fc0af00ca4f2cc9f8243a27d536703bf1460c381978493aa\": container with ID starting with 27a56c8f2bc88df1fc0af00ca4f2cc9f8243a27d536703bf1460c381978493aa not found: ID does not exist" containerID="27a56c8f2bc88df1fc0af00ca4f2cc9f8243a27d536703bf1460c381978493aa" Mar 13 14:21:12 crc kubenswrapper[4907]: I0313 14:21:12.702858 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"27a56c8f2bc88df1fc0af00ca4f2cc9f8243a27d536703bf1460c381978493aa"} err="failed to get container status \"27a56c8f2bc88df1fc0af00ca4f2cc9f8243a27d536703bf1460c381978493aa\": rpc error: code = NotFound desc = could not find container \"27a56c8f2bc88df1fc0af00ca4f2cc9f8243a27d536703bf1460c381978493aa\": container with ID starting with 27a56c8f2bc88df1fc0af00ca4f2cc9f8243a27d536703bf1460c381978493aa not found: ID does not exist" Mar 13 14:21:12 crc kubenswrapper[4907]: I0313 14:21:12.706770 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-l9xpb"] Mar 13 14:21:12 crc kubenswrapper[4907]: I0313 14:21:12.713679 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-l9xpb"] Mar 13 14:21:13 crc kubenswrapper[4907]: I0313 14:21:13.679337 4907 generic.go:334] "Generic (PLEG): container finished" podID="385226e5-9bb7-4f71-888c-fdd2c3a59b6e" containerID="810cdbdfbf983697c54bd5c7dbb827e3963e513b598bbd0dac35273b14be8d25" exitCode=0 Mar 13 14:21:13 crc kubenswrapper[4907]: I0313 14:21:13.679398 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1qkjpl" event={"ID":"385226e5-9bb7-4f71-888c-fdd2c3a59b6e","Type":"ContainerDied","Data":"810cdbdfbf983697c54bd5c7dbb827e3963e513b598bbd0dac35273b14be8d25"} Mar 13 14:21:13 crc kubenswrapper[4907]: I0313 14:21:13.793144 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4c8c0266-5452-4abb-92c5-c536df94dd41" path="/var/lib/kubelet/pods/4c8c0266-5452-4abb-92c5-c536df94dd41/volumes" Mar 13 14:21:14 crc kubenswrapper[4907]: I0313 14:21:14.697559 4907 generic.go:334] "Generic (PLEG): container finished" podID="385226e5-9bb7-4f71-888c-fdd2c3a59b6e" containerID="d6aa27c66a086be71679587fc879df03add237621c2207cd9ce3d8a5d0e7296d" exitCode=0 Mar 13 14:21:14 crc kubenswrapper[4907]: I0313 14:21:14.697644 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1qkjpl" event={"ID":"385226e5-9bb7-4f71-888c-fdd2c3a59b6e","Type":"ContainerDied","Data":"d6aa27c66a086be71679587fc879df03add237621c2207cd9ce3d8a5d0e7296d"} Mar 13 14:21:15 crc kubenswrapper[4907]: I0313 14:21:15.982065 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1qkjpl" Mar 13 14:21:16 crc kubenswrapper[4907]: I0313 14:21:16.134446 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4sr9w\" (UniqueName: \"kubernetes.io/projected/385226e5-9bb7-4f71-888c-fdd2c3a59b6e-kube-api-access-4sr9w\") pod \"385226e5-9bb7-4f71-888c-fdd2c3a59b6e\" (UID: \"385226e5-9bb7-4f71-888c-fdd2c3a59b6e\") " Mar 13 14:21:16 crc kubenswrapper[4907]: I0313 14:21:16.134584 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/385226e5-9bb7-4f71-888c-fdd2c3a59b6e-bundle\") pod \"385226e5-9bb7-4f71-888c-fdd2c3a59b6e\" (UID: \"385226e5-9bb7-4f71-888c-fdd2c3a59b6e\") " Mar 13 14:21:16 crc kubenswrapper[4907]: I0313 14:21:16.134706 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/385226e5-9bb7-4f71-888c-fdd2c3a59b6e-util\") pod \"385226e5-9bb7-4f71-888c-fdd2c3a59b6e\" (UID: \"385226e5-9bb7-4f71-888c-fdd2c3a59b6e\") " Mar 13 14:21:16 crc kubenswrapper[4907]: I0313 14:21:16.137824 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/385226e5-9bb7-4f71-888c-fdd2c3a59b6e-bundle" (OuterVolumeSpecName: "bundle") pod "385226e5-9bb7-4f71-888c-fdd2c3a59b6e" (UID: "385226e5-9bb7-4f71-888c-fdd2c3a59b6e"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:21:16 crc kubenswrapper[4907]: I0313 14:21:16.140016 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/385226e5-9bb7-4f71-888c-fdd2c3a59b6e-kube-api-access-4sr9w" (OuterVolumeSpecName: "kube-api-access-4sr9w") pod "385226e5-9bb7-4f71-888c-fdd2c3a59b6e" (UID: "385226e5-9bb7-4f71-888c-fdd2c3a59b6e"). InnerVolumeSpecName "kube-api-access-4sr9w". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:21:16 crc kubenswrapper[4907]: I0313 14:21:16.153475 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/385226e5-9bb7-4f71-888c-fdd2c3a59b6e-util" (OuterVolumeSpecName: "util") pod "385226e5-9bb7-4f71-888c-fdd2c3a59b6e" (UID: "385226e5-9bb7-4f71-888c-fdd2c3a59b6e"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:21:16 crc kubenswrapper[4907]: I0313 14:21:16.236944 4907 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/385226e5-9bb7-4f71-888c-fdd2c3a59b6e-util\") on node \"crc\" DevicePath \"\"" Mar 13 14:21:16 crc kubenswrapper[4907]: I0313 14:21:16.236993 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4sr9w\" (UniqueName: \"kubernetes.io/projected/385226e5-9bb7-4f71-888c-fdd2c3a59b6e-kube-api-access-4sr9w\") on node \"crc\" DevicePath \"\"" Mar 13 14:21:16 crc kubenswrapper[4907]: I0313 14:21:16.237014 4907 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/385226e5-9bb7-4f71-888c-fdd2c3a59b6e-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:21:16 crc kubenswrapper[4907]: I0313 14:21:16.713842 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1qkjpl" event={"ID":"385226e5-9bb7-4f71-888c-fdd2c3a59b6e","Type":"ContainerDied","Data":"ee977785926063e5f131d49c4b9de5e11f45ee3e668dd114d5625a14c04dadc5"} Mar 13 14:21:16 crc kubenswrapper[4907]: I0313 14:21:16.713919 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ee977785926063e5f131d49c4b9de5e11f45ee3e668dd114d5625a14c04dadc5" Mar 13 14:21:16 crc kubenswrapper[4907]: I0313 14:21:16.714031 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1qkjpl" Mar 13 14:21:18 crc kubenswrapper[4907]: I0313 14:21:18.041478 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 14:21:18 crc kubenswrapper[4907]: I0313 14:21:18.041857 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 14:21:25 crc kubenswrapper[4907]: I0313 14:21:25.948257 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-cf9f9b596-sd4ls"] Mar 13 14:21:25 crc kubenswrapper[4907]: E0313 14:21:25.948815 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="385226e5-9bb7-4f71-888c-fdd2c3a59b6e" containerName="util" Mar 13 14:21:25 crc kubenswrapper[4907]: I0313 14:21:25.948827 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="385226e5-9bb7-4f71-888c-fdd2c3a59b6e" containerName="util" Mar 13 14:21:25 crc kubenswrapper[4907]: E0313 14:21:25.948840 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="385226e5-9bb7-4f71-888c-fdd2c3a59b6e" containerName="pull" Mar 13 14:21:25 crc kubenswrapper[4907]: I0313 14:21:25.948845 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="385226e5-9bb7-4f71-888c-fdd2c3a59b6e" containerName="pull" Mar 13 14:21:25 crc kubenswrapper[4907]: E0313 14:21:25.948853 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="385226e5-9bb7-4f71-888c-fdd2c3a59b6e" containerName="extract" Mar 13 14:21:25 crc kubenswrapper[4907]: I0313 14:21:25.948859 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="385226e5-9bb7-4f71-888c-fdd2c3a59b6e" containerName="extract" Mar 13 14:21:25 crc kubenswrapper[4907]: E0313 14:21:25.948872 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c8c0266-5452-4abb-92c5-c536df94dd41" containerName="console" Mar 13 14:21:25 crc kubenswrapper[4907]: I0313 14:21:25.948878 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c8c0266-5452-4abb-92c5-c536df94dd41" containerName="console" Mar 13 14:21:25 crc kubenswrapper[4907]: I0313 14:21:25.948972 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c8c0266-5452-4abb-92c5-c536df94dd41" containerName="console" Mar 13 14:21:25 crc kubenswrapper[4907]: I0313 14:21:25.948986 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="385226e5-9bb7-4f71-888c-fdd2c3a59b6e" containerName="extract" Mar 13 14:21:25 crc kubenswrapper[4907]: I0313 14:21:25.949313 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-cf9f9b596-sd4ls" Mar 13 14:21:25 crc kubenswrapper[4907]: I0313 14:21:25.952932 4907 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-zzb7m" Mar 13 14:21:25 crc kubenswrapper[4907]: I0313 14:21:25.953159 4907 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Mar 13 14:21:25 crc kubenswrapper[4907]: I0313 14:21:25.956717 4907 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Mar 13 14:21:25 crc kubenswrapper[4907]: I0313 14:21:25.956937 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Mar 13 14:21:25 crc kubenswrapper[4907]: I0313 14:21:25.963356 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Mar 13 14:21:25 crc kubenswrapper[4907]: I0313 14:21:25.969383 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-cf9f9b596-sd4ls"] Mar 13 14:21:26 crc kubenswrapper[4907]: I0313 14:21:26.094143 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-88mvb\" (UniqueName: \"kubernetes.io/projected/1c19fee6-f4bd-4488-acd9-23a4835edd83-kube-api-access-88mvb\") pod \"metallb-operator-controller-manager-cf9f9b596-sd4ls\" (UID: \"1c19fee6-f4bd-4488-acd9-23a4835edd83\") " pod="metallb-system/metallb-operator-controller-manager-cf9f9b596-sd4ls" Mar 13 14:21:26 crc kubenswrapper[4907]: I0313 14:21:26.094199 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/1c19fee6-f4bd-4488-acd9-23a4835edd83-apiservice-cert\") pod \"metallb-operator-controller-manager-cf9f9b596-sd4ls\" (UID: \"1c19fee6-f4bd-4488-acd9-23a4835edd83\") " pod="metallb-system/metallb-operator-controller-manager-cf9f9b596-sd4ls" Mar 13 14:21:26 crc kubenswrapper[4907]: I0313 14:21:26.094248 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/1c19fee6-f4bd-4488-acd9-23a4835edd83-webhook-cert\") pod \"metallb-operator-controller-manager-cf9f9b596-sd4ls\" (UID: \"1c19fee6-f4bd-4488-acd9-23a4835edd83\") " pod="metallb-system/metallb-operator-controller-manager-cf9f9b596-sd4ls" Mar 13 14:21:26 crc kubenswrapper[4907]: I0313 14:21:26.167226 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-bf6b45999-8mfb6"] Mar 13 14:21:26 crc kubenswrapper[4907]: I0313 14:21:26.167852 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-bf6b45999-8mfb6" Mar 13 14:21:26 crc kubenswrapper[4907]: I0313 14:21:26.169947 4907 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Mar 13 14:21:26 crc kubenswrapper[4907]: I0313 14:21:26.170222 4907 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Mar 13 14:21:26 crc kubenswrapper[4907]: I0313 14:21:26.170343 4907 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-vtpkb" Mar 13 14:21:26 crc kubenswrapper[4907]: I0313 14:21:26.192490 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-bf6b45999-8mfb6"] Mar 13 14:21:26 crc kubenswrapper[4907]: I0313 14:21:26.194771 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/1c19fee6-f4bd-4488-acd9-23a4835edd83-webhook-cert\") pod \"metallb-operator-controller-manager-cf9f9b596-sd4ls\" (UID: \"1c19fee6-f4bd-4488-acd9-23a4835edd83\") " pod="metallb-system/metallb-operator-controller-manager-cf9f9b596-sd4ls" Mar 13 14:21:26 crc kubenswrapper[4907]: I0313 14:21:26.194831 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-88mvb\" (UniqueName: \"kubernetes.io/projected/1c19fee6-f4bd-4488-acd9-23a4835edd83-kube-api-access-88mvb\") pod \"metallb-operator-controller-manager-cf9f9b596-sd4ls\" (UID: \"1c19fee6-f4bd-4488-acd9-23a4835edd83\") " pod="metallb-system/metallb-operator-controller-manager-cf9f9b596-sd4ls" Mar 13 14:21:26 crc kubenswrapper[4907]: I0313 14:21:26.194865 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/1c19fee6-f4bd-4488-acd9-23a4835edd83-apiservice-cert\") pod \"metallb-operator-controller-manager-cf9f9b596-sd4ls\" (UID: \"1c19fee6-f4bd-4488-acd9-23a4835edd83\") " pod="metallb-system/metallb-operator-controller-manager-cf9f9b596-sd4ls" Mar 13 14:21:26 crc kubenswrapper[4907]: I0313 14:21:26.200437 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/1c19fee6-f4bd-4488-acd9-23a4835edd83-apiservice-cert\") pod \"metallb-operator-controller-manager-cf9f9b596-sd4ls\" (UID: \"1c19fee6-f4bd-4488-acd9-23a4835edd83\") " pod="metallb-system/metallb-operator-controller-manager-cf9f9b596-sd4ls" Mar 13 14:21:26 crc kubenswrapper[4907]: I0313 14:21:26.218985 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-88mvb\" (UniqueName: \"kubernetes.io/projected/1c19fee6-f4bd-4488-acd9-23a4835edd83-kube-api-access-88mvb\") pod \"metallb-operator-controller-manager-cf9f9b596-sd4ls\" (UID: \"1c19fee6-f4bd-4488-acd9-23a4835edd83\") " pod="metallb-system/metallb-operator-controller-manager-cf9f9b596-sd4ls" Mar 13 14:21:26 crc kubenswrapper[4907]: I0313 14:21:26.221685 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/1c19fee6-f4bd-4488-acd9-23a4835edd83-webhook-cert\") pod \"metallb-operator-controller-manager-cf9f9b596-sd4ls\" (UID: \"1c19fee6-f4bd-4488-acd9-23a4835edd83\") " pod="metallb-system/metallb-operator-controller-manager-cf9f9b596-sd4ls" Mar 13 14:21:26 crc kubenswrapper[4907]: I0313 14:21:26.267326 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-cf9f9b596-sd4ls" Mar 13 14:21:26 crc kubenswrapper[4907]: I0313 14:21:26.295737 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/211e8a4b-dd3b-4669-b426-f0ac30bb3ca2-apiservice-cert\") pod \"metallb-operator-webhook-server-bf6b45999-8mfb6\" (UID: \"211e8a4b-dd3b-4669-b426-f0ac30bb3ca2\") " pod="metallb-system/metallb-operator-webhook-server-bf6b45999-8mfb6" Mar 13 14:21:26 crc kubenswrapper[4907]: I0313 14:21:26.295801 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/211e8a4b-dd3b-4669-b426-f0ac30bb3ca2-webhook-cert\") pod \"metallb-operator-webhook-server-bf6b45999-8mfb6\" (UID: \"211e8a4b-dd3b-4669-b426-f0ac30bb3ca2\") " pod="metallb-system/metallb-operator-webhook-server-bf6b45999-8mfb6" Mar 13 14:21:26 crc kubenswrapper[4907]: I0313 14:21:26.295858 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-njlcd\" (UniqueName: \"kubernetes.io/projected/211e8a4b-dd3b-4669-b426-f0ac30bb3ca2-kube-api-access-njlcd\") pod \"metallb-operator-webhook-server-bf6b45999-8mfb6\" (UID: \"211e8a4b-dd3b-4669-b426-f0ac30bb3ca2\") " pod="metallb-system/metallb-operator-webhook-server-bf6b45999-8mfb6" Mar 13 14:21:26 crc kubenswrapper[4907]: I0313 14:21:26.397046 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-njlcd\" (UniqueName: \"kubernetes.io/projected/211e8a4b-dd3b-4669-b426-f0ac30bb3ca2-kube-api-access-njlcd\") pod \"metallb-operator-webhook-server-bf6b45999-8mfb6\" (UID: \"211e8a4b-dd3b-4669-b426-f0ac30bb3ca2\") " pod="metallb-system/metallb-operator-webhook-server-bf6b45999-8mfb6" Mar 13 14:21:26 crc kubenswrapper[4907]: I0313 14:21:26.397401 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/211e8a4b-dd3b-4669-b426-f0ac30bb3ca2-apiservice-cert\") pod \"metallb-operator-webhook-server-bf6b45999-8mfb6\" (UID: \"211e8a4b-dd3b-4669-b426-f0ac30bb3ca2\") " pod="metallb-system/metallb-operator-webhook-server-bf6b45999-8mfb6" Mar 13 14:21:26 crc kubenswrapper[4907]: I0313 14:21:26.397430 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/211e8a4b-dd3b-4669-b426-f0ac30bb3ca2-webhook-cert\") pod \"metallb-operator-webhook-server-bf6b45999-8mfb6\" (UID: \"211e8a4b-dd3b-4669-b426-f0ac30bb3ca2\") " pod="metallb-system/metallb-operator-webhook-server-bf6b45999-8mfb6" Mar 13 14:21:26 crc kubenswrapper[4907]: I0313 14:21:26.400745 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/211e8a4b-dd3b-4669-b426-f0ac30bb3ca2-webhook-cert\") pod \"metallb-operator-webhook-server-bf6b45999-8mfb6\" (UID: \"211e8a4b-dd3b-4669-b426-f0ac30bb3ca2\") " pod="metallb-system/metallb-operator-webhook-server-bf6b45999-8mfb6" Mar 13 14:21:26 crc kubenswrapper[4907]: I0313 14:21:26.403448 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/211e8a4b-dd3b-4669-b426-f0ac30bb3ca2-apiservice-cert\") pod \"metallb-operator-webhook-server-bf6b45999-8mfb6\" (UID: \"211e8a4b-dd3b-4669-b426-f0ac30bb3ca2\") " pod="metallb-system/metallb-operator-webhook-server-bf6b45999-8mfb6" Mar 13 14:21:26 crc kubenswrapper[4907]: I0313 14:21:26.416199 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-njlcd\" (UniqueName: \"kubernetes.io/projected/211e8a4b-dd3b-4669-b426-f0ac30bb3ca2-kube-api-access-njlcd\") pod \"metallb-operator-webhook-server-bf6b45999-8mfb6\" (UID: \"211e8a4b-dd3b-4669-b426-f0ac30bb3ca2\") " pod="metallb-system/metallb-operator-webhook-server-bf6b45999-8mfb6" Mar 13 14:21:26 crc kubenswrapper[4907]: I0313 14:21:26.480528 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-bf6b45999-8mfb6" Mar 13 14:21:26 crc kubenswrapper[4907]: I0313 14:21:26.690376 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-bf6b45999-8mfb6"] Mar 13 14:21:26 crc kubenswrapper[4907]: I0313 14:21:26.746000 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-cf9f9b596-sd4ls"] Mar 13 14:21:26 crc kubenswrapper[4907]: W0313 14:21:26.752405 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1c19fee6_f4bd_4488_acd9_23a4835edd83.slice/crio-5a45331b0fbfdd4da374f224485bc37c8953dfddabdb352dc960db3fe447c1a0 WatchSource:0}: Error finding container 5a45331b0fbfdd4da374f224485bc37c8953dfddabdb352dc960db3fe447c1a0: Status 404 returned error can't find the container with id 5a45331b0fbfdd4da374f224485bc37c8953dfddabdb352dc960db3fe447c1a0 Mar 13 14:21:26 crc kubenswrapper[4907]: I0313 14:21:26.767243 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-cf9f9b596-sd4ls" event={"ID":"1c19fee6-f4bd-4488-acd9-23a4835edd83","Type":"ContainerStarted","Data":"5a45331b0fbfdd4da374f224485bc37c8953dfddabdb352dc960db3fe447c1a0"} Mar 13 14:21:26 crc kubenswrapper[4907]: I0313 14:21:26.768321 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-bf6b45999-8mfb6" event={"ID":"211e8a4b-dd3b-4669-b426-f0ac30bb3ca2","Type":"ContainerStarted","Data":"01fd754435af4816d44cee7c7358e62423e9beb748c1d066d0c36016623e5098"} Mar 13 14:21:31 crc kubenswrapper[4907]: I0313 14:21:31.806086 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-cf9f9b596-sd4ls" event={"ID":"1c19fee6-f4bd-4488-acd9-23a4835edd83","Type":"ContainerStarted","Data":"aaf611434f75db6dfcd0b5f6ac99f50e8b3151fa4b3b2d139729f6b42c5a5d75"} Mar 13 14:21:31 crc kubenswrapper[4907]: I0313 14:21:31.806690 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-cf9f9b596-sd4ls" Mar 13 14:21:31 crc kubenswrapper[4907]: I0313 14:21:31.807604 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-bf6b45999-8mfb6" event={"ID":"211e8a4b-dd3b-4669-b426-f0ac30bb3ca2","Type":"ContainerStarted","Data":"25dafd6957026f609842bb395e5b42b5f4f7b10c5a859c6dae3f8f705a216d24"} Mar 13 14:21:31 crc kubenswrapper[4907]: I0313 14:21:31.807760 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-bf6b45999-8mfb6" Mar 13 14:21:31 crc kubenswrapper[4907]: I0313 14:21:31.857253 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-cf9f9b596-sd4ls" podStartSLOduration=4.119773106 podStartE2EDuration="6.857234375s" podCreationTimestamp="2026-03-13 14:21:25 +0000 UTC" firstStartedPulling="2026-03-13 14:21:26.755629383 +0000 UTC m=+985.655417082" lastFinishedPulling="2026-03-13 14:21:29.493090662 +0000 UTC m=+988.392878351" observedRunningTime="2026-03-13 14:21:31.853025979 +0000 UTC m=+990.752813688" watchObservedRunningTime="2026-03-13 14:21:31.857234375 +0000 UTC m=+990.757022084" Mar 13 14:21:31 crc kubenswrapper[4907]: I0313 14:21:31.873353 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-bf6b45999-8mfb6" podStartSLOduration=1.414381437 podStartE2EDuration="5.87333694s" podCreationTimestamp="2026-03-13 14:21:26 +0000 UTC" firstStartedPulling="2026-03-13 14:21:26.697105605 +0000 UTC m=+985.596893294" lastFinishedPulling="2026-03-13 14:21:31.156061108 +0000 UTC m=+990.055848797" observedRunningTime="2026-03-13 14:21:31.872435025 +0000 UTC m=+990.772222704" watchObservedRunningTime="2026-03-13 14:21:31.87333694 +0000 UTC m=+990.773124629" Mar 13 14:21:46 crc kubenswrapper[4907]: I0313 14:21:46.504960 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-bf6b45999-8mfb6" Mar 13 14:21:48 crc kubenswrapper[4907]: I0313 14:21:48.042007 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 14:21:48 crc kubenswrapper[4907]: I0313 14:21:48.042604 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 14:22:00 crc kubenswrapper[4907]: I0313 14:22:00.136015 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556862-sczxs"] Mar 13 14:22:00 crc kubenswrapper[4907]: I0313 14:22:00.137508 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556862-sczxs" Mar 13 14:22:00 crc kubenswrapper[4907]: I0313 14:22:00.141257 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 14:22:00 crc kubenswrapper[4907]: I0313 14:22:00.141809 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 14:22:00 crc kubenswrapper[4907]: I0313 14:22:00.143278 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 14:22:00 crc kubenswrapper[4907]: I0313 14:22:00.147197 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556862-sczxs"] Mar 13 14:22:00 crc kubenswrapper[4907]: I0313 14:22:00.250543 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s8bmc\" (UniqueName: \"kubernetes.io/projected/72e27ed7-be4f-4d90-ac02-fa7d64585d6d-kube-api-access-s8bmc\") pod \"auto-csr-approver-29556862-sczxs\" (UID: \"72e27ed7-be4f-4d90-ac02-fa7d64585d6d\") " pod="openshift-infra/auto-csr-approver-29556862-sczxs" Mar 13 14:22:00 crc kubenswrapper[4907]: I0313 14:22:00.352383 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s8bmc\" (UniqueName: \"kubernetes.io/projected/72e27ed7-be4f-4d90-ac02-fa7d64585d6d-kube-api-access-s8bmc\") pod \"auto-csr-approver-29556862-sczxs\" (UID: \"72e27ed7-be4f-4d90-ac02-fa7d64585d6d\") " pod="openshift-infra/auto-csr-approver-29556862-sczxs" Mar 13 14:22:00 crc kubenswrapper[4907]: I0313 14:22:00.389965 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s8bmc\" (UniqueName: \"kubernetes.io/projected/72e27ed7-be4f-4d90-ac02-fa7d64585d6d-kube-api-access-s8bmc\") pod \"auto-csr-approver-29556862-sczxs\" (UID: \"72e27ed7-be4f-4d90-ac02-fa7d64585d6d\") " pod="openshift-infra/auto-csr-approver-29556862-sczxs" Mar 13 14:22:00 crc kubenswrapper[4907]: I0313 14:22:00.467851 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556862-sczxs" Mar 13 14:22:00 crc kubenswrapper[4907]: I0313 14:22:00.672562 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556862-sczxs"] Mar 13 14:22:00 crc kubenswrapper[4907]: W0313 14:22:00.675873 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod72e27ed7_be4f_4d90_ac02_fa7d64585d6d.slice/crio-caaccae846433276178ae1d1425b6650200856ae6f5f00b62a4c34787d3ca1da WatchSource:0}: Error finding container caaccae846433276178ae1d1425b6650200856ae6f5f00b62a4c34787d3ca1da: Status 404 returned error can't find the container with id caaccae846433276178ae1d1425b6650200856ae6f5f00b62a4c34787d3ca1da Mar 13 14:22:00 crc kubenswrapper[4907]: I0313 14:22:00.996866 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556862-sczxs" event={"ID":"72e27ed7-be4f-4d90-ac02-fa7d64585d6d","Type":"ContainerStarted","Data":"caaccae846433276178ae1d1425b6650200856ae6f5f00b62a4c34787d3ca1da"} Mar 13 14:22:03 crc kubenswrapper[4907]: I0313 14:22:03.023330 4907 generic.go:334] "Generic (PLEG): container finished" podID="72e27ed7-be4f-4d90-ac02-fa7d64585d6d" containerID="ff874f38ed077cec3c5651ce4ba0b1fcf1996e7c127b7d6b128f6c5552e26bed" exitCode=0 Mar 13 14:22:03 crc kubenswrapper[4907]: I0313 14:22:03.023387 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556862-sczxs" event={"ID":"72e27ed7-be4f-4d90-ac02-fa7d64585d6d","Type":"ContainerDied","Data":"ff874f38ed077cec3c5651ce4ba0b1fcf1996e7c127b7d6b128f6c5552e26bed"} Mar 13 14:22:04 crc kubenswrapper[4907]: I0313 14:22:04.344014 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556862-sczxs" Mar 13 14:22:04 crc kubenswrapper[4907]: I0313 14:22:04.505213 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s8bmc\" (UniqueName: \"kubernetes.io/projected/72e27ed7-be4f-4d90-ac02-fa7d64585d6d-kube-api-access-s8bmc\") pod \"72e27ed7-be4f-4d90-ac02-fa7d64585d6d\" (UID: \"72e27ed7-be4f-4d90-ac02-fa7d64585d6d\") " Mar 13 14:22:04 crc kubenswrapper[4907]: I0313 14:22:04.515130 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/72e27ed7-be4f-4d90-ac02-fa7d64585d6d-kube-api-access-s8bmc" (OuterVolumeSpecName: "kube-api-access-s8bmc") pod "72e27ed7-be4f-4d90-ac02-fa7d64585d6d" (UID: "72e27ed7-be4f-4d90-ac02-fa7d64585d6d"). InnerVolumeSpecName "kube-api-access-s8bmc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:22:04 crc kubenswrapper[4907]: I0313 14:22:04.607418 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s8bmc\" (UniqueName: \"kubernetes.io/projected/72e27ed7-be4f-4d90-ac02-fa7d64585d6d-kube-api-access-s8bmc\") on node \"crc\" DevicePath \"\"" Mar 13 14:22:05 crc kubenswrapper[4907]: I0313 14:22:05.035868 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556862-sczxs" event={"ID":"72e27ed7-be4f-4d90-ac02-fa7d64585d6d","Type":"ContainerDied","Data":"caaccae846433276178ae1d1425b6650200856ae6f5f00b62a4c34787d3ca1da"} Mar 13 14:22:05 crc kubenswrapper[4907]: I0313 14:22:05.035927 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="caaccae846433276178ae1d1425b6650200856ae6f5f00b62a4c34787d3ca1da" Mar 13 14:22:05 crc kubenswrapper[4907]: I0313 14:22:05.035942 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556862-sczxs" Mar 13 14:22:05 crc kubenswrapper[4907]: I0313 14:22:05.417211 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556856-8grxd"] Mar 13 14:22:05 crc kubenswrapper[4907]: I0313 14:22:05.421016 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556856-8grxd"] Mar 13 14:22:05 crc kubenswrapper[4907]: I0313 14:22:05.792034 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1604551e-a53d-483c-a03e-3c2159768001" path="/var/lib/kubelet/pods/1604551e-a53d-483c-a03e-3c2159768001/volumes" Mar 13 14:22:06 crc kubenswrapper[4907]: I0313 14:22:06.269414 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-cf9f9b596-sd4ls" Mar 13 14:22:06 crc kubenswrapper[4907]: I0313 14:22:06.895849 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-bcc4b6f68-jlsw9"] Mar 13 14:22:06 crc kubenswrapper[4907]: E0313 14:22:06.896141 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72e27ed7-be4f-4d90-ac02-fa7d64585d6d" containerName="oc" Mar 13 14:22:06 crc kubenswrapper[4907]: I0313 14:22:06.896154 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="72e27ed7-be4f-4d90-ac02-fa7d64585d6d" containerName="oc" Mar 13 14:22:06 crc kubenswrapper[4907]: I0313 14:22:06.896264 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="72e27ed7-be4f-4d90-ac02-fa7d64585d6d" containerName="oc" Mar 13 14:22:06 crc kubenswrapper[4907]: I0313 14:22:06.896669 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-jlsw9" Mar 13 14:22:06 crc kubenswrapper[4907]: I0313 14:22:06.899510 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-6r77v"] Mar 13 14:22:06 crc kubenswrapper[4907]: I0313 14:22:06.901685 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-6r77v" Mar 13 14:22:06 crc kubenswrapper[4907]: I0313 14:22:06.909020 4907 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Mar 13 14:22:06 crc kubenswrapper[4907]: I0313 14:22:06.909221 4907 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-472l6" Mar 13 14:22:06 crc kubenswrapper[4907]: I0313 14:22:06.909775 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Mar 13 14:22:06 crc kubenswrapper[4907]: I0313 14:22:06.909905 4907 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Mar 13 14:22:06 crc kubenswrapper[4907]: I0313 14:22:06.919665 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-bcc4b6f68-jlsw9"] Mar 13 14:22:06 crc kubenswrapper[4907]: I0313 14:22:06.991770 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-fkplv"] Mar 13 14:22:06 crc kubenswrapper[4907]: I0313 14:22:06.993201 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-fkplv" Mar 13 14:22:06 crc kubenswrapper[4907]: I0313 14:22:06.997210 4907 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Mar 13 14:22:06 crc kubenswrapper[4907]: I0313 14:22:06.997907 4907 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:06.998683 4907 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-dsr7f" Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.001058 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.006627 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-7bb4cc7c98-2g4kt"] Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.007714 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-7bb4cc7c98-2g4kt" Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.009470 4907 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.018938 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-7bb4cc7c98-2g4kt"] Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.038981 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/f5203ff6-d473-484d-ab8f-414e9ca804d4-frr-sockets\") pod \"frr-k8s-6r77v\" (UID: \"f5203ff6-d473-484d-ab8f-414e9ca804d4\") " pod="metallb-system/frr-k8s-6r77v" Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.039038 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/f5203ff6-d473-484d-ab8f-414e9ca804d4-frr-conf\") pod \"frr-k8s-6r77v\" (UID: \"f5203ff6-d473-484d-ab8f-414e9ca804d4\") " pod="metallb-system/frr-k8s-6r77v" Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.039068 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/f5203ff6-d473-484d-ab8f-414e9ca804d4-reloader\") pod \"frr-k8s-6r77v\" (UID: \"f5203ff6-d473-484d-ab8f-414e9ca804d4\") " pod="metallb-system/frr-k8s-6r77v" Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.039092 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7l8tg\" (UniqueName: \"kubernetes.io/projected/f5203ff6-d473-484d-ab8f-414e9ca804d4-kube-api-access-7l8tg\") pod \"frr-k8s-6r77v\" (UID: \"f5203ff6-d473-484d-ab8f-414e9ca804d4\") " pod="metallb-system/frr-k8s-6r77v" Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.039110 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/21498ef9-03ec-44de-ade5-cc56f504499b-cert\") pod \"frr-k8s-webhook-server-bcc4b6f68-jlsw9\" (UID: \"21498ef9-03ec-44de-ade5-cc56f504499b\") " pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-jlsw9" Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.039129 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f5203ff6-d473-484d-ab8f-414e9ca804d4-metrics-certs\") pod \"frr-k8s-6r77v\" (UID: \"f5203ff6-d473-484d-ab8f-414e9ca804d4\") " pod="metallb-system/frr-k8s-6r77v" Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.039157 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/f5203ff6-d473-484d-ab8f-414e9ca804d4-frr-startup\") pod \"frr-k8s-6r77v\" (UID: \"f5203ff6-d473-484d-ab8f-414e9ca804d4\") " pod="metallb-system/frr-k8s-6r77v" Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.039175 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x42bp\" (UniqueName: \"kubernetes.io/projected/21498ef9-03ec-44de-ade5-cc56f504499b-kube-api-access-x42bp\") pod \"frr-k8s-webhook-server-bcc4b6f68-jlsw9\" (UID: \"21498ef9-03ec-44de-ade5-cc56f504499b\") " pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-jlsw9" Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.039203 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/f5203ff6-d473-484d-ab8f-414e9ca804d4-metrics\") pod \"frr-k8s-6r77v\" (UID: \"f5203ff6-d473-484d-ab8f-414e9ca804d4\") " pod="metallb-system/frr-k8s-6r77v" Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.140521 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/f5203ff6-d473-484d-ab8f-414e9ca804d4-frr-conf\") pod \"frr-k8s-6r77v\" (UID: \"f5203ff6-d473-484d-ab8f-414e9ca804d4\") " pod="metallb-system/frr-k8s-6r77v" Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.140585 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5564ca73-f820-49a4-affa-fd718d1ad7af-metrics-certs\") pod \"speaker-fkplv\" (UID: \"5564ca73-f820-49a4-affa-fd718d1ad7af\") " pod="metallb-system/speaker-fkplv" Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.140625 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/f5203ff6-d473-484d-ab8f-414e9ca804d4-reloader\") pod \"frr-k8s-6r77v\" (UID: \"f5203ff6-d473-484d-ab8f-414e9ca804d4\") " pod="metallb-system/frr-k8s-6r77v" Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.140661 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7l8tg\" (UniqueName: \"kubernetes.io/projected/f5203ff6-d473-484d-ab8f-414e9ca804d4-kube-api-access-7l8tg\") pod \"frr-k8s-6r77v\" (UID: \"f5203ff6-d473-484d-ab8f-414e9ca804d4\") " pod="metallb-system/frr-k8s-6r77v" Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.140683 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/21498ef9-03ec-44de-ade5-cc56f504499b-cert\") pod \"frr-k8s-webhook-server-bcc4b6f68-jlsw9\" (UID: \"21498ef9-03ec-44de-ade5-cc56f504499b\") " pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-jlsw9" Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.140963 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/f5203ff6-d473-484d-ab8f-414e9ca804d4-frr-conf\") pod \"frr-k8s-6r77v\" (UID: \"f5203ff6-d473-484d-ab8f-414e9ca804d4\") " pod="metallb-system/frr-k8s-6r77v" Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.141014 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/f5203ff6-d473-484d-ab8f-414e9ca804d4-reloader\") pod \"frr-k8s-6r77v\" (UID: \"f5203ff6-d473-484d-ab8f-414e9ca804d4\") " pod="metallb-system/frr-k8s-6r77v" Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.141527 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f5203ff6-d473-484d-ab8f-414e9ca804d4-metrics-certs\") pod \"frr-k8s-6r77v\" (UID: \"f5203ff6-d473-484d-ab8f-414e9ca804d4\") " pod="metallb-system/frr-k8s-6r77v" Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.141582 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/5564ca73-f820-49a4-affa-fd718d1ad7af-metallb-excludel2\") pod \"speaker-fkplv\" (UID: \"5564ca73-f820-49a4-affa-fd718d1ad7af\") " pod="metallb-system/speaker-fkplv" Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.141609 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-twrdg\" (UniqueName: \"kubernetes.io/projected/255dfb62-fe3e-440f-82ed-1ff604426a9b-kube-api-access-twrdg\") pod \"controller-7bb4cc7c98-2g4kt\" (UID: \"255dfb62-fe3e-440f-82ed-1ff604426a9b\") " pod="metallb-system/controller-7bb4cc7c98-2g4kt" Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.141636 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/5564ca73-f820-49a4-affa-fd718d1ad7af-memberlist\") pod \"speaker-fkplv\" (UID: \"5564ca73-f820-49a4-affa-fd718d1ad7af\") " pod="metallb-system/speaker-fkplv" Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.141671 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/f5203ff6-d473-484d-ab8f-414e9ca804d4-frr-startup\") pod \"frr-k8s-6r77v\" (UID: \"f5203ff6-d473-484d-ab8f-414e9ca804d4\") " pod="metallb-system/frr-k8s-6r77v" Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.141694 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zhdph\" (UniqueName: \"kubernetes.io/projected/5564ca73-f820-49a4-affa-fd718d1ad7af-kube-api-access-zhdph\") pod \"speaker-fkplv\" (UID: \"5564ca73-f820-49a4-affa-fd718d1ad7af\") " pod="metallb-system/speaker-fkplv" Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.141727 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x42bp\" (UniqueName: \"kubernetes.io/projected/21498ef9-03ec-44de-ade5-cc56f504499b-kube-api-access-x42bp\") pod \"frr-k8s-webhook-server-bcc4b6f68-jlsw9\" (UID: \"21498ef9-03ec-44de-ade5-cc56f504499b\") " pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-jlsw9" Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.141768 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/f5203ff6-d473-484d-ab8f-414e9ca804d4-metrics\") pod \"frr-k8s-6r77v\" (UID: \"f5203ff6-d473-484d-ab8f-414e9ca804d4\") " pod="metallb-system/frr-k8s-6r77v" Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.141790 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/f5203ff6-d473-484d-ab8f-414e9ca804d4-frr-sockets\") pod \"frr-k8s-6r77v\" (UID: \"f5203ff6-d473-484d-ab8f-414e9ca804d4\") " pod="metallb-system/frr-k8s-6r77v" Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.141818 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/255dfb62-fe3e-440f-82ed-1ff604426a9b-metrics-certs\") pod \"controller-7bb4cc7c98-2g4kt\" (UID: \"255dfb62-fe3e-440f-82ed-1ff604426a9b\") " pod="metallb-system/controller-7bb4cc7c98-2g4kt" Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.141839 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/255dfb62-fe3e-440f-82ed-1ff604426a9b-cert\") pod \"controller-7bb4cc7c98-2g4kt\" (UID: \"255dfb62-fe3e-440f-82ed-1ff604426a9b\") " pod="metallb-system/controller-7bb4cc7c98-2g4kt" Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.142295 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/f5203ff6-d473-484d-ab8f-414e9ca804d4-metrics\") pod \"frr-k8s-6r77v\" (UID: \"f5203ff6-d473-484d-ab8f-414e9ca804d4\") " pod="metallb-system/frr-k8s-6r77v" Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.142389 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/f5203ff6-d473-484d-ab8f-414e9ca804d4-frr-startup\") pod \"frr-k8s-6r77v\" (UID: \"f5203ff6-d473-484d-ab8f-414e9ca804d4\") " pod="metallb-system/frr-k8s-6r77v" Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.142505 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/f5203ff6-d473-484d-ab8f-414e9ca804d4-frr-sockets\") pod \"frr-k8s-6r77v\" (UID: \"f5203ff6-d473-484d-ab8f-414e9ca804d4\") " pod="metallb-system/frr-k8s-6r77v" Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.157163 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/21498ef9-03ec-44de-ade5-cc56f504499b-cert\") pod \"frr-k8s-webhook-server-bcc4b6f68-jlsw9\" (UID: \"21498ef9-03ec-44de-ade5-cc56f504499b\") " pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-jlsw9" Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.158228 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f5203ff6-d473-484d-ab8f-414e9ca804d4-metrics-certs\") pod \"frr-k8s-6r77v\" (UID: \"f5203ff6-d473-484d-ab8f-414e9ca804d4\") " pod="metallb-system/frr-k8s-6r77v" Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.162265 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7l8tg\" (UniqueName: \"kubernetes.io/projected/f5203ff6-d473-484d-ab8f-414e9ca804d4-kube-api-access-7l8tg\") pod \"frr-k8s-6r77v\" (UID: \"f5203ff6-d473-484d-ab8f-414e9ca804d4\") " pod="metallb-system/frr-k8s-6r77v" Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.162335 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x42bp\" (UniqueName: \"kubernetes.io/projected/21498ef9-03ec-44de-ade5-cc56f504499b-kube-api-access-x42bp\") pod \"frr-k8s-webhook-server-bcc4b6f68-jlsw9\" (UID: \"21498ef9-03ec-44de-ade5-cc56f504499b\") " pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-jlsw9" Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.234227 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-jlsw9" Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.243319 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-twrdg\" (UniqueName: \"kubernetes.io/projected/255dfb62-fe3e-440f-82ed-1ff604426a9b-kube-api-access-twrdg\") pod \"controller-7bb4cc7c98-2g4kt\" (UID: \"255dfb62-fe3e-440f-82ed-1ff604426a9b\") " pod="metallb-system/controller-7bb4cc7c98-2g4kt" Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.243395 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/5564ca73-f820-49a4-affa-fd718d1ad7af-metallb-excludel2\") pod \"speaker-fkplv\" (UID: \"5564ca73-f820-49a4-affa-fd718d1ad7af\") " pod="metallb-system/speaker-fkplv" Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.243422 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/5564ca73-f820-49a4-affa-fd718d1ad7af-memberlist\") pod \"speaker-fkplv\" (UID: \"5564ca73-f820-49a4-affa-fd718d1ad7af\") " pod="metallb-system/speaker-fkplv" Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.243455 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zhdph\" (UniqueName: \"kubernetes.io/projected/5564ca73-f820-49a4-affa-fd718d1ad7af-kube-api-access-zhdph\") pod \"speaker-fkplv\" (UID: \"5564ca73-f820-49a4-affa-fd718d1ad7af\") " pod="metallb-system/speaker-fkplv" Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.243512 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/255dfb62-fe3e-440f-82ed-1ff604426a9b-metrics-certs\") pod \"controller-7bb4cc7c98-2g4kt\" (UID: \"255dfb62-fe3e-440f-82ed-1ff604426a9b\") " pod="metallb-system/controller-7bb4cc7c98-2g4kt" Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.243535 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/255dfb62-fe3e-440f-82ed-1ff604426a9b-cert\") pod \"controller-7bb4cc7c98-2g4kt\" (UID: \"255dfb62-fe3e-440f-82ed-1ff604426a9b\") " pod="metallb-system/controller-7bb4cc7c98-2g4kt" Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.243566 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5564ca73-f820-49a4-affa-fd718d1ad7af-metrics-certs\") pod \"speaker-fkplv\" (UID: \"5564ca73-f820-49a4-affa-fd718d1ad7af\") " pod="metallb-system/speaker-fkplv" Mar 13 14:22:07 crc kubenswrapper[4907]: E0313 14:22:07.243886 4907 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Mar 13 14:22:07 crc kubenswrapper[4907]: E0313 14:22:07.243974 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5564ca73-f820-49a4-affa-fd718d1ad7af-memberlist podName:5564ca73-f820-49a4-affa-fd718d1ad7af nodeName:}" failed. No retries permitted until 2026-03-13 14:22:07.743954407 +0000 UTC m=+1026.643742096 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/5564ca73-f820-49a4-affa-fd718d1ad7af-memberlist") pod "speaker-fkplv" (UID: "5564ca73-f820-49a4-affa-fd718d1ad7af") : secret "metallb-memberlist" not found Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.244939 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-6r77v" Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.245266 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/5564ca73-f820-49a4-affa-fd718d1ad7af-metallb-excludel2\") pod \"speaker-fkplv\" (UID: \"5564ca73-f820-49a4-affa-fd718d1ad7af\") " pod="metallb-system/speaker-fkplv" Mar 13 14:22:07 crc kubenswrapper[4907]: E0313 14:22:07.246486 4907 secret.go:188] Couldn't get secret metallb-system/controller-certs-secret: secret "controller-certs-secret" not found Mar 13 14:22:07 crc kubenswrapper[4907]: E0313 14:22:07.246639 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/255dfb62-fe3e-440f-82ed-1ff604426a9b-metrics-certs podName:255dfb62-fe3e-440f-82ed-1ff604426a9b nodeName:}" failed. No retries permitted until 2026-03-13 14:22:07.74662049 +0000 UTC m=+1026.646408189 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/255dfb62-fe3e-440f-82ed-1ff604426a9b-metrics-certs") pod "controller-7bb4cc7c98-2g4kt" (UID: "255dfb62-fe3e-440f-82ed-1ff604426a9b") : secret "controller-certs-secret" not found Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.249407 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5564ca73-f820-49a4-affa-fd718d1ad7af-metrics-certs\") pod \"speaker-fkplv\" (UID: \"5564ca73-f820-49a4-affa-fd718d1ad7af\") " pod="metallb-system/speaker-fkplv" Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.275205 4907 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.285616 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-twrdg\" (UniqueName: \"kubernetes.io/projected/255dfb62-fe3e-440f-82ed-1ff604426a9b-kube-api-access-twrdg\") pod \"controller-7bb4cc7c98-2g4kt\" (UID: \"255dfb62-fe3e-440f-82ed-1ff604426a9b\") " pod="metallb-system/controller-7bb4cc7c98-2g4kt" Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.296555 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/255dfb62-fe3e-440f-82ed-1ff604426a9b-cert\") pod \"controller-7bb4cc7c98-2g4kt\" (UID: \"255dfb62-fe3e-440f-82ed-1ff604426a9b\") " pod="metallb-system/controller-7bb4cc7c98-2g4kt" Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.322453 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zhdph\" (UniqueName: \"kubernetes.io/projected/5564ca73-f820-49a4-affa-fd718d1ad7af-kube-api-access-zhdph\") pod \"speaker-fkplv\" (UID: \"5564ca73-f820-49a4-affa-fd718d1ad7af\") " pod="metallb-system/speaker-fkplv" Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.515839 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-bcc4b6f68-jlsw9"] Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.750157 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/255dfb62-fe3e-440f-82ed-1ff604426a9b-metrics-certs\") pod \"controller-7bb4cc7c98-2g4kt\" (UID: \"255dfb62-fe3e-440f-82ed-1ff604426a9b\") " pod="metallb-system/controller-7bb4cc7c98-2g4kt" Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.750259 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/5564ca73-f820-49a4-affa-fd718d1ad7af-memberlist\") pod \"speaker-fkplv\" (UID: \"5564ca73-f820-49a4-affa-fd718d1ad7af\") " pod="metallb-system/speaker-fkplv" Mar 13 14:22:07 crc kubenswrapper[4907]: E0313 14:22:07.750370 4907 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Mar 13 14:22:07 crc kubenswrapper[4907]: E0313 14:22:07.750431 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5564ca73-f820-49a4-affa-fd718d1ad7af-memberlist podName:5564ca73-f820-49a4-affa-fd718d1ad7af nodeName:}" failed. No retries permitted until 2026-03-13 14:22:08.750416612 +0000 UTC m=+1027.650204301 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/5564ca73-f820-49a4-affa-fd718d1ad7af-memberlist") pod "speaker-fkplv" (UID: "5564ca73-f820-49a4-affa-fd718d1ad7af") : secret "metallb-memberlist" not found Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.756532 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/255dfb62-fe3e-440f-82ed-1ff604426a9b-metrics-certs\") pod \"controller-7bb4cc7c98-2g4kt\" (UID: \"255dfb62-fe3e-440f-82ed-1ff604426a9b\") " pod="metallb-system/controller-7bb4cc7c98-2g4kt" Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.862864 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-d4jn9"] Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.864583 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-d4jn9" Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.878273 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-d4jn9"] Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.927581 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-7bb4cc7c98-2g4kt" Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.953242 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-95pxs\" (UniqueName: \"kubernetes.io/projected/4211d5e0-93f1-4c48-a60e-6c89ae200a59-kube-api-access-95pxs\") pod \"community-operators-d4jn9\" (UID: \"4211d5e0-93f1-4c48-a60e-6c89ae200a59\") " pod="openshift-marketplace/community-operators-d4jn9" Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.953305 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4211d5e0-93f1-4c48-a60e-6c89ae200a59-utilities\") pod \"community-operators-d4jn9\" (UID: \"4211d5e0-93f1-4c48-a60e-6c89ae200a59\") " pod="openshift-marketplace/community-operators-d4jn9" Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.953330 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4211d5e0-93f1-4c48-a60e-6c89ae200a59-catalog-content\") pod \"community-operators-d4jn9\" (UID: \"4211d5e0-93f1-4c48-a60e-6c89ae200a59\") " pod="openshift-marketplace/community-operators-d4jn9" Mar 13 14:22:08 crc kubenswrapper[4907]: I0313 14:22:08.055003 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-95pxs\" (UniqueName: \"kubernetes.io/projected/4211d5e0-93f1-4c48-a60e-6c89ae200a59-kube-api-access-95pxs\") pod \"community-operators-d4jn9\" (UID: \"4211d5e0-93f1-4c48-a60e-6c89ae200a59\") " pod="openshift-marketplace/community-operators-d4jn9" Mar 13 14:22:08 crc kubenswrapper[4907]: I0313 14:22:08.055077 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4211d5e0-93f1-4c48-a60e-6c89ae200a59-utilities\") pod \"community-operators-d4jn9\" (UID: \"4211d5e0-93f1-4c48-a60e-6c89ae200a59\") " pod="openshift-marketplace/community-operators-d4jn9" Mar 13 14:22:08 crc kubenswrapper[4907]: I0313 14:22:08.055106 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4211d5e0-93f1-4c48-a60e-6c89ae200a59-catalog-content\") pod \"community-operators-d4jn9\" (UID: \"4211d5e0-93f1-4c48-a60e-6c89ae200a59\") " pod="openshift-marketplace/community-operators-d4jn9" Mar 13 14:22:08 crc kubenswrapper[4907]: I0313 14:22:08.055629 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4211d5e0-93f1-4c48-a60e-6c89ae200a59-catalog-content\") pod \"community-operators-d4jn9\" (UID: \"4211d5e0-93f1-4c48-a60e-6c89ae200a59\") " pod="openshift-marketplace/community-operators-d4jn9" Mar 13 14:22:08 crc kubenswrapper[4907]: I0313 14:22:08.055683 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4211d5e0-93f1-4c48-a60e-6c89ae200a59-utilities\") pod \"community-operators-d4jn9\" (UID: \"4211d5e0-93f1-4c48-a60e-6c89ae200a59\") " pod="openshift-marketplace/community-operators-d4jn9" Mar 13 14:22:08 crc kubenswrapper[4907]: I0313 14:22:08.065871 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-jlsw9" event={"ID":"21498ef9-03ec-44de-ade5-cc56f504499b","Type":"ContainerStarted","Data":"e3c1b667e890611fca7f8facb65f2255ec3d081d5d6caadf21a54230dd63cae4"} Mar 13 14:22:08 crc kubenswrapper[4907]: I0313 14:22:08.068147 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-6r77v" event={"ID":"f5203ff6-d473-484d-ab8f-414e9ca804d4","Type":"ContainerStarted","Data":"56f5bcfba3b8a1ee28a14907c426f784464a16911d4abf852784c13a3c742e3f"} Mar 13 14:22:08 crc kubenswrapper[4907]: I0313 14:22:08.076778 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-95pxs\" (UniqueName: \"kubernetes.io/projected/4211d5e0-93f1-4c48-a60e-6c89ae200a59-kube-api-access-95pxs\") pod \"community-operators-d4jn9\" (UID: \"4211d5e0-93f1-4c48-a60e-6c89ae200a59\") " pod="openshift-marketplace/community-operators-d4jn9" Mar 13 14:22:08 crc kubenswrapper[4907]: I0313 14:22:08.161616 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-7bb4cc7c98-2g4kt"] Mar 13 14:22:08 crc kubenswrapper[4907]: W0313 14:22:08.170954 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod255dfb62_fe3e_440f_82ed_1ff604426a9b.slice/crio-9bf9ea49769c0e80dc1e8af3b47bb1a01df01b53a39b033db4d913049dbaa5dd WatchSource:0}: Error finding container 9bf9ea49769c0e80dc1e8af3b47bb1a01df01b53a39b033db4d913049dbaa5dd: Status 404 returned error can't find the container with id 9bf9ea49769c0e80dc1e8af3b47bb1a01df01b53a39b033db4d913049dbaa5dd Mar 13 14:22:08 crc kubenswrapper[4907]: I0313 14:22:08.180895 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-d4jn9" Mar 13 14:22:08 crc kubenswrapper[4907]: I0313 14:22:08.527186 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-d4jn9"] Mar 13 14:22:08 crc kubenswrapper[4907]: I0313 14:22:08.767784 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/5564ca73-f820-49a4-affa-fd718d1ad7af-memberlist\") pod \"speaker-fkplv\" (UID: \"5564ca73-f820-49a4-affa-fd718d1ad7af\") " pod="metallb-system/speaker-fkplv" Mar 13 14:22:08 crc kubenswrapper[4907]: I0313 14:22:08.775575 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/5564ca73-f820-49a4-affa-fd718d1ad7af-memberlist\") pod \"speaker-fkplv\" (UID: \"5564ca73-f820-49a4-affa-fd718d1ad7af\") " pod="metallb-system/speaker-fkplv" Mar 13 14:22:08 crc kubenswrapper[4907]: I0313 14:22:08.814139 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-fkplv" Mar 13 14:22:09 crc kubenswrapper[4907]: I0313 14:22:09.077147 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-7bb4cc7c98-2g4kt" event={"ID":"255dfb62-fe3e-440f-82ed-1ff604426a9b","Type":"ContainerStarted","Data":"546c6c32d1aa7efc921fd1cd99597c6e7498e802373b3bd72c2dfb820ad5bbad"} Mar 13 14:22:09 crc kubenswrapper[4907]: I0313 14:22:09.077466 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-7bb4cc7c98-2g4kt" Mar 13 14:22:09 crc kubenswrapper[4907]: I0313 14:22:09.077479 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-7bb4cc7c98-2g4kt" event={"ID":"255dfb62-fe3e-440f-82ed-1ff604426a9b","Type":"ContainerStarted","Data":"bde2e3680ec5095a34b683394b0a19b7d7cfc0c2a683b3c8d5154da577f8bd79"} Mar 13 14:22:09 crc kubenswrapper[4907]: I0313 14:22:09.077489 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-7bb4cc7c98-2g4kt" event={"ID":"255dfb62-fe3e-440f-82ed-1ff604426a9b","Type":"ContainerStarted","Data":"9bf9ea49769c0e80dc1e8af3b47bb1a01df01b53a39b033db4d913049dbaa5dd"} Mar 13 14:22:09 crc kubenswrapper[4907]: I0313 14:22:09.080989 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-fkplv" event={"ID":"5564ca73-f820-49a4-affa-fd718d1ad7af","Type":"ContainerStarted","Data":"f211dd70f0a3af32a37f299297d692f784b97297cafc168f9d7b88a1639001e6"} Mar 13 14:22:09 crc kubenswrapper[4907]: I0313 14:22:09.081028 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-fkplv" event={"ID":"5564ca73-f820-49a4-affa-fd718d1ad7af","Type":"ContainerStarted","Data":"174f46432910e0a6e62d1e4d69e213487934101226dbd3e79892f20e63afe1b5"} Mar 13 14:22:09 crc kubenswrapper[4907]: I0313 14:22:09.082552 4907 generic.go:334] "Generic (PLEG): container finished" podID="4211d5e0-93f1-4c48-a60e-6c89ae200a59" containerID="4cc7ca19806ba6c30546ebc2e565b8d92a5693e0b133afb20dfc043acb245e70" exitCode=0 Mar 13 14:22:09 crc kubenswrapper[4907]: I0313 14:22:09.082599 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d4jn9" event={"ID":"4211d5e0-93f1-4c48-a60e-6c89ae200a59","Type":"ContainerDied","Data":"4cc7ca19806ba6c30546ebc2e565b8d92a5693e0b133afb20dfc043acb245e70"} Mar 13 14:22:09 crc kubenswrapper[4907]: I0313 14:22:09.082624 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d4jn9" event={"ID":"4211d5e0-93f1-4c48-a60e-6c89ae200a59","Type":"ContainerStarted","Data":"9ed471a362cc5821dd762507d4b70cdc1ef3e27ed3258ebfcd19ca5d91f1ce3f"} Mar 13 14:22:09 crc kubenswrapper[4907]: I0313 14:22:09.094287 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-7bb4cc7c98-2g4kt" podStartSLOduration=3.0942682 podStartE2EDuration="3.0942682s" podCreationTimestamp="2026-03-13 14:22:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:22:09.090772823 +0000 UTC m=+1027.990560532" watchObservedRunningTime="2026-03-13 14:22:09.0942682 +0000 UTC m=+1027.994055889" Mar 13 14:22:10 crc kubenswrapper[4907]: I0313 14:22:10.094171 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-fkplv" event={"ID":"5564ca73-f820-49a4-affa-fd718d1ad7af","Type":"ContainerStarted","Data":"6b4f1ac5209d94c3433caa6ad4775427996e7d7ed104c04a12e543aa7f15299f"} Mar 13 14:22:10 crc kubenswrapper[4907]: I0313 14:22:10.094608 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-fkplv" Mar 13 14:22:10 crc kubenswrapper[4907]: I0313 14:22:10.096985 4907 generic.go:334] "Generic (PLEG): container finished" podID="4211d5e0-93f1-4c48-a60e-6c89ae200a59" containerID="1006ba55273fa47807d756205c115736def78e9d78f5d0bc2e7b68771fa457bf" exitCode=0 Mar 13 14:22:10 crc kubenswrapper[4907]: I0313 14:22:10.097335 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d4jn9" event={"ID":"4211d5e0-93f1-4c48-a60e-6c89ae200a59","Type":"ContainerDied","Data":"1006ba55273fa47807d756205c115736def78e9d78f5d0bc2e7b68771fa457bf"} Mar 13 14:22:10 crc kubenswrapper[4907]: I0313 14:22:10.118861 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-fkplv" podStartSLOduration=4.118843463 podStartE2EDuration="4.118843463s" podCreationTimestamp="2026-03-13 14:22:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:22:10.115255284 +0000 UTC m=+1029.015042983" watchObservedRunningTime="2026-03-13 14:22:10.118843463 +0000 UTC m=+1029.018631152" Mar 13 14:22:11 crc kubenswrapper[4907]: I0313 14:22:11.109572 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d4jn9" event={"ID":"4211d5e0-93f1-4c48-a60e-6c89ae200a59","Type":"ContainerStarted","Data":"a91b547a2dfe97aa6285764503c56a2bb34d57df00449c14921d3d722377c73a"} Mar 13 14:22:11 crc kubenswrapper[4907]: I0313 14:22:11.129530 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-d4jn9" podStartSLOduration=2.679355529 podStartE2EDuration="4.129512544s" podCreationTimestamp="2026-03-13 14:22:07 +0000 UTC" firstStartedPulling="2026-03-13 14:22:09.089807756 +0000 UTC m=+1027.989595445" lastFinishedPulling="2026-03-13 14:22:10.539964781 +0000 UTC m=+1029.439752460" observedRunningTime="2026-03-13 14:22:11.129140663 +0000 UTC m=+1030.028928352" watchObservedRunningTime="2026-03-13 14:22:11.129512544 +0000 UTC m=+1030.029300233" Mar 13 14:22:15 crc kubenswrapper[4907]: I0313 14:22:15.146648 4907 generic.go:334] "Generic (PLEG): container finished" podID="f5203ff6-d473-484d-ab8f-414e9ca804d4" containerID="ef64bf9c95353ba32a6b642cc6315c18c1791829c19b951c20bd6c3207c661ff" exitCode=0 Mar 13 14:22:15 crc kubenswrapper[4907]: I0313 14:22:15.146792 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-6r77v" event={"ID":"f5203ff6-d473-484d-ab8f-414e9ca804d4","Type":"ContainerDied","Data":"ef64bf9c95353ba32a6b642cc6315c18c1791829c19b951c20bd6c3207c661ff"} Mar 13 14:22:15 crc kubenswrapper[4907]: I0313 14:22:15.152846 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-jlsw9" event={"ID":"21498ef9-03ec-44de-ade5-cc56f504499b","Type":"ContainerStarted","Data":"05914453e76dad3d66b07b4cb9ed0eeb79c3f0bd773c8b86deba22bd595b08ad"} Mar 13 14:22:15 crc kubenswrapper[4907]: I0313 14:22:15.153000 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-jlsw9" Mar 13 14:22:15 crc kubenswrapper[4907]: I0313 14:22:15.191590 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-jlsw9" podStartSLOduration=1.9440115919999998 podStartE2EDuration="9.191560298s" podCreationTimestamp="2026-03-13 14:22:06 +0000 UTC" firstStartedPulling="2026-03-13 14:22:07.522412181 +0000 UTC m=+1026.422199860" lastFinishedPulling="2026-03-13 14:22:14.769960847 +0000 UTC m=+1033.669748566" observedRunningTime="2026-03-13 14:22:15.184634056 +0000 UTC m=+1034.084421745" watchObservedRunningTime="2026-03-13 14:22:15.191560298 +0000 UTC m=+1034.091347987" Mar 13 14:22:16 crc kubenswrapper[4907]: I0313 14:22:16.162158 4907 generic.go:334] "Generic (PLEG): container finished" podID="f5203ff6-d473-484d-ab8f-414e9ca804d4" containerID="40f2f957691e4f361abb6febbbedd135054fe0358dc3de7631f107f4850803fb" exitCode=0 Mar 13 14:22:16 crc kubenswrapper[4907]: I0313 14:22:16.162227 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-6r77v" event={"ID":"f5203ff6-d473-484d-ab8f-414e9ca804d4","Type":"ContainerDied","Data":"40f2f957691e4f361abb6febbbedd135054fe0358dc3de7631f107f4850803fb"} Mar 13 14:22:17 crc kubenswrapper[4907]: I0313 14:22:17.170080 4907 generic.go:334] "Generic (PLEG): container finished" podID="f5203ff6-d473-484d-ab8f-414e9ca804d4" containerID="67fa2c802fc206f29dd1e5d6af6be86dda8bb45ad7adda57ab07a6246fe5da5c" exitCode=0 Mar 13 14:22:17 crc kubenswrapper[4907]: I0313 14:22:17.170157 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-6r77v" event={"ID":"f5203ff6-d473-484d-ab8f-414e9ca804d4","Type":"ContainerDied","Data":"67fa2c802fc206f29dd1e5d6af6be86dda8bb45ad7adda57ab07a6246fe5da5c"} Mar 13 14:22:18 crc kubenswrapper[4907]: I0313 14:22:18.041354 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 14:22:18 crc kubenswrapper[4907]: I0313 14:22:18.041708 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 14:22:18 crc kubenswrapper[4907]: I0313 14:22:18.041752 4907 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" Mar 13 14:22:18 crc kubenswrapper[4907]: I0313 14:22:18.042409 4907 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"abe5ce1bc0297c29b962bd8d44bae1a474036514922a5df8e114b5dafeb79cce"} pod="openshift-machine-config-operator/machine-config-daemon-hm56j" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 14:22:18 crc kubenswrapper[4907]: I0313 14:22:18.042467 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" containerID="cri-o://abe5ce1bc0297c29b962bd8d44bae1a474036514922a5df8e114b5dafeb79cce" gracePeriod=600 Mar 13 14:22:18 crc kubenswrapper[4907]: I0313 14:22:18.181860 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-d4jn9" Mar 13 14:22:18 crc kubenswrapper[4907]: I0313 14:22:18.181927 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-d4jn9" Mar 13 14:22:18 crc kubenswrapper[4907]: I0313 14:22:18.197020 4907 generic.go:334] "Generic (PLEG): container finished" podID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerID="abe5ce1bc0297c29b962bd8d44bae1a474036514922a5df8e114b5dafeb79cce" exitCode=0 Mar 13 14:22:18 crc kubenswrapper[4907]: I0313 14:22:18.197102 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerDied","Data":"abe5ce1bc0297c29b962bd8d44bae1a474036514922a5df8e114b5dafeb79cce"} Mar 13 14:22:18 crc kubenswrapper[4907]: I0313 14:22:18.197215 4907 scope.go:117] "RemoveContainer" containerID="da83df5fde38cf429de6b21ade144e7c74078db3c329650c84ad8d669e1beb0e" Mar 13 14:22:18 crc kubenswrapper[4907]: I0313 14:22:18.204188 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-6r77v" event={"ID":"f5203ff6-d473-484d-ab8f-414e9ca804d4","Type":"ContainerStarted","Data":"23c7904ad4905889eb3b476b6012e61cd0db81357dc282484c1273f664551e4a"} Mar 13 14:22:18 crc kubenswrapper[4907]: I0313 14:22:18.204261 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-6r77v" event={"ID":"f5203ff6-d473-484d-ab8f-414e9ca804d4","Type":"ContainerStarted","Data":"a59c2cd2a86e447386b98c8ae0b549d25e103c6e0aa54f07054070409d8496ee"} Mar 13 14:22:18 crc kubenswrapper[4907]: I0313 14:22:18.204275 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-6r77v" event={"ID":"f5203ff6-d473-484d-ab8f-414e9ca804d4","Type":"ContainerStarted","Data":"a977125b67b9294bb56e9b61199b7602c12da5c9c2332514a161007f18a9303c"} Mar 13 14:22:18 crc kubenswrapper[4907]: I0313 14:22:18.204309 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-6r77v" event={"ID":"f5203ff6-d473-484d-ab8f-414e9ca804d4","Type":"ContainerStarted","Data":"dbf34e412a3edcde991c53a218c3ca5c91ab2414ac507c8adb456ed9fa61e44e"} Mar 13 14:22:18 crc kubenswrapper[4907]: I0313 14:22:18.204324 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-6r77v" event={"ID":"f5203ff6-d473-484d-ab8f-414e9ca804d4","Type":"ContainerStarted","Data":"333fbdf93c4f3c0f5b37ac3c5311ed6b367753cf5d440e8f958c6e712bb2e4d1"} Mar 13 14:22:18 crc kubenswrapper[4907]: I0313 14:22:18.236135 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-d4jn9" Mar 13 14:22:19 crc kubenswrapper[4907]: I0313 14:22:19.219919 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-6r77v" event={"ID":"f5203ff6-d473-484d-ab8f-414e9ca804d4","Type":"ContainerStarted","Data":"02fa6003694dc80306fe15a5f7194247a3867e3bda8fe6e237d269d53553a6a0"} Mar 13 14:22:19 crc kubenswrapper[4907]: I0313 14:22:19.220478 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-6r77v" Mar 13 14:22:19 crc kubenswrapper[4907]: I0313 14:22:19.223528 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerStarted","Data":"ac423129eeb1fa8eb8df325acb5ccb91e119992cd551dd9d8c11d304cb95b736"} Mar 13 14:22:19 crc kubenswrapper[4907]: I0313 14:22:19.259739 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-6r77v" podStartSLOduration=5.991184115 podStartE2EDuration="13.25971355s" podCreationTimestamp="2026-03-13 14:22:06 +0000 UTC" firstStartedPulling="2026-03-13 14:22:07.480587326 +0000 UTC m=+1026.380375015" lastFinishedPulling="2026-03-13 14:22:14.749116761 +0000 UTC m=+1033.648904450" observedRunningTime="2026-03-13 14:22:19.254451896 +0000 UTC m=+1038.154239595" watchObservedRunningTime="2026-03-13 14:22:19.25971355 +0000 UTC m=+1038.159501239" Mar 13 14:22:19 crc kubenswrapper[4907]: I0313 14:22:19.333793 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-d4jn9" Mar 13 14:22:19 crc kubenswrapper[4907]: I0313 14:22:19.380586 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-d4jn9"] Mar 13 14:22:21 crc kubenswrapper[4907]: I0313 14:22:21.239123 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-d4jn9" podUID="4211d5e0-93f1-4c48-a60e-6c89ae200a59" containerName="registry-server" containerID="cri-o://a91b547a2dfe97aa6285764503c56a2bb34d57df00449c14921d3d722377c73a" gracePeriod=2 Mar 13 14:22:22 crc kubenswrapper[4907]: I0313 14:22:22.245475 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-6r77v" Mar 13 14:22:22 crc kubenswrapper[4907]: I0313 14:22:22.250010 4907 generic.go:334] "Generic (PLEG): container finished" podID="4211d5e0-93f1-4c48-a60e-6c89ae200a59" containerID="a91b547a2dfe97aa6285764503c56a2bb34d57df00449c14921d3d722377c73a" exitCode=0 Mar 13 14:22:22 crc kubenswrapper[4907]: I0313 14:22:22.250069 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d4jn9" event={"ID":"4211d5e0-93f1-4c48-a60e-6c89ae200a59","Type":"ContainerDied","Data":"a91b547a2dfe97aa6285764503c56a2bb34d57df00449c14921d3d722377c73a"} Mar 13 14:22:22 crc kubenswrapper[4907]: I0313 14:22:22.288661 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-6r77v" Mar 13 14:22:22 crc kubenswrapper[4907]: I0313 14:22:22.487464 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-d4jn9" Mar 13 14:22:22 crc kubenswrapper[4907]: I0313 14:22:22.614340 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4211d5e0-93f1-4c48-a60e-6c89ae200a59-catalog-content\") pod \"4211d5e0-93f1-4c48-a60e-6c89ae200a59\" (UID: \"4211d5e0-93f1-4c48-a60e-6c89ae200a59\") " Mar 13 14:22:22 crc kubenswrapper[4907]: I0313 14:22:22.614855 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-95pxs\" (UniqueName: \"kubernetes.io/projected/4211d5e0-93f1-4c48-a60e-6c89ae200a59-kube-api-access-95pxs\") pod \"4211d5e0-93f1-4c48-a60e-6c89ae200a59\" (UID: \"4211d5e0-93f1-4c48-a60e-6c89ae200a59\") " Mar 13 14:22:22 crc kubenswrapper[4907]: I0313 14:22:22.614947 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4211d5e0-93f1-4c48-a60e-6c89ae200a59-utilities\") pod \"4211d5e0-93f1-4c48-a60e-6c89ae200a59\" (UID: \"4211d5e0-93f1-4c48-a60e-6c89ae200a59\") " Mar 13 14:22:22 crc kubenswrapper[4907]: I0313 14:22:22.615826 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4211d5e0-93f1-4c48-a60e-6c89ae200a59-utilities" (OuterVolumeSpecName: "utilities") pod "4211d5e0-93f1-4c48-a60e-6c89ae200a59" (UID: "4211d5e0-93f1-4c48-a60e-6c89ae200a59"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:22:22 crc kubenswrapper[4907]: I0313 14:22:22.620581 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4211d5e0-93f1-4c48-a60e-6c89ae200a59-kube-api-access-95pxs" (OuterVolumeSpecName: "kube-api-access-95pxs") pod "4211d5e0-93f1-4c48-a60e-6c89ae200a59" (UID: "4211d5e0-93f1-4c48-a60e-6c89ae200a59"). InnerVolumeSpecName "kube-api-access-95pxs". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:22:22 crc kubenswrapper[4907]: I0313 14:22:22.672862 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4211d5e0-93f1-4c48-a60e-6c89ae200a59-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4211d5e0-93f1-4c48-a60e-6c89ae200a59" (UID: "4211d5e0-93f1-4c48-a60e-6c89ae200a59"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:22:22 crc kubenswrapper[4907]: I0313 14:22:22.716373 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4211d5e0-93f1-4c48-a60e-6c89ae200a59-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 14:22:22 crc kubenswrapper[4907]: I0313 14:22:22.716431 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-95pxs\" (UniqueName: \"kubernetes.io/projected/4211d5e0-93f1-4c48-a60e-6c89ae200a59-kube-api-access-95pxs\") on node \"crc\" DevicePath \"\"" Mar 13 14:22:22 crc kubenswrapper[4907]: I0313 14:22:22.716452 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4211d5e0-93f1-4c48-a60e-6c89ae200a59-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 14:22:23 crc kubenswrapper[4907]: I0313 14:22:23.264641 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-d4jn9" Mar 13 14:22:23 crc kubenswrapper[4907]: I0313 14:22:23.264637 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d4jn9" event={"ID":"4211d5e0-93f1-4c48-a60e-6c89ae200a59","Type":"ContainerDied","Data":"9ed471a362cc5821dd762507d4b70cdc1ef3e27ed3258ebfcd19ca5d91f1ce3f"} Mar 13 14:22:23 crc kubenswrapper[4907]: I0313 14:22:23.264810 4907 scope.go:117] "RemoveContainer" containerID="a91b547a2dfe97aa6285764503c56a2bb34d57df00449c14921d3d722377c73a" Mar 13 14:22:23 crc kubenswrapper[4907]: I0313 14:22:23.289439 4907 scope.go:117] "RemoveContainer" containerID="1006ba55273fa47807d756205c115736def78e9d78f5d0bc2e7b68771fa457bf" Mar 13 14:22:23 crc kubenswrapper[4907]: I0313 14:22:23.326346 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-d4jn9"] Mar 13 14:22:23 crc kubenswrapper[4907]: I0313 14:22:23.333595 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-d4jn9"] Mar 13 14:22:23 crc kubenswrapper[4907]: I0313 14:22:23.346312 4907 scope.go:117] "RemoveContainer" containerID="4cc7ca19806ba6c30546ebc2e565b8d92a5693e0b133afb20dfc043acb245e70" Mar 13 14:22:23 crc kubenswrapper[4907]: I0313 14:22:23.799528 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4211d5e0-93f1-4c48-a60e-6c89ae200a59" path="/var/lib/kubelet/pods/4211d5e0-93f1-4c48-a60e-6c89ae200a59/volumes" Mar 13 14:22:27 crc kubenswrapper[4907]: I0313 14:22:27.242455 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-jlsw9" Mar 13 14:22:27 crc kubenswrapper[4907]: I0313 14:22:27.247213 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-6r77v" Mar 13 14:22:27 crc kubenswrapper[4907]: I0313 14:22:27.932761 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-7bb4cc7c98-2g4kt" Mar 13 14:22:28 crc kubenswrapper[4907]: I0313 14:22:28.818287 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-fkplv" Mar 13 14:22:30 crc kubenswrapper[4907]: I0313 14:22:30.302437 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5snd4g"] Mar 13 14:22:30 crc kubenswrapper[4907]: E0313 14:22:30.302652 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4211d5e0-93f1-4c48-a60e-6c89ae200a59" containerName="registry-server" Mar 13 14:22:30 crc kubenswrapper[4907]: I0313 14:22:30.302664 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="4211d5e0-93f1-4c48-a60e-6c89ae200a59" containerName="registry-server" Mar 13 14:22:30 crc kubenswrapper[4907]: E0313 14:22:30.302674 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4211d5e0-93f1-4c48-a60e-6c89ae200a59" containerName="extract-utilities" Mar 13 14:22:30 crc kubenswrapper[4907]: I0313 14:22:30.302680 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="4211d5e0-93f1-4c48-a60e-6c89ae200a59" containerName="extract-utilities" Mar 13 14:22:30 crc kubenswrapper[4907]: E0313 14:22:30.302689 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4211d5e0-93f1-4c48-a60e-6c89ae200a59" containerName="extract-content" Mar 13 14:22:30 crc kubenswrapper[4907]: I0313 14:22:30.302695 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="4211d5e0-93f1-4c48-a60e-6c89ae200a59" containerName="extract-content" Mar 13 14:22:30 crc kubenswrapper[4907]: I0313 14:22:30.302791 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="4211d5e0-93f1-4c48-a60e-6c89ae200a59" containerName="registry-server" Mar 13 14:22:30 crc kubenswrapper[4907]: I0313 14:22:30.303586 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5snd4g" Mar 13 14:22:30 crc kubenswrapper[4907]: I0313 14:22:30.305526 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Mar 13 14:22:30 crc kubenswrapper[4907]: I0313 14:22:30.313630 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5snd4g"] Mar 13 14:22:30 crc kubenswrapper[4907]: I0313 14:22:30.352438 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1ba163f7-0dce-42a8-ab55-15aba1940065-bundle\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5snd4g\" (UID: \"1ba163f7-0dce-42a8-ab55-15aba1940065\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5snd4g" Mar 13 14:22:30 crc kubenswrapper[4907]: I0313 14:22:30.352779 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xg5pk\" (UniqueName: \"kubernetes.io/projected/1ba163f7-0dce-42a8-ab55-15aba1940065-kube-api-access-xg5pk\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5snd4g\" (UID: \"1ba163f7-0dce-42a8-ab55-15aba1940065\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5snd4g" Mar 13 14:22:30 crc kubenswrapper[4907]: I0313 14:22:30.352809 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1ba163f7-0dce-42a8-ab55-15aba1940065-util\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5snd4g\" (UID: \"1ba163f7-0dce-42a8-ab55-15aba1940065\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5snd4g" Mar 13 14:22:30 crc kubenswrapper[4907]: I0313 14:22:30.453758 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1ba163f7-0dce-42a8-ab55-15aba1940065-bundle\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5snd4g\" (UID: \"1ba163f7-0dce-42a8-ab55-15aba1940065\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5snd4g" Mar 13 14:22:30 crc kubenswrapper[4907]: I0313 14:22:30.453806 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xg5pk\" (UniqueName: \"kubernetes.io/projected/1ba163f7-0dce-42a8-ab55-15aba1940065-kube-api-access-xg5pk\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5snd4g\" (UID: \"1ba163f7-0dce-42a8-ab55-15aba1940065\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5snd4g" Mar 13 14:22:30 crc kubenswrapper[4907]: I0313 14:22:30.453838 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1ba163f7-0dce-42a8-ab55-15aba1940065-util\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5snd4g\" (UID: \"1ba163f7-0dce-42a8-ab55-15aba1940065\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5snd4g" Mar 13 14:22:30 crc kubenswrapper[4907]: I0313 14:22:30.454353 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1ba163f7-0dce-42a8-ab55-15aba1940065-util\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5snd4g\" (UID: \"1ba163f7-0dce-42a8-ab55-15aba1940065\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5snd4g" Mar 13 14:22:30 crc kubenswrapper[4907]: I0313 14:22:30.454346 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1ba163f7-0dce-42a8-ab55-15aba1940065-bundle\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5snd4g\" (UID: \"1ba163f7-0dce-42a8-ab55-15aba1940065\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5snd4g" Mar 13 14:22:30 crc kubenswrapper[4907]: I0313 14:22:30.473297 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xg5pk\" (UniqueName: \"kubernetes.io/projected/1ba163f7-0dce-42a8-ab55-15aba1940065-kube-api-access-xg5pk\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5snd4g\" (UID: \"1ba163f7-0dce-42a8-ab55-15aba1940065\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5snd4g" Mar 13 14:22:30 crc kubenswrapper[4907]: I0313 14:22:30.637580 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5snd4g" Mar 13 14:22:30 crc kubenswrapper[4907]: I0313 14:22:30.873632 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5snd4g"] Mar 13 14:22:31 crc kubenswrapper[4907]: I0313 14:22:31.321529 4907 generic.go:334] "Generic (PLEG): container finished" podID="1ba163f7-0dce-42a8-ab55-15aba1940065" containerID="9a7f8034ae10429297ca47eef0ac8cf7185b4b0681d362f7ac3ab16e8b94f648" exitCode=0 Mar 13 14:22:31 crc kubenswrapper[4907]: I0313 14:22:31.321566 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5snd4g" event={"ID":"1ba163f7-0dce-42a8-ab55-15aba1940065","Type":"ContainerDied","Data":"9a7f8034ae10429297ca47eef0ac8cf7185b4b0681d362f7ac3ab16e8b94f648"} Mar 13 14:22:31 crc kubenswrapper[4907]: I0313 14:22:31.321590 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5snd4g" event={"ID":"1ba163f7-0dce-42a8-ab55-15aba1940065","Type":"ContainerStarted","Data":"015cef428accbaa410a9449f28eee6690287054668a794e7e0447f58adb4d487"} Mar 13 14:22:34 crc kubenswrapper[4907]: I0313 14:22:34.339747 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5snd4g" event={"ID":"1ba163f7-0dce-42a8-ab55-15aba1940065","Type":"ContainerStarted","Data":"79c7a18f632a438161553f4588a9282777d07b1799c22f74e34574606d71fc57"} Mar 13 14:22:35 crc kubenswrapper[4907]: I0313 14:22:35.348613 4907 generic.go:334] "Generic (PLEG): container finished" podID="1ba163f7-0dce-42a8-ab55-15aba1940065" containerID="79c7a18f632a438161553f4588a9282777d07b1799c22f74e34574606d71fc57" exitCode=0 Mar 13 14:22:35 crc kubenswrapper[4907]: I0313 14:22:35.348700 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5snd4g" event={"ID":"1ba163f7-0dce-42a8-ab55-15aba1940065","Type":"ContainerDied","Data":"79c7a18f632a438161553f4588a9282777d07b1799c22f74e34574606d71fc57"} Mar 13 14:22:36 crc kubenswrapper[4907]: I0313 14:22:36.356962 4907 generic.go:334] "Generic (PLEG): container finished" podID="1ba163f7-0dce-42a8-ab55-15aba1940065" containerID="5ad71ff77a738ebdca98e6804ade91d8b524277872fa6f4892e2b3d140b3443f" exitCode=0 Mar 13 14:22:36 crc kubenswrapper[4907]: I0313 14:22:36.357082 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5snd4g" event={"ID":"1ba163f7-0dce-42a8-ab55-15aba1940065","Type":"ContainerDied","Data":"5ad71ff77a738ebdca98e6804ade91d8b524277872fa6f4892e2b3d140b3443f"} Mar 13 14:22:37 crc kubenswrapper[4907]: I0313 14:22:37.629598 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5snd4g" Mar 13 14:22:37 crc kubenswrapper[4907]: I0313 14:22:37.669995 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1ba163f7-0dce-42a8-ab55-15aba1940065-util\") pod \"1ba163f7-0dce-42a8-ab55-15aba1940065\" (UID: \"1ba163f7-0dce-42a8-ab55-15aba1940065\") " Mar 13 14:22:37 crc kubenswrapper[4907]: I0313 14:22:37.670084 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1ba163f7-0dce-42a8-ab55-15aba1940065-bundle\") pod \"1ba163f7-0dce-42a8-ab55-15aba1940065\" (UID: \"1ba163f7-0dce-42a8-ab55-15aba1940065\") " Mar 13 14:22:37 crc kubenswrapper[4907]: I0313 14:22:37.670185 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xg5pk\" (UniqueName: \"kubernetes.io/projected/1ba163f7-0dce-42a8-ab55-15aba1940065-kube-api-access-xg5pk\") pod \"1ba163f7-0dce-42a8-ab55-15aba1940065\" (UID: \"1ba163f7-0dce-42a8-ab55-15aba1940065\") " Mar 13 14:22:37 crc kubenswrapper[4907]: I0313 14:22:37.672929 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1ba163f7-0dce-42a8-ab55-15aba1940065-bundle" (OuterVolumeSpecName: "bundle") pod "1ba163f7-0dce-42a8-ab55-15aba1940065" (UID: "1ba163f7-0dce-42a8-ab55-15aba1940065"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:22:37 crc kubenswrapper[4907]: I0313 14:22:37.678766 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1ba163f7-0dce-42a8-ab55-15aba1940065-kube-api-access-xg5pk" (OuterVolumeSpecName: "kube-api-access-xg5pk") pod "1ba163f7-0dce-42a8-ab55-15aba1940065" (UID: "1ba163f7-0dce-42a8-ab55-15aba1940065"). InnerVolumeSpecName "kube-api-access-xg5pk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:22:37 crc kubenswrapper[4907]: I0313 14:22:37.684056 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1ba163f7-0dce-42a8-ab55-15aba1940065-util" (OuterVolumeSpecName: "util") pod "1ba163f7-0dce-42a8-ab55-15aba1940065" (UID: "1ba163f7-0dce-42a8-ab55-15aba1940065"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:22:37 crc kubenswrapper[4907]: I0313 14:22:37.772213 4907 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1ba163f7-0dce-42a8-ab55-15aba1940065-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:22:37 crc kubenswrapper[4907]: I0313 14:22:37.772275 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xg5pk\" (UniqueName: \"kubernetes.io/projected/1ba163f7-0dce-42a8-ab55-15aba1940065-kube-api-access-xg5pk\") on node \"crc\" DevicePath \"\"" Mar 13 14:22:37 crc kubenswrapper[4907]: I0313 14:22:37.772292 4907 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1ba163f7-0dce-42a8-ab55-15aba1940065-util\") on node \"crc\" DevicePath \"\"" Mar 13 14:22:38 crc kubenswrapper[4907]: I0313 14:22:38.369339 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5snd4g" event={"ID":"1ba163f7-0dce-42a8-ab55-15aba1940065","Type":"ContainerDied","Data":"015cef428accbaa410a9449f28eee6690287054668a794e7e0447f58adb4d487"} Mar 13 14:22:38 crc kubenswrapper[4907]: I0313 14:22:38.369394 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="015cef428accbaa410a9449f28eee6690287054668a794e7e0447f58adb4d487" Mar 13 14:22:38 crc kubenswrapper[4907]: I0313 14:22:38.369414 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5snd4g" Mar 13 14:22:43 crc kubenswrapper[4907]: I0313 14:22:43.422519 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-jzdbj"] Mar 13 14:22:43 crc kubenswrapper[4907]: E0313 14:22:43.423518 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ba163f7-0dce-42a8-ab55-15aba1940065" containerName="util" Mar 13 14:22:43 crc kubenswrapper[4907]: I0313 14:22:43.423531 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ba163f7-0dce-42a8-ab55-15aba1940065" containerName="util" Mar 13 14:22:43 crc kubenswrapper[4907]: E0313 14:22:43.423553 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ba163f7-0dce-42a8-ab55-15aba1940065" containerName="pull" Mar 13 14:22:43 crc kubenswrapper[4907]: I0313 14:22:43.423560 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ba163f7-0dce-42a8-ab55-15aba1940065" containerName="pull" Mar 13 14:22:43 crc kubenswrapper[4907]: E0313 14:22:43.423571 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ba163f7-0dce-42a8-ab55-15aba1940065" containerName="extract" Mar 13 14:22:43 crc kubenswrapper[4907]: I0313 14:22:43.423579 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ba163f7-0dce-42a8-ab55-15aba1940065" containerName="extract" Mar 13 14:22:43 crc kubenswrapper[4907]: I0313 14:22:43.423708 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ba163f7-0dce-42a8-ab55-15aba1940065" containerName="extract" Mar 13 14:22:43 crc kubenswrapper[4907]: I0313 14:22:43.424232 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-jzdbj" Mar 13 14:22:43 crc kubenswrapper[4907]: I0313 14:22:43.426753 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager-operator"/"openshift-service-ca.crt" Mar 13 14:22:43 crc kubenswrapper[4907]: I0313 14:22:43.426824 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager-operator"/"kube-root-ca.crt" Mar 13 14:22:43 crc kubenswrapper[4907]: I0313 14:22:43.427107 4907 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager-operator"/"cert-manager-operator-controller-manager-dockercfg-mjvjd" Mar 13 14:22:43 crc kubenswrapper[4907]: I0313 14:22:43.445628 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-jzdbj"] Mar 13 14:22:43 crc kubenswrapper[4907]: I0313 14:22:43.567901 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/7ff33217-23de-4659-8503-ba87078b122c-tmp\") pod \"cert-manager-operator-controller-manager-66c8bdd694-jzdbj\" (UID: \"7ff33217-23de-4659-8503-ba87078b122c\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-jzdbj" Mar 13 14:22:43 crc kubenswrapper[4907]: I0313 14:22:43.567972 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5vdsn\" (UniqueName: \"kubernetes.io/projected/7ff33217-23de-4659-8503-ba87078b122c-kube-api-access-5vdsn\") pod \"cert-manager-operator-controller-manager-66c8bdd694-jzdbj\" (UID: \"7ff33217-23de-4659-8503-ba87078b122c\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-jzdbj" Mar 13 14:22:43 crc kubenswrapper[4907]: I0313 14:22:43.669820 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/7ff33217-23de-4659-8503-ba87078b122c-tmp\") pod \"cert-manager-operator-controller-manager-66c8bdd694-jzdbj\" (UID: \"7ff33217-23de-4659-8503-ba87078b122c\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-jzdbj" Mar 13 14:22:43 crc kubenswrapper[4907]: I0313 14:22:43.669899 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5vdsn\" (UniqueName: \"kubernetes.io/projected/7ff33217-23de-4659-8503-ba87078b122c-kube-api-access-5vdsn\") pod \"cert-manager-operator-controller-manager-66c8bdd694-jzdbj\" (UID: \"7ff33217-23de-4659-8503-ba87078b122c\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-jzdbj" Mar 13 14:22:43 crc kubenswrapper[4907]: I0313 14:22:43.670642 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/7ff33217-23de-4659-8503-ba87078b122c-tmp\") pod \"cert-manager-operator-controller-manager-66c8bdd694-jzdbj\" (UID: \"7ff33217-23de-4659-8503-ba87078b122c\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-jzdbj" Mar 13 14:22:43 crc kubenswrapper[4907]: I0313 14:22:43.694227 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5vdsn\" (UniqueName: \"kubernetes.io/projected/7ff33217-23de-4659-8503-ba87078b122c-kube-api-access-5vdsn\") pod \"cert-manager-operator-controller-manager-66c8bdd694-jzdbj\" (UID: \"7ff33217-23de-4659-8503-ba87078b122c\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-jzdbj" Mar 13 14:22:43 crc kubenswrapper[4907]: I0313 14:22:43.747341 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-jzdbj" Mar 13 14:22:44 crc kubenswrapper[4907]: I0313 14:22:44.237955 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-jzdbj"] Mar 13 14:22:44 crc kubenswrapper[4907]: W0313 14:22:44.246520 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7ff33217_23de_4659_8503_ba87078b122c.slice/crio-89165b1c30c4ec07deb4d7feafe9b15d6512ed2dfa67ff11476f2079553d3701 WatchSource:0}: Error finding container 89165b1c30c4ec07deb4d7feafe9b15d6512ed2dfa67ff11476f2079553d3701: Status 404 returned error can't find the container with id 89165b1c30c4ec07deb4d7feafe9b15d6512ed2dfa67ff11476f2079553d3701 Mar 13 14:22:44 crc kubenswrapper[4907]: I0313 14:22:44.403838 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-jzdbj" event={"ID":"7ff33217-23de-4659-8503-ba87078b122c","Type":"ContainerStarted","Data":"89165b1c30c4ec07deb4d7feafe9b15d6512ed2dfa67ff11476f2079553d3701"} Mar 13 14:22:48 crc kubenswrapper[4907]: I0313 14:22:48.431225 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-jzdbj" event={"ID":"7ff33217-23de-4659-8503-ba87078b122c","Type":"ContainerStarted","Data":"244d6d5079fbb4cedc0ad864a08b16f344431675a94efbb7435358a38cc6d135"} Mar 13 14:22:48 crc kubenswrapper[4907]: I0313 14:22:48.454052 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-jzdbj" podStartSLOduration=2.21627938 podStartE2EDuration="5.454020321s" podCreationTimestamp="2026-03-13 14:22:43 +0000 UTC" firstStartedPulling="2026-03-13 14:22:44.248243039 +0000 UTC m=+1063.148030728" lastFinishedPulling="2026-03-13 14:22:47.48598398 +0000 UTC m=+1066.385771669" observedRunningTime="2026-03-13 14:22:48.448745764 +0000 UTC m=+1067.348533453" watchObservedRunningTime="2026-03-13 14:22:48.454020321 +0000 UTC m=+1067.353808020" Mar 13 14:22:51 crc kubenswrapper[4907]: I0313 14:22:51.421817 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-6888856db4-tk9wv"] Mar 13 14:22:51 crc kubenswrapper[4907]: I0313 14:22:51.423462 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-6888856db4-tk9wv" Mar 13 14:22:51 crc kubenswrapper[4907]: I0313 14:22:51.426658 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Mar 13 14:22:51 crc kubenswrapper[4907]: I0313 14:22:51.429279 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Mar 13 14:22:51 crc kubenswrapper[4907]: I0313 14:22:51.433189 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-6888856db4-tk9wv"] Mar 13 14:22:51 crc kubenswrapper[4907]: I0313 14:22:51.433373 4907 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-nm4p6" Mar 13 14:22:51 crc kubenswrapper[4907]: I0313 14:22:51.481223 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/3634e065-182f-49b7-af9c-01ea4779b1a6-bound-sa-token\") pod \"cert-manager-webhook-6888856db4-tk9wv\" (UID: \"3634e065-182f-49b7-af9c-01ea4779b1a6\") " pod="cert-manager/cert-manager-webhook-6888856db4-tk9wv" Mar 13 14:22:51 crc kubenswrapper[4907]: I0313 14:22:51.481284 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kgmqf\" (UniqueName: \"kubernetes.io/projected/3634e065-182f-49b7-af9c-01ea4779b1a6-kube-api-access-kgmqf\") pod \"cert-manager-webhook-6888856db4-tk9wv\" (UID: \"3634e065-182f-49b7-af9c-01ea4779b1a6\") " pod="cert-manager/cert-manager-webhook-6888856db4-tk9wv" Mar 13 14:22:51 crc kubenswrapper[4907]: I0313 14:22:51.582354 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/3634e065-182f-49b7-af9c-01ea4779b1a6-bound-sa-token\") pod \"cert-manager-webhook-6888856db4-tk9wv\" (UID: \"3634e065-182f-49b7-af9c-01ea4779b1a6\") " pod="cert-manager/cert-manager-webhook-6888856db4-tk9wv" Mar 13 14:22:51 crc kubenswrapper[4907]: I0313 14:22:51.582431 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kgmqf\" (UniqueName: \"kubernetes.io/projected/3634e065-182f-49b7-af9c-01ea4779b1a6-kube-api-access-kgmqf\") pod \"cert-manager-webhook-6888856db4-tk9wv\" (UID: \"3634e065-182f-49b7-af9c-01ea4779b1a6\") " pod="cert-manager/cert-manager-webhook-6888856db4-tk9wv" Mar 13 14:22:51 crc kubenswrapper[4907]: I0313 14:22:51.607326 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/3634e065-182f-49b7-af9c-01ea4779b1a6-bound-sa-token\") pod \"cert-manager-webhook-6888856db4-tk9wv\" (UID: \"3634e065-182f-49b7-af9c-01ea4779b1a6\") " pod="cert-manager/cert-manager-webhook-6888856db4-tk9wv" Mar 13 14:22:51 crc kubenswrapper[4907]: I0313 14:22:51.607328 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kgmqf\" (UniqueName: \"kubernetes.io/projected/3634e065-182f-49b7-af9c-01ea4779b1a6-kube-api-access-kgmqf\") pod \"cert-manager-webhook-6888856db4-tk9wv\" (UID: \"3634e065-182f-49b7-af9c-01ea4779b1a6\") " pod="cert-manager/cert-manager-webhook-6888856db4-tk9wv" Mar 13 14:22:51 crc kubenswrapper[4907]: I0313 14:22:51.745460 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-6888856db4-tk9wv" Mar 13 14:22:52 crc kubenswrapper[4907]: I0313 14:22:52.245934 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-6888856db4-tk9wv"] Mar 13 14:22:52 crc kubenswrapper[4907]: I0313 14:22:52.453914 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-6888856db4-tk9wv" event={"ID":"3634e065-182f-49b7-af9c-01ea4779b1a6","Type":"ContainerStarted","Data":"093ceea01d9d08e0fe67f760877688892a998bd62331ee65bcfaba6a5f019313"} Mar 13 14:22:54 crc kubenswrapper[4907]: I0313 14:22:54.846987 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-5545bd876-v92wj"] Mar 13 14:22:54 crc kubenswrapper[4907]: I0313 14:22:54.847654 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-5545bd876-v92wj" Mar 13 14:22:54 crc kubenswrapper[4907]: I0313 14:22:54.851217 4907 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-tk6cd" Mar 13 14:22:54 crc kubenswrapper[4907]: I0313 14:22:54.859547 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-5545bd876-v92wj"] Mar 13 14:22:54 crc kubenswrapper[4907]: I0313 14:22:54.927994 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e1799412-2b44-413c-8c9d-b362432fd536-bound-sa-token\") pod \"cert-manager-cainjector-5545bd876-v92wj\" (UID: \"e1799412-2b44-413c-8c9d-b362432fd536\") " pod="cert-manager/cert-manager-cainjector-5545bd876-v92wj" Mar 13 14:22:54 crc kubenswrapper[4907]: I0313 14:22:54.928113 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jlr7g\" (UniqueName: \"kubernetes.io/projected/e1799412-2b44-413c-8c9d-b362432fd536-kube-api-access-jlr7g\") pod \"cert-manager-cainjector-5545bd876-v92wj\" (UID: \"e1799412-2b44-413c-8c9d-b362432fd536\") " pod="cert-manager/cert-manager-cainjector-5545bd876-v92wj" Mar 13 14:22:55 crc kubenswrapper[4907]: I0313 14:22:55.030101 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e1799412-2b44-413c-8c9d-b362432fd536-bound-sa-token\") pod \"cert-manager-cainjector-5545bd876-v92wj\" (UID: \"e1799412-2b44-413c-8c9d-b362432fd536\") " pod="cert-manager/cert-manager-cainjector-5545bd876-v92wj" Mar 13 14:22:55 crc kubenswrapper[4907]: I0313 14:22:55.030207 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jlr7g\" (UniqueName: \"kubernetes.io/projected/e1799412-2b44-413c-8c9d-b362432fd536-kube-api-access-jlr7g\") pod \"cert-manager-cainjector-5545bd876-v92wj\" (UID: \"e1799412-2b44-413c-8c9d-b362432fd536\") " pod="cert-manager/cert-manager-cainjector-5545bd876-v92wj" Mar 13 14:22:55 crc kubenswrapper[4907]: I0313 14:22:55.054110 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e1799412-2b44-413c-8c9d-b362432fd536-bound-sa-token\") pod \"cert-manager-cainjector-5545bd876-v92wj\" (UID: \"e1799412-2b44-413c-8c9d-b362432fd536\") " pod="cert-manager/cert-manager-cainjector-5545bd876-v92wj" Mar 13 14:22:55 crc kubenswrapper[4907]: I0313 14:22:55.054999 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jlr7g\" (UniqueName: \"kubernetes.io/projected/e1799412-2b44-413c-8c9d-b362432fd536-kube-api-access-jlr7g\") pod \"cert-manager-cainjector-5545bd876-v92wj\" (UID: \"e1799412-2b44-413c-8c9d-b362432fd536\") " pod="cert-manager/cert-manager-cainjector-5545bd876-v92wj" Mar 13 14:22:55 crc kubenswrapper[4907]: I0313 14:22:55.223209 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-5545bd876-v92wj" Mar 13 14:22:55 crc kubenswrapper[4907]: I0313 14:22:55.631684 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-5545bd876-v92wj"] Mar 13 14:22:57 crc kubenswrapper[4907]: I0313 14:22:57.555484 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-5545bd876-v92wj" event={"ID":"e1799412-2b44-413c-8c9d-b362432fd536","Type":"ContainerStarted","Data":"6d73d860d2e2c77085d118696a4d100a2ba5c83d6d87b67fd063290b86d671f3"} Mar 13 14:22:57 crc kubenswrapper[4907]: I0313 14:22:57.556663 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-6888856db4-tk9wv" event={"ID":"3634e065-182f-49b7-af9c-01ea4779b1a6","Type":"ContainerStarted","Data":"ce77e2ddb9406e42c13eea05c925689ed9ac92bd180f84bf11d1e3f23c548581"} Mar 13 14:22:57 crc kubenswrapper[4907]: I0313 14:22:57.557035 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-6888856db4-tk9wv" Mar 13 14:22:57 crc kubenswrapper[4907]: I0313 14:22:57.590809 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-6888856db4-tk9wv" podStartSLOduration=1.86427175 podStartE2EDuration="6.590793842s" podCreationTimestamp="2026-03-13 14:22:51 +0000 UTC" firstStartedPulling="2026-03-13 14:22:52.26275769 +0000 UTC m=+1071.162545379" lastFinishedPulling="2026-03-13 14:22:56.989279782 +0000 UTC m=+1075.889067471" observedRunningTime="2026-03-13 14:22:57.586380569 +0000 UTC m=+1076.486168258" watchObservedRunningTime="2026-03-13 14:22:57.590793842 +0000 UTC m=+1076.490581531" Mar 13 14:22:58 crc kubenswrapper[4907]: I0313 14:22:58.562934 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-5545bd876-v92wj" event={"ID":"e1799412-2b44-413c-8c9d-b362432fd536","Type":"ContainerStarted","Data":"f705f01b0f23fb4e01b414400e8950c0fdce958711b3dd3e51bab6e21366b850"} Mar 13 14:22:58 crc kubenswrapper[4907]: I0313 14:22:58.581212 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-5545bd876-v92wj" podStartSLOduration=3.517988513 podStartE2EDuration="4.581191407s" podCreationTimestamp="2026-03-13 14:22:54 +0000 UTC" firstStartedPulling="2026-03-13 14:22:56.926660227 +0000 UTC m=+1075.826447916" lastFinishedPulling="2026-03-13 14:22:57.989863091 +0000 UTC m=+1076.889650810" observedRunningTime="2026-03-13 14:22:58.575859428 +0000 UTC m=+1077.475647117" watchObservedRunningTime="2026-03-13 14:22:58.581191407 +0000 UTC m=+1077.480979096" Mar 13 14:23:02 crc kubenswrapper[4907]: I0313 14:23:02.592679 4907 scope.go:117] "RemoveContainer" containerID="ee5ef069f4132418de863818eaddffd619fd31484fffe36ce7827c7d59840bbd" Mar 13 14:23:04 crc kubenswrapper[4907]: I0313 14:23:04.771799 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-545d4d4674-rn7sp"] Mar 13 14:23:04 crc kubenswrapper[4907]: I0313 14:23:04.773429 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-545d4d4674-rn7sp" Mar 13 14:23:04 crc kubenswrapper[4907]: I0313 14:23:04.778974 4907 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-jvdvb" Mar 13 14:23:04 crc kubenswrapper[4907]: I0313 14:23:04.787409 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-545d4d4674-rn7sp"] Mar 13 14:23:04 crc kubenswrapper[4907]: I0313 14:23:04.888966 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8a238ffb-ab19-472f-bb14-580e63f0fb7d-bound-sa-token\") pod \"cert-manager-545d4d4674-rn7sp\" (UID: \"8a238ffb-ab19-472f-bb14-580e63f0fb7d\") " pod="cert-manager/cert-manager-545d4d4674-rn7sp" Mar 13 14:23:04 crc kubenswrapper[4907]: I0313 14:23:04.889241 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tx9wq\" (UniqueName: \"kubernetes.io/projected/8a238ffb-ab19-472f-bb14-580e63f0fb7d-kube-api-access-tx9wq\") pod \"cert-manager-545d4d4674-rn7sp\" (UID: \"8a238ffb-ab19-472f-bb14-580e63f0fb7d\") " pod="cert-manager/cert-manager-545d4d4674-rn7sp" Mar 13 14:23:04 crc kubenswrapper[4907]: I0313 14:23:04.990449 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tx9wq\" (UniqueName: \"kubernetes.io/projected/8a238ffb-ab19-472f-bb14-580e63f0fb7d-kube-api-access-tx9wq\") pod \"cert-manager-545d4d4674-rn7sp\" (UID: \"8a238ffb-ab19-472f-bb14-580e63f0fb7d\") " pod="cert-manager/cert-manager-545d4d4674-rn7sp" Mar 13 14:23:04 crc kubenswrapper[4907]: I0313 14:23:04.990535 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8a238ffb-ab19-472f-bb14-580e63f0fb7d-bound-sa-token\") pod \"cert-manager-545d4d4674-rn7sp\" (UID: \"8a238ffb-ab19-472f-bb14-580e63f0fb7d\") " pod="cert-manager/cert-manager-545d4d4674-rn7sp" Mar 13 14:23:05 crc kubenswrapper[4907]: I0313 14:23:05.009840 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tx9wq\" (UniqueName: \"kubernetes.io/projected/8a238ffb-ab19-472f-bb14-580e63f0fb7d-kube-api-access-tx9wq\") pod \"cert-manager-545d4d4674-rn7sp\" (UID: \"8a238ffb-ab19-472f-bb14-580e63f0fb7d\") " pod="cert-manager/cert-manager-545d4d4674-rn7sp" Mar 13 14:23:05 crc kubenswrapper[4907]: I0313 14:23:05.010830 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8a238ffb-ab19-472f-bb14-580e63f0fb7d-bound-sa-token\") pod \"cert-manager-545d4d4674-rn7sp\" (UID: \"8a238ffb-ab19-472f-bb14-580e63f0fb7d\") " pod="cert-manager/cert-manager-545d4d4674-rn7sp" Mar 13 14:23:05 crc kubenswrapper[4907]: I0313 14:23:05.122797 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-545d4d4674-rn7sp" Mar 13 14:23:05 crc kubenswrapper[4907]: I0313 14:23:05.549188 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-545d4d4674-rn7sp"] Mar 13 14:23:05 crc kubenswrapper[4907]: W0313 14:23:05.561271 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8a238ffb_ab19_472f_bb14_580e63f0fb7d.slice/crio-584f553807d7e58f6c1ea9e6145adedf6646747aea75f7a1e516745d22377d5d WatchSource:0}: Error finding container 584f553807d7e58f6c1ea9e6145adedf6646747aea75f7a1e516745d22377d5d: Status 404 returned error can't find the container with id 584f553807d7e58f6c1ea9e6145adedf6646747aea75f7a1e516745d22377d5d Mar 13 14:23:05 crc kubenswrapper[4907]: I0313 14:23:05.605231 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-545d4d4674-rn7sp" event={"ID":"8a238ffb-ab19-472f-bb14-580e63f0fb7d","Type":"ContainerStarted","Data":"584f553807d7e58f6c1ea9e6145adedf6646747aea75f7a1e516745d22377d5d"} Mar 13 14:23:06 crc kubenswrapper[4907]: I0313 14:23:06.612853 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-545d4d4674-rn7sp" event={"ID":"8a238ffb-ab19-472f-bb14-580e63f0fb7d","Type":"ContainerStarted","Data":"dd446bcce87574a0ca09e0574b1f618887d39ad0c4423f66b4b23d04fccda7c1"} Mar 13 14:23:06 crc kubenswrapper[4907]: I0313 14:23:06.626823 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-545d4d4674-rn7sp" podStartSLOduration=2.626804645 podStartE2EDuration="2.626804645s" podCreationTimestamp="2026-03-13 14:23:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:23:06.62590826 +0000 UTC m=+1085.525695979" watchObservedRunningTime="2026-03-13 14:23:06.626804645 +0000 UTC m=+1085.526592334" Mar 13 14:23:06 crc kubenswrapper[4907]: I0313 14:23:06.772472 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-6888856db4-tk9wv" Mar 13 14:23:09 crc kubenswrapper[4907]: I0313 14:23:09.971152 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-sp5p9"] Mar 13 14:23:09 crc kubenswrapper[4907]: I0313 14:23:09.972598 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-sp5p9" Mar 13 14:23:09 crc kubenswrapper[4907]: I0313 14:23:09.975369 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-5ljjn" Mar 13 14:23:09 crc kubenswrapper[4907]: I0313 14:23:09.975924 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Mar 13 14:23:09 crc kubenswrapper[4907]: I0313 14:23:09.978116 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Mar 13 14:23:10 crc kubenswrapper[4907]: I0313 14:23:10.028385 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-sp5p9"] Mar 13 14:23:10 crc kubenswrapper[4907]: I0313 14:23:10.065640 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bnxsp\" (UniqueName: \"kubernetes.io/projected/919e211f-e4a5-4af6-8ae8-81e2500ab50b-kube-api-access-bnxsp\") pod \"openstack-operator-index-sp5p9\" (UID: \"919e211f-e4a5-4af6-8ae8-81e2500ab50b\") " pod="openstack-operators/openstack-operator-index-sp5p9" Mar 13 14:23:10 crc kubenswrapper[4907]: I0313 14:23:10.166757 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bnxsp\" (UniqueName: \"kubernetes.io/projected/919e211f-e4a5-4af6-8ae8-81e2500ab50b-kube-api-access-bnxsp\") pod \"openstack-operator-index-sp5p9\" (UID: \"919e211f-e4a5-4af6-8ae8-81e2500ab50b\") " pod="openstack-operators/openstack-operator-index-sp5p9" Mar 13 14:23:10 crc kubenswrapper[4907]: I0313 14:23:10.188608 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bnxsp\" (UniqueName: \"kubernetes.io/projected/919e211f-e4a5-4af6-8ae8-81e2500ab50b-kube-api-access-bnxsp\") pod \"openstack-operator-index-sp5p9\" (UID: \"919e211f-e4a5-4af6-8ae8-81e2500ab50b\") " pod="openstack-operators/openstack-operator-index-sp5p9" Mar 13 14:23:10 crc kubenswrapper[4907]: I0313 14:23:10.295375 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-sp5p9" Mar 13 14:23:10 crc kubenswrapper[4907]: I0313 14:23:10.515090 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-sp5p9"] Mar 13 14:23:10 crc kubenswrapper[4907]: I0313 14:23:10.656256 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-sp5p9" event={"ID":"919e211f-e4a5-4af6-8ae8-81e2500ab50b","Type":"ContainerStarted","Data":"0f9b7376b543a8fb6b597bc980448ce89cc3f22d712c728980f748145ec9f64b"} Mar 13 14:23:11 crc kubenswrapper[4907]: I0313 14:23:11.669577 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-sp5p9" event={"ID":"919e211f-e4a5-4af6-8ae8-81e2500ab50b","Type":"ContainerStarted","Data":"35fa0a9a1b21ee3f2bcb0895ae1cf3a5cf49c10c24083a3e95693b10237b2528"} Mar 13 14:23:11 crc kubenswrapper[4907]: I0313 14:23:11.687224 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-sp5p9" podStartSLOduration=1.9511680120000001 podStartE2EDuration="2.68720042s" podCreationTimestamp="2026-03-13 14:23:09 +0000 UTC" firstStartedPulling="2026-03-13 14:23:10.522080156 +0000 UTC m=+1089.421867845" lastFinishedPulling="2026-03-13 14:23:11.258112564 +0000 UTC m=+1090.157900253" observedRunningTime="2026-03-13 14:23:11.681974894 +0000 UTC m=+1090.581762593" watchObservedRunningTime="2026-03-13 14:23:11.68720042 +0000 UTC m=+1090.586988149" Mar 13 14:23:13 crc kubenswrapper[4907]: I0313 14:23:13.350818 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-sp5p9"] Mar 13 14:23:13 crc kubenswrapper[4907]: I0313 14:23:13.679323 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-sp5p9" podUID="919e211f-e4a5-4af6-8ae8-81e2500ab50b" containerName="registry-server" containerID="cri-o://35fa0a9a1b21ee3f2bcb0895ae1cf3a5cf49c10c24083a3e95693b10237b2528" gracePeriod=2 Mar 13 14:23:13 crc kubenswrapper[4907]: I0313 14:23:13.961069 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-z98nv"] Mar 13 14:23:13 crc kubenswrapper[4907]: I0313 14:23:13.962683 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-z98nv" Mar 13 14:23:13 crc kubenswrapper[4907]: I0313 14:23:13.985640 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-z98nv"] Mar 13 14:23:14 crc kubenswrapper[4907]: I0313 14:23:14.016369 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-prsvc\" (UniqueName: \"kubernetes.io/projected/c23245c8-0fba-483e-b126-09349d413030-kube-api-access-prsvc\") pod \"openstack-operator-index-z98nv\" (UID: \"c23245c8-0fba-483e-b126-09349d413030\") " pod="openstack-operators/openstack-operator-index-z98nv" Mar 13 14:23:14 crc kubenswrapper[4907]: I0313 14:23:14.086121 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-sp5p9" Mar 13 14:23:14 crc kubenswrapper[4907]: I0313 14:23:14.117298 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-prsvc\" (UniqueName: \"kubernetes.io/projected/c23245c8-0fba-483e-b126-09349d413030-kube-api-access-prsvc\") pod \"openstack-operator-index-z98nv\" (UID: \"c23245c8-0fba-483e-b126-09349d413030\") " pod="openstack-operators/openstack-operator-index-z98nv" Mar 13 14:23:14 crc kubenswrapper[4907]: I0313 14:23:14.139827 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-prsvc\" (UniqueName: \"kubernetes.io/projected/c23245c8-0fba-483e-b126-09349d413030-kube-api-access-prsvc\") pod \"openstack-operator-index-z98nv\" (UID: \"c23245c8-0fba-483e-b126-09349d413030\") " pod="openstack-operators/openstack-operator-index-z98nv" Mar 13 14:23:14 crc kubenswrapper[4907]: I0313 14:23:14.218808 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bnxsp\" (UniqueName: \"kubernetes.io/projected/919e211f-e4a5-4af6-8ae8-81e2500ab50b-kube-api-access-bnxsp\") pod \"919e211f-e4a5-4af6-8ae8-81e2500ab50b\" (UID: \"919e211f-e4a5-4af6-8ae8-81e2500ab50b\") " Mar 13 14:23:14 crc kubenswrapper[4907]: I0313 14:23:14.223417 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/919e211f-e4a5-4af6-8ae8-81e2500ab50b-kube-api-access-bnxsp" (OuterVolumeSpecName: "kube-api-access-bnxsp") pod "919e211f-e4a5-4af6-8ae8-81e2500ab50b" (UID: "919e211f-e4a5-4af6-8ae8-81e2500ab50b"). InnerVolumeSpecName "kube-api-access-bnxsp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:23:14 crc kubenswrapper[4907]: I0313 14:23:14.305353 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-z98nv" Mar 13 14:23:14 crc kubenswrapper[4907]: I0313 14:23:14.320176 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bnxsp\" (UniqueName: \"kubernetes.io/projected/919e211f-e4a5-4af6-8ae8-81e2500ab50b-kube-api-access-bnxsp\") on node \"crc\" DevicePath \"\"" Mar 13 14:23:14 crc kubenswrapper[4907]: I0313 14:23:14.688083 4907 generic.go:334] "Generic (PLEG): container finished" podID="919e211f-e4a5-4af6-8ae8-81e2500ab50b" containerID="35fa0a9a1b21ee3f2bcb0895ae1cf3a5cf49c10c24083a3e95693b10237b2528" exitCode=0 Mar 13 14:23:14 crc kubenswrapper[4907]: I0313 14:23:14.688561 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-sp5p9" Mar 13 14:23:14 crc kubenswrapper[4907]: I0313 14:23:14.688578 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-sp5p9" event={"ID":"919e211f-e4a5-4af6-8ae8-81e2500ab50b","Type":"ContainerDied","Data":"35fa0a9a1b21ee3f2bcb0895ae1cf3a5cf49c10c24083a3e95693b10237b2528"} Mar 13 14:23:14 crc kubenswrapper[4907]: I0313 14:23:14.688706 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-sp5p9" event={"ID":"919e211f-e4a5-4af6-8ae8-81e2500ab50b","Type":"ContainerDied","Data":"0f9b7376b543a8fb6b597bc980448ce89cc3f22d712c728980f748145ec9f64b"} Mar 13 14:23:14 crc kubenswrapper[4907]: I0313 14:23:14.688756 4907 scope.go:117] "RemoveContainer" containerID="35fa0a9a1b21ee3f2bcb0895ae1cf3a5cf49c10c24083a3e95693b10237b2528" Mar 13 14:23:14 crc kubenswrapper[4907]: I0313 14:23:14.717837 4907 scope.go:117] "RemoveContainer" containerID="35fa0a9a1b21ee3f2bcb0895ae1cf3a5cf49c10c24083a3e95693b10237b2528" Mar 13 14:23:14 crc kubenswrapper[4907]: E0313 14:23:14.718748 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"35fa0a9a1b21ee3f2bcb0895ae1cf3a5cf49c10c24083a3e95693b10237b2528\": container with ID starting with 35fa0a9a1b21ee3f2bcb0895ae1cf3a5cf49c10c24083a3e95693b10237b2528 not found: ID does not exist" containerID="35fa0a9a1b21ee3f2bcb0895ae1cf3a5cf49c10c24083a3e95693b10237b2528" Mar 13 14:23:14 crc kubenswrapper[4907]: I0313 14:23:14.718821 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"35fa0a9a1b21ee3f2bcb0895ae1cf3a5cf49c10c24083a3e95693b10237b2528"} err="failed to get container status \"35fa0a9a1b21ee3f2bcb0895ae1cf3a5cf49c10c24083a3e95693b10237b2528\": rpc error: code = NotFound desc = could not find container \"35fa0a9a1b21ee3f2bcb0895ae1cf3a5cf49c10c24083a3e95693b10237b2528\": container with ID starting with 35fa0a9a1b21ee3f2bcb0895ae1cf3a5cf49c10c24083a3e95693b10237b2528 not found: ID does not exist" Mar 13 14:23:14 crc kubenswrapper[4907]: I0313 14:23:14.722482 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-sp5p9"] Mar 13 14:23:14 crc kubenswrapper[4907]: I0313 14:23:14.728793 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-sp5p9"] Mar 13 14:23:14 crc kubenswrapper[4907]: I0313 14:23:14.763223 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-z98nv"] Mar 13 14:23:14 crc kubenswrapper[4907]: W0313 14:23:14.769732 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc23245c8_0fba_483e_b126_09349d413030.slice/crio-e06b2c4cd29ed453c67abc329e6e12a8b3bb3370b0dd96514d62c69eca51be86 WatchSource:0}: Error finding container e06b2c4cd29ed453c67abc329e6e12a8b3bb3370b0dd96514d62c69eca51be86: Status 404 returned error can't find the container with id e06b2c4cd29ed453c67abc329e6e12a8b3bb3370b0dd96514d62c69eca51be86 Mar 13 14:23:15 crc kubenswrapper[4907]: I0313 14:23:15.700174 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-z98nv" event={"ID":"c23245c8-0fba-483e-b126-09349d413030","Type":"ContainerStarted","Data":"5b6ec09d34114f19900bf9c8b582572254b25c91acb3e2a5a15a991df195c79b"} Mar 13 14:23:15 crc kubenswrapper[4907]: I0313 14:23:15.700836 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-z98nv" event={"ID":"c23245c8-0fba-483e-b126-09349d413030","Type":"ContainerStarted","Data":"e06b2c4cd29ed453c67abc329e6e12a8b3bb3370b0dd96514d62c69eca51be86"} Mar 13 14:23:15 crc kubenswrapper[4907]: I0313 14:23:15.725625 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-z98nv" podStartSLOduration=2.236400147 podStartE2EDuration="2.725584888s" podCreationTimestamp="2026-03-13 14:23:13 +0000 UTC" firstStartedPulling="2026-03-13 14:23:14.774060526 +0000 UTC m=+1093.673848215" lastFinishedPulling="2026-03-13 14:23:15.263245277 +0000 UTC m=+1094.163032956" observedRunningTime="2026-03-13 14:23:15.712782092 +0000 UTC m=+1094.612569781" watchObservedRunningTime="2026-03-13 14:23:15.725584888 +0000 UTC m=+1094.625372617" Mar 13 14:23:15 crc kubenswrapper[4907]: I0313 14:23:15.793604 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="919e211f-e4a5-4af6-8ae8-81e2500ab50b" path="/var/lib/kubelet/pods/919e211f-e4a5-4af6-8ae8-81e2500ab50b/volumes" Mar 13 14:23:20 crc kubenswrapper[4907]: I0313 14:23:20.366704 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-4tdsc"] Mar 13 14:23:20 crc kubenswrapper[4907]: E0313 14:23:20.368179 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="919e211f-e4a5-4af6-8ae8-81e2500ab50b" containerName="registry-server" Mar 13 14:23:20 crc kubenswrapper[4907]: I0313 14:23:20.368207 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="919e211f-e4a5-4af6-8ae8-81e2500ab50b" containerName="registry-server" Mar 13 14:23:20 crc kubenswrapper[4907]: I0313 14:23:20.368427 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="919e211f-e4a5-4af6-8ae8-81e2500ab50b" containerName="registry-server" Mar 13 14:23:20 crc kubenswrapper[4907]: I0313 14:23:20.370489 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4tdsc" Mar 13 14:23:20 crc kubenswrapper[4907]: I0313 14:23:20.376369 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4tdsc"] Mar 13 14:23:20 crc kubenswrapper[4907]: I0313 14:23:20.432480 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m474c\" (UniqueName: \"kubernetes.io/projected/0a48608e-b0d0-4ff5-963f-0fc5fae809d5-kube-api-access-m474c\") pod \"redhat-marketplace-4tdsc\" (UID: \"0a48608e-b0d0-4ff5-963f-0fc5fae809d5\") " pod="openshift-marketplace/redhat-marketplace-4tdsc" Mar 13 14:23:20 crc kubenswrapper[4907]: I0313 14:23:20.433065 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a48608e-b0d0-4ff5-963f-0fc5fae809d5-catalog-content\") pod \"redhat-marketplace-4tdsc\" (UID: \"0a48608e-b0d0-4ff5-963f-0fc5fae809d5\") " pod="openshift-marketplace/redhat-marketplace-4tdsc" Mar 13 14:23:20 crc kubenswrapper[4907]: I0313 14:23:20.433148 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a48608e-b0d0-4ff5-963f-0fc5fae809d5-utilities\") pod \"redhat-marketplace-4tdsc\" (UID: \"0a48608e-b0d0-4ff5-963f-0fc5fae809d5\") " pod="openshift-marketplace/redhat-marketplace-4tdsc" Mar 13 14:23:20 crc kubenswrapper[4907]: I0313 14:23:20.534776 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a48608e-b0d0-4ff5-963f-0fc5fae809d5-utilities\") pod \"redhat-marketplace-4tdsc\" (UID: \"0a48608e-b0d0-4ff5-963f-0fc5fae809d5\") " pod="openshift-marketplace/redhat-marketplace-4tdsc" Mar 13 14:23:20 crc kubenswrapper[4907]: I0313 14:23:20.534919 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m474c\" (UniqueName: \"kubernetes.io/projected/0a48608e-b0d0-4ff5-963f-0fc5fae809d5-kube-api-access-m474c\") pod \"redhat-marketplace-4tdsc\" (UID: \"0a48608e-b0d0-4ff5-963f-0fc5fae809d5\") " pod="openshift-marketplace/redhat-marketplace-4tdsc" Mar 13 14:23:20 crc kubenswrapper[4907]: I0313 14:23:20.535066 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a48608e-b0d0-4ff5-963f-0fc5fae809d5-catalog-content\") pod \"redhat-marketplace-4tdsc\" (UID: \"0a48608e-b0d0-4ff5-963f-0fc5fae809d5\") " pod="openshift-marketplace/redhat-marketplace-4tdsc" Mar 13 14:23:20 crc kubenswrapper[4907]: I0313 14:23:20.535690 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a48608e-b0d0-4ff5-963f-0fc5fae809d5-utilities\") pod \"redhat-marketplace-4tdsc\" (UID: \"0a48608e-b0d0-4ff5-963f-0fc5fae809d5\") " pod="openshift-marketplace/redhat-marketplace-4tdsc" Mar 13 14:23:20 crc kubenswrapper[4907]: I0313 14:23:20.535824 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a48608e-b0d0-4ff5-963f-0fc5fae809d5-catalog-content\") pod \"redhat-marketplace-4tdsc\" (UID: \"0a48608e-b0d0-4ff5-963f-0fc5fae809d5\") " pod="openshift-marketplace/redhat-marketplace-4tdsc" Mar 13 14:23:20 crc kubenswrapper[4907]: I0313 14:23:20.567918 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m474c\" (UniqueName: \"kubernetes.io/projected/0a48608e-b0d0-4ff5-963f-0fc5fae809d5-kube-api-access-m474c\") pod \"redhat-marketplace-4tdsc\" (UID: \"0a48608e-b0d0-4ff5-963f-0fc5fae809d5\") " pod="openshift-marketplace/redhat-marketplace-4tdsc" Mar 13 14:23:20 crc kubenswrapper[4907]: I0313 14:23:20.702832 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4tdsc" Mar 13 14:23:21 crc kubenswrapper[4907]: I0313 14:23:21.196838 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4tdsc"] Mar 13 14:23:21 crc kubenswrapper[4907]: I0313 14:23:21.765196 4907 generic.go:334] "Generic (PLEG): container finished" podID="0a48608e-b0d0-4ff5-963f-0fc5fae809d5" containerID="e917c4d1488ff6e34a1eae85a4179e14a1fb7da267d8c17bc59f188d9f93a711" exitCode=0 Mar 13 14:23:21 crc kubenswrapper[4907]: I0313 14:23:21.765283 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4tdsc" event={"ID":"0a48608e-b0d0-4ff5-963f-0fc5fae809d5","Type":"ContainerDied","Data":"e917c4d1488ff6e34a1eae85a4179e14a1fb7da267d8c17bc59f188d9f93a711"} Mar 13 14:23:21 crc kubenswrapper[4907]: I0313 14:23:21.765338 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4tdsc" event={"ID":"0a48608e-b0d0-4ff5-963f-0fc5fae809d5","Type":"ContainerStarted","Data":"3ee4542adf3dfb6a230d733a6eecb4aea1254a119b039f5276071b36fe25d5af"} Mar 13 14:23:23 crc kubenswrapper[4907]: I0313 14:23:23.785488 4907 generic.go:334] "Generic (PLEG): container finished" podID="0a48608e-b0d0-4ff5-963f-0fc5fae809d5" containerID="859e57ce5122694ec07507ac13d55d3387cef8b8951ec996ebb84737ad268a3b" exitCode=0 Mar 13 14:23:23 crc kubenswrapper[4907]: I0313 14:23:23.797733 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4tdsc" event={"ID":"0a48608e-b0d0-4ff5-963f-0fc5fae809d5","Type":"ContainerDied","Data":"859e57ce5122694ec07507ac13d55d3387cef8b8951ec996ebb84737ad268a3b"} Mar 13 14:23:24 crc kubenswrapper[4907]: I0313 14:23:24.305567 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-z98nv" Mar 13 14:23:24 crc kubenswrapper[4907]: I0313 14:23:24.305686 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-z98nv" Mar 13 14:23:24 crc kubenswrapper[4907]: I0313 14:23:24.353288 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-z98nv" Mar 13 14:23:24 crc kubenswrapper[4907]: I0313 14:23:24.797002 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4tdsc" event={"ID":"0a48608e-b0d0-4ff5-963f-0fc5fae809d5","Type":"ContainerStarted","Data":"a2bf587b73502d20df6faba1a74fb50eb91bfbf9b9fa5e9342562ea96082f40d"} Mar 13 14:23:24 crc kubenswrapper[4907]: I0313 14:23:24.848415 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-z98nv" Mar 13 14:23:24 crc kubenswrapper[4907]: I0313 14:23:24.852497 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-4tdsc" podStartSLOduration=2.285607132 podStartE2EDuration="4.852451159s" podCreationTimestamp="2026-03-13 14:23:20 +0000 UTC" firstStartedPulling="2026-03-13 14:23:21.767909439 +0000 UTC m=+1100.667697128" lastFinishedPulling="2026-03-13 14:23:24.334753456 +0000 UTC m=+1103.234541155" observedRunningTime="2026-03-13 14:23:24.827251607 +0000 UTC m=+1103.727039336" watchObservedRunningTime="2026-03-13 14:23:24.852451159 +0000 UTC m=+1103.752238868" Mar 13 14:23:25 crc kubenswrapper[4907]: I0313 14:23:25.602798 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/a25da44700f2da9e511716e9ea5c2d7cc3bdaf6532fb20e8f09ee07298p5ddv"] Mar 13 14:23:25 crc kubenswrapper[4907]: I0313 14:23:25.603942 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/a25da44700f2da9e511716e9ea5c2d7cc3bdaf6532fb20e8f09ee07298p5ddv" Mar 13 14:23:25 crc kubenswrapper[4907]: I0313 14:23:25.608874 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-kzqgh" Mar 13 14:23:25 crc kubenswrapper[4907]: I0313 14:23:25.617754 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/a25da44700f2da9e511716e9ea5c2d7cc3bdaf6532fb20e8f09ee07298p5ddv"] Mar 13 14:23:25 crc kubenswrapper[4907]: I0313 14:23:25.723524 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8wlxv\" (UniqueName: \"kubernetes.io/projected/29e80c51-f1f5-4850-8e81-ee52fe060a43-kube-api-access-8wlxv\") pod \"a25da44700f2da9e511716e9ea5c2d7cc3bdaf6532fb20e8f09ee07298p5ddv\" (UID: \"29e80c51-f1f5-4850-8e81-ee52fe060a43\") " pod="openstack-operators/a25da44700f2da9e511716e9ea5c2d7cc3bdaf6532fb20e8f09ee07298p5ddv" Mar 13 14:23:25 crc kubenswrapper[4907]: I0313 14:23:25.723641 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/29e80c51-f1f5-4850-8e81-ee52fe060a43-util\") pod \"a25da44700f2da9e511716e9ea5c2d7cc3bdaf6532fb20e8f09ee07298p5ddv\" (UID: \"29e80c51-f1f5-4850-8e81-ee52fe060a43\") " pod="openstack-operators/a25da44700f2da9e511716e9ea5c2d7cc3bdaf6532fb20e8f09ee07298p5ddv" Mar 13 14:23:25 crc kubenswrapper[4907]: I0313 14:23:25.723694 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/29e80c51-f1f5-4850-8e81-ee52fe060a43-bundle\") pod \"a25da44700f2da9e511716e9ea5c2d7cc3bdaf6532fb20e8f09ee07298p5ddv\" (UID: \"29e80c51-f1f5-4850-8e81-ee52fe060a43\") " pod="openstack-operators/a25da44700f2da9e511716e9ea5c2d7cc3bdaf6532fb20e8f09ee07298p5ddv" Mar 13 14:23:25 crc kubenswrapper[4907]: I0313 14:23:25.825693 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/29e80c51-f1f5-4850-8e81-ee52fe060a43-util\") pod \"a25da44700f2da9e511716e9ea5c2d7cc3bdaf6532fb20e8f09ee07298p5ddv\" (UID: \"29e80c51-f1f5-4850-8e81-ee52fe060a43\") " pod="openstack-operators/a25da44700f2da9e511716e9ea5c2d7cc3bdaf6532fb20e8f09ee07298p5ddv" Mar 13 14:23:25 crc kubenswrapper[4907]: I0313 14:23:25.826362 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/29e80c51-f1f5-4850-8e81-ee52fe060a43-bundle\") pod \"a25da44700f2da9e511716e9ea5c2d7cc3bdaf6532fb20e8f09ee07298p5ddv\" (UID: \"29e80c51-f1f5-4850-8e81-ee52fe060a43\") " pod="openstack-operators/a25da44700f2da9e511716e9ea5c2d7cc3bdaf6532fb20e8f09ee07298p5ddv" Mar 13 14:23:25 crc kubenswrapper[4907]: I0313 14:23:25.826627 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8wlxv\" (UniqueName: \"kubernetes.io/projected/29e80c51-f1f5-4850-8e81-ee52fe060a43-kube-api-access-8wlxv\") pod \"a25da44700f2da9e511716e9ea5c2d7cc3bdaf6532fb20e8f09ee07298p5ddv\" (UID: \"29e80c51-f1f5-4850-8e81-ee52fe060a43\") " pod="openstack-operators/a25da44700f2da9e511716e9ea5c2d7cc3bdaf6532fb20e8f09ee07298p5ddv" Mar 13 14:23:25 crc kubenswrapper[4907]: I0313 14:23:25.827085 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/29e80c51-f1f5-4850-8e81-ee52fe060a43-util\") pod \"a25da44700f2da9e511716e9ea5c2d7cc3bdaf6532fb20e8f09ee07298p5ddv\" (UID: \"29e80c51-f1f5-4850-8e81-ee52fe060a43\") " pod="openstack-operators/a25da44700f2da9e511716e9ea5c2d7cc3bdaf6532fb20e8f09ee07298p5ddv" Mar 13 14:23:25 crc kubenswrapper[4907]: I0313 14:23:25.827726 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/29e80c51-f1f5-4850-8e81-ee52fe060a43-bundle\") pod \"a25da44700f2da9e511716e9ea5c2d7cc3bdaf6532fb20e8f09ee07298p5ddv\" (UID: \"29e80c51-f1f5-4850-8e81-ee52fe060a43\") " pod="openstack-operators/a25da44700f2da9e511716e9ea5c2d7cc3bdaf6532fb20e8f09ee07298p5ddv" Mar 13 14:23:25 crc kubenswrapper[4907]: I0313 14:23:25.855726 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8wlxv\" (UniqueName: \"kubernetes.io/projected/29e80c51-f1f5-4850-8e81-ee52fe060a43-kube-api-access-8wlxv\") pod \"a25da44700f2da9e511716e9ea5c2d7cc3bdaf6532fb20e8f09ee07298p5ddv\" (UID: \"29e80c51-f1f5-4850-8e81-ee52fe060a43\") " pod="openstack-operators/a25da44700f2da9e511716e9ea5c2d7cc3bdaf6532fb20e8f09ee07298p5ddv" Mar 13 14:23:25 crc kubenswrapper[4907]: I0313 14:23:25.952762 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/a25da44700f2da9e511716e9ea5c2d7cc3bdaf6532fb20e8f09ee07298p5ddv" Mar 13 14:23:26 crc kubenswrapper[4907]: I0313 14:23:26.434168 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/a25da44700f2da9e511716e9ea5c2d7cc3bdaf6532fb20e8f09ee07298p5ddv"] Mar 13 14:23:26 crc kubenswrapper[4907]: I0313 14:23:26.812787 4907 generic.go:334] "Generic (PLEG): container finished" podID="29e80c51-f1f5-4850-8e81-ee52fe060a43" containerID="56614d64f72203351cf4d876cc8dfb5b9282687bb09eeafa3abb1fdbd52371d1" exitCode=0 Mar 13 14:23:26 crc kubenswrapper[4907]: I0313 14:23:26.812848 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/a25da44700f2da9e511716e9ea5c2d7cc3bdaf6532fb20e8f09ee07298p5ddv" event={"ID":"29e80c51-f1f5-4850-8e81-ee52fe060a43","Type":"ContainerDied","Data":"56614d64f72203351cf4d876cc8dfb5b9282687bb09eeafa3abb1fdbd52371d1"} Mar 13 14:23:26 crc kubenswrapper[4907]: I0313 14:23:26.813582 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/a25da44700f2da9e511716e9ea5c2d7cc3bdaf6532fb20e8f09ee07298p5ddv" event={"ID":"29e80c51-f1f5-4850-8e81-ee52fe060a43","Type":"ContainerStarted","Data":"f7a55efc6d71185f02a71437939fed4071aac8483a065cea96bb2638c19b8a80"} Mar 13 14:23:29 crc kubenswrapper[4907]: I0313 14:23:29.832627 4907 generic.go:334] "Generic (PLEG): container finished" podID="29e80c51-f1f5-4850-8e81-ee52fe060a43" containerID="99a13d1aea20b5d81fd740254033bb43e952a711596b576c4767e172d46118a5" exitCode=0 Mar 13 14:23:29 crc kubenswrapper[4907]: I0313 14:23:29.832687 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/a25da44700f2da9e511716e9ea5c2d7cc3bdaf6532fb20e8f09ee07298p5ddv" event={"ID":"29e80c51-f1f5-4850-8e81-ee52fe060a43","Type":"ContainerDied","Data":"99a13d1aea20b5d81fd740254033bb43e952a711596b576c4767e172d46118a5"} Mar 13 14:23:30 crc kubenswrapper[4907]: I0313 14:23:30.703733 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-4tdsc" Mar 13 14:23:30 crc kubenswrapper[4907]: I0313 14:23:30.704096 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-4tdsc" Mar 13 14:23:30 crc kubenswrapper[4907]: I0313 14:23:30.760430 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-4tdsc" Mar 13 14:23:30 crc kubenswrapper[4907]: I0313 14:23:30.842965 4907 generic.go:334] "Generic (PLEG): container finished" podID="29e80c51-f1f5-4850-8e81-ee52fe060a43" containerID="ca026a81b43b3f27ef8a94825e51af847ca34ff14e2cf7bf77947bc3fc682487" exitCode=0 Mar 13 14:23:30 crc kubenswrapper[4907]: I0313 14:23:30.843033 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/a25da44700f2da9e511716e9ea5c2d7cc3bdaf6532fb20e8f09ee07298p5ddv" event={"ID":"29e80c51-f1f5-4850-8e81-ee52fe060a43","Type":"ContainerDied","Data":"ca026a81b43b3f27ef8a94825e51af847ca34ff14e2cf7bf77947bc3fc682487"} Mar 13 14:23:30 crc kubenswrapper[4907]: I0313 14:23:30.891162 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-4tdsc" Mar 13 14:23:32 crc kubenswrapper[4907]: I0313 14:23:32.115313 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/a25da44700f2da9e511716e9ea5c2d7cc3bdaf6532fb20e8f09ee07298p5ddv" Mar 13 14:23:32 crc kubenswrapper[4907]: I0313 14:23:32.117475 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/29e80c51-f1f5-4850-8e81-ee52fe060a43-bundle\") pod \"29e80c51-f1f5-4850-8e81-ee52fe060a43\" (UID: \"29e80c51-f1f5-4850-8e81-ee52fe060a43\") " Mar 13 14:23:32 crc kubenswrapper[4907]: I0313 14:23:32.117621 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8wlxv\" (UniqueName: \"kubernetes.io/projected/29e80c51-f1f5-4850-8e81-ee52fe060a43-kube-api-access-8wlxv\") pod \"29e80c51-f1f5-4850-8e81-ee52fe060a43\" (UID: \"29e80c51-f1f5-4850-8e81-ee52fe060a43\") " Mar 13 14:23:32 crc kubenswrapper[4907]: I0313 14:23:32.117713 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/29e80c51-f1f5-4850-8e81-ee52fe060a43-util\") pod \"29e80c51-f1f5-4850-8e81-ee52fe060a43\" (UID: \"29e80c51-f1f5-4850-8e81-ee52fe060a43\") " Mar 13 14:23:32 crc kubenswrapper[4907]: I0313 14:23:32.118850 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/29e80c51-f1f5-4850-8e81-ee52fe060a43-bundle" (OuterVolumeSpecName: "bundle") pod "29e80c51-f1f5-4850-8e81-ee52fe060a43" (UID: "29e80c51-f1f5-4850-8e81-ee52fe060a43"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:23:32 crc kubenswrapper[4907]: I0313 14:23:32.125157 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/29e80c51-f1f5-4850-8e81-ee52fe060a43-kube-api-access-8wlxv" (OuterVolumeSpecName: "kube-api-access-8wlxv") pod "29e80c51-f1f5-4850-8e81-ee52fe060a43" (UID: "29e80c51-f1f5-4850-8e81-ee52fe060a43"). InnerVolumeSpecName "kube-api-access-8wlxv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:23:32 crc kubenswrapper[4907]: I0313 14:23:32.219063 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8wlxv\" (UniqueName: \"kubernetes.io/projected/29e80c51-f1f5-4850-8e81-ee52fe060a43-kube-api-access-8wlxv\") on node \"crc\" DevicePath \"\"" Mar 13 14:23:32 crc kubenswrapper[4907]: I0313 14:23:32.219103 4907 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/29e80c51-f1f5-4850-8e81-ee52fe060a43-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:23:32 crc kubenswrapper[4907]: I0313 14:23:32.299233 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/29e80c51-f1f5-4850-8e81-ee52fe060a43-util" (OuterVolumeSpecName: "util") pod "29e80c51-f1f5-4850-8e81-ee52fe060a43" (UID: "29e80c51-f1f5-4850-8e81-ee52fe060a43"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:23:32 crc kubenswrapper[4907]: I0313 14:23:32.321570 4907 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/29e80c51-f1f5-4850-8e81-ee52fe060a43-util\") on node \"crc\" DevicePath \"\"" Mar 13 14:23:32 crc kubenswrapper[4907]: I0313 14:23:32.353063 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4tdsc"] Mar 13 14:23:32 crc kubenswrapper[4907]: I0313 14:23:32.865766 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/a25da44700f2da9e511716e9ea5c2d7cc3bdaf6532fb20e8f09ee07298p5ddv" event={"ID":"29e80c51-f1f5-4850-8e81-ee52fe060a43","Type":"ContainerDied","Data":"f7a55efc6d71185f02a71437939fed4071aac8483a065cea96bb2638c19b8a80"} Mar 13 14:23:32 crc kubenswrapper[4907]: I0313 14:23:32.865843 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f7a55efc6d71185f02a71437939fed4071aac8483a065cea96bb2638c19b8a80" Mar 13 14:23:32 crc kubenswrapper[4907]: I0313 14:23:32.865867 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/a25da44700f2da9e511716e9ea5c2d7cc3bdaf6532fb20e8f09ee07298p5ddv" Mar 13 14:23:32 crc kubenswrapper[4907]: I0313 14:23:32.866077 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-4tdsc" podUID="0a48608e-b0d0-4ff5-963f-0fc5fae809d5" containerName="registry-server" containerID="cri-o://a2bf587b73502d20df6faba1a74fb50eb91bfbf9b9fa5e9342562ea96082f40d" gracePeriod=2 Mar 13 14:23:33 crc kubenswrapper[4907]: I0313 14:23:33.251695 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4tdsc" Mar 13 14:23:33 crc kubenswrapper[4907]: I0313 14:23:33.334951 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m474c\" (UniqueName: \"kubernetes.io/projected/0a48608e-b0d0-4ff5-963f-0fc5fae809d5-kube-api-access-m474c\") pod \"0a48608e-b0d0-4ff5-963f-0fc5fae809d5\" (UID: \"0a48608e-b0d0-4ff5-963f-0fc5fae809d5\") " Mar 13 14:23:33 crc kubenswrapper[4907]: I0313 14:23:33.335025 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a48608e-b0d0-4ff5-963f-0fc5fae809d5-catalog-content\") pod \"0a48608e-b0d0-4ff5-963f-0fc5fae809d5\" (UID: \"0a48608e-b0d0-4ff5-963f-0fc5fae809d5\") " Mar 13 14:23:33 crc kubenswrapper[4907]: I0313 14:23:33.335097 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a48608e-b0d0-4ff5-963f-0fc5fae809d5-utilities\") pod \"0a48608e-b0d0-4ff5-963f-0fc5fae809d5\" (UID: \"0a48608e-b0d0-4ff5-963f-0fc5fae809d5\") " Mar 13 14:23:33 crc kubenswrapper[4907]: I0313 14:23:33.336131 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0a48608e-b0d0-4ff5-963f-0fc5fae809d5-utilities" (OuterVolumeSpecName: "utilities") pod "0a48608e-b0d0-4ff5-963f-0fc5fae809d5" (UID: "0a48608e-b0d0-4ff5-963f-0fc5fae809d5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:23:33 crc kubenswrapper[4907]: I0313 14:23:33.345288 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0a48608e-b0d0-4ff5-963f-0fc5fae809d5-kube-api-access-m474c" (OuterVolumeSpecName: "kube-api-access-m474c") pod "0a48608e-b0d0-4ff5-963f-0fc5fae809d5" (UID: "0a48608e-b0d0-4ff5-963f-0fc5fae809d5"). InnerVolumeSpecName "kube-api-access-m474c". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:23:33 crc kubenswrapper[4907]: I0313 14:23:33.372187 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0a48608e-b0d0-4ff5-963f-0fc5fae809d5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0a48608e-b0d0-4ff5-963f-0fc5fae809d5" (UID: "0a48608e-b0d0-4ff5-963f-0fc5fae809d5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:23:33 crc kubenswrapper[4907]: I0313 14:23:33.436238 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m474c\" (UniqueName: \"kubernetes.io/projected/0a48608e-b0d0-4ff5-963f-0fc5fae809d5-kube-api-access-m474c\") on node \"crc\" DevicePath \"\"" Mar 13 14:23:33 crc kubenswrapper[4907]: I0313 14:23:33.436282 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a48608e-b0d0-4ff5-963f-0fc5fae809d5-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 14:23:33 crc kubenswrapper[4907]: I0313 14:23:33.436295 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a48608e-b0d0-4ff5-963f-0fc5fae809d5-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 14:23:33 crc kubenswrapper[4907]: I0313 14:23:33.873294 4907 generic.go:334] "Generic (PLEG): container finished" podID="0a48608e-b0d0-4ff5-963f-0fc5fae809d5" containerID="a2bf587b73502d20df6faba1a74fb50eb91bfbf9b9fa5e9342562ea96082f40d" exitCode=0 Mar 13 14:23:33 crc kubenswrapper[4907]: I0313 14:23:33.873505 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4tdsc" event={"ID":"0a48608e-b0d0-4ff5-963f-0fc5fae809d5","Type":"ContainerDied","Data":"a2bf587b73502d20df6faba1a74fb50eb91bfbf9b9fa5e9342562ea96082f40d"} Mar 13 14:23:33 crc kubenswrapper[4907]: I0313 14:23:33.873579 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4tdsc" Mar 13 14:23:33 crc kubenswrapper[4907]: I0313 14:23:33.873591 4907 scope.go:117] "RemoveContainer" containerID="a2bf587b73502d20df6faba1a74fb50eb91bfbf9b9fa5e9342562ea96082f40d" Mar 13 14:23:33 crc kubenswrapper[4907]: I0313 14:23:33.873576 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4tdsc" event={"ID":"0a48608e-b0d0-4ff5-963f-0fc5fae809d5","Type":"ContainerDied","Data":"3ee4542adf3dfb6a230d733a6eecb4aea1254a119b039f5276071b36fe25d5af"} Mar 13 14:23:33 crc kubenswrapper[4907]: I0313 14:23:33.891412 4907 scope.go:117] "RemoveContainer" containerID="859e57ce5122694ec07507ac13d55d3387cef8b8951ec996ebb84737ad268a3b" Mar 13 14:23:33 crc kubenswrapper[4907]: I0313 14:23:33.892152 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4tdsc"] Mar 13 14:23:33 crc kubenswrapper[4907]: I0313 14:23:33.896871 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-4tdsc"] Mar 13 14:23:33 crc kubenswrapper[4907]: I0313 14:23:33.912234 4907 scope.go:117] "RemoveContainer" containerID="e917c4d1488ff6e34a1eae85a4179e14a1fb7da267d8c17bc59f188d9f93a711" Mar 13 14:23:33 crc kubenswrapper[4907]: I0313 14:23:33.926330 4907 scope.go:117] "RemoveContainer" containerID="a2bf587b73502d20df6faba1a74fb50eb91bfbf9b9fa5e9342562ea96082f40d" Mar 13 14:23:33 crc kubenswrapper[4907]: E0313 14:23:33.927430 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a2bf587b73502d20df6faba1a74fb50eb91bfbf9b9fa5e9342562ea96082f40d\": container with ID starting with a2bf587b73502d20df6faba1a74fb50eb91bfbf9b9fa5e9342562ea96082f40d not found: ID does not exist" containerID="a2bf587b73502d20df6faba1a74fb50eb91bfbf9b9fa5e9342562ea96082f40d" Mar 13 14:23:33 crc kubenswrapper[4907]: I0313 14:23:33.927460 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a2bf587b73502d20df6faba1a74fb50eb91bfbf9b9fa5e9342562ea96082f40d"} err="failed to get container status \"a2bf587b73502d20df6faba1a74fb50eb91bfbf9b9fa5e9342562ea96082f40d\": rpc error: code = NotFound desc = could not find container \"a2bf587b73502d20df6faba1a74fb50eb91bfbf9b9fa5e9342562ea96082f40d\": container with ID starting with a2bf587b73502d20df6faba1a74fb50eb91bfbf9b9fa5e9342562ea96082f40d not found: ID does not exist" Mar 13 14:23:33 crc kubenswrapper[4907]: I0313 14:23:33.927483 4907 scope.go:117] "RemoveContainer" containerID="859e57ce5122694ec07507ac13d55d3387cef8b8951ec996ebb84737ad268a3b" Mar 13 14:23:33 crc kubenswrapper[4907]: E0313 14:23:33.927811 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"859e57ce5122694ec07507ac13d55d3387cef8b8951ec996ebb84737ad268a3b\": container with ID starting with 859e57ce5122694ec07507ac13d55d3387cef8b8951ec996ebb84737ad268a3b not found: ID does not exist" containerID="859e57ce5122694ec07507ac13d55d3387cef8b8951ec996ebb84737ad268a3b" Mar 13 14:23:33 crc kubenswrapper[4907]: I0313 14:23:33.927834 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"859e57ce5122694ec07507ac13d55d3387cef8b8951ec996ebb84737ad268a3b"} err="failed to get container status \"859e57ce5122694ec07507ac13d55d3387cef8b8951ec996ebb84737ad268a3b\": rpc error: code = NotFound desc = could not find container \"859e57ce5122694ec07507ac13d55d3387cef8b8951ec996ebb84737ad268a3b\": container with ID starting with 859e57ce5122694ec07507ac13d55d3387cef8b8951ec996ebb84737ad268a3b not found: ID does not exist" Mar 13 14:23:33 crc kubenswrapper[4907]: I0313 14:23:33.927851 4907 scope.go:117] "RemoveContainer" containerID="e917c4d1488ff6e34a1eae85a4179e14a1fb7da267d8c17bc59f188d9f93a711" Mar 13 14:23:33 crc kubenswrapper[4907]: E0313 14:23:33.929339 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e917c4d1488ff6e34a1eae85a4179e14a1fb7da267d8c17bc59f188d9f93a711\": container with ID starting with e917c4d1488ff6e34a1eae85a4179e14a1fb7da267d8c17bc59f188d9f93a711 not found: ID does not exist" containerID="e917c4d1488ff6e34a1eae85a4179e14a1fb7da267d8c17bc59f188d9f93a711" Mar 13 14:23:33 crc kubenswrapper[4907]: I0313 14:23:33.929363 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e917c4d1488ff6e34a1eae85a4179e14a1fb7da267d8c17bc59f188d9f93a711"} err="failed to get container status \"e917c4d1488ff6e34a1eae85a4179e14a1fb7da267d8c17bc59f188d9f93a711\": rpc error: code = NotFound desc = could not find container \"e917c4d1488ff6e34a1eae85a4179e14a1fb7da267d8c17bc59f188d9f93a711\": container with ID starting with e917c4d1488ff6e34a1eae85a4179e14a1fb7da267d8c17bc59f188d9f93a711 not found: ID does not exist" Mar 13 14:23:35 crc kubenswrapper[4907]: I0313 14:23:35.791762 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0a48608e-b0d0-4ff5-963f-0fc5fae809d5" path="/var/lib/kubelet/pods/0a48608e-b0d0-4ff5-963f-0fc5fae809d5/volumes" Mar 13 14:23:36 crc kubenswrapper[4907]: I0313 14:23:36.647713 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-init-6dc56d8cd6-hxpjr"] Mar 13 14:23:36 crc kubenswrapper[4907]: E0313 14:23:36.648420 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29e80c51-f1f5-4850-8e81-ee52fe060a43" containerName="util" Mar 13 14:23:36 crc kubenswrapper[4907]: I0313 14:23:36.648468 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="29e80c51-f1f5-4850-8e81-ee52fe060a43" containerName="util" Mar 13 14:23:36 crc kubenswrapper[4907]: E0313 14:23:36.648497 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a48608e-b0d0-4ff5-963f-0fc5fae809d5" containerName="extract-content" Mar 13 14:23:36 crc kubenswrapper[4907]: I0313 14:23:36.648511 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a48608e-b0d0-4ff5-963f-0fc5fae809d5" containerName="extract-content" Mar 13 14:23:36 crc kubenswrapper[4907]: E0313 14:23:36.648537 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29e80c51-f1f5-4850-8e81-ee52fe060a43" containerName="pull" Mar 13 14:23:36 crc kubenswrapper[4907]: I0313 14:23:36.648552 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="29e80c51-f1f5-4850-8e81-ee52fe060a43" containerName="pull" Mar 13 14:23:36 crc kubenswrapper[4907]: E0313 14:23:36.648574 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29e80c51-f1f5-4850-8e81-ee52fe060a43" containerName="extract" Mar 13 14:23:36 crc kubenswrapper[4907]: I0313 14:23:36.648589 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="29e80c51-f1f5-4850-8e81-ee52fe060a43" containerName="extract" Mar 13 14:23:36 crc kubenswrapper[4907]: E0313 14:23:36.648618 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a48608e-b0d0-4ff5-963f-0fc5fae809d5" containerName="extract-utilities" Mar 13 14:23:36 crc kubenswrapper[4907]: I0313 14:23:36.648638 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a48608e-b0d0-4ff5-963f-0fc5fae809d5" containerName="extract-utilities" Mar 13 14:23:36 crc kubenswrapper[4907]: E0313 14:23:36.648698 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a48608e-b0d0-4ff5-963f-0fc5fae809d5" containerName="registry-server" Mar 13 14:23:36 crc kubenswrapper[4907]: I0313 14:23:36.648712 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a48608e-b0d0-4ff5-963f-0fc5fae809d5" containerName="registry-server" Mar 13 14:23:36 crc kubenswrapper[4907]: I0313 14:23:36.648992 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="29e80c51-f1f5-4850-8e81-ee52fe060a43" containerName="extract" Mar 13 14:23:36 crc kubenswrapper[4907]: I0313 14:23:36.649040 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a48608e-b0d0-4ff5-963f-0fc5fae809d5" containerName="registry-server" Mar 13 14:23:36 crc kubenswrapper[4907]: I0313 14:23:36.649668 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-init-6dc56d8cd6-hxpjr" Mar 13 14:23:36 crc kubenswrapper[4907]: I0313 14:23:36.652005 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-init-dockercfg-slwjf" Mar 13 14:23:36 crc kubenswrapper[4907]: I0313 14:23:36.667189 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-init-6dc56d8cd6-hxpjr"] Mar 13 14:23:36 crc kubenswrapper[4907]: I0313 14:23:36.678502 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kpcx5\" (UniqueName: \"kubernetes.io/projected/dc51eaa2-313d-4256-a145-3b73b02e2dfe-kube-api-access-kpcx5\") pod \"openstack-operator-controller-init-6dc56d8cd6-hxpjr\" (UID: \"dc51eaa2-313d-4256-a145-3b73b02e2dfe\") " pod="openstack-operators/openstack-operator-controller-init-6dc56d8cd6-hxpjr" Mar 13 14:23:36 crc kubenswrapper[4907]: I0313 14:23:36.779858 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kpcx5\" (UniqueName: \"kubernetes.io/projected/dc51eaa2-313d-4256-a145-3b73b02e2dfe-kube-api-access-kpcx5\") pod \"openstack-operator-controller-init-6dc56d8cd6-hxpjr\" (UID: \"dc51eaa2-313d-4256-a145-3b73b02e2dfe\") " pod="openstack-operators/openstack-operator-controller-init-6dc56d8cd6-hxpjr" Mar 13 14:23:36 crc kubenswrapper[4907]: I0313 14:23:36.803057 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kpcx5\" (UniqueName: \"kubernetes.io/projected/dc51eaa2-313d-4256-a145-3b73b02e2dfe-kube-api-access-kpcx5\") pod \"openstack-operator-controller-init-6dc56d8cd6-hxpjr\" (UID: \"dc51eaa2-313d-4256-a145-3b73b02e2dfe\") " pod="openstack-operators/openstack-operator-controller-init-6dc56d8cd6-hxpjr" Mar 13 14:23:36 crc kubenswrapper[4907]: I0313 14:23:36.973817 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-init-6dc56d8cd6-hxpjr" Mar 13 14:23:37 crc kubenswrapper[4907]: I0313 14:23:37.424341 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-init-6dc56d8cd6-hxpjr"] Mar 13 14:23:37 crc kubenswrapper[4907]: I0313 14:23:37.904404 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-init-6dc56d8cd6-hxpjr" event={"ID":"dc51eaa2-313d-4256-a145-3b73b02e2dfe","Type":"ContainerStarted","Data":"024e1e4036febf3a5b23ce7ce85464428178f1fb3ec1537da29c5d516178dad4"} Mar 13 14:23:41 crc kubenswrapper[4907]: I0313 14:23:41.935434 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-init-6dc56d8cd6-hxpjr" event={"ID":"dc51eaa2-313d-4256-a145-3b73b02e2dfe","Type":"ContainerStarted","Data":"c6bd1731c8cc406b0d04b52d5d560717b4f7b481fb734a5174b56fb30c9b9f82"} Mar 13 14:23:41 crc kubenswrapper[4907]: I0313 14:23:41.936228 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-init-6dc56d8cd6-hxpjr" Mar 13 14:23:41 crc kubenswrapper[4907]: I0313 14:23:41.972692 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-init-6dc56d8cd6-hxpjr" podStartSLOduration=2.303056543 podStartE2EDuration="5.972675457s" podCreationTimestamp="2026-03-13 14:23:36 +0000 UTC" firstStartedPulling="2026-03-13 14:23:37.433175397 +0000 UTC m=+1116.332963086" lastFinishedPulling="2026-03-13 14:23:41.102794311 +0000 UTC m=+1120.002582000" observedRunningTime="2026-03-13 14:23:41.967548574 +0000 UTC m=+1120.867336263" watchObservedRunningTime="2026-03-13 14:23:41.972675457 +0000 UTC m=+1120.872463146" Mar 13 14:23:46 crc kubenswrapper[4907]: I0313 14:23:46.975990 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-init-6dc56d8cd6-hxpjr" Mar 13 14:24:00 crc kubenswrapper[4907]: I0313 14:24:00.136794 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556864-768cg"] Mar 13 14:24:00 crc kubenswrapper[4907]: I0313 14:24:00.138089 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556864-768cg" Mar 13 14:24:00 crc kubenswrapper[4907]: I0313 14:24:00.140411 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 14:24:00 crc kubenswrapper[4907]: I0313 14:24:00.140635 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 14:24:00 crc kubenswrapper[4907]: I0313 14:24:00.141059 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 14:24:00 crc kubenswrapper[4907]: I0313 14:24:00.142357 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556864-768cg"] Mar 13 14:24:00 crc kubenswrapper[4907]: I0313 14:24:00.194347 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2nhqz\" (UniqueName: \"kubernetes.io/projected/32f92b7e-159a-4824-93e7-ea6d49961db8-kube-api-access-2nhqz\") pod \"auto-csr-approver-29556864-768cg\" (UID: \"32f92b7e-159a-4824-93e7-ea6d49961db8\") " pod="openshift-infra/auto-csr-approver-29556864-768cg" Mar 13 14:24:00 crc kubenswrapper[4907]: I0313 14:24:00.295197 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2nhqz\" (UniqueName: \"kubernetes.io/projected/32f92b7e-159a-4824-93e7-ea6d49961db8-kube-api-access-2nhqz\") pod \"auto-csr-approver-29556864-768cg\" (UID: \"32f92b7e-159a-4824-93e7-ea6d49961db8\") " pod="openshift-infra/auto-csr-approver-29556864-768cg" Mar 13 14:24:00 crc kubenswrapper[4907]: I0313 14:24:00.311038 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2nhqz\" (UniqueName: \"kubernetes.io/projected/32f92b7e-159a-4824-93e7-ea6d49961db8-kube-api-access-2nhqz\") pod \"auto-csr-approver-29556864-768cg\" (UID: \"32f92b7e-159a-4824-93e7-ea6d49961db8\") " pod="openshift-infra/auto-csr-approver-29556864-768cg" Mar 13 14:24:00 crc kubenswrapper[4907]: I0313 14:24:00.454507 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556864-768cg" Mar 13 14:24:00 crc kubenswrapper[4907]: I0313 14:24:00.937232 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556864-768cg"] Mar 13 14:24:01 crc kubenswrapper[4907]: I0313 14:24:01.051545 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556864-768cg" event={"ID":"32f92b7e-159a-4824-93e7-ea6d49961db8","Type":"ContainerStarted","Data":"6cc6b642475e740e71d15bc948ee2b3420da216f5037a2a22c34dcc911c783d7"} Mar 13 14:24:03 crc kubenswrapper[4907]: I0313 14:24:03.073688 4907 generic.go:334] "Generic (PLEG): container finished" podID="32f92b7e-159a-4824-93e7-ea6d49961db8" containerID="a53552e35726d3cc22de750fc04d8a15ada7be07f868ed527291e30242034c7d" exitCode=0 Mar 13 14:24:03 crc kubenswrapper[4907]: I0313 14:24:03.073764 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556864-768cg" event={"ID":"32f92b7e-159a-4824-93e7-ea6d49961db8","Type":"ContainerDied","Data":"a53552e35726d3cc22de750fc04d8a15ada7be07f868ed527291e30242034c7d"} Mar 13 14:24:04 crc kubenswrapper[4907]: I0313 14:24:04.366332 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556864-768cg" Mar 13 14:24:04 crc kubenswrapper[4907]: I0313 14:24:04.471849 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2nhqz\" (UniqueName: \"kubernetes.io/projected/32f92b7e-159a-4824-93e7-ea6d49961db8-kube-api-access-2nhqz\") pod \"32f92b7e-159a-4824-93e7-ea6d49961db8\" (UID: \"32f92b7e-159a-4824-93e7-ea6d49961db8\") " Mar 13 14:24:04 crc kubenswrapper[4907]: I0313 14:24:04.477496 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/32f92b7e-159a-4824-93e7-ea6d49961db8-kube-api-access-2nhqz" (OuterVolumeSpecName: "kube-api-access-2nhqz") pod "32f92b7e-159a-4824-93e7-ea6d49961db8" (UID: "32f92b7e-159a-4824-93e7-ea6d49961db8"). InnerVolumeSpecName "kube-api-access-2nhqz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:24:04 crc kubenswrapper[4907]: I0313 14:24:04.573155 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2nhqz\" (UniqueName: \"kubernetes.io/projected/32f92b7e-159a-4824-93e7-ea6d49961db8-kube-api-access-2nhqz\") on node \"crc\" DevicePath \"\"" Mar 13 14:24:05 crc kubenswrapper[4907]: I0313 14:24:05.087502 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556864-768cg" event={"ID":"32f92b7e-159a-4824-93e7-ea6d49961db8","Type":"ContainerDied","Data":"6cc6b642475e740e71d15bc948ee2b3420da216f5037a2a22c34dcc911c783d7"} Mar 13 14:24:05 crc kubenswrapper[4907]: I0313 14:24:05.087802 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6cc6b642475e740e71d15bc948ee2b3420da216f5037a2a22c34dcc911c783d7" Mar 13 14:24:05 crc kubenswrapper[4907]: I0313 14:24:05.087580 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556864-768cg" Mar 13 14:24:05 crc kubenswrapper[4907]: I0313 14:24:05.408744 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556858-bk6rj"] Mar 13 14:24:05 crc kubenswrapper[4907]: I0313 14:24:05.412917 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556858-bk6rj"] Mar 13 14:24:05 crc kubenswrapper[4907]: I0313 14:24:05.793858 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="03fa7dad-f13d-4e5c-8b9c-60b2ef2d272a" path="/var/lib/kubelet/pods/03fa7dad-f13d-4e5c-8b9c-60b2ef2d272a/volumes" Mar 13 14:24:05 crc kubenswrapper[4907]: I0313 14:24:05.884686 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-d47688694-tl8t4"] Mar 13 14:24:05 crc kubenswrapper[4907]: E0313 14:24:05.885028 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32f92b7e-159a-4824-93e7-ea6d49961db8" containerName="oc" Mar 13 14:24:05 crc kubenswrapper[4907]: I0313 14:24:05.885052 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="32f92b7e-159a-4824-93e7-ea6d49961db8" containerName="oc" Mar 13 14:24:05 crc kubenswrapper[4907]: I0313 14:24:05.885214 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="32f92b7e-159a-4824-93e7-ea6d49961db8" containerName="oc" Mar 13 14:24:05 crc kubenswrapper[4907]: I0313 14:24:05.885766 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-d47688694-tl8t4" Mar 13 14:24:05 crc kubenswrapper[4907]: I0313 14:24:05.887680 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-xftd5" Mar 13 14:24:05 crc kubenswrapper[4907]: I0313 14:24:05.899073 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-984cd4dcf-w497d"] Mar 13 14:24:05 crc kubenswrapper[4907]: I0313 14:24:05.899833 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-984cd4dcf-w497d" Mar 13 14:24:05 crc kubenswrapper[4907]: I0313 14:24:05.904323 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-f2bvz" Mar 13 14:24:05 crc kubenswrapper[4907]: I0313 14:24:05.909864 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-984cd4dcf-w497d"] Mar 13 14:24:05 crc kubenswrapper[4907]: I0313 14:24:05.915559 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-d47688694-tl8t4"] Mar 13 14:24:05 crc kubenswrapper[4907]: I0313 14:24:05.924149 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-66d56f6ff4-7cbb2"] Mar 13 14:24:05 crc kubenswrapper[4907]: I0313 14:24:05.924978 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-66d56f6ff4-7cbb2" Mar 13 14:24:05 crc kubenswrapper[4907]: I0313 14:24:05.927794 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-hfwlj" Mar 13 14:24:05 crc kubenswrapper[4907]: I0313 14:24:05.932480 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-5964f64c48-gll76"] Mar 13 14:24:05 crc kubenswrapper[4907]: I0313 14:24:05.933296 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-5964f64c48-gll76" Mar 13 14:24:05 crc kubenswrapper[4907]: I0313 14:24:05.937322 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-84xsv" Mar 13 14:24:05 crc kubenswrapper[4907]: I0313 14:24:05.953349 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-5964f64c48-gll76"] Mar 13 14:24:05 crc kubenswrapper[4907]: I0313 14:24:05.959963 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-66d56f6ff4-7cbb2"] Mar 13 14:24:05 crc kubenswrapper[4907]: I0313 14:24:05.987851 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t4zll\" (UniqueName: \"kubernetes.io/projected/863c77e9-4022-4ab9-8ac7-a4006d22813c-kube-api-access-t4zll\") pod \"barbican-operator-controller-manager-d47688694-tl8t4\" (UID: \"863c77e9-4022-4ab9-8ac7-a4006d22813c\") " pod="openstack-operators/barbican-operator-controller-manager-d47688694-tl8t4" Mar 13 14:24:05 crc kubenswrapper[4907]: I0313 14:24:05.988062 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5gv2c\" (UniqueName: \"kubernetes.io/projected/b7890be5-5773-4562-b1ab-8e647aa16256-kube-api-access-5gv2c\") pod \"glance-operator-controller-manager-5964f64c48-gll76\" (UID: \"b7890be5-5773-4562-b1ab-8e647aa16256\") " pod="openstack-operators/glance-operator-controller-manager-5964f64c48-gll76" Mar 13 14:24:05 crc kubenswrapper[4907]: I0313 14:24:05.988133 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ltpg6\" (UniqueName: \"kubernetes.io/projected/77b1eb71-a455-4ec3-8e17-a847575cf0b0-kube-api-access-ltpg6\") pod \"cinder-operator-controller-manager-984cd4dcf-w497d\" (UID: \"77b1eb71-a455-4ec3-8e17-a847575cf0b0\") " pod="openstack-operators/cinder-operator-controller-manager-984cd4dcf-w497d" Mar 13 14:24:05 crc kubenswrapper[4907]: I0313 14:24:05.988209 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tk68h\" (UniqueName: \"kubernetes.io/projected/8789d213-46a4-4649-add6-51ebe4412c27-kube-api-access-tk68h\") pod \"designate-operator-controller-manager-66d56f6ff4-7cbb2\" (UID: \"8789d213-46a4-4649-add6-51ebe4412c27\") " pod="openstack-operators/designate-operator-controller-manager-66d56f6ff4-7cbb2" Mar 13 14:24:05 crc kubenswrapper[4907]: I0313 14:24:05.988799 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-77b6666d85-874m2"] Mar 13 14:24:05 crc kubenswrapper[4907]: I0313 14:24:05.989540 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-77b6666d85-874m2" Mar 13 14:24:05 crc kubenswrapper[4907]: I0313 14:24:05.992243 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-jfmkp" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.016252 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-77b6666d85-874m2"] Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.023140 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-6d9d6b584d-r7bhq"] Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.023894 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-6d9d6b584d-r7bhq" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.027318 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-6tth2" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.029183 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-54dc5b8f8d-lp7kj"] Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.029940 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-54dc5b8f8d-lp7kj" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.033792 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-5tgz9" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.034091 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.045335 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-5bc894d9b-w5829"] Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.046269 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-5bc894d9b-w5829" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.048489 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-swjxd" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.048653 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-54dc5b8f8d-lp7kj"] Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.060677 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-6d9d6b584d-r7bhq"] Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.076196 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-5bc894d9b-w5829"] Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.087051 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-684f77d66d-7rdzd"] Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.087934 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-684f77d66d-7rdzd" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.092258 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-6tq54" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.093259 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5gv2c\" (UniqueName: \"kubernetes.io/projected/b7890be5-5773-4562-b1ab-8e647aa16256-kube-api-access-5gv2c\") pod \"glance-operator-controller-manager-5964f64c48-gll76\" (UID: \"b7890be5-5773-4562-b1ab-8e647aa16256\") " pod="openstack-operators/glance-operator-controller-manager-5964f64c48-gll76" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.093290 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/64d55221-60c5-4521-ac84-beed4e9b3993-cert\") pod \"infra-operator-controller-manager-54dc5b8f8d-lp7kj\" (UID: \"64d55221-60c5-4521-ac84-beed4e9b3993\") " pod="openstack-operators/infra-operator-controller-manager-54dc5b8f8d-lp7kj" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.093316 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ltpg6\" (UniqueName: \"kubernetes.io/projected/77b1eb71-a455-4ec3-8e17-a847575cf0b0-kube-api-access-ltpg6\") pod \"cinder-operator-controller-manager-984cd4dcf-w497d\" (UID: \"77b1eb71-a455-4ec3-8e17-a847575cf0b0\") " pod="openstack-operators/cinder-operator-controller-manager-984cd4dcf-w497d" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.093337 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4whsv\" (UniqueName: \"kubernetes.io/projected/64d55221-60c5-4521-ac84-beed4e9b3993-kube-api-access-4whsv\") pod \"infra-operator-controller-manager-54dc5b8f8d-lp7kj\" (UID: \"64d55221-60c5-4521-ac84-beed4e9b3993\") " pod="openstack-operators/infra-operator-controller-manager-54dc5b8f8d-lp7kj" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.093355 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dmv9x\" (UniqueName: \"kubernetes.io/projected/8ecaf4b8-808e-44b9-8ae7-6ef9519c4a66-kube-api-access-dmv9x\") pod \"horizon-operator-controller-manager-6d9d6b584d-r7bhq\" (UID: \"8ecaf4b8-808e-44b9-8ae7-6ef9519c4a66\") " pod="openstack-operators/horizon-operator-controller-manager-6d9d6b584d-r7bhq" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.093388 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tk68h\" (UniqueName: \"kubernetes.io/projected/8789d213-46a4-4649-add6-51ebe4412c27-kube-api-access-tk68h\") pod \"designate-operator-controller-manager-66d56f6ff4-7cbb2\" (UID: \"8789d213-46a4-4649-add6-51ebe4412c27\") " pod="openstack-operators/designate-operator-controller-manager-66d56f6ff4-7cbb2" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.093422 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-64424\" (UniqueName: \"kubernetes.io/projected/d3cdd2ca-3e81-480f-876e-2b975d8bc9d8-kube-api-access-64424\") pod \"heat-operator-controller-manager-77b6666d85-874m2\" (UID: \"d3cdd2ca-3e81-480f-876e-2b975d8bc9d8\") " pod="openstack-operators/heat-operator-controller-manager-77b6666d85-874m2" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.093450 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t4zll\" (UniqueName: \"kubernetes.io/projected/863c77e9-4022-4ab9-8ac7-a4006d22813c-kube-api-access-t4zll\") pod \"barbican-operator-controller-manager-d47688694-tl8t4\" (UID: \"863c77e9-4022-4ab9-8ac7-a4006d22813c\") " pod="openstack-operators/barbican-operator-controller-manager-d47688694-tl8t4" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.093478 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-59tqw\" (UniqueName: \"kubernetes.io/projected/19c320fa-6619-45cd-a73e-65834173b786-kube-api-access-59tqw\") pod \"ironic-operator-controller-manager-5bc894d9b-w5829\" (UID: \"19c320fa-6619-45cd-a73e-65834173b786\") " pod="openstack-operators/ironic-operator-controller-manager-5bc894d9b-w5829" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.162513 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5gv2c\" (UniqueName: \"kubernetes.io/projected/b7890be5-5773-4562-b1ab-8e647aa16256-kube-api-access-5gv2c\") pod \"glance-operator-controller-manager-5964f64c48-gll76\" (UID: \"b7890be5-5773-4562-b1ab-8e647aa16256\") " pod="openstack-operators/glance-operator-controller-manager-5964f64c48-gll76" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.175043 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-684f77d66d-7rdzd"] Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.185968 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tk68h\" (UniqueName: \"kubernetes.io/projected/8789d213-46a4-4649-add6-51ebe4412c27-kube-api-access-tk68h\") pod \"designate-operator-controller-manager-66d56f6ff4-7cbb2\" (UID: \"8789d213-46a4-4649-add6-51ebe4412c27\") " pod="openstack-operators/designate-operator-controller-manager-66d56f6ff4-7cbb2" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.192679 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ltpg6\" (UniqueName: \"kubernetes.io/projected/77b1eb71-a455-4ec3-8e17-a847575cf0b0-kube-api-access-ltpg6\") pod \"cinder-operator-controller-manager-984cd4dcf-w497d\" (UID: \"77b1eb71-a455-4ec3-8e17-a847575cf0b0\") " pod="openstack-operators/cinder-operator-controller-manager-984cd4dcf-w497d" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.208258 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t4zll\" (UniqueName: \"kubernetes.io/projected/863c77e9-4022-4ab9-8ac7-a4006d22813c-kube-api-access-t4zll\") pod \"barbican-operator-controller-manager-d47688694-tl8t4\" (UID: \"863c77e9-4022-4ab9-8ac7-a4006d22813c\") " pod="openstack-operators/barbican-operator-controller-manager-d47688694-tl8t4" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.212741 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-64424\" (UniqueName: \"kubernetes.io/projected/d3cdd2ca-3e81-480f-876e-2b975d8bc9d8-kube-api-access-64424\") pod \"heat-operator-controller-manager-77b6666d85-874m2\" (UID: \"d3cdd2ca-3e81-480f-876e-2b975d8bc9d8\") " pod="openstack-operators/heat-operator-controller-manager-77b6666d85-874m2" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.212805 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n7mn9\" (UniqueName: \"kubernetes.io/projected/7d620a90-0e93-41dc-ab39-fc6ea8c461e7-kube-api-access-n7mn9\") pod \"keystone-operator-controller-manager-684f77d66d-7rdzd\" (UID: \"7d620a90-0e93-41dc-ab39-fc6ea8c461e7\") " pod="openstack-operators/keystone-operator-controller-manager-684f77d66d-7rdzd" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.212837 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-59tqw\" (UniqueName: \"kubernetes.io/projected/19c320fa-6619-45cd-a73e-65834173b786-kube-api-access-59tqw\") pod \"ironic-operator-controller-manager-5bc894d9b-w5829\" (UID: \"19c320fa-6619-45cd-a73e-65834173b786\") " pod="openstack-operators/ironic-operator-controller-manager-5bc894d9b-w5829" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.212870 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/64d55221-60c5-4521-ac84-beed4e9b3993-cert\") pod \"infra-operator-controller-manager-54dc5b8f8d-lp7kj\" (UID: \"64d55221-60c5-4521-ac84-beed4e9b3993\") " pod="openstack-operators/infra-operator-controller-manager-54dc5b8f8d-lp7kj" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.212914 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4whsv\" (UniqueName: \"kubernetes.io/projected/64d55221-60c5-4521-ac84-beed4e9b3993-kube-api-access-4whsv\") pod \"infra-operator-controller-manager-54dc5b8f8d-lp7kj\" (UID: \"64d55221-60c5-4521-ac84-beed4e9b3993\") " pod="openstack-operators/infra-operator-controller-manager-54dc5b8f8d-lp7kj" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.212932 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dmv9x\" (UniqueName: \"kubernetes.io/projected/8ecaf4b8-808e-44b9-8ae7-6ef9519c4a66-kube-api-access-dmv9x\") pod \"horizon-operator-controller-manager-6d9d6b584d-r7bhq\" (UID: \"8ecaf4b8-808e-44b9-8ae7-6ef9519c4a66\") " pod="openstack-operators/horizon-operator-controller-manager-6d9d6b584d-r7bhq" Mar 13 14:24:06 crc kubenswrapper[4907]: E0313 14:24:06.213553 4907 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Mar 13 14:24:06 crc kubenswrapper[4907]: E0313 14:24:06.213603 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/64d55221-60c5-4521-ac84-beed4e9b3993-cert podName:64d55221-60c5-4521-ac84-beed4e9b3993 nodeName:}" failed. No retries permitted until 2026-03-13 14:24:06.713587441 +0000 UTC m=+1145.613375130 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/64d55221-60c5-4521-ac84-beed4e9b3993-cert") pod "infra-operator-controller-manager-54dc5b8f8d-lp7kj" (UID: "64d55221-60c5-4521-ac84-beed4e9b3993") : secret "infra-operator-webhook-server-cert" not found Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.214017 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-d47688694-tl8t4" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.229148 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-984cd4dcf-w497d" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.263723 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-66d56f6ff4-7cbb2" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.264500 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-57b484b4df-m8j7m"] Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.264930 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-64424\" (UniqueName: \"kubernetes.io/projected/d3cdd2ca-3e81-480f-876e-2b975d8bc9d8-kube-api-access-64424\") pod \"heat-operator-controller-manager-77b6666d85-874m2\" (UID: \"d3cdd2ca-3e81-480f-876e-2b975d8bc9d8\") " pod="openstack-operators/heat-operator-controller-manager-77b6666d85-874m2" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.265012 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4whsv\" (UniqueName: \"kubernetes.io/projected/64d55221-60c5-4521-ac84-beed4e9b3993-kube-api-access-4whsv\") pod \"infra-operator-controller-manager-54dc5b8f8d-lp7kj\" (UID: \"64d55221-60c5-4521-ac84-beed4e9b3993\") " pod="openstack-operators/infra-operator-controller-manager-54dc5b8f8d-lp7kj" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.265279 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dmv9x\" (UniqueName: \"kubernetes.io/projected/8ecaf4b8-808e-44b9-8ae7-6ef9519c4a66-kube-api-access-dmv9x\") pod \"horizon-operator-controller-manager-6d9d6b584d-r7bhq\" (UID: \"8ecaf4b8-808e-44b9-8ae7-6ef9519c4a66\") " pod="openstack-operators/horizon-operator-controller-manager-6d9d6b584d-r7bhq" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.265586 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-57b484b4df-m8j7m" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.268074 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-5b6b6b4c9f-5m2vw"] Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.268793 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-59tqw\" (UniqueName: \"kubernetes.io/projected/19c320fa-6619-45cd-a73e-65834173b786-kube-api-access-59tqw\") pod \"ironic-operator-controller-manager-5bc894d9b-w5829\" (UID: \"19c320fa-6619-45cd-a73e-65834173b786\") " pod="openstack-operators/ironic-operator-controller-manager-5bc894d9b-w5829" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.269110 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-5b6b6b4c9f-5m2vw" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.272934 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-5964f64c48-gll76" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.278160 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-x5qcj" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.280767 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-hj4pg" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.283243 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-776c5696bf-gvgdl"] Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.285710 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-776c5696bf-gvgdl" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.287126 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-fwlbf" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.304085 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-57b484b4df-m8j7m"] Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.311554 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-5b6b6b4c9f-5m2vw"] Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.314180 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b25c5\" (UniqueName: \"kubernetes.io/projected/29201cb7-c056-4e4e-b0da-cc775c7ef11f-kube-api-access-b25c5\") pod \"neutron-operator-controller-manager-776c5696bf-gvgdl\" (UID: \"29201cb7-c056-4e4e-b0da-cc775c7ef11f\") " pod="openstack-operators/neutron-operator-controller-manager-776c5696bf-gvgdl" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.314241 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n7mn9\" (UniqueName: \"kubernetes.io/projected/7d620a90-0e93-41dc-ab39-fc6ea8c461e7-kube-api-access-n7mn9\") pod \"keystone-operator-controller-manager-684f77d66d-7rdzd\" (UID: \"7d620a90-0e93-41dc-ab39-fc6ea8c461e7\") " pod="openstack-operators/keystone-operator-controller-manager-684f77d66d-7rdzd" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.314288 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nmthx\" (UniqueName: \"kubernetes.io/projected/e8134e91-4b9d-4e3c-9aec-eba18acabdce-kube-api-access-nmthx\") pod \"mariadb-operator-controller-manager-5b6b6b4c9f-5m2vw\" (UID: \"e8134e91-4b9d-4e3c-9aec-eba18acabdce\") " pod="openstack-operators/mariadb-operator-controller-manager-5b6b6b4c9f-5m2vw" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.314314 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7cq7h\" (UniqueName: \"kubernetes.io/projected/c55515f1-509e-4a4e-813d-fbc466fca489-kube-api-access-7cq7h\") pod \"manila-operator-controller-manager-57b484b4df-m8j7m\" (UID: \"c55515f1-509e-4a4e-813d-fbc466fca489\") " pod="openstack-operators/manila-operator-controller-manager-57b484b4df-m8j7m" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.317251 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-77b6666d85-874m2" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.323733 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-776c5696bf-gvgdl"] Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.337447 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-sk2l9"] Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.337844 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n7mn9\" (UniqueName: \"kubernetes.io/projected/7d620a90-0e93-41dc-ab39-fc6ea8c461e7-kube-api-access-n7mn9\") pod \"keystone-operator-controller-manager-684f77d66d-7rdzd\" (UID: \"7d620a90-0e93-41dc-ab39-fc6ea8c461e7\") " pod="openstack-operators/keystone-operator-controller-manager-684f77d66d-7rdzd" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.338255 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-sk2l9" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.340061 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-67nq5" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.343660 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-7f84474648-ql4kd"] Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.344510 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-7f84474648-ql4kd" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.345796 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-6d9d6b584d-r7bhq" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.346590 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-kcfs6" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.352332 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-sk2l9"] Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.376272 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-7f84474648-ql4kd"] Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.385930 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-5bc894d9b-w5829" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.394092 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-6f7958d774rr6kf"] Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.394990 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6f7958d774rr6kf" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.399200 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.399583 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-q8r6n" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.415513 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-684f77d66d-7rdzd" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.416579 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b25c5\" (UniqueName: \"kubernetes.io/projected/29201cb7-c056-4e4e-b0da-cc775c7ef11f-kube-api-access-b25c5\") pod \"neutron-operator-controller-manager-776c5696bf-gvgdl\" (UID: \"29201cb7-c056-4e4e-b0da-cc775c7ef11f\") " pod="openstack-operators/neutron-operator-controller-manager-776c5696bf-gvgdl" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.416613 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2gkcl\" (UniqueName: \"kubernetes.io/projected/4ab8f569-de56-4fea-9be0-07908063522a-kube-api-access-2gkcl\") pod \"nova-operator-controller-manager-7f84474648-ql4kd\" (UID: \"4ab8f569-de56-4fea-9be0-07908063522a\") " pod="openstack-operators/nova-operator-controller-manager-7f84474648-ql4kd" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.416641 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5xtmd\" (UniqueName: \"kubernetes.io/projected/f915558a-c0ab-4d8e-a427-e56a7588e382-kube-api-access-5xtmd\") pod \"openstack-baremetal-operator-controller-manager-6f7958d774rr6kf\" (UID: \"f915558a-c0ab-4d8e-a427-e56a7588e382\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6f7958d774rr6kf" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.416677 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f915558a-c0ab-4d8e-a427-e56a7588e382-cert\") pod \"openstack-baremetal-operator-controller-manager-6f7958d774rr6kf\" (UID: \"f915558a-c0ab-4d8e-a427-e56a7588e382\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6f7958d774rr6kf" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.416710 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nmthx\" (UniqueName: \"kubernetes.io/projected/e8134e91-4b9d-4e3c-9aec-eba18acabdce-kube-api-access-nmthx\") pod \"mariadb-operator-controller-manager-5b6b6b4c9f-5m2vw\" (UID: \"e8134e91-4b9d-4e3c-9aec-eba18acabdce\") " pod="openstack-operators/mariadb-operator-controller-manager-5b6b6b4c9f-5m2vw" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.416735 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7cq7h\" (UniqueName: \"kubernetes.io/projected/c55515f1-509e-4a4e-813d-fbc466fca489-kube-api-access-7cq7h\") pod \"manila-operator-controller-manager-57b484b4df-m8j7m\" (UID: \"c55515f1-509e-4a4e-813d-fbc466fca489\") " pod="openstack-operators/manila-operator-controller-manager-57b484b4df-m8j7m" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.416801 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2lwf7\" (UniqueName: \"kubernetes.io/projected/5f3d2dd8-6ee2-415a-8cc6-55f6e09cb473-kube-api-access-2lwf7\") pod \"octavia-operator-controller-manager-5f4f55cb5c-sk2l9\" (UID: \"5f3d2dd8-6ee2-415a-8cc6-55f6e09cb473\") " pod="openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-sk2l9" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.428573 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-6f7958d774rr6kf"] Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.449773 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b25c5\" (UniqueName: \"kubernetes.io/projected/29201cb7-c056-4e4e-b0da-cc775c7ef11f-kube-api-access-b25c5\") pod \"neutron-operator-controller-manager-776c5696bf-gvgdl\" (UID: \"29201cb7-c056-4e4e-b0da-cc775c7ef11f\") " pod="openstack-operators/neutron-operator-controller-manager-776c5696bf-gvgdl" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.449828 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nmthx\" (UniqueName: \"kubernetes.io/projected/e8134e91-4b9d-4e3c-9aec-eba18acabdce-kube-api-access-nmthx\") pod \"mariadb-operator-controller-manager-5b6b6b4c9f-5m2vw\" (UID: \"e8134e91-4b9d-4e3c-9aec-eba18acabdce\") " pod="openstack-operators/mariadb-operator-controller-manager-5b6b6b4c9f-5m2vw" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.456265 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-bbc5b68f9-cqmhw"] Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.457101 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-bbc5b68f9-cqmhw" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.460440 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-znvp6" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.462824 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7cq7h\" (UniqueName: \"kubernetes.io/projected/c55515f1-509e-4a4e-813d-fbc466fca489-kube-api-access-7cq7h\") pod \"manila-operator-controller-manager-57b484b4df-m8j7m\" (UID: \"c55515f1-509e-4a4e-813d-fbc466fca489\") " pod="openstack-operators/manila-operator-controller-manager-57b484b4df-m8j7m" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.485406 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-574d45c66c-b99f7"] Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.486403 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-574d45c66c-b99f7" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.488712 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-jwrkw" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.496757 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-574d45c66c-b99f7"] Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.520033 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2lwf7\" (UniqueName: \"kubernetes.io/projected/5f3d2dd8-6ee2-415a-8cc6-55f6e09cb473-kube-api-access-2lwf7\") pod \"octavia-operator-controller-manager-5f4f55cb5c-sk2l9\" (UID: \"5f3d2dd8-6ee2-415a-8cc6-55f6e09cb473\") " pod="openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-sk2l9" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.520287 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2gkcl\" (UniqueName: \"kubernetes.io/projected/4ab8f569-de56-4fea-9be0-07908063522a-kube-api-access-2gkcl\") pod \"nova-operator-controller-manager-7f84474648-ql4kd\" (UID: \"4ab8f569-de56-4fea-9be0-07908063522a\") " pod="openstack-operators/nova-operator-controller-manager-7f84474648-ql4kd" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.520318 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5xtmd\" (UniqueName: \"kubernetes.io/projected/f915558a-c0ab-4d8e-a427-e56a7588e382-kube-api-access-5xtmd\") pod \"openstack-baremetal-operator-controller-manager-6f7958d774rr6kf\" (UID: \"f915558a-c0ab-4d8e-a427-e56a7588e382\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6f7958d774rr6kf" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.520353 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f915558a-c0ab-4d8e-a427-e56a7588e382-cert\") pod \"openstack-baremetal-operator-controller-manager-6f7958d774rr6kf\" (UID: \"f915558a-c0ab-4d8e-a427-e56a7588e382\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6f7958d774rr6kf" Mar 13 14:24:06 crc kubenswrapper[4907]: E0313 14:24:06.520523 4907 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 13 14:24:06 crc kubenswrapper[4907]: E0313 14:24:06.520574 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f915558a-c0ab-4d8e-a427-e56a7588e382-cert podName:f915558a-c0ab-4d8e-a427-e56a7588e382 nodeName:}" failed. No retries permitted until 2026-03-13 14:24:07.020556034 +0000 UTC m=+1145.920343733 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/f915558a-c0ab-4d8e-a427-e56a7588e382-cert") pod "openstack-baremetal-operator-controller-manager-6f7958d774rr6kf" (UID: "f915558a-c0ab-4d8e-a427-e56a7588e382") : secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.536945 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-bbc5b68f9-cqmhw"] Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.547438 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5xtmd\" (UniqueName: \"kubernetes.io/projected/f915558a-c0ab-4d8e-a427-e56a7588e382-kube-api-access-5xtmd\") pod \"openstack-baremetal-operator-controller-manager-6f7958d774rr6kf\" (UID: \"f915558a-c0ab-4d8e-a427-e56a7588e382\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6f7958d774rr6kf" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.552939 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-7f9cc5dd44-sd6nr"] Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.560928 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-7f9cc5dd44-sd6nr" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.554485 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2lwf7\" (UniqueName: \"kubernetes.io/projected/5f3d2dd8-6ee2-415a-8cc6-55f6e09cb473-kube-api-access-2lwf7\") pod \"octavia-operator-controller-manager-5f4f55cb5c-sk2l9\" (UID: \"5f3d2dd8-6ee2-415a-8cc6-55f6e09cb473\") " pod="openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-sk2l9" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.562675 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2gkcl\" (UniqueName: \"kubernetes.io/projected/4ab8f569-de56-4fea-9be0-07908063522a-kube-api-access-2gkcl\") pod \"nova-operator-controller-manager-7f84474648-ql4kd\" (UID: \"4ab8f569-de56-4fea-9be0-07908063522a\") " pod="openstack-operators/nova-operator-controller-manager-7f84474648-ql4kd" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.572060 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-6kktj" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.586944 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-7f9cc5dd44-sd6nr"] Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.592542 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-57b484b4df-m8j7m" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.610421 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-6854b8b9d9-dj95f"] Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.611182 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-6854b8b9d9-dj95f" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.622426 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mzlzx\" (UniqueName: \"kubernetes.io/projected/fdddbc99-1a71-4250-ab2a-6f426e744423-kube-api-access-mzlzx\") pod \"ovn-operator-controller-manager-bbc5b68f9-cqmhw\" (UID: \"fdddbc99-1a71-4250-ab2a-6f426e744423\") " pod="openstack-operators/ovn-operator-controller-manager-bbc5b68f9-cqmhw" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.622528 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d5trp\" (UniqueName: \"kubernetes.io/projected/0b3db736-b3d4-494f-9755-9103150d267f-kube-api-access-d5trp\") pod \"placement-operator-controller-manager-574d45c66c-b99f7\" (UID: \"0b3db736-b3d4-494f-9755-9103150d267f\") " pod="openstack-operators/placement-operator-controller-manager-574d45c66c-b99f7" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.629873 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-5b6b6b4c9f-5m2vw" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.631778 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-pcq8d" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.636259 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-776c5696bf-gvgdl" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.665738 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-6854b8b9d9-dj95f"] Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.665901 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-sk2l9" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.679274 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-7f84474648-ql4kd" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.689933 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-5c5cb9c4d7-vzpcd"] Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.690781 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-vzpcd" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.691751 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5c5cb9c4d7-vzpcd"] Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.710588 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-bjcb6" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.717818 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-hjp4p"] Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.718798 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-hjp4p" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.725099 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-sp8zt" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.730591 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mwz7r\" (UniqueName: \"kubernetes.io/projected/d139e1bb-8ae8-4760-a1b6-3f7d96f3e9a1-kube-api-access-mwz7r\") pod \"telemetry-operator-controller-manager-6854b8b9d9-dj95f\" (UID: \"d139e1bb-8ae8-4760-a1b6-3f7d96f3e9a1\") " pod="openstack-operators/telemetry-operator-controller-manager-6854b8b9d9-dj95f" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.730631 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hpzf7\" (UniqueName: \"kubernetes.io/projected/92718823-f344-4d7b-beff-e75f1574f1e2-kube-api-access-hpzf7\") pod \"swift-operator-controller-manager-7f9cc5dd44-sd6nr\" (UID: \"92718823-f344-4d7b-beff-e75f1574f1e2\") " pod="openstack-operators/swift-operator-controller-manager-7f9cc5dd44-sd6nr" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.730657 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/64d55221-60c5-4521-ac84-beed4e9b3993-cert\") pod \"infra-operator-controller-manager-54dc5b8f8d-lp7kj\" (UID: \"64d55221-60c5-4521-ac84-beed4e9b3993\") " pod="openstack-operators/infra-operator-controller-manager-54dc5b8f8d-lp7kj" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.730700 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d5trp\" (UniqueName: \"kubernetes.io/projected/0b3db736-b3d4-494f-9755-9103150d267f-kube-api-access-d5trp\") pod \"placement-operator-controller-manager-574d45c66c-b99f7\" (UID: \"0b3db736-b3d4-494f-9755-9103150d267f\") " pod="openstack-operators/placement-operator-controller-manager-574d45c66c-b99f7" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.730747 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mzlzx\" (UniqueName: \"kubernetes.io/projected/fdddbc99-1a71-4250-ab2a-6f426e744423-kube-api-access-mzlzx\") pod \"ovn-operator-controller-manager-bbc5b68f9-cqmhw\" (UID: \"fdddbc99-1a71-4250-ab2a-6f426e744423\") " pod="openstack-operators/ovn-operator-controller-manager-bbc5b68f9-cqmhw" Mar 13 14:24:06 crc kubenswrapper[4907]: E0313 14:24:06.731355 4907 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Mar 13 14:24:06 crc kubenswrapper[4907]: E0313 14:24:06.731778 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/64d55221-60c5-4521-ac84-beed4e9b3993-cert podName:64d55221-60c5-4521-ac84-beed4e9b3993 nodeName:}" failed. No retries permitted until 2026-03-13 14:24:07.731756759 +0000 UTC m=+1146.631544448 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/64d55221-60c5-4521-ac84-beed4e9b3993-cert") pod "infra-operator-controller-manager-54dc5b8f8d-lp7kj" (UID: "64d55221-60c5-4521-ac84-beed4e9b3993") : secret "infra-operator-webhook-server-cert" not found Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.743386 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-hjp4p"] Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.791646 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d5trp\" (UniqueName: \"kubernetes.io/projected/0b3db736-b3d4-494f-9755-9103150d267f-kube-api-access-d5trp\") pod \"placement-operator-controller-manager-574d45c66c-b99f7\" (UID: \"0b3db736-b3d4-494f-9755-9103150d267f\") " pod="openstack-operators/placement-operator-controller-manager-574d45c66c-b99f7" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.793420 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mzlzx\" (UniqueName: \"kubernetes.io/projected/fdddbc99-1a71-4250-ab2a-6f426e744423-kube-api-access-mzlzx\") pod \"ovn-operator-controller-manager-bbc5b68f9-cqmhw\" (UID: \"fdddbc99-1a71-4250-ab2a-6f426e744423\") " pod="openstack-operators/ovn-operator-controller-manager-bbc5b68f9-cqmhw" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.834782 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hpzf7\" (UniqueName: \"kubernetes.io/projected/92718823-f344-4d7b-beff-e75f1574f1e2-kube-api-access-hpzf7\") pod \"swift-operator-controller-manager-7f9cc5dd44-sd6nr\" (UID: \"92718823-f344-4d7b-beff-e75f1574f1e2\") " pod="openstack-operators/swift-operator-controller-manager-7f9cc5dd44-sd6nr" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.834861 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5h7lx\" (UniqueName: \"kubernetes.io/projected/a89fbf9b-b391-426b-9f9d-91cdba71602b-kube-api-access-5h7lx\") pod \"watcher-operator-controller-manager-6c4d75f7f9-hjp4p\" (UID: \"a89fbf9b-b391-426b-9f9d-91cdba71602b\") " pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-hjp4p" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.834925 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b4b47\" (UniqueName: \"kubernetes.io/projected/a2b0825f-f2f1-4b3f-b7b0-5a476f99b4fe-kube-api-access-b4b47\") pod \"test-operator-controller-manager-5c5cb9c4d7-vzpcd\" (UID: \"a2b0825f-f2f1-4b3f-b7b0-5a476f99b4fe\") " pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-vzpcd" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.835020 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mwz7r\" (UniqueName: \"kubernetes.io/projected/d139e1bb-8ae8-4760-a1b6-3f7d96f3e9a1-kube-api-access-mwz7r\") pod \"telemetry-operator-controller-manager-6854b8b9d9-dj95f\" (UID: \"d139e1bb-8ae8-4760-a1b6-3f7d96f3e9a1\") " pod="openstack-operators/telemetry-operator-controller-manager-6854b8b9d9-dj95f" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.836318 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-bbc5b68f9-cqmhw" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.866607 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-574d45c66c-b99f7" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.878211 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hpzf7\" (UniqueName: \"kubernetes.io/projected/92718823-f344-4d7b-beff-e75f1574f1e2-kube-api-access-hpzf7\") pod \"swift-operator-controller-manager-7f9cc5dd44-sd6nr\" (UID: \"92718823-f344-4d7b-beff-e75f1574f1e2\") " pod="openstack-operators/swift-operator-controller-manager-7f9cc5dd44-sd6nr" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.879064 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mwz7r\" (UniqueName: \"kubernetes.io/projected/d139e1bb-8ae8-4760-a1b6-3f7d96f3e9a1-kube-api-access-mwz7r\") pod \"telemetry-operator-controller-manager-6854b8b9d9-dj95f\" (UID: \"d139e1bb-8ae8-4760-a1b6-3f7d96f3e9a1\") " pod="openstack-operators/telemetry-operator-controller-manager-6854b8b9d9-dj95f" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.904613 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-6484b7b757-qldxp"] Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.905780 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-6484b7b757-qldxp" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.910085 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.910290 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.910438 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-xm9nq" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.910791 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-7f9cc5dd44-sd6nr" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.922388 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-6484b7b757-qldxp"] Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.927098 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-d47688694-tl8t4"] Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.938328 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-xtvzb"] Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.939232 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5h7lx\" (UniqueName: \"kubernetes.io/projected/a89fbf9b-b391-426b-9f9d-91cdba71602b-kube-api-access-5h7lx\") pod \"watcher-operator-controller-manager-6c4d75f7f9-hjp4p\" (UID: \"a89fbf9b-b391-426b-9f9d-91cdba71602b\") " pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-hjp4p" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.947196 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b4b47\" (UniqueName: \"kubernetes.io/projected/a2b0825f-f2f1-4b3f-b7b0-5a476f99b4fe-kube-api-access-b4b47\") pod \"test-operator-controller-manager-5c5cb9c4d7-vzpcd\" (UID: \"a2b0825f-f2f1-4b3f-b7b0-5a476f99b4fe\") " pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-vzpcd" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.950314 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-xtvzb"] Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.950682 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-xtvzb" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.964767 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-nn9ml" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.966563 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-6854b8b9d9-dj95f" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.967330 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b4b47\" (UniqueName: \"kubernetes.io/projected/a2b0825f-f2f1-4b3f-b7b0-5a476f99b4fe-kube-api-access-b4b47\") pod \"test-operator-controller-manager-5c5cb9c4d7-vzpcd\" (UID: \"a2b0825f-f2f1-4b3f-b7b0-5a476f99b4fe\") " pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-vzpcd" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.991420 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5h7lx\" (UniqueName: \"kubernetes.io/projected/a89fbf9b-b391-426b-9f9d-91cdba71602b-kube-api-access-5h7lx\") pod \"watcher-operator-controller-manager-6c4d75f7f9-hjp4p\" (UID: \"a89fbf9b-b391-426b-9f9d-91cdba71602b\") " pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-hjp4p" Mar 13 14:24:07 crc kubenswrapper[4907]: I0313 14:24:07.047636 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kp585\" (UniqueName: \"kubernetes.io/projected/44ecb24b-4c5c-4cf6-96e9-4d8b52469d3d-kube-api-access-kp585\") pod \"rabbitmq-cluster-operator-manager-668c99d594-xtvzb\" (UID: \"44ecb24b-4c5c-4cf6-96e9-4d8b52469d3d\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-xtvzb" Mar 13 14:24:07 crc kubenswrapper[4907]: I0313 14:24:07.047684 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f915558a-c0ab-4d8e-a427-e56a7588e382-cert\") pod \"openstack-baremetal-operator-controller-manager-6f7958d774rr6kf\" (UID: \"f915558a-c0ab-4d8e-a427-e56a7588e382\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6f7958d774rr6kf" Mar 13 14:24:07 crc kubenswrapper[4907]: I0313 14:24:07.047757 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/9ac47edf-03ab-4e93-b238-2aa00765ab06-webhook-certs\") pod \"openstack-operator-controller-manager-6484b7b757-qldxp\" (UID: \"9ac47edf-03ab-4e93-b238-2aa00765ab06\") " pod="openstack-operators/openstack-operator-controller-manager-6484b7b757-qldxp" Mar 13 14:24:07 crc kubenswrapper[4907]: I0313 14:24:07.047780 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9ac47edf-03ab-4e93-b238-2aa00765ab06-metrics-certs\") pod \"openstack-operator-controller-manager-6484b7b757-qldxp\" (UID: \"9ac47edf-03ab-4e93-b238-2aa00765ab06\") " pod="openstack-operators/openstack-operator-controller-manager-6484b7b757-qldxp" Mar 13 14:24:07 crc kubenswrapper[4907]: I0313 14:24:07.047804 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bfzck\" (UniqueName: \"kubernetes.io/projected/9ac47edf-03ab-4e93-b238-2aa00765ab06-kube-api-access-bfzck\") pod \"openstack-operator-controller-manager-6484b7b757-qldxp\" (UID: \"9ac47edf-03ab-4e93-b238-2aa00765ab06\") " pod="openstack-operators/openstack-operator-controller-manager-6484b7b757-qldxp" Mar 13 14:24:07 crc kubenswrapper[4907]: E0313 14:24:07.047961 4907 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 13 14:24:07 crc kubenswrapper[4907]: E0313 14:24:07.048014 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f915558a-c0ab-4d8e-a427-e56a7588e382-cert podName:f915558a-c0ab-4d8e-a427-e56a7588e382 nodeName:}" failed. No retries permitted until 2026-03-13 14:24:08.04799598 +0000 UTC m=+1146.947783669 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/f915558a-c0ab-4d8e-a427-e56a7588e382-cert") pod "openstack-baremetal-operator-controller-manager-6f7958d774rr6kf" (UID: "f915558a-c0ab-4d8e-a427-e56a7588e382") : secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 13 14:24:07 crc kubenswrapper[4907]: I0313 14:24:07.086912 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-vzpcd" Mar 13 14:24:07 crc kubenswrapper[4907]: I0313 14:24:07.095848 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-hjp4p" Mar 13 14:24:07 crc kubenswrapper[4907]: I0313 14:24:07.109605 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-d47688694-tl8t4" event={"ID":"863c77e9-4022-4ab9-8ac7-a4006d22813c","Type":"ContainerStarted","Data":"bef85b99ba0d7433d5ec268d18ee5ad44e31169909fdb11b1722a7603775c913"} Mar 13 14:24:07 crc kubenswrapper[4907]: I0313 14:24:07.134669 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-66d56f6ff4-7cbb2"] Mar 13 14:24:07 crc kubenswrapper[4907]: I0313 14:24:07.148736 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/9ac47edf-03ab-4e93-b238-2aa00765ab06-webhook-certs\") pod \"openstack-operator-controller-manager-6484b7b757-qldxp\" (UID: \"9ac47edf-03ab-4e93-b238-2aa00765ab06\") " pod="openstack-operators/openstack-operator-controller-manager-6484b7b757-qldxp" Mar 13 14:24:07 crc kubenswrapper[4907]: I0313 14:24:07.148780 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9ac47edf-03ab-4e93-b238-2aa00765ab06-metrics-certs\") pod \"openstack-operator-controller-manager-6484b7b757-qldxp\" (UID: \"9ac47edf-03ab-4e93-b238-2aa00765ab06\") " pod="openstack-operators/openstack-operator-controller-manager-6484b7b757-qldxp" Mar 13 14:24:07 crc kubenswrapper[4907]: I0313 14:24:07.148812 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bfzck\" (UniqueName: \"kubernetes.io/projected/9ac47edf-03ab-4e93-b238-2aa00765ab06-kube-api-access-bfzck\") pod \"openstack-operator-controller-manager-6484b7b757-qldxp\" (UID: \"9ac47edf-03ab-4e93-b238-2aa00765ab06\") " pod="openstack-operators/openstack-operator-controller-manager-6484b7b757-qldxp" Mar 13 14:24:07 crc kubenswrapper[4907]: I0313 14:24:07.148841 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kp585\" (UniqueName: \"kubernetes.io/projected/44ecb24b-4c5c-4cf6-96e9-4d8b52469d3d-kube-api-access-kp585\") pod \"rabbitmq-cluster-operator-manager-668c99d594-xtvzb\" (UID: \"44ecb24b-4c5c-4cf6-96e9-4d8b52469d3d\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-xtvzb" Mar 13 14:24:07 crc kubenswrapper[4907]: E0313 14:24:07.150605 4907 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Mar 13 14:24:07 crc kubenswrapper[4907]: E0313 14:24:07.150678 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9ac47edf-03ab-4e93-b238-2aa00765ab06-webhook-certs podName:9ac47edf-03ab-4e93-b238-2aa00765ab06 nodeName:}" failed. No retries permitted until 2026-03-13 14:24:07.65066066 +0000 UTC m=+1146.550448349 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/9ac47edf-03ab-4e93-b238-2aa00765ab06-webhook-certs") pod "openstack-operator-controller-manager-6484b7b757-qldxp" (UID: "9ac47edf-03ab-4e93-b238-2aa00765ab06") : secret "webhook-server-cert" not found Mar 13 14:24:07 crc kubenswrapper[4907]: E0313 14:24:07.151534 4907 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Mar 13 14:24:07 crc kubenswrapper[4907]: E0313 14:24:07.151567 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9ac47edf-03ab-4e93-b238-2aa00765ab06-metrics-certs podName:9ac47edf-03ab-4e93-b238-2aa00765ab06 nodeName:}" failed. No retries permitted until 2026-03-13 14:24:07.651557886 +0000 UTC m=+1146.551345575 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/9ac47edf-03ab-4e93-b238-2aa00765ab06-metrics-certs") pod "openstack-operator-controller-manager-6484b7b757-qldxp" (UID: "9ac47edf-03ab-4e93-b238-2aa00765ab06") : secret "metrics-server-cert" not found Mar 13 14:24:07 crc kubenswrapper[4907]: I0313 14:24:07.181365 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bfzck\" (UniqueName: \"kubernetes.io/projected/9ac47edf-03ab-4e93-b238-2aa00765ab06-kube-api-access-bfzck\") pod \"openstack-operator-controller-manager-6484b7b757-qldxp\" (UID: \"9ac47edf-03ab-4e93-b238-2aa00765ab06\") " pod="openstack-operators/openstack-operator-controller-manager-6484b7b757-qldxp" Mar 13 14:24:07 crc kubenswrapper[4907]: I0313 14:24:07.186525 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kp585\" (UniqueName: \"kubernetes.io/projected/44ecb24b-4c5c-4cf6-96e9-4d8b52469d3d-kube-api-access-kp585\") pod \"rabbitmq-cluster-operator-manager-668c99d594-xtvzb\" (UID: \"44ecb24b-4c5c-4cf6-96e9-4d8b52469d3d\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-xtvzb" Mar 13 14:24:07 crc kubenswrapper[4907]: I0313 14:24:07.298961 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-xtvzb" Mar 13 14:24:07 crc kubenswrapper[4907]: I0313 14:24:07.376390 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-5964f64c48-gll76"] Mar 13 14:24:07 crc kubenswrapper[4907]: I0313 14:24:07.390062 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-984cd4dcf-w497d"] Mar 13 14:24:07 crc kubenswrapper[4907]: I0313 14:24:07.656953 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-5b6b6b4c9f-5m2vw"] Mar 13 14:24:07 crc kubenswrapper[4907]: I0313 14:24:07.661717 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-684f77d66d-7rdzd"] Mar 13 14:24:07 crc kubenswrapper[4907]: I0313 14:24:07.664599 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/9ac47edf-03ab-4e93-b238-2aa00765ab06-webhook-certs\") pod \"openstack-operator-controller-manager-6484b7b757-qldxp\" (UID: \"9ac47edf-03ab-4e93-b238-2aa00765ab06\") " pod="openstack-operators/openstack-operator-controller-manager-6484b7b757-qldxp" Mar 13 14:24:07 crc kubenswrapper[4907]: I0313 14:24:07.664645 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9ac47edf-03ab-4e93-b238-2aa00765ab06-metrics-certs\") pod \"openstack-operator-controller-manager-6484b7b757-qldxp\" (UID: \"9ac47edf-03ab-4e93-b238-2aa00765ab06\") " pod="openstack-operators/openstack-operator-controller-manager-6484b7b757-qldxp" Mar 13 14:24:07 crc kubenswrapper[4907]: E0313 14:24:07.664892 4907 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Mar 13 14:24:07 crc kubenswrapper[4907]: E0313 14:24:07.664916 4907 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Mar 13 14:24:07 crc kubenswrapper[4907]: E0313 14:24:07.664954 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9ac47edf-03ab-4e93-b238-2aa00765ab06-webhook-certs podName:9ac47edf-03ab-4e93-b238-2aa00765ab06 nodeName:}" failed. No retries permitted until 2026-03-13 14:24:08.66493421 +0000 UTC m=+1147.564721959 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/9ac47edf-03ab-4e93-b238-2aa00765ab06-webhook-certs") pod "openstack-operator-controller-manager-6484b7b757-qldxp" (UID: "9ac47edf-03ab-4e93-b238-2aa00765ab06") : secret "webhook-server-cert" not found Mar 13 14:24:07 crc kubenswrapper[4907]: E0313 14:24:07.664975 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9ac47edf-03ab-4e93-b238-2aa00765ab06-metrics-certs podName:9ac47edf-03ab-4e93-b238-2aa00765ab06 nodeName:}" failed. No retries permitted until 2026-03-13 14:24:08.664965941 +0000 UTC m=+1147.564753740 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/9ac47edf-03ab-4e93-b238-2aa00765ab06-metrics-certs") pod "openstack-operator-controller-manager-6484b7b757-qldxp" (UID: "9ac47edf-03ab-4e93-b238-2aa00765ab06") : secret "metrics-server-cert" not found Mar 13 14:24:07 crc kubenswrapper[4907]: I0313 14:24:07.672805 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-5bc894d9b-w5829"] Mar 13 14:24:07 crc kubenswrapper[4907]: W0313 14:24:07.689591 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4ab8f569_de56_4fea_9be0_07908063522a.slice/crio-2bed19672f429f39772b19fc86747e4dbbce4b48f9ab5aac93c9a3800c55fd88 WatchSource:0}: Error finding container 2bed19672f429f39772b19fc86747e4dbbce4b48f9ab5aac93c9a3800c55fd88: Status 404 returned error can't find the container with id 2bed19672f429f39772b19fc86747e4dbbce4b48f9ab5aac93c9a3800c55fd88 Mar 13 14:24:07 crc kubenswrapper[4907]: I0313 14:24:07.689661 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-7f84474648-ql4kd"] Mar 13 14:24:07 crc kubenswrapper[4907]: I0313 14:24:07.697443 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-sk2l9"] Mar 13 14:24:07 crc kubenswrapper[4907]: I0313 14:24:07.702846 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-77b6666d85-874m2"] Mar 13 14:24:07 crc kubenswrapper[4907]: I0313 14:24:07.708430 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-6d9d6b584d-r7bhq"] Mar 13 14:24:07 crc kubenswrapper[4907]: W0313 14:24:07.726930 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod19c320fa_6619_45cd_a73e_65834173b786.slice/crio-d18a8174c620b5a6ee3028caa18c2a577a0e97ec87adc0db628b0d1c7e9036a8 WatchSource:0}: Error finding container d18a8174c620b5a6ee3028caa18c2a577a0e97ec87adc0db628b0d1c7e9036a8: Status 404 returned error can't find the container with id d18a8174c620b5a6ee3028caa18c2a577a0e97ec87adc0db628b0d1c7e9036a8 Mar 13 14:24:07 crc kubenswrapper[4907]: W0313 14:24:07.727857 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8ecaf4b8_808e_44b9_8ae7_6ef9519c4a66.slice/crio-d023d3167efff6988171f93e3761488c1ce0729bd3cb6e8aa0a553e1a18fe33e WatchSource:0}: Error finding container d023d3167efff6988171f93e3761488c1ce0729bd3cb6e8aa0a553e1a18fe33e: Status 404 returned error can't find the container with id d023d3167efff6988171f93e3761488c1ce0729bd3cb6e8aa0a553e1a18fe33e Mar 13 14:24:07 crc kubenswrapper[4907]: I0313 14:24:07.767834 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/64d55221-60c5-4521-ac84-beed4e9b3993-cert\") pod \"infra-operator-controller-manager-54dc5b8f8d-lp7kj\" (UID: \"64d55221-60c5-4521-ac84-beed4e9b3993\") " pod="openstack-operators/infra-operator-controller-manager-54dc5b8f8d-lp7kj" Mar 13 14:24:07 crc kubenswrapper[4907]: E0313 14:24:07.768989 4907 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Mar 13 14:24:07 crc kubenswrapper[4907]: E0313 14:24:07.769053 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/64d55221-60c5-4521-ac84-beed4e9b3993-cert podName:64d55221-60c5-4521-ac84-beed4e9b3993 nodeName:}" failed. No retries permitted until 2026-03-13 14:24:09.76903617 +0000 UTC m=+1148.668823859 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/64d55221-60c5-4521-ac84-beed4e9b3993-cert") pod "infra-operator-controller-manager-54dc5b8f8d-lp7kj" (UID: "64d55221-60c5-4521-ac84-beed4e9b3993") : secret "infra-operator-webhook-server-cert" not found Mar 13 14:24:07 crc kubenswrapper[4907]: I0313 14:24:07.822495 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-6854b8b9d9-dj95f"] Mar 13 14:24:07 crc kubenswrapper[4907]: I0313 14:24:07.829546 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-57b484b4df-m8j7m"] Mar 13 14:24:07 crc kubenswrapper[4907]: I0313 14:24:07.847036 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-574d45c66c-b99f7"] Mar 13 14:24:07 crc kubenswrapper[4907]: I0313 14:24:07.851610 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-776c5696bf-gvgdl"] Mar 13 14:24:07 crc kubenswrapper[4907]: W0313 14:24:07.858133 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc55515f1_509e_4a4e_813d_fbc466fca489.slice/crio-f3fff2cc89f13cc8f5f42b57d76dd61682015600446bfbf3151c19fd65c90e59 WatchSource:0}: Error finding container f3fff2cc89f13cc8f5f42b57d76dd61682015600446bfbf3151c19fd65c90e59: Status 404 returned error can't find the container with id f3fff2cc89f13cc8f5f42b57d76dd61682015600446bfbf3151c19fd65c90e59 Mar 13 14:24:07 crc kubenswrapper[4907]: E0313 14:24:07.868360 4907 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:e7e865363955c670e41b6c042c4f87abceff78f5495ba5c5c82988baad45c978,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-d5trp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-574d45c66c-b99f7_openstack-operators(0b3db736-b3d4-494f-9755-9103150d267f): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Mar 13 14:24:07 crc kubenswrapper[4907]: E0313 14:24:07.869797 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/placement-operator-controller-manager-574d45c66c-b99f7" podUID="0b3db736-b3d4-494f-9755-9103150d267f" Mar 13 14:24:08 crc kubenswrapper[4907]: I0313 14:24:08.049381 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5c5cb9c4d7-vzpcd"] Mar 13 14:24:08 crc kubenswrapper[4907]: I0313 14:24:08.071983 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-7f9cc5dd44-sd6nr"] Mar 13 14:24:08 crc kubenswrapper[4907]: I0313 14:24:08.085287 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-hjp4p"] Mar 13 14:24:08 crc kubenswrapper[4907]: I0313 14:24:08.086897 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f915558a-c0ab-4d8e-a427-e56a7588e382-cert\") pod \"openstack-baremetal-operator-controller-manager-6f7958d774rr6kf\" (UID: \"f915558a-c0ab-4d8e-a427-e56a7588e382\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6f7958d774rr6kf" Mar 13 14:24:08 crc kubenswrapper[4907]: E0313 14:24:08.087187 4907 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 13 14:24:08 crc kubenswrapper[4907]: E0313 14:24:08.087258 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f915558a-c0ab-4d8e-a427-e56a7588e382-cert podName:f915558a-c0ab-4d8e-a427-e56a7588e382 nodeName:}" failed. No retries permitted until 2026-03-13 14:24:10.087234075 +0000 UTC m=+1148.987021774 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/f915558a-c0ab-4d8e-a427-e56a7588e382-cert") pod "openstack-baremetal-operator-controller-manager-6f7958d774rr6kf" (UID: "f915558a-c0ab-4d8e-a427-e56a7588e382") : secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 13 14:24:08 crc kubenswrapper[4907]: E0313 14:24:08.101371 4907 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:72db77c98e7bca64d469b4dc316e9c8d329681f825d19ef8f333437fb1c6d3f5,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-hpzf7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-7f9cc5dd44-sd6nr_openstack-operators(92718823-f344-4d7b-beff-e75f1574f1e2): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Mar 13 14:24:08 crc kubenswrapper[4907]: E0313 14:24:08.102722 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/swift-operator-controller-manager-7f9cc5dd44-sd6nr" podUID="92718823-f344-4d7b-beff-e75f1574f1e2" Mar 13 14:24:08 crc kubenswrapper[4907]: E0313 14:24:08.103653 4907 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:d9c55e8c6304a0e32289b5e8c69a87ea59b9968918a5c85b7c384633df82c807,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-5h7lx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-6c4d75f7f9-hjp4p_openstack-operators(a89fbf9b-b391-426b-9f9d-91cdba71602b): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Mar 13 14:24:08 crc kubenswrapper[4907]: W0313 14:24:08.103993 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfdddbc99_1a71_4250_ab2a_6f426e744423.slice/crio-23437d6b081e76de7b1c3ba9d56430532e367089e5a7f5e6435afb9e118f4503 WatchSource:0}: Error finding container 23437d6b081e76de7b1c3ba9d56430532e367089e5a7f5e6435afb9e118f4503: Status 404 returned error can't find the container with id 23437d6b081e76de7b1c3ba9d56430532e367089e5a7f5e6435afb9e118f4503 Mar 13 14:24:08 crc kubenswrapper[4907]: E0313 14:24:08.105206 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-hjp4p" podUID="a89fbf9b-b391-426b-9f9d-91cdba71602b" Mar 13 14:24:08 crc kubenswrapper[4907]: W0313 14:24:08.105845 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod44ecb24b_4c5c_4cf6_96e9_4d8b52469d3d.slice/crio-a8124cbf931bb327fe1039e05d89d2da5fa96c7a2fbfa266e53a0d33a1439c5d WatchSource:0}: Error finding container a8124cbf931bb327fe1039e05d89d2da5fa96c7a2fbfa266e53a0d33a1439c5d: Status 404 returned error can't find the container with id a8124cbf931bb327fe1039e05d89d2da5fa96c7a2fbfa266e53a0d33a1439c5d Mar 13 14:24:08 crc kubenswrapper[4907]: E0313 14:24:08.106337 4907 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ovn-operator@sha256:2f63ddf5c95c6c82f6e04bc9f7f20d56dc003614647726ab00276239eec40b7f,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-mzlzx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-bbc5b68f9-cqmhw_openstack-operators(fdddbc99-1a71-4250-ab2a-6f426e744423): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Mar 13 14:24:08 crc kubenswrapper[4907]: E0313 14:24:08.107474 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/ovn-operator-controller-manager-bbc5b68f9-cqmhw" podUID="fdddbc99-1a71-4250-ab2a-6f426e744423" Mar 13 14:24:08 crc kubenswrapper[4907]: E0313 14:24:08.107635 4907 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-kp585,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-xtvzb_openstack-operators(44ecb24b-4c5c-4cf6-96e9-4d8b52469d3d): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Mar 13 14:24:08 crc kubenswrapper[4907]: I0313 14:24:08.108273 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-bbc5b68f9-cqmhw"] Mar 13 14:24:08 crc kubenswrapper[4907]: E0313 14:24:08.110311 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-xtvzb" podUID="44ecb24b-4c5c-4cf6-96e9-4d8b52469d3d" Mar 13 14:24:08 crc kubenswrapper[4907]: I0313 14:24:08.132493 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-xtvzb" event={"ID":"44ecb24b-4c5c-4cf6-96e9-4d8b52469d3d","Type":"ContainerStarted","Data":"a8124cbf931bb327fe1039e05d89d2da5fa96c7a2fbfa266e53a0d33a1439c5d"} Mar 13 14:24:08 crc kubenswrapper[4907]: I0313 14:24:08.135035 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-xtvzb"] Mar 13 14:24:08 crc kubenswrapper[4907]: E0313 14:24:08.135205 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-xtvzb" podUID="44ecb24b-4c5c-4cf6-96e9-4d8b52469d3d" Mar 13 14:24:08 crc kubenswrapper[4907]: I0313 14:24:08.136043 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-984cd4dcf-w497d" event={"ID":"77b1eb71-a455-4ec3-8e17-a847575cf0b0","Type":"ContainerStarted","Data":"8133a4150e9ff6586b24c5809c839ec3151e63009c4e9d4f463d2426457e2024"} Mar 13 14:24:08 crc kubenswrapper[4907]: I0313 14:24:08.140267 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-6d9d6b584d-r7bhq" event={"ID":"8ecaf4b8-808e-44b9-8ae7-6ef9519c4a66","Type":"ContainerStarted","Data":"d023d3167efff6988171f93e3761488c1ce0729bd3cb6e8aa0a553e1a18fe33e"} Mar 13 14:24:08 crc kubenswrapper[4907]: I0313 14:24:08.142790 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-574d45c66c-b99f7" event={"ID":"0b3db736-b3d4-494f-9755-9103150d267f","Type":"ContainerStarted","Data":"31ac27c8d988f95e7b59f94abe8de0c0755f0ea00708ce916ac4e0f6c75f2dd0"} Mar 13 14:24:08 crc kubenswrapper[4907]: E0313 14:24:08.144118 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:e7e865363955c670e41b6c042c4f87abceff78f5495ba5c5c82988baad45c978\\\"\"" pod="openstack-operators/placement-operator-controller-manager-574d45c66c-b99f7" podUID="0b3db736-b3d4-494f-9755-9103150d267f" Mar 13 14:24:08 crc kubenswrapper[4907]: I0313 14:24:08.169649 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-5bc894d9b-w5829" event={"ID":"19c320fa-6619-45cd-a73e-65834173b786","Type":"ContainerStarted","Data":"d18a8174c620b5a6ee3028caa18c2a577a0e97ec87adc0db628b0d1c7e9036a8"} Mar 13 14:24:08 crc kubenswrapper[4907]: I0313 14:24:08.173546 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-6854b8b9d9-dj95f" event={"ID":"d139e1bb-8ae8-4760-a1b6-3f7d96f3e9a1","Type":"ContainerStarted","Data":"249124a101607411664b00fc3ece797bc08d43cce738dd15d70fdceaa982f6d3"} Mar 13 14:24:08 crc kubenswrapper[4907]: I0313 14:24:08.188089 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-57b484b4df-m8j7m" event={"ID":"c55515f1-509e-4a4e-813d-fbc466fca489","Type":"ContainerStarted","Data":"f3fff2cc89f13cc8f5f42b57d76dd61682015600446bfbf3151c19fd65c90e59"} Mar 13 14:24:08 crc kubenswrapper[4907]: I0313 14:24:08.194143 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-vzpcd" event={"ID":"a2b0825f-f2f1-4b3f-b7b0-5a476f99b4fe","Type":"ContainerStarted","Data":"079300236dd122d951f49a408fe129275af832b44c6f7e64416f2a0469ccba77"} Mar 13 14:24:08 crc kubenswrapper[4907]: I0313 14:24:08.195847 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-7f84474648-ql4kd" event={"ID":"4ab8f569-de56-4fea-9be0-07908063522a","Type":"ContainerStarted","Data":"2bed19672f429f39772b19fc86747e4dbbce4b48f9ab5aac93c9a3800c55fd88"} Mar 13 14:24:08 crc kubenswrapper[4907]: I0313 14:24:08.197298 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-5964f64c48-gll76" event={"ID":"b7890be5-5773-4562-b1ab-8e647aa16256","Type":"ContainerStarted","Data":"24a301c9f2dbd19e528ed3398be77fecd32947246f3a9d49c7babaca70800a42"} Mar 13 14:24:08 crc kubenswrapper[4907]: I0313 14:24:08.198480 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-684f77d66d-7rdzd" event={"ID":"7d620a90-0e93-41dc-ab39-fc6ea8c461e7","Type":"ContainerStarted","Data":"439611099e1dbdaaa0b92adbf75b260031ff4a85f58ee41465805b2d5cfbd289"} Mar 13 14:24:08 crc kubenswrapper[4907]: I0313 14:24:08.205805 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-77b6666d85-874m2" event={"ID":"d3cdd2ca-3e81-480f-876e-2b975d8bc9d8","Type":"ContainerStarted","Data":"7effad99e4ddab39739a3036988a870d9b248b765b7764f466306ec247ad5974"} Mar 13 14:24:08 crc kubenswrapper[4907]: I0313 14:24:08.208043 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-hjp4p" event={"ID":"a89fbf9b-b391-426b-9f9d-91cdba71602b","Type":"ContainerStarted","Data":"6734e50dcc6089686b369fd6c9ffe2bbbb6b6683f81309eeb5d5494d6365b007"} Mar 13 14:24:08 crc kubenswrapper[4907]: I0313 14:24:08.209137 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-776c5696bf-gvgdl" event={"ID":"29201cb7-c056-4e4e-b0da-cc775c7ef11f","Type":"ContainerStarted","Data":"730ba7fd8e2c2f7a3080b0b3b70054d68651814baacb28b06daf6c30ac9c2710"} Mar 13 14:24:08 crc kubenswrapper[4907]: E0313 14:24:08.210529 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:d9c55e8c6304a0e32289b5e8c69a87ea59b9968918a5c85b7c384633df82c807\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-hjp4p" podUID="a89fbf9b-b391-426b-9f9d-91cdba71602b" Mar 13 14:24:08 crc kubenswrapper[4907]: I0313 14:24:08.216399 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-66d56f6ff4-7cbb2" event={"ID":"8789d213-46a4-4649-add6-51ebe4412c27","Type":"ContainerStarted","Data":"542f7823b45dccbadb548ebf3a6b53de33234194035260a112d7af8fa2a6c417"} Mar 13 14:24:08 crc kubenswrapper[4907]: I0313 14:24:08.218679 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-7f9cc5dd44-sd6nr" event={"ID":"92718823-f344-4d7b-beff-e75f1574f1e2","Type":"ContainerStarted","Data":"471369223f2f7aeca31c0e6f7d51eb344c95995296092c5fe268a7081154592c"} Mar 13 14:24:08 crc kubenswrapper[4907]: I0313 14:24:08.220142 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-5b6b6b4c9f-5m2vw" event={"ID":"e8134e91-4b9d-4e3c-9aec-eba18acabdce","Type":"ContainerStarted","Data":"5b5bc5bef33ef19b061bdb6b2f2e5536bd5cb5c80c42fa20343f63c0e712ee5b"} Mar 13 14:24:08 crc kubenswrapper[4907]: E0313 14:24:08.220255 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:72db77c98e7bca64d469b4dc316e9c8d329681f825d19ef8f333437fb1c6d3f5\\\"\"" pod="openstack-operators/swift-operator-controller-manager-7f9cc5dd44-sd6nr" podUID="92718823-f344-4d7b-beff-e75f1574f1e2" Mar 13 14:24:08 crc kubenswrapper[4907]: I0313 14:24:08.221559 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-bbc5b68f9-cqmhw" event={"ID":"fdddbc99-1a71-4250-ab2a-6f426e744423","Type":"ContainerStarted","Data":"23437d6b081e76de7b1c3ba9d56430532e367089e5a7f5e6435afb9e118f4503"} Mar 13 14:24:08 crc kubenswrapper[4907]: E0313 14:24:08.222673 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:2f63ddf5c95c6c82f6e04bc9f7f20d56dc003614647726ab00276239eec40b7f\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-bbc5b68f9-cqmhw" podUID="fdddbc99-1a71-4250-ab2a-6f426e744423" Mar 13 14:24:08 crc kubenswrapper[4907]: I0313 14:24:08.223198 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-sk2l9" event={"ID":"5f3d2dd8-6ee2-415a-8cc6-55f6e09cb473","Type":"ContainerStarted","Data":"d870537e08a01962b5adaee2ef2e6fa2ced689caa2a1f2199d1293e680e16ad7"} Mar 13 14:24:08 crc kubenswrapper[4907]: I0313 14:24:08.701483 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/9ac47edf-03ab-4e93-b238-2aa00765ab06-webhook-certs\") pod \"openstack-operator-controller-manager-6484b7b757-qldxp\" (UID: \"9ac47edf-03ab-4e93-b238-2aa00765ab06\") " pod="openstack-operators/openstack-operator-controller-manager-6484b7b757-qldxp" Mar 13 14:24:08 crc kubenswrapper[4907]: I0313 14:24:08.701803 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9ac47edf-03ab-4e93-b238-2aa00765ab06-metrics-certs\") pod \"openstack-operator-controller-manager-6484b7b757-qldxp\" (UID: \"9ac47edf-03ab-4e93-b238-2aa00765ab06\") " pod="openstack-operators/openstack-operator-controller-manager-6484b7b757-qldxp" Mar 13 14:24:08 crc kubenswrapper[4907]: E0313 14:24:08.702036 4907 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Mar 13 14:24:08 crc kubenswrapper[4907]: E0313 14:24:08.702112 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9ac47edf-03ab-4e93-b238-2aa00765ab06-webhook-certs podName:9ac47edf-03ab-4e93-b238-2aa00765ab06 nodeName:}" failed. No retries permitted until 2026-03-13 14:24:10.702094757 +0000 UTC m=+1149.601882436 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/9ac47edf-03ab-4e93-b238-2aa00765ab06-webhook-certs") pod "openstack-operator-controller-manager-6484b7b757-qldxp" (UID: "9ac47edf-03ab-4e93-b238-2aa00765ab06") : secret "webhook-server-cert" not found Mar 13 14:24:08 crc kubenswrapper[4907]: E0313 14:24:08.703930 4907 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Mar 13 14:24:08 crc kubenswrapper[4907]: E0313 14:24:08.703988 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9ac47edf-03ab-4e93-b238-2aa00765ab06-metrics-certs podName:9ac47edf-03ab-4e93-b238-2aa00765ab06 nodeName:}" failed. No retries permitted until 2026-03-13 14:24:10.70397205 +0000 UTC m=+1149.603759739 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/9ac47edf-03ab-4e93-b238-2aa00765ab06-metrics-certs") pod "openstack-operator-controller-manager-6484b7b757-qldxp" (UID: "9ac47edf-03ab-4e93-b238-2aa00765ab06") : secret "metrics-server-cert" not found Mar 13 14:24:09 crc kubenswrapper[4907]: E0313 14:24:09.240350 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:d9c55e8c6304a0e32289b5e8c69a87ea59b9968918a5c85b7c384633df82c807\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-hjp4p" podUID="a89fbf9b-b391-426b-9f9d-91cdba71602b" Mar 13 14:24:09 crc kubenswrapper[4907]: E0313 14:24:09.240675 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:e7e865363955c670e41b6c042c4f87abceff78f5495ba5c5c82988baad45c978\\\"\"" pod="openstack-operators/placement-operator-controller-manager-574d45c66c-b99f7" podUID="0b3db736-b3d4-494f-9755-9103150d267f" Mar 13 14:24:09 crc kubenswrapper[4907]: E0313 14:24:09.240713 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-xtvzb" podUID="44ecb24b-4c5c-4cf6-96e9-4d8b52469d3d" Mar 13 14:24:09 crc kubenswrapper[4907]: E0313 14:24:09.240740 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:72db77c98e7bca64d469b4dc316e9c8d329681f825d19ef8f333437fb1c6d3f5\\\"\"" pod="openstack-operators/swift-operator-controller-manager-7f9cc5dd44-sd6nr" podUID="92718823-f344-4d7b-beff-e75f1574f1e2" Mar 13 14:24:09 crc kubenswrapper[4907]: E0313 14:24:09.250084 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:2f63ddf5c95c6c82f6e04bc9f7f20d56dc003614647726ab00276239eec40b7f\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-bbc5b68f9-cqmhw" podUID="fdddbc99-1a71-4250-ab2a-6f426e744423" Mar 13 14:24:09 crc kubenswrapper[4907]: I0313 14:24:09.826637 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/64d55221-60c5-4521-ac84-beed4e9b3993-cert\") pod \"infra-operator-controller-manager-54dc5b8f8d-lp7kj\" (UID: \"64d55221-60c5-4521-ac84-beed4e9b3993\") " pod="openstack-operators/infra-operator-controller-manager-54dc5b8f8d-lp7kj" Mar 13 14:24:09 crc kubenswrapper[4907]: E0313 14:24:09.826859 4907 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Mar 13 14:24:09 crc kubenswrapper[4907]: E0313 14:24:09.826926 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/64d55221-60c5-4521-ac84-beed4e9b3993-cert podName:64d55221-60c5-4521-ac84-beed4e9b3993 nodeName:}" failed. No retries permitted until 2026-03-13 14:24:13.826912147 +0000 UTC m=+1152.726699836 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/64d55221-60c5-4521-ac84-beed4e9b3993-cert") pod "infra-operator-controller-manager-54dc5b8f8d-lp7kj" (UID: "64d55221-60c5-4521-ac84-beed4e9b3993") : secret "infra-operator-webhook-server-cert" not found Mar 13 14:24:10 crc kubenswrapper[4907]: I0313 14:24:10.135847 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f915558a-c0ab-4d8e-a427-e56a7588e382-cert\") pod \"openstack-baremetal-operator-controller-manager-6f7958d774rr6kf\" (UID: \"f915558a-c0ab-4d8e-a427-e56a7588e382\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6f7958d774rr6kf" Mar 13 14:24:10 crc kubenswrapper[4907]: E0313 14:24:10.136586 4907 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 13 14:24:10 crc kubenswrapper[4907]: E0313 14:24:10.136648 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f915558a-c0ab-4d8e-a427-e56a7588e382-cert podName:f915558a-c0ab-4d8e-a427-e56a7588e382 nodeName:}" failed. No retries permitted until 2026-03-13 14:24:14.136625616 +0000 UTC m=+1153.036413315 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/f915558a-c0ab-4d8e-a427-e56a7588e382-cert") pod "openstack-baremetal-operator-controller-manager-6f7958d774rr6kf" (UID: "f915558a-c0ab-4d8e-a427-e56a7588e382") : secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 13 14:24:10 crc kubenswrapper[4907]: I0313 14:24:10.747575 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9ac47edf-03ab-4e93-b238-2aa00765ab06-metrics-certs\") pod \"openstack-operator-controller-manager-6484b7b757-qldxp\" (UID: \"9ac47edf-03ab-4e93-b238-2aa00765ab06\") " pod="openstack-operators/openstack-operator-controller-manager-6484b7b757-qldxp" Mar 13 14:24:10 crc kubenswrapper[4907]: I0313 14:24:10.747753 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/9ac47edf-03ab-4e93-b238-2aa00765ab06-webhook-certs\") pod \"openstack-operator-controller-manager-6484b7b757-qldxp\" (UID: \"9ac47edf-03ab-4e93-b238-2aa00765ab06\") " pod="openstack-operators/openstack-operator-controller-manager-6484b7b757-qldxp" Mar 13 14:24:10 crc kubenswrapper[4907]: E0313 14:24:10.748007 4907 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Mar 13 14:24:10 crc kubenswrapper[4907]: E0313 14:24:10.748063 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9ac47edf-03ab-4e93-b238-2aa00765ab06-webhook-certs podName:9ac47edf-03ab-4e93-b238-2aa00765ab06 nodeName:}" failed. No retries permitted until 2026-03-13 14:24:14.748048022 +0000 UTC m=+1153.647835711 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/9ac47edf-03ab-4e93-b238-2aa00765ab06-webhook-certs") pod "openstack-operator-controller-manager-6484b7b757-qldxp" (UID: "9ac47edf-03ab-4e93-b238-2aa00765ab06") : secret "webhook-server-cert" not found Mar 13 14:24:10 crc kubenswrapper[4907]: E0313 14:24:10.748104 4907 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Mar 13 14:24:10 crc kubenswrapper[4907]: E0313 14:24:10.748123 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9ac47edf-03ab-4e93-b238-2aa00765ab06-metrics-certs podName:9ac47edf-03ab-4e93-b238-2aa00765ab06 nodeName:}" failed. No retries permitted until 2026-03-13 14:24:14.748117114 +0000 UTC m=+1153.647904803 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/9ac47edf-03ab-4e93-b238-2aa00765ab06-metrics-certs") pod "openstack-operator-controller-manager-6484b7b757-qldxp" (UID: "9ac47edf-03ab-4e93-b238-2aa00765ab06") : secret "metrics-server-cert" not found Mar 13 14:24:13 crc kubenswrapper[4907]: I0313 14:24:13.894822 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/64d55221-60c5-4521-ac84-beed4e9b3993-cert\") pod \"infra-operator-controller-manager-54dc5b8f8d-lp7kj\" (UID: \"64d55221-60c5-4521-ac84-beed4e9b3993\") " pod="openstack-operators/infra-operator-controller-manager-54dc5b8f8d-lp7kj" Mar 13 14:24:13 crc kubenswrapper[4907]: E0313 14:24:13.895220 4907 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Mar 13 14:24:13 crc kubenswrapper[4907]: E0313 14:24:13.895266 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/64d55221-60c5-4521-ac84-beed4e9b3993-cert podName:64d55221-60c5-4521-ac84-beed4e9b3993 nodeName:}" failed. No retries permitted until 2026-03-13 14:24:21.89525322 +0000 UTC m=+1160.795040909 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/64d55221-60c5-4521-ac84-beed4e9b3993-cert") pod "infra-operator-controller-manager-54dc5b8f8d-lp7kj" (UID: "64d55221-60c5-4521-ac84-beed4e9b3993") : secret "infra-operator-webhook-server-cert" not found Mar 13 14:24:14 crc kubenswrapper[4907]: I0313 14:24:14.236451 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f915558a-c0ab-4d8e-a427-e56a7588e382-cert\") pod \"openstack-baremetal-operator-controller-manager-6f7958d774rr6kf\" (UID: \"f915558a-c0ab-4d8e-a427-e56a7588e382\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6f7958d774rr6kf" Mar 13 14:24:14 crc kubenswrapper[4907]: E0313 14:24:14.236635 4907 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 13 14:24:14 crc kubenswrapper[4907]: E0313 14:24:14.236738 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f915558a-c0ab-4d8e-a427-e56a7588e382-cert podName:f915558a-c0ab-4d8e-a427-e56a7588e382 nodeName:}" failed. No retries permitted until 2026-03-13 14:24:22.236708674 +0000 UTC m=+1161.136496363 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/f915558a-c0ab-4d8e-a427-e56a7588e382-cert") pod "openstack-baremetal-operator-controller-manager-6f7958d774rr6kf" (UID: "f915558a-c0ab-4d8e-a427-e56a7588e382") : secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 13 14:24:14 crc kubenswrapper[4907]: I0313 14:24:14.844722 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/9ac47edf-03ab-4e93-b238-2aa00765ab06-webhook-certs\") pod \"openstack-operator-controller-manager-6484b7b757-qldxp\" (UID: \"9ac47edf-03ab-4e93-b238-2aa00765ab06\") " pod="openstack-operators/openstack-operator-controller-manager-6484b7b757-qldxp" Mar 13 14:24:14 crc kubenswrapper[4907]: I0313 14:24:14.844786 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9ac47edf-03ab-4e93-b238-2aa00765ab06-metrics-certs\") pod \"openstack-operator-controller-manager-6484b7b757-qldxp\" (UID: \"9ac47edf-03ab-4e93-b238-2aa00765ab06\") " pod="openstack-operators/openstack-operator-controller-manager-6484b7b757-qldxp" Mar 13 14:24:14 crc kubenswrapper[4907]: E0313 14:24:14.844927 4907 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Mar 13 14:24:14 crc kubenswrapper[4907]: E0313 14:24:14.844994 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9ac47edf-03ab-4e93-b238-2aa00765ab06-webhook-certs podName:9ac47edf-03ab-4e93-b238-2aa00765ab06 nodeName:}" failed. No retries permitted until 2026-03-13 14:24:22.844976502 +0000 UTC m=+1161.744764191 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/9ac47edf-03ab-4e93-b238-2aa00765ab06-webhook-certs") pod "openstack-operator-controller-manager-6484b7b757-qldxp" (UID: "9ac47edf-03ab-4e93-b238-2aa00765ab06") : secret "webhook-server-cert" not found Mar 13 14:24:14 crc kubenswrapper[4907]: E0313 14:24:14.845032 4907 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Mar 13 14:24:14 crc kubenswrapper[4907]: E0313 14:24:14.845110 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9ac47edf-03ab-4e93-b238-2aa00765ab06-metrics-certs podName:9ac47edf-03ab-4e93-b238-2aa00765ab06 nodeName:}" failed. No retries permitted until 2026-03-13 14:24:22.845092415 +0000 UTC m=+1161.744880104 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/9ac47edf-03ab-4e93-b238-2aa00765ab06-metrics-certs") pod "openstack-operator-controller-manager-6484b7b757-qldxp" (UID: "9ac47edf-03ab-4e93-b238-2aa00765ab06") : secret "metrics-server-cert" not found Mar 13 14:24:18 crc kubenswrapper[4907]: I0313 14:24:18.041473 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 14:24:18 crc kubenswrapper[4907]: I0313 14:24:18.041977 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 14:24:20 crc kubenswrapper[4907]: E0313 14:24:20.901052 4907 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/neutron-operator@sha256:5fe5351a3de5e1267112d52cd81477a01d47f90be713cc5439c76543a4c33721" Mar 13 14:24:20 crc kubenswrapper[4907]: E0313 14:24:20.901708 4907 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/neutron-operator@sha256:5fe5351a3de5e1267112d52cd81477a01d47f90be713cc5439c76543a4c33721,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-b25c5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-776c5696bf-gvgdl_openstack-operators(29201cb7-c056-4e4e-b0da-cc775c7ef11f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 13 14:24:20 crc kubenswrapper[4907]: E0313 14:24:20.903132 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/neutron-operator-controller-manager-776c5696bf-gvgdl" podUID="29201cb7-c056-4e4e-b0da-cc775c7ef11f" Mar 13 14:24:21 crc kubenswrapper[4907]: E0313 14:24:21.357320 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:5fe5351a3de5e1267112d52cd81477a01d47f90be713cc5439c76543a4c33721\\\"\"" pod="openstack-operators/neutron-operator-controller-manager-776c5696bf-gvgdl" podUID="29201cb7-c056-4e4e-b0da-cc775c7ef11f" Mar 13 14:24:21 crc kubenswrapper[4907]: E0313 14:24:21.666391 4907 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/mariadb-operator@sha256:a26d062af19b3bc6dc6633171f1eff8eec33e8e925465d4968a0b9a36012a7e7" Mar 13 14:24:21 crc kubenswrapper[4907]: E0313 14:24:21.666569 4907 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/mariadb-operator@sha256:a26d062af19b3bc6dc6633171f1eff8eec33e8e925465d4968a0b9a36012a7e7,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-nmthx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod mariadb-operator-controller-manager-5b6b6b4c9f-5m2vw_openstack-operators(e8134e91-4b9d-4e3c-9aec-eba18acabdce): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 13 14:24:21 crc kubenswrapper[4907]: E0313 14:24:21.667924 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/mariadb-operator-controller-manager-5b6b6b4c9f-5m2vw" podUID="e8134e91-4b9d-4e3c-9aec-eba18acabdce" Mar 13 14:24:21 crc kubenswrapper[4907]: I0313 14:24:21.964996 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/64d55221-60c5-4521-ac84-beed4e9b3993-cert\") pod \"infra-operator-controller-manager-54dc5b8f8d-lp7kj\" (UID: \"64d55221-60c5-4521-ac84-beed4e9b3993\") " pod="openstack-operators/infra-operator-controller-manager-54dc5b8f8d-lp7kj" Mar 13 14:24:21 crc kubenswrapper[4907]: E0313 14:24:21.965652 4907 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Mar 13 14:24:21 crc kubenswrapper[4907]: E0313 14:24:21.965701 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/64d55221-60c5-4521-ac84-beed4e9b3993-cert podName:64d55221-60c5-4521-ac84-beed4e9b3993 nodeName:}" failed. No retries permitted until 2026-03-13 14:24:37.96568585 +0000 UTC m=+1176.865473539 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/64d55221-60c5-4521-ac84-beed4e9b3993-cert") pod "infra-operator-controller-manager-54dc5b8f8d-lp7kj" (UID: "64d55221-60c5-4521-ac84-beed4e9b3993") : secret "infra-operator-webhook-server-cert" not found Mar 13 14:24:22 crc kubenswrapper[4907]: I0313 14:24:22.270014 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f915558a-c0ab-4d8e-a427-e56a7588e382-cert\") pod \"openstack-baremetal-operator-controller-manager-6f7958d774rr6kf\" (UID: \"f915558a-c0ab-4d8e-a427-e56a7588e382\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6f7958d774rr6kf" Mar 13 14:24:22 crc kubenswrapper[4907]: E0313 14:24:22.270296 4907 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 13 14:24:22 crc kubenswrapper[4907]: E0313 14:24:22.270358 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f915558a-c0ab-4d8e-a427-e56a7588e382-cert podName:f915558a-c0ab-4d8e-a427-e56a7588e382 nodeName:}" failed. No retries permitted until 2026-03-13 14:24:38.270343409 +0000 UTC m=+1177.170131098 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/f915558a-c0ab-4d8e-a427-e56a7588e382-cert") pod "openstack-baremetal-operator-controller-manager-6f7958d774rr6kf" (UID: "f915558a-c0ab-4d8e-a427-e56a7588e382") : secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 13 14:24:22 crc kubenswrapper[4907]: E0313 14:24:22.340428 4907 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/manila-operator@sha256:dd62e104225ea255af5a32828af4c21e1dfb50fbdf35cd41d07d1326f9017a40" Mar 13 14:24:22 crc kubenswrapper[4907]: E0313 14:24:22.340587 4907 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/manila-operator@sha256:dd62e104225ea255af5a32828af4c21e1dfb50fbdf35cd41d07d1326f9017a40,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-7cq7h,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod manila-operator-controller-manager-57b484b4df-m8j7m_openstack-operators(c55515f1-509e-4a4e-813d-fbc466fca489): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 13 14:24:22 crc kubenswrapper[4907]: E0313 14:24:22.342493 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/manila-operator-controller-manager-57b484b4df-m8j7m" podUID="c55515f1-509e-4a4e-813d-fbc466fca489" Mar 13 14:24:22 crc kubenswrapper[4907]: E0313 14:24:22.366418 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/manila-operator@sha256:dd62e104225ea255af5a32828af4c21e1dfb50fbdf35cd41d07d1326f9017a40\\\"\"" pod="openstack-operators/manila-operator-controller-manager-57b484b4df-m8j7m" podUID="c55515f1-509e-4a4e-813d-fbc466fca489" Mar 13 14:24:22 crc kubenswrapper[4907]: E0313 14:24:22.366633 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/mariadb-operator@sha256:a26d062af19b3bc6dc6633171f1eff8eec33e8e925465d4968a0b9a36012a7e7\\\"\"" pod="openstack-operators/mariadb-operator-controller-manager-5b6b6b4c9f-5m2vw" podUID="e8134e91-4b9d-4e3c-9aec-eba18acabdce" Mar 13 14:24:22 crc kubenswrapper[4907]: I0313 14:24:22.879281 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/9ac47edf-03ab-4e93-b238-2aa00765ab06-webhook-certs\") pod \"openstack-operator-controller-manager-6484b7b757-qldxp\" (UID: \"9ac47edf-03ab-4e93-b238-2aa00765ab06\") " pod="openstack-operators/openstack-operator-controller-manager-6484b7b757-qldxp" Mar 13 14:24:22 crc kubenswrapper[4907]: I0313 14:24:22.879640 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9ac47edf-03ab-4e93-b238-2aa00765ab06-metrics-certs\") pod \"openstack-operator-controller-manager-6484b7b757-qldxp\" (UID: \"9ac47edf-03ab-4e93-b238-2aa00765ab06\") " pod="openstack-operators/openstack-operator-controller-manager-6484b7b757-qldxp" Mar 13 14:24:22 crc kubenswrapper[4907]: E0313 14:24:22.879500 4907 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Mar 13 14:24:22 crc kubenswrapper[4907]: E0313 14:24:22.879973 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9ac47edf-03ab-4e93-b238-2aa00765ab06-webhook-certs podName:9ac47edf-03ab-4e93-b238-2aa00765ab06 nodeName:}" failed. No retries permitted until 2026-03-13 14:24:38.879949903 +0000 UTC m=+1177.779737592 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/9ac47edf-03ab-4e93-b238-2aa00765ab06-webhook-certs") pod "openstack-operator-controller-manager-6484b7b757-qldxp" (UID: "9ac47edf-03ab-4e93-b238-2aa00765ab06") : secret "webhook-server-cert" not found Mar 13 14:24:22 crc kubenswrapper[4907]: E0313 14:24:22.879993 4907 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Mar 13 14:24:22 crc kubenswrapper[4907]: E0313 14:24:22.880037 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9ac47edf-03ab-4e93-b238-2aa00765ab06-metrics-certs podName:9ac47edf-03ab-4e93-b238-2aa00765ab06 nodeName:}" failed. No retries permitted until 2026-03-13 14:24:38.880025606 +0000 UTC m=+1177.779813395 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/9ac47edf-03ab-4e93-b238-2aa00765ab06-metrics-certs") pod "openstack-operator-controller-manager-6484b7b757-qldxp" (UID: "9ac47edf-03ab-4e93-b238-2aa00765ab06") : secret "metrics-server-cert" not found Mar 13 14:24:24 crc kubenswrapper[4907]: E0313 14:24:24.356564 4907 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/nova-operator@sha256:bbe772fa051f782c9dcc3c34ce43495e1116aa9089a760c10068790baa9b25ff" Mar 13 14:24:24 crc kubenswrapper[4907]: E0313 14:24:24.356739 4907 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:bbe772fa051f782c9dcc3c34ce43495e1116aa9089a760c10068790baa9b25ff,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-2gkcl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-7f84474648-ql4kd_openstack-operators(4ab8f569-de56-4fea-9be0-07908063522a): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 13 14:24:24 crc kubenswrapper[4907]: E0313 14:24:24.357932 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/nova-operator-controller-manager-7f84474648-ql4kd" podUID="4ab8f569-de56-4fea-9be0-07908063522a" Mar 13 14:24:24 crc kubenswrapper[4907]: E0313 14:24:24.378895 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:bbe772fa051f782c9dcc3c34ce43495e1116aa9089a760c10068790baa9b25ff\\\"\"" pod="openstack-operators/nova-operator-controller-manager-7f84474648-ql4kd" podUID="4ab8f569-de56-4fea-9be0-07908063522a" Mar 13 14:24:29 crc kubenswrapper[4907]: I0313 14:24:29.409912 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-sk2l9" event={"ID":"5f3d2dd8-6ee2-415a-8cc6-55f6e09cb473","Type":"ContainerStarted","Data":"05abb31bbf4ec7fe38a62ff30c059ce6b087be181a0e28042622a070932727fd"} Mar 13 14:24:29 crc kubenswrapper[4907]: I0313 14:24:29.410595 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-sk2l9" Mar 13 14:24:29 crc kubenswrapper[4907]: I0313 14:24:29.411998 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-684f77d66d-7rdzd" event={"ID":"7d620a90-0e93-41dc-ab39-fc6ea8c461e7","Type":"ContainerStarted","Data":"93faad90b79e0bc4524f13f90e4ca5cc1287f3e4911bbcede638bdc9c21a3ee8"} Mar 13 14:24:29 crc kubenswrapper[4907]: I0313 14:24:29.412580 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-684f77d66d-7rdzd" Mar 13 14:24:29 crc kubenswrapper[4907]: I0313 14:24:29.414507 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-vzpcd" event={"ID":"a2b0825f-f2f1-4b3f-b7b0-5a476f99b4fe","Type":"ContainerStarted","Data":"512d1ab3188bf1032e497da8cbbffdc3e5d0fb10c97420ee700178ce6c2df88a"} Mar 13 14:24:29 crc kubenswrapper[4907]: I0313 14:24:29.414936 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-vzpcd" Mar 13 14:24:29 crc kubenswrapper[4907]: I0313 14:24:29.416013 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-5964f64c48-gll76" event={"ID":"b7890be5-5773-4562-b1ab-8e647aa16256","Type":"ContainerStarted","Data":"43d4c09378c5dcd8bef4c18832d79b482b75c17adb040b1d95fb1f2993ee3ab2"} Mar 13 14:24:29 crc kubenswrapper[4907]: I0313 14:24:29.416332 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-5964f64c48-gll76" Mar 13 14:24:29 crc kubenswrapper[4907]: I0313 14:24:29.417391 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-7f9cc5dd44-sd6nr" event={"ID":"92718823-f344-4d7b-beff-e75f1574f1e2","Type":"ContainerStarted","Data":"1426ea8e86e38c9303fa0dc9e6313ceb1f59abf0760282c44b6ec01bd4aae436"} Mar 13 14:24:29 crc kubenswrapper[4907]: I0313 14:24:29.417720 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-7f9cc5dd44-sd6nr" Mar 13 14:24:29 crc kubenswrapper[4907]: I0313 14:24:29.419299 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-574d45c66c-b99f7" event={"ID":"0b3db736-b3d4-494f-9755-9103150d267f","Type":"ContainerStarted","Data":"7d0eab1124c4f5223e3ff4691e59dd0f8e154658825f67875ee41d5b34768f3e"} Mar 13 14:24:29 crc kubenswrapper[4907]: I0313 14:24:29.419605 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-574d45c66c-b99f7" Mar 13 14:24:29 crc kubenswrapper[4907]: I0313 14:24:29.421673 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-6d9d6b584d-r7bhq" event={"ID":"8ecaf4b8-808e-44b9-8ae7-6ef9519c4a66","Type":"ContainerStarted","Data":"75ae9afda13bc4d203b15ff875e1f989dc0aee5879dcf40a5388c3239c039ecb"} Mar 13 14:24:29 crc kubenswrapper[4907]: I0313 14:24:29.421821 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-6d9d6b584d-r7bhq" Mar 13 14:24:29 crc kubenswrapper[4907]: I0313 14:24:29.423025 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-bbc5b68f9-cqmhw" event={"ID":"fdddbc99-1a71-4250-ab2a-6f426e744423","Type":"ContainerStarted","Data":"225ace33198a3da486e62fa067504b35a227f923a438459073f4ea5dc42a43c6"} Mar 13 14:24:29 crc kubenswrapper[4907]: I0313 14:24:29.423175 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-bbc5b68f9-cqmhw" Mar 13 14:24:29 crc kubenswrapper[4907]: I0313 14:24:29.424383 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-hjp4p" event={"ID":"a89fbf9b-b391-426b-9f9d-91cdba71602b","Type":"ContainerStarted","Data":"7102de9beb3b8815117d2ed9366d7536e6f3f834b813e68b06eb8d8f1bffbf25"} Mar 13 14:24:29 crc kubenswrapper[4907]: I0313 14:24:29.424595 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-hjp4p" Mar 13 14:24:29 crc kubenswrapper[4907]: I0313 14:24:29.426165 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-xtvzb" event={"ID":"44ecb24b-4c5c-4cf6-96e9-4d8b52469d3d","Type":"ContainerStarted","Data":"0ae4f540d42cbc607b9ece7623709fa969fa343a1eabbbf6fe6be4e3fd1c241c"} Mar 13 14:24:29 crc kubenswrapper[4907]: I0313 14:24:29.427563 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-984cd4dcf-w497d" event={"ID":"77b1eb71-a455-4ec3-8e17-a847575cf0b0","Type":"ContainerStarted","Data":"d5eb9803af60e9700eca1bfdb4a88a53f1da388443d89a4bcb492cf22e743924"} Mar 13 14:24:29 crc kubenswrapper[4907]: I0313 14:24:29.427706 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-984cd4dcf-w497d" Mar 13 14:24:29 crc kubenswrapper[4907]: I0313 14:24:29.429526 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-77b6666d85-874m2" event={"ID":"d3cdd2ca-3e81-480f-876e-2b975d8bc9d8","Type":"ContainerStarted","Data":"e60ac8a501dd0be585475c0e81ac1273c6a26ae808f805e897cc13551694b004"} Mar 13 14:24:29 crc kubenswrapper[4907]: I0313 14:24:29.429661 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-77b6666d85-874m2" Mar 13 14:24:29 crc kubenswrapper[4907]: I0313 14:24:29.431420 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-d47688694-tl8t4" event={"ID":"863c77e9-4022-4ab9-8ac7-a4006d22813c","Type":"ContainerStarted","Data":"35717b532a5717fb681c62aebea52a6a1b7aca58bed964725f877360feefee9f"} Mar 13 14:24:29 crc kubenswrapper[4907]: I0313 14:24:29.431558 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-d47688694-tl8t4" Mar 13 14:24:29 crc kubenswrapper[4907]: I0313 14:24:29.433095 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-5bc894d9b-w5829" event={"ID":"19c320fa-6619-45cd-a73e-65834173b786","Type":"ContainerStarted","Data":"090c2fbf2baac7dfb5c922f84512bb5616df8cf7281e1d6a388d38f2f4a6a2f9"} Mar 13 14:24:29 crc kubenswrapper[4907]: I0313 14:24:29.433226 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-5bc894d9b-w5829" Mar 13 14:24:29 crc kubenswrapper[4907]: I0313 14:24:29.434628 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-6854b8b9d9-dj95f" event={"ID":"d139e1bb-8ae8-4760-a1b6-3f7d96f3e9a1","Type":"ContainerStarted","Data":"55330794c339c40fc7cd21675097faf304a29511453d047470018a57a2abdc0f"} Mar 13 14:24:29 crc kubenswrapper[4907]: I0313 14:24:29.434745 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-6854b8b9d9-dj95f" Mar 13 14:24:29 crc kubenswrapper[4907]: I0313 14:24:29.436362 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-66d56f6ff4-7cbb2" event={"ID":"8789d213-46a4-4649-add6-51ebe4412c27","Type":"ContainerStarted","Data":"898c1a1ae3b709eac2d6916ccecd30e0d7e2e968f67dc58e937fdda34ccc2478"} Mar 13 14:24:29 crc kubenswrapper[4907]: I0313 14:24:29.437217 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-66d56f6ff4-7cbb2" Mar 13 14:24:29 crc kubenswrapper[4907]: I0313 14:24:29.505506 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-sk2l9" podStartSLOduration=4.895257022 podStartE2EDuration="23.505477645s" podCreationTimestamp="2026-03-13 14:24:06 +0000 UTC" firstStartedPulling="2026-03-13 14:24:07.696858468 +0000 UTC m=+1146.596646167" lastFinishedPulling="2026-03-13 14:24:26.307079101 +0000 UTC m=+1165.206866790" observedRunningTime="2026-03-13 14:24:29.450874274 +0000 UTC m=+1168.350661963" watchObservedRunningTime="2026-03-13 14:24:29.505477645 +0000 UTC m=+1168.405265324" Mar 13 14:24:29 crc kubenswrapper[4907]: I0313 14:24:29.585641 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-6854b8b9d9-dj95f" podStartSLOduration=7.113427147 podStartE2EDuration="23.585611679s" podCreationTimestamp="2026-03-13 14:24:06 +0000 UTC" firstStartedPulling="2026-03-13 14:24:07.846129748 +0000 UTC m=+1146.745917437" lastFinishedPulling="2026-03-13 14:24:24.31831428 +0000 UTC m=+1163.218101969" observedRunningTime="2026-03-13 14:24:29.511649198 +0000 UTC m=+1168.411436887" watchObservedRunningTime="2026-03-13 14:24:29.585611679 +0000 UTC m=+1168.485399368" Mar 13 14:24:29 crc kubenswrapper[4907]: I0313 14:24:29.747944 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-5964f64c48-gll76" podStartSLOduration=7.231655757 podStartE2EDuration="24.74792568s" podCreationTimestamp="2026-03-13 14:24:05 +0000 UTC" firstStartedPulling="2026-03-13 14:24:07.405503211 +0000 UTC m=+1146.305290900" lastFinishedPulling="2026-03-13 14:24:24.921773134 +0000 UTC m=+1163.821560823" observedRunningTime="2026-03-13 14:24:29.595635148 +0000 UTC m=+1168.495422837" watchObservedRunningTime="2026-03-13 14:24:29.74792568 +0000 UTC m=+1168.647713369" Mar 13 14:24:29 crc kubenswrapper[4907]: I0313 14:24:29.748762 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-7f9cc5dd44-sd6nr" podStartSLOduration=3.23411504 podStartE2EDuration="23.748757394s" podCreationTimestamp="2026-03-13 14:24:06 +0000 UTC" firstStartedPulling="2026-03-13 14:24:08.101212625 +0000 UTC m=+1147.001000314" lastFinishedPulling="2026-03-13 14:24:28.615854979 +0000 UTC m=+1167.515642668" observedRunningTime="2026-03-13 14:24:29.68255461 +0000 UTC m=+1168.582342299" watchObservedRunningTime="2026-03-13 14:24:29.748757394 +0000 UTC m=+1168.648545083" Mar 13 14:24:30 crc kubenswrapper[4907]: I0313 14:24:30.129325 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-vzpcd" podStartSLOduration=7.266576323 podStartE2EDuration="24.129308727s" podCreationTimestamp="2026-03-13 14:24:06 +0000 UTC" firstStartedPulling="2026-03-13 14:24:08.05760486 +0000 UTC m=+1146.957392549" lastFinishedPulling="2026-03-13 14:24:24.920337264 +0000 UTC m=+1163.820124953" observedRunningTime="2026-03-13 14:24:29.834851613 +0000 UTC m=+1168.734639302" watchObservedRunningTime="2026-03-13 14:24:30.129308727 +0000 UTC m=+1169.029096416" Mar 13 14:24:30 crc kubenswrapper[4907]: I0313 14:24:30.250709 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-574d45c66c-b99f7" podStartSLOduration=3.419635161 podStartE2EDuration="24.25068595s" podCreationTimestamp="2026-03-13 14:24:06 +0000 UTC" firstStartedPulling="2026-03-13 14:24:07.868166552 +0000 UTC m=+1146.767954241" lastFinishedPulling="2026-03-13 14:24:28.699217321 +0000 UTC m=+1167.599005030" observedRunningTime="2026-03-13 14:24:30.24427116 +0000 UTC m=+1169.144058849" watchObservedRunningTime="2026-03-13 14:24:30.25068595 +0000 UTC m=+1169.150473639" Mar 13 14:24:30 crc kubenswrapper[4907]: I0313 14:24:30.251410 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-684f77d66d-7rdzd" podStartSLOduration=7.009688866 podStartE2EDuration="24.251403779s" podCreationTimestamp="2026-03-13 14:24:06 +0000 UTC" firstStartedPulling="2026-03-13 14:24:07.679619398 +0000 UTC m=+1146.579407087" lastFinishedPulling="2026-03-13 14:24:24.921334311 +0000 UTC m=+1163.821122000" observedRunningTime="2026-03-13 14:24:30.141099865 +0000 UTC m=+1169.040887554" watchObservedRunningTime="2026-03-13 14:24:30.251403779 +0000 UTC m=+1169.151191468" Mar 13 14:24:30 crc kubenswrapper[4907]: I0313 14:24:30.373509 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-hjp4p" podStartSLOduration=4.297184929 podStartE2EDuration="24.37349029s" podCreationTimestamp="2026-03-13 14:24:06 +0000 UTC" firstStartedPulling="2026-03-13 14:24:08.10352641 +0000 UTC m=+1147.003314089" lastFinishedPulling="2026-03-13 14:24:28.179831761 +0000 UTC m=+1167.079619450" observedRunningTime="2026-03-13 14:24:30.359562242 +0000 UTC m=+1169.259349931" watchObservedRunningTime="2026-03-13 14:24:30.37349029 +0000 UTC m=+1169.273277979" Mar 13 14:24:30 crc kubenswrapper[4907]: I0313 14:24:30.527920 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-984cd4dcf-w497d" podStartSLOduration=8.012141324 podStartE2EDuration="25.527902163s" podCreationTimestamp="2026-03-13 14:24:05 +0000 UTC" firstStartedPulling="2026-03-13 14:24:07.403790733 +0000 UTC m=+1146.303578422" lastFinishedPulling="2026-03-13 14:24:24.919551572 +0000 UTC m=+1163.819339261" observedRunningTime="2026-03-13 14:24:30.458097038 +0000 UTC m=+1169.357884727" watchObservedRunningTime="2026-03-13 14:24:30.527902163 +0000 UTC m=+1169.427689852" Mar 13 14:24:30 crc kubenswrapper[4907]: I0313 14:24:30.569601 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-d47688694-tl8t4" podStartSLOduration=7.634635386 podStartE2EDuration="25.569583994s" podCreationTimestamp="2026-03-13 14:24:05 +0000 UTC" firstStartedPulling="2026-03-13 14:24:06.98626655 +0000 UTC m=+1145.886054229" lastFinishedPulling="2026-03-13 14:24:24.921215148 +0000 UTC m=+1163.821002837" observedRunningTime="2026-03-13 14:24:30.567315591 +0000 UTC m=+1169.467103291" watchObservedRunningTime="2026-03-13 14:24:30.569583994 +0000 UTC m=+1169.469371683" Mar 13 14:24:30 crc kubenswrapper[4907]: I0313 14:24:30.569713 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-bbc5b68f9-cqmhw" podStartSLOduration=4.044845458 podStartE2EDuration="24.569709907s" podCreationTimestamp="2026-03-13 14:24:06 +0000 UTC" firstStartedPulling="2026-03-13 14:24:08.106246495 +0000 UTC m=+1147.006034184" lastFinishedPulling="2026-03-13 14:24:28.631110944 +0000 UTC m=+1167.530898633" observedRunningTime="2026-03-13 14:24:30.534740993 +0000 UTC m=+1169.434528682" watchObservedRunningTime="2026-03-13 14:24:30.569709907 +0000 UTC m=+1169.469497596" Mar 13 14:24:30 crc kubenswrapper[4907]: I0313 14:24:30.622028 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-77b6666d85-874m2" podStartSLOduration=8.443666728 podStartE2EDuration="25.622006865s" podCreationTimestamp="2026-03-13 14:24:05 +0000 UTC" firstStartedPulling="2026-03-13 14:24:07.742992604 +0000 UTC m=+1146.642780293" lastFinishedPulling="2026-03-13 14:24:24.921332741 +0000 UTC m=+1163.821120430" observedRunningTime="2026-03-13 14:24:30.618129047 +0000 UTC m=+1169.517916746" watchObservedRunningTime="2026-03-13 14:24:30.622006865 +0000 UTC m=+1169.521794554" Mar 13 14:24:30 crc kubenswrapper[4907]: I0313 14:24:30.669216 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-5bc894d9b-w5829" podStartSLOduration=7.839257027 podStartE2EDuration="25.669198429s" podCreationTimestamp="2026-03-13 14:24:05 +0000 UTC" firstStartedPulling="2026-03-13 14:24:07.742903842 +0000 UTC m=+1146.642691531" lastFinishedPulling="2026-03-13 14:24:25.572845244 +0000 UTC m=+1164.472632933" observedRunningTime="2026-03-13 14:24:30.664145179 +0000 UTC m=+1169.563932868" watchObservedRunningTime="2026-03-13 14:24:30.669198429 +0000 UTC m=+1169.568986118" Mar 13 14:24:30 crc kubenswrapper[4907]: I0313 14:24:30.691498 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-xtvzb" podStartSLOduration=4.061683358 podStartE2EDuration="24.69148052s" podCreationTimestamp="2026-03-13 14:24:06 +0000 UTC" firstStartedPulling="2026-03-13 14:24:08.107552832 +0000 UTC m=+1147.007340521" lastFinishedPulling="2026-03-13 14:24:28.737349994 +0000 UTC m=+1167.637137683" observedRunningTime="2026-03-13 14:24:30.689493725 +0000 UTC m=+1169.589281424" watchObservedRunningTime="2026-03-13 14:24:30.69148052 +0000 UTC m=+1169.591268209" Mar 13 14:24:30 crc kubenswrapper[4907]: I0313 14:24:30.722136 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-66d56f6ff4-7cbb2" podStartSLOduration=7.984207326 podStartE2EDuration="25.722111694s" podCreationTimestamp="2026-03-13 14:24:05 +0000 UTC" firstStartedPulling="2026-03-13 14:24:07.182699673 +0000 UTC m=+1146.082487362" lastFinishedPulling="2026-03-13 14:24:24.920604041 +0000 UTC m=+1163.820391730" observedRunningTime="2026-03-13 14:24:30.719894793 +0000 UTC m=+1169.619682482" watchObservedRunningTime="2026-03-13 14:24:30.722111694 +0000 UTC m=+1169.621899373" Mar 13 14:24:32 crc kubenswrapper[4907]: I0313 14:24:32.805284 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-6d9d6b584d-r7bhq" podStartSLOduration=11.229193399 podStartE2EDuration="27.805270126s" podCreationTimestamp="2026-03-13 14:24:05 +0000 UTC" firstStartedPulling="2026-03-13 14:24:07.744627169 +0000 UTC m=+1146.644414858" lastFinishedPulling="2026-03-13 14:24:24.320703896 +0000 UTC m=+1163.220491585" observedRunningTime="2026-03-13 14:24:30.756706978 +0000 UTC m=+1169.656494677" watchObservedRunningTime="2026-03-13 14:24:32.805270126 +0000 UTC m=+1171.705057815" Mar 13 14:24:36 crc kubenswrapper[4907]: I0313 14:24:36.219133 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-d47688694-tl8t4" Mar 13 14:24:36 crc kubenswrapper[4907]: I0313 14:24:36.239754 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-984cd4dcf-w497d" Mar 13 14:24:36 crc kubenswrapper[4907]: I0313 14:24:36.266581 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-66d56f6ff4-7cbb2" Mar 13 14:24:36 crc kubenswrapper[4907]: I0313 14:24:36.279595 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-5964f64c48-gll76" Mar 13 14:24:36 crc kubenswrapper[4907]: I0313 14:24:36.324657 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-77b6666d85-874m2" Mar 13 14:24:36 crc kubenswrapper[4907]: I0313 14:24:36.352516 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-6d9d6b584d-r7bhq" Mar 13 14:24:36 crc kubenswrapper[4907]: I0313 14:24:36.396386 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-5bc894d9b-w5829" Mar 13 14:24:36 crc kubenswrapper[4907]: I0313 14:24:36.421719 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-684f77d66d-7rdzd" Mar 13 14:24:36 crc kubenswrapper[4907]: I0313 14:24:36.669108 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-sk2l9" Mar 13 14:24:36 crc kubenswrapper[4907]: I0313 14:24:36.839367 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-bbc5b68f9-cqmhw" Mar 13 14:24:36 crc kubenswrapper[4907]: I0313 14:24:36.869250 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-574d45c66c-b99f7" Mar 13 14:24:36 crc kubenswrapper[4907]: I0313 14:24:36.915969 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-7f9cc5dd44-sd6nr" Mar 13 14:24:36 crc kubenswrapper[4907]: I0313 14:24:36.969743 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-6854b8b9d9-dj95f" Mar 13 14:24:37 crc kubenswrapper[4907]: I0313 14:24:37.088899 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-vzpcd" Mar 13 14:24:37 crc kubenswrapper[4907]: I0313 14:24:37.098763 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-hjp4p" Mar 13 14:24:38 crc kubenswrapper[4907]: I0313 14:24:38.036349 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/64d55221-60c5-4521-ac84-beed4e9b3993-cert\") pod \"infra-operator-controller-manager-54dc5b8f8d-lp7kj\" (UID: \"64d55221-60c5-4521-ac84-beed4e9b3993\") " pod="openstack-operators/infra-operator-controller-manager-54dc5b8f8d-lp7kj" Mar 13 14:24:38 crc kubenswrapper[4907]: I0313 14:24:38.047612 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/64d55221-60c5-4521-ac84-beed4e9b3993-cert\") pod \"infra-operator-controller-manager-54dc5b8f8d-lp7kj\" (UID: \"64d55221-60c5-4521-ac84-beed4e9b3993\") " pod="openstack-operators/infra-operator-controller-manager-54dc5b8f8d-lp7kj" Mar 13 14:24:38 crc kubenswrapper[4907]: I0313 14:24:38.170773 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-54dc5b8f8d-lp7kj" Mar 13 14:24:38 crc kubenswrapper[4907]: I0313 14:24:38.340023 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f915558a-c0ab-4d8e-a427-e56a7588e382-cert\") pod \"openstack-baremetal-operator-controller-manager-6f7958d774rr6kf\" (UID: \"f915558a-c0ab-4d8e-a427-e56a7588e382\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6f7958d774rr6kf" Mar 13 14:24:38 crc kubenswrapper[4907]: I0313 14:24:38.348129 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f915558a-c0ab-4d8e-a427-e56a7588e382-cert\") pod \"openstack-baremetal-operator-controller-manager-6f7958d774rr6kf\" (UID: \"f915558a-c0ab-4d8e-a427-e56a7588e382\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6f7958d774rr6kf" Mar 13 14:24:38 crc kubenswrapper[4907]: I0313 14:24:38.403628 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-54dc5b8f8d-lp7kj"] Mar 13 14:24:38 crc kubenswrapper[4907]: I0313 14:24:38.501948 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-54dc5b8f8d-lp7kj" event={"ID":"64d55221-60c5-4521-ac84-beed4e9b3993","Type":"ContainerStarted","Data":"0be6213c7eab4cfb969c5684d05680002c055fb080ed20ba4cb83fbda2a9bdb9"} Mar 13 14:24:38 crc kubenswrapper[4907]: I0313 14:24:38.537394 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6f7958d774rr6kf" Mar 13 14:24:38 crc kubenswrapper[4907]: I0313 14:24:38.789276 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-6f7958d774rr6kf"] Mar 13 14:24:38 crc kubenswrapper[4907]: W0313 14:24:38.801963 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf915558a_c0ab_4d8e_a427_e56a7588e382.slice/crio-75c38ddcdad2111a6bf9a09fd763d9ba0db0b2f0bd3aeb1f22fdd968692242fe WatchSource:0}: Error finding container 75c38ddcdad2111a6bf9a09fd763d9ba0db0b2f0bd3aeb1f22fdd968692242fe: Status 404 returned error can't find the container with id 75c38ddcdad2111a6bf9a09fd763d9ba0db0b2f0bd3aeb1f22fdd968692242fe Mar 13 14:24:38 crc kubenswrapper[4907]: I0313 14:24:38.957461 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/9ac47edf-03ab-4e93-b238-2aa00765ab06-webhook-certs\") pod \"openstack-operator-controller-manager-6484b7b757-qldxp\" (UID: \"9ac47edf-03ab-4e93-b238-2aa00765ab06\") " pod="openstack-operators/openstack-operator-controller-manager-6484b7b757-qldxp" Mar 13 14:24:38 crc kubenswrapper[4907]: I0313 14:24:38.957497 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9ac47edf-03ab-4e93-b238-2aa00765ab06-metrics-certs\") pod \"openstack-operator-controller-manager-6484b7b757-qldxp\" (UID: \"9ac47edf-03ab-4e93-b238-2aa00765ab06\") " pod="openstack-operators/openstack-operator-controller-manager-6484b7b757-qldxp" Mar 13 14:24:38 crc kubenswrapper[4907]: I0313 14:24:38.964224 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/9ac47edf-03ab-4e93-b238-2aa00765ab06-webhook-certs\") pod \"openstack-operator-controller-manager-6484b7b757-qldxp\" (UID: \"9ac47edf-03ab-4e93-b238-2aa00765ab06\") " pod="openstack-operators/openstack-operator-controller-manager-6484b7b757-qldxp" Mar 13 14:24:38 crc kubenswrapper[4907]: I0313 14:24:38.964982 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9ac47edf-03ab-4e93-b238-2aa00765ab06-metrics-certs\") pod \"openstack-operator-controller-manager-6484b7b757-qldxp\" (UID: \"9ac47edf-03ab-4e93-b238-2aa00765ab06\") " pod="openstack-operators/openstack-operator-controller-manager-6484b7b757-qldxp" Mar 13 14:24:39 crc kubenswrapper[4907]: I0313 14:24:39.045369 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-6484b7b757-qldxp" Mar 13 14:24:39 crc kubenswrapper[4907]: I0313 14:24:39.490415 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-6484b7b757-qldxp"] Mar 13 14:24:39 crc kubenswrapper[4907]: I0313 14:24:39.508021 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6f7958d774rr6kf" event={"ID":"f915558a-c0ab-4d8e-a427-e56a7588e382","Type":"ContainerStarted","Data":"75c38ddcdad2111a6bf9a09fd763d9ba0db0b2f0bd3aeb1f22fdd968692242fe"} Mar 13 14:24:41 crc kubenswrapper[4907]: W0313 14:24:41.691823 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9ac47edf_03ab_4e93_b238_2aa00765ab06.slice/crio-4dbfcfd327b82f91e60a91bb090ca1b38f55a0c6e523350af485793c5f2a5ed1 WatchSource:0}: Error finding container 4dbfcfd327b82f91e60a91bb090ca1b38f55a0c6e523350af485793c5f2a5ed1: Status 404 returned error can't find the container with id 4dbfcfd327b82f91e60a91bb090ca1b38f55a0c6e523350af485793c5f2a5ed1 Mar 13 14:24:42 crc kubenswrapper[4907]: I0313 14:24:42.530562 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-5b6b6b4c9f-5m2vw" event={"ID":"e8134e91-4b9d-4e3c-9aec-eba18acabdce","Type":"ContainerStarted","Data":"bedb2a7348ed35c5b50d4408da49c01fe6c8c250dacc06b7cb57beaed77557f4"} Mar 13 14:24:42 crc kubenswrapper[4907]: I0313 14:24:42.531060 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-5b6b6b4c9f-5m2vw" Mar 13 14:24:42 crc kubenswrapper[4907]: I0313 14:24:42.531852 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-7f84474648-ql4kd" event={"ID":"4ab8f569-de56-4fea-9be0-07908063522a","Type":"ContainerStarted","Data":"7cfcdad4ee4313cbae8d6e85cbba93863fe24e32bb46b602f6c5107a517a932b"} Mar 13 14:24:42 crc kubenswrapper[4907]: I0313 14:24:42.532107 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-7f84474648-ql4kd" Mar 13 14:24:42 crc kubenswrapper[4907]: I0313 14:24:42.532906 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-6484b7b757-qldxp" event={"ID":"9ac47edf-03ab-4e93-b238-2aa00765ab06","Type":"ContainerStarted","Data":"e00335a5fbe716a9217486cf5b472f2201008b6b99dbf7c88c2718632c7cd36a"} Mar 13 14:24:42 crc kubenswrapper[4907]: I0313 14:24:42.532954 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-6484b7b757-qldxp" event={"ID":"9ac47edf-03ab-4e93-b238-2aa00765ab06","Type":"ContainerStarted","Data":"4dbfcfd327b82f91e60a91bb090ca1b38f55a0c6e523350af485793c5f2a5ed1"} Mar 13 14:24:42 crc kubenswrapper[4907]: I0313 14:24:42.533016 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-6484b7b757-qldxp" Mar 13 14:24:42 crc kubenswrapper[4907]: I0313 14:24:42.534248 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-776c5696bf-gvgdl" event={"ID":"29201cb7-c056-4e4e-b0da-cc775c7ef11f","Type":"ContainerStarted","Data":"74583fc9170a3e56782dd10b6fbd0dbdb6d034e5d6a17b6fb0a275c783ef4c8e"} Mar 13 14:24:42 crc kubenswrapper[4907]: I0313 14:24:42.534427 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-776c5696bf-gvgdl" Mar 13 14:24:42 crc kubenswrapper[4907]: I0313 14:24:42.536370 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-57b484b4df-m8j7m" event={"ID":"c55515f1-509e-4a4e-813d-fbc466fca489","Type":"ContainerStarted","Data":"d2c7db2e6567e3827e93800958d2e7f4cf9db08c46c28bc01c5ce56ac4b44b3d"} Mar 13 14:24:42 crc kubenswrapper[4907]: I0313 14:24:42.536654 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-57b484b4df-m8j7m" Mar 13 14:24:42 crc kubenswrapper[4907]: I0313 14:24:42.557606 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-5b6b6b4c9f-5m2vw" podStartSLOduration=2.423232236 podStartE2EDuration="36.557589066s" podCreationTimestamp="2026-03-13 14:24:06 +0000 UTC" firstStartedPulling="2026-03-13 14:24:07.664877628 +0000 UTC m=+1146.564665317" lastFinishedPulling="2026-03-13 14:24:41.799234468 +0000 UTC m=+1180.699022147" observedRunningTime="2026-03-13 14:24:42.554282245 +0000 UTC m=+1181.454069934" watchObservedRunningTime="2026-03-13 14:24:42.557589066 +0000 UTC m=+1181.457376755" Mar 13 14:24:42 crc kubenswrapper[4907]: I0313 14:24:42.579355 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-57b484b4df-m8j7m" podStartSLOduration=2.713793994 podStartE2EDuration="36.579340353s" podCreationTimestamp="2026-03-13 14:24:06 +0000 UTC" firstStartedPulling="2026-03-13 14:24:07.860966812 +0000 UTC m=+1146.760754501" lastFinishedPulling="2026-03-13 14:24:41.726513171 +0000 UTC m=+1180.626300860" observedRunningTime="2026-03-13 14:24:42.574960261 +0000 UTC m=+1181.474747950" watchObservedRunningTime="2026-03-13 14:24:42.579340353 +0000 UTC m=+1181.479128032" Mar 13 14:24:42 crc kubenswrapper[4907]: I0313 14:24:42.603872 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-7f84474648-ql4kd" podStartSLOduration=2.396819621 podStartE2EDuration="36.603857585s" podCreationTimestamp="2026-03-13 14:24:06 +0000 UTC" firstStartedPulling="2026-03-13 14:24:07.693148966 +0000 UTC m=+1146.592936655" lastFinishedPulling="2026-03-13 14:24:41.90018693 +0000 UTC m=+1180.799974619" observedRunningTime="2026-03-13 14:24:42.600095731 +0000 UTC m=+1181.499883420" watchObservedRunningTime="2026-03-13 14:24:42.603857585 +0000 UTC m=+1181.503645274" Mar 13 14:24:42 crc kubenswrapper[4907]: I0313 14:24:42.616700 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-776c5696bf-gvgdl" podStartSLOduration=2.74386588 podStartE2EDuration="36.616681023s" podCreationTimestamp="2026-03-13 14:24:06 +0000 UTC" firstStartedPulling="2026-03-13 14:24:07.859420278 +0000 UTC m=+1146.759207967" lastFinishedPulling="2026-03-13 14:24:41.732235411 +0000 UTC m=+1180.632023110" observedRunningTime="2026-03-13 14:24:42.614834512 +0000 UTC m=+1181.514622201" watchObservedRunningTime="2026-03-13 14:24:42.616681023 +0000 UTC m=+1181.516468712" Mar 13 14:24:42 crc kubenswrapper[4907]: I0313 14:24:42.654988 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-6484b7b757-qldxp" podStartSLOduration=36.65497235 podStartE2EDuration="36.65497235s" podCreationTimestamp="2026-03-13 14:24:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:24:42.653872489 +0000 UTC m=+1181.553660178" watchObservedRunningTime="2026-03-13 14:24:42.65497235 +0000 UTC m=+1181.554760039" Mar 13 14:24:44 crc kubenswrapper[4907]: I0313 14:24:44.562120 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6f7958d774rr6kf" event={"ID":"f915558a-c0ab-4d8e-a427-e56a7588e382","Type":"ContainerStarted","Data":"c490d1d54cec4cc6b65cac26ba1a01b62916a5722f82287e8f4992cf2722d681"} Mar 13 14:24:44 crc kubenswrapper[4907]: I0313 14:24:44.562495 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6f7958d774rr6kf" Mar 13 14:24:44 crc kubenswrapper[4907]: I0313 14:24:44.600796 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6f7958d774rr6kf" podStartSLOduration=33.758992032 podStartE2EDuration="38.600777705s" podCreationTimestamp="2026-03-13 14:24:06 +0000 UTC" firstStartedPulling="2026-03-13 14:24:38.804011993 +0000 UTC m=+1177.703799692" lastFinishedPulling="2026-03-13 14:24:43.645797676 +0000 UTC m=+1182.545585365" observedRunningTime="2026-03-13 14:24:44.593530922 +0000 UTC m=+1183.493318621" watchObservedRunningTime="2026-03-13 14:24:44.600777705 +0000 UTC m=+1183.500565394" Mar 13 14:24:45 crc kubenswrapper[4907]: I0313 14:24:45.571921 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-54dc5b8f8d-lp7kj" event={"ID":"64d55221-60c5-4521-ac84-beed4e9b3993","Type":"ContainerStarted","Data":"1a0c55227e3c894cc4a86677ca0790696d229c5bb1f62e53708a5514a6737c05"} Mar 13 14:24:45 crc kubenswrapper[4907]: I0313 14:24:45.572316 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-54dc5b8f8d-lp7kj" Mar 13 14:24:45 crc kubenswrapper[4907]: I0313 14:24:45.588355 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-54dc5b8f8d-lp7kj" podStartSLOduration=34.18932223 podStartE2EDuration="40.58832268s" podCreationTimestamp="2026-03-13 14:24:05 +0000 UTC" firstStartedPulling="2026-03-13 14:24:38.412107225 +0000 UTC m=+1177.311894914" lastFinishedPulling="2026-03-13 14:24:44.811107675 +0000 UTC m=+1183.710895364" observedRunningTime="2026-03-13 14:24:45.587506327 +0000 UTC m=+1184.487294016" watchObservedRunningTime="2026-03-13 14:24:45.58832268 +0000 UTC m=+1184.488110369" Mar 13 14:24:48 crc kubenswrapper[4907]: I0313 14:24:48.041518 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 14:24:48 crc kubenswrapper[4907]: I0313 14:24:48.041599 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 14:24:48 crc kubenswrapper[4907]: I0313 14:24:48.545389 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6f7958d774rr6kf" Mar 13 14:24:49 crc kubenswrapper[4907]: I0313 14:24:49.050932 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-6484b7b757-qldxp" Mar 13 14:24:56 crc kubenswrapper[4907]: I0313 14:24:56.597400 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-57b484b4df-m8j7m" Mar 13 14:24:56 crc kubenswrapper[4907]: I0313 14:24:56.643240 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-776c5696bf-gvgdl" Mar 13 14:24:56 crc kubenswrapper[4907]: I0313 14:24:56.643734 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-5b6b6b4c9f-5m2vw" Mar 13 14:24:56 crc kubenswrapper[4907]: I0313 14:24:56.681757 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-7f84474648-ql4kd" Mar 13 14:24:58 crc kubenswrapper[4907]: I0313 14:24:58.176404 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-54dc5b8f8d-lp7kj" Mar 13 14:25:02 crc kubenswrapper[4907]: I0313 14:25:02.688567 4907 scope.go:117] "RemoveContainer" containerID="7377ca75e3097f224f1949f7536a8f7a1c19b44224ea656c9d4dd6e3d82e1652" Mar 13 14:25:13 crc kubenswrapper[4907]: I0313 14:25:13.888627 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5448ff6dc7-pgq4q"] Mar 13 14:25:13 crc kubenswrapper[4907]: I0313 14:25:13.891283 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5448ff6dc7-pgq4q" Mar 13 14:25:13 crc kubenswrapper[4907]: I0313 14:25:13.894189 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Mar 13 14:25:13 crc kubenswrapper[4907]: I0313 14:25:13.894541 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Mar 13 14:25:13 crc kubenswrapper[4907]: I0313 14:25:13.894536 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Mar 13 14:25:13 crc kubenswrapper[4907]: I0313 14:25:13.894615 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-wgpxh" Mar 13 14:25:13 crc kubenswrapper[4907]: I0313 14:25:13.895769 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/152380e7-1c7a-4c89-a14e-dd179009f6af-config\") pod \"dnsmasq-dns-5448ff6dc7-pgq4q\" (UID: \"152380e7-1c7a-4c89-a14e-dd179009f6af\") " pod="openstack/dnsmasq-dns-5448ff6dc7-pgq4q" Mar 13 14:25:13 crc kubenswrapper[4907]: I0313 14:25:13.895835 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qxpcr\" (UniqueName: \"kubernetes.io/projected/152380e7-1c7a-4c89-a14e-dd179009f6af-kube-api-access-qxpcr\") pod \"dnsmasq-dns-5448ff6dc7-pgq4q\" (UID: \"152380e7-1c7a-4c89-a14e-dd179009f6af\") " pod="openstack/dnsmasq-dns-5448ff6dc7-pgq4q" Mar 13 14:25:13 crc kubenswrapper[4907]: I0313 14:25:13.906170 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5448ff6dc7-pgq4q"] Mar 13 14:25:13 crc kubenswrapper[4907]: I0313 14:25:13.961226 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-64696987c5-gdk7n"] Mar 13 14:25:13 crc kubenswrapper[4907]: I0313 14:25:13.962619 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64696987c5-gdk7n" Mar 13 14:25:13 crc kubenswrapper[4907]: I0313 14:25:13.968831 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-64696987c5-gdk7n"] Mar 13 14:25:13 crc kubenswrapper[4907]: I0313 14:25:13.969734 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Mar 13 14:25:14 crc kubenswrapper[4907]: I0313 14:25:14.005693 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/152380e7-1c7a-4c89-a14e-dd179009f6af-config\") pod \"dnsmasq-dns-5448ff6dc7-pgq4q\" (UID: \"152380e7-1c7a-4c89-a14e-dd179009f6af\") " pod="openstack/dnsmasq-dns-5448ff6dc7-pgq4q" Mar 13 14:25:14 crc kubenswrapper[4907]: I0313 14:25:14.005746 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qxpcr\" (UniqueName: \"kubernetes.io/projected/152380e7-1c7a-4c89-a14e-dd179009f6af-kube-api-access-qxpcr\") pod \"dnsmasq-dns-5448ff6dc7-pgq4q\" (UID: \"152380e7-1c7a-4c89-a14e-dd179009f6af\") " pod="openstack/dnsmasq-dns-5448ff6dc7-pgq4q" Mar 13 14:25:14 crc kubenswrapper[4907]: I0313 14:25:14.006806 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/152380e7-1c7a-4c89-a14e-dd179009f6af-config\") pod \"dnsmasq-dns-5448ff6dc7-pgq4q\" (UID: \"152380e7-1c7a-4c89-a14e-dd179009f6af\") " pod="openstack/dnsmasq-dns-5448ff6dc7-pgq4q" Mar 13 14:25:14 crc kubenswrapper[4907]: I0313 14:25:14.046553 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qxpcr\" (UniqueName: \"kubernetes.io/projected/152380e7-1c7a-4c89-a14e-dd179009f6af-kube-api-access-qxpcr\") pod \"dnsmasq-dns-5448ff6dc7-pgq4q\" (UID: \"152380e7-1c7a-4c89-a14e-dd179009f6af\") " pod="openstack/dnsmasq-dns-5448ff6dc7-pgq4q" Mar 13 14:25:14 crc kubenswrapper[4907]: I0313 14:25:14.107497 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-br82q\" (UniqueName: \"kubernetes.io/projected/874829e6-8bd5-4040-a493-317a2af4fa35-kube-api-access-br82q\") pod \"dnsmasq-dns-64696987c5-gdk7n\" (UID: \"874829e6-8bd5-4040-a493-317a2af4fa35\") " pod="openstack/dnsmasq-dns-64696987c5-gdk7n" Mar 13 14:25:14 crc kubenswrapper[4907]: I0313 14:25:14.107635 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/874829e6-8bd5-4040-a493-317a2af4fa35-config\") pod \"dnsmasq-dns-64696987c5-gdk7n\" (UID: \"874829e6-8bd5-4040-a493-317a2af4fa35\") " pod="openstack/dnsmasq-dns-64696987c5-gdk7n" Mar 13 14:25:14 crc kubenswrapper[4907]: I0313 14:25:14.107707 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/874829e6-8bd5-4040-a493-317a2af4fa35-dns-svc\") pod \"dnsmasq-dns-64696987c5-gdk7n\" (UID: \"874829e6-8bd5-4040-a493-317a2af4fa35\") " pod="openstack/dnsmasq-dns-64696987c5-gdk7n" Mar 13 14:25:14 crc kubenswrapper[4907]: I0313 14:25:14.207953 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5448ff6dc7-pgq4q" Mar 13 14:25:14 crc kubenswrapper[4907]: I0313 14:25:14.208603 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/874829e6-8bd5-4040-a493-317a2af4fa35-config\") pod \"dnsmasq-dns-64696987c5-gdk7n\" (UID: \"874829e6-8bd5-4040-a493-317a2af4fa35\") " pod="openstack/dnsmasq-dns-64696987c5-gdk7n" Mar 13 14:25:14 crc kubenswrapper[4907]: I0313 14:25:14.208669 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/874829e6-8bd5-4040-a493-317a2af4fa35-dns-svc\") pod \"dnsmasq-dns-64696987c5-gdk7n\" (UID: \"874829e6-8bd5-4040-a493-317a2af4fa35\") " pod="openstack/dnsmasq-dns-64696987c5-gdk7n" Mar 13 14:25:14 crc kubenswrapper[4907]: I0313 14:25:14.208704 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-br82q\" (UniqueName: \"kubernetes.io/projected/874829e6-8bd5-4040-a493-317a2af4fa35-kube-api-access-br82q\") pod \"dnsmasq-dns-64696987c5-gdk7n\" (UID: \"874829e6-8bd5-4040-a493-317a2af4fa35\") " pod="openstack/dnsmasq-dns-64696987c5-gdk7n" Mar 13 14:25:14 crc kubenswrapper[4907]: I0313 14:25:14.209434 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/874829e6-8bd5-4040-a493-317a2af4fa35-config\") pod \"dnsmasq-dns-64696987c5-gdk7n\" (UID: \"874829e6-8bd5-4040-a493-317a2af4fa35\") " pod="openstack/dnsmasq-dns-64696987c5-gdk7n" Mar 13 14:25:14 crc kubenswrapper[4907]: I0313 14:25:14.209535 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/874829e6-8bd5-4040-a493-317a2af4fa35-dns-svc\") pod \"dnsmasq-dns-64696987c5-gdk7n\" (UID: \"874829e6-8bd5-4040-a493-317a2af4fa35\") " pod="openstack/dnsmasq-dns-64696987c5-gdk7n" Mar 13 14:25:14 crc kubenswrapper[4907]: I0313 14:25:14.243773 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-br82q\" (UniqueName: \"kubernetes.io/projected/874829e6-8bd5-4040-a493-317a2af4fa35-kube-api-access-br82q\") pod \"dnsmasq-dns-64696987c5-gdk7n\" (UID: \"874829e6-8bd5-4040-a493-317a2af4fa35\") " pod="openstack/dnsmasq-dns-64696987c5-gdk7n" Mar 13 14:25:14 crc kubenswrapper[4907]: I0313 14:25:14.276575 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64696987c5-gdk7n" Mar 13 14:25:14 crc kubenswrapper[4907]: I0313 14:25:14.588303 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-64696987c5-gdk7n"] Mar 13 14:25:14 crc kubenswrapper[4907]: W0313 14:25:14.691033 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod152380e7_1c7a_4c89_a14e_dd179009f6af.slice/crio-5fce8f6cb27b89ad9b7731f40d4a1c6f21ae57078b5536d25290f96168055e8b WatchSource:0}: Error finding container 5fce8f6cb27b89ad9b7731f40d4a1c6f21ae57078b5536d25290f96168055e8b: Status 404 returned error can't find the container with id 5fce8f6cb27b89ad9b7731f40d4a1c6f21ae57078b5536d25290f96168055e8b Mar 13 14:25:14 crc kubenswrapper[4907]: I0313 14:25:14.692828 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5448ff6dc7-pgq4q"] Mar 13 14:25:14 crc kubenswrapper[4907]: I0313 14:25:14.789421 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64696987c5-gdk7n" event={"ID":"874829e6-8bd5-4040-a493-317a2af4fa35","Type":"ContainerStarted","Data":"d7f24309913163b7f0ac4df94b2d551597928123506fa9a0c942b473d75c1940"} Mar 13 14:25:14 crc kubenswrapper[4907]: I0313 14:25:14.790512 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5448ff6dc7-pgq4q" event={"ID":"152380e7-1c7a-4c89-a14e-dd179009f6af","Type":"ContainerStarted","Data":"5fce8f6cb27b89ad9b7731f40d4a1c6f21ae57078b5536d25290f96168055e8b"} Mar 13 14:25:16 crc kubenswrapper[4907]: I0313 14:25:16.735637 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5448ff6dc7-pgq4q"] Mar 13 14:25:16 crc kubenswrapper[4907]: I0313 14:25:16.776671 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-658f55c9f5-dp8jf"] Mar 13 14:25:16 crc kubenswrapper[4907]: I0313 14:25:16.778289 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-658f55c9f5-dp8jf" Mar 13 14:25:16 crc kubenswrapper[4907]: I0313 14:25:16.793016 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-658f55c9f5-dp8jf"] Mar 13 14:25:16 crc kubenswrapper[4907]: I0313 14:25:16.960829 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b921f7f7-cd29-4ac8-82e5-c69a8f9e2bc0-config\") pod \"dnsmasq-dns-658f55c9f5-dp8jf\" (UID: \"b921f7f7-cd29-4ac8-82e5-c69a8f9e2bc0\") " pod="openstack/dnsmasq-dns-658f55c9f5-dp8jf" Mar 13 14:25:16 crc kubenswrapper[4907]: I0313 14:25:16.960971 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-684gn\" (UniqueName: \"kubernetes.io/projected/b921f7f7-cd29-4ac8-82e5-c69a8f9e2bc0-kube-api-access-684gn\") pod \"dnsmasq-dns-658f55c9f5-dp8jf\" (UID: \"b921f7f7-cd29-4ac8-82e5-c69a8f9e2bc0\") " pod="openstack/dnsmasq-dns-658f55c9f5-dp8jf" Mar 13 14:25:16 crc kubenswrapper[4907]: I0313 14:25:16.961098 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b921f7f7-cd29-4ac8-82e5-c69a8f9e2bc0-dns-svc\") pod \"dnsmasq-dns-658f55c9f5-dp8jf\" (UID: \"b921f7f7-cd29-4ac8-82e5-c69a8f9e2bc0\") " pod="openstack/dnsmasq-dns-658f55c9f5-dp8jf" Mar 13 14:25:17 crc kubenswrapper[4907]: I0313 14:25:17.062250 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-684gn\" (UniqueName: \"kubernetes.io/projected/b921f7f7-cd29-4ac8-82e5-c69a8f9e2bc0-kube-api-access-684gn\") pod \"dnsmasq-dns-658f55c9f5-dp8jf\" (UID: \"b921f7f7-cd29-4ac8-82e5-c69a8f9e2bc0\") " pod="openstack/dnsmasq-dns-658f55c9f5-dp8jf" Mar 13 14:25:17 crc kubenswrapper[4907]: I0313 14:25:17.062399 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b921f7f7-cd29-4ac8-82e5-c69a8f9e2bc0-dns-svc\") pod \"dnsmasq-dns-658f55c9f5-dp8jf\" (UID: \"b921f7f7-cd29-4ac8-82e5-c69a8f9e2bc0\") " pod="openstack/dnsmasq-dns-658f55c9f5-dp8jf" Mar 13 14:25:17 crc kubenswrapper[4907]: I0313 14:25:17.062430 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b921f7f7-cd29-4ac8-82e5-c69a8f9e2bc0-config\") pod \"dnsmasq-dns-658f55c9f5-dp8jf\" (UID: \"b921f7f7-cd29-4ac8-82e5-c69a8f9e2bc0\") " pod="openstack/dnsmasq-dns-658f55c9f5-dp8jf" Mar 13 14:25:17 crc kubenswrapper[4907]: I0313 14:25:17.063542 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b921f7f7-cd29-4ac8-82e5-c69a8f9e2bc0-config\") pod \"dnsmasq-dns-658f55c9f5-dp8jf\" (UID: \"b921f7f7-cd29-4ac8-82e5-c69a8f9e2bc0\") " pod="openstack/dnsmasq-dns-658f55c9f5-dp8jf" Mar 13 14:25:17 crc kubenswrapper[4907]: I0313 14:25:17.064069 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b921f7f7-cd29-4ac8-82e5-c69a8f9e2bc0-dns-svc\") pod \"dnsmasq-dns-658f55c9f5-dp8jf\" (UID: \"b921f7f7-cd29-4ac8-82e5-c69a8f9e2bc0\") " pod="openstack/dnsmasq-dns-658f55c9f5-dp8jf" Mar 13 14:25:17 crc kubenswrapper[4907]: I0313 14:25:17.105108 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-684gn\" (UniqueName: \"kubernetes.io/projected/b921f7f7-cd29-4ac8-82e5-c69a8f9e2bc0-kube-api-access-684gn\") pod \"dnsmasq-dns-658f55c9f5-dp8jf\" (UID: \"b921f7f7-cd29-4ac8-82e5-c69a8f9e2bc0\") " pod="openstack/dnsmasq-dns-658f55c9f5-dp8jf" Mar 13 14:25:17 crc kubenswrapper[4907]: I0313 14:25:17.108468 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-658f55c9f5-dp8jf" Mar 13 14:25:17 crc kubenswrapper[4907]: I0313 14:25:17.180357 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-64696987c5-gdk7n"] Mar 13 14:25:17 crc kubenswrapper[4907]: I0313 14:25:17.227460 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-54b5dffb47-wttzd"] Mar 13 14:25:17 crc kubenswrapper[4907]: I0313 14:25:17.228532 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54b5dffb47-wttzd" Mar 13 14:25:17 crc kubenswrapper[4907]: I0313 14:25:17.245922 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-54b5dffb47-wttzd"] Mar 13 14:25:17 crc kubenswrapper[4907]: I0313 14:25:17.370933 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/779b1958-f7ed-4c53-a4ff-a6c4a803bebf-dns-svc\") pod \"dnsmasq-dns-54b5dffb47-wttzd\" (UID: \"779b1958-f7ed-4c53-a4ff-a6c4a803bebf\") " pod="openstack/dnsmasq-dns-54b5dffb47-wttzd" Mar 13 14:25:17 crc kubenswrapper[4907]: I0313 14:25:17.370989 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/779b1958-f7ed-4c53-a4ff-a6c4a803bebf-config\") pod \"dnsmasq-dns-54b5dffb47-wttzd\" (UID: \"779b1958-f7ed-4c53-a4ff-a6c4a803bebf\") " pod="openstack/dnsmasq-dns-54b5dffb47-wttzd" Mar 13 14:25:17 crc kubenswrapper[4907]: I0313 14:25:17.371035 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bmczb\" (UniqueName: \"kubernetes.io/projected/779b1958-f7ed-4c53-a4ff-a6c4a803bebf-kube-api-access-bmczb\") pod \"dnsmasq-dns-54b5dffb47-wttzd\" (UID: \"779b1958-f7ed-4c53-a4ff-a6c4a803bebf\") " pod="openstack/dnsmasq-dns-54b5dffb47-wttzd" Mar 13 14:25:17 crc kubenswrapper[4907]: I0313 14:25:17.478372 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/779b1958-f7ed-4c53-a4ff-a6c4a803bebf-dns-svc\") pod \"dnsmasq-dns-54b5dffb47-wttzd\" (UID: \"779b1958-f7ed-4c53-a4ff-a6c4a803bebf\") " pod="openstack/dnsmasq-dns-54b5dffb47-wttzd" Mar 13 14:25:17 crc kubenswrapper[4907]: I0313 14:25:17.478431 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/779b1958-f7ed-4c53-a4ff-a6c4a803bebf-config\") pod \"dnsmasq-dns-54b5dffb47-wttzd\" (UID: \"779b1958-f7ed-4c53-a4ff-a6c4a803bebf\") " pod="openstack/dnsmasq-dns-54b5dffb47-wttzd" Mar 13 14:25:17 crc kubenswrapper[4907]: I0313 14:25:17.478466 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bmczb\" (UniqueName: \"kubernetes.io/projected/779b1958-f7ed-4c53-a4ff-a6c4a803bebf-kube-api-access-bmczb\") pod \"dnsmasq-dns-54b5dffb47-wttzd\" (UID: \"779b1958-f7ed-4c53-a4ff-a6c4a803bebf\") " pod="openstack/dnsmasq-dns-54b5dffb47-wttzd" Mar 13 14:25:17 crc kubenswrapper[4907]: I0313 14:25:17.479951 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/779b1958-f7ed-4c53-a4ff-a6c4a803bebf-dns-svc\") pod \"dnsmasq-dns-54b5dffb47-wttzd\" (UID: \"779b1958-f7ed-4c53-a4ff-a6c4a803bebf\") " pod="openstack/dnsmasq-dns-54b5dffb47-wttzd" Mar 13 14:25:17 crc kubenswrapper[4907]: I0313 14:25:17.480496 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/779b1958-f7ed-4c53-a4ff-a6c4a803bebf-config\") pod \"dnsmasq-dns-54b5dffb47-wttzd\" (UID: \"779b1958-f7ed-4c53-a4ff-a6c4a803bebf\") " pod="openstack/dnsmasq-dns-54b5dffb47-wttzd" Mar 13 14:25:17 crc kubenswrapper[4907]: I0313 14:25:17.498903 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bmczb\" (UniqueName: \"kubernetes.io/projected/779b1958-f7ed-4c53-a4ff-a6c4a803bebf-kube-api-access-bmczb\") pod \"dnsmasq-dns-54b5dffb47-wttzd\" (UID: \"779b1958-f7ed-4c53-a4ff-a6c4a803bebf\") " pod="openstack/dnsmasq-dns-54b5dffb47-wttzd" Mar 13 14:25:17 crc kubenswrapper[4907]: I0313 14:25:17.551267 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54b5dffb47-wttzd" Mar 13 14:25:17 crc kubenswrapper[4907]: I0313 14:25:17.652772 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-658f55c9f5-dp8jf"] Mar 13 14:25:17 crc kubenswrapper[4907]: I0313 14:25:17.825476 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-658f55c9f5-dp8jf" event={"ID":"b921f7f7-cd29-4ac8-82e5-c69a8f9e2bc0","Type":"ContainerStarted","Data":"00577cc935dfba6c223104c8f045d104f6284aba771cb8fe435ed068373d1547"} Mar 13 14:25:17 crc kubenswrapper[4907]: I0313 14:25:17.948232 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 13 14:25:17 crc kubenswrapper[4907]: I0313 14:25:17.949281 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Mar 13 14:25:17 crc kubenswrapper[4907]: I0313 14:25:17.954740 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Mar 13 14:25:17 crc kubenswrapper[4907]: I0313 14:25:17.955021 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Mar 13 14:25:17 crc kubenswrapper[4907]: I0313 14:25:17.955168 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Mar 13 14:25:17 crc kubenswrapper[4907]: I0313 14:25:17.955326 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-swzzq" Mar 13 14:25:17 crc kubenswrapper[4907]: I0313 14:25:17.956665 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Mar 13 14:25:17 crc kubenswrapper[4907]: I0313 14:25:17.958076 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Mar 13 14:25:17 crc kubenswrapper[4907]: I0313 14:25:17.958250 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Mar 13 14:25:17 crc kubenswrapper[4907]: I0313 14:25:17.960402 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.041651 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.041703 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.041749 4907 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.042304 4907 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ac423129eeb1fa8eb8df325acb5ccb91e119992cd551dd9d8c11d304cb95b736"} pod="openshift-machine-config-operator/machine-config-daemon-hm56j" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.042356 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" containerID="cri-o://ac423129eeb1fa8eb8df325acb5ccb91e119992cd551dd9d8c11d304cb95b736" gracePeriod=600 Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.087613 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/286c82e8-b74a-49d7-a355-ac074aace10e-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"286c82e8-b74a-49d7-a355-ac074aace10e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.087659 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/286c82e8-b74a-49d7-a355-ac074aace10e-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"286c82e8-b74a-49d7-a355-ac074aace10e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.087680 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w8snq\" (UniqueName: \"kubernetes.io/projected/286c82e8-b74a-49d7-a355-ac074aace10e-kube-api-access-w8snq\") pod \"rabbitmq-cell1-server-0\" (UID: \"286c82e8-b74a-49d7-a355-ac074aace10e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.087710 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/286c82e8-b74a-49d7-a355-ac074aace10e-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"286c82e8-b74a-49d7-a355-ac074aace10e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.087733 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/286c82e8-b74a-49d7-a355-ac074aace10e-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"286c82e8-b74a-49d7-a355-ac074aace10e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.087765 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/286c82e8-b74a-49d7-a355-ac074aace10e-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"286c82e8-b74a-49d7-a355-ac074aace10e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.087782 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/286c82e8-b74a-49d7-a355-ac074aace10e-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"286c82e8-b74a-49d7-a355-ac074aace10e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.087807 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/286c82e8-b74a-49d7-a355-ac074aace10e-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"286c82e8-b74a-49d7-a355-ac074aace10e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.087825 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/286c82e8-b74a-49d7-a355-ac074aace10e-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"286c82e8-b74a-49d7-a355-ac074aace10e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.087859 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/286c82e8-b74a-49d7-a355-ac074aace10e-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"286c82e8-b74a-49d7-a355-ac074aace10e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.087894 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"286c82e8-b74a-49d7-a355-ac074aace10e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.091740 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-54b5dffb47-wttzd"] Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.188858 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/286c82e8-b74a-49d7-a355-ac074aace10e-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"286c82e8-b74a-49d7-a355-ac074aace10e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.189148 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/286c82e8-b74a-49d7-a355-ac074aace10e-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"286c82e8-b74a-49d7-a355-ac074aace10e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.189180 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/286c82e8-b74a-49d7-a355-ac074aace10e-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"286c82e8-b74a-49d7-a355-ac074aace10e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.189201 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/286c82e8-b74a-49d7-a355-ac074aace10e-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"286c82e8-b74a-49d7-a355-ac074aace10e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.189219 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/286c82e8-b74a-49d7-a355-ac074aace10e-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"286c82e8-b74a-49d7-a355-ac074aace10e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.189237 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/286c82e8-b74a-49d7-a355-ac074aace10e-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"286c82e8-b74a-49d7-a355-ac074aace10e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.189286 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/286c82e8-b74a-49d7-a355-ac074aace10e-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"286c82e8-b74a-49d7-a355-ac074aace10e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.189304 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"286c82e8-b74a-49d7-a355-ac074aace10e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.189362 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/286c82e8-b74a-49d7-a355-ac074aace10e-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"286c82e8-b74a-49d7-a355-ac074aace10e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.189379 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/286c82e8-b74a-49d7-a355-ac074aace10e-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"286c82e8-b74a-49d7-a355-ac074aace10e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.189395 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w8snq\" (UniqueName: \"kubernetes.io/projected/286c82e8-b74a-49d7-a355-ac074aace10e-kube-api-access-w8snq\") pod \"rabbitmq-cell1-server-0\" (UID: \"286c82e8-b74a-49d7-a355-ac074aace10e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.190180 4907 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"286c82e8-b74a-49d7-a355-ac074aace10e\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/rabbitmq-cell1-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.190655 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/286c82e8-b74a-49d7-a355-ac074aace10e-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"286c82e8-b74a-49d7-a355-ac074aace10e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.190693 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/286c82e8-b74a-49d7-a355-ac074aace10e-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"286c82e8-b74a-49d7-a355-ac074aace10e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.191204 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/286c82e8-b74a-49d7-a355-ac074aace10e-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"286c82e8-b74a-49d7-a355-ac074aace10e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.191689 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/286c82e8-b74a-49d7-a355-ac074aace10e-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"286c82e8-b74a-49d7-a355-ac074aace10e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.191804 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/286c82e8-b74a-49d7-a355-ac074aace10e-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"286c82e8-b74a-49d7-a355-ac074aace10e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.196529 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/286c82e8-b74a-49d7-a355-ac074aace10e-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"286c82e8-b74a-49d7-a355-ac074aace10e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.209626 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w8snq\" (UniqueName: \"kubernetes.io/projected/286c82e8-b74a-49d7-a355-ac074aace10e-kube-api-access-w8snq\") pod \"rabbitmq-cell1-server-0\" (UID: \"286c82e8-b74a-49d7-a355-ac074aace10e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.209790 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/286c82e8-b74a-49d7-a355-ac074aace10e-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"286c82e8-b74a-49d7-a355-ac074aace10e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.212846 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/286c82e8-b74a-49d7-a355-ac074aace10e-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"286c82e8-b74a-49d7-a355-ac074aace10e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.213440 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/286c82e8-b74a-49d7-a355-ac074aace10e-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"286c82e8-b74a-49d7-a355-ac074aace10e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.226498 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"286c82e8-b74a-49d7-a355-ac074aace10e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.274188 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.347579 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.354904 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.360206 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.360428 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-msttx" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.360615 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.360765 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.361109 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.361252 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.361397 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.370623 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.496039 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/e3c34454-315e-4821-ab25-b0f331a0d521-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"e3c34454-315e-4821-ab25-b0f331a0d521\") " pod="openstack/rabbitmq-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.496100 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e3c34454-315e-4821-ab25-b0f331a0d521-config-data\") pod \"rabbitmq-server-0\" (UID: \"e3c34454-315e-4821-ab25-b0f331a0d521\") " pod="openstack/rabbitmq-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.496133 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/e3c34454-315e-4821-ab25-b0f331a0d521-server-conf\") pod \"rabbitmq-server-0\" (UID: \"e3c34454-315e-4821-ab25-b0f331a0d521\") " pod="openstack/rabbitmq-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.496171 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/e3c34454-315e-4821-ab25-b0f331a0d521-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"e3c34454-315e-4821-ab25-b0f331a0d521\") " pod="openstack/rabbitmq-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.496222 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/e3c34454-315e-4821-ab25-b0f331a0d521-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"e3c34454-315e-4821-ab25-b0f331a0d521\") " pod="openstack/rabbitmq-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.496250 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/e3c34454-315e-4821-ab25-b0f331a0d521-pod-info\") pod \"rabbitmq-server-0\" (UID: \"e3c34454-315e-4821-ab25-b0f331a0d521\") " pod="openstack/rabbitmq-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.496283 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-server-0\" (UID: \"e3c34454-315e-4821-ab25-b0f331a0d521\") " pod="openstack/rabbitmq-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.496310 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/e3c34454-315e-4821-ab25-b0f331a0d521-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"e3c34454-315e-4821-ab25-b0f331a0d521\") " pod="openstack/rabbitmq-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.496341 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/e3c34454-315e-4821-ab25-b0f331a0d521-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"e3c34454-315e-4821-ab25-b0f331a0d521\") " pod="openstack/rabbitmq-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.496361 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mnz7l\" (UniqueName: \"kubernetes.io/projected/e3c34454-315e-4821-ab25-b0f331a0d521-kube-api-access-mnz7l\") pod \"rabbitmq-server-0\" (UID: \"e3c34454-315e-4821-ab25-b0f331a0d521\") " pod="openstack/rabbitmq-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.496401 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/e3c34454-315e-4821-ab25-b0f331a0d521-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"e3c34454-315e-4821-ab25-b0f331a0d521\") " pod="openstack/rabbitmq-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.597749 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-server-0\" (UID: \"e3c34454-315e-4821-ab25-b0f331a0d521\") " pod="openstack/rabbitmq-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.597787 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/e3c34454-315e-4821-ab25-b0f331a0d521-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"e3c34454-315e-4821-ab25-b0f331a0d521\") " pod="openstack/rabbitmq-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.597821 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/e3c34454-315e-4821-ab25-b0f331a0d521-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"e3c34454-315e-4821-ab25-b0f331a0d521\") " pod="openstack/rabbitmq-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.597837 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mnz7l\" (UniqueName: \"kubernetes.io/projected/e3c34454-315e-4821-ab25-b0f331a0d521-kube-api-access-mnz7l\") pod \"rabbitmq-server-0\" (UID: \"e3c34454-315e-4821-ab25-b0f331a0d521\") " pod="openstack/rabbitmq-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.597867 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/e3c34454-315e-4821-ab25-b0f331a0d521-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"e3c34454-315e-4821-ab25-b0f331a0d521\") " pod="openstack/rabbitmq-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.597909 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/e3c34454-315e-4821-ab25-b0f331a0d521-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"e3c34454-315e-4821-ab25-b0f331a0d521\") " pod="openstack/rabbitmq-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.597927 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e3c34454-315e-4821-ab25-b0f331a0d521-config-data\") pod \"rabbitmq-server-0\" (UID: \"e3c34454-315e-4821-ab25-b0f331a0d521\") " pod="openstack/rabbitmq-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.597946 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/e3c34454-315e-4821-ab25-b0f331a0d521-server-conf\") pod \"rabbitmq-server-0\" (UID: \"e3c34454-315e-4821-ab25-b0f331a0d521\") " pod="openstack/rabbitmq-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.597970 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/e3c34454-315e-4821-ab25-b0f331a0d521-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"e3c34454-315e-4821-ab25-b0f331a0d521\") " pod="openstack/rabbitmq-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.598003 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/e3c34454-315e-4821-ab25-b0f331a0d521-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"e3c34454-315e-4821-ab25-b0f331a0d521\") " pod="openstack/rabbitmq-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.598023 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/e3c34454-315e-4821-ab25-b0f331a0d521-pod-info\") pod \"rabbitmq-server-0\" (UID: \"e3c34454-315e-4821-ab25-b0f331a0d521\") " pod="openstack/rabbitmq-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.599021 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/e3c34454-315e-4821-ab25-b0f331a0d521-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"e3c34454-315e-4821-ab25-b0f331a0d521\") " pod="openstack/rabbitmq-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.599121 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/e3c34454-315e-4821-ab25-b0f331a0d521-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"e3c34454-315e-4821-ab25-b0f331a0d521\") " pod="openstack/rabbitmq-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.599155 4907 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-server-0\" (UID: \"e3c34454-315e-4821-ab25-b0f331a0d521\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/rabbitmq-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.599660 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/e3c34454-315e-4821-ab25-b0f331a0d521-server-conf\") pod \"rabbitmq-server-0\" (UID: \"e3c34454-315e-4821-ab25-b0f331a0d521\") " pod="openstack/rabbitmq-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.603416 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/e3c34454-315e-4821-ab25-b0f331a0d521-pod-info\") pod \"rabbitmq-server-0\" (UID: \"e3c34454-315e-4821-ab25-b0f331a0d521\") " pod="openstack/rabbitmq-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.603502 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/e3c34454-315e-4821-ab25-b0f331a0d521-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"e3c34454-315e-4821-ab25-b0f331a0d521\") " pod="openstack/rabbitmq-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.603526 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e3c34454-315e-4821-ab25-b0f331a0d521-config-data\") pod \"rabbitmq-server-0\" (UID: \"e3c34454-315e-4821-ab25-b0f331a0d521\") " pod="openstack/rabbitmq-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.603561 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/e3c34454-315e-4821-ab25-b0f331a0d521-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"e3c34454-315e-4821-ab25-b0f331a0d521\") " pod="openstack/rabbitmq-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.608870 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/e3c34454-315e-4821-ab25-b0f331a0d521-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"e3c34454-315e-4821-ab25-b0f331a0d521\") " pod="openstack/rabbitmq-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.613809 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/e3c34454-315e-4821-ab25-b0f331a0d521-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"e3c34454-315e-4821-ab25-b0f331a0d521\") " pod="openstack/rabbitmq-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.616911 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mnz7l\" (UniqueName: \"kubernetes.io/projected/e3c34454-315e-4821-ab25-b0f331a0d521-kube-api-access-mnz7l\") pod \"rabbitmq-server-0\" (UID: \"e3c34454-315e-4821-ab25-b0f331a0d521\") " pod="openstack/rabbitmq-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.617487 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-server-0\" (UID: \"e3c34454-315e-4821-ab25-b0f331a0d521\") " pod="openstack/rabbitmq-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.694734 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.844304 4907 generic.go:334] "Generic (PLEG): container finished" podID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerID="ac423129eeb1fa8eb8df325acb5ccb91e119992cd551dd9d8c11d304cb95b736" exitCode=0 Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.844346 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerDied","Data":"ac423129eeb1fa8eb8df325acb5ccb91e119992cd551dd9d8c11d304cb95b736"} Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.844378 4907 scope.go:117] "RemoveContainer" containerID="abe5ce1bc0297c29b962bd8d44bae1a474036514922a5df8e114b5dafeb79cce" Mar 13 14:25:19 crc kubenswrapper[4907]: I0313 14:25:19.415834 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Mar 13 14:25:19 crc kubenswrapper[4907]: I0313 14:25:19.417278 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Mar 13 14:25:19 crc kubenswrapper[4907]: I0313 14:25:19.420430 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Mar 13 14:25:19 crc kubenswrapper[4907]: I0313 14:25:19.420862 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-tptn8" Mar 13 14:25:19 crc kubenswrapper[4907]: I0313 14:25:19.421725 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Mar 13 14:25:19 crc kubenswrapper[4907]: I0313 14:25:19.421863 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Mar 13 14:25:19 crc kubenswrapper[4907]: I0313 14:25:19.425947 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Mar 13 14:25:19 crc kubenswrapper[4907]: I0313 14:25:19.445677 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Mar 13 14:25:19 crc kubenswrapper[4907]: I0313 14:25:19.512452 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-0\" (UID: \"c4411bd2-c555-433a-9015-f623948b1401\") " pod="openstack/openstack-galera-0" Mar 13 14:25:19 crc kubenswrapper[4907]: I0313 14:25:19.512510 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/c4411bd2-c555-433a-9015-f623948b1401-config-data-default\") pod \"openstack-galera-0\" (UID: \"c4411bd2-c555-433a-9015-f623948b1401\") " pod="openstack/openstack-galera-0" Mar 13 14:25:19 crc kubenswrapper[4907]: I0313 14:25:19.512557 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c4411bd2-c555-433a-9015-f623948b1401-operator-scripts\") pod \"openstack-galera-0\" (UID: \"c4411bd2-c555-433a-9015-f623948b1401\") " pod="openstack/openstack-galera-0" Mar 13 14:25:19 crc kubenswrapper[4907]: I0313 14:25:19.512592 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pcwkz\" (UniqueName: \"kubernetes.io/projected/c4411bd2-c555-433a-9015-f623948b1401-kube-api-access-pcwkz\") pod \"openstack-galera-0\" (UID: \"c4411bd2-c555-433a-9015-f623948b1401\") " pod="openstack/openstack-galera-0" Mar 13 14:25:19 crc kubenswrapper[4907]: I0313 14:25:19.512612 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4411bd2-c555-433a-9015-f623948b1401-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"c4411bd2-c555-433a-9015-f623948b1401\") " pod="openstack/openstack-galera-0" Mar 13 14:25:19 crc kubenswrapper[4907]: I0313 14:25:19.512671 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/c4411bd2-c555-433a-9015-f623948b1401-kolla-config\") pod \"openstack-galera-0\" (UID: \"c4411bd2-c555-433a-9015-f623948b1401\") " pod="openstack/openstack-galera-0" Mar 13 14:25:19 crc kubenswrapper[4907]: I0313 14:25:19.512690 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/c4411bd2-c555-433a-9015-f623948b1401-config-data-generated\") pod \"openstack-galera-0\" (UID: \"c4411bd2-c555-433a-9015-f623948b1401\") " pod="openstack/openstack-galera-0" Mar 13 14:25:19 crc kubenswrapper[4907]: I0313 14:25:19.512721 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/c4411bd2-c555-433a-9015-f623948b1401-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"c4411bd2-c555-433a-9015-f623948b1401\") " pod="openstack/openstack-galera-0" Mar 13 14:25:19 crc kubenswrapper[4907]: I0313 14:25:19.614643 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c4411bd2-c555-433a-9015-f623948b1401-operator-scripts\") pod \"openstack-galera-0\" (UID: \"c4411bd2-c555-433a-9015-f623948b1401\") " pod="openstack/openstack-galera-0" Mar 13 14:25:19 crc kubenswrapper[4907]: I0313 14:25:19.614721 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pcwkz\" (UniqueName: \"kubernetes.io/projected/c4411bd2-c555-433a-9015-f623948b1401-kube-api-access-pcwkz\") pod \"openstack-galera-0\" (UID: \"c4411bd2-c555-433a-9015-f623948b1401\") " pod="openstack/openstack-galera-0" Mar 13 14:25:19 crc kubenswrapper[4907]: I0313 14:25:19.614747 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4411bd2-c555-433a-9015-f623948b1401-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"c4411bd2-c555-433a-9015-f623948b1401\") " pod="openstack/openstack-galera-0" Mar 13 14:25:19 crc kubenswrapper[4907]: I0313 14:25:19.614815 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/c4411bd2-c555-433a-9015-f623948b1401-kolla-config\") pod \"openstack-galera-0\" (UID: \"c4411bd2-c555-433a-9015-f623948b1401\") " pod="openstack/openstack-galera-0" Mar 13 14:25:19 crc kubenswrapper[4907]: I0313 14:25:19.614839 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/c4411bd2-c555-433a-9015-f623948b1401-config-data-generated\") pod \"openstack-galera-0\" (UID: \"c4411bd2-c555-433a-9015-f623948b1401\") " pod="openstack/openstack-galera-0" Mar 13 14:25:19 crc kubenswrapper[4907]: I0313 14:25:19.614870 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/c4411bd2-c555-433a-9015-f623948b1401-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"c4411bd2-c555-433a-9015-f623948b1401\") " pod="openstack/openstack-galera-0" Mar 13 14:25:19 crc kubenswrapper[4907]: I0313 14:25:19.614944 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-0\" (UID: \"c4411bd2-c555-433a-9015-f623948b1401\") " pod="openstack/openstack-galera-0" Mar 13 14:25:19 crc kubenswrapper[4907]: I0313 14:25:19.614972 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/c4411bd2-c555-433a-9015-f623948b1401-config-data-default\") pod \"openstack-galera-0\" (UID: \"c4411bd2-c555-433a-9015-f623948b1401\") " pod="openstack/openstack-galera-0" Mar 13 14:25:19 crc kubenswrapper[4907]: I0313 14:25:19.616101 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/c4411bd2-c555-433a-9015-f623948b1401-config-data-default\") pod \"openstack-galera-0\" (UID: \"c4411bd2-c555-433a-9015-f623948b1401\") " pod="openstack/openstack-galera-0" Mar 13 14:25:19 crc kubenswrapper[4907]: I0313 14:25:19.617411 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c4411bd2-c555-433a-9015-f623948b1401-operator-scripts\") pod \"openstack-galera-0\" (UID: \"c4411bd2-c555-433a-9015-f623948b1401\") " pod="openstack/openstack-galera-0" Mar 13 14:25:19 crc kubenswrapper[4907]: I0313 14:25:19.619597 4907 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-0\" (UID: \"c4411bd2-c555-433a-9015-f623948b1401\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/openstack-galera-0" Mar 13 14:25:19 crc kubenswrapper[4907]: I0313 14:25:19.619671 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/c4411bd2-c555-433a-9015-f623948b1401-kolla-config\") pod \"openstack-galera-0\" (UID: \"c4411bd2-c555-433a-9015-f623948b1401\") " pod="openstack/openstack-galera-0" Mar 13 14:25:19 crc kubenswrapper[4907]: I0313 14:25:19.620208 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/c4411bd2-c555-433a-9015-f623948b1401-config-data-generated\") pod \"openstack-galera-0\" (UID: \"c4411bd2-c555-433a-9015-f623948b1401\") " pod="openstack/openstack-galera-0" Mar 13 14:25:19 crc kubenswrapper[4907]: I0313 14:25:19.628459 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/c4411bd2-c555-433a-9015-f623948b1401-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"c4411bd2-c555-433a-9015-f623948b1401\") " pod="openstack/openstack-galera-0" Mar 13 14:25:19 crc kubenswrapper[4907]: I0313 14:25:19.646748 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4411bd2-c555-433a-9015-f623948b1401-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"c4411bd2-c555-433a-9015-f623948b1401\") " pod="openstack/openstack-galera-0" Mar 13 14:25:19 crc kubenswrapper[4907]: I0313 14:25:19.647550 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pcwkz\" (UniqueName: \"kubernetes.io/projected/c4411bd2-c555-433a-9015-f623948b1401-kube-api-access-pcwkz\") pod \"openstack-galera-0\" (UID: \"c4411bd2-c555-433a-9015-f623948b1401\") " pod="openstack/openstack-galera-0" Mar 13 14:25:19 crc kubenswrapper[4907]: I0313 14:25:19.655864 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-0\" (UID: \"c4411bd2-c555-433a-9015-f623948b1401\") " pod="openstack/openstack-galera-0" Mar 13 14:25:19 crc kubenswrapper[4907]: I0313 14:25:19.746444 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Mar 13 14:25:20 crc kubenswrapper[4907]: I0313 14:25:20.832832 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Mar 13 14:25:20 crc kubenswrapper[4907]: I0313 14:25:20.834355 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Mar 13 14:25:20 crc kubenswrapper[4907]: I0313 14:25:20.835991 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Mar 13 14:25:20 crc kubenswrapper[4907]: I0313 14:25:20.837777 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Mar 13 14:25:20 crc kubenswrapper[4907]: I0313 14:25:20.838091 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Mar 13 14:25:20 crc kubenswrapper[4907]: I0313 14:25:20.838213 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-szj6q" Mar 13 14:25:20 crc kubenswrapper[4907]: I0313 14:25:20.843041 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Mar 13 14:25:20 crc kubenswrapper[4907]: I0313 14:25:20.931773 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/58ab9689-3ab2-42f2-ad56-beb22e29f8da-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"58ab9689-3ab2-42f2-ad56-beb22e29f8da\") " pod="openstack/openstack-cell1-galera-0" Mar 13 14:25:20 crc kubenswrapper[4907]: I0313 14:25:20.932195 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/58ab9689-3ab2-42f2-ad56-beb22e29f8da-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"58ab9689-3ab2-42f2-ad56-beb22e29f8da\") " pod="openstack/openstack-cell1-galera-0" Mar 13 14:25:20 crc kubenswrapper[4907]: I0313 14:25:20.932266 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-cell1-galera-0\" (UID: \"58ab9689-3ab2-42f2-ad56-beb22e29f8da\") " pod="openstack/openstack-cell1-galera-0" Mar 13 14:25:20 crc kubenswrapper[4907]: I0313 14:25:20.932302 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kvwz4\" (UniqueName: \"kubernetes.io/projected/58ab9689-3ab2-42f2-ad56-beb22e29f8da-kube-api-access-kvwz4\") pod \"openstack-cell1-galera-0\" (UID: \"58ab9689-3ab2-42f2-ad56-beb22e29f8da\") " pod="openstack/openstack-cell1-galera-0" Mar 13 14:25:20 crc kubenswrapper[4907]: I0313 14:25:20.932332 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/58ab9689-3ab2-42f2-ad56-beb22e29f8da-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"58ab9689-3ab2-42f2-ad56-beb22e29f8da\") " pod="openstack/openstack-cell1-galera-0" Mar 13 14:25:20 crc kubenswrapper[4907]: I0313 14:25:20.932382 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/58ab9689-3ab2-42f2-ad56-beb22e29f8da-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"58ab9689-3ab2-42f2-ad56-beb22e29f8da\") " pod="openstack/openstack-cell1-galera-0" Mar 13 14:25:20 crc kubenswrapper[4907]: I0313 14:25:20.932409 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58ab9689-3ab2-42f2-ad56-beb22e29f8da-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"58ab9689-3ab2-42f2-ad56-beb22e29f8da\") " pod="openstack/openstack-cell1-galera-0" Mar 13 14:25:20 crc kubenswrapper[4907]: I0313 14:25:20.932509 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/58ab9689-3ab2-42f2-ad56-beb22e29f8da-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"58ab9689-3ab2-42f2-ad56-beb22e29f8da\") " pod="openstack/openstack-cell1-galera-0" Mar 13 14:25:20 crc kubenswrapper[4907]: I0313 14:25:20.953571 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Mar 13 14:25:20 crc kubenswrapper[4907]: I0313 14:25:20.955741 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Mar 13 14:25:20 crc kubenswrapper[4907]: I0313 14:25:20.959024 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Mar 13 14:25:20 crc kubenswrapper[4907]: I0313 14:25:20.959384 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-zr52r" Mar 13 14:25:20 crc kubenswrapper[4907]: I0313 14:25:20.959624 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Mar 13 14:25:20 crc kubenswrapper[4907]: I0313 14:25:20.966507 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Mar 13 14:25:21 crc kubenswrapper[4907]: I0313 14:25:21.034057 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/58ab9689-3ab2-42f2-ad56-beb22e29f8da-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"58ab9689-3ab2-42f2-ad56-beb22e29f8da\") " pod="openstack/openstack-cell1-galera-0" Mar 13 14:25:21 crc kubenswrapper[4907]: I0313 14:25:21.034098 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cacb281b-ee3b-42b6-8754-d07e6ea6aa4b-combined-ca-bundle\") pod \"memcached-0\" (UID: \"cacb281b-ee3b-42b6-8754-d07e6ea6aa4b\") " pod="openstack/memcached-0" Mar 13 14:25:21 crc kubenswrapper[4907]: I0313 14:25:21.034126 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/58ab9689-3ab2-42f2-ad56-beb22e29f8da-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"58ab9689-3ab2-42f2-ad56-beb22e29f8da\") " pod="openstack/openstack-cell1-galera-0" Mar 13 14:25:21 crc kubenswrapper[4907]: I0313 14:25:21.034142 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58ab9689-3ab2-42f2-ad56-beb22e29f8da-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"58ab9689-3ab2-42f2-ad56-beb22e29f8da\") " pod="openstack/openstack-cell1-galera-0" Mar 13 14:25:21 crc kubenswrapper[4907]: I0313 14:25:21.034160 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cacb281b-ee3b-42b6-8754-d07e6ea6aa4b-config-data\") pod \"memcached-0\" (UID: \"cacb281b-ee3b-42b6-8754-d07e6ea6aa4b\") " pod="openstack/memcached-0" Mar 13 14:25:21 crc kubenswrapper[4907]: I0313 14:25:21.034202 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/cacb281b-ee3b-42b6-8754-d07e6ea6aa4b-memcached-tls-certs\") pod \"memcached-0\" (UID: \"cacb281b-ee3b-42b6-8754-d07e6ea6aa4b\") " pod="openstack/memcached-0" Mar 13 14:25:21 crc kubenswrapper[4907]: I0313 14:25:21.034224 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/58ab9689-3ab2-42f2-ad56-beb22e29f8da-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"58ab9689-3ab2-42f2-ad56-beb22e29f8da\") " pod="openstack/openstack-cell1-galera-0" Mar 13 14:25:21 crc kubenswrapper[4907]: I0313 14:25:21.034254 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/cacb281b-ee3b-42b6-8754-d07e6ea6aa4b-kolla-config\") pod \"memcached-0\" (UID: \"cacb281b-ee3b-42b6-8754-d07e6ea6aa4b\") " pod="openstack/memcached-0" Mar 13 14:25:21 crc kubenswrapper[4907]: I0313 14:25:21.034271 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/58ab9689-3ab2-42f2-ad56-beb22e29f8da-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"58ab9689-3ab2-42f2-ad56-beb22e29f8da\") " pod="openstack/openstack-cell1-galera-0" Mar 13 14:25:21 crc kubenswrapper[4907]: I0313 14:25:21.034298 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/58ab9689-3ab2-42f2-ad56-beb22e29f8da-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"58ab9689-3ab2-42f2-ad56-beb22e29f8da\") " pod="openstack/openstack-cell1-galera-0" Mar 13 14:25:21 crc kubenswrapper[4907]: I0313 14:25:21.034335 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-cell1-galera-0\" (UID: \"58ab9689-3ab2-42f2-ad56-beb22e29f8da\") " pod="openstack/openstack-cell1-galera-0" Mar 13 14:25:21 crc kubenswrapper[4907]: I0313 14:25:21.034359 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vdj7f\" (UniqueName: \"kubernetes.io/projected/cacb281b-ee3b-42b6-8754-d07e6ea6aa4b-kube-api-access-vdj7f\") pod \"memcached-0\" (UID: \"cacb281b-ee3b-42b6-8754-d07e6ea6aa4b\") " pod="openstack/memcached-0" Mar 13 14:25:21 crc kubenswrapper[4907]: I0313 14:25:21.034376 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kvwz4\" (UniqueName: \"kubernetes.io/projected/58ab9689-3ab2-42f2-ad56-beb22e29f8da-kube-api-access-kvwz4\") pod \"openstack-cell1-galera-0\" (UID: \"58ab9689-3ab2-42f2-ad56-beb22e29f8da\") " pod="openstack/openstack-cell1-galera-0" Mar 13 14:25:21 crc kubenswrapper[4907]: I0313 14:25:21.035202 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/58ab9689-3ab2-42f2-ad56-beb22e29f8da-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"58ab9689-3ab2-42f2-ad56-beb22e29f8da\") " pod="openstack/openstack-cell1-galera-0" Mar 13 14:25:21 crc kubenswrapper[4907]: I0313 14:25:21.036238 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/58ab9689-3ab2-42f2-ad56-beb22e29f8da-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"58ab9689-3ab2-42f2-ad56-beb22e29f8da\") " pod="openstack/openstack-cell1-galera-0" Mar 13 14:25:21 crc kubenswrapper[4907]: I0313 14:25:21.036754 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/58ab9689-3ab2-42f2-ad56-beb22e29f8da-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"58ab9689-3ab2-42f2-ad56-beb22e29f8da\") " pod="openstack/openstack-cell1-galera-0" Mar 13 14:25:21 crc kubenswrapper[4907]: I0313 14:25:21.036843 4907 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-cell1-galera-0\" (UID: \"58ab9689-3ab2-42f2-ad56-beb22e29f8da\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/openstack-cell1-galera-0" Mar 13 14:25:21 crc kubenswrapper[4907]: I0313 14:25:21.037716 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/58ab9689-3ab2-42f2-ad56-beb22e29f8da-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"58ab9689-3ab2-42f2-ad56-beb22e29f8da\") " pod="openstack/openstack-cell1-galera-0" Mar 13 14:25:21 crc kubenswrapper[4907]: I0313 14:25:21.041707 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/58ab9689-3ab2-42f2-ad56-beb22e29f8da-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"58ab9689-3ab2-42f2-ad56-beb22e29f8da\") " pod="openstack/openstack-cell1-galera-0" Mar 13 14:25:21 crc kubenswrapper[4907]: I0313 14:25:21.056012 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kvwz4\" (UniqueName: \"kubernetes.io/projected/58ab9689-3ab2-42f2-ad56-beb22e29f8da-kube-api-access-kvwz4\") pod \"openstack-cell1-galera-0\" (UID: \"58ab9689-3ab2-42f2-ad56-beb22e29f8da\") " pod="openstack/openstack-cell1-galera-0" Mar 13 14:25:21 crc kubenswrapper[4907]: I0313 14:25:21.057019 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-cell1-galera-0\" (UID: \"58ab9689-3ab2-42f2-ad56-beb22e29f8da\") " pod="openstack/openstack-cell1-galera-0" Mar 13 14:25:21 crc kubenswrapper[4907]: I0313 14:25:21.058515 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58ab9689-3ab2-42f2-ad56-beb22e29f8da-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"58ab9689-3ab2-42f2-ad56-beb22e29f8da\") " pod="openstack/openstack-cell1-galera-0" Mar 13 14:25:21 crc kubenswrapper[4907]: I0313 14:25:21.140200 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cacb281b-ee3b-42b6-8754-d07e6ea6aa4b-config-data\") pod \"memcached-0\" (UID: \"cacb281b-ee3b-42b6-8754-d07e6ea6aa4b\") " pod="openstack/memcached-0" Mar 13 14:25:21 crc kubenswrapper[4907]: I0313 14:25:21.140276 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/cacb281b-ee3b-42b6-8754-d07e6ea6aa4b-memcached-tls-certs\") pod \"memcached-0\" (UID: \"cacb281b-ee3b-42b6-8754-d07e6ea6aa4b\") " pod="openstack/memcached-0" Mar 13 14:25:21 crc kubenswrapper[4907]: I0313 14:25:21.140303 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/cacb281b-ee3b-42b6-8754-d07e6ea6aa4b-kolla-config\") pod \"memcached-0\" (UID: \"cacb281b-ee3b-42b6-8754-d07e6ea6aa4b\") " pod="openstack/memcached-0" Mar 13 14:25:21 crc kubenswrapper[4907]: I0313 14:25:21.140360 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vdj7f\" (UniqueName: \"kubernetes.io/projected/cacb281b-ee3b-42b6-8754-d07e6ea6aa4b-kube-api-access-vdj7f\") pod \"memcached-0\" (UID: \"cacb281b-ee3b-42b6-8754-d07e6ea6aa4b\") " pod="openstack/memcached-0" Mar 13 14:25:21 crc kubenswrapper[4907]: I0313 14:25:21.140388 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cacb281b-ee3b-42b6-8754-d07e6ea6aa4b-combined-ca-bundle\") pod \"memcached-0\" (UID: \"cacb281b-ee3b-42b6-8754-d07e6ea6aa4b\") " pod="openstack/memcached-0" Mar 13 14:25:21 crc kubenswrapper[4907]: I0313 14:25:21.142943 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/cacb281b-ee3b-42b6-8754-d07e6ea6aa4b-kolla-config\") pod \"memcached-0\" (UID: \"cacb281b-ee3b-42b6-8754-d07e6ea6aa4b\") " pod="openstack/memcached-0" Mar 13 14:25:21 crc kubenswrapper[4907]: I0313 14:25:21.142946 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cacb281b-ee3b-42b6-8754-d07e6ea6aa4b-config-data\") pod \"memcached-0\" (UID: \"cacb281b-ee3b-42b6-8754-d07e6ea6aa4b\") " pod="openstack/memcached-0" Mar 13 14:25:21 crc kubenswrapper[4907]: I0313 14:25:21.143438 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cacb281b-ee3b-42b6-8754-d07e6ea6aa4b-combined-ca-bundle\") pod \"memcached-0\" (UID: \"cacb281b-ee3b-42b6-8754-d07e6ea6aa4b\") " pod="openstack/memcached-0" Mar 13 14:25:21 crc kubenswrapper[4907]: I0313 14:25:21.145635 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/cacb281b-ee3b-42b6-8754-d07e6ea6aa4b-memcached-tls-certs\") pod \"memcached-0\" (UID: \"cacb281b-ee3b-42b6-8754-d07e6ea6aa4b\") " pod="openstack/memcached-0" Mar 13 14:25:21 crc kubenswrapper[4907]: I0313 14:25:21.155808 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Mar 13 14:25:21 crc kubenswrapper[4907]: I0313 14:25:21.162502 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vdj7f\" (UniqueName: \"kubernetes.io/projected/cacb281b-ee3b-42b6-8754-d07e6ea6aa4b-kube-api-access-vdj7f\") pod \"memcached-0\" (UID: \"cacb281b-ee3b-42b6-8754-d07e6ea6aa4b\") " pod="openstack/memcached-0" Mar 13 14:25:21 crc kubenswrapper[4907]: I0313 14:25:21.271710 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Mar 13 14:25:21 crc kubenswrapper[4907]: W0313 14:25:21.428686 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod779b1958_f7ed_4c53_a4ff_a6c4a803bebf.slice/crio-54b19297eaef464c5a401c18dcda022f8c38c1cecb9978d5b544939e48c81aa1 WatchSource:0}: Error finding container 54b19297eaef464c5a401c18dcda022f8c38c1cecb9978d5b544939e48c81aa1: Status 404 returned error can't find the container with id 54b19297eaef464c5a401c18dcda022f8c38c1cecb9978d5b544939e48c81aa1 Mar 13 14:25:21 crc kubenswrapper[4907]: I0313 14:25:21.908081 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54b5dffb47-wttzd" event={"ID":"779b1958-f7ed-4c53-a4ff-a6c4a803bebf","Type":"ContainerStarted","Data":"54b19297eaef464c5a401c18dcda022f8c38c1cecb9978d5b544939e48c81aa1"} Mar 13 14:25:22 crc kubenswrapper[4907]: I0313 14:25:22.989499 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Mar 13 14:25:22 crc kubenswrapper[4907]: I0313 14:25:22.990691 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Mar 13 14:25:23 crc kubenswrapper[4907]: I0313 14:25:23.000682 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Mar 13 14:25:23 crc kubenswrapper[4907]: I0313 14:25:23.046417 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-48tmv" Mar 13 14:25:23 crc kubenswrapper[4907]: I0313 14:25:23.077129 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zg7qk\" (UniqueName: \"kubernetes.io/projected/a7f6877c-1e9e-4e17-803d-90efa7d66469-kube-api-access-zg7qk\") pod \"kube-state-metrics-0\" (UID: \"a7f6877c-1e9e-4e17-803d-90efa7d66469\") " pod="openstack/kube-state-metrics-0" Mar 13 14:25:23 crc kubenswrapper[4907]: I0313 14:25:23.178099 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zg7qk\" (UniqueName: \"kubernetes.io/projected/a7f6877c-1e9e-4e17-803d-90efa7d66469-kube-api-access-zg7qk\") pod \"kube-state-metrics-0\" (UID: \"a7f6877c-1e9e-4e17-803d-90efa7d66469\") " pod="openstack/kube-state-metrics-0" Mar 13 14:25:23 crc kubenswrapper[4907]: I0313 14:25:23.206939 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zg7qk\" (UniqueName: \"kubernetes.io/projected/a7f6877c-1e9e-4e17-803d-90efa7d66469-kube-api-access-zg7qk\") pod \"kube-state-metrics-0\" (UID: \"a7f6877c-1e9e-4e17-803d-90efa7d66469\") " pod="openstack/kube-state-metrics-0" Mar 13 14:25:23 crc kubenswrapper[4907]: I0313 14:25:23.363653 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Mar 13 14:25:25 crc kubenswrapper[4907]: I0313 14:25:25.110579 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Mar 13 14:25:26 crc kubenswrapper[4907]: I0313 14:25:26.516595 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Mar 13 14:25:26 crc kubenswrapper[4907]: I0313 14:25:26.517753 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Mar 13 14:25:26 crc kubenswrapper[4907]: I0313 14:25:26.520398 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Mar 13 14:25:26 crc kubenswrapper[4907]: I0313 14:25:26.520744 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-tsp64" Mar 13 14:25:26 crc kubenswrapper[4907]: I0313 14:25:26.521836 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Mar 13 14:25:26 crc kubenswrapper[4907]: I0313 14:25:26.521961 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Mar 13 14:25:26 crc kubenswrapper[4907]: I0313 14:25:26.522360 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Mar 13 14:25:26 crc kubenswrapper[4907]: I0313 14:25:26.542253 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Mar 13 14:25:26 crc kubenswrapper[4907]: I0313 14:25:26.644556 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/056dd756-0d7b-471b-9929-f622d05ad606-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"056dd756-0d7b-471b-9929-f622d05ad606\") " pod="openstack/ovsdbserver-nb-0" Mar 13 14:25:26 crc kubenswrapper[4907]: I0313 14:25:26.644596 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ckz4p\" (UniqueName: \"kubernetes.io/projected/056dd756-0d7b-471b-9929-f622d05ad606-kube-api-access-ckz4p\") pod \"ovsdbserver-nb-0\" (UID: \"056dd756-0d7b-471b-9929-f622d05ad606\") " pod="openstack/ovsdbserver-nb-0" Mar 13 14:25:26 crc kubenswrapper[4907]: I0313 14:25:26.644626 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/056dd756-0d7b-471b-9929-f622d05ad606-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"056dd756-0d7b-471b-9929-f622d05ad606\") " pod="openstack/ovsdbserver-nb-0" Mar 13 14:25:26 crc kubenswrapper[4907]: I0313 14:25:26.644645 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/056dd756-0d7b-471b-9929-f622d05ad606-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"056dd756-0d7b-471b-9929-f622d05ad606\") " pod="openstack/ovsdbserver-nb-0" Mar 13 14:25:26 crc kubenswrapper[4907]: I0313 14:25:26.644660 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/056dd756-0d7b-471b-9929-f622d05ad606-config\") pod \"ovsdbserver-nb-0\" (UID: \"056dd756-0d7b-471b-9929-f622d05ad606\") " pod="openstack/ovsdbserver-nb-0" Mar 13 14:25:26 crc kubenswrapper[4907]: I0313 14:25:26.644679 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"ovsdbserver-nb-0\" (UID: \"056dd756-0d7b-471b-9929-f622d05ad606\") " pod="openstack/ovsdbserver-nb-0" Mar 13 14:25:26 crc kubenswrapper[4907]: I0313 14:25:26.644920 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/056dd756-0d7b-471b-9929-f622d05ad606-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"056dd756-0d7b-471b-9929-f622d05ad606\") " pod="openstack/ovsdbserver-nb-0" Mar 13 14:25:26 crc kubenswrapper[4907]: I0313 14:25:26.645019 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/056dd756-0d7b-471b-9929-f622d05ad606-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"056dd756-0d7b-471b-9929-f622d05ad606\") " pod="openstack/ovsdbserver-nb-0" Mar 13 14:25:26 crc kubenswrapper[4907]: I0313 14:25:26.748866 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/056dd756-0d7b-471b-9929-f622d05ad606-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"056dd756-0d7b-471b-9929-f622d05ad606\") " pod="openstack/ovsdbserver-nb-0" Mar 13 14:25:26 crc kubenswrapper[4907]: I0313 14:25:26.749205 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/056dd756-0d7b-471b-9929-f622d05ad606-config\") pod \"ovsdbserver-nb-0\" (UID: \"056dd756-0d7b-471b-9929-f622d05ad606\") " pod="openstack/ovsdbserver-nb-0" Mar 13 14:25:26 crc kubenswrapper[4907]: I0313 14:25:26.749232 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"ovsdbserver-nb-0\" (UID: \"056dd756-0d7b-471b-9929-f622d05ad606\") " pod="openstack/ovsdbserver-nb-0" Mar 13 14:25:26 crc kubenswrapper[4907]: I0313 14:25:26.749270 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/056dd756-0d7b-471b-9929-f622d05ad606-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"056dd756-0d7b-471b-9929-f622d05ad606\") " pod="openstack/ovsdbserver-nb-0" Mar 13 14:25:26 crc kubenswrapper[4907]: I0313 14:25:26.749310 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/056dd756-0d7b-471b-9929-f622d05ad606-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"056dd756-0d7b-471b-9929-f622d05ad606\") " pod="openstack/ovsdbserver-nb-0" Mar 13 14:25:26 crc kubenswrapper[4907]: I0313 14:25:26.749372 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/056dd756-0d7b-471b-9929-f622d05ad606-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"056dd756-0d7b-471b-9929-f622d05ad606\") " pod="openstack/ovsdbserver-nb-0" Mar 13 14:25:26 crc kubenswrapper[4907]: I0313 14:25:26.749386 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ckz4p\" (UniqueName: \"kubernetes.io/projected/056dd756-0d7b-471b-9929-f622d05ad606-kube-api-access-ckz4p\") pod \"ovsdbserver-nb-0\" (UID: \"056dd756-0d7b-471b-9929-f622d05ad606\") " pod="openstack/ovsdbserver-nb-0" Mar 13 14:25:26 crc kubenswrapper[4907]: I0313 14:25:26.749411 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/056dd756-0d7b-471b-9929-f622d05ad606-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"056dd756-0d7b-471b-9929-f622d05ad606\") " pod="openstack/ovsdbserver-nb-0" Mar 13 14:25:26 crc kubenswrapper[4907]: I0313 14:25:26.750666 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/056dd756-0d7b-471b-9929-f622d05ad606-config\") pod \"ovsdbserver-nb-0\" (UID: \"056dd756-0d7b-471b-9929-f622d05ad606\") " pod="openstack/ovsdbserver-nb-0" Mar 13 14:25:26 crc kubenswrapper[4907]: I0313 14:25:26.750713 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/056dd756-0d7b-471b-9929-f622d05ad606-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"056dd756-0d7b-471b-9929-f622d05ad606\") " pod="openstack/ovsdbserver-nb-0" Mar 13 14:25:26 crc kubenswrapper[4907]: I0313 14:25:26.750965 4907 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"ovsdbserver-nb-0\" (UID: \"056dd756-0d7b-471b-9929-f622d05ad606\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/ovsdbserver-nb-0" Mar 13 14:25:26 crc kubenswrapper[4907]: I0313 14:25:26.751326 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/056dd756-0d7b-471b-9929-f622d05ad606-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"056dd756-0d7b-471b-9929-f622d05ad606\") " pod="openstack/ovsdbserver-nb-0" Mar 13 14:25:26 crc kubenswrapper[4907]: I0313 14:25:26.757003 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/056dd756-0d7b-471b-9929-f622d05ad606-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"056dd756-0d7b-471b-9929-f622d05ad606\") " pod="openstack/ovsdbserver-nb-0" Mar 13 14:25:26 crc kubenswrapper[4907]: I0313 14:25:26.757595 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/056dd756-0d7b-471b-9929-f622d05ad606-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"056dd756-0d7b-471b-9929-f622d05ad606\") " pod="openstack/ovsdbserver-nb-0" Mar 13 14:25:26 crc kubenswrapper[4907]: I0313 14:25:26.761303 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/056dd756-0d7b-471b-9929-f622d05ad606-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"056dd756-0d7b-471b-9929-f622d05ad606\") " pod="openstack/ovsdbserver-nb-0" Mar 13 14:25:26 crc kubenswrapper[4907]: I0313 14:25:26.772172 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ckz4p\" (UniqueName: \"kubernetes.io/projected/056dd756-0d7b-471b-9929-f622d05ad606-kube-api-access-ckz4p\") pod \"ovsdbserver-nb-0\" (UID: \"056dd756-0d7b-471b-9929-f622d05ad606\") " pod="openstack/ovsdbserver-nb-0" Mar 13 14:25:26 crc kubenswrapper[4907]: I0313 14:25:26.772435 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"ovsdbserver-nb-0\" (UID: \"056dd756-0d7b-471b-9929-f622d05ad606\") " pod="openstack/ovsdbserver-nb-0" Mar 13 14:25:26 crc kubenswrapper[4907]: I0313 14:25:26.837343 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Mar 13 14:25:27 crc kubenswrapper[4907]: I0313 14:25:27.099127 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-bw2xb"] Mar 13 14:25:27 crc kubenswrapper[4907]: I0313 14:25:27.100163 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-bw2xb" Mar 13 14:25:27 crc kubenswrapper[4907]: I0313 14:25:27.105247 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-4m476" Mar 13 14:25:27 crc kubenswrapper[4907]: I0313 14:25:27.105470 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Mar 13 14:25:27 crc kubenswrapper[4907]: I0313 14:25:27.105625 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Mar 13 14:25:27 crc kubenswrapper[4907]: I0313 14:25:27.109297 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-vp6f5"] Mar 13 14:25:27 crc kubenswrapper[4907]: I0313 14:25:27.110830 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-vp6f5" Mar 13 14:25:27 crc kubenswrapper[4907]: I0313 14:25:27.116205 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-vp6f5"] Mar 13 14:25:27 crc kubenswrapper[4907]: I0313 14:25:27.122395 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-bw2xb"] Mar 13 14:25:27 crc kubenswrapper[4907]: I0313 14:25:27.155301 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c1b70392-1240-40d9-8128-e7abe29c8398-scripts\") pod \"ovn-controller-ovs-vp6f5\" (UID: \"c1b70392-1240-40d9-8128-e7abe29c8398\") " pod="openstack/ovn-controller-ovs-vp6f5" Mar 13 14:25:27 crc kubenswrapper[4907]: I0313 14:25:27.155348 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f8119b2-e38a-494d-967f-5198b83512c7-combined-ca-bundle\") pod \"ovn-controller-bw2xb\" (UID: \"6f8119b2-e38a-494d-967f-5198b83512c7\") " pod="openstack/ovn-controller-bw2xb" Mar 13 14:25:27 crc kubenswrapper[4907]: I0313 14:25:27.155382 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/6f8119b2-e38a-494d-967f-5198b83512c7-ovn-controller-tls-certs\") pod \"ovn-controller-bw2xb\" (UID: \"6f8119b2-e38a-494d-967f-5198b83512c7\") " pod="openstack/ovn-controller-bw2xb" Mar 13 14:25:27 crc kubenswrapper[4907]: I0313 14:25:27.155476 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6f8119b2-e38a-494d-967f-5198b83512c7-var-run\") pod \"ovn-controller-bw2xb\" (UID: \"6f8119b2-e38a-494d-967f-5198b83512c7\") " pod="openstack/ovn-controller-bw2xb" Mar 13 14:25:27 crc kubenswrapper[4907]: I0313 14:25:27.155500 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/c1b70392-1240-40d9-8128-e7abe29c8398-var-log\") pod \"ovn-controller-ovs-vp6f5\" (UID: \"c1b70392-1240-40d9-8128-e7abe29c8398\") " pod="openstack/ovn-controller-ovs-vp6f5" Mar 13 14:25:27 crc kubenswrapper[4907]: I0313 14:25:27.155521 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/c1b70392-1240-40d9-8128-e7abe29c8398-var-run\") pod \"ovn-controller-ovs-vp6f5\" (UID: \"c1b70392-1240-40d9-8128-e7abe29c8398\") " pod="openstack/ovn-controller-ovs-vp6f5" Mar 13 14:25:27 crc kubenswrapper[4907]: I0313 14:25:27.155539 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/c1b70392-1240-40d9-8128-e7abe29c8398-etc-ovs\") pod \"ovn-controller-ovs-vp6f5\" (UID: \"c1b70392-1240-40d9-8128-e7abe29c8398\") " pod="openstack/ovn-controller-ovs-vp6f5" Mar 13 14:25:27 crc kubenswrapper[4907]: I0313 14:25:27.155557 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4qvcj\" (UniqueName: \"kubernetes.io/projected/c1b70392-1240-40d9-8128-e7abe29c8398-kube-api-access-4qvcj\") pod \"ovn-controller-ovs-vp6f5\" (UID: \"c1b70392-1240-40d9-8128-e7abe29c8398\") " pod="openstack/ovn-controller-ovs-vp6f5" Mar 13 14:25:27 crc kubenswrapper[4907]: I0313 14:25:27.155587 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v2qq5\" (UniqueName: \"kubernetes.io/projected/6f8119b2-e38a-494d-967f-5198b83512c7-kube-api-access-v2qq5\") pod \"ovn-controller-bw2xb\" (UID: \"6f8119b2-e38a-494d-967f-5198b83512c7\") " pod="openstack/ovn-controller-bw2xb" Mar 13 14:25:27 crc kubenswrapper[4907]: I0313 14:25:27.155608 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/6f8119b2-e38a-494d-967f-5198b83512c7-var-log-ovn\") pod \"ovn-controller-bw2xb\" (UID: \"6f8119b2-e38a-494d-967f-5198b83512c7\") " pod="openstack/ovn-controller-bw2xb" Mar 13 14:25:27 crc kubenswrapper[4907]: I0313 14:25:27.155621 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6f8119b2-e38a-494d-967f-5198b83512c7-scripts\") pod \"ovn-controller-bw2xb\" (UID: \"6f8119b2-e38a-494d-967f-5198b83512c7\") " pod="openstack/ovn-controller-bw2xb" Mar 13 14:25:27 crc kubenswrapper[4907]: I0313 14:25:27.155636 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/c1b70392-1240-40d9-8128-e7abe29c8398-var-lib\") pod \"ovn-controller-ovs-vp6f5\" (UID: \"c1b70392-1240-40d9-8128-e7abe29c8398\") " pod="openstack/ovn-controller-ovs-vp6f5" Mar 13 14:25:27 crc kubenswrapper[4907]: I0313 14:25:27.155653 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/6f8119b2-e38a-494d-967f-5198b83512c7-var-run-ovn\") pod \"ovn-controller-bw2xb\" (UID: \"6f8119b2-e38a-494d-967f-5198b83512c7\") " pod="openstack/ovn-controller-bw2xb" Mar 13 14:25:27 crc kubenswrapper[4907]: I0313 14:25:27.257534 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c1b70392-1240-40d9-8128-e7abe29c8398-scripts\") pod \"ovn-controller-ovs-vp6f5\" (UID: \"c1b70392-1240-40d9-8128-e7abe29c8398\") " pod="openstack/ovn-controller-ovs-vp6f5" Mar 13 14:25:27 crc kubenswrapper[4907]: I0313 14:25:27.257594 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f8119b2-e38a-494d-967f-5198b83512c7-combined-ca-bundle\") pod \"ovn-controller-bw2xb\" (UID: \"6f8119b2-e38a-494d-967f-5198b83512c7\") " pod="openstack/ovn-controller-bw2xb" Mar 13 14:25:27 crc kubenswrapper[4907]: I0313 14:25:27.257639 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/6f8119b2-e38a-494d-967f-5198b83512c7-ovn-controller-tls-certs\") pod \"ovn-controller-bw2xb\" (UID: \"6f8119b2-e38a-494d-967f-5198b83512c7\") " pod="openstack/ovn-controller-bw2xb" Mar 13 14:25:27 crc kubenswrapper[4907]: I0313 14:25:27.257675 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6f8119b2-e38a-494d-967f-5198b83512c7-var-run\") pod \"ovn-controller-bw2xb\" (UID: \"6f8119b2-e38a-494d-967f-5198b83512c7\") " pod="openstack/ovn-controller-bw2xb" Mar 13 14:25:27 crc kubenswrapper[4907]: I0313 14:25:27.257712 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/c1b70392-1240-40d9-8128-e7abe29c8398-var-log\") pod \"ovn-controller-ovs-vp6f5\" (UID: \"c1b70392-1240-40d9-8128-e7abe29c8398\") " pod="openstack/ovn-controller-ovs-vp6f5" Mar 13 14:25:27 crc kubenswrapper[4907]: I0313 14:25:27.257740 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/c1b70392-1240-40d9-8128-e7abe29c8398-var-run\") pod \"ovn-controller-ovs-vp6f5\" (UID: \"c1b70392-1240-40d9-8128-e7abe29c8398\") " pod="openstack/ovn-controller-ovs-vp6f5" Mar 13 14:25:27 crc kubenswrapper[4907]: I0313 14:25:27.257764 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/c1b70392-1240-40d9-8128-e7abe29c8398-etc-ovs\") pod \"ovn-controller-ovs-vp6f5\" (UID: \"c1b70392-1240-40d9-8128-e7abe29c8398\") " pod="openstack/ovn-controller-ovs-vp6f5" Mar 13 14:25:27 crc kubenswrapper[4907]: I0313 14:25:27.257785 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4qvcj\" (UniqueName: \"kubernetes.io/projected/c1b70392-1240-40d9-8128-e7abe29c8398-kube-api-access-4qvcj\") pod \"ovn-controller-ovs-vp6f5\" (UID: \"c1b70392-1240-40d9-8128-e7abe29c8398\") " pod="openstack/ovn-controller-ovs-vp6f5" Mar 13 14:25:27 crc kubenswrapper[4907]: I0313 14:25:27.257828 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v2qq5\" (UniqueName: \"kubernetes.io/projected/6f8119b2-e38a-494d-967f-5198b83512c7-kube-api-access-v2qq5\") pod \"ovn-controller-bw2xb\" (UID: \"6f8119b2-e38a-494d-967f-5198b83512c7\") " pod="openstack/ovn-controller-bw2xb" Mar 13 14:25:27 crc kubenswrapper[4907]: I0313 14:25:27.257858 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/6f8119b2-e38a-494d-967f-5198b83512c7-var-log-ovn\") pod \"ovn-controller-bw2xb\" (UID: \"6f8119b2-e38a-494d-967f-5198b83512c7\") " pod="openstack/ovn-controller-bw2xb" Mar 13 14:25:27 crc kubenswrapper[4907]: I0313 14:25:27.257912 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6f8119b2-e38a-494d-967f-5198b83512c7-scripts\") pod \"ovn-controller-bw2xb\" (UID: \"6f8119b2-e38a-494d-967f-5198b83512c7\") " pod="openstack/ovn-controller-bw2xb" Mar 13 14:25:27 crc kubenswrapper[4907]: I0313 14:25:27.257935 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/c1b70392-1240-40d9-8128-e7abe29c8398-var-lib\") pod \"ovn-controller-ovs-vp6f5\" (UID: \"c1b70392-1240-40d9-8128-e7abe29c8398\") " pod="openstack/ovn-controller-ovs-vp6f5" Mar 13 14:25:27 crc kubenswrapper[4907]: I0313 14:25:27.257963 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/6f8119b2-e38a-494d-967f-5198b83512c7-var-run-ovn\") pod \"ovn-controller-bw2xb\" (UID: \"6f8119b2-e38a-494d-967f-5198b83512c7\") " pod="openstack/ovn-controller-bw2xb" Mar 13 14:25:27 crc kubenswrapper[4907]: I0313 14:25:27.258892 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/6f8119b2-e38a-494d-967f-5198b83512c7-var-run-ovn\") pod \"ovn-controller-bw2xb\" (UID: \"6f8119b2-e38a-494d-967f-5198b83512c7\") " pod="openstack/ovn-controller-bw2xb" Mar 13 14:25:27 crc kubenswrapper[4907]: I0313 14:25:27.261238 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/c1b70392-1240-40d9-8128-e7abe29c8398-etc-ovs\") pod \"ovn-controller-ovs-vp6f5\" (UID: \"c1b70392-1240-40d9-8128-e7abe29c8398\") " pod="openstack/ovn-controller-ovs-vp6f5" Mar 13 14:25:27 crc kubenswrapper[4907]: I0313 14:25:27.261323 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c1b70392-1240-40d9-8128-e7abe29c8398-scripts\") pod \"ovn-controller-ovs-vp6f5\" (UID: \"c1b70392-1240-40d9-8128-e7abe29c8398\") " pod="openstack/ovn-controller-ovs-vp6f5" Mar 13 14:25:27 crc kubenswrapper[4907]: I0313 14:25:27.261784 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6f8119b2-e38a-494d-967f-5198b83512c7-var-run\") pod \"ovn-controller-bw2xb\" (UID: \"6f8119b2-e38a-494d-967f-5198b83512c7\") " pod="openstack/ovn-controller-bw2xb" Mar 13 14:25:27 crc kubenswrapper[4907]: I0313 14:25:27.261931 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/6f8119b2-e38a-494d-967f-5198b83512c7-var-log-ovn\") pod \"ovn-controller-bw2xb\" (UID: \"6f8119b2-e38a-494d-967f-5198b83512c7\") " pod="openstack/ovn-controller-bw2xb" Mar 13 14:25:27 crc kubenswrapper[4907]: I0313 14:25:27.262320 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/c1b70392-1240-40d9-8128-e7abe29c8398-var-lib\") pod \"ovn-controller-ovs-vp6f5\" (UID: \"c1b70392-1240-40d9-8128-e7abe29c8398\") " pod="openstack/ovn-controller-ovs-vp6f5" Mar 13 14:25:27 crc kubenswrapper[4907]: I0313 14:25:27.262372 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/c1b70392-1240-40d9-8128-e7abe29c8398-var-run\") pod \"ovn-controller-ovs-vp6f5\" (UID: \"c1b70392-1240-40d9-8128-e7abe29c8398\") " pod="openstack/ovn-controller-ovs-vp6f5" Mar 13 14:25:27 crc kubenswrapper[4907]: I0313 14:25:27.262480 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/c1b70392-1240-40d9-8128-e7abe29c8398-var-log\") pod \"ovn-controller-ovs-vp6f5\" (UID: \"c1b70392-1240-40d9-8128-e7abe29c8398\") " pod="openstack/ovn-controller-ovs-vp6f5" Mar 13 14:25:27 crc kubenswrapper[4907]: I0313 14:25:27.264379 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f8119b2-e38a-494d-967f-5198b83512c7-combined-ca-bundle\") pod \"ovn-controller-bw2xb\" (UID: \"6f8119b2-e38a-494d-967f-5198b83512c7\") " pod="openstack/ovn-controller-bw2xb" Mar 13 14:25:27 crc kubenswrapper[4907]: I0313 14:25:27.264944 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/6f8119b2-e38a-494d-967f-5198b83512c7-ovn-controller-tls-certs\") pod \"ovn-controller-bw2xb\" (UID: \"6f8119b2-e38a-494d-967f-5198b83512c7\") " pod="openstack/ovn-controller-bw2xb" Mar 13 14:25:27 crc kubenswrapper[4907]: I0313 14:25:27.270055 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6f8119b2-e38a-494d-967f-5198b83512c7-scripts\") pod \"ovn-controller-bw2xb\" (UID: \"6f8119b2-e38a-494d-967f-5198b83512c7\") " pod="openstack/ovn-controller-bw2xb" Mar 13 14:25:27 crc kubenswrapper[4907]: I0313 14:25:27.286324 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v2qq5\" (UniqueName: \"kubernetes.io/projected/6f8119b2-e38a-494d-967f-5198b83512c7-kube-api-access-v2qq5\") pod \"ovn-controller-bw2xb\" (UID: \"6f8119b2-e38a-494d-967f-5198b83512c7\") " pod="openstack/ovn-controller-bw2xb" Mar 13 14:25:27 crc kubenswrapper[4907]: I0313 14:25:27.289651 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4qvcj\" (UniqueName: \"kubernetes.io/projected/c1b70392-1240-40d9-8128-e7abe29c8398-kube-api-access-4qvcj\") pod \"ovn-controller-ovs-vp6f5\" (UID: \"c1b70392-1240-40d9-8128-e7abe29c8398\") " pod="openstack/ovn-controller-ovs-vp6f5" Mar 13 14:25:27 crc kubenswrapper[4907]: I0313 14:25:27.426265 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-bw2xb" Mar 13 14:25:27 crc kubenswrapper[4907]: I0313 14:25:27.432397 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-vp6f5" Mar 13 14:25:29 crc kubenswrapper[4907]: I0313 14:25:29.818624 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Mar 13 14:25:29 crc kubenswrapper[4907]: I0313 14:25:29.820072 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Mar 13 14:25:29 crc kubenswrapper[4907]: I0313 14:25:29.822077 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Mar 13 14:25:29 crc kubenswrapper[4907]: I0313 14:25:29.822142 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-k4qll" Mar 13 14:25:29 crc kubenswrapper[4907]: I0313 14:25:29.822577 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Mar 13 14:25:29 crc kubenswrapper[4907]: I0313 14:25:29.822622 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Mar 13 14:25:29 crc kubenswrapper[4907]: I0313 14:25:29.831342 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Mar 13 14:25:29 crc kubenswrapper[4907]: I0313 14:25:29.897580 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c2938e39-f1d8-4cdc-a32c-5d57b8f2034f-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"c2938e39-f1d8-4cdc-a32c-5d57b8f2034f\") " pod="openstack/ovsdbserver-sb-0" Mar 13 14:25:29 crc kubenswrapper[4907]: I0313 14:25:29.897631 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2938e39-f1d8-4cdc-a32c-5d57b8f2034f-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"c2938e39-f1d8-4cdc-a32c-5d57b8f2034f\") " pod="openstack/ovsdbserver-sb-0" Mar 13 14:25:29 crc kubenswrapper[4907]: I0313 14:25:29.897690 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c2938e39-f1d8-4cdc-a32c-5d57b8f2034f-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"c2938e39-f1d8-4cdc-a32c-5d57b8f2034f\") " pod="openstack/ovsdbserver-sb-0" Mar 13 14:25:29 crc kubenswrapper[4907]: I0313 14:25:29.897779 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c2938e39-f1d8-4cdc-a32c-5d57b8f2034f-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"c2938e39-f1d8-4cdc-a32c-5d57b8f2034f\") " pod="openstack/ovsdbserver-sb-0" Mar 13 14:25:29 crc kubenswrapper[4907]: I0313 14:25:29.897800 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c2938e39-f1d8-4cdc-a32c-5d57b8f2034f-config\") pod \"ovsdbserver-sb-0\" (UID: \"c2938e39-f1d8-4cdc-a32c-5d57b8f2034f\") " pod="openstack/ovsdbserver-sb-0" Mar 13 14:25:29 crc kubenswrapper[4907]: I0313 14:25:29.897821 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pxn9s\" (UniqueName: \"kubernetes.io/projected/c2938e39-f1d8-4cdc-a32c-5d57b8f2034f-kube-api-access-pxn9s\") pod \"ovsdbserver-sb-0\" (UID: \"c2938e39-f1d8-4cdc-a32c-5d57b8f2034f\") " pod="openstack/ovsdbserver-sb-0" Mar 13 14:25:29 crc kubenswrapper[4907]: I0313 14:25:29.897987 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c2938e39-f1d8-4cdc-a32c-5d57b8f2034f-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"c2938e39-f1d8-4cdc-a32c-5d57b8f2034f\") " pod="openstack/ovsdbserver-sb-0" Mar 13 14:25:29 crc kubenswrapper[4907]: I0313 14:25:29.898055 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ovsdbserver-sb-0\" (UID: \"c2938e39-f1d8-4cdc-a32c-5d57b8f2034f\") " pod="openstack/ovsdbserver-sb-0" Mar 13 14:25:30 crc kubenswrapper[4907]: I0313 14:25:29.998962 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c2938e39-f1d8-4cdc-a32c-5d57b8f2034f-config\") pod \"ovsdbserver-sb-0\" (UID: \"c2938e39-f1d8-4cdc-a32c-5d57b8f2034f\") " pod="openstack/ovsdbserver-sb-0" Mar 13 14:25:30 crc kubenswrapper[4907]: I0313 14:25:29.999372 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pxn9s\" (UniqueName: \"kubernetes.io/projected/c2938e39-f1d8-4cdc-a32c-5d57b8f2034f-kube-api-access-pxn9s\") pod \"ovsdbserver-sb-0\" (UID: \"c2938e39-f1d8-4cdc-a32c-5d57b8f2034f\") " pod="openstack/ovsdbserver-sb-0" Mar 13 14:25:30 crc kubenswrapper[4907]: I0313 14:25:29.999419 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c2938e39-f1d8-4cdc-a32c-5d57b8f2034f-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"c2938e39-f1d8-4cdc-a32c-5d57b8f2034f\") " pod="openstack/ovsdbserver-sb-0" Mar 13 14:25:30 crc kubenswrapper[4907]: I0313 14:25:29.999446 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ovsdbserver-sb-0\" (UID: \"c2938e39-f1d8-4cdc-a32c-5d57b8f2034f\") " pod="openstack/ovsdbserver-sb-0" Mar 13 14:25:30 crc kubenswrapper[4907]: I0313 14:25:29.999480 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c2938e39-f1d8-4cdc-a32c-5d57b8f2034f-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"c2938e39-f1d8-4cdc-a32c-5d57b8f2034f\") " pod="openstack/ovsdbserver-sb-0" Mar 13 14:25:30 crc kubenswrapper[4907]: I0313 14:25:29.999506 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2938e39-f1d8-4cdc-a32c-5d57b8f2034f-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"c2938e39-f1d8-4cdc-a32c-5d57b8f2034f\") " pod="openstack/ovsdbserver-sb-0" Mar 13 14:25:30 crc kubenswrapper[4907]: I0313 14:25:29.999557 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c2938e39-f1d8-4cdc-a32c-5d57b8f2034f-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"c2938e39-f1d8-4cdc-a32c-5d57b8f2034f\") " pod="openstack/ovsdbserver-sb-0" Mar 13 14:25:30 crc kubenswrapper[4907]: I0313 14:25:29.999583 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c2938e39-f1d8-4cdc-a32c-5d57b8f2034f-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"c2938e39-f1d8-4cdc-a32c-5d57b8f2034f\") " pod="openstack/ovsdbserver-sb-0" Mar 13 14:25:30 crc kubenswrapper[4907]: I0313 14:25:30.000120 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c2938e39-f1d8-4cdc-a32c-5d57b8f2034f-config\") pod \"ovsdbserver-sb-0\" (UID: \"c2938e39-f1d8-4cdc-a32c-5d57b8f2034f\") " pod="openstack/ovsdbserver-sb-0" Mar 13 14:25:30 crc kubenswrapper[4907]: I0313 14:25:30.000360 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c2938e39-f1d8-4cdc-a32c-5d57b8f2034f-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"c2938e39-f1d8-4cdc-a32c-5d57b8f2034f\") " pod="openstack/ovsdbserver-sb-0" Mar 13 14:25:30 crc kubenswrapper[4907]: I0313 14:25:30.000624 4907 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ovsdbserver-sb-0\" (UID: \"c2938e39-f1d8-4cdc-a32c-5d57b8f2034f\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/ovsdbserver-sb-0" Mar 13 14:25:30 crc kubenswrapper[4907]: I0313 14:25:30.007124 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c2938e39-f1d8-4cdc-a32c-5d57b8f2034f-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"c2938e39-f1d8-4cdc-a32c-5d57b8f2034f\") " pod="openstack/ovsdbserver-sb-0" Mar 13 14:25:30 crc kubenswrapper[4907]: I0313 14:25:30.015980 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c2938e39-f1d8-4cdc-a32c-5d57b8f2034f-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"c2938e39-f1d8-4cdc-a32c-5d57b8f2034f\") " pod="openstack/ovsdbserver-sb-0" Mar 13 14:25:30 crc kubenswrapper[4907]: I0313 14:25:30.025681 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2938e39-f1d8-4cdc-a32c-5d57b8f2034f-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"c2938e39-f1d8-4cdc-a32c-5d57b8f2034f\") " pod="openstack/ovsdbserver-sb-0" Mar 13 14:25:30 crc kubenswrapper[4907]: I0313 14:25:30.029304 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c2938e39-f1d8-4cdc-a32c-5d57b8f2034f-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"c2938e39-f1d8-4cdc-a32c-5d57b8f2034f\") " pod="openstack/ovsdbserver-sb-0" Mar 13 14:25:30 crc kubenswrapper[4907]: I0313 14:25:30.029487 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pxn9s\" (UniqueName: \"kubernetes.io/projected/c2938e39-f1d8-4cdc-a32c-5d57b8f2034f-kube-api-access-pxn9s\") pod \"ovsdbserver-sb-0\" (UID: \"c2938e39-f1d8-4cdc-a32c-5d57b8f2034f\") " pod="openstack/ovsdbserver-sb-0" Mar 13 14:25:30 crc kubenswrapper[4907]: I0313 14:25:30.053282 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ovsdbserver-sb-0\" (UID: \"c2938e39-f1d8-4cdc-a32c-5d57b8f2034f\") " pod="openstack/ovsdbserver-sb-0" Mar 13 14:25:30 crc kubenswrapper[4907]: I0313 14:25:30.153842 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Mar 13 14:25:30 crc kubenswrapper[4907]: I0313 14:25:30.391719 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Mar 13 14:25:30 crc kubenswrapper[4907]: I0313 14:25:30.499411 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Mar 13 14:25:31 crc kubenswrapper[4907]: I0313 14:25:31.044737 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"e3c34454-315e-4821-ab25-b0f331a0d521","Type":"ContainerStarted","Data":"0dddf7e485bfeda5a825f940bee1a34df372878429700a5d41310e0d0412399b"} Mar 13 14:25:31 crc kubenswrapper[4907]: I0313 14:25:31.045710 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"58ab9689-3ab2-42f2-ad56-beb22e29f8da","Type":"ContainerStarted","Data":"be11df2bf36396e7ed2d074eed97b3e79df7e43e0e8a1a8dec1ef8739586a724"} Mar 13 14:25:31 crc kubenswrapper[4907]: W0313 14:25:31.058867 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc4411bd2_c555_433a_9015_f623948b1401.slice/crio-da35a91fc58ce0787798d501f92eec9bbafae6711281a9caca5018e3a65c78c8 WatchSource:0}: Error finding container da35a91fc58ce0787798d501f92eec9bbafae6711281a9caca5018e3a65c78c8: Status 404 returned error can't find the container with id da35a91fc58ce0787798d501f92eec9bbafae6711281a9caca5018e3a65c78c8 Mar 13 14:25:31 crc kubenswrapper[4907]: E0313 14:25:31.111627 4907 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:fbb5be29e9e4fa11f0743e7f74f2e80dcc7445d24770709ea0e038147f752c51" Mar 13 14:25:31 crc kubenswrapper[4907]: E0313 14:25:31.111866 4907 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:fbb5be29e9e4fa11f0743e7f74f2e80dcc7445d24770709ea0e038147f752c51,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-qxpcr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-5448ff6dc7-pgq4q_openstack(152380e7-1c7a-4c89-a14e-dd179009f6af): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 13 14:25:31 crc kubenswrapper[4907]: E0313 14:25:31.115359 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-5448ff6dc7-pgq4q" podUID="152380e7-1c7a-4c89-a14e-dd179009f6af" Mar 13 14:25:31 crc kubenswrapper[4907]: E0313 14:25:31.318009 4907 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:fbb5be29e9e4fa11f0743e7f74f2e80dcc7445d24770709ea0e038147f752c51" Mar 13 14:25:31 crc kubenswrapper[4907]: E0313 14:25:31.318695 4907 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:fbb5be29e9e4fa11f0743e7f74f2e80dcc7445d24770709ea0e038147f752c51,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-br82q,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-64696987c5-gdk7n_openstack(874829e6-8bd5-4040-a493-317a2af4fa35): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 13 14:25:31 crc kubenswrapper[4907]: E0313 14:25:31.321338 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-64696987c5-gdk7n" podUID="874829e6-8bd5-4040-a493-317a2af4fa35" Mar 13 14:25:31 crc kubenswrapper[4907]: I0313 14:25:31.510213 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Mar 13 14:25:31 crc kubenswrapper[4907]: I0313 14:25:31.571193 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 13 14:25:31 crc kubenswrapper[4907]: I0313 14:25:31.587805 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Mar 13 14:25:31 crc kubenswrapper[4907]: I0313 14:25:31.888098 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-vp6f5"] Mar 13 14:25:31 crc kubenswrapper[4907]: W0313 14:25:31.953756 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6f8119b2_e38a_494d_967f_5198b83512c7.slice/crio-4f06fa5373719502d5e6526674b7a5bdf213bd13582c27f9af62c582539db3b6 WatchSource:0}: Error finding container 4f06fa5373719502d5e6526674b7a5bdf213bd13582c27f9af62c582539db3b6: Status 404 returned error can't find the container with id 4f06fa5373719502d5e6526674b7a5bdf213bd13582c27f9af62c582539db3b6 Mar 13 14:25:31 crc kubenswrapper[4907]: I0313 14:25:31.958247 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-bw2xb"] Mar 13 14:25:32 crc kubenswrapper[4907]: W0313 14:25:32.007774 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod056dd756_0d7b_471b_9929_f622d05ad606.slice/crio-8f94280adbcbea9567f862d27736043270ac488194f6b0213cdb79c3aa5a9dee WatchSource:0}: Error finding container 8f94280adbcbea9567f862d27736043270ac488194f6b0213cdb79c3aa5a9dee: Status 404 returned error can't find the container with id 8f94280adbcbea9567f862d27736043270ac488194f6b0213cdb79c3aa5a9dee Mar 13 14:25:32 crc kubenswrapper[4907]: I0313 14:25:32.011879 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Mar 13 14:25:32 crc kubenswrapper[4907]: I0313 14:25:32.088061 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"a7f6877c-1e9e-4e17-803d-90efa7d66469","Type":"ContainerStarted","Data":"3893671130f88340bc9f4b2c1c7ae2e64ebc3c35392832b178f2caf4f860f9fd"} Mar 13 14:25:32 crc kubenswrapper[4907]: I0313 14:25:32.090267 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"286c82e8-b74a-49d7-a355-ac074aace10e","Type":"ContainerStarted","Data":"84aa3724ad3336159381b40fcc1f3bf130caa7609c9affeddec48f79002eb26c"} Mar 13 14:25:32 crc kubenswrapper[4907]: I0313 14:25:32.093464 4907 generic.go:334] "Generic (PLEG): container finished" podID="b921f7f7-cd29-4ac8-82e5-c69a8f9e2bc0" containerID="69b8616a20bf2d26e83bca32550971f5d3206d2a8a1805adafdfe85310db55b3" exitCode=0 Mar 13 14:25:32 crc kubenswrapper[4907]: I0313 14:25:32.093547 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-658f55c9f5-dp8jf" event={"ID":"b921f7f7-cd29-4ac8-82e5-c69a8f9e2bc0","Type":"ContainerDied","Data":"69b8616a20bf2d26e83bca32550971f5d3206d2a8a1805adafdfe85310db55b3"} Mar 13 14:25:32 crc kubenswrapper[4907]: I0313 14:25:32.100807 4907 generic.go:334] "Generic (PLEG): container finished" podID="779b1958-f7ed-4c53-a4ff-a6c4a803bebf" containerID="cd0379440a1db9f081229f5637156055839f155515dc1039d28132ecd64ab32a" exitCode=0 Mar 13 14:25:32 crc kubenswrapper[4907]: I0313 14:25:32.105047 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54b5dffb47-wttzd" event={"ID":"779b1958-f7ed-4c53-a4ff-a6c4a803bebf","Type":"ContainerDied","Data":"cd0379440a1db9f081229f5637156055839f155515dc1039d28132ecd64ab32a"} Mar 13 14:25:32 crc kubenswrapper[4907]: I0313 14:25:32.110220 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Mar 13 14:25:32 crc kubenswrapper[4907]: I0313 14:25:32.116020 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerStarted","Data":"1f79a88cd4431d2b6772ef1dc1fb18ce20d965a3a936a7157e03bc28ca239c57"} Mar 13 14:25:32 crc kubenswrapper[4907]: I0313 14:25:32.125171 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"c4411bd2-c555-433a-9015-f623948b1401","Type":"ContainerStarted","Data":"da35a91fc58ce0787798d501f92eec9bbafae6711281a9caca5018e3a65c78c8"} Mar 13 14:25:32 crc kubenswrapper[4907]: W0313 14:25:32.131065 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc2938e39_f1d8_4cdc_a32c_5d57b8f2034f.slice/crio-c836a111eed36848627fde431a8ac117cdf68714d43b9b328bad3f2315e96e0e WatchSource:0}: Error finding container c836a111eed36848627fde431a8ac117cdf68714d43b9b328bad3f2315e96e0e: Status 404 returned error can't find the container with id c836a111eed36848627fde431a8ac117cdf68714d43b9b328bad3f2315e96e0e Mar 13 14:25:32 crc kubenswrapper[4907]: I0313 14:25:32.136948 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"cacb281b-ee3b-42b6-8754-d07e6ea6aa4b","Type":"ContainerStarted","Data":"c174ea292d3346503480cf55d66ad04b5cc26c89fac549263ee565359ea5a081"} Mar 13 14:25:32 crc kubenswrapper[4907]: I0313 14:25:32.139185 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-bw2xb" event={"ID":"6f8119b2-e38a-494d-967f-5198b83512c7","Type":"ContainerStarted","Data":"4f06fa5373719502d5e6526674b7a5bdf213bd13582c27f9af62c582539db3b6"} Mar 13 14:25:32 crc kubenswrapper[4907]: I0313 14:25:32.141434 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"056dd756-0d7b-471b-9929-f622d05ad606","Type":"ContainerStarted","Data":"8f94280adbcbea9567f862d27736043270ac488194f6b0213cdb79c3aa5a9dee"} Mar 13 14:25:32 crc kubenswrapper[4907]: I0313 14:25:32.147091 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-vp6f5" event={"ID":"c1b70392-1240-40d9-8128-e7abe29c8398","Type":"ContainerStarted","Data":"99abe9d38052154162343b52b5e60bd6560b49e824682c932c7a2a64adbb5d61"} Mar 13 14:25:32 crc kubenswrapper[4907]: I0313 14:25:32.502195 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5448ff6dc7-pgq4q" Mar 13 14:25:32 crc kubenswrapper[4907]: I0313 14:25:32.570497 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64696987c5-gdk7n" Mar 13 14:25:32 crc kubenswrapper[4907]: I0313 14:25:32.579456 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/152380e7-1c7a-4c89-a14e-dd179009f6af-config\") pod \"152380e7-1c7a-4c89-a14e-dd179009f6af\" (UID: \"152380e7-1c7a-4c89-a14e-dd179009f6af\") " Mar 13 14:25:32 crc kubenswrapper[4907]: I0313 14:25:32.579924 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qxpcr\" (UniqueName: \"kubernetes.io/projected/152380e7-1c7a-4c89-a14e-dd179009f6af-kube-api-access-qxpcr\") pod \"152380e7-1c7a-4c89-a14e-dd179009f6af\" (UID: \"152380e7-1c7a-4c89-a14e-dd179009f6af\") " Mar 13 14:25:32 crc kubenswrapper[4907]: I0313 14:25:32.579979 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/152380e7-1c7a-4c89-a14e-dd179009f6af-config" (OuterVolumeSpecName: "config") pod "152380e7-1c7a-4c89-a14e-dd179009f6af" (UID: "152380e7-1c7a-4c89-a14e-dd179009f6af"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:25:32 crc kubenswrapper[4907]: I0313 14:25:32.580703 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/152380e7-1c7a-4c89-a14e-dd179009f6af-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:25:32 crc kubenswrapper[4907]: I0313 14:25:32.589071 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/152380e7-1c7a-4c89-a14e-dd179009f6af-kube-api-access-qxpcr" (OuterVolumeSpecName: "kube-api-access-qxpcr") pod "152380e7-1c7a-4c89-a14e-dd179009f6af" (UID: "152380e7-1c7a-4c89-a14e-dd179009f6af"). InnerVolumeSpecName "kube-api-access-qxpcr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:25:32 crc kubenswrapper[4907]: I0313 14:25:32.681462 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-br82q\" (UniqueName: \"kubernetes.io/projected/874829e6-8bd5-4040-a493-317a2af4fa35-kube-api-access-br82q\") pod \"874829e6-8bd5-4040-a493-317a2af4fa35\" (UID: \"874829e6-8bd5-4040-a493-317a2af4fa35\") " Mar 13 14:25:32 crc kubenswrapper[4907]: I0313 14:25:32.681551 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/874829e6-8bd5-4040-a493-317a2af4fa35-config\") pod \"874829e6-8bd5-4040-a493-317a2af4fa35\" (UID: \"874829e6-8bd5-4040-a493-317a2af4fa35\") " Mar 13 14:25:32 crc kubenswrapper[4907]: I0313 14:25:32.681608 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/874829e6-8bd5-4040-a493-317a2af4fa35-dns-svc\") pod \"874829e6-8bd5-4040-a493-317a2af4fa35\" (UID: \"874829e6-8bd5-4040-a493-317a2af4fa35\") " Mar 13 14:25:32 crc kubenswrapper[4907]: I0313 14:25:32.682274 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qxpcr\" (UniqueName: \"kubernetes.io/projected/152380e7-1c7a-4c89-a14e-dd179009f6af-kube-api-access-qxpcr\") on node \"crc\" DevicePath \"\"" Mar 13 14:25:32 crc kubenswrapper[4907]: I0313 14:25:32.682278 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/874829e6-8bd5-4040-a493-317a2af4fa35-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "874829e6-8bd5-4040-a493-317a2af4fa35" (UID: "874829e6-8bd5-4040-a493-317a2af4fa35"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:25:32 crc kubenswrapper[4907]: I0313 14:25:32.682306 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/874829e6-8bd5-4040-a493-317a2af4fa35-config" (OuterVolumeSpecName: "config") pod "874829e6-8bd5-4040-a493-317a2af4fa35" (UID: "874829e6-8bd5-4040-a493-317a2af4fa35"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:25:32 crc kubenswrapper[4907]: I0313 14:25:32.685463 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/874829e6-8bd5-4040-a493-317a2af4fa35-kube-api-access-br82q" (OuterVolumeSpecName: "kube-api-access-br82q") pod "874829e6-8bd5-4040-a493-317a2af4fa35" (UID: "874829e6-8bd5-4040-a493-317a2af4fa35"). InnerVolumeSpecName "kube-api-access-br82q". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:25:32 crc kubenswrapper[4907]: I0313 14:25:32.784177 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-br82q\" (UniqueName: \"kubernetes.io/projected/874829e6-8bd5-4040-a493-317a2af4fa35-kube-api-access-br82q\") on node \"crc\" DevicePath \"\"" Mar 13 14:25:32 crc kubenswrapper[4907]: I0313 14:25:32.784406 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/874829e6-8bd5-4040-a493-317a2af4fa35-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:25:32 crc kubenswrapper[4907]: I0313 14:25:32.784416 4907 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/874829e6-8bd5-4040-a493-317a2af4fa35-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 13 14:25:33 crc kubenswrapper[4907]: I0313 14:25:33.162299 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5448ff6dc7-pgq4q" Mar 13 14:25:33 crc kubenswrapper[4907]: I0313 14:25:33.162294 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5448ff6dc7-pgq4q" event={"ID":"152380e7-1c7a-4c89-a14e-dd179009f6af","Type":"ContainerDied","Data":"5fce8f6cb27b89ad9b7731f40d4a1c6f21ae57078b5536d25290f96168055e8b"} Mar 13 14:25:33 crc kubenswrapper[4907]: I0313 14:25:33.165931 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54b5dffb47-wttzd" event={"ID":"779b1958-f7ed-4c53-a4ff-a6c4a803bebf","Type":"ContainerStarted","Data":"923c944515f9f62d3a3c73bba3a0205c7dccd6753e7da2435a2be1308a3d0b7c"} Mar 13 14:25:33 crc kubenswrapper[4907]: I0313 14:25:33.167370 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"c2938e39-f1d8-4cdc-a32c-5d57b8f2034f","Type":"ContainerStarted","Data":"c836a111eed36848627fde431a8ac117cdf68714d43b9b328bad3f2315e96e0e"} Mar 13 14:25:33 crc kubenswrapper[4907]: I0313 14:25:33.168799 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64696987c5-gdk7n" event={"ID":"874829e6-8bd5-4040-a493-317a2af4fa35","Type":"ContainerDied","Data":"d7f24309913163b7f0ac4df94b2d551597928123506fa9a0c942b473d75c1940"} Mar 13 14:25:33 crc kubenswrapper[4907]: I0313 14:25:33.168830 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64696987c5-gdk7n" Mar 13 14:25:33 crc kubenswrapper[4907]: I0313 14:25:33.171842 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-658f55c9f5-dp8jf" event={"ID":"b921f7f7-cd29-4ac8-82e5-c69a8f9e2bc0","Type":"ContainerStarted","Data":"9cfe7128fc9a7f1b95d9545bb2adc26bb8cdaafa4e69d6b740d1e1ef7d29828e"} Mar 13 14:25:33 crc kubenswrapper[4907]: I0313 14:25:33.187867 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-54b5dffb47-wttzd" podStartSLOduration=6.205187355 podStartE2EDuration="16.187837705s" podCreationTimestamp="2026-03-13 14:25:17 +0000 UTC" firstStartedPulling="2026-03-13 14:25:21.437749451 +0000 UTC m=+1220.337537140" lastFinishedPulling="2026-03-13 14:25:31.420399801 +0000 UTC m=+1230.320187490" observedRunningTime="2026-03-13 14:25:33.18586266 +0000 UTC m=+1232.085650359" watchObservedRunningTime="2026-03-13 14:25:33.187837705 +0000 UTC m=+1232.087625394" Mar 13 14:25:33 crc kubenswrapper[4907]: I0313 14:25:33.211313 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-658f55c9f5-dp8jf" podStartSLOduration=3.465015686 podStartE2EDuration="17.211292199s" podCreationTimestamp="2026-03-13 14:25:16 +0000 UTC" firstStartedPulling="2026-03-13 14:25:17.664366596 +0000 UTC m=+1216.564154285" lastFinishedPulling="2026-03-13 14:25:31.410643119 +0000 UTC m=+1230.310430798" observedRunningTime="2026-03-13 14:25:33.203688886 +0000 UTC m=+1232.103476595" watchObservedRunningTime="2026-03-13 14:25:33.211292199 +0000 UTC m=+1232.111079888" Mar 13 14:25:33 crc kubenswrapper[4907]: I0313 14:25:33.270295 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-64696987c5-gdk7n"] Mar 13 14:25:33 crc kubenswrapper[4907]: I0313 14:25:33.284703 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-64696987c5-gdk7n"] Mar 13 14:25:33 crc kubenswrapper[4907]: I0313 14:25:33.299284 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5448ff6dc7-pgq4q"] Mar 13 14:25:33 crc kubenswrapper[4907]: I0313 14:25:33.305708 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5448ff6dc7-pgq4q"] Mar 13 14:25:33 crc kubenswrapper[4907]: I0313 14:25:33.810157 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="152380e7-1c7a-4c89-a14e-dd179009f6af" path="/var/lib/kubelet/pods/152380e7-1c7a-4c89-a14e-dd179009f6af/volumes" Mar 13 14:25:33 crc kubenswrapper[4907]: I0313 14:25:33.810591 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="874829e6-8bd5-4040-a493-317a2af4fa35" path="/var/lib/kubelet/pods/874829e6-8bd5-4040-a493-317a2af4fa35/volumes" Mar 13 14:25:34 crc kubenswrapper[4907]: I0313 14:25:34.178959 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-54b5dffb47-wttzd" Mar 13 14:25:34 crc kubenswrapper[4907]: I0313 14:25:34.179005 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-658f55c9f5-dp8jf" Mar 13 14:25:37 crc kubenswrapper[4907]: I0313 14:25:37.111538 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-658f55c9f5-dp8jf" Mar 13 14:25:37 crc kubenswrapper[4907]: I0313 14:25:37.553101 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-54b5dffb47-wttzd" Mar 13 14:25:37 crc kubenswrapper[4907]: I0313 14:25:37.621790 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-658f55c9f5-dp8jf"] Mar 13 14:25:37 crc kubenswrapper[4907]: I0313 14:25:37.622214 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-658f55c9f5-dp8jf" podUID="b921f7f7-cd29-4ac8-82e5-c69a8f9e2bc0" containerName="dnsmasq-dns" containerID="cri-o://9cfe7128fc9a7f1b95d9545bb2adc26bb8cdaafa4e69d6b740d1e1ef7d29828e" gracePeriod=10 Mar 13 14:25:38 crc kubenswrapper[4907]: I0313 14:25:38.216003 4907 generic.go:334] "Generic (PLEG): container finished" podID="b921f7f7-cd29-4ac8-82e5-c69a8f9e2bc0" containerID="9cfe7128fc9a7f1b95d9545bb2adc26bb8cdaafa4e69d6b740d1e1ef7d29828e" exitCode=0 Mar 13 14:25:38 crc kubenswrapper[4907]: I0313 14:25:38.216050 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-658f55c9f5-dp8jf" event={"ID":"b921f7f7-cd29-4ac8-82e5-c69a8f9e2bc0","Type":"ContainerDied","Data":"9cfe7128fc9a7f1b95d9545bb2adc26bb8cdaafa4e69d6b740d1e1ef7d29828e"} Mar 13 14:25:40 crc kubenswrapper[4907]: I0313 14:25:40.216612 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-658f55c9f5-dp8jf" Mar 13 14:25:40 crc kubenswrapper[4907]: I0313 14:25:40.278422 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-658f55c9f5-dp8jf" event={"ID":"b921f7f7-cd29-4ac8-82e5-c69a8f9e2bc0","Type":"ContainerDied","Data":"00577cc935dfba6c223104c8f045d104f6284aba771cb8fe435ed068373d1547"} Mar 13 14:25:40 crc kubenswrapper[4907]: I0313 14:25:40.278466 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-658f55c9f5-dp8jf" Mar 13 14:25:40 crc kubenswrapper[4907]: I0313 14:25:40.278476 4907 scope.go:117] "RemoveContainer" containerID="9cfe7128fc9a7f1b95d9545bb2adc26bb8cdaafa4e69d6b740d1e1ef7d29828e" Mar 13 14:25:40 crc kubenswrapper[4907]: I0313 14:25:40.311909 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b921f7f7-cd29-4ac8-82e5-c69a8f9e2bc0-config\") pod \"b921f7f7-cd29-4ac8-82e5-c69a8f9e2bc0\" (UID: \"b921f7f7-cd29-4ac8-82e5-c69a8f9e2bc0\") " Mar 13 14:25:40 crc kubenswrapper[4907]: I0313 14:25:40.312338 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-684gn\" (UniqueName: \"kubernetes.io/projected/b921f7f7-cd29-4ac8-82e5-c69a8f9e2bc0-kube-api-access-684gn\") pod \"b921f7f7-cd29-4ac8-82e5-c69a8f9e2bc0\" (UID: \"b921f7f7-cd29-4ac8-82e5-c69a8f9e2bc0\") " Mar 13 14:25:40 crc kubenswrapper[4907]: I0313 14:25:40.312478 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b921f7f7-cd29-4ac8-82e5-c69a8f9e2bc0-dns-svc\") pod \"b921f7f7-cd29-4ac8-82e5-c69a8f9e2bc0\" (UID: \"b921f7f7-cd29-4ac8-82e5-c69a8f9e2bc0\") " Mar 13 14:25:40 crc kubenswrapper[4907]: I0313 14:25:40.316202 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b921f7f7-cd29-4ac8-82e5-c69a8f9e2bc0-kube-api-access-684gn" (OuterVolumeSpecName: "kube-api-access-684gn") pod "b921f7f7-cd29-4ac8-82e5-c69a8f9e2bc0" (UID: "b921f7f7-cd29-4ac8-82e5-c69a8f9e2bc0"). InnerVolumeSpecName "kube-api-access-684gn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:25:40 crc kubenswrapper[4907]: I0313 14:25:40.361130 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b921f7f7-cd29-4ac8-82e5-c69a8f9e2bc0-config" (OuterVolumeSpecName: "config") pod "b921f7f7-cd29-4ac8-82e5-c69a8f9e2bc0" (UID: "b921f7f7-cd29-4ac8-82e5-c69a8f9e2bc0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:25:40 crc kubenswrapper[4907]: I0313 14:25:40.369234 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b921f7f7-cd29-4ac8-82e5-c69a8f9e2bc0-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b921f7f7-cd29-4ac8-82e5-c69a8f9e2bc0" (UID: "b921f7f7-cd29-4ac8-82e5-c69a8f9e2bc0"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:25:40 crc kubenswrapper[4907]: I0313 14:25:40.413600 4907 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b921f7f7-cd29-4ac8-82e5-c69a8f9e2bc0-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 13 14:25:40 crc kubenswrapper[4907]: I0313 14:25:40.413627 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b921f7f7-cd29-4ac8-82e5-c69a8f9e2bc0-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:25:40 crc kubenswrapper[4907]: I0313 14:25:40.413637 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-684gn\" (UniqueName: \"kubernetes.io/projected/b921f7f7-cd29-4ac8-82e5-c69a8f9e2bc0-kube-api-access-684gn\") on node \"crc\" DevicePath \"\"" Mar 13 14:25:40 crc kubenswrapper[4907]: I0313 14:25:40.545639 4907 scope.go:117] "RemoveContainer" containerID="69b8616a20bf2d26e83bca32550971f5d3206d2a8a1805adafdfe85310db55b3" Mar 13 14:25:40 crc kubenswrapper[4907]: I0313 14:25:40.613620 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-658f55c9f5-dp8jf"] Mar 13 14:25:40 crc kubenswrapper[4907]: I0313 14:25:40.623586 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-658f55c9f5-dp8jf"] Mar 13 14:25:41 crc kubenswrapper[4907]: I0313 14:25:41.314743 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"c2938e39-f1d8-4cdc-a32c-5d57b8f2034f","Type":"ContainerStarted","Data":"432aefb6d80f0522ac7d6d516626fd7ea34147a98d43c3ddba905f2a1bb074ad"} Mar 13 14:25:41 crc kubenswrapper[4907]: I0313 14:25:41.316484 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"c4411bd2-c555-433a-9015-f623948b1401","Type":"ContainerStarted","Data":"1953abb8d3f9054081c314893bde9d173fee7430bf61ccd6ae11b908a7002729"} Mar 13 14:25:41 crc kubenswrapper[4907]: I0313 14:25:41.321865 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"cacb281b-ee3b-42b6-8754-d07e6ea6aa4b","Type":"ContainerStarted","Data":"65475a4b4006963b7fec1fa34278a44b7bf08cbcf9c088a764d7d29c870c630c"} Mar 13 14:25:41 crc kubenswrapper[4907]: I0313 14:25:41.322465 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Mar 13 14:25:41 crc kubenswrapper[4907]: I0313 14:25:41.324712 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"58ab9689-3ab2-42f2-ad56-beb22e29f8da","Type":"ContainerStarted","Data":"62c42ba7e98c9bfe072975553c668c323179c088fa7a92a9812e075126aebb86"} Mar 13 14:25:41 crc kubenswrapper[4907]: I0313 14:25:41.386725 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=13.087687758 podStartE2EDuration="21.38670885s" podCreationTimestamp="2026-03-13 14:25:20 +0000 UTC" firstStartedPulling="2026-03-13 14:25:31.523425281 +0000 UTC m=+1230.423212980" lastFinishedPulling="2026-03-13 14:25:39.822446383 +0000 UTC m=+1238.722234072" observedRunningTime="2026-03-13 14:25:41.37977035 +0000 UTC m=+1240.279558039" watchObservedRunningTime="2026-03-13 14:25:41.38670885 +0000 UTC m=+1240.286496539" Mar 13 14:25:41 crc kubenswrapper[4907]: I0313 14:25:41.806875 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b921f7f7-cd29-4ac8-82e5-c69a8f9e2bc0" path="/var/lib/kubelet/pods/b921f7f7-cd29-4ac8-82e5-c69a8f9e2bc0/volumes" Mar 13 14:25:42 crc kubenswrapper[4907]: I0313 14:25:42.340109 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-bw2xb" event={"ID":"6f8119b2-e38a-494d-967f-5198b83512c7","Type":"ContainerStarted","Data":"df7f17b832ba26f166592b0696a663ded9a2e03678601f86699f09b76e050e58"} Mar 13 14:25:42 crc kubenswrapper[4907]: I0313 14:25:42.348160 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"286c82e8-b74a-49d7-a355-ac074aace10e","Type":"ContainerStarted","Data":"275b76780d99e94e2c958194ce29198632d04fd0fe0f0668ff5025ad387fc2f7"} Mar 13 14:25:42 crc kubenswrapper[4907]: I0313 14:25:42.354566 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"056dd756-0d7b-471b-9929-f622d05ad606","Type":"ContainerStarted","Data":"46f9fcec5ce705120f46ecb4deec4665885423d12b23cabce7f929fa5941d7ab"} Mar 13 14:25:42 crc kubenswrapper[4907]: I0313 14:25:42.356976 4907 generic.go:334] "Generic (PLEG): container finished" podID="c1b70392-1240-40d9-8128-e7abe29c8398" containerID="4cf539e8b8cde03127504ad0dcb18c074f6f0c3dab53805c59e991c594c32a31" exitCode=0 Mar 13 14:25:42 crc kubenswrapper[4907]: I0313 14:25:42.357023 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-vp6f5" event={"ID":"c1b70392-1240-40d9-8128-e7abe29c8398","Type":"ContainerDied","Data":"4cf539e8b8cde03127504ad0dcb18c074f6f0c3dab53805c59e991c594c32a31"} Mar 13 14:25:42 crc kubenswrapper[4907]: I0313 14:25:42.368660 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"a7f6877c-1e9e-4e17-803d-90efa7d66469","Type":"ContainerStarted","Data":"f5792262fabc2fac4148b46830f3f0bdf740121dbf4200684e158ad3432a779c"} Mar 13 14:25:42 crc kubenswrapper[4907]: I0313 14:25:42.487985 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=10.968026545 podStartE2EDuration="20.487961632s" podCreationTimestamp="2026-03-13 14:25:22 +0000 UTC" firstStartedPulling="2026-03-13 14:25:31.605871698 +0000 UTC m=+1230.505659387" lastFinishedPulling="2026-03-13 14:25:41.125806785 +0000 UTC m=+1240.025594474" observedRunningTime="2026-03-13 14:25:42.474832922 +0000 UTC m=+1241.374620611" watchObservedRunningTime="2026-03-13 14:25:42.487961632 +0000 UTC m=+1241.387749321" Mar 13 14:25:43 crc kubenswrapper[4907]: I0313 14:25:43.364604 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Mar 13 14:25:43 crc kubenswrapper[4907]: I0313 14:25:43.378898 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"e3c34454-315e-4821-ab25-b0f331a0d521","Type":"ContainerStarted","Data":"987e5ce0ff0b4c3691f9df5efc31a14e40c988230f0cf14b4b06e00cd7129b38"} Mar 13 14:25:43 crc kubenswrapper[4907]: I0313 14:25:43.383968 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-vp6f5" event={"ID":"c1b70392-1240-40d9-8128-e7abe29c8398","Type":"ContainerStarted","Data":"265b6b6fe5f95195cdc86c65456998ad327fc7ff3a6641300cd6ac00798eff4f"} Mar 13 14:25:43 crc kubenswrapper[4907]: I0313 14:25:43.384024 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-vp6f5" event={"ID":"c1b70392-1240-40d9-8128-e7abe29c8398","Type":"ContainerStarted","Data":"6b0b25079cc099dabe922ac89fa41ae76d2f01797aac9f273b042e242459c574"} Mar 13 14:25:43 crc kubenswrapper[4907]: I0313 14:25:43.384572 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-vp6f5" Mar 13 14:25:43 crc kubenswrapper[4907]: I0313 14:25:43.384756 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-vp6f5" Mar 13 14:25:43 crc kubenswrapper[4907]: I0313 14:25:43.435789 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-bw2xb" podStartSLOduration=7.898204665 podStartE2EDuration="16.435768591s" podCreationTimestamp="2026-03-13 14:25:27 +0000 UTC" firstStartedPulling="2026-03-13 14:25:31.990189186 +0000 UTC m=+1230.889976875" lastFinishedPulling="2026-03-13 14:25:40.527753112 +0000 UTC m=+1239.427540801" observedRunningTime="2026-03-13 14:25:43.431262247 +0000 UTC m=+1242.331049936" watchObservedRunningTime="2026-03-13 14:25:43.435768591 +0000 UTC m=+1242.335556280" Mar 13 14:25:43 crc kubenswrapper[4907]: I0313 14:25:43.449443 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-vp6f5" podStartSLOduration=8.118672503 podStartE2EDuration="16.449426726s" podCreationTimestamp="2026-03-13 14:25:27 +0000 UTC" firstStartedPulling="2026-03-13 14:25:31.898159222 +0000 UTC m=+1230.797946911" lastFinishedPulling="2026-03-13 14:25:40.228913435 +0000 UTC m=+1239.128701134" observedRunningTime="2026-03-13 14:25:43.448239533 +0000 UTC m=+1242.348027232" watchObservedRunningTime="2026-03-13 14:25:43.449426726 +0000 UTC m=+1242.349214415" Mar 13 14:25:45 crc kubenswrapper[4907]: I0313 14:25:45.400598 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"056dd756-0d7b-471b-9929-f622d05ad606","Type":"ContainerStarted","Data":"f294f1dfa23d177ca2e8f5511b6bd495bd3146c4baee6d3b95ccd81e1b4496fb"} Mar 13 14:25:45 crc kubenswrapper[4907]: I0313 14:25:45.404294 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"c2938e39-f1d8-4cdc-a32c-5d57b8f2034f","Type":"ContainerStarted","Data":"1ff45617937d399a62e2b1c7ef156ccf7052d51f57fba51d267daa3e49810ce6"} Mar 13 14:25:45 crc kubenswrapper[4907]: I0313 14:25:45.423474 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=7.513472542 podStartE2EDuration="20.423452586s" podCreationTimestamp="2026-03-13 14:25:25 +0000 UTC" firstStartedPulling="2026-03-13 14:25:32.033185113 +0000 UTC m=+1230.932972802" lastFinishedPulling="2026-03-13 14:25:44.943165157 +0000 UTC m=+1243.842952846" observedRunningTime="2026-03-13 14:25:45.420465834 +0000 UTC m=+1244.320253533" watchObservedRunningTime="2026-03-13 14:25:45.423452586 +0000 UTC m=+1244.323240275" Mar 13 14:25:45 crc kubenswrapper[4907]: I0313 14:25:45.442318 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=4.623879986 podStartE2EDuration="17.442298613s" podCreationTimestamp="2026-03-13 14:25:28 +0000 UTC" firstStartedPulling="2026-03-13 14:25:32.137002996 +0000 UTC m=+1231.036790685" lastFinishedPulling="2026-03-13 14:25:44.955421623 +0000 UTC m=+1243.855209312" observedRunningTime="2026-03-13 14:25:45.437546603 +0000 UTC m=+1244.337334302" watchObservedRunningTime="2026-03-13 14:25:45.442298613 +0000 UTC m=+1244.342086302" Mar 13 14:25:46 crc kubenswrapper[4907]: I0313 14:25:46.273704 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Mar 13 14:25:46 crc kubenswrapper[4907]: I0313 14:25:46.413413 4907 generic.go:334] "Generic (PLEG): container finished" podID="58ab9689-3ab2-42f2-ad56-beb22e29f8da" containerID="62c42ba7e98c9bfe072975553c668c323179c088fa7a92a9812e075126aebb86" exitCode=0 Mar 13 14:25:46 crc kubenswrapper[4907]: I0313 14:25:46.413481 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"58ab9689-3ab2-42f2-ad56-beb22e29f8da","Type":"ContainerDied","Data":"62c42ba7e98c9bfe072975553c668c323179c088fa7a92a9812e075126aebb86"} Mar 13 14:25:46 crc kubenswrapper[4907]: I0313 14:25:46.415461 4907 generic.go:334] "Generic (PLEG): container finished" podID="c4411bd2-c555-433a-9015-f623948b1401" containerID="1953abb8d3f9054081c314893bde9d173fee7430bf61ccd6ae11b908a7002729" exitCode=0 Mar 13 14:25:46 crc kubenswrapper[4907]: I0313 14:25:46.415494 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"c4411bd2-c555-433a-9015-f623948b1401","Type":"ContainerDied","Data":"1953abb8d3f9054081c314893bde9d173fee7430bf61ccd6ae11b908a7002729"} Mar 13 14:25:46 crc kubenswrapper[4907]: I0313 14:25:46.842694 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Mar 13 14:25:47 crc kubenswrapper[4907]: I0313 14:25:47.424848 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"c4411bd2-c555-433a-9015-f623948b1401","Type":"ContainerStarted","Data":"b5e68fc91f35838327b362dd7ecd6d2f5ee1173bc7ae4fafcb541ac0c96b44fa"} Mar 13 14:25:47 crc kubenswrapper[4907]: I0313 14:25:47.426347 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-bw2xb" Mar 13 14:25:47 crc kubenswrapper[4907]: I0313 14:25:47.427083 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"58ab9689-3ab2-42f2-ad56-beb22e29f8da","Type":"ContainerStarted","Data":"af7b1c6225d2e94f5a2f9827f7b02dc18651ab20c1d53b4a73e6d032e463897a"} Mar 13 14:25:47 crc kubenswrapper[4907]: I0313 14:25:47.498490 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=20.146814831 podStartE2EDuration="29.49846879s" podCreationTimestamp="2026-03-13 14:25:18 +0000 UTC" firstStartedPulling="2026-03-13 14:25:31.077097965 +0000 UTC m=+1229.976885654" lastFinishedPulling="2026-03-13 14:25:40.428751924 +0000 UTC m=+1239.328539613" observedRunningTime="2026-03-13 14:25:47.455842029 +0000 UTC m=+1246.355629718" watchObservedRunningTime="2026-03-13 14:25:47.49846879 +0000 UTC m=+1246.398256479" Mar 13 14:25:47 crc kubenswrapper[4907]: I0313 14:25:47.503574 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=18.43645419 podStartE2EDuration="28.503557659s" podCreationTimestamp="2026-03-13 14:25:19 +0000 UTC" firstStartedPulling="2026-03-13 14:25:30.069412459 +0000 UTC m=+1228.969200148" lastFinishedPulling="2026-03-13 14:25:40.136515928 +0000 UTC m=+1239.036303617" observedRunningTime="2026-03-13 14:25:47.503129767 +0000 UTC m=+1246.402917456" watchObservedRunningTime="2026-03-13 14:25:47.503557659 +0000 UTC m=+1246.403345348" Mar 13 14:25:47 crc kubenswrapper[4907]: I0313 14:25:47.837778 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Mar 13 14:25:47 crc kubenswrapper[4907]: I0313 14:25:47.880924 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Mar 13 14:25:48 crc kubenswrapper[4907]: I0313 14:25:48.155583 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Mar 13 14:25:48 crc kubenswrapper[4907]: I0313 14:25:48.200973 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Mar 13 14:25:48 crc kubenswrapper[4907]: I0313 14:25:48.435277 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Mar 13 14:25:48 crc kubenswrapper[4907]: I0313 14:25:48.478309 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Mar 13 14:25:48 crc kubenswrapper[4907]: I0313 14:25:48.495333 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Mar 13 14:25:48 crc kubenswrapper[4907]: I0313 14:25:48.736291 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7988f9db49-2mh7s"] Mar 13 14:25:48 crc kubenswrapper[4907]: E0313 14:25:48.736641 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b921f7f7-cd29-4ac8-82e5-c69a8f9e2bc0" containerName="dnsmasq-dns" Mar 13 14:25:48 crc kubenswrapper[4907]: I0313 14:25:48.736654 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="b921f7f7-cd29-4ac8-82e5-c69a8f9e2bc0" containerName="dnsmasq-dns" Mar 13 14:25:48 crc kubenswrapper[4907]: E0313 14:25:48.736673 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b921f7f7-cd29-4ac8-82e5-c69a8f9e2bc0" containerName="init" Mar 13 14:25:48 crc kubenswrapper[4907]: I0313 14:25:48.736679 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="b921f7f7-cd29-4ac8-82e5-c69a8f9e2bc0" containerName="init" Mar 13 14:25:48 crc kubenswrapper[4907]: I0313 14:25:48.736913 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="b921f7f7-cd29-4ac8-82e5-c69a8f9e2bc0" containerName="dnsmasq-dns" Mar 13 14:25:48 crc kubenswrapper[4907]: I0313 14:25:48.737739 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7988f9db49-2mh7s" Mar 13 14:25:48 crc kubenswrapper[4907]: I0313 14:25:48.740234 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Mar 13 14:25:48 crc kubenswrapper[4907]: I0313 14:25:48.750942 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7988f9db49-2mh7s"] Mar 13 14:25:48 crc kubenswrapper[4907]: I0313 14:25:48.788129 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-nhnkq"] Mar 13 14:25:48 crc kubenswrapper[4907]: I0313 14:25:48.789507 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-nhnkq" Mar 13 14:25:48 crc kubenswrapper[4907]: I0313 14:25:48.792562 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Mar 13 14:25:48 crc kubenswrapper[4907]: I0313 14:25:48.858234 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-nhnkq"] Mar 13 14:25:48 crc kubenswrapper[4907]: I0313 14:25:48.877673 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/37978cc1-fcba-4032-a8b1-6632b61692ff-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-nhnkq\" (UID: \"37978cc1-fcba-4032-a8b1-6632b61692ff\") " pod="openstack/ovn-controller-metrics-nhnkq" Mar 13 14:25:48 crc kubenswrapper[4907]: I0313 14:25:48.877730 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5e9f0d4f-790a-444e-9090-d9c6fa952c2b-ovsdbserver-sb\") pod \"dnsmasq-dns-7988f9db49-2mh7s\" (UID: \"5e9f0d4f-790a-444e-9090-d9c6fa952c2b\") " pod="openstack/dnsmasq-dns-7988f9db49-2mh7s" Mar 13 14:25:48 crc kubenswrapper[4907]: I0313 14:25:48.877758 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/37978cc1-fcba-4032-a8b1-6632b61692ff-ovs-rundir\") pod \"ovn-controller-metrics-nhnkq\" (UID: \"37978cc1-fcba-4032-a8b1-6632b61692ff\") " pod="openstack/ovn-controller-metrics-nhnkq" Mar 13 14:25:48 crc kubenswrapper[4907]: I0313 14:25:48.877782 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5e9f0d4f-790a-444e-9090-d9c6fa952c2b-dns-svc\") pod \"dnsmasq-dns-7988f9db49-2mh7s\" (UID: \"5e9f0d4f-790a-444e-9090-d9c6fa952c2b\") " pod="openstack/dnsmasq-dns-7988f9db49-2mh7s" Mar 13 14:25:48 crc kubenswrapper[4907]: I0313 14:25:48.877845 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ndzf5\" (UniqueName: \"kubernetes.io/projected/37978cc1-fcba-4032-a8b1-6632b61692ff-kube-api-access-ndzf5\") pod \"ovn-controller-metrics-nhnkq\" (UID: \"37978cc1-fcba-4032-a8b1-6632b61692ff\") " pod="openstack/ovn-controller-metrics-nhnkq" Mar 13 14:25:48 crc kubenswrapper[4907]: I0313 14:25:48.877861 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/37978cc1-fcba-4032-a8b1-6632b61692ff-ovn-rundir\") pod \"ovn-controller-metrics-nhnkq\" (UID: \"37978cc1-fcba-4032-a8b1-6632b61692ff\") " pod="openstack/ovn-controller-metrics-nhnkq" Mar 13 14:25:48 crc kubenswrapper[4907]: I0313 14:25:48.877878 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5e9f0d4f-790a-444e-9090-d9c6fa952c2b-config\") pod \"dnsmasq-dns-7988f9db49-2mh7s\" (UID: \"5e9f0d4f-790a-444e-9090-d9c6fa952c2b\") " pod="openstack/dnsmasq-dns-7988f9db49-2mh7s" Mar 13 14:25:48 crc kubenswrapper[4907]: I0313 14:25:48.877939 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z85pw\" (UniqueName: \"kubernetes.io/projected/5e9f0d4f-790a-444e-9090-d9c6fa952c2b-kube-api-access-z85pw\") pod \"dnsmasq-dns-7988f9db49-2mh7s\" (UID: \"5e9f0d4f-790a-444e-9090-d9c6fa952c2b\") " pod="openstack/dnsmasq-dns-7988f9db49-2mh7s" Mar 13 14:25:48 crc kubenswrapper[4907]: I0313 14:25:48.877964 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37978cc1-fcba-4032-a8b1-6632b61692ff-combined-ca-bundle\") pod \"ovn-controller-metrics-nhnkq\" (UID: \"37978cc1-fcba-4032-a8b1-6632b61692ff\") " pod="openstack/ovn-controller-metrics-nhnkq" Mar 13 14:25:48 crc kubenswrapper[4907]: I0313 14:25:48.877990 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/37978cc1-fcba-4032-a8b1-6632b61692ff-config\") pod \"ovn-controller-metrics-nhnkq\" (UID: \"37978cc1-fcba-4032-a8b1-6632b61692ff\") " pod="openstack/ovn-controller-metrics-nhnkq" Mar 13 14:25:48 crc kubenswrapper[4907]: I0313 14:25:48.911277 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Mar 13 14:25:48 crc kubenswrapper[4907]: I0313 14:25:48.912492 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Mar 13 14:25:48 crc kubenswrapper[4907]: I0313 14:25:48.920605 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Mar 13 14:25:48 crc kubenswrapper[4907]: I0313 14:25:48.920835 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Mar 13 14:25:48 crc kubenswrapper[4907]: I0313 14:25:48.920997 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-c5q7n" Mar 13 14:25:48 crc kubenswrapper[4907]: I0313 14:25:48.921196 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Mar 13 14:25:48 crc kubenswrapper[4907]: I0313 14:25:48.937797 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7988f9db49-2mh7s"] Mar 13 14:25:48 crc kubenswrapper[4907]: E0313 14:25:48.938507 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[config dns-svc kube-api-access-z85pw ovsdbserver-sb], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/dnsmasq-dns-7988f9db49-2mh7s" podUID="5e9f0d4f-790a-444e-9090-d9c6fa952c2b" Mar 13 14:25:48 crc kubenswrapper[4907]: I0313 14:25:48.952567 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Mar 13 14:25:48 crc kubenswrapper[4907]: I0313 14:25:48.981725 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ndzf5\" (UniqueName: \"kubernetes.io/projected/37978cc1-fcba-4032-a8b1-6632b61692ff-kube-api-access-ndzf5\") pod \"ovn-controller-metrics-nhnkq\" (UID: \"37978cc1-fcba-4032-a8b1-6632b61692ff\") " pod="openstack/ovn-controller-metrics-nhnkq" Mar 13 14:25:48 crc kubenswrapper[4907]: I0313 14:25:48.981768 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/37978cc1-fcba-4032-a8b1-6632b61692ff-ovn-rundir\") pod \"ovn-controller-metrics-nhnkq\" (UID: \"37978cc1-fcba-4032-a8b1-6632b61692ff\") " pod="openstack/ovn-controller-metrics-nhnkq" Mar 13 14:25:48 crc kubenswrapper[4907]: I0313 14:25:48.981790 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5e9f0d4f-790a-444e-9090-d9c6fa952c2b-config\") pod \"dnsmasq-dns-7988f9db49-2mh7s\" (UID: \"5e9f0d4f-790a-444e-9090-d9c6fa952c2b\") " pod="openstack/dnsmasq-dns-7988f9db49-2mh7s" Mar 13 14:25:48 crc kubenswrapper[4907]: I0313 14:25:48.981832 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z85pw\" (UniqueName: \"kubernetes.io/projected/5e9f0d4f-790a-444e-9090-d9c6fa952c2b-kube-api-access-z85pw\") pod \"dnsmasq-dns-7988f9db49-2mh7s\" (UID: \"5e9f0d4f-790a-444e-9090-d9c6fa952c2b\") " pod="openstack/dnsmasq-dns-7988f9db49-2mh7s" Mar 13 14:25:48 crc kubenswrapper[4907]: I0313 14:25:48.981850 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37978cc1-fcba-4032-a8b1-6632b61692ff-combined-ca-bundle\") pod \"ovn-controller-metrics-nhnkq\" (UID: \"37978cc1-fcba-4032-a8b1-6632b61692ff\") " pod="openstack/ovn-controller-metrics-nhnkq" Mar 13 14:25:48 crc kubenswrapper[4907]: I0313 14:25:48.981875 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/37978cc1-fcba-4032-a8b1-6632b61692ff-config\") pod \"ovn-controller-metrics-nhnkq\" (UID: \"37978cc1-fcba-4032-a8b1-6632b61692ff\") " pod="openstack/ovn-controller-metrics-nhnkq" Mar 13 14:25:48 crc kubenswrapper[4907]: I0313 14:25:48.981919 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/37978cc1-fcba-4032-a8b1-6632b61692ff-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-nhnkq\" (UID: \"37978cc1-fcba-4032-a8b1-6632b61692ff\") " pod="openstack/ovn-controller-metrics-nhnkq" Mar 13 14:25:48 crc kubenswrapper[4907]: I0313 14:25:48.981944 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5e9f0d4f-790a-444e-9090-d9c6fa952c2b-ovsdbserver-sb\") pod \"dnsmasq-dns-7988f9db49-2mh7s\" (UID: \"5e9f0d4f-790a-444e-9090-d9c6fa952c2b\") " pod="openstack/dnsmasq-dns-7988f9db49-2mh7s" Mar 13 14:25:48 crc kubenswrapper[4907]: I0313 14:25:48.981966 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/37978cc1-fcba-4032-a8b1-6632b61692ff-ovs-rundir\") pod \"ovn-controller-metrics-nhnkq\" (UID: \"37978cc1-fcba-4032-a8b1-6632b61692ff\") " pod="openstack/ovn-controller-metrics-nhnkq" Mar 13 14:25:48 crc kubenswrapper[4907]: I0313 14:25:48.981986 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5e9f0d4f-790a-444e-9090-d9c6fa952c2b-dns-svc\") pod \"dnsmasq-dns-7988f9db49-2mh7s\" (UID: \"5e9f0d4f-790a-444e-9090-d9c6fa952c2b\") " pod="openstack/dnsmasq-dns-7988f9db49-2mh7s" Mar 13 14:25:48 crc kubenswrapper[4907]: I0313 14:25:48.982359 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5d944d7b75-sk7j6"] Mar 13 14:25:48 crc kubenswrapper[4907]: I0313 14:25:48.983258 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5e9f0d4f-790a-444e-9090-d9c6fa952c2b-dns-svc\") pod \"dnsmasq-dns-7988f9db49-2mh7s\" (UID: \"5e9f0d4f-790a-444e-9090-d9c6fa952c2b\") " pod="openstack/dnsmasq-dns-7988f9db49-2mh7s" Mar 13 14:25:48 crc kubenswrapper[4907]: I0313 14:25:48.983461 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/37978cc1-fcba-4032-a8b1-6632b61692ff-ovn-rundir\") pod \"ovn-controller-metrics-nhnkq\" (UID: \"37978cc1-fcba-4032-a8b1-6632b61692ff\") " pod="openstack/ovn-controller-metrics-nhnkq" Mar 13 14:25:48 crc kubenswrapper[4907]: I0313 14:25:48.984337 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5e9f0d4f-790a-444e-9090-d9c6fa952c2b-config\") pod \"dnsmasq-dns-7988f9db49-2mh7s\" (UID: \"5e9f0d4f-790a-444e-9090-d9c6fa952c2b\") " pod="openstack/dnsmasq-dns-7988f9db49-2mh7s" Mar 13 14:25:48 crc kubenswrapper[4907]: I0313 14:25:48.985520 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d944d7b75-sk7j6" Mar 13 14:25:48 crc kubenswrapper[4907]: I0313 14:25:48.989049 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Mar 13 14:25:48 crc kubenswrapper[4907]: I0313 14:25:48.989567 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37978cc1-fcba-4032-a8b1-6632b61692ff-combined-ca-bundle\") pod \"ovn-controller-metrics-nhnkq\" (UID: \"37978cc1-fcba-4032-a8b1-6632b61692ff\") " pod="openstack/ovn-controller-metrics-nhnkq" Mar 13 14:25:48 crc kubenswrapper[4907]: I0313 14:25:48.989700 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/37978cc1-fcba-4032-a8b1-6632b61692ff-ovs-rundir\") pod \"ovn-controller-metrics-nhnkq\" (UID: \"37978cc1-fcba-4032-a8b1-6632b61692ff\") " pod="openstack/ovn-controller-metrics-nhnkq" Mar 13 14:25:48 crc kubenswrapper[4907]: I0313 14:25:48.990226 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/37978cc1-fcba-4032-a8b1-6632b61692ff-config\") pod \"ovn-controller-metrics-nhnkq\" (UID: \"37978cc1-fcba-4032-a8b1-6632b61692ff\") " pod="openstack/ovn-controller-metrics-nhnkq" Mar 13 14:25:48 crc kubenswrapper[4907]: I0313 14:25:48.991113 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5e9f0d4f-790a-444e-9090-d9c6fa952c2b-ovsdbserver-sb\") pod \"dnsmasq-dns-7988f9db49-2mh7s\" (UID: \"5e9f0d4f-790a-444e-9090-d9c6fa952c2b\") " pod="openstack/dnsmasq-dns-7988f9db49-2mh7s" Mar 13 14:25:48 crc kubenswrapper[4907]: I0313 14:25:48.992427 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/37978cc1-fcba-4032-a8b1-6632b61692ff-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-nhnkq\" (UID: \"37978cc1-fcba-4032-a8b1-6632b61692ff\") " pod="openstack/ovn-controller-metrics-nhnkq" Mar 13 14:25:49 crc kubenswrapper[4907]: I0313 14:25:49.008755 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z85pw\" (UniqueName: \"kubernetes.io/projected/5e9f0d4f-790a-444e-9090-d9c6fa952c2b-kube-api-access-z85pw\") pod \"dnsmasq-dns-7988f9db49-2mh7s\" (UID: \"5e9f0d4f-790a-444e-9090-d9c6fa952c2b\") " pod="openstack/dnsmasq-dns-7988f9db49-2mh7s" Mar 13 14:25:49 crc kubenswrapper[4907]: I0313 14:25:49.008762 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ndzf5\" (UniqueName: \"kubernetes.io/projected/37978cc1-fcba-4032-a8b1-6632b61692ff-kube-api-access-ndzf5\") pod \"ovn-controller-metrics-nhnkq\" (UID: \"37978cc1-fcba-4032-a8b1-6632b61692ff\") " pod="openstack/ovn-controller-metrics-nhnkq" Mar 13 14:25:49 crc kubenswrapper[4907]: I0313 14:25:49.024903 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5d944d7b75-sk7j6"] Mar 13 14:25:49 crc kubenswrapper[4907]: I0313 14:25:49.082975 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/26827945-75f0-4867-ba04-31ff6428e06a-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"26827945-75f0-4867-ba04-31ff6428e06a\") " pod="openstack/ovn-northd-0" Mar 13 14:25:49 crc kubenswrapper[4907]: I0313 14:25:49.083185 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b-ovsdbserver-sb\") pod \"dnsmasq-dns-5d944d7b75-sk7j6\" (UID: \"ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b\") " pod="openstack/dnsmasq-dns-5d944d7b75-sk7j6" Mar 13 14:25:49 crc kubenswrapper[4907]: I0313 14:25:49.083286 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/26827945-75f0-4867-ba04-31ff6428e06a-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"26827945-75f0-4867-ba04-31ff6428e06a\") " pod="openstack/ovn-northd-0" Mar 13 14:25:49 crc kubenswrapper[4907]: I0313 14:25:49.083379 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b-dns-svc\") pod \"dnsmasq-dns-5d944d7b75-sk7j6\" (UID: \"ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b\") " pod="openstack/dnsmasq-dns-5d944d7b75-sk7j6" Mar 13 14:25:49 crc kubenswrapper[4907]: I0313 14:25:49.083457 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b-config\") pod \"dnsmasq-dns-5d944d7b75-sk7j6\" (UID: \"ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b\") " pod="openstack/dnsmasq-dns-5d944d7b75-sk7j6" Mar 13 14:25:49 crc kubenswrapper[4907]: I0313 14:25:49.083535 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7kqs6\" (UniqueName: \"kubernetes.io/projected/ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b-kube-api-access-7kqs6\") pod \"dnsmasq-dns-5d944d7b75-sk7j6\" (UID: \"ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b\") " pod="openstack/dnsmasq-dns-5d944d7b75-sk7j6" Mar 13 14:25:49 crc kubenswrapper[4907]: I0313 14:25:49.083621 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26827945-75f0-4867-ba04-31ff6428e06a-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"26827945-75f0-4867-ba04-31ff6428e06a\") " pod="openstack/ovn-northd-0" Mar 13 14:25:49 crc kubenswrapper[4907]: I0313 14:25:49.083700 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/26827945-75f0-4867-ba04-31ff6428e06a-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"26827945-75f0-4867-ba04-31ff6428e06a\") " pod="openstack/ovn-northd-0" Mar 13 14:25:49 crc kubenswrapper[4907]: I0313 14:25:49.083773 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/26827945-75f0-4867-ba04-31ff6428e06a-scripts\") pod \"ovn-northd-0\" (UID: \"26827945-75f0-4867-ba04-31ff6428e06a\") " pod="openstack/ovn-northd-0" Mar 13 14:25:49 crc kubenswrapper[4907]: I0313 14:25:49.083856 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lgj9j\" (UniqueName: \"kubernetes.io/projected/26827945-75f0-4867-ba04-31ff6428e06a-kube-api-access-lgj9j\") pod \"ovn-northd-0\" (UID: \"26827945-75f0-4867-ba04-31ff6428e06a\") " pod="openstack/ovn-northd-0" Mar 13 14:25:49 crc kubenswrapper[4907]: I0313 14:25:49.083988 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b-ovsdbserver-nb\") pod \"dnsmasq-dns-5d944d7b75-sk7j6\" (UID: \"ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b\") " pod="openstack/dnsmasq-dns-5d944d7b75-sk7j6" Mar 13 14:25:49 crc kubenswrapper[4907]: I0313 14:25:49.084069 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/26827945-75f0-4867-ba04-31ff6428e06a-config\") pod \"ovn-northd-0\" (UID: \"26827945-75f0-4867-ba04-31ff6428e06a\") " pod="openstack/ovn-northd-0" Mar 13 14:25:49 crc kubenswrapper[4907]: I0313 14:25:49.117238 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-nhnkq" Mar 13 14:25:49 crc kubenswrapper[4907]: I0313 14:25:49.185715 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b-config\") pod \"dnsmasq-dns-5d944d7b75-sk7j6\" (UID: \"ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b\") " pod="openstack/dnsmasq-dns-5d944d7b75-sk7j6" Mar 13 14:25:49 crc kubenswrapper[4907]: I0313 14:25:49.185782 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7kqs6\" (UniqueName: \"kubernetes.io/projected/ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b-kube-api-access-7kqs6\") pod \"dnsmasq-dns-5d944d7b75-sk7j6\" (UID: \"ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b\") " pod="openstack/dnsmasq-dns-5d944d7b75-sk7j6" Mar 13 14:25:49 crc kubenswrapper[4907]: I0313 14:25:49.185825 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26827945-75f0-4867-ba04-31ff6428e06a-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"26827945-75f0-4867-ba04-31ff6428e06a\") " pod="openstack/ovn-northd-0" Mar 13 14:25:49 crc kubenswrapper[4907]: I0313 14:25:49.185849 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/26827945-75f0-4867-ba04-31ff6428e06a-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"26827945-75f0-4867-ba04-31ff6428e06a\") " pod="openstack/ovn-northd-0" Mar 13 14:25:49 crc kubenswrapper[4907]: I0313 14:25:49.185920 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/26827945-75f0-4867-ba04-31ff6428e06a-scripts\") pod \"ovn-northd-0\" (UID: \"26827945-75f0-4867-ba04-31ff6428e06a\") " pod="openstack/ovn-northd-0" Mar 13 14:25:49 crc kubenswrapper[4907]: I0313 14:25:49.185948 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lgj9j\" (UniqueName: \"kubernetes.io/projected/26827945-75f0-4867-ba04-31ff6428e06a-kube-api-access-lgj9j\") pod \"ovn-northd-0\" (UID: \"26827945-75f0-4867-ba04-31ff6428e06a\") " pod="openstack/ovn-northd-0" Mar 13 14:25:49 crc kubenswrapper[4907]: I0313 14:25:49.186005 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b-ovsdbserver-nb\") pod \"dnsmasq-dns-5d944d7b75-sk7j6\" (UID: \"ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b\") " pod="openstack/dnsmasq-dns-5d944d7b75-sk7j6" Mar 13 14:25:49 crc kubenswrapper[4907]: I0313 14:25:49.186030 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/26827945-75f0-4867-ba04-31ff6428e06a-config\") pod \"ovn-northd-0\" (UID: \"26827945-75f0-4867-ba04-31ff6428e06a\") " pod="openstack/ovn-northd-0" Mar 13 14:25:49 crc kubenswrapper[4907]: I0313 14:25:49.186104 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/26827945-75f0-4867-ba04-31ff6428e06a-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"26827945-75f0-4867-ba04-31ff6428e06a\") " pod="openstack/ovn-northd-0" Mar 13 14:25:49 crc kubenswrapper[4907]: I0313 14:25:49.186121 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b-ovsdbserver-sb\") pod \"dnsmasq-dns-5d944d7b75-sk7j6\" (UID: \"ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b\") " pod="openstack/dnsmasq-dns-5d944d7b75-sk7j6" Mar 13 14:25:49 crc kubenswrapper[4907]: I0313 14:25:49.186142 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/26827945-75f0-4867-ba04-31ff6428e06a-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"26827945-75f0-4867-ba04-31ff6428e06a\") " pod="openstack/ovn-northd-0" Mar 13 14:25:49 crc kubenswrapper[4907]: I0313 14:25:49.186170 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b-dns-svc\") pod \"dnsmasq-dns-5d944d7b75-sk7j6\" (UID: \"ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b\") " pod="openstack/dnsmasq-dns-5d944d7b75-sk7j6" Mar 13 14:25:49 crc kubenswrapper[4907]: I0313 14:25:49.187089 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b-dns-svc\") pod \"dnsmasq-dns-5d944d7b75-sk7j6\" (UID: \"ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b\") " pod="openstack/dnsmasq-dns-5d944d7b75-sk7j6" Mar 13 14:25:49 crc kubenswrapper[4907]: I0313 14:25:49.187353 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b-ovsdbserver-sb\") pod \"dnsmasq-dns-5d944d7b75-sk7j6\" (UID: \"ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b\") " pod="openstack/dnsmasq-dns-5d944d7b75-sk7j6" Mar 13 14:25:49 crc kubenswrapper[4907]: I0313 14:25:49.187832 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/26827945-75f0-4867-ba04-31ff6428e06a-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"26827945-75f0-4867-ba04-31ff6428e06a\") " pod="openstack/ovn-northd-0" Mar 13 14:25:49 crc kubenswrapper[4907]: I0313 14:25:49.187972 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b-config\") pod \"dnsmasq-dns-5d944d7b75-sk7j6\" (UID: \"ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b\") " pod="openstack/dnsmasq-dns-5d944d7b75-sk7j6" Mar 13 14:25:49 crc kubenswrapper[4907]: I0313 14:25:49.188210 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b-ovsdbserver-nb\") pod \"dnsmasq-dns-5d944d7b75-sk7j6\" (UID: \"ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b\") " pod="openstack/dnsmasq-dns-5d944d7b75-sk7j6" Mar 13 14:25:49 crc kubenswrapper[4907]: I0313 14:25:49.188478 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/26827945-75f0-4867-ba04-31ff6428e06a-scripts\") pod \"ovn-northd-0\" (UID: \"26827945-75f0-4867-ba04-31ff6428e06a\") " pod="openstack/ovn-northd-0" Mar 13 14:25:49 crc kubenswrapper[4907]: I0313 14:25:49.188553 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/26827945-75f0-4867-ba04-31ff6428e06a-config\") pod \"ovn-northd-0\" (UID: \"26827945-75f0-4867-ba04-31ff6428e06a\") " pod="openstack/ovn-northd-0" Mar 13 14:25:49 crc kubenswrapper[4907]: I0313 14:25:49.191264 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/26827945-75f0-4867-ba04-31ff6428e06a-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"26827945-75f0-4867-ba04-31ff6428e06a\") " pod="openstack/ovn-northd-0" Mar 13 14:25:49 crc kubenswrapper[4907]: I0313 14:25:49.192539 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/26827945-75f0-4867-ba04-31ff6428e06a-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"26827945-75f0-4867-ba04-31ff6428e06a\") " pod="openstack/ovn-northd-0" Mar 13 14:25:49 crc kubenswrapper[4907]: I0313 14:25:49.206727 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26827945-75f0-4867-ba04-31ff6428e06a-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"26827945-75f0-4867-ba04-31ff6428e06a\") " pod="openstack/ovn-northd-0" Mar 13 14:25:49 crc kubenswrapper[4907]: I0313 14:25:49.210877 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lgj9j\" (UniqueName: \"kubernetes.io/projected/26827945-75f0-4867-ba04-31ff6428e06a-kube-api-access-lgj9j\") pod \"ovn-northd-0\" (UID: \"26827945-75f0-4867-ba04-31ff6428e06a\") " pod="openstack/ovn-northd-0" Mar 13 14:25:49 crc kubenswrapper[4907]: I0313 14:25:49.229610 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7kqs6\" (UniqueName: \"kubernetes.io/projected/ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b-kube-api-access-7kqs6\") pod \"dnsmasq-dns-5d944d7b75-sk7j6\" (UID: \"ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b\") " pod="openstack/dnsmasq-dns-5d944d7b75-sk7j6" Mar 13 14:25:49 crc kubenswrapper[4907]: I0313 14:25:49.250452 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Mar 13 14:25:49 crc kubenswrapper[4907]: I0313 14:25:49.386985 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d944d7b75-sk7j6" Mar 13 14:25:49 crc kubenswrapper[4907]: I0313 14:25:49.442615 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7988f9db49-2mh7s" Mar 13 14:25:49 crc kubenswrapper[4907]: I0313 14:25:49.464819 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7988f9db49-2mh7s" Mar 13 14:25:49 crc kubenswrapper[4907]: I0313 14:25:49.590748 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5e9f0d4f-790a-444e-9090-d9c6fa952c2b-config\") pod \"5e9f0d4f-790a-444e-9090-d9c6fa952c2b\" (UID: \"5e9f0d4f-790a-444e-9090-d9c6fa952c2b\") " Mar 13 14:25:49 crc kubenswrapper[4907]: I0313 14:25:49.590897 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5e9f0d4f-790a-444e-9090-d9c6fa952c2b-ovsdbserver-sb\") pod \"5e9f0d4f-790a-444e-9090-d9c6fa952c2b\" (UID: \"5e9f0d4f-790a-444e-9090-d9c6fa952c2b\") " Mar 13 14:25:49 crc kubenswrapper[4907]: I0313 14:25:49.591007 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z85pw\" (UniqueName: \"kubernetes.io/projected/5e9f0d4f-790a-444e-9090-d9c6fa952c2b-kube-api-access-z85pw\") pod \"5e9f0d4f-790a-444e-9090-d9c6fa952c2b\" (UID: \"5e9f0d4f-790a-444e-9090-d9c6fa952c2b\") " Mar 13 14:25:49 crc kubenswrapper[4907]: I0313 14:25:49.591040 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5e9f0d4f-790a-444e-9090-d9c6fa952c2b-dns-svc\") pod \"5e9f0d4f-790a-444e-9090-d9c6fa952c2b\" (UID: \"5e9f0d4f-790a-444e-9090-d9c6fa952c2b\") " Mar 13 14:25:49 crc kubenswrapper[4907]: I0313 14:25:49.591526 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5e9f0d4f-790a-444e-9090-d9c6fa952c2b-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "5e9f0d4f-790a-444e-9090-d9c6fa952c2b" (UID: "5e9f0d4f-790a-444e-9090-d9c6fa952c2b"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:25:49 crc kubenswrapper[4907]: I0313 14:25:49.591526 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5e9f0d4f-790a-444e-9090-d9c6fa952c2b-config" (OuterVolumeSpecName: "config") pod "5e9f0d4f-790a-444e-9090-d9c6fa952c2b" (UID: "5e9f0d4f-790a-444e-9090-d9c6fa952c2b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:25:49 crc kubenswrapper[4907]: I0313 14:25:49.591776 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5e9f0d4f-790a-444e-9090-d9c6fa952c2b-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:25:49 crc kubenswrapper[4907]: I0313 14:25:49.592110 4907 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5e9f0d4f-790a-444e-9090-d9c6fa952c2b-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 13 14:25:49 crc kubenswrapper[4907]: I0313 14:25:49.592481 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5e9f0d4f-790a-444e-9090-d9c6fa952c2b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "5e9f0d4f-790a-444e-9090-d9c6fa952c2b" (UID: "5e9f0d4f-790a-444e-9090-d9c6fa952c2b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:25:49 crc kubenswrapper[4907]: I0313 14:25:49.596073 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e9f0d4f-790a-444e-9090-d9c6fa952c2b-kube-api-access-z85pw" (OuterVolumeSpecName: "kube-api-access-z85pw") pod "5e9f0d4f-790a-444e-9090-d9c6fa952c2b" (UID: "5e9f0d4f-790a-444e-9090-d9c6fa952c2b"). InnerVolumeSpecName "kube-api-access-z85pw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:25:49 crc kubenswrapper[4907]: I0313 14:25:49.609625 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-nhnkq"] Mar 13 14:25:49 crc kubenswrapper[4907]: I0313 14:25:49.694122 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z85pw\" (UniqueName: \"kubernetes.io/projected/5e9f0d4f-790a-444e-9090-d9c6fa952c2b-kube-api-access-z85pw\") on node \"crc\" DevicePath \"\"" Mar 13 14:25:49 crc kubenswrapper[4907]: I0313 14:25:49.694229 4907 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5e9f0d4f-790a-444e-9090-d9c6fa952c2b-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 13 14:25:49 crc kubenswrapper[4907]: I0313 14:25:49.747171 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Mar 13 14:25:49 crc kubenswrapper[4907]: I0313 14:25:49.747215 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Mar 13 14:25:49 crc kubenswrapper[4907]: I0313 14:25:49.833775 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Mar 13 14:25:49 crc kubenswrapper[4907]: I0313 14:25:49.883673 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5d944d7b75-sk7j6"] Mar 13 14:25:49 crc kubenswrapper[4907]: W0313 14:25:49.885200 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod26827945_75f0_4867_ba04_31ff6428e06a.slice/crio-59513110b8b488ed882794dedcca28bd1c64f5a6610bf84e8eaeafbea7ce6daa WatchSource:0}: Error finding container 59513110b8b488ed882794dedcca28bd1c64f5a6610bf84e8eaeafbea7ce6daa: Status 404 returned error can't find the container with id 59513110b8b488ed882794dedcca28bd1c64f5a6610bf84e8eaeafbea7ce6daa Mar 13 14:25:50 crc kubenswrapper[4907]: I0313 14:25:50.450938 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-nhnkq" event={"ID":"37978cc1-fcba-4032-a8b1-6632b61692ff","Type":"ContainerStarted","Data":"5a1bb100c70a1f42d543b5e88b530ca72ed03e93dea7325eb0c88d0809f831ad"} Mar 13 14:25:50 crc kubenswrapper[4907]: I0313 14:25:50.453308 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d944d7b75-sk7j6" event={"ID":"ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b","Type":"ContainerStarted","Data":"d38876707ec2c1ee40fc15622432cfe811af1e53b494065c48deee03085ef192"} Mar 13 14:25:50 crc kubenswrapper[4907]: I0313 14:25:50.453558 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"26827945-75f0-4867-ba04-31ff6428e06a","Type":"ContainerStarted","Data":"59513110b8b488ed882794dedcca28bd1c64f5a6610bf84e8eaeafbea7ce6daa"} Mar 13 14:25:50 crc kubenswrapper[4907]: I0313 14:25:50.453447 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7988f9db49-2mh7s" Mar 13 14:25:50 crc kubenswrapper[4907]: I0313 14:25:50.513357 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7988f9db49-2mh7s"] Mar 13 14:25:50 crc kubenswrapper[4907]: I0313 14:25:50.539029 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7988f9db49-2mh7s"] Mar 13 14:25:51 crc kubenswrapper[4907]: I0313 14:25:51.156904 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Mar 13 14:25:51 crc kubenswrapper[4907]: I0313 14:25:51.157244 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Mar 13 14:25:51 crc kubenswrapper[4907]: I0313 14:25:51.791992 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5e9f0d4f-790a-444e-9090-d9c6fa952c2b" path="/var/lib/kubelet/pods/5e9f0d4f-790a-444e-9090-d9c6fa952c2b/volumes" Mar 13 14:25:52 crc kubenswrapper[4907]: I0313 14:25:52.466639 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-nhnkq" event={"ID":"37978cc1-fcba-4032-a8b1-6632b61692ff","Type":"ContainerStarted","Data":"62edff4734da4ab22bc7673f1d2c0e9a9db082a3ca22b12fadc41486071fe66a"} Mar 13 14:25:53 crc kubenswrapper[4907]: I0313 14:25:53.160127 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5d944d7b75-sk7j6"] Mar 13 14:25:53 crc kubenswrapper[4907]: I0313 14:25:53.192145 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7b9fd7d84c-wkdxf"] Mar 13 14:25:53 crc kubenswrapper[4907]: I0313 14:25:53.193305 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7b9fd7d84c-wkdxf" Mar 13 14:25:53 crc kubenswrapper[4907]: I0313 14:25:53.203124 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7b9fd7d84c-wkdxf"] Mar 13 14:25:53 crc kubenswrapper[4907]: I0313 14:25:53.262621 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/89cd945f-311f-47d1-982a-641f062e4f57-ovsdbserver-nb\") pod \"dnsmasq-dns-7b9fd7d84c-wkdxf\" (UID: \"89cd945f-311f-47d1-982a-641f062e4f57\") " pod="openstack/dnsmasq-dns-7b9fd7d84c-wkdxf" Mar 13 14:25:53 crc kubenswrapper[4907]: I0313 14:25:53.263016 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lnkf5\" (UniqueName: \"kubernetes.io/projected/89cd945f-311f-47d1-982a-641f062e4f57-kube-api-access-lnkf5\") pod \"dnsmasq-dns-7b9fd7d84c-wkdxf\" (UID: \"89cd945f-311f-47d1-982a-641f062e4f57\") " pod="openstack/dnsmasq-dns-7b9fd7d84c-wkdxf" Mar 13 14:25:53 crc kubenswrapper[4907]: I0313 14:25:53.263127 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/89cd945f-311f-47d1-982a-641f062e4f57-dns-svc\") pod \"dnsmasq-dns-7b9fd7d84c-wkdxf\" (UID: \"89cd945f-311f-47d1-982a-641f062e4f57\") " pod="openstack/dnsmasq-dns-7b9fd7d84c-wkdxf" Mar 13 14:25:53 crc kubenswrapper[4907]: I0313 14:25:53.263236 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/89cd945f-311f-47d1-982a-641f062e4f57-config\") pod \"dnsmasq-dns-7b9fd7d84c-wkdxf\" (UID: \"89cd945f-311f-47d1-982a-641f062e4f57\") " pod="openstack/dnsmasq-dns-7b9fd7d84c-wkdxf" Mar 13 14:25:53 crc kubenswrapper[4907]: I0313 14:25:53.263403 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/89cd945f-311f-47d1-982a-641f062e4f57-ovsdbserver-sb\") pod \"dnsmasq-dns-7b9fd7d84c-wkdxf\" (UID: \"89cd945f-311f-47d1-982a-641f062e4f57\") " pod="openstack/dnsmasq-dns-7b9fd7d84c-wkdxf" Mar 13 14:25:53 crc kubenswrapper[4907]: I0313 14:25:53.364980 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/89cd945f-311f-47d1-982a-641f062e4f57-ovsdbserver-nb\") pod \"dnsmasq-dns-7b9fd7d84c-wkdxf\" (UID: \"89cd945f-311f-47d1-982a-641f062e4f57\") " pod="openstack/dnsmasq-dns-7b9fd7d84c-wkdxf" Mar 13 14:25:53 crc kubenswrapper[4907]: I0313 14:25:53.365043 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lnkf5\" (UniqueName: \"kubernetes.io/projected/89cd945f-311f-47d1-982a-641f062e4f57-kube-api-access-lnkf5\") pod \"dnsmasq-dns-7b9fd7d84c-wkdxf\" (UID: \"89cd945f-311f-47d1-982a-641f062e4f57\") " pod="openstack/dnsmasq-dns-7b9fd7d84c-wkdxf" Mar 13 14:25:53 crc kubenswrapper[4907]: I0313 14:25:53.365072 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/89cd945f-311f-47d1-982a-641f062e4f57-dns-svc\") pod \"dnsmasq-dns-7b9fd7d84c-wkdxf\" (UID: \"89cd945f-311f-47d1-982a-641f062e4f57\") " pod="openstack/dnsmasq-dns-7b9fd7d84c-wkdxf" Mar 13 14:25:53 crc kubenswrapper[4907]: I0313 14:25:53.365096 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/89cd945f-311f-47d1-982a-641f062e4f57-config\") pod \"dnsmasq-dns-7b9fd7d84c-wkdxf\" (UID: \"89cd945f-311f-47d1-982a-641f062e4f57\") " pod="openstack/dnsmasq-dns-7b9fd7d84c-wkdxf" Mar 13 14:25:53 crc kubenswrapper[4907]: I0313 14:25:53.365144 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/89cd945f-311f-47d1-982a-641f062e4f57-ovsdbserver-sb\") pod \"dnsmasq-dns-7b9fd7d84c-wkdxf\" (UID: \"89cd945f-311f-47d1-982a-641f062e4f57\") " pod="openstack/dnsmasq-dns-7b9fd7d84c-wkdxf" Mar 13 14:25:53 crc kubenswrapper[4907]: I0313 14:25:53.366147 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/89cd945f-311f-47d1-982a-641f062e4f57-ovsdbserver-sb\") pod \"dnsmasq-dns-7b9fd7d84c-wkdxf\" (UID: \"89cd945f-311f-47d1-982a-641f062e4f57\") " pod="openstack/dnsmasq-dns-7b9fd7d84c-wkdxf" Mar 13 14:25:53 crc kubenswrapper[4907]: I0313 14:25:53.366595 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/89cd945f-311f-47d1-982a-641f062e4f57-dns-svc\") pod \"dnsmasq-dns-7b9fd7d84c-wkdxf\" (UID: \"89cd945f-311f-47d1-982a-641f062e4f57\") " pod="openstack/dnsmasq-dns-7b9fd7d84c-wkdxf" Mar 13 14:25:53 crc kubenswrapper[4907]: I0313 14:25:53.367281 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/89cd945f-311f-47d1-982a-641f062e4f57-config\") pod \"dnsmasq-dns-7b9fd7d84c-wkdxf\" (UID: \"89cd945f-311f-47d1-982a-641f062e4f57\") " pod="openstack/dnsmasq-dns-7b9fd7d84c-wkdxf" Mar 13 14:25:53 crc kubenswrapper[4907]: I0313 14:25:53.367648 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/89cd945f-311f-47d1-982a-641f062e4f57-ovsdbserver-nb\") pod \"dnsmasq-dns-7b9fd7d84c-wkdxf\" (UID: \"89cd945f-311f-47d1-982a-641f062e4f57\") " pod="openstack/dnsmasq-dns-7b9fd7d84c-wkdxf" Mar 13 14:25:53 crc kubenswrapper[4907]: I0313 14:25:53.368143 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Mar 13 14:25:53 crc kubenswrapper[4907]: I0313 14:25:53.401834 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lnkf5\" (UniqueName: \"kubernetes.io/projected/89cd945f-311f-47d1-982a-641f062e4f57-kube-api-access-lnkf5\") pod \"dnsmasq-dns-7b9fd7d84c-wkdxf\" (UID: \"89cd945f-311f-47d1-982a-641f062e4f57\") " pod="openstack/dnsmasq-dns-7b9fd7d84c-wkdxf" Mar 13 14:25:53 crc kubenswrapper[4907]: I0313 14:25:53.475684 4907 generic.go:334] "Generic (PLEG): container finished" podID="ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b" containerID="e663df1a6e0053a10caca381de5f2223c40515b9e82b5695a1a4d758c5c706ee" exitCode=0 Mar 13 14:25:53 crc kubenswrapper[4907]: I0313 14:25:53.475823 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d944d7b75-sk7j6" event={"ID":"ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b","Type":"ContainerDied","Data":"e663df1a6e0053a10caca381de5f2223c40515b9e82b5695a1a4d758c5c706ee"} Mar 13 14:25:53 crc kubenswrapper[4907]: I0313 14:25:53.492209 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-nhnkq" podStartSLOduration=5.492192908 podStartE2EDuration="5.492192908s" podCreationTimestamp="2026-03-13 14:25:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:25:53.488807854 +0000 UTC m=+1252.388595543" watchObservedRunningTime="2026-03-13 14:25:53.492192908 +0000 UTC m=+1252.391980597" Mar 13 14:25:53 crc kubenswrapper[4907]: I0313 14:25:53.513889 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7b9fd7d84c-wkdxf" Mar 13 14:25:53 crc kubenswrapper[4907]: I0313 14:25:53.913961 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Mar 13 14:25:53 crc kubenswrapper[4907]: I0313 14:25:53.994434 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Mar 13 14:25:54 crc kubenswrapper[4907]: I0313 14:25:54.199299 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7b9fd7d84c-wkdxf"] Mar 13 14:25:54 crc kubenswrapper[4907]: W0313 14:25:54.202068 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod89cd945f_311f_47d1_982a_641f062e4f57.slice/crio-b45d10603df2d7ee66cbd50a38052eed85880ecd3d8f94e817d4236b2c83c476 WatchSource:0}: Error finding container b45d10603df2d7ee66cbd50a38052eed85880ecd3d8f94e817d4236b2c83c476: Status 404 returned error can't find the container with id b45d10603df2d7ee66cbd50a38052eed85880ecd3d8f94e817d4236b2c83c476 Mar 13 14:25:54 crc kubenswrapper[4907]: I0313 14:25:54.291999 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Mar 13 14:25:54 crc kubenswrapper[4907]: I0313 14:25:54.302037 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Mar 13 14:25:54 crc kubenswrapper[4907]: I0313 14:25:54.304552 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Mar 13 14:25:54 crc kubenswrapper[4907]: I0313 14:25:54.304593 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-mrlx5" Mar 13 14:25:54 crc kubenswrapper[4907]: I0313 14:25:54.304698 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Mar 13 14:25:54 crc kubenswrapper[4907]: I0313 14:25:54.304991 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Mar 13 14:25:54 crc kubenswrapper[4907]: I0313 14:25:54.320557 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Mar 13 14:25:54 crc kubenswrapper[4907]: I0313 14:25:54.481199 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"swift-storage-0\" (UID: \"14f05f31-8185-43b0-be69-bcf8d5388ea5\") " pod="openstack/swift-storage-0" Mar 13 14:25:54 crc kubenswrapper[4907]: I0313 14:25:54.481270 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/14f05f31-8185-43b0-be69-bcf8d5388ea5-lock\") pod \"swift-storage-0\" (UID: \"14f05f31-8185-43b0-be69-bcf8d5388ea5\") " pod="openstack/swift-storage-0" Mar 13 14:25:54 crc kubenswrapper[4907]: I0313 14:25:54.481316 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/14f05f31-8185-43b0-be69-bcf8d5388ea5-cache\") pod \"swift-storage-0\" (UID: \"14f05f31-8185-43b0-be69-bcf8d5388ea5\") " pod="openstack/swift-storage-0" Mar 13 14:25:54 crc kubenswrapper[4907]: I0313 14:25:54.481345 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/14f05f31-8185-43b0-be69-bcf8d5388ea5-etc-swift\") pod \"swift-storage-0\" (UID: \"14f05f31-8185-43b0-be69-bcf8d5388ea5\") " pod="openstack/swift-storage-0" Mar 13 14:25:54 crc kubenswrapper[4907]: I0313 14:25:54.481381 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7ps87\" (UniqueName: \"kubernetes.io/projected/14f05f31-8185-43b0-be69-bcf8d5388ea5-kube-api-access-7ps87\") pod \"swift-storage-0\" (UID: \"14f05f31-8185-43b0-be69-bcf8d5388ea5\") " pod="openstack/swift-storage-0" Mar 13 14:25:54 crc kubenswrapper[4907]: I0313 14:25:54.481417 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14f05f31-8185-43b0-be69-bcf8d5388ea5-combined-ca-bundle\") pod \"swift-storage-0\" (UID: \"14f05f31-8185-43b0-be69-bcf8d5388ea5\") " pod="openstack/swift-storage-0" Mar 13 14:25:54 crc kubenswrapper[4907]: I0313 14:25:54.484383 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d944d7b75-sk7j6" event={"ID":"ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b","Type":"ContainerStarted","Data":"0da5a6353eaf69e9bcfb1c6797b8931d8c2f75f259f947daed552ac9ac517adc"} Mar 13 14:25:54 crc kubenswrapper[4907]: I0313 14:25:54.484498 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5d944d7b75-sk7j6" podUID="ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b" containerName="dnsmasq-dns" containerID="cri-o://0da5a6353eaf69e9bcfb1c6797b8931d8c2f75f259f947daed552ac9ac517adc" gracePeriod=10 Mar 13 14:25:54 crc kubenswrapper[4907]: I0313 14:25:54.484547 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5d944d7b75-sk7j6" Mar 13 14:25:54 crc kubenswrapper[4907]: I0313 14:25:54.486122 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"26827945-75f0-4867-ba04-31ff6428e06a","Type":"ContainerStarted","Data":"59dc65eade55429a5d17c956c987a8209b624ac898dd355bca1d93bdf510b508"} Mar 13 14:25:54 crc kubenswrapper[4907]: I0313 14:25:54.486381 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"26827945-75f0-4867-ba04-31ff6428e06a","Type":"ContainerStarted","Data":"e23ef8c27779c528c50bff3fd731469de16d17d8b86a266ab0618151d6dd930b"} Mar 13 14:25:54 crc kubenswrapper[4907]: I0313 14:25:54.486397 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Mar 13 14:25:54 crc kubenswrapper[4907]: I0313 14:25:54.488271 4907 generic.go:334] "Generic (PLEG): container finished" podID="89cd945f-311f-47d1-982a-641f062e4f57" containerID="b1318d08ab870fbe30d1438d26843f78b5e12e9bf7ba05a68e44fb095db6996e" exitCode=0 Mar 13 14:25:54 crc kubenswrapper[4907]: I0313 14:25:54.488299 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b9fd7d84c-wkdxf" event={"ID":"89cd945f-311f-47d1-982a-641f062e4f57","Type":"ContainerDied","Data":"b1318d08ab870fbe30d1438d26843f78b5e12e9bf7ba05a68e44fb095db6996e"} Mar 13 14:25:54 crc kubenswrapper[4907]: I0313 14:25:54.488325 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b9fd7d84c-wkdxf" event={"ID":"89cd945f-311f-47d1-982a-641f062e4f57","Type":"ContainerStarted","Data":"b45d10603df2d7ee66cbd50a38052eed85880ecd3d8f94e817d4236b2c83c476"} Mar 13 14:25:54 crc kubenswrapper[4907]: I0313 14:25:54.511320 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5d944d7b75-sk7j6" podStartSLOduration=6.5112991529999995 podStartE2EDuration="6.511299153s" podCreationTimestamp="2026-03-13 14:25:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:25:54.498623265 +0000 UTC m=+1253.398410974" watchObservedRunningTime="2026-03-13 14:25:54.511299153 +0000 UTC m=+1253.411086842" Mar 13 14:25:54 crc kubenswrapper[4907]: I0313 14:25:54.549419 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=2.499962866 podStartE2EDuration="6.54939769s" podCreationTimestamp="2026-03-13 14:25:48 +0000 UTC" firstStartedPulling="2026-03-13 14:25:49.90000902 +0000 UTC m=+1248.799796709" lastFinishedPulling="2026-03-13 14:25:53.949443844 +0000 UTC m=+1252.849231533" observedRunningTime="2026-03-13 14:25:54.540202977 +0000 UTC m=+1253.439990666" watchObservedRunningTime="2026-03-13 14:25:54.54939769 +0000 UTC m=+1253.449185379" Mar 13 14:25:54 crc kubenswrapper[4907]: I0313 14:25:54.582491 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"swift-storage-0\" (UID: \"14f05f31-8185-43b0-be69-bcf8d5388ea5\") " pod="openstack/swift-storage-0" Mar 13 14:25:54 crc kubenswrapper[4907]: I0313 14:25:54.582610 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/14f05f31-8185-43b0-be69-bcf8d5388ea5-lock\") pod \"swift-storage-0\" (UID: \"14f05f31-8185-43b0-be69-bcf8d5388ea5\") " pod="openstack/swift-storage-0" Mar 13 14:25:54 crc kubenswrapper[4907]: I0313 14:25:54.582706 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/14f05f31-8185-43b0-be69-bcf8d5388ea5-cache\") pod \"swift-storage-0\" (UID: \"14f05f31-8185-43b0-be69-bcf8d5388ea5\") " pod="openstack/swift-storage-0" Mar 13 14:25:54 crc kubenswrapper[4907]: I0313 14:25:54.582725 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/14f05f31-8185-43b0-be69-bcf8d5388ea5-etc-swift\") pod \"swift-storage-0\" (UID: \"14f05f31-8185-43b0-be69-bcf8d5388ea5\") " pod="openstack/swift-storage-0" Mar 13 14:25:54 crc kubenswrapper[4907]: I0313 14:25:54.582762 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7ps87\" (UniqueName: \"kubernetes.io/projected/14f05f31-8185-43b0-be69-bcf8d5388ea5-kube-api-access-7ps87\") pod \"swift-storage-0\" (UID: \"14f05f31-8185-43b0-be69-bcf8d5388ea5\") " pod="openstack/swift-storage-0" Mar 13 14:25:54 crc kubenswrapper[4907]: I0313 14:25:54.582784 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14f05f31-8185-43b0-be69-bcf8d5388ea5-combined-ca-bundle\") pod \"swift-storage-0\" (UID: \"14f05f31-8185-43b0-be69-bcf8d5388ea5\") " pod="openstack/swift-storage-0" Mar 13 14:25:54 crc kubenswrapper[4907]: I0313 14:25:54.583489 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/14f05f31-8185-43b0-be69-bcf8d5388ea5-cache\") pod \"swift-storage-0\" (UID: \"14f05f31-8185-43b0-be69-bcf8d5388ea5\") " pod="openstack/swift-storage-0" Mar 13 14:25:54 crc kubenswrapper[4907]: E0313 14:25:54.583595 4907 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Mar 13 14:25:54 crc kubenswrapper[4907]: E0313 14:25:54.583613 4907 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Mar 13 14:25:54 crc kubenswrapper[4907]: E0313 14:25:54.583661 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/14f05f31-8185-43b0-be69-bcf8d5388ea5-etc-swift podName:14f05f31-8185-43b0-be69-bcf8d5388ea5 nodeName:}" failed. No retries permitted until 2026-03-13 14:25:55.08364227 +0000 UTC m=+1253.983429959 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/14f05f31-8185-43b0-be69-bcf8d5388ea5-etc-swift") pod "swift-storage-0" (UID: "14f05f31-8185-43b0-be69-bcf8d5388ea5") : configmap "swift-ring-files" not found Mar 13 14:25:54 crc kubenswrapper[4907]: I0313 14:25:54.583713 4907 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"swift-storage-0\" (UID: \"14f05f31-8185-43b0-be69-bcf8d5388ea5\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/swift-storage-0" Mar 13 14:25:54 crc kubenswrapper[4907]: I0313 14:25:54.583896 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/14f05f31-8185-43b0-be69-bcf8d5388ea5-lock\") pod \"swift-storage-0\" (UID: \"14f05f31-8185-43b0-be69-bcf8d5388ea5\") " pod="openstack/swift-storage-0" Mar 13 14:25:54 crc kubenswrapper[4907]: I0313 14:25:54.589605 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14f05f31-8185-43b0-be69-bcf8d5388ea5-combined-ca-bundle\") pod \"swift-storage-0\" (UID: \"14f05f31-8185-43b0-be69-bcf8d5388ea5\") " pod="openstack/swift-storage-0" Mar 13 14:25:54 crc kubenswrapper[4907]: I0313 14:25:54.604671 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7ps87\" (UniqueName: \"kubernetes.io/projected/14f05f31-8185-43b0-be69-bcf8d5388ea5-kube-api-access-7ps87\") pod \"swift-storage-0\" (UID: \"14f05f31-8185-43b0-be69-bcf8d5388ea5\") " pod="openstack/swift-storage-0" Mar 13 14:25:54 crc kubenswrapper[4907]: I0313 14:25:54.606787 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"swift-storage-0\" (UID: \"14f05f31-8185-43b0-be69-bcf8d5388ea5\") " pod="openstack/swift-storage-0" Mar 13 14:25:54 crc kubenswrapper[4907]: I0313 14:25:54.871025 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d944d7b75-sk7j6" Mar 13 14:25:54 crc kubenswrapper[4907]: I0313 14:25:54.990962 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b-ovsdbserver-sb\") pod \"ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b\" (UID: \"ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b\") " Mar 13 14:25:54 crc kubenswrapper[4907]: I0313 14:25:54.991082 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b-dns-svc\") pod \"ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b\" (UID: \"ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b\") " Mar 13 14:25:54 crc kubenswrapper[4907]: I0313 14:25:54.991133 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b-ovsdbserver-nb\") pod \"ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b\" (UID: \"ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b\") " Mar 13 14:25:54 crc kubenswrapper[4907]: I0313 14:25:54.991203 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b-config\") pod \"ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b\" (UID: \"ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b\") " Mar 13 14:25:54 crc kubenswrapper[4907]: I0313 14:25:54.991230 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7kqs6\" (UniqueName: \"kubernetes.io/projected/ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b-kube-api-access-7kqs6\") pod \"ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b\" (UID: \"ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b\") " Mar 13 14:25:54 crc kubenswrapper[4907]: I0313 14:25:54.997576 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b-kube-api-access-7kqs6" (OuterVolumeSpecName: "kube-api-access-7kqs6") pod "ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b" (UID: "ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b"). InnerVolumeSpecName "kube-api-access-7kqs6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:25:55 crc kubenswrapper[4907]: I0313 14:25:55.032319 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b" (UID: "ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:25:55 crc kubenswrapper[4907]: I0313 14:25:55.037073 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b" (UID: "ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:25:55 crc kubenswrapper[4907]: I0313 14:25:55.040938 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b" (UID: "ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:25:55 crc kubenswrapper[4907]: I0313 14:25:55.043265 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b-config" (OuterVolumeSpecName: "config") pod "ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b" (UID: "ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:25:55 crc kubenswrapper[4907]: I0313 14:25:55.093399 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/14f05f31-8185-43b0-be69-bcf8d5388ea5-etc-swift\") pod \"swift-storage-0\" (UID: \"14f05f31-8185-43b0-be69-bcf8d5388ea5\") " pod="openstack/swift-storage-0" Mar 13 14:25:55 crc kubenswrapper[4907]: I0313 14:25:55.093546 4907 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 13 14:25:55 crc kubenswrapper[4907]: I0313 14:25:55.093563 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:25:55 crc kubenswrapper[4907]: I0313 14:25:55.093574 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7kqs6\" (UniqueName: \"kubernetes.io/projected/ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b-kube-api-access-7kqs6\") on node \"crc\" DevicePath \"\"" Mar 13 14:25:55 crc kubenswrapper[4907]: I0313 14:25:55.093586 4907 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 13 14:25:55 crc kubenswrapper[4907]: E0313 14:25:55.093589 4907 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Mar 13 14:25:55 crc kubenswrapper[4907]: E0313 14:25:55.093617 4907 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Mar 13 14:25:55 crc kubenswrapper[4907]: E0313 14:25:55.093664 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/14f05f31-8185-43b0-be69-bcf8d5388ea5-etc-swift podName:14f05f31-8185-43b0-be69-bcf8d5388ea5 nodeName:}" failed. No retries permitted until 2026-03-13 14:25:56.093649736 +0000 UTC m=+1254.993437425 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/14f05f31-8185-43b0-be69-bcf8d5388ea5-etc-swift") pod "swift-storage-0" (UID: "14f05f31-8185-43b0-be69-bcf8d5388ea5") : configmap "swift-ring-files" not found Mar 13 14:25:55 crc kubenswrapper[4907]: I0313 14:25:55.093597 4907 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 13 14:25:55 crc kubenswrapper[4907]: I0313 14:25:55.279416 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Mar 13 14:25:55 crc kubenswrapper[4907]: I0313 14:25:55.481992 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Mar 13 14:25:55 crc kubenswrapper[4907]: I0313 14:25:55.496299 4907 generic.go:334] "Generic (PLEG): container finished" podID="ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b" containerID="0da5a6353eaf69e9bcfb1c6797b8931d8c2f75f259f947daed552ac9ac517adc" exitCode=0 Mar 13 14:25:55 crc kubenswrapper[4907]: I0313 14:25:55.496359 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d944d7b75-sk7j6" event={"ID":"ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b","Type":"ContainerDied","Data":"0da5a6353eaf69e9bcfb1c6797b8931d8c2f75f259f947daed552ac9ac517adc"} Mar 13 14:25:55 crc kubenswrapper[4907]: I0313 14:25:55.496367 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d944d7b75-sk7j6" Mar 13 14:25:55 crc kubenswrapper[4907]: I0313 14:25:55.496384 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d944d7b75-sk7j6" event={"ID":"ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b","Type":"ContainerDied","Data":"d38876707ec2c1ee40fc15622432cfe811af1e53b494065c48deee03085ef192"} Mar 13 14:25:55 crc kubenswrapper[4907]: I0313 14:25:55.496402 4907 scope.go:117] "RemoveContainer" containerID="0da5a6353eaf69e9bcfb1c6797b8931d8c2f75f259f947daed552ac9ac517adc" Mar 13 14:25:55 crc kubenswrapper[4907]: I0313 14:25:55.498775 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b9fd7d84c-wkdxf" event={"ID":"89cd945f-311f-47d1-982a-641f062e4f57","Type":"ContainerStarted","Data":"ad9755b55eba4bb5e259a681410af3ffba2c66af2583b34cd43ac29178757a96"} Mar 13 14:25:55 crc kubenswrapper[4907]: I0313 14:25:55.517589 4907 scope.go:117] "RemoveContainer" containerID="e663df1a6e0053a10caca381de5f2223c40515b9e82b5695a1a4d758c5c706ee" Mar 13 14:25:55 crc kubenswrapper[4907]: I0313 14:25:55.542010 4907 scope.go:117] "RemoveContainer" containerID="0da5a6353eaf69e9bcfb1c6797b8931d8c2f75f259f947daed552ac9ac517adc" Mar 13 14:25:55 crc kubenswrapper[4907]: E0313 14:25:55.543860 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0da5a6353eaf69e9bcfb1c6797b8931d8c2f75f259f947daed552ac9ac517adc\": container with ID starting with 0da5a6353eaf69e9bcfb1c6797b8931d8c2f75f259f947daed552ac9ac517adc not found: ID does not exist" containerID="0da5a6353eaf69e9bcfb1c6797b8931d8c2f75f259f947daed552ac9ac517adc" Mar 13 14:25:55 crc kubenswrapper[4907]: I0313 14:25:55.543935 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0da5a6353eaf69e9bcfb1c6797b8931d8c2f75f259f947daed552ac9ac517adc"} err="failed to get container status \"0da5a6353eaf69e9bcfb1c6797b8931d8c2f75f259f947daed552ac9ac517adc\": rpc error: code = NotFound desc = could not find container \"0da5a6353eaf69e9bcfb1c6797b8931d8c2f75f259f947daed552ac9ac517adc\": container with ID starting with 0da5a6353eaf69e9bcfb1c6797b8931d8c2f75f259f947daed552ac9ac517adc not found: ID does not exist" Mar 13 14:25:55 crc kubenswrapper[4907]: I0313 14:25:55.543965 4907 scope.go:117] "RemoveContainer" containerID="e663df1a6e0053a10caca381de5f2223c40515b9e82b5695a1a4d758c5c706ee" Mar 13 14:25:55 crc kubenswrapper[4907]: E0313 14:25:55.547165 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e663df1a6e0053a10caca381de5f2223c40515b9e82b5695a1a4d758c5c706ee\": container with ID starting with e663df1a6e0053a10caca381de5f2223c40515b9e82b5695a1a4d758c5c706ee not found: ID does not exist" containerID="e663df1a6e0053a10caca381de5f2223c40515b9e82b5695a1a4d758c5c706ee" Mar 13 14:25:55 crc kubenswrapper[4907]: I0313 14:25:55.547197 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e663df1a6e0053a10caca381de5f2223c40515b9e82b5695a1a4d758c5c706ee"} err="failed to get container status \"e663df1a6e0053a10caca381de5f2223c40515b9e82b5695a1a4d758c5c706ee\": rpc error: code = NotFound desc = could not find container \"e663df1a6e0053a10caca381de5f2223c40515b9e82b5695a1a4d758c5c706ee\": container with ID starting with e663df1a6e0053a10caca381de5f2223c40515b9e82b5695a1a4d758c5c706ee not found: ID does not exist" Mar 13 14:25:55 crc kubenswrapper[4907]: I0313 14:25:55.578919 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7b9fd7d84c-wkdxf" podStartSLOduration=2.578895302 podStartE2EDuration="2.578895302s" podCreationTimestamp="2026-03-13 14:25:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:25:55.560094086 +0000 UTC m=+1254.459881775" watchObservedRunningTime="2026-03-13 14:25:55.578895302 +0000 UTC m=+1254.478683011" Mar 13 14:25:55 crc kubenswrapper[4907]: I0313 14:25:55.586966 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5d944d7b75-sk7j6"] Mar 13 14:25:55 crc kubenswrapper[4907]: I0313 14:25:55.593609 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5d944d7b75-sk7j6"] Mar 13 14:25:55 crc kubenswrapper[4907]: I0313 14:25:55.791182 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b" path="/var/lib/kubelet/pods/ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b/volumes" Mar 13 14:25:56 crc kubenswrapper[4907]: I0313 14:25:56.109758 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/14f05f31-8185-43b0-be69-bcf8d5388ea5-etc-swift\") pod \"swift-storage-0\" (UID: \"14f05f31-8185-43b0-be69-bcf8d5388ea5\") " pod="openstack/swift-storage-0" Mar 13 14:25:56 crc kubenswrapper[4907]: E0313 14:25:56.109987 4907 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Mar 13 14:25:56 crc kubenswrapper[4907]: E0313 14:25:56.110203 4907 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Mar 13 14:25:56 crc kubenswrapper[4907]: E0313 14:25:56.110279 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/14f05f31-8185-43b0-be69-bcf8d5388ea5-etc-swift podName:14f05f31-8185-43b0-be69-bcf8d5388ea5 nodeName:}" failed. No retries permitted until 2026-03-13 14:25:58.110255613 +0000 UTC m=+1257.010043342 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/14f05f31-8185-43b0-be69-bcf8d5388ea5-etc-swift") pod "swift-storage-0" (UID: "14f05f31-8185-43b0-be69-bcf8d5388ea5") : configmap "swift-ring-files" not found Mar 13 14:25:56 crc kubenswrapper[4907]: I0313 14:25:56.508593 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7b9fd7d84c-wkdxf" Mar 13 14:25:56 crc kubenswrapper[4907]: I0313 14:25:56.521073 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-sgpkl"] Mar 13 14:25:56 crc kubenswrapper[4907]: E0313 14:25:56.521403 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b" containerName="init" Mar 13 14:25:56 crc kubenswrapper[4907]: I0313 14:25:56.521418 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b" containerName="init" Mar 13 14:25:56 crc kubenswrapper[4907]: E0313 14:25:56.521436 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b" containerName="dnsmasq-dns" Mar 13 14:25:56 crc kubenswrapper[4907]: I0313 14:25:56.521443 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b" containerName="dnsmasq-dns" Mar 13 14:25:56 crc kubenswrapper[4907]: I0313 14:25:56.521616 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b" containerName="dnsmasq-dns" Mar 13 14:25:56 crc kubenswrapper[4907]: I0313 14:25:56.522840 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-sgpkl" Mar 13 14:25:56 crc kubenswrapper[4907]: I0313 14:25:56.531105 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-17a8-account-create-update-xv8sm"] Mar 13 14:25:56 crc kubenswrapper[4907]: I0313 14:25:56.532135 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-17a8-account-create-update-xv8sm" Mar 13 14:25:56 crc kubenswrapper[4907]: I0313 14:25:56.535873 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Mar 13 14:25:56 crc kubenswrapper[4907]: I0313 14:25:56.551840 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-sgpkl"] Mar 13 14:25:56 crc kubenswrapper[4907]: I0313 14:25:56.561081 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-17a8-account-create-update-xv8sm"] Mar 13 14:25:56 crc kubenswrapper[4907]: I0313 14:25:56.618696 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4q8qf\" (UniqueName: \"kubernetes.io/projected/4d7d2639-4229-4286-be5c-d8b15ed91d17-kube-api-access-4q8qf\") pod \"glance-db-create-sgpkl\" (UID: \"4d7d2639-4229-4286-be5c-d8b15ed91d17\") " pod="openstack/glance-db-create-sgpkl" Mar 13 14:25:56 crc kubenswrapper[4907]: I0313 14:25:56.618787 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4d7d2639-4229-4286-be5c-d8b15ed91d17-operator-scripts\") pod \"glance-db-create-sgpkl\" (UID: \"4d7d2639-4229-4286-be5c-d8b15ed91d17\") " pod="openstack/glance-db-create-sgpkl" Mar 13 14:25:56 crc kubenswrapper[4907]: I0313 14:25:56.619023 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zkz6n\" (UniqueName: \"kubernetes.io/projected/cbb6c089-0112-43f4-8731-c2d68932795a-kube-api-access-zkz6n\") pod \"glance-17a8-account-create-update-xv8sm\" (UID: \"cbb6c089-0112-43f4-8731-c2d68932795a\") " pod="openstack/glance-17a8-account-create-update-xv8sm" Mar 13 14:25:56 crc kubenswrapper[4907]: I0313 14:25:56.619405 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cbb6c089-0112-43f4-8731-c2d68932795a-operator-scripts\") pod \"glance-17a8-account-create-update-xv8sm\" (UID: \"cbb6c089-0112-43f4-8731-c2d68932795a\") " pod="openstack/glance-17a8-account-create-update-xv8sm" Mar 13 14:25:56 crc kubenswrapper[4907]: I0313 14:25:56.721282 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cbb6c089-0112-43f4-8731-c2d68932795a-operator-scripts\") pod \"glance-17a8-account-create-update-xv8sm\" (UID: \"cbb6c089-0112-43f4-8731-c2d68932795a\") " pod="openstack/glance-17a8-account-create-update-xv8sm" Mar 13 14:25:56 crc kubenswrapper[4907]: I0313 14:25:56.721375 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4q8qf\" (UniqueName: \"kubernetes.io/projected/4d7d2639-4229-4286-be5c-d8b15ed91d17-kube-api-access-4q8qf\") pod \"glance-db-create-sgpkl\" (UID: \"4d7d2639-4229-4286-be5c-d8b15ed91d17\") " pod="openstack/glance-db-create-sgpkl" Mar 13 14:25:56 crc kubenswrapper[4907]: I0313 14:25:56.721412 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4d7d2639-4229-4286-be5c-d8b15ed91d17-operator-scripts\") pod \"glance-db-create-sgpkl\" (UID: \"4d7d2639-4229-4286-be5c-d8b15ed91d17\") " pod="openstack/glance-db-create-sgpkl" Mar 13 14:25:56 crc kubenswrapper[4907]: I0313 14:25:56.721457 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zkz6n\" (UniqueName: \"kubernetes.io/projected/cbb6c089-0112-43f4-8731-c2d68932795a-kube-api-access-zkz6n\") pod \"glance-17a8-account-create-update-xv8sm\" (UID: \"cbb6c089-0112-43f4-8731-c2d68932795a\") " pod="openstack/glance-17a8-account-create-update-xv8sm" Mar 13 14:25:56 crc kubenswrapper[4907]: I0313 14:25:56.722301 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cbb6c089-0112-43f4-8731-c2d68932795a-operator-scripts\") pod \"glance-17a8-account-create-update-xv8sm\" (UID: \"cbb6c089-0112-43f4-8731-c2d68932795a\") " pod="openstack/glance-17a8-account-create-update-xv8sm" Mar 13 14:25:56 crc kubenswrapper[4907]: I0313 14:25:56.722413 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4d7d2639-4229-4286-be5c-d8b15ed91d17-operator-scripts\") pod \"glance-db-create-sgpkl\" (UID: \"4d7d2639-4229-4286-be5c-d8b15ed91d17\") " pod="openstack/glance-db-create-sgpkl" Mar 13 14:25:56 crc kubenswrapper[4907]: I0313 14:25:56.743130 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zkz6n\" (UniqueName: \"kubernetes.io/projected/cbb6c089-0112-43f4-8731-c2d68932795a-kube-api-access-zkz6n\") pod \"glance-17a8-account-create-update-xv8sm\" (UID: \"cbb6c089-0112-43f4-8731-c2d68932795a\") " pod="openstack/glance-17a8-account-create-update-xv8sm" Mar 13 14:25:56 crc kubenswrapper[4907]: I0313 14:25:56.749484 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4q8qf\" (UniqueName: \"kubernetes.io/projected/4d7d2639-4229-4286-be5c-d8b15ed91d17-kube-api-access-4q8qf\") pod \"glance-db-create-sgpkl\" (UID: \"4d7d2639-4229-4286-be5c-d8b15ed91d17\") " pod="openstack/glance-db-create-sgpkl" Mar 13 14:25:56 crc kubenswrapper[4907]: I0313 14:25:56.852278 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-sgpkl" Mar 13 14:25:56 crc kubenswrapper[4907]: I0313 14:25:56.861795 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-17a8-account-create-update-xv8sm" Mar 13 14:25:57 crc kubenswrapper[4907]: I0313 14:25:57.278104 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-sgpkl"] Mar 13 14:25:57 crc kubenswrapper[4907]: W0313 14:25:57.280157 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4d7d2639_4229_4286_be5c_d8b15ed91d17.slice/crio-64d9b417d3e9f250fd98eb8db5c1fbf579e78ade9d4cbc4644beb2dad8d75fbc WatchSource:0}: Error finding container 64d9b417d3e9f250fd98eb8db5c1fbf579e78ade9d4cbc4644beb2dad8d75fbc: Status 404 returned error can't find the container with id 64d9b417d3e9f250fd98eb8db5c1fbf579e78ade9d4cbc4644beb2dad8d75fbc Mar 13 14:25:57 crc kubenswrapper[4907]: I0313 14:25:57.354903 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-17a8-account-create-update-xv8sm"] Mar 13 14:25:57 crc kubenswrapper[4907]: W0313 14:25:57.364626 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcbb6c089_0112_43f4_8731_c2d68932795a.slice/crio-99873ca15b9fb22afcd6ec2e9a14245394d9db6cee5f1f289e90f95b20ea6652 WatchSource:0}: Error finding container 99873ca15b9fb22afcd6ec2e9a14245394d9db6cee5f1f289e90f95b20ea6652: Status 404 returned error can't find the container with id 99873ca15b9fb22afcd6ec2e9a14245394d9db6cee5f1f289e90f95b20ea6652 Mar 13 14:25:57 crc kubenswrapper[4907]: I0313 14:25:57.520688 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-sgpkl" event={"ID":"4d7d2639-4229-4286-be5c-d8b15ed91d17","Type":"ContainerStarted","Data":"3c90645adf755c84a8db83dd2e03403f66a250afbd541346b9261208e83c56ff"} Mar 13 14:25:57 crc kubenswrapper[4907]: I0313 14:25:57.520731 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-sgpkl" event={"ID":"4d7d2639-4229-4286-be5c-d8b15ed91d17","Type":"ContainerStarted","Data":"64d9b417d3e9f250fd98eb8db5c1fbf579e78ade9d4cbc4644beb2dad8d75fbc"} Mar 13 14:25:57 crc kubenswrapper[4907]: I0313 14:25:57.524330 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-17a8-account-create-update-xv8sm" event={"ID":"cbb6c089-0112-43f4-8731-c2d68932795a","Type":"ContainerStarted","Data":"b2e917cc4c559e06be2ed2a503af651b2cead394bb06409f6e1c888788df817a"} Mar 13 14:25:57 crc kubenswrapper[4907]: I0313 14:25:57.524396 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-17a8-account-create-update-xv8sm" event={"ID":"cbb6c089-0112-43f4-8731-c2d68932795a","Type":"ContainerStarted","Data":"99873ca15b9fb22afcd6ec2e9a14245394d9db6cee5f1f289e90f95b20ea6652"} Mar 13 14:25:57 crc kubenswrapper[4907]: I0313 14:25:57.539874 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-create-sgpkl" podStartSLOduration=1.539851932 podStartE2EDuration="1.539851932s" podCreationTimestamp="2026-03-13 14:25:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:25:57.534348452 +0000 UTC m=+1256.434136141" watchObservedRunningTime="2026-03-13 14:25:57.539851932 +0000 UTC m=+1256.439639641" Mar 13 14:25:57 crc kubenswrapper[4907]: I0313 14:25:57.553424 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-17a8-account-create-update-xv8sm" podStartSLOduration=1.553405905 podStartE2EDuration="1.553405905s" podCreationTimestamp="2026-03-13 14:25:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:25:57.552416557 +0000 UTC m=+1256.452204236" watchObservedRunningTime="2026-03-13 14:25:57.553405905 +0000 UTC m=+1256.453193594" Mar 13 14:25:58 crc kubenswrapper[4907]: I0313 14:25:58.163322 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/14f05f31-8185-43b0-be69-bcf8d5388ea5-etc-swift\") pod \"swift-storage-0\" (UID: \"14f05f31-8185-43b0-be69-bcf8d5388ea5\") " pod="openstack/swift-storage-0" Mar 13 14:25:58 crc kubenswrapper[4907]: E0313 14:25:58.163586 4907 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Mar 13 14:25:58 crc kubenswrapper[4907]: E0313 14:25:58.163829 4907 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Mar 13 14:25:58 crc kubenswrapper[4907]: E0313 14:25:58.163926 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/14f05f31-8185-43b0-be69-bcf8d5388ea5-etc-swift podName:14f05f31-8185-43b0-be69-bcf8d5388ea5 nodeName:}" failed. No retries permitted until 2026-03-13 14:26:02.163875069 +0000 UTC m=+1261.063662768 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/14f05f31-8185-43b0-be69-bcf8d5388ea5-etc-swift") pod "swift-storage-0" (UID: "14f05f31-8185-43b0-be69-bcf8d5388ea5") : configmap "swift-ring-files" not found Mar 13 14:25:58 crc kubenswrapper[4907]: I0313 14:25:58.287069 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-9fxfh"] Mar 13 14:25:58 crc kubenswrapper[4907]: I0313 14:25:58.288931 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-9fxfh" Mar 13 14:25:58 crc kubenswrapper[4907]: I0313 14:25:58.296213 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Mar 13 14:25:58 crc kubenswrapper[4907]: I0313 14:25:58.296296 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Mar 13 14:25:58 crc kubenswrapper[4907]: I0313 14:25:58.296443 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Mar 13 14:25:58 crc kubenswrapper[4907]: I0313 14:25:58.309288 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-9fxfh"] Mar 13 14:25:58 crc kubenswrapper[4907]: I0313 14:25:58.357805 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/root-account-create-update-5lhdv"] Mar 13 14:25:58 crc kubenswrapper[4907]: I0313 14:25:58.358768 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-5lhdv" Mar 13 14:25:58 crc kubenswrapper[4907]: I0313 14:25:58.361266 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-mariadb-root-db-secret" Mar 13 14:25:58 crc kubenswrapper[4907]: I0313 14:25:58.366521 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/e2585497-6837-4d17-8d51-c3d7879fdb46-ring-data-devices\") pod \"swift-ring-rebalance-9fxfh\" (UID: \"e2585497-6837-4d17-8d51-c3d7879fdb46\") " pod="openstack/swift-ring-rebalance-9fxfh" Mar 13 14:25:58 crc kubenswrapper[4907]: I0313 14:25:58.366571 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-msr86\" (UniqueName: \"kubernetes.io/projected/e2585497-6837-4d17-8d51-c3d7879fdb46-kube-api-access-msr86\") pod \"swift-ring-rebalance-9fxfh\" (UID: \"e2585497-6837-4d17-8d51-c3d7879fdb46\") " pod="openstack/swift-ring-rebalance-9fxfh" Mar 13 14:25:58 crc kubenswrapper[4907]: I0313 14:25:58.366593 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/e2585497-6837-4d17-8d51-c3d7879fdb46-dispersionconf\") pod \"swift-ring-rebalance-9fxfh\" (UID: \"e2585497-6837-4d17-8d51-c3d7879fdb46\") " pod="openstack/swift-ring-rebalance-9fxfh" Mar 13 14:25:58 crc kubenswrapper[4907]: I0313 14:25:58.366628 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/e2585497-6837-4d17-8d51-c3d7879fdb46-etc-swift\") pod \"swift-ring-rebalance-9fxfh\" (UID: \"e2585497-6837-4d17-8d51-c3d7879fdb46\") " pod="openstack/swift-ring-rebalance-9fxfh" Mar 13 14:25:58 crc kubenswrapper[4907]: I0313 14:25:58.366695 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2585497-6837-4d17-8d51-c3d7879fdb46-combined-ca-bundle\") pod \"swift-ring-rebalance-9fxfh\" (UID: \"e2585497-6837-4d17-8d51-c3d7879fdb46\") " pod="openstack/swift-ring-rebalance-9fxfh" Mar 13 14:25:58 crc kubenswrapper[4907]: I0313 14:25:58.366728 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/e2585497-6837-4d17-8d51-c3d7879fdb46-swiftconf\") pod \"swift-ring-rebalance-9fxfh\" (UID: \"e2585497-6837-4d17-8d51-c3d7879fdb46\") " pod="openstack/swift-ring-rebalance-9fxfh" Mar 13 14:25:58 crc kubenswrapper[4907]: I0313 14:25:58.366758 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e2585497-6837-4d17-8d51-c3d7879fdb46-scripts\") pod \"swift-ring-rebalance-9fxfh\" (UID: \"e2585497-6837-4d17-8d51-c3d7879fdb46\") " pod="openstack/swift-ring-rebalance-9fxfh" Mar 13 14:25:58 crc kubenswrapper[4907]: I0313 14:25:58.370388 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-5lhdv"] Mar 13 14:25:58 crc kubenswrapper[4907]: I0313 14:25:58.467765 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/e2585497-6837-4d17-8d51-c3d7879fdb46-etc-swift\") pod \"swift-ring-rebalance-9fxfh\" (UID: \"e2585497-6837-4d17-8d51-c3d7879fdb46\") " pod="openstack/swift-ring-rebalance-9fxfh" Mar 13 14:25:58 crc kubenswrapper[4907]: I0313 14:25:58.467859 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2585497-6837-4d17-8d51-c3d7879fdb46-combined-ca-bundle\") pod \"swift-ring-rebalance-9fxfh\" (UID: \"e2585497-6837-4d17-8d51-c3d7879fdb46\") " pod="openstack/swift-ring-rebalance-9fxfh" Mar 13 14:25:58 crc kubenswrapper[4907]: I0313 14:25:58.467910 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/e2585497-6837-4d17-8d51-c3d7879fdb46-swiftconf\") pod \"swift-ring-rebalance-9fxfh\" (UID: \"e2585497-6837-4d17-8d51-c3d7879fdb46\") " pod="openstack/swift-ring-rebalance-9fxfh" Mar 13 14:25:58 crc kubenswrapper[4907]: I0313 14:25:58.467974 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zqzfb\" (UniqueName: \"kubernetes.io/projected/f12f5638-4cb2-4424-86d8-5c0c829d82d5-kube-api-access-zqzfb\") pod \"root-account-create-update-5lhdv\" (UID: \"f12f5638-4cb2-4424-86d8-5c0c829d82d5\") " pod="openstack/root-account-create-update-5lhdv" Mar 13 14:25:58 crc kubenswrapper[4907]: I0313 14:25:58.468029 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e2585497-6837-4d17-8d51-c3d7879fdb46-scripts\") pod \"swift-ring-rebalance-9fxfh\" (UID: \"e2585497-6837-4d17-8d51-c3d7879fdb46\") " pod="openstack/swift-ring-rebalance-9fxfh" Mar 13 14:25:58 crc kubenswrapper[4907]: I0313 14:25:58.468079 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/e2585497-6837-4d17-8d51-c3d7879fdb46-ring-data-devices\") pod \"swift-ring-rebalance-9fxfh\" (UID: \"e2585497-6837-4d17-8d51-c3d7879fdb46\") " pod="openstack/swift-ring-rebalance-9fxfh" Mar 13 14:25:58 crc kubenswrapper[4907]: I0313 14:25:58.468106 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f12f5638-4cb2-4424-86d8-5c0c829d82d5-operator-scripts\") pod \"root-account-create-update-5lhdv\" (UID: \"f12f5638-4cb2-4424-86d8-5c0c829d82d5\") " pod="openstack/root-account-create-update-5lhdv" Mar 13 14:25:58 crc kubenswrapper[4907]: I0313 14:25:58.468146 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-msr86\" (UniqueName: \"kubernetes.io/projected/e2585497-6837-4d17-8d51-c3d7879fdb46-kube-api-access-msr86\") pod \"swift-ring-rebalance-9fxfh\" (UID: \"e2585497-6837-4d17-8d51-c3d7879fdb46\") " pod="openstack/swift-ring-rebalance-9fxfh" Mar 13 14:25:58 crc kubenswrapper[4907]: I0313 14:25:58.468165 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/e2585497-6837-4d17-8d51-c3d7879fdb46-dispersionconf\") pod \"swift-ring-rebalance-9fxfh\" (UID: \"e2585497-6837-4d17-8d51-c3d7879fdb46\") " pod="openstack/swift-ring-rebalance-9fxfh" Mar 13 14:25:58 crc kubenswrapper[4907]: I0313 14:25:58.468304 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/e2585497-6837-4d17-8d51-c3d7879fdb46-etc-swift\") pod \"swift-ring-rebalance-9fxfh\" (UID: \"e2585497-6837-4d17-8d51-c3d7879fdb46\") " pod="openstack/swift-ring-rebalance-9fxfh" Mar 13 14:25:58 crc kubenswrapper[4907]: I0313 14:25:58.468683 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e2585497-6837-4d17-8d51-c3d7879fdb46-scripts\") pod \"swift-ring-rebalance-9fxfh\" (UID: \"e2585497-6837-4d17-8d51-c3d7879fdb46\") " pod="openstack/swift-ring-rebalance-9fxfh" Mar 13 14:25:58 crc kubenswrapper[4907]: I0313 14:25:58.468800 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/e2585497-6837-4d17-8d51-c3d7879fdb46-ring-data-devices\") pod \"swift-ring-rebalance-9fxfh\" (UID: \"e2585497-6837-4d17-8d51-c3d7879fdb46\") " pod="openstack/swift-ring-rebalance-9fxfh" Mar 13 14:25:58 crc kubenswrapper[4907]: I0313 14:25:58.473243 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2585497-6837-4d17-8d51-c3d7879fdb46-combined-ca-bundle\") pod \"swift-ring-rebalance-9fxfh\" (UID: \"e2585497-6837-4d17-8d51-c3d7879fdb46\") " pod="openstack/swift-ring-rebalance-9fxfh" Mar 13 14:25:58 crc kubenswrapper[4907]: I0313 14:25:58.473363 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/e2585497-6837-4d17-8d51-c3d7879fdb46-swiftconf\") pod \"swift-ring-rebalance-9fxfh\" (UID: \"e2585497-6837-4d17-8d51-c3d7879fdb46\") " pod="openstack/swift-ring-rebalance-9fxfh" Mar 13 14:25:58 crc kubenswrapper[4907]: I0313 14:25:58.474033 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/e2585497-6837-4d17-8d51-c3d7879fdb46-dispersionconf\") pod \"swift-ring-rebalance-9fxfh\" (UID: \"e2585497-6837-4d17-8d51-c3d7879fdb46\") " pod="openstack/swift-ring-rebalance-9fxfh" Mar 13 14:25:58 crc kubenswrapper[4907]: I0313 14:25:58.483447 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-msr86\" (UniqueName: \"kubernetes.io/projected/e2585497-6837-4d17-8d51-c3d7879fdb46-kube-api-access-msr86\") pod \"swift-ring-rebalance-9fxfh\" (UID: \"e2585497-6837-4d17-8d51-c3d7879fdb46\") " pod="openstack/swift-ring-rebalance-9fxfh" Mar 13 14:25:58 crc kubenswrapper[4907]: I0313 14:25:58.532215 4907 generic.go:334] "Generic (PLEG): container finished" podID="cbb6c089-0112-43f4-8731-c2d68932795a" containerID="b2e917cc4c559e06be2ed2a503af651b2cead394bb06409f6e1c888788df817a" exitCode=0 Mar 13 14:25:58 crc kubenswrapper[4907]: I0313 14:25:58.532278 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-17a8-account-create-update-xv8sm" event={"ID":"cbb6c089-0112-43f4-8731-c2d68932795a","Type":"ContainerDied","Data":"b2e917cc4c559e06be2ed2a503af651b2cead394bb06409f6e1c888788df817a"} Mar 13 14:25:58 crc kubenswrapper[4907]: I0313 14:25:58.535028 4907 generic.go:334] "Generic (PLEG): container finished" podID="4d7d2639-4229-4286-be5c-d8b15ed91d17" containerID="3c90645adf755c84a8db83dd2e03403f66a250afbd541346b9261208e83c56ff" exitCode=0 Mar 13 14:25:58 crc kubenswrapper[4907]: I0313 14:25:58.535059 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-sgpkl" event={"ID":"4d7d2639-4229-4286-be5c-d8b15ed91d17","Type":"ContainerDied","Data":"3c90645adf755c84a8db83dd2e03403f66a250afbd541346b9261208e83c56ff"} Mar 13 14:25:58 crc kubenswrapper[4907]: I0313 14:25:58.569863 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zqzfb\" (UniqueName: \"kubernetes.io/projected/f12f5638-4cb2-4424-86d8-5c0c829d82d5-kube-api-access-zqzfb\") pod \"root-account-create-update-5lhdv\" (UID: \"f12f5638-4cb2-4424-86d8-5c0c829d82d5\") " pod="openstack/root-account-create-update-5lhdv" Mar 13 14:25:58 crc kubenswrapper[4907]: I0313 14:25:58.569971 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f12f5638-4cb2-4424-86d8-5c0c829d82d5-operator-scripts\") pod \"root-account-create-update-5lhdv\" (UID: \"f12f5638-4cb2-4424-86d8-5c0c829d82d5\") " pod="openstack/root-account-create-update-5lhdv" Mar 13 14:25:58 crc kubenswrapper[4907]: I0313 14:25:58.570701 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f12f5638-4cb2-4424-86d8-5c0c829d82d5-operator-scripts\") pod \"root-account-create-update-5lhdv\" (UID: \"f12f5638-4cb2-4424-86d8-5c0c829d82d5\") " pod="openstack/root-account-create-update-5lhdv" Mar 13 14:25:58 crc kubenswrapper[4907]: I0313 14:25:58.588092 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zqzfb\" (UniqueName: \"kubernetes.io/projected/f12f5638-4cb2-4424-86d8-5c0c829d82d5-kube-api-access-zqzfb\") pod \"root-account-create-update-5lhdv\" (UID: \"f12f5638-4cb2-4424-86d8-5c0c829d82d5\") " pod="openstack/root-account-create-update-5lhdv" Mar 13 14:25:58 crc kubenswrapper[4907]: I0313 14:25:58.613331 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-9fxfh" Mar 13 14:25:58 crc kubenswrapper[4907]: I0313 14:25:58.722599 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-5lhdv" Mar 13 14:25:59 crc kubenswrapper[4907]: W0313 14:25:59.031103 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode2585497_6837_4d17_8d51_c3d7879fdb46.slice/crio-c44bcf911a2f35735f5bae70947997de7ad5c7bedca91be9607b903b91ea86c5 WatchSource:0}: Error finding container c44bcf911a2f35735f5bae70947997de7ad5c7bedca91be9607b903b91ea86c5: Status 404 returned error can't find the container with id c44bcf911a2f35735f5bae70947997de7ad5c7bedca91be9607b903b91ea86c5 Mar 13 14:25:59 crc kubenswrapper[4907]: I0313 14:25:59.036395 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-9fxfh"] Mar 13 14:25:59 crc kubenswrapper[4907]: I0313 14:25:59.178443 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-5lhdv"] Mar 13 14:25:59 crc kubenswrapper[4907]: I0313 14:25:59.547157 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-9fxfh" event={"ID":"e2585497-6837-4d17-8d51-c3d7879fdb46","Type":"ContainerStarted","Data":"c44bcf911a2f35735f5bae70947997de7ad5c7bedca91be9607b903b91ea86c5"} Mar 13 14:25:59 crc kubenswrapper[4907]: I0313 14:25:59.551267 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-5lhdv" event={"ID":"f12f5638-4cb2-4424-86d8-5c0c829d82d5","Type":"ContainerStarted","Data":"c688bfe6d6ea48e683001f8125ee585ba89819e96749b44bbe5f071b99df1226"} Mar 13 14:25:59 crc kubenswrapper[4907]: I0313 14:25:59.551315 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-5lhdv" event={"ID":"f12f5638-4cb2-4424-86d8-5c0c829d82d5","Type":"ContainerStarted","Data":"d2489c29658426420882959df704422f457e13151fc0758486e6c026a736b39b"} Mar 13 14:25:59 crc kubenswrapper[4907]: I0313 14:25:59.577148 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/root-account-create-update-5lhdv" podStartSLOduration=1.577073028 podStartE2EDuration="1.577073028s" podCreationTimestamp="2026-03-13 14:25:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:25:59.56948906 +0000 UTC m=+1258.469276759" watchObservedRunningTime="2026-03-13 14:25:59.577073028 +0000 UTC m=+1258.476860717" Mar 13 14:26:00 crc kubenswrapper[4907]: I0313 14:26:00.133760 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556866-fwg48"] Mar 13 14:26:00 crc kubenswrapper[4907]: I0313 14:26:00.135032 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556866-fwg48" Mar 13 14:26:00 crc kubenswrapper[4907]: I0313 14:26:00.136786 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 14:26:00 crc kubenswrapper[4907]: I0313 14:26:00.136975 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 14:26:00 crc kubenswrapper[4907]: I0313 14:26:00.137909 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 14:26:00 crc kubenswrapper[4907]: I0313 14:26:00.140198 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556866-fwg48"] Mar 13 14:26:00 crc kubenswrapper[4907]: I0313 14:26:00.201675 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f5d44\" (UniqueName: \"kubernetes.io/projected/7d930b98-c1af-4e47-a663-86afa9484856-kube-api-access-f5d44\") pod \"auto-csr-approver-29556866-fwg48\" (UID: \"7d930b98-c1af-4e47-a663-86afa9484856\") " pod="openshift-infra/auto-csr-approver-29556866-fwg48" Mar 13 14:26:00 crc kubenswrapper[4907]: I0313 14:26:00.210615 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-17a8-account-create-update-xv8sm" Mar 13 14:26:00 crc kubenswrapper[4907]: I0313 14:26:00.217731 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-sgpkl" Mar 13 14:26:00 crc kubenswrapper[4907]: I0313 14:26:00.303224 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4q8qf\" (UniqueName: \"kubernetes.io/projected/4d7d2639-4229-4286-be5c-d8b15ed91d17-kube-api-access-4q8qf\") pod \"4d7d2639-4229-4286-be5c-d8b15ed91d17\" (UID: \"4d7d2639-4229-4286-be5c-d8b15ed91d17\") " Mar 13 14:26:00 crc kubenswrapper[4907]: I0313 14:26:00.303408 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4d7d2639-4229-4286-be5c-d8b15ed91d17-operator-scripts\") pod \"4d7d2639-4229-4286-be5c-d8b15ed91d17\" (UID: \"4d7d2639-4229-4286-be5c-d8b15ed91d17\") " Mar 13 14:26:00 crc kubenswrapper[4907]: I0313 14:26:00.303467 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkz6n\" (UniqueName: \"kubernetes.io/projected/cbb6c089-0112-43f4-8731-c2d68932795a-kube-api-access-zkz6n\") pod \"cbb6c089-0112-43f4-8731-c2d68932795a\" (UID: \"cbb6c089-0112-43f4-8731-c2d68932795a\") " Mar 13 14:26:00 crc kubenswrapper[4907]: I0313 14:26:00.303499 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cbb6c089-0112-43f4-8731-c2d68932795a-operator-scripts\") pod \"cbb6c089-0112-43f4-8731-c2d68932795a\" (UID: \"cbb6c089-0112-43f4-8731-c2d68932795a\") " Mar 13 14:26:00 crc kubenswrapper[4907]: I0313 14:26:00.303943 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4d7d2639-4229-4286-be5c-d8b15ed91d17-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "4d7d2639-4229-4286-be5c-d8b15ed91d17" (UID: "4d7d2639-4229-4286-be5c-d8b15ed91d17"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:26:00 crc kubenswrapper[4907]: I0313 14:26:00.304292 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cbb6c089-0112-43f4-8731-c2d68932795a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "cbb6c089-0112-43f4-8731-c2d68932795a" (UID: "cbb6c089-0112-43f4-8731-c2d68932795a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:26:00 crc kubenswrapper[4907]: I0313 14:26:00.304459 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f5d44\" (UniqueName: \"kubernetes.io/projected/7d930b98-c1af-4e47-a663-86afa9484856-kube-api-access-f5d44\") pod \"auto-csr-approver-29556866-fwg48\" (UID: \"7d930b98-c1af-4e47-a663-86afa9484856\") " pod="openshift-infra/auto-csr-approver-29556866-fwg48" Mar 13 14:26:00 crc kubenswrapper[4907]: I0313 14:26:00.304540 4907 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4d7d2639-4229-4286-be5c-d8b15ed91d17-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:00 crc kubenswrapper[4907]: I0313 14:26:00.304556 4907 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cbb6c089-0112-43f4-8731-c2d68932795a-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:00 crc kubenswrapper[4907]: I0313 14:26:00.309135 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cbb6c089-0112-43f4-8731-c2d68932795a-kube-api-access-zkz6n" (OuterVolumeSpecName: "kube-api-access-zkz6n") pod "cbb6c089-0112-43f4-8731-c2d68932795a" (UID: "cbb6c089-0112-43f4-8731-c2d68932795a"). InnerVolumeSpecName "kube-api-access-zkz6n". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:26:00 crc kubenswrapper[4907]: I0313 14:26:00.309466 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d7d2639-4229-4286-be5c-d8b15ed91d17-kube-api-access-4q8qf" (OuterVolumeSpecName: "kube-api-access-4q8qf") pod "4d7d2639-4229-4286-be5c-d8b15ed91d17" (UID: "4d7d2639-4229-4286-be5c-d8b15ed91d17"). InnerVolumeSpecName "kube-api-access-4q8qf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:26:00 crc kubenswrapper[4907]: I0313 14:26:00.322191 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f5d44\" (UniqueName: \"kubernetes.io/projected/7d930b98-c1af-4e47-a663-86afa9484856-kube-api-access-f5d44\") pod \"auto-csr-approver-29556866-fwg48\" (UID: \"7d930b98-c1af-4e47-a663-86afa9484856\") " pod="openshift-infra/auto-csr-approver-29556866-fwg48" Mar 13 14:26:00 crc kubenswrapper[4907]: I0313 14:26:00.406062 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkz6n\" (UniqueName: \"kubernetes.io/projected/cbb6c089-0112-43f4-8731-c2d68932795a-kube-api-access-zkz6n\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:00 crc kubenswrapper[4907]: I0313 14:26:00.406098 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4q8qf\" (UniqueName: \"kubernetes.io/projected/4d7d2639-4229-4286-be5c-d8b15ed91d17-kube-api-access-4q8qf\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:00 crc kubenswrapper[4907]: I0313 14:26:00.538405 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556866-fwg48" Mar 13 14:26:00 crc kubenswrapper[4907]: I0313 14:26:00.568570 4907 generic.go:334] "Generic (PLEG): container finished" podID="f12f5638-4cb2-4424-86d8-5c0c829d82d5" containerID="c688bfe6d6ea48e683001f8125ee585ba89819e96749b44bbe5f071b99df1226" exitCode=0 Mar 13 14:26:00 crc kubenswrapper[4907]: I0313 14:26:00.568667 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-5lhdv" event={"ID":"f12f5638-4cb2-4424-86d8-5c0c829d82d5","Type":"ContainerDied","Data":"c688bfe6d6ea48e683001f8125ee585ba89819e96749b44bbe5f071b99df1226"} Mar 13 14:26:00 crc kubenswrapper[4907]: I0313 14:26:00.574831 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-17a8-account-create-update-xv8sm" event={"ID":"cbb6c089-0112-43f4-8731-c2d68932795a","Type":"ContainerDied","Data":"99873ca15b9fb22afcd6ec2e9a14245394d9db6cee5f1f289e90f95b20ea6652"} Mar 13 14:26:00 crc kubenswrapper[4907]: I0313 14:26:00.574869 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="99873ca15b9fb22afcd6ec2e9a14245394d9db6cee5f1f289e90f95b20ea6652" Mar 13 14:26:00 crc kubenswrapper[4907]: I0313 14:26:00.575006 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-17a8-account-create-update-xv8sm" Mar 13 14:26:00 crc kubenswrapper[4907]: I0313 14:26:00.578813 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-sgpkl" event={"ID":"4d7d2639-4229-4286-be5c-d8b15ed91d17","Type":"ContainerDied","Data":"64d9b417d3e9f250fd98eb8db5c1fbf579e78ade9d4cbc4644beb2dad8d75fbc"} Mar 13 14:26:00 crc kubenswrapper[4907]: I0313 14:26:00.578840 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-sgpkl" Mar 13 14:26:00 crc kubenswrapper[4907]: I0313 14:26:00.578857 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="64d9b417d3e9f250fd98eb8db5c1fbf579e78ade9d4cbc4644beb2dad8d75fbc" Mar 13 14:26:01 crc kubenswrapper[4907]: I0313 14:26:01.761696 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-4v4ps"] Mar 13 14:26:01 crc kubenswrapper[4907]: E0313 14:26:01.762514 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d7d2639-4229-4286-be5c-d8b15ed91d17" containerName="mariadb-database-create" Mar 13 14:26:01 crc kubenswrapper[4907]: I0313 14:26:01.762533 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d7d2639-4229-4286-be5c-d8b15ed91d17" containerName="mariadb-database-create" Mar 13 14:26:01 crc kubenswrapper[4907]: E0313 14:26:01.762566 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cbb6c089-0112-43f4-8731-c2d68932795a" containerName="mariadb-account-create-update" Mar 13 14:26:01 crc kubenswrapper[4907]: I0313 14:26:01.762574 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="cbb6c089-0112-43f4-8731-c2d68932795a" containerName="mariadb-account-create-update" Mar 13 14:26:01 crc kubenswrapper[4907]: I0313 14:26:01.762769 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="cbb6c089-0112-43f4-8731-c2d68932795a" containerName="mariadb-account-create-update" Mar 13 14:26:01 crc kubenswrapper[4907]: I0313 14:26:01.762785 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d7d2639-4229-4286-be5c-d8b15ed91d17" containerName="mariadb-database-create" Mar 13 14:26:01 crc kubenswrapper[4907]: I0313 14:26:01.763425 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-4v4ps" Mar 13 14:26:01 crc kubenswrapper[4907]: I0313 14:26:01.766029 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Mar 13 14:26:01 crc kubenswrapper[4907]: I0313 14:26:01.766286 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-fc9j2" Mar 13 14:26:01 crc kubenswrapper[4907]: I0313 14:26:01.775990 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-4v4ps"] Mar 13 14:26:01 crc kubenswrapper[4907]: I0313 14:26:01.827503 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02262347-eaec-462e-a3ef-58c670885d99-combined-ca-bundle\") pod \"glance-db-sync-4v4ps\" (UID: \"02262347-eaec-462e-a3ef-58c670885d99\") " pod="openstack/glance-db-sync-4v4ps" Mar 13 14:26:01 crc kubenswrapper[4907]: I0313 14:26:01.827874 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02262347-eaec-462e-a3ef-58c670885d99-config-data\") pod \"glance-db-sync-4v4ps\" (UID: \"02262347-eaec-462e-a3ef-58c670885d99\") " pod="openstack/glance-db-sync-4v4ps" Mar 13 14:26:01 crc kubenswrapper[4907]: I0313 14:26:01.828401 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-65qwz\" (UniqueName: \"kubernetes.io/projected/02262347-eaec-462e-a3ef-58c670885d99-kube-api-access-65qwz\") pod \"glance-db-sync-4v4ps\" (UID: \"02262347-eaec-462e-a3ef-58c670885d99\") " pod="openstack/glance-db-sync-4v4ps" Mar 13 14:26:01 crc kubenswrapper[4907]: I0313 14:26:01.828826 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/02262347-eaec-462e-a3ef-58c670885d99-db-sync-config-data\") pod \"glance-db-sync-4v4ps\" (UID: \"02262347-eaec-462e-a3ef-58c670885d99\") " pod="openstack/glance-db-sync-4v4ps" Mar 13 14:26:01 crc kubenswrapper[4907]: I0313 14:26:01.930301 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/02262347-eaec-462e-a3ef-58c670885d99-db-sync-config-data\") pod \"glance-db-sync-4v4ps\" (UID: \"02262347-eaec-462e-a3ef-58c670885d99\") " pod="openstack/glance-db-sync-4v4ps" Mar 13 14:26:01 crc kubenswrapper[4907]: I0313 14:26:01.930392 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02262347-eaec-462e-a3ef-58c670885d99-combined-ca-bundle\") pod \"glance-db-sync-4v4ps\" (UID: \"02262347-eaec-462e-a3ef-58c670885d99\") " pod="openstack/glance-db-sync-4v4ps" Mar 13 14:26:01 crc kubenswrapper[4907]: I0313 14:26:01.930568 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02262347-eaec-462e-a3ef-58c670885d99-config-data\") pod \"glance-db-sync-4v4ps\" (UID: \"02262347-eaec-462e-a3ef-58c670885d99\") " pod="openstack/glance-db-sync-4v4ps" Mar 13 14:26:01 crc kubenswrapper[4907]: I0313 14:26:01.930593 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-65qwz\" (UniqueName: \"kubernetes.io/projected/02262347-eaec-462e-a3ef-58c670885d99-kube-api-access-65qwz\") pod \"glance-db-sync-4v4ps\" (UID: \"02262347-eaec-462e-a3ef-58c670885d99\") " pod="openstack/glance-db-sync-4v4ps" Mar 13 14:26:01 crc kubenswrapper[4907]: I0313 14:26:01.937985 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/02262347-eaec-462e-a3ef-58c670885d99-db-sync-config-data\") pod \"glance-db-sync-4v4ps\" (UID: \"02262347-eaec-462e-a3ef-58c670885d99\") " pod="openstack/glance-db-sync-4v4ps" Mar 13 14:26:01 crc kubenswrapper[4907]: I0313 14:26:01.938074 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02262347-eaec-462e-a3ef-58c670885d99-combined-ca-bundle\") pod \"glance-db-sync-4v4ps\" (UID: \"02262347-eaec-462e-a3ef-58c670885d99\") " pod="openstack/glance-db-sync-4v4ps" Mar 13 14:26:01 crc kubenswrapper[4907]: I0313 14:26:01.940764 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02262347-eaec-462e-a3ef-58c670885d99-config-data\") pod \"glance-db-sync-4v4ps\" (UID: \"02262347-eaec-462e-a3ef-58c670885d99\") " pod="openstack/glance-db-sync-4v4ps" Mar 13 14:26:01 crc kubenswrapper[4907]: I0313 14:26:01.949509 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-65qwz\" (UniqueName: \"kubernetes.io/projected/02262347-eaec-462e-a3ef-58c670885d99-kube-api-access-65qwz\") pod \"glance-db-sync-4v4ps\" (UID: \"02262347-eaec-462e-a3ef-58c670885d99\") " pod="openstack/glance-db-sync-4v4ps" Mar 13 14:26:02 crc kubenswrapper[4907]: I0313 14:26:02.097099 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-4v4ps" Mar 13 14:26:02 crc kubenswrapper[4907]: I0313 14:26:02.212718 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-hg688"] Mar 13 14:26:02 crc kubenswrapper[4907]: I0313 14:26:02.213782 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-hg688" Mar 13 14:26:02 crc kubenswrapper[4907]: I0313 14:26:02.225871 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-hg688"] Mar 13 14:26:02 crc kubenswrapper[4907]: I0313 14:26:02.236391 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/14f05f31-8185-43b0-be69-bcf8d5388ea5-etc-swift\") pod \"swift-storage-0\" (UID: \"14f05f31-8185-43b0-be69-bcf8d5388ea5\") " pod="openstack/swift-storage-0" Mar 13 14:26:02 crc kubenswrapper[4907]: E0313 14:26:02.236629 4907 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Mar 13 14:26:02 crc kubenswrapper[4907]: E0313 14:26:02.236646 4907 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Mar 13 14:26:02 crc kubenswrapper[4907]: E0313 14:26:02.236692 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/14f05f31-8185-43b0-be69-bcf8d5388ea5-etc-swift podName:14f05f31-8185-43b0-be69-bcf8d5388ea5 nodeName:}" failed. No retries permitted until 2026-03-13 14:26:10.236676855 +0000 UTC m=+1269.136464544 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/14f05f31-8185-43b0-be69-bcf8d5388ea5-etc-swift") pod "swift-storage-0" (UID: "14f05f31-8185-43b0-be69-bcf8d5388ea5") : configmap "swift-ring-files" not found Mar 13 14:26:02 crc kubenswrapper[4907]: I0313 14:26:02.314967 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-04a8-account-create-update-ws5ht"] Mar 13 14:26:02 crc kubenswrapper[4907]: I0313 14:26:02.316099 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-04a8-account-create-update-ws5ht" Mar 13 14:26:02 crc kubenswrapper[4907]: I0313 14:26:02.319469 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Mar 13 14:26:02 crc kubenswrapper[4907]: I0313 14:26:02.331738 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-04a8-account-create-update-ws5ht"] Mar 13 14:26:02 crc kubenswrapper[4907]: I0313 14:26:02.337636 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9c985ff9-3545-4544-acc8-ca9eb2bfad6a-operator-scripts\") pod \"keystone-db-create-hg688\" (UID: \"9c985ff9-3545-4544-acc8-ca9eb2bfad6a\") " pod="openstack/keystone-db-create-hg688" Mar 13 14:26:02 crc kubenswrapper[4907]: I0313 14:26:02.337746 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qcmkf\" (UniqueName: \"kubernetes.io/projected/9c985ff9-3545-4544-acc8-ca9eb2bfad6a-kube-api-access-qcmkf\") pod \"keystone-db-create-hg688\" (UID: \"9c985ff9-3545-4544-acc8-ca9eb2bfad6a\") " pod="openstack/keystone-db-create-hg688" Mar 13 14:26:02 crc kubenswrapper[4907]: I0313 14:26:02.439559 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qcmkf\" (UniqueName: \"kubernetes.io/projected/9c985ff9-3545-4544-acc8-ca9eb2bfad6a-kube-api-access-qcmkf\") pod \"keystone-db-create-hg688\" (UID: \"9c985ff9-3545-4544-acc8-ca9eb2bfad6a\") " pod="openstack/keystone-db-create-hg688" Mar 13 14:26:02 crc kubenswrapper[4907]: I0313 14:26:02.439628 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b4bzw\" (UniqueName: \"kubernetes.io/projected/37ba2d41-5de4-48cb-aad6-59df13307ed1-kube-api-access-b4bzw\") pod \"keystone-04a8-account-create-update-ws5ht\" (UID: \"37ba2d41-5de4-48cb-aad6-59df13307ed1\") " pod="openstack/keystone-04a8-account-create-update-ws5ht" Mar 13 14:26:02 crc kubenswrapper[4907]: I0313 14:26:02.439873 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/37ba2d41-5de4-48cb-aad6-59df13307ed1-operator-scripts\") pod \"keystone-04a8-account-create-update-ws5ht\" (UID: \"37ba2d41-5de4-48cb-aad6-59df13307ed1\") " pod="openstack/keystone-04a8-account-create-update-ws5ht" Mar 13 14:26:02 crc kubenswrapper[4907]: I0313 14:26:02.440034 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9c985ff9-3545-4544-acc8-ca9eb2bfad6a-operator-scripts\") pod \"keystone-db-create-hg688\" (UID: \"9c985ff9-3545-4544-acc8-ca9eb2bfad6a\") " pod="openstack/keystone-db-create-hg688" Mar 13 14:26:02 crc kubenswrapper[4907]: I0313 14:26:02.440843 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9c985ff9-3545-4544-acc8-ca9eb2bfad6a-operator-scripts\") pod \"keystone-db-create-hg688\" (UID: \"9c985ff9-3545-4544-acc8-ca9eb2bfad6a\") " pod="openstack/keystone-db-create-hg688" Mar 13 14:26:02 crc kubenswrapper[4907]: I0313 14:26:02.456522 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qcmkf\" (UniqueName: \"kubernetes.io/projected/9c985ff9-3545-4544-acc8-ca9eb2bfad6a-kube-api-access-qcmkf\") pod \"keystone-db-create-hg688\" (UID: \"9c985ff9-3545-4544-acc8-ca9eb2bfad6a\") " pod="openstack/keystone-db-create-hg688" Mar 13 14:26:02 crc kubenswrapper[4907]: I0313 14:26:02.526410 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-cnnxs"] Mar 13 14:26:02 crc kubenswrapper[4907]: I0313 14:26:02.527557 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-cnnxs" Mar 13 14:26:02 crc kubenswrapper[4907]: I0313 14:26:02.532632 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-hg688" Mar 13 14:26:02 crc kubenswrapper[4907]: I0313 14:26:02.537196 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-cnnxs"] Mar 13 14:26:02 crc kubenswrapper[4907]: I0313 14:26:02.542857 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/37ba2d41-5de4-48cb-aad6-59df13307ed1-operator-scripts\") pod \"keystone-04a8-account-create-update-ws5ht\" (UID: \"37ba2d41-5de4-48cb-aad6-59df13307ed1\") " pod="openstack/keystone-04a8-account-create-update-ws5ht" Mar 13 14:26:02 crc kubenswrapper[4907]: I0313 14:26:02.543011 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b4bzw\" (UniqueName: \"kubernetes.io/projected/37ba2d41-5de4-48cb-aad6-59df13307ed1-kube-api-access-b4bzw\") pod \"keystone-04a8-account-create-update-ws5ht\" (UID: \"37ba2d41-5de4-48cb-aad6-59df13307ed1\") " pod="openstack/keystone-04a8-account-create-update-ws5ht" Mar 13 14:26:02 crc kubenswrapper[4907]: I0313 14:26:02.543781 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/37ba2d41-5de4-48cb-aad6-59df13307ed1-operator-scripts\") pod \"keystone-04a8-account-create-update-ws5ht\" (UID: \"37ba2d41-5de4-48cb-aad6-59df13307ed1\") " pod="openstack/keystone-04a8-account-create-update-ws5ht" Mar 13 14:26:02 crc kubenswrapper[4907]: I0313 14:26:02.543844 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-9c64-account-create-update-v65tv"] Mar 13 14:26:02 crc kubenswrapper[4907]: I0313 14:26:02.545079 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-9c64-account-create-update-v65tv" Mar 13 14:26:02 crc kubenswrapper[4907]: I0313 14:26:02.547233 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Mar 13 14:26:02 crc kubenswrapper[4907]: I0313 14:26:02.552315 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-9c64-account-create-update-v65tv"] Mar 13 14:26:02 crc kubenswrapper[4907]: I0313 14:26:02.563158 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b4bzw\" (UniqueName: \"kubernetes.io/projected/37ba2d41-5de4-48cb-aad6-59df13307ed1-kube-api-access-b4bzw\") pod \"keystone-04a8-account-create-update-ws5ht\" (UID: \"37ba2d41-5de4-48cb-aad6-59df13307ed1\") " pod="openstack/keystone-04a8-account-create-update-ws5ht" Mar 13 14:26:02 crc kubenswrapper[4907]: I0313 14:26:02.608284 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-5lhdv" event={"ID":"f12f5638-4cb2-4424-86d8-5c0c829d82d5","Type":"ContainerDied","Data":"d2489c29658426420882959df704422f457e13151fc0758486e6c026a736b39b"} Mar 13 14:26:02 crc kubenswrapper[4907]: I0313 14:26:02.608321 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d2489c29658426420882959df704422f457e13151fc0758486e6c026a736b39b" Mar 13 14:26:02 crc kubenswrapper[4907]: I0313 14:26:02.634017 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-5lhdv" Mar 13 14:26:02 crc kubenswrapper[4907]: I0313 14:26:02.637006 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-04a8-account-create-update-ws5ht" Mar 13 14:26:02 crc kubenswrapper[4907]: I0313 14:26:02.644314 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h5wpg\" (UniqueName: \"kubernetes.io/projected/3db3f0b6-9cae-40fa-b54d-3ed06c568c9c-kube-api-access-h5wpg\") pod \"placement-9c64-account-create-update-v65tv\" (UID: \"3db3f0b6-9cae-40fa-b54d-3ed06c568c9c\") " pod="openstack/placement-9c64-account-create-update-v65tv" Mar 13 14:26:02 crc kubenswrapper[4907]: I0313 14:26:02.644407 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3db3f0b6-9cae-40fa-b54d-3ed06c568c9c-operator-scripts\") pod \"placement-9c64-account-create-update-v65tv\" (UID: \"3db3f0b6-9cae-40fa-b54d-3ed06c568c9c\") " pod="openstack/placement-9c64-account-create-update-v65tv" Mar 13 14:26:02 crc kubenswrapper[4907]: I0313 14:26:02.644437 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3f2b7163-b5d4-4575-a72c-e6ad1b3137df-operator-scripts\") pod \"placement-db-create-cnnxs\" (UID: \"3f2b7163-b5d4-4575-a72c-e6ad1b3137df\") " pod="openstack/placement-db-create-cnnxs" Mar 13 14:26:02 crc kubenswrapper[4907]: I0313 14:26:02.644518 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5cr4l\" (UniqueName: \"kubernetes.io/projected/3f2b7163-b5d4-4575-a72c-e6ad1b3137df-kube-api-access-5cr4l\") pod \"placement-db-create-cnnxs\" (UID: \"3f2b7163-b5d4-4575-a72c-e6ad1b3137df\") " pod="openstack/placement-db-create-cnnxs" Mar 13 14:26:02 crc kubenswrapper[4907]: I0313 14:26:02.745595 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f12f5638-4cb2-4424-86d8-5c0c829d82d5-operator-scripts\") pod \"f12f5638-4cb2-4424-86d8-5c0c829d82d5\" (UID: \"f12f5638-4cb2-4424-86d8-5c0c829d82d5\") " Mar 13 14:26:02 crc kubenswrapper[4907]: I0313 14:26:02.745962 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zqzfb\" (UniqueName: \"kubernetes.io/projected/f12f5638-4cb2-4424-86d8-5c0c829d82d5-kube-api-access-zqzfb\") pod \"f12f5638-4cb2-4424-86d8-5c0c829d82d5\" (UID: \"f12f5638-4cb2-4424-86d8-5c0c829d82d5\") " Mar 13 14:26:02 crc kubenswrapper[4907]: I0313 14:26:02.746200 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3db3f0b6-9cae-40fa-b54d-3ed06c568c9c-operator-scripts\") pod \"placement-9c64-account-create-update-v65tv\" (UID: \"3db3f0b6-9cae-40fa-b54d-3ed06c568c9c\") " pod="openstack/placement-9c64-account-create-update-v65tv" Mar 13 14:26:02 crc kubenswrapper[4907]: I0313 14:26:02.746230 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3f2b7163-b5d4-4575-a72c-e6ad1b3137df-operator-scripts\") pod \"placement-db-create-cnnxs\" (UID: \"3f2b7163-b5d4-4575-a72c-e6ad1b3137df\") " pod="openstack/placement-db-create-cnnxs" Mar 13 14:26:02 crc kubenswrapper[4907]: I0313 14:26:02.746271 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f12f5638-4cb2-4424-86d8-5c0c829d82d5-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f12f5638-4cb2-4424-86d8-5c0c829d82d5" (UID: "f12f5638-4cb2-4424-86d8-5c0c829d82d5"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:26:02 crc kubenswrapper[4907]: I0313 14:26:02.746298 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5cr4l\" (UniqueName: \"kubernetes.io/projected/3f2b7163-b5d4-4575-a72c-e6ad1b3137df-kube-api-access-5cr4l\") pod \"placement-db-create-cnnxs\" (UID: \"3f2b7163-b5d4-4575-a72c-e6ad1b3137df\") " pod="openstack/placement-db-create-cnnxs" Mar 13 14:26:02 crc kubenswrapper[4907]: I0313 14:26:02.746354 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h5wpg\" (UniqueName: \"kubernetes.io/projected/3db3f0b6-9cae-40fa-b54d-3ed06c568c9c-kube-api-access-h5wpg\") pod \"placement-9c64-account-create-update-v65tv\" (UID: \"3db3f0b6-9cae-40fa-b54d-3ed06c568c9c\") " pod="openstack/placement-9c64-account-create-update-v65tv" Mar 13 14:26:02 crc kubenswrapper[4907]: I0313 14:26:02.746416 4907 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f12f5638-4cb2-4424-86d8-5c0c829d82d5-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:02 crc kubenswrapper[4907]: I0313 14:26:02.747054 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3db3f0b6-9cae-40fa-b54d-3ed06c568c9c-operator-scripts\") pod \"placement-9c64-account-create-update-v65tv\" (UID: \"3db3f0b6-9cae-40fa-b54d-3ed06c568c9c\") " pod="openstack/placement-9c64-account-create-update-v65tv" Mar 13 14:26:02 crc kubenswrapper[4907]: I0313 14:26:02.747383 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3f2b7163-b5d4-4575-a72c-e6ad1b3137df-operator-scripts\") pod \"placement-db-create-cnnxs\" (UID: \"3f2b7163-b5d4-4575-a72c-e6ad1b3137df\") " pod="openstack/placement-db-create-cnnxs" Mar 13 14:26:02 crc kubenswrapper[4907]: I0313 14:26:02.751498 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f12f5638-4cb2-4424-86d8-5c0c829d82d5-kube-api-access-zqzfb" (OuterVolumeSpecName: "kube-api-access-zqzfb") pod "f12f5638-4cb2-4424-86d8-5c0c829d82d5" (UID: "f12f5638-4cb2-4424-86d8-5c0c829d82d5"). InnerVolumeSpecName "kube-api-access-zqzfb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:26:02 crc kubenswrapper[4907]: I0313 14:26:02.767343 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5cr4l\" (UniqueName: \"kubernetes.io/projected/3f2b7163-b5d4-4575-a72c-e6ad1b3137df-kube-api-access-5cr4l\") pod \"placement-db-create-cnnxs\" (UID: \"3f2b7163-b5d4-4575-a72c-e6ad1b3137df\") " pod="openstack/placement-db-create-cnnxs" Mar 13 14:26:02 crc kubenswrapper[4907]: I0313 14:26:02.770482 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h5wpg\" (UniqueName: \"kubernetes.io/projected/3db3f0b6-9cae-40fa-b54d-3ed06c568c9c-kube-api-access-h5wpg\") pod \"placement-9c64-account-create-update-v65tv\" (UID: \"3db3f0b6-9cae-40fa-b54d-3ed06c568c9c\") " pod="openstack/placement-9c64-account-create-update-v65tv" Mar 13 14:26:02 crc kubenswrapper[4907]: I0313 14:26:02.847728 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zqzfb\" (UniqueName: \"kubernetes.io/projected/f12f5638-4cb2-4424-86d8-5c0c829d82d5-kube-api-access-zqzfb\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:02 crc kubenswrapper[4907]: I0313 14:26:02.848045 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-cnnxs" Mar 13 14:26:02 crc kubenswrapper[4907]: I0313 14:26:02.864157 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-9c64-account-create-update-v65tv" Mar 13 14:26:03 crc kubenswrapper[4907]: I0313 14:26:03.094721 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556866-fwg48"] Mar 13 14:26:03 crc kubenswrapper[4907]: W0313 14:26:03.106397 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7d930b98_c1af_4e47_a663_86afa9484856.slice/crio-c521a07f9fcca8ad2de643a0b0fa242b546a5a944f92129f2cbcb3aea70c71e9 WatchSource:0}: Error finding container c521a07f9fcca8ad2de643a0b0fa242b546a5a944f92129f2cbcb3aea70c71e9: Status 404 returned error can't find the container with id c521a07f9fcca8ad2de643a0b0fa242b546a5a944f92129f2cbcb3aea70c71e9 Mar 13 14:26:03 crc kubenswrapper[4907]: I0313 14:26:03.171240 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-hg688"] Mar 13 14:26:03 crc kubenswrapper[4907]: I0313 14:26:03.266053 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-4v4ps"] Mar 13 14:26:03 crc kubenswrapper[4907]: I0313 14:26:03.306895 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-04a8-account-create-update-ws5ht"] Mar 13 14:26:03 crc kubenswrapper[4907]: W0313 14:26:03.310637 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod37ba2d41_5de4_48cb_aad6_59df13307ed1.slice/crio-85ee0b6a029ba5152ff5abc51ef7d2c859d0da76a5bac6cb02c6728c11a62a9f WatchSource:0}: Error finding container 85ee0b6a029ba5152ff5abc51ef7d2c859d0da76a5bac6cb02c6728c11a62a9f: Status 404 returned error can't find the container with id 85ee0b6a029ba5152ff5abc51ef7d2c859d0da76a5bac6cb02c6728c11a62a9f Mar 13 14:26:03 crc kubenswrapper[4907]: I0313 14:26:03.387811 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-cnnxs"] Mar 13 14:26:03 crc kubenswrapper[4907]: I0313 14:26:03.394640 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-9c64-account-create-update-v65tv"] Mar 13 14:26:03 crc kubenswrapper[4907]: W0313 14:26:03.401065 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3f2b7163_b5d4_4575_a72c_e6ad1b3137df.slice/crio-03e0154228fb7cb7a65f349825e100c422b39b776d8d1f870768bf174633ebf0 WatchSource:0}: Error finding container 03e0154228fb7cb7a65f349825e100c422b39b776d8d1f870768bf174633ebf0: Status 404 returned error can't find the container with id 03e0154228fb7cb7a65f349825e100c422b39b776d8d1f870768bf174633ebf0 Mar 13 14:26:03 crc kubenswrapper[4907]: W0313 14:26:03.404027 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3db3f0b6_9cae_40fa_b54d_3ed06c568c9c.slice/crio-46e329b2292a23b4e7db5c2587420d8b2ff9fb0168254af0462a3f9acca4b553 WatchSource:0}: Error finding container 46e329b2292a23b4e7db5c2587420d8b2ff9fb0168254af0462a3f9acca4b553: Status 404 returned error can't find the container with id 46e329b2292a23b4e7db5c2587420d8b2ff9fb0168254af0462a3f9acca4b553 Mar 13 14:26:03 crc kubenswrapper[4907]: I0313 14:26:03.516620 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7b9fd7d84c-wkdxf" Mar 13 14:26:03 crc kubenswrapper[4907]: I0313 14:26:03.583310 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-54b5dffb47-wttzd"] Mar 13 14:26:03 crc kubenswrapper[4907]: I0313 14:26:03.583551 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-54b5dffb47-wttzd" podUID="779b1958-f7ed-4c53-a4ff-a6c4a803bebf" containerName="dnsmasq-dns" containerID="cri-o://923c944515f9f62d3a3c73bba3a0205c7dccd6753e7da2435a2be1308a3d0b7c" gracePeriod=10 Mar 13 14:26:03 crc kubenswrapper[4907]: I0313 14:26:03.621248 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-4v4ps" event={"ID":"02262347-eaec-462e-a3ef-58c670885d99","Type":"ContainerStarted","Data":"e0469ec6298d86f707976b7020c89f6d29a4be70c9a77f315ceb22e64cbb4824"} Mar 13 14:26:03 crc kubenswrapper[4907]: I0313 14:26:03.623040 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-cnnxs" event={"ID":"3f2b7163-b5d4-4575-a72c-e6ad1b3137df","Type":"ContainerStarted","Data":"d5c1cbc11101f46dd814717af56c35f8ffa0c1cd1640b4a7f3a2d4236359a48e"} Mar 13 14:26:03 crc kubenswrapper[4907]: I0313 14:26:03.623134 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-cnnxs" event={"ID":"3f2b7163-b5d4-4575-a72c-e6ad1b3137df","Type":"ContainerStarted","Data":"03e0154228fb7cb7a65f349825e100c422b39b776d8d1f870768bf174633ebf0"} Mar 13 14:26:03 crc kubenswrapper[4907]: I0313 14:26:03.627143 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-9fxfh" event={"ID":"e2585497-6837-4d17-8d51-c3d7879fdb46","Type":"ContainerStarted","Data":"ae6341ae2f61b11a2ad0c4f86bf51fc609fce628a9af3e20ce77aa5f1d917a69"} Mar 13 14:26:03 crc kubenswrapper[4907]: I0313 14:26:03.632557 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-hg688" event={"ID":"9c985ff9-3545-4544-acc8-ca9eb2bfad6a","Type":"ContainerStarted","Data":"93ab187126d08591f8a6823b5ca150c5522a7db6aea7e1ca8cc868ee0d426800"} Mar 13 14:26:03 crc kubenswrapper[4907]: I0313 14:26:03.632596 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-hg688" event={"ID":"9c985ff9-3545-4544-acc8-ca9eb2bfad6a","Type":"ContainerStarted","Data":"485fdae7de5b84d520754700103a7b2d1a2699988fd5bac3363e984c2ccb2b0e"} Mar 13 14:26:03 crc kubenswrapper[4907]: I0313 14:26:03.634856 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-9c64-account-create-update-v65tv" event={"ID":"3db3f0b6-9cae-40fa-b54d-3ed06c568c9c","Type":"ContainerStarted","Data":"1a454f90b2465a04739b60faa1ea54a4c62cae0b7a6742d83f6b477594d8a8a9"} Mar 13 14:26:03 crc kubenswrapper[4907]: I0313 14:26:03.634927 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-9c64-account-create-update-v65tv" event={"ID":"3db3f0b6-9cae-40fa-b54d-3ed06c568c9c","Type":"ContainerStarted","Data":"46e329b2292a23b4e7db5c2587420d8b2ff9fb0168254af0462a3f9acca4b553"} Mar 13 14:26:03 crc kubenswrapper[4907]: I0313 14:26:03.635967 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556866-fwg48" event={"ID":"7d930b98-c1af-4e47-a663-86afa9484856","Type":"ContainerStarted","Data":"c521a07f9fcca8ad2de643a0b0fa242b546a5a944f92129f2cbcb3aea70c71e9"} Mar 13 14:26:03 crc kubenswrapper[4907]: I0313 14:26:03.639568 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-5lhdv" Mar 13 14:26:03 crc kubenswrapper[4907]: I0313 14:26:03.643466 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-04a8-account-create-update-ws5ht" event={"ID":"37ba2d41-5de4-48cb-aad6-59df13307ed1","Type":"ContainerStarted","Data":"3d70e191441a97faf79353905c86d9af8e9fd167c63431a9f984ad9d2a36445e"} Mar 13 14:26:03 crc kubenswrapper[4907]: I0313 14:26:03.643514 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-04a8-account-create-update-ws5ht" event={"ID":"37ba2d41-5de4-48cb-aad6-59df13307ed1","Type":"ContainerStarted","Data":"85ee0b6a029ba5152ff5abc51ef7d2c859d0da76a5bac6cb02c6728c11a62a9f"} Mar 13 14:26:03 crc kubenswrapper[4907]: I0313 14:26:03.663231 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-create-cnnxs" podStartSLOduration=1.66321323 podStartE2EDuration="1.66321323s" podCreationTimestamp="2026-03-13 14:26:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:26:03.643929541 +0000 UTC m=+1262.543717230" watchObservedRunningTime="2026-03-13 14:26:03.66321323 +0000 UTC m=+1262.563000919" Mar 13 14:26:03 crc kubenswrapper[4907]: I0313 14:26:03.675590 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-9fxfh" podStartSLOduration=2.003299873 podStartE2EDuration="5.67557094s" podCreationTimestamp="2026-03-13 14:25:58 +0000 UTC" firstStartedPulling="2026-03-13 14:25:59.040315998 +0000 UTC m=+1257.940103687" lastFinishedPulling="2026-03-13 14:26:02.712587065 +0000 UTC m=+1261.612374754" observedRunningTime="2026-03-13 14:26:03.668424994 +0000 UTC m=+1262.568212683" watchObservedRunningTime="2026-03-13 14:26:03.67557094 +0000 UTC m=+1262.575358629" Mar 13 14:26:03 crc kubenswrapper[4907]: I0313 14:26:03.686898 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-9c64-account-create-update-v65tv" podStartSLOduration=1.68687187 podStartE2EDuration="1.68687187s" podCreationTimestamp="2026-03-13 14:26:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:26:03.684409472 +0000 UTC m=+1262.584197161" watchObservedRunningTime="2026-03-13 14:26:03.68687187 +0000 UTC m=+1262.586659559" Mar 13 14:26:03 crc kubenswrapper[4907]: I0313 14:26:03.710453 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-04a8-account-create-update-ws5ht" podStartSLOduration=1.710435197 podStartE2EDuration="1.710435197s" podCreationTimestamp="2026-03-13 14:26:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:26:03.701661006 +0000 UTC m=+1262.601448695" watchObservedRunningTime="2026-03-13 14:26:03.710435197 +0000 UTC m=+1262.610222886" Mar 13 14:26:03 crc kubenswrapper[4907]: I0313 14:26:03.727851 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-create-hg688" podStartSLOduration=1.7278339649999999 podStartE2EDuration="1.727833965s" podCreationTimestamp="2026-03-13 14:26:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:26:03.724407521 +0000 UTC m=+1262.624195210" watchObservedRunningTime="2026-03-13 14:26:03.727833965 +0000 UTC m=+1262.627621654" Mar 13 14:26:04 crc kubenswrapper[4907]: I0313 14:26:04.072596 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54b5dffb47-wttzd" Mar 13 14:26:04 crc kubenswrapper[4907]: I0313 14:26:04.172190 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/779b1958-f7ed-4c53-a4ff-a6c4a803bebf-dns-svc\") pod \"779b1958-f7ed-4c53-a4ff-a6c4a803bebf\" (UID: \"779b1958-f7ed-4c53-a4ff-a6c4a803bebf\") " Mar 13 14:26:04 crc kubenswrapper[4907]: I0313 14:26:04.172247 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/779b1958-f7ed-4c53-a4ff-a6c4a803bebf-config\") pod \"779b1958-f7ed-4c53-a4ff-a6c4a803bebf\" (UID: \"779b1958-f7ed-4c53-a4ff-a6c4a803bebf\") " Mar 13 14:26:04 crc kubenswrapper[4907]: I0313 14:26:04.172274 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bmczb\" (UniqueName: \"kubernetes.io/projected/779b1958-f7ed-4c53-a4ff-a6c4a803bebf-kube-api-access-bmczb\") pod \"779b1958-f7ed-4c53-a4ff-a6c4a803bebf\" (UID: \"779b1958-f7ed-4c53-a4ff-a6c4a803bebf\") " Mar 13 14:26:04 crc kubenswrapper[4907]: I0313 14:26:04.177737 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/779b1958-f7ed-4c53-a4ff-a6c4a803bebf-kube-api-access-bmczb" (OuterVolumeSpecName: "kube-api-access-bmczb") pod "779b1958-f7ed-4c53-a4ff-a6c4a803bebf" (UID: "779b1958-f7ed-4c53-a4ff-a6c4a803bebf"). InnerVolumeSpecName "kube-api-access-bmczb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:26:04 crc kubenswrapper[4907]: I0313 14:26:04.230438 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/779b1958-f7ed-4c53-a4ff-a6c4a803bebf-config" (OuterVolumeSpecName: "config") pod "779b1958-f7ed-4c53-a4ff-a6c4a803bebf" (UID: "779b1958-f7ed-4c53-a4ff-a6c4a803bebf"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:26:04 crc kubenswrapper[4907]: I0313 14:26:04.246378 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/779b1958-f7ed-4c53-a4ff-a6c4a803bebf-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "779b1958-f7ed-4c53-a4ff-a6c4a803bebf" (UID: "779b1958-f7ed-4c53-a4ff-a6c4a803bebf"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:26:04 crc kubenswrapper[4907]: I0313 14:26:04.274582 4907 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/779b1958-f7ed-4c53-a4ff-a6c4a803bebf-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:04 crc kubenswrapper[4907]: I0313 14:26:04.274621 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/779b1958-f7ed-4c53-a4ff-a6c4a803bebf-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:04 crc kubenswrapper[4907]: I0313 14:26:04.274634 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bmczb\" (UniqueName: \"kubernetes.io/projected/779b1958-f7ed-4c53-a4ff-a6c4a803bebf-kube-api-access-bmczb\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:04 crc kubenswrapper[4907]: I0313 14:26:04.649923 4907 generic.go:334] "Generic (PLEG): container finished" podID="3db3f0b6-9cae-40fa-b54d-3ed06c568c9c" containerID="1a454f90b2465a04739b60faa1ea54a4c62cae0b7a6742d83f6b477594d8a8a9" exitCode=0 Mar 13 14:26:04 crc kubenswrapper[4907]: I0313 14:26:04.649973 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-9c64-account-create-update-v65tv" event={"ID":"3db3f0b6-9cae-40fa-b54d-3ed06c568c9c","Type":"ContainerDied","Data":"1a454f90b2465a04739b60faa1ea54a4c62cae0b7a6742d83f6b477594d8a8a9"} Mar 13 14:26:04 crc kubenswrapper[4907]: I0313 14:26:04.653177 4907 generic.go:334] "Generic (PLEG): container finished" podID="9c985ff9-3545-4544-acc8-ca9eb2bfad6a" containerID="93ab187126d08591f8a6823b5ca150c5522a7db6aea7e1ca8cc868ee0d426800" exitCode=0 Mar 13 14:26:04 crc kubenswrapper[4907]: I0313 14:26:04.653260 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-hg688" event={"ID":"9c985ff9-3545-4544-acc8-ca9eb2bfad6a","Type":"ContainerDied","Data":"93ab187126d08591f8a6823b5ca150c5522a7db6aea7e1ca8cc868ee0d426800"} Mar 13 14:26:04 crc kubenswrapper[4907]: I0313 14:26:04.654751 4907 generic.go:334] "Generic (PLEG): container finished" podID="37ba2d41-5de4-48cb-aad6-59df13307ed1" containerID="3d70e191441a97faf79353905c86d9af8e9fd167c63431a9f984ad9d2a36445e" exitCode=0 Mar 13 14:26:04 crc kubenswrapper[4907]: I0313 14:26:04.654795 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-04a8-account-create-update-ws5ht" event={"ID":"37ba2d41-5de4-48cb-aad6-59df13307ed1","Type":"ContainerDied","Data":"3d70e191441a97faf79353905c86d9af8e9fd167c63431a9f984ad9d2a36445e"} Mar 13 14:26:04 crc kubenswrapper[4907]: I0313 14:26:04.656685 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556866-fwg48" event={"ID":"7d930b98-c1af-4e47-a663-86afa9484856","Type":"ContainerStarted","Data":"37d0b4abf880f9988e205a5d09debf0e634cb45e2049ee542b43a256eccbec25"} Mar 13 14:26:04 crc kubenswrapper[4907]: I0313 14:26:04.663444 4907 generic.go:334] "Generic (PLEG): container finished" podID="3f2b7163-b5d4-4575-a72c-e6ad1b3137df" containerID="d5c1cbc11101f46dd814717af56c35f8ffa0c1cd1640b4a7f3a2d4236359a48e" exitCode=0 Mar 13 14:26:04 crc kubenswrapper[4907]: I0313 14:26:04.663660 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-cnnxs" event={"ID":"3f2b7163-b5d4-4575-a72c-e6ad1b3137df","Type":"ContainerDied","Data":"d5c1cbc11101f46dd814717af56c35f8ffa0c1cd1640b4a7f3a2d4236359a48e"} Mar 13 14:26:04 crc kubenswrapper[4907]: I0313 14:26:04.667919 4907 generic.go:334] "Generic (PLEG): container finished" podID="779b1958-f7ed-4c53-a4ff-a6c4a803bebf" containerID="923c944515f9f62d3a3c73bba3a0205c7dccd6753e7da2435a2be1308a3d0b7c" exitCode=0 Mar 13 14:26:04 crc kubenswrapper[4907]: I0313 14:26:04.667983 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54b5dffb47-wttzd" Mar 13 14:26:04 crc kubenswrapper[4907]: I0313 14:26:04.668006 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54b5dffb47-wttzd" event={"ID":"779b1958-f7ed-4c53-a4ff-a6c4a803bebf","Type":"ContainerDied","Data":"923c944515f9f62d3a3c73bba3a0205c7dccd6753e7da2435a2be1308a3d0b7c"} Mar 13 14:26:04 crc kubenswrapper[4907]: I0313 14:26:04.668044 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54b5dffb47-wttzd" event={"ID":"779b1958-f7ed-4c53-a4ff-a6c4a803bebf","Type":"ContainerDied","Data":"54b19297eaef464c5a401c18dcda022f8c38c1cecb9978d5b544939e48c81aa1"} Mar 13 14:26:04 crc kubenswrapper[4907]: I0313 14:26:04.668060 4907 scope.go:117] "RemoveContainer" containerID="923c944515f9f62d3a3c73bba3a0205c7dccd6753e7da2435a2be1308a3d0b7c" Mar 13 14:26:04 crc kubenswrapper[4907]: I0313 14:26:04.692368 4907 scope.go:117] "RemoveContainer" containerID="cd0379440a1db9f081229f5637156055839f155515dc1039d28132ecd64ab32a" Mar 13 14:26:04 crc kubenswrapper[4907]: I0313 14:26:04.729679 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29556866-fwg48" podStartSLOduration=3.871471579 podStartE2EDuration="4.729612175s" podCreationTimestamp="2026-03-13 14:26:00 +0000 UTC" firstStartedPulling="2026-03-13 14:26:03.108751744 +0000 UTC m=+1262.008539433" lastFinishedPulling="2026-03-13 14:26:03.96689234 +0000 UTC m=+1262.866680029" observedRunningTime="2026-03-13 14:26:04.716073774 +0000 UTC m=+1263.615861463" watchObservedRunningTime="2026-03-13 14:26:04.729612175 +0000 UTC m=+1263.629399864" Mar 13 14:26:04 crc kubenswrapper[4907]: I0313 14:26:04.742924 4907 scope.go:117] "RemoveContainer" containerID="923c944515f9f62d3a3c73bba3a0205c7dccd6753e7da2435a2be1308a3d0b7c" Mar 13 14:26:04 crc kubenswrapper[4907]: E0313 14:26:04.746161 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"923c944515f9f62d3a3c73bba3a0205c7dccd6753e7da2435a2be1308a3d0b7c\": container with ID starting with 923c944515f9f62d3a3c73bba3a0205c7dccd6753e7da2435a2be1308a3d0b7c not found: ID does not exist" containerID="923c944515f9f62d3a3c73bba3a0205c7dccd6753e7da2435a2be1308a3d0b7c" Mar 13 14:26:04 crc kubenswrapper[4907]: I0313 14:26:04.746403 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"923c944515f9f62d3a3c73bba3a0205c7dccd6753e7da2435a2be1308a3d0b7c"} err="failed to get container status \"923c944515f9f62d3a3c73bba3a0205c7dccd6753e7da2435a2be1308a3d0b7c\": rpc error: code = NotFound desc = could not find container \"923c944515f9f62d3a3c73bba3a0205c7dccd6753e7da2435a2be1308a3d0b7c\": container with ID starting with 923c944515f9f62d3a3c73bba3a0205c7dccd6753e7da2435a2be1308a3d0b7c not found: ID does not exist" Mar 13 14:26:04 crc kubenswrapper[4907]: I0313 14:26:04.746507 4907 scope.go:117] "RemoveContainer" containerID="cd0379440a1db9f081229f5637156055839f155515dc1039d28132ecd64ab32a" Mar 13 14:26:04 crc kubenswrapper[4907]: E0313 14:26:04.747295 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cd0379440a1db9f081229f5637156055839f155515dc1039d28132ecd64ab32a\": container with ID starting with cd0379440a1db9f081229f5637156055839f155515dc1039d28132ecd64ab32a not found: ID does not exist" containerID="cd0379440a1db9f081229f5637156055839f155515dc1039d28132ecd64ab32a" Mar 13 14:26:04 crc kubenswrapper[4907]: I0313 14:26:04.747341 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cd0379440a1db9f081229f5637156055839f155515dc1039d28132ecd64ab32a"} err="failed to get container status \"cd0379440a1db9f081229f5637156055839f155515dc1039d28132ecd64ab32a\": rpc error: code = NotFound desc = could not find container \"cd0379440a1db9f081229f5637156055839f155515dc1039d28132ecd64ab32a\": container with ID starting with cd0379440a1db9f081229f5637156055839f155515dc1039d28132ecd64ab32a not found: ID does not exist" Mar 13 14:26:04 crc kubenswrapper[4907]: I0313 14:26:04.763678 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-54b5dffb47-wttzd"] Mar 13 14:26:04 crc kubenswrapper[4907]: I0313 14:26:04.770550 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-54b5dffb47-wttzd"] Mar 13 14:26:04 crc kubenswrapper[4907]: I0313 14:26:04.815999 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/root-account-create-update-5lhdv"] Mar 13 14:26:04 crc kubenswrapper[4907]: I0313 14:26:04.823622 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/root-account-create-update-5lhdv"] Mar 13 14:26:05 crc kubenswrapper[4907]: I0313 14:26:05.679754 4907 generic.go:334] "Generic (PLEG): container finished" podID="7d930b98-c1af-4e47-a663-86afa9484856" containerID="37d0b4abf880f9988e205a5d09debf0e634cb45e2049ee542b43a256eccbec25" exitCode=0 Mar 13 14:26:05 crc kubenswrapper[4907]: I0313 14:26:05.679852 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556866-fwg48" event={"ID":"7d930b98-c1af-4e47-a663-86afa9484856","Type":"ContainerDied","Data":"37d0b4abf880f9988e205a5d09debf0e634cb45e2049ee542b43a256eccbec25"} Mar 13 14:26:05 crc kubenswrapper[4907]: I0313 14:26:05.791356 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="779b1958-f7ed-4c53-a4ff-a6c4a803bebf" path="/var/lib/kubelet/pods/779b1958-f7ed-4c53-a4ff-a6c4a803bebf/volumes" Mar 13 14:26:05 crc kubenswrapper[4907]: I0313 14:26:05.792397 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f12f5638-4cb2-4424-86d8-5c0c829d82d5" path="/var/lib/kubelet/pods/f12f5638-4cb2-4424-86d8-5c0c829d82d5/volumes" Mar 13 14:26:06 crc kubenswrapper[4907]: I0313 14:26:06.032119 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-9c64-account-create-update-v65tv" Mar 13 14:26:06 crc kubenswrapper[4907]: I0313 14:26:06.106282 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3db3f0b6-9cae-40fa-b54d-3ed06c568c9c-operator-scripts\") pod \"3db3f0b6-9cae-40fa-b54d-3ed06c568c9c\" (UID: \"3db3f0b6-9cae-40fa-b54d-3ed06c568c9c\") " Mar 13 14:26:06 crc kubenswrapper[4907]: I0313 14:26:06.106789 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h5wpg\" (UniqueName: \"kubernetes.io/projected/3db3f0b6-9cae-40fa-b54d-3ed06c568c9c-kube-api-access-h5wpg\") pod \"3db3f0b6-9cae-40fa-b54d-3ed06c568c9c\" (UID: \"3db3f0b6-9cae-40fa-b54d-3ed06c568c9c\") " Mar 13 14:26:06 crc kubenswrapper[4907]: I0313 14:26:06.107389 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3db3f0b6-9cae-40fa-b54d-3ed06c568c9c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "3db3f0b6-9cae-40fa-b54d-3ed06c568c9c" (UID: "3db3f0b6-9cae-40fa-b54d-3ed06c568c9c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:26:06 crc kubenswrapper[4907]: I0313 14:26:06.114118 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3db3f0b6-9cae-40fa-b54d-3ed06c568c9c-kube-api-access-h5wpg" (OuterVolumeSpecName: "kube-api-access-h5wpg") pod "3db3f0b6-9cae-40fa-b54d-3ed06c568c9c" (UID: "3db3f0b6-9cae-40fa-b54d-3ed06c568c9c"). InnerVolumeSpecName "kube-api-access-h5wpg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:26:06 crc kubenswrapper[4907]: I0313 14:26:06.191961 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-cnnxs" Mar 13 14:26:06 crc kubenswrapper[4907]: I0313 14:26:06.197468 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-04a8-account-create-update-ws5ht" Mar 13 14:26:06 crc kubenswrapper[4907]: I0313 14:26:06.209345 4907 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3db3f0b6-9cae-40fa-b54d-3ed06c568c9c-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:06 crc kubenswrapper[4907]: I0313 14:26:06.209372 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h5wpg\" (UniqueName: \"kubernetes.io/projected/3db3f0b6-9cae-40fa-b54d-3ed06c568c9c-kube-api-access-h5wpg\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:06 crc kubenswrapper[4907]: I0313 14:26:06.209827 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-hg688" Mar 13 14:26:06 crc kubenswrapper[4907]: I0313 14:26:06.310277 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3f2b7163-b5d4-4575-a72c-e6ad1b3137df-operator-scripts\") pod \"3f2b7163-b5d4-4575-a72c-e6ad1b3137df\" (UID: \"3f2b7163-b5d4-4575-a72c-e6ad1b3137df\") " Mar 13 14:26:06 crc kubenswrapper[4907]: I0313 14:26:06.310343 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9c985ff9-3545-4544-acc8-ca9eb2bfad6a-operator-scripts\") pod \"9c985ff9-3545-4544-acc8-ca9eb2bfad6a\" (UID: \"9c985ff9-3545-4544-acc8-ca9eb2bfad6a\") " Mar 13 14:26:06 crc kubenswrapper[4907]: I0313 14:26:06.310399 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5cr4l\" (UniqueName: \"kubernetes.io/projected/3f2b7163-b5d4-4575-a72c-e6ad1b3137df-kube-api-access-5cr4l\") pod \"3f2b7163-b5d4-4575-a72c-e6ad1b3137df\" (UID: \"3f2b7163-b5d4-4575-a72c-e6ad1b3137df\") " Mar 13 14:26:06 crc kubenswrapper[4907]: I0313 14:26:06.310508 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b4bzw\" (UniqueName: \"kubernetes.io/projected/37ba2d41-5de4-48cb-aad6-59df13307ed1-kube-api-access-b4bzw\") pod \"37ba2d41-5de4-48cb-aad6-59df13307ed1\" (UID: \"37ba2d41-5de4-48cb-aad6-59df13307ed1\") " Mar 13 14:26:06 crc kubenswrapper[4907]: I0313 14:26:06.310535 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qcmkf\" (UniqueName: \"kubernetes.io/projected/9c985ff9-3545-4544-acc8-ca9eb2bfad6a-kube-api-access-qcmkf\") pod \"9c985ff9-3545-4544-acc8-ca9eb2bfad6a\" (UID: \"9c985ff9-3545-4544-acc8-ca9eb2bfad6a\") " Mar 13 14:26:06 crc kubenswrapper[4907]: I0313 14:26:06.310569 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/37ba2d41-5de4-48cb-aad6-59df13307ed1-operator-scripts\") pod \"37ba2d41-5de4-48cb-aad6-59df13307ed1\" (UID: \"37ba2d41-5de4-48cb-aad6-59df13307ed1\") " Mar 13 14:26:06 crc kubenswrapper[4907]: I0313 14:26:06.311178 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9c985ff9-3545-4544-acc8-ca9eb2bfad6a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "9c985ff9-3545-4544-acc8-ca9eb2bfad6a" (UID: "9c985ff9-3545-4544-acc8-ca9eb2bfad6a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:26:06 crc kubenswrapper[4907]: I0313 14:26:06.311857 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3f2b7163-b5d4-4575-a72c-e6ad1b3137df-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "3f2b7163-b5d4-4575-a72c-e6ad1b3137df" (UID: "3f2b7163-b5d4-4575-a72c-e6ad1b3137df"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:26:06 crc kubenswrapper[4907]: I0313 14:26:06.312099 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/37ba2d41-5de4-48cb-aad6-59df13307ed1-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "37ba2d41-5de4-48cb-aad6-59df13307ed1" (UID: "37ba2d41-5de4-48cb-aad6-59df13307ed1"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:26:06 crc kubenswrapper[4907]: I0313 14:26:06.314113 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3f2b7163-b5d4-4575-a72c-e6ad1b3137df-kube-api-access-5cr4l" (OuterVolumeSpecName: "kube-api-access-5cr4l") pod "3f2b7163-b5d4-4575-a72c-e6ad1b3137df" (UID: "3f2b7163-b5d4-4575-a72c-e6ad1b3137df"). InnerVolumeSpecName "kube-api-access-5cr4l". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:26:06 crc kubenswrapper[4907]: I0313 14:26:06.315134 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c985ff9-3545-4544-acc8-ca9eb2bfad6a-kube-api-access-qcmkf" (OuterVolumeSpecName: "kube-api-access-qcmkf") pod "9c985ff9-3545-4544-acc8-ca9eb2bfad6a" (UID: "9c985ff9-3545-4544-acc8-ca9eb2bfad6a"). InnerVolumeSpecName "kube-api-access-qcmkf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:26:06 crc kubenswrapper[4907]: I0313 14:26:06.315598 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/37ba2d41-5de4-48cb-aad6-59df13307ed1-kube-api-access-b4bzw" (OuterVolumeSpecName: "kube-api-access-b4bzw") pod "37ba2d41-5de4-48cb-aad6-59df13307ed1" (UID: "37ba2d41-5de4-48cb-aad6-59df13307ed1"). InnerVolumeSpecName "kube-api-access-b4bzw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:26:06 crc kubenswrapper[4907]: I0313 14:26:06.412280 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b4bzw\" (UniqueName: \"kubernetes.io/projected/37ba2d41-5de4-48cb-aad6-59df13307ed1-kube-api-access-b4bzw\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:06 crc kubenswrapper[4907]: I0313 14:26:06.412318 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qcmkf\" (UniqueName: \"kubernetes.io/projected/9c985ff9-3545-4544-acc8-ca9eb2bfad6a-kube-api-access-qcmkf\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:06 crc kubenswrapper[4907]: I0313 14:26:06.412327 4907 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/37ba2d41-5de4-48cb-aad6-59df13307ed1-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:06 crc kubenswrapper[4907]: I0313 14:26:06.412336 4907 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3f2b7163-b5d4-4575-a72c-e6ad1b3137df-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:06 crc kubenswrapper[4907]: I0313 14:26:06.412345 4907 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9c985ff9-3545-4544-acc8-ca9eb2bfad6a-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:06 crc kubenswrapper[4907]: I0313 14:26:06.412355 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5cr4l\" (UniqueName: \"kubernetes.io/projected/3f2b7163-b5d4-4575-a72c-e6ad1b3137df-kube-api-access-5cr4l\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:06 crc kubenswrapper[4907]: I0313 14:26:06.688901 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-hg688" Mar 13 14:26:06 crc kubenswrapper[4907]: I0313 14:26:06.688904 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-hg688" event={"ID":"9c985ff9-3545-4544-acc8-ca9eb2bfad6a","Type":"ContainerDied","Data":"485fdae7de5b84d520754700103a7b2d1a2699988fd5bac3363e984c2ccb2b0e"} Mar 13 14:26:06 crc kubenswrapper[4907]: I0313 14:26:06.688952 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="485fdae7de5b84d520754700103a7b2d1a2699988fd5bac3363e984c2ccb2b0e" Mar 13 14:26:06 crc kubenswrapper[4907]: I0313 14:26:06.690835 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-04a8-account-create-update-ws5ht" Mar 13 14:26:06 crc kubenswrapper[4907]: I0313 14:26:06.690814 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-04a8-account-create-update-ws5ht" event={"ID":"37ba2d41-5de4-48cb-aad6-59df13307ed1","Type":"ContainerDied","Data":"85ee0b6a029ba5152ff5abc51ef7d2c859d0da76a5bac6cb02c6728c11a62a9f"} Mar 13 14:26:06 crc kubenswrapper[4907]: I0313 14:26:06.690913 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="85ee0b6a029ba5152ff5abc51ef7d2c859d0da76a5bac6cb02c6728c11a62a9f" Mar 13 14:26:06 crc kubenswrapper[4907]: I0313 14:26:06.693142 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-cnnxs" event={"ID":"3f2b7163-b5d4-4575-a72c-e6ad1b3137df","Type":"ContainerDied","Data":"03e0154228fb7cb7a65f349825e100c422b39b776d8d1f870768bf174633ebf0"} Mar 13 14:26:06 crc kubenswrapper[4907]: I0313 14:26:06.693755 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="03e0154228fb7cb7a65f349825e100c422b39b776d8d1f870768bf174633ebf0" Mar 13 14:26:06 crc kubenswrapper[4907]: I0313 14:26:06.693159 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-cnnxs" Mar 13 14:26:06 crc kubenswrapper[4907]: I0313 14:26:06.694515 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-9c64-account-create-update-v65tv" Mar 13 14:26:06 crc kubenswrapper[4907]: I0313 14:26:06.694536 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-9c64-account-create-update-v65tv" event={"ID":"3db3f0b6-9cae-40fa-b54d-3ed06c568c9c","Type":"ContainerDied","Data":"46e329b2292a23b4e7db5c2587420d8b2ff9fb0168254af0462a3f9acca4b553"} Mar 13 14:26:06 crc kubenswrapper[4907]: I0313 14:26:06.694568 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="46e329b2292a23b4e7db5c2587420d8b2ff9fb0168254af0462a3f9acca4b553" Mar 13 14:26:07 crc kubenswrapper[4907]: I0313 14:26:07.024926 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556866-fwg48" Mar 13 14:26:07 crc kubenswrapper[4907]: I0313 14:26:07.121323 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f5d44\" (UniqueName: \"kubernetes.io/projected/7d930b98-c1af-4e47-a663-86afa9484856-kube-api-access-f5d44\") pod \"7d930b98-c1af-4e47-a663-86afa9484856\" (UID: \"7d930b98-c1af-4e47-a663-86afa9484856\") " Mar 13 14:26:07 crc kubenswrapper[4907]: I0313 14:26:07.138387 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7d930b98-c1af-4e47-a663-86afa9484856-kube-api-access-f5d44" (OuterVolumeSpecName: "kube-api-access-f5d44") pod "7d930b98-c1af-4e47-a663-86afa9484856" (UID: "7d930b98-c1af-4e47-a663-86afa9484856"). InnerVolumeSpecName "kube-api-access-f5d44". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:26:07 crc kubenswrapper[4907]: I0313 14:26:07.225098 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f5d44\" (UniqueName: \"kubernetes.io/projected/7d930b98-c1af-4e47-a663-86afa9484856-kube-api-access-f5d44\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:07 crc kubenswrapper[4907]: I0313 14:26:07.724017 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556866-fwg48" event={"ID":"7d930b98-c1af-4e47-a663-86afa9484856","Type":"ContainerDied","Data":"c521a07f9fcca8ad2de643a0b0fa242b546a5a944f92129f2cbcb3aea70c71e9"} Mar 13 14:26:07 crc kubenswrapper[4907]: I0313 14:26:07.724284 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c521a07f9fcca8ad2de643a0b0fa242b546a5a944f92129f2cbcb3aea70c71e9" Mar 13 14:26:07 crc kubenswrapper[4907]: I0313 14:26:07.724100 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556866-fwg48" Mar 13 14:26:08 crc kubenswrapper[4907]: I0313 14:26:08.081710 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556860-ljdj7"] Mar 13 14:26:08 crc kubenswrapper[4907]: I0313 14:26:08.088377 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556860-ljdj7"] Mar 13 14:26:08 crc kubenswrapper[4907]: I0313 14:26:08.447178 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/root-account-create-update-r2rlh"] Mar 13 14:26:08 crc kubenswrapper[4907]: E0313 14:26:08.447511 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="779b1958-f7ed-4c53-a4ff-a6c4a803bebf" containerName="dnsmasq-dns" Mar 13 14:26:08 crc kubenswrapper[4907]: I0313 14:26:08.447525 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="779b1958-f7ed-4c53-a4ff-a6c4a803bebf" containerName="dnsmasq-dns" Mar 13 14:26:08 crc kubenswrapper[4907]: E0313 14:26:08.447539 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="779b1958-f7ed-4c53-a4ff-a6c4a803bebf" containerName="init" Mar 13 14:26:08 crc kubenswrapper[4907]: I0313 14:26:08.447547 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="779b1958-f7ed-4c53-a4ff-a6c4a803bebf" containerName="init" Mar 13 14:26:08 crc kubenswrapper[4907]: E0313 14:26:08.447563 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37ba2d41-5de4-48cb-aad6-59df13307ed1" containerName="mariadb-account-create-update" Mar 13 14:26:08 crc kubenswrapper[4907]: I0313 14:26:08.447571 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="37ba2d41-5de4-48cb-aad6-59df13307ed1" containerName="mariadb-account-create-update" Mar 13 14:26:08 crc kubenswrapper[4907]: E0313 14:26:08.447610 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f12f5638-4cb2-4424-86d8-5c0c829d82d5" containerName="mariadb-account-create-update" Mar 13 14:26:08 crc kubenswrapper[4907]: I0313 14:26:08.447622 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="f12f5638-4cb2-4424-86d8-5c0c829d82d5" containerName="mariadb-account-create-update" Mar 13 14:26:08 crc kubenswrapper[4907]: E0313 14:26:08.447635 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c985ff9-3545-4544-acc8-ca9eb2bfad6a" containerName="mariadb-database-create" Mar 13 14:26:08 crc kubenswrapper[4907]: I0313 14:26:08.447642 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c985ff9-3545-4544-acc8-ca9eb2bfad6a" containerName="mariadb-database-create" Mar 13 14:26:08 crc kubenswrapper[4907]: E0313 14:26:08.447663 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f2b7163-b5d4-4575-a72c-e6ad1b3137df" containerName="mariadb-database-create" Mar 13 14:26:08 crc kubenswrapper[4907]: I0313 14:26:08.447670 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f2b7163-b5d4-4575-a72c-e6ad1b3137df" containerName="mariadb-database-create" Mar 13 14:26:08 crc kubenswrapper[4907]: E0313 14:26:08.447682 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3db3f0b6-9cae-40fa-b54d-3ed06c568c9c" containerName="mariadb-account-create-update" Mar 13 14:26:08 crc kubenswrapper[4907]: I0313 14:26:08.447689 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="3db3f0b6-9cae-40fa-b54d-3ed06c568c9c" containerName="mariadb-account-create-update" Mar 13 14:26:08 crc kubenswrapper[4907]: E0313 14:26:08.447703 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d930b98-c1af-4e47-a663-86afa9484856" containerName="oc" Mar 13 14:26:08 crc kubenswrapper[4907]: I0313 14:26:08.447710 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d930b98-c1af-4e47-a663-86afa9484856" containerName="oc" Mar 13 14:26:08 crc kubenswrapper[4907]: I0313 14:26:08.447932 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="37ba2d41-5de4-48cb-aad6-59df13307ed1" containerName="mariadb-account-create-update" Mar 13 14:26:08 crc kubenswrapper[4907]: I0313 14:26:08.447951 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="779b1958-f7ed-4c53-a4ff-a6c4a803bebf" containerName="dnsmasq-dns" Mar 13 14:26:08 crc kubenswrapper[4907]: I0313 14:26:08.447963 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="f12f5638-4cb2-4424-86d8-5c0c829d82d5" containerName="mariadb-account-create-update" Mar 13 14:26:08 crc kubenswrapper[4907]: I0313 14:26:08.448006 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="3db3f0b6-9cae-40fa-b54d-3ed06c568c9c" containerName="mariadb-account-create-update" Mar 13 14:26:08 crc kubenswrapper[4907]: I0313 14:26:08.448018 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="3f2b7163-b5d4-4575-a72c-e6ad1b3137df" containerName="mariadb-database-create" Mar 13 14:26:08 crc kubenswrapper[4907]: I0313 14:26:08.448033 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c985ff9-3545-4544-acc8-ca9eb2bfad6a" containerName="mariadb-database-create" Mar 13 14:26:08 crc kubenswrapper[4907]: I0313 14:26:08.448041 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d930b98-c1af-4e47-a663-86afa9484856" containerName="oc" Mar 13 14:26:08 crc kubenswrapper[4907]: I0313 14:26:08.448718 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-r2rlh" Mar 13 14:26:08 crc kubenswrapper[4907]: I0313 14:26:08.450797 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-mariadb-root-db-secret" Mar 13 14:26:08 crc kubenswrapper[4907]: I0313 14:26:08.456094 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-r2rlh"] Mar 13 14:26:08 crc kubenswrapper[4907]: I0313 14:26:08.649522 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pgg7x\" (UniqueName: \"kubernetes.io/projected/b3affff5-a720-4d5e-8f17-e8ec23eff0c9-kube-api-access-pgg7x\") pod \"root-account-create-update-r2rlh\" (UID: \"b3affff5-a720-4d5e-8f17-e8ec23eff0c9\") " pod="openstack/root-account-create-update-r2rlh" Mar 13 14:26:08 crc kubenswrapper[4907]: I0313 14:26:08.649599 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b3affff5-a720-4d5e-8f17-e8ec23eff0c9-operator-scripts\") pod \"root-account-create-update-r2rlh\" (UID: \"b3affff5-a720-4d5e-8f17-e8ec23eff0c9\") " pod="openstack/root-account-create-update-r2rlh" Mar 13 14:26:08 crc kubenswrapper[4907]: I0313 14:26:08.750862 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pgg7x\" (UniqueName: \"kubernetes.io/projected/b3affff5-a720-4d5e-8f17-e8ec23eff0c9-kube-api-access-pgg7x\") pod \"root-account-create-update-r2rlh\" (UID: \"b3affff5-a720-4d5e-8f17-e8ec23eff0c9\") " pod="openstack/root-account-create-update-r2rlh" Mar 13 14:26:08 crc kubenswrapper[4907]: I0313 14:26:08.751599 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b3affff5-a720-4d5e-8f17-e8ec23eff0c9-operator-scripts\") pod \"root-account-create-update-r2rlh\" (UID: \"b3affff5-a720-4d5e-8f17-e8ec23eff0c9\") " pod="openstack/root-account-create-update-r2rlh" Mar 13 14:26:08 crc kubenswrapper[4907]: I0313 14:26:08.752621 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b3affff5-a720-4d5e-8f17-e8ec23eff0c9-operator-scripts\") pod \"root-account-create-update-r2rlh\" (UID: \"b3affff5-a720-4d5e-8f17-e8ec23eff0c9\") " pod="openstack/root-account-create-update-r2rlh" Mar 13 14:26:08 crc kubenswrapper[4907]: I0313 14:26:08.767541 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pgg7x\" (UniqueName: \"kubernetes.io/projected/b3affff5-a720-4d5e-8f17-e8ec23eff0c9-kube-api-access-pgg7x\") pod \"root-account-create-update-r2rlh\" (UID: \"b3affff5-a720-4d5e-8f17-e8ec23eff0c9\") " pod="openstack/root-account-create-update-r2rlh" Mar 13 14:26:08 crc kubenswrapper[4907]: I0313 14:26:08.768250 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-r2rlh" Mar 13 14:26:09 crc kubenswrapper[4907]: I0313 14:26:09.344322 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Mar 13 14:26:09 crc kubenswrapper[4907]: I0313 14:26:09.804548 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="257ee294-461c-4344-9dcb-b44f1497f8bf" path="/var/lib/kubelet/pods/257ee294-461c-4344-9dcb-b44f1497f8bf/volumes" Mar 13 14:26:10 crc kubenswrapper[4907]: I0313 14:26:10.291710 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/14f05f31-8185-43b0-be69-bcf8d5388ea5-etc-swift\") pod \"swift-storage-0\" (UID: \"14f05f31-8185-43b0-be69-bcf8d5388ea5\") " pod="openstack/swift-storage-0" Mar 13 14:26:10 crc kubenswrapper[4907]: E0313 14:26:10.291903 4907 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Mar 13 14:26:10 crc kubenswrapper[4907]: E0313 14:26:10.292184 4907 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Mar 13 14:26:10 crc kubenswrapper[4907]: E0313 14:26:10.292246 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/14f05f31-8185-43b0-be69-bcf8d5388ea5-etc-swift podName:14f05f31-8185-43b0-be69-bcf8d5388ea5 nodeName:}" failed. No retries permitted until 2026-03-13 14:26:26.292227834 +0000 UTC m=+1285.192015523 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/14f05f31-8185-43b0-be69-bcf8d5388ea5-etc-swift") pod "swift-storage-0" (UID: "14f05f31-8185-43b0-be69-bcf8d5388ea5") : configmap "swift-ring-files" not found Mar 13 14:26:10 crc kubenswrapper[4907]: I0313 14:26:10.770806 4907 generic.go:334] "Generic (PLEG): container finished" podID="e2585497-6837-4d17-8d51-c3d7879fdb46" containerID="ae6341ae2f61b11a2ad0c4f86bf51fc609fce628a9af3e20ce77aa5f1d917a69" exitCode=0 Mar 13 14:26:10 crc kubenswrapper[4907]: I0313 14:26:10.770852 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-9fxfh" event={"ID":"e2585497-6837-4d17-8d51-c3d7879fdb46","Type":"ContainerDied","Data":"ae6341ae2f61b11a2ad0c4f86bf51fc609fce628a9af3e20ce77aa5f1d917a69"} Mar 13 14:26:12 crc kubenswrapper[4907]: I0313 14:26:12.472746 4907 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-bw2xb" podUID="6f8119b2-e38a-494d-967f-5198b83512c7" containerName="ovn-controller" probeResult="failure" output=< Mar 13 14:26:12 crc kubenswrapper[4907]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Mar 13 14:26:12 crc kubenswrapper[4907]: > Mar 13 14:26:12 crc kubenswrapper[4907]: I0313 14:26:12.481224 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-vp6f5" Mar 13 14:26:14 crc kubenswrapper[4907]: I0313 14:26:14.803361 4907 generic.go:334] "Generic (PLEG): container finished" podID="286c82e8-b74a-49d7-a355-ac074aace10e" containerID="275b76780d99e94e2c958194ce29198632d04fd0fe0f0668ff5025ad387fc2f7" exitCode=0 Mar 13 14:26:14 crc kubenswrapper[4907]: I0313 14:26:14.803470 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"286c82e8-b74a-49d7-a355-ac074aace10e","Type":"ContainerDied","Data":"275b76780d99e94e2c958194ce29198632d04fd0fe0f0668ff5025ad387fc2f7"} Mar 13 14:26:15 crc kubenswrapper[4907]: I0313 14:26:15.813676 4907 generic.go:334] "Generic (PLEG): container finished" podID="e3c34454-315e-4821-ab25-b0f331a0d521" containerID="987e5ce0ff0b4c3691f9df5efc31a14e40c988230f0cf14b4b06e00cd7129b38" exitCode=0 Mar 13 14:26:15 crc kubenswrapper[4907]: I0313 14:26:15.813777 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"e3c34454-315e-4821-ab25-b0f331a0d521","Type":"ContainerDied","Data":"987e5ce0ff0b4c3691f9df5efc31a14e40c988230f0cf14b4b06e00cd7129b38"} Mar 13 14:26:17 crc kubenswrapper[4907]: I0313 14:26:17.467903 4907 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-bw2xb" podUID="6f8119b2-e38a-494d-967f-5198b83512c7" containerName="ovn-controller" probeResult="failure" output=< Mar 13 14:26:17 crc kubenswrapper[4907]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Mar 13 14:26:17 crc kubenswrapper[4907]: > Mar 13 14:26:17 crc kubenswrapper[4907]: I0313 14:26:17.477975 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-vp6f5" Mar 13 14:26:17 crc kubenswrapper[4907]: I0313 14:26:17.682038 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-bw2xb-config-gdlfq"] Mar 13 14:26:17 crc kubenswrapper[4907]: I0313 14:26:17.683015 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-bw2xb-config-gdlfq" Mar 13 14:26:17 crc kubenswrapper[4907]: I0313 14:26:17.685032 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Mar 13 14:26:17 crc kubenswrapper[4907]: I0313 14:26:17.694851 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-bw2xb-config-gdlfq"] Mar 13 14:26:17 crc kubenswrapper[4907]: I0313 14:26:17.725618 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/8e734847-721a-487c-ad5d-e947f9eb2a5b-var-run-ovn\") pod \"ovn-controller-bw2xb-config-gdlfq\" (UID: \"8e734847-721a-487c-ad5d-e947f9eb2a5b\") " pod="openstack/ovn-controller-bw2xb-config-gdlfq" Mar 13 14:26:17 crc kubenswrapper[4907]: I0313 14:26:17.725666 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/8e734847-721a-487c-ad5d-e947f9eb2a5b-var-run\") pod \"ovn-controller-bw2xb-config-gdlfq\" (UID: \"8e734847-721a-487c-ad5d-e947f9eb2a5b\") " pod="openstack/ovn-controller-bw2xb-config-gdlfq" Mar 13 14:26:17 crc kubenswrapper[4907]: I0313 14:26:17.725684 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8e734847-721a-487c-ad5d-e947f9eb2a5b-scripts\") pod \"ovn-controller-bw2xb-config-gdlfq\" (UID: \"8e734847-721a-487c-ad5d-e947f9eb2a5b\") " pod="openstack/ovn-controller-bw2xb-config-gdlfq" Mar 13 14:26:17 crc kubenswrapper[4907]: I0313 14:26:17.725708 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/8e734847-721a-487c-ad5d-e947f9eb2a5b-var-log-ovn\") pod \"ovn-controller-bw2xb-config-gdlfq\" (UID: \"8e734847-721a-487c-ad5d-e947f9eb2a5b\") " pod="openstack/ovn-controller-bw2xb-config-gdlfq" Mar 13 14:26:17 crc kubenswrapper[4907]: I0313 14:26:17.725806 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9hp7z\" (UniqueName: \"kubernetes.io/projected/8e734847-721a-487c-ad5d-e947f9eb2a5b-kube-api-access-9hp7z\") pod \"ovn-controller-bw2xb-config-gdlfq\" (UID: \"8e734847-721a-487c-ad5d-e947f9eb2a5b\") " pod="openstack/ovn-controller-bw2xb-config-gdlfq" Mar 13 14:26:17 crc kubenswrapper[4907]: I0313 14:26:17.725854 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/8e734847-721a-487c-ad5d-e947f9eb2a5b-additional-scripts\") pod \"ovn-controller-bw2xb-config-gdlfq\" (UID: \"8e734847-721a-487c-ad5d-e947f9eb2a5b\") " pod="openstack/ovn-controller-bw2xb-config-gdlfq" Mar 13 14:26:17 crc kubenswrapper[4907]: I0313 14:26:17.827564 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9hp7z\" (UniqueName: \"kubernetes.io/projected/8e734847-721a-487c-ad5d-e947f9eb2a5b-kube-api-access-9hp7z\") pod \"ovn-controller-bw2xb-config-gdlfq\" (UID: \"8e734847-721a-487c-ad5d-e947f9eb2a5b\") " pod="openstack/ovn-controller-bw2xb-config-gdlfq" Mar 13 14:26:17 crc kubenswrapper[4907]: I0313 14:26:17.827643 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/8e734847-721a-487c-ad5d-e947f9eb2a5b-additional-scripts\") pod \"ovn-controller-bw2xb-config-gdlfq\" (UID: \"8e734847-721a-487c-ad5d-e947f9eb2a5b\") " pod="openstack/ovn-controller-bw2xb-config-gdlfq" Mar 13 14:26:17 crc kubenswrapper[4907]: I0313 14:26:17.827687 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/8e734847-721a-487c-ad5d-e947f9eb2a5b-var-run-ovn\") pod \"ovn-controller-bw2xb-config-gdlfq\" (UID: \"8e734847-721a-487c-ad5d-e947f9eb2a5b\") " pod="openstack/ovn-controller-bw2xb-config-gdlfq" Mar 13 14:26:17 crc kubenswrapper[4907]: I0313 14:26:17.827711 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/8e734847-721a-487c-ad5d-e947f9eb2a5b-var-run\") pod \"ovn-controller-bw2xb-config-gdlfq\" (UID: \"8e734847-721a-487c-ad5d-e947f9eb2a5b\") " pod="openstack/ovn-controller-bw2xb-config-gdlfq" Mar 13 14:26:17 crc kubenswrapper[4907]: I0313 14:26:17.827727 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8e734847-721a-487c-ad5d-e947f9eb2a5b-scripts\") pod \"ovn-controller-bw2xb-config-gdlfq\" (UID: \"8e734847-721a-487c-ad5d-e947f9eb2a5b\") " pod="openstack/ovn-controller-bw2xb-config-gdlfq" Mar 13 14:26:17 crc kubenswrapper[4907]: I0313 14:26:17.827756 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/8e734847-721a-487c-ad5d-e947f9eb2a5b-var-log-ovn\") pod \"ovn-controller-bw2xb-config-gdlfq\" (UID: \"8e734847-721a-487c-ad5d-e947f9eb2a5b\") " pod="openstack/ovn-controller-bw2xb-config-gdlfq" Mar 13 14:26:17 crc kubenswrapper[4907]: I0313 14:26:17.828094 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/8e734847-721a-487c-ad5d-e947f9eb2a5b-var-log-ovn\") pod \"ovn-controller-bw2xb-config-gdlfq\" (UID: \"8e734847-721a-487c-ad5d-e947f9eb2a5b\") " pod="openstack/ovn-controller-bw2xb-config-gdlfq" Mar 13 14:26:17 crc kubenswrapper[4907]: I0313 14:26:17.828097 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/8e734847-721a-487c-ad5d-e947f9eb2a5b-var-run-ovn\") pod \"ovn-controller-bw2xb-config-gdlfq\" (UID: \"8e734847-721a-487c-ad5d-e947f9eb2a5b\") " pod="openstack/ovn-controller-bw2xb-config-gdlfq" Mar 13 14:26:17 crc kubenswrapper[4907]: I0313 14:26:17.828205 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/8e734847-721a-487c-ad5d-e947f9eb2a5b-var-run\") pod \"ovn-controller-bw2xb-config-gdlfq\" (UID: \"8e734847-721a-487c-ad5d-e947f9eb2a5b\") " pod="openstack/ovn-controller-bw2xb-config-gdlfq" Mar 13 14:26:17 crc kubenswrapper[4907]: I0313 14:26:17.828993 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/8e734847-721a-487c-ad5d-e947f9eb2a5b-additional-scripts\") pod \"ovn-controller-bw2xb-config-gdlfq\" (UID: \"8e734847-721a-487c-ad5d-e947f9eb2a5b\") " pod="openstack/ovn-controller-bw2xb-config-gdlfq" Mar 13 14:26:17 crc kubenswrapper[4907]: I0313 14:26:17.830017 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8e734847-721a-487c-ad5d-e947f9eb2a5b-scripts\") pod \"ovn-controller-bw2xb-config-gdlfq\" (UID: \"8e734847-721a-487c-ad5d-e947f9eb2a5b\") " pod="openstack/ovn-controller-bw2xb-config-gdlfq" Mar 13 14:26:17 crc kubenswrapper[4907]: I0313 14:26:17.845315 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9hp7z\" (UniqueName: \"kubernetes.io/projected/8e734847-721a-487c-ad5d-e947f9eb2a5b-kube-api-access-9hp7z\") pod \"ovn-controller-bw2xb-config-gdlfq\" (UID: \"8e734847-721a-487c-ad5d-e947f9eb2a5b\") " pod="openstack/ovn-controller-bw2xb-config-gdlfq" Mar 13 14:26:18 crc kubenswrapper[4907]: I0313 14:26:18.004090 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-bw2xb-config-gdlfq" Mar 13 14:26:19 crc kubenswrapper[4907]: E0313 14:26:19.874211 4907 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-glance-api@sha256:dae5e39780d5a15eed030c7009f8e5317139d447558ac83f038497be594be120" Mar 13 14:26:19 crc kubenswrapper[4907]: E0313 14:26:19.874636 4907 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:glance-db-sync,Image:quay.io/podified-antelope-centos9/openstack-glance-api@sha256:dae5e39780d5a15eed030c7009f8e5317139d447558ac83f038497be594be120,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/glance/glance.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-65qwz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42415,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42415,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-db-sync-4v4ps_openstack(02262347-eaec-462e-a3ef-58c670885d99): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 13 14:26:19 crc kubenswrapper[4907]: E0313 14:26:19.875851 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/glance-db-sync-4v4ps" podUID="02262347-eaec-462e-a3ef-58c670885d99" Mar 13 14:26:19 crc kubenswrapper[4907]: I0313 14:26:19.998155 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-9fxfh" Mar 13 14:26:20 crc kubenswrapper[4907]: I0313 14:26:20.062603 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/e2585497-6837-4d17-8d51-c3d7879fdb46-etc-swift\") pod \"e2585497-6837-4d17-8d51-c3d7879fdb46\" (UID: \"e2585497-6837-4d17-8d51-c3d7879fdb46\") " Mar 13 14:26:20 crc kubenswrapper[4907]: I0313 14:26:20.064291 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e2585497-6837-4d17-8d51-c3d7879fdb46-scripts\") pod \"e2585497-6837-4d17-8d51-c3d7879fdb46\" (UID: \"e2585497-6837-4d17-8d51-c3d7879fdb46\") " Mar 13 14:26:20 crc kubenswrapper[4907]: I0313 14:26:20.064236 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e2585497-6837-4d17-8d51-c3d7879fdb46-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "e2585497-6837-4d17-8d51-c3d7879fdb46" (UID: "e2585497-6837-4d17-8d51-c3d7879fdb46"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:26:20 crc kubenswrapper[4907]: I0313 14:26:20.064372 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2585497-6837-4d17-8d51-c3d7879fdb46-combined-ca-bundle\") pod \"e2585497-6837-4d17-8d51-c3d7879fdb46\" (UID: \"e2585497-6837-4d17-8d51-c3d7879fdb46\") " Mar 13 14:26:20 crc kubenswrapper[4907]: I0313 14:26:20.065959 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-msr86\" (UniqueName: \"kubernetes.io/projected/e2585497-6837-4d17-8d51-c3d7879fdb46-kube-api-access-msr86\") pod \"e2585497-6837-4d17-8d51-c3d7879fdb46\" (UID: \"e2585497-6837-4d17-8d51-c3d7879fdb46\") " Mar 13 14:26:20 crc kubenswrapper[4907]: I0313 14:26:20.066017 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/e2585497-6837-4d17-8d51-c3d7879fdb46-swiftconf\") pod \"e2585497-6837-4d17-8d51-c3d7879fdb46\" (UID: \"e2585497-6837-4d17-8d51-c3d7879fdb46\") " Mar 13 14:26:20 crc kubenswrapper[4907]: I0313 14:26:20.066091 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/e2585497-6837-4d17-8d51-c3d7879fdb46-ring-data-devices\") pod \"e2585497-6837-4d17-8d51-c3d7879fdb46\" (UID: \"e2585497-6837-4d17-8d51-c3d7879fdb46\") " Mar 13 14:26:20 crc kubenswrapper[4907]: I0313 14:26:20.066141 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/e2585497-6837-4d17-8d51-c3d7879fdb46-dispersionconf\") pod \"e2585497-6837-4d17-8d51-c3d7879fdb46\" (UID: \"e2585497-6837-4d17-8d51-c3d7879fdb46\") " Mar 13 14:26:20 crc kubenswrapper[4907]: I0313 14:26:20.066796 4907 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/e2585497-6837-4d17-8d51-c3d7879fdb46-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:20 crc kubenswrapper[4907]: I0313 14:26:20.068006 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e2585497-6837-4d17-8d51-c3d7879fdb46-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "e2585497-6837-4d17-8d51-c3d7879fdb46" (UID: "e2585497-6837-4d17-8d51-c3d7879fdb46"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:26:20 crc kubenswrapper[4907]: I0313 14:26:20.071135 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e2585497-6837-4d17-8d51-c3d7879fdb46-kube-api-access-msr86" (OuterVolumeSpecName: "kube-api-access-msr86") pod "e2585497-6837-4d17-8d51-c3d7879fdb46" (UID: "e2585497-6837-4d17-8d51-c3d7879fdb46"). InnerVolumeSpecName "kube-api-access-msr86". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:26:20 crc kubenswrapper[4907]: I0313 14:26:20.077729 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e2585497-6837-4d17-8d51-c3d7879fdb46-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "e2585497-6837-4d17-8d51-c3d7879fdb46" (UID: "e2585497-6837-4d17-8d51-c3d7879fdb46"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:26:20 crc kubenswrapper[4907]: I0313 14:26:20.097125 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e2585497-6837-4d17-8d51-c3d7879fdb46-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e2585497-6837-4d17-8d51-c3d7879fdb46" (UID: "e2585497-6837-4d17-8d51-c3d7879fdb46"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:26:20 crc kubenswrapper[4907]: I0313 14:26:20.097721 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e2585497-6837-4d17-8d51-c3d7879fdb46-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "e2585497-6837-4d17-8d51-c3d7879fdb46" (UID: "e2585497-6837-4d17-8d51-c3d7879fdb46"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:26:20 crc kubenswrapper[4907]: I0313 14:26:20.107800 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e2585497-6837-4d17-8d51-c3d7879fdb46-scripts" (OuterVolumeSpecName: "scripts") pod "e2585497-6837-4d17-8d51-c3d7879fdb46" (UID: "e2585497-6837-4d17-8d51-c3d7879fdb46"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:26:20 crc kubenswrapper[4907]: I0313 14:26:20.169835 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-msr86\" (UniqueName: \"kubernetes.io/projected/e2585497-6837-4d17-8d51-c3d7879fdb46-kube-api-access-msr86\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:20 crc kubenswrapper[4907]: I0313 14:26:20.170107 4907 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/e2585497-6837-4d17-8d51-c3d7879fdb46-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:20 crc kubenswrapper[4907]: I0313 14:26:20.170120 4907 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/e2585497-6837-4d17-8d51-c3d7879fdb46-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:20 crc kubenswrapper[4907]: I0313 14:26:20.170128 4907 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/e2585497-6837-4d17-8d51-c3d7879fdb46-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:20 crc kubenswrapper[4907]: I0313 14:26:20.170136 4907 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e2585497-6837-4d17-8d51-c3d7879fdb46-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:20 crc kubenswrapper[4907]: I0313 14:26:20.170145 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2585497-6837-4d17-8d51-c3d7879fdb46-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:20 crc kubenswrapper[4907]: I0313 14:26:20.332141 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-bw2xb-config-gdlfq"] Mar 13 14:26:20 crc kubenswrapper[4907]: I0313 14:26:20.338656 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-r2rlh"] Mar 13 14:26:20 crc kubenswrapper[4907]: W0313 14:26:20.342726 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8e734847_721a_487c_ad5d_e947f9eb2a5b.slice/crio-f9b5f7112a5f19ae7b3025c0088305694e18947d40457ccaa65e79dd6ee2919b WatchSource:0}: Error finding container f9b5f7112a5f19ae7b3025c0088305694e18947d40457ccaa65e79dd6ee2919b: Status 404 returned error can't find the container with id f9b5f7112a5f19ae7b3025c0088305694e18947d40457ccaa65e79dd6ee2919b Mar 13 14:26:20 crc kubenswrapper[4907]: W0313 14:26:20.343405 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb3affff5_a720_4d5e_8f17_e8ec23eff0c9.slice/crio-d555d68605167e787613a0ff52207f2638755acf69400396ea72ba52afda2d81 WatchSource:0}: Error finding container d555d68605167e787613a0ff52207f2638755acf69400396ea72ba52afda2d81: Status 404 returned error can't find the container with id d555d68605167e787613a0ff52207f2638755acf69400396ea72ba52afda2d81 Mar 13 14:26:20 crc kubenswrapper[4907]: I0313 14:26:20.860350 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"e3c34454-315e-4821-ab25-b0f331a0d521","Type":"ContainerStarted","Data":"c5d00dd71aba57d230e51645212e11968b2acd8a912b082abd6373002fb907b4"} Mar 13 14:26:20 crc kubenswrapper[4907]: I0313 14:26:20.861009 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Mar 13 14:26:20 crc kubenswrapper[4907]: I0313 14:26:20.863363 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"286c82e8-b74a-49d7-a355-ac074aace10e","Type":"ContainerStarted","Data":"d23153977130306af0e21b41c03dbbc96a217b011704059af7d65c0a4153cf43"} Mar 13 14:26:20 crc kubenswrapper[4907]: I0313 14:26:20.863620 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Mar 13 14:26:20 crc kubenswrapper[4907]: I0313 14:26:20.883263 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-bw2xb-config-gdlfq" event={"ID":"8e734847-721a-487c-ad5d-e947f9eb2a5b","Type":"ContainerStarted","Data":"25b2791fa541284bf13b0da85be9466979de89ae098c8a967456ebbd1cd18235"} Mar 13 14:26:20 crc kubenswrapper[4907]: I0313 14:26:20.883373 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-bw2xb-config-gdlfq" event={"ID":"8e734847-721a-487c-ad5d-e947f9eb2a5b","Type":"ContainerStarted","Data":"f9b5f7112a5f19ae7b3025c0088305694e18947d40457ccaa65e79dd6ee2919b"} Mar 13 14:26:20 crc kubenswrapper[4907]: I0313 14:26:20.890044 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=54.508031654 podStartE2EDuration="1m3.890018007s" podCreationTimestamp="2026-03-13 14:25:17 +0000 UTC" firstStartedPulling="2026-03-13 14:25:31.046998857 +0000 UTC m=+1229.946786546" lastFinishedPulling="2026-03-13 14:25:40.42898521 +0000 UTC m=+1239.328772899" observedRunningTime="2026-03-13 14:26:20.885361749 +0000 UTC m=+1279.785149458" watchObservedRunningTime="2026-03-13 14:26:20.890018007 +0000 UTC m=+1279.789805696" Mar 13 14:26:20 crc kubenswrapper[4907]: I0313 14:26:20.896543 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-9fxfh" event={"ID":"e2585497-6837-4d17-8d51-c3d7879fdb46","Type":"ContainerDied","Data":"c44bcf911a2f35735f5bae70947997de7ad5c7bedca91be9607b903b91ea86c5"} Mar 13 14:26:20 crc kubenswrapper[4907]: I0313 14:26:20.896593 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c44bcf911a2f35735f5bae70947997de7ad5c7bedca91be9607b903b91ea86c5" Mar 13 14:26:20 crc kubenswrapper[4907]: I0313 14:26:20.896618 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-9fxfh" Mar 13 14:26:20 crc kubenswrapper[4907]: I0313 14:26:20.898967 4907 generic.go:334] "Generic (PLEG): container finished" podID="b3affff5-a720-4d5e-8f17-e8ec23eff0c9" containerID="e6d88705fdc67e86fd0214f4c6bbf1c37e140065fdd712f700392f915f879795" exitCode=0 Mar 13 14:26:20 crc kubenswrapper[4907]: I0313 14:26:20.899199 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-r2rlh" event={"ID":"b3affff5-a720-4d5e-8f17-e8ec23eff0c9","Type":"ContainerDied","Data":"e6d88705fdc67e86fd0214f4c6bbf1c37e140065fdd712f700392f915f879795"} Mar 13 14:26:20 crc kubenswrapper[4907]: I0313 14:26:20.899240 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-r2rlh" event={"ID":"b3affff5-a720-4d5e-8f17-e8ec23eff0c9","Type":"ContainerStarted","Data":"d555d68605167e787613a0ff52207f2638755acf69400396ea72ba52afda2d81"} Mar 13 14:26:20 crc kubenswrapper[4907]: E0313 14:26:20.905575 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-glance-api@sha256:dae5e39780d5a15eed030c7009f8e5317139d447558ac83f038497be594be120\\\"\"" pod="openstack/glance-db-sync-4v4ps" podUID="02262347-eaec-462e-a3ef-58c670885d99" Mar 13 14:26:20 crc kubenswrapper[4907]: I0313 14:26:20.924042 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-bw2xb-config-gdlfq" podStartSLOduration=3.924020841 podStartE2EDuration="3.924020841s" podCreationTimestamp="2026-03-13 14:26:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:26:20.90285762 +0000 UTC m=+1279.802645309" watchObservedRunningTime="2026-03-13 14:26:20.924020841 +0000 UTC m=+1279.823808530" Mar 13 14:26:20 crc kubenswrapper[4907]: I0313 14:26:20.931657 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=56.717069131 podStartE2EDuration="1m4.93163531s" podCreationTimestamp="2026-03-13 14:25:16 +0000 UTC" firstStartedPulling="2026-03-13 14:25:31.607641697 +0000 UTC m=+1230.507429386" lastFinishedPulling="2026-03-13 14:25:39.822207876 +0000 UTC m=+1238.721995565" observedRunningTime="2026-03-13 14:26:20.922445088 +0000 UTC m=+1279.822232777" watchObservedRunningTime="2026-03-13 14:26:20.93163531 +0000 UTC m=+1279.831422999" Mar 13 14:26:21 crc kubenswrapper[4907]: I0313 14:26:21.907544 4907 generic.go:334] "Generic (PLEG): container finished" podID="8e734847-721a-487c-ad5d-e947f9eb2a5b" containerID="25b2791fa541284bf13b0da85be9466979de89ae098c8a967456ebbd1cd18235" exitCode=0 Mar 13 14:26:21 crc kubenswrapper[4907]: I0313 14:26:21.907636 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-bw2xb-config-gdlfq" event={"ID":"8e734847-721a-487c-ad5d-e947f9eb2a5b","Type":"ContainerDied","Data":"25b2791fa541284bf13b0da85be9466979de89ae098c8a967456ebbd1cd18235"} Mar 13 14:26:22 crc kubenswrapper[4907]: I0313 14:26:22.192063 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-r2rlh" Mar 13 14:26:22 crc kubenswrapper[4907]: I0313 14:26:22.323599 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pgg7x\" (UniqueName: \"kubernetes.io/projected/b3affff5-a720-4d5e-8f17-e8ec23eff0c9-kube-api-access-pgg7x\") pod \"b3affff5-a720-4d5e-8f17-e8ec23eff0c9\" (UID: \"b3affff5-a720-4d5e-8f17-e8ec23eff0c9\") " Mar 13 14:26:22 crc kubenswrapper[4907]: I0313 14:26:22.323709 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b3affff5-a720-4d5e-8f17-e8ec23eff0c9-operator-scripts\") pod \"b3affff5-a720-4d5e-8f17-e8ec23eff0c9\" (UID: \"b3affff5-a720-4d5e-8f17-e8ec23eff0c9\") " Mar 13 14:26:22 crc kubenswrapper[4907]: I0313 14:26:22.324233 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b3affff5-a720-4d5e-8f17-e8ec23eff0c9-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b3affff5-a720-4d5e-8f17-e8ec23eff0c9" (UID: "b3affff5-a720-4d5e-8f17-e8ec23eff0c9"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:26:22 crc kubenswrapper[4907]: I0313 14:26:22.328738 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b3affff5-a720-4d5e-8f17-e8ec23eff0c9-kube-api-access-pgg7x" (OuterVolumeSpecName: "kube-api-access-pgg7x") pod "b3affff5-a720-4d5e-8f17-e8ec23eff0c9" (UID: "b3affff5-a720-4d5e-8f17-e8ec23eff0c9"). InnerVolumeSpecName "kube-api-access-pgg7x". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:26:22 crc kubenswrapper[4907]: I0313 14:26:22.425732 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pgg7x\" (UniqueName: \"kubernetes.io/projected/b3affff5-a720-4d5e-8f17-e8ec23eff0c9-kube-api-access-pgg7x\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:22 crc kubenswrapper[4907]: I0313 14:26:22.425776 4907 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b3affff5-a720-4d5e-8f17-e8ec23eff0c9-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:22 crc kubenswrapper[4907]: I0313 14:26:22.463832 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-bw2xb" Mar 13 14:26:22 crc kubenswrapper[4907]: I0313 14:26:22.919039 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-r2rlh" Mar 13 14:26:22 crc kubenswrapper[4907]: I0313 14:26:22.919046 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-r2rlh" event={"ID":"b3affff5-a720-4d5e-8f17-e8ec23eff0c9","Type":"ContainerDied","Data":"d555d68605167e787613a0ff52207f2638755acf69400396ea72ba52afda2d81"} Mar 13 14:26:22 crc kubenswrapper[4907]: I0313 14:26:22.919096 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d555d68605167e787613a0ff52207f2638755acf69400396ea72ba52afda2d81" Mar 13 14:26:23 crc kubenswrapper[4907]: I0313 14:26:23.214539 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-bw2xb-config-gdlfq" Mar 13 14:26:23 crc kubenswrapper[4907]: I0313 14:26:23.340416 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/8e734847-721a-487c-ad5d-e947f9eb2a5b-var-run\") pod \"8e734847-721a-487c-ad5d-e947f9eb2a5b\" (UID: \"8e734847-721a-487c-ad5d-e947f9eb2a5b\") " Mar 13 14:26:23 crc kubenswrapper[4907]: I0313 14:26:23.340512 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8e734847-721a-487c-ad5d-e947f9eb2a5b-var-run" (OuterVolumeSpecName: "var-run") pod "8e734847-721a-487c-ad5d-e947f9eb2a5b" (UID: "8e734847-721a-487c-ad5d-e947f9eb2a5b"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 14:26:23 crc kubenswrapper[4907]: I0313 14:26:23.340529 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9hp7z\" (UniqueName: \"kubernetes.io/projected/8e734847-721a-487c-ad5d-e947f9eb2a5b-kube-api-access-9hp7z\") pod \"8e734847-721a-487c-ad5d-e947f9eb2a5b\" (UID: \"8e734847-721a-487c-ad5d-e947f9eb2a5b\") " Mar 13 14:26:23 crc kubenswrapper[4907]: I0313 14:26:23.340623 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/8e734847-721a-487c-ad5d-e947f9eb2a5b-additional-scripts\") pod \"8e734847-721a-487c-ad5d-e947f9eb2a5b\" (UID: \"8e734847-721a-487c-ad5d-e947f9eb2a5b\") " Mar 13 14:26:23 crc kubenswrapper[4907]: I0313 14:26:23.340668 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/8e734847-721a-487c-ad5d-e947f9eb2a5b-var-log-ovn\") pod \"8e734847-721a-487c-ad5d-e947f9eb2a5b\" (UID: \"8e734847-721a-487c-ad5d-e947f9eb2a5b\") " Mar 13 14:26:23 crc kubenswrapper[4907]: I0313 14:26:23.340730 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8e734847-721a-487c-ad5d-e947f9eb2a5b-scripts\") pod \"8e734847-721a-487c-ad5d-e947f9eb2a5b\" (UID: \"8e734847-721a-487c-ad5d-e947f9eb2a5b\") " Mar 13 14:26:23 crc kubenswrapper[4907]: I0313 14:26:23.340791 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/8e734847-721a-487c-ad5d-e947f9eb2a5b-var-run-ovn\") pod \"8e734847-721a-487c-ad5d-e947f9eb2a5b\" (UID: \"8e734847-721a-487c-ad5d-e947f9eb2a5b\") " Mar 13 14:26:23 crc kubenswrapper[4907]: I0313 14:26:23.340790 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8e734847-721a-487c-ad5d-e947f9eb2a5b-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "8e734847-721a-487c-ad5d-e947f9eb2a5b" (UID: "8e734847-721a-487c-ad5d-e947f9eb2a5b"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 14:26:23 crc kubenswrapper[4907]: I0313 14:26:23.340909 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8e734847-721a-487c-ad5d-e947f9eb2a5b-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "8e734847-721a-487c-ad5d-e947f9eb2a5b" (UID: "8e734847-721a-487c-ad5d-e947f9eb2a5b"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 14:26:23 crc kubenswrapper[4907]: I0313 14:26:23.341084 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8e734847-721a-487c-ad5d-e947f9eb2a5b-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "8e734847-721a-487c-ad5d-e947f9eb2a5b" (UID: "8e734847-721a-487c-ad5d-e947f9eb2a5b"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:26:23 crc kubenswrapper[4907]: I0313 14:26:23.341426 4907 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/8e734847-721a-487c-ad5d-e947f9eb2a5b-var-run-ovn\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:23 crc kubenswrapper[4907]: I0313 14:26:23.341459 4907 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/8e734847-721a-487c-ad5d-e947f9eb2a5b-var-run\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:23 crc kubenswrapper[4907]: I0313 14:26:23.341473 4907 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/8e734847-721a-487c-ad5d-e947f9eb2a5b-additional-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:23 crc kubenswrapper[4907]: I0313 14:26:23.341487 4907 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/8e734847-721a-487c-ad5d-e947f9eb2a5b-var-log-ovn\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:23 crc kubenswrapper[4907]: I0313 14:26:23.341610 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8e734847-721a-487c-ad5d-e947f9eb2a5b-scripts" (OuterVolumeSpecName: "scripts") pod "8e734847-721a-487c-ad5d-e947f9eb2a5b" (UID: "8e734847-721a-487c-ad5d-e947f9eb2a5b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:26:23 crc kubenswrapper[4907]: I0313 14:26:23.358171 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8e734847-721a-487c-ad5d-e947f9eb2a5b-kube-api-access-9hp7z" (OuterVolumeSpecName: "kube-api-access-9hp7z") pod "8e734847-721a-487c-ad5d-e947f9eb2a5b" (UID: "8e734847-721a-487c-ad5d-e947f9eb2a5b"). InnerVolumeSpecName "kube-api-access-9hp7z". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:26:23 crc kubenswrapper[4907]: I0313 14:26:23.434352 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-bw2xb-config-gdlfq"] Mar 13 14:26:23 crc kubenswrapper[4907]: I0313 14:26:23.441020 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-bw2xb-config-gdlfq"] Mar 13 14:26:23 crc kubenswrapper[4907]: I0313 14:26:23.442717 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9hp7z\" (UniqueName: \"kubernetes.io/projected/8e734847-721a-487c-ad5d-e947f9eb2a5b-kube-api-access-9hp7z\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:23 crc kubenswrapper[4907]: I0313 14:26:23.442746 4907 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8e734847-721a-487c-ad5d-e947f9eb2a5b-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:23 crc kubenswrapper[4907]: I0313 14:26:23.791927 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8e734847-721a-487c-ad5d-e947f9eb2a5b" path="/var/lib/kubelet/pods/8e734847-721a-487c-ad5d-e947f9eb2a5b/volumes" Mar 13 14:26:23 crc kubenswrapper[4907]: I0313 14:26:23.927634 4907 scope.go:117] "RemoveContainer" containerID="25b2791fa541284bf13b0da85be9466979de89ae098c8a967456ebbd1cd18235" Mar 13 14:26:23 crc kubenswrapper[4907]: I0313 14:26:23.927774 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-bw2xb-config-gdlfq" Mar 13 14:26:24 crc kubenswrapper[4907]: I0313 14:26:24.825962 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/root-account-create-update-r2rlh"] Mar 13 14:26:24 crc kubenswrapper[4907]: I0313 14:26:24.833671 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/root-account-create-update-r2rlh"] Mar 13 14:26:25 crc kubenswrapper[4907]: I0313 14:26:25.791507 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b3affff5-a720-4d5e-8f17-e8ec23eff0c9" path="/var/lib/kubelet/pods/b3affff5-a720-4d5e-8f17-e8ec23eff0c9/volumes" Mar 13 14:26:26 crc kubenswrapper[4907]: I0313 14:26:26.386452 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/14f05f31-8185-43b0-be69-bcf8d5388ea5-etc-swift\") pod \"swift-storage-0\" (UID: \"14f05f31-8185-43b0-be69-bcf8d5388ea5\") " pod="openstack/swift-storage-0" Mar 13 14:26:26 crc kubenswrapper[4907]: I0313 14:26:26.393420 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/14f05f31-8185-43b0-be69-bcf8d5388ea5-etc-swift\") pod \"swift-storage-0\" (UID: \"14f05f31-8185-43b0-be69-bcf8d5388ea5\") " pod="openstack/swift-storage-0" Mar 13 14:26:26 crc kubenswrapper[4907]: I0313 14:26:26.461809 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Mar 13 14:26:27 crc kubenswrapper[4907]: I0313 14:26:27.288909 4907 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 13 14:26:27 crc kubenswrapper[4907]: I0313 14:26:27.290222 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Mar 13 14:26:27 crc kubenswrapper[4907]: I0313 14:26:27.966606 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"14f05f31-8185-43b0-be69-bcf8d5388ea5","Type":"ContainerStarted","Data":"8ab726bec0bb062d6ac5d70e873cfd78800b9921e7163b9cb0ca74467cc3c7f4"} Mar 13 14:26:28 crc kubenswrapper[4907]: I0313 14:26:28.976010 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"14f05f31-8185-43b0-be69-bcf8d5388ea5","Type":"ContainerStarted","Data":"8168297a51ad087d7375035164e54e4a0d24e6fd56cf85e922e11d80f7f3d091"} Mar 13 14:26:29 crc kubenswrapper[4907]: I0313 14:26:29.845819 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/root-account-create-update-r6vk2"] Mar 13 14:26:29 crc kubenswrapper[4907]: E0313 14:26:29.846413 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e734847-721a-487c-ad5d-e947f9eb2a5b" containerName="ovn-config" Mar 13 14:26:29 crc kubenswrapper[4907]: I0313 14:26:29.848246 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e734847-721a-487c-ad5d-e947f9eb2a5b" containerName="ovn-config" Mar 13 14:26:29 crc kubenswrapper[4907]: E0313 14:26:29.848297 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3affff5-a720-4d5e-8f17-e8ec23eff0c9" containerName="mariadb-account-create-update" Mar 13 14:26:29 crc kubenswrapper[4907]: I0313 14:26:29.848305 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3affff5-a720-4d5e-8f17-e8ec23eff0c9" containerName="mariadb-account-create-update" Mar 13 14:26:29 crc kubenswrapper[4907]: E0313 14:26:29.848318 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2585497-6837-4d17-8d51-c3d7879fdb46" containerName="swift-ring-rebalance" Mar 13 14:26:29 crc kubenswrapper[4907]: I0313 14:26:29.848325 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2585497-6837-4d17-8d51-c3d7879fdb46" containerName="swift-ring-rebalance" Mar 13 14:26:29 crc kubenswrapper[4907]: I0313 14:26:29.848495 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="b3affff5-a720-4d5e-8f17-e8ec23eff0c9" containerName="mariadb-account-create-update" Mar 13 14:26:29 crc kubenswrapper[4907]: I0313 14:26:29.848516 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="e2585497-6837-4d17-8d51-c3d7879fdb46" containerName="swift-ring-rebalance" Mar 13 14:26:29 crc kubenswrapper[4907]: I0313 14:26:29.848529 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="8e734847-721a-487c-ad5d-e947f9eb2a5b" containerName="ovn-config" Mar 13 14:26:29 crc kubenswrapper[4907]: I0313 14:26:29.849083 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-r6vk2" Mar 13 14:26:29 crc kubenswrapper[4907]: I0313 14:26:29.851597 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-mariadb-root-db-secret" Mar 13 14:26:29 crc kubenswrapper[4907]: I0313 14:26:29.857019 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-r6vk2"] Mar 13 14:26:29 crc kubenswrapper[4907]: I0313 14:26:29.940379 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d1b4fd44-4a11-4521-bfd0-16c5436368fa-operator-scripts\") pod \"root-account-create-update-r6vk2\" (UID: \"d1b4fd44-4a11-4521-bfd0-16c5436368fa\") " pod="openstack/root-account-create-update-r6vk2" Mar 13 14:26:29 crc kubenswrapper[4907]: I0313 14:26:29.940478 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-grms5\" (UniqueName: \"kubernetes.io/projected/d1b4fd44-4a11-4521-bfd0-16c5436368fa-kube-api-access-grms5\") pod \"root-account-create-update-r6vk2\" (UID: \"d1b4fd44-4a11-4521-bfd0-16c5436368fa\") " pod="openstack/root-account-create-update-r6vk2" Mar 13 14:26:29 crc kubenswrapper[4907]: I0313 14:26:29.989092 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"14f05f31-8185-43b0-be69-bcf8d5388ea5","Type":"ContainerStarted","Data":"c7370f4c24ed7b82c59bc2c1e3ae152c65420b5dd6227a66eb64b5b83832e410"} Mar 13 14:26:29 crc kubenswrapper[4907]: I0313 14:26:29.989140 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"14f05f31-8185-43b0-be69-bcf8d5388ea5","Type":"ContainerStarted","Data":"f8bc1f3dd2242db8254b664f00b3789940491fa59b38a7fc033c225b233b776e"} Mar 13 14:26:29 crc kubenswrapper[4907]: I0313 14:26:29.989153 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"14f05f31-8185-43b0-be69-bcf8d5388ea5","Type":"ContainerStarted","Data":"9b6f16a5610e8f2c08a40a29bfba0c00e1ff883cb29b5290604713aad1dab566"} Mar 13 14:26:30 crc kubenswrapper[4907]: I0313 14:26:30.042765 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d1b4fd44-4a11-4521-bfd0-16c5436368fa-operator-scripts\") pod \"root-account-create-update-r6vk2\" (UID: \"d1b4fd44-4a11-4521-bfd0-16c5436368fa\") " pod="openstack/root-account-create-update-r6vk2" Mar 13 14:26:30 crc kubenswrapper[4907]: I0313 14:26:30.042832 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-grms5\" (UniqueName: \"kubernetes.io/projected/d1b4fd44-4a11-4521-bfd0-16c5436368fa-kube-api-access-grms5\") pod \"root-account-create-update-r6vk2\" (UID: \"d1b4fd44-4a11-4521-bfd0-16c5436368fa\") " pod="openstack/root-account-create-update-r6vk2" Mar 13 14:26:30 crc kubenswrapper[4907]: I0313 14:26:30.044589 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d1b4fd44-4a11-4521-bfd0-16c5436368fa-operator-scripts\") pod \"root-account-create-update-r6vk2\" (UID: \"d1b4fd44-4a11-4521-bfd0-16c5436368fa\") " pod="openstack/root-account-create-update-r6vk2" Mar 13 14:26:30 crc kubenswrapper[4907]: I0313 14:26:30.064831 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-grms5\" (UniqueName: \"kubernetes.io/projected/d1b4fd44-4a11-4521-bfd0-16c5436368fa-kube-api-access-grms5\") pod \"root-account-create-update-r6vk2\" (UID: \"d1b4fd44-4a11-4521-bfd0-16c5436368fa\") " pod="openstack/root-account-create-update-r6vk2" Mar 13 14:26:30 crc kubenswrapper[4907]: I0313 14:26:30.168017 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-r6vk2" Mar 13 14:26:30 crc kubenswrapper[4907]: I0313 14:26:30.431260 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-r6vk2"] Mar 13 14:26:30 crc kubenswrapper[4907]: I0313 14:26:30.997095 4907 generic.go:334] "Generic (PLEG): container finished" podID="d1b4fd44-4a11-4521-bfd0-16c5436368fa" containerID="3e1e04ee01cb4c593a590da900f137153aed8f7c58309d1a24f19dd94c9af998" exitCode=0 Mar 13 14:26:30 crc kubenswrapper[4907]: I0313 14:26:30.997281 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-r6vk2" event={"ID":"d1b4fd44-4a11-4521-bfd0-16c5436368fa","Type":"ContainerDied","Data":"3e1e04ee01cb4c593a590da900f137153aed8f7c58309d1a24f19dd94c9af998"} Mar 13 14:26:31 crc kubenswrapper[4907]: I0313 14:26:30.997724 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-r6vk2" event={"ID":"d1b4fd44-4a11-4521-bfd0-16c5436368fa","Type":"ContainerStarted","Data":"2471a799ca22d8e83409150033938766a0e509f65d736ee99f5f36e665621536"} Mar 13 14:26:31 crc kubenswrapper[4907]: I0313 14:26:31.001769 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"14f05f31-8185-43b0-be69-bcf8d5388ea5","Type":"ContainerStarted","Data":"2bc2275a73f47c1a27db775d1cc1c3ae9c20c01f9ddad7a35bf652467b66ea80"} Mar 13 14:26:31 crc kubenswrapper[4907]: I0313 14:26:31.001818 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"14f05f31-8185-43b0-be69-bcf8d5388ea5","Type":"ContainerStarted","Data":"b7f60cadabc1fd2d193073423cce7296e305ff6325001c7ec9823dd267a1787a"} Mar 13 14:26:32 crc kubenswrapper[4907]: I0313 14:26:32.012398 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"14f05f31-8185-43b0-be69-bcf8d5388ea5","Type":"ContainerStarted","Data":"dbcb2b472a15fbcc053532418c2f7a654827914050e539781030bc0749d6218f"} Mar 13 14:26:32 crc kubenswrapper[4907]: I0313 14:26:32.012795 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"14f05f31-8185-43b0-be69-bcf8d5388ea5","Type":"ContainerStarted","Data":"7a94b1f20e2cb964aaf2fe5c9f23f0cd193291a14edd5411717f7cad81ed9808"} Mar 13 14:26:32 crc kubenswrapper[4907]: I0313 14:26:32.303672 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-r6vk2" Mar 13 14:26:32 crc kubenswrapper[4907]: I0313 14:26:32.383535 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d1b4fd44-4a11-4521-bfd0-16c5436368fa-operator-scripts\") pod \"d1b4fd44-4a11-4521-bfd0-16c5436368fa\" (UID: \"d1b4fd44-4a11-4521-bfd0-16c5436368fa\") " Mar 13 14:26:32 crc kubenswrapper[4907]: I0313 14:26:32.383666 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-grms5\" (UniqueName: \"kubernetes.io/projected/d1b4fd44-4a11-4521-bfd0-16c5436368fa-kube-api-access-grms5\") pod \"d1b4fd44-4a11-4521-bfd0-16c5436368fa\" (UID: \"d1b4fd44-4a11-4521-bfd0-16c5436368fa\") " Mar 13 14:26:32 crc kubenswrapper[4907]: I0313 14:26:32.384250 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d1b4fd44-4a11-4521-bfd0-16c5436368fa-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d1b4fd44-4a11-4521-bfd0-16c5436368fa" (UID: "d1b4fd44-4a11-4521-bfd0-16c5436368fa"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:26:32 crc kubenswrapper[4907]: I0313 14:26:32.388974 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d1b4fd44-4a11-4521-bfd0-16c5436368fa-kube-api-access-grms5" (OuterVolumeSpecName: "kube-api-access-grms5") pod "d1b4fd44-4a11-4521-bfd0-16c5436368fa" (UID: "d1b4fd44-4a11-4521-bfd0-16c5436368fa"). InnerVolumeSpecName "kube-api-access-grms5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:26:32 crc kubenswrapper[4907]: I0313 14:26:32.485625 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-grms5\" (UniqueName: \"kubernetes.io/projected/d1b4fd44-4a11-4521-bfd0-16c5436368fa-kube-api-access-grms5\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:32 crc kubenswrapper[4907]: I0313 14:26:32.485946 4907 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d1b4fd44-4a11-4521-bfd0-16c5436368fa-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:33 crc kubenswrapper[4907]: I0313 14:26:33.040111 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-r6vk2" event={"ID":"d1b4fd44-4a11-4521-bfd0-16c5436368fa","Type":"ContainerDied","Data":"2471a799ca22d8e83409150033938766a0e509f65d736ee99f5f36e665621536"} Mar 13 14:26:33 crc kubenswrapper[4907]: I0313 14:26:33.040159 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2471a799ca22d8e83409150033938766a0e509f65d736ee99f5f36e665621536" Mar 13 14:26:33 crc kubenswrapper[4907]: I0313 14:26:33.040213 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-r6vk2" Mar 13 14:26:33 crc kubenswrapper[4907]: I0313 14:26:33.061270 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"14f05f31-8185-43b0-be69-bcf8d5388ea5","Type":"ContainerStarted","Data":"8765bf4cde625366ddd57fa87118a09b18b3fbbf1d2ded31c12344e9c08dadb2"} Mar 13 14:26:33 crc kubenswrapper[4907]: I0313 14:26:33.061315 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"14f05f31-8185-43b0-be69-bcf8d5388ea5","Type":"ContainerStarted","Data":"6f5cf167a5953dd1bdb0f436b238215d64975f16e188a1dacd546c8d2adfba14"} Mar 13 14:26:33 crc kubenswrapper[4907]: I0313 14:26:33.061329 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"14f05f31-8185-43b0-be69-bcf8d5388ea5","Type":"ContainerStarted","Data":"c1b4fe0c395990393cfbace1600ea9872fde43535722ed738332ba0b26f178a4"} Mar 13 14:26:33 crc kubenswrapper[4907]: I0313 14:26:33.061342 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"14f05f31-8185-43b0-be69-bcf8d5388ea5","Type":"ContainerStarted","Data":"5a6b392bf5f4cbd99ba2fc751327322f8960f4c73aba15cb9a501d4c9c3edd94"} Mar 13 14:26:33 crc kubenswrapper[4907]: I0313 14:26:33.061354 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"14f05f31-8185-43b0-be69-bcf8d5388ea5","Type":"ContainerStarted","Data":"c81c73db839bc006e1b7fdce225dd05a9a50870ea60ca389ea146e1e69d195b7"} Mar 13 14:26:34 crc kubenswrapper[4907]: I0313 14:26:34.076850 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"14f05f31-8185-43b0-be69-bcf8d5388ea5","Type":"ContainerStarted","Data":"0696dc2db616f4616e43dd171ba757d11bce09169b35b516964d1af58bad39ea"} Mar 13 14:26:34 crc kubenswrapper[4907]: I0313 14:26:34.077245 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"14f05f31-8185-43b0-be69-bcf8d5388ea5","Type":"ContainerStarted","Data":"1e5ba68b21b276693f03089f51ca3955dd88adcaacae52727d03a999b543d411"} Mar 13 14:26:34 crc kubenswrapper[4907]: I0313 14:26:34.113198 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=36.27268048 podStartE2EDuration="41.113178399s" podCreationTimestamp="2026-03-13 14:25:53 +0000 UTC" firstStartedPulling="2026-03-13 14:26:27.288573833 +0000 UTC m=+1286.188361532" lastFinishedPulling="2026-03-13 14:26:32.129071762 +0000 UTC m=+1291.028859451" observedRunningTime="2026-03-13 14:26:34.110913706 +0000 UTC m=+1293.010701425" watchObservedRunningTime="2026-03-13 14:26:34.113178399 +0000 UTC m=+1293.012966098" Mar 13 14:26:34 crc kubenswrapper[4907]: I0313 14:26:34.396971 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-67754df655-k2v49"] Mar 13 14:26:34 crc kubenswrapper[4907]: E0313 14:26:34.397584 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1b4fd44-4a11-4521-bfd0-16c5436368fa" containerName="mariadb-account-create-update" Mar 13 14:26:34 crc kubenswrapper[4907]: I0313 14:26:34.397652 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1b4fd44-4a11-4521-bfd0-16c5436368fa" containerName="mariadb-account-create-update" Mar 13 14:26:34 crc kubenswrapper[4907]: I0313 14:26:34.397851 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1b4fd44-4a11-4521-bfd0-16c5436368fa" containerName="mariadb-account-create-update" Mar 13 14:26:34 crc kubenswrapper[4907]: I0313 14:26:34.398746 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67754df655-k2v49" Mar 13 14:26:34 crc kubenswrapper[4907]: I0313 14:26:34.400931 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Mar 13 14:26:34 crc kubenswrapper[4907]: I0313 14:26:34.413128 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-67754df655-k2v49"] Mar 13 14:26:34 crc kubenswrapper[4907]: I0313 14:26:34.520241 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c065ce46-d7f3-4e50-b30c-83a1592b6ead-ovsdbserver-sb\") pod \"dnsmasq-dns-67754df655-k2v49\" (UID: \"c065ce46-d7f3-4e50-b30c-83a1592b6ead\") " pod="openstack/dnsmasq-dns-67754df655-k2v49" Mar 13 14:26:34 crc kubenswrapper[4907]: I0313 14:26:34.520310 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c065ce46-d7f3-4e50-b30c-83a1592b6ead-ovsdbserver-nb\") pod \"dnsmasq-dns-67754df655-k2v49\" (UID: \"c065ce46-d7f3-4e50-b30c-83a1592b6ead\") " pod="openstack/dnsmasq-dns-67754df655-k2v49" Mar 13 14:26:34 crc kubenswrapper[4907]: I0313 14:26:34.520343 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c065ce46-d7f3-4e50-b30c-83a1592b6ead-config\") pod \"dnsmasq-dns-67754df655-k2v49\" (UID: \"c065ce46-d7f3-4e50-b30c-83a1592b6ead\") " pod="openstack/dnsmasq-dns-67754df655-k2v49" Mar 13 14:26:34 crc kubenswrapper[4907]: I0313 14:26:34.520377 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c065ce46-d7f3-4e50-b30c-83a1592b6ead-dns-swift-storage-0\") pod \"dnsmasq-dns-67754df655-k2v49\" (UID: \"c065ce46-d7f3-4e50-b30c-83a1592b6ead\") " pod="openstack/dnsmasq-dns-67754df655-k2v49" Mar 13 14:26:34 crc kubenswrapper[4907]: I0313 14:26:34.520395 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vpx9g\" (UniqueName: \"kubernetes.io/projected/c065ce46-d7f3-4e50-b30c-83a1592b6ead-kube-api-access-vpx9g\") pod \"dnsmasq-dns-67754df655-k2v49\" (UID: \"c065ce46-d7f3-4e50-b30c-83a1592b6ead\") " pod="openstack/dnsmasq-dns-67754df655-k2v49" Mar 13 14:26:34 crc kubenswrapper[4907]: I0313 14:26:34.520412 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c065ce46-d7f3-4e50-b30c-83a1592b6ead-dns-svc\") pod \"dnsmasq-dns-67754df655-k2v49\" (UID: \"c065ce46-d7f3-4e50-b30c-83a1592b6ead\") " pod="openstack/dnsmasq-dns-67754df655-k2v49" Mar 13 14:26:34 crc kubenswrapper[4907]: I0313 14:26:34.622062 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c065ce46-d7f3-4e50-b30c-83a1592b6ead-ovsdbserver-sb\") pod \"dnsmasq-dns-67754df655-k2v49\" (UID: \"c065ce46-d7f3-4e50-b30c-83a1592b6ead\") " pod="openstack/dnsmasq-dns-67754df655-k2v49" Mar 13 14:26:34 crc kubenswrapper[4907]: I0313 14:26:34.622141 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c065ce46-d7f3-4e50-b30c-83a1592b6ead-ovsdbserver-nb\") pod \"dnsmasq-dns-67754df655-k2v49\" (UID: \"c065ce46-d7f3-4e50-b30c-83a1592b6ead\") " pod="openstack/dnsmasq-dns-67754df655-k2v49" Mar 13 14:26:34 crc kubenswrapper[4907]: I0313 14:26:34.622176 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c065ce46-d7f3-4e50-b30c-83a1592b6ead-config\") pod \"dnsmasq-dns-67754df655-k2v49\" (UID: \"c065ce46-d7f3-4e50-b30c-83a1592b6ead\") " pod="openstack/dnsmasq-dns-67754df655-k2v49" Mar 13 14:26:34 crc kubenswrapper[4907]: I0313 14:26:34.622215 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c065ce46-d7f3-4e50-b30c-83a1592b6ead-dns-swift-storage-0\") pod \"dnsmasq-dns-67754df655-k2v49\" (UID: \"c065ce46-d7f3-4e50-b30c-83a1592b6ead\") " pod="openstack/dnsmasq-dns-67754df655-k2v49" Mar 13 14:26:34 crc kubenswrapper[4907]: I0313 14:26:34.622264 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vpx9g\" (UniqueName: \"kubernetes.io/projected/c065ce46-d7f3-4e50-b30c-83a1592b6ead-kube-api-access-vpx9g\") pod \"dnsmasq-dns-67754df655-k2v49\" (UID: \"c065ce46-d7f3-4e50-b30c-83a1592b6ead\") " pod="openstack/dnsmasq-dns-67754df655-k2v49" Mar 13 14:26:34 crc kubenswrapper[4907]: I0313 14:26:34.622290 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c065ce46-d7f3-4e50-b30c-83a1592b6ead-dns-svc\") pod \"dnsmasq-dns-67754df655-k2v49\" (UID: \"c065ce46-d7f3-4e50-b30c-83a1592b6ead\") " pod="openstack/dnsmasq-dns-67754df655-k2v49" Mar 13 14:26:34 crc kubenswrapper[4907]: I0313 14:26:34.623212 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c065ce46-d7f3-4e50-b30c-83a1592b6ead-ovsdbserver-sb\") pod \"dnsmasq-dns-67754df655-k2v49\" (UID: \"c065ce46-d7f3-4e50-b30c-83a1592b6ead\") " pod="openstack/dnsmasq-dns-67754df655-k2v49" Mar 13 14:26:34 crc kubenswrapper[4907]: I0313 14:26:34.623217 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c065ce46-d7f3-4e50-b30c-83a1592b6ead-dns-svc\") pod \"dnsmasq-dns-67754df655-k2v49\" (UID: \"c065ce46-d7f3-4e50-b30c-83a1592b6ead\") " pod="openstack/dnsmasq-dns-67754df655-k2v49" Mar 13 14:26:34 crc kubenswrapper[4907]: I0313 14:26:34.623802 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c065ce46-d7f3-4e50-b30c-83a1592b6ead-config\") pod \"dnsmasq-dns-67754df655-k2v49\" (UID: \"c065ce46-d7f3-4e50-b30c-83a1592b6ead\") " pod="openstack/dnsmasq-dns-67754df655-k2v49" Mar 13 14:26:34 crc kubenswrapper[4907]: I0313 14:26:34.623915 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c065ce46-d7f3-4e50-b30c-83a1592b6ead-dns-swift-storage-0\") pod \"dnsmasq-dns-67754df655-k2v49\" (UID: \"c065ce46-d7f3-4e50-b30c-83a1592b6ead\") " pod="openstack/dnsmasq-dns-67754df655-k2v49" Mar 13 14:26:34 crc kubenswrapper[4907]: I0313 14:26:34.624038 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c065ce46-d7f3-4e50-b30c-83a1592b6ead-ovsdbserver-nb\") pod \"dnsmasq-dns-67754df655-k2v49\" (UID: \"c065ce46-d7f3-4e50-b30c-83a1592b6ead\") " pod="openstack/dnsmasq-dns-67754df655-k2v49" Mar 13 14:26:34 crc kubenswrapper[4907]: I0313 14:26:34.641135 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vpx9g\" (UniqueName: \"kubernetes.io/projected/c065ce46-d7f3-4e50-b30c-83a1592b6ead-kube-api-access-vpx9g\") pod \"dnsmasq-dns-67754df655-k2v49\" (UID: \"c065ce46-d7f3-4e50-b30c-83a1592b6ead\") " pod="openstack/dnsmasq-dns-67754df655-k2v49" Mar 13 14:26:34 crc kubenswrapper[4907]: I0313 14:26:34.718253 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67754df655-k2v49" Mar 13 14:26:35 crc kubenswrapper[4907]: I0313 14:26:35.157264 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-67754df655-k2v49"] Mar 13 14:26:36 crc kubenswrapper[4907]: I0313 14:26:36.098490 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-4v4ps" event={"ID":"02262347-eaec-462e-a3ef-58c670885d99","Type":"ContainerStarted","Data":"2ad298467d04021b9b1a343e12a26a79c8a6605114bfe6ee4b9cfd24470e49c4"} Mar 13 14:26:36 crc kubenswrapper[4907]: I0313 14:26:36.100551 4907 generic.go:334] "Generic (PLEG): container finished" podID="c065ce46-d7f3-4e50-b30c-83a1592b6ead" containerID="fc5cca7a53563949cc0a2a52bb6bc5ab553013ec2df261991d02d7a2bd1a082d" exitCode=0 Mar 13 14:26:36 crc kubenswrapper[4907]: I0313 14:26:36.100608 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67754df655-k2v49" event={"ID":"c065ce46-d7f3-4e50-b30c-83a1592b6ead","Type":"ContainerDied","Data":"fc5cca7a53563949cc0a2a52bb6bc5ab553013ec2df261991d02d7a2bd1a082d"} Mar 13 14:26:36 crc kubenswrapper[4907]: I0313 14:26:36.100647 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67754df655-k2v49" event={"ID":"c065ce46-d7f3-4e50-b30c-83a1592b6ead","Type":"ContainerStarted","Data":"1dc89b2197970bff5fe57426501341e047b5f1d0b4ba93884702cc3864ea9041"} Mar 13 14:26:36 crc kubenswrapper[4907]: I0313 14:26:36.145838 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-4v4ps" podStartSLOduration=3.156574488 podStartE2EDuration="35.145814487s" podCreationTimestamp="2026-03-13 14:26:01 +0000 UTC" firstStartedPulling="2026-03-13 14:26:03.2710231 +0000 UTC m=+1262.170810789" lastFinishedPulling="2026-03-13 14:26:35.260263099 +0000 UTC m=+1294.160050788" observedRunningTime="2026-03-13 14:26:36.121278114 +0000 UTC m=+1295.021065803" watchObservedRunningTime="2026-03-13 14:26:36.145814487 +0000 UTC m=+1295.045602186" Mar 13 14:26:37 crc kubenswrapper[4907]: I0313 14:26:37.109992 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67754df655-k2v49" event={"ID":"c065ce46-d7f3-4e50-b30c-83a1592b6ead","Type":"ContainerStarted","Data":"fd0a2c793e62b5b128e16861d049a22d36d9baa0de3410f3961ca59044ad1213"} Mar 13 14:26:37 crc kubenswrapper[4907]: I0313 14:26:37.110369 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-67754df655-k2v49" Mar 13 14:26:37 crc kubenswrapper[4907]: I0313 14:26:37.140686 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-67754df655-k2v49" podStartSLOduration=3.140657178 podStartE2EDuration="3.140657178s" podCreationTimestamp="2026-03-13 14:26:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:26:37.132839033 +0000 UTC m=+1296.032626802" watchObservedRunningTime="2026-03-13 14:26:37.140657178 +0000 UTC m=+1296.040444897" Mar 13 14:26:38 crc kubenswrapper[4907]: I0313 14:26:38.278097 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Mar 13 14:26:38 crc kubenswrapper[4907]: I0313 14:26:38.698068 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.120125 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-vbcql"] Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.121328 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-vbcql" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.143458 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-vbcql"] Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.215554 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-47e2-account-create-update-r4ccm"] Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.216828 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-47e2-account-create-update-r4ccm" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.217556 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-chjsd\" (UniqueName: \"kubernetes.io/projected/e1068e89-10df-4d4c-a213-d9e9643fab38-kube-api-access-chjsd\") pod \"cinder-db-create-vbcql\" (UID: \"e1068e89-10df-4d4c-a213-d9e9643fab38\") " pod="openstack/cinder-db-create-vbcql" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.217616 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e1068e89-10df-4d4c-a213-d9e9643fab38-operator-scripts\") pod \"cinder-db-create-vbcql\" (UID: \"e1068e89-10df-4d4c-a213-d9e9643fab38\") " pod="openstack/cinder-db-create-vbcql" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.219470 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.230057 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-htpvb"] Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.231497 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-htpvb" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.241226 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-47e2-account-create-update-r4ccm"] Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.250441 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-htpvb"] Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.319498 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e1068e89-10df-4d4c-a213-d9e9643fab38-operator-scripts\") pod \"cinder-db-create-vbcql\" (UID: \"e1068e89-10df-4d4c-a213-d9e9643fab38\") " pod="openstack/cinder-db-create-vbcql" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.319578 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a89525da-db62-49c2-b548-bab4c974dfdd-operator-scripts\") pod \"barbican-db-create-htpvb\" (UID: \"a89525da-db62-49c2-b548-bab4c974dfdd\") " pod="openstack/barbican-db-create-htpvb" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.319607 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kgvgp\" (UniqueName: \"kubernetes.io/projected/3fe8116b-9e68-41e2-aad2-cdd7d5e51c83-kube-api-access-kgvgp\") pod \"cinder-47e2-account-create-update-r4ccm\" (UID: \"3fe8116b-9e68-41e2-aad2-cdd7d5e51c83\") " pod="openstack/cinder-47e2-account-create-update-r4ccm" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.319639 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3fe8116b-9e68-41e2-aad2-cdd7d5e51c83-operator-scripts\") pod \"cinder-47e2-account-create-update-r4ccm\" (UID: \"3fe8116b-9e68-41e2-aad2-cdd7d5e51c83\") " pod="openstack/cinder-47e2-account-create-update-r4ccm" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.319677 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-chjsd\" (UniqueName: \"kubernetes.io/projected/e1068e89-10df-4d4c-a213-d9e9643fab38-kube-api-access-chjsd\") pod \"cinder-db-create-vbcql\" (UID: \"e1068e89-10df-4d4c-a213-d9e9643fab38\") " pod="openstack/cinder-db-create-vbcql" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.319714 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qv2ks\" (UniqueName: \"kubernetes.io/projected/a89525da-db62-49c2-b548-bab4c974dfdd-kube-api-access-qv2ks\") pod \"barbican-db-create-htpvb\" (UID: \"a89525da-db62-49c2-b548-bab4c974dfdd\") " pod="openstack/barbican-db-create-htpvb" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.320338 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e1068e89-10df-4d4c-a213-d9e9643fab38-operator-scripts\") pod \"cinder-db-create-vbcql\" (UID: \"e1068e89-10df-4d4c-a213-d9e9643fab38\") " pod="openstack/cinder-db-create-vbcql" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.348628 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-chjsd\" (UniqueName: \"kubernetes.io/projected/e1068e89-10df-4d4c-a213-d9e9643fab38-kube-api-access-chjsd\") pod \"cinder-db-create-vbcql\" (UID: \"e1068e89-10df-4d4c-a213-d9e9643fab38\") " pod="openstack/cinder-db-create-vbcql" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.411289 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-x9b2b"] Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.412544 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-x9b2b" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.422603 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a89525da-db62-49c2-b548-bab4c974dfdd-operator-scripts\") pod \"barbican-db-create-htpvb\" (UID: \"a89525da-db62-49c2-b548-bab4c974dfdd\") " pod="openstack/barbican-db-create-htpvb" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.422687 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kgvgp\" (UniqueName: \"kubernetes.io/projected/3fe8116b-9e68-41e2-aad2-cdd7d5e51c83-kube-api-access-kgvgp\") pod \"cinder-47e2-account-create-update-r4ccm\" (UID: \"3fe8116b-9e68-41e2-aad2-cdd7d5e51c83\") " pod="openstack/cinder-47e2-account-create-update-r4ccm" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.422746 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3fe8116b-9e68-41e2-aad2-cdd7d5e51c83-operator-scripts\") pod \"cinder-47e2-account-create-update-r4ccm\" (UID: \"3fe8116b-9e68-41e2-aad2-cdd7d5e51c83\") " pod="openstack/cinder-47e2-account-create-update-r4ccm" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.422831 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qv2ks\" (UniqueName: \"kubernetes.io/projected/a89525da-db62-49c2-b548-bab4c974dfdd-kube-api-access-qv2ks\") pod \"barbican-db-create-htpvb\" (UID: \"a89525da-db62-49c2-b548-bab4c974dfdd\") " pod="openstack/barbican-db-create-htpvb" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.424819 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3fe8116b-9e68-41e2-aad2-cdd7d5e51c83-operator-scripts\") pod \"cinder-47e2-account-create-update-r4ccm\" (UID: \"3fe8116b-9e68-41e2-aad2-cdd7d5e51c83\") " pod="openstack/cinder-47e2-account-create-update-r4ccm" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.439419 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a89525da-db62-49c2-b548-bab4c974dfdd-operator-scripts\") pod \"barbican-db-create-htpvb\" (UID: \"a89525da-db62-49c2-b548-bab4c974dfdd\") " pod="openstack/barbican-db-create-htpvb" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.444429 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-vbcql" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.445638 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-x9b2b"] Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.484669 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kgvgp\" (UniqueName: \"kubernetes.io/projected/3fe8116b-9e68-41e2-aad2-cdd7d5e51c83-kube-api-access-kgvgp\") pod \"cinder-47e2-account-create-update-r4ccm\" (UID: \"3fe8116b-9e68-41e2-aad2-cdd7d5e51c83\") " pod="openstack/cinder-47e2-account-create-update-r4ccm" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.485907 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-kcjn8"] Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.487381 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-kcjn8" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.488026 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qv2ks\" (UniqueName: \"kubernetes.io/projected/a89525da-db62-49c2-b548-bab4c974dfdd-kube-api-access-qv2ks\") pod \"barbican-db-create-htpvb\" (UID: \"a89525da-db62-49c2-b548-bab4c974dfdd\") " pod="openstack/barbican-db-create-htpvb" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.489973 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.490237 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.490596 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.490875 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-p9m6m" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.497072 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-b501-account-create-update-dzlqr"] Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.498336 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-b501-account-create-update-dzlqr" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.500092 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.504676 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-b501-account-create-update-dzlqr"] Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.515401 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-kcjn8"] Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.524389 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2bab6ed6-5335-40cc-bdff-3d7461876fb8-operator-scripts\") pod \"neutron-db-create-x9b2b\" (UID: \"2bab6ed6-5335-40cc-bdff-3d7461876fb8\") " pod="openstack/neutron-db-create-x9b2b" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.524450 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lwp6d\" (UniqueName: \"kubernetes.io/projected/2bab6ed6-5335-40cc-bdff-3d7461876fb8-kube-api-access-lwp6d\") pod \"neutron-db-create-x9b2b\" (UID: \"2bab6ed6-5335-40cc-bdff-3d7461876fb8\") " pod="openstack/neutron-db-create-x9b2b" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.532990 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-47e2-account-create-update-r4ccm" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.545772 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-htpvb" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.626299 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/98748d4d-f934-49ed-91d9-7a9e467b2753-operator-scripts\") pod \"barbican-b501-account-create-update-dzlqr\" (UID: \"98748d4d-f934-49ed-91d9-7a9e467b2753\") " pod="openstack/barbican-b501-account-create-update-dzlqr" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.626764 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g78kx\" (UniqueName: \"kubernetes.io/projected/480d23fe-2980-46b3-9e86-e74f2d051357-kube-api-access-g78kx\") pod \"keystone-db-sync-kcjn8\" (UID: \"480d23fe-2980-46b3-9e86-e74f2d051357\") " pod="openstack/keystone-db-sync-kcjn8" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.626838 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/480d23fe-2980-46b3-9e86-e74f2d051357-combined-ca-bundle\") pod \"keystone-db-sync-kcjn8\" (UID: \"480d23fe-2980-46b3-9e86-e74f2d051357\") " pod="openstack/keystone-db-sync-kcjn8" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.627083 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2bab6ed6-5335-40cc-bdff-3d7461876fb8-operator-scripts\") pod \"neutron-db-create-x9b2b\" (UID: \"2bab6ed6-5335-40cc-bdff-3d7461876fb8\") " pod="openstack/neutron-db-create-x9b2b" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.627162 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lwp6d\" (UniqueName: \"kubernetes.io/projected/2bab6ed6-5335-40cc-bdff-3d7461876fb8-kube-api-access-lwp6d\") pod \"neutron-db-create-x9b2b\" (UID: \"2bab6ed6-5335-40cc-bdff-3d7461876fb8\") " pod="openstack/neutron-db-create-x9b2b" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.627195 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ddmjb\" (UniqueName: \"kubernetes.io/projected/98748d4d-f934-49ed-91d9-7a9e467b2753-kube-api-access-ddmjb\") pod \"barbican-b501-account-create-update-dzlqr\" (UID: \"98748d4d-f934-49ed-91d9-7a9e467b2753\") " pod="openstack/barbican-b501-account-create-update-dzlqr" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.627298 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/480d23fe-2980-46b3-9e86-e74f2d051357-config-data\") pod \"keystone-db-sync-kcjn8\" (UID: \"480d23fe-2980-46b3-9e86-e74f2d051357\") " pod="openstack/keystone-db-sync-kcjn8" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.628421 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2bab6ed6-5335-40cc-bdff-3d7461876fb8-operator-scripts\") pod \"neutron-db-create-x9b2b\" (UID: \"2bab6ed6-5335-40cc-bdff-3d7461876fb8\") " pod="openstack/neutron-db-create-x9b2b" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.641041 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-b69e-account-create-update-ptpcx"] Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.642544 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-b69e-account-create-update-ptpcx" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.662778 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.684212 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-b69e-account-create-update-ptpcx"] Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.703465 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lwp6d\" (UniqueName: \"kubernetes.io/projected/2bab6ed6-5335-40cc-bdff-3d7461876fb8-kube-api-access-lwp6d\") pod \"neutron-db-create-x9b2b\" (UID: \"2bab6ed6-5335-40cc-bdff-3d7461876fb8\") " pod="openstack/neutron-db-create-x9b2b" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.728454 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ddmjb\" (UniqueName: \"kubernetes.io/projected/98748d4d-f934-49ed-91d9-7a9e467b2753-kube-api-access-ddmjb\") pod \"barbican-b501-account-create-update-dzlqr\" (UID: \"98748d4d-f934-49ed-91d9-7a9e467b2753\") " pod="openstack/barbican-b501-account-create-update-dzlqr" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.728537 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/480d23fe-2980-46b3-9e86-e74f2d051357-config-data\") pod \"keystone-db-sync-kcjn8\" (UID: \"480d23fe-2980-46b3-9e86-e74f2d051357\") " pod="openstack/keystone-db-sync-kcjn8" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.728561 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d2gv5\" (UniqueName: \"kubernetes.io/projected/0983a1bd-a83c-4658-a405-09b8ab0d0002-kube-api-access-d2gv5\") pod \"neutron-b69e-account-create-update-ptpcx\" (UID: \"0983a1bd-a83c-4658-a405-09b8ab0d0002\") " pod="openstack/neutron-b69e-account-create-update-ptpcx" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.728597 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/98748d4d-f934-49ed-91d9-7a9e467b2753-operator-scripts\") pod \"barbican-b501-account-create-update-dzlqr\" (UID: \"98748d4d-f934-49ed-91d9-7a9e467b2753\") " pod="openstack/barbican-b501-account-create-update-dzlqr" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.728622 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0983a1bd-a83c-4658-a405-09b8ab0d0002-operator-scripts\") pod \"neutron-b69e-account-create-update-ptpcx\" (UID: \"0983a1bd-a83c-4658-a405-09b8ab0d0002\") " pod="openstack/neutron-b69e-account-create-update-ptpcx" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.728641 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g78kx\" (UniqueName: \"kubernetes.io/projected/480d23fe-2980-46b3-9e86-e74f2d051357-kube-api-access-g78kx\") pod \"keystone-db-sync-kcjn8\" (UID: \"480d23fe-2980-46b3-9e86-e74f2d051357\") " pod="openstack/keystone-db-sync-kcjn8" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.728675 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/480d23fe-2980-46b3-9e86-e74f2d051357-combined-ca-bundle\") pod \"keystone-db-sync-kcjn8\" (UID: \"480d23fe-2980-46b3-9e86-e74f2d051357\") " pod="openstack/keystone-db-sync-kcjn8" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.729418 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/98748d4d-f934-49ed-91d9-7a9e467b2753-operator-scripts\") pod \"barbican-b501-account-create-update-dzlqr\" (UID: \"98748d4d-f934-49ed-91d9-7a9e467b2753\") " pod="openstack/barbican-b501-account-create-update-dzlqr" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.735363 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/480d23fe-2980-46b3-9e86-e74f2d051357-combined-ca-bundle\") pod \"keystone-db-sync-kcjn8\" (UID: \"480d23fe-2980-46b3-9e86-e74f2d051357\") " pod="openstack/keystone-db-sync-kcjn8" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.735984 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/480d23fe-2980-46b3-9e86-e74f2d051357-config-data\") pod \"keystone-db-sync-kcjn8\" (UID: \"480d23fe-2980-46b3-9e86-e74f2d051357\") " pod="openstack/keystone-db-sync-kcjn8" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.739219 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-x9b2b" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.754523 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g78kx\" (UniqueName: \"kubernetes.io/projected/480d23fe-2980-46b3-9e86-e74f2d051357-kube-api-access-g78kx\") pod \"keystone-db-sync-kcjn8\" (UID: \"480d23fe-2980-46b3-9e86-e74f2d051357\") " pod="openstack/keystone-db-sync-kcjn8" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.762532 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ddmjb\" (UniqueName: \"kubernetes.io/projected/98748d4d-f934-49ed-91d9-7a9e467b2753-kube-api-access-ddmjb\") pod \"barbican-b501-account-create-update-dzlqr\" (UID: \"98748d4d-f934-49ed-91d9-7a9e467b2753\") " pod="openstack/barbican-b501-account-create-update-dzlqr" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.830143 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d2gv5\" (UniqueName: \"kubernetes.io/projected/0983a1bd-a83c-4658-a405-09b8ab0d0002-kube-api-access-d2gv5\") pod \"neutron-b69e-account-create-update-ptpcx\" (UID: \"0983a1bd-a83c-4658-a405-09b8ab0d0002\") " pod="openstack/neutron-b69e-account-create-update-ptpcx" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.830217 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0983a1bd-a83c-4658-a405-09b8ab0d0002-operator-scripts\") pod \"neutron-b69e-account-create-update-ptpcx\" (UID: \"0983a1bd-a83c-4658-a405-09b8ab0d0002\") " pod="openstack/neutron-b69e-account-create-update-ptpcx" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.832312 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0983a1bd-a83c-4658-a405-09b8ab0d0002-operator-scripts\") pod \"neutron-b69e-account-create-update-ptpcx\" (UID: \"0983a1bd-a83c-4658-a405-09b8ab0d0002\") " pod="openstack/neutron-b69e-account-create-update-ptpcx" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.853058 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d2gv5\" (UniqueName: \"kubernetes.io/projected/0983a1bd-a83c-4658-a405-09b8ab0d0002-kube-api-access-d2gv5\") pod \"neutron-b69e-account-create-update-ptpcx\" (UID: \"0983a1bd-a83c-4658-a405-09b8ab0d0002\") " pod="openstack/neutron-b69e-account-create-update-ptpcx" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.964952 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-kcjn8" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.982351 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-b501-account-create-update-dzlqr" Mar 13 14:26:41 crc kubenswrapper[4907]: I0313 14:26:41.003092 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-b69e-account-create-update-ptpcx" Mar 13 14:26:41 crc kubenswrapper[4907]: I0313 14:26:41.051118 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-vbcql"] Mar 13 14:26:41 crc kubenswrapper[4907]: I0313 14:26:41.181390 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-47e2-account-create-update-r4ccm"] Mar 13 14:26:41 crc kubenswrapper[4907]: I0313 14:26:41.184225 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-vbcql" event={"ID":"e1068e89-10df-4d4c-a213-d9e9643fab38","Type":"ContainerStarted","Data":"800ea320e37f9ac3fbdc6c79cbb80396598682384d93ce308ede98d248a09cc4"} Mar 13 14:26:41 crc kubenswrapper[4907]: I0313 14:26:41.211675 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-htpvb"] Mar 13 14:26:41 crc kubenswrapper[4907]: W0313 14:26:41.242715 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda89525da_db62_49c2_b548_bab4c974dfdd.slice/crio-265dfd02ff62990e5c06a12bdcd62f177d8601a264d83ff1c46d7bdee007b20d WatchSource:0}: Error finding container 265dfd02ff62990e5c06a12bdcd62f177d8601a264d83ff1c46d7bdee007b20d: Status 404 returned error can't find the container with id 265dfd02ff62990e5c06a12bdcd62f177d8601a264d83ff1c46d7bdee007b20d Mar 13 14:26:41 crc kubenswrapper[4907]: I0313 14:26:41.269592 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-x9b2b"] Mar 13 14:26:41 crc kubenswrapper[4907]: W0313 14:26:41.284063 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2bab6ed6_5335_40cc_bdff_3d7461876fb8.slice/crio-ce59465e7c0c37d5e78037bb1b277bb43931b1fc4a590e4e9d4e42dd7fdc27d3 WatchSource:0}: Error finding container ce59465e7c0c37d5e78037bb1b277bb43931b1fc4a590e4e9d4e42dd7fdc27d3: Status 404 returned error can't find the container with id ce59465e7c0c37d5e78037bb1b277bb43931b1fc4a590e4e9d4e42dd7fdc27d3 Mar 13 14:26:41 crc kubenswrapper[4907]: I0313 14:26:41.517999 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-b69e-account-create-update-ptpcx"] Mar 13 14:26:41 crc kubenswrapper[4907]: W0313 14:26:41.527091 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0983a1bd_a83c_4658_a405_09b8ab0d0002.slice/crio-807b6a972511f96035dc8e7a5888a9cb76102f79128af800532f9871f36ec567 WatchSource:0}: Error finding container 807b6a972511f96035dc8e7a5888a9cb76102f79128af800532f9871f36ec567: Status 404 returned error can't find the container with id 807b6a972511f96035dc8e7a5888a9cb76102f79128af800532f9871f36ec567 Mar 13 14:26:41 crc kubenswrapper[4907]: I0313 14:26:41.594557 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-b501-account-create-update-dzlqr"] Mar 13 14:26:41 crc kubenswrapper[4907]: I0313 14:26:41.629836 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-kcjn8"] Mar 13 14:26:42 crc kubenswrapper[4907]: I0313 14:26:42.198530 4907 generic.go:334] "Generic (PLEG): container finished" podID="3fe8116b-9e68-41e2-aad2-cdd7d5e51c83" containerID="8f4cb6476a276775b25f72fa46f11e3f550f531236e5cf19234f8b274637da17" exitCode=0 Mar 13 14:26:42 crc kubenswrapper[4907]: I0313 14:26:42.198871 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-47e2-account-create-update-r4ccm" event={"ID":"3fe8116b-9e68-41e2-aad2-cdd7d5e51c83","Type":"ContainerDied","Data":"8f4cb6476a276775b25f72fa46f11e3f550f531236e5cf19234f8b274637da17"} Mar 13 14:26:42 crc kubenswrapper[4907]: I0313 14:26:42.198914 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-47e2-account-create-update-r4ccm" event={"ID":"3fe8116b-9e68-41e2-aad2-cdd7d5e51c83","Type":"ContainerStarted","Data":"31fc0aff0eaffc735f077ff239eae73496c2aa5d04daeeff77d6a777c8e6f3f3"} Mar 13 14:26:42 crc kubenswrapper[4907]: I0313 14:26:42.200453 4907 generic.go:334] "Generic (PLEG): container finished" podID="98748d4d-f934-49ed-91d9-7a9e467b2753" containerID="cdd8592044b8b42302c000b60e593bcd7aa7352aa0bfd95a951e13ddb4d1cef1" exitCode=0 Mar 13 14:26:42 crc kubenswrapper[4907]: I0313 14:26:42.200490 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-b501-account-create-update-dzlqr" event={"ID":"98748d4d-f934-49ed-91d9-7a9e467b2753","Type":"ContainerDied","Data":"cdd8592044b8b42302c000b60e593bcd7aa7352aa0bfd95a951e13ddb4d1cef1"} Mar 13 14:26:42 crc kubenswrapper[4907]: I0313 14:26:42.200505 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-b501-account-create-update-dzlqr" event={"ID":"98748d4d-f934-49ed-91d9-7a9e467b2753","Type":"ContainerStarted","Data":"5a33722749134dbbc6eb786e1dd47e935db2b7e8fc7b1822f79e1da556bcd6c2"} Mar 13 14:26:42 crc kubenswrapper[4907]: I0313 14:26:42.203361 4907 generic.go:334] "Generic (PLEG): container finished" podID="a89525da-db62-49c2-b548-bab4c974dfdd" containerID="766d1e6536dea7ad49da9b30346790d8d4eb902b9e7e7b3c7bf460cfd415e154" exitCode=0 Mar 13 14:26:42 crc kubenswrapper[4907]: I0313 14:26:42.203417 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-htpvb" event={"ID":"a89525da-db62-49c2-b548-bab4c974dfdd","Type":"ContainerDied","Data":"766d1e6536dea7ad49da9b30346790d8d4eb902b9e7e7b3c7bf460cfd415e154"} Mar 13 14:26:42 crc kubenswrapper[4907]: I0313 14:26:42.203440 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-htpvb" event={"ID":"a89525da-db62-49c2-b548-bab4c974dfdd","Type":"ContainerStarted","Data":"265dfd02ff62990e5c06a12bdcd62f177d8601a264d83ff1c46d7bdee007b20d"} Mar 13 14:26:42 crc kubenswrapper[4907]: I0313 14:26:42.206535 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-kcjn8" event={"ID":"480d23fe-2980-46b3-9e86-e74f2d051357","Type":"ContainerStarted","Data":"35b6cb9082e770b5ec87125f38e9a5f5d08513cff5fef490907772e3d539df90"} Mar 13 14:26:42 crc kubenswrapper[4907]: I0313 14:26:42.208310 4907 generic.go:334] "Generic (PLEG): container finished" podID="e1068e89-10df-4d4c-a213-d9e9643fab38" containerID="0a3f47d3069fe894c10b414d1b636502b45151a11f2acdaa45abe3e56ee5a7a1" exitCode=0 Mar 13 14:26:42 crc kubenswrapper[4907]: I0313 14:26:42.208360 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-vbcql" event={"ID":"e1068e89-10df-4d4c-a213-d9e9643fab38","Type":"ContainerDied","Data":"0a3f47d3069fe894c10b414d1b636502b45151a11f2acdaa45abe3e56ee5a7a1"} Mar 13 14:26:42 crc kubenswrapper[4907]: I0313 14:26:42.210605 4907 generic.go:334] "Generic (PLEG): container finished" podID="2bab6ed6-5335-40cc-bdff-3d7461876fb8" containerID="71b8e65c6bc18f48a724ed12b9dcf478104abf3f0beba637548e1962f43f12bb" exitCode=0 Mar 13 14:26:42 crc kubenswrapper[4907]: I0313 14:26:42.210686 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-x9b2b" event={"ID":"2bab6ed6-5335-40cc-bdff-3d7461876fb8","Type":"ContainerDied","Data":"71b8e65c6bc18f48a724ed12b9dcf478104abf3f0beba637548e1962f43f12bb"} Mar 13 14:26:42 crc kubenswrapper[4907]: I0313 14:26:42.210714 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-x9b2b" event={"ID":"2bab6ed6-5335-40cc-bdff-3d7461876fb8","Type":"ContainerStarted","Data":"ce59465e7c0c37d5e78037bb1b277bb43931b1fc4a590e4e9d4e42dd7fdc27d3"} Mar 13 14:26:42 crc kubenswrapper[4907]: I0313 14:26:42.218261 4907 generic.go:334] "Generic (PLEG): container finished" podID="0983a1bd-a83c-4658-a405-09b8ab0d0002" containerID="e3a0771ba1e8115d911c740ed70ee5fca37f11ff18ec41bfab54c46258690b83" exitCode=0 Mar 13 14:26:42 crc kubenswrapper[4907]: I0313 14:26:42.218317 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-b69e-account-create-update-ptpcx" event={"ID":"0983a1bd-a83c-4658-a405-09b8ab0d0002","Type":"ContainerDied","Data":"e3a0771ba1e8115d911c740ed70ee5fca37f11ff18ec41bfab54c46258690b83"} Mar 13 14:26:42 crc kubenswrapper[4907]: I0313 14:26:42.218343 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-b69e-account-create-update-ptpcx" event={"ID":"0983a1bd-a83c-4658-a405-09b8ab0d0002","Type":"ContainerStarted","Data":"807b6a972511f96035dc8e7a5888a9cb76102f79128af800532f9871f36ec567"} Mar 13 14:26:44 crc kubenswrapper[4907]: I0313 14:26:44.235941 4907 generic.go:334] "Generic (PLEG): container finished" podID="02262347-eaec-462e-a3ef-58c670885d99" containerID="2ad298467d04021b9b1a343e12a26a79c8a6605114bfe6ee4b9cfd24470e49c4" exitCode=0 Mar 13 14:26:44 crc kubenswrapper[4907]: I0313 14:26:44.236052 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-4v4ps" event={"ID":"02262347-eaec-462e-a3ef-58c670885d99","Type":"ContainerDied","Data":"2ad298467d04021b9b1a343e12a26a79c8a6605114bfe6ee4b9cfd24470e49c4"} Mar 13 14:26:44 crc kubenswrapper[4907]: I0313 14:26:44.720064 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-67754df655-k2v49" Mar 13 14:26:44 crc kubenswrapper[4907]: I0313 14:26:44.808530 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7b9fd7d84c-wkdxf"] Mar 13 14:26:44 crc kubenswrapper[4907]: I0313 14:26:44.808753 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7b9fd7d84c-wkdxf" podUID="89cd945f-311f-47d1-982a-641f062e4f57" containerName="dnsmasq-dns" containerID="cri-o://ad9755b55eba4bb5e259a681410af3ffba2c66af2583b34cd43ac29178757a96" gracePeriod=10 Mar 13 14:26:45 crc kubenswrapper[4907]: I0313 14:26:45.256559 4907 generic.go:334] "Generic (PLEG): container finished" podID="89cd945f-311f-47d1-982a-641f062e4f57" containerID="ad9755b55eba4bb5e259a681410af3ffba2c66af2583b34cd43ac29178757a96" exitCode=0 Mar 13 14:26:45 crc kubenswrapper[4907]: I0313 14:26:45.256644 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b9fd7d84c-wkdxf" event={"ID":"89cd945f-311f-47d1-982a-641f062e4f57","Type":"ContainerDied","Data":"ad9755b55eba4bb5e259a681410af3ffba2c66af2583b34cd43ac29178757a96"} Mar 13 14:26:46 crc kubenswrapper[4907]: I0313 14:26:46.996723 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-b501-account-create-update-dzlqr" Mar 13 14:26:47 crc kubenswrapper[4907]: I0313 14:26:47.004938 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-x9b2b" Mar 13 14:26:47 crc kubenswrapper[4907]: I0313 14:26:47.020855 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-vbcql" Mar 13 14:26:47 crc kubenswrapper[4907]: I0313 14:26:47.030322 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-htpvb" Mar 13 14:26:47 crc kubenswrapper[4907]: I0313 14:26:47.041981 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-b69e-account-create-update-ptpcx" Mar 13 14:26:47 crc kubenswrapper[4907]: I0313 14:26:47.063183 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-47e2-account-create-update-r4ccm" Mar 13 14:26:47 crc kubenswrapper[4907]: I0313 14:26:47.088119 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lwp6d\" (UniqueName: \"kubernetes.io/projected/2bab6ed6-5335-40cc-bdff-3d7461876fb8-kube-api-access-lwp6d\") pod \"2bab6ed6-5335-40cc-bdff-3d7461876fb8\" (UID: \"2bab6ed6-5335-40cc-bdff-3d7461876fb8\") " Mar 13 14:26:47 crc kubenswrapper[4907]: I0313 14:26:47.088176 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/98748d4d-f934-49ed-91d9-7a9e467b2753-operator-scripts\") pod \"98748d4d-f934-49ed-91d9-7a9e467b2753\" (UID: \"98748d4d-f934-49ed-91d9-7a9e467b2753\") " Mar 13 14:26:47 crc kubenswrapper[4907]: I0313 14:26:47.088246 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ddmjb\" (UniqueName: \"kubernetes.io/projected/98748d4d-f934-49ed-91d9-7a9e467b2753-kube-api-access-ddmjb\") pod \"98748d4d-f934-49ed-91d9-7a9e467b2753\" (UID: \"98748d4d-f934-49ed-91d9-7a9e467b2753\") " Mar 13 14:26:47 crc kubenswrapper[4907]: I0313 14:26:47.088308 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2bab6ed6-5335-40cc-bdff-3d7461876fb8-operator-scripts\") pod \"2bab6ed6-5335-40cc-bdff-3d7461876fb8\" (UID: \"2bab6ed6-5335-40cc-bdff-3d7461876fb8\") " Mar 13 14:26:47 crc kubenswrapper[4907]: I0313 14:26:47.089640 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/98748d4d-f934-49ed-91d9-7a9e467b2753-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "98748d4d-f934-49ed-91d9-7a9e467b2753" (UID: "98748d4d-f934-49ed-91d9-7a9e467b2753"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:26:47 crc kubenswrapper[4907]: I0313 14:26:47.090695 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2bab6ed6-5335-40cc-bdff-3d7461876fb8-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2bab6ed6-5335-40cc-bdff-3d7461876fb8" (UID: "2bab6ed6-5335-40cc-bdff-3d7461876fb8"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:26:47 crc kubenswrapper[4907]: I0313 14:26:47.102767 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2bab6ed6-5335-40cc-bdff-3d7461876fb8-kube-api-access-lwp6d" (OuterVolumeSpecName: "kube-api-access-lwp6d") pod "2bab6ed6-5335-40cc-bdff-3d7461876fb8" (UID: "2bab6ed6-5335-40cc-bdff-3d7461876fb8"). InnerVolumeSpecName "kube-api-access-lwp6d". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:26:47 crc kubenswrapper[4907]: I0313 14:26:47.104507 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/98748d4d-f934-49ed-91d9-7a9e467b2753-kube-api-access-ddmjb" (OuterVolumeSpecName: "kube-api-access-ddmjb") pod "98748d4d-f934-49ed-91d9-7a9e467b2753" (UID: "98748d4d-f934-49ed-91d9-7a9e467b2753"). InnerVolumeSpecName "kube-api-access-ddmjb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:26:47 crc kubenswrapper[4907]: I0313 14:26:47.191075 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3fe8116b-9e68-41e2-aad2-cdd7d5e51c83-operator-scripts\") pod \"3fe8116b-9e68-41e2-aad2-cdd7d5e51c83\" (UID: \"3fe8116b-9e68-41e2-aad2-cdd7d5e51c83\") " Mar 13 14:26:47 crc kubenswrapper[4907]: I0313 14:26:47.191119 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a89525da-db62-49c2-b548-bab4c974dfdd-operator-scripts\") pod \"a89525da-db62-49c2-b548-bab4c974dfdd\" (UID: \"a89525da-db62-49c2-b548-bab4c974dfdd\") " Mar 13 14:26:47 crc kubenswrapper[4907]: I0313 14:26:47.191206 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d2gv5\" (UniqueName: \"kubernetes.io/projected/0983a1bd-a83c-4658-a405-09b8ab0d0002-kube-api-access-d2gv5\") pod \"0983a1bd-a83c-4658-a405-09b8ab0d0002\" (UID: \"0983a1bd-a83c-4658-a405-09b8ab0d0002\") " Mar 13 14:26:47 crc kubenswrapper[4907]: I0313 14:26:47.191244 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0983a1bd-a83c-4658-a405-09b8ab0d0002-operator-scripts\") pod \"0983a1bd-a83c-4658-a405-09b8ab0d0002\" (UID: \"0983a1bd-a83c-4658-a405-09b8ab0d0002\") " Mar 13 14:26:47 crc kubenswrapper[4907]: I0313 14:26:47.191289 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e1068e89-10df-4d4c-a213-d9e9643fab38-operator-scripts\") pod \"e1068e89-10df-4d4c-a213-d9e9643fab38\" (UID: \"e1068e89-10df-4d4c-a213-d9e9643fab38\") " Mar 13 14:26:47 crc kubenswrapper[4907]: I0313 14:26:47.191328 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-chjsd\" (UniqueName: \"kubernetes.io/projected/e1068e89-10df-4d4c-a213-d9e9643fab38-kube-api-access-chjsd\") pod \"e1068e89-10df-4d4c-a213-d9e9643fab38\" (UID: \"e1068e89-10df-4d4c-a213-d9e9643fab38\") " Mar 13 14:26:47 crc kubenswrapper[4907]: I0313 14:26:47.191353 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kgvgp\" (UniqueName: \"kubernetes.io/projected/3fe8116b-9e68-41e2-aad2-cdd7d5e51c83-kube-api-access-kgvgp\") pod \"3fe8116b-9e68-41e2-aad2-cdd7d5e51c83\" (UID: \"3fe8116b-9e68-41e2-aad2-cdd7d5e51c83\") " Mar 13 14:26:47 crc kubenswrapper[4907]: I0313 14:26:47.191384 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qv2ks\" (UniqueName: \"kubernetes.io/projected/a89525da-db62-49c2-b548-bab4c974dfdd-kube-api-access-qv2ks\") pod \"a89525da-db62-49c2-b548-bab4c974dfdd\" (UID: \"a89525da-db62-49c2-b548-bab4c974dfdd\") " Mar 13 14:26:47 crc kubenswrapper[4907]: I0313 14:26:47.194018 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lwp6d\" (UniqueName: \"kubernetes.io/projected/2bab6ed6-5335-40cc-bdff-3d7461876fb8-kube-api-access-lwp6d\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:47 crc kubenswrapper[4907]: I0313 14:26:47.194045 4907 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/98748d4d-f934-49ed-91d9-7a9e467b2753-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:47 crc kubenswrapper[4907]: I0313 14:26:47.194057 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ddmjb\" (UniqueName: \"kubernetes.io/projected/98748d4d-f934-49ed-91d9-7a9e467b2753-kube-api-access-ddmjb\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:47 crc kubenswrapper[4907]: I0313 14:26:47.194069 4907 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2bab6ed6-5335-40cc-bdff-3d7461876fb8-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:47 crc kubenswrapper[4907]: I0313 14:26:47.194249 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3fe8116b-9e68-41e2-aad2-cdd7d5e51c83-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "3fe8116b-9e68-41e2-aad2-cdd7d5e51c83" (UID: "3fe8116b-9e68-41e2-aad2-cdd7d5e51c83"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:26:47 crc kubenswrapper[4907]: I0313 14:26:47.194471 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a89525da-db62-49c2-b548-bab4c974dfdd-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a89525da-db62-49c2-b548-bab4c974dfdd" (UID: "a89525da-db62-49c2-b548-bab4c974dfdd"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:26:47 crc kubenswrapper[4907]: I0313 14:26:47.194596 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e1068e89-10df-4d4c-a213-d9e9643fab38-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e1068e89-10df-4d4c-a213-d9e9643fab38" (UID: "e1068e89-10df-4d4c-a213-d9e9643fab38"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:26:47 crc kubenswrapper[4907]: I0313 14:26:47.194971 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0983a1bd-a83c-4658-a405-09b8ab0d0002-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "0983a1bd-a83c-4658-a405-09b8ab0d0002" (UID: "0983a1bd-a83c-4658-a405-09b8ab0d0002"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:26:47 crc kubenswrapper[4907]: I0313 14:26:47.199709 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3fe8116b-9e68-41e2-aad2-cdd7d5e51c83-kube-api-access-kgvgp" (OuterVolumeSpecName: "kube-api-access-kgvgp") pod "3fe8116b-9e68-41e2-aad2-cdd7d5e51c83" (UID: "3fe8116b-9e68-41e2-aad2-cdd7d5e51c83"). InnerVolumeSpecName "kube-api-access-kgvgp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:26:47 crc kubenswrapper[4907]: I0313 14:26:47.204428 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0983a1bd-a83c-4658-a405-09b8ab0d0002-kube-api-access-d2gv5" (OuterVolumeSpecName: "kube-api-access-d2gv5") pod "0983a1bd-a83c-4658-a405-09b8ab0d0002" (UID: "0983a1bd-a83c-4658-a405-09b8ab0d0002"). InnerVolumeSpecName "kube-api-access-d2gv5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:26:47 crc kubenswrapper[4907]: I0313 14:26:47.210805 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a89525da-db62-49c2-b548-bab4c974dfdd-kube-api-access-qv2ks" (OuterVolumeSpecName: "kube-api-access-qv2ks") pod "a89525da-db62-49c2-b548-bab4c974dfdd" (UID: "a89525da-db62-49c2-b548-bab4c974dfdd"). InnerVolumeSpecName "kube-api-access-qv2ks". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:26:47 crc kubenswrapper[4907]: I0313 14:26:47.232215 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e1068e89-10df-4d4c-a213-d9e9643fab38-kube-api-access-chjsd" (OuterVolumeSpecName: "kube-api-access-chjsd") pod "e1068e89-10df-4d4c-a213-d9e9643fab38" (UID: "e1068e89-10df-4d4c-a213-d9e9643fab38"). InnerVolumeSpecName "kube-api-access-chjsd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:26:47 crc kubenswrapper[4907]: I0313 14:26:47.280762 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-47e2-account-create-update-r4ccm" Mar 13 14:26:47 crc kubenswrapper[4907]: I0313 14:26:47.280994 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-47e2-account-create-update-r4ccm" event={"ID":"3fe8116b-9e68-41e2-aad2-cdd7d5e51c83","Type":"ContainerDied","Data":"31fc0aff0eaffc735f077ff239eae73496c2aa5d04daeeff77d6a777c8e6f3f3"} Mar 13 14:26:47 crc kubenswrapper[4907]: I0313 14:26:47.281030 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="31fc0aff0eaffc735f077ff239eae73496c2aa5d04daeeff77d6a777c8e6f3f3" Mar 13 14:26:47 crc kubenswrapper[4907]: I0313 14:26:47.283943 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-b501-account-create-update-dzlqr" event={"ID":"98748d4d-f934-49ed-91d9-7a9e467b2753","Type":"ContainerDied","Data":"5a33722749134dbbc6eb786e1dd47e935db2b7e8fc7b1822f79e1da556bcd6c2"} Mar 13 14:26:47 crc kubenswrapper[4907]: I0313 14:26:47.283968 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5a33722749134dbbc6eb786e1dd47e935db2b7e8fc7b1822f79e1da556bcd6c2" Mar 13 14:26:47 crc kubenswrapper[4907]: I0313 14:26:47.284106 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-b501-account-create-update-dzlqr" Mar 13 14:26:47 crc kubenswrapper[4907]: I0313 14:26:47.285689 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-htpvb" event={"ID":"a89525da-db62-49c2-b548-bab4c974dfdd","Type":"ContainerDied","Data":"265dfd02ff62990e5c06a12bdcd62f177d8601a264d83ff1c46d7bdee007b20d"} Mar 13 14:26:47 crc kubenswrapper[4907]: I0313 14:26:47.285711 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="265dfd02ff62990e5c06a12bdcd62f177d8601a264d83ff1c46d7bdee007b20d" Mar 13 14:26:47 crc kubenswrapper[4907]: I0313 14:26:47.285783 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-htpvb" Mar 13 14:26:47 crc kubenswrapper[4907]: I0313 14:26:47.294364 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-vbcql" event={"ID":"e1068e89-10df-4d4c-a213-d9e9643fab38","Type":"ContainerDied","Data":"800ea320e37f9ac3fbdc6c79cbb80396598682384d93ce308ede98d248a09cc4"} Mar 13 14:26:47 crc kubenswrapper[4907]: I0313 14:26:47.294453 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="800ea320e37f9ac3fbdc6c79cbb80396598682384d93ce308ede98d248a09cc4" Mar 13 14:26:47 crc kubenswrapper[4907]: I0313 14:26:47.294390 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-vbcql" Mar 13 14:26:47 crc kubenswrapper[4907]: I0313 14:26:47.295750 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d2gv5\" (UniqueName: \"kubernetes.io/projected/0983a1bd-a83c-4658-a405-09b8ab0d0002-kube-api-access-d2gv5\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:47 crc kubenswrapper[4907]: I0313 14:26:47.295773 4907 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0983a1bd-a83c-4658-a405-09b8ab0d0002-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:47 crc kubenswrapper[4907]: I0313 14:26:47.295787 4907 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e1068e89-10df-4d4c-a213-d9e9643fab38-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:47 crc kubenswrapper[4907]: I0313 14:26:47.295800 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-chjsd\" (UniqueName: \"kubernetes.io/projected/e1068e89-10df-4d4c-a213-d9e9643fab38-kube-api-access-chjsd\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:47 crc kubenswrapper[4907]: I0313 14:26:47.295813 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kgvgp\" (UniqueName: \"kubernetes.io/projected/3fe8116b-9e68-41e2-aad2-cdd7d5e51c83-kube-api-access-kgvgp\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:47 crc kubenswrapper[4907]: I0313 14:26:47.295829 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qv2ks\" (UniqueName: \"kubernetes.io/projected/a89525da-db62-49c2-b548-bab4c974dfdd-kube-api-access-qv2ks\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:47 crc kubenswrapper[4907]: I0313 14:26:47.295847 4907 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a89525da-db62-49c2-b548-bab4c974dfdd-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:47 crc kubenswrapper[4907]: I0313 14:26:47.295859 4907 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3fe8116b-9e68-41e2-aad2-cdd7d5e51c83-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:47 crc kubenswrapper[4907]: I0313 14:26:47.298835 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-x9b2b" Mar 13 14:26:47 crc kubenswrapper[4907]: I0313 14:26:47.298836 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-x9b2b" event={"ID":"2bab6ed6-5335-40cc-bdff-3d7461876fb8","Type":"ContainerDied","Data":"ce59465e7c0c37d5e78037bb1b277bb43931b1fc4a590e4e9d4e42dd7fdc27d3"} Mar 13 14:26:47 crc kubenswrapper[4907]: I0313 14:26:47.299827 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ce59465e7c0c37d5e78037bb1b277bb43931b1fc4a590e4e9d4e42dd7fdc27d3" Mar 13 14:26:47 crc kubenswrapper[4907]: I0313 14:26:47.317221 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-b69e-account-create-update-ptpcx" event={"ID":"0983a1bd-a83c-4658-a405-09b8ab0d0002","Type":"ContainerDied","Data":"807b6a972511f96035dc8e7a5888a9cb76102f79128af800532f9871f36ec567"} Mar 13 14:26:47 crc kubenswrapper[4907]: I0313 14:26:47.317270 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="807b6a972511f96035dc8e7a5888a9cb76102f79128af800532f9871f36ec567" Mar 13 14:26:47 crc kubenswrapper[4907]: I0313 14:26:47.317295 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-b69e-account-create-update-ptpcx" Mar 13 14:26:48 crc kubenswrapper[4907]: I0313 14:26:48.121261 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7b9fd7d84c-wkdxf" Mar 13 14:26:48 crc kubenswrapper[4907]: I0313 14:26:48.139669 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-4v4ps" Mar 13 14:26:48 crc kubenswrapper[4907]: I0313 14:26:48.212166 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lnkf5\" (UniqueName: \"kubernetes.io/projected/89cd945f-311f-47d1-982a-641f062e4f57-kube-api-access-lnkf5\") pod \"89cd945f-311f-47d1-982a-641f062e4f57\" (UID: \"89cd945f-311f-47d1-982a-641f062e4f57\") " Mar 13 14:26:48 crc kubenswrapper[4907]: I0313 14:26:48.212231 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/89cd945f-311f-47d1-982a-641f062e4f57-ovsdbserver-sb\") pod \"89cd945f-311f-47d1-982a-641f062e4f57\" (UID: \"89cd945f-311f-47d1-982a-641f062e4f57\") " Mar 13 14:26:48 crc kubenswrapper[4907]: I0313 14:26:48.212343 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/89cd945f-311f-47d1-982a-641f062e4f57-dns-svc\") pod \"89cd945f-311f-47d1-982a-641f062e4f57\" (UID: \"89cd945f-311f-47d1-982a-641f062e4f57\") " Mar 13 14:26:48 crc kubenswrapper[4907]: I0313 14:26:48.212385 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/89cd945f-311f-47d1-982a-641f062e4f57-ovsdbserver-nb\") pod \"89cd945f-311f-47d1-982a-641f062e4f57\" (UID: \"89cd945f-311f-47d1-982a-641f062e4f57\") " Mar 13 14:26:48 crc kubenswrapper[4907]: I0313 14:26:48.212448 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/89cd945f-311f-47d1-982a-641f062e4f57-config\") pod \"89cd945f-311f-47d1-982a-641f062e4f57\" (UID: \"89cd945f-311f-47d1-982a-641f062e4f57\") " Mar 13 14:26:48 crc kubenswrapper[4907]: I0313 14:26:48.229073 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/89cd945f-311f-47d1-982a-641f062e4f57-kube-api-access-lnkf5" (OuterVolumeSpecName: "kube-api-access-lnkf5") pod "89cd945f-311f-47d1-982a-641f062e4f57" (UID: "89cd945f-311f-47d1-982a-641f062e4f57"). InnerVolumeSpecName "kube-api-access-lnkf5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:26:48 crc kubenswrapper[4907]: I0313 14:26:48.256482 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/89cd945f-311f-47d1-982a-641f062e4f57-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "89cd945f-311f-47d1-982a-641f062e4f57" (UID: "89cd945f-311f-47d1-982a-641f062e4f57"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:26:48 crc kubenswrapper[4907]: I0313 14:26:48.257177 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/89cd945f-311f-47d1-982a-641f062e4f57-config" (OuterVolumeSpecName: "config") pod "89cd945f-311f-47d1-982a-641f062e4f57" (UID: "89cd945f-311f-47d1-982a-641f062e4f57"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:26:48 crc kubenswrapper[4907]: I0313 14:26:48.265075 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/89cd945f-311f-47d1-982a-641f062e4f57-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "89cd945f-311f-47d1-982a-641f062e4f57" (UID: "89cd945f-311f-47d1-982a-641f062e4f57"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:26:48 crc kubenswrapper[4907]: I0313 14:26:48.278276 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/89cd945f-311f-47d1-982a-641f062e4f57-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "89cd945f-311f-47d1-982a-641f062e4f57" (UID: "89cd945f-311f-47d1-982a-641f062e4f57"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:26:48 crc kubenswrapper[4907]: I0313 14:26:48.325710 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/02262347-eaec-462e-a3ef-58c670885d99-db-sync-config-data\") pod \"02262347-eaec-462e-a3ef-58c670885d99\" (UID: \"02262347-eaec-462e-a3ef-58c670885d99\") " Mar 13 14:26:48 crc kubenswrapper[4907]: I0313 14:26:48.325857 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02262347-eaec-462e-a3ef-58c670885d99-combined-ca-bundle\") pod \"02262347-eaec-462e-a3ef-58c670885d99\" (UID: \"02262347-eaec-462e-a3ef-58c670885d99\") " Mar 13 14:26:48 crc kubenswrapper[4907]: I0313 14:26:48.325920 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02262347-eaec-462e-a3ef-58c670885d99-config-data\") pod \"02262347-eaec-462e-a3ef-58c670885d99\" (UID: \"02262347-eaec-462e-a3ef-58c670885d99\") " Mar 13 14:26:48 crc kubenswrapper[4907]: I0313 14:26:48.326166 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-65qwz\" (UniqueName: \"kubernetes.io/projected/02262347-eaec-462e-a3ef-58c670885d99-kube-api-access-65qwz\") pod \"02262347-eaec-462e-a3ef-58c670885d99\" (UID: \"02262347-eaec-462e-a3ef-58c670885d99\") " Mar 13 14:26:48 crc kubenswrapper[4907]: I0313 14:26:48.326611 4907 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/89cd945f-311f-47d1-982a-641f062e4f57-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:48 crc kubenswrapper[4907]: I0313 14:26:48.326657 4907 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/89cd945f-311f-47d1-982a-641f062e4f57-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:48 crc kubenswrapper[4907]: I0313 14:26:48.326672 4907 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/89cd945f-311f-47d1-982a-641f062e4f57-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:48 crc kubenswrapper[4907]: I0313 14:26:48.326705 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/89cd945f-311f-47d1-982a-641f062e4f57-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:48 crc kubenswrapper[4907]: I0313 14:26:48.326739 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lnkf5\" (UniqueName: \"kubernetes.io/projected/89cd945f-311f-47d1-982a-641f062e4f57-kube-api-access-lnkf5\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:48 crc kubenswrapper[4907]: I0313 14:26:48.327636 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b9fd7d84c-wkdxf" event={"ID":"89cd945f-311f-47d1-982a-641f062e4f57","Type":"ContainerDied","Data":"b45d10603df2d7ee66cbd50a38052eed85880ecd3d8f94e817d4236b2c83c476"} Mar 13 14:26:48 crc kubenswrapper[4907]: I0313 14:26:48.327681 4907 scope.go:117] "RemoveContainer" containerID="ad9755b55eba4bb5e259a681410af3ffba2c66af2583b34cd43ac29178757a96" Mar 13 14:26:48 crc kubenswrapper[4907]: I0313 14:26:48.327816 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7b9fd7d84c-wkdxf" Mar 13 14:26:48 crc kubenswrapper[4907]: I0313 14:26:48.329220 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02262347-eaec-462e-a3ef-58c670885d99-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "02262347-eaec-462e-a3ef-58c670885d99" (UID: "02262347-eaec-462e-a3ef-58c670885d99"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:26:48 crc kubenswrapper[4907]: I0313 14:26:48.331602 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/02262347-eaec-462e-a3ef-58c670885d99-kube-api-access-65qwz" (OuterVolumeSpecName: "kube-api-access-65qwz") pod "02262347-eaec-462e-a3ef-58c670885d99" (UID: "02262347-eaec-462e-a3ef-58c670885d99"). InnerVolumeSpecName "kube-api-access-65qwz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:26:48 crc kubenswrapper[4907]: I0313 14:26:48.334259 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-4v4ps" event={"ID":"02262347-eaec-462e-a3ef-58c670885d99","Type":"ContainerDied","Data":"e0469ec6298d86f707976b7020c89f6d29a4be70c9a77f315ceb22e64cbb4824"} Mar 13 14:26:48 crc kubenswrapper[4907]: I0313 14:26:48.334336 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e0469ec6298d86f707976b7020c89f6d29a4be70c9a77f315ceb22e64cbb4824" Mar 13 14:26:48 crc kubenswrapper[4907]: I0313 14:26:48.334350 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-4v4ps" Mar 13 14:26:48 crc kubenswrapper[4907]: I0313 14:26:48.350408 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02262347-eaec-462e-a3ef-58c670885d99-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "02262347-eaec-462e-a3ef-58c670885d99" (UID: "02262347-eaec-462e-a3ef-58c670885d99"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:26:48 crc kubenswrapper[4907]: I0313 14:26:48.365619 4907 scope.go:117] "RemoveContainer" containerID="b1318d08ab870fbe30d1438d26843f78b5e12e9bf7ba05a68e44fb095db6996e" Mar 13 14:26:48 crc kubenswrapper[4907]: I0313 14:26:48.367080 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7b9fd7d84c-wkdxf"] Mar 13 14:26:48 crc kubenswrapper[4907]: I0313 14:26:48.369162 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02262347-eaec-462e-a3ef-58c670885d99-config-data" (OuterVolumeSpecName: "config-data") pod "02262347-eaec-462e-a3ef-58c670885d99" (UID: "02262347-eaec-462e-a3ef-58c670885d99"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:26:48 crc kubenswrapper[4907]: I0313 14:26:48.376563 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7b9fd7d84c-wkdxf"] Mar 13 14:26:48 crc kubenswrapper[4907]: I0313 14:26:48.428371 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-65qwz\" (UniqueName: \"kubernetes.io/projected/02262347-eaec-462e-a3ef-58c670885d99-kube-api-access-65qwz\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:48 crc kubenswrapper[4907]: I0313 14:26:48.428632 4907 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/02262347-eaec-462e-a3ef-58c670885d99-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:48 crc kubenswrapper[4907]: I0313 14:26:48.428715 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02262347-eaec-462e-a3ef-58c670885d99-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:48 crc kubenswrapper[4907]: I0313 14:26:48.428799 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02262347-eaec-462e-a3ef-58c670885d99-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:49 crc kubenswrapper[4907]: I0313 14:26:49.346554 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-kcjn8" event={"ID":"480d23fe-2980-46b3-9e86-e74f2d051357","Type":"ContainerStarted","Data":"fad76ed01fe4f9787942e4f47ff8ffafe18d1d8c62a2533bd080f0d2d246c933"} Mar 13 14:26:49 crc kubenswrapper[4907]: I0313 14:26:49.366343 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-kcjn8" podStartSLOduration=2.737317143 podStartE2EDuration="9.366322105s" podCreationTimestamp="2026-03-13 14:26:40 +0000 UTC" firstStartedPulling="2026-03-13 14:26:41.671021732 +0000 UTC m=+1300.570809421" lastFinishedPulling="2026-03-13 14:26:48.300026694 +0000 UTC m=+1307.199814383" observedRunningTime="2026-03-13 14:26:49.363072176 +0000 UTC m=+1308.262859885" watchObservedRunningTime="2026-03-13 14:26:49.366322105 +0000 UTC m=+1308.266109814" Mar 13 14:26:49 crc kubenswrapper[4907]: I0313 14:26:49.676090 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6f88567fd9-f24rf"] Mar 13 14:26:49 crc kubenswrapper[4907]: E0313 14:26:49.676672 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89cd945f-311f-47d1-982a-641f062e4f57" containerName="init" Mar 13 14:26:49 crc kubenswrapper[4907]: I0313 14:26:49.676689 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="89cd945f-311f-47d1-982a-641f062e4f57" containerName="init" Mar 13 14:26:49 crc kubenswrapper[4907]: E0313 14:26:49.676697 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a89525da-db62-49c2-b548-bab4c974dfdd" containerName="mariadb-database-create" Mar 13 14:26:49 crc kubenswrapper[4907]: I0313 14:26:49.676703 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="a89525da-db62-49c2-b548-bab4c974dfdd" containerName="mariadb-database-create" Mar 13 14:26:49 crc kubenswrapper[4907]: E0313 14:26:49.676717 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2bab6ed6-5335-40cc-bdff-3d7461876fb8" containerName="mariadb-database-create" Mar 13 14:26:49 crc kubenswrapper[4907]: I0313 14:26:49.676725 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="2bab6ed6-5335-40cc-bdff-3d7461876fb8" containerName="mariadb-database-create" Mar 13 14:26:49 crc kubenswrapper[4907]: E0313 14:26:49.676735 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98748d4d-f934-49ed-91d9-7a9e467b2753" containerName="mariadb-account-create-update" Mar 13 14:26:49 crc kubenswrapper[4907]: I0313 14:26:49.676741 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="98748d4d-f934-49ed-91d9-7a9e467b2753" containerName="mariadb-account-create-update" Mar 13 14:26:49 crc kubenswrapper[4907]: E0313 14:26:49.676753 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0983a1bd-a83c-4658-a405-09b8ab0d0002" containerName="mariadb-account-create-update" Mar 13 14:26:49 crc kubenswrapper[4907]: I0313 14:26:49.676759 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="0983a1bd-a83c-4658-a405-09b8ab0d0002" containerName="mariadb-account-create-update" Mar 13 14:26:49 crc kubenswrapper[4907]: E0313 14:26:49.676770 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89cd945f-311f-47d1-982a-641f062e4f57" containerName="dnsmasq-dns" Mar 13 14:26:49 crc kubenswrapper[4907]: I0313 14:26:49.676776 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="89cd945f-311f-47d1-982a-641f062e4f57" containerName="dnsmasq-dns" Mar 13 14:26:49 crc kubenswrapper[4907]: E0313 14:26:49.676787 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1068e89-10df-4d4c-a213-d9e9643fab38" containerName="mariadb-database-create" Mar 13 14:26:49 crc kubenswrapper[4907]: I0313 14:26:49.676793 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1068e89-10df-4d4c-a213-d9e9643fab38" containerName="mariadb-database-create" Mar 13 14:26:49 crc kubenswrapper[4907]: E0313 14:26:49.676805 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02262347-eaec-462e-a3ef-58c670885d99" containerName="glance-db-sync" Mar 13 14:26:49 crc kubenswrapper[4907]: I0313 14:26:49.676811 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="02262347-eaec-462e-a3ef-58c670885d99" containerName="glance-db-sync" Mar 13 14:26:49 crc kubenswrapper[4907]: E0313 14:26:49.676831 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3fe8116b-9e68-41e2-aad2-cdd7d5e51c83" containerName="mariadb-account-create-update" Mar 13 14:26:49 crc kubenswrapper[4907]: I0313 14:26:49.676837 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="3fe8116b-9e68-41e2-aad2-cdd7d5e51c83" containerName="mariadb-account-create-update" Mar 13 14:26:49 crc kubenswrapper[4907]: I0313 14:26:49.677025 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="89cd945f-311f-47d1-982a-641f062e4f57" containerName="dnsmasq-dns" Mar 13 14:26:49 crc kubenswrapper[4907]: I0313 14:26:49.677036 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="a89525da-db62-49c2-b548-bab4c974dfdd" containerName="mariadb-database-create" Mar 13 14:26:49 crc kubenswrapper[4907]: I0313 14:26:49.677048 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="0983a1bd-a83c-4658-a405-09b8ab0d0002" containerName="mariadb-account-create-update" Mar 13 14:26:49 crc kubenswrapper[4907]: I0313 14:26:49.677059 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="98748d4d-f934-49ed-91d9-7a9e467b2753" containerName="mariadb-account-create-update" Mar 13 14:26:49 crc kubenswrapper[4907]: I0313 14:26:49.677073 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="3fe8116b-9e68-41e2-aad2-cdd7d5e51c83" containerName="mariadb-account-create-update" Mar 13 14:26:49 crc kubenswrapper[4907]: I0313 14:26:49.677083 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="2bab6ed6-5335-40cc-bdff-3d7461876fb8" containerName="mariadb-database-create" Mar 13 14:26:49 crc kubenswrapper[4907]: I0313 14:26:49.677093 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="02262347-eaec-462e-a3ef-58c670885d99" containerName="glance-db-sync" Mar 13 14:26:49 crc kubenswrapper[4907]: I0313 14:26:49.677100 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1068e89-10df-4d4c-a213-d9e9643fab38" containerName="mariadb-database-create" Mar 13 14:26:49 crc kubenswrapper[4907]: I0313 14:26:49.677908 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f88567fd9-f24rf" Mar 13 14:26:49 crc kubenswrapper[4907]: I0313 14:26:49.695106 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6f88567fd9-f24rf"] Mar 13 14:26:49 crc kubenswrapper[4907]: I0313 14:26:49.792908 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="89cd945f-311f-47d1-982a-641f062e4f57" path="/var/lib/kubelet/pods/89cd945f-311f-47d1-982a-641f062e4f57/volumes" Mar 13 14:26:49 crc kubenswrapper[4907]: I0313 14:26:49.853115 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7aba6c70-a80a-4bce-9216-887cb7a2061c-ovsdbserver-nb\") pod \"dnsmasq-dns-6f88567fd9-f24rf\" (UID: \"7aba6c70-a80a-4bce-9216-887cb7a2061c\") " pod="openstack/dnsmasq-dns-6f88567fd9-f24rf" Mar 13 14:26:49 crc kubenswrapper[4907]: I0313 14:26:49.853302 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2dk2z\" (UniqueName: \"kubernetes.io/projected/7aba6c70-a80a-4bce-9216-887cb7a2061c-kube-api-access-2dk2z\") pod \"dnsmasq-dns-6f88567fd9-f24rf\" (UID: \"7aba6c70-a80a-4bce-9216-887cb7a2061c\") " pod="openstack/dnsmasq-dns-6f88567fd9-f24rf" Mar 13 14:26:49 crc kubenswrapper[4907]: I0313 14:26:49.853473 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7aba6c70-a80a-4bce-9216-887cb7a2061c-config\") pod \"dnsmasq-dns-6f88567fd9-f24rf\" (UID: \"7aba6c70-a80a-4bce-9216-887cb7a2061c\") " pod="openstack/dnsmasq-dns-6f88567fd9-f24rf" Mar 13 14:26:49 crc kubenswrapper[4907]: I0313 14:26:49.853611 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7aba6c70-a80a-4bce-9216-887cb7a2061c-dns-svc\") pod \"dnsmasq-dns-6f88567fd9-f24rf\" (UID: \"7aba6c70-a80a-4bce-9216-887cb7a2061c\") " pod="openstack/dnsmasq-dns-6f88567fd9-f24rf" Mar 13 14:26:49 crc kubenswrapper[4907]: I0313 14:26:49.853724 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7aba6c70-a80a-4bce-9216-887cb7a2061c-ovsdbserver-sb\") pod \"dnsmasq-dns-6f88567fd9-f24rf\" (UID: \"7aba6c70-a80a-4bce-9216-887cb7a2061c\") " pod="openstack/dnsmasq-dns-6f88567fd9-f24rf" Mar 13 14:26:49 crc kubenswrapper[4907]: I0313 14:26:49.853936 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7aba6c70-a80a-4bce-9216-887cb7a2061c-dns-swift-storage-0\") pod \"dnsmasq-dns-6f88567fd9-f24rf\" (UID: \"7aba6c70-a80a-4bce-9216-887cb7a2061c\") " pod="openstack/dnsmasq-dns-6f88567fd9-f24rf" Mar 13 14:26:49 crc kubenswrapper[4907]: I0313 14:26:49.955769 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7aba6c70-a80a-4bce-9216-887cb7a2061c-dns-svc\") pod \"dnsmasq-dns-6f88567fd9-f24rf\" (UID: \"7aba6c70-a80a-4bce-9216-887cb7a2061c\") " pod="openstack/dnsmasq-dns-6f88567fd9-f24rf" Mar 13 14:26:49 crc kubenswrapper[4907]: I0313 14:26:49.956223 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7aba6c70-a80a-4bce-9216-887cb7a2061c-ovsdbserver-sb\") pod \"dnsmasq-dns-6f88567fd9-f24rf\" (UID: \"7aba6c70-a80a-4bce-9216-887cb7a2061c\") " pod="openstack/dnsmasq-dns-6f88567fd9-f24rf" Mar 13 14:26:49 crc kubenswrapper[4907]: I0313 14:26:49.956365 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7aba6c70-a80a-4bce-9216-887cb7a2061c-dns-swift-storage-0\") pod \"dnsmasq-dns-6f88567fd9-f24rf\" (UID: \"7aba6c70-a80a-4bce-9216-887cb7a2061c\") " pod="openstack/dnsmasq-dns-6f88567fd9-f24rf" Mar 13 14:26:49 crc kubenswrapper[4907]: I0313 14:26:49.956482 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7aba6c70-a80a-4bce-9216-887cb7a2061c-ovsdbserver-nb\") pod \"dnsmasq-dns-6f88567fd9-f24rf\" (UID: \"7aba6c70-a80a-4bce-9216-887cb7a2061c\") " pod="openstack/dnsmasq-dns-6f88567fd9-f24rf" Mar 13 14:26:49 crc kubenswrapper[4907]: I0313 14:26:49.956597 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2dk2z\" (UniqueName: \"kubernetes.io/projected/7aba6c70-a80a-4bce-9216-887cb7a2061c-kube-api-access-2dk2z\") pod \"dnsmasq-dns-6f88567fd9-f24rf\" (UID: \"7aba6c70-a80a-4bce-9216-887cb7a2061c\") " pod="openstack/dnsmasq-dns-6f88567fd9-f24rf" Mar 13 14:26:49 crc kubenswrapper[4907]: I0313 14:26:49.956757 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7aba6c70-a80a-4bce-9216-887cb7a2061c-config\") pod \"dnsmasq-dns-6f88567fd9-f24rf\" (UID: \"7aba6c70-a80a-4bce-9216-887cb7a2061c\") " pod="openstack/dnsmasq-dns-6f88567fd9-f24rf" Mar 13 14:26:49 crc kubenswrapper[4907]: I0313 14:26:49.956843 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7aba6c70-a80a-4bce-9216-887cb7a2061c-dns-svc\") pod \"dnsmasq-dns-6f88567fd9-f24rf\" (UID: \"7aba6c70-a80a-4bce-9216-887cb7a2061c\") " pod="openstack/dnsmasq-dns-6f88567fd9-f24rf" Mar 13 14:26:49 crc kubenswrapper[4907]: I0313 14:26:49.957429 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7aba6c70-a80a-4bce-9216-887cb7a2061c-ovsdbserver-nb\") pod \"dnsmasq-dns-6f88567fd9-f24rf\" (UID: \"7aba6c70-a80a-4bce-9216-887cb7a2061c\") " pod="openstack/dnsmasq-dns-6f88567fd9-f24rf" Mar 13 14:26:49 crc kubenswrapper[4907]: I0313 14:26:49.957456 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7aba6c70-a80a-4bce-9216-887cb7a2061c-dns-swift-storage-0\") pod \"dnsmasq-dns-6f88567fd9-f24rf\" (UID: \"7aba6c70-a80a-4bce-9216-887cb7a2061c\") " pod="openstack/dnsmasq-dns-6f88567fd9-f24rf" Mar 13 14:26:49 crc kubenswrapper[4907]: I0313 14:26:49.957794 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7aba6c70-a80a-4bce-9216-887cb7a2061c-config\") pod \"dnsmasq-dns-6f88567fd9-f24rf\" (UID: \"7aba6c70-a80a-4bce-9216-887cb7a2061c\") " pod="openstack/dnsmasq-dns-6f88567fd9-f24rf" Mar 13 14:26:49 crc kubenswrapper[4907]: I0313 14:26:49.958114 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7aba6c70-a80a-4bce-9216-887cb7a2061c-ovsdbserver-sb\") pod \"dnsmasq-dns-6f88567fd9-f24rf\" (UID: \"7aba6c70-a80a-4bce-9216-887cb7a2061c\") " pod="openstack/dnsmasq-dns-6f88567fd9-f24rf" Mar 13 14:26:50 crc kubenswrapper[4907]: I0313 14:26:50.011782 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2dk2z\" (UniqueName: \"kubernetes.io/projected/7aba6c70-a80a-4bce-9216-887cb7a2061c-kube-api-access-2dk2z\") pod \"dnsmasq-dns-6f88567fd9-f24rf\" (UID: \"7aba6c70-a80a-4bce-9216-887cb7a2061c\") " pod="openstack/dnsmasq-dns-6f88567fd9-f24rf" Mar 13 14:26:50 crc kubenswrapper[4907]: I0313 14:26:50.299492 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f88567fd9-f24rf" Mar 13 14:26:50 crc kubenswrapper[4907]: I0313 14:26:50.781404 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6f88567fd9-f24rf"] Mar 13 14:26:51 crc kubenswrapper[4907]: I0313 14:26:51.393028 4907 generic.go:334] "Generic (PLEG): container finished" podID="7aba6c70-a80a-4bce-9216-887cb7a2061c" containerID="b20445b78457705b4b60b2201ff58c281da6086f771d56be4ceb3ab98dc8b1f5" exitCode=0 Mar 13 14:26:51 crc kubenswrapper[4907]: I0313 14:26:51.393124 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f88567fd9-f24rf" event={"ID":"7aba6c70-a80a-4bce-9216-887cb7a2061c","Type":"ContainerDied","Data":"b20445b78457705b4b60b2201ff58c281da6086f771d56be4ceb3ab98dc8b1f5"} Mar 13 14:26:51 crc kubenswrapper[4907]: I0313 14:26:51.393340 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f88567fd9-f24rf" event={"ID":"7aba6c70-a80a-4bce-9216-887cb7a2061c","Type":"ContainerStarted","Data":"6f3c01688bba973a49cd38ead6f32071203a13f1c437e87590a2062f43c42d3e"} Mar 13 14:26:52 crc kubenswrapper[4907]: I0313 14:26:52.402456 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f88567fd9-f24rf" event={"ID":"7aba6c70-a80a-4bce-9216-887cb7a2061c","Type":"ContainerStarted","Data":"3fd8ece48af6f3186fe1a67ce35dde9ab752271f826eb2d327ed3e6092299012"} Mar 13 14:26:52 crc kubenswrapper[4907]: I0313 14:26:52.402854 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6f88567fd9-f24rf" Mar 13 14:26:52 crc kubenswrapper[4907]: I0313 14:26:52.426408 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6f88567fd9-f24rf" podStartSLOduration=3.42638791 podStartE2EDuration="3.42638791s" podCreationTimestamp="2026-03-13 14:26:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:26:52.424469808 +0000 UTC m=+1311.324257497" watchObservedRunningTime="2026-03-13 14:26:52.42638791 +0000 UTC m=+1311.326175599" Mar 13 14:26:54 crc kubenswrapper[4907]: I0313 14:26:54.431821 4907 generic.go:334] "Generic (PLEG): container finished" podID="480d23fe-2980-46b3-9e86-e74f2d051357" containerID="fad76ed01fe4f9787942e4f47ff8ffafe18d1d8c62a2533bd080f0d2d246c933" exitCode=0 Mar 13 14:26:54 crc kubenswrapper[4907]: I0313 14:26:54.431932 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-kcjn8" event={"ID":"480d23fe-2980-46b3-9e86-e74f2d051357","Type":"ContainerDied","Data":"fad76ed01fe4f9787942e4f47ff8ffafe18d1d8c62a2533bd080f0d2d246c933"} Mar 13 14:26:55 crc kubenswrapper[4907]: I0313 14:26:55.887999 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-kcjn8" Mar 13 14:26:56 crc kubenswrapper[4907]: I0313 14:26:56.018142 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g78kx\" (UniqueName: \"kubernetes.io/projected/480d23fe-2980-46b3-9e86-e74f2d051357-kube-api-access-g78kx\") pod \"480d23fe-2980-46b3-9e86-e74f2d051357\" (UID: \"480d23fe-2980-46b3-9e86-e74f2d051357\") " Mar 13 14:26:56 crc kubenswrapper[4907]: I0313 14:26:56.018326 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/480d23fe-2980-46b3-9e86-e74f2d051357-combined-ca-bundle\") pod \"480d23fe-2980-46b3-9e86-e74f2d051357\" (UID: \"480d23fe-2980-46b3-9e86-e74f2d051357\") " Mar 13 14:26:56 crc kubenswrapper[4907]: I0313 14:26:56.018517 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/480d23fe-2980-46b3-9e86-e74f2d051357-config-data\") pod \"480d23fe-2980-46b3-9e86-e74f2d051357\" (UID: \"480d23fe-2980-46b3-9e86-e74f2d051357\") " Mar 13 14:26:56 crc kubenswrapper[4907]: I0313 14:26:56.024212 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/480d23fe-2980-46b3-9e86-e74f2d051357-kube-api-access-g78kx" (OuterVolumeSpecName: "kube-api-access-g78kx") pod "480d23fe-2980-46b3-9e86-e74f2d051357" (UID: "480d23fe-2980-46b3-9e86-e74f2d051357"). InnerVolumeSpecName "kube-api-access-g78kx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:26:56 crc kubenswrapper[4907]: I0313 14:26:56.041336 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/480d23fe-2980-46b3-9e86-e74f2d051357-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "480d23fe-2980-46b3-9e86-e74f2d051357" (UID: "480d23fe-2980-46b3-9e86-e74f2d051357"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:26:56 crc kubenswrapper[4907]: I0313 14:26:56.062107 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/480d23fe-2980-46b3-9e86-e74f2d051357-config-data" (OuterVolumeSpecName: "config-data") pod "480d23fe-2980-46b3-9e86-e74f2d051357" (UID: "480d23fe-2980-46b3-9e86-e74f2d051357"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:26:56 crc kubenswrapper[4907]: I0313 14:26:56.120632 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g78kx\" (UniqueName: \"kubernetes.io/projected/480d23fe-2980-46b3-9e86-e74f2d051357-kube-api-access-g78kx\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:56 crc kubenswrapper[4907]: I0313 14:26:56.120666 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/480d23fe-2980-46b3-9e86-e74f2d051357-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:56 crc kubenswrapper[4907]: I0313 14:26:56.120677 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/480d23fe-2980-46b3-9e86-e74f2d051357-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:56 crc kubenswrapper[4907]: I0313 14:26:56.456331 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-kcjn8" event={"ID":"480d23fe-2980-46b3-9e86-e74f2d051357","Type":"ContainerDied","Data":"35b6cb9082e770b5ec87125f38e9a5f5d08513cff5fef490907772e3d539df90"} Mar 13 14:26:56 crc kubenswrapper[4907]: I0313 14:26:56.456381 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="35b6cb9082e770b5ec87125f38e9a5f5d08513cff5fef490907772e3d539df90" Mar 13 14:26:56 crc kubenswrapper[4907]: I0313 14:26:56.456473 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-kcjn8" Mar 13 14:26:56 crc kubenswrapper[4907]: I0313 14:26:56.763929 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-4tskm"] Mar 13 14:26:56 crc kubenswrapper[4907]: E0313 14:26:56.764358 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="480d23fe-2980-46b3-9e86-e74f2d051357" containerName="keystone-db-sync" Mar 13 14:26:56 crc kubenswrapper[4907]: I0313 14:26:56.764373 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="480d23fe-2980-46b3-9e86-e74f2d051357" containerName="keystone-db-sync" Mar 13 14:26:56 crc kubenswrapper[4907]: I0313 14:26:56.764589 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="480d23fe-2980-46b3-9e86-e74f2d051357" containerName="keystone-db-sync" Mar 13 14:26:56 crc kubenswrapper[4907]: I0313 14:26:56.765275 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-4tskm" Mar 13 14:26:56 crc kubenswrapper[4907]: I0313 14:26:56.770938 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Mar 13 14:26:56 crc kubenswrapper[4907]: I0313 14:26:56.771304 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Mar 13 14:26:56 crc kubenswrapper[4907]: I0313 14:26:56.771360 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Mar 13 14:26:56 crc kubenswrapper[4907]: I0313 14:26:56.771453 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Mar 13 14:26:56 crc kubenswrapper[4907]: I0313 14:26:56.773313 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-p9m6m" Mar 13 14:26:56 crc kubenswrapper[4907]: I0313 14:26:56.799378 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6f88567fd9-f24rf"] Mar 13 14:26:56 crc kubenswrapper[4907]: I0313 14:26:56.799634 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6f88567fd9-f24rf" podUID="7aba6c70-a80a-4bce-9216-887cb7a2061c" containerName="dnsmasq-dns" containerID="cri-o://3fd8ece48af6f3186fe1a67ce35dde9ab752271f826eb2d327ed3e6092299012" gracePeriod=10 Mar 13 14:26:56 crc kubenswrapper[4907]: I0313 14:26:56.801995 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6f88567fd9-f24rf" Mar 13 14:26:56 crc kubenswrapper[4907]: I0313 14:26:56.816515 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-4tskm"] Mar 13 14:26:56 crc kubenswrapper[4907]: I0313 14:26:56.864577 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5cb4dcfdd7-k6pkx"] Mar 13 14:26:56 crc kubenswrapper[4907]: I0313 14:26:56.865925 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5cb4dcfdd7-k6pkx" Mar 13 14:26:56 crc kubenswrapper[4907]: I0313 14:26:56.895233 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5cb4dcfdd7-k6pkx"] Mar 13 14:26:56 crc kubenswrapper[4907]: I0313 14:26:56.937666 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4a97d1cc-3762-47fe-a8fe-ed9e95b466a4-fernet-keys\") pod \"keystone-bootstrap-4tskm\" (UID: \"4a97d1cc-3762-47fe-a8fe-ed9e95b466a4\") " pod="openstack/keystone-bootstrap-4tskm" Mar 13 14:26:56 crc kubenswrapper[4907]: I0313 14:26:56.937736 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4a97d1cc-3762-47fe-a8fe-ed9e95b466a4-scripts\") pod \"keystone-bootstrap-4tskm\" (UID: \"4a97d1cc-3762-47fe-a8fe-ed9e95b466a4\") " pod="openstack/keystone-bootstrap-4tskm" Mar 13 14:26:56 crc kubenswrapper[4907]: I0313 14:26:56.937872 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a97d1cc-3762-47fe-a8fe-ed9e95b466a4-combined-ca-bundle\") pod \"keystone-bootstrap-4tskm\" (UID: \"4a97d1cc-3762-47fe-a8fe-ed9e95b466a4\") " pod="openstack/keystone-bootstrap-4tskm" Mar 13 14:26:56 crc kubenswrapper[4907]: I0313 14:26:56.937960 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a97d1cc-3762-47fe-a8fe-ed9e95b466a4-config-data\") pod \"keystone-bootstrap-4tskm\" (UID: \"4a97d1cc-3762-47fe-a8fe-ed9e95b466a4\") " pod="openstack/keystone-bootstrap-4tskm" Mar 13 14:26:56 crc kubenswrapper[4907]: I0313 14:26:56.937992 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pvhgq\" (UniqueName: \"kubernetes.io/projected/4a97d1cc-3762-47fe-a8fe-ed9e95b466a4-kube-api-access-pvhgq\") pod \"keystone-bootstrap-4tskm\" (UID: \"4a97d1cc-3762-47fe-a8fe-ed9e95b466a4\") " pod="openstack/keystone-bootstrap-4tskm" Mar 13 14:26:56 crc kubenswrapper[4907]: I0313 14:26:56.938077 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/4a97d1cc-3762-47fe-a8fe-ed9e95b466a4-credential-keys\") pod \"keystone-bootstrap-4tskm\" (UID: \"4a97d1cc-3762-47fe-a8fe-ed9e95b466a4\") " pod="openstack/keystone-bootstrap-4tskm" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.050861 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/4a97d1cc-3762-47fe-a8fe-ed9e95b466a4-credential-keys\") pod \"keystone-bootstrap-4tskm\" (UID: \"4a97d1cc-3762-47fe-a8fe-ed9e95b466a4\") " pod="openstack/keystone-bootstrap-4tskm" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.050956 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c495e208-654b-477a-869e-87a5c2566519-config\") pod \"dnsmasq-dns-5cb4dcfdd7-k6pkx\" (UID: \"c495e208-654b-477a-869e-87a5c2566519\") " pod="openstack/dnsmasq-dns-5cb4dcfdd7-k6pkx" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.050984 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4a97d1cc-3762-47fe-a8fe-ed9e95b466a4-fernet-keys\") pod \"keystone-bootstrap-4tskm\" (UID: \"4a97d1cc-3762-47fe-a8fe-ed9e95b466a4\") " pod="openstack/keystone-bootstrap-4tskm" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.051015 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4a97d1cc-3762-47fe-a8fe-ed9e95b466a4-scripts\") pod \"keystone-bootstrap-4tskm\" (UID: \"4a97d1cc-3762-47fe-a8fe-ed9e95b466a4\") " pod="openstack/keystone-bootstrap-4tskm" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.051083 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a97d1cc-3762-47fe-a8fe-ed9e95b466a4-combined-ca-bundle\") pod \"keystone-bootstrap-4tskm\" (UID: \"4a97d1cc-3762-47fe-a8fe-ed9e95b466a4\") " pod="openstack/keystone-bootstrap-4tskm" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.051108 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c495e208-654b-477a-869e-87a5c2566519-ovsdbserver-nb\") pod \"dnsmasq-dns-5cb4dcfdd7-k6pkx\" (UID: \"c495e208-654b-477a-869e-87a5c2566519\") " pod="openstack/dnsmasq-dns-5cb4dcfdd7-k6pkx" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.051135 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c495e208-654b-477a-869e-87a5c2566519-dns-swift-storage-0\") pod \"dnsmasq-dns-5cb4dcfdd7-k6pkx\" (UID: \"c495e208-654b-477a-869e-87a5c2566519\") " pod="openstack/dnsmasq-dns-5cb4dcfdd7-k6pkx" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.051172 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a97d1cc-3762-47fe-a8fe-ed9e95b466a4-config-data\") pod \"keystone-bootstrap-4tskm\" (UID: \"4a97d1cc-3762-47fe-a8fe-ed9e95b466a4\") " pod="openstack/keystone-bootstrap-4tskm" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.051193 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-db278\" (UniqueName: \"kubernetes.io/projected/c495e208-654b-477a-869e-87a5c2566519-kube-api-access-db278\") pod \"dnsmasq-dns-5cb4dcfdd7-k6pkx\" (UID: \"c495e208-654b-477a-869e-87a5c2566519\") " pod="openstack/dnsmasq-dns-5cb4dcfdd7-k6pkx" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.051219 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c495e208-654b-477a-869e-87a5c2566519-ovsdbserver-sb\") pod \"dnsmasq-dns-5cb4dcfdd7-k6pkx\" (UID: \"c495e208-654b-477a-869e-87a5c2566519\") " pod="openstack/dnsmasq-dns-5cb4dcfdd7-k6pkx" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.051241 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c495e208-654b-477a-869e-87a5c2566519-dns-svc\") pod \"dnsmasq-dns-5cb4dcfdd7-k6pkx\" (UID: \"c495e208-654b-477a-869e-87a5c2566519\") " pod="openstack/dnsmasq-dns-5cb4dcfdd7-k6pkx" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.051261 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pvhgq\" (UniqueName: \"kubernetes.io/projected/4a97d1cc-3762-47fe-a8fe-ed9e95b466a4-kube-api-access-pvhgq\") pod \"keystone-bootstrap-4tskm\" (UID: \"4a97d1cc-3762-47fe-a8fe-ed9e95b466a4\") " pod="openstack/keystone-bootstrap-4tskm" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.059661 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a97d1cc-3762-47fe-a8fe-ed9e95b466a4-combined-ca-bundle\") pod \"keystone-bootstrap-4tskm\" (UID: \"4a97d1cc-3762-47fe-a8fe-ed9e95b466a4\") " pod="openstack/keystone-bootstrap-4tskm" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.062234 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a97d1cc-3762-47fe-a8fe-ed9e95b466a4-config-data\") pod \"keystone-bootstrap-4tskm\" (UID: \"4a97d1cc-3762-47fe-a8fe-ed9e95b466a4\") " pod="openstack/keystone-bootstrap-4tskm" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.062742 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4a97d1cc-3762-47fe-a8fe-ed9e95b466a4-fernet-keys\") pod \"keystone-bootstrap-4tskm\" (UID: \"4a97d1cc-3762-47fe-a8fe-ed9e95b466a4\") " pod="openstack/keystone-bootstrap-4tskm" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.070404 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4a97d1cc-3762-47fe-a8fe-ed9e95b466a4-scripts\") pod \"keystone-bootstrap-4tskm\" (UID: \"4a97d1cc-3762-47fe-a8fe-ed9e95b466a4\") " pod="openstack/keystone-bootstrap-4tskm" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.080679 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.085051 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pvhgq\" (UniqueName: \"kubernetes.io/projected/4a97d1cc-3762-47fe-a8fe-ed9e95b466a4-kube-api-access-pvhgq\") pod \"keystone-bootstrap-4tskm\" (UID: \"4a97d1cc-3762-47fe-a8fe-ed9e95b466a4\") " pod="openstack/keystone-bootstrap-4tskm" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.087287 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/4a97d1cc-3762-47fe-a8fe-ed9e95b466a4-credential-keys\") pod \"keystone-bootstrap-4tskm\" (UID: \"4a97d1cc-3762-47fe-a8fe-ed9e95b466a4\") " pod="openstack/keystone-bootstrap-4tskm" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.087761 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.114311 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.114627 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.152785 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c495e208-654b-477a-869e-87a5c2566519-config\") pod \"dnsmasq-dns-5cb4dcfdd7-k6pkx\" (UID: \"c495e208-654b-477a-869e-87a5c2566519\") " pod="openstack/dnsmasq-dns-5cb4dcfdd7-k6pkx" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.153778 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c495e208-654b-477a-869e-87a5c2566519-config\") pod \"dnsmasq-dns-5cb4dcfdd7-k6pkx\" (UID: \"c495e208-654b-477a-869e-87a5c2566519\") " pod="openstack/dnsmasq-dns-5cb4dcfdd7-k6pkx" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.153838 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c495e208-654b-477a-869e-87a5c2566519-ovsdbserver-nb\") pod \"dnsmasq-dns-5cb4dcfdd7-k6pkx\" (UID: \"c495e208-654b-477a-869e-87a5c2566519\") " pod="openstack/dnsmasq-dns-5cb4dcfdd7-k6pkx" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.153864 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c495e208-654b-477a-869e-87a5c2566519-dns-swift-storage-0\") pod \"dnsmasq-dns-5cb4dcfdd7-k6pkx\" (UID: \"c495e208-654b-477a-869e-87a5c2566519\") " pod="openstack/dnsmasq-dns-5cb4dcfdd7-k6pkx" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.153923 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-db278\" (UniqueName: \"kubernetes.io/projected/c495e208-654b-477a-869e-87a5c2566519-kube-api-access-db278\") pod \"dnsmasq-dns-5cb4dcfdd7-k6pkx\" (UID: \"c495e208-654b-477a-869e-87a5c2566519\") " pod="openstack/dnsmasq-dns-5cb4dcfdd7-k6pkx" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.153946 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c495e208-654b-477a-869e-87a5c2566519-ovsdbserver-sb\") pod \"dnsmasq-dns-5cb4dcfdd7-k6pkx\" (UID: \"c495e208-654b-477a-869e-87a5c2566519\") " pod="openstack/dnsmasq-dns-5cb4dcfdd7-k6pkx" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.153968 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c495e208-654b-477a-869e-87a5c2566519-dns-svc\") pod \"dnsmasq-dns-5cb4dcfdd7-k6pkx\" (UID: \"c495e208-654b-477a-869e-87a5c2566519\") " pod="openstack/dnsmasq-dns-5cb4dcfdd7-k6pkx" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.153996 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c495e208-654b-477a-869e-87a5c2566519-ovsdbserver-nb\") pod \"dnsmasq-dns-5cb4dcfdd7-k6pkx\" (UID: \"c495e208-654b-477a-869e-87a5c2566519\") " pod="openstack/dnsmasq-dns-5cb4dcfdd7-k6pkx" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.154668 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c495e208-654b-477a-869e-87a5c2566519-ovsdbserver-sb\") pod \"dnsmasq-dns-5cb4dcfdd7-k6pkx\" (UID: \"c495e208-654b-477a-869e-87a5c2566519\") " pod="openstack/dnsmasq-dns-5cb4dcfdd7-k6pkx" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.154837 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c495e208-654b-477a-869e-87a5c2566519-dns-swift-storage-0\") pod \"dnsmasq-dns-5cb4dcfdd7-k6pkx\" (UID: \"c495e208-654b-477a-869e-87a5c2566519\") " pod="openstack/dnsmasq-dns-5cb4dcfdd7-k6pkx" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.155212 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c495e208-654b-477a-869e-87a5c2566519-dns-svc\") pod \"dnsmasq-dns-5cb4dcfdd7-k6pkx\" (UID: \"c495e208-654b-477a-869e-87a5c2566519\") " pod="openstack/dnsmasq-dns-5cb4dcfdd7-k6pkx" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.180020 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.210907 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-db278\" (UniqueName: \"kubernetes.io/projected/c495e208-654b-477a-869e-87a5c2566519-kube-api-access-db278\") pod \"dnsmasq-dns-5cb4dcfdd7-k6pkx\" (UID: \"c495e208-654b-477a-869e-87a5c2566519\") " pod="openstack/dnsmasq-dns-5cb4dcfdd7-k6pkx" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.212495 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-7q999"] Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.213483 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-7q999" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.220372 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5cb4dcfdd7-k6pkx" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.222168 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-7q999"] Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.228860 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-rz6qh"] Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.230091 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-rz6qh" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.235686 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.235987 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.236174 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-skqjb" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.236391 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-d7phq" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.236440 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.241372 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-jt88g"] Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.242840 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-jt88g" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.248127 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.248357 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-tvcdf" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.248505 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.251570 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-rz6qh"] Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.256001 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-jt88g"] Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.256812 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5-run-httpd\") pod \"ceilometer-0\" (UID: \"58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5\") " pod="openstack/ceilometer-0" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.256852 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2xbmw\" (UniqueName: \"kubernetes.io/projected/58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5-kube-api-access-2xbmw\") pod \"ceilometer-0\" (UID: \"58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5\") " pod="openstack/ceilometer-0" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.256872 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5-log-httpd\") pod \"ceilometer-0\" (UID: \"58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5\") " pod="openstack/ceilometer-0" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.256919 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5\") " pod="openstack/ceilometer-0" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.256950 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5-scripts\") pod \"ceilometer-0\" (UID: \"58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5\") " pod="openstack/ceilometer-0" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.256988 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5\") " pod="openstack/ceilometer-0" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.257004 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5-config-data\") pod \"ceilometer-0\" (UID: \"58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5\") " pod="openstack/ceilometer-0" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.304677 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-jjqr9"] Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.309346 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-jjqr9" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.314453 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5cb4dcfdd7-k6pkx"] Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.323985 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.324080 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.324347 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-pg28b" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.332460 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-jjqr9"] Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.358022 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2xbmw\" (UniqueName: \"kubernetes.io/projected/58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5-kube-api-access-2xbmw\") pod \"ceilometer-0\" (UID: \"58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5\") " pod="openstack/ceilometer-0" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.358070 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/870ec55c-fc6f-4b16-8c49-a3cda0d0d010-combined-ca-bundle\") pod \"neutron-db-sync-jt88g\" (UID: \"870ec55c-fc6f-4b16-8c49-a3cda0d0d010\") " pod="openstack/neutron-db-sync-jt88g" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.358094 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5-log-httpd\") pod \"ceilometer-0\" (UID: \"58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5\") " pod="openstack/ceilometer-0" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.358119 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/870ec55c-fc6f-4b16-8c49-a3cda0d0d010-config\") pod \"neutron-db-sync-jt88g\" (UID: \"870ec55c-fc6f-4b16-8c49-a3cda0d0d010\") " pod="openstack/neutron-db-sync-jt88g" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.358144 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-578zc\" (UniqueName: \"kubernetes.io/projected/32b531c6-64c1-4137-b82e-3c10789325e6-kube-api-access-578zc\") pod \"barbican-db-sync-rz6qh\" (UID: \"32b531c6-64c1-4137-b82e-3c10789325e6\") " pod="openstack/barbican-db-sync-rz6qh" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.358169 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5\") " pod="openstack/ceilometer-0" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.358201 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wq68t\" (UniqueName: \"kubernetes.io/projected/870ec55c-fc6f-4b16-8c49-a3cda0d0d010-kube-api-access-wq68t\") pod \"neutron-db-sync-jt88g\" (UID: \"870ec55c-fc6f-4b16-8c49-a3cda0d0d010\") " pod="openstack/neutron-db-sync-jt88g" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.358219 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5-scripts\") pod \"ceilometer-0\" (UID: \"58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5\") " pod="openstack/ceilometer-0" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.358238 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/594b61fa-12b7-47b6-8af3-17f024e0d54d-db-sync-config-data\") pod \"cinder-db-sync-7q999\" (UID: \"594b61fa-12b7-47b6-8af3-17f024e0d54d\") " pod="openstack/cinder-db-sync-7q999" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.358258 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wpft2\" (UniqueName: \"kubernetes.io/projected/594b61fa-12b7-47b6-8af3-17f024e0d54d-kube-api-access-wpft2\") pod \"cinder-db-sync-7q999\" (UID: \"594b61fa-12b7-47b6-8af3-17f024e0d54d\") " pod="openstack/cinder-db-sync-7q999" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.358293 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/594b61fa-12b7-47b6-8af3-17f024e0d54d-config-data\") pod \"cinder-db-sync-7q999\" (UID: \"594b61fa-12b7-47b6-8af3-17f024e0d54d\") " pod="openstack/cinder-db-sync-7q999" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.358311 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5\") " pod="openstack/ceilometer-0" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.358331 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5-config-data\") pod \"ceilometer-0\" (UID: \"58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5\") " pod="openstack/ceilometer-0" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.358358 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/594b61fa-12b7-47b6-8af3-17f024e0d54d-scripts\") pod \"cinder-db-sync-7q999\" (UID: \"594b61fa-12b7-47b6-8af3-17f024e0d54d\") " pod="openstack/cinder-db-sync-7q999" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.358394 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5-run-httpd\") pod \"ceilometer-0\" (UID: \"58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5\") " pod="openstack/ceilometer-0" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.358414 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32b531c6-64c1-4137-b82e-3c10789325e6-combined-ca-bundle\") pod \"barbican-db-sync-rz6qh\" (UID: \"32b531c6-64c1-4137-b82e-3c10789325e6\") " pod="openstack/barbican-db-sync-rz6qh" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.358430 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/594b61fa-12b7-47b6-8af3-17f024e0d54d-combined-ca-bundle\") pod \"cinder-db-sync-7q999\" (UID: \"594b61fa-12b7-47b6-8af3-17f024e0d54d\") " pod="openstack/cinder-db-sync-7q999" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.358447 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/594b61fa-12b7-47b6-8af3-17f024e0d54d-etc-machine-id\") pod \"cinder-db-sync-7q999\" (UID: \"594b61fa-12b7-47b6-8af3-17f024e0d54d\") " pod="openstack/cinder-db-sync-7q999" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.358474 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/32b531c6-64c1-4137-b82e-3c10789325e6-db-sync-config-data\") pod \"barbican-db-sync-rz6qh\" (UID: \"32b531c6-64c1-4137-b82e-3c10789325e6\") " pod="openstack/barbican-db-sync-rz6qh" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.359372 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5-log-httpd\") pod \"ceilometer-0\" (UID: \"58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5\") " pod="openstack/ceilometer-0" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.366172 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5-run-httpd\") pod \"ceilometer-0\" (UID: \"58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5\") " pod="openstack/ceilometer-0" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.369345 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5\") " pod="openstack/ceilometer-0" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.371813 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5-scripts\") pod \"ceilometer-0\" (UID: \"58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5\") " pod="openstack/ceilometer-0" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.372794 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5-config-data\") pod \"ceilometer-0\" (UID: \"58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5\") " pod="openstack/ceilometer-0" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.374415 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5\") " pod="openstack/ceilometer-0" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.383777 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-4tskm" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.408520 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2xbmw\" (UniqueName: \"kubernetes.io/projected/58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5-kube-api-access-2xbmw\") pod \"ceilometer-0\" (UID: \"58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5\") " pod="openstack/ceilometer-0" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.434248 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-759cc7f497-7nv78"] Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.435600 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-759cc7f497-7nv78" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.458567 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-759cc7f497-7nv78"] Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.459286 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/32b531c6-64c1-4137-b82e-3c10789325e6-db-sync-config-data\") pod \"barbican-db-sync-rz6qh\" (UID: \"32b531c6-64c1-4137-b82e-3c10789325e6\") " pod="openstack/barbican-db-sync-rz6qh" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.459309 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/870ec55c-fc6f-4b16-8c49-a3cda0d0d010-combined-ca-bundle\") pod \"neutron-db-sync-jt88g\" (UID: \"870ec55c-fc6f-4b16-8c49-a3cda0d0d010\") " pod="openstack/neutron-db-sync-jt88g" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.459334 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b144897c-8640-4c5c-afa1-68b44dfd4f98-config-data\") pod \"placement-db-sync-jjqr9\" (UID: \"b144897c-8640-4c5c-afa1-68b44dfd4f98\") " pod="openstack/placement-db-sync-jjqr9" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.459358 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/870ec55c-fc6f-4b16-8c49-a3cda0d0d010-config\") pod \"neutron-db-sync-jt88g\" (UID: \"870ec55c-fc6f-4b16-8c49-a3cda0d0d010\") " pod="openstack/neutron-db-sync-jt88g" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.459380 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-578zc\" (UniqueName: \"kubernetes.io/projected/32b531c6-64c1-4137-b82e-3c10789325e6-kube-api-access-578zc\") pod \"barbican-db-sync-rz6qh\" (UID: \"32b531c6-64c1-4137-b82e-3c10789325e6\") " pod="openstack/barbican-db-sync-rz6qh" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.459404 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b144897c-8640-4c5c-afa1-68b44dfd4f98-logs\") pod \"placement-db-sync-jjqr9\" (UID: \"b144897c-8640-4c5c-afa1-68b44dfd4f98\") " pod="openstack/placement-db-sync-jjqr9" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.459437 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wq68t\" (UniqueName: \"kubernetes.io/projected/870ec55c-fc6f-4b16-8c49-a3cda0d0d010-kube-api-access-wq68t\") pod \"neutron-db-sync-jt88g\" (UID: \"870ec55c-fc6f-4b16-8c49-a3cda0d0d010\") " pod="openstack/neutron-db-sync-jt88g" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.459457 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b144897c-8640-4c5c-afa1-68b44dfd4f98-combined-ca-bundle\") pod \"placement-db-sync-jjqr9\" (UID: \"b144897c-8640-4c5c-afa1-68b44dfd4f98\") " pod="openstack/placement-db-sync-jjqr9" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.459474 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/594b61fa-12b7-47b6-8af3-17f024e0d54d-db-sync-config-data\") pod \"cinder-db-sync-7q999\" (UID: \"594b61fa-12b7-47b6-8af3-17f024e0d54d\") " pod="openstack/cinder-db-sync-7q999" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.459496 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wpft2\" (UniqueName: \"kubernetes.io/projected/594b61fa-12b7-47b6-8af3-17f024e0d54d-kube-api-access-wpft2\") pod \"cinder-db-sync-7q999\" (UID: \"594b61fa-12b7-47b6-8af3-17f024e0d54d\") " pod="openstack/cinder-db-sync-7q999" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.459530 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b144897c-8640-4c5c-afa1-68b44dfd4f98-scripts\") pod \"placement-db-sync-jjqr9\" (UID: \"b144897c-8640-4c5c-afa1-68b44dfd4f98\") " pod="openstack/placement-db-sync-jjqr9" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.459547 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/594b61fa-12b7-47b6-8af3-17f024e0d54d-config-data\") pod \"cinder-db-sync-7q999\" (UID: \"594b61fa-12b7-47b6-8af3-17f024e0d54d\") " pod="openstack/cinder-db-sync-7q999" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.459573 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/594b61fa-12b7-47b6-8af3-17f024e0d54d-scripts\") pod \"cinder-db-sync-7q999\" (UID: \"594b61fa-12b7-47b6-8af3-17f024e0d54d\") " pod="openstack/cinder-db-sync-7q999" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.459631 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wcv67\" (UniqueName: \"kubernetes.io/projected/b144897c-8640-4c5c-afa1-68b44dfd4f98-kube-api-access-wcv67\") pod \"placement-db-sync-jjqr9\" (UID: \"b144897c-8640-4c5c-afa1-68b44dfd4f98\") " pod="openstack/placement-db-sync-jjqr9" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.459657 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32b531c6-64c1-4137-b82e-3c10789325e6-combined-ca-bundle\") pod \"barbican-db-sync-rz6qh\" (UID: \"32b531c6-64c1-4137-b82e-3c10789325e6\") " pod="openstack/barbican-db-sync-rz6qh" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.459674 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/594b61fa-12b7-47b6-8af3-17f024e0d54d-combined-ca-bundle\") pod \"cinder-db-sync-7q999\" (UID: \"594b61fa-12b7-47b6-8af3-17f024e0d54d\") " pod="openstack/cinder-db-sync-7q999" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.459689 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/594b61fa-12b7-47b6-8af3-17f024e0d54d-etc-machine-id\") pod \"cinder-db-sync-7q999\" (UID: \"594b61fa-12b7-47b6-8af3-17f024e0d54d\") " pod="openstack/cinder-db-sync-7q999" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.459763 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/594b61fa-12b7-47b6-8af3-17f024e0d54d-etc-machine-id\") pod \"cinder-db-sync-7q999\" (UID: \"594b61fa-12b7-47b6-8af3-17f024e0d54d\") " pod="openstack/cinder-db-sync-7q999" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.471224 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32b531c6-64c1-4137-b82e-3c10789325e6-combined-ca-bundle\") pod \"barbican-db-sync-rz6qh\" (UID: \"32b531c6-64c1-4137-b82e-3c10789325e6\") " pod="openstack/barbican-db-sync-rz6qh" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.473816 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/594b61fa-12b7-47b6-8af3-17f024e0d54d-scripts\") pod \"cinder-db-sync-7q999\" (UID: \"594b61fa-12b7-47b6-8af3-17f024e0d54d\") " pod="openstack/cinder-db-sync-7q999" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.477749 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/594b61fa-12b7-47b6-8af3-17f024e0d54d-config-data\") pod \"cinder-db-sync-7q999\" (UID: \"594b61fa-12b7-47b6-8af3-17f024e0d54d\") " pod="openstack/cinder-db-sync-7q999" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.492349 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/870ec55c-fc6f-4b16-8c49-a3cda0d0d010-combined-ca-bundle\") pod \"neutron-db-sync-jt88g\" (UID: \"870ec55c-fc6f-4b16-8c49-a3cda0d0d010\") " pod="openstack/neutron-db-sync-jt88g" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.507687 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/870ec55c-fc6f-4b16-8c49-a3cda0d0d010-config\") pod \"neutron-db-sync-jt88g\" (UID: \"870ec55c-fc6f-4b16-8c49-a3cda0d0d010\") " pod="openstack/neutron-db-sync-jt88g" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.512513 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/594b61fa-12b7-47b6-8af3-17f024e0d54d-combined-ca-bundle\") pod \"cinder-db-sync-7q999\" (UID: \"594b61fa-12b7-47b6-8af3-17f024e0d54d\") " pod="openstack/cinder-db-sync-7q999" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.512922 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.515376 4907 generic.go:334] "Generic (PLEG): container finished" podID="7aba6c70-a80a-4bce-9216-887cb7a2061c" containerID="3fd8ece48af6f3186fe1a67ce35dde9ab752271f826eb2d327ed3e6092299012" exitCode=0 Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.515898 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/594b61fa-12b7-47b6-8af3-17f024e0d54d-db-sync-config-data\") pod \"cinder-db-sync-7q999\" (UID: \"594b61fa-12b7-47b6-8af3-17f024e0d54d\") " pod="openstack/cinder-db-sync-7q999" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.515893 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f88567fd9-f24rf" event={"ID":"7aba6c70-a80a-4bce-9216-887cb7a2061c","Type":"ContainerDied","Data":"3fd8ece48af6f3186fe1a67ce35dde9ab752271f826eb2d327ed3e6092299012"} Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.516085 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-578zc\" (UniqueName: \"kubernetes.io/projected/32b531c6-64c1-4137-b82e-3c10789325e6-kube-api-access-578zc\") pod \"barbican-db-sync-rz6qh\" (UID: \"32b531c6-64c1-4137-b82e-3c10789325e6\") " pod="openstack/barbican-db-sync-rz6qh" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.523532 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/32b531c6-64c1-4137-b82e-3c10789325e6-db-sync-config-data\") pod \"barbican-db-sync-rz6qh\" (UID: \"32b531c6-64c1-4137-b82e-3c10789325e6\") " pod="openstack/barbican-db-sync-rz6qh" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.523581 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wq68t\" (UniqueName: \"kubernetes.io/projected/870ec55c-fc6f-4b16-8c49-a3cda0d0d010-kube-api-access-wq68t\") pod \"neutron-db-sync-jt88g\" (UID: \"870ec55c-fc6f-4b16-8c49-a3cda0d0d010\") " pod="openstack/neutron-db-sync-jt88g" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.539292 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wpft2\" (UniqueName: \"kubernetes.io/projected/594b61fa-12b7-47b6-8af3-17f024e0d54d-kube-api-access-wpft2\") pod \"cinder-db-sync-7q999\" (UID: \"594b61fa-12b7-47b6-8af3-17f024e0d54d\") " pod="openstack/cinder-db-sync-7q999" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.560851 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wcv67\" (UniqueName: \"kubernetes.io/projected/b144897c-8640-4c5c-afa1-68b44dfd4f98-kube-api-access-wcv67\") pod \"placement-db-sync-jjqr9\" (UID: \"b144897c-8640-4c5c-afa1-68b44dfd4f98\") " pod="openstack/placement-db-sync-jjqr9" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.560945 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a661b9e1-c551-4b36-87a9-833df73bd6b2-dns-swift-storage-0\") pod \"dnsmasq-dns-759cc7f497-7nv78\" (UID: \"a661b9e1-c551-4b36-87a9-833df73bd6b2\") " pod="openstack/dnsmasq-dns-759cc7f497-7nv78" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.560969 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b144897c-8640-4c5c-afa1-68b44dfd4f98-config-data\") pod \"placement-db-sync-jjqr9\" (UID: \"b144897c-8640-4c5c-afa1-68b44dfd4f98\") " pod="openstack/placement-db-sync-jjqr9" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.560997 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a661b9e1-c551-4b36-87a9-833df73bd6b2-config\") pod \"dnsmasq-dns-759cc7f497-7nv78\" (UID: \"a661b9e1-c551-4b36-87a9-833df73bd6b2\") " pod="openstack/dnsmasq-dns-759cc7f497-7nv78" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.561034 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b144897c-8640-4c5c-afa1-68b44dfd4f98-logs\") pod \"placement-db-sync-jjqr9\" (UID: \"b144897c-8640-4c5c-afa1-68b44dfd4f98\") " pod="openstack/placement-db-sync-jjqr9" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.561052 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ntd2v\" (UniqueName: \"kubernetes.io/projected/a661b9e1-c551-4b36-87a9-833df73bd6b2-kube-api-access-ntd2v\") pod \"dnsmasq-dns-759cc7f497-7nv78\" (UID: \"a661b9e1-c551-4b36-87a9-833df73bd6b2\") " pod="openstack/dnsmasq-dns-759cc7f497-7nv78" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.561121 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b144897c-8640-4c5c-afa1-68b44dfd4f98-combined-ca-bundle\") pod \"placement-db-sync-jjqr9\" (UID: \"b144897c-8640-4c5c-afa1-68b44dfd4f98\") " pod="openstack/placement-db-sync-jjqr9" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.561164 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a661b9e1-c551-4b36-87a9-833df73bd6b2-ovsdbserver-sb\") pod \"dnsmasq-dns-759cc7f497-7nv78\" (UID: \"a661b9e1-c551-4b36-87a9-833df73bd6b2\") " pod="openstack/dnsmasq-dns-759cc7f497-7nv78" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.561203 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b144897c-8640-4c5c-afa1-68b44dfd4f98-scripts\") pod \"placement-db-sync-jjqr9\" (UID: \"b144897c-8640-4c5c-afa1-68b44dfd4f98\") " pod="openstack/placement-db-sync-jjqr9" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.561218 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a661b9e1-c551-4b36-87a9-833df73bd6b2-ovsdbserver-nb\") pod \"dnsmasq-dns-759cc7f497-7nv78\" (UID: \"a661b9e1-c551-4b36-87a9-833df73bd6b2\") " pod="openstack/dnsmasq-dns-759cc7f497-7nv78" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.561243 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a661b9e1-c551-4b36-87a9-833df73bd6b2-dns-svc\") pod \"dnsmasq-dns-759cc7f497-7nv78\" (UID: \"a661b9e1-c551-4b36-87a9-833df73bd6b2\") " pod="openstack/dnsmasq-dns-759cc7f497-7nv78" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.567513 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b144897c-8640-4c5c-afa1-68b44dfd4f98-config-data\") pod \"placement-db-sync-jjqr9\" (UID: \"b144897c-8640-4c5c-afa1-68b44dfd4f98\") " pod="openstack/placement-db-sync-jjqr9" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.567670 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b144897c-8640-4c5c-afa1-68b44dfd4f98-logs\") pod \"placement-db-sync-jjqr9\" (UID: \"b144897c-8640-4c5c-afa1-68b44dfd4f98\") " pod="openstack/placement-db-sync-jjqr9" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.569255 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b144897c-8640-4c5c-afa1-68b44dfd4f98-scripts\") pod \"placement-db-sync-jjqr9\" (UID: \"b144897c-8640-4c5c-afa1-68b44dfd4f98\") " pod="openstack/placement-db-sync-jjqr9" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.574927 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b144897c-8640-4c5c-afa1-68b44dfd4f98-combined-ca-bundle\") pod \"placement-db-sync-jjqr9\" (UID: \"b144897c-8640-4c5c-afa1-68b44dfd4f98\") " pod="openstack/placement-db-sync-jjqr9" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.583529 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wcv67\" (UniqueName: \"kubernetes.io/projected/b144897c-8640-4c5c-afa1-68b44dfd4f98-kube-api-access-wcv67\") pod \"placement-db-sync-jjqr9\" (UID: \"b144897c-8640-4c5c-afa1-68b44dfd4f98\") " pod="openstack/placement-db-sync-jjqr9" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.594222 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f88567fd9-f24rf" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.641589 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-7q999" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.674864 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a661b9e1-c551-4b36-87a9-833df73bd6b2-ovsdbserver-nb\") pod \"dnsmasq-dns-759cc7f497-7nv78\" (UID: \"a661b9e1-c551-4b36-87a9-833df73bd6b2\") " pod="openstack/dnsmasq-dns-759cc7f497-7nv78" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.674953 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a661b9e1-c551-4b36-87a9-833df73bd6b2-dns-svc\") pod \"dnsmasq-dns-759cc7f497-7nv78\" (UID: \"a661b9e1-c551-4b36-87a9-833df73bd6b2\") " pod="openstack/dnsmasq-dns-759cc7f497-7nv78" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.675048 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a661b9e1-c551-4b36-87a9-833df73bd6b2-dns-swift-storage-0\") pod \"dnsmasq-dns-759cc7f497-7nv78\" (UID: \"a661b9e1-c551-4b36-87a9-833df73bd6b2\") " pod="openstack/dnsmasq-dns-759cc7f497-7nv78" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.675085 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a661b9e1-c551-4b36-87a9-833df73bd6b2-config\") pod \"dnsmasq-dns-759cc7f497-7nv78\" (UID: \"a661b9e1-c551-4b36-87a9-833df73bd6b2\") " pod="openstack/dnsmasq-dns-759cc7f497-7nv78" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.675130 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ntd2v\" (UniqueName: \"kubernetes.io/projected/a661b9e1-c551-4b36-87a9-833df73bd6b2-kube-api-access-ntd2v\") pod \"dnsmasq-dns-759cc7f497-7nv78\" (UID: \"a661b9e1-c551-4b36-87a9-833df73bd6b2\") " pod="openstack/dnsmasq-dns-759cc7f497-7nv78" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.675199 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a661b9e1-c551-4b36-87a9-833df73bd6b2-ovsdbserver-sb\") pod \"dnsmasq-dns-759cc7f497-7nv78\" (UID: \"a661b9e1-c551-4b36-87a9-833df73bd6b2\") " pod="openstack/dnsmasq-dns-759cc7f497-7nv78" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.676162 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a661b9e1-c551-4b36-87a9-833df73bd6b2-ovsdbserver-sb\") pod \"dnsmasq-dns-759cc7f497-7nv78\" (UID: \"a661b9e1-c551-4b36-87a9-833df73bd6b2\") " pod="openstack/dnsmasq-dns-759cc7f497-7nv78" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.676224 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a661b9e1-c551-4b36-87a9-833df73bd6b2-dns-swift-storage-0\") pod \"dnsmasq-dns-759cc7f497-7nv78\" (UID: \"a661b9e1-c551-4b36-87a9-833df73bd6b2\") " pod="openstack/dnsmasq-dns-759cc7f497-7nv78" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.676830 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a661b9e1-c551-4b36-87a9-833df73bd6b2-config\") pod \"dnsmasq-dns-759cc7f497-7nv78\" (UID: \"a661b9e1-c551-4b36-87a9-833df73bd6b2\") " pod="openstack/dnsmasq-dns-759cc7f497-7nv78" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.677015 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a661b9e1-c551-4b36-87a9-833df73bd6b2-ovsdbserver-nb\") pod \"dnsmasq-dns-759cc7f497-7nv78\" (UID: \"a661b9e1-c551-4b36-87a9-833df73bd6b2\") " pod="openstack/dnsmasq-dns-759cc7f497-7nv78" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.677772 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a661b9e1-c551-4b36-87a9-833df73bd6b2-dns-svc\") pod \"dnsmasq-dns-759cc7f497-7nv78\" (UID: \"a661b9e1-c551-4b36-87a9-833df73bd6b2\") " pod="openstack/dnsmasq-dns-759cc7f497-7nv78" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.683060 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-rz6qh" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.694786 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ntd2v\" (UniqueName: \"kubernetes.io/projected/a661b9e1-c551-4b36-87a9-833df73bd6b2-kube-api-access-ntd2v\") pod \"dnsmasq-dns-759cc7f497-7nv78\" (UID: \"a661b9e1-c551-4b36-87a9-833df73bd6b2\") " pod="openstack/dnsmasq-dns-759cc7f497-7nv78" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.710645 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-jt88g" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.721140 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-jjqr9" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.754045 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-759cc7f497-7nv78" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.778776 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7aba6c70-a80a-4bce-9216-887cb7a2061c-ovsdbserver-nb\") pod \"7aba6c70-a80a-4bce-9216-887cb7a2061c\" (UID: \"7aba6c70-a80a-4bce-9216-887cb7a2061c\") " Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.778828 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2dk2z\" (UniqueName: \"kubernetes.io/projected/7aba6c70-a80a-4bce-9216-887cb7a2061c-kube-api-access-2dk2z\") pod \"7aba6c70-a80a-4bce-9216-887cb7a2061c\" (UID: \"7aba6c70-a80a-4bce-9216-887cb7a2061c\") " Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.778932 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7aba6c70-a80a-4bce-9216-887cb7a2061c-config\") pod \"7aba6c70-a80a-4bce-9216-887cb7a2061c\" (UID: \"7aba6c70-a80a-4bce-9216-887cb7a2061c\") " Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.778976 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7aba6c70-a80a-4bce-9216-887cb7a2061c-ovsdbserver-sb\") pod \"7aba6c70-a80a-4bce-9216-887cb7a2061c\" (UID: \"7aba6c70-a80a-4bce-9216-887cb7a2061c\") " Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.779047 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7aba6c70-a80a-4bce-9216-887cb7a2061c-dns-svc\") pod \"7aba6c70-a80a-4bce-9216-887cb7a2061c\" (UID: \"7aba6c70-a80a-4bce-9216-887cb7a2061c\") " Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.779068 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7aba6c70-a80a-4bce-9216-887cb7a2061c-dns-swift-storage-0\") pod \"7aba6c70-a80a-4bce-9216-887cb7a2061c\" (UID: \"7aba6c70-a80a-4bce-9216-887cb7a2061c\") " Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.787112 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7aba6c70-a80a-4bce-9216-887cb7a2061c-kube-api-access-2dk2z" (OuterVolumeSpecName: "kube-api-access-2dk2z") pod "7aba6c70-a80a-4bce-9216-887cb7a2061c" (UID: "7aba6c70-a80a-4bce-9216-887cb7a2061c"). InnerVolumeSpecName "kube-api-access-2dk2z". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.875819 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7aba6c70-a80a-4bce-9216-887cb7a2061c-config" (OuterVolumeSpecName: "config") pod "7aba6c70-a80a-4bce-9216-887cb7a2061c" (UID: "7aba6c70-a80a-4bce-9216-887cb7a2061c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.881632 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7aba6c70-a80a-4bce-9216-887cb7a2061c-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.881674 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2dk2z\" (UniqueName: \"kubernetes.io/projected/7aba6c70-a80a-4bce-9216-887cb7a2061c-kube-api-access-2dk2z\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.886972 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7aba6c70-a80a-4bce-9216-887cb7a2061c-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "7aba6c70-a80a-4bce-9216-887cb7a2061c" (UID: "7aba6c70-a80a-4bce-9216-887cb7a2061c"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.892778 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7aba6c70-a80a-4bce-9216-887cb7a2061c-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "7aba6c70-a80a-4bce-9216-887cb7a2061c" (UID: "7aba6c70-a80a-4bce-9216-887cb7a2061c"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.900815 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7aba6c70-a80a-4bce-9216-887cb7a2061c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7aba6c70-a80a-4bce-9216-887cb7a2061c" (UID: "7aba6c70-a80a-4bce-9216-887cb7a2061c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.952963 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Mar 13 14:26:57 crc kubenswrapper[4907]: E0313 14:26:57.953408 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7aba6c70-a80a-4bce-9216-887cb7a2061c" containerName="dnsmasq-dns" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.953424 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="7aba6c70-a80a-4bce-9216-887cb7a2061c" containerName="dnsmasq-dns" Mar 13 14:26:57 crc kubenswrapper[4907]: E0313 14:26:57.953443 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7aba6c70-a80a-4bce-9216-887cb7a2061c" containerName="init" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.953451 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="7aba6c70-a80a-4bce-9216-887cb7a2061c" containerName="init" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.953654 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="7aba6c70-a80a-4bce-9216-887cb7a2061c" containerName="dnsmasq-dns" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.954682 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.960511 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.960856 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.961089 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.961838 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-fc9j2" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.972042 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5cb4dcfdd7-k6pkx"] Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.976859 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7aba6c70-a80a-4bce-9216-887cb7a2061c-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "7aba6c70-a80a-4bce-9216-887cb7a2061c" (UID: "7aba6c70-a80a-4bce-9216-887cb7a2061c"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.978136 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.983386 4907 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7aba6c70-a80a-4bce-9216-887cb7a2061c-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.983424 4907 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7aba6c70-a80a-4bce-9216-887cb7a2061c-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.983439 4907 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7aba6c70-a80a-4bce-9216-887cb7a2061c-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.983451 4907 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7aba6c70-a80a-4bce-9216-887cb7a2061c-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.032990 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.036095 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.037717 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.041129 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.054220 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.076802 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-4tskm"] Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.085627 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ce2a993a-c706-400f-967d-35a9c38d5937-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"ce2a993a-c706-400f-967d-35a9c38d5937\") " pod="openstack/glance-default-external-api-0" Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.085901 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"ce2a993a-c706-400f-967d-35a9c38d5937\") " pod="openstack/glance-default-external-api-0" Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.085926 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce2a993a-c706-400f-967d-35a9c38d5937-config-data\") pod \"glance-default-external-api-0\" (UID: \"ce2a993a-c706-400f-967d-35a9c38d5937\") " pod="openstack/glance-default-external-api-0" Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.085953 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce2a993a-c706-400f-967d-35a9c38d5937-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"ce2a993a-c706-400f-967d-35a9c38d5937\") " pod="openstack/glance-default-external-api-0" Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.085985 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ce2a993a-c706-400f-967d-35a9c38d5937-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"ce2a993a-c706-400f-967d-35a9c38d5937\") " pod="openstack/glance-default-external-api-0" Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.086035 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jxx26\" (UniqueName: \"kubernetes.io/projected/ce2a993a-c706-400f-967d-35a9c38d5937-kube-api-access-jxx26\") pod \"glance-default-external-api-0\" (UID: \"ce2a993a-c706-400f-967d-35a9c38d5937\") " pod="openstack/glance-default-external-api-0" Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.086069 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ce2a993a-c706-400f-967d-35a9c38d5937-logs\") pod \"glance-default-external-api-0\" (UID: \"ce2a993a-c706-400f-967d-35a9c38d5937\") " pod="openstack/glance-default-external-api-0" Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.086092 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ce2a993a-c706-400f-967d-35a9c38d5937-scripts\") pod \"glance-default-external-api-0\" (UID: \"ce2a993a-c706-400f-967d-35a9c38d5937\") " pod="openstack/glance-default-external-api-0" Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.187505 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.187561 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ce2a993a-c706-400f-967d-35a9c38d5937-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"ce2a993a-c706-400f-967d-35a9c38d5937\") " pod="openstack/glance-default-external-api-0" Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.187594 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"ce2a993a-c706-400f-967d-35a9c38d5937\") " pod="openstack/glance-default-external-api-0" Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.187611 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce2a993a-c706-400f-967d-35a9c38d5937-config-data\") pod \"glance-default-external-api-0\" (UID: \"ce2a993a-c706-400f-967d-35a9c38d5937\") " pod="openstack/glance-default-external-api-0" Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.187628 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5-scripts\") pod \"glance-default-internal-api-0\" (UID: \"8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.187650 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce2a993a-c706-400f-967d-35a9c38d5937-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"ce2a993a-c706-400f-967d-35a9c38d5937\") " pod="openstack/glance-default-external-api-0" Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.191530 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5-logs\") pod \"glance-default-internal-api-0\" (UID: \"8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.191648 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ce2a993a-c706-400f-967d-35a9c38d5937-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"ce2a993a-c706-400f-967d-35a9c38d5937\") " pod="openstack/glance-default-external-api-0" Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.191691 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.191720 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5-config-data\") pod \"glance-default-internal-api-0\" (UID: \"8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.191800 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jxx26\" (UniqueName: \"kubernetes.io/projected/ce2a993a-c706-400f-967d-35a9c38d5937-kube-api-access-jxx26\") pod \"glance-default-external-api-0\" (UID: \"ce2a993a-c706-400f-967d-35a9c38d5937\") " pod="openstack/glance-default-external-api-0" Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.191820 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.191839 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k48mp\" (UniqueName: \"kubernetes.io/projected/8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5-kube-api-access-k48mp\") pod \"glance-default-internal-api-0\" (UID: \"8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.192013 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ce2a993a-c706-400f-967d-35a9c38d5937-logs\") pod \"glance-default-external-api-0\" (UID: \"ce2a993a-c706-400f-967d-35a9c38d5937\") " pod="openstack/glance-default-external-api-0" Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.192065 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ce2a993a-c706-400f-967d-35a9c38d5937-scripts\") pod \"glance-default-external-api-0\" (UID: \"ce2a993a-c706-400f-967d-35a9c38d5937\") " pod="openstack/glance-default-external-api-0" Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.192087 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.194232 4907 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"ce2a993a-c706-400f-967d-35a9c38d5937\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/glance-default-external-api-0" Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.212133 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ce2a993a-c706-400f-967d-35a9c38d5937-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"ce2a993a-c706-400f-967d-35a9c38d5937\") " pod="openstack/glance-default-external-api-0" Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.213074 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ce2a993a-c706-400f-967d-35a9c38d5937-logs\") pod \"glance-default-external-api-0\" (UID: \"ce2a993a-c706-400f-967d-35a9c38d5937\") " pod="openstack/glance-default-external-api-0" Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.218144 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ce2a993a-c706-400f-967d-35a9c38d5937-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"ce2a993a-c706-400f-967d-35a9c38d5937\") " pod="openstack/glance-default-external-api-0" Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.256198 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ce2a993a-c706-400f-967d-35a9c38d5937-scripts\") pod \"glance-default-external-api-0\" (UID: \"ce2a993a-c706-400f-967d-35a9c38d5937\") " pod="openstack/glance-default-external-api-0" Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.278624 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jxx26\" (UniqueName: \"kubernetes.io/projected/ce2a993a-c706-400f-967d-35a9c38d5937-kube-api-access-jxx26\") pod \"glance-default-external-api-0\" (UID: \"ce2a993a-c706-400f-967d-35a9c38d5937\") " pod="openstack/glance-default-external-api-0" Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.283570 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce2a993a-c706-400f-967d-35a9c38d5937-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"ce2a993a-c706-400f-967d-35a9c38d5937\") " pod="openstack/glance-default-external-api-0" Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.286937 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce2a993a-c706-400f-967d-35a9c38d5937-config-data\") pod \"glance-default-external-api-0\" (UID: \"ce2a993a-c706-400f-967d-35a9c38d5937\") " pod="openstack/glance-default-external-api-0" Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.300252 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.300398 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5-scripts\") pod \"glance-default-internal-api-0\" (UID: \"8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.300440 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5-logs\") pod \"glance-default-internal-api-0\" (UID: \"8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.300556 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.300624 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5-config-data\") pod \"glance-default-internal-api-0\" (UID: \"8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.300693 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.300720 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k48mp\" (UniqueName: \"kubernetes.io/projected/8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5-kube-api-access-k48mp\") pod \"glance-default-internal-api-0\" (UID: \"8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.300790 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.302141 4907 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/glance-default-internal-api-0" Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.303106 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.308625 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5-scripts\") pod \"glance-default-internal-api-0\" (UID: \"8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.308913 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5-logs\") pod \"glance-default-internal-api-0\" (UID: \"8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.314776 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.325210 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5-config-data\") pod \"glance-default-internal-api-0\" (UID: \"8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.325856 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.327851 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.328111 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"ce2a993a-c706-400f-967d-35a9c38d5937\") " pod="openstack/glance-default-external-api-0" Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.342968 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k48mp\" (UniqueName: \"kubernetes.io/projected/8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5-kube-api-access-k48mp\") pod \"glance-default-internal-api-0\" (UID: \"8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.400189 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.531703 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-4tskm" event={"ID":"4a97d1cc-3762-47fe-a8fe-ed9e95b466a4","Type":"ContainerStarted","Data":"8a34f94ac114c7c6f4deeb646d486da44df7c2de5f8d5ce22726005779d1271f"} Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.532093 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-4tskm" event={"ID":"4a97d1cc-3762-47fe-a8fe-ed9e95b466a4","Type":"ContainerStarted","Data":"a2858c30fa2ae4911b2a7afe4d6fb52407d5d3328c466096a4d4cc1761524119"} Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.535681 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5","Type":"ContainerStarted","Data":"cf9fd571ce09eea426c8836be80b3c4ef776ec41dabdcd7886212b3709f230ce"} Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.552327 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f88567fd9-f24rf" event={"ID":"7aba6c70-a80a-4bce-9216-887cb7a2061c","Type":"ContainerDied","Data":"6f3c01688bba973a49cd38ead6f32071203a13f1c437e87590a2062f43c42d3e"} Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.552406 4907 scope.go:117] "RemoveContainer" containerID="3fd8ece48af6f3186fe1a67ce35dde9ab752271f826eb2d327ed3e6092299012" Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.552506 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f88567fd9-f24rf" Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.559129 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-4tskm" podStartSLOduration=2.559112485 podStartE2EDuration="2.559112485s" podCreationTimestamp="2026-03-13 14:26:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:26:58.552384451 +0000 UTC m=+1317.452172140" watchObservedRunningTime="2026-03-13 14:26:58.559112485 +0000 UTC m=+1317.458900174" Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.560334 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.572869 4907 generic.go:334] "Generic (PLEG): container finished" podID="c495e208-654b-477a-869e-87a5c2566519" containerID="12156f70b991f6a7b1a9f418d96256568ed61931e620f9a12a50e33bb305271b" exitCode=0 Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.572996 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5cb4dcfdd7-k6pkx" event={"ID":"c495e208-654b-477a-869e-87a5c2566519","Type":"ContainerDied","Data":"12156f70b991f6a7b1a9f418d96256568ed61931e620f9a12a50e33bb305271b"} Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.573126 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5cb4dcfdd7-k6pkx" event={"ID":"c495e208-654b-477a-869e-87a5c2566519","Type":"ContainerStarted","Data":"6116fd540adb884533031b5d7e277d5df41a49b6c0e586ce5363f34faed7d7b7"} Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.591435 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.624714 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6f88567fd9-f24rf"] Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.624993 4907 scope.go:117] "RemoveContainer" containerID="b20445b78457705b4b60b2201ff58c281da6086f771d56be4ceb3ab98dc8b1f5" Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.633450 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6f88567fd9-f24rf"] Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.717171 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-rz6qh"] Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.728309 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-7q999"] Mar 13 14:26:58 crc kubenswrapper[4907]: W0313 14:26:58.729410 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod32b531c6_64c1_4137_b82e_3c10789325e6.slice/crio-4b7673a5ff0b4e402bcd5b85ece6329b6bba3e9faaac7ef9c519ffe96fec5929 WatchSource:0}: Error finding container 4b7673a5ff0b4e402bcd5b85ece6329b6bba3e9faaac7ef9c519ffe96fec5929: Status 404 returned error can't find the container with id 4b7673a5ff0b4e402bcd5b85ece6329b6bba3e9faaac7ef9c519ffe96fec5929 Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.735643 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-jt88g"] Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.763229 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-jjqr9"] Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.789651 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-759cc7f497-7nv78"] Mar 13 14:26:58 crc kubenswrapper[4907]: W0313 14:26:58.821513 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda661b9e1_c551_4b36_87a9_833df73bd6b2.slice/crio-eaee566543e3557ec83d2d26f27876f1ccb4095ddf1c90e6680a5df6c9c175d0 WatchSource:0}: Error finding container eaee566543e3557ec83d2d26f27876f1ccb4095ddf1c90e6680a5df6c9c175d0: Status 404 returned error can't find the container with id eaee566543e3557ec83d2d26f27876f1ccb4095ddf1c90e6680a5df6c9c175d0 Mar 13 14:26:59 crc kubenswrapper[4907]: I0313 14:26:59.054079 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5cb4dcfdd7-k6pkx" Mar 13 14:26:59 crc kubenswrapper[4907]: I0313 14:26:59.182589 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 13 14:26:59 crc kubenswrapper[4907]: I0313 14:26:59.227346 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-db278\" (UniqueName: \"kubernetes.io/projected/c495e208-654b-477a-869e-87a5c2566519-kube-api-access-db278\") pod \"c495e208-654b-477a-869e-87a5c2566519\" (UID: \"c495e208-654b-477a-869e-87a5c2566519\") " Mar 13 14:26:59 crc kubenswrapper[4907]: I0313 14:26:59.227478 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c495e208-654b-477a-869e-87a5c2566519-dns-svc\") pod \"c495e208-654b-477a-869e-87a5c2566519\" (UID: \"c495e208-654b-477a-869e-87a5c2566519\") " Mar 13 14:26:59 crc kubenswrapper[4907]: I0313 14:26:59.227783 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c495e208-654b-477a-869e-87a5c2566519-ovsdbserver-nb\") pod \"c495e208-654b-477a-869e-87a5c2566519\" (UID: \"c495e208-654b-477a-869e-87a5c2566519\") " Mar 13 14:26:59 crc kubenswrapper[4907]: I0313 14:26:59.227829 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c495e208-654b-477a-869e-87a5c2566519-dns-swift-storage-0\") pod \"c495e208-654b-477a-869e-87a5c2566519\" (UID: \"c495e208-654b-477a-869e-87a5c2566519\") " Mar 13 14:26:59 crc kubenswrapper[4907]: I0313 14:26:59.227854 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c495e208-654b-477a-869e-87a5c2566519-ovsdbserver-sb\") pod \"c495e208-654b-477a-869e-87a5c2566519\" (UID: \"c495e208-654b-477a-869e-87a5c2566519\") " Mar 13 14:26:59 crc kubenswrapper[4907]: I0313 14:26:59.227951 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c495e208-654b-477a-869e-87a5c2566519-config\") pod \"c495e208-654b-477a-869e-87a5c2566519\" (UID: \"c495e208-654b-477a-869e-87a5c2566519\") " Mar 13 14:26:59 crc kubenswrapper[4907]: I0313 14:26:59.235472 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c495e208-654b-477a-869e-87a5c2566519-kube-api-access-db278" (OuterVolumeSpecName: "kube-api-access-db278") pod "c495e208-654b-477a-869e-87a5c2566519" (UID: "c495e208-654b-477a-869e-87a5c2566519"). InnerVolumeSpecName "kube-api-access-db278". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:26:59 crc kubenswrapper[4907]: I0313 14:26:59.247798 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 13 14:26:59 crc kubenswrapper[4907]: I0313 14:26:59.278531 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c495e208-654b-477a-869e-87a5c2566519-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "c495e208-654b-477a-869e-87a5c2566519" (UID: "c495e208-654b-477a-869e-87a5c2566519"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:26:59 crc kubenswrapper[4907]: I0313 14:26:59.281279 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c495e208-654b-477a-869e-87a5c2566519-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "c495e208-654b-477a-869e-87a5c2566519" (UID: "c495e208-654b-477a-869e-87a5c2566519"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:26:59 crc kubenswrapper[4907]: I0313 14:26:59.308535 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c495e208-654b-477a-869e-87a5c2566519-config" (OuterVolumeSpecName: "config") pod "c495e208-654b-477a-869e-87a5c2566519" (UID: "c495e208-654b-477a-869e-87a5c2566519"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:26:59 crc kubenswrapper[4907]: I0313 14:26:59.314648 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c495e208-654b-477a-869e-87a5c2566519-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "c495e208-654b-477a-869e-87a5c2566519" (UID: "c495e208-654b-477a-869e-87a5c2566519"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:26:59 crc kubenswrapper[4907]: I0313 14:26:59.351661 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 13 14:26:59 crc kubenswrapper[4907]: I0313 14:26:59.357437 4907 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c495e208-654b-477a-869e-87a5c2566519-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:59 crc kubenswrapper[4907]: I0313 14:26:59.357478 4907 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c495e208-654b-477a-869e-87a5c2566519-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:59 crc kubenswrapper[4907]: I0313 14:26:59.357491 4907 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c495e208-654b-477a-869e-87a5c2566519-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:59 crc kubenswrapper[4907]: I0313 14:26:59.357503 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c495e208-654b-477a-869e-87a5c2566519-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:59 crc kubenswrapper[4907]: I0313 14:26:59.357514 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-db278\" (UniqueName: \"kubernetes.io/projected/c495e208-654b-477a-869e-87a5c2566519-kube-api-access-db278\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:59 crc kubenswrapper[4907]: I0313 14:26:59.357581 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c495e208-654b-477a-869e-87a5c2566519-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c495e208-654b-477a-869e-87a5c2566519" (UID: "c495e208-654b-477a-869e-87a5c2566519"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:26:59 crc kubenswrapper[4907]: I0313 14:26:59.436692 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 13 14:26:59 crc kubenswrapper[4907]: I0313 14:26:59.460395 4907 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c495e208-654b-477a-869e-87a5c2566519-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:59 crc kubenswrapper[4907]: I0313 14:26:59.586341 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5","Type":"ContainerStarted","Data":"e365b5ca2986747d106a442f14cd719005e64c8ba4070ee9236f7119ccd3db0a"} Mar 13 14:26:59 crc kubenswrapper[4907]: I0313 14:26:59.589720 4907 generic.go:334] "Generic (PLEG): container finished" podID="a661b9e1-c551-4b36-87a9-833df73bd6b2" containerID="4e3ebec78f62227a1eec084cd7d40d9bb86b61261221bf385169185e602cfad5" exitCode=0 Mar 13 14:26:59 crc kubenswrapper[4907]: I0313 14:26:59.589844 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-759cc7f497-7nv78" event={"ID":"a661b9e1-c551-4b36-87a9-833df73bd6b2","Type":"ContainerDied","Data":"4e3ebec78f62227a1eec084cd7d40d9bb86b61261221bf385169185e602cfad5"} Mar 13 14:26:59 crc kubenswrapper[4907]: I0313 14:26:59.589910 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-759cc7f497-7nv78" event={"ID":"a661b9e1-c551-4b36-87a9-833df73bd6b2","Type":"ContainerStarted","Data":"eaee566543e3557ec83d2d26f27876f1ccb4095ddf1c90e6680a5df6c9c175d0"} Mar 13 14:26:59 crc kubenswrapper[4907]: I0313 14:26:59.596380 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-jt88g" event={"ID":"870ec55c-fc6f-4b16-8c49-a3cda0d0d010","Type":"ContainerStarted","Data":"c90c150f12af47e1b50f63c9ddceacb4d1551893cbeff831172fa51757c596cb"} Mar 13 14:26:59 crc kubenswrapper[4907]: I0313 14:26:59.596426 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-jt88g" event={"ID":"870ec55c-fc6f-4b16-8c49-a3cda0d0d010","Type":"ContainerStarted","Data":"61484e5c865e06dbc491f0726df5e379526c160f176a2435939232b317293ca1"} Mar 13 14:26:59 crc kubenswrapper[4907]: I0313 14:26:59.609637 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5cb4dcfdd7-k6pkx" event={"ID":"c495e208-654b-477a-869e-87a5c2566519","Type":"ContainerDied","Data":"6116fd540adb884533031b5d7e277d5df41a49b6c0e586ce5363f34faed7d7b7"} Mar 13 14:26:59 crc kubenswrapper[4907]: I0313 14:26:59.609690 4907 scope.go:117] "RemoveContainer" containerID="12156f70b991f6a7b1a9f418d96256568ed61931e620f9a12a50e33bb305271b" Mar 13 14:26:59 crc kubenswrapper[4907]: I0313 14:26:59.609796 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5cb4dcfdd7-k6pkx" Mar 13 14:26:59 crc kubenswrapper[4907]: I0313 14:26:59.616918 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-rz6qh" event={"ID":"32b531c6-64c1-4137-b82e-3c10789325e6","Type":"ContainerStarted","Data":"4b7673a5ff0b4e402bcd5b85ece6329b6bba3e9faaac7ef9c519ffe96fec5929"} Mar 13 14:26:59 crc kubenswrapper[4907]: I0313 14:26:59.637543 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-7q999" event={"ID":"594b61fa-12b7-47b6-8af3-17f024e0d54d","Type":"ContainerStarted","Data":"247e75fb8f70b076f7d10c60ec2902246989846bb77cff82137c6d0beb413598"} Mar 13 14:26:59 crc kubenswrapper[4907]: I0313 14:26:59.640578 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-jjqr9" event={"ID":"b144897c-8640-4c5c-afa1-68b44dfd4f98","Type":"ContainerStarted","Data":"d144653276b0a0015c06098911900d038b6dd5b20a40a40339bc3b434d2953a5"} Mar 13 14:26:59 crc kubenswrapper[4907]: I0313 14:26:59.644770 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-jt88g" podStartSLOduration=2.644756852 podStartE2EDuration="2.644756852s" podCreationTimestamp="2026-03-13 14:26:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:26:59.631989244 +0000 UTC m=+1318.531776933" watchObservedRunningTime="2026-03-13 14:26:59.644756852 +0000 UTC m=+1318.544544541" Mar 13 14:26:59 crc kubenswrapper[4907]: I0313 14:26:59.692624 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5cb4dcfdd7-k6pkx"] Mar 13 14:26:59 crc kubenswrapper[4907]: I0313 14:26:59.705187 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5cb4dcfdd7-k6pkx"] Mar 13 14:26:59 crc kubenswrapper[4907]: I0313 14:26:59.939224 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7aba6c70-a80a-4bce-9216-887cb7a2061c" path="/var/lib/kubelet/pods/7aba6c70-a80a-4bce-9216-887cb7a2061c/volumes" Mar 13 14:26:59 crc kubenswrapper[4907]: I0313 14:26:59.940058 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c495e208-654b-477a-869e-87a5c2566519" path="/var/lib/kubelet/pods/c495e208-654b-477a-869e-87a5c2566519/volumes" Mar 13 14:27:00 crc kubenswrapper[4907]: I0313 14:27:00.319757 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 13 14:27:00 crc kubenswrapper[4907]: I0313 14:27:00.683432 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"ce2a993a-c706-400f-967d-35a9c38d5937","Type":"ContainerStarted","Data":"70f4055057ae74811d3e519a30dbdf9559ebe320aa1f7b52ebbfd8cfd47c2673"} Mar 13 14:27:00 crc kubenswrapper[4907]: I0313 14:27:00.686960 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5","Type":"ContainerStarted","Data":"d56b748bd6482c49c5b182589b671f235af50f54159cfd34d117c3100423cb0f"} Mar 13 14:27:00 crc kubenswrapper[4907]: I0313 14:27:00.699661 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-759cc7f497-7nv78" event={"ID":"a661b9e1-c551-4b36-87a9-833df73bd6b2","Type":"ContainerStarted","Data":"46c2ff1e9e66c6fc449295a888dc400f857863d7df33a386b4b1e1cf42f6a3f4"} Mar 13 14:27:00 crc kubenswrapper[4907]: I0313 14:27:00.700505 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-759cc7f497-7nv78" Mar 13 14:27:00 crc kubenswrapper[4907]: I0313 14:27:00.730253 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-759cc7f497-7nv78" podStartSLOduration=3.730231944 podStartE2EDuration="3.730231944s" podCreationTimestamp="2026-03-13 14:26:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:27:00.724065206 +0000 UTC m=+1319.623852895" watchObservedRunningTime="2026-03-13 14:27:00.730231944 +0000 UTC m=+1319.630019633" Mar 13 14:27:01 crc kubenswrapper[4907]: I0313 14:27:01.739419 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5","Type":"ContainerStarted","Data":"9d6a810392296b49581d1d82cea96d2320cf1660821d93127c921446e01396f1"} Mar 13 14:27:01 crc kubenswrapper[4907]: I0313 14:27:01.743302 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"ce2a993a-c706-400f-967d-35a9c38d5937","Type":"ContainerStarted","Data":"512291676abb06ed48b9f163f1a7202f5fd9b599abc8f6b456c67f77c909e778"} Mar 13 14:27:02 crc kubenswrapper[4907]: I0313 14:27:02.754227 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"ce2a993a-c706-400f-967d-35a9c38d5937","Type":"ContainerStarted","Data":"c604cb7d84d59e89fd67f7b2707cc5e5fce549a56d035662c24b9cb3f930d149"} Mar 13 14:27:02 crc kubenswrapper[4907]: I0313 14:27:02.754307 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5" containerName="glance-log" containerID="cri-o://d56b748bd6482c49c5b182589b671f235af50f54159cfd34d117c3100423cb0f" gracePeriod=30 Mar 13 14:27:02 crc kubenswrapper[4907]: I0313 14:27:02.755128 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="ce2a993a-c706-400f-967d-35a9c38d5937" containerName="glance-log" containerID="cri-o://512291676abb06ed48b9f163f1a7202f5fd9b599abc8f6b456c67f77c909e778" gracePeriod=30 Mar 13 14:27:02 crc kubenswrapper[4907]: I0313 14:27:02.755240 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="ce2a993a-c706-400f-967d-35a9c38d5937" containerName="glance-httpd" containerID="cri-o://c604cb7d84d59e89fd67f7b2707cc5e5fce549a56d035662c24b9cb3f930d149" gracePeriod=30 Mar 13 14:27:02 crc kubenswrapper[4907]: I0313 14:27:02.755331 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5" containerName="glance-httpd" containerID="cri-o://9d6a810392296b49581d1d82cea96d2320cf1660821d93127c921446e01396f1" gracePeriod=30 Mar 13 14:27:02 crc kubenswrapper[4907]: I0313 14:27:02.784073 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=6.784019912 podStartE2EDuration="6.784019912s" podCreationTimestamp="2026-03-13 14:26:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:27:02.778303467 +0000 UTC m=+1321.678091156" watchObservedRunningTime="2026-03-13 14:27:02.784019912 +0000 UTC m=+1321.683807601" Mar 13 14:27:02 crc kubenswrapper[4907]: I0313 14:27:02.804058 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=6.804042417 podStartE2EDuration="6.804042417s" podCreationTimestamp="2026-03-13 14:26:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:27:02.800962223 +0000 UTC m=+1321.700749912" watchObservedRunningTime="2026-03-13 14:27:02.804042417 +0000 UTC m=+1321.703830106" Mar 13 14:27:02 crc kubenswrapper[4907]: I0313 14:27:02.897865 4907 scope.go:117] "RemoveContainer" containerID="dd6efe3aa3a79d161b96b8d84e7fa19f50f23cdeadb3268906a3667faac57368" Mar 13 14:27:03 crc kubenswrapper[4907]: I0313 14:27:03.764416 4907 generic.go:334] "Generic (PLEG): container finished" podID="4a97d1cc-3762-47fe-a8fe-ed9e95b466a4" containerID="8a34f94ac114c7c6f4deeb646d486da44df7c2de5f8d5ce22726005779d1271f" exitCode=0 Mar 13 14:27:03 crc kubenswrapper[4907]: I0313 14:27:03.764524 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-4tskm" event={"ID":"4a97d1cc-3762-47fe-a8fe-ed9e95b466a4","Type":"ContainerDied","Data":"8a34f94ac114c7c6f4deeb646d486da44df7c2de5f8d5ce22726005779d1271f"} Mar 13 14:27:03 crc kubenswrapper[4907]: I0313 14:27:03.766897 4907 generic.go:334] "Generic (PLEG): container finished" podID="ce2a993a-c706-400f-967d-35a9c38d5937" containerID="c604cb7d84d59e89fd67f7b2707cc5e5fce549a56d035662c24b9cb3f930d149" exitCode=0 Mar 13 14:27:03 crc kubenswrapper[4907]: I0313 14:27:03.766915 4907 generic.go:334] "Generic (PLEG): container finished" podID="ce2a993a-c706-400f-967d-35a9c38d5937" containerID="512291676abb06ed48b9f163f1a7202f5fd9b599abc8f6b456c67f77c909e778" exitCode=143 Mar 13 14:27:03 crc kubenswrapper[4907]: I0313 14:27:03.766922 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"ce2a993a-c706-400f-967d-35a9c38d5937","Type":"ContainerDied","Data":"c604cb7d84d59e89fd67f7b2707cc5e5fce549a56d035662c24b9cb3f930d149"} Mar 13 14:27:03 crc kubenswrapper[4907]: I0313 14:27:03.766957 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"ce2a993a-c706-400f-967d-35a9c38d5937","Type":"ContainerDied","Data":"512291676abb06ed48b9f163f1a7202f5fd9b599abc8f6b456c67f77c909e778"} Mar 13 14:27:03 crc kubenswrapper[4907]: I0313 14:27:03.769807 4907 generic.go:334] "Generic (PLEG): container finished" podID="8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5" containerID="9d6a810392296b49581d1d82cea96d2320cf1660821d93127c921446e01396f1" exitCode=0 Mar 13 14:27:03 crc kubenswrapper[4907]: I0313 14:27:03.769827 4907 generic.go:334] "Generic (PLEG): container finished" podID="8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5" containerID="d56b748bd6482c49c5b182589b671f235af50f54159cfd34d117c3100423cb0f" exitCode=143 Mar 13 14:27:03 crc kubenswrapper[4907]: I0313 14:27:03.769844 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5","Type":"ContainerDied","Data":"9d6a810392296b49581d1d82cea96d2320cf1660821d93127c921446e01396f1"} Mar 13 14:27:03 crc kubenswrapper[4907]: I0313 14:27:03.769862 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5","Type":"ContainerDied","Data":"d56b748bd6482c49c5b182589b671f235af50f54159cfd34d117c3100423cb0f"} Mar 13 14:27:07 crc kubenswrapper[4907]: I0313 14:27:07.756146 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-759cc7f497-7nv78" Mar 13 14:27:07 crc kubenswrapper[4907]: I0313 14:27:07.848226 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-67754df655-k2v49"] Mar 13 14:27:07 crc kubenswrapper[4907]: I0313 14:27:07.848506 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-67754df655-k2v49" podUID="c065ce46-d7f3-4e50-b30c-83a1592b6ead" containerName="dnsmasq-dns" containerID="cri-o://fd0a2c793e62b5b128e16861d049a22d36d9baa0de3410f3961ca59044ad1213" gracePeriod=10 Mar 13 14:27:08 crc kubenswrapper[4907]: I0313 14:27:08.847669 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-4tskm" event={"ID":"4a97d1cc-3762-47fe-a8fe-ed9e95b466a4","Type":"ContainerDied","Data":"a2858c30fa2ae4911b2a7afe4d6fb52407d5d3328c466096a4d4cc1761524119"} Mar 13 14:27:08 crc kubenswrapper[4907]: I0313 14:27:08.848096 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a2858c30fa2ae4911b2a7afe4d6fb52407d5d3328c466096a4d4cc1761524119" Mar 13 14:27:08 crc kubenswrapper[4907]: I0313 14:27:08.850616 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 13 14:27:08 crc kubenswrapper[4907]: I0313 14:27:08.851008 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5","Type":"ContainerDied","Data":"e365b5ca2986747d106a442f14cd719005e64c8ba4070ee9236f7119ccd3db0a"} Mar 13 14:27:08 crc kubenswrapper[4907]: I0313 14:27:08.851081 4907 scope.go:117] "RemoveContainer" containerID="9d6a810392296b49581d1d82cea96d2320cf1660821d93127c921446e01396f1" Mar 13 14:27:08 crc kubenswrapper[4907]: I0313 14:27:08.854642 4907 generic.go:334] "Generic (PLEG): container finished" podID="c065ce46-d7f3-4e50-b30c-83a1592b6ead" containerID="fd0a2c793e62b5b128e16861d049a22d36d9baa0de3410f3961ca59044ad1213" exitCode=0 Mar 13 14:27:08 crc kubenswrapper[4907]: I0313 14:27:08.854715 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67754df655-k2v49" event={"ID":"c065ce46-d7f3-4e50-b30c-83a1592b6ead","Type":"ContainerDied","Data":"fd0a2c793e62b5b128e16861d049a22d36d9baa0de3410f3961ca59044ad1213"} Mar 13 14:27:08 crc kubenswrapper[4907]: I0313 14:27:08.859456 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-4tskm" Mar 13 14:27:08 crc kubenswrapper[4907]: I0313 14:27:08.963012 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/4a97d1cc-3762-47fe-a8fe-ed9e95b466a4-credential-keys\") pod \"4a97d1cc-3762-47fe-a8fe-ed9e95b466a4\" (UID: \"4a97d1cc-3762-47fe-a8fe-ed9e95b466a4\") " Mar 13 14:27:08 crc kubenswrapper[4907]: I0313 14:27:08.963086 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4a97d1cc-3762-47fe-a8fe-ed9e95b466a4-fernet-keys\") pod \"4a97d1cc-3762-47fe-a8fe-ed9e95b466a4\" (UID: \"4a97d1cc-3762-47fe-a8fe-ed9e95b466a4\") " Mar 13 14:27:08 crc kubenswrapper[4907]: I0313 14:27:08.963109 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a97d1cc-3762-47fe-a8fe-ed9e95b466a4-combined-ca-bundle\") pod \"4a97d1cc-3762-47fe-a8fe-ed9e95b466a4\" (UID: \"4a97d1cc-3762-47fe-a8fe-ed9e95b466a4\") " Mar 13 14:27:08 crc kubenswrapper[4907]: I0313 14:27:08.963125 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5-config-data\") pod \"8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5\" (UID: \"8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5\") " Mar 13 14:27:08 crc kubenswrapper[4907]: I0313 14:27:08.963143 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5\" (UID: \"8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5\") " Mar 13 14:27:08 crc kubenswrapper[4907]: I0313 14:27:08.963163 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k48mp\" (UniqueName: \"kubernetes.io/projected/8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5-kube-api-access-k48mp\") pod \"8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5\" (UID: \"8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5\") " Mar 13 14:27:08 crc kubenswrapper[4907]: I0313 14:27:08.963183 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a97d1cc-3762-47fe-a8fe-ed9e95b466a4-config-data\") pod \"4a97d1cc-3762-47fe-a8fe-ed9e95b466a4\" (UID: \"4a97d1cc-3762-47fe-a8fe-ed9e95b466a4\") " Mar 13 14:27:08 crc kubenswrapper[4907]: I0313 14:27:08.963198 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5-combined-ca-bundle\") pod \"8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5\" (UID: \"8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5\") " Mar 13 14:27:08 crc kubenswrapper[4907]: I0313 14:27:08.963218 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5-scripts\") pod \"8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5\" (UID: \"8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5\") " Mar 13 14:27:08 crc kubenswrapper[4907]: I0313 14:27:08.963236 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5-internal-tls-certs\") pod \"8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5\" (UID: \"8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5\") " Mar 13 14:27:08 crc kubenswrapper[4907]: I0313 14:27:08.963265 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4a97d1cc-3762-47fe-a8fe-ed9e95b466a4-scripts\") pod \"4a97d1cc-3762-47fe-a8fe-ed9e95b466a4\" (UID: \"4a97d1cc-3762-47fe-a8fe-ed9e95b466a4\") " Mar 13 14:27:08 crc kubenswrapper[4907]: I0313 14:27:08.963345 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pvhgq\" (UniqueName: \"kubernetes.io/projected/4a97d1cc-3762-47fe-a8fe-ed9e95b466a4-kube-api-access-pvhgq\") pod \"4a97d1cc-3762-47fe-a8fe-ed9e95b466a4\" (UID: \"4a97d1cc-3762-47fe-a8fe-ed9e95b466a4\") " Mar 13 14:27:08 crc kubenswrapper[4907]: I0313 14:27:08.963380 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5-logs\") pod \"8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5\" (UID: \"8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5\") " Mar 13 14:27:08 crc kubenswrapper[4907]: I0313 14:27:08.963414 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5-httpd-run\") pod \"8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5\" (UID: \"8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5\") " Mar 13 14:27:08 crc kubenswrapper[4907]: I0313 14:27:08.966285 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5-logs" (OuterVolumeSpecName: "logs") pod "8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5" (UID: "8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:27:08 crc kubenswrapper[4907]: I0313 14:27:08.966516 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5" (UID: "8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:27:08 crc kubenswrapper[4907]: I0313 14:27:08.971692 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a97d1cc-3762-47fe-a8fe-ed9e95b466a4-scripts" (OuterVolumeSpecName: "scripts") pod "4a97d1cc-3762-47fe-a8fe-ed9e95b466a4" (UID: "4a97d1cc-3762-47fe-a8fe-ed9e95b466a4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:27:08 crc kubenswrapper[4907]: I0313 14:27:08.971968 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5-kube-api-access-k48mp" (OuterVolumeSpecName: "kube-api-access-k48mp") pod "8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5" (UID: "8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5"). InnerVolumeSpecName "kube-api-access-k48mp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:27:08 crc kubenswrapper[4907]: I0313 14:27:08.972512 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a97d1cc-3762-47fe-a8fe-ed9e95b466a4-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "4a97d1cc-3762-47fe-a8fe-ed9e95b466a4" (UID: "4a97d1cc-3762-47fe-a8fe-ed9e95b466a4"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:27:08 crc kubenswrapper[4907]: I0313 14:27:08.973017 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a97d1cc-3762-47fe-a8fe-ed9e95b466a4-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "4a97d1cc-3762-47fe-a8fe-ed9e95b466a4" (UID: "4a97d1cc-3762-47fe-a8fe-ed9e95b466a4"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:27:08 crc kubenswrapper[4907]: I0313 14:27:08.973152 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4a97d1cc-3762-47fe-a8fe-ed9e95b466a4-kube-api-access-pvhgq" (OuterVolumeSpecName: "kube-api-access-pvhgq") pod "4a97d1cc-3762-47fe-a8fe-ed9e95b466a4" (UID: "4a97d1cc-3762-47fe-a8fe-ed9e95b466a4"). InnerVolumeSpecName "kube-api-access-pvhgq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:27:08 crc kubenswrapper[4907]: I0313 14:27:08.973513 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5-scripts" (OuterVolumeSpecName: "scripts") pod "8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5" (UID: "8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:27:08 crc kubenswrapper[4907]: I0313 14:27:08.977574 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "glance") pod "8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5" (UID: "8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Mar 13 14:27:08 crc kubenswrapper[4907]: I0313 14:27:08.997562 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a97d1cc-3762-47fe-a8fe-ed9e95b466a4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4a97d1cc-3762-47fe-a8fe-ed9e95b466a4" (UID: "4a97d1cc-3762-47fe-a8fe-ed9e95b466a4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:27:08 crc kubenswrapper[4907]: I0313 14:27:08.998132 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a97d1cc-3762-47fe-a8fe-ed9e95b466a4-config-data" (OuterVolumeSpecName: "config-data") pod "4a97d1cc-3762-47fe-a8fe-ed9e95b466a4" (UID: "4a97d1cc-3762-47fe-a8fe-ed9e95b466a4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:27:09 crc kubenswrapper[4907]: I0313 14:27:09.011022 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5" (UID: "8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:27:09 crc kubenswrapper[4907]: I0313 14:27:09.036663 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5" (UID: "8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:27:09 crc kubenswrapper[4907]: I0313 14:27:09.036997 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5-config-data" (OuterVolumeSpecName: "config-data") pod "8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5" (UID: "8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:27:09 crc kubenswrapper[4907]: I0313 14:27:09.065979 4907 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:09 crc kubenswrapper[4907]: I0313 14:27:09.066018 4907 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:09 crc kubenswrapper[4907]: I0313 14:27:09.066031 4907 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4a97d1cc-3762-47fe-a8fe-ed9e95b466a4-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:09 crc kubenswrapper[4907]: I0313 14:27:09.066145 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pvhgq\" (UniqueName: \"kubernetes.io/projected/4a97d1cc-3762-47fe-a8fe-ed9e95b466a4-kube-api-access-pvhgq\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:09 crc kubenswrapper[4907]: I0313 14:27:09.066157 4907 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5-logs\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:09 crc kubenswrapper[4907]: I0313 14:27:09.066167 4907 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5-httpd-run\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:09 crc kubenswrapper[4907]: I0313 14:27:09.066178 4907 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/4a97d1cc-3762-47fe-a8fe-ed9e95b466a4-credential-keys\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:09 crc kubenswrapper[4907]: I0313 14:27:09.066189 4907 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4a97d1cc-3762-47fe-a8fe-ed9e95b466a4-fernet-keys\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:09 crc kubenswrapper[4907]: I0313 14:27:09.066227 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a97d1cc-3762-47fe-a8fe-ed9e95b466a4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:09 crc kubenswrapper[4907]: I0313 14:27:09.066239 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:09 crc kubenswrapper[4907]: I0313 14:27:09.066305 4907 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Mar 13 14:27:09 crc kubenswrapper[4907]: I0313 14:27:09.066320 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k48mp\" (UniqueName: \"kubernetes.io/projected/8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5-kube-api-access-k48mp\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:09 crc kubenswrapper[4907]: I0313 14:27:09.066329 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:09 crc kubenswrapper[4907]: I0313 14:27:09.066338 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a97d1cc-3762-47fe-a8fe-ed9e95b466a4-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:09 crc kubenswrapper[4907]: I0313 14:27:09.095700 4907 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Mar 13 14:27:09 crc kubenswrapper[4907]: I0313 14:27:09.168174 4907 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:09 crc kubenswrapper[4907]: I0313 14:27:09.719648 4907 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-67754df655-k2v49" podUID="c065ce46-d7f3-4e50-b30c-83a1592b6ead" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.133:5353: connect: connection refused" Mar 13 14:27:09 crc kubenswrapper[4907]: I0313 14:27:09.869399 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-4tskm" Mar 13 14:27:09 crc kubenswrapper[4907]: I0313 14:27:09.869454 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 13 14:27:09 crc kubenswrapper[4907]: I0313 14:27:09.894396 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 13 14:27:09 crc kubenswrapper[4907]: I0313 14:27:09.903872 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 13 14:27:09 crc kubenswrapper[4907]: I0313 14:27:09.951083 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 13 14:27:09 crc kubenswrapper[4907]: E0313 14:27:09.951503 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c495e208-654b-477a-869e-87a5c2566519" containerName="init" Mar 13 14:27:09 crc kubenswrapper[4907]: I0313 14:27:09.951524 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="c495e208-654b-477a-869e-87a5c2566519" containerName="init" Mar 13 14:27:09 crc kubenswrapper[4907]: E0313 14:27:09.951542 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5" containerName="glance-log" Mar 13 14:27:09 crc kubenswrapper[4907]: I0313 14:27:09.951550 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5" containerName="glance-log" Mar 13 14:27:09 crc kubenswrapper[4907]: E0313 14:27:09.951565 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a97d1cc-3762-47fe-a8fe-ed9e95b466a4" containerName="keystone-bootstrap" Mar 13 14:27:09 crc kubenswrapper[4907]: I0313 14:27:09.951574 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a97d1cc-3762-47fe-a8fe-ed9e95b466a4" containerName="keystone-bootstrap" Mar 13 14:27:09 crc kubenswrapper[4907]: E0313 14:27:09.951600 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5" containerName="glance-httpd" Mar 13 14:27:09 crc kubenswrapper[4907]: I0313 14:27:09.951608 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5" containerName="glance-httpd" Mar 13 14:27:09 crc kubenswrapper[4907]: I0313 14:27:09.951788 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="4a97d1cc-3762-47fe-a8fe-ed9e95b466a4" containerName="keystone-bootstrap" Mar 13 14:27:09 crc kubenswrapper[4907]: I0313 14:27:09.951798 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="c495e208-654b-477a-869e-87a5c2566519" containerName="init" Mar 13 14:27:09 crc kubenswrapper[4907]: I0313 14:27:09.951807 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5" containerName="glance-log" Mar 13 14:27:09 crc kubenswrapper[4907]: I0313 14:27:09.951828 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5" containerName="glance-httpd" Mar 13 14:27:09 crc kubenswrapper[4907]: I0313 14:27:09.952954 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 13 14:27:09 crc kubenswrapper[4907]: I0313 14:27:09.955450 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Mar 13 14:27:09 crc kubenswrapper[4907]: I0313 14:27:09.955829 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Mar 13 14:27:09 crc kubenswrapper[4907]: I0313 14:27:09.975445 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 13 14:27:09 crc kubenswrapper[4907]: I0313 14:27:09.986298 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-4tskm"] Mar 13 14:27:09 crc kubenswrapper[4907]: I0313 14:27:09.994816 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-4tskm"] Mar 13 14:27:10 crc kubenswrapper[4907]: I0313 14:27:10.056058 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-crrww"] Mar 13 14:27:10 crc kubenswrapper[4907]: I0313 14:27:10.057127 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-crrww" Mar 13 14:27:10 crc kubenswrapper[4907]: I0313 14:27:10.059019 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Mar 13 14:27:10 crc kubenswrapper[4907]: I0313 14:27:10.059124 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Mar 13 14:27:10 crc kubenswrapper[4907]: I0313 14:27:10.059304 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Mar 13 14:27:10 crc kubenswrapper[4907]: I0313 14:27:10.060841 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-p9m6m" Mar 13 14:27:10 crc kubenswrapper[4907]: I0313 14:27:10.062961 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Mar 13 14:27:10 crc kubenswrapper[4907]: I0313 14:27:10.074569 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-crrww"] Mar 13 14:27:10 crc kubenswrapper[4907]: I0313 14:27:10.092842 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"5c36a352-151f-4e93-8094-3855bfed532e\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:27:10 crc kubenswrapper[4907]: I0313 14:27:10.092913 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f7gh4\" (UniqueName: \"kubernetes.io/projected/5c36a352-151f-4e93-8094-3855bfed532e-kube-api-access-f7gh4\") pod \"glance-default-internal-api-0\" (UID: \"5c36a352-151f-4e93-8094-3855bfed532e\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:27:10 crc kubenswrapper[4907]: I0313 14:27:10.092936 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c36a352-151f-4e93-8094-3855bfed532e-config-data\") pod \"glance-default-internal-api-0\" (UID: \"5c36a352-151f-4e93-8094-3855bfed532e\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:27:10 crc kubenswrapper[4907]: I0313 14:27:10.092954 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c36a352-151f-4e93-8094-3855bfed532e-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"5c36a352-151f-4e93-8094-3855bfed532e\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:27:10 crc kubenswrapper[4907]: I0313 14:27:10.092981 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5c36a352-151f-4e93-8094-3855bfed532e-scripts\") pod \"glance-default-internal-api-0\" (UID: \"5c36a352-151f-4e93-8094-3855bfed532e\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:27:10 crc kubenswrapper[4907]: I0313 14:27:10.093015 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5c36a352-151f-4e93-8094-3855bfed532e-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"5c36a352-151f-4e93-8094-3855bfed532e\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:27:10 crc kubenswrapper[4907]: I0313 14:27:10.093038 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5c36a352-151f-4e93-8094-3855bfed532e-logs\") pod \"glance-default-internal-api-0\" (UID: \"5c36a352-151f-4e93-8094-3855bfed532e\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:27:10 crc kubenswrapper[4907]: I0313 14:27:10.093061 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5c36a352-151f-4e93-8094-3855bfed532e-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"5c36a352-151f-4e93-8094-3855bfed532e\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:27:10 crc kubenswrapper[4907]: I0313 14:27:10.195001 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"5c36a352-151f-4e93-8094-3855bfed532e\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:27:10 crc kubenswrapper[4907]: I0313 14:27:10.195056 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34a0ea40-bdaa-4eef-8a28-263d8db74f10-config-data\") pod \"keystone-bootstrap-crrww\" (UID: \"34a0ea40-bdaa-4eef-8a28-263d8db74f10\") " pod="openstack/keystone-bootstrap-crrww" Mar 13 14:27:10 crc kubenswrapper[4907]: I0313 14:27:10.195100 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f7gh4\" (UniqueName: \"kubernetes.io/projected/5c36a352-151f-4e93-8094-3855bfed532e-kube-api-access-f7gh4\") pod \"glance-default-internal-api-0\" (UID: \"5c36a352-151f-4e93-8094-3855bfed532e\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:27:10 crc kubenswrapper[4907]: I0313 14:27:10.195131 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c36a352-151f-4e93-8094-3855bfed532e-config-data\") pod \"glance-default-internal-api-0\" (UID: \"5c36a352-151f-4e93-8094-3855bfed532e\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:27:10 crc kubenswrapper[4907]: I0313 14:27:10.195157 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c36a352-151f-4e93-8094-3855bfed532e-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"5c36a352-151f-4e93-8094-3855bfed532e\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:27:10 crc kubenswrapper[4907]: I0313 14:27:10.195187 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5c36a352-151f-4e93-8094-3855bfed532e-scripts\") pod \"glance-default-internal-api-0\" (UID: \"5c36a352-151f-4e93-8094-3855bfed532e\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:27:10 crc kubenswrapper[4907]: I0313 14:27:10.195239 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/34a0ea40-bdaa-4eef-8a28-263d8db74f10-fernet-keys\") pod \"keystone-bootstrap-crrww\" (UID: \"34a0ea40-bdaa-4eef-8a28-263d8db74f10\") " pod="openstack/keystone-bootstrap-crrww" Mar 13 14:27:10 crc kubenswrapper[4907]: I0313 14:27:10.195263 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5c36a352-151f-4e93-8094-3855bfed532e-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"5c36a352-151f-4e93-8094-3855bfed532e\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:27:10 crc kubenswrapper[4907]: I0313 14:27:10.195292 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8grrt\" (UniqueName: \"kubernetes.io/projected/34a0ea40-bdaa-4eef-8a28-263d8db74f10-kube-api-access-8grrt\") pod \"keystone-bootstrap-crrww\" (UID: \"34a0ea40-bdaa-4eef-8a28-263d8db74f10\") " pod="openstack/keystone-bootstrap-crrww" Mar 13 14:27:10 crc kubenswrapper[4907]: I0313 14:27:10.195312 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5c36a352-151f-4e93-8094-3855bfed532e-logs\") pod \"glance-default-internal-api-0\" (UID: \"5c36a352-151f-4e93-8094-3855bfed532e\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:27:10 crc kubenswrapper[4907]: I0313 14:27:10.195345 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5c36a352-151f-4e93-8094-3855bfed532e-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"5c36a352-151f-4e93-8094-3855bfed532e\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:27:10 crc kubenswrapper[4907]: I0313 14:27:10.195372 4907 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"5c36a352-151f-4e93-8094-3855bfed532e\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/glance-default-internal-api-0" Mar 13 14:27:10 crc kubenswrapper[4907]: I0313 14:27:10.195381 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34a0ea40-bdaa-4eef-8a28-263d8db74f10-combined-ca-bundle\") pod \"keystone-bootstrap-crrww\" (UID: \"34a0ea40-bdaa-4eef-8a28-263d8db74f10\") " pod="openstack/keystone-bootstrap-crrww" Mar 13 14:27:10 crc kubenswrapper[4907]: I0313 14:27:10.195494 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/34a0ea40-bdaa-4eef-8a28-263d8db74f10-credential-keys\") pod \"keystone-bootstrap-crrww\" (UID: \"34a0ea40-bdaa-4eef-8a28-263d8db74f10\") " pod="openstack/keystone-bootstrap-crrww" Mar 13 14:27:10 crc kubenswrapper[4907]: I0313 14:27:10.195539 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/34a0ea40-bdaa-4eef-8a28-263d8db74f10-scripts\") pod \"keystone-bootstrap-crrww\" (UID: \"34a0ea40-bdaa-4eef-8a28-263d8db74f10\") " pod="openstack/keystone-bootstrap-crrww" Mar 13 14:27:10 crc kubenswrapper[4907]: I0313 14:27:10.296975 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/34a0ea40-bdaa-4eef-8a28-263d8db74f10-scripts\") pod \"keystone-bootstrap-crrww\" (UID: \"34a0ea40-bdaa-4eef-8a28-263d8db74f10\") " pod="openstack/keystone-bootstrap-crrww" Mar 13 14:27:10 crc kubenswrapper[4907]: I0313 14:27:10.297076 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34a0ea40-bdaa-4eef-8a28-263d8db74f10-config-data\") pod \"keystone-bootstrap-crrww\" (UID: \"34a0ea40-bdaa-4eef-8a28-263d8db74f10\") " pod="openstack/keystone-bootstrap-crrww" Mar 13 14:27:10 crc kubenswrapper[4907]: I0313 14:27:10.297168 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/34a0ea40-bdaa-4eef-8a28-263d8db74f10-fernet-keys\") pod \"keystone-bootstrap-crrww\" (UID: \"34a0ea40-bdaa-4eef-8a28-263d8db74f10\") " pod="openstack/keystone-bootstrap-crrww" Mar 13 14:27:10 crc kubenswrapper[4907]: I0313 14:27:10.297192 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8grrt\" (UniqueName: \"kubernetes.io/projected/34a0ea40-bdaa-4eef-8a28-263d8db74f10-kube-api-access-8grrt\") pod \"keystone-bootstrap-crrww\" (UID: \"34a0ea40-bdaa-4eef-8a28-263d8db74f10\") " pod="openstack/keystone-bootstrap-crrww" Mar 13 14:27:10 crc kubenswrapper[4907]: I0313 14:27:10.297238 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34a0ea40-bdaa-4eef-8a28-263d8db74f10-combined-ca-bundle\") pod \"keystone-bootstrap-crrww\" (UID: \"34a0ea40-bdaa-4eef-8a28-263d8db74f10\") " pod="openstack/keystone-bootstrap-crrww" Mar 13 14:27:10 crc kubenswrapper[4907]: I0313 14:27:10.297254 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/34a0ea40-bdaa-4eef-8a28-263d8db74f10-credential-keys\") pod \"keystone-bootstrap-crrww\" (UID: \"34a0ea40-bdaa-4eef-8a28-263d8db74f10\") " pod="openstack/keystone-bootstrap-crrww" Mar 13 14:27:10 crc kubenswrapper[4907]: I0313 14:27:10.304433 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/34a0ea40-bdaa-4eef-8a28-263d8db74f10-scripts\") pod \"keystone-bootstrap-crrww\" (UID: \"34a0ea40-bdaa-4eef-8a28-263d8db74f10\") " pod="openstack/keystone-bootstrap-crrww" Mar 13 14:27:10 crc kubenswrapper[4907]: I0313 14:27:10.304539 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34a0ea40-bdaa-4eef-8a28-263d8db74f10-config-data\") pod \"keystone-bootstrap-crrww\" (UID: \"34a0ea40-bdaa-4eef-8a28-263d8db74f10\") " pod="openstack/keystone-bootstrap-crrww" Mar 13 14:27:10 crc kubenswrapper[4907]: I0313 14:27:10.304546 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/34a0ea40-bdaa-4eef-8a28-263d8db74f10-fernet-keys\") pod \"keystone-bootstrap-crrww\" (UID: \"34a0ea40-bdaa-4eef-8a28-263d8db74f10\") " pod="openstack/keystone-bootstrap-crrww" Mar 13 14:27:10 crc kubenswrapper[4907]: I0313 14:27:10.306014 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34a0ea40-bdaa-4eef-8a28-263d8db74f10-combined-ca-bundle\") pod \"keystone-bootstrap-crrww\" (UID: \"34a0ea40-bdaa-4eef-8a28-263d8db74f10\") " pod="openstack/keystone-bootstrap-crrww" Mar 13 14:27:10 crc kubenswrapper[4907]: I0313 14:27:10.307173 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/34a0ea40-bdaa-4eef-8a28-263d8db74f10-credential-keys\") pod \"keystone-bootstrap-crrww\" (UID: \"34a0ea40-bdaa-4eef-8a28-263d8db74f10\") " pod="openstack/keystone-bootstrap-crrww" Mar 13 14:27:10 crc kubenswrapper[4907]: I0313 14:27:10.317388 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8grrt\" (UniqueName: \"kubernetes.io/projected/34a0ea40-bdaa-4eef-8a28-263d8db74f10-kube-api-access-8grrt\") pod \"keystone-bootstrap-crrww\" (UID: \"34a0ea40-bdaa-4eef-8a28-263d8db74f10\") " pod="openstack/keystone-bootstrap-crrww" Mar 13 14:27:10 crc kubenswrapper[4907]: I0313 14:27:10.323409 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c36a352-151f-4e93-8094-3855bfed532e-config-data\") pod \"glance-default-internal-api-0\" (UID: \"5c36a352-151f-4e93-8094-3855bfed532e\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:27:10 crc kubenswrapper[4907]: I0313 14:27:10.323450 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5c36a352-151f-4e93-8094-3855bfed532e-logs\") pod \"glance-default-internal-api-0\" (UID: \"5c36a352-151f-4e93-8094-3855bfed532e\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:27:10 crc kubenswrapper[4907]: I0313 14:27:10.323627 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c36a352-151f-4e93-8094-3855bfed532e-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"5c36a352-151f-4e93-8094-3855bfed532e\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:27:10 crc kubenswrapper[4907]: I0313 14:27:10.325190 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5c36a352-151f-4e93-8094-3855bfed532e-scripts\") pod \"glance-default-internal-api-0\" (UID: \"5c36a352-151f-4e93-8094-3855bfed532e\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:27:10 crc kubenswrapper[4907]: I0313 14:27:10.327142 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5c36a352-151f-4e93-8094-3855bfed532e-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"5c36a352-151f-4e93-8094-3855bfed532e\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:27:10 crc kubenswrapper[4907]: I0313 14:27:10.329361 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5c36a352-151f-4e93-8094-3855bfed532e-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"5c36a352-151f-4e93-8094-3855bfed532e\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:27:10 crc kubenswrapper[4907]: I0313 14:27:10.330520 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f7gh4\" (UniqueName: \"kubernetes.io/projected/5c36a352-151f-4e93-8094-3855bfed532e-kube-api-access-f7gh4\") pod \"glance-default-internal-api-0\" (UID: \"5c36a352-151f-4e93-8094-3855bfed532e\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:27:10 crc kubenswrapper[4907]: I0313 14:27:10.354987 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"5c36a352-151f-4e93-8094-3855bfed532e\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:27:10 crc kubenswrapper[4907]: I0313 14:27:10.413990 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-crrww" Mar 13 14:27:10 crc kubenswrapper[4907]: I0313 14:27:10.572594 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 13 14:27:11 crc kubenswrapper[4907]: I0313 14:27:11.638701 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 13 14:27:11 crc kubenswrapper[4907]: I0313 14:27:11.719787 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ce2a993a-c706-400f-967d-35a9c38d5937-scripts\") pod \"ce2a993a-c706-400f-967d-35a9c38d5937\" (UID: \"ce2a993a-c706-400f-967d-35a9c38d5937\") " Mar 13 14:27:11 crc kubenswrapper[4907]: I0313 14:27:11.719839 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce2a993a-c706-400f-967d-35a9c38d5937-config-data\") pod \"ce2a993a-c706-400f-967d-35a9c38d5937\" (UID: \"ce2a993a-c706-400f-967d-35a9c38d5937\") " Mar 13 14:27:11 crc kubenswrapper[4907]: I0313 14:27:11.719867 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ce2a993a-c706-400f-967d-35a9c38d5937-logs\") pod \"ce2a993a-c706-400f-967d-35a9c38d5937\" (UID: \"ce2a993a-c706-400f-967d-35a9c38d5937\") " Mar 13 14:27:11 crc kubenswrapper[4907]: I0313 14:27:11.719908 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ce2a993a-c706-400f-967d-35a9c38d5937-public-tls-certs\") pod \"ce2a993a-c706-400f-967d-35a9c38d5937\" (UID: \"ce2a993a-c706-400f-967d-35a9c38d5937\") " Mar 13 14:27:11 crc kubenswrapper[4907]: I0313 14:27:11.719926 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jxx26\" (UniqueName: \"kubernetes.io/projected/ce2a993a-c706-400f-967d-35a9c38d5937-kube-api-access-jxx26\") pod \"ce2a993a-c706-400f-967d-35a9c38d5937\" (UID: \"ce2a993a-c706-400f-967d-35a9c38d5937\") " Mar 13 14:27:11 crc kubenswrapper[4907]: I0313 14:27:11.719962 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ce2a993a-c706-400f-967d-35a9c38d5937-httpd-run\") pod \"ce2a993a-c706-400f-967d-35a9c38d5937\" (UID: \"ce2a993a-c706-400f-967d-35a9c38d5937\") " Mar 13 14:27:11 crc kubenswrapper[4907]: I0313 14:27:11.720043 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce2a993a-c706-400f-967d-35a9c38d5937-combined-ca-bundle\") pod \"ce2a993a-c706-400f-967d-35a9c38d5937\" (UID: \"ce2a993a-c706-400f-967d-35a9c38d5937\") " Mar 13 14:27:11 crc kubenswrapper[4907]: I0313 14:27:11.720062 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ce2a993a-c706-400f-967d-35a9c38d5937\" (UID: \"ce2a993a-c706-400f-967d-35a9c38d5937\") " Mar 13 14:27:11 crc kubenswrapper[4907]: I0313 14:27:11.721382 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ce2a993a-c706-400f-967d-35a9c38d5937-logs" (OuterVolumeSpecName: "logs") pod "ce2a993a-c706-400f-967d-35a9c38d5937" (UID: "ce2a993a-c706-400f-967d-35a9c38d5937"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:27:11 crc kubenswrapper[4907]: I0313 14:27:11.721816 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ce2a993a-c706-400f-967d-35a9c38d5937-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "ce2a993a-c706-400f-967d-35a9c38d5937" (UID: "ce2a993a-c706-400f-967d-35a9c38d5937"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:27:11 crc kubenswrapper[4907]: I0313 14:27:11.725937 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ce2a993a-c706-400f-967d-35a9c38d5937-kube-api-access-jxx26" (OuterVolumeSpecName: "kube-api-access-jxx26") pod "ce2a993a-c706-400f-967d-35a9c38d5937" (UID: "ce2a993a-c706-400f-967d-35a9c38d5937"). InnerVolumeSpecName "kube-api-access-jxx26". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:27:11 crc kubenswrapper[4907]: I0313 14:27:11.726028 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce2a993a-c706-400f-967d-35a9c38d5937-scripts" (OuterVolumeSpecName: "scripts") pod "ce2a993a-c706-400f-967d-35a9c38d5937" (UID: "ce2a993a-c706-400f-967d-35a9c38d5937"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:27:11 crc kubenswrapper[4907]: I0313 14:27:11.754901 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "glance") pod "ce2a993a-c706-400f-967d-35a9c38d5937" (UID: "ce2a993a-c706-400f-967d-35a9c38d5937"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Mar 13 14:27:11 crc kubenswrapper[4907]: I0313 14:27:11.772190 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce2a993a-c706-400f-967d-35a9c38d5937-config-data" (OuterVolumeSpecName: "config-data") pod "ce2a993a-c706-400f-967d-35a9c38d5937" (UID: "ce2a993a-c706-400f-967d-35a9c38d5937"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:27:11 crc kubenswrapper[4907]: I0313 14:27:11.773081 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce2a993a-c706-400f-967d-35a9c38d5937-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ce2a993a-c706-400f-967d-35a9c38d5937" (UID: "ce2a993a-c706-400f-967d-35a9c38d5937"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:27:11 crc kubenswrapper[4907]: I0313 14:27:11.787112 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce2a993a-c706-400f-967d-35a9c38d5937-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "ce2a993a-c706-400f-967d-35a9c38d5937" (UID: "ce2a993a-c706-400f-967d-35a9c38d5937"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:27:11 crc kubenswrapper[4907]: I0313 14:27:11.794226 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4a97d1cc-3762-47fe-a8fe-ed9e95b466a4" path="/var/lib/kubelet/pods/4a97d1cc-3762-47fe-a8fe-ed9e95b466a4/volumes" Mar 13 14:27:11 crc kubenswrapper[4907]: I0313 14:27:11.795213 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5" path="/var/lib/kubelet/pods/8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5/volumes" Mar 13 14:27:11 crc kubenswrapper[4907]: I0313 14:27:11.822736 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce2a993a-c706-400f-967d-35a9c38d5937-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:11 crc kubenswrapper[4907]: I0313 14:27:11.822798 4907 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Mar 13 14:27:11 crc kubenswrapper[4907]: I0313 14:27:11.822811 4907 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ce2a993a-c706-400f-967d-35a9c38d5937-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:11 crc kubenswrapper[4907]: I0313 14:27:11.822825 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce2a993a-c706-400f-967d-35a9c38d5937-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:11 crc kubenswrapper[4907]: I0313 14:27:11.822836 4907 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ce2a993a-c706-400f-967d-35a9c38d5937-logs\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:11 crc kubenswrapper[4907]: I0313 14:27:11.822846 4907 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ce2a993a-c706-400f-967d-35a9c38d5937-public-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:11 crc kubenswrapper[4907]: I0313 14:27:11.822856 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jxx26\" (UniqueName: \"kubernetes.io/projected/ce2a993a-c706-400f-967d-35a9c38d5937-kube-api-access-jxx26\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:11 crc kubenswrapper[4907]: I0313 14:27:11.822867 4907 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ce2a993a-c706-400f-967d-35a9c38d5937-httpd-run\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:11 crc kubenswrapper[4907]: I0313 14:27:11.843522 4907 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Mar 13 14:27:11 crc kubenswrapper[4907]: I0313 14:27:11.899519 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"ce2a993a-c706-400f-967d-35a9c38d5937","Type":"ContainerDied","Data":"70f4055057ae74811d3e519a30dbdf9559ebe320aa1f7b52ebbfd8cfd47c2673"} Mar 13 14:27:11 crc kubenswrapper[4907]: I0313 14:27:11.899620 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 13 14:27:11 crc kubenswrapper[4907]: I0313 14:27:11.924129 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 13 14:27:11 crc kubenswrapper[4907]: I0313 14:27:11.924530 4907 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:11 crc kubenswrapper[4907]: I0313 14:27:11.936983 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 13 14:27:11 crc kubenswrapper[4907]: I0313 14:27:11.983134 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Mar 13 14:27:11 crc kubenswrapper[4907]: E0313 14:27:11.983575 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce2a993a-c706-400f-967d-35a9c38d5937" containerName="glance-httpd" Mar 13 14:27:11 crc kubenswrapper[4907]: I0313 14:27:11.983597 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce2a993a-c706-400f-967d-35a9c38d5937" containerName="glance-httpd" Mar 13 14:27:11 crc kubenswrapper[4907]: E0313 14:27:11.983639 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce2a993a-c706-400f-967d-35a9c38d5937" containerName="glance-log" Mar 13 14:27:11 crc kubenswrapper[4907]: I0313 14:27:11.983648 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce2a993a-c706-400f-967d-35a9c38d5937" containerName="glance-log" Mar 13 14:27:11 crc kubenswrapper[4907]: I0313 14:27:11.983813 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce2a993a-c706-400f-967d-35a9c38d5937" containerName="glance-httpd" Mar 13 14:27:11 crc kubenswrapper[4907]: I0313 14:27:11.983843 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce2a993a-c706-400f-967d-35a9c38d5937" containerName="glance-log" Mar 13 14:27:11 crc kubenswrapper[4907]: I0313 14:27:11.984743 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 13 14:27:11 crc kubenswrapper[4907]: I0313 14:27:11.987456 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Mar 13 14:27:11 crc kubenswrapper[4907]: I0313 14:27:11.987604 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Mar 13 14:27:11 crc kubenswrapper[4907]: I0313 14:27:11.992105 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 13 14:27:12 crc kubenswrapper[4907]: I0313 14:27:12.129853 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"2c9261af-9eaa-4623-9f1b-719b58a9d3a2\") " pod="openstack/glance-default-external-api-0" Mar 13 14:27:12 crc kubenswrapper[4907]: I0313 14:27:12.129927 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7fdtd\" (UniqueName: \"kubernetes.io/projected/2c9261af-9eaa-4623-9f1b-719b58a9d3a2-kube-api-access-7fdtd\") pod \"glance-default-external-api-0\" (UID: \"2c9261af-9eaa-4623-9f1b-719b58a9d3a2\") " pod="openstack/glance-default-external-api-0" Mar 13 14:27:12 crc kubenswrapper[4907]: I0313 14:27:12.129984 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2c9261af-9eaa-4623-9f1b-719b58a9d3a2-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"2c9261af-9eaa-4623-9f1b-719b58a9d3a2\") " pod="openstack/glance-default-external-api-0" Mar 13 14:27:12 crc kubenswrapper[4907]: I0313 14:27:12.130129 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2c9261af-9eaa-4623-9f1b-719b58a9d3a2-scripts\") pod \"glance-default-external-api-0\" (UID: \"2c9261af-9eaa-4623-9f1b-719b58a9d3a2\") " pod="openstack/glance-default-external-api-0" Mar 13 14:27:12 crc kubenswrapper[4907]: I0313 14:27:12.130164 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2c9261af-9eaa-4623-9f1b-719b58a9d3a2-logs\") pod \"glance-default-external-api-0\" (UID: \"2c9261af-9eaa-4623-9f1b-719b58a9d3a2\") " pod="openstack/glance-default-external-api-0" Mar 13 14:27:12 crc kubenswrapper[4907]: I0313 14:27:12.130202 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c9261af-9eaa-4623-9f1b-719b58a9d3a2-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"2c9261af-9eaa-4623-9f1b-719b58a9d3a2\") " pod="openstack/glance-default-external-api-0" Mar 13 14:27:12 crc kubenswrapper[4907]: I0313 14:27:12.130308 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2c9261af-9eaa-4623-9f1b-719b58a9d3a2-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"2c9261af-9eaa-4623-9f1b-719b58a9d3a2\") " pod="openstack/glance-default-external-api-0" Mar 13 14:27:12 crc kubenswrapper[4907]: I0313 14:27:12.130431 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c9261af-9eaa-4623-9f1b-719b58a9d3a2-config-data\") pod \"glance-default-external-api-0\" (UID: \"2c9261af-9eaa-4623-9f1b-719b58a9d3a2\") " pod="openstack/glance-default-external-api-0" Mar 13 14:27:12 crc kubenswrapper[4907]: I0313 14:27:12.232593 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c9261af-9eaa-4623-9f1b-719b58a9d3a2-config-data\") pod \"glance-default-external-api-0\" (UID: \"2c9261af-9eaa-4623-9f1b-719b58a9d3a2\") " pod="openstack/glance-default-external-api-0" Mar 13 14:27:12 crc kubenswrapper[4907]: I0313 14:27:12.232717 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"2c9261af-9eaa-4623-9f1b-719b58a9d3a2\") " pod="openstack/glance-default-external-api-0" Mar 13 14:27:12 crc kubenswrapper[4907]: I0313 14:27:12.232740 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7fdtd\" (UniqueName: \"kubernetes.io/projected/2c9261af-9eaa-4623-9f1b-719b58a9d3a2-kube-api-access-7fdtd\") pod \"glance-default-external-api-0\" (UID: \"2c9261af-9eaa-4623-9f1b-719b58a9d3a2\") " pod="openstack/glance-default-external-api-0" Mar 13 14:27:12 crc kubenswrapper[4907]: I0313 14:27:12.232777 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2c9261af-9eaa-4623-9f1b-719b58a9d3a2-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"2c9261af-9eaa-4623-9f1b-719b58a9d3a2\") " pod="openstack/glance-default-external-api-0" Mar 13 14:27:12 crc kubenswrapper[4907]: I0313 14:27:12.232820 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2c9261af-9eaa-4623-9f1b-719b58a9d3a2-scripts\") pod \"glance-default-external-api-0\" (UID: \"2c9261af-9eaa-4623-9f1b-719b58a9d3a2\") " pod="openstack/glance-default-external-api-0" Mar 13 14:27:12 crc kubenswrapper[4907]: I0313 14:27:12.232850 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2c9261af-9eaa-4623-9f1b-719b58a9d3a2-logs\") pod \"glance-default-external-api-0\" (UID: \"2c9261af-9eaa-4623-9f1b-719b58a9d3a2\") " pod="openstack/glance-default-external-api-0" Mar 13 14:27:12 crc kubenswrapper[4907]: I0313 14:27:12.232939 4907 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"2c9261af-9eaa-4623-9f1b-719b58a9d3a2\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/glance-default-external-api-0" Mar 13 14:27:12 crc kubenswrapper[4907]: I0313 14:27:12.235864 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2c9261af-9eaa-4623-9f1b-719b58a9d3a2-logs\") pod \"glance-default-external-api-0\" (UID: \"2c9261af-9eaa-4623-9f1b-719b58a9d3a2\") " pod="openstack/glance-default-external-api-0" Mar 13 14:27:12 crc kubenswrapper[4907]: I0313 14:27:12.236022 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c9261af-9eaa-4623-9f1b-719b58a9d3a2-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"2c9261af-9eaa-4623-9f1b-719b58a9d3a2\") " pod="openstack/glance-default-external-api-0" Mar 13 14:27:12 crc kubenswrapper[4907]: I0313 14:27:12.236076 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2c9261af-9eaa-4623-9f1b-719b58a9d3a2-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"2c9261af-9eaa-4623-9f1b-719b58a9d3a2\") " pod="openstack/glance-default-external-api-0" Mar 13 14:27:12 crc kubenswrapper[4907]: I0313 14:27:12.236587 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2c9261af-9eaa-4623-9f1b-719b58a9d3a2-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"2c9261af-9eaa-4623-9f1b-719b58a9d3a2\") " pod="openstack/glance-default-external-api-0" Mar 13 14:27:12 crc kubenswrapper[4907]: I0313 14:27:12.238453 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2c9261af-9eaa-4623-9f1b-719b58a9d3a2-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"2c9261af-9eaa-4623-9f1b-719b58a9d3a2\") " pod="openstack/glance-default-external-api-0" Mar 13 14:27:12 crc kubenswrapper[4907]: I0313 14:27:12.240179 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c9261af-9eaa-4623-9f1b-719b58a9d3a2-config-data\") pod \"glance-default-external-api-0\" (UID: \"2c9261af-9eaa-4623-9f1b-719b58a9d3a2\") " pod="openstack/glance-default-external-api-0" Mar 13 14:27:12 crc kubenswrapper[4907]: I0313 14:27:12.243456 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2c9261af-9eaa-4623-9f1b-719b58a9d3a2-scripts\") pod \"glance-default-external-api-0\" (UID: \"2c9261af-9eaa-4623-9f1b-719b58a9d3a2\") " pod="openstack/glance-default-external-api-0" Mar 13 14:27:12 crc kubenswrapper[4907]: I0313 14:27:12.247040 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c9261af-9eaa-4623-9f1b-719b58a9d3a2-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"2c9261af-9eaa-4623-9f1b-719b58a9d3a2\") " pod="openstack/glance-default-external-api-0" Mar 13 14:27:12 crc kubenswrapper[4907]: I0313 14:27:12.249267 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7fdtd\" (UniqueName: \"kubernetes.io/projected/2c9261af-9eaa-4623-9f1b-719b58a9d3a2-kube-api-access-7fdtd\") pod \"glance-default-external-api-0\" (UID: \"2c9261af-9eaa-4623-9f1b-719b58a9d3a2\") " pod="openstack/glance-default-external-api-0" Mar 13 14:27:12 crc kubenswrapper[4907]: I0313 14:27:12.262295 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"2c9261af-9eaa-4623-9f1b-719b58a9d3a2\") " pod="openstack/glance-default-external-api-0" Mar 13 14:27:12 crc kubenswrapper[4907]: I0313 14:27:12.305651 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 13 14:27:13 crc kubenswrapper[4907]: I0313 14:27:13.791521 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ce2a993a-c706-400f-967d-35a9c38d5937" path="/var/lib/kubelet/pods/ce2a993a-c706-400f-967d-35a9c38d5937/volumes" Mar 13 14:27:18 crc kubenswrapper[4907]: I0313 14:27:18.662997 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67754df655-k2v49" Mar 13 14:27:18 crc kubenswrapper[4907]: I0313 14:27:18.750170 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vpx9g\" (UniqueName: \"kubernetes.io/projected/c065ce46-d7f3-4e50-b30c-83a1592b6ead-kube-api-access-vpx9g\") pod \"c065ce46-d7f3-4e50-b30c-83a1592b6ead\" (UID: \"c065ce46-d7f3-4e50-b30c-83a1592b6ead\") " Mar 13 14:27:18 crc kubenswrapper[4907]: I0313 14:27:18.750241 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c065ce46-d7f3-4e50-b30c-83a1592b6ead-ovsdbserver-sb\") pod \"c065ce46-d7f3-4e50-b30c-83a1592b6ead\" (UID: \"c065ce46-d7f3-4e50-b30c-83a1592b6ead\") " Mar 13 14:27:18 crc kubenswrapper[4907]: I0313 14:27:18.750315 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c065ce46-d7f3-4e50-b30c-83a1592b6ead-ovsdbserver-nb\") pod \"c065ce46-d7f3-4e50-b30c-83a1592b6ead\" (UID: \"c065ce46-d7f3-4e50-b30c-83a1592b6ead\") " Mar 13 14:27:18 crc kubenswrapper[4907]: I0313 14:27:18.750400 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c065ce46-d7f3-4e50-b30c-83a1592b6ead-dns-swift-storage-0\") pod \"c065ce46-d7f3-4e50-b30c-83a1592b6ead\" (UID: \"c065ce46-d7f3-4e50-b30c-83a1592b6ead\") " Mar 13 14:27:18 crc kubenswrapper[4907]: I0313 14:27:18.750461 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c065ce46-d7f3-4e50-b30c-83a1592b6ead-dns-svc\") pod \"c065ce46-d7f3-4e50-b30c-83a1592b6ead\" (UID: \"c065ce46-d7f3-4e50-b30c-83a1592b6ead\") " Mar 13 14:27:18 crc kubenswrapper[4907]: I0313 14:27:18.750497 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c065ce46-d7f3-4e50-b30c-83a1592b6ead-config\") pod \"c065ce46-d7f3-4e50-b30c-83a1592b6ead\" (UID: \"c065ce46-d7f3-4e50-b30c-83a1592b6ead\") " Mar 13 14:27:18 crc kubenswrapper[4907]: I0313 14:27:18.756040 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c065ce46-d7f3-4e50-b30c-83a1592b6ead-kube-api-access-vpx9g" (OuterVolumeSpecName: "kube-api-access-vpx9g") pod "c065ce46-d7f3-4e50-b30c-83a1592b6ead" (UID: "c065ce46-d7f3-4e50-b30c-83a1592b6ead"). InnerVolumeSpecName "kube-api-access-vpx9g". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:27:18 crc kubenswrapper[4907]: I0313 14:27:18.800904 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c065ce46-d7f3-4e50-b30c-83a1592b6ead-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "c065ce46-d7f3-4e50-b30c-83a1592b6ead" (UID: "c065ce46-d7f3-4e50-b30c-83a1592b6ead"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:27:18 crc kubenswrapper[4907]: I0313 14:27:18.809026 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c065ce46-d7f3-4e50-b30c-83a1592b6ead-config" (OuterVolumeSpecName: "config") pod "c065ce46-d7f3-4e50-b30c-83a1592b6ead" (UID: "c065ce46-d7f3-4e50-b30c-83a1592b6ead"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:27:18 crc kubenswrapper[4907]: I0313 14:27:18.809042 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c065ce46-d7f3-4e50-b30c-83a1592b6ead-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c065ce46-d7f3-4e50-b30c-83a1592b6ead" (UID: "c065ce46-d7f3-4e50-b30c-83a1592b6ead"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:27:18 crc kubenswrapper[4907]: I0313 14:27:18.813361 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c065ce46-d7f3-4e50-b30c-83a1592b6ead-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "c065ce46-d7f3-4e50-b30c-83a1592b6ead" (UID: "c065ce46-d7f3-4e50-b30c-83a1592b6ead"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:27:18 crc kubenswrapper[4907]: I0313 14:27:18.815247 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c065ce46-d7f3-4e50-b30c-83a1592b6ead-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "c065ce46-d7f3-4e50-b30c-83a1592b6ead" (UID: "c065ce46-d7f3-4e50-b30c-83a1592b6ead"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:27:18 crc kubenswrapper[4907]: I0313 14:27:18.853168 4907 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c065ce46-d7f3-4e50-b30c-83a1592b6ead-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:18 crc kubenswrapper[4907]: I0313 14:27:18.853223 4907 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c065ce46-d7f3-4e50-b30c-83a1592b6ead-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:18 crc kubenswrapper[4907]: I0313 14:27:18.853234 4907 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c065ce46-d7f3-4e50-b30c-83a1592b6ead-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:18 crc kubenswrapper[4907]: I0313 14:27:18.853289 4907 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c065ce46-d7f3-4e50-b30c-83a1592b6ead-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:18 crc kubenswrapper[4907]: I0313 14:27:18.853321 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c065ce46-d7f3-4e50-b30c-83a1592b6ead-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:18 crc kubenswrapper[4907]: I0313 14:27:18.853335 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vpx9g\" (UniqueName: \"kubernetes.io/projected/c065ce46-d7f3-4e50-b30c-83a1592b6ead-kube-api-access-vpx9g\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:18 crc kubenswrapper[4907]: I0313 14:27:18.969816 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67754df655-k2v49" event={"ID":"c065ce46-d7f3-4e50-b30c-83a1592b6ead","Type":"ContainerDied","Data":"1dc89b2197970bff5fe57426501341e047b5f1d0b4ba93884702cc3864ea9041"} Mar 13 14:27:18 crc kubenswrapper[4907]: I0313 14:27:18.969953 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67754df655-k2v49" Mar 13 14:27:19 crc kubenswrapper[4907]: I0313 14:27:19.003112 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-67754df655-k2v49"] Mar 13 14:27:19 crc kubenswrapper[4907]: I0313 14:27:19.011413 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-67754df655-k2v49"] Mar 13 14:27:19 crc kubenswrapper[4907]: I0313 14:27:19.719549 4907 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-67754df655-k2v49" podUID="c065ce46-d7f3-4e50-b30c-83a1592b6ead" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.133:5353: i/o timeout" Mar 13 14:27:19 crc kubenswrapper[4907]: I0313 14:27:19.792292 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c065ce46-d7f3-4e50-b30c-83a1592b6ead" path="/var/lib/kubelet/pods/c065ce46-d7f3-4e50-b30c-83a1592b6ead/volumes" Mar 13 14:27:20 crc kubenswrapper[4907]: I0313 14:27:20.931463 4907 scope.go:117] "RemoveContainer" containerID="d56b748bd6482c49c5b182589b671f235af50f54159cfd34d117c3100423cb0f" Mar 13 14:27:20 crc kubenswrapper[4907]: E0313 14:27:20.964361 4907 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:574a17f0877c175128a764f2b37fc02456649c8514689125718ce6ca974bfb6b" Mar 13 14:27:20 crc kubenswrapper[4907]: E0313 14:27:20.964497 4907 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:574a17f0877c175128a764f2b37fc02456649c8514689125718ce6ca974bfb6b,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-wpft2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-7q999_openstack(594b61fa-12b7-47b6-8af3-17f024e0d54d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 13 14:27:20 crc kubenswrapper[4907]: E0313 14:27:20.965644 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-7q999" podUID="594b61fa-12b7-47b6-8af3-17f024e0d54d" Mar 13 14:27:20 crc kubenswrapper[4907]: E0313 14:27:20.994704 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:574a17f0877c175128a764f2b37fc02456649c8514689125718ce6ca974bfb6b\\\"\"" pod="openstack/cinder-db-sync-7q999" podUID="594b61fa-12b7-47b6-8af3-17f024e0d54d" Mar 13 14:27:21 crc kubenswrapper[4907]: I0313 14:27:21.146806 4907 scope.go:117] "RemoveContainer" containerID="c604cb7d84d59e89fd67f7b2707cc5e5fce549a56d035662c24b9cb3f930d149" Mar 13 14:27:21 crc kubenswrapper[4907]: I0313 14:27:21.201298 4907 scope.go:117] "RemoveContainer" containerID="512291676abb06ed48b9f163f1a7202f5fd9b599abc8f6b456c67f77c909e778" Mar 13 14:27:21 crc kubenswrapper[4907]: I0313 14:27:21.250705 4907 scope.go:117] "RemoveContainer" containerID="fd0a2c793e62b5b128e16861d049a22d36d9baa0de3410f3961ca59044ad1213" Mar 13 14:27:21 crc kubenswrapper[4907]: I0313 14:27:21.286679 4907 scope.go:117] "RemoveContainer" containerID="fc5cca7a53563949cc0a2a52bb6bc5ab553013ec2df261991d02d7a2bd1a082d" Mar 13 14:27:21 crc kubenswrapper[4907]: I0313 14:27:21.412607 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-crrww"] Mar 13 14:27:21 crc kubenswrapper[4907]: W0313 14:27:21.415533 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod34a0ea40_bdaa_4eef_8a28_263d8db74f10.slice/crio-35ce55e83a38917576782260db545e824c3f1f351d75a47429e1e2d417a299bb WatchSource:0}: Error finding container 35ce55e83a38917576782260db545e824c3f1f351d75a47429e1e2d417a299bb: Status 404 returned error can't find the container with id 35ce55e83a38917576782260db545e824c3f1f351d75a47429e1e2d417a299bb Mar 13 14:27:21 crc kubenswrapper[4907]: I0313 14:27:21.586947 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 13 14:27:21 crc kubenswrapper[4907]: W0313 14:27:21.589063 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5c36a352_151f_4e93_8094_3855bfed532e.slice/crio-4cb91fefa1e0ea486010258682ba6df5f6a7f72f9a8eaedd21f297ae14bacf23 WatchSource:0}: Error finding container 4cb91fefa1e0ea486010258682ba6df5f6a7f72f9a8eaedd21f297ae14bacf23: Status 404 returned error can't find the container with id 4cb91fefa1e0ea486010258682ba6df5f6a7f72f9a8eaedd21f297ae14bacf23 Mar 13 14:27:21 crc kubenswrapper[4907]: I0313 14:27:21.692018 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 13 14:27:22 crc kubenswrapper[4907]: I0313 14:27:22.008646 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-rz6qh" event={"ID":"32b531c6-64c1-4137-b82e-3c10789325e6","Type":"ContainerStarted","Data":"5b6a2afb9b4ad5f576a34ca212d94796e12f98db7c7c3a2d4bb1600fc3d60034"} Mar 13 14:27:22 crc kubenswrapper[4907]: I0313 14:27:22.015436 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-jjqr9" event={"ID":"b144897c-8640-4c5c-afa1-68b44dfd4f98","Type":"ContainerStarted","Data":"70df9585a66270db2bec365cac2492109be5b468d9fc1bab776dec0ef7304eaa"} Mar 13 14:27:22 crc kubenswrapper[4907]: I0313 14:27:22.027532 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-crrww" event={"ID":"34a0ea40-bdaa-4eef-8a28-263d8db74f10","Type":"ContainerStarted","Data":"f407f7b1e1671cab73872a73a57d80595693b20473de90bda4d6de97abc60cd5"} Mar 13 14:27:22 crc kubenswrapper[4907]: I0313 14:27:22.027573 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-crrww" event={"ID":"34a0ea40-bdaa-4eef-8a28-263d8db74f10","Type":"ContainerStarted","Data":"35ce55e83a38917576782260db545e824c3f1f351d75a47429e1e2d417a299bb"} Mar 13 14:27:22 crc kubenswrapper[4907]: I0313 14:27:22.047367 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5","Type":"ContainerStarted","Data":"b9aab9a1290561a45c480e877db9222806553cdd6cfb12daf299ab965ad2ec5c"} Mar 13 14:27:22 crc kubenswrapper[4907]: I0313 14:27:22.050407 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-rz6qh" podStartSLOduration=2.827143362 podStartE2EDuration="25.050385457s" podCreationTimestamp="2026-03-13 14:26:57 +0000 UTC" firstStartedPulling="2026-03-13 14:26:58.736334635 +0000 UTC m=+1317.636122324" lastFinishedPulling="2026-03-13 14:27:20.95957673 +0000 UTC m=+1339.859364419" observedRunningTime="2026-03-13 14:27:22.040599181 +0000 UTC m=+1340.940386870" watchObservedRunningTime="2026-03-13 14:27:22.050385457 +0000 UTC m=+1340.950173156" Mar 13 14:27:22 crc kubenswrapper[4907]: I0313 14:27:22.050602 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2c9261af-9eaa-4623-9f1b-719b58a9d3a2","Type":"ContainerStarted","Data":"2cde1712ac814253467a7ebeac79ad28f1b6bc8f06f7979a4876a959f01886bd"} Mar 13 14:27:22 crc kubenswrapper[4907]: I0313 14:27:22.051945 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5c36a352-151f-4e93-8094-3855bfed532e","Type":"ContainerStarted","Data":"4cb91fefa1e0ea486010258682ba6df5f6a7f72f9a8eaedd21f297ae14bacf23"} Mar 13 14:27:22 crc kubenswrapper[4907]: I0313 14:27:22.065713 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-crrww" podStartSLOduration=12.065692364 podStartE2EDuration="12.065692364s" podCreationTimestamp="2026-03-13 14:27:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:27:22.06115099 +0000 UTC m=+1340.960938679" watchObservedRunningTime="2026-03-13 14:27:22.065692364 +0000 UTC m=+1340.965480053" Mar 13 14:27:22 crc kubenswrapper[4907]: I0313 14:27:22.080796 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-jjqr9" podStartSLOduration=2.953097157 podStartE2EDuration="25.080778144s" podCreationTimestamp="2026-03-13 14:26:57 +0000 UTC" firstStartedPulling="2026-03-13 14:26:58.808536038 +0000 UTC m=+1317.708323727" lastFinishedPulling="2026-03-13 14:27:20.936217025 +0000 UTC m=+1339.836004714" observedRunningTime="2026-03-13 14:27:22.078459401 +0000 UTC m=+1340.978247090" watchObservedRunningTime="2026-03-13 14:27:22.080778144 +0000 UTC m=+1340.980565833" Mar 13 14:27:23 crc kubenswrapper[4907]: I0313 14:27:23.067209 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5","Type":"ContainerStarted","Data":"af56a2f9ef345a9cf4a63dafe33dfb02aa9518339a5d2ef4a11fbf9ec8ab7512"} Mar 13 14:27:23 crc kubenswrapper[4907]: I0313 14:27:23.070969 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2c9261af-9eaa-4623-9f1b-719b58a9d3a2","Type":"ContainerStarted","Data":"6af4b6388e126ccd4e3d2cc04b6747e5283033aaa5d424f9e2907953b3f55ea5"} Mar 13 14:27:23 crc kubenswrapper[4907]: I0313 14:27:23.074454 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5c36a352-151f-4e93-8094-3855bfed532e","Type":"ContainerStarted","Data":"4364cee9bed454e59183eb0004b9dfc4ab15dc5a345acbfd2a54032342758450"} Mar 13 14:27:24 crc kubenswrapper[4907]: I0313 14:27:24.084566 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2c9261af-9eaa-4623-9f1b-719b58a9d3a2","Type":"ContainerStarted","Data":"22707601406e7d79a6ae3140c68d6ec6d23c0168f7d4a1523dfe2e817903318f"} Mar 13 14:27:24 crc kubenswrapper[4907]: I0313 14:27:24.087193 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5c36a352-151f-4e93-8094-3855bfed532e","Type":"ContainerStarted","Data":"456e65b2b22522c83f37e28d27a38f3af4c42d273550f8c8c1246de8d14f0a2f"} Mar 13 14:27:24 crc kubenswrapper[4907]: I0313 14:27:24.131336 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=15.131318393 podStartE2EDuration="15.131318393s" podCreationTimestamp="2026-03-13 14:27:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:27:24.13011225 +0000 UTC m=+1343.029899939" watchObservedRunningTime="2026-03-13 14:27:24.131318393 +0000 UTC m=+1343.031106082" Mar 13 14:27:24 crc kubenswrapper[4907]: I0313 14:27:24.131700 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=13.131695053 podStartE2EDuration="13.131695053s" podCreationTimestamp="2026-03-13 14:27:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:27:24.106489078 +0000 UTC m=+1343.006276767" watchObservedRunningTime="2026-03-13 14:27:24.131695053 +0000 UTC m=+1343.031482742" Mar 13 14:27:25 crc kubenswrapper[4907]: I0313 14:27:25.112003 4907 generic.go:334] "Generic (PLEG): container finished" podID="32b531c6-64c1-4137-b82e-3c10789325e6" containerID="5b6a2afb9b4ad5f576a34ca212d94796e12f98db7c7c3a2d4bb1600fc3d60034" exitCode=0 Mar 13 14:27:25 crc kubenswrapper[4907]: I0313 14:27:25.112183 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-rz6qh" event={"ID":"32b531c6-64c1-4137-b82e-3c10789325e6","Type":"ContainerDied","Data":"5b6a2afb9b4ad5f576a34ca212d94796e12f98db7c7c3a2d4bb1600fc3d60034"} Mar 13 14:27:25 crc kubenswrapper[4907]: I0313 14:27:25.117089 4907 generic.go:334] "Generic (PLEG): container finished" podID="b144897c-8640-4c5c-afa1-68b44dfd4f98" containerID="70df9585a66270db2bec365cac2492109be5b468d9fc1bab776dec0ef7304eaa" exitCode=0 Mar 13 14:27:25 crc kubenswrapper[4907]: I0313 14:27:25.117169 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-jjqr9" event={"ID":"b144897c-8640-4c5c-afa1-68b44dfd4f98","Type":"ContainerDied","Data":"70df9585a66270db2bec365cac2492109be5b468d9fc1bab776dec0ef7304eaa"} Mar 13 14:27:25 crc kubenswrapper[4907]: I0313 14:27:25.119875 4907 generic.go:334] "Generic (PLEG): container finished" podID="34a0ea40-bdaa-4eef-8a28-263d8db74f10" containerID="f407f7b1e1671cab73872a73a57d80595693b20473de90bda4d6de97abc60cd5" exitCode=0 Mar 13 14:27:25 crc kubenswrapper[4907]: I0313 14:27:25.121255 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-crrww" event={"ID":"34a0ea40-bdaa-4eef-8a28-263d8db74f10","Type":"ContainerDied","Data":"f407f7b1e1671cab73872a73a57d80595693b20473de90bda4d6de97abc60cd5"} Mar 13 14:27:26 crc kubenswrapper[4907]: I0313 14:27:26.130716 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5","Type":"ContainerStarted","Data":"a15bc8c355c09d50b27215622b3ea5aaebf1df184e4f387e2405fe44a886e49b"} Mar 13 14:27:26 crc kubenswrapper[4907]: I0313 14:27:26.577135 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-crrww" Mar 13 14:27:26 crc kubenswrapper[4907]: I0313 14:27:26.582952 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-jjqr9" Mar 13 14:27:26 crc kubenswrapper[4907]: I0313 14:27:26.588626 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-rz6qh" Mar 13 14:27:26 crc kubenswrapper[4907]: I0313 14:27:26.688383 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/34a0ea40-bdaa-4eef-8a28-263d8db74f10-fernet-keys\") pod \"34a0ea40-bdaa-4eef-8a28-263d8db74f10\" (UID: \"34a0ea40-bdaa-4eef-8a28-263d8db74f10\") " Mar 13 14:27:26 crc kubenswrapper[4907]: I0313 14:27:26.688444 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/34a0ea40-bdaa-4eef-8a28-263d8db74f10-credential-keys\") pod \"34a0ea40-bdaa-4eef-8a28-263d8db74f10\" (UID: \"34a0ea40-bdaa-4eef-8a28-263d8db74f10\") " Mar 13 14:27:26 crc kubenswrapper[4907]: I0313 14:27:26.688522 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b144897c-8640-4c5c-afa1-68b44dfd4f98-scripts\") pod \"b144897c-8640-4c5c-afa1-68b44dfd4f98\" (UID: \"b144897c-8640-4c5c-afa1-68b44dfd4f98\") " Mar 13 14:27:26 crc kubenswrapper[4907]: I0313 14:27:26.688572 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b144897c-8640-4c5c-afa1-68b44dfd4f98-config-data\") pod \"b144897c-8640-4c5c-afa1-68b44dfd4f98\" (UID: \"b144897c-8640-4c5c-afa1-68b44dfd4f98\") " Mar 13 14:27:26 crc kubenswrapper[4907]: I0313 14:27:26.688606 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34a0ea40-bdaa-4eef-8a28-263d8db74f10-config-data\") pod \"34a0ea40-bdaa-4eef-8a28-263d8db74f10\" (UID: \"34a0ea40-bdaa-4eef-8a28-263d8db74f10\") " Mar 13 14:27:26 crc kubenswrapper[4907]: I0313 14:27:26.688643 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b144897c-8640-4c5c-afa1-68b44dfd4f98-logs\") pod \"b144897c-8640-4c5c-afa1-68b44dfd4f98\" (UID: \"b144897c-8640-4c5c-afa1-68b44dfd4f98\") " Mar 13 14:27:26 crc kubenswrapper[4907]: I0313 14:27:26.688666 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34a0ea40-bdaa-4eef-8a28-263d8db74f10-combined-ca-bundle\") pod \"34a0ea40-bdaa-4eef-8a28-263d8db74f10\" (UID: \"34a0ea40-bdaa-4eef-8a28-263d8db74f10\") " Mar 13 14:27:26 crc kubenswrapper[4907]: I0313 14:27:26.688694 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8grrt\" (UniqueName: \"kubernetes.io/projected/34a0ea40-bdaa-4eef-8a28-263d8db74f10-kube-api-access-8grrt\") pod \"34a0ea40-bdaa-4eef-8a28-263d8db74f10\" (UID: \"34a0ea40-bdaa-4eef-8a28-263d8db74f10\") " Mar 13 14:27:26 crc kubenswrapper[4907]: I0313 14:27:26.688723 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b144897c-8640-4c5c-afa1-68b44dfd4f98-combined-ca-bundle\") pod \"b144897c-8640-4c5c-afa1-68b44dfd4f98\" (UID: \"b144897c-8640-4c5c-afa1-68b44dfd4f98\") " Mar 13 14:27:26 crc kubenswrapper[4907]: I0313 14:27:26.688845 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/34a0ea40-bdaa-4eef-8a28-263d8db74f10-scripts\") pod \"34a0ea40-bdaa-4eef-8a28-263d8db74f10\" (UID: \"34a0ea40-bdaa-4eef-8a28-263d8db74f10\") " Mar 13 14:27:26 crc kubenswrapper[4907]: I0313 14:27:26.688917 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wcv67\" (UniqueName: \"kubernetes.io/projected/b144897c-8640-4c5c-afa1-68b44dfd4f98-kube-api-access-wcv67\") pod \"b144897c-8640-4c5c-afa1-68b44dfd4f98\" (UID: \"b144897c-8640-4c5c-afa1-68b44dfd4f98\") " Mar 13 14:27:26 crc kubenswrapper[4907]: I0313 14:27:26.688943 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/32b531c6-64c1-4137-b82e-3c10789325e6-db-sync-config-data\") pod \"32b531c6-64c1-4137-b82e-3c10789325e6\" (UID: \"32b531c6-64c1-4137-b82e-3c10789325e6\") " Mar 13 14:27:26 crc kubenswrapper[4907]: I0313 14:27:26.690298 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b144897c-8640-4c5c-afa1-68b44dfd4f98-logs" (OuterVolumeSpecName: "logs") pod "b144897c-8640-4c5c-afa1-68b44dfd4f98" (UID: "b144897c-8640-4c5c-afa1-68b44dfd4f98"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:27:26 crc kubenswrapper[4907]: I0313 14:27:26.695685 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34a0ea40-bdaa-4eef-8a28-263d8db74f10-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "34a0ea40-bdaa-4eef-8a28-263d8db74f10" (UID: "34a0ea40-bdaa-4eef-8a28-263d8db74f10"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:27:26 crc kubenswrapper[4907]: I0313 14:27:26.695733 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32b531c6-64c1-4137-b82e-3c10789325e6-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "32b531c6-64c1-4137-b82e-3c10789325e6" (UID: "32b531c6-64c1-4137-b82e-3c10789325e6"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:27:26 crc kubenswrapper[4907]: I0313 14:27:26.695802 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b144897c-8640-4c5c-afa1-68b44dfd4f98-kube-api-access-wcv67" (OuterVolumeSpecName: "kube-api-access-wcv67") pod "b144897c-8640-4c5c-afa1-68b44dfd4f98" (UID: "b144897c-8640-4c5c-afa1-68b44dfd4f98"). InnerVolumeSpecName "kube-api-access-wcv67". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:27:26 crc kubenswrapper[4907]: I0313 14:27:26.696117 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b144897c-8640-4c5c-afa1-68b44dfd4f98-scripts" (OuterVolumeSpecName: "scripts") pod "b144897c-8640-4c5c-afa1-68b44dfd4f98" (UID: "b144897c-8640-4c5c-afa1-68b44dfd4f98"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:27:26 crc kubenswrapper[4907]: I0313 14:27:26.696470 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/34a0ea40-bdaa-4eef-8a28-263d8db74f10-kube-api-access-8grrt" (OuterVolumeSpecName: "kube-api-access-8grrt") pod "34a0ea40-bdaa-4eef-8a28-263d8db74f10" (UID: "34a0ea40-bdaa-4eef-8a28-263d8db74f10"). InnerVolumeSpecName "kube-api-access-8grrt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:27:26 crc kubenswrapper[4907]: I0313 14:27:26.696556 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34a0ea40-bdaa-4eef-8a28-263d8db74f10-scripts" (OuterVolumeSpecName: "scripts") pod "34a0ea40-bdaa-4eef-8a28-263d8db74f10" (UID: "34a0ea40-bdaa-4eef-8a28-263d8db74f10"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:27:26 crc kubenswrapper[4907]: I0313 14:27:26.697384 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34a0ea40-bdaa-4eef-8a28-263d8db74f10-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "34a0ea40-bdaa-4eef-8a28-263d8db74f10" (UID: "34a0ea40-bdaa-4eef-8a28-263d8db74f10"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:27:26 crc kubenswrapper[4907]: I0313 14:27:26.717042 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b144897c-8640-4c5c-afa1-68b44dfd4f98-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b144897c-8640-4c5c-afa1-68b44dfd4f98" (UID: "b144897c-8640-4c5c-afa1-68b44dfd4f98"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:27:26 crc kubenswrapper[4907]: I0313 14:27:26.720092 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34a0ea40-bdaa-4eef-8a28-263d8db74f10-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "34a0ea40-bdaa-4eef-8a28-263d8db74f10" (UID: "34a0ea40-bdaa-4eef-8a28-263d8db74f10"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:27:26 crc kubenswrapper[4907]: I0313 14:27:26.722012 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b144897c-8640-4c5c-afa1-68b44dfd4f98-config-data" (OuterVolumeSpecName: "config-data") pod "b144897c-8640-4c5c-afa1-68b44dfd4f98" (UID: "b144897c-8640-4c5c-afa1-68b44dfd4f98"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:27:26 crc kubenswrapper[4907]: I0313 14:27:26.722171 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34a0ea40-bdaa-4eef-8a28-263d8db74f10-config-data" (OuterVolumeSpecName: "config-data") pod "34a0ea40-bdaa-4eef-8a28-263d8db74f10" (UID: "34a0ea40-bdaa-4eef-8a28-263d8db74f10"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:27:26 crc kubenswrapper[4907]: I0313 14:27:26.791432 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32b531c6-64c1-4137-b82e-3c10789325e6-combined-ca-bundle\") pod \"32b531c6-64c1-4137-b82e-3c10789325e6\" (UID: \"32b531c6-64c1-4137-b82e-3c10789325e6\") " Mar 13 14:27:26 crc kubenswrapper[4907]: I0313 14:27:26.791596 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-578zc\" (UniqueName: \"kubernetes.io/projected/32b531c6-64c1-4137-b82e-3c10789325e6-kube-api-access-578zc\") pod \"32b531c6-64c1-4137-b82e-3c10789325e6\" (UID: \"32b531c6-64c1-4137-b82e-3c10789325e6\") " Mar 13 14:27:26 crc kubenswrapper[4907]: I0313 14:27:26.792104 4907 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/34a0ea40-bdaa-4eef-8a28-263d8db74f10-fernet-keys\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:26 crc kubenswrapper[4907]: I0313 14:27:26.792123 4907 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/34a0ea40-bdaa-4eef-8a28-263d8db74f10-credential-keys\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:26 crc kubenswrapper[4907]: I0313 14:27:26.792137 4907 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b144897c-8640-4c5c-afa1-68b44dfd4f98-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:26 crc kubenswrapper[4907]: I0313 14:27:26.792149 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b144897c-8640-4c5c-afa1-68b44dfd4f98-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:26 crc kubenswrapper[4907]: I0313 14:27:26.792161 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34a0ea40-bdaa-4eef-8a28-263d8db74f10-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:26 crc kubenswrapper[4907]: I0313 14:27:26.792173 4907 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b144897c-8640-4c5c-afa1-68b44dfd4f98-logs\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:26 crc kubenswrapper[4907]: I0313 14:27:26.793241 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34a0ea40-bdaa-4eef-8a28-263d8db74f10-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:26 crc kubenswrapper[4907]: I0313 14:27:26.793258 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8grrt\" (UniqueName: \"kubernetes.io/projected/34a0ea40-bdaa-4eef-8a28-263d8db74f10-kube-api-access-8grrt\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:26 crc kubenswrapper[4907]: I0313 14:27:26.793270 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b144897c-8640-4c5c-afa1-68b44dfd4f98-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:26 crc kubenswrapper[4907]: I0313 14:27:26.793281 4907 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/34a0ea40-bdaa-4eef-8a28-263d8db74f10-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:26 crc kubenswrapper[4907]: I0313 14:27:26.793295 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wcv67\" (UniqueName: \"kubernetes.io/projected/b144897c-8640-4c5c-afa1-68b44dfd4f98-kube-api-access-wcv67\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:26 crc kubenswrapper[4907]: I0313 14:27:26.793307 4907 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/32b531c6-64c1-4137-b82e-3c10789325e6-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:26 crc kubenswrapper[4907]: I0313 14:27:26.802411 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/32b531c6-64c1-4137-b82e-3c10789325e6-kube-api-access-578zc" (OuterVolumeSpecName: "kube-api-access-578zc") pod "32b531c6-64c1-4137-b82e-3c10789325e6" (UID: "32b531c6-64c1-4137-b82e-3c10789325e6"). InnerVolumeSpecName "kube-api-access-578zc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:27:26 crc kubenswrapper[4907]: I0313 14:27:26.823647 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32b531c6-64c1-4137-b82e-3c10789325e6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "32b531c6-64c1-4137-b82e-3c10789325e6" (UID: "32b531c6-64c1-4137-b82e-3c10789325e6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:27:26 crc kubenswrapper[4907]: I0313 14:27:26.895101 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-578zc\" (UniqueName: \"kubernetes.io/projected/32b531c6-64c1-4137-b82e-3c10789325e6-kube-api-access-578zc\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:26 crc kubenswrapper[4907]: I0313 14:27:26.895140 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32b531c6-64c1-4137-b82e-3c10789325e6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.143748 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-rz6qh" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.143747 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-rz6qh" event={"ID":"32b531c6-64c1-4137-b82e-3c10789325e6","Type":"ContainerDied","Data":"4b7673a5ff0b4e402bcd5b85ece6329b6bba3e9faaac7ef9c519ffe96fec5929"} Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.144222 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4b7673a5ff0b4e402bcd5b85ece6329b6bba3e9faaac7ef9c519ffe96fec5929" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.145565 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-jjqr9" event={"ID":"b144897c-8640-4c5c-afa1-68b44dfd4f98","Type":"ContainerDied","Data":"d144653276b0a0015c06098911900d038b6dd5b20a40a40339bc3b434d2953a5"} Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.145597 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d144653276b0a0015c06098911900d038b6dd5b20a40a40339bc3b434d2953a5" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.145612 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-jjqr9" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.151161 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-crrww" event={"ID":"34a0ea40-bdaa-4eef-8a28-263d8db74f10","Type":"ContainerDied","Data":"35ce55e83a38917576782260db545e824c3f1f351d75a47429e1e2d417a299bb"} Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.151198 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="35ce55e83a38917576782260db545e824c3f1f351d75a47429e1e2d417a299bb" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.151318 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-crrww" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.153425 4907 generic.go:334] "Generic (PLEG): container finished" podID="870ec55c-fc6f-4b16-8c49-a3cda0d0d010" containerID="c90c150f12af47e1b50f63c9ddceacb4d1551893cbeff831172fa51757c596cb" exitCode=0 Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.153455 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-jt88g" event={"ID":"870ec55c-fc6f-4b16-8c49-a3cda0d0d010","Type":"ContainerDied","Data":"c90c150f12af47e1b50f63c9ddceacb4d1551893cbeff831172fa51757c596cb"} Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.318468 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-85b9d64669-dxnf6"] Mar 13 14:27:27 crc kubenswrapper[4907]: E0313 14:27:27.318896 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b144897c-8640-4c5c-afa1-68b44dfd4f98" containerName="placement-db-sync" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.318919 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="b144897c-8640-4c5c-afa1-68b44dfd4f98" containerName="placement-db-sync" Mar 13 14:27:27 crc kubenswrapper[4907]: E0313 14:27:27.318945 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34a0ea40-bdaa-4eef-8a28-263d8db74f10" containerName="keystone-bootstrap" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.318953 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="34a0ea40-bdaa-4eef-8a28-263d8db74f10" containerName="keystone-bootstrap" Mar 13 14:27:27 crc kubenswrapper[4907]: E0313 14:27:27.318975 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32b531c6-64c1-4137-b82e-3c10789325e6" containerName="barbican-db-sync" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.318985 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="32b531c6-64c1-4137-b82e-3c10789325e6" containerName="barbican-db-sync" Mar 13 14:27:27 crc kubenswrapper[4907]: E0313 14:27:27.319005 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c065ce46-d7f3-4e50-b30c-83a1592b6ead" containerName="dnsmasq-dns" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.319014 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="c065ce46-d7f3-4e50-b30c-83a1592b6ead" containerName="dnsmasq-dns" Mar 13 14:27:27 crc kubenswrapper[4907]: E0313 14:27:27.319033 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c065ce46-d7f3-4e50-b30c-83a1592b6ead" containerName="init" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.319041 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="c065ce46-d7f3-4e50-b30c-83a1592b6ead" containerName="init" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.319250 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="34a0ea40-bdaa-4eef-8a28-263d8db74f10" containerName="keystone-bootstrap" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.319276 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="32b531c6-64c1-4137-b82e-3c10789325e6" containerName="barbican-db-sync" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.319293 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="b144897c-8640-4c5c-afa1-68b44dfd4f98" containerName="placement-db-sync" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.319308 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="c065ce46-d7f3-4e50-b30c-83a1592b6ead" containerName="dnsmasq-dns" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.320041 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-85b9d64669-dxnf6" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.325648 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.325917 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.325937 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.325989 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.326144 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.326323 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-p9m6m" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.342960 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-85b9d64669-dxnf6"] Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.356127 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-67d6ccd9c4-ms98l"] Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.364907 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-67d6ccd9c4-ms98l" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.374390 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-pg28b" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.377331 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.377553 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.377696 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.377943 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.409382 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-c685c5475-rvkf2"] Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.419830 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-c685c5475-rvkf2" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.430562 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.458648 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-skqjb" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.458866 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.520964 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/45d14b2b-1d0e-4e15-b837-0d2fb6aed785-config-data\") pod \"placement-67d6ccd9c4-ms98l\" (UID: \"45d14b2b-1d0e-4e15-b837-0d2fb6aed785\") " pod="openstack/placement-67d6ccd9c4-ms98l" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.521039 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/45d14b2b-1d0e-4e15-b837-0d2fb6aed785-internal-tls-certs\") pod \"placement-67d6ccd9c4-ms98l\" (UID: \"45d14b2b-1d0e-4e15-b837-0d2fb6aed785\") " pod="openstack/placement-67d6ccd9c4-ms98l" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.521076 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4sstr\" (UniqueName: \"kubernetes.io/projected/45d14b2b-1d0e-4e15-b837-0d2fb6aed785-kube-api-access-4sstr\") pod \"placement-67d6ccd9c4-ms98l\" (UID: \"45d14b2b-1d0e-4e15-b837-0d2fb6aed785\") " pod="openstack/placement-67d6ccd9c4-ms98l" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.521125 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/45d14b2b-1d0e-4e15-b837-0d2fb6aed785-public-tls-certs\") pod \"placement-67d6ccd9c4-ms98l\" (UID: \"45d14b2b-1d0e-4e15-b837-0d2fb6aed785\") " pod="openstack/placement-67d6ccd9c4-ms98l" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.521174 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8a738a66-54bc-4f5e-86d5-6e3004d8f265-fernet-keys\") pod \"keystone-85b9d64669-dxnf6\" (UID: \"8a738a66-54bc-4f5e-86d5-6e3004d8f265\") " pod="openstack/keystone-85b9d64669-dxnf6" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.521201 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/45d14b2b-1d0e-4e15-b837-0d2fb6aed785-scripts\") pod \"placement-67d6ccd9c4-ms98l\" (UID: \"45d14b2b-1d0e-4e15-b837-0d2fb6aed785\") " pod="openstack/placement-67d6ccd9c4-ms98l" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.521225 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45d14b2b-1d0e-4e15-b837-0d2fb6aed785-combined-ca-bundle\") pod \"placement-67d6ccd9c4-ms98l\" (UID: \"45d14b2b-1d0e-4e15-b837-0d2fb6aed785\") " pod="openstack/placement-67d6ccd9c4-ms98l" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.521250 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a738a66-54bc-4f5e-86d5-6e3004d8f265-public-tls-certs\") pod \"keystone-85b9d64669-dxnf6\" (UID: \"8a738a66-54bc-4f5e-86d5-6e3004d8f265\") " pod="openstack/keystone-85b9d64669-dxnf6" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.521313 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/8a738a66-54bc-4f5e-86d5-6e3004d8f265-credential-keys\") pod \"keystone-85b9d64669-dxnf6\" (UID: \"8a738a66-54bc-4f5e-86d5-6e3004d8f265\") " pod="openstack/keystone-85b9d64669-dxnf6" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.521371 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a738a66-54bc-4f5e-86d5-6e3004d8f265-combined-ca-bundle\") pod \"keystone-85b9d64669-dxnf6\" (UID: \"8a738a66-54bc-4f5e-86d5-6e3004d8f265\") " pod="openstack/keystone-85b9d64669-dxnf6" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.521407 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/45d14b2b-1d0e-4e15-b837-0d2fb6aed785-logs\") pod \"placement-67d6ccd9c4-ms98l\" (UID: \"45d14b2b-1d0e-4e15-b837-0d2fb6aed785\") " pod="openstack/placement-67d6ccd9c4-ms98l" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.521430 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pvwdb\" (UniqueName: \"kubernetes.io/projected/8a738a66-54bc-4f5e-86d5-6e3004d8f265-kube-api-access-pvwdb\") pod \"keystone-85b9d64669-dxnf6\" (UID: \"8a738a66-54bc-4f5e-86d5-6e3004d8f265\") " pod="openstack/keystone-85b9d64669-dxnf6" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.521470 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a738a66-54bc-4f5e-86d5-6e3004d8f265-scripts\") pod \"keystone-85b9d64669-dxnf6\" (UID: \"8a738a66-54bc-4f5e-86d5-6e3004d8f265\") " pod="openstack/keystone-85b9d64669-dxnf6" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.521497 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a738a66-54bc-4f5e-86d5-6e3004d8f265-internal-tls-certs\") pod \"keystone-85b9d64669-dxnf6\" (UID: \"8a738a66-54bc-4f5e-86d5-6e3004d8f265\") " pod="openstack/keystone-85b9d64669-dxnf6" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.521529 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a738a66-54bc-4f5e-86d5-6e3004d8f265-config-data\") pod \"keystone-85b9d64669-dxnf6\" (UID: \"8a738a66-54bc-4f5e-86d5-6e3004d8f265\") " pod="openstack/keystone-85b9d64669-dxnf6" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.531822 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-c685c5475-rvkf2"] Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.541600 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-54c69f4bdb-lt67r"] Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.549014 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-54c69f4bdb-lt67r" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.551925 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.576394 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-54c69f4bdb-lt67r"] Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.596522 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-67d6ccd9c4-ms98l"] Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.626771 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a738a66-54bc-4f5e-86d5-6e3004d8f265-public-tls-certs\") pod \"keystone-85b9d64669-dxnf6\" (UID: \"8a738a66-54bc-4f5e-86d5-6e3004d8f265\") " pod="openstack/keystone-85b9d64669-dxnf6" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.626831 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ca4928ba-1277-4918-8b92-3eda2b276577-logs\") pod \"barbican-worker-c685c5475-rvkf2\" (UID: \"ca4928ba-1277-4918-8b92-3eda2b276577\") " pod="openstack/barbican-worker-c685c5475-rvkf2" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.626850 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/8a738a66-54bc-4f5e-86d5-6e3004d8f265-credential-keys\") pod \"keystone-85b9d64669-dxnf6\" (UID: \"8a738a66-54bc-4f5e-86d5-6e3004d8f265\") " pod="openstack/keystone-85b9d64669-dxnf6" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.626876 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a738a66-54bc-4f5e-86d5-6e3004d8f265-combined-ca-bundle\") pod \"keystone-85b9d64669-dxnf6\" (UID: \"8a738a66-54bc-4f5e-86d5-6e3004d8f265\") " pod="openstack/keystone-85b9d64669-dxnf6" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.626915 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/45d14b2b-1d0e-4e15-b837-0d2fb6aed785-logs\") pod \"placement-67d6ccd9c4-ms98l\" (UID: \"45d14b2b-1d0e-4e15-b837-0d2fb6aed785\") " pod="openstack/placement-67d6ccd9c4-ms98l" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.626935 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pvwdb\" (UniqueName: \"kubernetes.io/projected/8a738a66-54bc-4f5e-86d5-6e3004d8f265-kube-api-access-pvwdb\") pod \"keystone-85b9d64669-dxnf6\" (UID: \"8a738a66-54bc-4f5e-86d5-6e3004d8f265\") " pod="openstack/keystone-85b9d64669-dxnf6" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.626962 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a738a66-54bc-4f5e-86d5-6e3004d8f265-scripts\") pod \"keystone-85b9d64669-dxnf6\" (UID: \"8a738a66-54bc-4f5e-86d5-6e3004d8f265\") " pod="openstack/keystone-85b9d64669-dxnf6" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.626981 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a738a66-54bc-4f5e-86d5-6e3004d8f265-internal-tls-certs\") pod \"keystone-85b9d64669-dxnf6\" (UID: \"8a738a66-54bc-4f5e-86d5-6e3004d8f265\") " pod="openstack/keystone-85b9d64669-dxnf6" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.626999 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a738a66-54bc-4f5e-86d5-6e3004d8f265-config-data\") pod \"keystone-85b9d64669-dxnf6\" (UID: \"8a738a66-54bc-4f5e-86d5-6e3004d8f265\") " pod="openstack/keystone-85b9d64669-dxnf6" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.627018 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/45d14b2b-1d0e-4e15-b837-0d2fb6aed785-config-data\") pod \"placement-67d6ccd9c4-ms98l\" (UID: \"45d14b2b-1d0e-4e15-b837-0d2fb6aed785\") " pod="openstack/placement-67d6ccd9c4-ms98l" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.627040 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca4928ba-1277-4918-8b92-3eda2b276577-config-data\") pod \"barbican-worker-c685c5475-rvkf2\" (UID: \"ca4928ba-1277-4918-8b92-3eda2b276577\") " pod="openstack/barbican-worker-c685c5475-rvkf2" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.627060 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9vftx\" (UniqueName: \"kubernetes.io/projected/ca4928ba-1277-4918-8b92-3eda2b276577-kube-api-access-9vftx\") pod \"barbican-worker-c685c5475-rvkf2\" (UID: \"ca4928ba-1277-4918-8b92-3eda2b276577\") " pod="openstack/barbican-worker-c685c5475-rvkf2" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.627082 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/45d14b2b-1d0e-4e15-b837-0d2fb6aed785-internal-tls-certs\") pod \"placement-67d6ccd9c4-ms98l\" (UID: \"45d14b2b-1d0e-4e15-b837-0d2fb6aed785\") " pod="openstack/placement-67d6ccd9c4-ms98l" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.627102 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4sstr\" (UniqueName: \"kubernetes.io/projected/45d14b2b-1d0e-4e15-b837-0d2fb6aed785-kube-api-access-4sstr\") pod \"placement-67d6ccd9c4-ms98l\" (UID: \"45d14b2b-1d0e-4e15-b837-0d2fb6aed785\") " pod="openstack/placement-67d6ccd9c4-ms98l" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.627123 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ca4928ba-1277-4918-8b92-3eda2b276577-config-data-custom\") pod \"barbican-worker-c685c5475-rvkf2\" (UID: \"ca4928ba-1277-4918-8b92-3eda2b276577\") " pod="openstack/barbican-worker-c685c5475-rvkf2" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.627147 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca4928ba-1277-4918-8b92-3eda2b276577-combined-ca-bundle\") pod \"barbican-worker-c685c5475-rvkf2\" (UID: \"ca4928ba-1277-4918-8b92-3eda2b276577\") " pod="openstack/barbican-worker-c685c5475-rvkf2" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.627163 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/45d14b2b-1d0e-4e15-b837-0d2fb6aed785-public-tls-certs\") pod \"placement-67d6ccd9c4-ms98l\" (UID: \"45d14b2b-1d0e-4e15-b837-0d2fb6aed785\") " pod="openstack/placement-67d6ccd9c4-ms98l" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.627192 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8a738a66-54bc-4f5e-86d5-6e3004d8f265-fernet-keys\") pod \"keystone-85b9d64669-dxnf6\" (UID: \"8a738a66-54bc-4f5e-86d5-6e3004d8f265\") " pod="openstack/keystone-85b9d64669-dxnf6" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.627210 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/45d14b2b-1d0e-4e15-b837-0d2fb6aed785-scripts\") pod \"placement-67d6ccd9c4-ms98l\" (UID: \"45d14b2b-1d0e-4e15-b837-0d2fb6aed785\") " pod="openstack/placement-67d6ccd9c4-ms98l" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.627227 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45d14b2b-1d0e-4e15-b837-0d2fb6aed785-combined-ca-bundle\") pod \"placement-67d6ccd9c4-ms98l\" (UID: \"45d14b2b-1d0e-4e15-b837-0d2fb6aed785\") " pod="openstack/placement-67d6ccd9c4-ms98l" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.630069 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/45d14b2b-1d0e-4e15-b837-0d2fb6aed785-logs\") pod \"placement-67d6ccd9c4-ms98l\" (UID: \"45d14b2b-1d0e-4e15-b837-0d2fb6aed785\") " pod="openstack/placement-67d6ccd9c4-ms98l" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.627856 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-79fd7f986f-t9wlb"] Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.635643 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/45d14b2b-1d0e-4e15-b837-0d2fb6aed785-public-tls-certs\") pod \"placement-67d6ccd9c4-ms98l\" (UID: \"45d14b2b-1d0e-4e15-b837-0d2fb6aed785\") " pod="openstack/placement-67d6ccd9c4-ms98l" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.636440 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a738a66-54bc-4f5e-86d5-6e3004d8f265-internal-tls-certs\") pod \"keystone-85b9d64669-dxnf6\" (UID: \"8a738a66-54bc-4f5e-86d5-6e3004d8f265\") " pod="openstack/keystone-85b9d64669-dxnf6" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.638674 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a738a66-54bc-4f5e-86d5-6e3004d8f265-scripts\") pod \"keystone-85b9d64669-dxnf6\" (UID: \"8a738a66-54bc-4f5e-86d5-6e3004d8f265\") " pod="openstack/keystone-85b9d64669-dxnf6" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.639274 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45d14b2b-1d0e-4e15-b837-0d2fb6aed785-combined-ca-bundle\") pod \"placement-67d6ccd9c4-ms98l\" (UID: \"45d14b2b-1d0e-4e15-b837-0d2fb6aed785\") " pod="openstack/placement-67d6ccd9c4-ms98l" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.640633 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/45d14b2b-1d0e-4e15-b837-0d2fb6aed785-config-data\") pod \"placement-67d6ccd9c4-ms98l\" (UID: \"45d14b2b-1d0e-4e15-b837-0d2fb6aed785\") " pod="openstack/placement-67d6ccd9c4-ms98l" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.640679 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79fd7f986f-t9wlb" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.644610 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/45d14b2b-1d0e-4e15-b837-0d2fb6aed785-scripts\") pod \"placement-67d6ccd9c4-ms98l\" (UID: \"45d14b2b-1d0e-4e15-b837-0d2fb6aed785\") " pod="openstack/placement-67d6ccd9c4-ms98l" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.645015 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a738a66-54bc-4f5e-86d5-6e3004d8f265-public-tls-certs\") pod \"keystone-85b9d64669-dxnf6\" (UID: \"8a738a66-54bc-4f5e-86d5-6e3004d8f265\") " pod="openstack/keystone-85b9d64669-dxnf6" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.645523 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/45d14b2b-1d0e-4e15-b837-0d2fb6aed785-internal-tls-certs\") pod \"placement-67d6ccd9c4-ms98l\" (UID: \"45d14b2b-1d0e-4e15-b837-0d2fb6aed785\") " pod="openstack/placement-67d6ccd9c4-ms98l" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.648387 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/8a738a66-54bc-4f5e-86d5-6e3004d8f265-credential-keys\") pod \"keystone-85b9d64669-dxnf6\" (UID: \"8a738a66-54bc-4f5e-86d5-6e3004d8f265\") " pod="openstack/keystone-85b9d64669-dxnf6" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.648826 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a738a66-54bc-4f5e-86d5-6e3004d8f265-config-data\") pod \"keystone-85b9d64669-dxnf6\" (UID: \"8a738a66-54bc-4f5e-86d5-6e3004d8f265\") " pod="openstack/keystone-85b9d64669-dxnf6" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.651449 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8a738a66-54bc-4f5e-86d5-6e3004d8f265-fernet-keys\") pod \"keystone-85b9d64669-dxnf6\" (UID: \"8a738a66-54bc-4f5e-86d5-6e3004d8f265\") " pod="openstack/keystone-85b9d64669-dxnf6" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.657398 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pvwdb\" (UniqueName: \"kubernetes.io/projected/8a738a66-54bc-4f5e-86d5-6e3004d8f265-kube-api-access-pvwdb\") pod \"keystone-85b9d64669-dxnf6\" (UID: \"8a738a66-54bc-4f5e-86d5-6e3004d8f265\") " pod="openstack/keystone-85b9d64669-dxnf6" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.663149 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4sstr\" (UniqueName: \"kubernetes.io/projected/45d14b2b-1d0e-4e15-b837-0d2fb6aed785-kube-api-access-4sstr\") pod \"placement-67d6ccd9c4-ms98l\" (UID: \"45d14b2b-1d0e-4e15-b837-0d2fb6aed785\") " pod="openstack/placement-67d6ccd9c4-ms98l" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.664673 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a738a66-54bc-4f5e-86d5-6e3004d8f265-combined-ca-bundle\") pod \"keystone-85b9d64669-dxnf6\" (UID: \"8a738a66-54bc-4f5e-86d5-6e3004d8f265\") " pod="openstack/keystone-85b9d64669-dxnf6" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.670094 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-79fd7f986f-t9wlb"] Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.710048 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-5d49fd688d-njvfm"] Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.713473 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5d49fd688d-njvfm" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.725868 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-5d49fd688d-njvfm"] Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.726045 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.728943 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2d754127-d8df-48de-b77a-46aa7898c45f-ovsdbserver-nb\") pod \"dnsmasq-dns-79fd7f986f-t9wlb\" (UID: \"2d754127-d8df-48de-b77a-46aa7898c45f\") " pod="openstack/dnsmasq-dns-79fd7f986f-t9wlb" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.728987 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ca4928ba-1277-4918-8b92-3eda2b276577-config-data-custom\") pod \"barbican-worker-c685c5475-rvkf2\" (UID: \"ca4928ba-1277-4918-8b92-3eda2b276577\") " pod="openstack/barbican-worker-c685c5475-rvkf2" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.729020 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca4928ba-1277-4918-8b92-3eda2b276577-combined-ca-bundle\") pod \"barbican-worker-c685c5475-rvkf2\" (UID: \"ca4928ba-1277-4918-8b92-3eda2b276577\") " pod="openstack/barbican-worker-c685c5475-rvkf2" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.729041 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2d754127-d8df-48de-b77a-46aa7898c45f-config\") pod \"dnsmasq-dns-79fd7f986f-t9wlb\" (UID: \"2d754127-d8df-48de-b77a-46aa7898c45f\") " pod="openstack/dnsmasq-dns-79fd7f986f-t9wlb" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.729069 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2d754127-d8df-48de-b77a-46aa7898c45f-dns-swift-storage-0\") pod \"dnsmasq-dns-79fd7f986f-t9wlb\" (UID: \"2d754127-d8df-48de-b77a-46aa7898c45f\") " pod="openstack/dnsmasq-dns-79fd7f986f-t9wlb" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.729102 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qttjx\" (UniqueName: \"kubernetes.io/projected/fcbb1687-db0b-4fc6-937c-89102831237b-kube-api-access-qttjx\") pod \"barbican-keystone-listener-54c69f4bdb-lt67r\" (UID: \"fcbb1687-db0b-4fc6-937c-89102831237b\") " pod="openstack/barbican-keystone-listener-54c69f4bdb-lt67r" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.729135 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ca4928ba-1277-4918-8b92-3eda2b276577-logs\") pod \"barbican-worker-c685c5475-rvkf2\" (UID: \"ca4928ba-1277-4918-8b92-3eda2b276577\") " pod="openstack/barbican-worker-c685c5475-rvkf2" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.729156 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fcbb1687-db0b-4fc6-937c-89102831237b-combined-ca-bundle\") pod \"barbican-keystone-listener-54c69f4bdb-lt67r\" (UID: \"fcbb1687-db0b-4fc6-937c-89102831237b\") " pod="openstack/barbican-keystone-listener-54c69f4bdb-lt67r" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.729175 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2d754127-d8df-48de-b77a-46aa7898c45f-dns-svc\") pod \"dnsmasq-dns-79fd7f986f-t9wlb\" (UID: \"2d754127-d8df-48de-b77a-46aa7898c45f\") " pod="openstack/dnsmasq-dns-79fd7f986f-t9wlb" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.729198 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fcbb1687-db0b-4fc6-937c-89102831237b-config-data\") pod \"barbican-keystone-listener-54c69f4bdb-lt67r\" (UID: \"fcbb1687-db0b-4fc6-937c-89102831237b\") " pod="openstack/barbican-keystone-listener-54c69f4bdb-lt67r" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.729233 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fcbb1687-db0b-4fc6-937c-89102831237b-logs\") pod \"barbican-keystone-listener-54c69f4bdb-lt67r\" (UID: \"fcbb1687-db0b-4fc6-937c-89102831237b\") " pod="openstack/barbican-keystone-listener-54c69f4bdb-lt67r" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.729266 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dhzwg\" (UniqueName: \"kubernetes.io/projected/2d754127-d8df-48de-b77a-46aa7898c45f-kube-api-access-dhzwg\") pod \"dnsmasq-dns-79fd7f986f-t9wlb\" (UID: \"2d754127-d8df-48de-b77a-46aa7898c45f\") " pod="openstack/dnsmasq-dns-79fd7f986f-t9wlb" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.729289 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2d754127-d8df-48de-b77a-46aa7898c45f-ovsdbserver-sb\") pod \"dnsmasq-dns-79fd7f986f-t9wlb\" (UID: \"2d754127-d8df-48de-b77a-46aa7898c45f\") " pod="openstack/dnsmasq-dns-79fd7f986f-t9wlb" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.729312 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca4928ba-1277-4918-8b92-3eda2b276577-config-data\") pod \"barbican-worker-c685c5475-rvkf2\" (UID: \"ca4928ba-1277-4918-8b92-3eda2b276577\") " pod="openstack/barbican-worker-c685c5475-rvkf2" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.729331 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9vftx\" (UniqueName: \"kubernetes.io/projected/ca4928ba-1277-4918-8b92-3eda2b276577-kube-api-access-9vftx\") pod \"barbican-worker-c685c5475-rvkf2\" (UID: \"ca4928ba-1277-4918-8b92-3eda2b276577\") " pod="openstack/barbican-worker-c685c5475-rvkf2" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.729348 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fcbb1687-db0b-4fc6-937c-89102831237b-config-data-custom\") pod \"barbican-keystone-listener-54c69f4bdb-lt67r\" (UID: \"fcbb1687-db0b-4fc6-937c-89102831237b\") " pod="openstack/barbican-keystone-listener-54c69f4bdb-lt67r" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.735250 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ca4928ba-1277-4918-8b92-3eda2b276577-logs\") pod \"barbican-worker-c685c5475-rvkf2\" (UID: \"ca4928ba-1277-4918-8b92-3eda2b276577\") " pod="openstack/barbican-worker-c685c5475-rvkf2" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.740856 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ca4928ba-1277-4918-8b92-3eda2b276577-config-data-custom\") pod \"barbican-worker-c685c5475-rvkf2\" (UID: \"ca4928ba-1277-4918-8b92-3eda2b276577\") " pod="openstack/barbican-worker-c685c5475-rvkf2" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.741633 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-67d6ccd9c4-ms98l" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.742227 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca4928ba-1277-4918-8b92-3eda2b276577-config-data\") pod \"barbican-worker-c685c5475-rvkf2\" (UID: \"ca4928ba-1277-4918-8b92-3eda2b276577\") " pod="openstack/barbican-worker-c685c5475-rvkf2" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.743590 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca4928ba-1277-4918-8b92-3eda2b276577-combined-ca-bundle\") pod \"barbican-worker-c685c5475-rvkf2\" (UID: \"ca4928ba-1277-4918-8b92-3eda2b276577\") " pod="openstack/barbican-worker-c685c5475-rvkf2" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.793806 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9vftx\" (UniqueName: \"kubernetes.io/projected/ca4928ba-1277-4918-8b92-3eda2b276577-kube-api-access-9vftx\") pod \"barbican-worker-c685c5475-rvkf2\" (UID: \"ca4928ba-1277-4918-8b92-3eda2b276577\") " pod="openstack/barbican-worker-c685c5475-rvkf2" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.832690 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/070e2b75-05b1-42a4-a34e-950d45d65920-config-data\") pod \"barbican-api-5d49fd688d-njvfm\" (UID: \"070e2b75-05b1-42a4-a34e-950d45d65920\") " pod="openstack/barbican-api-5d49fd688d-njvfm" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.832747 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fcbb1687-db0b-4fc6-937c-89102831237b-combined-ca-bundle\") pod \"barbican-keystone-listener-54c69f4bdb-lt67r\" (UID: \"fcbb1687-db0b-4fc6-937c-89102831237b\") " pod="openstack/barbican-keystone-listener-54c69f4bdb-lt67r" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.832768 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2d754127-d8df-48de-b77a-46aa7898c45f-dns-svc\") pod \"dnsmasq-dns-79fd7f986f-t9wlb\" (UID: \"2d754127-d8df-48de-b77a-46aa7898c45f\") " pod="openstack/dnsmasq-dns-79fd7f986f-t9wlb" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.832785 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fcbb1687-db0b-4fc6-937c-89102831237b-config-data\") pod \"barbican-keystone-listener-54c69f4bdb-lt67r\" (UID: \"fcbb1687-db0b-4fc6-937c-89102831237b\") " pod="openstack/barbican-keystone-listener-54c69f4bdb-lt67r" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.832812 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/070e2b75-05b1-42a4-a34e-950d45d65920-logs\") pod \"barbican-api-5d49fd688d-njvfm\" (UID: \"070e2b75-05b1-42a4-a34e-950d45d65920\") " pod="openstack/barbican-api-5d49fd688d-njvfm" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.832833 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bfwxk\" (UniqueName: \"kubernetes.io/projected/070e2b75-05b1-42a4-a34e-950d45d65920-kube-api-access-bfwxk\") pod \"barbican-api-5d49fd688d-njvfm\" (UID: \"070e2b75-05b1-42a4-a34e-950d45d65920\") " pod="openstack/barbican-api-5d49fd688d-njvfm" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.832859 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fcbb1687-db0b-4fc6-937c-89102831237b-logs\") pod \"barbican-keystone-listener-54c69f4bdb-lt67r\" (UID: \"fcbb1687-db0b-4fc6-937c-89102831237b\") " pod="openstack/barbican-keystone-listener-54c69f4bdb-lt67r" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.832925 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dhzwg\" (UniqueName: \"kubernetes.io/projected/2d754127-d8df-48de-b77a-46aa7898c45f-kube-api-access-dhzwg\") pod \"dnsmasq-dns-79fd7f986f-t9wlb\" (UID: \"2d754127-d8df-48de-b77a-46aa7898c45f\") " pod="openstack/dnsmasq-dns-79fd7f986f-t9wlb" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.832950 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2d754127-d8df-48de-b77a-46aa7898c45f-ovsdbserver-sb\") pod \"dnsmasq-dns-79fd7f986f-t9wlb\" (UID: \"2d754127-d8df-48de-b77a-46aa7898c45f\") " pod="openstack/dnsmasq-dns-79fd7f986f-t9wlb" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.832981 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fcbb1687-db0b-4fc6-937c-89102831237b-config-data-custom\") pod \"barbican-keystone-listener-54c69f4bdb-lt67r\" (UID: \"fcbb1687-db0b-4fc6-937c-89102831237b\") " pod="openstack/barbican-keystone-listener-54c69f4bdb-lt67r" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.833004 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2d754127-d8df-48de-b77a-46aa7898c45f-ovsdbserver-nb\") pod \"dnsmasq-dns-79fd7f986f-t9wlb\" (UID: \"2d754127-d8df-48de-b77a-46aa7898c45f\") " pod="openstack/dnsmasq-dns-79fd7f986f-t9wlb" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.833034 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2d754127-d8df-48de-b77a-46aa7898c45f-config\") pod \"dnsmasq-dns-79fd7f986f-t9wlb\" (UID: \"2d754127-d8df-48de-b77a-46aa7898c45f\") " pod="openstack/dnsmasq-dns-79fd7f986f-t9wlb" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.833056 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/070e2b75-05b1-42a4-a34e-950d45d65920-config-data-custom\") pod \"barbican-api-5d49fd688d-njvfm\" (UID: \"070e2b75-05b1-42a4-a34e-950d45d65920\") " pod="openstack/barbican-api-5d49fd688d-njvfm" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.833071 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/070e2b75-05b1-42a4-a34e-950d45d65920-combined-ca-bundle\") pod \"barbican-api-5d49fd688d-njvfm\" (UID: \"070e2b75-05b1-42a4-a34e-950d45d65920\") " pod="openstack/barbican-api-5d49fd688d-njvfm" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.833092 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2d754127-d8df-48de-b77a-46aa7898c45f-dns-swift-storage-0\") pod \"dnsmasq-dns-79fd7f986f-t9wlb\" (UID: \"2d754127-d8df-48de-b77a-46aa7898c45f\") " pod="openstack/dnsmasq-dns-79fd7f986f-t9wlb" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.833125 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qttjx\" (UniqueName: \"kubernetes.io/projected/fcbb1687-db0b-4fc6-937c-89102831237b-kube-api-access-qttjx\") pod \"barbican-keystone-listener-54c69f4bdb-lt67r\" (UID: \"fcbb1687-db0b-4fc6-937c-89102831237b\") " pod="openstack/barbican-keystone-listener-54c69f4bdb-lt67r" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.833482 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fcbb1687-db0b-4fc6-937c-89102831237b-logs\") pod \"barbican-keystone-listener-54c69f4bdb-lt67r\" (UID: \"fcbb1687-db0b-4fc6-937c-89102831237b\") " pod="openstack/barbican-keystone-listener-54c69f4bdb-lt67r" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.837718 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fcbb1687-db0b-4fc6-937c-89102831237b-config-data-custom\") pod \"barbican-keystone-listener-54c69f4bdb-lt67r\" (UID: \"fcbb1687-db0b-4fc6-937c-89102831237b\") " pod="openstack/barbican-keystone-listener-54c69f4bdb-lt67r" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.838069 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2d754127-d8df-48de-b77a-46aa7898c45f-dns-svc\") pod \"dnsmasq-dns-79fd7f986f-t9wlb\" (UID: \"2d754127-d8df-48de-b77a-46aa7898c45f\") " pod="openstack/dnsmasq-dns-79fd7f986f-t9wlb" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.842040 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fcbb1687-db0b-4fc6-937c-89102831237b-config-data\") pod \"barbican-keystone-listener-54c69f4bdb-lt67r\" (UID: \"fcbb1687-db0b-4fc6-937c-89102831237b\") " pod="openstack/barbican-keystone-listener-54c69f4bdb-lt67r" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.842548 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2d754127-d8df-48de-b77a-46aa7898c45f-ovsdbserver-nb\") pod \"dnsmasq-dns-79fd7f986f-t9wlb\" (UID: \"2d754127-d8df-48de-b77a-46aa7898c45f\") " pod="openstack/dnsmasq-dns-79fd7f986f-t9wlb" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.843292 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2d754127-d8df-48de-b77a-46aa7898c45f-config\") pod \"dnsmasq-dns-79fd7f986f-t9wlb\" (UID: \"2d754127-d8df-48de-b77a-46aa7898c45f\") " pod="openstack/dnsmasq-dns-79fd7f986f-t9wlb" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.844144 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2d754127-d8df-48de-b77a-46aa7898c45f-ovsdbserver-sb\") pod \"dnsmasq-dns-79fd7f986f-t9wlb\" (UID: \"2d754127-d8df-48de-b77a-46aa7898c45f\") " pod="openstack/dnsmasq-dns-79fd7f986f-t9wlb" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.844943 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2d754127-d8df-48de-b77a-46aa7898c45f-dns-swift-storage-0\") pod \"dnsmasq-dns-79fd7f986f-t9wlb\" (UID: \"2d754127-d8df-48de-b77a-46aa7898c45f\") " pod="openstack/dnsmasq-dns-79fd7f986f-t9wlb" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.871394 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fcbb1687-db0b-4fc6-937c-89102831237b-combined-ca-bundle\") pod \"barbican-keystone-listener-54c69f4bdb-lt67r\" (UID: \"fcbb1687-db0b-4fc6-937c-89102831237b\") " pod="openstack/barbican-keystone-listener-54c69f4bdb-lt67r" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.874740 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-76bbcdf7d9-j6hw5"] Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.880333 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qttjx\" (UniqueName: \"kubernetes.io/projected/fcbb1687-db0b-4fc6-937c-89102831237b-kube-api-access-qttjx\") pod \"barbican-keystone-listener-54c69f4bdb-lt67r\" (UID: \"fcbb1687-db0b-4fc6-937c-89102831237b\") " pod="openstack/barbican-keystone-listener-54c69f4bdb-lt67r" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.881167 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-76bbcdf7d9-j6hw5"] Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.881249 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-76bbcdf7d9-j6hw5" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.905538 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-7d7bc9f7d-fks76"] Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.907038 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-7d7bc9f7d-fks76" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.917141 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-7d7bc9f7d-fks76"] Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.945142 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-85b9d64669-dxnf6" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.954824 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dhzwg\" (UniqueName: \"kubernetes.io/projected/2d754127-d8df-48de-b77a-46aa7898c45f-kube-api-access-dhzwg\") pod \"dnsmasq-dns-79fd7f986f-t9wlb\" (UID: \"2d754127-d8df-48de-b77a-46aa7898c45f\") " pod="openstack/dnsmasq-dns-79fd7f986f-t9wlb" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.961463 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/070e2b75-05b1-42a4-a34e-950d45d65920-config-data-custom\") pod \"barbican-api-5d49fd688d-njvfm\" (UID: \"070e2b75-05b1-42a4-a34e-950d45d65920\") " pod="openstack/barbican-api-5d49fd688d-njvfm" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.961515 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/070e2b75-05b1-42a4-a34e-950d45d65920-combined-ca-bundle\") pod \"barbican-api-5d49fd688d-njvfm\" (UID: \"070e2b75-05b1-42a4-a34e-950d45d65920\") " pod="openstack/barbican-api-5d49fd688d-njvfm" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.961691 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/070e2b75-05b1-42a4-a34e-950d45d65920-config-data\") pod \"barbican-api-5d49fd688d-njvfm\" (UID: \"070e2b75-05b1-42a4-a34e-950d45d65920\") " pod="openstack/barbican-api-5d49fd688d-njvfm" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.961784 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/070e2b75-05b1-42a4-a34e-950d45d65920-logs\") pod \"barbican-api-5d49fd688d-njvfm\" (UID: \"070e2b75-05b1-42a4-a34e-950d45d65920\") " pod="openstack/barbican-api-5d49fd688d-njvfm" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.961846 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bfwxk\" (UniqueName: \"kubernetes.io/projected/070e2b75-05b1-42a4-a34e-950d45d65920-kube-api-access-bfwxk\") pod \"barbican-api-5d49fd688d-njvfm\" (UID: \"070e2b75-05b1-42a4-a34e-950d45d65920\") " pod="openstack/barbican-api-5d49fd688d-njvfm" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.965310 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/070e2b75-05b1-42a4-a34e-950d45d65920-logs\") pod \"barbican-api-5d49fd688d-njvfm\" (UID: \"070e2b75-05b1-42a4-a34e-950d45d65920\") " pod="openstack/barbican-api-5d49fd688d-njvfm" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.981466 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/070e2b75-05b1-42a4-a34e-950d45d65920-config-data\") pod \"barbican-api-5d49fd688d-njvfm\" (UID: \"070e2b75-05b1-42a4-a34e-950d45d65920\") " pod="openstack/barbican-api-5d49fd688d-njvfm" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.982194 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/070e2b75-05b1-42a4-a34e-950d45d65920-config-data-custom\") pod \"barbican-api-5d49fd688d-njvfm\" (UID: \"070e2b75-05b1-42a4-a34e-950d45d65920\") " pod="openstack/barbican-api-5d49fd688d-njvfm" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.982702 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/070e2b75-05b1-42a4-a34e-950d45d65920-combined-ca-bundle\") pod \"barbican-api-5d49fd688d-njvfm\" (UID: \"070e2b75-05b1-42a4-a34e-950d45d65920\") " pod="openstack/barbican-api-5d49fd688d-njvfm" Mar 13 14:27:28 crc kubenswrapper[4907]: I0313 14:27:28.005825 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79fd7f986f-t9wlb" Mar 13 14:27:28 crc kubenswrapper[4907]: I0313 14:27:28.025748 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bfwxk\" (UniqueName: \"kubernetes.io/projected/070e2b75-05b1-42a4-a34e-950d45d65920-kube-api-access-bfwxk\") pod \"barbican-api-5d49fd688d-njvfm\" (UID: \"070e2b75-05b1-42a4-a34e-950d45d65920\") " pod="openstack/barbican-api-5d49fd688d-njvfm" Mar 13 14:27:28 crc kubenswrapper[4907]: I0313 14:27:28.043586 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5d49fd688d-njvfm" Mar 13 14:27:28 crc kubenswrapper[4907]: I0313 14:27:28.053280 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-74984bd8b4-x42pq"] Mar 13 14:27:28 crc kubenswrapper[4907]: I0313 14:27:28.054613 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-74984bd8b4-x42pq" Mar 13 14:27:28 crc kubenswrapper[4907]: I0313 14:27:28.063639 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25b5fafe-65c8-48ba-bc34-83442f8ace4c-combined-ca-bundle\") pod \"barbican-worker-76bbcdf7d9-j6hw5\" (UID: \"25b5fafe-65c8-48ba-bc34-83442f8ace4c\") " pod="openstack/barbican-worker-76bbcdf7d9-j6hw5" Mar 13 14:27:28 crc kubenswrapper[4907]: I0313 14:27:28.063712 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/25b5fafe-65c8-48ba-bc34-83442f8ace4c-config-data\") pod \"barbican-worker-76bbcdf7d9-j6hw5\" (UID: \"25b5fafe-65c8-48ba-bc34-83442f8ace4c\") " pod="openstack/barbican-worker-76bbcdf7d9-j6hw5" Mar 13 14:27:28 crc kubenswrapper[4907]: I0313 14:27:28.063740 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65531461-cea2-4b2c-a9b6-8fd9e9bffb27-combined-ca-bundle\") pod \"barbican-keystone-listener-7d7bc9f7d-fks76\" (UID: \"65531461-cea2-4b2c-a9b6-8fd9e9bffb27\") " pod="openstack/barbican-keystone-listener-7d7bc9f7d-fks76" Mar 13 14:27:28 crc kubenswrapper[4907]: I0313 14:27:28.063774 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65531461-cea2-4b2c-a9b6-8fd9e9bffb27-config-data\") pod \"barbican-keystone-listener-7d7bc9f7d-fks76\" (UID: \"65531461-cea2-4b2c-a9b6-8fd9e9bffb27\") " pod="openstack/barbican-keystone-listener-7d7bc9f7d-fks76" Mar 13 14:27:28 crc kubenswrapper[4907]: I0313 14:27:28.063848 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/25b5fafe-65c8-48ba-bc34-83442f8ace4c-logs\") pod \"barbican-worker-76bbcdf7d9-j6hw5\" (UID: \"25b5fafe-65c8-48ba-bc34-83442f8ace4c\") " pod="openstack/barbican-worker-76bbcdf7d9-j6hw5" Mar 13 14:27:28 crc kubenswrapper[4907]: I0313 14:27:28.063960 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/25b5fafe-65c8-48ba-bc34-83442f8ace4c-config-data-custom\") pod \"barbican-worker-76bbcdf7d9-j6hw5\" (UID: \"25b5fafe-65c8-48ba-bc34-83442f8ace4c\") " pod="openstack/barbican-worker-76bbcdf7d9-j6hw5" Mar 13 14:27:28 crc kubenswrapper[4907]: I0313 14:27:28.063986 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/65531461-cea2-4b2c-a9b6-8fd9e9bffb27-config-data-custom\") pod \"barbican-keystone-listener-7d7bc9f7d-fks76\" (UID: \"65531461-cea2-4b2c-a9b6-8fd9e9bffb27\") " pod="openstack/barbican-keystone-listener-7d7bc9f7d-fks76" Mar 13 14:27:28 crc kubenswrapper[4907]: I0313 14:27:28.064009 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/65531461-cea2-4b2c-a9b6-8fd9e9bffb27-logs\") pod \"barbican-keystone-listener-7d7bc9f7d-fks76\" (UID: \"65531461-cea2-4b2c-a9b6-8fd9e9bffb27\") " pod="openstack/barbican-keystone-listener-7d7bc9f7d-fks76" Mar 13 14:27:28 crc kubenswrapper[4907]: I0313 14:27:28.064038 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-99bgf\" (UniqueName: \"kubernetes.io/projected/25b5fafe-65c8-48ba-bc34-83442f8ace4c-kube-api-access-99bgf\") pod \"barbican-worker-76bbcdf7d9-j6hw5\" (UID: \"25b5fafe-65c8-48ba-bc34-83442f8ace4c\") " pod="openstack/barbican-worker-76bbcdf7d9-j6hw5" Mar 13 14:27:28 crc kubenswrapper[4907]: I0313 14:27:28.064077 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lz7hb\" (UniqueName: \"kubernetes.io/projected/65531461-cea2-4b2c-a9b6-8fd9e9bffb27-kube-api-access-lz7hb\") pod \"barbican-keystone-listener-7d7bc9f7d-fks76\" (UID: \"65531461-cea2-4b2c-a9b6-8fd9e9bffb27\") " pod="openstack/barbican-keystone-listener-7d7bc9f7d-fks76" Mar 13 14:27:28 crc kubenswrapper[4907]: I0313 14:27:28.083516 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-c685c5475-rvkf2" Mar 13 14:27:28 crc kubenswrapper[4907]: I0313 14:27:28.094551 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-74984bd8b4-x42pq"] Mar 13 14:27:28 crc kubenswrapper[4907]: I0313 14:27:28.165417 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/25b5fafe-65c8-48ba-bc34-83442f8ace4c-logs\") pod \"barbican-worker-76bbcdf7d9-j6hw5\" (UID: \"25b5fafe-65c8-48ba-bc34-83442f8ace4c\") " pod="openstack/barbican-worker-76bbcdf7d9-j6hw5" Mar 13 14:27:28 crc kubenswrapper[4907]: I0313 14:27:28.165842 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/65531461-cea2-4b2c-a9b6-8fd9e9bffb27-config-data-custom\") pod \"barbican-keystone-listener-7d7bc9f7d-fks76\" (UID: \"65531461-cea2-4b2c-a9b6-8fd9e9bffb27\") " pod="openstack/barbican-keystone-listener-7d7bc9f7d-fks76" Mar 13 14:27:28 crc kubenswrapper[4907]: I0313 14:27:28.165864 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/65531461-cea2-4b2c-a9b6-8fd9e9bffb27-logs\") pod \"barbican-keystone-listener-7d7bc9f7d-fks76\" (UID: \"65531461-cea2-4b2c-a9b6-8fd9e9bffb27\") " pod="openstack/barbican-keystone-listener-7d7bc9f7d-fks76" Mar 13 14:27:28 crc kubenswrapper[4907]: I0313 14:27:28.165911 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/25b5fafe-65c8-48ba-bc34-83442f8ace4c-config-data-custom\") pod \"barbican-worker-76bbcdf7d9-j6hw5\" (UID: \"25b5fafe-65c8-48ba-bc34-83442f8ace4c\") " pod="openstack/barbican-worker-76bbcdf7d9-j6hw5" Mar 13 14:27:28 crc kubenswrapper[4907]: I0313 14:27:28.165946 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/032183a7-c5db-4528-a211-803050007b68-config-data\") pod \"barbican-api-74984bd8b4-x42pq\" (UID: \"032183a7-c5db-4528-a211-803050007b68\") " pod="openstack/barbican-api-74984bd8b4-x42pq" Mar 13 14:27:28 crc kubenswrapper[4907]: I0313 14:27:28.165966 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-99bgf\" (UniqueName: \"kubernetes.io/projected/25b5fafe-65c8-48ba-bc34-83442f8ace4c-kube-api-access-99bgf\") pod \"barbican-worker-76bbcdf7d9-j6hw5\" (UID: \"25b5fafe-65c8-48ba-bc34-83442f8ace4c\") " pod="openstack/barbican-worker-76bbcdf7d9-j6hw5" Mar 13 14:27:28 crc kubenswrapper[4907]: I0313 14:27:28.165993 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/032183a7-c5db-4528-a211-803050007b68-config-data-custom\") pod \"barbican-api-74984bd8b4-x42pq\" (UID: \"032183a7-c5db-4528-a211-803050007b68\") " pod="openstack/barbican-api-74984bd8b4-x42pq" Mar 13 14:27:28 crc kubenswrapper[4907]: I0313 14:27:28.166029 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lz7hb\" (UniqueName: \"kubernetes.io/projected/65531461-cea2-4b2c-a9b6-8fd9e9bffb27-kube-api-access-lz7hb\") pod \"barbican-keystone-listener-7d7bc9f7d-fks76\" (UID: \"65531461-cea2-4b2c-a9b6-8fd9e9bffb27\") " pod="openstack/barbican-keystone-listener-7d7bc9f7d-fks76" Mar 13 14:27:28 crc kubenswrapper[4907]: I0313 14:27:28.166087 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/032183a7-c5db-4528-a211-803050007b68-combined-ca-bundle\") pod \"barbican-api-74984bd8b4-x42pq\" (UID: \"032183a7-c5db-4528-a211-803050007b68\") " pod="openstack/barbican-api-74984bd8b4-x42pq" Mar 13 14:27:28 crc kubenswrapper[4907]: I0313 14:27:28.166111 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25b5fafe-65c8-48ba-bc34-83442f8ace4c-combined-ca-bundle\") pod \"barbican-worker-76bbcdf7d9-j6hw5\" (UID: \"25b5fafe-65c8-48ba-bc34-83442f8ace4c\") " pod="openstack/barbican-worker-76bbcdf7d9-j6hw5" Mar 13 14:27:28 crc kubenswrapper[4907]: I0313 14:27:28.166159 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/25b5fafe-65c8-48ba-bc34-83442f8ace4c-config-data\") pod \"barbican-worker-76bbcdf7d9-j6hw5\" (UID: \"25b5fafe-65c8-48ba-bc34-83442f8ace4c\") " pod="openstack/barbican-worker-76bbcdf7d9-j6hw5" Mar 13 14:27:28 crc kubenswrapper[4907]: I0313 14:27:28.166182 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65531461-cea2-4b2c-a9b6-8fd9e9bffb27-combined-ca-bundle\") pod \"barbican-keystone-listener-7d7bc9f7d-fks76\" (UID: \"65531461-cea2-4b2c-a9b6-8fd9e9bffb27\") " pod="openstack/barbican-keystone-listener-7d7bc9f7d-fks76" Mar 13 14:27:28 crc kubenswrapper[4907]: I0313 14:27:28.166208 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65531461-cea2-4b2c-a9b6-8fd9e9bffb27-config-data\") pod \"barbican-keystone-listener-7d7bc9f7d-fks76\" (UID: \"65531461-cea2-4b2c-a9b6-8fd9e9bffb27\") " pod="openstack/barbican-keystone-listener-7d7bc9f7d-fks76" Mar 13 14:27:28 crc kubenswrapper[4907]: I0313 14:27:28.166224 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tfsh2\" (UniqueName: \"kubernetes.io/projected/032183a7-c5db-4528-a211-803050007b68-kube-api-access-tfsh2\") pod \"barbican-api-74984bd8b4-x42pq\" (UID: \"032183a7-c5db-4528-a211-803050007b68\") " pod="openstack/barbican-api-74984bd8b4-x42pq" Mar 13 14:27:28 crc kubenswrapper[4907]: I0313 14:27:28.166252 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/032183a7-c5db-4528-a211-803050007b68-logs\") pod \"barbican-api-74984bd8b4-x42pq\" (UID: \"032183a7-c5db-4528-a211-803050007b68\") " pod="openstack/barbican-api-74984bd8b4-x42pq" Mar 13 14:27:28 crc kubenswrapper[4907]: I0313 14:27:28.170916 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/25b5fafe-65c8-48ba-bc34-83442f8ace4c-logs\") pod \"barbican-worker-76bbcdf7d9-j6hw5\" (UID: \"25b5fafe-65c8-48ba-bc34-83442f8ace4c\") " pod="openstack/barbican-worker-76bbcdf7d9-j6hw5" Mar 13 14:27:28 crc kubenswrapper[4907]: I0313 14:27:28.171759 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/65531461-cea2-4b2c-a9b6-8fd9e9bffb27-config-data-custom\") pod \"barbican-keystone-listener-7d7bc9f7d-fks76\" (UID: \"65531461-cea2-4b2c-a9b6-8fd9e9bffb27\") " pod="openstack/barbican-keystone-listener-7d7bc9f7d-fks76" Mar 13 14:27:28 crc kubenswrapper[4907]: I0313 14:27:28.172010 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/65531461-cea2-4b2c-a9b6-8fd9e9bffb27-logs\") pod \"barbican-keystone-listener-7d7bc9f7d-fks76\" (UID: \"65531461-cea2-4b2c-a9b6-8fd9e9bffb27\") " pod="openstack/barbican-keystone-listener-7d7bc9f7d-fks76" Mar 13 14:27:28 crc kubenswrapper[4907]: I0313 14:27:28.176611 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/25b5fafe-65c8-48ba-bc34-83442f8ace4c-config-data\") pod \"barbican-worker-76bbcdf7d9-j6hw5\" (UID: \"25b5fafe-65c8-48ba-bc34-83442f8ace4c\") " pod="openstack/barbican-worker-76bbcdf7d9-j6hw5" Mar 13 14:27:28 crc kubenswrapper[4907]: I0313 14:27:28.177039 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-54c69f4bdb-lt67r" Mar 13 14:27:28 crc kubenswrapper[4907]: I0313 14:27:28.185557 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25b5fafe-65c8-48ba-bc34-83442f8ace4c-combined-ca-bundle\") pod \"barbican-worker-76bbcdf7d9-j6hw5\" (UID: \"25b5fafe-65c8-48ba-bc34-83442f8ace4c\") " pod="openstack/barbican-worker-76bbcdf7d9-j6hw5" Mar 13 14:27:28 crc kubenswrapper[4907]: I0313 14:27:28.186894 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65531461-cea2-4b2c-a9b6-8fd9e9bffb27-config-data\") pod \"barbican-keystone-listener-7d7bc9f7d-fks76\" (UID: \"65531461-cea2-4b2c-a9b6-8fd9e9bffb27\") " pod="openstack/barbican-keystone-listener-7d7bc9f7d-fks76" Mar 13 14:27:28 crc kubenswrapper[4907]: I0313 14:27:28.215475 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lz7hb\" (UniqueName: \"kubernetes.io/projected/65531461-cea2-4b2c-a9b6-8fd9e9bffb27-kube-api-access-lz7hb\") pod \"barbican-keystone-listener-7d7bc9f7d-fks76\" (UID: \"65531461-cea2-4b2c-a9b6-8fd9e9bffb27\") " pod="openstack/barbican-keystone-listener-7d7bc9f7d-fks76" Mar 13 14:27:28 crc kubenswrapper[4907]: I0313 14:27:28.218043 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65531461-cea2-4b2c-a9b6-8fd9e9bffb27-combined-ca-bundle\") pod \"barbican-keystone-listener-7d7bc9f7d-fks76\" (UID: \"65531461-cea2-4b2c-a9b6-8fd9e9bffb27\") " pod="openstack/barbican-keystone-listener-7d7bc9f7d-fks76" Mar 13 14:27:28 crc kubenswrapper[4907]: I0313 14:27:28.218625 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/25b5fafe-65c8-48ba-bc34-83442f8ace4c-config-data-custom\") pod \"barbican-worker-76bbcdf7d9-j6hw5\" (UID: \"25b5fafe-65c8-48ba-bc34-83442f8ace4c\") " pod="openstack/barbican-worker-76bbcdf7d9-j6hw5" Mar 13 14:27:28 crc kubenswrapper[4907]: I0313 14:27:28.222542 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-99bgf\" (UniqueName: \"kubernetes.io/projected/25b5fafe-65c8-48ba-bc34-83442f8ace4c-kube-api-access-99bgf\") pod \"barbican-worker-76bbcdf7d9-j6hw5\" (UID: \"25b5fafe-65c8-48ba-bc34-83442f8ace4c\") " pod="openstack/barbican-worker-76bbcdf7d9-j6hw5" Mar 13 14:27:28 crc kubenswrapper[4907]: I0313 14:27:28.267730 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/032183a7-c5db-4528-a211-803050007b68-combined-ca-bundle\") pod \"barbican-api-74984bd8b4-x42pq\" (UID: \"032183a7-c5db-4528-a211-803050007b68\") " pod="openstack/barbican-api-74984bd8b4-x42pq" Mar 13 14:27:28 crc kubenswrapper[4907]: I0313 14:27:28.267825 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tfsh2\" (UniqueName: \"kubernetes.io/projected/032183a7-c5db-4528-a211-803050007b68-kube-api-access-tfsh2\") pod \"barbican-api-74984bd8b4-x42pq\" (UID: \"032183a7-c5db-4528-a211-803050007b68\") " pod="openstack/barbican-api-74984bd8b4-x42pq" Mar 13 14:27:28 crc kubenswrapper[4907]: I0313 14:27:28.267860 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/032183a7-c5db-4528-a211-803050007b68-logs\") pod \"barbican-api-74984bd8b4-x42pq\" (UID: \"032183a7-c5db-4528-a211-803050007b68\") " pod="openstack/barbican-api-74984bd8b4-x42pq" Mar 13 14:27:28 crc kubenswrapper[4907]: I0313 14:27:28.267955 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/032183a7-c5db-4528-a211-803050007b68-config-data\") pod \"barbican-api-74984bd8b4-x42pq\" (UID: \"032183a7-c5db-4528-a211-803050007b68\") " pod="openstack/barbican-api-74984bd8b4-x42pq" Mar 13 14:27:28 crc kubenswrapper[4907]: I0313 14:27:28.267983 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/032183a7-c5db-4528-a211-803050007b68-config-data-custom\") pod \"barbican-api-74984bd8b4-x42pq\" (UID: \"032183a7-c5db-4528-a211-803050007b68\") " pod="openstack/barbican-api-74984bd8b4-x42pq" Mar 13 14:27:28 crc kubenswrapper[4907]: I0313 14:27:28.269382 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/032183a7-c5db-4528-a211-803050007b68-logs\") pod \"barbican-api-74984bd8b4-x42pq\" (UID: \"032183a7-c5db-4528-a211-803050007b68\") " pod="openstack/barbican-api-74984bd8b4-x42pq" Mar 13 14:27:28 crc kubenswrapper[4907]: I0313 14:27:28.274538 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/032183a7-c5db-4528-a211-803050007b68-config-data-custom\") pod \"barbican-api-74984bd8b4-x42pq\" (UID: \"032183a7-c5db-4528-a211-803050007b68\") " pod="openstack/barbican-api-74984bd8b4-x42pq" Mar 13 14:27:28 crc kubenswrapper[4907]: I0313 14:27:28.281388 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/032183a7-c5db-4528-a211-803050007b68-config-data\") pod \"barbican-api-74984bd8b4-x42pq\" (UID: \"032183a7-c5db-4528-a211-803050007b68\") " pod="openstack/barbican-api-74984bd8b4-x42pq" Mar 13 14:27:28 crc kubenswrapper[4907]: I0313 14:27:28.289167 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/032183a7-c5db-4528-a211-803050007b68-combined-ca-bundle\") pod \"barbican-api-74984bd8b4-x42pq\" (UID: \"032183a7-c5db-4528-a211-803050007b68\") " pod="openstack/barbican-api-74984bd8b4-x42pq" Mar 13 14:27:28 crc kubenswrapper[4907]: I0313 14:27:28.290015 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tfsh2\" (UniqueName: \"kubernetes.io/projected/032183a7-c5db-4528-a211-803050007b68-kube-api-access-tfsh2\") pod \"barbican-api-74984bd8b4-x42pq\" (UID: \"032183a7-c5db-4528-a211-803050007b68\") " pod="openstack/barbican-api-74984bd8b4-x42pq" Mar 13 14:27:28 crc kubenswrapper[4907]: I0313 14:27:28.393323 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-76bbcdf7d9-j6hw5" Mar 13 14:27:28 crc kubenswrapper[4907]: I0313 14:27:28.425418 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-7d7bc9f7d-fks76" Mar 13 14:27:28 crc kubenswrapper[4907]: I0313 14:27:28.438096 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-74984bd8b4-x42pq" Mar 13 14:27:28 crc kubenswrapper[4907]: I0313 14:27:28.897100 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-67d6ccd9c4-ms98l"] Mar 13 14:27:28 crc kubenswrapper[4907]: I0313 14:27:28.924251 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-85b9d64669-dxnf6"] Mar 13 14:27:28 crc kubenswrapper[4907]: W0313 14:27:28.932960 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8a738a66_54bc_4f5e_86d5_6e3004d8f265.slice/crio-50cd2791b5f7c237a1259945e1c40427e424019970480fa895cfa345fab258b4 WatchSource:0}: Error finding container 50cd2791b5f7c237a1259945e1c40427e424019970480fa895cfa345fab258b4: Status 404 returned error can't find the container with id 50cd2791b5f7c237a1259945e1c40427e424019970480fa895cfa345fab258b4 Mar 13 14:27:28 crc kubenswrapper[4907]: I0313 14:27:28.963517 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-c685c5475-rvkf2"] Mar 13 14:27:29 crc kubenswrapper[4907]: W0313 14:27:28.995522 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podca4928ba_1277_4918_8b92_3eda2b276577.slice/crio-6c4a87798f44345ddbfe9a9bbadc067d17cd8c04974906e2bc3559decad063c2 WatchSource:0}: Error finding container 6c4a87798f44345ddbfe9a9bbadc067d17cd8c04974906e2bc3559decad063c2: Status 404 returned error can't find the container with id 6c4a87798f44345ddbfe9a9bbadc067d17cd8c04974906e2bc3559decad063c2 Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.095251 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-54c69f4bdb-lt67r"] Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.105843 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-79fd7f986f-t9wlb"] Mar 13 14:27:29 crc kubenswrapper[4907]: W0313 14:27:29.115042 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2d754127_d8df_48de_b77a_46aa7898c45f.slice/crio-90e9289f409305f490b74370d98aff58f19c0c80ed8fdee54b693082f3b3ca50 WatchSource:0}: Error finding container 90e9289f409305f490b74370d98aff58f19c0c80ed8fdee54b693082f3b3ca50: Status 404 returned error can't find the container with id 90e9289f409305f490b74370d98aff58f19c0c80ed8fdee54b693082f3b3ca50 Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.117315 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-jt88g" Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.123226 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-5d49fd688d-njvfm"] Mar 13 14:27:29 crc kubenswrapper[4907]: W0313 14:27:29.130112 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod070e2b75_05b1_42a4_a34e_950d45d65920.slice/crio-eaad4530bf1fa0795ee15397e0a9c4c262de8913b337803ac449643d1177e1b7 WatchSource:0}: Error finding container eaad4530bf1fa0795ee15397e0a9c4c262de8913b337803ac449643d1177e1b7: Status 404 returned error can't find the container with id eaad4530bf1fa0795ee15397e0a9c4c262de8913b337803ac449643d1177e1b7 Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.203172 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wq68t\" (UniqueName: \"kubernetes.io/projected/870ec55c-fc6f-4b16-8c49-a3cda0d0d010-kube-api-access-wq68t\") pod \"870ec55c-fc6f-4b16-8c49-a3cda0d0d010\" (UID: \"870ec55c-fc6f-4b16-8c49-a3cda0d0d010\") " Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.203277 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/870ec55c-fc6f-4b16-8c49-a3cda0d0d010-config\") pod \"870ec55c-fc6f-4b16-8c49-a3cda0d0d010\" (UID: \"870ec55c-fc6f-4b16-8c49-a3cda0d0d010\") " Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.210550 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/870ec55c-fc6f-4b16-8c49-a3cda0d0d010-kube-api-access-wq68t" (OuterVolumeSpecName: "kube-api-access-wq68t") pod "870ec55c-fc6f-4b16-8c49-a3cda0d0d010" (UID: "870ec55c-fc6f-4b16-8c49-a3cda0d0d010"). InnerVolumeSpecName "kube-api-access-wq68t". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.284302 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/870ec55c-fc6f-4b16-8c49-a3cda0d0d010-config" (OuterVolumeSpecName: "config") pod "870ec55c-fc6f-4b16-8c49-a3cda0d0d010" (UID: "870ec55c-fc6f-4b16-8c49-a3cda0d0d010"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.289918 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-74984bd8b4-x42pq"] Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.310415 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-7d7bc9f7d-fks76"] Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.311976 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/870ec55c-fc6f-4b16-8c49-a3cda0d0d010-combined-ca-bundle\") pod \"870ec55c-fc6f-4b16-8c49-a3cda0d0d010\" (UID: \"870ec55c-fc6f-4b16-8c49-a3cda0d0d010\") " Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.312650 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wq68t\" (UniqueName: \"kubernetes.io/projected/870ec55c-fc6f-4b16-8c49-a3cda0d0d010-kube-api-access-wq68t\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.312663 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/870ec55c-fc6f-4b16-8c49-a3cda0d0d010-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.321169 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-76bbcdf7d9-j6hw5"] Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.333390 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-jt88g" event={"ID":"870ec55c-fc6f-4b16-8c49-a3cda0d0d010","Type":"ContainerDied","Data":"61484e5c865e06dbc491f0726df5e379526c160f176a2435939232b317293ca1"} Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.333443 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="61484e5c865e06dbc491f0726df5e379526c160f176a2435939232b317293ca1" Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.333510 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-jt88g" Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.356586 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-85b9d64669-dxnf6" event={"ID":"8a738a66-54bc-4f5e-86d5-6e3004d8f265","Type":"ContainerStarted","Data":"50cd2791b5f7c237a1259945e1c40427e424019970480fa895cfa345fab258b4"} Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.381361 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-54c69f4bdb-lt67r" event={"ID":"fcbb1687-db0b-4fc6-937c-89102831237b","Type":"ContainerStarted","Data":"7a66d01a94d06b4565675b1fcb3708a34133ebf672d54a557137e862ae4fd1c3"} Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.399193 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-c685c5475-rvkf2" event={"ID":"ca4928ba-1277-4918-8b92-3eda2b276577","Type":"ContainerStarted","Data":"6c4a87798f44345ddbfe9a9bbadc067d17cd8c04974906e2bc3559decad063c2"} Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.413042 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/870ec55c-fc6f-4b16-8c49-a3cda0d0d010-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "870ec55c-fc6f-4b16-8c49-a3cda0d0d010" (UID: "870ec55c-fc6f-4b16-8c49-a3cda0d0d010"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.415317 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/870ec55c-fc6f-4b16-8c49-a3cda0d0d010-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.434294 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5d49fd688d-njvfm" event={"ID":"070e2b75-05b1-42a4-a34e-950d45d65920","Type":"ContainerStarted","Data":"eaad4530bf1fa0795ee15397e0a9c4c262de8913b337803ac449643d1177e1b7"} Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.438011 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-79fd7f986f-t9wlb"] Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.445159 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-67d6ccd9c4-ms98l" event={"ID":"45d14b2b-1d0e-4e15-b837-0d2fb6aed785","Type":"ContainerStarted","Data":"e2ad3bad04a843173cf63dc1e67628c441f9aeb3514e5f9bf8afda3eba7b52f6"} Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.445206 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-67d6ccd9c4-ms98l" event={"ID":"45d14b2b-1d0e-4e15-b837-0d2fb6aed785","Type":"ContainerStarted","Data":"0497ac4d8503d6e45cbe04af5eff7d939b60b6a6232917e037fb11fdbb11b8dd"} Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.476629 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79fd7f986f-t9wlb" event={"ID":"2d754127-d8df-48de-b77a-46aa7898c45f","Type":"ContainerStarted","Data":"90e9289f409305f490b74370d98aff58f19c0c80ed8fdee54b693082f3b3ca50"} Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.480854 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7fc46d7df7-77dng"] Mar 13 14:27:29 crc kubenswrapper[4907]: E0313 14:27:29.481297 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="870ec55c-fc6f-4b16-8c49-a3cda0d0d010" containerName="neutron-db-sync" Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.481310 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="870ec55c-fc6f-4b16-8c49-a3cda0d0d010" containerName="neutron-db-sync" Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.481497 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="870ec55c-fc6f-4b16-8c49-a3cda0d0d010" containerName="neutron-db-sync" Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.482472 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fc46d7df7-77dng" Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.519008 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/49bc69d8-fe85-4994-a5cf-c51472a124e9-dns-swift-storage-0\") pod \"dnsmasq-dns-7fc46d7df7-77dng\" (UID: \"49bc69d8-fe85-4994-a5cf-c51472a124e9\") " pod="openstack/dnsmasq-dns-7fc46d7df7-77dng" Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.519051 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/49bc69d8-fe85-4994-a5cf-c51472a124e9-dns-svc\") pod \"dnsmasq-dns-7fc46d7df7-77dng\" (UID: \"49bc69d8-fe85-4994-a5cf-c51472a124e9\") " pod="openstack/dnsmasq-dns-7fc46d7df7-77dng" Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.519127 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/49bc69d8-fe85-4994-a5cf-c51472a124e9-ovsdbserver-sb\") pod \"dnsmasq-dns-7fc46d7df7-77dng\" (UID: \"49bc69d8-fe85-4994-a5cf-c51472a124e9\") " pod="openstack/dnsmasq-dns-7fc46d7df7-77dng" Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.519158 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r77nl\" (UniqueName: \"kubernetes.io/projected/49bc69d8-fe85-4994-a5cf-c51472a124e9-kube-api-access-r77nl\") pod \"dnsmasq-dns-7fc46d7df7-77dng\" (UID: \"49bc69d8-fe85-4994-a5cf-c51472a124e9\") " pod="openstack/dnsmasq-dns-7fc46d7df7-77dng" Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.519280 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/49bc69d8-fe85-4994-a5cf-c51472a124e9-ovsdbserver-nb\") pod \"dnsmasq-dns-7fc46d7df7-77dng\" (UID: \"49bc69d8-fe85-4994-a5cf-c51472a124e9\") " pod="openstack/dnsmasq-dns-7fc46d7df7-77dng" Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.519322 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/49bc69d8-fe85-4994-a5cf-c51472a124e9-config\") pod \"dnsmasq-dns-7fc46d7df7-77dng\" (UID: \"49bc69d8-fe85-4994-a5cf-c51472a124e9\") " pod="openstack/dnsmasq-dns-7fc46d7df7-77dng" Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.545127 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7fc46d7df7-77dng"] Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.621011 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r77nl\" (UniqueName: \"kubernetes.io/projected/49bc69d8-fe85-4994-a5cf-c51472a124e9-kube-api-access-r77nl\") pod \"dnsmasq-dns-7fc46d7df7-77dng\" (UID: \"49bc69d8-fe85-4994-a5cf-c51472a124e9\") " pod="openstack/dnsmasq-dns-7fc46d7df7-77dng" Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.621056 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/49bc69d8-fe85-4994-a5cf-c51472a124e9-ovsdbserver-nb\") pod \"dnsmasq-dns-7fc46d7df7-77dng\" (UID: \"49bc69d8-fe85-4994-a5cf-c51472a124e9\") " pod="openstack/dnsmasq-dns-7fc46d7df7-77dng" Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.621081 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/49bc69d8-fe85-4994-a5cf-c51472a124e9-config\") pod \"dnsmasq-dns-7fc46d7df7-77dng\" (UID: \"49bc69d8-fe85-4994-a5cf-c51472a124e9\") " pod="openstack/dnsmasq-dns-7fc46d7df7-77dng" Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.621137 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/49bc69d8-fe85-4994-a5cf-c51472a124e9-dns-swift-storage-0\") pod \"dnsmasq-dns-7fc46d7df7-77dng\" (UID: \"49bc69d8-fe85-4994-a5cf-c51472a124e9\") " pod="openstack/dnsmasq-dns-7fc46d7df7-77dng" Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.621155 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/49bc69d8-fe85-4994-a5cf-c51472a124e9-dns-svc\") pod \"dnsmasq-dns-7fc46d7df7-77dng\" (UID: \"49bc69d8-fe85-4994-a5cf-c51472a124e9\") " pod="openstack/dnsmasq-dns-7fc46d7df7-77dng" Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.621236 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/49bc69d8-fe85-4994-a5cf-c51472a124e9-ovsdbserver-sb\") pod \"dnsmasq-dns-7fc46d7df7-77dng\" (UID: \"49bc69d8-fe85-4994-a5cf-c51472a124e9\") " pod="openstack/dnsmasq-dns-7fc46d7df7-77dng" Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.621972 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/49bc69d8-fe85-4994-a5cf-c51472a124e9-ovsdbserver-sb\") pod \"dnsmasq-dns-7fc46d7df7-77dng\" (UID: \"49bc69d8-fe85-4994-a5cf-c51472a124e9\") " pod="openstack/dnsmasq-dns-7fc46d7df7-77dng" Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.624976 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/49bc69d8-fe85-4994-a5cf-c51472a124e9-config\") pod \"dnsmasq-dns-7fc46d7df7-77dng\" (UID: \"49bc69d8-fe85-4994-a5cf-c51472a124e9\") " pod="openstack/dnsmasq-dns-7fc46d7df7-77dng" Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.625604 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/49bc69d8-fe85-4994-a5cf-c51472a124e9-ovsdbserver-nb\") pod \"dnsmasq-dns-7fc46d7df7-77dng\" (UID: \"49bc69d8-fe85-4994-a5cf-c51472a124e9\") " pod="openstack/dnsmasq-dns-7fc46d7df7-77dng" Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.626467 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/49bc69d8-fe85-4994-a5cf-c51472a124e9-dns-svc\") pod \"dnsmasq-dns-7fc46d7df7-77dng\" (UID: \"49bc69d8-fe85-4994-a5cf-c51472a124e9\") " pod="openstack/dnsmasq-dns-7fc46d7df7-77dng" Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.629516 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/49bc69d8-fe85-4994-a5cf-c51472a124e9-dns-swift-storage-0\") pod \"dnsmasq-dns-7fc46d7df7-77dng\" (UID: \"49bc69d8-fe85-4994-a5cf-c51472a124e9\") " pod="openstack/dnsmasq-dns-7fc46d7df7-77dng" Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.663267 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r77nl\" (UniqueName: \"kubernetes.io/projected/49bc69d8-fe85-4994-a5cf-c51472a124e9-kube-api-access-r77nl\") pod \"dnsmasq-dns-7fc46d7df7-77dng\" (UID: \"49bc69d8-fe85-4994-a5cf-c51472a124e9\") " pod="openstack/dnsmasq-dns-7fc46d7df7-77dng" Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.677638 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-944cbc5bb-fj29g"] Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.679331 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-944cbc5bb-fj29g" Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.687757 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-tvcdf" Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.688072 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.688226 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.688843 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.731740 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-944cbc5bb-fj29g"] Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.833453 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/10a74a61-272c-4c54-9b2a-6379b77c9984-httpd-config\") pod \"neutron-944cbc5bb-fj29g\" (UID: \"10a74a61-272c-4c54-9b2a-6379b77c9984\") " pod="openstack/neutron-944cbc5bb-fj29g" Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.834423 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10a74a61-272c-4c54-9b2a-6379b77c9984-combined-ca-bundle\") pod \"neutron-944cbc5bb-fj29g\" (UID: \"10a74a61-272c-4c54-9b2a-6379b77c9984\") " pod="openstack/neutron-944cbc5bb-fj29g" Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.834565 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/10a74a61-272c-4c54-9b2a-6379b77c9984-ovndb-tls-certs\") pod \"neutron-944cbc5bb-fj29g\" (UID: \"10a74a61-272c-4c54-9b2a-6379b77c9984\") " pod="openstack/neutron-944cbc5bb-fj29g" Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.834775 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/10a74a61-272c-4c54-9b2a-6379b77c9984-config\") pod \"neutron-944cbc5bb-fj29g\" (UID: \"10a74a61-272c-4c54-9b2a-6379b77c9984\") " pod="openstack/neutron-944cbc5bb-fj29g" Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.835778 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gsjwt\" (UniqueName: \"kubernetes.io/projected/10a74a61-272c-4c54-9b2a-6379b77c9984-kube-api-access-gsjwt\") pod \"neutron-944cbc5bb-fj29g\" (UID: \"10a74a61-272c-4c54-9b2a-6379b77c9984\") " pod="openstack/neutron-944cbc5bb-fj29g" Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.871221 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fc46d7df7-77dng" Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.939580 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/10a74a61-272c-4c54-9b2a-6379b77c9984-ovndb-tls-certs\") pod \"neutron-944cbc5bb-fj29g\" (UID: \"10a74a61-272c-4c54-9b2a-6379b77c9984\") " pod="openstack/neutron-944cbc5bb-fj29g" Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.939681 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/10a74a61-272c-4c54-9b2a-6379b77c9984-config\") pod \"neutron-944cbc5bb-fj29g\" (UID: \"10a74a61-272c-4c54-9b2a-6379b77c9984\") " pod="openstack/neutron-944cbc5bb-fj29g" Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.939701 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gsjwt\" (UniqueName: \"kubernetes.io/projected/10a74a61-272c-4c54-9b2a-6379b77c9984-kube-api-access-gsjwt\") pod \"neutron-944cbc5bb-fj29g\" (UID: \"10a74a61-272c-4c54-9b2a-6379b77c9984\") " pod="openstack/neutron-944cbc5bb-fj29g" Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.939799 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/10a74a61-272c-4c54-9b2a-6379b77c9984-httpd-config\") pod \"neutron-944cbc5bb-fj29g\" (UID: \"10a74a61-272c-4c54-9b2a-6379b77c9984\") " pod="openstack/neutron-944cbc5bb-fj29g" Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.939925 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10a74a61-272c-4c54-9b2a-6379b77c9984-combined-ca-bundle\") pod \"neutron-944cbc5bb-fj29g\" (UID: \"10a74a61-272c-4c54-9b2a-6379b77c9984\") " pod="openstack/neutron-944cbc5bb-fj29g" Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.946316 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10a74a61-272c-4c54-9b2a-6379b77c9984-combined-ca-bundle\") pod \"neutron-944cbc5bb-fj29g\" (UID: \"10a74a61-272c-4c54-9b2a-6379b77c9984\") " pod="openstack/neutron-944cbc5bb-fj29g" Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.948589 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/10a74a61-272c-4c54-9b2a-6379b77c9984-httpd-config\") pod \"neutron-944cbc5bb-fj29g\" (UID: \"10a74a61-272c-4c54-9b2a-6379b77c9984\") " pod="openstack/neutron-944cbc5bb-fj29g" Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.951991 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/10a74a61-272c-4c54-9b2a-6379b77c9984-ovndb-tls-certs\") pod \"neutron-944cbc5bb-fj29g\" (UID: \"10a74a61-272c-4c54-9b2a-6379b77c9984\") " pod="openstack/neutron-944cbc5bb-fj29g" Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.953615 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/10a74a61-272c-4c54-9b2a-6379b77c9984-config\") pod \"neutron-944cbc5bb-fj29g\" (UID: \"10a74a61-272c-4c54-9b2a-6379b77c9984\") " pod="openstack/neutron-944cbc5bb-fj29g" Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.974439 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gsjwt\" (UniqueName: \"kubernetes.io/projected/10a74a61-272c-4c54-9b2a-6379b77c9984-kube-api-access-gsjwt\") pod \"neutron-944cbc5bb-fj29g\" (UID: \"10a74a61-272c-4c54-9b2a-6379b77c9984\") " pod="openstack/neutron-944cbc5bb-fj29g" Mar 13 14:27:30 crc kubenswrapper[4907]: I0313 14:27:30.023923 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-944cbc5bb-fj29g" Mar 13 14:27:30 crc kubenswrapper[4907]: I0313 14:27:30.330441 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7fc46d7df7-77dng"] Mar 13 14:27:30 crc kubenswrapper[4907]: I0313 14:27:30.494720 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-67d6ccd9c4-ms98l" event={"ID":"45d14b2b-1d0e-4e15-b837-0d2fb6aed785","Type":"ContainerStarted","Data":"38b798d627fc6c621c04314f998e3ab08f943e11ebed179dd2fc6641c6c999a2"} Mar 13 14:27:30 crc kubenswrapper[4907]: I0313 14:27:30.495388 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-67d6ccd9c4-ms98l" Mar 13 14:27:30 crc kubenswrapper[4907]: I0313 14:27:30.495418 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-67d6ccd9c4-ms98l" Mar 13 14:27:30 crc kubenswrapper[4907]: I0313 14:27:30.501366 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-76bbcdf7d9-j6hw5" event={"ID":"25b5fafe-65c8-48ba-bc34-83442f8ace4c","Type":"ContainerStarted","Data":"57bab78a65adf0bc133213358e1f66cc1c97a4b0aefeb088fc087562e2ca2999"} Mar 13 14:27:30 crc kubenswrapper[4907]: I0313 14:27:30.504379 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fc46d7df7-77dng" event={"ID":"49bc69d8-fe85-4994-a5cf-c51472a124e9","Type":"ContainerStarted","Data":"74e19d820a0370c065d553cb0dbb43ed632fe5214db6fe0a8a5f25c3d63f84ae"} Mar 13 14:27:30 crc kubenswrapper[4907]: I0313 14:27:30.507085 4907 generic.go:334] "Generic (PLEG): container finished" podID="2d754127-d8df-48de-b77a-46aa7898c45f" containerID="9ea73af353594fa5d86ca1191ad584d04081c81707270b01c9c5415f0d4cc909" exitCode=0 Mar 13 14:27:30 crc kubenswrapper[4907]: I0313 14:27:30.507154 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79fd7f986f-t9wlb" event={"ID":"2d754127-d8df-48de-b77a-46aa7898c45f","Type":"ContainerDied","Data":"9ea73af353594fa5d86ca1191ad584d04081c81707270b01c9c5415f0d4cc909"} Mar 13 14:27:30 crc kubenswrapper[4907]: I0313 14:27:30.508974 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-74984bd8b4-x42pq" event={"ID":"032183a7-c5db-4528-a211-803050007b68","Type":"ContainerStarted","Data":"be905f5a0aa412e370260df6ece658648121619b19bad67aee3af44172b10250"} Mar 13 14:27:30 crc kubenswrapper[4907]: I0313 14:27:30.509009 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-74984bd8b4-x42pq" event={"ID":"032183a7-c5db-4528-a211-803050007b68","Type":"ContainerStarted","Data":"3dc5784d660cfabfa43e3df490d1b2200ffa14fc3ff516c9488bcc67f04af55b"} Mar 13 14:27:30 crc kubenswrapper[4907]: I0313 14:27:30.509023 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-74984bd8b4-x42pq" event={"ID":"032183a7-c5db-4528-a211-803050007b68","Type":"ContainerStarted","Data":"3fa9f3f934d41bda9d2c0367bd05516b9fcfddd3de52bfe4494f0529db4eafd7"} Mar 13 14:27:30 crc kubenswrapper[4907]: I0313 14:27:30.510392 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-85b9d64669-dxnf6" event={"ID":"8a738a66-54bc-4f5e-86d5-6e3004d8f265","Type":"ContainerStarted","Data":"e6a1933783ac7223012561b093ce055e734ac2c3c802390f6f0bf14354694b72"} Mar 13 14:27:30 crc kubenswrapper[4907]: I0313 14:27:30.510538 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-85b9d64669-dxnf6" Mar 13 14:27:30 crc kubenswrapper[4907]: I0313 14:27:30.511623 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7d7bc9f7d-fks76" event={"ID":"65531461-cea2-4b2c-a9b6-8fd9e9bffb27","Type":"ContainerStarted","Data":"d83bd2c5278d976f4863dc597e16a40e27298e4624048379804b2a946c5145cc"} Mar 13 14:27:30 crc kubenswrapper[4907]: I0313 14:27:30.513153 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5d49fd688d-njvfm" event={"ID":"070e2b75-05b1-42a4-a34e-950d45d65920","Type":"ContainerStarted","Data":"8ce806c1853f002c1ca5aee8d1ea69c43b2c38c4c666d7df49a773673406c573"} Mar 13 14:27:30 crc kubenswrapper[4907]: I0313 14:27:30.513199 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5d49fd688d-njvfm" event={"ID":"070e2b75-05b1-42a4-a34e-950d45d65920","Type":"ContainerStarted","Data":"57e7ca150948c3586cf56019f626e80fe0c0e49f021cce99a8bbf1bcfa081624"} Mar 13 14:27:30 crc kubenswrapper[4907]: I0313 14:27:30.513311 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5d49fd688d-njvfm" Mar 13 14:27:30 crc kubenswrapper[4907]: I0313 14:27:30.539862 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-67d6ccd9c4-ms98l" podStartSLOduration=3.539841278 podStartE2EDuration="3.539841278s" podCreationTimestamp="2026-03-13 14:27:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:27:30.525780136 +0000 UTC m=+1349.425567825" watchObservedRunningTime="2026-03-13 14:27:30.539841278 +0000 UTC m=+1349.439628967" Mar 13 14:27:30 crc kubenswrapper[4907]: I0313 14:27:30.572461 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-85b9d64669-dxnf6" podStartSLOduration=3.572435484 podStartE2EDuration="3.572435484s" podCreationTimestamp="2026-03-13 14:27:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:27:30.562618588 +0000 UTC m=+1349.462406277" watchObservedRunningTime="2026-03-13 14:27:30.572435484 +0000 UTC m=+1349.472223173" Mar 13 14:27:30 crc kubenswrapper[4907]: I0313 14:27:30.574415 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Mar 13 14:27:30 crc kubenswrapper[4907]: I0313 14:27:30.574469 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Mar 13 14:27:30 crc kubenswrapper[4907]: I0313 14:27:30.608818 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-5d49fd688d-njvfm" podStartSLOduration=3.608799904 podStartE2EDuration="3.608799904s" podCreationTimestamp="2026-03-13 14:27:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:27:30.597497176 +0000 UTC m=+1349.497284875" watchObservedRunningTime="2026-03-13 14:27:30.608799904 +0000 UTC m=+1349.508587593" Mar 13 14:27:30 crc kubenswrapper[4907]: I0313 14:27:30.660298 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Mar 13 14:27:30 crc kubenswrapper[4907]: I0313 14:27:30.660407 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Mar 13 14:27:30 crc kubenswrapper[4907]: I0313 14:27:30.804396 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-944cbc5bb-fj29g"] Mar 13 14:27:30 crc kubenswrapper[4907]: I0313 14:27:30.972858 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79fd7f986f-t9wlb" Mar 13 14:27:31 crc kubenswrapper[4907]: I0313 14:27:31.068660 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2d754127-d8df-48de-b77a-46aa7898c45f-config\") pod \"2d754127-d8df-48de-b77a-46aa7898c45f\" (UID: \"2d754127-d8df-48de-b77a-46aa7898c45f\") " Mar 13 14:27:31 crc kubenswrapper[4907]: I0313 14:27:31.068819 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2d754127-d8df-48de-b77a-46aa7898c45f-dns-swift-storage-0\") pod \"2d754127-d8df-48de-b77a-46aa7898c45f\" (UID: \"2d754127-d8df-48de-b77a-46aa7898c45f\") " Mar 13 14:27:31 crc kubenswrapper[4907]: I0313 14:27:31.068870 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2d754127-d8df-48de-b77a-46aa7898c45f-dns-svc\") pod \"2d754127-d8df-48de-b77a-46aa7898c45f\" (UID: \"2d754127-d8df-48de-b77a-46aa7898c45f\") " Mar 13 14:27:31 crc kubenswrapper[4907]: I0313 14:27:31.071085 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2d754127-d8df-48de-b77a-46aa7898c45f-ovsdbserver-sb\") pod \"2d754127-d8df-48de-b77a-46aa7898c45f\" (UID: \"2d754127-d8df-48de-b77a-46aa7898c45f\") " Mar 13 14:27:31 crc kubenswrapper[4907]: I0313 14:27:31.071205 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2d754127-d8df-48de-b77a-46aa7898c45f-ovsdbserver-nb\") pod \"2d754127-d8df-48de-b77a-46aa7898c45f\" (UID: \"2d754127-d8df-48de-b77a-46aa7898c45f\") " Mar 13 14:27:31 crc kubenswrapper[4907]: I0313 14:27:31.071319 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dhzwg\" (UniqueName: \"kubernetes.io/projected/2d754127-d8df-48de-b77a-46aa7898c45f-kube-api-access-dhzwg\") pod \"2d754127-d8df-48de-b77a-46aa7898c45f\" (UID: \"2d754127-d8df-48de-b77a-46aa7898c45f\") " Mar 13 14:27:31 crc kubenswrapper[4907]: I0313 14:27:31.079035 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2d754127-d8df-48de-b77a-46aa7898c45f-kube-api-access-dhzwg" (OuterVolumeSpecName: "kube-api-access-dhzwg") pod "2d754127-d8df-48de-b77a-46aa7898c45f" (UID: "2d754127-d8df-48de-b77a-46aa7898c45f"). InnerVolumeSpecName "kube-api-access-dhzwg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:27:31 crc kubenswrapper[4907]: I0313 14:27:31.119374 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2d754127-d8df-48de-b77a-46aa7898c45f-config" (OuterVolumeSpecName: "config") pod "2d754127-d8df-48de-b77a-46aa7898c45f" (UID: "2d754127-d8df-48de-b77a-46aa7898c45f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:27:31 crc kubenswrapper[4907]: I0313 14:27:31.133979 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2d754127-d8df-48de-b77a-46aa7898c45f-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "2d754127-d8df-48de-b77a-46aa7898c45f" (UID: "2d754127-d8df-48de-b77a-46aa7898c45f"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:27:31 crc kubenswrapper[4907]: I0313 14:27:31.134291 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2d754127-d8df-48de-b77a-46aa7898c45f-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "2d754127-d8df-48de-b77a-46aa7898c45f" (UID: "2d754127-d8df-48de-b77a-46aa7898c45f"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:27:31 crc kubenswrapper[4907]: I0313 14:27:31.132808 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2d754127-d8df-48de-b77a-46aa7898c45f-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "2d754127-d8df-48de-b77a-46aa7898c45f" (UID: "2d754127-d8df-48de-b77a-46aa7898c45f"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:27:31 crc kubenswrapper[4907]: I0313 14:27:31.159772 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2d754127-d8df-48de-b77a-46aa7898c45f-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "2d754127-d8df-48de-b77a-46aa7898c45f" (UID: "2d754127-d8df-48de-b77a-46aa7898c45f"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:27:31 crc kubenswrapper[4907]: I0313 14:27:31.177356 4907 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2d754127-d8df-48de-b77a-46aa7898c45f-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:31 crc kubenswrapper[4907]: I0313 14:27:31.177398 4907 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2d754127-d8df-48de-b77a-46aa7898c45f-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:31 crc kubenswrapper[4907]: I0313 14:27:31.177413 4907 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2d754127-d8df-48de-b77a-46aa7898c45f-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:31 crc kubenswrapper[4907]: I0313 14:27:31.177424 4907 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2d754127-d8df-48de-b77a-46aa7898c45f-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:31 crc kubenswrapper[4907]: I0313 14:27:31.177437 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dhzwg\" (UniqueName: \"kubernetes.io/projected/2d754127-d8df-48de-b77a-46aa7898c45f-kube-api-access-dhzwg\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:31 crc kubenswrapper[4907]: I0313 14:27:31.177449 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2d754127-d8df-48de-b77a-46aa7898c45f-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:31 crc kubenswrapper[4907]: I0313 14:27:31.522477 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-944cbc5bb-fj29g" event={"ID":"10a74a61-272c-4c54-9b2a-6379b77c9984","Type":"ContainerStarted","Data":"b6a3399278c9114cdb41d239ecf9266ca04769841425ab359d3ec15a0ea2369a"} Mar 13 14:27:31 crc kubenswrapper[4907]: I0313 14:27:31.527637 4907 generic.go:334] "Generic (PLEG): container finished" podID="49bc69d8-fe85-4994-a5cf-c51472a124e9" containerID="577048fe225fd9933f67497a9e39e7a61a8a9a2ce395fd1bd05afbdcf17490ca" exitCode=0 Mar 13 14:27:31 crc kubenswrapper[4907]: I0313 14:27:31.527684 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fc46d7df7-77dng" event={"ID":"49bc69d8-fe85-4994-a5cf-c51472a124e9","Type":"ContainerDied","Data":"577048fe225fd9933f67497a9e39e7a61a8a9a2ce395fd1bd05afbdcf17490ca"} Mar 13 14:27:31 crc kubenswrapper[4907]: I0313 14:27:31.530376 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79fd7f986f-t9wlb" Mar 13 14:27:31 crc kubenswrapper[4907]: I0313 14:27:31.530419 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79fd7f986f-t9wlb" event={"ID":"2d754127-d8df-48de-b77a-46aa7898c45f","Type":"ContainerDied","Data":"90e9289f409305f490b74370d98aff58f19c0c80ed8fdee54b693082f3b3ca50"} Mar 13 14:27:31 crc kubenswrapper[4907]: I0313 14:27:31.530484 4907 scope.go:117] "RemoveContainer" containerID="9ea73af353594fa5d86ca1191ad584d04081c81707270b01c9c5415f0d4cc909" Mar 13 14:27:31 crc kubenswrapper[4907]: I0313 14:27:31.531432 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5d49fd688d-njvfm" Mar 13 14:27:31 crc kubenswrapper[4907]: I0313 14:27:31.531469 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Mar 13 14:27:31 crc kubenswrapper[4907]: I0313 14:27:31.531481 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Mar 13 14:27:31 crc kubenswrapper[4907]: I0313 14:27:31.580238 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-74984bd8b4-x42pq" podStartSLOduration=4.580218064 podStartE2EDuration="4.580218064s" podCreationTimestamp="2026-03-13 14:27:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:27:31.566056919 +0000 UTC m=+1350.465844628" watchObservedRunningTime="2026-03-13 14:27:31.580218064 +0000 UTC m=+1350.480005753" Mar 13 14:27:31 crc kubenswrapper[4907]: I0313 14:27:31.629926 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-79fd7f986f-t9wlb"] Mar 13 14:27:31 crc kubenswrapper[4907]: I0313 14:27:31.640608 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-79fd7f986f-t9wlb"] Mar 13 14:27:31 crc kubenswrapper[4907]: I0313 14:27:31.800589 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2d754127-d8df-48de-b77a-46aa7898c45f" path="/var/lib/kubelet/pods/2d754127-d8df-48de-b77a-46aa7898c45f/volumes" Mar 13 14:27:32 crc kubenswrapper[4907]: I0313 14:27:32.308257 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Mar 13 14:27:32 crc kubenswrapper[4907]: I0313 14:27:32.308650 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Mar 13 14:27:32 crc kubenswrapper[4907]: I0313 14:27:32.350037 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Mar 13 14:27:32 crc kubenswrapper[4907]: I0313 14:27:32.365583 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Mar 13 14:27:32 crc kubenswrapper[4907]: I0313 14:27:32.552716 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Mar 13 14:27:32 crc kubenswrapper[4907]: I0313 14:27:32.552746 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Mar 13 14:27:33 crc kubenswrapper[4907]: I0313 14:27:33.438430 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-74984bd8b4-x42pq" Mar 13 14:27:33 crc kubenswrapper[4907]: I0313 14:27:33.438979 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-74984bd8b4-x42pq" Mar 13 14:27:33 crc kubenswrapper[4907]: I0313 14:27:33.571470 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fc46d7df7-77dng" event={"ID":"49bc69d8-fe85-4994-a5cf-c51472a124e9","Type":"ContainerStarted","Data":"10817f472001c46a816f4b77aa0c7ef3e99533487d552c19bd292e7357506e9f"} Mar 13 14:27:33 crc kubenswrapper[4907]: I0313 14:27:33.571545 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7fc46d7df7-77dng" Mar 13 14:27:33 crc kubenswrapper[4907]: I0313 14:27:33.575450 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7d7bc9f7d-fks76" event={"ID":"65531461-cea2-4b2c-a9b6-8fd9e9bffb27","Type":"ContainerStarted","Data":"3ccd8eb5dd33f0cece989e6411df4852cb55e2cb0b138f16332e7a07ccf5c865"} Mar 13 14:27:33 crc kubenswrapper[4907]: I0313 14:27:33.577610 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-54c69f4bdb-lt67r" event={"ID":"fcbb1687-db0b-4fc6-937c-89102831237b","Type":"ContainerStarted","Data":"6ad9081a27708afd67b77cedc2aa19d22a6afb40f65f24b3a9da7ca3a519bf38"} Mar 13 14:27:33 crc kubenswrapper[4907]: I0313 14:27:33.579294 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-c685c5475-rvkf2" event={"ID":"ca4928ba-1277-4918-8b92-3eda2b276577","Type":"ContainerStarted","Data":"bb7f9aa2bb828fff3a17bc00a23c15ddb0d4b9b7ae4d6e61a19fc6386d089dbb"} Mar 13 14:27:33 crc kubenswrapper[4907]: I0313 14:27:33.580930 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-76bbcdf7d9-j6hw5" event={"ID":"25b5fafe-65c8-48ba-bc34-83442f8ace4c","Type":"ContainerStarted","Data":"b5356b2175cd794b4fc6692eb61775ae05e66fb791f7dbd913a5d208e5880c25"} Mar 13 14:27:33 crc kubenswrapper[4907]: I0313 14:27:33.582981 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-944cbc5bb-fj29g" event={"ID":"10a74a61-272c-4c54-9b2a-6379b77c9984","Type":"ContainerStarted","Data":"b624bd9d00cbcadf5cdc9520721c5131708177fac0d6006945f6f3ffc27df6cd"} Mar 13 14:27:33 crc kubenswrapper[4907]: I0313 14:27:33.583022 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-944cbc5bb-fj29g" event={"ID":"10a74a61-272c-4c54-9b2a-6379b77c9984","Type":"ContainerStarted","Data":"a7d63138378c0ce32dc24735fd0ab7db352755405ba476da24cb2065858f3b75"} Mar 13 14:27:33 crc kubenswrapper[4907]: I0313 14:27:33.583307 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-944cbc5bb-fj29g" Mar 13 14:27:33 crc kubenswrapper[4907]: I0313 14:27:33.601425 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7fc46d7df7-77dng" podStartSLOduration=4.601405385 podStartE2EDuration="4.601405385s" podCreationTimestamp="2026-03-13 14:27:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:27:33.595768522 +0000 UTC m=+1352.495556211" watchObservedRunningTime="2026-03-13 14:27:33.601405385 +0000 UTC m=+1352.501193074" Mar 13 14:27:33 crc kubenswrapper[4907]: I0313 14:27:33.653440 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-944cbc5bb-fj29g" podStartSLOduration=4.65341697 podStartE2EDuration="4.65341697s" podCreationTimestamp="2026-03-13 14:27:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:27:33.617595455 +0000 UTC m=+1352.517383144" watchObservedRunningTime="2026-03-13 14:27:33.65341697 +0000 UTC m=+1352.553204659" Mar 13 14:27:33 crc kubenswrapper[4907]: I0313 14:27:33.841893 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-7bf587b8d9-pvvbz"] Mar 13 14:27:33 crc kubenswrapper[4907]: E0313 14:27:33.842267 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d754127-d8df-48de-b77a-46aa7898c45f" containerName="init" Mar 13 14:27:33 crc kubenswrapper[4907]: I0313 14:27:33.842284 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d754127-d8df-48de-b77a-46aa7898c45f" containerName="init" Mar 13 14:27:33 crc kubenswrapper[4907]: I0313 14:27:33.842472 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d754127-d8df-48de-b77a-46aa7898c45f" containerName="init" Mar 13 14:27:33 crc kubenswrapper[4907]: I0313 14:27:33.843406 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7bf587b8d9-pvvbz" Mar 13 14:27:33 crc kubenswrapper[4907]: I0313 14:27:33.846593 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Mar 13 14:27:33 crc kubenswrapper[4907]: I0313 14:27:33.846842 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Mar 13 14:27:33 crc kubenswrapper[4907]: I0313 14:27:33.860979 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-7bf587b8d9-pvvbz"] Mar 13 14:27:33 crc kubenswrapper[4907]: I0313 14:27:33.941733 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/001078cc-c01e-4173-9740-4f9ff082c593-ovndb-tls-certs\") pod \"neutron-7bf587b8d9-pvvbz\" (UID: \"001078cc-c01e-4173-9740-4f9ff082c593\") " pod="openstack/neutron-7bf587b8d9-pvvbz" Mar 13 14:27:33 crc kubenswrapper[4907]: I0313 14:27:33.941826 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/001078cc-c01e-4173-9740-4f9ff082c593-combined-ca-bundle\") pod \"neutron-7bf587b8d9-pvvbz\" (UID: \"001078cc-c01e-4173-9740-4f9ff082c593\") " pod="openstack/neutron-7bf587b8d9-pvvbz" Mar 13 14:27:33 crc kubenswrapper[4907]: I0313 14:27:33.941861 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/001078cc-c01e-4173-9740-4f9ff082c593-public-tls-certs\") pod \"neutron-7bf587b8d9-pvvbz\" (UID: \"001078cc-c01e-4173-9740-4f9ff082c593\") " pod="openstack/neutron-7bf587b8d9-pvvbz" Mar 13 14:27:33 crc kubenswrapper[4907]: I0313 14:27:33.941942 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-67wmh\" (UniqueName: \"kubernetes.io/projected/001078cc-c01e-4173-9740-4f9ff082c593-kube-api-access-67wmh\") pod \"neutron-7bf587b8d9-pvvbz\" (UID: \"001078cc-c01e-4173-9740-4f9ff082c593\") " pod="openstack/neutron-7bf587b8d9-pvvbz" Mar 13 14:27:33 crc kubenswrapper[4907]: I0313 14:27:33.941987 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/001078cc-c01e-4173-9740-4f9ff082c593-internal-tls-certs\") pod \"neutron-7bf587b8d9-pvvbz\" (UID: \"001078cc-c01e-4173-9740-4f9ff082c593\") " pod="openstack/neutron-7bf587b8d9-pvvbz" Mar 13 14:27:33 crc kubenswrapper[4907]: I0313 14:27:33.942162 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/001078cc-c01e-4173-9740-4f9ff082c593-httpd-config\") pod \"neutron-7bf587b8d9-pvvbz\" (UID: \"001078cc-c01e-4173-9740-4f9ff082c593\") " pod="openstack/neutron-7bf587b8d9-pvvbz" Mar 13 14:27:33 crc kubenswrapper[4907]: I0313 14:27:33.942270 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/001078cc-c01e-4173-9740-4f9ff082c593-config\") pod \"neutron-7bf587b8d9-pvvbz\" (UID: \"001078cc-c01e-4173-9740-4f9ff082c593\") " pod="openstack/neutron-7bf587b8d9-pvvbz" Mar 13 14:27:34 crc kubenswrapper[4907]: I0313 14:27:34.044911 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/001078cc-c01e-4173-9740-4f9ff082c593-httpd-config\") pod \"neutron-7bf587b8d9-pvvbz\" (UID: \"001078cc-c01e-4173-9740-4f9ff082c593\") " pod="openstack/neutron-7bf587b8d9-pvvbz" Mar 13 14:27:34 crc kubenswrapper[4907]: I0313 14:27:34.045070 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/001078cc-c01e-4173-9740-4f9ff082c593-config\") pod \"neutron-7bf587b8d9-pvvbz\" (UID: \"001078cc-c01e-4173-9740-4f9ff082c593\") " pod="openstack/neutron-7bf587b8d9-pvvbz" Mar 13 14:27:34 crc kubenswrapper[4907]: I0313 14:27:34.045341 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/001078cc-c01e-4173-9740-4f9ff082c593-ovndb-tls-certs\") pod \"neutron-7bf587b8d9-pvvbz\" (UID: \"001078cc-c01e-4173-9740-4f9ff082c593\") " pod="openstack/neutron-7bf587b8d9-pvvbz" Mar 13 14:27:34 crc kubenswrapper[4907]: I0313 14:27:34.045385 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/001078cc-c01e-4173-9740-4f9ff082c593-combined-ca-bundle\") pod \"neutron-7bf587b8d9-pvvbz\" (UID: \"001078cc-c01e-4173-9740-4f9ff082c593\") " pod="openstack/neutron-7bf587b8d9-pvvbz" Mar 13 14:27:34 crc kubenswrapper[4907]: I0313 14:27:34.045420 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/001078cc-c01e-4173-9740-4f9ff082c593-public-tls-certs\") pod \"neutron-7bf587b8d9-pvvbz\" (UID: \"001078cc-c01e-4173-9740-4f9ff082c593\") " pod="openstack/neutron-7bf587b8d9-pvvbz" Mar 13 14:27:34 crc kubenswrapper[4907]: I0313 14:27:34.047968 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-67wmh\" (UniqueName: \"kubernetes.io/projected/001078cc-c01e-4173-9740-4f9ff082c593-kube-api-access-67wmh\") pod \"neutron-7bf587b8d9-pvvbz\" (UID: \"001078cc-c01e-4173-9740-4f9ff082c593\") " pod="openstack/neutron-7bf587b8d9-pvvbz" Mar 13 14:27:34 crc kubenswrapper[4907]: I0313 14:27:34.048070 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/001078cc-c01e-4173-9740-4f9ff082c593-internal-tls-certs\") pod \"neutron-7bf587b8d9-pvvbz\" (UID: \"001078cc-c01e-4173-9740-4f9ff082c593\") " pod="openstack/neutron-7bf587b8d9-pvvbz" Mar 13 14:27:34 crc kubenswrapper[4907]: I0313 14:27:34.053678 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/001078cc-c01e-4173-9740-4f9ff082c593-ovndb-tls-certs\") pod \"neutron-7bf587b8d9-pvvbz\" (UID: \"001078cc-c01e-4173-9740-4f9ff082c593\") " pod="openstack/neutron-7bf587b8d9-pvvbz" Mar 13 14:27:34 crc kubenswrapper[4907]: I0313 14:27:34.054235 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/001078cc-c01e-4173-9740-4f9ff082c593-httpd-config\") pod \"neutron-7bf587b8d9-pvvbz\" (UID: \"001078cc-c01e-4173-9740-4f9ff082c593\") " pod="openstack/neutron-7bf587b8d9-pvvbz" Mar 13 14:27:34 crc kubenswrapper[4907]: I0313 14:27:34.054643 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/001078cc-c01e-4173-9740-4f9ff082c593-config\") pod \"neutron-7bf587b8d9-pvvbz\" (UID: \"001078cc-c01e-4173-9740-4f9ff082c593\") " pod="openstack/neutron-7bf587b8d9-pvvbz" Mar 13 14:27:34 crc kubenswrapper[4907]: I0313 14:27:34.055760 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/001078cc-c01e-4173-9740-4f9ff082c593-public-tls-certs\") pod \"neutron-7bf587b8d9-pvvbz\" (UID: \"001078cc-c01e-4173-9740-4f9ff082c593\") " pod="openstack/neutron-7bf587b8d9-pvvbz" Mar 13 14:27:34 crc kubenswrapper[4907]: I0313 14:27:34.059716 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/001078cc-c01e-4173-9740-4f9ff082c593-combined-ca-bundle\") pod \"neutron-7bf587b8d9-pvvbz\" (UID: \"001078cc-c01e-4173-9740-4f9ff082c593\") " pod="openstack/neutron-7bf587b8d9-pvvbz" Mar 13 14:27:34 crc kubenswrapper[4907]: I0313 14:27:34.077725 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/001078cc-c01e-4173-9740-4f9ff082c593-internal-tls-certs\") pod \"neutron-7bf587b8d9-pvvbz\" (UID: \"001078cc-c01e-4173-9740-4f9ff082c593\") " pod="openstack/neutron-7bf587b8d9-pvvbz" Mar 13 14:27:34 crc kubenswrapper[4907]: I0313 14:27:34.091723 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Mar 13 14:27:34 crc kubenswrapper[4907]: I0313 14:27:34.092107 4907 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 13 14:27:34 crc kubenswrapper[4907]: I0313 14:27:34.093473 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-67wmh\" (UniqueName: \"kubernetes.io/projected/001078cc-c01e-4173-9740-4f9ff082c593-kube-api-access-67wmh\") pod \"neutron-7bf587b8d9-pvvbz\" (UID: \"001078cc-c01e-4173-9740-4f9ff082c593\") " pod="openstack/neutron-7bf587b8d9-pvvbz" Mar 13 14:27:34 crc kubenswrapper[4907]: I0313 14:27:34.189133 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7bf587b8d9-pvvbz" Mar 13 14:27:34 crc kubenswrapper[4907]: I0313 14:27:34.604971 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-54c69f4bdb-lt67r" event={"ID":"fcbb1687-db0b-4fc6-937c-89102831237b","Type":"ContainerStarted","Data":"9fe21d916b51d235fb7ba28d7a6934f6f8ffaa95ccce23ab15efee2b6e69acd5"} Mar 13 14:27:34 crc kubenswrapper[4907]: I0313 14:27:34.629218 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-c685c5475-rvkf2" event={"ID":"ca4928ba-1277-4918-8b92-3eda2b276577","Type":"ContainerStarted","Data":"2ef6cd9eb88308a93f1d78fb9954045701e6f121a3c20317549b8db40b280bd4"} Mar 13 14:27:34 crc kubenswrapper[4907]: I0313 14:27:34.637367 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-76bbcdf7d9-j6hw5" event={"ID":"25b5fafe-65c8-48ba-bc34-83442f8ace4c","Type":"ContainerStarted","Data":"be9d4c930500ca917844d7687055220642421756a075ce078989248af8d81176"} Mar 13 14:27:34 crc kubenswrapper[4907]: I0313 14:27:34.675422 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7d7bc9f7d-fks76" event={"ID":"65531461-cea2-4b2c-a9b6-8fd9e9bffb27","Type":"ContainerStarted","Data":"7117af57f620e2d772f6e2ca6d92edfdc283b3eb640ab414d9483ba1758f36d4"} Mar 13 14:27:34 crc kubenswrapper[4907]: I0313 14:27:34.703502 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-76bbcdf7d9-j6hw5" podStartSLOduration=4.525446015 podStartE2EDuration="7.703478959s" podCreationTimestamp="2026-03-13 14:27:27 +0000 UTC" firstStartedPulling="2026-03-13 14:27:29.451100348 +0000 UTC m=+1348.350888037" lastFinishedPulling="2026-03-13 14:27:32.629133292 +0000 UTC m=+1351.528920981" observedRunningTime="2026-03-13 14:27:34.699736307 +0000 UTC m=+1353.599524016" watchObservedRunningTime="2026-03-13 14:27:34.703478959 +0000 UTC m=+1353.603266648" Mar 13 14:27:34 crc kubenswrapper[4907]: I0313 14:27:34.718233 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-54c69f4bdb-lt67r" podStartSLOduration=4.201840112 podStartE2EDuration="7.718209029s" podCreationTimestamp="2026-03-13 14:27:27 +0000 UTC" firstStartedPulling="2026-03-13 14:27:29.100366668 +0000 UTC m=+1348.000154357" lastFinishedPulling="2026-03-13 14:27:32.616735585 +0000 UTC m=+1351.516523274" observedRunningTime="2026-03-13 14:27:34.652292496 +0000 UTC m=+1353.552080185" watchObservedRunningTime="2026-03-13 14:27:34.718209029 +0000 UTC m=+1353.617996718" Mar 13 14:27:34 crc kubenswrapper[4907]: I0313 14:27:34.751843 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-c685c5475-rvkf2" podStartSLOduration=4.11569887 podStartE2EDuration="7.751821763s" podCreationTimestamp="2026-03-13 14:27:27 +0000 UTC" firstStartedPulling="2026-03-13 14:27:28.999928067 +0000 UTC m=+1347.899715756" lastFinishedPulling="2026-03-13 14:27:32.63605096 +0000 UTC m=+1351.535838649" observedRunningTime="2026-03-13 14:27:34.749385828 +0000 UTC m=+1353.649173527" watchObservedRunningTime="2026-03-13 14:27:34.751821763 +0000 UTC m=+1353.651609462" Mar 13 14:27:34 crc kubenswrapper[4907]: I0313 14:27:34.774342 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-worker-c685c5475-rvkf2"] Mar 13 14:27:34 crc kubenswrapper[4907]: I0313 14:27:34.780920 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-7d7bc9f7d-fks76" podStartSLOduration=4.553303091 podStartE2EDuration="7.780899384s" podCreationTimestamp="2026-03-13 14:27:27 +0000 UTC" firstStartedPulling="2026-03-13 14:27:29.400143181 +0000 UTC m=+1348.299930870" lastFinishedPulling="2026-03-13 14:27:32.627739474 +0000 UTC m=+1351.527527163" observedRunningTime="2026-03-13 14:27:34.779247779 +0000 UTC m=+1353.679035478" watchObservedRunningTime="2026-03-13 14:27:34.780899384 +0000 UTC m=+1353.680687063" Mar 13 14:27:34 crc kubenswrapper[4907]: I0313 14:27:34.819131 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-keystone-listener-54c69f4bdb-lt67r"] Mar 13 14:27:34 crc kubenswrapper[4907]: I0313 14:27:34.885998 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Mar 13 14:27:34 crc kubenswrapper[4907]: I0313 14:27:34.961826 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-7bf587b8d9-pvvbz"] Mar 13 14:27:35 crc kubenswrapper[4907]: I0313 14:27:35.686985 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7bf587b8d9-pvvbz" event={"ID":"001078cc-c01e-4173-9740-4f9ff082c593","Type":"ContainerStarted","Data":"8f41b5a31898b937ed3fab2872e25ea71023ac05d2c81d4552d7e5ae8d96095e"} Mar 13 14:27:35 crc kubenswrapper[4907]: I0313 14:27:35.687263 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7bf587b8d9-pvvbz" event={"ID":"001078cc-c01e-4173-9740-4f9ff082c593","Type":"ContainerStarted","Data":"502cbc857cf878844cea8a77f9b3d530353b13301737ac4a045e3f57f1cad50e"} Mar 13 14:27:35 crc kubenswrapper[4907]: I0313 14:27:35.712710 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-5d49fd688d-njvfm"] Mar 13 14:27:35 crc kubenswrapper[4907]: I0313 14:27:35.712997 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-5d49fd688d-njvfm" podUID="070e2b75-05b1-42a4-a34e-950d45d65920" containerName="barbican-api-log" containerID="cri-o://57e7ca150948c3586cf56019f626e80fe0c0e49f021cce99a8bbf1bcfa081624" gracePeriod=30 Mar 13 14:27:35 crc kubenswrapper[4907]: I0313 14:27:35.713056 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-5d49fd688d-njvfm" podUID="070e2b75-05b1-42a4-a34e-950d45d65920" containerName="barbican-api" containerID="cri-o://8ce806c1853f002c1ca5aee8d1ea69c43b2c38c4c666d7df49a773673406c573" gracePeriod=30 Mar 13 14:27:35 crc kubenswrapper[4907]: I0313 14:27:35.746718 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-7fc86597fb-mt9pk"] Mar 13 14:27:35 crc kubenswrapper[4907]: I0313 14:27:35.747260 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-5d49fd688d-njvfm" podUID="070e2b75-05b1-42a4-a34e-950d45d65920" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.160:9311/healthcheck\": EOF" Mar 13 14:27:35 crc kubenswrapper[4907]: I0313 14:27:35.747502 4907 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5d49fd688d-njvfm" podUID="070e2b75-05b1-42a4-a34e-950d45d65920" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.160:9311/healthcheck\": EOF" Mar 13 14:27:35 crc kubenswrapper[4907]: I0313 14:27:35.747502 4907 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5d49fd688d-njvfm" podUID="070e2b75-05b1-42a4-a34e-950d45d65920" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.160:9311/healthcheck\": EOF" Mar 13 14:27:35 crc kubenswrapper[4907]: I0313 14:27:35.752577 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7fc86597fb-mt9pk" Mar 13 14:27:35 crc kubenswrapper[4907]: I0313 14:27:35.756480 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Mar 13 14:27:35 crc kubenswrapper[4907]: I0313 14:27:35.756808 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Mar 13 14:27:35 crc kubenswrapper[4907]: I0313 14:27:35.767991 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-7fc86597fb-mt9pk"] Mar 13 14:27:35 crc kubenswrapper[4907]: I0313 14:27:35.795936 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1b2f62d4-0584-4cc3-81ed-e067d8db23b1-logs\") pod \"barbican-api-7fc86597fb-mt9pk\" (UID: \"1b2f62d4-0584-4cc3-81ed-e067d8db23b1\") " pod="openstack/barbican-api-7fc86597fb-mt9pk" Mar 13 14:27:35 crc kubenswrapper[4907]: I0313 14:27:35.796023 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b2f62d4-0584-4cc3-81ed-e067d8db23b1-config-data\") pod \"barbican-api-7fc86597fb-mt9pk\" (UID: \"1b2f62d4-0584-4cc3-81ed-e067d8db23b1\") " pod="openstack/barbican-api-7fc86597fb-mt9pk" Mar 13 14:27:35 crc kubenswrapper[4907]: I0313 14:27:35.796271 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1b2f62d4-0584-4cc3-81ed-e067d8db23b1-config-data-custom\") pod \"barbican-api-7fc86597fb-mt9pk\" (UID: \"1b2f62d4-0584-4cc3-81ed-e067d8db23b1\") " pod="openstack/barbican-api-7fc86597fb-mt9pk" Mar 13 14:27:35 crc kubenswrapper[4907]: I0313 14:27:35.796318 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b2f62d4-0584-4cc3-81ed-e067d8db23b1-combined-ca-bundle\") pod \"barbican-api-7fc86597fb-mt9pk\" (UID: \"1b2f62d4-0584-4cc3-81ed-e067d8db23b1\") " pod="openstack/barbican-api-7fc86597fb-mt9pk" Mar 13 14:27:35 crc kubenswrapper[4907]: I0313 14:27:35.796366 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ktf7c\" (UniqueName: \"kubernetes.io/projected/1b2f62d4-0584-4cc3-81ed-e067d8db23b1-kube-api-access-ktf7c\") pod \"barbican-api-7fc86597fb-mt9pk\" (UID: \"1b2f62d4-0584-4cc3-81ed-e067d8db23b1\") " pod="openstack/barbican-api-7fc86597fb-mt9pk" Mar 13 14:27:35 crc kubenswrapper[4907]: I0313 14:27:35.796485 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1b2f62d4-0584-4cc3-81ed-e067d8db23b1-public-tls-certs\") pod \"barbican-api-7fc86597fb-mt9pk\" (UID: \"1b2f62d4-0584-4cc3-81ed-e067d8db23b1\") " pod="openstack/barbican-api-7fc86597fb-mt9pk" Mar 13 14:27:35 crc kubenswrapper[4907]: I0313 14:27:35.796558 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1b2f62d4-0584-4cc3-81ed-e067d8db23b1-internal-tls-certs\") pod \"barbican-api-7fc86597fb-mt9pk\" (UID: \"1b2f62d4-0584-4cc3-81ed-e067d8db23b1\") " pod="openstack/barbican-api-7fc86597fb-mt9pk" Mar 13 14:27:35 crc kubenswrapper[4907]: I0313 14:27:35.878353 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Mar 13 14:27:35 crc kubenswrapper[4907]: I0313 14:27:35.878471 4907 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 13 14:27:35 crc kubenswrapper[4907]: I0313 14:27:35.885706 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Mar 13 14:27:35 crc kubenswrapper[4907]: I0313 14:27:35.899124 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1b2f62d4-0584-4cc3-81ed-e067d8db23b1-public-tls-certs\") pod \"barbican-api-7fc86597fb-mt9pk\" (UID: \"1b2f62d4-0584-4cc3-81ed-e067d8db23b1\") " pod="openstack/barbican-api-7fc86597fb-mt9pk" Mar 13 14:27:35 crc kubenswrapper[4907]: I0313 14:27:35.899190 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1b2f62d4-0584-4cc3-81ed-e067d8db23b1-internal-tls-certs\") pod \"barbican-api-7fc86597fb-mt9pk\" (UID: \"1b2f62d4-0584-4cc3-81ed-e067d8db23b1\") " pod="openstack/barbican-api-7fc86597fb-mt9pk" Mar 13 14:27:35 crc kubenswrapper[4907]: I0313 14:27:35.899240 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1b2f62d4-0584-4cc3-81ed-e067d8db23b1-logs\") pod \"barbican-api-7fc86597fb-mt9pk\" (UID: \"1b2f62d4-0584-4cc3-81ed-e067d8db23b1\") " pod="openstack/barbican-api-7fc86597fb-mt9pk" Mar 13 14:27:35 crc kubenswrapper[4907]: I0313 14:27:35.899274 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b2f62d4-0584-4cc3-81ed-e067d8db23b1-config-data\") pod \"barbican-api-7fc86597fb-mt9pk\" (UID: \"1b2f62d4-0584-4cc3-81ed-e067d8db23b1\") " pod="openstack/barbican-api-7fc86597fb-mt9pk" Mar 13 14:27:35 crc kubenswrapper[4907]: I0313 14:27:35.899398 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1b2f62d4-0584-4cc3-81ed-e067d8db23b1-config-data-custom\") pod \"barbican-api-7fc86597fb-mt9pk\" (UID: \"1b2f62d4-0584-4cc3-81ed-e067d8db23b1\") " pod="openstack/barbican-api-7fc86597fb-mt9pk" Mar 13 14:27:35 crc kubenswrapper[4907]: I0313 14:27:35.899425 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b2f62d4-0584-4cc3-81ed-e067d8db23b1-combined-ca-bundle\") pod \"barbican-api-7fc86597fb-mt9pk\" (UID: \"1b2f62d4-0584-4cc3-81ed-e067d8db23b1\") " pod="openstack/barbican-api-7fc86597fb-mt9pk" Mar 13 14:27:35 crc kubenswrapper[4907]: I0313 14:27:35.899456 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ktf7c\" (UniqueName: \"kubernetes.io/projected/1b2f62d4-0584-4cc3-81ed-e067d8db23b1-kube-api-access-ktf7c\") pod \"barbican-api-7fc86597fb-mt9pk\" (UID: \"1b2f62d4-0584-4cc3-81ed-e067d8db23b1\") " pod="openstack/barbican-api-7fc86597fb-mt9pk" Mar 13 14:27:35 crc kubenswrapper[4907]: I0313 14:27:35.908125 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1b2f62d4-0584-4cc3-81ed-e067d8db23b1-logs\") pod \"barbican-api-7fc86597fb-mt9pk\" (UID: \"1b2f62d4-0584-4cc3-81ed-e067d8db23b1\") " pod="openstack/barbican-api-7fc86597fb-mt9pk" Mar 13 14:27:35 crc kubenswrapper[4907]: I0313 14:27:35.912596 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b2f62d4-0584-4cc3-81ed-e067d8db23b1-config-data\") pod \"barbican-api-7fc86597fb-mt9pk\" (UID: \"1b2f62d4-0584-4cc3-81ed-e067d8db23b1\") " pod="openstack/barbican-api-7fc86597fb-mt9pk" Mar 13 14:27:35 crc kubenswrapper[4907]: I0313 14:27:35.913050 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b2f62d4-0584-4cc3-81ed-e067d8db23b1-combined-ca-bundle\") pod \"barbican-api-7fc86597fb-mt9pk\" (UID: \"1b2f62d4-0584-4cc3-81ed-e067d8db23b1\") " pod="openstack/barbican-api-7fc86597fb-mt9pk" Mar 13 14:27:35 crc kubenswrapper[4907]: I0313 14:27:35.913331 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1b2f62d4-0584-4cc3-81ed-e067d8db23b1-internal-tls-certs\") pod \"barbican-api-7fc86597fb-mt9pk\" (UID: \"1b2f62d4-0584-4cc3-81ed-e067d8db23b1\") " pod="openstack/barbican-api-7fc86597fb-mt9pk" Mar 13 14:27:35 crc kubenswrapper[4907]: I0313 14:27:35.915197 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1b2f62d4-0584-4cc3-81ed-e067d8db23b1-public-tls-certs\") pod \"barbican-api-7fc86597fb-mt9pk\" (UID: \"1b2f62d4-0584-4cc3-81ed-e067d8db23b1\") " pod="openstack/barbican-api-7fc86597fb-mt9pk" Mar 13 14:27:35 crc kubenswrapper[4907]: I0313 14:27:35.934621 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ktf7c\" (UniqueName: \"kubernetes.io/projected/1b2f62d4-0584-4cc3-81ed-e067d8db23b1-kube-api-access-ktf7c\") pod \"barbican-api-7fc86597fb-mt9pk\" (UID: \"1b2f62d4-0584-4cc3-81ed-e067d8db23b1\") " pod="openstack/barbican-api-7fc86597fb-mt9pk" Mar 13 14:27:35 crc kubenswrapper[4907]: I0313 14:27:35.936537 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1b2f62d4-0584-4cc3-81ed-e067d8db23b1-config-data-custom\") pod \"barbican-api-7fc86597fb-mt9pk\" (UID: \"1b2f62d4-0584-4cc3-81ed-e067d8db23b1\") " pod="openstack/barbican-api-7fc86597fb-mt9pk" Mar 13 14:27:36 crc kubenswrapper[4907]: I0313 14:27:36.160599 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7fc86597fb-mt9pk" Mar 13 14:27:36 crc kubenswrapper[4907]: I0313 14:27:36.724040 4907 generic.go:334] "Generic (PLEG): container finished" podID="070e2b75-05b1-42a4-a34e-950d45d65920" containerID="57e7ca150948c3586cf56019f626e80fe0c0e49f021cce99a8bbf1bcfa081624" exitCode=143 Mar 13 14:27:36 crc kubenswrapper[4907]: I0313 14:27:36.724418 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5d49fd688d-njvfm" event={"ID":"070e2b75-05b1-42a4-a34e-950d45d65920","Type":"ContainerDied","Data":"57e7ca150948c3586cf56019f626e80fe0c0e49f021cce99a8bbf1bcfa081624"} Mar 13 14:27:36 crc kubenswrapper[4907]: I0313 14:27:36.737807 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-keystone-listener-54c69f4bdb-lt67r" podUID="fcbb1687-db0b-4fc6-937c-89102831237b" containerName="barbican-keystone-listener-log" containerID="cri-o://6ad9081a27708afd67b77cedc2aa19d22a6afb40f65f24b3a9da7ca3a519bf38" gracePeriod=30 Mar 13 14:27:36 crc kubenswrapper[4907]: I0313 14:27:36.738941 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-worker-c685c5475-rvkf2" podUID="ca4928ba-1277-4918-8b92-3eda2b276577" containerName="barbican-worker-log" containerID="cri-o://bb7f9aa2bb828fff3a17bc00a23c15ddb0d4b9b7ae4d6e61a19fc6386d089dbb" gracePeriod=30 Mar 13 14:27:36 crc kubenswrapper[4907]: I0313 14:27:36.739072 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7bf587b8d9-pvvbz" event={"ID":"001078cc-c01e-4173-9740-4f9ff082c593","Type":"ContainerStarted","Data":"3e3eb6f5fca9c3693edb3599f1d3dd3d706131518c37c580379168a8a1b7e5e0"} Mar 13 14:27:36 crc kubenswrapper[4907]: I0313 14:27:36.739459 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-keystone-listener-54c69f4bdb-lt67r" podUID="fcbb1687-db0b-4fc6-937c-89102831237b" containerName="barbican-keystone-listener" containerID="cri-o://9fe21d916b51d235fb7ba28d7a6934f6f8ffaa95ccce23ab15efee2b6e69acd5" gracePeriod=30 Mar 13 14:27:36 crc kubenswrapper[4907]: I0313 14:27:36.739540 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-worker-c685c5475-rvkf2" podUID="ca4928ba-1277-4918-8b92-3eda2b276577" containerName="barbican-worker" containerID="cri-o://2ef6cd9eb88308a93f1d78fb9954045701e6f121a3c20317549b8db40b280bd4" gracePeriod=30 Mar 13 14:27:36 crc kubenswrapper[4907]: I0313 14:27:36.777145 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-7bf587b8d9-pvvbz" podStartSLOduration=3.777128897 podStartE2EDuration="3.777128897s" podCreationTimestamp="2026-03-13 14:27:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:27:36.775219385 +0000 UTC m=+1355.675007074" watchObservedRunningTime="2026-03-13 14:27:36.777128897 +0000 UTC m=+1355.676916576" Mar 13 14:27:36 crc kubenswrapper[4907]: I0313 14:27:36.815202 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-7fc86597fb-mt9pk"] Mar 13 14:27:37 crc kubenswrapper[4907]: I0313 14:27:37.752502 4907 generic.go:334] "Generic (PLEG): container finished" podID="fcbb1687-db0b-4fc6-937c-89102831237b" containerID="9fe21d916b51d235fb7ba28d7a6934f6f8ffaa95ccce23ab15efee2b6e69acd5" exitCode=0 Mar 13 14:27:37 crc kubenswrapper[4907]: I0313 14:27:37.752770 4907 generic.go:334] "Generic (PLEG): container finished" podID="fcbb1687-db0b-4fc6-937c-89102831237b" containerID="6ad9081a27708afd67b77cedc2aa19d22a6afb40f65f24b3a9da7ca3a519bf38" exitCode=143 Mar 13 14:27:37 crc kubenswrapper[4907]: I0313 14:27:37.752592 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-54c69f4bdb-lt67r" event={"ID":"fcbb1687-db0b-4fc6-937c-89102831237b","Type":"ContainerDied","Data":"9fe21d916b51d235fb7ba28d7a6934f6f8ffaa95ccce23ab15efee2b6e69acd5"} Mar 13 14:27:37 crc kubenswrapper[4907]: I0313 14:27:37.752818 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-54c69f4bdb-lt67r" event={"ID":"fcbb1687-db0b-4fc6-937c-89102831237b","Type":"ContainerDied","Data":"6ad9081a27708afd67b77cedc2aa19d22a6afb40f65f24b3a9da7ca3a519bf38"} Mar 13 14:27:37 crc kubenswrapper[4907]: I0313 14:27:37.754405 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7fc86597fb-mt9pk" event={"ID":"1b2f62d4-0584-4cc3-81ed-e067d8db23b1","Type":"ContainerStarted","Data":"97dce4dda51605af3970178d1030468b7ba5970d94173d175f92a65d74507fff"} Mar 13 14:27:37 crc kubenswrapper[4907]: I0313 14:27:37.761159 4907 generic.go:334] "Generic (PLEG): container finished" podID="ca4928ba-1277-4918-8b92-3eda2b276577" containerID="2ef6cd9eb88308a93f1d78fb9954045701e6f121a3c20317549b8db40b280bd4" exitCode=0 Mar 13 14:27:37 crc kubenswrapper[4907]: I0313 14:27:37.761193 4907 generic.go:334] "Generic (PLEG): container finished" podID="ca4928ba-1277-4918-8b92-3eda2b276577" containerID="bb7f9aa2bb828fff3a17bc00a23c15ddb0d4b9b7ae4d6e61a19fc6386d089dbb" exitCode=143 Mar 13 14:27:37 crc kubenswrapper[4907]: I0313 14:27:37.761237 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-c685c5475-rvkf2" event={"ID":"ca4928ba-1277-4918-8b92-3eda2b276577","Type":"ContainerDied","Data":"2ef6cd9eb88308a93f1d78fb9954045701e6f121a3c20317549b8db40b280bd4"} Mar 13 14:27:37 crc kubenswrapper[4907]: I0313 14:27:37.761293 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-c685c5475-rvkf2" event={"ID":"ca4928ba-1277-4918-8b92-3eda2b276577","Type":"ContainerDied","Data":"bb7f9aa2bb828fff3a17bc00a23c15ddb0d4b9b7ae4d6e61a19fc6386d089dbb"} Mar 13 14:27:37 crc kubenswrapper[4907]: I0313 14:27:37.769053 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-7q999" event={"ID":"594b61fa-12b7-47b6-8af3-17f024e0d54d","Type":"ContainerStarted","Data":"5feda018c451c4b446f95e35e5721969c1c913a7887bdf916e5ea2193df476ec"} Mar 13 14:27:37 crc kubenswrapper[4907]: I0313 14:27:37.769325 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-7bf587b8d9-pvvbz" Mar 13 14:27:37 crc kubenswrapper[4907]: I0313 14:27:37.805415 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-7q999" podStartSLOduration=3.103559929 podStartE2EDuration="40.805393152s" podCreationTimestamp="2026-03-13 14:26:57 +0000 UTC" firstStartedPulling="2026-03-13 14:26:58.763220846 +0000 UTC m=+1317.663008535" lastFinishedPulling="2026-03-13 14:27:36.465054059 +0000 UTC m=+1355.364841758" observedRunningTime="2026-03-13 14:27:37.793297374 +0000 UTC m=+1356.693085063" watchObservedRunningTime="2026-03-13 14:27:37.805393152 +0000 UTC m=+1356.705180841" Mar 13 14:27:39 crc kubenswrapper[4907]: I0313 14:27:39.873057 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7fc46d7df7-77dng" Mar 13 14:27:39 crc kubenswrapper[4907]: I0313 14:27:39.959121 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-759cc7f497-7nv78"] Mar 13 14:27:39 crc kubenswrapper[4907]: I0313 14:27:39.959680 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-759cc7f497-7nv78" podUID="a661b9e1-c551-4b36-87a9-833df73bd6b2" containerName="dnsmasq-dns" containerID="cri-o://46c2ff1e9e66c6fc449295a888dc400f857863d7df33a386b4b1e1cf42f6a3f4" gracePeriod=10 Mar 13 14:27:40 crc kubenswrapper[4907]: I0313 14:27:40.093693 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-74984bd8b4-x42pq" Mar 13 14:27:40 crc kubenswrapper[4907]: I0313 14:27:40.258635 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-74984bd8b4-x42pq" Mar 13 14:27:40 crc kubenswrapper[4907]: I0313 14:27:40.817544 4907 generic.go:334] "Generic (PLEG): container finished" podID="a661b9e1-c551-4b36-87a9-833df73bd6b2" containerID="46c2ff1e9e66c6fc449295a888dc400f857863d7df33a386b4b1e1cf42f6a3f4" exitCode=0 Mar 13 14:27:40 crc kubenswrapper[4907]: I0313 14:27:40.817608 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-759cc7f497-7nv78" event={"ID":"a661b9e1-c551-4b36-87a9-833df73bd6b2","Type":"ContainerDied","Data":"46c2ff1e9e66c6fc449295a888dc400f857863d7df33a386b4b1e1cf42f6a3f4"} Mar 13 14:27:42 crc kubenswrapper[4907]: I0313 14:27:42.153109 4907 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5d49fd688d-njvfm" podUID="070e2b75-05b1-42a4-a34e-950d45d65920" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.160:9311/healthcheck\": read tcp 10.217.0.2:42076->10.217.0.160:9311: read: connection reset by peer" Mar 13 14:27:42 crc kubenswrapper[4907]: I0313 14:27:42.153143 4907 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5d49fd688d-njvfm" podUID="070e2b75-05b1-42a4-a34e-950d45d65920" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.160:9311/healthcheck\": read tcp 10.217.0.2:42066->10.217.0.160:9311: read: connection reset by peer" Mar 13 14:27:42 crc kubenswrapper[4907]: I0313 14:27:42.755706 4907 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-759cc7f497-7nv78" podUID="a661b9e1-c551-4b36-87a9-833df73bd6b2" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.149:5353: connect: connection refused" Mar 13 14:27:42 crc kubenswrapper[4907]: I0313 14:27:42.838244 4907 generic.go:334] "Generic (PLEG): container finished" podID="070e2b75-05b1-42a4-a34e-950d45d65920" containerID="8ce806c1853f002c1ca5aee8d1ea69c43b2c38c4c666d7df49a773673406c573" exitCode=0 Mar 13 14:27:42 crc kubenswrapper[4907]: I0313 14:27:42.838289 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5d49fd688d-njvfm" event={"ID":"070e2b75-05b1-42a4-a34e-950d45d65920","Type":"ContainerDied","Data":"8ce806c1853f002c1ca5aee8d1ea69c43b2c38c4c666d7df49a773673406c573"} Mar 13 14:27:43 crc kubenswrapper[4907]: I0313 14:27:43.044922 4907 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5d49fd688d-njvfm" podUID="070e2b75-05b1-42a4-a34e-950d45d65920" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.160:9311/healthcheck\": dial tcp 10.217.0.160:9311: connect: connection refused" Mar 13 14:27:43 crc kubenswrapper[4907]: I0313 14:27:43.045148 4907 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5d49fd688d-njvfm" podUID="070e2b75-05b1-42a4-a34e-950d45d65920" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.160:9311/healthcheck\": dial tcp 10.217.0.160:9311: connect: connection refused" Mar 13 14:27:43 crc kubenswrapper[4907]: I0313 14:27:43.867447 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-54c69f4bdb-lt67r" event={"ID":"fcbb1687-db0b-4fc6-937c-89102831237b","Type":"ContainerDied","Data":"7a66d01a94d06b4565675b1fcb3708a34133ebf672d54a557137e862ae4fd1c3"} Mar 13 14:27:43 crc kubenswrapper[4907]: I0313 14:27:43.867749 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7a66d01a94d06b4565675b1fcb3708a34133ebf672d54a557137e862ae4fd1c3" Mar 13 14:27:43 crc kubenswrapper[4907]: I0313 14:27:43.874430 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7fc86597fb-mt9pk" event={"ID":"1b2f62d4-0584-4cc3-81ed-e067d8db23b1","Type":"ContainerStarted","Data":"f21714a9f3b46d6b404fb8600bc2b71af7d623654fb77b4f61140d6330569927"} Mar 13 14:27:43 crc kubenswrapper[4907]: I0313 14:27:43.881626 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-c685c5475-rvkf2" event={"ID":"ca4928ba-1277-4918-8b92-3eda2b276577","Type":"ContainerDied","Data":"6c4a87798f44345ddbfe9a9bbadc067d17cd8c04974906e2bc3559decad063c2"} Mar 13 14:27:43 crc kubenswrapper[4907]: I0313 14:27:43.881708 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6c4a87798f44345ddbfe9a9bbadc067d17cd8c04974906e2bc3559decad063c2" Mar 13 14:27:43 crc kubenswrapper[4907]: I0313 14:27:43.884978 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5d49fd688d-njvfm" event={"ID":"070e2b75-05b1-42a4-a34e-950d45d65920","Type":"ContainerDied","Data":"eaad4530bf1fa0795ee15397e0a9c4c262de8913b337803ac449643d1177e1b7"} Mar 13 14:27:43 crc kubenswrapper[4907]: I0313 14:27:43.885023 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eaad4530bf1fa0795ee15397e0a9c4c262de8913b337803ac449643d1177e1b7" Mar 13 14:27:43 crc kubenswrapper[4907]: I0313 14:27:43.887282 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-759cc7f497-7nv78" event={"ID":"a661b9e1-c551-4b36-87a9-833df73bd6b2","Type":"ContainerDied","Data":"eaee566543e3557ec83d2d26f27876f1ccb4095ddf1c90e6680a5df6c9c175d0"} Mar 13 14:27:43 crc kubenswrapper[4907]: I0313 14:27:43.887319 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eaee566543e3557ec83d2d26f27876f1ccb4095ddf1c90e6680a5df6c9c175d0" Mar 13 14:27:43 crc kubenswrapper[4907]: I0313 14:27:43.953572 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-759cc7f497-7nv78" Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.073121 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a661b9e1-c551-4b36-87a9-833df73bd6b2-ovsdbserver-sb\") pod \"a661b9e1-c551-4b36-87a9-833df73bd6b2\" (UID: \"a661b9e1-c551-4b36-87a9-833df73bd6b2\") " Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.073622 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ntd2v\" (UniqueName: \"kubernetes.io/projected/a661b9e1-c551-4b36-87a9-833df73bd6b2-kube-api-access-ntd2v\") pod \"a661b9e1-c551-4b36-87a9-833df73bd6b2\" (UID: \"a661b9e1-c551-4b36-87a9-833df73bd6b2\") " Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.074244 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a661b9e1-c551-4b36-87a9-833df73bd6b2-dns-swift-storage-0\") pod \"a661b9e1-c551-4b36-87a9-833df73bd6b2\" (UID: \"a661b9e1-c551-4b36-87a9-833df73bd6b2\") " Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.074287 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a661b9e1-c551-4b36-87a9-833df73bd6b2-dns-svc\") pod \"a661b9e1-c551-4b36-87a9-833df73bd6b2\" (UID: \"a661b9e1-c551-4b36-87a9-833df73bd6b2\") " Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.074413 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a661b9e1-c551-4b36-87a9-833df73bd6b2-config\") pod \"a661b9e1-c551-4b36-87a9-833df73bd6b2\" (UID: \"a661b9e1-c551-4b36-87a9-833df73bd6b2\") " Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.074527 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a661b9e1-c551-4b36-87a9-833df73bd6b2-ovsdbserver-nb\") pod \"a661b9e1-c551-4b36-87a9-833df73bd6b2\" (UID: \"a661b9e1-c551-4b36-87a9-833df73bd6b2\") " Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.083194 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a661b9e1-c551-4b36-87a9-833df73bd6b2-kube-api-access-ntd2v" (OuterVolumeSpecName: "kube-api-access-ntd2v") pod "a661b9e1-c551-4b36-87a9-833df73bd6b2" (UID: "a661b9e1-c551-4b36-87a9-833df73bd6b2"). InnerVolumeSpecName "kube-api-access-ntd2v". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.098833 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5d49fd688d-njvfm" Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.132787 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a661b9e1-c551-4b36-87a9-833df73bd6b2-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a661b9e1-c551-4b36-87a9-833df73bd6b2" (UID: "a661b9e1-c551-4b36-87a9-833df73bd6b2"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.139396 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a661b9e1-c551-4b36-87a9-833df73bd6b2-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "a661b9e1-c551-4b36-87a9-833df73bd6b2" (UID: "a661b9e1-c551-4b36-87a9-833df73bd6b2"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.139720 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a661b9e1-c551-4b36-87a9-833df73bd6b2-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "a661b9e1-c551-4b36-87a9-833df73bd6b2" (UID: "a661b9e1-c551-4b36-87a9-833df73bd6b2"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.141317 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a661b9e1-c551-4b36-87a9-833df73bd6b2-config" (OuterVolumeSpecName: "config") pod "a661b9e1-c551-4b36-87a9-833df73bd6b2" (UID: "a661b9e1-c551-4b36-87a9-833df73bd6b2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.158668 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a661b9e1-c551-4b36-87a9-833df73bd6b2-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "a661b9e1-c551-4b36-87a9-833df73bd6b2" (UID: "a661b9e1-c551-4b36-87a9-833df73bd6b2"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.177203 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/070e2b75-05b1-42a4-a34e-950d45d65920-combined-ca-bundle\") pod \"070e2b75-05b1-42a4-a34e-950d45d65920\" (UID: \"070e2b75-05b1-42a4-a34e-950d45d65920\") " Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.177409 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/070e2b75-05b1-42a4-a34e-950d45d65920-config-data\") pod \"070e2b75-05b1-42a4-a34e-950d45d65920\" (UID: \"070e2b75-05b1-42a4-a34e-950d45d65920\") " Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.177468 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/070e2b75-05b1-42a4-a34e-950d45d65920-logs\") pod \"070e2b75-05b1-42a4-a34e-950d45d65920\" (UID: \"070e2b75-05b1-42a4-a34e-950d45d65920\") " Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.177525 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bfwxk\" (UniqueName: \"kubernetes.io/projected/070e2b75-05b1-42a4-a34e-950d45d65920-kube-api-access-bfwxk\") pod \"070e2b75-05b1-42a4-a34e-950d45d65920\" (UID: \"070e2b75-05b1-42a4-a34e-950d45d65920\") " Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.177552 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/070e2b75-05b1-42a4-a34e-950d45d65920-config-data-custom\") pod \"070e2b75-05b1-42a4-a34e-950d45d65920\" (UID: \"070e2b75-05b1-42a4-a34e-950d45d65920\") " Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.177994 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a661b9e1-c551-4b36-87a9-833df73bd6b2-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.178014 4907 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a661b9e1-c551-4b36-87a9-833df73bd6b2-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.178025 4907 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a661b9e1-c551-4b36-87a9-833df73bd6b2-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.178035 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ntd2v\" (UniqueName: \"kubernetes.io/projected/a661b9e1-c551-4b36-87a9-833df73bd6b2-kube-api-access-ntd2v\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.178045 4907 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a661b9e1-c551-4b36-87a9-833df73bd6b2-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.178054 4907 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a661b9e1-c551-4b36-87a9-833df73bd6b2-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.178099 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/070e2b75-05b1-42a4-a34e-950d45d65920-logs" (OuterVolumeSpecName: "logs") pod "070e2b75-05b1-42a4-a34e-950d45d65920" (UID: "070e2b75-05b1-42a4-a34e-950d45d65920"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.180311 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/070e2b75-05b1-42a4-a34e-950d45d65920-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "070e2b75-05b1-42a4-a34e-950d45d65920" (UID: "070e2b75-05b1-42a4-a34e-950d45d65920"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.185068 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/070e2b75-05b1-42a4-a34e-950d45d65920-kube-api-access-bfwxk" (OuterVolumeSpecName: "kube-api-access-bfwxk") pod "070e2b75-05b1-42a4-a34e-950d45d65920" (UID: "070e2b75-05b1-42a4-a34e-950d45d65920"). InnerVolumeSpecName "kube-api-access-bfwxk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.208742 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/070e2b75-05b1-42a4-a34e-950d45d65920-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "070e2b75-05b1-42a4-a34e-950d45d65920" (UID: "070e2b75-05b1-42a4-a34e-950d45d65920"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.222487 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-54c69f4bdb-lt67r" Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.241986 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/070e2b75-05b1-42a4-a34e-950d45d65920-config-data" (OuterVolumeSpecName: "config-data") pod "070e2b75-05b1-42a4-a34e-950d45d65920" (UID: "070e2b75-05b1-42a4-a34e-950d45d65920"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.247211 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-c685c5475-rvkf2" Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.279254 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/070e2b75-05b1-42a4-a34e-950d45d65920-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.279288 4907 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/070e2b75-05b1-42a4-a34e-950d45d65920-logs\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.279299 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bfwxk\" (UniqueName: \"kubernetes.io/projected/070e2b75-05b1-42a4-a34e-950d45d65920-kube-api-access-bfwxk\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.279308 4907 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/070e2b75-05b1-42a4-a34e-950d45d65920-config-data-custom\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.279318 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/070e2b75-05b1-42a4-a34e-950d45d65920-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.380931 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qttjx\" (UniqueName: \"kubernetes.io/projected/fcbb1687-db0b-4fc6-937c-89102831237b-kube-api-access-qttjx\") pod \"fcbb1687-db0b-4fc6-937c-89102831237b\" (UID: \"fcbb1687-db0b-4fc6-937c-89102831237b\") " Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.381245 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fcbb1687-db0b-4fc6-937c-89102831237b-combined-ca-bundle\") pod \"fcbb1687-db0b-4fc6-937c-89102831237b\" (UID: \"fcbb1687-db0b-4fc6-937c-89102831237b\") " Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.381363 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ca4928ba-1277-4918-8b92-3eda2b276577-config-data-custom\") pod \"ca4928ba-1277-4918-8b92-3eda2b276577\" (UID: \"ca4928ba-1277-4918-8b92-3eda2b276577\") " Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.381507 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca4928ba-1277-4918-8b92-3eda2b276577-config-data\") pod \"ca4928ba-1277-4918-8b92-3eda2b276577\" (UID: \"ca4928ba-1277-4918-8b92-3eda2b276577\") " Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.381595 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ca4928ba-1277-4918-8b92-3eda2b276577-logs\") pod \"ca4928ba-1277-4918-8b92-3eda2b276577\" (UID: \"ca4928ba-1277-4918-8b92-3eda2b276577\") " Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.381770 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fcbb1687-db0b-4fc6-937c-89102831237b-config-data-custom\") pod \"fcbb1687-db0b-4fc6-937c-89102831237b\" (UID: \"fcbb1687-db0b-4fc6-937c-89102831237b\") " Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.381911 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ca4928ba-1277-4918-8b92-3eda2b276577-logs" (OuterVolumeSpecName: "logs") pod "ca4928ba-1277-4918-8b92-3eda2b276577" (UID: "ca4928ba-1277-4918-8b92-3eda2b276577"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.381918 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca4928ba-1277-4918-8b92-3eda2b276577-combined-ca-bundle\") pod \"ca4928ba-1277-4918-8b92-3eda2b276577\" (UID: \"ca4928ba-1277-4918-8b92-3eda2b276577\") " Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.381987 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fcbb1687-db0b-4fc6-937c-89102831237b-logs\") pod \"fcbb1687-db0b-4fc6-937c-89102831237b\" (UID: \"fcbb1687-db0b-4fc6-937c-89102831237b\") " Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.382022 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fcbb1687-db0b-4fc6-937c-89102831237b-config-data\") pod \"fcbb1687-db0b-4fc6-937c-89102831237b\" (UID: \"fcbb1687-db0b-4fc6-937c-89102831237b\") " Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.382094 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9vftx\" (UniqueName: \"kubernetes.io/projected/ca4928ba-1277-4918-8b92-3eda2b276577-kube-api-access-9vftx\") pod \"ca4928ba-1277-4918-8b92-3eda2b276577\" (UID: \"ca4928ba-1277-4918-8b92-3eda2b276577\") " Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.382319 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fcbb1687-db0b-4fc6-937c-89102831237b-logs" (OuterVolumeSpecName: "logs") pod "fcbb1687-db0b-4fc6-937c-89102831237b" (UID: "fcbb1687-db0b-4fc6-937c-89102831237b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.382580 4907 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fcbb1687-db0b-4fc6-937c-89102831237b-logs\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.382605 4907 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ca4928ba-1277-4918-8b92-3eda2b276577-logs\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.386405 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca4928ba-1277-4918-8b92-3eda2b276577-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "ca4928ba-1277-4918-8b92-3eda2b276577" (UID: "ca4928ba-1277-4918-8b92-3eda2b276577"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.386427 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fcbb1687-db0b-4fc6-937c-89102831237b-kube-api-access-qttjx" (OuterVolumeSpecName: "kube-api-access-qttjx") pod "fcbb1687-db0b-4fc6-937c-89102831237b" (UID: "fcbb1687-db0b-4fc6-937c-89102831237b"). InnerVolumeSpecName "kube-api-access-qttjx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.388664 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fcbb1687-db0b-4fc6-937c-89102831237b-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "fcbb1687-db0b-4fc6-937c-89102831237b" (UID: "fcbb1687-db0b-4fc6-937c-89102831237b"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.389112 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ca4928ba-1277-4918-8b92-3eda2b276577-kube-api-access-9vftx" (OuterVolumeSpecName: "kube-api-access-9vftx") pod "ca4928ba-1277-4918-8b92-3eda2b276577" (UID: "ca4928ba-1277-4918-8b92-3eda2b276577"). InnerVolumeSpecName "kube-api-access-9vftx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.423416 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fcbb1687-db0b-4fc6-937c-89102831237b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fcbb1687-db0b-4fc6-937c-89102831237b" (UID: "fcbb1687-db0b-4fc6-937c-89102831237b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.430958 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca4928ba-1277-4918-8b92-3eda2b276577-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ca4928ba-1277-4918-8b92-3eda2b276577" (UID: "ca4928ba-1277-4918-8b92-3eda2b276577"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.455682 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fcbb1687-db0b-4fc6-937c-89102831237b-config-data" (OuterVolumeSpecName: "config-data") pod "fcbb1687-db0b-4fc6-937c-89102831237b" (UID: "fcbb1687-db0b-4fc6-937c-89102831237b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.459901 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca4928ba-1277-4918-8b92-3eda2b276577-config-data" (OuterVolumeSpecName: "config-data") pod "ca4928ba-1277-4918-8b92-3eda2b276577" (UID: "ca4928ba-1277-4918-8b92-3eda2b276577"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.484526 4907 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fcbb1687-db0b-4fc6-937c-89102831237b-config-data-custom\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.484582 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca4928ba-1277-4918-8b92-3eda2b276577-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.484593 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fcbb1687-db0b-4fc6-937c-89102831237b-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.484603 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9vftx\" (UniqueName: \"kubernetes.io/projected/ca4928ba-1277-4918-8b92-3eda2b276577-kube-api-access-9vftx\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.484617 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qttjx\" (UniqueName: \"kubernetes.io/projected/fcbb1687-db0b-4fc6-937c-89102831237b-kube-api-access-qttjx\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.484650 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fcbb1687-db0b-4fc6-937c-89102831237b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.484663 4907 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ca4928ba-1277-4918-8b92-3eda2b276577-config-data-custom\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.484674 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca4928ba-1277-4918-8b92-3eda2b276577-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.897282 4907 generic.go:334] "Generic (PLEG): container finished" podID="594b61fa-12b7-47b6-8af3-17f024e0d54d" containerID="5feda018c451c4b446f95e35e5721969c1c913a7887bdf916e5ea2193df476ec" exitCode=0 Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.897369 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-7q999" event={"ID":"594b61fa-12b7-47b6-8af3-17f024e0d54d","Type":"ContainerDied","Data":"5feda018c451c4b446f95e35e5721969c1c913a7887bdf916e5ea2193df476ec"} Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.899765 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7fc86597fb-mt9pk" event={"ID":"1b2f62d4-0584-4cc3-81ed-e067d8db23b1","Type":"ContainerStarted","Data":"22aa297a2f364f3884c53d61c34c94086546e7771e899b38ffebaf99570e4f07"} Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.899909 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-7fc86597fb-mt9pk" Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.900014 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-7fc86597fb-mt9pk" Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.902489 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-54c69f4bdb-lt67r" Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.902512 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-759cc7f497-7nv78" Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.902522 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-c685c5475-rvkf2" Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.902547 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5","Type":"ContainerStarted","Data":"dd889dbfbaa8a48ab7c7b9c0cf914f3824e4417de226256cac660ed3cfc140b1"} Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.902779 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5d49fd688d-njvfm" Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.902816 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5" containerName="ceilometer-central-agent" containerID="cri-o://b9aab9a1290561a45c480e877db9222806553cdd6cfb12daf299ab965ad2ec5c" gracePeriod=30 Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.902935 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5" containerName="ceilometer-notification-agent" containerID="cri-o://af56a2f9ef345a9cf4a63dafe33dfb02aa9518339a5d2ef4a11fbf9ec8ab7512" gracePeriod=30 Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.902933 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5" containerName="proxy-httpd" containerID="cri-o://dd889dbfbaa8a48ab7c7b9c0cf914f3824e4417de226256cac660ed3cfc140b1" gracePeriod=30 Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.902909 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5" containerName="sg-core" containerID="cri-o://a15bc8c355c09d50b27215622b3ea5aaebf1df184e4f387e2405fe44a886e49b" gracePeriod=30 Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.902963 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.960752 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-7fc86597fb-mt9pk" podStartSLOduration=9.96070637 podStartE2EDuration="9.96070637s" podCreationTimestamp="2026-03-13 14:27:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:27:44.950039259 +0000 UTC m=+1363.849826948" watchObservedRunningTime="2026-03-13 14:27:44.96070637 +0000 UTC m=+1363.860494069" Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.981583 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.417911262 podStartE2EDuration="48.981562307s" podCreationTimestamp="2026-03-13 14:26:56 +0000 UTC" firstStartedPulling="2026-03-13 14:26:58.370075573 +0000 UTC m=+1317.269863262" lastFinishedPulling="2026-03-13 14:27:43.933726628 +0000 UTC m=+1362.833514307" observedRunningTime="2026-03-13 14:27:44.972362516 +0000 UTC m=+1363.872150225" watchObservedRunningTime="2026-03-13 14:27:44.981562307 +0000 UTC m=+1363.881349996" Mar 13 14:27:45 crc kubenswrapper[4907]: I0313 14:27:45.009220 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-worker-c685c5475-rvkf2"] Mar 13 14:27:45 crc kubenswrapper[4907]: I0313 14:27:45.021726 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-worker-c685c5475-rvkf2"] Mar 13 14:27:45 crc kubenswrapper[4907]: I0313 14:27:45.039116 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-keystone-listener-54c69f4bdb-lt67r"] Mar 13 14:27:45 crc kubenswrapper[4907]: I0313 14:27:45.050389 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-keystone-listener-54c69f4bdb-lt67r"] Mar 13 14:27:45 crc kubenswrapper[4907]: I0313 14:27:45.059821 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-759cc7f497-7nv78"] Mar 13 14:27:45 crc kubenswrapper[4907]: I0313 14:27:45.068098 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-759cc7f497-7nv78"] Mar 13 14:27:45 crc kubenswrapper[4907]: I0313 14:27:45.077333 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-5d49fd688d-njvfm"] Mar 13 14:27:45 crc kubenswrapper[4907]: I0313 14:27:45.086358 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-5d49fd688d-njvfm"] Mar 13 14:27:45 crc kubenswrapper[4907]: I0313 14:27:45.806716 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="070e2b75-05b1-42a4-a34e-950d45d65920" path="/var/lib/kubelet/pods/070e2b75-05b1-42a4-a34e-950d45d65920/volumes" Mar 13 14:27:45 crc kubenswrapper[4907]: I0313 14:27:45.807330 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a661b9e1-c551-4b36-87a9-833df73bd6b2" path="/var/lib/kubelet/pods/a661b9e1-c551-4b36-87a9-833df73bd6b2/volumes" Mar 13 14:27:45 crc kubenswrapper[4907]: I0313 14:27:45.807945 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ca4928ba-1277-4918-8b92-3eda2b276577" path="/var/lib/kubelet/pods/ca4928ba-1277-4918-8b92-3eda2b276577/volumes" Mar 13 14:27:45 crc kubenswrapper[4907]: I0313 14:27:45.808988 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fcbb1687-db0b-4fc6-937c-89102831237b" path="/var/lib/kubelet/pods/fcbb1687-db0b-4fc6-937c-89102831237b/volumes" Mar 13 14:27:45 crc kubenswrapper[4907]: I0313 14:27:45.914169 4907 generic.go:334] "Generic (PLEG): container finished" podID="58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5" containerID="dd889dbfbaa8a48ab7c7b9c0cf914f3824e4417de226256cac660ed3cfc140b1" exitCode=0 Mar 13 14:27:45 crc kubenswrapper[4907]: I0313 14:27:45.914223 4907 generic.go:334] "Generic (PLEG): container finished" podID="58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5" containerID="a15bc8c355c09d50b27215622b3ea5aaebf1df184e4f387e2405fe44a886e49b" exitCode=2 Mar 13 14:27:45 crc kubenswrapper[4907]: I0313 14:27:45.914235 4907 generic.go:334] "Generic (PLEG): container finished" podID="58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5" containerID="b9aab9a1290561a45c480e877db9222806553cdd6cfb12daf299ab965ad2ec5c" exitCode=0 Mar 13 14:27:45 crc kubenswrapper[4907]: I0313 14:27:45.914258 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5","Type":"ContainerDied","Data":"dd889dbfbaa8a48ab7c7b9c0cf914f3824e4417de226256cac660ed3cfc140b1"} Mar 13 14:27:45 crc kubenswrapper[4907]: I0313 14:27:45.914320 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5","Type":"ContainerDied","Data":"a15bc8c355c09d50b27215622b3ea5aaebf1df184e4f387e2405fe44a886e49b"} Mar 13 14:27:45 crc kubenswrapper[4907]: I0313 14:27:45.914337 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5","Type":"ContainerDied","Data":"b9aab9a1290561a45c480e877db9222806553cdd6cfb12daf299ab965ad2ec5c"} Mar 13 14:27:46 crc kubenswrapper[4907]: I0313 14:27:46.254287 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-7q999" Mar 13 14:27:46 crc kubenswrapper[4907]: I0313 14:27:46.418613 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/594b61fa-12b7-47b6-8af3-17f024e0d54d-combined-ca-bundle\") pod \"594b61fa-12b7-47b6-8af3-17f024e0d54d\" (UID: \"594b61fa-12b7-47b6-8af3-17f024e0d54d\") " Mar 13 14:27:46 crc kubenswrapper[4907]: I0313 14:27:46.418667 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/594b61fa-12b7-47b6-8af3-17f024e0d54d-db-sync-config-data\") pod \"594b61fa-12b7-47b6-8af3-17f024e0d54d\" (UID: \"594b61fa-12b7-47b6-8af3-17f024e0d54d\") " Mar 13 14:27:46 crc kubenswrapper[4907]: I0313 14:27:46.418925 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/594b61fa-12b7-47b6-8af3-17f024e0d54d-etc-machine-id\") pod \"594b61fa-12b7-47b6-8af3-17f024e0d54d\" (UID: \"594b61fa-12b7-47b6-8af3-17f024e0d54d\") " Mar 13 14:27:46 crc kubenswrapper[4907]: I0313 14:27:46.419039 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/594b61fa-12b7-47b6-8af3-17f024e0d54d-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "594b61fa-12b7-47b6-8af3-17f024e0d54d" (UID: "594b61fa-12b7-47b6-8af3-17f024e0d54d"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 14:27:46 crc kubenswrapper[4907]: I0313 14:27:46.419120 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/594b61fa-12b7-47b6-8af3-17f024e0d54d-scripts\") pod \"594b61fa-12b7-47b6-8af3-17f024e0d54d\" (UID: \"594b61fa-12b7-47b6-8af3-17f024e0d54d\") " Mar 13 14:27:46 crc kubenswrapper[4907]: I0313 14:27:46.419191 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wpft2\" (UniqueName: \"kubernetes.io/projected/594b61fa-12b7-47b6-8af3-17f024e0d54d-kube-api-access-wpft2\") pod \"594b61fa-12b7-47b6-8af3-17f024e0d54d\" (UID: \"594b61fa-12b7-47b6-8af3-17f024e0d54d\") " Mar 13 14:27:46 crc kubenswrapper[4907]: I0313 14:27:46.419233 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/594b61fa-12b7-47b6-8af3-17f024e0d54d-config-data\") pod \"594b61fa-12b7-47b6-8af3-17f024e0d54d\" (UID: \"594b61fa-12b7-47b6-8af3-17f024e0d54d\") " Mar 13 14:27:46 crc kubenswrapper[4907]: I0313 14:27:46.419713 4907 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/594b61fa-12b7-47b6-8af3-17f024e0d54d-etc-machine-id\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:46 crc kubenswrapper[4907]: I0313 14:27:46.425936 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/594b61fa-12b7-47b6-8af3-17f024e0d54d-scripts" (OuterVolumeSpecName: "scripts") pod "594b61fa-12b7-47b6-8af3-17f024e0d54d" (UID: "594b61fa-12b7-47b6-8af3-17f024e0d54d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:27:46 crc kubenswrapper[4907]: I0313 14:27:46.425980 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/594b61fa-12b7-47b6-8af3-17f024e0d54d-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "594b61fa-12b7-47b6-8af3-17f024e0d54d" (UID: "594b61fa-12b7-47b6-8af3-17f024e0d54d"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:27:46 crc kubenswrapper[4907]: I0313 14:27:46.427063 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/594b61fa-12b7-47b6-8af3-17f024e0d54d-kube-api-access-wpft2" (OuterVolumeSpecName: "kube-api-access-wpft2") pod "594b61fa-12b7-47b6-8af3-17f024e0d54d" (UID: "594b61fa-12b7-47b6-8af3-17f024e0d54d"). InnerVolumeSpecName "kube-api-access-wpft2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:27:46 crc kubenswrapper[4907]: I0313 14:27:46.451700 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/594b61fa-12b7-47b6-8af3-17f024e0d54d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "594b61fa-12b7-47b6-8af3-17f024e0d54d" (UID: "594b61fa-12b7-47b6-8af3-17f024e0d54d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:27:46 crc kubenswrapper[4907]: I0313 14:27:46.473125 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/594b61fa-12b7-47b6-8af3-17f024e0d54d-config-data" (OuterVolumeSpecName: "config-data") pod "594b61fa-12b7-47b6-8af3-17f024e0d54d" (UID: "594b61fa-12b7-47b6-8af3-17f024e0d54d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:27:46 crc kubenswrapper[4907]: I0313 14:27:46.522589 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/594b61fa-12b7-47b6-8af3-17f024e0d54d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:46 crc kubenswrapper[4907]: I0313 14:27:46.522664 4907 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/594b61fa-12b7-47b6-8af3-17f024e0d54d-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:46 crc kubenswrapper[4907]: I0313 14:27:46.522682 4907 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/594b61fa-12b7-47b6-8af3-17f024e0d54d-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:46 crc kubenswrapper[4907]: I0313 14:27:46.522697 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wpft2\" (UniqueName: \"kubernetes.io/projected/594b61fa-12b7-47b6-8af3-17f024e0d54d-kube-api-access-wpft2\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:46 crc kubenswrapper[4907]: I0313 14:27:46.522712 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/594b61fa-12b7-47b6-8af3-17f024e0d54d-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:46 crc kubenswrapper[4907]: I0313 14:27:46.872841 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 14:27:47 crc kubenswrapper[4907]: I0313 14:27:47.748925 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5-log-httpd\") pod \"58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5\" (UID: \"58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5\") " Mar 13 14:27:47 crc kubenswrapper[4907]: I0313 14:27:47.749007 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5-config-data\") pod \"58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5\" (UID: \"58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5\") " Mar 13 14:27:47 crc kubenswrapper[4907]: I0313 14:27:47.749032 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5-scripts\") pod \"58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5\" (UID: \"58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5\") " Mar 13 14:27:47 crc kubenswrapper[4907]: I0313 14:27:47.749097 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2xbmw\" (UniqueName: \"kubernetes.io/projected/58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5-kube-api-access-2xbmw\") pod \"58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5\" (UID: \"58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5\") " Mar 13 14:27:47 crc kubenswrapper[4907]: I0313 14:27:47.749131 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5-sg-core-conf-yaml\") pod \"58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5\" (UID: \"58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5\") " Mar 13 14:27:47 crc kubenswrapper[4907]: I0313 14:27:47.749203 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5-run-httpd\") pod \"58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5\" (UID: \"58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5\") " Mar 13 14:27:47 crc kubenswrapper[4907]: I0313 14:27:47.749338 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5-combined-ca-bundle\") pod \"58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5\" (UID: \"58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5\") " Mar 13 14:27:47 crc kubenswrapper[4907]: I0313 14:27:47.752645 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5" (UID: "58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:27:47 crc kubenswrapper[4907]: I0313 14:27:47.756717 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5" (UID: "58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:27:47 crc kubenswrapper[4907]: I0313 14:27:47.781246 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5-kube-api-access-2xbmw" (OuterVolumeSpecName: "kube-api-access-2xbmw") pod "58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5" (UID: "58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5"). InnerVolumeSpecName "kube-api-access-2xbmw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:27:47 crc kubenswrapper[4907]: I0313 14:27:47.781341 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5-scripts" (OuterVolumeSpecName: "scripts") pod "58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5" (UID: "58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:27:47 crc kubenswrapper[4907]: I0313 14:27:47.888342 4907 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:47 crc kubenswrapper[4907]: I0313 14:27:47.888650 4907 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:47 crc kubenswrapper[4907]: I0313 14:27:47.888659 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2xbmw\" (UniqueName: \"kubernetes.io/projected/58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5-kube-api-access-2xbmw\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:47 crc kubenswrapper[4907]: I0313 14:27:47.888668 4907 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:47 crc kubenswrapper[4907]: I0313 14:27:47.907671 4907 generic.go:334] "Generic (PLEG): container finished" podID="58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5" containerID="af56a2f9ef345a9cf4a63dafe33dfb02aa9518339a5d2ef4a11fbf9ec8ab7512" exitCode=0 Mar 13 14:27:47 crc kubenswrapper[4907]: I0313 14:27:47.907805 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 14:27:47 crc kubenswrapper[4907]: I0313 14:27:47.927712 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5" (UID: "58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:27:47 crc kubenswrapper[4907]: I0313 14:27:47.941691 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-7q999" Mar 13 14:27:47 crc kubenswrapper[4907]: I0313 14:27:47.995239 4907 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.006909 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5","Type":"ContainerDied","Data":"af56a2f9ef345a9cf4a63dafe33dfb02aa9518339a5d2ef4a11fbf9ec8ab7512"} Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.006969 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5","Type":"ContainerDied","Data":"cf9fd571ce09eea426c8836be80b3c4ef776ec41dabdcd7886212b3709f230ce"} Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.006986 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-7q999" event={"ID":"594b61fa-12b7-47b6-8af3-17f024e0d54d","Type":"ContainerDied","Data":"247e75fb8f70b076f7d10c60ec2902246989846bb77cff82137c6d0beb413598"} Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.007003 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="247e75fb8f70b076f7d10c60ec2902246989846bb77cff82137c6d0beb413598" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.007022 4907 scope.go:117] "RemoveContainer" containerID="dd889dbfbaa8a48ab7c7b9c0cf914f3824e4417de226256cac660ed3cfc140b1" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.034042 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5" (UID: "58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.046059 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.046122 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.075173 4907 scope.go:117] "RemoveContainer" containerID="a15bc8c355c09d50b27215622b3ea5aaebf1df184e4f387e2405fe44a886e49b" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.100450 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.137265 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5-config-data" (OuterVolumeSpecName: "config-data") pod "58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5" (UID: "58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.142959 4907 scope.go:117] "RemoveContainer" containerID="af56a2f9ef345a9cf4a63dafe33dfb02aa9518339a5d2ef4a11fbf9ec8ab7512" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.158650 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Mar 13 14:27:48 crc kubenswrapper[4907]: E0313 14:27:48.159080 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="070e2b75-05b1-42a4-a34e-950d45d65920" containerName="barbican-api" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.159098 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="070e2b75-05b1-42a4-a34e-950d45d65920" containerName="barbican-api" Mar 13 14:27:48 crc kubenswrapper[4907]: E0313 14:27:48.159113 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca4928ba-1277-4918-8b92-3eda2b276577" containerName="barbican-worker" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.159120 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca4928ba-1277-4918-8b92-3eda2b276577" containerName="barbican-worker" Mar 13 14:27:48 crc kubenswrapper[4907]: E0313 14:27:48.159132 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5" containerName="ceilometer-central-agent" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.159142 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5" containerName="ceilometer-central-agent" Mar 13 14:27:48 crc kubenswrapper[4907]: E0313 14:27:48.159150 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="594b61fa-12b7-47b6-8af3-17f024e0d54d" containerName="cinder-db-sync" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.159158 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="594b61fa-12b7-47b6-8af3-17f024e0d54d" containerName="cinder-db-sync" Mar 13 14:27:48 crc kubenswrapper[4907]: E0313 14:27:48.159168 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="070e2b75-05b1-42a4-a34e-950d45d65920" containerName="barbican-api-log" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.159174 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="070e2b75-05b1-42a4-a34e-950d45d65920" containerName="barbican-api-log" Mar 13 14:27:48 crc kubenswrapper[4907]: E0313 14:27:48.159188 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fcbb1687-db0b-4fc6-937c-89102831237b" containerName="barbican-keystone-listener-log" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.159195 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="fcbb1687-db0b-4fc6-937c-89102831237b" containerName="barbican-keystone-listener-log" Mar 13 14:27:48 crc kubenswrapper[4907]: E0313 14:27:48.159209 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5" containerName="sg-core" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.159216 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5" containerName="sg-core" Mar 13 14:27:48 crc kubenswrapper[4907]: E0313 14:27:48.159230 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5" containerName="proxy-httpd" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.159237 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5" containerName="proxy-httpd" Mar 13 14:27:48 crc kubenswrapper[4907]: E0313 14:27:48.159251 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a661b9e1-c551-4b36-87a9-833df73bd6b2" containerName="dnsmasq-dns" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.159259 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="a661b9e1-c551-4b36-87a9-833df73bd6b2" containerName="dnsmasq-dns" Mar 13 14:27:48 crc kubenswrapper[4907]: E0313 14:27:48.159269 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a661b9e1-c551-4b36-87a9-833df73bd6b2" containerName="init" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.159275 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="a661b9e1-c551-4b36-87a9-833df73bd6b2" containerName="init" Mar 13 14:27:48 crc kubenswrapper[4907]: E0313 14:27:48.159300 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca4928ba-1277-4918-8b92-3eda2b276577" containerName="barbican-worker-log" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.159307 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca4928ba-1277-4918-8b92-3eda2b276577" containerName="barbican-worker-log" Mar 13 14:27:48 crc kubenswrapper[4907]: E0313 14:27:48.159326 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5" containerName="ceilometer-notification-agent" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.159333 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5" containerName="ceilometer-notification-agent" Mar 13 14:27:48 crc kubenswrapper[4907]: E0313 14:27:48.159343 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fcbb1687-db0b-4fc6-937c-89102831237b" containerName="barbican-keystone-listener" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.159352 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="fcbb1687-db0b-4fc6-937c-89102831237b" containerName="barbican-keystone-listener" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.159546 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="fcbb1687-db0b-4fc6-937c-89102831237b" containerName="barbican-keystone-listener" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.159561 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="fcbb1687-db0b-4fc6-937c-89102831237b" containerName="barbican-keystone-listener-log" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.159572 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="a661b9e1-c551-4b36-87a9-833df73bd6b2" containerName="dnsmasq-dns" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.159583 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="594b61fa-12b7-47b6-8af3-17f024e0d54d" containerName="cinder-db-sync" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.159601 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="ca4928ba-1277-4918-8b92-3eda2b276577" containerName="barbican-worker-log" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.159613 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="070e2b75-05b1-42a4-a34e-950d45d65920" containerName="barbican-api-log" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.165009 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5" containerName="sg-core" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.165023 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5" containerName="ceilometer-notification-agent" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.165038 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="070e2b75-05b1-42a4-a34e-950d45d65920" containerName="barbican-api" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.165055 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="ca4928ba-1277-4918-8b92-3eda2b276577" containerName="barbican-worker" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.165070 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5" containerName="ceilometer-central-agent" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.165079 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5" containerName="proxy-httpd" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.166157 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.178165 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.178483 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.178608 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.178819 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-d7phq" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.185758 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.201614 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.271687 4907 scope.go:117] "RemoveContainer" containerID="b9aab9a1290561a45c480e877db9222806553cdd6cfb12daf299ab965ad2ec5c" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.303448 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r9xxm\" (UniqueName: \"kubernetes.io/projected/e7fd8554-798f-4e3a-acf1-8425c15e2665-kube-api-access-r9xxm\") pod \"cinder-scheduler-0\" (UID: \"e7fd8554-798f-4e3a-acf1-8425c15e2665\") " pod="openstack/cinder-scheduler-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.303517 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e7fd8554-798f-4e3a-acf1-8425c15e2665-config-data\") pod \"cinder-scheduler-0\" (UID: \"e7fd8554-798f-4e3a-acf1-8425c15e2665\") " pod="openstack/cinder-scheduler-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.303544 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e7fd8554-798f-4e3a-acf1-8425c15e2665-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"e7fd8554-798f-4e3a-acf1-8425c15e2665\") " pod="openstack/cinder-scheduler-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.303574 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e7fd8554-798f-4e3a-acf1-8425c15e2665-scripts\") pod \"cinder-scheduler-0\" (UID: \"e7fd8554-798f-4e3a-acf1-8425c15e2665\") " pod="openstack/cinder-scheduler-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.303610 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e7fd8554-798f-4e3a-acf1-8425c15e2665-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"e7fd8554-798f-4e3a-acf1-8425c15e2665\") " pod="openstack/cinder-scheduler-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.303655 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7fd8554-798f-4e3a-acf1-8425c15e2665-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"e7fd8554-798f-4e3a-acf1-8425c15e2665\") " pod="openstack/cinder-scheduler-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.324968 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.335523 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-58b85ccffc-jt5lv"] Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.337256 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58b85ccffc-jt5lv" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.351415 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.366283 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-58b85ccffc-jt5lv"] Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.375156 4907 scope.go:117] "RemoveContainer" containerID="dd889dbfbaa8a48ab7c7b9c0cf914f3824e4417de226256cac660ed3cfc140b1" Mar 13 14:27:48 crc kubenswrapper[4907]: E0313 14:27:48.377058 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dd889dbfbaa8a48ab7c7b9c0cf914f3824e4417de226256cac660ed3cfc140b1\": container with ID starting with dd889dbfbaa8a48ab7c7b9c0cf914f3824e4417de226256cac660ed3cfc140b1 not found: ID does not exist" containerID="dd889dbfbaa8a48ab7c7b9c0cf914f3824e4417de226256cac660ed3cfc140b1" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.377101 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd889dbfbaa8a48ab7c7b9c0cf914f3824e4417de226256cac660ed3cfc140b1"} err="failed to get container status \"dd889dbfbaa8a48ab7c7b9c0cf914f3824e4417de226256cac660ed3cfc140b1\": rpc error: code = NotFound desc = could not find container \"dd889dbfbaa8a48ab7c7b9c0cf914f3824e4417de226256cac660ed3cfc140b1\": container with ID starting with dd889dbfbaa8a48ab7c7b9c0cf914f3824e4417de226256cac660ed3cfc140b1 not found: ID does not exist" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.377127 4907 scope.go:117] "RemoveContainer" containerID="a15bc8c355c09d50b27215622b3ea5aaebf1df184e4f387e2405fe44a886e49b" Mar 13 14:27:48 crc kubenswrapper[4907]: E0313 14:27:48.377399 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a15bc8c355c09d50b27215622b3ea5aaebf1df184e4f387e2405fe44a886e49b\": container with ID starting with a15bc8c355c09d50b27215622b3ea5aaebf1df184e4f387e2405fe44a886e49b not found: ID does not exist" containerID="a15bc8c355c09d50b27215622b3ea5aaebf1df184e4f387e2405fe44a886e49b" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.377417 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a15bc8c355c09d50b27215622b3ea5aaebf1df184e4f387e2405fe44a886e49b"} err="failed to get container status \"a15bc8c355c09d50b27215622b3ea5aaebf1df184e4f387e2405fe44a886e49b\": rpc error: code = NotFound desc = could not find container \"a15bc8c355c09d50b27215622b3ea5aaebf1df184e4f387e2405fe44a886e49b\": container with ID starting with a15bc8c355c09d50b27215622b3ea5aaebf1df184e4f387e2405fe44a886e49b not found: ID does not exist" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.377428 4907 scope.go:117] "RemoveContainer" containerID="af56a2f9ef345a9cf4a63dafe33dfb02aa9518339a5d2ef4a11fbf9ec8ab7512" Mar 13 14:27:48 crc kubenswrapper[4907]: E0313 14:27:48.380133 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"af56a2f9ef345a9cf4a63dafe33dfb02aa9518339a5d2ef4a11fbf9ec8ab7512\": container with ID starting with af56a2f9ef345a9cf4a63dafe33dfb02aa9518339a5d2ef4a11fbf9ec8ab7512 not found: ID does not exist" containerID="af56a2f9ef345a9cf4a63dafe33dfb02aa9518339a5d2ef4a11fbf9ec8ab7512" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.380169 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af56a2f9ef345a9cf4a63dafe33dfb02aa9518339a5d2ef4a11fbf9ec8ab7512"} err="failed to get container status \"af56a2f9ef345a9cf4a63dafe33dfb02aa9518339a5d2ef4a11fbf9ec8ab7512\": rpc error: code = NotFound desc = could not find container \"af56a2f9ef345a9cf4a63dafe33dfb02aa9518339a5d2ef4a11fbf9ec8ab7512\": container with ID starting with af56a2f9ef345a9cf4a63dafe33dfb02aa9518339a5d2ef4a11fbf9ec8ab7512 not found: ID does not exist" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.380196 4907 scope.go:117] "RemoveContainer" containerID="b9aab9a1290561a45c480e877db9222806553cdd6cfb12daf299ab965ad2ec5c" Mar 13 14:27:48 crc kubenswrapper[4907]: E0313 14:27:48.391295 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b9aab9a1290561a45c480e877db9222806553cdd6cfb12daf299ab965ad2ec5c\": container with ID starting with b9aab9a1290561a45c480e877db9222806553cdd6cfb12daf299ab965ad2ec5c not found: ID does not exist" containerID="b9aab9a1290561a45c480e877db9222806553cdd6cfb12daf299ab965ad2ec5c" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.391333 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b9aab9a1290561a45c480e877db9222806553cdd6cfb12daf299ab965ad2ec5c"} err="failed to get container status \"b9aab9a1290561a45c480e877db9222806553cdd6cfb12daf299ab965ad2ec5c\": rpc error: code = NotFound desc = could not find container \"b9aab9a1290561a45c480e877db9222806553cdd6cfb12daf299ab965ad2ec5c\": container with ID starting with b9aab9a1290561a45c480e877db9222806553cdd6cfb12daf299ab965ad2ec5c not found: ID does not exist" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.391743 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.394241 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.399789 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.400039 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.411621 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e7fd8554-798f-4e3a-acf1-8425c15e2665-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"e7fd8554-798f-4e3a-acf1-8425c15e2665\") " pod="openstack/cinder-scheduler-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.411720 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24-dns-svc\") pod \"dnsmasq-dns-58b85ccffc-jt5lv\" (UID: \"eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24\") " pod="openstack/dnsmasq-dns-58b85ccffc-jt5lv" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.411768 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5zvgh\" (UniqueName: \"kubernetes.io/projected/eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24-kube-api-access-5zvgh\") pod \"dnsmasq-dns-58b85ccffc-jt5lv\" (UID: \"eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24\") " pod="openstack/dnsmasq-dns-58b85ccffc-jt5lv" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.411848 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7fd8554-798f-4e3a-acf1-8425c15e2665-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"e7fd8554-798f-4e3a-acf1-8425c15e2665\") " pod="openstack/cinder-scheduler-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.411947 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r9xxm\" (UniqueName: \"kubernetes.io/projected/e7fd8554-798f-4e3a-acf1-8425c15e2665-kube-api-access-r9xxm\") pod \"cinder-scheduler-0\" (UID: \"e7fd8554-798f-4e3a-acf1-8425c15e2665\") " pod="openstack/cinder-scheduler-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.412054 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24-config\") pod \"dnsmasq-dns-58b85ccffc-jt5lv\" (UID: \"eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24\") " pod="openstack/dnsmasq-dns-58b85ccffc-jt5lv" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.412089 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24-ovsdbserver-sb\") pod \"dnsmasq-dns-58b85ccffc-jt5lv\" (UID: \"eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24\") " pod="openstack/dnsmasq-dns-58b85ccffc-jt5lv" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.412123 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e7fd8554-798f-4e3a-acf1-8425c15e2665-config-data\") pod \"cinder-scheduler-0\" (UID: \"e7fd8554-798f-4e3a-acf1-8425c15e2665\") " pod="openstack/cinder-scheduler-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.412173 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24-ovsdbserver-nb\") pod \"dnsmasq-dns-58b85ccffc-jt5lv\" (UID: \"eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24\") " pod="openstack/dnsmasq-dns-58b85ccffc-jt5lv" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.412199 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e7fd8554-798f-4e3a-acf1-8425c15e2665-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"e7fd8554-798f-4e3a-acf1-8425c15e2665\") " pod="openstack/cinder-scheduler-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.412221 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24-dns-swift-storage-0\") pod \"dnsmasq-dns-58b85ccffc-jt5lv\" (UID: \"eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24\") " pod="openstack/dnsmasq-dns-58b85ccffc-jt5lv" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.412295 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e7fd8554-798f-4e3a-acf1-8425c15e2665-scripts\") pod \"cinder-scheduler-0\" (UID: \"e7fd8554-798f-4e3a-acf1-8425c15e2665\") " pod="openstack/cinder-scheduler-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.413899 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.414809 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e7fd8554-798f-4e3a-acf1-8425c15e2665-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"e7fd8554-798f-4e3a-acf1-8425c15e2665\") " pod="openstack/cinder-scheduler-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.431735 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7fd8554-798f-4e3a-acf1-8425c15e2665-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"e7fd8554-798f-4e3a-acf1-8425c15e2665\") " pod="openstack/cinder-scheduler-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.432097 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e7fd8554-798f-4e3a-acf1-8425c15e2665-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"e7fd8554-798f-4e3a-acf1-8425c15e2665\") " pod="openstack/cinder-scheduler-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.432781 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e7fd8554-798f-4e3a-acf1-8425c15e2665-config-data\") pod \"cinder-scheduler-0\" (UID: \"e7fd8554-798f-4e3a-acf1-8425c15e2665\") " pod="openstack/cinder-scheduler-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.435786 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e7fd8554-798f-4e3a-acf1-8425c15e2665-scripts\") pod \"cinder-scheduler-0\" (UID: \"e7fd8554-798f-4e3a-acf1-8425c15e2665\") " pod="openstack/cinder-scheduler-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.444011 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r9xxm\" (UniqueName: \"kubernetes.io/projected/e7fd8554-798f-4e3a-acf1-8425c15e2665-kube-api-access-r9xxm\") pod \"cinder-scheduler-0\" (UID: \"e7fd8554-798f-4e3a-acf1-8425c15e2665\") " pod="openstack/cinder-scheduler-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.466214 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.467979 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.477330 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.507208 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.517557 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f744eee5-064f-464c-86fc-5130071b313d-scripts\") pod \"ceilometer-0\" (UID: \"f744eee5-064f-464c-86fc-5130071b313d\") " pod="openstack/ceilometer-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.517673 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dtpdc\" (UniqueName: \"kubernetes.io/projected/f744eee5-064f-464c-86fc-5130071b313d-kube-api-access-dtpdc\") pod \"ceilometer-0\" (UID: \"f744eee5-064f-464c-86fc-5130071b313d\") " pod="openstack/ceilometer-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.517723 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24-config\") pod \"dnsmasq-dns-58b85ccffc-jt5lv\" (UID: \"eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24\") " pod="openstack/dnsmasq-dns-58b85ccffc-jt5lv" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.517762 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24-ovsdbserver-sb\") pod \"dnsmasq-dns-58b85ccffc-jt5lv\" (UID: \"eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24\") " pod="openstack/dnsmasq-dns-58b85ccffc-jt5lv" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.517794 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f744eee5-064f-464c-86fc-5130071b313d-config-data\") pod \"ceilometer-0\" (UID: \"f744eee5-064f-464c-86fc-5130071b313d\") " pod="openstack/ceilometer-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.518916 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24-ovsdbserver-sb\") pod \"dnsmasq-dns-58b85ccffc-jt5lv\" (UID: \"eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24\") " pod="openstack/dnsmasq-dns-58b85ccffc-jt5lv" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.519104 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24-ovsdbserver-nb\") pod \"dnsmasq-dns-58b85ccffc-jt5lv\" (UID: \"eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24\") " pod="openstack/dnsmasq-dns-58b85ccffc-jt5lv" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.519137 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24-dns-swift-storage-0\") pod \"dnsmasq-dns-58b85ccffc-jt5lv\" (UID: \"eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24\") " pod="openstack/dnsmasq-dns-58b85ccffc-jt5lv" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.519173 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f744eee5-064f-464c-86fc-5130071b313d-log-httpd\") pod \"ceilometer-0\" (UID: \"f744eee5-064f-464c-86fc-5130071b313d\") " pod="openstack/ceilometer-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.519194 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f744eee5-064f-464c-86fc-5130071b313d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f744eee5-064f-464c-86fc-5130071b313d\") " pod="openstack/ceilometer-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.519253 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f744eee5-064f-464c-86fc-5130071b313d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f744eee5-064f-464c-86fc-5130071b313d\") " pod="openstack/ceilometer-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.519275 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24-dns-svc\") pod \"dnsmasq-dns-58b85ccffc-jt5lv\" (UID: \"eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24\") " pod="openstack/dnsmasq-dns-58b85ccffc-jt5lv" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.519300 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f744eee5-064f-464c-86fc-5130071b313d-run-httpd\") pod \"ceilometer-0\" (UID: \"f744eee5-064f-464c-86fc-5130071b313d\") " pod="openstack/ceilometer-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.519323 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5zvgh\" (UniqueName: \"kubernetes.io/projected/eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24-kube-api-access-5zvgh\") pod \"dnsmasq-dns-58b85ccffc-jt5lv\" (UID: \"eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24\") " pod="openstack/dnsmasq-dns-58b85ccffc-jt5lv" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.520234 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24-ovsdbserver-nb\") pod \"dnsmasq-dns-58b85ccffc-jt5lv\" (UID: \"eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24\") " pod="openstack/dnsmasq-dns-58b85ccffc-jt5lv" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.520763 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24-dns-swift-storage-0\") pod \"dnsmasq-dns-58b85ccffc-jt5lv\" (UID: \"eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24\") " pod="openstack/dnsmasq-dns-58b85ccffc-jt5lv" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.521343 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24-dns-svc\") pod \"dnsmasq-dns-58b85ccffc-jt5lv\" (UID: \"eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24\") " pod="openstack/dnsmasq-dns-58b85ccffc-jt5lv" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.521288 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24-config\") pod \"dnsmasq-dns-58b85ccffc-jt5lv\" (UID: \"eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24\") " pod="openstack/dnsmasq-dns-58b85ccffc-jt5lv" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.541490 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5zvgh\" (UniqueName: \"kubernetes.io/projected/eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24-kube-api-access-5zvgh\") pod \"dnsmasq-dns-58b85ccffc-jt5lv\" (UID: \"eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24\") " pod="openstack/dnsmasq-dns-58b85ccffc-jt5lv" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.567812 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.621129 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xlvfk\" (UniqueName: \"kubernetes.io/projected/ca2411de-0fb9-4702-aae0-808e5d759667-kube-api-access-xlvfk\") pod \"cinder-api-0\" (UID: \"ca2411de-0fb9-4702-aae0-808e5d759667\") " pod="openstack/cinder-api-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.621181 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dtpdc\" (UniqueName: \"kubernetes.io/projected/f744eee5-064f-464c-86fc-5130071b313d-kube-api-access-dtpdc\") pod \"ceilometer-0\" (UID: \"f744eee5-064f-464c-86fc-5130071b313d\") " pod="openstack/ceilometer-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.621214 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f744eee5-064f-464c-86fc-5130071b313d-config-data\") pod \"ceilometer-0\" (UID: \"f744eee5-064f-464c-86fc-5130071b313d\") " pod="openstack/ceilometer-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.621254 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f744eee5-064f-464c-86fc-5130071b313d-log-httpd\") pod \"ceilometer-0\" (UID: \"f744eee5-064f-464c-86fc-5130071b313d\") " pod="openstack/ceilometer-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.621268 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f744eee5-064f-464c-86fc-5130071b313d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f744eee5-064f-464c-86fc-5130071b313d\") " pod="openstack/ceilometer-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.621316 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f744eee5-064f-464c-86fc-5130071b313d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f744eee5-064f-464c-86fc-5130071b313d\") " pod="openstack/ceilometer-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.621333 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca2411de-0fb9-4702-aae0-808e5d759667-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"ca2411de-0fb9-4702-aae0-808e5d759667\") " pod="openstack/cinder-api-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.621349 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f744eee5-064f-464c-86fc-5130071b313d-run-httpd\") pod \"ceilometer-0\" (UID: \"f744eee5-064f-464c-86fc-5130071b313d\") " pod="openstack/ceilometer-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.621365 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ca2411de-0fb9-4702-aae0-808e5d759667-logs\") pod \"cinder-api-0\" (UID: \"ca2411de-0fb9-4702-aae0-808e5d759667\") " pod="openstack/cinder-api-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.621389 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca2411de-0fb9-4702-aae0-808e5d759667-config-data\") pod \"cinder-api-0\" (UID: \"ca2411de-0fb9-4702-aae0-808e5d759667\") " pod="openstack/cinder-api-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.621411 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ca2411de-0fb9-4702-aae0-808e5d759667-config-data-custom\") pod \"cinder-api-0\" (UID: \"ca2411de-0fb9-4702-aae0-808e5d759667\") " pod="openstack/cinder-api-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.621432 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f744eee5-064f-464c-86fc-5130071b313d-scripts\") pod \"ceilometer-0\" (UID: \"f744eee5-064f-464c-86fc-5130071b313d\") " pod="openstack/ceilometer-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.621447 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ca2411de-0fb9-4702-aae0-808e5d759667-scripts\") pod \"cinder-api-0\" (UID: \"ca2411de-0fb9-4702-aae0-808e5d759667\") " pod="openstack/cinder-api-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.621488 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ca2411de-0fb9-4702-aae0-808e5d759667-etc-machine-id\") pod \"cinder-api-0\" (UID: \"ca2411de-0fb9-4702-aae0-808e5d759667\") " pod="openstack/cinder-api-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.625579 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f744eee5-064f-464c-86fc-5130071b313d-log-httpd\") pod \"ceilometer-0\" (UID: \"f744eee5-064f-464c-86fc-5130071b313d\") " pod="openstack/ceilometer-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.630197 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f744eee5-064f-464c-86fc-5130071b313d-run-httpd\") pod \"ceilometer-0\" (UID: \"f744eee5-064f-464c-86fc-5130071b313d\") " pod="openstack/ceilometer-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.631097 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f744eee5-064f-464c-86fc-5130071b313d-config-data\") pod \"ceilometer-0\" (UID: \"f744eee5-064f-464c-86fc-5130071b313d\") " pod="openstack/ceilometer-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.634413 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f744eee5-064f-464c-86fc-5130071b313d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f744eee5-064f-464c-86fc-5130071b313d\") " pod="openstack/ceilometer-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.634500 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f744eee5-064f-464c-86fc-5130071b313d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f744eee5-064f-464c-86fc-5130071b313d\") " pod="openstack/ceilometer-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.634802 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f744eee5-064f-464c-86fc-5130071b313d-scripts\") pod \"ceilometer-0\" (UID: \"f744eee5-064f-464c-86fc-5130071b313d\") " pod="openstack/ceilometer-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.638965 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dtpdc\" (UniqueName: \"kubernetes.io/projected/f744eee5-064f-464c-86fc-5130071b313d-kube-api-access-dtpdc\") pod \"ceilometer-0\" (UID: \"f744eee5-064f-464c-86fc-5130071b313d\") " pod="openstack/ceilometer-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.693360 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58b85ccffc-jt5lv" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.723262 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ca2411de-0fb9-4702-aae0-808e5d759667-scripts\") pod \"cinder-api-0\" (UID: \"ca2411de-0fb9-4702-aae0-808e5d759667\") " pod="openstack/cinder-api-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.723680 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ca2411de-0fb9-4702-aae0-808e5d759667-etc-machine-id\") pod \"cinder-api-0\" (UID: \"ca2411de-0fb9-4702-aae0-808e5d759667\") " pod="openstack/cinder-api-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.723714 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xlvfk\" (UniqueName: \"kubernetes.io/projected/ca2411de-0fb9-4702-aae0-808e5d759667-kube-api-access-xlvfk\") pod \"cinder-api-0\" (UID: \"ca2411de-0fb9-4702-aae0-808e5d759667\") " pod="openstack/cinder-api-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.723815 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca2411de-0fb9-4702-aae0-808e5d759667-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"ca2411de-0fb9-4702-aae0-808e5d759667\") " pod="openstack/cinder-api-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.723841 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ca2411de-0fb9-4702-aae0-808e5d759667-logs\") pod \"cinder-api-0\" (UID: \"ca2411de-0fb9-4702-aae0-808e5d759667\") " pod="openstack/cinder-api-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.723868 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca2411de-0fb9-4702-aae0-808e5d759667-config-data\") pod \"cinder-api-0\" (UID: \"ca2411de-0fb9-4702-aae0-808e5d759667\") " pod="openstack/cinder-api-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.723911 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ca2411de-0fb9-4702-aae0-808e5d759667-config-data-custom\") pod \"cinder-api-0\" (UID: \"ca2411de-0fb9-4702-aae0-808e5d759667\") " pod="openstack/cinder-api-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.726188 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ca2411de-0fb9-4702-aae0-808e5d759667-etc-machine-id\") pod \"cinder-api-0\" (UID: \"ca2411de-0fb9-4702-aae0-808e5d759667\") " pod="openstack/cinder-api-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.726245 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ca2411de-0fb9-4702-aae0-808e5d759667-logs\") pod \"cinder-api-0\" (UID: \"ca2411de-0fb9-4702-aae0-808e5d759667\") " pod="openstack/cinder-api-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.730101 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ca2411de-0fb9-4702-aae0-808e5d759667-config-data-custom\") pod \"cinder-api-0\" (UID: \"ca2411de-0fb9-4702-aae0-808e5d759667\") " pod="openstack/cinder-api-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.732484 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca2411de-0fb9-4702-aae0-808e5d759667-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"ca2411de-0fb9-4702-aae0-808e5d759667\") " pod="openstack/cinder-api-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.736326 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ca2411de-0fb9-4702-aae0-808e5d759667-scripts\") pod \"cinder-api-0\" (UID: \"ca2411de-0fb9-4702-aae0-808e5d759667\") " pod="openstack/cinder-api-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.736907 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca2411de-0fb9-4702-aae0-808e5d759667-config-data\") pod \"cinder-api-0\" (UID: \"ca2411de-0fb9-4702-aae0-808e5d759667\") " pod="openstack/cinder-api-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.758075 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xlvfk\" (UniqueName: \"kubernetes.io/projected/ca2411de-0fb9-4702-aae0-808e5d759667-kube-api-access-xlvfk\") pod \"cinder-api-0\" (UID: \"ca2411de-0fb9-4702-aae0-808e5d759667\") " pod="openstack/cinder-api-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.831583 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.890291 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Mar 13 14:27:49 crc kubenswrapper[4907]: I0313 14:27:49.060566 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Mar 13 14:27:49 crc kubenswrapper[4907]: W0313 14:27:49.072338 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode7fd8554_798f_4e3a_acf1_8425c15e2665.slice/crio-58e8239ab4329277f0169cd2f4714646b1acec175437b4fe78554e7e213348e5 WatchSource:0}: Error finding container 58e8239ab4329277f0169cd2f4714646b1acec175437b4fe78554e7e213348e5: Status 404 returned error can't find the container with id 58e8239ab4329277f0169cd2f4714646b1acec175437b4fe78554e7e213348e5 Mar 13 14:27:49 crc kubenswrapper[4907]: I0313 14:27:49.262807 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-58b85ccffc-jt5lv"] Mar 13 14:27:49 crc kubenswrapper[4907]: W0313 14:27:49.263684 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeed444b5_3ec0_44b6_b0a0_3e3e5a66fe24.slice/crio-a66bc3530699af62851033529fd1a73503f7565c8f14e919806b488f53d8de33 WatchSource:0}: Error finding container a66bc3530699af62851033529fd1a73503f7565c8f14e919806b488f53d8de33: Status 404 returned error can't find the container with id a66bc3530699af62851033529fd1a73503f7565c8f14e919806b488f53d8de33 Mar 13 14:27:49 crc kubenswrapper[4907]: I0313 14:27:49.439171 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Mar 13 14:27:49 crc kubenswrapper[4907]: I0313 14:27:49.455824 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 13 14:27:49 crc kubenswrapper[4907]: I0313 14:27:49.808708 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5" path="/var/lib/kubelet/pods/58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5/volumes" Mar 13 14:27:50 crc kubenswrapper[4907]: I0313 14:27:50.004402 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f744eee5-064f-464c-86fc-5130071b313d","Type":"ContainerStarted","Data":"ad91253ab4bed9131d491de9a5560dd1867583453c957940978da876fc7db3cf"} Mar 13 14:27:50 crc kubenswrapper[4907]: I0313 14:27:50.018094 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"ca2411de-0fb9-4702-aae0-808e5d759667","Type":"ContainerStarted","Data":"bbe16c55906016b19d3449613b6670869e1b477f29a2b60cc4d61f1e4a53e85e"} Mar 13 14:27:50 crc kubenswrapper[4907]: I0313 14:27:50.029319 4907 generic.go:334] "Generic (PLEG): container finished" podID="eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24" containerID="4f693ef5bd7b7f8e96bb97b86b9be101cd4aab9df400b462e0a79f5818a5b73b" exitCode=0 Mar 13 14:27:50 crc kubenswrapper[4907]: I0313 14:27:50.029385 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58b85ccffc-jt5lv" event={"ID":"eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24","Type":"ContainerDied","Data":"4f693ef5bd7b7f8e96bb97b86b9be101cd4aab9df400b462e0a79f5818a5b73b"} Mar 13 14:27:50 crc kubenswrapper[4907]: I0313 14:27:50.029410 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58b85ccffc-jt5lv" event={"ID":"eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24","Type":"ContainerStarted","Data":"a66bc3530699af62851033529fd1a73503f7565c8f14e919806b488f53d8de33"} Mar 13 14:27:50 crc kubenswrapper[4907]: I0313 14:27:50.042750 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e7fd8554-798f-4e3a-acf1-8425c15e2665","Type":"ContainerStarted","Data":"58e8239ab4329277f0169cd2f4714646b1acec175437b4fe78554e7e213348e5"} Mar 13 14:27:51 crc kubenswrapper[4907]: I0313 14:27:51.079763 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Mar 13 14:27:51 crc kubenswrapper[4907]: I0313 14:27:51.082211 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f744eee5-064f-464c-86fc-5130071b313d","Type":"ContainerStarted","Data":"041e4dec51c360b9cb2fc26fee14b909f22ef44631885cdd0d5fc5f23a7ef2e5"} Mar 13 14:27:51 crc kubenswrapper[4907]: I0313 14:27:51.082237 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f744eee5-064f-464c-86fc-5130071b313d","Type":"ContainerStarted","Data":"2ba377587d3943d9d1a24298ede2bf3a69a7589b2b93375a231da14a6aa5fa8e"} Mar 13 14:27:51 crc kubenswrapper[4907]: I0313 14:27:51.084313 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"ca2411de-0fb9-4702-aae0-808e5d759667","Type":"ContainerStarted","Data":"e994224e3dd0dfe7ba5a1ff2805c3b640032239385a154735e1317449e6ca000"} Mar 13 14:27:51 crc kubenswrapper[4907]: I0313 14:27:51.085781 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58b85ccffc-jt5lv" event={"ID":"eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24","Type":"ContainerStarted","Data":"6b06c4d709787d5a6f17229ea98351d8fef051feb5e0a2046eb23fd42751cf63"} Mar 13 14:27:51 crc kubenswrapper[4907]: I0313 14:27:51.087051 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-58b85ccffc-jt5lv" Mar 13 14:27:51 crc kubenswrapper[4907]: I0313 14:27:51.097403 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e7fd8554-798f-4e3a-acf1-8425c15e2665","Type":"ContainerStarted","Data":"0a62f5f3cb31fd39481383f9dca4b2b20b3144ace647921ed05a9b4b9214a38d"} Mar 13 14:27:51 crc kubenswrapper[4907]: I0313 14:27:51.103694 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-58b85ccffc-jt5lv" podStartSLOduration=3.103677053 podStartE2EDuration="3.103677053s" podCreationTimestamp="2026-03-13 14:27:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:27:51.102543802 +0000 UTC m=+1370.002331491" watchObservedRunningTime="2026-03-13 14:27:51.103677053 +0000 UTC m=+1370.003464742" Mar 13 14:27:52 crc kubenswrapper[4907]: I0313 14:27:52.111156 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"ca2411de-0fb9-4702-aae0-808e5d759667","Type":"ContainerStarted","Data":"1a5d6bde83c26d0dd9c7eb5214a4d0bd35bb82277b346a9cbb3e5efef6ed9c8d"} Mar 13 14:27:52 crc kubenswrapper[4907]: I0313 14:27:52.111711 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Mar 13 14:27:52 crc kubenswrapper[4907]: I0313 14:27:52.111390 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="ca2411de-0fb9-4702-aae0-808e5d759667" containerName="cinder-api" containerID="cri-o://1a5d6bde83c26d0dd9c7eb5214a4d0bd35bb82277b346a9cbb3e5efef6ed9c8d" gracePeriod=30 Mar 13 14:27:52 crc kubenswrapper[4907]: I0313 14:27:52.111305 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="ca2411de-0fb9-4702-aae0-808e5d759667" containerName="cinder-api-log" containerID="cri-o://e994224e3dd0dfe7ba5a1ff2805c3b640032239385a154735e1317449e6ca000" gracePeriod=30 Mar 13 14:27:52 crc kubenswrapper[4907]: I0313 14:27:52.117035 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e7fd8554-798f-4e3a-acf1-8425c15e2665","Type":"ContainerStarted","Data":"2b64ad381c9b4854fe824a41af04da817955826edfb608941d06803d7f8f4d59"} Mar 13 14:27:52 crc kubenswrapper[4907]: I0313 14:27:52.120618 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f744eee5-064f-464c-86fc-5130071b313d","Type":"ContainerStarted","Data":"e47ab86336baef65bd0ea7edbd7c31a75a6d68df9c09a6a282295545b620225a"} Mar 13 14:27:52 crc kubenswrapper[4907]: I0313 14:27:52.140480 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=4.140462571 podStartE2EDuration="4.140462571s" podCreationTimestamp="2026-03-13 14:27:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:27:52.130586333 +0000 UTC m=+1371.030374022" watchObservedRunningTime="2026-03-13 14:27:52.140462571 +0000 UTC m=+1371.040250260" Mar 13 14:27:52 crc kubenswrapper[4907]: I0313 14:27:52.156712 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.570057367 podStartE2EDuration="4.156692862s" podCreationTimestamp="2026-03-13 14:27:48 +0000 UTC" firstStartedPulling="2026-03-13 14:27:49.078811942 +0000 UTC m=+1367.978599641" lastFinishedPulling="2026-03-13 14:27:49.665447447 +0000 UTC m=+1368.565235136" observedRunningTime="2026-03-13 14:27:52.155261883 +0000 UTC m=+1371.055049582" watchObservedRunningTime="2026-03-13 14:27:52.156692862 +0000 UTC m=+1371.056480551" Mar 13 14:27:52 crc kubenswrapper[4907]: I0313 14:27:52.770668 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Mar 13 14:27:52 crc kubenswrapper[4907]: I0313 14:27:52.910807 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ca2411de-0fb9-4702-aae0-808e5d759667-scripts\") pod \"ca2411de-0fb9-4702-aae0-808e5d759667\" (UID: \"ca2411de-0fb9-4702-aae0-808e5d759667\") " Mar 13 14:27:52 crc kubenswrapper[4907]: I0313 14:27:52.910867 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca2411de-0fb9-4702-aae0-808e5d759667-combined-ca-bundle\") pod \"ca2411de-0fb9-4702-aae0-808e5d759667\" (UID: \"ca2411de-0fb9-4702-aae0-808e5d759667\") " Mar 13 14:27:52 crc kubenswrapper[4907]: I0313 14:27:52.910935 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ca2411de-0fb9-4702-aae0-808e5d759667-logs\") pod \"ca2411de-0fb9-4702-aae0-808e5d759667\" (UID: \"ca2411de-0fb9-4702-aae0-808e5d759667\") " Mar 13 14:27:52 crc kubenswrapper[4907]: I0313 14:27:52.911032 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ca2411de-0fb9-4702-aae0-808e5d759667-config-data-custom\") pod \"ca2411de-0fb9-4702-aae0-808e5d759667\" (UID: \"ca2411de-0fb9-4702-aae0-808e5d759667\") " Mar 13 14:27:52 crc kubenswrapper[4907]: I0313 14:27:52.911083 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xlvfk\" (UniqueName: \"kubernetes.io/projected/ca2411de-0fb9-4702-aae0-808e5d759667-kube-api-access-xlvfk\") pod \"ca2411de-0fb9-4702-aae0-808e5d759667\" (UID: \"ca2411de-0fb9-4702-aae0-808e5d759667\") " Mar 13 14:27:52 crc kubenswrapper[4907]: I0313 14:27:52.911432 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ca2411de-0fb9-4702-aae0-808e5d759667-logs" (OuterVolumeSpecName: "logs") pod "ca2411de-0fb9-4702-aae0-808e5d759667" (UID: "ca2411de-0fb9-4702-aae0-808e5d759667"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:27:52 crc kubenswrapper[4907]: I0313 14:27:52.911835 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ca2411de-0fb9-4702-aae0-808e5d759667-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "ca2411de-0fb9-4702-aae0-808e5d759667" (UID: "ca2411de-0fb9-4702-aae0-808e5d759667"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 14:27:52 crc kubenswrapper[4907]: I0313 14:27:52.911779 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ca2411de-0fb9-4702-aae0-808e5d759667-etc-machine-id\") pod \"ca2411de-0fb9-4702-aae0-808e5d759667\" (UID: \"ca2411de-0fb9-4702-aae0-808e5d759667\") " Mar 13 14:27:52 crc kubenswrapper[4907]: I0313 14:27:52.911967 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca2411de-0fb9-4702-aae0-808e5d759667-config-data\") pod \"ca2411de-0fb9-4702-aae0-808e5d759667\" (UID: \"ca2411de-0fb9-4702-aae0-808e5d759667\") " Mar 13 14:27:52 crc kubenswrapper[4907]: I0313 14:27:52.912435 4907 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ca2411de-0fb9-4702-aae0-808e5d759667-logs\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:52 crc kubenswrapper[4907]: I0313 14:27:52.912450 4907 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ca2411de-0fb9-4702-aae0-808e5d759667-etc-machine-id\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:52 crc kubenswrapper[4907]: I0313 14:27:52.930317 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ca2411de-0fb9-4702-aae0-808e5d759667-kube-api-access-xlvfk" (OuterVolumeSpecName: "kube-api-access-xlvfk") pod "ca2411de-0fb9-4702-aae0-808e5d759667" (UID: "ca2411de-0fb9-4702-aae0-808e5d759667"). InnerVolumeSpecName "kube-api-access-xlvfk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:27:52 crc kubenswrapper[4907]: I0313 14:27:52.931549 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca2411de-0fb9-4702-aae0-808e5d759667-scripts" (OuterVolumeSpecName: "scripts") pod "ca2411de-0fb9-4702-aae0-808e5d759667" (UID: "ca2411de-0fb9-4702-aae0-808e5d759667"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:27:52 crc kubenswrapper[4907]: I0313 14:27:52.952046 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca2411de-0fb9-4702-aae0-808e5d759667-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "ca2411de-0fb9-4702-aae0-808e5d759667" (UID: "ca2411de-0fb9-4702-aae0-808e5d759667"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.019341 4907 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ca2411de-0fb9-4702-aae0-808e5d759667-config-data-custom\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.019574 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xlvfk\" (UniqueName: \"kubernetes.io/projected/ca2411de-0fb9-4702-aae0-808e5d759667-kube-api-access-xlvfk\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.019667 4907 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ca2411de-0fb9-4702-aae0-808e5d759667-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.033069 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca2411de-0fb9-4702-aae0-808e5d759667-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ca2411de-0fb9-4702-aae0-808e5d759667" (UID: "ca2411de-0fb9-4702-aae0-808e5d759667"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.041139 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca2411de-0fb9-4702-aae0-808e5d759667-config-data" (OuterVolumeSpecName: "config-data") pod "ca2411de-0fb9-4702-aae0-808e5d759667" (UID: "ca2411de-0fb9-4702-aae0-808e5d759667"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.121792 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca2411de-0fb9-4702-aae0-808e5d759667-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.121823 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca2411de-0fb9-4702-aae0-808e5d759667-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.130355 4907 generic.go:334] "Generic (PLEG): container finished" podID="ca2411de-0fb9-4702-aae0-808e5d759667" containerID="1a5d6bde83c26d0dd9c7eb5214a4d0bd35bb82277b346a9cbb3e5efef6ed9c8d" exitCode=0 Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.130386 4907 generic.go:334] "Generic (PLEG): container finished" podID="ca2411de-0fb9-4702-aae0-808e5d759667" containerID="e994224e3dd0dfe7ba5a1ff2805c3b640032239385a154735e1317449e6ca000" exitCode=143 Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.130432 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.130463 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"ca2411de-0fb9-4702-aae0-808e5d759667","Type":"ContainerDied","Data":"1a5d6bde83c26d0dd9c7eb5214a4d0bd35bb82277b346a9cbb3e5efef6ed9c8d"} Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.130509 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"ca2411de-0fb9-4702-aae0-808e5d759667","Type":"ContainerDied","Data":"e994224e3dd0dfe7ba5a1ff2805c3b640032239385a154735e1317449e6ca000"} Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.130568 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"ca2411de-0fb9-4702-aae0-808e5d759667","Type":"ContainerDied","Data":"bbe16c55906016b19d3449613b6670869e1b477f29a2b60cc4d61f1e4a53e85e"} Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.130608 4907 scope.go:117] "RemoveContainer" containerID="1a5d6bde83c26d0dd9c7eb5214a4d0bd35bb82277b346a9cbb3e5efef6ed9c8d" Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.159712 4907 scope.go:117] "RemoveContainer" containerID="e994224e3dd0dfe7ba5a1ff2805c3b640032239385a154735e1317449e6ca000" Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.165191 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.183334 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.191556 4907 scope.go:117] "RemoveContainer" containerID="1a5d6bde83c26d0dd9c7eb5214a4d0bd35bb82277b346a9cbb3e5efef6ed9c8d" Mar 13 14:27:53 crc kubenswrapper[4907]: E0313 14:27:53.194129 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1a5d6bde83c26d0dd9c7eb5214a4d0bd35bb82277b346a9cbb3e5efef6ed9c8d\": container with ID starting with 1a5d6bde83c26d0dd9c7eb5214a4d0bd35bb82277b346a9cbb3e5efef6ed9c8d not found: ID does not exist" containerID="1a5d6bde83c26d0dd9c7eb5214a4d0bd35bb82277b346a9cbb3e5efef6ed9c8d" Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.194183 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a5d6bde83c26d0dd9c7eb5214a4d0bd35bb82277b346a9cbb3e5efef6ed9c8d"} err="failed to get container status \"1a5d6bde83c26d0dd9c7eb5214a4d0bd35bb82277b346a9cbb3e5efef6ed9c8d\": rpc error: code = NotFound desc = could not find container \"1a5d6bde83c26d0dd9c7eb5214a4d0bd35bb82277b346a9cbb3e5efef6ed9c8d\": container with ID starting with 1a5d6bde83c26d0dd9c7eb5214a4d0bd35bb82277b346a9cbb3e5efef6ed9c8d not found: ID does not exist" Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.194216 4907 scope.go:117] "RemoveContainer" containerID="e994224e3dd0dfe7ba5a1ff2805c3b640032239385a154735e1317449e6ca000" Mar 13 14:27:53 crc kubenswrapper[4907]: E0313 14:27:53.194675 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e994224e3dd0dfe7ba5a1ff2805c3b640032239385a154735e1317449e6ca000\": container with ID starting with e994224e3dd0dfe7ba5a1ff2805c3b640032239385a154735e1317449e6ca000 not found: ID does not exist" containerID="e994224e3dd0dfe7ba5a1ff2805c3b640032239385a154735e1317449e6ca000" Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.194702 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e994224e3dd0dfe7ba5a1ff2805c3b640032239385a154735e1317449e6ca000"} err="failed to get container status \"e994224e3dd0dfe7ba5a1ff2805c3b640032239385a154735e1317449e6ca000\": rpc error: code = NotFound desc = could not find container \"e994224e3dd0dfe7ba5a1ff2805c3b640032239385a154735e1317449e6ca000\": container with ID starting with e994224e3dd0dfe7ba5a1ff2805c3b640032239385a154735e1317449e6ca000 not found: ID does not exist" Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.194720 4907 scope.go:117] "RemoveContainer" containerID="1a5d6bde83c26d0dd9c7eb5214a4d0bd35bb82277b346a9cbb3e5efef6ed9c8d" Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.194965 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a5d6bde83c26d0dd9c7eb5214a4d0bd35bb82277b346a9cbb3e5efef6ed9c8d"} err="failed to get container status \"1a5d6bde83c26d0dd9c7eb5214a4d0bd35bb82277b346a9cbb3e5efef6ed9c8d\": rpc error: code = NotFound desc = could not find container \"1a5d6bde83c26d0dd9c7eb5214a4d0bd35bb82277b346a9cbb3e5efef6ed9c8d\": container with ID starting with 1a5d6bde83c26d0dd9c7eb5214a4d0bd35bb82277b346a9cbb3e5efef6ed9c8d not found: ID does not exist" Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.194985 4907 scope.go:117] "RemoveContainer" containerID="e994224e3dd0dfe7ba5a1ff2805c3b640032239385a154735e1317449e6ca000" Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.195276 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e994224e3dd0dfe7ba5a1ff2805c3b640032239385a154735e1317449e6ca000"} err="failed to get container status \"e994224e3dd0dfe7ba5a1ff2805c3b640032239385a154735e1317449e6ca000\": rpc error: code = NotFound desc = could not find container \"e994224e3dd0dfe7ba5a1ff2805c3b640032239385a154735e1317449e6ca000\": container with ID starting with e994224e3dd0dfe7ba5a1ff2805c3b640032239385a154735e1317449e6ca000 not found: ID does not exist" Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.209716 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Mar 13 14:27:53 crc kubenswrapper[4907]: E0313 14:27:53.210498 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca2411de-0fb9-4702-aae0-808e5d759667" containerName="cinder-api" Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.210519 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca2411de-0fb9-4702-aae0-808e5d759667" containerName="cinder-api" Mar 13 14:27:53 crc kubenswrapper[4907]: E0313 14:27:53.210545 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca2411de-0fb9-4702-aae0-808e5d759667" containerName="cinder-api-log" Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.210555 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca2411de-0fb9-4702-aae0-808e5d759667" containerName="cinder-api-log" Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.210799 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="ca2411de-0fb9-4702-aae0-808e5d759667" containerName="cinder-api-log" Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.210831 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="ca2411de-0fb9-4702-aae0-808e5d759667" containerName="cinder-api" Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.212040 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.216005 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.217131 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.217365 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.226731 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76-config-data\") pod \"cinder-api-0\" (UID: \"41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76\") " pod="openstack/cinder-api-0" Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.227075 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dnjx6\" (UniqueName: \"kubernetes.io/projected/41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76-kube-api-access-dnjx6\") pod \"cinder-api-0\" (UID: \"41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76\") " pod="openstack/cinder-api-0" Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.227301 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76-etc-machine-id\") pod \"cinder-api-0\" (UID: \"41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76\") " pod="openstack/cinder-api-0" Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.227518 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76\") " pod="openstack/cinder-api-0" Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.227622 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76-scripts\") pod \"cinder-api-0\" (UID: \"41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76\") " pod="openstack/cinder-api-0" Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.227781 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76-config-data-custom\") pod \"cinder-api-0\" (UID: \"41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76\") " pod="openstack/cinder-api-0" Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.227974 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76-public-tls-certs\") pod \"cinder-api-0\" (UID: \"41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76\") " pod="openstack/cinder-api-0" Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.228125 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76\") " pod="openstack/cinder-api-0" Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.228253 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76-logs\") pod \"cinder-api-0\" (UID: \"41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76\") " pod="openstack/cinder-api-0" Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.233585 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.331365 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76-public-tls-certs\") pod \"cinder-api-0\" (UID: \"41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76\") " pod="openstack/cinder-api-0" Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.331471 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76\") " pod="openstack/cinder-api-0" Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.331551 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76-logs\") pod \"cinder-api-0\" (UID: \"41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76\") " pod="openstack/cinder-api-0" Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.332226 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76-config-data\") pod \"cinder-api-0\" (UID: \"41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76\") " pod="openstack/cinder-api-0" Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.332381 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dnjx6\" (UniqueName: \"kubernetes.io/projected/41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76-kube-api-access-dnjx6\") pod \"cinder-api-0\" (UID: \"41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76\") " pod="openstack/cinder-api-0" Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.332584 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76-etc-machine-id\") pod \"cinder-api-0\" (UID: \"41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76\") " pod="openstack/cinder-api-0" Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.333240 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76\") " pod="openstack/cinder-api-0" Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.333313 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76-scripts\") pod \"cinder-api-0\" (UID: \"41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76\") " pod="openstack/cinder-api-0" Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.333364 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76-etc-machine-id\") pod \"cinder-api-0\" (UID: \"41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76\") " pod="openstack/cinder-api-0" Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.333407 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76-config-data-custom\") pod \"cinder-api-0\" (UID: \"41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76\") " pod="openstack/cinder-api-0" Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.334403 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76-logs\") pod \"cinder-api-0\" (UID: \"41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76\") " pod="openstack/cinder-api-0" Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.334723 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76-public-tls-certs\") pod \"cinder-api-0\" (UID: \"41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76\") " pod="openstack/cinder-api-0" Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.342500 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76-config-data\") pod \"cinder-api-0\" (UID: \"41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76\") " pod="openstack/cinder-api-0" Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.343248 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76\") " pod="openstack/cinder-api-0" Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.346369 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76-scripts\") pod \"cinder-api-0\" (UID: \"41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76\") " pod="openstack/cinder-api-0" Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.347153 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76\") " pod="openstack/cinder-api-0" Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.351398 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dnjx6\" (UniqueName: \"kubernetes.io/projected/41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76-kube-api-access-dnjx6\") pod \"cinder-api-0\" (UID: \"41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76\") " pod="openstack/cinder-api-0" Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.353430 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76-config-data-custom\") pod \"cinder-api-0\" (UID: \"41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76\") " pod="openstack/cinder-api-0" Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.408143 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-7fc86597fb-mt9pk" Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.570379 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.583176 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.649657 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-7fc86597fb-mt9pk" Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.759466 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-74984bd8b4-x42pq"] Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.759773 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-74984bd8b4-x42pq" podUID="032183a7-c5db-4528-a211-803050007b68" containerName="barbican-api-log" containerID="cri-o://3dc5784d660cfabfa43e3df490d1b2200ffa14fc3ff516c9488bcc67f04af55b" gracePeriod=30 Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.760439 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-74984bd8b4-x42pq" podUID="032183a7-c5db-4528-a211-803050007b68" containerName="barbican-api" containerID="cri-o://be905f5a0aa412e370260df6ece658648121619b19bad67aee3af44172b10250" gracePeriod=30 Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.824097 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ca2411de-0fb9-4702-aae0-808e5d759667" path="/var/lib/kubelet/pods/ca2411de-0fb9-4702-aae0-808e5d759667/volumes" Mar 13 14:27:54 crc kubenswrapper[4907]: I0313 14:27:54.106462 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Mar 13 14:27:54 crc kubenswrapper[4907]: W0313 14:27:54.110208 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod41d7a0d1_a3ce_45f6_a8b7_fcf9de7f7a76.slice/crio-10db10f108be05500948726431cb9db777e566b5230954b5ba74892f495e9236 WatchSource:0}: Error finding container 10db10f108be05500948726431cb9db777e566b5230954b5ba74892f495e9236: Status 404 returned error can't find the container with id 10db10f108be05500948726431cb9db777e566b5230954b5ba74892f495e9236 Mar 13 14:27:54 crc kubenswrapper[4907]: I0313 14:27:54.148846 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f744eee5-064f-464c-86fc-5130071b313d","Type":"ContainerStarted","Data":"7a11049e1abd6d70a403607531097f57571aa3ac1adb2b5af76d4fe1cd5fb6bd"} Mar 13 14:27:54 crc kubenswrapper[4907]: I0313 14:27:54.150206 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Mar 13 14:27:54 crc kubenswrapper[4907]: I0313 14:27:54.151675 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76","Type":"ContainerStarted","Data":"10db10f108be05500948726431cb9db777e566b5230954b5ba74892f495e9236"} Mar 13 14:27:54 crc kubenswrapper[4907]: I0313 14:27:54.158999 4907 generic.go:334] "Generic (PLEG): container finished" podID="032183a7-c5db-4528-a211-803050007b68" containerID="3dc5784d660cfabfa43e3df490d1b2200ffa14fc3ff516c9488bcc67f04af55b" exitCode=143 Mar 13 14:27:54 crc kubenswrapper[4907]: I0313 14:27:54.159105 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-74984bd8b4-x42pq" event={"ID":"032183a7-c5db-4528-a211-803050007b68","Type":"ContainerDied","Data":"3dc5784d660cfabfa43e3df490d1b2200ffa14fc3ff516c9488bcc67f04af55b"} Mar 13 14:27:54 crc kubenswrapper[4907]: I0313 14:27:54.178495 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.250185069 podStartE2EDuration="6.178475129s" podCreationTimestamp="2026-03-13 14:27:48 +0000 UTC" firstStartedPulling="2026-03-13 14:27:49.448725132 +0000 UTC m=+1368.348512821" lastFinishedPulling="2026-03-13 14:27:53.377015202 +0000 UTC m=+1372.276802881" observedRunningTime="2026-03-13 14:27:54.170162593 +0000 UTC m=+1373.069950282" watchObservedRunningTime="2026-03-13 14:27:54.178475129 +0000 UTC m=+1373.078262818" Mar 13 14:27:55 crc kubenswrapper[4907]: I0313 14:27:55.171375 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76","Type":"ContainerStarted","Data":"2cbce92733c8c36aab8d17f31bd5085e394cb7fec21de278bf9301bd315d6bae"} Mar 13 14:27:56 crc kubenswrapper[4907]: I0313 14:27:56.181995 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76","Type":"ContainerStarted","Data":"a07507237e2303a227977bfc1acdae544607145ad994269ea3936668b9e9c650"} Mar 13 14:27:56 crc kubenswrapper[4907]: I0313 14:27:56.214597 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.214577906 podStartE2EDuration="3.214577906s" podCreationTimestamp="2026-03-13 14:27:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:27:56.213935209 +0000 UTC m=+1375.113722898" watchObservedRunningTime="2026-03-13 14:27:56.214577906 +0000 UTC m=+1375.114365605" Mar 13 14:27:57 crc kubenswrapper[4907]: I0313 14:27:57.192515 4907 generic.go:334] "Generic (PLEG): container finished" podID="032183a7-c5db-4528-a211-803050007b68" containerID="be905f5a0aa412e370260df6ece658648121619b19bad67aee3af44172b10250" exitCode=0 Mar 13 14:27:57 crc kubenswrapper[4907]: I0313 14:27:57.193923 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-74984bd8b4-x42pq" event={"ID":"032183a7-c5db-4528-a211-803050007b68","Type":"ContainerDied","Data":"be905f5a0aa412e370260df6ece658648121619b19bad67aee3af44172b10250"} Mar 13 14:27:57 crc kubenswrapper[4907]: I0313 14:27:57.193965 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Mar 13 14:27:57 crc kubenswrapper[4907]: I0313 14:27:57.361412 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-74984bd8b4-x42pq" Mar 13 14:27:57 crc kubenswrapper[4907]: I0313 14:27:57.532826 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/032183a7-c5db-4528-a211-803050007b68-config-data\") pod \"032183a7-c5db-4528-a211-803050007b68\" (UID: \"032183a7-c5db-4528-a211-803050007b68\") " Mar 13 14:27:57 crc kubenswrapper[4907]: I0313 14:27:57.533115 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/032183a7-c5db-4528-a211-803050007b68-config-data-custom\") pod \"032183a7-c5db-4528-a211-803050007b68\" (UID: \"032183a7-c5db-4528-a211-803050007b68\") " Mar 13 14:27:57 crc kubenswrapper[4907]: I0313 14:27:57.533177 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/032183a7-c5db-4528-a211-803050007b68-logs\") pod \"032183a7-c5db-4528-a211-803050007b68\" (UID: \"032183a7-c5db-4528-a211-803050007b68\") " Mar 13 14:27:57 crc kubenswrapper[4907]: I0313 14:27:57.533208 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tfsh2\" (UniqueName: \"kubernetes.io/projected/032183a7-c5db-4528-a211-803050007b68-kube-api-access-tfsh2\") pod \"032183a7-c5db-4528-a211-803050007b68\" (UID: \"032183a7-c5db-4528-a211-803050007b68\") " Mar 13 14:27:57 crc kubenswrapper[4907]: I0313 14:27:57.533235 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/032183a7-c5db-4528-a211-803050007b68-combined-ca-bundle\") pod \"032183a7-c5db-4528-a211-803050007b68\" (UID: \"032183a7-c5db-4528-a211-803050007b68\") " Mar 13 14:27:57 crc kubenswrapper[4907]: I0313 14:27:57.533874 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/032183a7-c5db-4528-a211-803050007b68-logs" (OuterVolumeSpecName: "logs") pod "032183a7-c5db-4528-a211-803050007b68" (UID: "032183a7-c5db-4528-a211-803050007b68"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:27:57 crc kubenswrapper[4907]: I0313 14:27:57.534296 4907 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/032183a7-c5db-4528-a211-803050007b68-logs\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:57 crc kubenswrapper[4907]: I0313 14:27:57.540311 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/032183a7-c5db-4528-a211-803050007b68-kube-api-access-tfsh2" (OuterVolumeSpecName: "kube-api-access-tfsh2") pod "032183a7-c5db-4528-a211-803050007b68" (UID: "032183a7-c5db-4528-a211-803050007b68"). InnerVolumeSpecName "kube-api-access-tfsh2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:27:57 crc kubenswrapper[4907]: I0313 14:27:57.547168 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/032183a7-c5db-4528-a211-803050007b68-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "032183a7-c5db-4528-a211-803050007b68" (UID: "032183a7-c5db-4528-a211-803050007b68"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:27:57 crc kubenswrapper[4907]: I0313 14:27:57.586722 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/032183a7-c5db-4528-a211-803050007b68-config-data" (OuterVolumeSpecName: "config-data") pod "032183a7-c5db-4528-a211-803050007b68" (UID: "032183a7-c5db-4528-a211-803050007b68"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:27:57 crc kubenswrapper[4907]: I0313 14:27:57.588347 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/032183a7-c5db-4528-a211-803050007b68-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "032183a7-c5db-4528-a211-803050007b68" (UID: "032183a7-c5db-4528-a211-803050007b68"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:27:57 crc kubenswrapper[4907]: I0313 14:27:57.635661 4907 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/032183a7-c5db-4528-a211-803050007b68-config-data-custom\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:57 crc kubenswrapper[4907]: I0313 14:27:57.635692 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tfsh2\" (UniqueName: \"kubernetes.io/projected/032183a7-c5db-4528-a211-803050007b68-kube-api-access-tfsh2\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:57 crc kubenswrapper[4907]: I0313 14:27:57.635704 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/032183a7-c5db-4528-a211-803050007b68-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:57 crc kubenswrapper[4907]: I0313 14:27:57.635714 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/032183a7-c5db-4528-a211-803050007b68-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:58 crc kubenswrapper[4907]: I0313 14:27:58.203664 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-74984bd8b4-x42pq" event={"ID":"032183a7-c5db-4528-a211-803050007b68","Type":"ContainerDied","Data":"3fa9f3f934d41bda9d2c0367bd05516b9fcfddd3de52bfe4494f0529db4eafd7"} Mar 13 14:27:58 crc kubenswrapper[4907]: I0313 14:27:58.203713 4907 scope.go:117] "RemoveContainer" containerID="be905f5a0aa412e370260df6ece658648121619b19bad67aee3af44172b10250" Mar 13 14:27:58 crc kubenswrapper[4907]: I0313 14:27:58.203727 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-74984bd8b4-x42pq" Mar 13 14:27:58 crc kubenswrapper[4907]: I0313 14:27:58.229959 4907 scope.go:117] "RemoveContainer" containerID="3dc5784d660cfabfa43e3df490d1b2200ffa14fc3ff516c9488bcc67f04af55b" Mar 13 14:27:58 crc kubenswrapper[4907]: I0313 14:27:58.230775 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-74984bd8b4-x42pq"] Mar 13 14:27:58 crc kubenswrapper[4907]: I0313 14:27:58.240556 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-74984bd8b4-x42pq"] Mar 13 14:27:58 crc kubenswrapper[4907]: I0313 14:27:58.697127 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-58b85ccffc-jt5lv" Mar 13 14:27:58 crc kubenswrapper[4907]: I0313 14:27:58.763068 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7fc46d7df7-77dng"] Mar 13 14:27:58 crc kubenswrapper[4907]: I0313 14:27:58.763644 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7fc46d7df7-77dng" podUID="49bc69d8-fe85-4994-a5cf-c51472a124e9" containerName="dnsmasq-dns" containerID="cri-o://10817f472001c46a816f4b77aa0c7ef3e99533487d552c19bd292e7357506e9f" gracePeriod=10 Mar 13 14:27:58 crc kubenswrapper[4907]: I0313 14:27:58.939469 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Mar 13 14:27:58 crc kubenswrapper[4907]: I0313 14:27:58.998523 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Mar 13 14:27:59 crc kubenswrapper[4907]: I0313 14:27:59.011675 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-67d6ccd9c4-ms98l" Mar 13 14:27:59 crc kubenswrapper[4907]: I0313 14:27:59.076287 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-67d6ccd9c4-ms98l" Mar 13 14:27:59 crc kubenswrapper[4907]: I0313 14:27:59.218615 4907 generic.go:334] "Generic (PLEG): container finished" podID="49bc69d8-fe85-4994-a5cf-c51472a124e9" containerID="10817f472001c46a816f4b77aa0c7ef3e99533487d552c19bd292e7357506e9f" exitCode=0 Mar 13 14:27:59 crc kubenswrapper[4907]: I0313 14:27:59.218702 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fc46d7df7-77dng" event={"ID":"49bc69d8-fe85-4994-a5cf-c51472a124e9","Type":"ContainerDied","Data":"10817f472001c46a816f4b77aa0c7ef3e99533487d552c19bd292e7357506e9f"} Mar 13 14:27:59 crc kubenswrapper[4907]: I0313 14:27:59.222038 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="e7fd8554-798f-4e3a-acf1-8425c15e2665" containerName="cinder-scheduler" containerID="cri-o://0a62f5f3cb31fd39481383f9dca4b2b20b3144ace647921ed05a9b4b9214a38d" gracePeriod=30 Mar 13 14:27:59 crc kubenswrapper[4907]: I0313 14:27:59.222592 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="e7fd8554-798f-4e3a-acf1-8425c15e2665" containerName="probe" containerID="cri-o://2b64ad381c9b4854fe824a41af04da817955826edfb608941d06803d7f8f4d59" gracePeriod=30 Mar 13 14:27:59 crc kubenswrapper[4907]: I0313 14:27:59.400491 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-85dc486cc4-fwrm6"] Mar 13 14:27:59 crc kubenswrapper[4907]: E0313 14:27:59.401143 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="032183a7-c5db-4528-a211-803050007b68" containerName="barbican-api" Mar 13 14:27:59 crc kubenswrapper[4907]: I0313 14:27:59.401160 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="032183a7-c5db-4528-a211-803050007b68" containerName="barbican-api" Mar 13 14:27:59 crc kubenswrapper[4907]: E0313 14:27:59.401173 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="032183a7-c5db-4528-a211-803050007b68" containerName="barbican-api-log" Mar 13 14:27:59 crc kubenswrapper[4907]: I0313 14:27:59.401179 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="032183a7-c5db-4528-a211-803050007b68" containerName="barbican-api-log" Mar 13 14:27:59 crc kubenswrapper[4907]: I0313 14:27:59.401327 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="032183a7-c5db-4528-a211-803050007b68" containerName="barbican-api" Mar 13 14:27:59 crc kubenswrapper[4907]: I0313 14:27:59.401350 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="032183a7-c5db-4528-a211-803050007b68" containerName="barbican-api-log" Mar 13 14:27:59 crc kubenswrapper[4907]: I0313 14:27:59.402224 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-85dc486cc4-fwrm6" Mar 13 14:27:59 crc kubenswrapper[4907]: I0313 14:27:59.423757 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-85dc486cc4-fwrm6"] Mar 13 14:27:59 crc kubenswrapper[4907]: I0313 14:27:59.428734 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fc46d7df7-77dng" Mar 13 14:27:59 crc kubenswrapper[4907]: I0313 14:27:59.479669 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e3ee069a-41fb-4cab-b650-9bb9c03ac271-logs\") pod \"placement-85dc486cc4-fwrm6\" (UID: \"e3ee069a-41fb-4cab-b650-9bb9c03ac271\") " pod="openstack/placement-85dc486cc4-fwrm6" Mar 13 14:27:59 crc kubenswrapper[4907]: I0313 14:27:59.479735 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3ee069a-41fb-4cab-b650-9bb9c03ac271-config-data\") pod \"placement-85dc486cc4-fwrm6\" (UID: \"e3ee069a-41fb-4cab-b650-9bb9c03ac271\") " pod="openstack/placement-85dc486cc4-fwrm6" Mar 13 14:27:59 crc kubenswrapper[4907]: I0313 14:27:59.479767 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e3ee069a-41fb-4cab-b650-9bb9c03ac271-internal-tls-certs\") pod \"placement-85dc486cc4-fwrm6\" (UID: \"e3ee069a-41fb-4cab-b650-9bb9c03ac271\") " pod="openstack/placement-85dc486cc4-fwrm6" Mar 13 14:27:59 crc kubenswrapper[4907]: I0313 14:27:59.479793 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gds2t\" (UniqueName: \"kubernetes.io/projected/e3ee069a-41fb-4cab-b650-9bb9c03ac271-kube-api-access-gds2t\") pod \"placement-85dc486cc4-fwrm6\" (UID: \"e3ee069a-41fb-4cab-b650-9bb9c03ac271\") " pod="openstack/placement-85dc486cc4-fwrm6" Mar 13 14:27:59 crc kubenswrapper[4907]: I0313 14:27:59.479854 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e3ee069a-41fb-4cab-b650-9bb9c03ac271-scripts\") pod \"placement-85dc486cc4-fwrm6\" (UID: \"e3ee069a-41fb-4cab-b650-9bb9c03ac271\") " pod="openstack/placement-85dc486cc4-fwrm6" Mar 13 14:27:59 crc kubenswrapper[4907]: I0313 14:27:59.479906 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3ee069a-41fb-4cab-b650-9bb9c03ac271-combined-ca-bundle\") pod \"placement-85dc486cc4-fwrm6\" (UID: \"e3ee069a-41fb-4cab-b650-9bb9c03ac271\") " pod="openstack/placement-85dc486cc4-fwrm6" Mar 13 14:27:59 crc kubenswrapper[4907]: I0313 14:27:59.479931 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e3ee069a-41fb-4cab-b650-9bb9c03ac271-public-tls-certs\") pod \"placement-85dc486cc4-fwrm6\" (UID: \"e3ee069a-41fb-4cab-b650-9bb9c03ac271\") " pod="openstack/placement-85dc486cc4-fwrm6" Mar 13 14:27:59 crc kubenswrapper[4907]: I0313 14:27:59.581254 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/49bc69d8-fe85-4994-a5cf-c51472a124e9-ovsdbserver-nb\") pod \"49bc69d8-fe85-4994-a5cf-c51472a124e9\" (UID: \"49bc69d8-fe85-4994-a5cf-c51472a124e9\") " Mar 13 14:27:59 crc kubenswrapper[4907]: I0313 14:27:59.581567 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/49bc69d8-fe85-4994-a5cf-c51472a124e9-config\") pod \"49bc69d8-fe85-4994-a5cf-c51472a124e9\" (UID: \"49bc69d8-fe85-4994-a5cf-c51472a124e9\") " Mar 13 14:27:59 crc kubenswrapper[4907]: I0313 14:27:59.582607 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/49bc69d8-fe85-4994-a5cf-c51472a124e9-dns-swift-storage-0\") pod \"49bc69d8-fe85-4994-a5cf-c51472a124e9\" (UID: \"49bc69d8-fe85-4994-a5cf-c51472a124e9\") " Mar 13 14:27:59 crc kubenswrapper[4907]: I0313 14:27:59.582721 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/49bc69d8-fe85-4994-a5cf-c51472a124e9-ovsdbserver-sb\") pod \"49bc69d8-fe85-4994-a5cf-c51472a124e9\" (UID: \"49bc69d8-fe85-4994-a5cf-c51472a124e9\") " Mar 13 14:27:59 crc kubenswrapper[4907]: I0313 14:27:59.582862 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r77nl\" (UniqueName: \"kubernetes.io/projected/49bc69d8-fe85-4994-a5cf-c51472a124e9-kube-api-access-r77nl\") pod \"49bc69d8-fe85-4994-a5cf-c51472a124e9\" (UID: \"49bc69d8-fe85-4994-a5cf-c51472a124e9\") " Mar 13 14:27:59 crc kubenswrapper[4907]: I0313 14:27:59.583090 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/49bc69d8-fe85-4994-a5cf-c51472a124e9-dns-svc\") pod \"49bc69d8-fe85-4994-a5cf-c51472a124e9\" (UID: \"49bc69d8-fe85-4994-a5cf-c51472a124e9\") " Mar 13 14:27:59 crc kubenswrapper[4907]: I0313 14:27:59.583364 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3ee069a-41fb-4cab-b650-9bb9c03ac271-config-data\") pod \"placement-85dc486cc4-fwrm6\" (UID: \"e3ee069a-41fb-4cab-b650-9bb9c03ac271\") " pod="openstack/placement-85dc486cc4-fwrm6" Mar 13 14:27:59 crc kubenswrapper[4907]: I0313 14:27:59.583486 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e3ee069a-41fb-4cab-b650-9bb9c03ac271-internal-tls-certs\") pod \"placement-85dc486cc4-fwrm6\" (UID: \"e3ee069a-41fb-4cab-b650-9bb9c03ac271\") " pod="openstack/placement-85dc486cc4-fwrm6" Mar 13 14:27:59 crc kubenswrapper[4907]: I0313 14:27:59.583597 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gds2t\" (UniqueName: \"kubernetes.io/projected/e3ee069a-41fb-4cab-b650-9bb9c03ac271-kube-api-access-gds2t\") pod \"placement-85dc486cc4-fwrm6\" (UID: \"e3ee069a-41fb-4cab-b650-9bb9c03ac271\") " pod="openstack/placement-85dc486cc4-fwrm6" Mar 13 14:27:59 crc kubenswrapper[4907]: I0313 14:27:59.583807 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e3ee069a-41fb-4cab-b650-9bb9c03ac271-scripts\") pod \"placement-85dc486cc4-fwrm6\" (UID: \"e3ee069a-41fb-4cab-b650-9bb9c03ac271\") " pod="openstack/placement-85dc486cc4-fwrm6" Mar 13 14:27:59 crc kubenswrapper[4907]: I0313 14:27:59.583997 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3ee069a-41fb-4cab-b650-9bb9c03ac271-combined-ca-bundle\") pod \"placement-85dc486cc4-fwrm6\" (UID: \"e3ee069a-41fb-4cab-b650-9bb9c03ac271\") " pod="openstack/placement-85dc486cc4-fwrm6" Mar 13 14:27:59 crc kubenswrapper[4907]: I0313 14:27:59.584106 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e3ee069a-41fb-4cab-b650-9bb9c03ac271-public-tls-certs\") pod \"placement-85dc486cc4-fwrm6\" (UID: \"e3ee069a-41fb-4cab-b650-9bb9c03ac271\") " pod="openstack/placement-85dc486cc4-fwrm6" Mar 13 14:27:59 crc kubenswrapper[4907]: I0313 14:27:59.584404 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e3ee069a-41fb-4cab-b650-9bb9c03ac271-logs\") pod \"placement-85dc486cc4-fwrm6\" (UID: \"e3ee069a-41fb-4cab-b650-9bb9c03ac271\") " pod="openstack/placement-85dc486cc4-fwrm6" Mar 13 14:27:59 crc kubenswrapper[4907]: I0313 14:27:59.584949 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e3ee069a-41fb-4cab-b650-9bb9c03ac271-logs\") pod \"placement-85dc486cc4-fwrm6\" (UID: \"e3ee069a-41fb-4cab-b650-9bb9c03ac271\") " pod="openstack/placement-85dc486cc4-fwrm6" Mar 13 14:27:59 crc kubenswrapper[4907]: I0313 14:27:59.595434 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e3ee069a-41fb-4cab-b650-9bb9c03ac271-internal-tls-certs\") pod \"placement-85dc486cc4-fwrm6\" (UID: \"e3ee069a-41fb-4cab-b650-9bb9c03ac271\") " pod="openstack/placement-85dc486cc4-fwrm6" Mar 13 14:27:59 crc kubenswrapper[4907]: I0313 14:27:59.599086 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e3ee069a-41fb-4cab-b650-9bb9c03ac271-scripts\") pod \"placement-85dc486cc4-fwrm6\" (UID: \"e3ee069a-41fb-4cab-b650-9bb9c03ac271\") " pod="openstack/placement-85dc486cc4-fwrm6" Mar 13 14:27:59 crc kubenswrapper[4907]: I0313 14:27:59.599779 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3ee069a-41fb-4cab-b650-9bb9c03ac271-combined-ca-bundle\") pod \"placement-85dc486cc4-fwrm6\" (UID: \"e3ee069a-41fb-4cab-b650-9bb9c03ac271\") " pod="openstack/placement-85dc486cc4-fwrm6" Mar 13 14:27:59 crc kubenswrapper[4907]: I0313 14:27:59.601720 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3ee069a-41fb-4cab-b650-9bb9c03ac271-config-data\") pod \"placement-85dc486cc4-fwrm6\" (UID: \"e3ee069a-41fb-4cab-b650-9bb9c03ac271\") " pod="openstack/placement-85dc486cc4-fwrm6" Mar 13 14:27:59 crc kubenswrapper[4907]: I0313 14:27:59.612477 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e3ee069a-41fb-4cab-b650-9bb9c03ac271-public-tls-certs\") pod \"placement-85dc486cc4-fwrm6\" (UID: \"e3ee069a-41fb-4cab-b650-9bb9c03ac271\") " pod="openstack/placement-85dc486cc4-fwrm6" Mar 13 14:27:59 crc kubenswrapper[4907]: I0313 14:27:59.613650 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49bc69d8-fe85-4994-a5cf-c51472a124e9-kube-api-access-r77nl" (OuterVolumeSpecName: "kube-api-access-r77nl") pod "49bc69d8-fe85-4994-a5cf-c51472a124e9" (UID: "49bc69d8-fe85-4994-a5cf-c51472a124e9"). InnerVolumeSpecName "kube-api-access-r77nl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:27:59 crc kubenswrapper[4907]: I0313 14:27:59.644699 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gds2t\" (UniqueName: \"kubernetes.io/projected/e3ee069a-41fb-4cab-b650-9bb9c03ac271-kube-api-access-gds2t\") pod \"placement-85dc486cc4-fwrm6\" (UID: \"e3ee069a-41fb-4cab-b650-9bb9c03ac271\") " pod="openstack/placement-85dc486cc4-fwrm6" Mar 13 14:27:59 crc kubenswrapper[4907]: I0313 14:27:59.690085 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r77nl\" (UniqueName: \"kubernetes.io/projected/49bc69d8-fe85-4994-a5cf-c51472a124e9-kube-api-access-r77nl\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:59 crc kubenswrapper[4907]: I0313 14:27:59.739706 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-85dc486cc4-fwrm6" Mar 13 14:27:59 crc kubenswrapper[4907]: I0313 14:27:59.762721 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49bc69d8-fe85-4994-a5cf-c51472a124e9-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "49bc69d8-fe85-4994-a5cf-c51472a124e9" (UID: "49bc69d8-fe85-4994-a5cf-c51472a124e9"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:27:59 crc kubenswrapper[4907]: I0313 14:27:59.766827 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49bc69d8-fe85-4994-a5cf-c51472a124e9-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "49bc69d8-fe85-4994-a5cf-c51472a124e9" (UID: "49bc69d8-fe85-4994-a5cf-c51472a124e9"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:27:59 crc kubenswrapper[4907]: I0313 14:27:59.793106 4907 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/49bc69d8-fe85-4994-a5cf-c51472a124e9-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:59 crc kubenswrapper[4907]: I0313 14:27:59.793134 4907 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/49bc69d8-fe85-4994-a5cf-c51472a124e9-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:59 crc kubenswrapper[4907]: I0313 14:27:59.796674 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="032183a7-c5db-4528-a211-803050007b68" path="/var/lib/kubelet/pods/032183a7-c5db-4528-a211-803050007b68/volumes" Mar 13 14:27:59 crc kubenswrapper[4907]: I0313 14:27:59.806336 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49bc69d8-fe85-4994-a5cf-c51472a124e9-config" (OuterVolumeSpecName: "config") pod "49bc69d8-fe85-4994-a5cf-c51472a124e9" (UID: "49bc69d8-fe85-4994-a5cf-c51472a124e9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:27:59 crc kubenswrapper[4907]: I0313 14:27:59.812754 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49bc69d8-fe85-4994-a5cf-c51472a124e9-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "49bc69d8-fe85-4994-a5cf-c51472a124e9" (UID: "49bc69d8-fe85-4994-a5cf-c51472a124e9"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:27:59 crc kubenswrapper[4907]: I0313 14:27:59.823931 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49bc69d8-fe85-4994-a5cf-c51472a124e9-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "49bc69d8-fe85-4994-a5cf-c51472a124e9" (UID: "49bc69d8-fe85-4994-a5cf-c51472a124e9"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:27:59 crc kubenswrapper[4907]: I0313 14:27:59.907528 4907 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/49bc69d8-fe85-4994-a5cf-c51472a124e9-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:59 crc kubenswrapper[4907]: I0313 14:27:59.907800 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/49bc69d8-fe85-4994-a5cf-c51472a124e9-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:59 crc kubenswrapper[4907]: I0313 14:27:59.907812 4907 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/49bc69d8-fe85-4994-a5cf-c51472a124e9-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:00 crc kubenswrapper[4907]: I0313 14:28:00.050283 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-944cbc5bb-fj29g" Mar 13 14:28:00 crc kubenswrapper[4907]: I0313 14:28:00.143954 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556868-pzml6"] Mar 13 14:28:00 crc kubenswrapper[4907]: E0313 14:28:00.144453 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49bc69d8-fe85-4994-a5cf-c51472a124e9" containerName="init" Mar 13 14:28:00 crc kubenswrapper[4907]: I0313 14:28:00.144482 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="49bc69d8-fe85-4994-a5cf-c51472a124e9" containerName="init" Mar 13 14:28:00 crc kubenswrapper[4907]: E0313 14:28:00.144506 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49bc69d8-fe85-4994-a5cf-c51472a124e9" containerName="dnsmasq-dns" Mar 13 14:28:00 crc kubenswrapper[4907]: I0313 14:28:00.144516 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="49bc69d8-fe85-4994-a5cf-c51472a124e9" containerName="dnsmasq-dns" Mar 13 14:28:00 crc kubenswrapper[4907]: I0313 14:28:00.144788 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="49bc69d8-fe85-4994-a5cf-c51472a124e9" containerName="dnsmasq-dns" Mar 13 14:28:00 crc kubenswrapper[4907]: I0313 14:28:00.145622 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556868-pzml6" Mar 13 14:28:00 crc kubenswrapper[4907]: I0313 14:28:00.149063 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 14:28:00 crc kubenswrapper[4907]: I0313 14:28:00.149074 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 14:28:00 crc kubenswrapper[4907]: I0313 14:28:00.149433 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 14:28:00 crc kubenswrapper[4907]: I0313 14:28:00.151960 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556868-pzml6"] Mar 13 14:28:00 crc kubenswrapper[4907]: I0313 14:28:00.257577 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-85b9d64669-dxnf6" Mar 13 14:28:00 crc kubenswrapper[4907]: I0313 14:28:00.258362 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fc46d7df7-77dng" event={"ID":"49bc69d8-fe85-4994-a5cf-c51472a124e9","Type":"ContainerDied","Data":"74e19d820a0370c065d553cb0dbb43ed632fe5214db6fe0a8a5f25c3d63f84ae"} Mar 13 14:28:00 crc kubenswrapper[4907]: I0313 14:28:00.258416 4907 scope.go:117] "RemoveContainer" containerID="10817f472001c46a816f4b77aa0c7ef3e99533487d552c19bd292e7357506e9f" Mar 13 14:28:00 crc kubenswrapper[4907]: I0313 14:28:00.258427 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fc46d7df7-77dng" Mar 13 14:28:00 crc kubenswrapper[4907]: I0313 14:28:00.266737 4907 generic.go:334] "Generic (PLEG): container finished" podID="e7fd8554-798f-4e3a-acf1-8425c15e2665" containerID="2b64ad381c9b4854fe824a41af04da817955826edfb608941d06803d7f8f4d59" exitCode=0 Mar 13 14:28:00 crc kubenswrapper[4907]: I0313 14:28:00.266785 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e7fd8554-798f-4e3a-acf1-8425c15e2665","Type":"ContainerDied","Data":"2b64ad381c9b4854fe824a41af04da817955826edfb608941d06803d7f8f4d59"} Mar 13 14:28:00 crc kubenswrapper[4907]: I0313 14:28:00.301631 4907 scope.go:117] "RemoveContainer" containerID="577048fe225fd9933f67497a9e39e7a61a8a9a2ce395fd1bd05afbdcf17490ca" Mar 13 14:28:00 crc kubenswrapper[4907]: I0313 14:28:00.304368 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-85dc486cc4-fwrm6"] Mar 13 14:28:00 crc kubenswrapper[4907]: W0313 14:28:00.311356 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode3ee069a_41fb_4cab_b650_9bb9c03ac271.slice/crio-80317f4f67489584d7a194b5c3eced56f03150f3c1e4e1db7ac581e0ad05f710 WatchSource:0}: Error finding container 80317f4f67489584d7a194b5c3eced56f03150f3c1e4e1db7ac581e0ad05f710: Status 404 returned error can't find the container with id 80317f4f67489584d7a194b5c3eced56f03150f3c1e4e1db7ac581e0ad05f710 Mar 13 14:28:00 crc kubenswrapper[4907]: I0313 14:28:00.313315 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7fc46d7df7-77dng"] Mar 13 14:28:00 crc kubenswrapper[4907]: I0313 14:28:00.315583 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xmrsg\" (UniqueName: \"kubernetes.io/projected/6301088c-1715-4d57-a9b8-1ea9f7128560-kube-api-access-xmrsg\") pod \"auto-csr-approver-29556868-pzml6\" (UID: \"6301088c-1715-4d57-a9b8-1ea9f7128560\") " pod="openshift-infra/auto-csr-approver-29556868-pzml6" Mar 13 14:28:00 crc kubenswrapper[4907]: I0313 14:28:00.325814 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7fc46d7df7-77dng"] Mar 13 14:28:00 crc kubenswrapper[4907]: I0313 14:28:00.417557 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xmrsg\" (UniqueName: \"kubernetes.io/projected/6301088c-1715-4d57-a9b8-1ea9f7128560-kube-api-access-xmrsg\") pod \"auto-csr-approver-29556868-pzml6\" (UID: \"6301088c-1715-4d57-a9b8-1ea9f7128560\") " pod="openshift-infra/auto-csr-approver-29556868-pzml6" Mar 13 14:28:00 crc kubenswrapper[4907]: I0313 14:28:00.438440 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xmrsg\" (UniqueName: \"kubernetes.io/projected/6301088c-1715-4d57-a9b8-1ea9f7128560-kube-api-access-xmrsg\") pod \"auto-csr-approver-29556868-pzml6\" (UID: \"6301088c-1715-4d57-a9b8-1ea9f7128560\") " pod="openshift-infra/auto-csr-approver-29556868-pzml6" Mar 13 14:28:00 crc kubenswrapper[4907]: I0313 14:28:00.470108 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556868-pzml6" Mar 13 14:28:00 crc kubenswrapper[4907]: I0313 14:28:00.913761 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556868-pzml6"] Mar 13 14:28:01 crc kubenswrapper[4907]: I0313 14:28:01.280056 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556868-pzml6" event={"ID":"6301088c-1715-4d57-a9b8-1ea9f7128560","Type":"ContainerStarted","Data":"baf4195c37e7df43009ff885852df0c99df75e43454c0fbdc4b11a230d7af5e3"} Mar 13 14:28:01 crc kubenswrapper[4907]: I0313 14:28:01.282410 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-85dc486cc4-fwrm6" event={"ID":"e3ee069a-41fb-4cab-b650-9bb9c03ac271","Type":"ContainerStarted","Data":"6d08d480e1be3a68c3d20d884244fa154245d26dcb3fb269a2aeb6806364407e"} Mar 13 14:28:01 crc kubenswrapper[4907]: I0313 14:28:01.282436 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-85dc486cc4-fwrm6" event={"ID":"e3ee069a-41fb-4cab-b650-9bb9c03ac271","Type":"ContainerStarted","Data":"c0c472fe526cce04d2912530f4de0e17fa4ad4be341dbd2a462b13899fc1ad89"} Mar 13 14:28:01 crc kubenswrapper[4907]: I0313 14:28:01.282451 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-85dc486cc4-fwrm6" event={"ID":"e3ee069a-41fb-4cab-b650-9bb9c03ac271","Type":"ContainerStarted","Data":"80317f4f67489584d7a194b5c3eced56f03150f3c1e4e1db7ac581e0ad05f710"} Mar 13 14:28:01 crc kubenswrapper[4907]: I0313 14:28:01.283765 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-85dc486cc4-fwrm6" Mar 13 14:28:01 crc kubenswrapper[4907]: I0313 14:28:01.283797 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-85dc486cc4-fwrm6" Mar 13 14:28:01 crc kubenswrapper[4907]: I0313 14:28:01.310076 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-85dc486cc4-fwrm6" podStartSLOduration=2.310056751 podStartE2EDuration="2.310056751s" podCreationTimestamp="2026-03-13 14:27:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:28:01.303869012 +0000 UTC m=+1380.203656701" watchObservedRunningTime="2026-03-13 14:28:01.310056751 +0000 UTC m=+1380.209844440" Mar 13 14:28:01 crc kubenswrapper[4907]: I0313 14:28:01.793505 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49bc69d8-fe85-4994-a5cf-c51472a124e9" path="/var/lib/kubelet/pods/49bc69d8-fe85-4994-a5cf-c51472a124e9/volumes" Mar 13 14:28:02 crc kubenswrapper[4907]: I0313 14:28:02.294749 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556868-pzml6" event={"ID":"6301088c-1715-4d57-a9b8-1ea9f7128560","Type":"ContainerStarted","Data":"52f2356a35bf1ec5bd881f141be785ea7a9d11be6b6e8da92a08cbad79935d8d"} Mar 13 14:28:02 crc kubenswrapper[4907]: I0313 14:28:02.313181 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29556868-pzml6" podStartSLOduration=1.452517895 podStartE2EDuration="2.313161092s" podCreationTimestamp="2026-03-13 14:28:00 +0000 UTC" firstStartedPulling="2026-03-13 14:28:00.918018248 +0000 UTC m=+1379.817805937" lastFinishedPulling="2026-03-13 14:28:01.778661445 +0000 UTC m=+1380.678449134" observedRunningTime="2026-03-13 14:28:02.305619977 +0000 UTC m=+1381.205407666" watchObservedRunningTime="2026-03-13 14:28:02.313161092 +0000 UTC m=+1381.212948781" Mar 13 14:28:02 crc kubenswrapper[4907]: I0313 14:28:02.649011 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Mar 13 14:28:02 crc kubenswrapper[4907]: I0313 14:28:02.651825 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Mar 13 14:28:02 crc kubenswrapper[4907]: I0313 14:28:02.655606 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Mar 13 14:28:02 crc kubenswrapper[4907]: I0313 14:28:02.655733 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-c7qcz" Mar 13 14:28:02 crc kubenswrapper[4907]: I0313 14:28:02.655622 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Mar 13 14:28:02 crc kubenswrapper[4907]: I0313 14:28:02.688086 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Mar 13 14:28:02 crc kubenswrapper[4907]: I0313 14:28:02.762031 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/87fb1bb3-8511-4d25-bf36-7ba0a278e101-openstack-config-secret\") pod \"openstackclient\" (UID: \"87fb1bb3-8511-4d25-bf36-7ba0a278e101\") " pod="openstack/openstackclient" Mar 13 14:28:02 crc kubenswrapper[4907]: I0313 14:28:02.762107 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c9hts\" (UniqueName: \"kubernetes.io/projected/87fb1bb3-8511-4d25-bf36-7ba0a278e101-kube-api-access-c9hts\") pod \"openstackclient\" (UID: \"87fb1bb3-8511-4d25-bf36-7ba0a278e101\") " pod="openstack/openstackclient" Mar 13 14:28:02 crc kubenswrapper[4907]: I0313 14:28:02.762404 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/87fb1bb3-8511-4d25-bf36-7ba0a278e101-openstack-config\") pod \"openstackclient\" (UID: \"87fb1bb3-8511-4d25-bf36-7ba0a278e101\") " pod="openstack/openstackclient" Mar 13 14:28:02 crc kubenswrapper[4907]: I0313 14:28:02.762497 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87fb1bb3-8511-4d25-bf36-7ba0a278e101-combined-ca-bundle\") pod \"openstackclient\" (UID: \"87fb1bb3-8511-4d25-bf36-7ba0a278e101\") " pod="openstack/openstackclient" Mar 13 14:28:02 crc kubenswrapper[4907]: I0313 14:28:02.864423 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/87fb1bb3-8511-4d25-bf36-7ba0a278e101-openstack-config\") pod \"openstackclient\" (UID: \"87fb1bb3-8511-4d25-bf36-7ba0a278e101\") " pod="openstack/openstackclient" Mar 13 14:28:02 crc kubenswrapper[4907]: I0313 14:28:02.864478 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87fb1bb3-8511-4d25-bf36-7ba0a278e101-combined-ca-bundle\") pod \"openstackclient\" (UID: \"87fb1bb3-8511-4d25-bf36-7ba0a278e101\") " pod="openstack/openstackclient" Mar 13 14:28:02 crc kubenswrapper[4907]: I0313 14:28:02.864665 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/87fb1bb3-8511-4d25-bf36-7ba0a278e101-openstack-config-secret\") pod \"openstackclient\" (UID: \"87fb1bb3-8511-4d25-bf36-7ba0a278e101\") " pod="openstack/openstackclient" Mar 13 14:28:02 crc kubenswrapper[4907]: I0313 14:28:02.864698 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c9hts\" (UniqueName: \"kubernetes.io/projected/87fb1bb3-8511-4d25-bf36-7ba0a278e101-kube-api-access-c9hts\") pod \"openstackclient\" (UID: \"87fb1bb3-8511-4d25-bf36-7ba0a278e101\") " pod="openstack/openstackclient" Mar 13 14:28:02 crc kubenswrapper[4907]: I0313 14:28:02.866989 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/87fb1bb3-8511-4d25-bf36-7ba0a278e101-openstack-config\") pod \"openstackclient\" (UID: \"87fb1bb3-8511-4d25-bf36-7ba0a278e101\") " pod="openstack/openstackclient" Mar 13 14:28:02 crc kubenswrapper[4907]: I0313 14:28:02.873150 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/87fb1bb3-8511-4d25-bf36-7ba0a278e101-openstack-config-secret\") pod \"openstackclient\" (UID: \"87fb1bb3-8511-4d25-bf36-7ba0a278e101\") " pod="openstack/openstackclient" Mar 13 14:28:02 crc kubenswrapper[4907]: I0313 14:28:02.885412 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87fb1bb3-8511-4d25-bf36-7ba0a278e101-combined-ca-bundle\") pod \"openstackclient\" (UID: \"87fb1bb3-8511-4d25-bf36-7ba0a278e101\") " pod="openstack/openstackclient" Mar 13 14:28:02 crc kubenswrapper[4907]: I0313 14:28:02.889569 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c9hts\" (UniqueName: \"kubernetes.io/projected/87fb1bb3-8511-4d25-bf36-7ba0a278e101-kube-api-access-c9hts\") pod \"openstackclient\" (UID: \"87fb1bb3-8511-4d25-bf36-7ba0a278e101\") " pod="openstack/openstackclient" Mar 13 14:28:02 crc kubenswrapper[4907]: I0313 14:28:02.951964 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Mar 13 14:28:02 crc kubenswrapper[4907]: I0313 14:28:02.952767 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Mar 13 14:28:02 crc kubenswrapper[4907]: I0313 14:28:02.974557 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Mar 13 14:28:02 crc kubenswrapper[4907]: I0313 14:28:02.987479 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Mar 13 14:28:02 crc kubenswrapper[4907]: I0313 14:28:02.988952 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Mar 13 14:28:02 crc kubenswrapper[4907]: I0313 14:28:02.997508 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Mar 13 14:28:03 crc kubenswrapper[4907]: E0313 14:28:03.104422 4907 log.go:32] "RunPodSandbox from runtime service failed" err=< Mar 13 14:28:03 crc kubenswrapper[4907]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openstackclient_openstack_87fb1bb3-8511-4d25-bf36-7ba0a278e101_0(90a3d91a8b93297c4e43bf3c51abc0a220d77d20297eda06a993ae96cc8b4616): error adding pod openstack_openstackclient to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"90a3d91a8b93297c4e43bf3c51abc0a220d77d20297eda06a993ae96cc8b4616" Netns:"/var/run/netns/5e0c96e2-3668-4751-a7e4-6ac9d06b067f" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openstack;K8S_POD_NAME=openstackclient;K8S_POD_INFRA_CONTAINER_ID=90a3d91a8b93297c4e43bf3c51abc0a220d77d20297eda06a993ae96cc8b4616;K8S_POD_UID=87fb1bb3-8511-4d25-bf36-7ba0a278e101" Path:"" ERRORED: error configuring pod [openstack/openstackclient] networking: Multus: [openstack/openstackclient/87fb1bb3-8511-4d25-bf36-7ba0a278e101]: expected pod UID "87fb1bb3-8511-4d25-bf36-7ba0a278e101" but got "a055ce7f-0538-4e5e-938a-eaf6d0aa2dec" from Kube API Mar 13 14:28:03 crc kubenswrapper[4907]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Mar 13 14:28:03 crc kubenswrapper[4907]: > Mar 13 14:28:03 crc kubenswrapper[4907]: E0313 14:28:03.104807 4907 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err=< Mar 13 14:28:03 crc kubenswrapper[4907]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openstackclient_openstack_87fb1bb3-8511-4d25-bf36-7ba0a278e101_0(90a3d91a8b93297c4e43bf3c51abc0a220d77d20297eda06a993ae96cc8b4616): error adding pod openstack_openstackclient to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"90a3d91a8b93297c4e43bf3c51abc0a220d77d20297eda06a993ae96cc8b4616" Netns:"/var/run/netns/5e0c96e2-3668-4751-a7e4-6ac9d06b067f" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openstack;K8S_POD_NAME=openstackclient;K8S_POD_INFRA_CONTAINER_ID=90a3d91a8b93297c4e43bf3c51abc0a220d77d20297eda06a993ae96cc8b4616;K8S_POD_UID=87fb1bb3-8511-4d25-bf36-7ba0a278e101" Path:"" ERRORED: error configuring pod [openstack/openstackclient] networking: Multus: [openstack/openstackclient/87fb1bb3-8511-4d25-bf36-7ba0a278e101]: expected pod UID "87fb1bb3-8511-4d25-bf36-7ba0a278e101" but got "a055ce7f-0538-4e5e-938a-eaf6d0aa2dec" from Kube API Mar 13 14:28:03 crc kubenswrapper[4907]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Mar 13 14:28:03 crc kubenswrapper[4907]: > pod="openstack/openstackclient" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.126803 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.170250 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/a055ce7f-0538-4e5e-938a-eaf6d0aa2dec-openstack-config-secret\") pod \"openstackclient\" (UID: \"a055ce7f-0538-4e5e-938a-eaf6d0aa2dec\") " pod="openstack/openstackclient" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.170297 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a055ce7f-0538-4e5e-938a-eaf6d0aa2dec-combined-ca-bundle\") pod \"openstackclient\" (UID: \"a055ce7f-0538-4e5e-938a-eaf6d0aa2dec\") " pod="openstack/openstackclient" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.170340 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/a055ce7f-0538-4e5e-938a-eaf6d0aa2dec-openstack-config\") pod \"openstackclient\" (UID: \"a055ce7f-0538-4e5e-938a-eaf6d0aa2dec\") " pod="openstack/openstackclient" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.170422 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fk82k\" (UniqueName: \"kubernetes.io/projected/a055ce7f-0538-4e5e-938a-eaf6d0aa2dec-kube-api-access-fk82k\") pod \"openstackclient\" (UID: \"a055ce7f-0538-4e5e-938a-eaf6d0aa2dec\") " pod="openstack/openstackclient" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.271914 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7fd8554-798f-4e3a-acf1-8425c15e2665-combined-ca-bundle\") pod \"e7fd8554-798f-4e3a-acf1-8425c15e2665\" (UID: \"e7fd8554-798f-4e3a-acf1-8425c15e2665\") " Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.272005 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r9xxm\" (UniqueName: \"kubernetes.io/projected/e7fd8554-798f-4e3a-acf1-8425c15e2665-kube-api-access-r9xxm\") pod \"e7fd8554-798f-4e3a-acf1-8425c15e2665\" (UID: \"e7fd8554-798f-4e3a-acf1-8425c15e2665\") " Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.272071 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e7fd8554-798f-4e3a-acf1-8425c15e2665-config-data-custom\") pod \"e7fd8554-798f-4e3a-acf1-8425c15e2665\" (UID: \"e7fd8554-798f-4e3a-acf1-8425c15e2665\") " Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.272195 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e7fd8554-798f-4e3a-acf1-8425c15e2665-etc-machine-id\") pod \"e7fd8554-798f-4e3a-acf1-8425c15e2665\" (UID: \"e7fd8554-798f-4e3a-acf1-8425c15e2665\") " Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.272290 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e7fd8554-798f-4e3a-acf1-8425c15e2665-config-data\") pod \"e7fd8554-798f-4e3a-acf1-8425c15e2665\" (UID: \"e7fd8554-798f-4e3a-acf1-8425c15e2665\") " Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.272320 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e7fd8554-798f-4e3a-acf1-8425c15e2665-scripts\") pod \"e7fd8554-798f-4e3a-acf1-8425c15e2665\" (UID: \"e7fd8554-798f-4e3a-acf1-8425c15e2665\") " Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.272619 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fk82k\" (UniqueName: \"kubernetes.io/projected/a055ce7f-0538-4e5e-938a-eaf6d0aa2dec-kube-api-access-fk82k\") pod \"openstackclient\" (UID: \"a055ce7f-0538-4e5e-938a-eaf6d0aa2dec\") " pod="openstack/openstackclient" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.272764 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e7fd8554-798f-4e3a-acf1-8425c15e2665-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "e7fd8554-798f-4e3a-acf1-8425c15e2665" (UID: "e7fd8554-798f-4e3a-acf1-8425c15e2665"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.274332 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/a055ce7f-0538-4e5e-938a-eaf6d0aa2dec-openstack-config-secret\") pod \"openstackclient\" (UID: \"a055ce7f-0538-4e5e-938a-eaf6d0aa2dec\") " pod="openstack/openstackclient" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.274441 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a055ce7f-0538-4e5e-938a-eaf6d0aa2dec-combined-ca-bundle\") pod \"openstackclient\" (UID: \"a055ce7f-0538-4e5e-938a-eaf6d0aa2dec\") " pod="openstack/openstackclient" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.274561 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/a055ce7f-0538-4e5e-938a-eaf6d0aa2dec-openstack-config\") pod \"openstackclient\" (UID: \"a055ce7f-0538-4e5e-938a-eaf6d0aa2dec\") " pod="openstack/openstackclient" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.278392 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/a055ce7f-0538-4e5e-938a-eaf6d0aa2dec-openstack-config-secret\") pod \"openstackclient\" (UID: \"a055ce7f-0538-4e5e-938a-eaf6d0aa2dec\") " pod="openstack/openstackclient" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.278714 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7fd8554-798f-4e3a-acf1-8425c15e2665-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "e7fd8554-798f-4e3a-acf1-8425c15e2665" (UID: "e7fd8554-798f-4e3a-acf1-8425c15e2665"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.280019 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7fd8554-798f-4e3a-acf1-8425c15e2665-kube-api-access-r9xxm" (OuterVolumeSpecName: "kube-api-access-r9xxm") pod "e7fd8554-798f-4e3a-acf1-8425c15e2665" (UID: "e7fd8554-798f-4e3a-acf1-8425c15e2665"). InnerVolumeSpecName "kube-api-access-r9xxm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.281122 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a055ce7f-0538-4e5e-938a-eaf6d0aa2dec-combined-ca-bundle\") pod \"openstackclient\" (UID: \"a055ce7f-0538-4e5e-938a-eaf6d0aa2dec\") " pod="openstack/openstackclient" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.283988 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7fd8554-798f-4e3a-acf1-8425c15e2665-scripts" (OuterVolumeSpecName: "scripts") pod "e7fd8554-798f-4e3a-acf1-8425c15e2665" (UID: "e7fd8554-798f-4e3a-acf1-8425c15e2665"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.284082 4907 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e7fd8554-798f-4e3a-acf1-8425c15e2665-etc-machine-id\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.284763 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/a055ce7f-0538-4e5e-938a-eaf6d0aa2dec-openstack-config\") pod \"openstackclient\" (UID: \"a055ce7f-0538-4e5e-938a-eaf6d0aa2dec\") " pod="openstack/openstackclient" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.307368 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fk82k\" (UniqueName: \"kubernetes.io/projected/a055ce7f-0538-4e5e-938a-eaf6d0aa2dec-kube-api-access-fk82k\") pod \"openstackclient\" (UID: \"a055ce7f-0538-4e5e-938a-eaf6d0aa2dec\") " pod="openstack/openstackclient" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.309693 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.319020 4907 generic.go:334] "Generic (PLEG): container finished" podID="e7fd8554-798f-4e3a-acf1-8425c15e2665" containerID="0a62f5f3cb31fd39481383f9dca4b2b20b3144ace647921ed05a9b4b9214a38d" exitCode=0 Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.319108 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e7fd8554-798f-4e3a-acf1-8425c15e2665","Type":"ContainerDied","Data":"0a62f5f3cb31fd39481383f9dca4b2b20b3144ace647921ed05a9b4b9214a38d"} Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.319140 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e7fd8554-798f-4e3a-acf1-8425c15e2665","Type":"ContainerDied","Data":"58e8239ab4329277f0169cd2f4714646b1acec175437b4fe78554e7e213348e5"} Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.319163 4907 scope.go:117] "RemoveContainer" containerID="2b64ad381c9b4854fe824a41af04da817955826edfb608941d06803d7f8f4d59" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.319320 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.356153 4907 generic.go:334] "Generic (PLEG): container finished" podID="6301088c-1715-4d57-a9b8-1ea9f7128560" containerID="52f2356a35bf1ec5bd881f141be785ea7a9d11be6b6e8da92a08cbad79935d8d" exitCode=0 Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.356928 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556868-pzml6" event={"ID":"6301088c-1715-4d57-a9b8-1ea9f7128560","Type":"ContainerDied","Data":"52f2356a35bf1ec5bd881f141be785ea7a9d11be6b6e8da92a08cbad79935d8d"} Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.357023 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.380390 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.387467 4907 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e7fd8554-798f-4e3a-acf1-8425c15e2665-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.387501 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r9xxm\" (UniqueName: \"kubernetes.io/projected/e7fd8554-798f-4e3a-acf1-8425c15e2665-kube-api-access-r9xxm\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.387514 4907 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e7fd8554-798f-4e3a-acf1-8425c15e2665-config-data-custom\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.395494 4907 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="87fb1bb3-8511-4d25-bf36-7ba0a278e101" podUID="a055ce7f-0538-4e5e-938a-eaf6d0aa2dec" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.397085 4907 scope.go:117] "RemoveContainer" containerID="0a62f5f3cb31fd39481383f9dca4b2b20b3144ace647921ed05a9b4b9214a38d" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.404031 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7fd8554-798f-4e3a-acf1-8425c15e2665-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e7fd8554-798f-4e3a-acf1-8425c15e2665" (UID: "e7fd8554-798f-4e3a-acf1-8425c15e2665"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.448555 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7fd8554-798f-4e3a-acf1-8425c15e2665-config-data" (OuterVolumeSpecName: "config-data") pod "e7fd8554-798f-4e3a-acf1-8425c15e2665" (UID: "e7fd8554-798f-4e3a-acf1-8425c15e2665"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.488396 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87fb1bb3-8511-4d25-bf36-7ba0a278e101-combined-ca-bundle\") pod \"87fb1bb3-8511-4d25-bf36-7ba0a278e101\" (UID: \"87fb1bb3-8511-4d25-bf36-7ba0a278e101\") " Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.488541 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/87fb1bb3-8511-4d25-bf36-7ba0a278e101-openstack-config\") pod \"87fb1bb3-8511-4d25-bf36-7ba0a278e101\" (UID: \"87fb1bb3-8511-4d25-bf36-7ba0a278e101\") " Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.488604 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c9hts\" (UniqueName: \"kubernetes.io/projected/87fb1bb3-8511-4d25-bf36-7ba0a278e101-kube-api-access-c9hts\") pod \"87fb1bb3-8511-4d25-bf36-7ba0a278e101\" (UID: \"87fb1bb3-8511-4d25-bf36-7ba0a278e101\") " Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.488641 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/87fb1bb3-8511-4d25-bf36-7ba0a278e101-openstack-config-secret\") pod \"87fb1bb3-8511-4d25-bf36-7ba0a278e101\" (UID: \"87fb1bb3-8511-4d25-bf36-7ba0a278e101\") " Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.489217 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7fd8554-798f-4e3a-acf1-8425c15e2665-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.489238 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e7fd8554-798f-4e3a-acf1-8425c15e2665-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.489499 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87fb1bb3-8511-4d25-bf36-7ba0a278e101-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "87fb1bb3-8511-4d25-bf36-7ba0a278e101" (UID: "87fb1bb3-8511-4d25-bf36-7ba0a278e101"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.493803 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87fb1bb3-8511-4d25-bf36-7ba0a278e101-kube-api-access-c9hts" (OuterVolumeSpecName: "kube-api-access-c9hts") pod "87fb1bb3-8511-4d25-bf36-7ba0a278e101" (UID: "87fb1bb3-8511-4d25-bf36-7ba0a278e101"). InnerVolumeSpecName "kube-api-access-c9hts". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.495673 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87fb1bb3-8511-4d25-bf36-7ba0a278e101-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "87fb1bb3-8511-4d25-bf36-7ba0a278e101" (UID: "87fb1bb3-8511-4d25-bf36-7ba0a278e101"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.495950 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87fb1bb3-8511-4d25-bf36-7ba0a278e101-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "87fb1bb3-8511-4d25-bf36-7ba0a278e101" (UID: "87fb1bb3-8511-4d25-bf36-7ba0a278e101"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.498574 4907 scope.go:117] "RemoveContainer" containerID="2b64ad381c9b4854fe824a41af04da817955826edfb608941d06803d7f8f4d59" Mar 13 14:28:03 crc kubenswrapper[4907]: E0313 14:28:03.500896 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2b64ad381c9b4854fe824a41af04da817955826edfb608941d06803d7f8f4d59\": container with ID starting with 2b64ad381c9b4854fe824a41af04da817955826edfb608941d06803d7f8f4d59 not found: ID does not exist" containerID="2b64ad381c9b4854fe824a41af04da817955826edfb608941d06803d7f8f4d59" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.500937 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b64ad381c9b4854fe824a41af04da817955826edfb608941d06803d7f8f4d59"} err="failed to get container status \"2b64ad381c9b4854fe824a41af04da817955826edfb608941d06803d7f8f4d59\": rpc error: code = NotFound desc = could not find container \"2b64ad381c9b4854fe824a41af04da817955826edfb608941d06803d7f8f4d59\": container with ID starting with 2b64ad381c9b4854fe824a41af04da817955826edfb608941d06803d7f8f4d59 not found: ID does not exist" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.500959 4907 scope.go:117] "RemoveContainer" containerID="0a62f5f3cb31fd39481383f9dca4b2b20b3144ace647921ed05a9b4b9214a38d" Mar 13 14:28:03 crc kubenswrapper[4907]: E0313 14:28:03.502458 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0a62f5f3cb31fd39481383f9dca4b2b20b3144ace647921ed05a9b4b9214a38d\": container with ID starting with 0a62f5f3cb31fd39481383f9dca4b2b20b3144ace647921ed05a9b4b9214a38d not found: ID does not exist" containerID="0a62f5f3cb31fd39481383f9dca4b2b20b3144ace647921ed05a9b4b9214a38d" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.502486 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0a62f5f3cb31fd39481383f9dca4b2b20b3144ace647921ed05a9b4b9214a38d"} err="failed to get container status \"0a62f5f3cb31fd39481383f9dca4b2b20b3144ace647921ed05a9b4b9214a38d\": rpc error: code = NotFound desc = could not find container \"0a62f5f3cb31fd39481383f9dca4b2b20b3144ace647921ed05a9b4b9214a38d\": container with ID starting with 0a62f5f3cb31fd39481383f9dca4b2b20b3144ace647921ed05a9b4b9214a38d not found: ID does not exist" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.592328 4907 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/87fb1bb3-8511-4d25-bf36-7ba0a278e101-openstack-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.592948 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c9hts\" (UniqueName: \"kubernetes.io/projected/87fb1bb3-8511-4d25-bf36-7ba0a278e101-kube-api-access-c9hts\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.592974 4907 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/87fb1bb3-8511-4d25-bf36-7ba0a278e101-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.592988 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87fb1bb3-8511-4d25-bf36-7ba0a278e101-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.658322 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.672218 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.681123 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Mar 13 14:28:03 crc kubenswrapper[4907]: E0313 14:28:03.681498 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7fd8554-798f-4e3a-acf1-8425c15e2665" containerName="cinder-scheduler" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.681515 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7fd8554-798f-4e3a-acf1-8425c15e2665" containerName="cinder-scheduler" Mar 13 14:28:03 crc kubenswrapper[4907]: E0313 14:28:03.681522 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7fd8554-798f-4e3a-acf1-8425c15e2665" containerName="probe" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.681528 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7fd8554-798f-4e3a-acf1-8425c15e2665" containerName="probe" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.681730 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="e7fd8554-798f-4e3a-acf1-8425c15e2665" containerName="cinder-scheduler" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.681753 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="e7fd8554-798f-4e3a-acf1-8425c15e2665" containerName="probe" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.682949 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.685446 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.711117 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.797373 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3793f53a-ca36-4b27-8444-d6dbfd860424-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"3793f53a-ca36-4b27-8444-d6dbfd860424\") " pod="openstack/cinder-scheduler-0" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.797457 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3793f53a-ca36-4b27-8444-d6dbfd860424-config-data\") pod \"cinder-scheduler-0\" (UID: \"3793f53a-ca36-4b27-8444-d6dbfd860424\") " pod="openstack/cinder-scheduler-0" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.797511 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3793f53a-ca36-4b27-8444-d6dbfd860424-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"3793f53a-ca36-4b27-8444-d6dbfd860424\") " pod="openstack/cinder-scheduler-0" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.797544 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3793f53a-ca36-4b27-8444-d6dbfd860424-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"3793f53a-ca36-4b27-8444-d6dbfd860424\") " pod="openstack/cinder-scheduler-0" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.797578 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8gkl4\" (UniqueName: \"kubernetes.io/projected/3793f53a-ca36-4b27-8444-d6dbfd860424-kube-api-access-8gkl4\") pod \"cinder-scheduler-0\" (UID: \"3793f53a-ca36-4b27-8444-d6dbfd860424\") " pod="openstack/cinder-scheduler-0" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.797636 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3793f53a-ca36-4b27-8444-d6dbfd860424-scripts\") pod \"cinder-scheduler-0\" (UID: \"3793f53a-ca36-4b27-8444-d6dbfd860424\") " pod="openstack/cinder-scheduler-0" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.803050 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87fb1bb3-8511-4d25-bf36-7ba0a278e101" path="/var/lib/kubelet/pods/87fb1bb3-8511-4d25-bf36-7ba0a278e101/volumes" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.803510 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7fd8554-798f-4e3a-acf1-8425c15e2665" path="/var/lib/kubelet/pods/e7fd8554-798f-4e3a-acf1-8425c15e2665/volumes" Mar 13 14:28:03 crc kubenswrapper[4907]: E0313 14:28:03.888646 4907 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode7fd8554_798f_4e3a_acf1_8425c15e2665.slice/crio-58e8239ab4329277f0169cd2f4714646b1acec175437b4fe78554e7e213348e5\": RecentStats: unable to find data in memory cache]" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.898762 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.899895 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3793f53a-ca36-4b27-8444-d6dbfd860424-scripts\") pod \"cinder-scheduler-0\" (UID: \"3793f53a-ca36-4b27-8444-d6dbfd860424\") " pod="openstack/cinder-scheduler-0" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.901303 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3793f53a-ca36-4b27-8444-d6dbfd860424-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"3793f53a-ca36-4b27-8444-d6dbfd860424\") " pod="openstack/cinder-scheduler-0" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.901442 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3793f53a-ca36-4b27-8444-d6dbfd860424-config-data\") pod \"cinder-scheduler-0\" (UID: \"3793f53a-ca36-4b27-8444-d6dbfd860424\") " pod="openstack/cinder-scheduler-0" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.901560 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3793f53a-ca36-4b27-8444-d6dbfd860424-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"3793f53a-ca36-4b27-8444-d6dbfd860424\") " pod="openstack/cinder-scheduler-0" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.902315 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3793f53a-ca36-4b27-8444-d6dbfd860424-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"3793f53a-ca36-4b27-8444-d6dbfd860424\") " pod="openstack/cinder-scheduler-0" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.902459 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8gkl4\" (UniqueName: \"kubernetes.io/projected/3793f53a-ca36-4b27-8444-d6dbfd860424-kube-api-access-8gkl4\") pod \"cinder-scheduler-0\" (UID: \"3793f53a-ca36-4b27-8444-d6dbfd860424\") " pod="openstack/cinder-scheduler-0" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.902635 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3793f53a-ca36-4b27-8444-d6dbfd860424-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"3793f53a-ca36-4b27-8444-d6dbfd860424\") " pod="openstack/cinder-scheduler-0" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.906454 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3793f53a-ca36-4b27-8444-d6dbfd860424-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"3793f53a-ca36-4b27-8444-d6dbfd860424\") " pod="openstack/cinder-scheduler-0" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.908117 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3793f53a-ca36-4b27-8444-d6dbfd860424-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"3793f53a-ca36-4b27-8444-d6dbfd860424\") " pod="openstack/cinder-scheduler-0" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.908175 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3793f53a-ca36-4b27-8444-d6dbfd860424-config-data\") pod \"cinder-scheduler-0\" (UID: \"3793f53a-ca36-4b27-8444-d6dbfd860424\") " pod="openstack/cinder-scheduler-0" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.911756 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3793f53a-ca36-4b27-8444-d6dbfd860424-scripts\") pod \"cinder-scheduler-0\" (UID: \"3793f53a-ca36-4b27-8444-d6dbfd860424\") " pod="openstack/cinder-scheduler-0" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.922199 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8gkl4\" (UniqueName: \"kubernetes.io/projected/3793f53a-ca36-4b27-8444-d6dbfd860424-kube-api-access-8gkl4\") pod \"cinder-scheduler-0\" (UID: \"3793f53a-ca36-4b27-8444-d6dbfd860424\") " pod="openstack/cinder-scheduler-0" Mar 13 14:28:04 crc kubenswrapper[4907]: I0313 14:28:04.012150 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Mar 13 14:28:04 crc kubenswrapper[4907]: I0313 14:28:04.231204 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-7bf587b8d9-pvvbz" Mar 13 14:28:04 crc kubenswrapper[4907]: I0313 14:28:04.320979 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-944cbc5bb-fj29g"] Mar 13 14:28:04 crc kubenswrapper[4907]: I0313 14:28:04.321359 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-944cbc5bb-fj29g" podUID="10a74a61-272c-4c54-9b2a-6379b77c9984" containerName="neutron-api" containerID="cri-o://a7d63138378c0ce32dc24735fd0ab7db352755405ba476da24cb2065858f3b75" gracePeriod=30 Mar 13 14:28:04 crc kubenswrapper[4907]: I0313 14:28:04.321947 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-944cbc5bb-fj29g" podUID="10a74a61-272c-4c54-9b2a-6379b77c9984" containerName="neutron-httpd" containerID="cri-o://b624bd9d00cbcadf5cdc9520721c5131708177fac0d6006945f6f3ffc27df6cd" gracePeriod=30 Mar 13 14:28:04 crc kubenswrapper[4907]: I0313 14:28:04.375390 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"a055ce7f-0538-4e5e-938a-eaf6d0aa2dec","Type":"ContainerStarted","Data":"7886c1c9bb89142f6b353653e6db0031015d0af7793820ba26eb48966adcdb36"} Mar 13 14:28:04 crc kubenswrapper[4907]: I0313 14:28:04.379285 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Mar 13 14:28:04 crc kubenswrapper[4907]: I0313 14:28:04.399680 4907 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="87fb1bb3-8511-4d25-bf36-7ba0a278e101" podUID="a055ce7f-0538-4e5e-938a-eaf6d0aa2dec" Mar 13 14:28:04 crc kubenswrapper[4907]: I0313 14:28:04.489518 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Mar 13 14:28:04 crc kubenswrapper[4907]: I0313 14:28:04.750274 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556868-pzml6" Mar 13 14:28:04 crc kubenswrapper[4907]: I0313 14:28:04.886327 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556862-sczxs"] Mar 13 14:28:04 crc kubenswrapper[4907]: I0313 14:28:04.895800 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556862-sczxs"] Mar 13 14:28:04 crc kubenswrapper[4907]: I0313 14:28:04.923035 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xmrsg\" (UniqueName: \"kubernetes.io/projected/6301088c-1715-4d57-a9b8-1ea9f7128560-kube-api-access-xmrsg\") pod \"6301088c-1715-4d57-a9b8-1ea9f7128560\" (UID: \"6301088c-1715-4d57-a9b8-1ea9f7128560\") " Mar 13 14:28:04 crc kubenswrapper[4907]: I0313 14:28:04.928864 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6301088c-1715-4d57-a9b8-1ea9f7128560-kube-api-access-xmrsg" (OuterVolumeSpecName: "kube-api-access-xmrsg") pod "6301088c-1715-4d57-a9b8-1ea9f7128560" (UID: "6301088c-1715-4d57-a9b8-1ea9f7128560"). InnerVolumeSpecName "kube-api-access-xmrsg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:28:05 crc kubenswrapper[4907]: I0313 14:28:05.026129 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xmrsg\" (UniqueName: \"kubernetes.io/projected/6301088c-1715-4d57-a9b8-1ea9f7128560-kube-api-access-xmrsg\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:05 crc kubenswrapper[4907]: I0313 14:28:05.399250 4907 generic.go:334] "Generic (PLEG): container finished" podID="10a74a61-272c-4c54-9b2a-6379b77c9984" containerID="b624bd9d00cbcadf5cdc9520721c5131708177fac0d6006945f6f3ffc27df6cd" exitCode=0 Mar 13 14:28:05 crc kubenswrapper[4907]: I0313 14:28:05.399625 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-944cbc5bb-fj29g" event={"ID":"10a74a61-272c-4c54-9b2a-6379b77c9984","Type":"ContainerDied","Data":"b624bd9d00cbcadf5cdc9520721c5131708177fac0d6006945f6f3ffc27df6cd"} Mar 13 14:28:05 crc kubenswrapper[4907]: I0313 14:28:05.403845 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556868-pzml6" event={"ID":"6301088c-1715-4d57-a9b8-1ea9f7128560","Type":"ContainerDied","Data":"baf4195c37e7df43009ff885852df0c99df75e43454c0fbdc4b11a230d7af5e3"} Mar 13 14:28:05 crc kubenswrapper[4907]: I0313 14:28:05.403991 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="baf4195c37e7df43009ff885852df0c99df75e43454c0fbdc4b11a230d7af5e3" Mar 13 14:28:05 crc kubenswrapper[4907]: I0313 14:28:05.404159 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556868-pzml6" Mar 13 14:28:05 crc kubenswrapper[4907]: I0313 14:28:05.425061 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"3793f53a-ca36-4b27-8444-d6dbfd860424","Type":"ContainerStarted","Data":"d484e26e1510dc0830d22229bd3bbd9bfa0e7d5cab421e8e5b87043a8df1abeb"} Mar 13 14:28:05 crc kubenswrapper[4907]: I0313 14:28:05.425139 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"3793f53a-ca36-4b27-8444-d6dbfd860424","Type":"ContainerStarted","Data":"8158569f0778590e4a7e4122294d9844f9402dad6a901826d7dd5af6e7fab659"} Mar 13 14:28:05 crc kubenswrapper[4907]: I0313 14:28:05.809014 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="72e27ed7-be4f-4d90-ac02-fa7d64585d6d" path="/var/lib/kubelet/pods/72e27ed7-be4f-4d90-ac02-fa7d64585d6d/volumes" Mar 13 14:28:06 crc kubenswrapper[4907]: I0313 14:28:06.108352 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Mar 13 14:28:06 crc kubenswrapper[4907]: I0313 14:28:06.447585 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"3793f53a-ca36-4b27-8444-d6dbfd860424","Type":"ContainerStarted","Data":"94bc55f3afe5f98100548e7593f16440afc25ced9f4af83fb6cd9f896253e749"} Mar 13 14:28:06 crc kubenswrapper[4907]: I0313 14:28:06.486176 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.486156596 podStartE2EDuration="3.486156596s" podCreationTimestamp="2026-03-13 14:28:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:28:06.481375467 +0000 UTC m=+1385.381163156" watchObservedRunningTime="2026-03-13 14:28:06.486156596 +0000 UTC m=+1385.385944295" Mar 13 14:28:07 crc kubenswrapper[4907]: I0313 14:28:07.978244 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 13 14:28:07 crc kubenswrapper[4907]: I0313 14:28:07.979061 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f744eee5-064f-464c-86fc-5130071b313d" containerName="ceilometer-central-agent" containerID="cri-o://2ba377587d3943d9d1a24298ede2bf3a69a7589b2b93375a231da14a6aa5fa8e" gracePeriod=30 Mar 13 14:28:07 crc kubenswrapper[4907]: I0313 14:28:07.979255 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f744eee5-064f-464c-86fc-5130071b313d" containerName="proxy-httpd" containerID="cri-o://7a11049e1abd6d70a403607531097f57571aa3ac1adb2b5af76d4fe1cd5fb6bd" gracePeriod=30 Mar 13 14:28:07 crc kubenswrapper[4907]: I0313 14:28:07.979300 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f744eee5-064f-464c-86fc-5130071b313d" containerName="sg-core" containerID="cri-o://e47ab86336baef65bd0ea7edbd7c31a75a6d68df9c09a6a282295545b620225a" gracePeriod=30 Mar 13 14:28:07 crc kubenswrapper[4907]: I0313 14:28:07.979439 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f744eee5-064f-464c-86fc-5130071b313d" containerName="ceilometer-notification-agent" containerID="cri-o://041e4dec51c360b9cb2fc26fee14b909f22ef44631885cdd0d5fc5f23a7ef2e5" gracePeriod=30 Mar 13 14:28:07 crc kubenswrapper[4907]: I0313 14:28:07.995247 4907 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="f744eee5-064f-464c-86fc-5130071b313d" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.170:3000/\": EOF" Mar 13 14:28:08 crc kubenswrapper[4907]: I0313 14:28:08.340669 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-7dfdbcdd5-dcb7n"] Mar 13 14:28:08 crc kubenswrapper[4907]: E0313 14:28:08.341677 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6301088c-1715-4d57-a9b8-1ea9f7128560" containerName="oc" Mar 13 14:28:08 crc kubenswrapper[4907]: I0313 14:28:08.344999 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="6301088c-1715-4d57-a9b8-1ea9f7128560" containerName="oc" Mar 13 14:28:08 crc kubenswrapper[4907]: I0313 14:28:08.345452 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="6301088c-1715-4d57-a9b8-1ea9f7128560" containerName="oc" Mar 13 14:28:08 crc kubenswrapper[4907]: I0313 14:28:08.346653 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-7dfdbcdd5-dcb7n" Mar 13 14:28:08 crc kubenswrapper[4907]: I0313 14:28:08.349659 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Mar 13 14:28:08 crc kubenswrapper[4907]: I0313 14:28:08.349847 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Mar 13 14:28:08 crc kubenswrapper[4907]: I0313 14:28:08.351895 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Mar 13 14:28:08 crc kubenswrapper[4907]: I0313 14:28:08.377215 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-7dfdbcdd5-dcb7n"] Mar 13 14:28:08 crc kubenswrapper[4907]: I0313 14:28:08.471968 4907 generic.go:334] "Generic (PLEG): container finished" podID="f744eee5-064f-464c-86fc-5130071b313d" containerID="7a11049e1abd6d70a403607531097f57571aa3ac1adb2b5af76d4fe1cd5fb6bd" exitCode=0 Mar 13 14:28:08 crc kubenswrapper[4907]: I0313 14:28:08.472006 4907 generic.go:334] "Generic (PLEG): container finished" podID="f744eee5-064f-464c-86fc-5130071b313d" containerID="e47ab86336baef65bd0ea7edbd7c31a75a6d68df9c09a6a282295545b620225a" exitCode=2 Mar 13 14:28:08 crc kubenswrapper[4907]: I0313 14:28:08.472017 4907 generic.go:334] "Generic (PLEG): container finished" podID="f744eee5-064f-464c-86fc-5130071b313d" containerID="2ba377587d3943d9d1a24298ede2bf3a69a7589b2b93375a231da14a6aa5fa8e" exitCode=0 Mar 13 14:28:08 crc kubenswrapper[4907]: I0313 14:28:08.472042 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f744eee5-064f-464c-86fc-5130071b313d","Type":"ContainerDied","Data":"7a11049e1abd6d70a403607531097f57571aa3ac1adb2b5af76d4fe1cd5fb6bd"} Mar 13 14:28:08 crc kubenswrapper[4907]: I0313 14:28:08.472090 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f744eee5-064f-464c-86fc-5130071b313d","Type":"ContainerDied","Data":"e47ab86336baef65bd0ea7edbd7c31a75a6d68df9c09a6a282295545b620225a"} Mar 13 14:28:08 crc kubenswrapper[4907]: I0313 14:28:08.472105 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f744eee5-064f-464c-86fc-5130071b313d","Type":"ContainerDied","Data":"2ba377587d3943d9d1a24298ede2bf3a69a7589b2b93375a231da14a6aa5fa8e"} Mar 13 14:28:08 crc kubenswrapper[4907]: I0313 14:28:08.507995 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7gbp6\" (UniqueName: \"kubernetes.io/projected/1a1494c5-e2d6-4d29-b161-97b720742d8d-kube-api-access-7gbp6\") pod \"swift-proxy-7dfdbcdd5-dcb7n\" (UID: \"1a1494c5-e2d6-4d29-b161-97b720742d8d\") " pod="openstack/swift-proxy-7dfdbcdd5-dcb7n" Mar 13 14:28:08 crc kubenswrapper[4907]: I0313 14:28:08.508040 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a1494c5-e2d6-4d29-b161-97b720742d8d-config-data\") pod \"swift-proxy-7dfdbcdd5-dcb7n\" (UID: \"1a1494c5-e2d6-4d29-b161-97b720742d8d\") " pod="openstack/swift-proxy-7dfdbcdd5-dcb7n" Mar 13 14:28:08 crc kubenswrapper[4907]: I0313 14:28:08.508105 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a1494c5-e2d6-4d29-b161-97b720742d8d-combined-ca-bundle\") pod \"swift-proxy-7dfdbcdd5-dcb7n\" (UID: \"1a1494c5-e2d6-4d29-b161-97b720742d8d\") " pod="openstack/swift-proxy-7dfdbcdd5-dcb7n" Mar 13 14:28:08 crc kubenswrapper[4907]: I0313 14:28:08.508309 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1a1494c5-e2d6-4d29-b161-97b720742d8d-internal-tls-certs\") pod \"swift-proxy-7dfdbcdd5-dcb7n\" (UID: \"1a1494c5-e2d6-4d29-b161-97b720742d8d\") " pod="openstack/swift-proxy-7dfdbcdd5-dcb7n" Mar 13 14:28:08 crc kubenswrapper[4907]: I0313 14:28:08.508417 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1a1494c5-e2d6-4d29-b161-97b720742d8d-run-httpd\") pod \"swift-proxy-7dfdbcdd5-dcb7n\" (UID: \"1a1494c5-e2d6-4d29-b161-97b720742d8d\") " pod="openstack/swift-proxy-7dfdbcdd5-dcb7n" Mar 13 14:28:08 crc kubenswrapper[4907]: I0313 14:28:08.508522 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1a1494c5-e2d6-4d29-b161-97b720742d8d-public-tls-certs\") pod \"swift-proxy-7dfdbcdd5-dcb7n\" (UID: \"1a1494c5-e2d6-4d29-b161-97b720742d8d\") " pod="openstack/swift-proxy-7dfdbcdd5-dcb7n" Mar 13 14:28:08 crc kubenswrapper[4907]: I0313 14:28:08.508576 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1a1494c5-e2d6-4d29-b161-97b720742d8d-log-httpd\") pod \"swift-proxy-7dfdbcdd5-dcb7n\" (UID: \"1a1494c5-e2d6-4d29-b161-97b720742d8d\") " pod="openstack/swift-proxy-7dfdbcdd5-dcb7n" Mar 13 14:28:08 crc kubenswrapper[4907]: I0313 14:28:08.508666 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/1a1494c5-e2d6-4d29-b161-97b720742d8d-etc-swift\") pod \"swift-proxy-7dfdbcdd5-dcb7n\" (UID: \"1a1494c5-e2d6-4d29-b161-97b720742d8d\") " pod="openstack/swift-proxy-7dfdbcdd5-dcb7n" Mar 13 14:28:08 crc kubenswrapper[4907]: I0313 14:28:08.609966 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a1494c5-e2d6-4d29-b161-97b720742d8d-config-data\") pod \"swift-proxy-7dfdbcdd5-dcb7n\" (UID: \"1a1494c5-e2d6-4d29-b161-97b720742d8d\") " pod="openstack/swift-proxy-7dfdbcdd5-dcb7n" Mar 13 14:28:08 crc kubenswrapper[4907]: I0313 14:28:08.610066 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a1494c5-e2d6-4d29-b161-97b720742d8d-combined-ca-bundle\") pod \"swift-proxy-7dfdbcdd5-dcb7n\" (UID: \"1a1494c5-e2d6-4d29-b161-97b720742d8d\") " pod="openstack/swift-proxy-7dfdbcdd5-dcb7n" Mar 13 14:28:08 crc kubenswrapper[4907]: I0313 14:28:08.610102 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1a1494c5-e2d6-4d29-b161-97b720742d8d-internal-tls-certs\") pod \"swift-proxy-7dfdbcdd5-dcb7n\" (UID: \"1a1494c5-e2d6-4d29-b161-97b720742d8d\") " pod="openstack/swift-proxy-7dfdbcdd5-dcb7n" Mar 13 14:28:08 crc kubenswrapper[4907]: I0313 14:28:08.610132 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1a1494c5-e2d6-4d29-b161-97b720742d8d-run-httpd\") pod \"swift-proxy-7dfdbcdd5-dcb7n\" (UID: \"1a1494c5-e2d6-4d29-b161-97b720742d8d\") " pod="openstack/swift-proxy-7dfdbcdd5-dcb7n" Mar 13 14:28:08 crc kubenswrapper[4907]: I0313 14:28:08.610166 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1a1494c5-e2d6-4d29-b161-97b720742d8d-public-tls-certs\") pod \"swift-proxy-7dfdbcdd5-dcb7n\" (UID: \"1a1494c5-e2d6-4d29-b161-97b720742d8d\") " pod="openstack/swift-proxy-7dfdbcdd5-dcb7n" Mar 13 14:28:08 crc kubenswrapper[4907]: I0313 14:28:08.610194 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1a1494c5-e2d6-4d29-b161-97b720742d8d-log-httpd\") pod \"swift-proxy-7dfdbcdd5-dcb7n\" (UID: \"1a1494c5-e2d6-4d29-b161-97b720742d8d\") " pod="openstack/swift-proxy-7dfdbcdd5-dcb7n" Mar 13 14:28:08 crc kubenswrapper[4907]: I0313 14:28:08.610227 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/1a1494c5-e2d6-4d29-b161-97b720742d8d-etc-swift\") pod \"swift-proxy-7dfdbcdd5-dcb7n\" (UID: \"1a1494c5-e2d6-4d29-b161-97b720742d8d\") " pod="openstack/swift-proxy-7dfdbcdd5-dcb7n" Mar 13 14:28:08 crc kubenswrapper[4907]: I0313 14:28:08.610290 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7gbp6\" (UniqueName: \"kubernetes.io/projected/1a1494c5-e2d6-4d29-b161-97b720742d8d-kube-api-access-7gbp6\") pod \"swift-proxy-7dfdbcdd5-dcb7n\" (UID: \"1a1494c5-e2d6-4d29-b161-97b720742d8d\") " pod="openstack/swift-proxy-7dfdbcdd5-dcb7n" Mar 13 14:28:08 crc kubenswrapper[4907]: I0313 14:28:08.615469 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1a1494c5-e2d6-4d29-b161-97b720742d8d-log-httpd\") pod \"swift-proxy-7dfdbcdd5-dcb7n\" (UID: \"1a1494c5-e2d6-4d29-b161-97b720742d8d\") " pod="openstack/swift-proxy-7dfdbcdd5-dcb7n" Mar 13 14:28:08 crc kubenswrapper[4907]: I0313 14:28:08.616028 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1a1494c5-e2d6-4d29-b161-97b720742d8d-run-httpd\") pod \"swift-proxy-7dfdbcdd5-dcb7n\" (UID: \"1a1494c5-e2d6-4d29-b161-97b720742d8d\") " pod="openstack/swift-proxy-7dfdbcdd5-dcb7n" Mar 13 14:28:08 crc kubenswrapper[4907]: I0313 14:28:08.620831 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1a1494c5-e2d6-4d29-b161-97b720742d8d-public-tls-certs\") pod \"swift-proxy-7dfdbcdd5-dcb7n\" (UID: \"1a1494c5-e2d6-4d29-b161-97b720742d8d\") " pod="openstack/swift-proxy-7dfdbcdd5-dcb7n" Mar 13 14:28:08 crc kubenswrapper[4907]: I0313 14:28:08.621316 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a1494c5-e2d6-4d29-b161-97b720742d8d-combined-ca-bundle\") pod \"swift-proxy-7dfdbcdd5-dcb7n\" (UID: \"1a1494c5-e2d6-4d29-b161-97b720742d8d\") " pod="openstack/swift-proxy-7dfdbcdd5-dcb7n" Mar 13 14:28:08 crc kubenswrapper[4907]: I0313 14:28:08.635075 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/1a1494c5-e2d6-4d29-b161-97b720742d8d-etc-swift\") pod \"swift-proxy-7dfdbcdd5-dcb7n\" (UID: \"1a1494c5-e2d6-4d29-b161-97b720742d8d\") " pod="openstack/swift-proxy-7dfdbcdd5-dcb7n" Mar 13 14:28:08 crc kubenswrapper[4907]: I0313 14:28:08.640972 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7gbp6\" (UniqueName: \"kubernetes.io/projected/1a1494c5-e2d6-4d29-b161-97b720742d8d-kube-api-access-7gbp6\") pod \"swift-proxy-7dfdbcdd5-dcb7n\" (UID: \"1a1494c5-e2d6-4d29-b161-97b720742d8d\") " pod="openstack/swift-proxy-7dfdbcdd5-dcb7n" Mar 13 14:28:08 crc kubenswrapper[4907]: I0313 14:28:08.642753 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1a1494c5-e2d6-4d29-b161-97b720742d8d-internal-tls-certs\") pod \"swift-proxy-7dfdbcdd5-dcb7n\" (UID: \"1a1494c5-e2d6-4d29-b161-97b720742d8d\") " pod="openstack/swift-proxy-7dfdbcdd5-dcb7n" Mar 13 14:28:08 crc kubenswrapper[4907]: I0313 14:28:08.645898 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a1494c5-e2d6-4d29-b161-97b720742d8d-config-data\") pod \"swift-proxy-7dfdbcdd5-dcb7n\" (UID: \"1a1494c5-e2d6-4d29-b161-97b720742d8d\") " pod="openstack/swift-proxy-7dfdbcdd5-dcb7n" Mar 13 14:28:08 crc kubenswrapper[4907]: I0313 14:28:08.672054 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-7dfdbcdd5-dcb7n" Mar 13 14:28:08 crc kubenswrapper[4907]: I0313 14:28:08.867728 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.013038 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.020259 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f744eee5-064f-464c-86fc-5130071b313d-sg-core-conf-yaml\") pod \"f744eee5-064f-464c-86fc-5130071b313d\" (UID: \"f744eee5-064f-464c-86fc-5130071b313d\") " Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.020373 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f744eee5-064f-464c-86fc-5130071b313d-run-httpd\") pod \"f744eee5-064f-464c-86fc-5130071b313d\" (UID: \"f744eee5-064f-464c-86fc-5130071b313d\") " Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.020412 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f744eee5-064f-464c-86fc-5130071b313d-combined-ca-bundle\") pod \"f744eee5-064f-464c-86fc-5130071b313d\" (UID: \"f744eee5-064f-464c-86fc-5130071b313d\") " Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.020607 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dtpdc\" (UniqueName: \"kubernetes.io/projected/f744eee5-064f-464c-86fc-5130071b313d-kube-api-access-dtpdc\") pod \"f744eee5-064f-464c-86fc-5130071b313d\" (UID: \"f744eee5-064f-464c-86fc-5130071b313d\") " Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.020640 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f744eee5-064f-464c-86fc-5130071b313d-scripts\") pod \"f744eee5-064f-464c-86fc-5130071b313d\" (UID: \"f744eee5-064f-464c-86fc-5130071b313d\") " Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.020689 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f744eee5-064f-464c-86fc-5130071b313d-log-httpd\") pod \"f744eee5-064f-464c-86fc-5130071b313d\" (UID: \"f744eee5-064f-464c-86fc-5130071b313d\") " Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.020724 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f744eee5-064f-464c-86fc-5130071b313d-config-data\") pod \"f744eee5-064f-464c-86fc-5130071b313d\" (UID: \"f744eee5-064f-464c-86fc-5130071b313d\") " Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.024211 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f744eee5-064f-464c-86fc-5130071b313d-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "f744eee5-064f-464c-86fc-5130071b313d" (UID: "f744eee5-064f-464c-86fc-5130071b313d"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.024652 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f744eee5-064f-464c-86fc-5130071b313d-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "f744eee5-064f-464c-86fc-5130071b313d" (UID: "f744eee5-064f-464c-86fc-5130071b313d"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.029851 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f744eee5-064f-464c-86fc-5130071b313d-kube-api-access-dtpdc" (OuterVolumeSpecName: "kube-api-access-dtpdc") pod "f744eee5-064f-464c-86fc-5130071b313d" (UID: "f744eee5-064f-464c-86fc-5130071b313d"). InnerVolumeSpecName "kube-api-access-dtpdc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.030287 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f744eee5-064f-464c-86fc-5130071b313d-scripts" (OuterVolumeSpecName: "scripts") pod "f744eee5-064f-464c-86fc-5130071b313d" (UID: "f744eee5-064f-464c-86fc-5130071b313d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.064022 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f744eee5-064f-464c-86fc-5130071b313d-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "f744eee5-064f-464c-86fc-5130071b313d" (UID: "f744eee5-064f-464c-86fc-5130071b313d"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.116441 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f744eee5-064f-464c-86fc-5130071b313d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f744eee5-064f-464c-86fc-5130071b313d" (UID: "f744eee5-064f-464c-86fc-5130071b313d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.128202 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dtpdc\" (UniqueName: \"kubernetes.io/projected/f744eee5-064f-464c-86fc-5130071b313d-kube-api-access-dtpdc\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.128228 4907 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f744eee5-064f-464c-86fc-5130071b313d-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.128241 4907 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f744eee5-064f-464c-86fc-5130071b313d-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.128253 4907 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f744eee5-064f-464c-86fc-5130071b313d-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.128263 4907 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f744eee5-064f-464c-86fc-5130071b313d-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.128274 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f744eee5-064f-464c-86fc-5130071b313d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.172171 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f744eee5-064f-464c-86fc-5130071b313d-config-data" (OuterVolumeSpecName: "config-data") pod "f744eee5-064f-464c-86fc-5130071b313d" (UID: "f744eee5-064f-464c-86fc-5130071b313d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.229968 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f744eee5-064f-464c-86fc-5130071b313d-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.363602 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-7dfdbcdd5-dcb7n"] Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.485368 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-7dfdbcdd5-dcb7n" event={"ID":"1a1494c5-e2d6-4d29-b161-97b720742d8d","Type":"ContainerStarted","Data":"3472978398de27f6988019a31485f47afeca847038681d998902a9798bd82c2a"} Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.488907 4907 generic.go:334] "Generic (PLEG): container finished" podID="f744eee5-064f-464c-86fc-5130071b313d" containerID="041e4dec51c360b9cb2fc26fee14b909f22ef44631885cdd0d5fc5f23a7ef2e5" exitCode=0 Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.488949 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f744eee5-064f-464c-86fc-5130071b313d","Type":"ContainerDied","Data":"041e4dec51c360b9cb2fc26fee14b909f22ef44631885cdd0d5fc5f23a7ef2e5"} Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.488976 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f744eee5-064f-464c-86fc-5130071b313d","Type":"ContainerDied","Data":"ad91253ab4bed9131d491de9a5560dd1867583453c957940978da876fc7db3cf"} Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.488991 4907 scope.go:117] "RemoveContainer" containerID="7a11049e1abd6d70a403607531097f57571aa3ac1adb2b5af76d4fe1cd5fb6bd" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.489120 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.538790 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.543217 4907 scope.go:117] "RemoveContainer" containerID="e47ab86336baef65bd0ea7edbd7c31a75a6d68df9c09a6a282295545b620225a" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.546653 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.568759 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 13 14:28:09 crc kubenswrapper[4907]: E0313 14:28:09.569116 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f744eee5-064f-464c-86fc-5130071b313d" containerName="sg-core" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.569133 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="f744eee5-064f-464c-86fc-5130071b313d" containerName="sg-core" Mar 13 14:28:09 crc kubenswrapper[4907]: E0313 14:28:09.569146 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f744eee5-064f-464c-86fc-5130071b313d" containerName="ceilometer-central-agent" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.569154 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="f744eee5-064f-464c-86fc-5130071b313d" containerName="ceilometer-central-agent" Mar 13 14:28:09 crc kubenswrapper[4907]: E0313 14:28:09.569175 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f744eee5-064f-464c-86fc-5130071b313d" containerName="proxy-httpd" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.569181 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="f744eee5-064f-464c-86fc-5130071b313d" containerName="proxy-httpd" Mar 13 14:28:09 crc kubenswrapper[4907]: E0313 14:28:09.569197 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f744eee5-064f-464c-86fc-5130071b313d" containerName="ceilometer-notification-agent" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.569203 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="f744eee5-064f-464c-86fc-5130071b313d" containerName="ceilometer-notification-agent" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.569354 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="f744eee5-064f-464c-86fc-5130071b313d" containerName="proxy-httpd" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.569370 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="f744eee5-064f-464c-86fc-5130071b313d" containerName="sg-core" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.569389 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="f744eee5-064f-464c-86fc-5130071b313d" containerName="ceilometer-central-agent" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.569401 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="f744eee5-064f-464c-86fc-5130071b313d" containerName="ceilometer-notification-agent" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.571138 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.577214 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.578444 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.584638 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.591848 4907 scope.go:117] "RemoveContainer" containerID="041e4dec51c360b9cb2fc26fee14b909f22ef44631885cdd0d5fc5f23a7ef2e5" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.623706 4907 scope.go:117] "RemoveContainer" containerID="2ba377587d3943d9d1a24298ede2bf3a69a7589b2b93375a231da14a6aa5fa8e" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.652990 4907 scope.go:117] "RemoveContainer" containerID="7a11049e1abd6d70a403607531097f57571aa3ac1adb2b5af76d4fe1cd5fb6bd" Mar 13 14:28:09 crc kubenswrapper[4907]: E0313 14:28:09.656972 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7a11049e1abd6d70a403607531097f57571aa3ac1adb2b5af76d4fe1cd5fb6bd\": container with ID starting with 7a11049e1abd6d70a403607531097f57571aa3ac1adb2b5af76d4fe1cd5fb6bd not found: ID does not exist" containerID="7a11049e1abd6d70a403607531097f57571aa3ac1adb2b5af76d4fe1cd5fb6bd" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.657008 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7a11049e1abd6d70a403607531097f57571aa3ac1adb2b5af76d4fe1cd5fb6bd"} err="failed to get container status \"7a11049e1abd6d70a403607531097f57571aa3ac1adb2b5af76d4fe1cd5fb6bd\": rpc error: code = NotFound desc = could not find container \"7a11049e1abd6d70a403607531097f57571aa3ac1adb2b5af76d4fe1cd5fb6bd\": container with ID starting with 7a11049e1abd6d70a403607531097f57571aa3ac1adb2b5af76d4fe1cd5fb6bd not found: ID does not exist" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.657029 4907 scope.go:117] "RemoveContainer" containerID="e47ab86336baef65bd0ea7edbd7c31a75a6d68df9c09a6a282295545b620225a" Mar 13 14:28:09 crc kubenswrapper[4907]: E0313 14:28:09.657424 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e47ab86336baef65bd0ea7edbd7c31a75a6d68df9c09a6a282295545b620225a\": container with ID starting with e47ab86336baef65bd0ea7edbd7c31a75a6d68df9c09a6a282295545b620225a not found: ID does not exist" containerID="e47ab86336baef65bd0ea7edbd7c31a75a6d68df9c09a6a282295545b620225a" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.657480 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e47ab86336baef65bd0ea7edbd7c31a75a6d68df9c09a6a282295545b620225a"} err="failed to get container status \"e47ab86336baef65bd0ea7edbd7c31a75a6d68df9c09a6a282295545b620225a\": rpc error: code = NotFound desc = could not find container \"e47ab86336baef65bd0ea7edbd7c31a75a6d68df9c09a6a282295545b620225a\": container with ID starting with e47ab86336baef65bd0ea7edbd7c31a75a6d68df9c09a6a282295545b620225a not found: ID does not exist" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.657547 4907 scope.go:117] "RemoveContainer" containerID="041e4dec51c360b9cb2fc26fee14b909f22ef44631885cdd0d5fc5f23a7ef2e5" Mar 13 14:28:09 crc kubenswrapper[4907]: E0313 14:28:09.661014 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"041e4dec51c360b9cb2fc26fee14b909f22ef44631885cdd0d5fc5f23a7ef2e5\": container with ID starting with 041e4dec51c360b9cb2fc26fee14b909f22ef44631885cdd0d5fc5f23a7ef2e5 not found: ID does not exist" containerID="041e4dec51c360b9cb2fc26fee14b909f22ef44631885cdd0d5fc5f23a7ef2e5" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.661070 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"041e4dec51c360b9cb2fc26fee14b909f22ef44631885cdd0d5fc5f23a7ef2e5"} err="failed to get container status \"041e4dec51c360b9cb2fc26fee14b909f22ef44631885cdd0d5fc5f23a7ef2e5\": rpc error: code = NotFound desc = could not find container \"041e4dec51c360b9cb2fc26fee14b909f22ef44631885cdd0d5fc5f23a7ef2e5\": container with ID starting with 041e4dec51c360b9cb2fc26fee14b909f22ef44631885cdd0d5fc5f23a7ef2e5 not found: ID does not exist" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.661099 4907 scope.go:117] "RemoveContainer" containerID="2ba377587d3943d9d1a24298ede2bf3a69a7589b2b93375a231da14a6aa5fa8e" Mar 13 14:28:09 crc kubenswrapper[4907]: E0313 14:28:09.661418 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2ba377587d3943d9d1a24298ede2bf3a69a7589b2b93375a231da14a6aa5fa8e\": container with ID starting with 2ba377587d3943d9d1a24298ede2bf3a69a7589b2b93375a231da14a6aa5fa8e not found: ID does not exist" containerID="2ba377587d3943d9d1a24298ede2bf3a69a7589b2b93375a231da14a6aa5fa8e" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.661447 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2ba377587d3943d9d1a24298ede2bf3a69a7589b2b93375a231da14a6aa5fa8e"} err="failed to get container status \"2ba377587d3943d9d1a24298ede2bf3a69a7589b2b93375a231da14a6aa5fa8e\": rpc error: code = NotFound desc = could not find container \"2ba377587d3943d9d1a24298ede2bf3a69a7589b2b93375a231da14a6aa5fa8e\": container with ID starting with 2ba377587d3943d9d1a24298ede2bf3a69a7589b2b93375a231da14a6aa5fa8e not found: ID does not exist" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.739062 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f7ca16f6-c9d9-4415-b8cf-b4753d784170-run-httpd\") pod \"ceilometer-0\" (UID: \"f7ca16f6-c9d9-4415-b8cf-b4753d784170\") " pod="openstack/ceilometer-0" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.739258 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gfnvm\" (UniqueName: \"kubernetes.io/projected/f7ca16f6-c9d9-4415-b8cf-b4753d784170-kube-api-access-gfnvm\") pod \"ceilometer-0\" (UID: \"f7ca16f6-c9d9-4415-b8cf-b4753d784170\") " pod="openstack/ceilometer-0" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.739305 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7ca16f6-c9d9-4415-b8cf-b4753d784170-config-data\") pod \"ceilometer-0\" (UID: \"f7ca16f6-c9d9-4415-b8cf-b4753d784170\") " pod="openstack/ceilometer-0" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.739324 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f7ca16f6-c9d9-4415-b8cf-b4753d784170-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f7ca16f6-c9d9-4415-b8cf-b4753d784170\") " pod="openstack/ceilometer-0" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.740547 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f7ca16f6-c9d9-4415-b8cf-b4753d784170-log-httpd\") pod \"ceilometer-0\" (UID: \"f7ca16f6-c9d9-4415-b8cf-b4753d784170\") " pod="openstack/ceilometer-0" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.740588 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7ca16f6-c9d9-4415-b8cf-b4753d784170-scripts\") pod \"ceilometer-0\" (UID: \"f7ca16f6-c9d9-4415-b8cf-b4753d784170\") " pod="openstack/ceilometer-0" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.740621 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7ca16f6-c9d9-4415-b8cf-b4753d784170-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f7ca16f6-c9d9-4415-b8cf-b4753d784170\") " pod="openstack/ceilometer-0" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.794758 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f744eee5-064f-464c-86fc-5130071b313d" path="/var/lib/kubelet/pods/f744eee5-064f-464c-86fc-5130071b313d/volumes" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.842837 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f7ca16f6-c9d9-4415-b8cf-b4753d784170-run-httpd\") pod \"ceilometer-0\" (UID: \"f7ca16f6-c9d9-4415-b8cf-b4753d784170\") " pod="openstack/ceilometer-0" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.842992 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gfnvm\" (UniqueName: \"kubernetes.io/projected/f7ca16f6-c9d9-4415-b8cf-b4753d784170-kube-api-access-gfnvm\") pod \"ceilometer-0\" (UID: \"f7ca16f6-c9d9-4415-b8cf-b4753d784170\") " pod="openstack/ceilometer-0" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.843028 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7ca16f6-c9d9-4415-b8cf-b4753d784170-config-data\") pod \"ceilometer-0\" (UID: \"f7ca16f6-c9d9-4415-b8cf-b4753d784170\") " pod="openstack/ceilometer-0" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.843055 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f7ca16f6-c9d9-4415-b8cf-b4753d784170-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f7ca16f6-c9d9-4415-b8cf-b4753d784170\") " pod="openstack/ceilometer-0" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.843101 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f7ca16f6-c9d9-4415-b8cf-b4753d784170-log-httpd\") pod \"ceilometer-0\" (UID: \"f7ca16f6-c9d9-4415-b8cf-b4753d784170\") " pod="openstack/ceilometer-0" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.843302 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7ca16f6-c9d9-4415-b8cf-b4753d784170-scripts\") pod \"ceilometer-0\" (UID: \"f7ca16f6-c9d9-4415-b8cf-b4753d784170\") " pod="openstack/ceilometer-0" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.843334 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7ca16f6-c9d9-4415-b8cf-b4753d784170-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f7ca16f6-c9d9-4415-b8cf-b4753d784170\") " pod="openstack/ceilometer-0" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.843417 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f7ca16f6-c9d9-4415-b8cf-b4753d784170-run-httpd\") pod \"ceilometer-0\" (UID: \"f7ca16f6-c9d9-4415-b8cf-b4753d784170\") " pod="openstack/ceilometer-0" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.843705 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f7ca16f6-c9d9-4415-b8cf-b4753d784170-log-httpd\") pod \"ceilometer-0\" (UID: \"f7ca16f6-c9d9-4415-b8cf-b4753d784170\") " pod="openstack/ceilometer-0" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.849095 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7ca16f6-c9d9-4415-b8cf-b4753d784170-config-data\") pod \"ceilometer-0\" (UID: \"f7ca16f6-c9d9-4415-b8cf-b4753d784170\") " pod="openstack/ceilometer-0" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.849688 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7ca16f6-c9d9-4415-b8cf-b4753d784170-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f7ca16f6-c9d9-4415-b8cf-b4753d784170\") " pod="openstack/ceilometer-0" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.852377 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f7ca16f6-c9d9-4415-b8cf-b4753d784170-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f7ca16f6-c9d9-4415-b8cf-b4753d784170\") " pod="openstack/ceilometer-0" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.859652 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7ca16f6-c9d9-4415-b8cf-b4753d784170-scripts\") pod \"ceilometer-0\" (UID: \"f7ca16f6-c9d9-4415-b8cf-b4753d784170\") " pod="openstack/ceilometer-0" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.863753 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gfnvm\" (UniqueName: \"kubernetes.io/projected/f7ca16f6-c9d9-4415-b8cf-b4753d784170-kube-api-access-gfnvm\") pod \"ceilometer-0\" (UID: \"f7ca16f6-c9d9-4415-b8cf-b4753d784170\") " pod="openstack/ceilometer-0" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.898589 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 14:28:10 crc kubenswrapper[4907]: I0313 14:28:10.497828 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 13 14:28:10 crc kubenswrapper[4907]: I0313 14:28:10.504328 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-7dfdbcdd5-dcb7n" event={"ID":"1a1494c5-e2d6-4d29-b161-97b720742d8d","Type":"ContainerStarted","Data":"c6c4a6a54081de702756857f0341351ea26c5f9aaa725027df2c989c6b1a6e18"} Mar 13 14:28:10 crc kubenswrapper[4907]: I0313 14:28:10.504380 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-7dfdbcdd5-dcb7n" event={"ID":"1a1494c5-e2d6-4d29-b161-97b720742d8d","Type":"ContainerStarted","Data":"676efd80769fe771754cb647e65098f243a686b0c64467414fa0d6ccfbece2d4"} Mar 13 14:28:10 crc kubenswrapper[4907]: I0313 14:28:10.506024 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-7dfdbcdd5-dcb7n" Mar 13 14:28:10 crc kubenswrapper[4907]: I0313 14:28:10.506139 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-7dfdbcdd5-dcb7n" Mar 13 14:28:10 crc kubenswrapper[4907]: I0313 14:28:10.532385 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-7dfdbcdd5-dcb7n" podStartSLOduration=2.5323603329999997 podStartE2EDuration="2.532360333s" podCreationTimestamp="2026-03-13 14:28:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:28:10.521713224 +0000 UTC m=+1389.421500913" watchObservedRunningTime="2026-03-13 14:28:10.532360333 +0000 UTC m=+1389.432148022" Mar 13 14:28:11 crc kubenswrapper[4907]: I0313 14:28:11.518429 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f7ca16f6-c9d9-4415-b8cf-b4753d784170","Type":"ContainerStarted","Data":"2e74becf0a06fc2c8e64ca7acbdb2fa95c901f1325aa7aeacdbfe4e9db0c932e"} Mar 13 14:28:11 crc kubenswrapper[4907]: I0313 14:28:11.731740 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 13 14:28:11 crc kubenswrapper[4907]: I0313 14:28:11.732103 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="2c9261af-9eaa-4623-9f1b-719b58a9d3a2" containerName="glance-log" containerID="cri-o://6af4b6388e126ccd4e3d2cc04b6747e5283033aaa5d424f9e2907953b3f55ea5" gracePeriod=30 Mar 13 14:28:11 crc kubenswrapper[4907]: I0313 14:28:11.732415 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="2c9261af-9eaa-4623-9f1b-719b58a9d3a2" containerName="glance-httpd" containerID="cri-o://22707601406e7d79a6ae3140c68d6ec6d23c0168f7d4a1523dfe2e817903318f" gracePeriod=30 Mar 13 14:28:12 crc kubenswrapper[4907]: I0313 14:28:12.505541 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 13 14:28:12 crc kubenswrapper[4907]: I0313 14:28:12.505796 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="5c36a352-151f-4e93-8094-3855bfed532e" containerName="glance-log" containerID="cri-o://4364cee9bed454e59183eb0004b9dfc4ab15dc5a345acbfd2a54032342758450" gracePeriod=30 Mar 13 14:28:12 crc kubenswrapper[4907]: I0313 14:28:12.506240 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="5c36a352-151f-4e93-8094-3855bfed532e" containerName="glance-httpd" containerID="cri-o://456e65b2b22522c83f37e28d27a38f3af4c42d273550f8c8c1246de8d14f0a2f" gracePeriod=30 Mar 13 14:28:12 crc kubenswrapper[4907]: I0313 14:28:12.534910 4907 generic.go:334] "Generic (PLEG): container finished" podID="2c9261af-9eaa-4623-9f1b-719b58a9d3a2" containerID="6af4b6388e126ccd4e3d2cc04b6747e5283033aaa5d424f9e2907953b3f55ea5" exitCode=143 Mar 13 14:28:12 crc kubenswrapper[4907]: I0313 14:28:12.535024 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2c9261af-9eaa-4623-9f1b-719b58a9d3a2","Type":"ContainerDied","Data":"6af4b6388e126ccd4e3d2cc04b6747e5283033aaa5d424f9e2907953b3f55ea5"} Mar 13 14:28:13 crc kubenswrapper[4907]: I0313 14:28:13.561055 4907 generic.go:334] "Generic (PLEG): container finished" podID="5c36a352-151f-4e93-8094-3855bfed532e" containerID="4364cee9bed454e59183eb0004b9dfc4ab15dc5a345acbfd2a54032342758450" exitCode=143 Mar 13 14:28:13 crc kubenswrapper[4907]: I0313 14:28:13.561134 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5c36a352-151f-4e93-8094-3855bfed532e","Type":"ContainerDied","Data":"4364cee9bed454e59183eb0004b9dfc4ab15dc5a345acbfd2a54032342758450"} Mar 13 14:28:13 crc kubenswrapper[4907]: I0313 14:28:13.721696 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 13 14:28:14 crc kubenswrapper[4907]: I0313 14:28:14.243159 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Mar 13 14:28:14 crc kubenswrapper[4907]: I0313 14:28:14.897412 4907 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-external-api-0" podUID="2c9261af-9eaa-4623-9f1b-719b58a9d3a2" containerName="glance-log" probeResult="failure" output="Get \"https://10.217.0.154:9292/healthcheck\": read tcp 10.217.0.2:41636->10.217.0.154:9292: read: connection reset by peer" Mar 13 14:28:14 crc kubenswrapper[4907]: I0313 14:28:14.900155 4907 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-external-api-0" podUID="2c9261af-9eaa-4623-9f1b-719b58a9d3a2" containerName="glance-httpd" probeResult="failure" output="Get \"https://10.217.0.154:9292/healthcheck\": read tcp 10.217.0.2:41622->10.217.0.154:9292: read: connection reset by peer" Mar 13 14:28:15 crc kubenswrapper[4907]: I0313 14:28:15.591377 4907 generic.go:334] "Generic (PLEG): container finished" podID="2c9261af-9eaa-4623-9f1b-719b58a9d3a2" containerID="22707601406e7d79a6ae3140c68d6ec6d23c0168f7d4a1523dfe2e817903318f" exitCode=0 Mar 13 14:28:15 crc kubenswrapper[4907]: I0313 14:28:15.591422 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2c9261af-9eaa-4623-9f1b-719b58a9d3a2","Type":"ContainerDied","Data":"22707601406e7d79a6ae3140c68d6ec6d23c0168f7d4a1523dfe2e817903318f"} Mar 13 14:28:16 crc kubenswrapper[4907]: I0313 14:28:16.604044 4907 generic.go:334] "Generic (PLEG): container finished" podID="5c36a352-151f-4e93-8094-3855bfed532e" containerID="456e65b2b22522c83f37e28d27a38f3af4c42d273550f8c8c1246de8d14f0a2f" exitCode=0 Mar 13 14:28:16 crc kubenswrapper[4907]: I0313 14:28:16.604096 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5c36a352-151f-4e93-8094-3855bfed532e","Type":"ContainerDied","Data":"456e65b2b22522c83f37e28d27a38f3af4c42d273550f8c8c1246de8d14f0a2f"} Mar 13 14:28:16 crc kubenswrapper[4907]: I0313 14:28:16.615934 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-g96mx"] Mar 13 14:28:16 crc kubenswrapper[4907]: I0313 14:28:16.617031 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-g96mx" Mar 13 14:28:16 crc kubenswrapper[4907]: I0313 14:28:16.654796 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-g96mx"] Mar 13 14:28:16 crc kubenswrapper[4907]: I0313 14:28:16.722900 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-vlp6r"] Mar 13 14:28:16 crc kubenswrapper[4907]: I0313 14:28:16.724510 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-vlp6r" Mar 13 14:28:16 crc kubenswrapper[4907]: I0313 14:28:16.730385 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-22bc-account-create-update-r969t"] Mar 13 14:28:16 crc kubenswrapper[4907]: I0313 14:28:16.731671 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-22bc-account-create-update-r969t" Mar 13 14:28:16 crc kubenswrapper[4907]: I0313 14:28:16.734085 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Mar 13 14:28:16 crc kubenswrapper[4907]: I0313 14:28:16.752354 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-vlp6r"] Mar 13 14:28:16 crc kubenswrapper[4907]: I0313 14:28:16.769443 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-22bc-account-create-update-r969t"] Mar 13 14:28:16 crc kubenswrapper[4907]: I0313 14:28:16.795479 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-swn2k\" (UniqueName: \"kubernetes.io/projected/9118baf0-df6a-41e5-98e5-3e4d8edfbaaa-kube-api-access-swn2k\") pod \"nova-cell0-db-create-vlp6r\" (UID: \"9118baf0-df6a-41e5-98e5-3e4d8edfbaaa\") " pod="openstack/nova-cell0-db-create-vlp6r" Mar 13 14:28:16 crc kubenswrapper[4907]: I0313 14:28:16.798700 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mjhh8\" (UniqueName: \"kubernetes.io/projected/251ab76e-b9df-406f-9b07-c7fa6e227feb-kube-api-access-mjhh8\") pod \"nova-api-db-create-g96mx\" (UID: \"251ab76e-b9df-406f-9b07-c7fa6e227feb\") " pod="openstack/nova-api-db-create-g96mx" Mar 13 14:28:16 crc kubenswrapper[4907]: I0313 14:28:16.798901 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4pm5m\" (UniqueName: \"kubernetes.io/projected/4eeb2fb2-22fb-4ea8-a39b-5f959877cfe6-kube-api-access-4pm5m\") pod \"nova-api-22bc-account-create-update-r969t\" (UID: \"4eeb2fb2-22fb-4ea8-a39b-5f959877cfe6\") " pod="openstack/nova-api-22bc-account-create-update-r969t" Mar 13 14:28:16 crc kubenswrapper[4907]: I0313 14:28:16.798989 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/251ab76e-b9df-406f-9b07-c7fa6e227feb-operator-scripts\") pod \"nova-api-db-create-g96mx\" (UID: \"251ab76e-b9df-406f-9b07-c7fa6e227feb\") " pod="openstack/nova-api-db-create-g96mx" Mar 13 14:28:16 crc kubenswrapper[4907]: I0313 14:28:16.799124 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9118baf0-df6a-41e5-98e5-3e4d8edfbaaa-operator-scripts\") pod \"nova-cell0-db-create-vlp6r\" (UID: \"9118baf0-df6a-41e5-98e5-3e4d8edfbaaa\") " pod="openstack/nova-cell0-db-create-vlp6r" Mar 13 14:28:16 crc kubenswrapper[4907]: I0313 14:28:16.799421 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4eeb2fb2-22fb-4ea8-a39b-5f959877cfe6-operator-scripts\") pod \"nova-api-22bc-account-create-update-r969t\" (UID: \"4eeb2fb2-22fb-4ea8-a39b-5f959877cfe6\") " pod="openstack/nova-api-22bc-account-create-update-r969t" Mar 13 14:28:16 crc kubenswrapper[4907]: I0313 14:28:16.839755 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-ldbt9"] Mar 13 14:28:16 crc kubenswrapper[4907]: I0313 14:28:16.842301 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-ldbt9" Mar 13 14:28:16 crc kubenswrapper[4907]: I0313 14:28:16.851238 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-ldbt9"] Mar 13 14:28:16 crc kubenswrapper[4907]: I0313 14:28:16.901321 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wj4tn\" (UniqueName: \"kubernetes.io/projected/5dcaf7c6-b4dd-4312-b68b-a833a827a2c2-kube-api-access-wj4tn\") pod \"nova-cell1-db-create-ldbt9\" (UID: \"5dcaf7c6-b4dd-4312-b68b-a833a827a2c2\") " pod="openstack/nova-cell1-db-create-ldbt9" Mar 13 14:28:16 crc kubenswrapper[4907]: I0313 14:28:16.901486 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mjhh8\" (UniqueName: \"kubernetes.io/projected/251ab76e-b9df-406f-9b07-c7fa6e227feb-kube-api-access-mjhh8\") pod \"nova-api-db-create-g96mx\" (UID: \"251ab76e-b9df-406f-9b07-c7fa6e227feb\") " pod="openstack/nova-api-db-create-g96mx" Mar 13 14:28:16 crc kubenswrapper[4907]: I0313 14:28:16.901638 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4pm5m\" (UniqueName: \"kubernetes.io/projected/4eeb2fb2-22fb-4ea8-a39b-5f959877cfe6-kube-api-access-4pm5m\") pod \"nova-api-22bc-account-create-update-r969t\" (UID: \"4eeb2fb2-22fb-4ea8-a39b-5f959877cfe6\") " pod="openstack/nova-api-22bc-account-create-update-r969t" Mar 13 14:28:16 crc kubenswrapper[4907]: I0313 14:28:16.901683 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/251ab76e-b9df-406f-9b07-c7fa6e227feb-operator-scripts\") pod \"nova-api-db-create-g96mx\" (UID: \"251ab76e-b9df-406f-9b07-c7fa6e227feb\") " pod="openstack/nova-api-db-create-g96mx" Mar 13 14:28:16 crc kubenswrapper[4907]: I0313 14:28:16.903074 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9118baf0-df6a-41e5-98e5-3e4d8edfbaaa-operator-scripts\") pod \"nova-cell0-db-create-vlp6r\" (UID: \"9118baf0-df6a-41e5-98e5-3e4d8edfbaaa\") " pod="openstack/nova-cell0-db-create-vlp6r" Mar 13 14:28:16 crc kubenswrapper[4907]: I0313 14:28:16.906501 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9118baf0-df6a-41e5-98e5-3e4d8edfbaaa-operator-scripts\") pod \"nova-cell0-db-create-vlp6r\" (UID: \"9118baf0-df6a-41e5-98e5-3e4d8edfbaaa\") " pod="openstack/nova-cell0-db-create-vlp6r" Mar 13 14:28:16 crc kubenswrapper[4907]: I0313 14:28:16.908480 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/251ab76e-b9df-406f-9b07-c7fa6e227feb-operator-scripts\") pod \"nova-api-db-create-g96mx\" (UID: \"251ab76e-b9df-406f-9b07-c7fa6e227feb\") " pod="openstack/nova-api-db-create-g96mx" Mar 13 14:28:16 crc kubenswrapper[4907]: I0313 14:28:16.920184 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4eeb2fb2-22fb-4ea8-a39b-5f959877cfe6-operator-scripts\") pod \"nova-api-22bc-account-create-update-r969t\" (UID: \"4eeb2fb2-22fb-4ea8-a39b-5f959877cfe6\") " pod="openstack/nova-api-22bc-account-create-update-r969t" Mar 13 14:28:16 crc kubenswrapper[4907]: I0313 14:28:16.920301 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5dcaf7c6-b4dd-4312-b68b-a833a827a2c2-operator-scripts\") pod \"nova-cell1-db-create-ldbt9\" (UID: \"5dcaf7c6-b4dd-4312-b68b-a833a827a2c2\") " pod="openstack/nova-cell1-db-create-ldbt9" Mar 13 14:28:16 crc kubenswrapper[4907]: I0313 14:28:16.920360 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-swn2k\" (UniqueName: \"kubernetes.io/projected/9118baf0-df6a-41e5-98e5-3e4d8edfbaaa-kube-api-access-swn2k\") pod \"nova-cell0-db-create-vlp6r\" (UID: \"9118baf0-df6a-41e5-98e5-3e4d8edfbaaa\") " pod="openstack/nova-cell0-db-create-vlp6r" Mar 13 14:28:16 crc kubenswrapper[4907]: I0313 14:28:16.923710 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4eeb2fb2-22fb-4ea8-a39b-5f959877cfe6-operator-scripts\") pod \"nova-api-22bc-account-create-update-r969t\" (UID: \"4eeb2fb2-22fb-4ea8-a39b-5f959877cfe6\") " pod="openstack/nova-api-22bc-account-create-update-r969t" Mar 13 14:28:16 crc kubenswrapper[4907]: I0313 14:28:16.927933 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4pm5m\" (UniqueName: \"kubernetes.io/projected/4eeb2fb2-22fb-4ea8-a39b-5f959877cfe6-kube-api-access-4pm5m\") pod \"nova-api-22bc-account-create-update-r969t\" (UID: \"4eeb2fb2-22fb-4ea8-a39b-5f959877cfe6\") " pod="openstack/nova-api-22bc-account-create-update-r969t" Mar 13 14:28:16 crc kubenswrapper[4907]: I0313 14:28:16.933765 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mjhh8\" (UniqueName: \"kubernetes.io/projected/251ab76e-b9df-406f-9b07-c7fa6e227feb-kube-api-access-mjhh8\") pod \"nova-api-db-create-g96mx\" (UID: \"251ab76e-b9df-406f-9b07-c7fa6e227feb\") " pod="openstack/nova-api-db-create-g96mx" Mar 13 14:28:16 crc kubenswrapper[4907]: I0313 14:28:16.941827 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-swn2k\" (UniqueName: \"kubernetes.io/projected/9118baf0-df6a-41e5-98e5-3e4d8edfbaaa-kube-api-access-swn2k\") pod \"nova-cell0-db-create-vlp6r\" (UID: \"9118baf0-df6a-41e5-98e5-3e4d8edfbaaa\") " pod="openstack/nova-cell0-db-create-vlp6r" Mar 13 14:28:16 crc kubenswrapper[4907]: I0313 14:28:16.943671 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-g96mx" Mar 13 14:28:16 crc kubenswrapper[4907]: I0313 14:28:16.949667 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-f148-account-create-update-wgw4b"] Mar 13 14:28:16 crc kubenswrapper[4907]: I0313 14:28:16.960746 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-f148-account-create-update-wgw4b" Mar 13 14:28:16 crc kubenswrapper[4907]: I0313 14:28:16.966243 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Mar 13 14:28:16 crc kubenswrapper[4907]: I0313 14:28:16.975366 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-f148-account-create-update-wgw4b"] Mar 13 14:28:17 crc kubenswrapper[4907]: I0313 14:28:17.022355 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5dcaf7c6-b4dd-4312-b68b-a833a827a2c2-operator-scripts\") pod \"nova-cell1-db-create-ldbt9\" (UID: \"5dcaf7c6-b4dd-4312-b68b-a833a827a2c2\") " pod="openstack/nova-cell1-db-create-ldbt9" Mar 13 14:28:17 crc kubenswrapper[4907]: I0313 14:28:17.022499 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wj4tn\" (UniqueName: \"kubernetes.io/projected/5dcaf7c6-b4dd-4312-b68b-a833a827a2c2-kube-api-access-wj4tn\") pod \"nova-cell1-db-create-ldbt9\" (UID: \"5dcaf7c6-b4dd-4312-b68b-a833a827a2c2\") " pod="openstack/nova-cell1-db-create-ldbt9" Mar 13 14:28:17 crc kubenswrapper[4907]: I0313 14:28:17.022544 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/85913727-bbcd-4343-9faf-a75f40b42dc8-operator-scripts\") pod \"nova-cell0-f148-account-create-update-wgw4b\" (UID: \"85913727-bbcd-4343-9faf-a75f40b42dc8\") " pod="openstack/nova-cell0-f148-account-create-update-wgw4b" Mar 13 14:28:17 crc kubenswrapper[4907]: I0313 14:28:17.022575 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cnlpw\" (UniqueName: \"kubernetes.io/projected/85913727-bbcd-4343-9faf-a75f40b42dc8-kube-api-access-cnlpw\") pod \"nova-cell0-f148-account-create-update-wgw4b\" (UID: \"85913727-bbcd-4343-9faf-a75f40b42dc8\") " pod="openstack/nova-cell0-f148-account-create-update-wgw4b" Mar 13 14:28:17 crc kubenswrapper[4907]: I0313 14:28:17.023417 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5dcaf7c6-b4dd-4312-b68b-a833a827a2c2-operator-scripts\") pod \"nova-cell1-db-create-ldbt9\" (UID: \"5dcaf7c6-b4dd-4312-b68b-a833a827a2c2\") " pod="openstack/nova-cell1-db-create-ldbt9" Mar 13 14:28:17 crc kubenswrapper[4907]: I0313 14:28:17.044166 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wj4tn\" (UniqueName: \"kubernetes.io/projected/5dcaf7c6-b4dd-4312-b68b-a833a827a2c2-kube-api-access-wj4tn\") pod \"nova-cell1-db-create-ldbt9\" (UID: \"5dcaf7c6-b4dd-4312-b68b-a833a827a2c2\") " pod="openstack/nova-cell1-db-create-ldbt9" Mar 13 14:28:17 crc kubenswrapper[4907]: I0313 14:28:17.056937 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-vlp6r" Mar 13 14:28:17 crc kubenswrapper[4907]: I0313 14:28:17.065848 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-22bc-account-create-update-r969t" Mar 13 14:28:17 crc kubenswrapper[4907]: I0313 14:28:17.128848 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/85913727-bbcd-4343-9faf-a75f40b42dc8-operator-scripts\") pod \"nova-cell0-f148-account-create-update-wgw4b\" (UID: \"85913727-bbcd-4343-9faf-a75f40b42dc8\") " pod="openstack/nova-cell0-f148-account-create-update-wgw4b" Mar 13 14:28:17 crc kubenswrapper[4907]: I0313 14:28:17.128908 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cnlpw\" (UniqueName: \"kubernetes.io/projected/85913727-bbcd-4343-9faf-a75f40b42dc8-kube-api-access-cnlpw\") pod \"nova-cell0-f148-account-create-update-wgw4b\" (UID: \"85913727-bbcd-4343-9faf-a75f40b42dc8\") " pod="openstack/nova-cell0-f148-account-create-update-wgw4b" Mar 13 14:28:17 crc kubenswrapper[4907]: I0313 14:28:17.129838 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/85913727-bbcd-4343-9faf-a75f40b42dc8-operator-scripts\") pod \"nova-cell0-f148-account-create-update-wgw4b\" (UID: \"85913727-bbcd-4343-9faf-a75f40b42dc8\") " pod="openstack/nova-cell0-f148-account-create-update-wgw4b" Mar 13 14:28:17 crc kubenswrapper[4907]: I0313 14:28:17.163091 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-ldbt9" Mar 13 14:28:17 crc kubenswrapper[4907]: I0313 14:28:17.166790 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cnlpw\" (UniqueName: \"kubernetes.io/projected/85913727-bbcd-4343-9faf-a75f40b42dc8-kube-api-access-cnlpw\") pod \"nova-cell0-f148-account-create-update-wgw4b\" (UID: \"85913727-bbcd-4343-9faf-a75f40b42dc8\") " pod="openstack/nova-cell0-f148-account-create-update-wgw4b" Mar 13 14:28:17 crc kubenswrapper[4907]: I0313 14:28:17.183476 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-f6fe-account-create-update-56786"] Mar 13 14:28:17 crc kubenswrapper[4907]: I0313 14:28:17.185026 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-f6fe-account-create-update-56786" Mar 13 14:28:17 crc kubenswrapper[4907]: I0313 14:28:17.190366 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Mar 13 14:28:17 crc kubenswrapper[4907]: I0313 14:28:17.193845 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-f6fe-account-create-update-56786"] Mar 13 14:28:17 crc kubenswrapper[4907]: I0313 14:28:17.333415 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7af00fde-8a92-4530-8fa8-9df7ff787c93-operator-scripts\") pod \"nova-cell1-f6fe-account-create-update-56786\" (UID: \"7af00fde-8a92-4530-8fa8-9df7ff787c93\") " pod="openstack/nova-cell1-f6fe-account-create-update-56786" Mar 13 14:28:17 crc kubenswrapper[4907]: I0313 14:28:17.335240 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dn4dc\" (UniqueName: \"kubernetes.io/projected/7af00fde-8a92-4530-8fa8-9df7ff787c93-kube-api-access-dn4dc\") pod \"nova-cell1-f6fe-account-create-update-56786\" (UID: \"7af00fde-8a92-4530-8fa8-9df7ff787c93\") " pod="openstack/nova-cell1-f6fe-account-create-update-56786" Mar 13 14:28:17 crc kubenswrapper[4907]: I0313 14:28:17.336178 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-f148-account-create-update-wgw4b" Mar 13 14:28:17 crc kubenswrapper[4907]: I0313 14:28:17.435984 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dn4dc\" (UniqueName: \"kubernetes.io/projected/7af00fde-8a92-4530-8fa8-9df7ff787c93-kube-api-access-dn4dc\") pod \"nova-cell1-f6fe-account-create-update-56786\" (UID: \"7af00fde-8a92-4530-8fa8-9df7ff787c93\") " pod="openstack/nova-cell1-f6fe-account-create-update-56786" Mar 13 14:28:17 crc kubenswrapper[4907]: I0313 14:28:17.436061 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7af00fde-8a92-4530-8fa8-9df7ff787c93-operator-scripts\") pod \"nova-cell1-f6fe-account-create-update-56786\" (UID: \"7af00fde-8a92-4530-8fa8-9df7ff787c93\") " pod="openstack/nova-cell1-f6fe-account-create-update-56786" Mar 13 14:28:17 crc kubenswrapper[4907]: I0313 14:28:17.436684 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7af00fde-8a92-4530-8fa8-9df7ff787c93-operator-scripts\") pod \"nova-cell1-f6fe-account-create-update-56786\" (UID: \"7af00fde-8a92-4530-8fa8-9df7ff787c93\") " pod="openstack/nova-cell1-f6fe-account-create-update-56786" Mar 13 14:28:17 crc kubenswrapper[4907]: I0313 14:28:17.458448 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dn4dc\" (UniqueName: \"kubernetes.io/projected/7af00fde-8a92-4530-8fa8-9df7ff787c93-kube-api-access-dn4dc\") pod \"nova-cell1-f6fe-account-create-update-56786\" (UID: \"7af00fde-8a92-4530-8fa8-9df7ff787c93\") " pod="openstack/nova-cell1-f6fe-account-create-update-56786" Mar 13 14:28:17 crc kubenswrapper[4907]: I0313 14:28:17.520169 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-f6fe-account-create-update-56786" Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.051146 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.051205 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.361999 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.470565 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c36a352-151f-4e93-8094-3855bfed532e-combined-ca-bundle\") pod \"5c36a352-151f-4e93-8094-3855bfed532e\" (UID: \"5c36a352-151f-4e93-8094-3855bfed532e\") " Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.470703 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5c36a352-151f-4e93-8094-3855bfed532e-internal-tls-certs\") pod \"5c36a352-151f-4e93-8094-3855bfed532e\" (UID: \"5c36a352-151f-4e93-8094-3855bfed532e\") " Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.470759 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5c36a352-151f-4e93-8094-3855bfed532e-scripts\") pod \"5c36a352-151f-4e93-8094-3855bfed532e\" (UID: \"5c36a352-151f-4e93-8094-3855bfed532e\") " Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.470830 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c36a352-151f-4e93-8094-3855bfed532e-config-data\") pod \"5c36a352-151f-4e93-8094-3855bfed532e\" (UID: \"5c36a352-151f-4e93-8094-3855bfed532e\") " Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.470859 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5c36a352-151f-4e93-8094-3855bfed532e-logs\") pod \"5c36a352-151f-4e93-8094-3855bfed532e\" (UID: \"5c36a352-151f-4e93-8094-3855bfed532e\") " Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.470941 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"5c36a352-151f-4e93-8094-3855bfed532e\" (UID: \"5c36a352-151f-4e93-8094-3855bfed532e\") " Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.470979 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5c36a352-151f-4e93-8094-3855bfed532e-httpd-run\") pod \"5c36a352-151f-4e93-8094-3855bfed532e\" (UID: \"5c36a352-151f-4e93-8094-3855bfed532e\") " Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.471014 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f7gh4\" (UniqueName: \"kubernetes.io/projected/5c36a352-151f-4e93-8094-3855bfed532e-kube-api-access-f7gh4\") pod \"5c36a352-151f-4e93-8094-3855bfed532e\" (UID: \"5c36a352-151f-4e93-8094-3855bfed532e\") " Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.472669 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5c36a352-151f-4e93-8094-3855bfed532e-logs" (OuterVolumeSpecName: "logs") pod "5c36a352-151f-4e93-8094-3855bfed532e" (UID: "5c36a352-151f-4e93-8094-3855bfed532e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.474131 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5c36a352-151f-4e93-8094-3855bfed532e-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "5c36a352-151f-4e93-8094-3855bfed532e" (UID: "5c36a352-151f-4e93-8094-3855bfed532e"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.490567 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c36a352-151f-4e93-8094-3855bfed532e-scripts" (OuterVolumeSpecName: "scripts") pod "5c36a352-151f-4e93-8094-3855bfed532e" (UID: "5c36a352-151f-4e93-8094-3855bfed532e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.514601 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "glance") pod "5c36a352-151f-4e93-8094-3855bfed532e" (UID: "5c36a352-151f-4e93-8094-3855bfed532e"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.524926 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5c36a352-151f-4e93-8094-3855bfed532e-kube-api-access-f7gh4" (OuterVolumeSpecName: "kube-api-access-f7gh4") pod "5c36a352-151f-4e93-8094-3855bfed532e" (UID: "5c36a352-151f-4e93-8094-3855bfed532e"). InnerVolumeSpecName "kube-api-access-f7gh4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.575604 4907 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5c36a352-151f-4e93-8094-3855bfed532e-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.575641 4907 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5c36a352-151f-4e93-8094-3855bfed532e-logs\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.575676 4907 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.575692 4907 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5c36a352-151f-4e93-8094-3855bfed532e-httpd-run\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.575704 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f7gh4\" (UniqueName: \"kubernetes.io/projected/5c36a352-151f-4e93-8094-3855bfed532e-kube-api-access-f7gh4\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.594944 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c36a352-151f-4e93-8094-3855bfed532e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5c36a352-151f-4e93-8094-3855bfed532e" (UID: "5c36a352-151f-4e93-8094-3855bfed532e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.630110 4907 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.663723 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"a055ce7f-0538-4e5e-938a-eaf6d0aa2dec","Type":"ContainerStarted","Data":"3e056a670aa7307d9d211e2123e0cead9b6cd31175ff09e1feddb508c47f1a88"} Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.671129 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c36a352-151f-4e93-8094-3855bfed532e-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "5c36a352-151f-4e93-8094-3855bfed532e" (UID: "5c36a352-151f-4e93-8094-3855bfed532e"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.684324 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.687496 4907 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.687782 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c36a352-151f-4e93-8094-3855bfed532e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.687923 4907 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5c36a352-151f-4e93-8094-3855bfed532e-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.702189 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.709656 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5c36a352-151f-4e93-8094-3855bfed532e","Type":"ContainerDied","Data":"4cb91fefa1e0ea486010258682ba6df5f6a7f72f9a8eaedd21f297ae14bacf23"} Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.710135 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-7dfdbcdd5-dcb7n" Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.710908 4907 scope.go:117] "RemoveContainer" containerID="456e65b2b22522c83f37e28d27a38f3af4c42d273550f8c8c1246de8d14f0a2f" Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.711146 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-7dfdbcdd5-dcb7n" Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.712923 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.761614914 podStartE2EDuration="16.712897554s" podCreationTimestamp="2026-03-13 14:28:02 +0000 UTC" firstStartedPulling="2026-03-13 14:28:03.903274809 +0000 UTC m=+1382.803062508" lastFinishedPulling="2026-03-13 14:28:17.854557459 +0000 UTC m=+1396.754345148" observedRunningTime="2026-03-13 14:28:18.685603681 +0000 UTC m=+1397.585391390" watchObservedRunningTime="2026-03-13 14:28:18.712897554 +0000 UTC m=+1397.612685273" Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.717161 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f7ca16f6-c9d9-4415-b8cf-b4753d784170","Type":"ContainerStarted","Data":"53d201dd158c6af9aadd5213934f6cc439d1b5e00b0a3e3af8f7f6824b4594d2"} Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.736482 4907 generic.go:334] "Generic (PLEG): container finished" podID="10a74a61-272c-4c54-9b2a-6379b77c9984" containerID="a7d63138378c0ce32dc24735fd0ab7db352755405ba476da24cb2065858f3b75" exitCode=0 Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.736573 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-944cbc5bb-fj29g" event={"ID":"10a74a61-272c-4c54-9b2a-6379b77c9984","Type":"ContainerDied","Data":"a7d63138378c0ce32dc24735fd0ab7db352755405ba476da24cb2065858f3b75"} Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.785907 4907 scope.go:117] "RemoveContainer" containerID="4364cee9bed454e59183eb0004b9dfc4ab15dc5a345acbfd2a54032342758450" Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.787606 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c36a352-151f-4e93-8094-3855bfed532e-config-data" (OuterVolumeSpecName: "config-data") pod "5c36a352-151f-4e93-8094-3855bfed532e" (UID: "5c36a352-151f-4e93-8094-3855bfed532e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.797924 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c9261af-9eaa-4623-9f1b-719b58a9d3a2-combined-ca-bundle\") pod \"2c9261af-9eaa-4623-9f1b-719b58a9d3a2\" (UID: \"2c9261af-9eaa-4623-9f1b-719b58a9d3a2\") " Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.797973 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2c9261af-9eaa-4623-9f1b-719b58a9d3a2-public-tls-certs\") pod \"2c9261af-9eaa-4623-9f1b-719b58a9d3a2\" (UID: \"2c9261af-9eaa-4623-9f1b-719b58a9d3a2\") " Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.798486 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2c9261af-9eaa-4623-9f1b-719b58a9d3a2-logs\") pod \"2c9261af-9eaa-4623-9f1b-719b58a9d3a2\" (UID: \"2c9261af-9eaa-4623-9f1b-719b58a9d3a2\") " Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.798546 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"2c9261af-9eaa-4623-9f1b-719b58a9d3a2\" (UID: \"2c9261af-9eaa-4623-9f1b-719b58a9d3a2\") " Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.798783 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7fdtd\" (UniqueName: \"kubernetes.io/projected/2c9261af-9eaa-4623-9f1b-719b58a9d3a2-kube-api-access-7fdtd\") pod \"2c9261af-9eaa-4623-9f1b-719b58a9d3a2\" (UID: \"2c9261af-9eaa-4623-9f1b-719b58a9d3a2\") " Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.798818 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2c9261af-9eaa-4623-9f1b-719b58a9d3a2-scripts\") pod \"2c9261af-9eaa-4623-9f1b-719b58a9d3a2\" (UID: \"2c9261af-9eaa-4623-9f1b-719b58a9d3a2\") " Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.798873 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2c9261af-9eaa-4623-9f1b-719b58a9d3a2-httpd-run\") pod \"2c9261af-9eaa-4623-9f1b-719b58a9d3a2\" (UID: \"2c9261af-9eaa-4623-9f1b-719b58a9d3a2\") " Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.798919 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c9261af-9eaa-4623-9f1b-719b58a9d3a2-config-data\") pod \"2c9261af-9eaa-4623-9f1b-719b58a9d3a2\" (UID: \"2c9261af-9eaa-4623-9f1b-719b58a9d3a2\") " Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.799734 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c36a352-151f-4e93-8094-3855bfed532e-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.802284 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2c9261af-9eaa-4623-9f1b-719b58a9d3a2-logs" (OuterVolumeSpecName: "logs") pod "2c9261af-9eaa-4623-9f1b-719b58a9d3a2" (UID: "2c9261af-9eaa-4623-9f1b-719b58a9d3a2"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.803035 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "glance") pod "2c9261af-9eaa-4623-9f1b-719b58a9d3a2" (UID: "2c9261af-9eaa-4623-9f1b-719b58a9d3a2"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.803199 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2c9261af-9eaa-4623-9f1b-719b58a9d3a2-kube-api-access-7fdtd" (OuterVolumeSpecName: "kube-api-access-7fdtd") pod "2c9261af-9eaa-4623-9f1b-719b58a9d3a2" (UID: "2c9261af-9eaa-4623-9f1b-719b58a9d3a2"). InnerVolumeSpecName "kube-api-access-7fdtd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.804026 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2c9261af-9eaa-4623-9f1b-719b58a9d3a2-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "2c9261af-9eaa-4623-9f1b-719b58a9d3a2" (UID: "2c9261af-9eaa-4623-9f1b-719b58a9d3a2"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.810490 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c9261af-9eaa-4623-9f1b-719b58a9d3a2-scripts" (OuterVolumeSpecName: "scripts") pod "2c9261af-9eaa-4623-9f1b-719b58a9d3a2" (UID: "2c9261af-9eaa-4623-9f1b-719b58a9d3a2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.845050 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c9261af-9eaa-4623-9f1b-719b58a9d3a2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2c9261af-9eaa-4623-9f1b-719b58a9d3a2" (UID: "2c9261af-9eaa-4623-9f1b-719b58a9d3a2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.865916 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c9261af-9eaa-4623-9f1b-719b58a9d3a2-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "2c9261af-9eaa-4623-9f1b-719b58a9d3a2" (UID: "2c9261af-9eaa-4623-9f1b-719b58a9d3a2"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.877026 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c9261af-9eaa-4623-9f1b-719b58a9d3a2-config-data" (OuterVolumeSpecName: "config-data") pod "2c9261af-9eaa-4623-9f1b-719b58a9d3a2" (UID: "2c9261af-9eaa-4623-9f1b-719b58a9d3a2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.902112 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7fdtd\" (UniqueName: \"kubernetes.io/projected/2c9261af-9eaa-4623-9f1b-719b58a9d3a2-kube-api-access-7fdtd\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.902156 4907 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2c9261af-9eaa-4623-9f1b-719b58a9d3a2-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.902169 4907 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2c9261af-9eaa-4623-9f1b-719b58a9d3a2-httpd-run\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.902183 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c9261af-9eaa-4623-9f1b-719b58a9d3a2-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.902194 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c9261af-9eaa-4623-9f1b-719b58a9d3a2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.902208 4907 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2c9261af-9eaa-4623-9f1b-719b58a9d3a2-public-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.902220 4907 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2c9261af-9eaa-4623-9f1b-719b58a9d3a2-logs\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.902257 4907 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.934705 4907 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.005944 4907 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.066936 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.115838 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.135935 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 13 14:28:19 crc kubenswrapper[4907]: E0313 14:28:19.136350 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c36a352-151f-4e93-8094-3855bfed532e" containerName="glance-httpd" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.136362 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c36a352-151f-4e93-8094-3855bfed532e" containerName="glance-httpd" Mar 13 14:28:19 crc kubenswrapper[4907]: E0313 14:28:19.136374 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c9261af-9eaa-4623-9f1b-719b58a9d3a2" containerName="glance-httpd" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.136380 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c9261af-9eaa-4623-9f1b-719b58a9d3a2" containerName="glance-httpd" Mar 13 14:28:19 crc kubenswrapper[4907]: E0313 14:28:19.136391 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c36a352-151f-4e93-8094-3855bfed532e" containerName="glance-log" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.136396 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c36a352-151f-4e93-8094-3855bfed532e" containerName="glance-log" Mar 13 14:28:19 crc kubenswrapper[4907]: E0313 14:28:19.136421 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c9261af-9eaa-4623-9f1b-719b58a9d3a2" containerName="glance-log" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.136427 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c9261af-9eaa-4623-9f1b-719b58a9d3a2" containerName="glance-log" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.136593 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c36a352-151f-4e93-8094-3855bfed532e" containerName="glance-log" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.136604 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c36a352-151f-4e93-8094-3855bfed532e" containerName="glance-httpd" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.136617 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c9261af-9eaa-4623-9f1b-719b58a9d3a2" containerName="glance-log" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.136628 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c9261af-9eaa-4623-9f1b-719b58a9d3a2" containerName="glance-httpd" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.137650 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.139166 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.141314 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.148937 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.175647 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-944cbc5bb-fj29g" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.210781 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gsjwt\" (UniqueName: \"kubernetes.io/projected/10a74a61-272c-4c54-9b2a-6379b77c9984-kube-api-access-gsjwt\") pod \"10a74a61-272c-4c54-9b2a-6379b77c9984\" (UID: \"10a74a61-272c-4c54-9b2a-6379b77c9984\") " Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.210845 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10a74a61-272c-4c54-9b2a-6379b77c9984-combined-ca-bundle\") pod \"10a74a61-272c-4c54-9b2a-6379b77c9984\" (UID: \"10a74a61-272c-4c54-9b2a-6379b77c9984\") " Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.210916 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/10a74a61-272c-4c54-9b2a-6379b77c9984-ovndb-tls-certs\") pod \"10a74a61-272c-4c54-9b2a-6379b77c9984\" (UID: \"10a74a61-272c-4c54-9b2a-6379b77c9984\") " Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.210955 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/10a74a61-272c-4c54-9b2a-6379b77c9984-httpd-config\") pod \"10a74a61-272c-4c54-9b2a-6379b77c9984\" (UID: \"10a74a61-272c-4c54-9b2a-6379b77c9984\") " Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.211033 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/10a74a61-272c-4c54-9b2a-6379b77c9984-config\") pod \"10a74a61-272c-4c54-9b2a-6379b77c9984\" (UID: \"10a74a61-272c-4c54-9b2a-6379b77c9984\") " Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.211456 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4251ae36-90ad-41ea-915e-862df60f5c07-scripts\") pod \"glance-default-internal-api-0\" (UID: \"4251ae36-90ad-41ea-915e-862df60f5c07\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.211701 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4251ae36-90ad-41ea-915e-862df60f5c07-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"4251ae36-90ad-41ea-915e-862df60f5c07\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.211753 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h2bph\" (UniqueName: \"kubernetes.io/projected/4251ae36-90ad-41ea-915e-862df60f5c07-kube-api-access-h2bph\") pod \"glance-default-internal-api-0\" (UID: \"4251ae36-90ad-41ea-915e-862df60f5c07\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.211808 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4251ae36-90ad-41ea-915e-862df60f5c07-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"4251ae36-90ad-41ea-915e-862df60f5c07\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.211836 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4251ae36-90ad-41ea-915e-862df60f5c07-logs\") pod \"glance-default-internal-api-0\" (UID: \"4251ae36-90ad-41ea-915e-862df60f5c07\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.211926 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4251ae36-90ad-41ea-915e-862df60f5c07-config-data\") pod \"glance-default-internal-api-0\" (UID: \"4251ae36-90ad-41ea-915e-862df60f5c07\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.211957 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4251ae36-90ad-41ea-915e-862df60f5c07-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"4251ae36-90ad-41ea-915e-862df60f5c07\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.212027 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"4251ae36-90ad-41ea-915e-862df60f5c07\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.233005 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/10a74a61-272c-4c54-9b2a-6379b77c9984-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "10a74a61-272c-4c54-9b2a-6379b77c9984" (UID: "10a74a61-272c-4c54-9b2a-6379b77c9984"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.235224 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/10a74a61-272c-4c54-9b2a-6379b77c9984-kube-api-access-gsjwt" (OuterVolumeSpecName: "kube-api-access-gsjwt") pod "10a74a61-272c-4c54-9b2a-6379b77c9984" (UID: "10a74a61-272c-4c54-9b2a-6379b77c9984"). InnerVolumeSpecName "kube-api-access-gsjwt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.245755 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-vlp6r"] Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.276573 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-g96mx"] Mar 13 14:28:19 crc kubenswrapper[4907]: W0313 14:28:19.276567 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4eeb2fb2_22fb_4ea8_a39b_5f959877cfe6.slice/crio-46336f79de33964f45a2c37528d8472d8d5d46341b5f9fc4bc15531e76f002ba WatchSource:0}: Error finding container 46336f79de33964f45a2c37528d8472d8d5d46341b5f9fc4bc15531e76f002ba: Status 404 returned error can't find the container with id 46336f79de33964f45a2c37528d8472d8d5d46341b5f9fc4bc15531e76f002ba Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.292359 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-f148-account-create-update-wgw4b"] Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.300597 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-f6fe-account-create-update-56786"] Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.313202 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4251ae36-90ad-41ea-915e-862df60f5c07-config-data\") pod \"glance-default-internal-api-0\" (UID: \"4251ae36-90ad-41ea-915e-862df60f5c07\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.313249 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4251ae36-90ad-41ea-915e-862df60f5c07-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"4251ae36-90ad-41ea-915e-862df60f5c07\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.313288 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"4251ae36-90ad-41ea-915e-862df60f5c07\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.313321 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4251ae36-90ad-41ea-915e-862df60f5c07-scripts\") pod \"glance-default-internal-api-0\" (UID: \"4251ae36-90ad-41ea-915e-862df60f5c07\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.313361 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4251ae36-90ad-41ea-915e-862df60f5c07-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"4251ae36-90ad-41ea-915e-862df60f5c07\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.313395 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h2bph\" (UniqueName: \"kubernetes.io/projected/4251ae36-90ad-41ea-915e-862df60f5c07-kube-api-access-h2bph\") pod \"glance-default-internal-api-0\" (UID: \"4251ae36-90ad-41ea-915e-862df60f5c07\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.313426 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4251ae36-90ad-41ea-915e-862df60f5c07-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"4251ae36-90ad-41ea-915e-862df60f5c07\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.313443 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4251ae36-90ad-41ea-915e-862df60f5c07-logs\") pod \"glance-default-internal-api-0\" (UID: \"4251ae36-90ad-41ea-915e-862df60f5c07\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.313513 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gsjwt\" (UniqueName: \"kubernetes.io/projected/10a74a61-272c-4c54-9b2a-6379b77c9984-kube-api-access-gsjwt\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.313523 4907 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/10a74a61-272c-4c54-9b2a-6379b77c9984-httpd-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.313962 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4251ae36-90ad-41ea-915e-862df60f5c07-logs\") pod \"glance-default-internal-api-0\" (UID: \"4251ae36-90ad-41ea-915e-862df60f5c07\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.320210 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4251ae36-90ad-41ea-915e-862df60f5c07-config-data\") pod \"glance-default-internal-api-0\" (UID: \"4251ae36-90ad-41ea-915e-862df60f5c07\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.326465 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4251ae36-90ad-41ea-915e-862df60f5c07-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"4251ae36-90ad-41ea-915e-862df60f5c07\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.326728 4907 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"4251ae36-90ad-41ea-915e-862df60f5c07\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/glance-default-internal-api-0" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.328970 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-22bc-account-create-update-r969t"] Mar 13 14:28:19 crc kubenswrapper[4907]: W0313 14:28:19.329155 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5dcaf7c6_b4dd_4312_b68b_a833a827a2c2.slice/crio-fa6acba78f266f6ff1774ad62b03c8a5d55736cb5f7697c6fcc3a1a5a0614819 WatchSource:0}: Error finding container fa6acba78f266f6ff1774ad62b03c8a5d55736cb5f7697c6fcc3a1a5a0614819: Status 404 returned error can't find the container with id fa6acba78f266f6ff1774ad62b03c8a5d55736cb5f7697c6fcc3a1a5a0614819 Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.331683 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4251ae36-90ad-41ea-915e-862df60f5c07-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"4251ae36-90ad-41ea-915e-862df60f5c07\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.333501 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4251ae36-90ad-41ea-915e-862df60f5c07-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"4251ae36-90ad-41ea-915e-862df60f5c07\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.337568 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4251ae36-90ad-41ea-915e-862df60f5c07-scripts\") pod \"glance-default-internal-api-0\" (UID: \"4251ae36-90ad-41ea-915e-862df60f5c07\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.351622 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h2bph\" (UniqueName: \"kubernetes.io/projected/4251ae36-90ad-41ea-915e-862df60f5c07-kube-api-access-h2bph\") pod \"glance-default-internal-api-0\" (UID: \"4251ae36-90ad-41ea-915e-862df60f5c07\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.389019 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/10a74a61-272c-4c54-9b2a-6379b77c9984-config" (OuterVolumeSpecName: "config") pod "10a74a61-272c-4c54-9b2a-6379b77c9984" (UID: "10a74a61-272c-4c54-9b2a-6379b77c9984"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.407201 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-ldbt9"] Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.414690 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/10a74a61-272c-4c54-9b2a-6379b77c9984-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.445645 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/10a74a61-272c-4c54-9b2a-6379b77c9984-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "10a74a61-272c-4c54-9b2a-6379b77c9984" (UID: "10a74a61-272c-4c54-9b2a-6379b77c9984"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.462549 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"4251ae36-90ad-41ea-915e-862df60f5c07\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.472076 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/10a74a61-272c-4c54-9b2a-6379b77c9984-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "10a74a61-272c-4c54-9b2a-6379b77c9984" (UID: "10a74a61-272c-4c54-9b2a-6379b77c9984"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.491685 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.516519 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10a74a61-272c-4c54-9b2a-6379b77c9984-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.517104 4907 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/10a74a61-272c-4c54-9b2a-6379b77c9984-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.749982 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-22bc-account-create-update-r969t" event={"ID":"4eeb2fb2-22fb-4ea8-a39b-5f959877cfe6","Type":"ContainerStarted","Data":"46336f79de33964f45a2c37528d8472d8d5d46341b5f9fc4bc15531e76f002ba"} Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.762508 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2c9261af-9eaa-4623-9f1b-719b58a9d3a2","Type":"ContainerDied","Data":"2cde1712ac814253467a7ebeac79ad28f1b6bc8f06f7979a4876a959f01886bd"} Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.762562 4907 scope.go:117] "RemoveContainer" containerID="22707601406e7d79a6ae3140c68d6ec6d23c0168f7d4a1523dfe2e817903318f" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.762723 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.843213 4907 scope.go:117] "RemoveContainer" containerID="6af4b6388e126ccd4e3d2cc04b6747e5283033aaa5d424f9e2907953b3f55ea5" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.845369 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5c36a352-151f-4e93-8094-3855bfed532e" path="/var/lib/kubelet/pods/5c36a352-151f-4e93-8094-3855bfed532e/volumes" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.851922 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-944cbc5bb-fj29g" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.871203 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f7ca16f6-c9d9-4415-b8cf-b4753d784170","Type":"ContainerStarted","Data":"306d706e8b95e9a38af5ba674a473f59462492d56df0ecb87578b8edfdde9aa6"} Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.871253 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.871275 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.871292 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-ldbt9" event={"ID":"5dcaf7c6-b4dd-4312-b68b-a833a827a2c2","Type":"ContainerStarted","Data":"fa6acba78f266f6ff1774ad62b03c8a5d55736cb5f7697c6fcc3a1a5a0614819"} Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.871315 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-vlp6r" event={"ID":"9118baf0-df6a-41e5-98e5-3e4d8edfbaaa","Type":"ContainerStarted","Data":"3f5c4dcc986bdd72e19fe37fafad59be240102706caf33d75e722b2f347bd3e3"} Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.871332 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-vlp6r" event={"ID":"9118baf0-df6a-41e5-98e5-3e4d8edfbaaa","Type":"ContainerStarted","Data":"b75ab2628fbe2653fbc966d7e517d3692b8792f5300df76a97698a45b1204d93"} Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.871343 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-944cbc5bb-fj29g" event={"ID":"10a74a61-272c-4c54-9b2a-6379b77c9984","Type":"ContainerDied","Data":"b6a3399278c9114cdb41d239ecf9266ca04769841425ab359d3ec15a0ea2369a"} Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.871358 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-g96mx" event={"ID":"251ab76e-b9df-406f-9b07-c7fa6e227feb","Type":"ContainerStarted","Data":"107f7252eaea9f43079c72f27a3a08aa9bdfea92d244d423f8695cb45030deb6"} Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.871371 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-g96mx" event={"ID":"251ab76e-b9df-406f-9b07-c7fa6e227feb","Type":"ContainerStarted","Data":"57e7cdfdc101c4e36f2572edec180e1f998c526cf20ca796f61ebb46ef192472"} Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.875696 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-db-create-ldbt9" podStartSLOduration=3.875669518 podStartE2EDuration="3.875669518s" podCreationTimestamp="2026-03-13 14:28:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:28:19.845110048 +0000 UTC m=+1398.744897727" watchObservedRunningTime="2026-03-13 14:28:19.875669518 +0000 UTC m=+1398.775457207" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.890357 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-f6fe-account-create-update-56786" event={"ID":"7af00fde-8a92-4530-8fa8-9df7ff787c93","Type":"ContainerStarted","Data":"13e164891acf22f2bb718fb70e94321cdc023003038bdb8dbc28a6f2b84707b6"} Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.890402 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-f6fe-account-create-update-56786" event={"ID":"7af00fde-8a92-4530-8fa8-9df7ff787c93","Type":"ContainerStarted","Data":"5c53d7426e0e1950f78a3c70586fa0b2e353687a87ba5dbe4ff6e6fc12256f18"} Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.920168 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-f148-account-create-update-wgw4b" event={"ID":"85913727-bbcd-4343-9faf-a75f40b42dc8","Type":"ContainerStarted","Data":"ca4cf56d486af5cc064a60fd28eb3806cfa9fe3d0bef16b65c3fd811e37a4cb4"} Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.943637 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Mar 13 14:28:19 crc kubenswrapper[4907]: E0313 14:28:19.944662 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10a74a61-272c-4c54-9b2a-6379b77c9984" containerName="neutron-api" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.944681 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="10a74a61-272c-4c54-9b2a-6379b77c9984" containerName="neutron-api" Mar 13 14:28:19 crc kubenswrapper[4907]: E0313 14:28:19.944713 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10a74a61-272c-4c54-9b2a-6379b77c9984" containerName="neutron-httpd" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.944723 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="10a74a61-272c-4c54-9b2a-6379b77c9984" containerName="neutron-httpd" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.945107 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="10a74a61-272c-4c54-9b2a-6379b77c9984" containerName="neutron-api" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.945141 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="10a74a61-272c-4c54-9b2a-6379b77c9984" containerName="neutron-httpd" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.946787 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.949408 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.964834 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.977746 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.991132 4907 scope.go:117] "RemoveContainer" containerID="b624bd9d00cbcadf5cdc9520721c5131708177fac0d6006945f6f3ffc27df6cd" Mar 13 14:28:20 crc kubenswrapper[4907]: I0313 14:28:20.023587 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-db-create-vlp6r" podStartSLOduration=4.02353542 podStartE2EDuration="4.02353542s" podCreationTimestamp="2026-03-13 14:28:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:28:19.883156582 +0000 UTC m=+1398.782944271" watchObservedRunningTime="2026-03-13 14:28:20.02353542 +0000 UTC m=+1398.923323119" Mar 13 14:28:20 crc kubenswrapper[4907]: I0313 14:28:20.060001 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-db-create-g96mx" podStartSLOduration=4.059974091 podStartE2EDuration="4.059974091s" podCreationTimestamp="2026-03-13 14:28:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:28:19.904661177 +0000 UTC m=+1398.804448876" watchObservedRunningTime="2026-03-13 14:28:20.059974091 +0000 UTC m=+1398.959761780" Mar 13 14:28:20 crc kubenswrapper[4907]: I0313 14:28:20.072070 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-f6fe-account-create-update-56786" podStartSLOduration=3.072045449 podStartE2EDuration="3.072045449s" podCreationTimestamp="2026-03-13 14:28:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:28:19.93933101 +0000 UTC m=+1398.839118699" watchObservedRunningTime="2026-03-13 14:28:20.072045449 +0000 UTC m=+1398.971833138" Mar 13 14:28:20 crc kubenswrapper[4907]: I0313 14:28:20.103134 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-944cbc5bb-fj29g"] Mar 13 14:28:20 crc kubenswrapper[4907]: I0313 14:28:20.109222 4907 scope.go:117] "RemoveContainer" containerID="a7d63138378c0ce32dc24735fd0ab7db352755405ba476da24cb2065858f3b75" Mar 13 14:28:20 crc kubenswrapper[4907]: I0313 14:28:20.130467 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b2e62fa-af35-4f35-8852-362ea97c17e7-config-data\") pod \"glance-default-external-api-0\" (UID: \"0b2e62fa-af35-4f35-8852-362ea97c17e7\") " pod="openstack/glance-default-external-api-0" Mar 13 14:28:20 crc kubenswrapper[4907]: I0313 14:28:20.130514 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0b2e62fa-af35-4f35-8852-362ea97c17e7-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"0b2e62fa-af35-4f35-8852-362ea97c17e7\") " pod="openstack/glance-default-external-api-0" Mar 13 14:28:20 crc kubenswrapper[4907]: I0313 14:28:20.130548 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"0b2e62fa-af35-4f35-8852-362ea97c17e7\") " pod="openstack/glance-default-external-api-0" Mar 13 14:28:20 crc kubenswrapper[4907]: I0313 14:28:20.130569 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b2e62fa-af35-4f35-8852-362ea97c17e7-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"0b2e62fa-af35-4f35-8852-362ea97c17e7\") " pod="openstack/glance-default-external-api-0" Mar 13 14:28:20 crc kubenswrapper[4907]: I0313 14:28:20.130616 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0b2e62fa-af35-4f35-8852-362ea97c17e7-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"0b2e62fa-af35-4f35-8852-362ea97c17e7\") " pod="openstack/glance-default-external-api-0" Mar 13 14:28:20 crc kubenswrapper[4907]: I0313 14:28:20.130637 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2jswj\" (UniqueName: \"kubernetes.io/projected/0b2e62fa-af35-4f35-8852-362ea97c17e7-kube-api-access-2jswj\") pod \"glance-default-external-api-0\" (UID: \"0b2e62fa-af35-4f35-8852-362ea97c17e7\") " pod="openstack/glance-default-external-api-0" Mar 13 14:28:20 crc kubenswrapper[4907]: I0313 14:28:20.130679 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0b2e62fa-af35-4f35-8852-362ea97c17e7-scripts\") pod \"glance-default-external-api-0\" (UID: \"0b2e62fa-af35-4f35-8852-362ea97c17e7\") " pod="openstack/glance-default-external-api-0" Mar 13 14:28:20 crc kubenswrapper[4907]: I0313 14:28:20.130752 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0b2e62fa-af35-4f35-8852-362ea97c17e7-logs\") pod \"glance-default-external-api-0\" (UID: \"0b2e62fa-af35-4f35-8852-362ea97c17e7\") " pod="openstack/glance-default-external-api-0" Mar 13 14:28:20 crc kubenswrapper[4907]: I0313 14:28:20.136534 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-944cbc5bb-fj29g"] Mar 13 14:28:20 crc kubenswrapper[4907]: I0313 14:28:20.162391 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-f148-account-create-update-wgw4b" podStartSLOduration=4.162371505 podStartE2EDuration="4.162371505s" podCreationTimestamp="2026-03-13 14:28:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:28:19.983522881 +0000 UTC m=+1398.883310580" watchObservedRunningTime="2026-03-13 14:28:20.162371505 +0000 UTC m=+1399.062159194" Mar 13 14:28:20 crc kubenswrapper[4907]: I0313 14:28:20.203352 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 13 14:28:20 crc kubenswrapper[4907]: I0313 14:28:20.232344 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b2e62fa-af35-4f35-8852-362ea97c17e7-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"0b2e62fa-af35-4f35-8852-362ea97c17e7\") " pod="openstack/glance-default-external-api-0" Mar 13 14:28:20 crc kubenswrapper[4907]: I0313 14:28:20.232436 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0b2e62fa-af35-4f35-8852-362ea97c17e7-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"0b2e62fa-af35-4f35-8852-362ea97c17e7\") " pod="openstack/glance-default-external-api-0" Mar 13 14:28:20 crc kubenswrapper[4907]: I0313 14:28:20.232463 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2jswj\" (UniqueName: \"kubernetes.io/projected/0b2e62fa-af35-4f35-8852-362ea97c17e7-kube-api-access-2jswj\") pod \"glance-default-external-api-0\" (UID: \"0b2e62fa-af35-4f35-8852-362ea97c17e7\") " pod="openstack/glance-default-external-api-0" Mar 13 14:28:20 crc kubenswrapper[4907]: I0313 14:28:20.232479 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0b2e62fa-af35-4f35-8852-362ea97c17e7-scripts\") pod \"glance-default-external-api-0\" (UID: \"0b2e62fa-af35-4f35-8852-362ea97c17e7\") " pod="openstack/glance-default-external-api-0" Mar 13 14:28:20 crc kubenswrapper[4907]: I0313 14:28:20.232546 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0b2e62fa-af35-4f35-8852-362ea97c17e7-logs\") pod \"glance-default-external-api-0\" (UID: \"0b2e62fa-af35-4f35-8852-362ea97c17e7\") " pod="openstack/glance-default-external-api-0" Mar 13 14:28:20 crc kubenswrapper[4907]: I0313 14:28:20.232619 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b2e62fa-af35-4f35-8852-362ea97c17e7-config-data\") pod \"glance-default-external-api-0\" (UID: \"0b2e62fa-af35-4f35-8852-362ea97c17e7\") " pod="openstack/glance-default-external-api-0" Mar 13 14:28:20 crc kubenswrapper[4907]: I0313 14:28:20.232648 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0b2e62fa-af35-4f35-8852-362ea97c17e7-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"0b2e62fa-af35-4f35-8852-362ea97c17e7\") " pod="openstack/glance-default-external-api-0" Mar 13 14:28:20 crc kubenswrapper[4907]: I0313 14:28:20.232681 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"0b2e62fa-af35-4f35-8852-362ea97c17e7\") " pod="openstack/glance-default-external-api-0" Mar 13 14:28:20 crc kubenswrapper[4907]: I0313 14:28:20.233153 4907 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"0b2e62fa-af35-4f35-8852-362ea97c17e7\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/glance-default-external-api-0" Mar 13 14:28:20 crc kubenswrapper[4907]: I0313 14:28:20.234673 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0b2e62fa-af35-4f35-8852-362ea97c17e7-logs\") pod \"glance-default-external-api-0\" (UID: \"0b2e62fa-af35-4f35-8852-362ea97c17e7\") " pod="openstack/glance-default-external-api-0" Mar 13 14:28:20 crc kubenswrapper[4907]: I0313 14:28:20.234783 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0b2e62fa-af35-4f35-8852-362ea97c17e7-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"0b2e62fa-af35-4f35-8852-362ea97c17e7\") " pod="openstack/glance-default-external-api-0" Mar 13 14:28:20 crc kubenswrapper[4907]: I0313 14:28:20.238483 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b2e62fa-af35-4f35-8852-362ea97c17e7-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"0b2e62fa-af35-4f35-8852-362ea97c17e7\") " pod="openstack/glance-default-external-api-0" Mar 13 14:28:20 crc kubenswrapper[4907]: I0313 14:28:20.238625 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b2e62fa-af35-4f35-8852-362ea97c17e7-config-data\") pod \"glance-default-external-api-0\" (UID: \"0b2e62fa-af35-4f35-8852-362ea97c17e7\") " pod="openstack/glance-default-external-api-0" Mar 13 14:28:20 crc kubenswrapper[4907]: I0313 14:28:20.238693 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0b2e62fa-af35-4f35-8852-362ea97c17e7-scripts\") pod \"glance-default-external-api-0\" (UID: \"0b2e62fa-af35-4f35-8852-362ea97c17e7\") " pod="openstack/glance-default-external-api-0" Mar 13 14:28:20 crc kubenswrapper[4907]: I0313 14:28:20.241543 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0b2e62fa-af35-4f35-8852-362ea97c17e7-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"0b2e62fa-af35-4f35-8852-362ea97c17e7\") " pod="openstack/glance-default-external-api-0" Mar 13 14:28:20 crc kubenswrapper[4907]: I0313 14:28:20.256728 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2jswj\" (UniqueName: \"kubernetes.io/projected/0b2e62fa-af35-4f35-8852-362ea97c17e7-kube-api-access-2jswj\") pod \"glance-default-external-api-0\" (UID: \"0b2e62fa-af35-4f35-8852-362ea97c17e7\") " pod="openstack/glance-default-external-api-0" Mar 13 14:28:20 crc kubenswrapper[4907]: I0313 14:28:20.285505 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"0b2e62fa-af35-4f35-8852-362ea97c17e7\") " pod="openstack/glance-default-external-api-0" Mar 13 14:28:20 crc kubenswrapper[4907]: I0313 14:28:20.347815 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 13 14:28:20 crc kubenswrapper[4907]: I0313 14:28:20.931765 4907 generic.go:334] "Generic (PLEG): container finished" podID="4eeb2fb2-22fb-4ea8-a39b-5f959877cfe6" containerID="e156e90b4d1cbea6f3a55f14df88e09d8aeafa1d0449cf8de460be879c418cf0" exitCode=0 Mar 13 14:28:20 crc kubenswrapper[4907]: I0313 14:28:20.932918 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-22bc-account-create-update-r969t" event={"ID":"4eeb2fb2-22fb-4ea8-a39b-5f959877cfe6","Type":"ContainerDied","Data":"e156e90b4d1cbea6f3a55f14df88e09d8aeafa1d0449cf8de460be879c418cf0"} Mar 13 14:28:20 crc kubenswrapper[4907]: I0313 14:28:20.934262 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"4251ae36-90ad-41ea-915e-862df60f5c07","Type":"ContainerStarted","Data":"d3f2a65d466df300e932d28e3a4a5fb0f68b272bc80ae37edb5a080aa7144cae"} Mar 13 14:28:20 crc kubenswrapper[4907]: I0313 14:28:20.937406 4907 generic.go:334] "Generic (PLEG): container finished" podID="251ab76e-b9df-406f-9b07-c7fa6e227feb" containerID="107f7252eaea9f43079c72f27a3a08aa9bdfea92d244d423f8695cb45030deb6" exitCode=0 Mar 13 14:28:20 crc kubenswrapper[4907]: I0313 14:28:20.937589 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-g96mx" event={"ID":"251ab76e-b9df-406f-9b07-c7fa6e227feb","Type":"ContainerDied","Data":"107f7252eaea9f43079c72f27a3a08aa9bdfea92d244d423f8695cb45030deb6"} Mar 13 14:28:20 crc kubenswrapper[4907]: I0313 14:28:20.939402 4907 generic.go:334] "Generic (PLEG): container finished" podID="7af00fde-8a92-4530-8fa8-9df7ff787c93" containerID="13e164891acf22f2bb718fb70e94321cdc023003038bdb8dbc28a6f2b84707b6" exitCode=0 Mar 13 14:28:20 crc kubenswrapper[4907]: I0313 14:28:20.939474 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-f6fe-account-create-update-56786" event={"ID":"7af00fde-8a92-4530-8fa8-9df7ff787c93","Type":"ContainerDied","Data":"13e164891acf22f2bb718fb70e94321cdc023003038bdb8dbc28a6f2b84707b6"} Mar 13 14:28:20 crc kubenswrapper[4907]: I0313 14:28:20.944233 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-f148-account-create-update-wgw4b" event={"ID":"85913727-bbcd-4343-9faf-a75f40b42dc8","Type":"ContainerStarted","Data":"598b6fbc5376df701a762ba6746b37137dc421681ebf472e4298098628662035"} Mar 13 14:28:20 crc kubenswrapper[4907]: I0313 14:28:20.954694 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-ldbt9" event={"ID":"5dcaf7c6-b4dd-4312-b68b-a833a827a2c2","Type":"ContainerStarted","Data":"ed14d79ba88da2a61d40c8ed633b8377f32a153aac0ee43765b80eb04fb89188"} Mar 13 14:28:20 crc kubenswrapper[4907]: I0313 14:28:20.958203 4907 generic.go:334] "Generic (PLEG): container finished" podID="9118baf0-df6a-41e5-98e5-3e4d8edfbaaa" containerID="3f5c4dcc986bdd72e19fe37fafad59be240102706caf33d75e722b2f347bd3e3" exitCode=0 Mar 13 14:28:20 crc kubenswrapper[4907]: I0313 14:28:20.958268 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-vlp6r" event={"ID":"9118baf0-df6a-41e5-98e5-3e4d8edfbaaa","Type":"ContainerDied","Data":"3f5c4dcc986bdd72e19fe37fafad59be240102706caf33d75e722b2f347bd3e3"} Mar 13 14:28:21 crc kubenswrapper[4907]: I0313 14:28:21.012686 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 13 14:28:21 crc kubenswrapper[4907]: I0313 14:28:21.139845 4907 scope.go:117] "RemoveContainer" containerID="ff874f38ed077cec3c5651ce4ba0b1fcf1996e7c127b7d6b128f6c5552e26bed" Mar 13 14:28:21 crc kubenswrapper[4907]: I0313 14:28:21.822951 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="10a74a61-272c-4c54-9b2a-6379b77c9984" path="/var/lib/kubelet/pods/10a74a61-272c-4c54-9b2a-6379b77c9984/volumes" Mar 13 14:28:21 crc kubenswrapper[4907]: I0313 14:28:21.823980 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2c9261af-9eaa-4623-9f1b-719b58a9d3a2" path="/var/lib/kubelet/pods/2c9261af-9eaa-4623-9f1b-719b58a9d3a2/volumes" Mar 13 14:28:21 crc kubenswrapper[4907]: I0313 14:28:21.973311 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"0b2e62fa-af35-4f35-8852-362ea97c17e7","Type":"ContainerStarted","Data":"eeaee513015ed79d41df919baf6b1bec0513b97e239abe794d310237db6cde32"} Mar 13 14:28:21 crc kubenswrapper[4907]: I0313 14:28:21.973361 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"0b2e62fa-af35-4f35-8852-362ea97c17e7","Type":"ContainerStarted","Data":"0ede32ea598e7a7a739c54b190ba597842e8cec1fea7546f52eefbb2b1fa0f83"} Mar 13 14:28:21 crc kubenswrapper[4907]: I0313 14:28:21.976721 4907 generic.go:334] "Generic (PLEG): container finished" podID="5dcaf7c6-b4dd-4312-b68b-a833a827a2c2" containerID="ed14d79ba88da2a61d40c8ed633b8377f32a153aac0ee43765b80eb04fb89188" exitCode=0 Mar 13 14:28:21 crc kubenswrapper[4907]: I0313 14:28:21.976793 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-ldbt9" event={"ID":"5dcaf7c6-b4dd-4312-b68b-a833a827a2c2","Type":"ContainerDied","Data":"ed14d79ba88da2a61d40c8ed633b8377f32a153aac0ee43765b80eb04fb89188"} Mar 13 14:28:21 crc kubenswrapper[4907]: I0313 14:28:21.985987 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f7ca16f6-c9d9-4415-b8cf-b4753d784170","Type":"ContainerStarted","Data":"7ca97160ecf2400b113e2721d2509a72b44110e201fc18a9c3d14fa335b64d1e"} Mar 13 14:28:21 crc kubenswrapper[4907]: I0313 14:28:21.989235 4907 generic.go:334] "Generic (PLEG): container finished" podID="85913727-bbcd-4343-9faf-a75f40b42dc8" containerID="598b6fbc5376df701a762ba6746b37137dc421681ebf472e4298098628662035" exitCode=0 Mar 13 14:28:21 crc kubenswrapper[4907]: I0313 14:28:21.989798 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-f148-account-create-update-wgw4b" event={"ID":"85913727-bbcd-4343-9faf-a75f40b42dc8","Type":"ContainerDied","Data":"598b6fbc5376df701a762ba6746b37137dc421681ebf472e4298098628662035"} Mar 13 14:28:22 crc kubenswrapper[4907]: I0313 14:28:22.008177 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"4251ae36-90ad-41ea-915e-862df60f5c07","Type":"ContainerStarted","Data":"70b1012f350567e4a184430af6b8324fc4a32cf4f648d2eb406bd3d58d2b10d8"} Mar 13 14:28:22 crc kubenswrapper[4907]: I0313 14:28:22.590525 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-22bc-account-create-update-r969t" Mar 13 14:28:22 crc kubenswrapper[4907]: I0313 14:28:22.691744 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4pm5m\" (UniqueName: \"kubernetes.io/projected/4eeb2fb2-22fb-4ea8-a39b-5f959877cfe6-kube-api-access-4pm5m\") pod \"4eeb2fb2-22fb-4ea8-a39b-5f959877cfe6\" (UID: \"4eeb2fb2-22fb-4ea8-a39b-5f959877cfe6\") " Mar 13 14:28:22 crc kubenswrapper[4907]: I0313 14:28:22.692142 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4eeb2fb2-22fb-4ea8-a39b-5f959877cfe6-operator-scripts\") pod \"4eeb2fb2-22fb-4ea8-a39b-5f959877cfe6\" (UID: \"4eeb2fb2-22fb-4ea8-a39b-5f959877cfe6\") " Mar 13 14:28:22 crc kubenswrapper[4907]: I0313 14:28:22.693349 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4eeb2fb2-22fb-4ea8-a39b-5f959877cfe6-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "4eeb2fb2-22fb-4ea8-a39b-5f959877cfe6" (UID: "4eeb2fb2-22fb-4ea8-a39b-5f959877cfe6"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:28:22 crc kubenswrapper[4907]: I0313 14:28:22.698992 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4eeb2fb2-22fb-4ea8-a39b-5f959877cfe6-kube-api-access-4pm5m" (OuterVolumeSpecName: "kube-api-access-4pm5m") pod "4eeb2fb2-22fb-4ea8-a39b-5f959877cfe6" (UID: "4eeb2fb2-22fb-4ea8-a39b-5f959877cfe6"). InnerVolumeSpecName "kube-api-access-4pm5m". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:28:22 crc kubenswrapper[4907]: I0313 14:28:22.795248 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4pm5m\" (UniqueName: \"kubernetes.io/projected/4eeb2fb2-22fb-4ea8-a39b-5f959877cfe6-kube-api-access-4pm5m\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:22 crc kubenswrapper[4907]: I0313 14:28:22.795280 4907 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4eeb2fb2-22fb-4ea8-a39b-5f959877cfe6-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:22 crc kubenswrapper[4907]: I0313 14:28:22.806126 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-f6fe-account-create-update-56786" Mar 13 14:28:22 crc kubenswrapper[4907]: I0313 14:28:22.832719 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-g96mx" Mar 13 14:28:22 crc kubenswrapper[4907]: I0313 14:28:22.834280 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-vlp6r" Mar 13 14:28:22 crc kubenswrapper[4907]: I0313 14:28:22.896594 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7af00fde-8a92-4530-8fa8-9df7ff787c93-operator-scripts\") pod \"7af00fde-8a92-4530-8fa8-9df7ff787c93\" (UID: \"7af00fde-8a92-4530-8fa8-9df7ff787c93\") " Mar 13 14:28:22 crc kubenswrapper[4907]: I0313 14:28:22.896789 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dn4dc\" (UniqueName: \"kubernetes.io/projected/7af00fde-8a92-4530-8fa8-9df7ff787c93-kube-api-access-dn4dc\") pod \"7af00fde-8a92-4530-8fa8-9df7ff787c93\" (UID: \"7af00fde-8a92-4530-8fa8-9df7ff787c93\") " Mar 13 14:28:22 crc kubenswrapper[4907]: I0313 14:28:22.897299 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7af00fde-8a92-4530-8fa8-9df7ff787c93-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "7af00fde-8a92-4530-8fa8-9df7ff787c93" (UID: "7af00fde-8a92-4530-8fa8-9df7ff787c93"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:28:22 crc kubenswrapper[4907]: I0313 14:28:22.902261 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7af00fde-8a92-4530-8fa8-9df7ff787c93-kube-api-access-dn4dc" (OuterVolumeSpecName: "kube-api-access-dn4dc") pod "7af00fde-8a92-4530-8fa8-9df7ff787c93" (UID: "7af00fde-8a92-4530-8fa8-9df7ff787c93"). InnerVolumeSpecName "kube-api-access-dn4dc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:28:22 crc kubenswrapper[4907]: I0313 14:28:22.998038 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-swn2k\" (UniqueName: \"kubernetes.io/projected/9118baf0-df6a-41e5-98e5-3e4d8edfbaaa-kube-api-access-swn2k\") pod \"9118baf0-df6a-41e5-98e5-3e4d8edfbaaa\" (UID: \"9118baf0-df6a-41e5-98e5-3e4d8edfbaaa\") " Mar 13 14:28:22 crc kubenswrapper[4907]: I0313 14:28:22.998400 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9118baf0-df6a-41e5-98e5-3e4d8edfbaaa-operator-scripts\") pod \"9118baf0-df6a-41e5-98e5-3e4d8edfbaaa\" (UID: \"9118baf0-df6a-41e5-98e5-3e4d8edfbaaa\") " Mar 13 14:28:22 crc kubenswrapper[4907]: I0313 14:28:22.998568 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/251ab76e-b9df-406f-9b07-c7fa6e227feb-operator-scripts\") pod \"251ab76e-b9df-406f-9b07-c7fa6e227feb\" (UID: \"251ab76e-b9df-406f-9b07-c7fa6e227feb\") " Mar 13 14:28:22 crc kubenswrapper[4907]: I0313 14:28:22.998609 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mjhh8\" (UniqueName: \"kubernetes.io/projected/251ab76e-b9df-406f-9b07-c7fa6e227feb-kube-api-access-mjhh8\") pod \"251ab76e-b9df-406f-9b07-c7fa6e227feb\" (UID: \"251ab76e-b9df-406f-9b07-c7fa6e227feb\") " Mar 13 14:28:22 crc kubenswrapper[4907]: I0313 14:28:22.998976 4907 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7af00fde-8a92-4530-8fa8-9df7ff787c93-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:22 crc kubenswrapper[4907]: I0313 14:28:22.999018 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dn4dc\" (UniqueName: \"kubernetes.io/projected/7af00fde-8a92-4530-8fa8-9df7ff787c93-kube-api-access-dn4dc\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:22 crc kubenswrapper[4907]: I0313 14:28:22.999177 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9118baf0-df6a-41e5-98e5-3e4d8edfbaaa-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "9118baf0-df6a-41e5-98e5-3e4d8edfbaaa" (UID: "9118baf0-df6a-41e5-98e5-3e4d8edfbaaa"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:28:22 crc kubenswrapper[4907]: I0313 14:28:22.999466 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/251ab76e-b9df-406f-9b07-c7fa6e227feb-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "251ab76e-b9df-406f-9b07-c7fa6e227feb" (UID: "251ab76e-b9df-406f-9b07-c7fa6e227feb"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:28:23 crc kubenswrapper[4907]: I0313 14:28:23.002705 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9118baf0-df6a-41e5-98e5-3e4d8edfbaaa-kube-api-access-swn2k" (OuterVolumeSpecName: "kube-api-access-swn2k") pod "9118baf0-df6a-41e5-98e5-3e4d8edfbaaa" (UID: "9118baf0-df6a-41e5-98e5-3e4d8edfbaaa"). InnerVolumeSpecName "kube-api-access-swn2k". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:28:23 crc kubenswrapper[4907]: I0313 14:28:23.006593 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/251ab76e-b9df-406f-9b07-c7fa6e227feb-kube-api-access-mjhh8" (OuterVolumeSpecName: "kube-api-access-mjhh8") pod "251ab76e-b9df-406f-9b07-c7fa6e227feb" (UID: "251ab76e-b9df-406f-9b07-c7fa6e227feb"). InnerVolumeSpecName "kube-api-access-mjhh8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:28:23 crc kubenswrapper[4907]: I0313 14:28:23.033682 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-f6fe-account-create-update-56786" event={"ID":"7af00fde-8a92-4530-8fa8-9df7ff787c93","Type":"ContainerDied","Data":"5c53d7426e0e1950f78a3c70586fa0b2e353687a87ba5dbe4ff6e6fc12256f18"} Mar 13 14:28:23 crc kubenswrapper[4907]: I0313 14:28:23.033718 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5c53d7426e0e1950f78a3c70586fa0b2e353687a87ba5dbe4ff6e6fc12256f18" Mar 13 14:28:23 crc kubenswrapper[4907]: I0313 14:28:23.033775 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-f6fe-account-create-update-56786" Mar 13 14:28:23 crc kubenswrapper[4907]: I0313 14:28:23.068853 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-vlp6r" event={"ID":"9118baf0-df6a-41e5-98e5-3e4d8edfbaaa","Type":"ContainerDied","Data":"b75ab2628fbe2653fbc966d7e517d3692b8792f5300df76a97698a45b1204d93"} Mar 13 14:28:23 crc kubenswrapper[4907]: I0313 14:28:23.068925 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b75ab2628fbe2653fbc966d7e517d3692b8792f5300df76a97698a45b1204d93" Mar 13 14:28:23 crc kubenswrapper[4907]: I0313 14:28:23.068993 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-vlp6r" Mar 13 14:28:23 crc kubenswrapper[4907]: I0313 14:28:23.096566 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"0b2e62fa-af35-4f35-8852-362ea97c17e7","Type":"ContainerStarted","Data":"0d9941f62564b14687332e4dbc1f81a8df8a858860923a1b915fe67a744f75f5"} Mar 13 14:28:23 crc kubenswrapper[4907]: I0313 14:28:23.100621 4907 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9118baf0-df6a-41e5-98e5-3e4d8edfbaaa-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:23 crc kubenswrapper[4907]: I0313 14:28:23.100653 4907 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/251ab76e-b9df-406f-9b07-c7fa6e227feb-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:23 crc kubenswrapper[4907]: I0313 14:28:23.100668 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mjhh8\" (UniqueName: \"kubernetes.io/projected/251ab76e-b9df-406f-9b07-c7fa6e227feb-kube-api-access-mjhh8\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:23 crc kubenswrapper[4907]: I0313 14:28:23.100709 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-swn2k\" (UniqueName: \"kubernetes.io/projected/9118baf0-df6a-41e5-98e5-3e4d8edfbaaa-kube-api-access-swn2k\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:23 crc kubenswrapper[4907]: I0313 14:28:23.115214 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-22bc-account-create-update-r969t" event={"ID":"4eeb2fb2-22fb-4ea8-a39b-5f959877cfe6","Type":"ContainerDied","Data":"46336f79de33964f45a2c37528d8472d8d5d46341b5f9fc4bc15531e76f002ba"} Mar 13 14:28:23 crc kubenswrapper[4907]: I0313 14:28:23.115477 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="46336f79de33964f45a2c37528d8472d8d5d46341b5f9fc4bc15531e76f002ba" Mar 13 14:28:23 crc kubenswrapper[4907]: I0313 14:28:23.115620 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-22bc-account-create-update-r969t" Mar 13 14:28:23 crc kubenswrapper[4907]: I0313 14:28:23.137794 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=4.137776329 podStartE2EDuration="4.137776329s" podCreationTimestamp="2026-03-13 14:28:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:28:23.133082051 +0000 UTC m=+1402.032869750" watchObservedRunningTime="2026-03-13 14:28:23.137776329 +0000 UTC m=+1402.037564018" Mar 13 14:28:23 crc kubenswrapper[4907]: I0313 14:28:23.141660 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"4251ae36-90ad-41ea-915e-862df60f5c07","Type":"ContainerStarted","Data":"479ee0fb693df48bcf5ef06770a45535f793e4153c556058631f42347318a544"} Mar 13 14:28:23 crc kubenswrapper[4907]: I0313 14:28:23.152259 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-g96mx" Mar 13 14:28:23 crc kubenswrapper[4907]: I0313 14:28:23.152518 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-g96mx" event={"ID":"251ab76e-b9df-406f-9b07-c7fa6e227feb","Type":"ContainerDied","Data":"57e7cdfdc101c4e36f2572edec180e1f998c526cf20ca796f61ebb46ef192472"} Mar 13 14:28:23 crc kubenswrapper[4907]: I0313 14:28:23.152596 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="57e7cdfdc101c4e36f2572edec180e1f998c526cf20ca796f61ebb46ef192472" Mar 13 14:28:23 crc kubenswrapper[4907]: I0313 14:28:23.192428 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=4.192410305 podStartE2EDuration="4.192410305s" podCreationTimestamp="2026-03-13 14:28:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:28:23.187486351 +0000 UTC m=+1402.087274040" watchObservedRunningTime="2026-03-13 14:28:23.192410305 +0000 UTC m=+1402.092197994" Mar 13 14:28:23 crc kubenswrapper[4907]: I0313 14:28:23.610374 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-ldbt9" Mar 13 14:28:23 crc kubenswrapper[4907]: I0313 14:28:23.713536 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-f148-account-create-update-wgw4b" Mar 13 14:28:23 crc kubenswrapper[4907]: I0313 14:28:23.720038 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wj4tn\" (UniqueName: \"kubernetes.io/projected/5dcaf7c6-b4dd-4312-b68b-a833a827a2c2-kube-api-access-wj4tn\") pod \"5dcaf7c6-b4dd-4312-b68b-a833a827a2c2\" (UID: \"5dcaf7c6-b4dd-4312-b68b-a833a827a2c2\") " Mar 13 14:28:23 crc kubenswrapper[4907]: I0313 14:28:23.720446 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5dcaf7c6-b4dd-4312-b68b-a833a827a2c2-operator-scripts\") pod \"5dcaf7c6-b4dd-4312-b68b-a833a827a2c2\" (UID: \"5dcaf7c6-b4dd-4312-b68b-a833a827a2c2\") " Mar 13 14:28:23 crc kubenswrapper[4907]: I0313 14:28:23.721154 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5dcaf7c6-b4dd-4312-b68b-a833a827a2c2-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "5dcaf7c6-b4dd-4312-b68b-a833a827a2c2" (UID: "5dcaf7c6-b4dd-4312-b68b-a833a827a2c2"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:28:23 crc kubenswrapper[4907]: I0313 14:28:23.721744 4907 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5dcaf7c6-b4dd-4312-b68b-a833a827a2c2-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:23 crc kubenswrapper[4907]: I0313 14:28:23.724545 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5dcaf7c6-b4dd-4312-b68b-a833a827a2c2-kube-api-access-wj4tn" (OuterVolumeSpecName: "kube-api-access-wj4tn") pod "5dcaf7c6-b4dd-4312-b68b-a833a827a2c2" (UID: "5dcaf7c6-b4dd-4312-b68b-a833a827a2c2"). InnerVolumeSpecName "kube-api-access-wj4tn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:28:23 crc kubenswrapper[4907]: I0313 14:28:23.823528 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cnlpw\" (UniqueName: \"kubernetes.io/projected/85913727-bbcd-4343-9faf-a75f40b42dc8-kube-api-access-cnlpw\") pod \"85913727-bbcd-4343-9faf-a75f40b42dc8\" (UID: \"85913727-bbcd-4343-9faf-a75f40b42dc8\") " Mar 13 14:28:23 crc kubenswrapper[4907]: I0313 14:28:23.823684 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/85913727-bbcd-4343-9faf-a75f40b42dc8-operator-scripts\") pod \"85913727-bbcd-4343-9faf-a75f40b42dc8\" (UID: \"85913727-bbcd-4343-9faf-a75f40b42dc8\") " Mar 13 14:28:23 crc kubenswrapper[4907]: I0313 14:28:23.824079 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wj4tn\" (UniqueName: \"kubernetes.io/projected/5dcaf7c6-b4dd-4312-b68b-a833a827a2c2-kube-api-access-wj4tn\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:23 crc kubenswrapper[4907]: I0313 14:28:23.824119 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/85913727-bbcd-4343-9faf-a75f40b42dc8-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "85913727-bbcd-4343-9faf-a75f40b42dc8" (UID: "85913727-bbcd-4343-9faf-a75f40b42dc8"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:28:23 crc kubenswrapper[4907]: I0313 14:28:23.827684 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/85913727-bbcd-4343-9faf-a75f40b42dc8-kube-api-access-cnlpw" (OuterVolumeSpecName: "kube-api-access-cnlpw") pod "85913727-bbcd-4343-9faf-a75f40b42dc8" (UID: "85913727-bbcd-4343-9faf-a75f40b42dc8"). InnerVolumeSpecName "kube-api-access-cnlpw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:28:23 crc kubenswrapper[4907]: I0313 14:28:23.926046 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cnlpw\" (UniqueName: \"kubernetes.io/projected/85913727-bbcd-4343-9faf-a75f40b42dc8-kube-api-access-cnlpw\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:23 crc kubenswrapper[4907]: I0313 14:28:23.926086 4907 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/85913727-bbcd-4343-9faf-a75f40b42dc8-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:24 crc kubenswrapper[4907]: I0313 14:28:24.171157 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-ldbt9" event={"ID":"5dcaf7c6-b4dd-4312-b68b-a833a827a2c2","Type":"ContainerDied","Data":"fa6acba78f266f6ff1774ad62b03c8a5d55736cb5f7697c6fcc3a1a5a0614819"} Mar 13 14:28:24 crc kubenswrapper[4907]: I0313 14:28:24.171199 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-ldbt9" Mar 13 14:28:24 crc kubenswrapper[4907]: I0313 14:28:24.171227 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fa6acba78f266f6ff1774ad62b03c8a5d55736cb5f7697c6fcc3a1a5a0614819" Mar 13 14:28:24 crc kubenswrapper[4907]: I0313 14:28:24.175551 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f7ca16f6-c9d9-4415-b8cf-b4753d784170","Type":"ContainerStarted","Data":"725fda3e17763bb514d9ba2d39dd7a4cfc02d5d585475411f3905020ac1f8e20"} Mar 13 14:28:24 crc kubenswrapper[4907]: I0313 14:28:24.175721 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f7ca16f6-c9d9-4415-b8cf-b4753d784170" containerName="ceilometer-central-agent" containerID="cri-o://53d201dd158c6af9aadd5213934f6cc439d1b5e00b0a3e3af8f7f6824b4594d2" gracePeriod=30 Mar 13 14:28:24 crc kubenswrapper[4907]: I0313 14:28:24.175841 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f7ca16f6-c9d9-4415-b8cf-b4753d784170" containerName="proxy-httpd" containerID="cri-o://725fda3e17763bb514d9ba2d39dd7a4cfc02d5d585475411f3905020ac1f8e20" gracePeriod=30 Mar 13 14:28:24 crc kubenswrapper[4907]: I0313 14:28:24.175919 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f7ca16f6-c9d9-4415-b8cf-b4753d784170" containerName="sg-core" containerID="cri-o://7ca97160ecf2400b113e2721d2509a72b44110e201fc18a9c3d14fa335b64d1e" gracePeriod=30 Mar 13 14:28:24 crc kubenswrapper[4907]: I0313 14:28:24.175958 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f7ca16f6-c9d9-4415-b8cf-b4753d784170" containerName="ceilometer-notification-agent" containerID="cri-o://306d706e8b95e9a38af5ba674a473f59462492d56df0ecb87578b8edfdde9aa6" gracePeriod=30 Mar 13 14:28:24 crc kubenswrapper[4907]: I0313 14:28:24.176192 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Mar 13 14:28:24 crc kubenswrapper[4907]: I0313 14:28:24.181212 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-f148-account-create-update-wgw4b" Mar 13 14:28:24 crc kubenswrapper[4907]: I0313 14:28:24.182224 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-f148-account-create-update-wgw4b" event={"ID":"85913727-bbcd-4343-9faf-a75f40b42dc8","Type":"ContainerDied","Data":"ca4cf56d486af5cc064a60fd28eb3806cfa9fe3d0bef16b65c3fd811e37a4cb4"} Mar 13 14:28:24 crc kubenswrapper[4907]: I0313 14:28:24.182292 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ca4cf56d486af5cc064a60fd28eb3806cfa9fe3d0bef16b65c3fd811e37a4cb4" Mar 13 14:28:24 crc kubenswrapper[4907]: I0313 14:28:24.211208 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.461854472 podStartE2EDuration="15.211189153s" podCreationTimestamp="2026-03-13 14:28:09 +0000 UTC" firstStartedPulling="2026-03-13 14:28:10.510785786 +0000 UTC m=+1389.410573475" lastFinishedPulling="2026-03-13 14:28:23.260120467 +0000 UTC m=+1402.159908156" observedRunningTime="2026-03-13 14:28:24.200235215 +0000 UTC m=+1403.100022904" watchObservedRunningTime="2026-03-13 14:28:24.211189153 +0000 UTC m=+1403.110976842" Mar 13 14:28:24 crc kubenswrapper[4907]: E0313 14:28:24.446388 4907 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf7ca16f6_c9d9_4415_b8cf_b4753d784170.slice/crio-725fda3e17763bb514d9ba2d39dd7a4cfc02d5d585475411f3905020ac1f8e20.scope\": RecentStats: unable to find data in memory cache]" Mar 13 14:28:24 crc kubenswrapper[4907]: I0313 14:28:24.931015 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.049904 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7ca16f6-c9d9-4415-b8cf-b4753d784170-scripts\") pod \"f7ca16f6-c9d9-4415-b8cf-b4753d784170\" (UID: \"f7ca16f6-c9d9-4415-b8cf-b4753d784170\") " Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.050238 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f7ca16f6-c9d9-4415-b8cf-b4753d784170-log-httpd\") pod \"f7ca16f6-c9d9-4415-b8cf-b4753d784170\" (UID: \"f7ca16f6-c9d9-4415-b8cf-b4753d784170\") " Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.050327 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gfnvm\" (UniqueName: \"kubernetes.io/projected/f7ca16f6-c9d9-4415-b8cf-b4753d784170-kube-api-access-gfnvm\") pod \"f7ca16f6-c9d9-4415-b8cf-b4753d784170\" (UID: \"f7ca16f6-c9d9-4415-b8cf-b4753d784170\") " Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.050398 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7ca16f6-c9d9-4415-b8cf-b4753d784170-config-data\") pod \"f7ca16f6-c9d9-4415-b8cf-b4753d784170\" (UID: \"f7ca16f6-c9d9-4415-b8cf-b4753d784170\") " Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.050545 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7ca16f6-c9d9-4415-b8cf-b4753d784170-combined-ca-bundle\") pod \"f7ca16f6-c9d9-4415-b8cf-b4753d784170\" (UID: \"f7ca16f6-c9d9-4415-b8cf-b4753d784170\") " Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.050621 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f7ca16f6-c9d9-4415-b8cf-b4753d784170-sg-core-conf-yaml\") pod \"f7ca16f6-c9d9-4415-b8cf-b4753d784170\" (UID: \"f7ca16f6-c9d9-4415-b8cf-b4753d784170\") " Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.050676 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f7ca16f6-c9d9-4415-b8cf-b4753d784170-run-httpd\") pod \"f7ca16f6-c9d9-4415-b8cf-b4753d784170\" (UID: \"f7ca16f6-c9d9-4415-b8cf-b4753d784170\") " Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.051051 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f7ca16f6-c9d9-4415-b8cf-b4753d784170-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "f7ca16f6-c9d9-4415-b8cf-b4753d784170" (UID: "f7ca16f6-c9d9-4415-b8cf-b4753d784170"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.051178 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f7ca16f6-c9d9-4415-b8cf-b4753d784170-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "f7ca16f6-c9d9-4415-b8cf-b4753d784170" (UID: "f7ca16f6-c9d9-4415-b8cf-b4753d784170"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.051444 4907 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f7ca16f6-c9d9-4415-b8cf-b4753d784170-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.051483 4907 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f7ca16f6-c9d9-4415-b8cf-b4753d784170-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.056937 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7ca16f6-c9d9-4415-b8cf-b4753d784170-scripts" (OuterVolumeSpecName: "scripts") pod "f7ca16f6-c9d9-4415-b8cf-b4753d784170" (UID: "f7ca16f6-c9d9-4415-b8cf-b4753d784170"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.065636 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f7ca16f6-c9d9-4415-b8cf-b4753d784170-kube-api-access-gfnvm" (OuterVolumeSpecName: "kube-api-access-gfnvm") pod "f7ca16f6-c9d9-4415-b8cf-b4753d784170" (UID: "f7ca16f6-c9d9-4415-b8cf-b4753d784170"). InnerVolumeSpecName "kube-api-access-gfnvm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.085649 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7ca16f6-c9d9-4415-b8cf-b4753d784170-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "f7ca16f6-c9d9-4415-b8cf-b4753d784170" (UID: "f7ca16f6-c9d9-4415-b8cf-b4753d784170"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.127849 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7ca16f6-c9d9-4415-b8cf-b4753d784170-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f7ca16f6-c9d9-4415-b8cf-b4753d784170" (UID: "f7ca16f6-c9d9-4415-b8cf-b4753d784170"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.154022 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7ca16f6-c9d9-4415-b8cf-b4753d784170-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.154056 4907 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f7ca16f6-c9d9-4415-b8cf-b4753d784170-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.154068 4907 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7ca16f6-c9d9-4415-b8cf-b4753d784170-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.154079 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gfnvm\" (UniqueName: \"kubernetes.io/projected/f7ca16f6-c9d9-4415-b8cf-b4753d784170-kube-api-access-gfnvm\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.156622 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7ca16f6-c9d9-4415-b8cf-b4753d784170-config-data" (OuterVolumeSpecName: "config-data") pod "f7ca16f6-c9d9-4415-b8cf-b4753d784170" (UID: "f7ca16f6-c9d9-4415-b8cf-b4753d784170"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.198306 4907 generic.go:334] "Generic (PLEG): container finished" podID="f7ca16f6-c9d9-4415-b8cf-b4753d784170" containerID="725fda3e17763bb514d9ba2d39dd7a4cfc02d5d585475411f3905020ac1f8e20" exitCode=0 Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.198338 4907 generic.go:334] "Generic (PLEG): container finished" podID="f7ca16f6-c9d9-4415-b8cf-b4753d784170" containerID="7ca97160ecf2400b113e2721d2509a72b44110e201fc18a9c3d14fa335b64d1e" exitCode=2 Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.198351 4907 generic.go:334] "Generic (PLEG): container finished" podID="f7ca16f6-c9d9-4415-b8cf-b4753d784170" containerID="306d706e8b95e9a38af5ba674a473f59462492d56df0ecb87578b8edfdde9aa6" exitCode=0 Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.198360 4907 generic.go:334] "Generic (PLEG): container finished" podID="f7ca16f6-c9d9-4415-b8cf-b4753d784170" containerID="53d201dd158c6af9aadd5213934f6cc439d1b5e00b0a3e3af8f7f6824b4594d2" exitCode=0 Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.198384 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f7ca16f6-c9d9-4415-b8cf-b4753d784170","Type":"ContainerDied","Data":"725fda3e17763bb514d9ba2d39dd7a4cfc02d5d585475411f3905020ac1f8e20"} Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.198410 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f7ca16f6-c9d9-4415-b8cf-b4753d784170","Type":"ContainerDied","Data":"7ca97160ecf2400b113e2721d2509a72b44110e201fc18a9c3d14fa335b64d1e"} Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.198422 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f7ca16f6-c9d9-4415-b8cf-b4753d784170","Type":"ContainerDied","Data":"306d706e8b95e9a38af5ba674a473f59462492d56df0ecb87578b8edfdde9aa6"} Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.198431 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f7ca16f6-c9d9-4415-b8cf-b4753d784170","Type":"ContainerDied","Data":"53d201dd158c6af9aadd5213934f6cc439d1b5e00b0a3e3af8f7f6824b4594d2"} Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.198439 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f7ca16f6-c9d9-4415-b8cf-b4753d784170","Type":"ContainerDied","Data":"2e74becf0a06fc2c8e64ca7acbdb2fa95c901f1325aa7aeacdbfe4e9db0c932e"} Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.198452 4907 scope.go:117] "RemoveContainer" containerID="725fda3e17763bb514d9ba2d39dd7a4cfc02d5d585475411f3905020ac1f8e20" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.199042 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.255661 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7ca16f6-c9d9-4415-b8cf-b4753d784170-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.259948 4907 scope.go:117] "RemoveContainer" containerID="7ca97160ecf2400b113e2721d2509a72b44110e201fc18a9c3d14fa335b64d1e" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.261027 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.294567 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.307631 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 13 14:28:25 crc kubenswrapper[4907]: E0313 14:28:25.308052 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9118baf0-df6a-41e5-98e5-3e4d8edfbaaa" containerName="mariadb-database-create" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.308070 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="9118baf0-df6a-41e5-98e5-3e4d8edfbaaa" containerName="mariadb-database-create" Mar 13 14:28:25 crc kubenswrapper[4907]: E0313 14:28:25.308083 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7af00fde-8a92-4530-8fa8-9df7ff787c93" containerName="mariadb-account-create-update" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.308090 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="7af00fde-8a92-4530-8fa8-9df7ff787c93" containerName="mariadb-account-create-update" Mar 13 14:28:25 crc kubenswrapper[4907]: E0313 14:28:25.308101 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7ca16f6-c9d9-4415-b8cf-b4753d784170" containerName="ceilometer-notification-agent" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.308109 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7ca16f6-c9d9-4415-b8cf-b4753d784170" containerName="ceilometer-notification-agent" Mar 13 14:28:25 crc kubenswrapper[4907]: E0313 14:28:25.308121 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85913727-bbcd-4343-9faf-a75f40b42dc8" containerName="mariadb-account-create-update" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.308128 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="85913727-bbcd-4343-9faf-a75f40b42dc8" containerName="mariadb-account-create-update" Mar 13 14:28:25 crc kubenswrapper[4907]: E0313 14:28:25.308145 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5dcaf7c6-b4dd-4312-b68b-a833a827a2c2" containerName="mariadb-database-create" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.308150 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="5dcaf7c6-b4dd-4312-b68b-a833a827a2c2" containerName="mariadb-database-create" Mar 13 14:28:25 crc kubenswrapper[4907]: E0313 14:28:25.308162 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7ca16f6-c9d9-4415-b8cf-b4753d784170" containerName="sg-core" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.308167 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7ca16f6-c9d9-4415-b8cf-b4753d784170" containerName="sg-core" Mar 13 14:28:25 crc kubenswrapper[4907]: E0313 14:28:25.308182 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4eeb2fb2-22fb-4ea8-a39b-5f959877cfe6" containerName="mariadb-account-create-update" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.308188 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="4eeb2fb2-22fb-4ea8-a39b-5f959877cfe6" containerName="mariadb-account-create-update" Mar 13 14:28:25 crc kubenswrapper[4907]: E0313 14:28:25.308201 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7ca16f6-c9d9-4415-b8cf-b4753d784170" containerName="ceilometer-central-agent" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.308208 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7ca16f6-c9d9-4415-b8cf-b4753d784170" containerName="ceilometer-central-agent" Mar 13 14:28:25 crc kubenswrapper[4907]: E0313 14:28:25.308217 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="251ab76e-b9df-406f-9b07-c7fa6e227feb" containerName="mariadb-database-create" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.308223 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="251ab76e-b9df-406f-9b07-c7fa6e227feb" containerName="mariadb-database-create" Mar 13 14:28:25 crc kubenswrapper[4907]: E0313 14:28:25.308236 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7ca16f6-c9d9-4415-b8cf-b4753d784170" containerName="proxy-httpd" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.308242 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7ca16f6-c9d9-4415-b8cf-b4753d784170" containerName="proxy-httpd" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.308398 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="7af00fde-8a92-4530-8fa8-9df7ff787c93" containerName="mariadb-account-create-update" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.308410 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7ca16f6-c9d9-4415-b8cf-b4753d784170" containerName="ceilometer-central-agent" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.308422 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="85913727-bbcd-4343-9faf-a75f40b42dc8" containerName="mariadb-account-create-update" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.308435 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="4eeb2fb2-22fb-4ea8-a39b-5f959877cfe6" containerName="mariadb-account-create-update" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.308442 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7ca16f6-c9d9-4415-b8cf-b4753d784170" containerName="ceilometer-notification-agent" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.308452 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="5dcaf7c6-b4dd-4312-b68b-a833a827a2c2" containerName="mariadb-database-create" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.308460 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="9118baf0-df6a-41e5-98e5-3e4d8edfbaaa" containerName="mariadb-database-create" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.308472 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7ca16f6-c9d9-4415-b8cf-b4753d784170" containerName="sg-core" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.308480 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="251ab76e-b9df-406f-9b07-c7fa6e227feb" containerName="mariadb-database-create" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.308488 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7ca16f6-c9d9-4415-b8cf-b4753d784170" containerName="proxy-httpd" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.310733 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.316579 4907 scope.go:117] "RemoveContainer" containerID="306d706e8b95e9a38af5ba674a473f59462492d56df0ecb87578b8edfdde9aa6" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.317102 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.317424 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.323415 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.353820 4907 scope.go:117] "RemoveContainer" containerID="53d201dd158c6af9aadd5213934f6cc439d1b5e00b0a3e3af8f7f6824b4594d2" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.358837 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6be28d7-de73-47ad-9b71-f7333bae7487-config-data\") pod \"ceilometer-0\" (UID: \"d6be28d7-de73-47ad-9b71-f7333bae7487\") " pod="openstack/ceilometer-0" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.358911 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d6be28d7-de73-47ad-9b71-f7333bae7487-run-httpd\") pod \"ceilometer-0\" (UID: \"d6be28d7-de73-47ad-9b71-f7333bae7487\") " pod="openstack/ceilometer-0" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.358954 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fldgc\" (UniqueName: \"kubernetes.io/projected/d6be28d7-de73-47ad-9b71-f7333bae7487-kube-api-access-fldgc\") pod \"ceilometer-0\" (UID: \"d6be28d7-de73-47ad-9b71-f7333bae7487\") " pod="openstack/ceilometer-0" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.358975 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d6be28d7-de73-47ad-9b71-f7333bae7487-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d6be28d7-de73-47ad-9b71-f7333bae7487\") " pod="openstack/ceilometer-0" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.359013 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d6be28d7-de73-47ad-9b71-f7333bae7487-log-httpd\") pod \"ceilometer-0\" (UID: \"d6be28d7-de73-47ad-9b71-f7333bae7487\") " pod="openstack/ceilometer-0" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.359080 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6be28d7-de73-47ad-9b71-f7333bae7487-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d6be28d7-de73-47ad-9b71-f7333bae7487\") " pod="openstack/ceilometer-0" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.359104 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d6be28d7-de73-47ad-9b71-f7333bae7487-scripts\") pod \"ceilometer-0\" (UID: \"d6be28d7-de73-47ad-9b71-f7333bae7487\") " pod="openstack/ceilometer-0" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.383818 4907 scope.go:117] "RemoveContainer" containerID="725fda3e17763bb514d9ba2d39dd7a4cfc02d5d585475411f3905020ac1f8e20" Mar 13 14:28:25 crc kubenswrapper[4907]: E0313 14:28:25.384287 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"725fda3e17763bb514d9ba2d39dd7a4cfc02d5d585475411f3905020ac1f8e20\": container with ID starting with 725fda3e17763bb514d9ba2d39dd7a4cfc02d5d585475411f3905020ac1f8e20 not found: ID does not exist" containerID="725fda3e17763bb514d9ba2d39dd7a4cfc02d5d585475411f3905020ac1f8e20" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.384313 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"725fda3e17763bb514d9ba2d39dd7a4cfc02d5d585475411f3905020ac1f8e20"} err="failed to get container status \"725fda3e17763bb514d9ba2d39dd7a4cfc02d5d585475411f3905020ac1f8e20\": rpc error: code = NotFound desc = could not find container \"725fda3e17763bb514d9ba2d39dd7a4cfc02d5d585475411f3905020ac1f8e20\": container with ID starting with 725fda3e17763bb514d9ba2d39dd7a4cfc02d5d585475411f3905020ac1f8e20 not found: ID does not exist" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.384337 4907 scope.go:117] "RemoveContainer" containerID="7ca97160ecf2400b113e2721d2509a72b44110e201fc18a9c3d14fa335b64d1e" Mar 13 14:28:25 crc kubenswrapper[4907]: E0313 14:28:25.384636 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7ca97160ecf2400b113e2721d2509a72b44110e201fc18a9c3d14fa335b64d1e\": container with ID starting with 7ca97160ecf2400b113e2721d2509a72b44110e201fc18a9c3d14fa335b64d1e not found: ID does not exist" containerID="7ca97160ecf2400b113e2721d2509a72b44110e201fc18a9c3d14fa335b64d1e" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.384671 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7ca97160ecf2400b113e2721d2509a72b44110e201fc18a9c3d14fa335b64d1e"} err="failed to get container status \"7ca97160ecf2400b113e2721d2509a72b44110e201fc18a9c3d14fa335b64d1e\": rpc error: code = NotFound desc = could not find container \"7ca97160ecf2400b113e2721d2509a72b44110e201fc18a9c3d14fa335b64d1e\": container with ID starting with 7ca97160ecf2400b113e2721d2509a72b44110e201fc18a9c3d14fa335b64d1e not found: ID does not exist" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.384696 4907 scope.go:117] "RemoveContainer" containerID="306d706e8b95e9a38af5ba674a473f59462492d56df0ecb87578b8edfdde9aa6" Mar 13 14:28:25 crc kubenswrapper[4907]: E0313 14:28:25.385055 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"306d706e8b95e9a38af5ba674a473f59462492d56df0ecb87578b8edfdde9aa6\": container with ID starting with 306d706e8b95e9a38af5ba674a473f59462492d56df0ecb87578b8edfdde9aa6 not found: ID does not exist" containerID="306d706e8b95e9a38af5ba674a473f59462492d56df0ecb87578b8edfdde9aa6" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.385138 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"306d706e8b95e9a38af5ba674a473f59462492d56df0ecb87578b8edfdde9aa6"} err="failed to get container status \"306d706e8b95e9a38af5ba674a473f59462492d56df0ecb87578b8edfdde9aa6\": rpc error: code = NotFound desc = could not find container \"306d706e8b95e9a38af5ba674a473f59462492d56df0ecb87578b8edfdde9aa6\": container with ID starting with 306d706e8b95e9a38af5ba674a473f59462492d56df0ecb87578b8edfdde9aa6 not found: ID does not exist" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.385212 4907 scope.go:117] "RemoveContainer" containerID="53d201dd158c6af9aadd5213934f6cc439d1b5e00b0a3e3af8f7f6824b4594d2" Mar 13 14:28:25 crc kubenswrapper[4907]: E0313 14:28:25.385515 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"53d201dd158c6af9aadd5213934f6cc439d1b5e00b0a3e3af8f7f6824b4594d2\": container with ID starting with 53d201dd158c6af9aadd5213934f6cc439d1b5e00b0a3e3af8f7f6824b4594d2 not found: ID does not exist" containerID="53d201dd158c6af9aadd5213934f6cc439d1b5e00b0a3e3af8f7f6824b4594d2" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.385594 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"53d201dd158c6af9aadd5213934f6cc439d1b5e00b0a3e3af8f7f6824b4594d2"} err="failed to get container status \"53d201dd158c6af9aadd5213934f6cc439d1b5e00b0a3e3af8f7f6824b4594d2\": rpc error: code = NotFound desc = could not find container \"53d201dd158c6af9aadd5213934f6cc439d1b5e00b0a3e3af8f7f6824b4594d2\": container with ID starting with 53d201dd158c6af9aadd5213934f6cc439d1b5e00b0a3e3af8f7f6824b4594d2 not found: ID does not exist" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.385663 4907 scope.go:117] "RemoveContainer" containerID="725fda3e17763bb514d9ba2d39dd7a4cfc02d5d585475411f3905020ac1f8e20" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.385919 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"725fda3e17763bb514d9ba2d39dd7a4cfc02d5d585475411f3905020ac1f8e20"} err="failed to get container status \"725fda3e17763bb514d9ba2d39dd7a4cfc02d5d585475411f3905020ac1f8e20\": rpc error: code = NotFound desc = could not find container \"725fda3e17763bb514d9ba2d39dd7a4cfc02d5d585475411f3905020ac1f8e20\": container with ID starting with 725fda3e17763bb514d9ba2d39dd7a4cfc02d5d585475411f3905020ac1f8e20 not found: ID does not exist" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.386001 4907 scope.go:117] "RemoveContainer" containerID="7ca97160ecf2400b113e2721d2509a72b44110e201fc18a9c3d14fa335b64d1e" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.386239 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7ca97160ecf2400b113e2721d2509a72b44110e201fc18a9c3d14fa335b64d1e"} err="failed to get container status \"7ca97160ecf2400b113e2721d2509a72b44110e201fc18a9c3d14fa335b64d1e\": rpc error: code = NotFound desc = could not find container \"7ca97160ecf2400b113e2721d2509a72b44110e201fc18a9c3d14fa335b64d1e\": container with ID starting with 7ca97160ecf2400b113e2721d2509a72b44110e201fc18a9c3d14fa335b64d1e not found: ID does not exist" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.386314 4907 scope.go:117] "RemoveContainer" containerID="306d706e8b95e9a38af5ba674a473f59462492d56df0ecb87578b8edfdde9aa6" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.386958 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"306d706e8b95e9a38af5ba674a473f59462492d56df0ecb87578b8edfdde9aa6"} err="failed to get container status \"306d706e8b95e9a38af5ba674a473f59462492d56df0ecb87578b8edfdde9aa6\": rpc error: code = NotFound desc = could not find container \"306d706e8b95e9a38af5ba674a473f59462492d56df0ecb87578b8edfdde9aa6\": container with ID starting with 306d706e8b95e9a38af5ba674a473f59462492d56df0ecb87578b8edfdde9aa6 not found: ID does not exist" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.387068 4907 scope.go:117] "RemoveContainer" containerID="53d201dd158c6af9aadd5213934f6cc439d1b5e00b0a3e3af8f7f6824b4594d2" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.387359 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"53d201dd158c6af9aadd5213934f6cc439d1b5e00b0a3e3af8f7f6824b4594d2"} err="failed to get container status \"53d201dd158c6af9aadd5213934f6cc439d1b5e00b0a3e3af8f7f6824b4594d2\": rpc error: code = NotFound desc = could not find container \"53d201dd158c6af9aadd5213934f6cc439d1b5e00b0a3e3af8f7f6824b4594d2\": container with ID starting with 53d201dd158c6af9aadd5213934f6cc439d1b5e00b0a3e3af8f7f6824b4594d2 not found: ID does not exist" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.387443 4907 scope.go:117] "RemoveContainer" containerID="725fda3e17763bb514d9ba2d39dd7a4cfc02d5d585475411f3905020ac1f8e20" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.387659 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"725fda3e17763bb514d9ba2d39dd7a4cfc02d5d585475411f3905020ac1f8e20"} err="failed to get container status \"725fda3e17763bb514d9ba2d39dd7a4cfc02d5d585475411f3905020ac1f8e20\": rpc error: code = NotFound desc = could not find container \"725fda3e17763bb514d9ba2d39dd7a4cfc02d5d585475411f3905020ac1f8e20\": container with ID starting with 725fda3e17763bb514d9ba2d39dd7a4cfc02d5d585475411f3905020ac1f8e20 not found: ID does not exist" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.387735 4907 scope.go:117] "RemoveContainer" containerID="7ca97160ecf2400b113e2721d2509a72b44110e201fc18a9c3d14fa335b64d1e" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.388090 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7ca97160ecf2400b113e2721d2509a72b44110e201fc18a9c3d14fa335b64d1e"} err="failed to get container status \"7ca97160ecf2400b113e2721d2509a72b44110e201fc18a9c3d14fa335b64d1e\": rpc error: code = NotFound desc = could not find container \"7ca97160ecf2400b113e2721d2509a72b44110e201fc18a9c3d14fa335b64d1e\": container with ID starting with 7ca97160ecf2400b113e2721d2509a72b44110e201fc18a9c3d14fa335b64d1e not found: ID does not exist" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.388204 4907 scope.go:117] "RemoveContainer" containerID="306d706e8b95e9a38af5ba674a473f59462492d56df0ecb87578b8edfdde9aa6" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.388451 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"306d706e8b95e9a38af5ba674a473f59462492d56df0ecb87578b8edfdde9aa6"} err="failed to get container status \"306d706e8b95e9a38af5ba674a473f59462492d56df0ecb87578b8edfdde9aa6\": rpc error: code = NotFound desc = could not find container \"306d706e8b95e9a38af5ba674a473f59462492d56df0ecb87578b8edfdde9aa6\": container with ID starting with 306d706e8b95e9a38af5ba674a473f59462492d56df0ecb87578b8edfdde9aa6 not found: ID does not exist" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.388530 4907 scope.go:117] "RemoveContainer" containerID="53d201dd158c6af9aadd5213934f6cc439d1b5e00b0a3e3af8f7f6824b4594d2" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.388744 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"53d201dd158c6af9aadd5213934f6cc439d1b5e00b0a3e3af8f7f6824b4594d2"} err="failed to get container status \"53d201dd158c6af9aadd5213934f6cc439d1b5e00b0a3e3af8f7f6824b4594d2\": rpc error: code = NotFound desc = could not find container \"53d201dd158c6af9aadd5213934f6cc439d1b5e00b0a3e3af8f7f6824b4594d2\": container with ID starting with 53d201dd158c6af9aadd5213934f6cc439d1b5e00b0a3e3af8f7f6824b4594d2 not found: ID does not exist" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.388825 4907 scope.go:117] "RemoveContainer" containerID="725fda3e17763bb514d9ba2d39dd7a4cfc02d5d585475411f3905020ac1f8e20" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.389050 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"725fda3e17763bb514d9ba2d39dd7a4cfc02d5d585475411f3905020ac1f8e20"} err="failed to get container status \"725fda3e17763bb514d9ba2d39dd7a4cfc02d5d585475411f3905020ac1f8e20\": rpc error: code = NotFound desc = could not find container \"725fda3e17763bb514d9ba2d39dd7a4cfc02d5d585475411f3905020ac1f8e20\": container with ID starting with 725fda3e17763bb514d9ba2d39dd7a4cfc02d5d585475411f3905020ac1f8e20 not found: ID does not exist" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.389136 4907 scope.go:117] "RemoveContainer" containerID="7ca97160ecf2400b113e2721d2509a72b44110e201fc18a9c3d14fa335b64d1e" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.389348 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7ca97160ecf2400b113e2721d2509a72b44110e201fc18a9c3d14fa335b64d1e"} err="failed to get container status \"7ca97160ecf2400b113e2721d2509a72b44110e201fc18a9c3d14fa335b64d1e\": rpc error: code = NotFound desc = could not find container \"7ca97160ecf2400b113e2721d2509a72b44110e201fc18a9c3d14fa335b64d1e\": container with ID starting with 7ca97160ecf2400b113e2721d2509a72b44110e201fc18a9c3d14fa335b64d1e not found: ID does not exist" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.389423 4907 scope.go:117] "RemoveContainer" containerID="306d706e8b95e9a38af5ba674a473f59462492d56df0ecb87578b8edfdde9aa6" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.389640 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"306d706e8b95e9a38af5ba674a473f59462492d56df0ecb87578b8edfdde9aa6"} err="failed to get container status \"306d706e8b95e9a38af5ba674a473f59462492d56df0ecb87578b8edfdde9aa6\": rpc error: code = NotFound desc = could not find container \"306d706e8b95e9a38af5ba674a473f59462492d56df0ecb87578b8edfdde9aa6\": container with ID starting with 306d706e8b95e9a38af5ba674a473f59462492d56df0ecb87578b8edfdde9aa6 not found: ID does not exist" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.389715 4907 scope.go:117] "RemoveContainer" containerID="53d201dd158c6af9aadd5213934f6cc439d1b5e00b0a3e3af8f7f6824b4594d2" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.389965 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"53d201dd158c6af9aadd5213934f6cc439d1b5e00b0a3e3af8f7f6824b4594d2"} err="failed to get container status \"53d201dd158c6af9aadd5213934f6cc439d1b5e00b0a3e3af8f7f6824b4594d2\": rpc error: code = NotFound desc = could not find container \"53d201dd158c6af9aadd5213934f6cc439d1b5e00b0a3e3af8f7f6824b4594d2\": container with ID starting with 53d201dd158c6af9aadd5213934f6cc439d1b5e00b0a3e3af8f7f6824b4594d2 not found: ID does not exist" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.461233 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d6be28d7-de73-47ad-9b71-f7333bae7487-log-httpd\") pod \"ceilometer-0\" (UID: \"d6be28d7-de73-47ad-9b71-f7333bae7487\") " pod="openstack/ceilometer-0" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.461419 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6be28d7-de73-47ad-9b71-f7333bae7487-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d6be28d7-de73-47ad-9b71-f7333bae7487\") " pod="openstack/ceilometer-0" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.461467 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d6be28d7-de73-47ad-9b71-f7333bae7487-scripts\") pod \"ceilometer-0\" (UID: \"d6be28d7-de73-47ad-9b71-f7333bae7487\") " pod="openstack/ceilometer-0" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.461605 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6be28d7-de73-47ad-9b71-f7333bae7487-config-data\") pod \"ceilometer-0\" (UID: \"d6be28d7-de73-47ad-9b71-f7333bae7487\") " pod="openstack/ceilometer-0" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.461660 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d6be28d7-de73-47ad-9b71-f7333bae7487-run-httpd\") pod \"ceilometer-0\" (UID: \"d6be28d7-de73-47ad-9b71-f7333bae7487\") " pod="openstack/ceilometer-0" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.461697 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fldgc\" (UniqueName: \"kubernetes.io/projected/d6be28d7-de73-47ad-9b71-f7333bae7487-kube-api-access-fldgc\") pod \"ceilometer-0\" (UID: \"d6be28d7-de73-47ad-9b71-f7333bae7487\") " pod="openstack/ceilometer-0" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.461726 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d6be28d7-de73-47ad-9b71-f7333bae7487-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d6be28d7-de73-47ad-9b71-f7333bae7487\") " pod="openstack/ceilometer-0" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.461757 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d6be28d7-de73-47ad-9b71-f7333bae7487-log-httpd\") pod \"ceilometer-0\" (UID: \"d6be28d7-de73-47ad-9b71-f7333bae7487\") " pod="openstack/ceilometer-0" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.462093 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d6be28d7-de73-47ad-9b71-f7333bae7487-run-httpd\") pod \"ceilometer-0\" (UID: \"d6be28d7-de73-47ad-9b71-f7333bae7487\") " pod="openstack/ceilometer-0" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.465847 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d6be28d7-de73-47ad-9b71-f7333bae7487-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d6be28d7-de73-47ad-9b71-f7333bae7487\") " pod="openstack/ceilometer-0" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.466347 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d6be28d7-de73-47ad-9b71-f7333bae7487-scripts\") pod \"ceilometer-0\" (UID: \"d6be28d7-de73-47ad-9b71-f7333bae7487\") " pod="openstack/ceilometer-0" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.466858 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6be28d7-de73-47ad-9b71-f7333bae7487-config-data\") pod \"ceilometer-0\" (UID: \"d6be28d7-de73-47ad-9b71-f7333bae7487\") " pod="openstack/ceilometer-0" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.469720 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6be28d7-de73-47ad-9b71-f7333bae7487-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d6be28d7-de73-47ad-9b71-f7333bae7487\") " pod="openstack/ceilometer-0" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.478484 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fldgc\" (UniqueName: \"kubernetes.io/projected/d6be28d7-de73-47ad-9b71-f7333bae7487-kube-api-access-fldgc\") pod \"ceilometer-0\" (UID: \"d6be28d7-de73-47ad-9b71-f7333bae7487\") " pod="openstack/ceilometer-0" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.654064 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.832804 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f7ca16f6-c9d9-4415-b8cf-b4753d784170" path="/var/lib/kubelet/pods/f7ca16f6-c9d9-4415-b8cf-b4753d784170/volumes" Mar 13 14:28:26 crc kubenswrapper[4907]: I0313 14:28:26.157559 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 13 14:28:26 crc kubenswrapper[4907]: W0313 14:28:26.161848 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd6be28d7_de73_47ad_9b71_f7333bae7487.slice/crio-1554f3d915bdc2d1f39f5b0de39037d95cef463ea1345bb0b4172dd3d3d6666a WatchSource:0}: Error finding container 1554f3d915bdc2d1f39f5b0de39037d95cef463ea1345bb0b4172dd3d3d6666a: Status 404 returned error can't find the container with id 1554f3d915bdc2d1f39f5b0de39037d95cef463ea1345bb0b4172dd3d3d6666a Mar 13 14:28:26 crc kubenswrapper[4907]: I0313 14:28:26.208835 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d6be28d7-de73-47ad-9b71-f7333bae7487","Type":"ContainerStarted","Data":"1554f3d915bdc2d1f39f5b0de39037d95cef463ea1345bb0b4172dd3d3d6666a"} Mar 13 14:28:27 crc kubenswrapper[4907]: I0313 14:28:27.221461 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d6be28d7-de73-47ad-9b71-f7333bae7487","Type":"ContainerStarted","Data":"d6e624a00a594aeaa75ca4ca30dbf27af1166a2309f4a98440260e91beadde21"} Mar 13 14:28:27 crc kubenswrapper[4907]: I0313 14:28:27.237230 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-z92h5"] Mar 13 14:28:27 crc kubenswrapper[4907]: I0313 14:28:27.238594 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-z92h5" Mar 13 14:28:27 crc kubenswrapper[4907]: I0313 14:28:27.240162 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-r4dpz" Mar 13 14:28:27 crc kubenswrapper[4907]: I0313 14:28:27.240500 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Mar 13 14:28:27 crc kubenswrapper[4907]: I0313 14:28:27.241677 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Mar 13 14:28:27 crc kubenswrapper[4907]: I0313 14:28:27.249785 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-z92h5"] Mar 13 14:28:27 crc kubenswrapper[4907]: I0313 14:28:27.320993 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dc5bv\" (UniqueName: \"kubernetes.io/projected/4d7ea7a2-ab25-4cb4-8749-10bc129cbd22-kube-api-access-dc5bv\") pod \"nova-cell0-conductor-db-sync-z92h5\" (UID: \"4d7ea7a2-ab25-4cb4-8749-10bc129cbd22\") " pod="openstack/nova-cell0-conductor-db-sync-z92h5" Mar 13 14:28:27 crc kubenswrapper[4907]: I0313 14:28:27.321295 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4d7ea7a2-ab25-4cb4-8749-10bc129cbd22-scripts\") pod \"nova-cell0-conductor-db-sync-z92h5\" (UID: \"4d7ea7a2-ab25-4cb4-8749-10bc129cbd22\") " pod="openstack/nova-cell0-conductor-db-sync-z92h5" Mar 13 14:28:27 crc kubenswrapper[4907]: I0313 14:28:27.321360 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d7ea7a2-ab25-4cb4-8749-10bc129cbd22-config-data\") pod \"nova-cell0-conductor-db-sync-z92h5\" (UID: \"4d7ea7a2-ab25-4cb4-8749-10bc129cbd22\") " pod="openstack/nova-cell0-conductor-db-sync-z92h5" Mar 13 14:28:27 crc kubenswrapper[4907]: I0313 14:28:27.321529 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d7ea7a2-ab25-4cb4-8749-10bc129cbd22-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-z92h5\" (UID: \"4d7ea7a2-ab25-4cb4-8749-10bc129cbd22\") " pod="openstack/nova-cell0-conductor-db-sync-z92h5" Mar 13 14:28:27 crc kubenswrapper[4907]: I0313 14:28:27.423853 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d7ea7a2-ab25-4cb4-8749-10bc129cbd22-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-z92h5\" (UID: \"4d7ea7a2-ab25-4cb4-8749-10bc129cbd22\") " pod="openstack/nova-cell0-conductor-db-sync-z92h5" Mar 13 14:28:27 crc kubenswrapper[4907]: I0313 14:28:27.423981 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dc5bv\" (UniqueName: \"kubernetes.io/projected/4d7ea7a2-ab25-4cb4-8749-10bc129cbd22-kube-api-access-dc5bv\") pod \"nova-cell0-conductor-db-sync-z92h5\" (UID: \"4d7ea7a2-ab25-4cb4-8749-10bc129cbd22\") " pod="openstack/nova-cell0-conductor-db-sync-z92h5" Mar 13 14:28:27 crc kubenswrapper[4907]: I0313 14:28:27.424138 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4d7ea7a2-ab25-4cb4-8749-10bc129cbd22-scripts\") pod \"nova-cell0-conductor-db-sync-z92h5\" (UID: \"4d7ea7a2-ab25-4cb4-8749-10bc129cbd22\") " pod="openstack/nova-cell0-conductor-db-sync-z92h5" Mar 13 14:28:27 crc kubenswrapper[4907]: I0313 14:28:27.424171 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d7ea7a2-ab25-4cb4-8749-10bc129cbd22-config-data\") pod \"nova-cell0-conductor-db-sync-z92h5\" (UID: \"4d7ea7a2-ab25-4cb4-8749-10bc129cbd22\") " pod="openstack/nova-cell0-conductor-db-sync-z92h5" Mar 13 14:28:27 crc kubenswrapper[4907]: I0313 14:28:27.432204 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d7ea7a2-ab25-4cb4-8749-10bc129cbd22-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-z92h5\" (UID: \"4d7ea7a2-ab25-4cb4-8749-10bc129cbd22\") " pod="openstack/nova-cell0-conductor-db-sync-z92h5" Mar 13 14:28:27 crc kubenswrapper[4907]: I0313 14:28:27.443292 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4d7ea7a2-ab25-4cb4-8749-10bc129cbd22-scripts\") pod \"nova-cell0-conductor-db-sync-z92h5\" (UID: \"4d7ea7a2-ab25-4cb4-8749-10bc129cbd22\") " pod="openstack/nova-cell0-conductor-db-sync-z92h5" Mar 13 14:28:27 crc kubenswrapper[4907]: I0313 14:28:27.451704 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d7ea7a2-ab25-4cb4-8749-10bc129cbd22-config-data\") pod \"nova-cell0-conductor-db-sync-z92h5\" (UID: \"4d7ea7a2-ab25-4cb4-8749-10bc129cbd22\") " pod="openstack/nova-cell0-conductor-db-sync-z92h5" Mar 13 14:28:27 crc kubenswrapper[4907]: I0313 14:28:27.453483 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dc5bv\" (UniqueName: \"kubernetes.io/projected/4d7ea7a2-ab25-4cb4-8749-10bc129cbd22-kube-api-access-dc5bv\") pod \"nova-cell0-conductor-db-sync-z92h5\" (UID: \"4d7ea7a2-ab25-4cb4-8749-10bc129cbd22\") " pod="openstack/nova-cell0-conductor-db-sync-z92h5" Mar 13 14:28:27 crc kubenswrapper[4907]: I0313 14:28:27.557717 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-z92h5" Mar 13 14:28:27 crc kubenswrapper[4907]: I0313 14:28:27.838349 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 13 14:28:28 crc kubenswrapper[4907]: I0313 14:28:28.076290 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-z92h5"] Mar 13 14:28:28 crc kubenswrapper[4907]: W0313 14:28:28.079392 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4d7ea7a2_ab25_4cb4_8749_10bc129cbd22.slice/crio-7d0624634945e5c5e075e91b43232f99c08cdffbe02fc1416ac047abb7f58463 WatchSource:0}: Error finding container 7d0624634945e5c5e075e91b43232f99c08cdffbe02fc1416ac047abb7f58463: Status 404 returned error can't find the container with id 7d0624634945e5c5e075e91b43232f99c08cdffbe02fc1416ac047abb7f58463 Mar 13 14:28:28 crc kubenswrapper[4907]: I0313 14:28:28.247226 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-z92h5" event={"ID":"4d7ea7a2-ab25-4cb4-8749-10bc129cbd22","Type":"ContainerStarted","Data":"7d0624634945e5c5e075e91b43232f99c08cdffbe02fc1416ac047abb7f58463"} Mar 13 14:28:28 crc kubenswrapper[4907]: I0313 14:28:28.248913 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d6be28d7-de73-47ad-9b71-f7333bae7487","Type":"ContainerStarted","Data":"af061011b60951ed9fad2d1d58dc7db5393d0bd0c5f54d1158edbedf15496e81"} Mar 13 14:28:29 crc kubenswrapper[4907]: I0313 14:28:29.265863 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d6be28d7-de73-47ad-9b71-f7333bae7487","Type":"ContainerStarted","Data":"7174b12084675db6df525f79c405a1d67339b8435a26c05853001bd9082cd33e"} Mar 13 14:28:29 crc kubenswrapper[4907]: I0313 14:28:29.493183 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Mar 13 14:28:29 crc kubenswrapper[4907]: I0313 14:28:29.493239 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Mar 13 14:28:29 crc kubenswrapper[4907]: I0313 14:28:29.546524 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Mar 13 14:28:29 crc kubenswrapper[4907]: I0313 14:28:29.553654 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Mar 13 14:28:30 crc kubenswrapper[4907]: I0313 14:28:30.282312 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d6be28d7-de73-47ad-9b71-f7333bae7487","Type":"ContainerStarted","Data":"c34a9a5b2a3bb1f13c00c35e532469604057426d78bfb43c38cf1c8fd94a7c99"} Mar 13 14:28:30 crc kubenswrapper[4907]: I0313 14:28:30.282863 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Mar 13 14:28:30 crc kubenswrapper[4907]: I0313 14:28:30.282896 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Mar 13 14:28:30 crc kubenswrapper[4907]: I0313 14:28:30.282895 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d6be28d7-de73-47ad-9b71-f7333bae7487" containerName="ceilometer-notification-agent" containerID="cri-o://af061011b60951ed9fad2d1d58dc7db5393d0bd0c5f54d1158edbedf15496e81" gracePeriod=30 Mar 13 14:28:30 crc kubenswrapper[4907]: I0313 14:28:30.282928 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d6be28d7-de73-47ad-9b71-f7333bae7487" containerName="sg-core" containerID="cri-o://7174b12084675db6df525f79c405a1d67339b8435a26c05853001bd9082cd33e" gracePeriod=30 Mar 13 14:28:30 crc kubenswrapper[4907]: I0313 14:28:30.282584 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d6be28d7-de73-47ad-9b71-f7333bae7487" containerName="ceilometer-central-agent" containerID="cri-o://d6e624a00a594aeaa75ca4ca30dbf27af1166a2309f4a98440260e91beadde21" gracePeriod=30 Mar 13 14:28:30 crc kubenswrapper[4907]: I0313 14:28:30.282811 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d6be28d7-de73-47ad-9b71-f7333bae7487" containerName="proxy-httpd" containerID="cri-o://c34a9a5b2a3bb1f13c00c35e532469604057426d78bfb43c38cf1c8fd94a7c99" gracePeriod=30 Mar 13 14:28:30 crc kubenswrapper[4907]: I0313 14:28:30.307752 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.768228928 podStartE2EDuration="5.307731263s" podCreationTimestamp="2026-03-13 14:28:25 +0000 UTC" firstStartedPulling="2026-03-13 14:28:26.164107428 +0000 UTC m=+1405.063895117" lastFinishedPulling="2026-03-13 14:28:29.703609753 +0000 UTC m=+1408.603397452" observedRunningTime="2026-03-13 14:28:30.303623792 +0000 UTC m=+1409.203411481" watchObservedRunningTime="2026-03-13 14:28:30.307731263 +0000 UTC m=+1409.207518952" Mar 13 14:28:30 crc kubenswrapper[4907]: I0313 14:28:30.348720 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Mar 13 14:28:30 crc kubenswrapper[4907]: I0313 14:28:30.348779 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Mar 13 14:28:30 crc kubenswrapper[4907]: I0313 14:28:30.381702 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Mar 13 14:28:30 crc kubenswrapper[4907]: I0313 14:28:30.404239 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Mar 13 14:28:31 crc kubenswrapper[4907]: I0313 14:28:31.090458 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-85dc486cc4-fwrm6" Mar 13 14:28:31 crc kubenswrapper[4907]: I0313 14:28:31.293716 4907 generic.go:334] "Generic (PLEG): container finished" podID="d6be28d7-de73-47ad-9b71-f7333bae7487" containerID="c34a9a5b2a3bb1f13c00c35e532469604057426d78bfb43c38cf1c8fd94a7c99" exitCode=0 Mar 13 14:28:31 crc kubenswrapper[4907]: I0313 14:28:31.294150 4907 generic.go:334] "Generic (PLEG): container finished" podID="d6be28d7-de73-47ad-9b71-f7333bae7487" containerID="7174b12084675db6df525f79c405a1d67339b8435a26c05853001bd9082cd33e" exitCode=2 Mar 13 14:28:31 crc kubenswrapper[4907]: I0313 14:28:31.294170 4907 generic.go:334] "Generic (PLEG): container finished" podID="d6be28d7-de73-47ad-9b71-f7333bae7487" containerID="af061011b60951ed9fad2d1d58dc7db5393d0bd0c5f54d1158edbedf15496e81" exitCode=0 Mar 13 14:28:31 crc kubenswrapper[4907]: I0313 14:28:31.293946 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d6be28d7-de73-47ad-9b71-f7333bae7487","Type":"ContainerDied","Data":"c34a9a5b2a3bb1f13c00c35e532469604057426d78bfb43c38cf1c8fd94a7c99"} Mar 13 14:28:31 crc kubenswrapper[4907]: I0313 14:28:31.295258 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d6be28d7-de73-47ad-9b71-f7333bae7487","Type":"ContainerDied","Data":"7174b12084675db6df525f79c405a1d67339b8435a26c05853001bd9082cd33e"} Mar 13 14:28:31 crc kubenswrapper[4907]: I0313 14:28:31.295277 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d6be28d7-de73-47ad-9b71-f7333bae7487","Type":"ContainerDied","Data":"af061011b60951ed9fad2d1d58dc7db5393d0bd0c5f54d1158edbedf15496e81"} Mar 13 14:28:31 crc kubenswrapper[4907]: I0313 14:28:31.295777 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Mar 13 14:28:31 crc kubenswrapper[4907]: I0313 14:28:31.295938 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Mar 13 14:28:31 crc kubenswrapper[4907]: I0313 14:28:31.324959 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-85dc486cc4-fwrm6" Mar 13 14:28:31 crc kubenswrapper[4907]: I0313 14:28:31.415934 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-67d6ccd9c4-ms98l"] Mar 13 14:28:31 crc kubenswrapper[4907]: I0313 14:28:31.416376 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement-67d6ccd9c4-ms98l" podUID="45d14b2b-1d0e-4e15-b837-0d2fb6aed785" containerName="placement-log" containerID="cri-o://e2ad3bad04a843173cf63dc1e67628c441f9aeb3514e5f9bf8afda3eba7b52f6" gracePeriod=30 Mar 13 14:28:31 crc kubenswrapper[4907]: I0313 14:28:31.416695 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement-67d6ccd9c4-ms98l" podUID="45d14b2b-1d0e-4e15-b837-0d2fb6aed785" containerName="placement-api" containerID="cri-o://38b798d627fc6c621c04314f998e3ab08f943e11ebed179dd2fc6641c6c999a2" gracePeriod=30 Mar 13 14:28:32 crc kubenswrapper[4907]: I0313 14:28:32.305974 4907 generic.go:334] "Generic (PLEG): container finished" podID="45d14b2b-1d0e-4e15-b837-0d2fb6aed785" containerID="e2ad3bad04a843173cf63dc1e67628c441f9aeb3514e5f9bf8afda3eba7b52f6" exitCode=143 Mar 13 14:28:32 crc kubenswrapper[4907]: I0313 14:28:32.306094 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-67d6ccd9c4-ms98l" event={"ID":"45d14b2b-1d0e-4e15-b837-0d2fb6aed785","Type":"ContainerDied","Data":"e2ad3bad04a843173cf63dc1e67628c441f9aeb3514e5f9bf8afda3eba7b52f6"} Mar 13 14:28:32 crc kubenswrapper[4907]: I0313 14:28:32.708782 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Mar 13 14:28:32 crc kubenswrapper[4907]: I0313 14:28:32.709246 4907 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 13 14:28:33 crc kubenswrapper[4907]: I0313 14:28:33.036537 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Mar 13 14:28:33 crc kubenswrapper[4907]: I0313 14:28:33.314479 4907 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 13 14:28:33 crc kubenswrapper[4907]: I0313 14:28:33.314505 4907 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 13 14:28:33 crc kubenswrapper[4907]: I0313 14:28:33.847513 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Mar 13 14:28:33 crc kubenswrapper[4907]: I0313 14:28:33.849974 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Mar 13 14:28:34 crc kubenswrapper[4907]: E0313 14:28:34.716720 4907 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod45d14b2b_1d0e_4e15_b837_0d2fb6aed785.slice/crio-conmon-38b798d627fc6c621c04314f998e3ab08f943e11ebed179dd2fc6641c6c999a2.scope\": RecentStats: unable to find data in memory cache]" Mar 13 14:28:35 crc kubenswrapper[4907]: I0313 14:28:35.338957 4907 generic.go:334] "Generic (PLEG): container finished" podID="45d14b2b-1d0e-4e15-b837-0d2fb6aed785" containerID="38b798d627fc6c621c04314f998e3ab08f943e11ebed179dd2fc6641c6c999a2" exitCode=0 Mar 13 14:28:35 crc kubenswrapper[4907]: I0313 14:28:35.340188 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-67d6ccd9c4-ms98l" event={"ID":"45d14b2b-1d0e-4e15-b837-0d2fb6aed785","Type":"ContainerDied","Data":"38b798d627fc6c621c04314f998e3ab08f943e11ebed179dd2fc6641c6c999a2"} Mar 13 14:28:38 crc kubenswrapper[4907]: I0313 14:28:38.400523 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-67d6ccd9c4-ms98l" Mar 13 14:28:38 crc kubenswrapper[4907]: I0313 14:28:38.410604 4907 generic.go:334] "Generic (PLEG): container finished" podID="d6be28d7-de73-47ad-9b71-f7333bae7487" containerID="d6e624a00a594aeaa75ca4ca30dbf27af1166a2309f4a98440260e91beadde21" exitCode=0 Mar 13 14:28:38 crc kubenswrapper[4907]: I0313 14:28:38.410662 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d6be28d7-de73-47ad-9b71-f7333bae7487","Type":"ContainerDied","Data":"d6e624a00a594aeaa75ca4ca30dbf27af1166a2309f4a98440260e91beadde21"} Mar 13 14:28:38 crc kubenswrapper[4907]: I0313 14:28:38.413158 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-67d6ccd9c4-ms98l" event={"ID":"45d14b2b-1d0e-4e15-b837-0d2fb6aed785","Type":"ContainerDied","Data":"0497ac4d8503d6e45cbe04af5eff7d939b60b6a6232917e037fb11fdbb11b8dd"} Mar 13 14:28:38 crc kubenswrapper[4907]: I0313 14:28:38.413183 4907 scope.go:117] "RemoveContainer" containerID="38b798d627fc6c621c04314f998e3ab08f943e11ebed179dd2fc6641c6c999a2" Mar 13 14:28:38 crc kubenswrapper[4907]: I0313 14:28:38.413314 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-67d6ccd9c4-ms98l" Mar 13 14:28:38 crc kubenswrapper[4907]: I0313 14:28:38.441726 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/45d14b2b-1d0e-4e15-b837-0d2fb6aed785-scripts\") pod \"45d14b2b-1d0e-4e15-b837-0d2fb6aed785\" (UID: \"45d14b2b-1d0e-4e15-b837-0d2fb6aed785\") " Mar 13 14:28:38 crc kubenswrapper[4907]: I0313 14:28:38.441805 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45d14b2b-1d0e-4e15-b837-0d2fb6aed785-combined-ca-bundle\") pod \"45d14b2b-1d0e-4e15-b837-0d2fb6aed785\" (UID: \"45d14b2b-1d0e-4e15-b837-0d2fb6aed785\") " Mar 13 14:28:38 crc kubenswrapper[4907]: I0313 14:28:38.441861 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4sstr\" (UniqueName: \"kubernetes.io/projected/45d14b2b-1d0e-4e15-b837-0d2fb6aed785-kube-api-access-4sstr\") pod \"45d14b2b-1d0e-4e15-b837-0d2fb6aed785\" (UID: \"45d14b2b-1d0e-4e15-b837-0d2fb6aed785\") " Mar 13 14:28:38 crc kubenswrapper[4907]: I0313 14:28:38.441918 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/45d14b2b-1d0e-4e15-b837-0d2fb6aed785-config-data\") pod \"45d14b2b-1d0e-4e15-b837-0d2fb6aed785\" (UID: \"45d14b2b-1d0e-4e15-b837-0d2fb6aed785\") " Mar 13 14:28:38 crc kubenswrapper[4907]: I0313 14:28:38.441960 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/45d14b2b-1d0e-4e15-b837-0d2fb6aed785-logs\") pod \"45d14b2b-1d0e-4e15-b837-0d2fb6aed785\" (UID: \"45d14b2b-1d0e-4e15-b837-0d2fb6aed785\") " Mar 13 14:28:38 crc kubenswrapper[4907]: I0313 14:28:38.441986 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/45d14b2b-1d0e-4e15-b837-0d2fb6aed785-internal-tls-certs\") pod \"45d14b2b-1d0e-4e15-b837-0d2fb6aed785\" (UID: \"45d14b2b-1d0e-4e15-b837-0d2fb6aed785\") " Mar 13 14:28:38 crc kubenswrapper[4907]: I0313 14:28:38.442034 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/45d14b2b-1d0e-4e15-b837-0d2fb6aed785-public-tls-certs\") pod \"45d14b2b-1d0e-4e15-b837-0d2fb6aed785\" (UID: \"45d14b2b-1d0e-4e15-b837-0d2fb6aed785\") " Mar 13 14:28:38 crc kubenswrapper[4907]: I0313 14:28:38.448672 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/45d14b2b-1d0e-4e15-b837-0d2fb6aed785-logs" (OuterVolumeSpecName: "logs") pod "45d14b2b-1d0e-4e15-b837-0d2fb6aed785" (UID: "45d14b2b-1d0e-4e15-b837-0d2fb6aed785"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:28:38 crc kubenswrapper[4907]: I0313 14:28:38.454025 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/45d14b2b-1d0e-4e15-b837-0d2fb6aed785-scripts" (OuterVolumeSpecName: "scripts") pod "45d14b2b-1d0e-4e15-b837-0d2fb6aed785" (UID: "45d14b2b-1d0e-4e15-b837-0d2fb6aed785"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:28:38 crc kubenswrapper[4907]: I0313 14:28:38.454449 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/45d14b2b-1d0e-4e15-b837-0d2fb6aed785-kube-api-access-4sstr" (OuterVolumeSpecName: "kube-api-access-4sstr") pod "45d14b2b-1d0e-4e15-b837-0d2fb6aed785" (UID: "45d14b2b-1d0e-4e15-b837-0d2fb6aed785"). InnerVolumeSpecName "kube-api-access-4sstr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:28:38 crc kubenswrapper[4907]: I0313 14:28:38.466454 4907 scope.go:117] "RemoveContainer" containerID="e2ad3bad04a843173cf63dc1e67628c441f9aeb3514e5f9bf8afda3eba7b52f6" Mar 13 14:28:38 crc kubenswrapper[4907]: I0313 14:28:38.508269 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/45d14b2b-1d0e-4e15-b837-0d2fb6aed785-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "45d14b2b-1d0e-4e15-b837-0d2fb6aed785" (UID: "45d14b2b-1d0e-4e15-b837-0d2fb6aed785"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:28:38 crc kubenswrapper[4907]: I0313 14:28:38.513140 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/45d14b2b-1d0e-4e15-b837-0d2fb6aed785-config-data" (OuterVolumeSpecName: "config-data") pod "45d14b2b-1d0e-4e15-b837-0d2fb6aed785" (UID: "45d14b2b-1d0e-4e15-b837-0d2fb6aed785"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:28:38 crc kubenswrapper[4907]: I0313 14:28:38.543744 4907 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/45d14b2b-1d0e-4e15-b837-0d2fb6aed785-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:38 crc kubenswrapper[4907]: I0313 14:28:38.543773 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45d14b2b-1d0e-4e15-b837-0d2fb6aed785-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:38 crc kubenswrapper[4907]: I0313 14:28:38.543783 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4sstr\" (UniqueName: \"kubernetes.io/projected/45d14b2b-1d0e-4e15-b837-0d2fb6aed785-kube-api-access-4sstr\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:38 crc kubenswrapper[4907]: I0313 14:28:38.543791 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/45d14b2b-1d0e-4e15-b837-0d2fb6aed785-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:38 crc kubenswrapper[4907]: I0313 14:28:38.543800 4907 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/45d14b2b-1d0e-4e15-b837-0d2fb6aed785-logs\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:38 crc kubenswrapper[4907]: I0313 14:28:38.571996 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 14:28:38 crc kubenswrapper[4907]: I0313 14:28:38.597123 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/45d14b2b-1d0e-4e15-b837-0d2fb6aed785-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "45d14b2b-1d0e-4e15-b837-0d2fb6aed785" (UID: "45d14b2b-1d0e-4e15-b837-0d2fb6aed785"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:28:38 crc kubenswrapper[4907]: I0313 14:28:38.616314 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/45d14b2b-1d0e-4e15-b837-0d2fb6aed785-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "45d14b2b-1d0e-4e15-b837-0d2fb6aed785" (UID: "45d14b2b-1d0e-4e15-b837-0d2fb6aed785"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:28:38 crc kubenswrapper[4907]: I0313 14:28:38.644398 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d6be28d7-de73-47ad-9b71-f7333bae7487-run-httpd\") pod \"d6be28d7-de73-47ad-9b71-f7333bae7487\" (UID: \"d6be28d7-de73-47ad-9b71-f7333bae7487\") " Mar 13 14:28:38 crc kubenswrapper[4907]: I0313 14:28:38.644454 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fldgc\" (UniqueName: \"kubernetes.io/projected/d6be28d7-de73-47ad-9b71-f7333bae7487-kube-api-access-fldgc\") pod \"d6be28d7-de73-47ad-9b71-f7333bae7487\" (UID: \"d6be28d7-de73-47ad-9b71-f7333bae7487\") " Mar 13 14:28:38 crc kubenswrapper[4907]: I0313 14:28:38.644512 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d6be28d7-de73-47ad-9b71-f7333bae7487-scripts\") pod \"d6be28d7-de73-47ad-9b71-f7333bae7487\" (UID: \"d6be28d7-de73-47ad-9b71-f7333bae7487\") " Mar 13 14:28:38 crc kubenswrapper[4907]: I0313 14:28:38.644549 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6be28d7-de73-47ad-9b71-f7333bae7487-config-data\") pod \"d6be28d7-de73-47ad-9b71-f7333bae7487\" (UID: \"d6be28d7-de73-47ad-9b71-f7333bae7487\") " Mar 13 14:28:38 crc kubenswrapper[4907]: I0313 14:28:38.644936 4907 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/45d14b2b-1d0e-4e15-b837-0d2fb6aed785-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:38 crc kubenswrapper[4907]: I0313 14:28:38.644956 4907 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/45d14b2b-1d0e-4e15-b837-0d2fb6aed785-public-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:38 crc kubenswrapper[4907]: I0313 14:28:38.645368 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d6be28d7-de73-47ad-9b71-f7333bae7487-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "d6be28d7-de73-47ad-9b71-f7333bae7487" (UID: "d6be28d7-de73-47ad-9b71-f7333bae7487"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:28:38 crc kubenswrapper[4907]: I0313 14:28:38.649966 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d6be28d7-de73-47ad-9b71-f7333bae7487-kube-api-access-fldgc" (OuterVolumeSpecName: "kube-api-access-fldgc") pod "d6be28d7-de73-47ad-9b71-f7333bae7487" (UID: "d6be28d7-de73-47ad-9b71-f7333bae7487"). InnerVolumeSpecName "kube-api-access-fldgc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:28:38 crc kubenswrapper[4907]: I0313 14:28:38.651786 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6be28d7-de73-47ad-9b71-f7333bae7487-scripts" (OuterVolumeSpecName: "scripts") pod "d6be28d7-de73-47ad-9b71-f7333bae7487" (UID: "d6be28d7-de73-47ad-9b71-f7333bae7487"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:28:38 crc kubenswrapper[4907]: I0313 14:28:38.744213 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6be28d7-de73-47ad-9b71-f7333bae7487-config-data" (OuterVolumeSpecName: "config-data") pod "d6be28d7-de73-47ad-9b71-f7333bae7487" (UID: "d6be28d7-de73-47ad-9b71-f7333bae7487"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:28:38 crc kubenswrapper[4907]: I0313 14:28:38.745996 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d6be28d7-de73-47ad-9b71-f7333bae7487-sg-core-conf-yaml\") pod \"d6be28d7-de73-47ad-9b71-f7333bae7487\" (UID: \"d6be28d7-de73-47ad-9b71-f7333bae7487\") " Mar 13 14:28:38 crc kubenswrapper[4907]: I0313 14:28:38.746038 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d6be28d7-de73-47ad-9b71-f7333bae7487-log-httpd\") pod \"d6be28d7-de73-47ad-9b71-f7333bae7487\" (UID: \"d6be28d7-de73-47ad-9b71-f7333bae7487\") " Mar 13 14:28:38 crc kubenswrapper[4907]: I0313 14:28:38.746060 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6be28d7-de73-47ad-9b71-f7333bae7487-combined-ca-bundle\") pod \"d6be28d7-de73-47ad-9b71-f7333bae7487\" (UID: \"d6be28d7-de73-47ad-9b71-f7333bae7487\") " Mar 13 14:28:38 crc kubenswrapper[4907]: I0313 14:28:38.746497 4907 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d6be28d7-de73-47ad-9b71-f7333bae7487-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:38 crc kubenswrapper[4907]: I0313 14:28:38.746516 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fldgc\" (UniqueName: \"kubernetes.io/projected/d6be28d7-de73-47ad-9b71-f7333bae7487-kube-api-access-fldgc\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:38 crc kubenswrapper[4907]: I0313 14:28:38.746526 4907 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d6be28d7-de73-47ad-9b71-f7333bae7487-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:38 crc kubenswrapper[4907]: I0313 14:28:38.746536 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6be28d7-de73-47ad-9b71-f7333bae7487-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:38 crc kubenswrapper[4907]: I0313 14:28:38.746669 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d6be28d7-de73-47ad-9b71-f7333bae7487-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "d6be28d7-de73-47ad-9b71-f7333bae7487" (UID: "d6be28d7-de73-47ad-9b71-f7333bae7487"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:28:38 crc kubenswrapper[4907]: I0313 14:28:38.749770 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-67d6ccd9c4-ms98l"] Mar 13 14:28:38 crc kubenswrapper[4907]: I0313 14:28:38.760189 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-67d6ccd9c4-ms98l"] Mar 13 14:28:38 crc kubenswrapper[4907]: I0313 14:28:38.768582 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6be28d7-de73-47ad-9b71-f7333bae7487-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "d6be28d7-de73-47ad-9b71-f7333bae7487" (UID: "d6be28d7-de73-47ad-9b71-f7333bae7487"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:28:38 crc kubenswrapper[4907]: I0313 14:28:38.812548 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6be28d7-de73-47ad-9b71-f7333bae7487-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d6be28d7-de73-47ad-9b71-f7333bae7487" (UID: "d6be28d7-de73-47ad-9b71-f7333bae7487"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:28:38 crc kubenswrapper[4907]: I0313 14:28:38.848689 4907 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d6be28d7-de73-47ad-9b71-f7333bae7487-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:38 crc kubenswrapper[4907]: I0313 14:28:38.848749 4907 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d6be28d7-de73-47ad-9b71-f7333bae7487-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:38 crc kubenswrapper[4907]: I0313 14:28:38.848760 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6be28d7-de73-47ad-9b71-f7333bae7487-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:39 crc kubenswrapper[4907]: I0313 14:28:39.426020 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d6be28d7-de73-47ad-9b71-f7333bae7487","Type":"ContainerDied","Data":"1554f3d915bdc2d1f39f5b0de39037d95cef463ea1345bb0b4172dd3d3d6666a"} Mar 13 14:28:39 crc kubenswrapper[4907]: I0313 14:28:39.426096 4907 scope.go:117] "RemoveContainer" containerID="c34a9a5b2a3bb1f13c00c35e532469604057426d78bfb43c38cf1c8fd94a7c99" Mar 13 14:28:39 crc kubenswrapper[4907]: I0313 14:28:39.426111 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 14:28:39 crc kubenswrapper[4907]: I0313 14:28:39.431203 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-z92h5" event={"ID":"4d7ea7a2-ab25-4cb4-8749-10bc129cbd22","Type":"ContainerStarted","Data":"d935db5c32cf1af5583418c37d51248a6d17b0514af1c3fa70469ca109747682"} Mar 13 14:28:39 crc kubenswrapper[4907]: I0313 14:28:39.460736 4907 scope.go:117] "RemoveContainer" containerID="7174b12084675db6df525f79c405a1d67339b8435a26c05853001bd9082cd33e" Mar 13 14:28:39 crc kubenswrapper[4907]: I0313 14:28:39.463866 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-z92h5" podStartSLOduration=2.129845158 podStartE2EDuration="12.463839136s" podCreationTimestamp="2026-03-13 14:28:27 +0000 UTC" firstStartedPulling="2026-03-13 14:28:28.081499466 +0000 UTC m=+1406.981287155" lastFinishedPulling="2026-03-13 14:28:38.415493444 +0000 UTC m=+1417.315281133" observedRunningTime="2026-03-13 14:28:39.456911378 +0000 UTC m=+1418.356699067" watchObservedRunningTime="2026-03-13 14:28:39.463839136 +0000 UTC m=+1418.363626815" Mar 13 14:28:39 crc kubenswrapper[4907]: I0313 14:28:39.481147 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 13 14:28:39 crc kubenswrapper[4907]: I0313 14:28:39.487777 4907 scope.go:117] "RemoveContainer" containerID="af061011b60951ed9fad2d1d58dc7db5393d0bd0c5f54d1158edbedf15496e81" Mar 13 14:28:39 crc kubenswrapper[4907]: I0313 14:28:39.498754 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Mar 13 14:28:39 crc kubenswrapper[4907]: I0313 14:28:39.505136 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 13 14:28:39 crc kubenswrapper[4907]: E0313 14:28:39.505511 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6be28d7-de73-47ad-9b71-f7333bae7487" containerName="ceilometer-central-agent" Mar 13 14:28:39 crc kubenswrapper[4907]: I0313 14:28:39.505527 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6be28d7-de73-47ad-9b71-f7333bae7487" containerName="ceilometer-central-agent" Mar 13 14:28:39 crc kubenswrapper[4907]: E0313 14:28:39.505555 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45d14b2b-1d0e-4e15-b837-0d2fb6aed785" containerName="placement-log" Mar 13 14:28:39 crc kubenswrapper[4907]: I0313 14:28:39.505562 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="45d14b2b-1d0e-4e15-b837-0d2fb6aed785" containerName="placement-log" Mar 13 14:28:39 crc kubenswrapper[4907]: E0313 14:28:39.505574 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6be28d7-de73-47ad-9b71-f7333bae7487" containerName="sg-core" Mar 13 14:28:39 crc kubenswrapper[4907]: I0313 14:28:39.505581 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6be28d7-de73-47ad-9b71-f7333bae7487" containerName="sg-core" Mar 13 14:28:39 crc kubenswrapper[4907]: E0313 14:28:39.505589 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6be28d7-de73-47ad-9b71-f7333bae7487" containerName="ceilometer-notification-agent" Mar 13 14:28:39 crc kubenswrapper[4907]: I0313 14:28:39.505595 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6be28d7-de73-47ad-9b71-f7333bae7487" containerName="ceilometer-notification-agent" Mar 13 14:28:39 crc kubenswrapper[4907]: E0313 14:28:39.505616 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6be28d7-de73-47ad-9b71-f7333bae7487" containerName="proxy-httpd" Mar 13 14:28:39 crc kubenswrapper[4907]: I0313 14:28:39.505621 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6be28d7-de73-47ad-9b71-f7333bae7487" containerName="proxy-httpd" Mar 13 14:28:39 crc kubenswrapper[4907]: E0313 14:28:39.505633 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45d14b2b-1d0e-4e15-b837-0d2fb6aed785" containerName="placement-api" Mar 13 14:28:39 crc kubenswrapper[4907]: I0313 14:28:39.505638 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="45d14b2b-1d0e-4e15-b837-0d2fb6aed785" containerName="placement-api" Mar 13 14:28:39 crc kubenswrapper[4907]: I0313 14:28:39.505791 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6be28d7-de73-47ad-9b71-f7333bae7487" containerName="proxy-httpd" Mar 13 14:28:39 crc kubenswrapper[4907]: I0313 14:28:39.505803 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="45d14b2b-1d0e-4e15-b837-0d2fb6aed785" containerName="placement-api" Mar 13 14:28:39 crc kubenswrapper[4907]: I0313 14:28:39.505814 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6be28d7-de73-47ad-9b71-f7333bae7487" containerName="ceilometer-notification-agent" Mar 13 14:28:39 crc kubenswrapper[4907]: I0313 14:28:39.505826 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6be28d7-de73-47ad-9b71-f7333bae7487" containerName="sg-core" Mar 13 14:28:39 crc kubenswrapper[4907]: I0313 14:28:39.505836 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6be28d7-de73-47ad-9b71-f7333bae7487" containerName="ceilometer-central-agent" Mar 13 14:28:39 crc kubenswrapper[4907]: I0313 14:28:39.505846 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="45d14b2b-1d0e-4e15-b837-0d2fb6aed785" containerName="placement-log" Mar 13 14:28:39 crc kubenswrapper[4907]: I0313 14:28:39.507502 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 14:28:39 crc kubenswrapper[4907]: I0313 14:28:39.516938 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 13 14:28:39 crc kubenswrapper[4907]: I0313 14:28:39.519276 4907 scope.go:117] "RemoveContainer" containerID="d6e624a00a594aeaa75ca4ca30dbf27af1166a2309f4a98440260e91beadde21" Mar 13 14:28:39 crc kubenswrapper[4907]: I0313 14:28:39.519334 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 13 14:28:39 crc kubenswrapper[4907]: I0313 14:28:39.519372 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 13 14:28:39 crc kubenswrapper[4907]: I0313 14:28:39.561233 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b8e62ba8-78aa-4acd-92f3-5adf43ca85ad-log-httpd\") pod \"ceilometer-0\" (UID: \"b8e62ba8-78aa-4acd-92f3-5adf43ca85ad\") " pod="openstack/ceilometer-0" Mar 13 14:28:39 crc kubenswrapper[4907]: I0313 14:28:39.561286 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8e62ba8-78aa-4acd-92f3-5adf43ca85ad-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b8e62ba8-78aa-4acd-92f3-5adf43ca85ad\") " pod="openstack/ceilometer-0" Mar 13 14:28:39 crc kubenswrapper[4907]: I0313 14:28:39.561313 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b8e62ba8-78aa-4acd-92f3-5adf43ca85ad-config-data\") pod \"ceilometer-0\" (UID: \"b8e62ba8-78aa-4acd-92f3-5adf43ca85ad\") " pod="openstack/ceilometer-0" Mar 13 14:28:39 crc kubenswrapper[4907]: I0313 14:28:39.561454 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b8e62ba8-78aa-4acd-92f3-5adf43ca85ad-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b8e62ba8-78aa-4acd-92f3-5adf43ca85ad\") " pod="openstack/ceilometer-0" Mar 13 14:28:39 crc kubenswrapper[4907]: I0313 14:28:39.561476 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hdhb7\" (UniqueName: \"kubernetes.io/projected/b8e62ba8-78aa-4acd-92f3-5adf43ca85ad-kube-api-access-hdhb7\") pod \"ceilometer-0\" (UID: \"b8e62ba8-78aa-4acd-92f3-5adf43ca85ad\") " pod="openstack/ceilometer-0" Mar 13 14:28:39 crc kubenswrapper[4907]: I0313 14:28:39.561496 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b8e62ba8-78aa-4acd-92f3-5adf43ca85ad-scripts\") pod \"ceilometer-0\" (UID: \"b8e62ba8-78aa-4acd-92f3-5adf43ca85ad\") " pod="openstack/ceilometer-0" Mar 13 14:28:39 crc kubenswrapper[4907]: I0313 14:28:39.561524 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b8e62ba8-78aa-4acd-92f3-5adf43ca85ad-run-httpd\") pod \"ceilometer-0\" (UID: \"b8e62ba8-78aa-4acd-92f3-5adf43ca85ad\") " pod="openstack/ceilometer-0" Mar 13 14:28:39 crc kubenswrapper[4907]: I0313 14:28:39.663055 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b8e62ba8-78aa-4acd-92f3-5adf43ca85ad-log-httpd\") pod \"ceilometer-0\" (UID: \"b8e62ba8-78aa-4acd-92f3-5adf43ca85ad\") " pod="openstack/ceilometer-0" Mar 13 14:28:39 crc kubenswrapper[4907]: I0313 14:28:39.663122 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8e62ba8-78aa-4acd-92f3-5adf43ca85ad-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b8e62ba8-78aa-4acd-92f3-5adf43ca85ad\") " pod="openstack/ceilometer-0" Mar 13 14:28:39 crc kubenswrapper[4907]: I0313 14:28:39.663144 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b8e62ba8-78aa-4acd-92f3-5adf43ca85ad-config-data\") pod \"ceilometer-0\" (UID: \"b8e62ba8-78aa-4acd-92f3-5adf43ca85ad\") " pod="openstack/ceilometer-0" Mar 13 14:28:39 crc kubenswrapper[4907]: I0313 14:28:39.663268 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b8e62ba8-78aa-4acd-92f3-5adf43ca85ad-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b8e62ba8-78aa-4acd-92f3-5adf43ca85ad\") " pod="openstack/ceilometer-0" Mar 13 14:28:39 crc kubenswrapper[4907]: I0313 14:28:39.663286 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hdhb7\" (UniqueName: \"kubernetes.io/projected/b8e62ba8-78aa-4acd-92f3-5adf43ca85ad-kube-api-access-hdhb7\") pod \"ceilometer-0\" (UID: \"b8e62ba8-78aa-4acd-92f3-5adf43ca85ad\") " pod="openstack/ceilometer-0" Mar 13 14:28:39 crc kubenswrapper[4907]: I0313 14:28:39.663311 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b8e62ba8-78aa-4acd-92f3-5adf43ca85ad-scripts\") pod \"ceilometer-0\" (UID: \"b8e62ba8-78aa-4acd-92f3-5adf43ca85ad\") " pod="openstack/ceilometer-0" Mar 13 14:28:39 crc kubenswrapper[4907]: I0313 14:28:39.663335 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b8e62ba8-78aa-4acd-92f3-5adf43ca85ad-run-httpd\") pod \"ceilometer-0\" (UID: \"b8e62ba8-78aa-4acd-92f3-5adf43ca85ad\") " pod="openstack/ceilometer-0" Mar 13 14:28:39 crc kubenswrapper[4907]: I0313 14:28:39.663985 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b8e62ba8-78aa-4acd-92f3-5adf43ca85ad-log-httpd\") pod \"ceilometer-0\" (UID: \"b8e62ba8-78aa-4acd-92f3-5adf43ca85ad\") " pod="openstack/ceilometer-0" Mar 13 14:28:39 crc kubenswrapper[4907]: I0313 14:28:39.664065 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b8e62ba8-78aa-4acd-92f3-5adf43ca85ad-run-httpd\") pod \"ceilometer-0\" (UID: \"b8e62ba8-78aa-4acd-92f3-5adf43ca85ad\") " pod="openstack/ceilometer-0" Mar 13 14:28:39 crc kubenswrapper[4907]: I0313 14:28:39.669493 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b8e62ba8-78aa-4acd-92f3-5adf43ca85ad-config-data\") pod \"ceilometer-0\" (UID: \"b8e62ba8-78aa-4acd-92f3-5adf43ca85ad\") " pod="openstack/ceilometer-0" Mar 13 14:28:39 crc kubenswrapper[4907]: I0313 14:28:39.670164 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b8e62ba8-78aa-4acd-92f3-5adf43ca85ad-scripts\") pod \"ceilometer-0\" (UID: \"b8e62ba8-78aa-4acd-92f3-5adf43ca85ad\") " pod="openstack/ceilometer-0" Mar 13 14:28:39 crc kubenswrapper[4907]: I0313 14:28:39.670898 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8e62ba8-78aa-4acd-92f3-5adf43ca85ad-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b8e62ba8-78aa-4acd-92f3-5adf43ca85ad\") " pod="openstack/ceilometer-0" Mar 13 14:28:39 crc kubenswrapper[4907]: I0313 14:28:39.678627 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b8e62ba8-78aa-4acd-92f3-5adf43ca85ad-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b8e62ba8-78aa-4acd-92f3-5adf43ca85ad\") " pod="openstack/ceilometer-0" Mar 13 14:28:39 crc kubenswrapper[4907]: I0313 14:28:39.681446 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hdhb7\" (UniqueName: \"kubernetes.io/projected/b8e62ba8-78aa-4acd-92f3-5adf43ca85ad-kube-api-access-hdhb7\") pod \"ceilometer-0\" (UID: \"b8e62ba8-78aa-4acd-92f3-5adf43ca85ad\") " pod="openstack/ceilometer-0" Mar 13 14:28:39 crc kubenswrapper[4907]: I0313 14:28:39.799044 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="45d14b2b-1d0e-4e15-b837-0d2fb6aed785" path="/var/lib/kubelet/pods/45d14b2b-1d0e-4e15-b837-0d2fb6aed785/volumes" Mar 13 14:28:39 crc kubenswrapper[4907]: I0313 14:28:39.799651 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d6be28d7-de73-47ad-9b71-f7333bae7487" path="/var/lib/kubelet/pods/d6be28d7-de73-47ad-9b71-f7333bae7487/volumes" Mar 13 14:28:39 crc kubenswrapper[4907]: I0313 14:28:39.843643 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 14:28:40 crc kubenswrapper[4907]: W0313 14:28:40.286160 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb8e62ba8_78aa_4acd_92f3_5adf43ca85ad.slice/crio-a83f85b89f55541dd31712e7fe8b295e79d7ff0a8e724ec9e202e87b0b0f97d7 WatchSource:0}: Error finding container a83f85b89f55541dd31712e7fe8b295e79d7ff0a8e724ec9e202e87b0b0f97d7: Status 404 returned error can't find the container with id a83f85b89f55541dd31712e7fe8b295e79d7ff0a8e724ec9e202e87b0b0f97d7 Mar 13 14:28:40 crc kubenswrapper[4907]: I0313 14:28:40.292917 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 13 14:28:40 crc kubenswrapper[4907]: I0313 14:28:40.445451 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b8e62ba8-78aa-4acd-92f3-5adf43ca85ad","Type":"ContainerStarted","Data":"a83f85b89f55541dd31712e7fe8b295e79d7ff0a8e724ec9e202e87b0b0f97d7"} Mar 13 14:28:41 crc kubenswrapper[4907]: I0313 14:28:41.454808 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b8e62ba8-78aa-4acd-92f3-5adf43ca85ad","Type":"ContainerStarted","Data":"59220582376a55be5a4d3d891ace1fbaa90206be7f65a48006e97d333f02f723"} Mar 13 14:28:42 crc kubenswrapper[4907]: I0313 14:28:42.467704 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b8e62ba8-78aa-4acd-92f3-5adf43ca85ad","Type":"ContainerStarted","Data":"f069ee7a6257ccd2ec10dd948c5d3a36fd6dadfff2902dc16abe2e27eff98763"} Mar 13 14:28:43 crc kubenswrapper[4907]: I0313 14:28:43.477392 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b8e62ba8-78aa-4acd-92f3-5adf43ca85ad","Type":"ContainerStarted","Data":"4b6c6996dbdabf0608354c555365a022e97acac7a151f533aa94e6c52bcba1bf"} Mar 13 14:28:45 crc kubenswrapper[4907]: I0313 14:28:45.495292 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b8e62ba8-78aa-4acd-92f3-5adf43ca85ad","Type":"ContainerStarted","Data":"6ece6a9813874bfabeea08b5871d5ce4883ed931c7898fc7cf4565120db9c082"} Mar 13 14:28:45 crc kubenswrapper[4907]: I0313 14:28:45.496966 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Mar 13 14:28:45 crc kubenswrapper[4907]: I0313 14:28:45.522347 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.485405578 podStartE2EDuration="6.522328702s" podCreationTimestamp="2026-03-13 14:28:39 +0000 UTC" firstStartedPulling="2026-03-13 14:28:40.289101731 +0000 UTC m=+1419.188889420" lastFinishedPulling="2026-03-13 14:28:44.326024855 +0000 UTC m=+1423.225812544" observedRunningTime="2026-03-13 14:28:45.521364465 +0000 UTC m=+1424.421152154" watchObservedRunningTime="2026-03-13 14:28:45.522328702 +0000 UTC m=+1424.422116391" Mar 13 14:28:47 crc kubenswrapper[4907]: I0313 14:28:47.207609 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 13 14:28:48 crc kubenswrapper[4907]: I0313 14:28:48.041153 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 14:28:48 crc kubenswrapper[4907]: I0313 14:28:48.041212 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 14:28:48 crc kubenswrapper[4907]: I0313 14:28:48.041263 4907 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" Mar 13 14:28:48 crc kubenswrapper[4907]: I0313 14:28:48.042101 4907 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1f79a88cd4431d2b6772ef1dc1fb18ce20d965a3a936a7157e03bc28ca239c57"} pod="openshift-machine-config-operator/machine-config-daemon-hm56j" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 14:28:48 crc kubenswrapper[4907]: I0313 14:28:48.042164 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" containerID="cri-o://1f79a88cd4431d2b6772ef1dc1fb18ce20d965a3a936a7157e03bc28ca239c57" gracePeriod=600 Mar 13 14:28:48 crc kubenswrapper[4907]: I0313 14:28:48.526004 4907 generic.go:334] "Generic (PLEG): container finished" podID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerID="1f79a88cd4431d2b6772ef1dc1fb18ce20d965a3a936a7157e03bc28ca239c57" exitCode=0 Mar 13 14:28:48 crc kubenswrapper[4907]: I0313 14:28:48.526705 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b8e62ba8-78aa-4acd-92f3-5adf43ca85ad" containerName="ceilometer-central-agent" containerID="cri-o://59220582376a55be5a4d3d891ace1fbaa90206be7f65a48006e97d333f02f723" gracePeriod=30 Mar 13 14:28:48 crc kubenswrapper[4907]: I0313 14:28:48.526069 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerDied","Data":"1f79a88cd4431d2b6772ef1dc1fb18ce20d965a3a936a7157e03bc28ca239c57"} Mar 13 14:28:48 crc kubenswrapper[4907]: I0313 14:28:48.526807 4907 scope.go:117] "RemoveContainer" containerID="ac423129eeb1fa8eb8df325acb5ccb91e119992cd551dd9d8c11d304cb95b736" Mar 13 14:28:48 crc kubenswrapper[4907]: I0313 14:28:48.527199 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b8e62ba8-78aa-4acd-92f3-5adf43ca85ad" containerName="proxy-httpd" containerID="cri-o://6ece6a9813874bfabeea08b5871d5ce4883ed931c7898fc7cf4565120db9c082" gracePeriod=30 Mar 13 14:28:48 crc kubenswrapper[4907]: I0313 14:28:48.527240 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b8e62ba8-78aa-4acd-92f3-5adf43ca85ad" containerName="sg-core" containerID="cri-o://4b6c6996dbdabf0608354c555365a022e97acac7a151f533aa94e6c52bcba1bf" gracePeriod=30 Mar 13 14:28:48 crc kubenswrapper[4907]: I0313 14:28:48.527272 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b8e62ba8-78aa-4acd-92f3-5adf43ca85ad" containerName="ceilometer-notification-agent" containerID="cri-o://f069ee7a6257ccd2ec10dd948c5d3a36fd6dadfff2902dc16abe2e27eff98763" gracePeriod=30 Mar 13 14:28:49 crc kubenswrapper[4907]: I0313 14:28:49.543246 4907 generic.go:334] "Generic (PLEG): container finished" podID="b8e62ba8-78aa-4acd-92f3-5adf43ca85ad" containerID="6ece6a9813874bfabeea08b5871d5ce4883ed931c7898fc7cf4565120db9c082" exitCode=0 Mar 13 14:28:49 crc kubenswrapper[4907]: I0313 14:28:49.543767 4907 generic.go:334] "Generic (PLEG): container finished" podID="b8e62ba8-78aa-4acd-92f3-5adf43ca85ad" containerID="4b6c6996dbdabf0608354c555365a022e97acac7a151f533aa94e6c52bcba1bf" exitCode=2 Mar 13 14:28:49 crc kubenswrapper[4907]: I0313 14:28:49.543449 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b8e62ba8-78aa-4acd-92f3-5adf43ca85ad","Type":"ContainerDied","Data":"6ece6a9813874bfabeea08b5871d5ce4883ed931c7898fc7cf4565120db9c082"} Mar 13 14:28:49 crc kubenswrapper[4907]: I0313 14:28:49.543839 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b8e62ba8-78aa-4acd-92f3-5adf43ca85ad","Type":"ContainerDied","Data":"4b6c6996dbdabf0608354c555365a022e97acac7a151f533aa94e6c52bcba1bf"} Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.258234 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.317782 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8e62ba8-78aa-4acd-92f3-5adf43ca85ad-combined-ca-bundle\") pod \"b8e62ba8-78aa-4acd-92f3-5adf43ca85ad\" (UID: \"b8e62ba8-78aa-4acd-92f3-5adf43ca85ad\") " Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.318026 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b8e62ba8-78aa-4acd-92f3-5adf43ca85ad-sg-core-conf-yaml\") pod \"b8e62ba8-78aa-4acd-92f3-5adf43ca85ad\" (UID: \"b8e62ba8-78aa-4acd-92f3-5adf43ca85ad\") " Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.318082 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hdhb7\" (UniqueName: \"kubernetes.io/projected/b8e62ba8-78aa-4acd-92f3-5adf43ca85ad-kube-api-access-hdhb7\") pod \"b8e62ba8-78aa-4acd-92f3-5adf43ca85ad\" (UID: \"b8e62ba8-78aa-4acd-92f3-5adf43ca85ad\") " Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.318117 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b8e62ba8-78aa-4acd-92f3-5adf43ca85ad-config-data\") pod \"b8e62ba8-78aa-4acd-92f3-5adf43ca85ad\" (UID: \"b8e62ba8-78aa-4acd-92f3-5adf43ca85ad\") " Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.318171 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b8e62ba8-78aa-4acd-92f3-5adf43ca85ad-log-httpd\") pod \"b8e62ba8-78aa-4acd-92f3-5adf43ca85ad\" (UID: \"b8e62ba8-78aa-4acd-92f3-5adf43ca85ad\") " Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.318365 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b8e62ba8-78aa-4acd-92f3-5adf43ca85ad-run-httpd\") pod \"b8e62ba8-78aa-4acd-92f3-5adf43ca85ad\" (UID: \"b8e62ba8-78aa-4acd-92f3-5adf43ca85ad\") " Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.318405 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b8e62ba8-78aa-4acd-92f3-5adf43ca85ad-scripts\") pod \"b8e62ba8-78aa-4acd-92f3-5adf43ca85ad\" (UID: \"b8e62ba8-78aa-4acd-92f3-5adf43ca85ad\") " Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.320926 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b8e62ba8-78aa-4acd-92f3-5adf43ca85ad-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "b8e62ba8-78aa-4acd-92f3-5adf43ca85ad" (UID: "b8e62ba8-78aa-4acd-92f3-5adf43ca85ad"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.321262 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b8e62ba8-78aa-4acd-92f3-5adf43ca85ad-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "b8e62ba8-78aa-4acd-92f3-5adf43ca85ad" (UID: "b8e62ba8-78aa-4acd-92f3-5adf43ca85ad"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.326262 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b8e62ba8-78aa-4acd-92f3-5adf43ca85ad-kube-api-access-hdhb7" (OuterVolumeSpecName: "kube-api-access-hdhb7") pod "b8e62ba8-78aa-4acd-92f3-5adf43ca85ad" (UID: "b8e62ba8-78aa-4acd-92f3-5adf43ca85ad"). InnerVolumeSpecName "kube-api-access-hdhb7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.326846 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8e62ba8-78aa-4acd-92f3-5adf43ca85ad-scripts" (OuterVolumeSpecName: "scripts") pod "b8e62ba8-78aa-4acd-92f3-5adf43ca85ad" (UID: "b8e62ba8-78aa-4acd-92f3-5adf43ca85ad"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.353553 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8e62ba8-78aa-4acd-92f3-5adf43ca85ad-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "b8e62ba8-78aa-4acd-92f3-5adf43ca85ad" (UID: "b8e62ba8-78aa-4acd-92f3-5adf43ca85ad"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.418383 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8e62ba8-78aa-4acd-92f3-5adf43ca85ad-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b8e62ba8-78aa-4acd-92f3-5adf43ca85ad" (UID: "b8e62ba8-78aa-4acd-92f3-5adf43ca85ad"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.421523 4907 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b8e62ba8-78aa-4acd-92f3-5adf43ca85ad-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.421577 4907 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b8e62ba8-78aa-4acd-92f3-5adf43ca85ad-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.421590 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8e62ba8-78aa-4acd-92f3-5adf43ca85ad-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.421604 4907 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b8e62ba8-78aa-4acd-92f3-5adf43ca85ad-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.421615 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hdhb7\" (UniqueName: \"kubernetes.io/projected/b8e62ba8-78aa-4acd-92f3-5adf43ca85ad-kube-api-access-hdhb7\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.421624 4907 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b8e62ba8-78aa-4acd-92f3-5adf43ca85ad-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.470146 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8e62ba8-78aa-4acd-92f3-5adf43ca85ad-config-data" (OuterVolumeSpecName: "config-data") pod "b8e62ba8-78aa-4acd-92f3-5adf43ca85ad" (UID: "b8e62ba8-78aa-4acd-92f3-5adf43ca85ad"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.523670 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b8e62ba8-78aa-4acd-92f3-5adf43ca85ad-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.553852 4907 generic.go:334] "Generic (PLEG): container finished" podID="b8e62ba8-78aa-4acd-92f3-5adf43ca85ad" containerID="f069ee7a6257ccd2ec10dd948c5d3a36fd6dadfff2902dc16abe2e27eff98763" exitCode=0 Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.553892 4907 generic.go:334] "Generic (PLEG): container finished" podID="b8e62ba8-78aa-4acd-92f3-5adf43ca85ad" containerID="59220582376a55be5a4d3d891ace1fbaa90206be7f65a48006e97d333f02f723" exitCode=0 Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.553948 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.553960 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b8e62ba8-78aa-4acd-92f3-5adf43ca85ad","Type":"ContainerDied","Data":"f069ee7a6257ccd2ec10dd948c5d3a36fd6dadfff2902dc16abe2e27eff98763"} Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.553990 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b8e62ba8-78aa-4acd-92f3-5adf43ca85ad","Type":"ContainerDied","Data":"59220582376a55be5a4d3d891ace1fbaa90206be7f65a48006e97d333f02f723"} Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.554002 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b8e62ba8-78aa-4acd-92f3-5adf43ca85ad","Type":"ContainerDied","Data":"a83f85b89f55541dd31712e7fe8b295e79d7ff0a8e724ec9e202e87b0b0f97d7"} Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.554018 4907 scope.go:117] "RemoveContainer" containerID="6ece6a9813874bfabeea08b5871d5ce4883ed931c7898fc7cf4565120db9c082" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.558169 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerStarted","Data":"a09a3bec467b7e275495edb5827150af756bcc256911e7e06cbae5765922b5e2"} Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.587115 4907 scope.go:117] "RemoveContainer" containerID="4b6c6996dbdabf0608354c555365a022e97acac7a151f533aa94e6c52bcba1bf" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.612303 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.620390 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.620686 4907 scope.go:117] "RemoveContainer" containerID="f069ee7a6257ccd2ec10dd948c5d3a36fd6dadfff2902dc16abe2e27eff98763" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.637008 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 13 14:28:50 crc kubenswrapper[4907]: E0313 14:28:50.637687 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8e62ba8-78aa-4acd-92f3-5adf43ca85ad" containerName="ceilometer-central-agent" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.637706 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8e62ba8-78aa-4acd-92f3-5adf43ca85ad" containerName="ceilometer-central-agent" Mar 13 14:28:50 crc kubenswrapper[4907]: E0313 14:28:50.637720 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8e62ba8-78aa-4acd-92f3-5adf43ca85ad" containerName="sg-core" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.637726 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8e62ba8-78aa-4acd-92f3-5adf43ca85ad" containerName="sg-core" Mar 13 14:28:50 crc kubenswrapper[4907]: E0313 14:28:50.637745 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8e62ba8-78aa-4acd-92f3-5adf43ca85ad" containerName="ceilometer-notification-agent" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.637752 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8e62ba8-78aa-4acd-92f3-5adf43ca85ad" containerName="ceilometer-notification-agent" Mar 13 14:28:50 crc kubenswrapper[4907]: E0313 14:28:50.637765 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8e62ba8-78aa-4acd-92f3-5adf43ca85ad" containerName="proxy-httpd" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.637772 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8e62ba8-78aa-4acd-92f3-5adf43ca85ad" containerName="proxy-httpd" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.637958 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="b8e62ba8-78aa-4acd-92f3-5adf43ca85ad" containerName="proxy-httpd" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.637976 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="b8e62ba8-78aa-4acd-92f3-5adf43ca85ad" containerName="ceilometer-notification-agent" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.637988 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="b8e62ba8-78aa-4acd-92f3-5adf43ca85ad" containerName="sg-core" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.638007 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="b8e62ba8-78aa-4acd-92f3-5adf43ca85ad" containerName="ceilometer-central-agent" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.639745 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.644451 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.644662 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.655681 4907 scope.go:117] "RemoveContainer" containerID="59220582376a55be5a4d3d891ace1fbaa90206be7f65a48006e97d333f02f723" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.655698 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.691439 4907 scope.go:117] "RemoveContainer" containerID="6ece6a9813874bfabeea08b5871d5ce4883ed931c7898fc7cf4565120db9c082" Mar 13 14:28:50 crc kubenswrapper[4907]: E0313 14:28:50.691764 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6ece6a9813874bfabeea08b5871d5ce4883ed931c7898fc7cf4565120db9c082\": container with ID starting with 6ece6a9813874bfabeea08b5871d5ce4883ed931c7898fc7cf4565120db9c082 not found: ID does not exist" containerID="6ece6a9813874bfabeea08b5871d5ce4883ed931c7898fc7cf4565120db9c082" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.691794 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6ece6a9813874bfabeea08b5871d5ce4883ed931c7898fc7cf4565120db9c082"} err="failed to get container status \"6ece6a9813874bfabeea08b5871d5ce4883ed931c7898fc7cf4565120db9c082\": rpc error: code = NotFound desc = could not find container \"6ece6a9813874bfabeea08b5871d5ce4883ed931c7898fc7cf4565120db9c082\": container with ID starting with 6ece6a9813874bfabeea08b5871d5ce4883ed931c7898fc7cf4565120db9c082 not found: ID does not exist" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.691815 4907 scope.go:117] "RemoveContainer" containerID="4b6c6996dbdabf0608354c555365a022e97acac7a151f533aa94e6c52bcba1bf" Mar 13 14:28:50 crc kubenswrapper[4907]: E0313 14:28:50.692289 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4b6c6996dbdabf0608354c555365a022e97acac7a151f533aa94e6c52bcba1bf\": container with ID starting with 4b6c6996dbdabf0608354c555365a022e97acac7a151f533aa94e6c52bcba1bf not found: ID does not exist" containerID="4b6c6996dbdabf0608354c555365a022e97acac7a151f533aa94e6c52bcba1bf" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.692309 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4b6c6996dbdabf0608354c555365a022e97acac7a151f533aa94e6c52bcba1bf"} err="failed to get container status \"4b6c6996dbdabf0608354c555365a022e97acac7a151f533aa94e6c52bcba1bf\": rpc error: code = NotFound desc = could not find container \"4b6c6996dbdabf0608354c555365a022e97acac7a151f533aa94e6c52bcba1bf\": container with ID starting with 4b6c6996dbdabf0608354c555365a022e97acac7a151f533aa94e6c52bcba1bf not found: ID does not exist" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.692339 4907 scope.go:117] "RemoveContainer" containerID="f069ee7a6257ccd2ec10dd948c5d3a36fd6dadfff2902dc16abe2e27eff98763" Mar 13 14:28:50 crc kubenswrapper[4907]: E0313 14:28:50.692691 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f069ee7a6257ccd2ec10dd948c5d3a36fd6dadfff2902dc16abe2e27eff98763\": container with ID starting with f069ee7a6257ccd2ec10dd948c5d3a36fd6dadfff2902dc16abe2e27eff98763 not found: ID does not exist" containerID="f069ee7a6257ccd2ec10dd948c5d3a36fd6dadfff2902dc16abe2e27eff98763" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.692713 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f069ee7a6257ccd2ec10dd948c5d3a36fd6dadfff2902dc16abe2e27eff98763"} err="failed to get container status \"f069ee7a6257ccd2ec10dd948c5d3a36fd6dadfff2902dc16abe2e27eff98763\": rpc error: code = NotFound desc = could not find container \"f069ee7a6257ccd2ec10dd948c5d3a36fd6dadfff2902dc16abe2e27eff98763\": container with ID starting with f069ee7a6257ccd2ec10dd948c5d3a36fd6dadfff2902dc16abe2e27eff98763 not found: ID does not exist" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.692736 4907 scope.go:117] "RemoveContainer" containerID="59220582376a55be5a4d3d891ace1fbaa90206be7f65a48006e97d333f02f723" Mar 13 14:28:50 crc kubenswrapper[4907]: E0313 14:28:50.693177 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"59220582376a55be5a4d3d891ace1fbaa90206be7f65a48006e97d333f02f723\": container with ID starting with 59220582376a55be5a4d3d891ace1fbaa90206be7f65a48006e97d333f02f723 not found: ID does not exist" containerID="59220582376a55be5a4d3d891ace1fbaa90206be7f65a48006e97d333f02f723" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.693198 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"59220582376a55be5a4d3d891ace1fbaa90206be7f65a48006e97d333f02f723"} err="failed to get container status \"59220582376a55be5a4d3d891ace1fbaa90206be7f65a48006e97d333f02f723\": rpc error: code = NotFound desc = could not find container \"59220582376a55be5a4d3d891ace1fbaa90206be7f65a48006e97d333f02f723\": container with ID starting with 59220582376a55be5a4d3d891ace1fbaa90206be7f65a48006e97d333f02f723 not found: ID does not exist" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.693209 4907 scope.go:117] "RemoveContainer" containerID="6ece6a9813874bfabeea08b5871d5ce4883ed931c7898fc7cf4565120db9c082" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.693503 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6ece6a9813874bfabeea08b5871d5ce4883ed931c7898fc7cf4565120db9c082"} err="failed to get container status \"6ece6a9813874bfabeea08b5871d5ce4883ed931c7898fc7cf4565120db9c082\": rpc error: code = NotFound desc = could not find container \"6ece6a9813874bfabeea08b5871d5ce4883ed931c7898fc7cf4565120db9c082\": container with ID starting with 6ece6a9813874bfabeea08b5871d5ce4883ed931c7898fc7cf4565120db9c082 not found: ID does not exist" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.693518 4907 scope.go:117] "RemoveContainer" containerID="4b6c6996dbdabf0608354c555365a022e97acac7a151f533aa94e6c52bcba1bf" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.693745 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4b6c6996dbdabf0608354c555365a022e97acac7a151f533aa94e6c52bcba1bf"} err="failed to get container status \"4b6c6996dbdabf0608354c555365a022e97acac7a151f533aa94e6c52bcba1bf\": rpc error: code = NotFound desc = could not find container \"4b6c6996dbdabf0608354c555365a022e97acac7a151f533aa94e6c52bcba1bf\": container with ID starting with 4b6c6996dbdabf0608354c555365a022e97acac7a151f533aa94e6c52bcba1bf not found: ID does not exist" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.693765 4907 scope.go:117] "RemoveContainer" containerID="f069ee7a6257ccd2ec10dd948c5d3a36fd6dadfff2902dc16abe2e27eff98763" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.697117 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f069ee7a6257ccd2ec10dd948c5d3a36fd6dadfff2902dc16abe2e27eff98763"} err="failed to get container status \"f069ee7a6257ccd2ec10dd948c5d3a36fd6dadfff2902dc16abe2e27eff98763\": rpc error: code = NotFound desc = could not find container \"f069ee7a6257ccd2ec10dd948c5d3a36fd6dadfff2902dc16abe2e27eff98763\": container with ID starting with f069ee7a6257ccd2ec10dd948c5d3a36fd6dadfff2902dc16abe2e27eff98763 not found: ID does not exist" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.697139 4907 scope.go:117] "RemoveContainer" containerID="59220582376a55be5a4d3d891ace1fbaa90206be7f65a48006e97d333f02f723" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.697675 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"59220582376a55be5a4d3d891ace1fbaa90206be7f65a48006e97d333f02f723"} err="failed to get container status \"59220582376a55be5a4d3d891ace1fbaa90206be7f65a48006e97d333f02f723\": rpc error: code = NotFound desc = could not find container \"59220582376a55be5a4d3d891ace1fbaa90206be7f65a48006e97d333f02f723\": container with ID starting with 59220582376a55be5a4d3d891ace1fbaa90206be7f65a48006e97d333f02f723 not found: ID does not exist" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.726734 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a85017ea-b984-471a-a9a8-22c02d94448b-config-data\") pod \"ceilometer-0\" (UID: \"a85017ea-b984-471a-a9a8-22c02d94448b\") " pod="openstack/ceilometer-0" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.726773 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a85017ea-b984-471a-a9a8-22c02d94448b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a85017ea-b984-471a-a9a8-22c02d94448b\") " pod="openstack/ceilometer-0" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.727020 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-462xb\" (UniqueName: \"kubernetes.io/projected/a85017ea-b984-471a-a9a8-22c02d94448b-kube-api-access-462xb\") pod \"ceilometer-0\" (UID: \"a85017ea-b984-471a-a9a8-22c02d94448b\") " pod="openstack/ceilometer-0" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.727099 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a85017ea-b984-471a-a9a8-22c02d94448b-scripts\") pod \"ceilometer-0\" (UID: \"a85017ea-b984-471a-a9a8-22c02d94448b\") " pod="openstack/ceilometer-0" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.727198 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a85017ea-b984-471a-a9a8-22c02d94448b-log-httpd\") pod \"ceilometer-0\" (UID: \"a85017ea-b984-471a-a9a8-22c02d94448b\") " pod="openstack/ceilometer-0" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.727248 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a85017ea-b984-471a-a9a8-22c02d94448b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a85017ea-b984-471a-a9a8-22c02d94448b\") " pod="openstack/ceilometer-0" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.727319 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a85017ea-b984-471a-a9a8-22c02d94448b-run-httpd\") pod \"ceilometer-0\" (UID: \"a85017ea-b984-471a-a9a8-22c02d94448b\") " pod="openstack/ceilometer-0" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.828970 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a85017ea-b984-471a-a9a8-22c02d94448b-log-httpd\") pod \"ceilometer-0\" (UID: \"a85017ea-b984-471a-a9a8-22c02d94448b\") " pod="openstack/ceilometer-0" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.829072 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a85017ea-b984-471a-a9a8-22c02d94448b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a85017ea-b984-471a-a9a8-22c02d94448b\") " pod="openstack/ceilometer-0" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.829241 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a85017ea-b984-471a-a9a8-22c02d94448b-run-httpd\") pod \"ceilometer-0\" (UID: \"a85017ea-b984-471a-a9a8-22c02d94448b\") " pod="openstack/ceilometer-0" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.829672 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a85017ea-b984-471a-a9a8-22c02d94448b-log-httpd\") pod \"ceilometer-0\" (UID: \"a85017ea-b984-471a-a9a8-22c02d94448b\") " pod="openstack/ceilometer-0" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.829738 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a85017ea-b984-471a-a9a8-22c02d94448b-run-httpd\") pod \"ceilometer-0\" (UID: \"a85017ea-b984-471a-a9a8-22c02d94448b\") " pod="openstack/ceilometer-0" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.829952 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a85017ea-b984-471a-a9a8-22c02d94448b-config-data\") pod \"ceilometer-0\" (UID: \"a85017ea-b984-471a-a9a8-22c02d94448b\") " pod="openstack/ceilometer-0" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.829997 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a85017ea-b984-471a-a9a8-22c02d94448b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a85017ea-b984-471a-a9a8-22c02d94448b\") " pod="openstack/ceilometer-0" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.831192 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-462xb\" (UniqueName: \"kubernetes.io/projected/a85017ea-b984-471a-a9a8-22c02d94448b-kube-api-access-462xb\") pod \"ceilometer-0\" (UID: \"a85017ea-b984-471a-a9a8-22c02d94448b\") " pod="openstack/ceilometer-0" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.831309 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a85017ea-b984-471a-a9a8-22c02d94448b-scripts\") pod \"ceilometer-0\" (UID: \"a85017ea-b984-471a-a9a8-22c02d94448b\") " pod="openstack/ceilometer-0" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.836984 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a85017ea-b984-471a-a9a8-22c02d94448b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a85017ea-b984-471a-a9a8-22c02d94448b\") " pod="openstack/ceilometer-0" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.838028 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a85017ea-b984-471a-a9a8-22c02d94448b-scripts\") pod \"ceilometer-0\" (UID: \"a85017ea-b984-471a-a9a8-22c02d94448b\") " pod="openstack/ceilometer-0" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.841501 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a85017ea-b984-471a-a9a8-22c02d94448b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a85017ea-b984-471a-a9a8-22c02d94448b\") " pod="openstack/ceilometer-0" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.847694 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a85017ea-b984-471a-a9a8-22c02d94448b-config-data\") pod \"ceilometer-0\" (UID: \"a85017ea-b984-471a-a9a8-22c02d94448b\") " pod="openstack/ceilometer-0" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.850076 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-462xb\" (UniqueName: \"kubernetes.io/projected/a85017ea-b984-471a-a9a8-22c02d94448b-kube-api-access-462xb\") pod \"ceilometer-0\" (UID: \"a85017ea-b984-471a-a9a8-22c02d94448b\") " pod="openstack/ceilometer-0" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.958424 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 14:28:51 crc kubenswrapper[4907]: I0313 14:28:51.395413 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 13 14:28:51 crc kubenswrapper[4907]: W0313 14:28:51.400484 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda85017ea_b984_471a_a9a8_22c02d94448b.slice/crio-5a88cd9ef6cae1bcbb8c65cf464b690c68ec2dfeb0b4737e783c308ca4a8ffd6 WatchSource:0}: Error finding container 5a88cd9ef6cae1bcbb8c65cf464b690c68ec2dfeb0b4737e783c308ca4a8ffd6: Status 404 returned error can't find the container with id 5a88cd9ef6cae1bcbb8c65cf464b690c68ec2dfeb0b4737e783c308ca4a8ffd6 Mar 13 14:28:51 crc kubenswrapper[4907]: I0313 14:28:51.570702 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a85017ea-b984-471a-a9a8-22c02d94448b","Type":"ContainerStarted","Data":"5a88cd9ef6cae1bcbb8c65cf464b690c68ec2dfeb0b4737e783c308ca4a8ffd6"} Mar 13 14:28:51 crc kubenswrapper[4907]: I0313 14:28:51.573700 4907 generic.go:334] "Generic (PLEG): container finished" podID="4d7ea7a2-ab25-4cb4-8749-10bc129cbd22" containerID="d935db5c32cf1af5583418c37d51248a6d17b0514af1c3fa70469ca109747682" exitCode=0 Mar 13 14:28:51 crc kubenswrapper[4907]: I0313 14:28:51.573796 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-z92h5" event={"ID":"4d7ea7a2-ab25-4cb4-8749-10bc129cbd22","Type":"ContainerDied","Data":"d935db5c32cf1af5583418c37d51248a6d17b0514af1c3fa70469ca109747682"} Mar 13 14:28:51 crc kubenswrapper[4907]: I0313 14:28:51.793074 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b8e62ba8-78aa-4acd-92f3-5adf43ca85ad" path="/var/lib/kubelet/pods/b8e62ba8-78aa-4acd-92f3-5adf43ca85ad/volumes" Mar 13 14:28:52 crc kubenswrapper[4907]: I0313 14:28:52.586851 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a85017ea-b984-471a-a9a8-22c02d94448b","Type":"ContainerStarted","Data":"62988d506f50133c1068883469efb51974597559ba58796a2aae59bf02bebb3a"} Mar 13 14:28:52 crc kubenswrapper[4907]: I0313 14:28:52.964939 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-z92h5" Mar 13 14:28:53 crc kubenswrapper[4907]: I0313 14:28:53.125840 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d7ea7a2-ab25-4cb4-8749-10bc129cbd22-config-data\") pod \"4d7ea7a2-ab25-4cb4-8749-10bc129cbd22\" (UID: \"4d7ea7a2-ab25-4cb4-8749-10bc129cbd22\") " Mar 13 14:28:53 crc kubenswrapper[4907]: I0313 14:28:53.125913 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d7ea7a2-ab25-4cb4-8749-10bc129cbd22-combined-ca-bundle\") pod \"4d7ea7a2-ab25-4cb4-8749-10bc129cbd22\" (UID: \"4d7ea7a2-ab25-4cb4-8749-10bc129cbd22\") " Mar 13 14:28:53 crc kubenswrapper[4907]: I0313 14:28:53.126172 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dc5bv\" (UniqueName: \"kubernetes.io/projected/4d7ea7a2-ab25-4cb4-8749-10bc129cbd22-kube-api-access-dc5bv\") pod \"4d7ea7a2-ab25-4cb4-8749-10bc129cbd22\" (UID: \"4d7ea7a2-ab25-4cb4-8749-10bc129cbd22\") " Mar 13 14:28:53 crc kubenswrapper[4907]: I0313 14:28:53.126209 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4d7ea7a2-ab25-4cb4-8749-10bc129cbd22-scripts\") pod \"4d7ea7a2-ab25-4cb4-8749-10bc129cbd22\" (UID: \"4d7ea7a2-ab25-4cb4-8749-10bc129cbd22\") " Mar 13 14:28:53 crc kubenswrapper[4907]: I0313 14:28:53.143075 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d7ea7a2-ab25-4cb4-8749-10bc129cbd22-kube-api-access-dc5bv" (OuterVolumeSpecName: "kube-api-access-dc5bv") pod "4d7ea7a2-ab25-4cb4-8749-10bc129cbd22" (UID: "4d7ea7a2-ab25-4cb4-8749-10bc129cbd22"). InnerVolumeSpecName "kube-api-access-dc5bv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:28:53 crc kubenswrapper[4907]: I0313 14:28:53.147237 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d7ea7a2-ab25-4cb4-8749-10bc129cbd22-scripts" (OuterVolumeSpecName: "scripts") pod "4d7ea7a2-ab25-4cb4-8749-10bc129cbd22" (UID: "4d7ea7a2-ab25-4cb4-8749-10bc129cbd22"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:28:53 crc kubenswrapper[4907]: I0313 14:28:53.197281 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d7ea7a2-ab25-4cb4-8749-10bc129cbd22-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4d7ea7a2-ab25-4cb4-8749-10bc129cbd22" (UID: "4d7ea7a2-ab25-4cb4-8749-10bc129cbd22"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:28:53 crc kubenswrapper[4907]: I0313 14:28:53.210268 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d7ea7a2-ab25-4cb4-8749-10bc129cbd22-config-data" (OuterVolumeSpecName: "config-data") pod "4d7ea7a2-ab25-4cb4-8749-10bc129cbd22" (UID: "4d7ea7a2-ab25-4cb4-8749-10bc129cbd22"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:28:53 crc kubenswrapper[4907]: I0313 14:28:53.228234 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dc5bv\" (UniqueName: \"kubernetes.io/projected/4d7ea7a2-ab25-4cb4-8749-10bc129cbd22-kube-api-access-dc5bv\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:53 crc kubenswrapper[4907]: I0313 14:28:53.228270 4907 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4d7ea7a2-ab25-4cb4-8749-10bc129cbd22-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:53 crc kubenswrapper[4907]: I0313 14:28:53.228279 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d7ea7a2-ab25-4cb4-8749-10bc129cbd22-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:53 crc kubenswrapper[4907]: I0313 14:28:53.228288 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d7ea7a2-ab25-4cb4-8749-10bc129cbd22-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:53 crc kubenswrapper[4907]: I0313 14:28:53.598034 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a85017ea-b984-471a-a9a8-22c02d94448b","Type":"ContainerStarted","Data":"f629b415ac0d40fdc7a871c081a5a0ceecff9578a74f62763cbbaae7d50396c7"} Mar 13 14:28:53 crc kubenswrapper[4907]: I0313 14:28:53.598396 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a85017ea-b984-471a-a9a8-22c02d94448b","Type":"ContainerStarted","Data":"3b55543c8226645f0d4a440e41c724886657fe5c5d9924185cd3344bad9ddc01"} Mar 13 14:28:53 crc kubenswrapper[4907]: I0313 14:28:53.600035 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-z92h5" event={"ID":"4d7ea7a2-ab25-4cb4-8749-10bc129cbd22","Type":"ContainerDied","Data":"7d0624634945e5c5e075e91b43232f99c08cdffbe02fc1416ac047abb7f58463"} Mar 13 14:28:53 crc kubenswrapper[4907]: I0313 14:28:53.600059 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7d0624634945e5c5e075e91b43232f99c08cdffbe02fc1416ac047abb7f58463" Mar 13 14:28:53 crc kubenswrapper[4907]: I0313 14:28:53.600109 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-z92h5" Mar 13 14:28:53 crc kubenswrapper[4907]: I0313 14:28:53.716468 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Mar 13 14:28:53 crc kubenswrapper[4907]: E0313 14:28:53.717237 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d7ea7a2-ab25-4cb4-8749-10bc129cbd22" containerName="nova-cell0-conductor-db-sync" Mar 13 14:28:53 crc kubenswrapper[4907]: I0313 14:28:53.717282 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d7ea7a2-ab25-4cb4-8749-10bc129cbd22" containerName="nova-cell0-conductor-db-sync" Mar 13 14:28:53 crc kubenswrapper[4907]: I0313 14:28:53.717670 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d7ea7a2-ab25-4cb4-8749-10bc129cbd22" containerName="nova-cell0-conductor-db-sync" Mar 13 14:28:53 crc kubenswrapper[4907]: I0313 14:28:53.718741 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Mar 13 14:28:53 crc kubenswrapper[4907]: I0313 14:28:53.721420 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-r4dpz" Mar 13 14:28:53 crc kubenswrapper[4907]: I0313 14:28:53.721697 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Mar 13 14:28:53 crc kubenswrapper[4907]: I0313 14:28:53.735147 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Mar 13 14:28:53 crc kubenswrapper[4907]: I0313 14:28:53.842134 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cdxcj\" (UniqueName: \"kubernetes.io/projected/364a702e-e6c6-42bc-8cee-4ba19ea2f1ec-kube-api-access-cdxcj\") pod \"nova-cell0-conductor-0\" (UID: \"364a702e-e6c6-42bc-8cee-4ba19ea2f1ec\") " pod="openstack/nova-cell0-conductor-0" Mar 13 14:28:53 crc kubenswrapper[4907]: I0313 14:28:53.842223 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/364a702e-e6c6-42bc-8cee-4ba19ea2f1ec-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"364a702e-e6c6-42bc-8cee-4ba19ea2f1ec\") " pod="openstack/nova-cell0-conductor-0" Mar 13 14:28:53 crc kubenswrapper[4907]: I0313 14:28:53.842315 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/364a702e-e6c6-42bc-8cee-4ba19ea2f1ec-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"364a702e-e6c6-42bc-8cee-4ba19ea2f1ec\") " pod="openstack/nova-cell0-conductor-0" Mar 13 14:28:53 crc kubenswrapper[4907]: I0313 14:28:53.943319 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/364a702e-e6c6-42bc-8cee-4ba19ea2f1ec-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"364a702e-e6c6-42bc-8cee-4ba19ea2f1ec\") " pod="openstack/nova-cell0-conductor-0" Mar 13 14:28:53 crc kubenswrapper[4907]: I0313 14:28:53.943398 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cdxcj\" (UniqueName: \"kubernetes.io/projected/364a702e-e6c6-42bc-8cee-4ba19ea2f1ec-kube-api-access-cdxcj\") pod \"nova-cell0-conductor-0\" (UID: \"364a702e-e6c6-42bc-8cee-4ba19ea2f1ec\") " pod="openstack/nova-cell0-conductor-0" Mar 13 14:28:53 crc kubenswrapper[4907]: I0313 14:28:53.943457 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/364a702e-e6c6-42bc-8cee-4ba19ea2f1ec-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"364a702e-e6c6-42bc-8cee-4ba19ea2f1ec\") " pod="openstack/nova-cell0-conductor-0" Mar 13 14:28:53 crc kubenswrapper[4907]: I0313 14:28:53.948211 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/364a702e-e6c6-42bc-8cee-4ba19ea2f1ec-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"364a702e-e6c6-42bc-8cee-4ba19ea2f1ec\") " pod="openstack/nova-cell0-conductor-0" Mar 13 14:28:53 crc kubenswrapper[4907]: I0313 14:28:53.948239 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/364a702e-e6c6-42bc-8cee-4ba19ea2f1ec-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"364a702e-e6c6-42bc-8cee-4ba19ea2f1ec\") " pod="openstack/nova-cell0-conductor-0" Mar 13 14:28:53 crc kubenswrapper[4907]: I0313 14:28:53.960509 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cdxcj\" (UniqueName: \"kubernetes.io/projected/364a702e-e6c6-42bc-8cee-4ba19ea2f1ec-kube-api-access-cdxcj\") pod \"nova-cell0-conductor-0\" (UID: \"364a702e-e6c6-42bc-8cee-4ba19ea2f1ec\") " pod="openstack/nova-cell0-conductor-0" Mar 13 14:28:54 crc kubenswrapper[4907]: I0313 14:28:54.051109 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Mar 13 14:28:54 crc kubenswrapper[4907]: I0313 14:28:54.529180 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Mar 13 14:28:54 crc kubenswrapper[4907]: I0313 14:28:54.611937 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"364a702e-e6c6-42bc-8cee-4ba19ea2f1ec","Type":"ContainerStarted","Data":"416b8c99953bb29e1da48539894292de008485b785183dda880949342468afcc"} Mar 13 14:28:55 crc kubenswrapper[4907]: I0313 14:28:55.624858 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"364a702e-e6c6-42bc-8cee-4ba19ea2f1ec","Type":"ContainerStarted","Data":"5ab7cf17ecb1804f362340163cfceb48e6f2c8900108799d01d1025c485c1e8f"} Mar 13 14:28:55 crc kubenswrapper[4907]: I0313 14:28:55.625396 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Mar 13 14:28:55 crc kubenswrapper[4907]: I0313 14:28:55.629514 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a85017ea-b984-471a-a9a8-22c02d94448b","Type":"ContainerStarted","Data":"c2a32ce2fcf985bbc86c96669646554e0cd290d1eaa4396d46e02db24cec19b9"} Mar 13 14:28:55 crc kubenswrapper[4907]: I0313 14:28:55.629776 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Mar 13 14:28:55 crc kubenswrapper[4907]: I0313 14:28:55.643970 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.6439525120000003 podStartE2EDuration="2.643952512s" podCreationTimestamp="2026-03-13 14:28:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:28:55.64095317 +0000 UTC m=+1434.540740859" watchObservedRunningTime="2026-03-13 14:28:55.643952512 +0000 UTC m=+1434.543740211" Mar 13 14:28:55 crc kubenswrapper[4907]: I0313 14:28:55.668932 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.14756681 podStartE2EDuration="5.668887839s" podCreationTimestamp="2026-03-13 14:28:50 +0000 UTC" firstStartedPulling="2026-03-13 14:28:51.403261757 +0000 UTC m=+1430.303049446" lastFinishedPulling="2026-03-13 14:28:54.924582786 +0000 UTC m=+1433.824370475" observedRunningTime="2026-03-13 14:28:55.664474679 +0000 UTC m=+1434.564262368" watchObservedRunningTime="2026-03-13 14:28:55.668887839 +0000 UTC m=+1434.568675528" Mar 13 14:28:59 crc kubenswrapper[4907]: I0313 14:28:59.080126 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Mar 13 14:28:59 crc kubenswrapper[4907]: I0313 14:28:59.543417 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-6g25t"] Mar 13 14:28:59 crc kubenswrapper[4907]: I0313 14:28:59.544690 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-6g25t" Mar 13 14:28:59 crc kubenswrapper[4907]: I0313 14:28:59.547063 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Mar 13 14:28:59 crc kubenswrapper[4907]: I0313 14:28:59.547650 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Mar 13 14:28:59 crc kubenswrapper[4907]: I0313 14:28:59.553242 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-6g25t"] Mar 13 14:28:59 crc kubenswrapper[4907]: I0313 14:28:59.656746 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jrsdn\" (UniqueName: \"kubernetes.io/projected/1b701a42-f649-4657-9250-49188b10d3b9-kube-api-access-jrsdn\") pod \"nova-cell0-cell-mapping-6g25t\" (UID: \"1b701a42-f649-4657-9250-49188b10d3b9\") " pod="openstack/nova-cell0-cell-mapping-6g25t" Mar 13 14:28:59 crc kubenswrapper[4907]: I0313 14:28:59.657123 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b701a42-f649-4657-9250-49188b10d3b9-config-data\") pod \"nova-cell0-cell-mapping-6g25t\" (UID: \"1b701a42-f649-4657-9250-49188b10d3b9\") " pod="openstack/nova-cell0-cell-mapping-6g25t" Mar 13 14:28:59 crc kubenswrapper[4907]: I0313 14:28:59.657152 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1b701a42-f649-4657-9250-49188b10d3b9-scripts\") pod \"nova-cell0-cell-mapping-6g25t\" (UID: \"1b701a42-f649-4657-9250-49188b10d3b9\") " pod="openstack/nova-cell0-cell-mapping-6g25t" Mar 13 14:28:59 crc kubenswrapper[4907]: I0313 14:28:59.657180 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b701a42-f649-4657-9250-49188b10d3b9-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-6g25t\" (UID: \"1b701a42-f649-4657-9250-49188b10d3b9\") " pod="openstack/nova-cell0-cell-mapping-6g25t" Mar 13 14:28:59 crc kubenswrapper[4907]: I0313 14:28:59.762098 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jrsdn\" (UniqueName: \"kubernetes.io/projected/1b701a42-f649-4657-9250-49188b10d3b9-kube-api-access-jrsdn\") pod \"nova-cell0-cell-mapping-6g25t\" (UID: \"1b701a42-f649-4657-9250-49188b10d3b9\") " pod="openstack/nova-cell0-cell-mapping-6g25t" Mar 13 14:28:59 crc kubenswrapper[4907]: I0313 14:28:59.762188 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b701a42-f649-4657-9250-49188b10d3b9-config-data\") pod \"nova-cell0-cell-mapping-6g25t\" (UID: \"1b701a42-f649-4657-9250-49188b10d3b9\") " pod="openstack/nova-cell0-cell-mapping-6g25t" Mar 13 14:28:59 crc kubenswrapper[4907]: I0313 14:28:59.762213 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1b701a42-f649-4657-9250-49188b10d3b9-scripts\") pod \"nova-cell0-cell-mapping-6g25t\" (UID: \"1b701a42-f649-4657-9250-49188b10d3b9\") " pod="openstack/nova-cell0-cell-mapping-6g25t" Mar 13 14:28:59 crc kubenswrapper[4907]: I0313 14:28:59.762241 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b701a42-f649-4657-9250-49188b10d3b9-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-6g25t\" (UID: \"1b701a42-f649-4657-9250-49188b10d3b9\") " pod="openstack/nova-cell0-cell-mapping-6g25t" Mar 13 14:28:59 crc kubenswrapper[4907]: I0313 14:28:59.770823 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b701a42-f649-4657-9250-49188b10d3b9-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-6g25t\" (UID: \"1b701a42-f649-4657-9250-49188b10d3b9\") " pod="openstack/nova-cell0-cell-mapping-6g25t" Mar 13 14:28:59 crc kubenswrapper[4907]: I0313 14:28:59.770899 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b701a42-f649-4657-9250-49188b10d3b9-config-data\") pod \"nova-cell0-cell-mapping-6g25t\" (UID: \"1b701a42-f649-4657-9250-49188b10d3b9\") " pod="openstack/nova-cell0-cell-mapping-6g25t" Mar 13 14:28:59 crc kubenswrapper[4907]: I0313 14:28:59.774667 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1b701a42-f649-4657-9250-49188b10d3b9-scripts\") pod \"nova-cell0-cell-mapping-6g25t\" (UID: \"1b701a42-f649-4657-9250-49188b10d3b9\") " pod="openstack/nova-cell0-cell-mapping-6g25t" Mar 13 14:28:59 crc kubenswrapper[4907]: I0313 14:28:59.779288 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 14:28:59 crc kubenswrapper[4907]: I0313 14:28:59.780384 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jrsdn\" (UniqueName: \"kubernetes.io/projected/1b701a42-f649-4657-9250-49188b10d3b9-kube-api-access-jrsdn\") pod \"nova-cell0-cell-mapping-6g25t\" (UID: \"1b701a42-f649-4657-9250-49188b10d3b9\") " pod="openstack/nova-cell0-cell-mapping-6g25t" Mar 13 14:28:59 crc kubenswrapper[4907]: I0313 14:28:59.780432 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 13 14:28:59 crc kubenswrapper[4907]: I0313 14:28:59.783181 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Mar 13 14:28:59 crc kubenswrapper[4907]: I0313 14:28:59.832869 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 14:28:59 crc kubenswrapper[4907]: I0313 14:28:59.864802 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-6g25t" Mar 13 14:28:59 crc kubenswrapper[4907]: I0313 14:28:59.894220 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 13 14:28:59 crc kubenswrapper[4907]: I0313 14:28:59.902181 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Mar 13 14:28:59 crc kubenswrapper[4907]: I0313 14:28:59.914741 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Mar 13 14:28:59 crc kubenswrapper[4907]: I0313 14:28:59.930180 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 13 14:28:59 crc kubenswrapper[4907]: I0313 14:28:59.954941 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Mar 13 14:28:59 crc kubenswrapper[4907]: I0313 14:28:59.956426 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 13 14:28:59 crc kubenswrapper[4907]: I0313 14:28:59.963395 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Mar 13 14:28:59 crc kubenswrapper[4907]: I0313 14:28:59.964809 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/636c92f5-234f-4de8-b37b-09be481708e1-config-data\") pod \"nova-scheduler-0\" (UID: \"636c92f5-234f-4de8-b37b-09be481708e1\") " pod="openstack/nova-scheduler-0" Mar 13 14:28:59 crc kubenswrapper[4907]: I0313 14:28:59.964912 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6c8mc\" (UniqueName: \"kubernetes.io/projected/636c92f5-234f-4de8-b37b-09be481708e1-kube-api-access-6c8mc\") pod \"nova-scheduler-0\" (UID: \"636c92f5-234f-4de8-b37b-09be481708e1\") " pod="openstack/nova-scheduler-0" Mar 13 14:28:59 crc kubenswrapper[4907]: I0313 14:28:59.965058 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/636c92f5-234f-4de8-b37b-09be481708e1-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"636c92f5-234f-4de8-b37b-09be481708e1\") " pod="openstack/nova-scheduler-0" Mar 13 14:28:59 crc kubenswrapper[4907]: I0313 14:28:59.984769 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.068704 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6c8mc\" (UniqueName: \"kubernetes.io/projected/636c92f5-234f-4de8-b37b-09be481708e1-kube-api-access-6c8mc\") pod \"nova-scheduler-0\" (UID: \"636c92f5-234f-4de8-b37b-09be481708e1\") " pod="openstack/nova-scheduler-0" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.069283 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a99d655e-f6db-4c83-a950-b935588b7df1-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"a99d655e-f6db-4c83-a950-b935588b7df1\") " pod="openstack/nova-metadata-0" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.069401 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-phktq\" (UniqueName: \"kubernetes.io/projected/1915a375-806d-43bb-8749-9524db071714-kube-api-access-phktq\") pod \"nova-cell1-novncproxy-0\" (UID: \"1915a375-806d-43bb-8749-9524db071714\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.069556 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a99d655e-f6db-4c83-a950-b935588b7df1-config-data\") pod \"nova-metadata-0\" (UID: \"a99d655e-f6db-4c83-a950-b935588b7df1\") " pod="openstack/nova-metadata-0" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.069669 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a99d655e-f6db-4c83-a950-b935588b7df1-logs\") pod \"nova-metadata-0\" (UID: \"a99d655e-f6db-4c83-a950-b935588b7df1\") " pod="openstack/nova-metadata-0" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.069756 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1915a375-806d-43bb-8749-9524db071714-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"1915a375-806d-43bb-8749-9524db071714\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.069837 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/636c92f5-234f-4de8-b37b-09be481708e1-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"636c92f5-234f-4de8-b37b-09be481708e1\") " pod="openstack/nova-scheduler-0" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.069937 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/636c92f5-234f-4de8-b37b-09be481708e1-config-data\") pod \"nova-scheduler-0\" (UID: \"636c92f5-234f-4de8-b37b-09be481708e1\") " pod="openstack/nova-scheduler-0" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.070052 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1915a375-806d-43bb-8749-9524db071714-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"1915a375-806d-43bb-8749-9524db071714\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.070143 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v4n9s\" (UniqueName: \"kubernetes.io/projected/a99d655e-f6db-4c83-a950-b935588b7df1-kube-api-access-v4n9s\") pod \"nova-metadata-0\" (UID: \"a99d655e-f6db-4c83-a950-b935588b7df1\") " pod="openstack/nova-metadata-0" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.074610 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.076322 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.097232 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/636c92f5-234f-4de8-b37b-09be481708e1-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"636c92f5-234f-4de8-b37b-09be481708e1\") " pod="openstack/nova-scheduler-0" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.097758 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.097987 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/636c92f5-234f-4de8-b37b-09be481708e1-config-data\") pod \"nova-scheduler-0\" (UID: \"636c92f5-234f-4de8-b37b-09be481708e1\") " pod="openstack/nova-scheduler-0" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.134925 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6c8mc\" (UniqueName: \"kubernetes.io/projected/636c92f5-234f-4de8-b37b-09be481708e1-kube-api-access-6c8mc\") pod \"nova-scheduler-0\" (UID: \"636c92f5-234f-4de8-b37b-09be481708e1\") " pod="openstack/nova-scheduler-0" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.150259 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.175121 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1915a375-806d-43bb-8749-9524db071714-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"1915a375-806d-43bb-8749-9524db071714\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.175193 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1915a375-806d-43bb-8749-9524db071714-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"1915a375-806d-43bb-8749-9524db071714\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.175216 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/522d9ac2-33a0-4499-8575-b9e252e1c859-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"522d9ac2-33a0-4499-8575-b9e252e1c859\") " pod="openstack/nova-api-0" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.175244 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v4n9s\" (UniqueName: \"kubernetes.io/projected/a99d655e-f6db-4c83-a950-b935588b7df1-kube-api-access-v4n9s\") pod \"nova-metadata-0\" (UID: \"a99d655e-f6db-4c83-a950-b935588b7df1\") " pod="openstack/nova-metadata-0" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.175290 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/522d9ac2-33a0-4499-8575-b9e252e1c859-logs\") pod \"nova-api-0\" (UID: \"522d9ac2-33a0-4499-8575-b9e252e1c859\") " pod="openstack/nova-api-0" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.175314 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/522d9ac2-33a0-4499-8575-b9e252e1c859-config-data\") pod \"nova-api-0\" (UID: \"522d9ac2-33a0-4499-8575-b9e252e1c859\") " pod="openstack/nova-api-0" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.175333 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a99d655e-f6db-4c83-a950-b935588b7df1-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"a99d655e-f6db-4c83-a950-b935588b7df1\") " pod="openstack/nova-metadata-0" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.175360 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-phktq\" (UniqueName: \"kubernetes.io/projected/1915a375-806d-43bb-8749-9524db071714-kube-api-access-phktq\") pod \"nova-cell1-novncproxy-0\" (UID: \"1915a375-806d-43bb-8749-9524db071714\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.175407 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a99d655e-f6db-4c83-a950-b935588b7df1-config-data\") pod \"nova-metadata-0\" (UID: \"a99d655e-f6db-4c83-a950-b935588b7df1\") " pod="openstack/nova-metadata-0" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.175436 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2p88n\" (UniqueName: \"kubernetes.io/projected/522d9ac2-33a0-4499-8575-b9e252e1c859-kube-api-access-2p88n\") pod \"nova-api-0\" (UID: \"522d9ac2-33a0-4499-8575-b9e252e1c859\") " pod="openstack/nova-api-0" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.175455 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a99d655e-f6db-4c83-a950-b935588b7df1-logs\") pod \"nova-metadata-0\" (UID: \"a99d655e-f6db-4c83-a950-b935588b7df1\") " pod="openstack/nova-metadata-0" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.175825 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a99d655e-f6db-4c83-a950-b935588b7df1-logs\") pod \"nova-metadata-0\" (UID: \"a99d655e-f6db-4c83-a950-b935588b7df1\") " pod="openstack/nova-metadata-0" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.192678 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1915a375-806d-43bb-8749-9524db071714-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"1915a375-806d-43bb-8749-9524db071714\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.192700 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a99d655e-f6db-4c83-a950-b935588b7df1-config-data\") pod \"nova-metadata-0\" (UID: \"a99d655e-f6db-4c83-a950-b935588b7df1\") " pod="openstack/nova-metadata-0" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.194250 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a99d655e-f6db-4c83-a950-b935588b7df1-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"a99d655e-f6db-4c83-a950-b935588b7df1\") " pod="openstack/nova-metadata-0" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.207447 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-phktq\" (UniqueName: \"kubernetes.io/projected/1915a375-806d-43bb-8749-9524db071714-kube-api-access-phktq\") pod \"nova-cell1-novncproxy-0\" (UID: \"1915a375-806d-43bb-8749-9524db071714\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.207875 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.213183 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1915a375-806d-43bb-8749-9524db071714-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"1915a375-806d-43bb-8749-9524db071714\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.279376 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/522d9ac2-33a0-4499-8575-b9e252e1c859-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"522d9ac2-33a0-4499-8575-b9e252e1c859\") " pod="openstack/nova-api-0" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.279489 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/522d9ac2-33a0-4499-8575-b9e252e1c859-logs\") pod \"nova-api-0\" (UID: \"522d9ac2-33a0-4499-8575-b9e252e1c859\") " pod="openstack/nova-api-0" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.279518 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/522d9ac2-33a0-4499-8575-b9e252e1c859-config-data\") pod \"nova-api-0\" (UID: \"522d9ac2-33a0-4499-8575-b9e252e1c859\") " pod="openstack/nova-api-0" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.279607 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2p88n\" (UniqueName: \"kubernetes.io/projected/522d9ac2-33a0-4499-8575-b9e252e1c859-kube-api-access-2p88n\") pod \"nova-api-0\" (UID: \"522d9ac2-33a0-4499-8575-b9e252e1c859\") " pod="openstack/nova-api-0" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.280900 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v4n9s\" (UniqueName: \"kubernetes.io/projected/a99d655e-f6db-4c83-a950-b935588b7df1-kube-api-access-v4n9s\") pod \"nova-metadata-0\" (UID: \"a99d655e-f6db-4c83-a950-b935588b7df1\") " pod="openstack/nova-metadata-0" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.281168 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/522d9ac2-33a0-4499-8575-b9e252e1c859-logs\") pod \"nova-api-0\" (UID: \"522d9ac2-33a0-4499-8575-b9e252e1c859\") " pod="openstack/nova-api-0" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.293029 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-69b4446475-ttj9k"] Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.294473 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-69b4446475-ttj9k" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.303915 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/522d9ac2-33a0-4499-8575-b9e252e1c859-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"522d9ac2-33a0-4499-8575-b9e252e1c859\") " pod="openstack/nova-api-0" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.304168 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/522d9ac2-33a0-4499-8575-b9e252e1c859-config-data\") pod \"nova-api-0\" (UID: \"522d9ac2-33a0-4499-8575-b9e252e1c859\") " pod="openstack/nova-api-0" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.305720 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-69b4446475-ttj9k"] Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.321782 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.333463 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2p88n\" (UniqueName: \"kubernetes.io/projected/522d9ac2-33a0-4499-8575-b9e252e1c859-kube-api-access-2p88n\") pod \"nova-api-0\" (UID: \"522d9ac2-33a0-4499-8575-b9e252e1c859\") " pod="openstack/nova-api-0" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.366342 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.381063 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f56f7387-a1c9-4250-8abf-6c0a3830970e-dns-svc\") pod \"dnsmasq-dns-69b4446475-ttj9k\" (UID: \"f56f7387-a1c9-4250-8abf-6c0a3830970e\") " pod="openstack/dnsmasq-dns-69b4446475-ttj9k" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.381123 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f56f7387-a1c9-4250-8abf-6c0a3830970e-dns-swift-storage-0\") pod \"dnsmasq-dns-69b4446475-ttj9k\" (UID: \"f56f7387-a1c9-4250-8abf-6c0a3830970e\") " pod="openstack/dnsmasq-dns-69b4446475-ttj9k" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.381151 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f56f7387-a1c9-4250-8abf-6c0a3830970e-config\") pod \"dnsmasq-dns-69b4446475-ttj9k\" (UID: \"f56f7387-a1c9-4250-8abf-6c0a3830970e\") " pod="openstack/dnsmasq-dns-69b4446475-ttj9k" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.381549 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f56f7387-a1c9-4250-8abf-6c0a3830970e-ovsdbserver-sb\") pod \"dnsmasq-dns-69b4446475-ttj9k\" (UID: \"f56f7387-a1c9-4250-8abf-6c0a3830970e\") " pod="openstack/dnsmasq-dns-69b4446475-ttj9k" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.381638 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f56f7387-a1c9-4250-8abf-6c0a3830970e-ovsdbserver-nb\") pod \"dnsmasq-dns-69b4446475-ttj9k\" (UID: \"f56f7387-a1c9-4250-8abf-6c0a3830970e\") " pod="openstack/dnsmasq-dns-69b4446475-ttj9k" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.381660 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8c8lc\" (UniqueName: \"kubernetes.io/projected/f56f7387-a1c9-4250-8abf-6c0a3830970e-kube-api-access-8c8lc\") pod \"dnsmasq-dns-69b4446475-ttj9k\" (UID: \"f56f7387-a1c9-4250-8abf-6c0a3830970e\") " pod="openstack/dnsmasq-dns-69b4446475-ttj9k" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.479245 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.483215 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f56f7387-a1c9-4250-8abf-6c0a3830970e-ovsdbserver-sb\") pod \"dnsmasq-dns-69b4446475-ttj9k\" (UID: \"f56f7387-a1c9-4250-8abf-6c0a3830970e\") " pod="openstack/dnsmasq-dns-69b4446475-ttj9k" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.483257 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f56f7387-a1c9-4250-8abf-6c0a3830970e-ovsdbserver-nb\") pod \"dnsmasq-dns-69b4446475-ttj9k\" (UID: \"f56f7387-a1c9-4250-8abf-6c0a3830970e\") " pod="openstack/dnsmasq-dns-69b4446475-ttj9k" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.483273 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8c8lc\" (UniqueName: \"kubernetes.io/projected/f56f7387-a1c9-4250-8abf-6c0a3830970e-kube-api-access-8c8lc\") pod \"dnsmasq-dns-69b4446475-ttj9k\" (UID: \"f56f7387-a1c9-4250-8abf-6c0a3830970e\") " pod="openstack/dnsmasq-dns-69b4446475-ttj9k" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.483314 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f56f7387-a1c9-4250-8abf-6c0a3830970e-dns-svc\") pod \"dnsmasq-dns-69b4446475-ttj9k\" (UID: \"f56f7387-a1c9-4250-8abf-6c0a3830970e\") " pod="openstack/dnsmasq-dns-69b4446475-ttj9k" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.483343 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f56f7387-a1c9-4250-8abf-6c0a3830970e-dns-swift-storage-0\") pod \"dnsmasq-dns-69b4446475-ttj9k\" (UID: \"f56f7387-a1c9-4250-8abf-6c0a3830970e\") " pod="openstack/dnsmasq-dns-69b4446475-ttj9k" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.483366 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f56f7387-a1c9-4250-8abf-6c0a3830970e-config\") pod \"dnsmasq-dns-69b4446475-ttj9k\" (UID: \"f56f7387-a1c9-4250-8abf-6c0a3830970e\") " pod="openstack/dnsmasq-dns-69b4446475-ttj9k" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.484240 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f56f7387-a1c9-4250-8abf-6c0a3830970e-config\") pod \"dnsmasq-dns-69b4446475-ttj9k\" (UID: \"f56f7387-a1c9-4250-8abf-6c0a3830970e\") " pod="openstack/dnsmasq-dns-69b4446475-ttj9k" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.484614 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f56f7387-a1c9-4250-8abf-6c0a3830970e-ovsdbserver-sb\") pod \"dnsmasq-dns-69b4446475-ttj9k\" (UID: \"f56f7387-a1c9-4250-8abf-6c0a3830970e\") " pod="openstack/dnsmasq-dns-69b4446475-ttj9k" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.485037 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f56f7387-a1c9-4250-8abf-6c0a3830970e-ovsdbserver-nb\") pod \"dnsmasq-dns-69b4446475-ttj9k\" (UID: \"f56f7387-a1c9-4250-8abf-6c0a3830970e\") " pod="openstack/dnsmasq-dns-69b4446475-ttj9k" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.485181 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f56f7387-a1c9-4250-8abf-6c0a3830970e-dns-svc\") pod \"dnsmasq-dns-69b4446475-ttj9k\" (UID: \"f56f7387-a1c9-4250-8abf-6c0a3830970e\") " pod="openstack/dnsmasq-dns-69b4446475-ttj9k" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.485570 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f56f7387-a1c9-4250-8abf-6c0a3830970e-dns-swift-storage-0\") pod \"dnsmasq-dns-69b4446475-ttj9k\" (UID: \"f56f7387-a1c9-4250-8abf-6c0a3830970e\") " pod="openstack/dnsmasq-dns-69b4446475-ttj9k" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.503060 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8c8lc\" (UniqueName: \"kubernetes.io/projected/f56f7387-a1c9-4250-8abf-6c0a3830970e-kube-api-access-8c8lc\") pod \"dnsmasq-dns-69b4446475-ttj9k\" (UID: \"f56f7387-a1c9-4250-8abf-6c0a3830970e\") " pod="openstack/dnsmasq-dns-69b4446475-ttj9k" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.609037 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-6g25t"] Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.647381 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-69b4446475-ttj9k" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.683788 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-6g25t" event={"ID":"1b701a42-f649-4657-9250-49188b10d3b9","Type":"ContainerStarted","Data":"63f23d24127f696c6ac8a975f8b88fae3bcf7cd62b9d4f19bb8d47d4b8611d9a"} Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.866073 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.936425 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-xw7n9"] Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.938381 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-xw7n9" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.944514 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.944769 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.953502 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-xw7n9"] Mar 13 14:29:01 crc kubenswrapper[4907]: I0313 14:29:01.005444 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 13 14:29:01 crc kubenswrapper[4907]: W0313 14:29:01.008114 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1915a375_806d_43bb_8749_9524db071714.slice/crio-878ef11d68a7add33ce200311e92840ebc8d43565aea3bfc5d4ea32961e387a8 WatchSource:0}: Error finding container 878ef11d68a7add33ce200311e92840ebc8d43565aea3bfc5d4ea32961e387a8: Status 404 returned error can't find the container with id 878ef11d68a7add33ce200311e92840ebc8d43565aea3bfc5d4ea32961e387a8 Mar 13 14:29:01 crc kubenswrapper[4907]: I0313 14:29:01.106933 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65556420-dc7f-49c8-af59-abafebbb4674-config-data\") pod \"nova-cell1-conductor-db-sync-xw7n9\" (UID: \"65556420-dc7f-49c8-af59-abafebbb4674\") " pod="openstack/nova-cell1-conductor-db-sync-xw7n9" Mar 13 14:29:01 crc kubenswrapper[4907]: I0313 14:29:01.107004 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65556420-dc7f-49c8-af59-abafebbb4674-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-xw7n9\" (UID: \"65556420-dc7f-49c8-af59-abafebbb4674\") " pod="openstack/nova-cell1-conductor-db-sync-xw7n9" Mar 13 14:29:01 crc kubenswrapper[4907]: I0313 14:29:01.107033 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zvml2\" (UniqueName: \"kubernetes.io/projected/65556420-dc7f-49c8-af59-abafebbb4674-kube-api-access-zvml2\") pod \"nova-cell1-conductor-db-sync-xw7n9\" (UID: \"65556420-dc7f-49c8-af59-abafebbb4674\") " pod="openstack/nova-cell1-conductor-db-sync-xw7n9" Mar 13 14:29:01 crc kubenswrapper[4907]: I0313 14:29:01.107056 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/65556420-dc7f-49c8-af59-abafebbb4674-scripts\") pod \"nova-cell1-conductor-db-sync-xw7n9\" (UID: \"65556420-dc7f-49c8-af59-abafebbb4674\") " pod="openstack/nova-cell1-conductor-db-sync-xw7n9" Mar 13 14:29:01 crc kubenswrapper[4907]: I0313 14:29:01.209840 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65556420-dc7f-49c8-af59-abafebbb4674-config-data\") pod \"nova-cell1-conductor-db-sync-xw7n9\" (UID: \"65556420-dc7f-49c8-af59-abafebbb4674\") " pod="openstack/nova-cell1-conductor-db-sync-xw7n9" Mar 13 14:29:01 crc kubenswrapper[4907]: I0313 14:29:01.212757 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65556420-dc7f-49c8-af59-abafebbb4674-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-xw7n9\" (UID: \"65556420-dc7f-49c8-af59-abafebbb4674\") " pod="openstack/nova-cell1-conductor-db-sync-xw7n9" Mar 13 14:29:01 crc kubenswrapper[4907]: I0313 14:29:01.212791 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zvml2\" (UniqueName: \"kubernetes.io/projected/65556420-dc7f-49c8-af59-abafebbb4674-kube-api-access-zvml2\") pod \"nova-cell1-conductor-db-sync-xw7n9\" (UID: \"65556420-dc7f-49c8-af59-abafebbb4674\") " pod="openstack/nova-cell1-conductor-db-sync-xw7n9" Mar 13 14:29:01 crc kubenswrapper[4907]: I0313 14:29:01.212857 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/65556420-dc7f-49c8-af59-abafebbb4674-scripts\") pod \"nova-cell1-conductor-db-sync-xw7n9\" (UID: \"65556420-dc7f-49c8-af59-abafebbb4674\") " pod="openstack/nova-cell1-conductor-db-sync-xw7n9" Mar 13 14:29:01 crc kubenswrapper[4907]: I0313 14:29:01.222232 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65556420-dc7f-49c8-af59-abafebbb4674-config-data\") pod \"nova-cell1-conductor-db-sync-xw7n9\" (UID: \"65556420-dc7f-49c8-af59-abafebbb4674\") " pod="openstack/nova-cell1-conductor-db-sync-xw7n9" Mar 13 14:29:01 crc kubenswrapper[4907]: I0313 14:29:01.224495 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/65556420-dc7f-49c8-af59-abafebbb4674-scripts\") pod \"nova-cell1-conductor-db-sync-xw7n9\" (UID: \"65556420-dc7f-49c8-af59-abafebbb4674\") " pod="openstack/nova-cell1-conductor-db-sync-xw7n9" Mar 13 14:29:01 crc kubenswrapper[4907]: I0313 14:29:01.225462 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65556420-dc7f-49c8-af59-abafebbb4674-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-xw7n9\" (UID: \"65556420-dc7f-49c8-af59-abafebbb4674\") " pod="openstack/nova-cell1-conductor-db-sync-xw7n9" Mar 13 14:29:01 crc kubenswrapper[4907]: I0313 14:29:01.227966 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 13 14:29:01 crc kubenswrapper[4907]: I0313 14:29:01.235749 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Mar 13 14:29:01 crc kubenswrapper[4907]: I0313 14:29:01.241643 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zvml2\" (UniqueName: \"kubernetes.io/projected/65556420-dc7f-49c8-af59-abafebbb4674-kube-api-access-zvml2\") pod \"nova-cell1-conductor-db-sync-xw7n9\" (UID: \"65556420-dc7f-49c8-af59-abafebbb4674\") " pod="openstack/nova-cell1-conductor-db-sync-xw7n9" Mar 13 14:29:01 crc kubenswrapper[4907]: I0313 14:29:01.270705 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-xw7n9" Mar 13 14:29:01 crc kubenswrapper[4907]: I0313 14:29:01.424684 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-69b4446475-ttj9k"] Mar 13 14:29:01 crc kubenswrapper[4907]: W0313 14:29:01.448149 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf56f7387_a1c9_4250_8abf_6c0a3830970e.slice/crio-c42518fd3cb49ffcc3d612c9faabeba7f93ee461bc56bb4d3b0a86fdeb96cbc5 WatchSource:0}: Error finding container c42518fd3cb49ffcc3d612c9faabeba7f93ee461bc56bb4d3b0a86fdeb96cbc5: Status 404 returned error can't find the container with id c42518fd3cb49ffcc3d612c9faabeba7f93ee461bc56bb4d3b0a86fdeb96cbc5 Mar 13 14:29:01 crc kubenswrapper[4907]: I0313 14:29:01.712819 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a99d655e-f6db-4c83-a950-b935588b7df1","Type":"ContainerStarted","Data":"f5827c9400a454954bce2169d1bdd16e49d4ec2488819960057446785871a720"} Mar 13 14:29:01 crc kubenswrapper[4907]: I0313 14:29:01.718863 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-69b4446475-ttj9k" event={"ID":"f56f7387-a1c9-4250-8abf-6c0a3830970e","Type":"ContainerStarted","Data":"c42518fd3cb49ffcc3d612c9faabeba7f93ee461bc56bb4d3b0a86fdeb96cbc5"} Mar 13 14:29:01 crc kubenswrapper[4907]: I0313 14:29:01.723170 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"1915a375-806d-43bb-8749-9524db071714","Type":"ContainerStarted","Data":"878ef11d68a7add33ce200311e92840ebc8d43565aea3bfc5d4ea32961e387a8"} Mar 13 14:29:01 crc kubenswrapper[4907]: I0313 14:29:01.724923 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"636c92f5-234f-4de8-b37b-09be481708e1","Type":"ContainerStarted","Data":"bc36702cf525d044a0ffd3094a7d069c598da5cdc98d37c88eb61db1ff303142"} Mar 13 14:29:01 crc kubenswrapper[4907]: I0313 14:29:01.732790 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"522d9ac2-33a0-4499-8575-b9e252e1c859","Type":"ContainerStarted","Data":"62294b7ca355e8d04aaa630081ec7abb20dd7d9bb3d9f97d4610e23fae479816"} Mar 13 14:29:01 crc kubenswrapper[4907]: I0313 14:29:01.738973 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-6g25t" event={"ID":"1b701a42-f649-4657-9250-49188b10d3b9","Type":"ContainerStarted","Data":"dd9f647265d73fb65018c8a51bfa4b5a5aa86cde573e29956aaf3479042709d4"} Mar 13 14:29:01 crc kubenswrapper[4907]: I0313 14:29:01.831607 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-6g25t" podStartSLOduration=2.8315857810000002 podStartE2EDuration="2.831585781s" podCreationTimestamp="2026-03-13 14:28:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:29:01.816097301 +0000 UTC m=+1440.715885010" watchObservedRunningTime="2026-03-13 14:29:01.831585781 +0000 UTC m=+1440.731373470" Mar 13 14:29:01 crc kubenswrapper[4907]: I0313 14:29:01.868576 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-xw7n9"] Mar 13 14:29:02 crc kubenswrapper[4907]: I0313 14:29:02.766078 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-xw7n9" event={"ID":"65556420-dc7f-49c8-af59-abafebbb4674","Type":"ContainerStarted","Data":"28a2d9fc7d22dcae7fbbc67bc1cf31d4d10a281e4e795b4b2a25e7a449ecff2c"} Mar 13 14:29:02 crc kubenswrapper[4907]: I0313 14:29:02.767231 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-xw7n9" event={"ID":"65556420-dc7f-49c8-af59-abafebbb4674","Type":"ContainerStarted","Data":"dc2a57efb2cfb970e90ab542c28a0ad4000fce19809d4298c8e7c3ed553951bb"} Mar 13 14:29:02 crc kubenswrapper[4907]: I0313 14:29:02.770960 4907 generic.go:334] "Generic (PLEG): container finished" podID="f56f7387-a1c9-4250-8abf-6c0a3830970e" containerID="d71398ef9c9b9dc145195170a4e15d1e48bb6494190d13130bb61281addfc624" exitCode=0 Mar 13 14:29:02 crc kubenswrapper[4907]: I0313 14:29:02.771027 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-69b4446475-ttj9k" event={"ID":"f56f7387-a1c9-4250-8abf-6c0a3830970e","Type":"ContainerDied","Data":"d71398ef9c9b9dc145195170a4e15d1e48bb6494190d13130bb61281addfc624"} Mar 13 14:29:02 crc kubenswrapper[4907]: I0313 14:29:02.786570 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-xw7n9" podStartSLOduration=2.786537005 podStartE2EDuration="2.786537005s" podCreationTimestamp="2026-03-13 14:29:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:29:02.78268486 +0000 UTC m=+1441.682472549" watchObservedRunningTime="2026-03-13 14:29:02.786537005 +0000 UTC m=+1441.686324694" Mar 13 14:29:04 crc kubenswrapper[4907]: I0313 14:29:04.468362 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Mar 13 14:29:04 crc kubenswrapper[4907]: I0313 14:29:04.540643 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 13 14:29:05 crc kubenswrapper[4907]: I0313 14:29:05.812104 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"636c92f5-234f-4de8-b37b-09be481708e1","Type":"ContainerStarted","Data":"0307fce5fca271fe0872451ae17b038a302f7cddd1ff8ef3cc5a23fc73b4c7d5"} Mar 13 14:29:05 crc kubenswrapper[4907]: I0313 14:29:05.814568 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"522d9ac2-33a0-4499-8575-b9e252e1c859","Type":"ContainerStarted","Data":"7dd20ac2524e3078e8bda1b7343498987597cfd7e20180905f0e9f71074bc2ec"} Mar 13 14:29:05 crc kubenswrapper[4907]: I0313 14:29:05.814615 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"522d9ac2-33a0-4499-8575-b9e252e1c859","Type":"ContainerStarted","Data":"7f405b45cecbda753272f1e3a9f4ad5fdb7025287b9661064da148b88958c60f"} Mar 13 14:29:05 crc kubenswrapper[4907]: I0313 14:29:05.817790 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a99d655e-f6db-4c83-a950-b935588b7df1","Type":"ContainerStarted","Data":"ab262217f9b17cb1ba1a68c95414fc65431f7517928e9e7fc8f78902f26ff961"} Mar 13 14:29:05 crc kubenswrapper[4907]: I0313 14:29:05.817823 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="a99d655e-f6db-4c83-a950-b935588b7df1" containerName="nova-metadata-log" containerID="cri-o://c48900e9ab50388f765d2b50bcf7cb048feb1b78e5669c08c203f15509de3dac" gracePeriod=30 Mar 13 14:29:05 crc kubenswrapper[4907]: I0313 14:29:05.817837 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a99d655e-f6db-4c83-a950-b935588b7df1","Type":"ContainerStarted","Data":"c48900e9ab50388f765d2b50bcf7cb048feb1b78e5669c08c203f15509de3dac"} Mar 13 14:29:05 crc kubenswrapper[4907]: I0313 14:29:05.817942 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="a99d655e-f6db-4c83-a950-b935588b7df1" containerName="nova-metadata-metadata" containerID="cri-o://ab262217f9b17cb1ba1a68c95414fc65431f7517928e9e7fc8f78902f26ff961" gracePeriod=30 Mar 13 14:29:05 crc kubenswrapper[4907]: I0313 14:29:05.829018 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-69b4446475-ttj9k" event={"ID":"f56f7387-a1c9-4250-8abf-6c0a3830970e","Type":"ContainerStarted","Data":"215938dfe070eab04984e52eb7b82e6d3f56a88ab88790cd28ad3b1a4bc7d77a"} Mar 13 14:29:05 crc kubenswrapper[4907]: I0313 14:29:05.829383 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-69b4446475-ttj9k" Mar 13 14:29:05 crc kubenswrapper[4907]: I0313 14:29:05.831356 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.929577322 podStartE2EDuration="6.831345623s" podCreationTimestamp="2026-03-13 14:28:59 +0000 UTC" firstStartedPulling="2026-03-13 14:29:00.88447607 +0000 UTC m=+1439.784263759" lastFinishedPulling="2026-03-13 14:29:04.786244371 +0000 UTC m=+1443.686032060" observedRunningTime="2026-03-13 14:29:05.831089606 +0000 UTC m=+1444.730877315" watchObservedRunningTime="2026-03-13 14:29:05.831345623 +0000 UTC m=+1444.731133312" Mar 13 14:29:05 crc kubenswrapper[4907]: I0313 14:29:05.836656 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"1915a375-806d-43bb-8749-9524db071714","Type":"ContainerStarted","Data":"45b20ff335a1fe9f36ec78c46577e6fb80b04fdde4e9de1eb52326676228f285"} Mar 13 14:29:05 crc kubenswrapper[4907]: I0313 14:29:05.836766 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="1915a375-806d-43bb-8749-9524db071714" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://45b20ff335a1fe9f36ec78c46577e6fb80b04fdde4e9de1eb52326676228f285" gracePeriod=30 Mar 13 14:29:05 crc kubenswrapper[4907]: I0313 14:29:05.859596 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-69b4446475-ttj9k" podStartSLOduration=5.8595774800000004 podStartE2EDuration="5.85957748s" podCreationTimestamp="2026-03-13 14:29:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:29:05.856748843 +0000 UTC m=+1444.756536532" watchObservedRunningTime="2026-03-13 14:29:05.85957748 +0000 UTC m=+1444.759365169" Mar 13 14:29:05 crc kubenswrapper[4907]: I0313 14:29:05.876678 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.337270115 podStartE2EDuration="6.876656364s" podCreationTimestamp="2026-03-13 14:28:59 +0000 UTC" firstStartedPulling="2026-03-13 14:29:01.249176095 +0000 UTC m=+1440.148963784" lastFinishedPulling="2026-03-13 14:29:04.788562354 +0000 UTC m=+1443.688350033" observedRunningTime="2026-03-13 14:29:05.874145625 +0000 UTC m=+1444.773933314" watchObservedRunningTime="2026-03-13 14:29:05.876656364 +0000 UTC m=+1444.776444053" Mar 13 14:29:05 crc kubenswrapper[4907]: I0313 14:29:05.921300 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.376429297 podStartE2EDuration="6.921280076s" podCreationTimestamp="2026-03-13 14:28:59 +0000 UTC" firstStartedPulling="2026-03-13 14:29:01.241386252 +0000 UTC m=+1440.141173941" lastFinishedPulling="2026-03-13 14:29:04.786237031 +0000 UTC m=+1443.686024720" observedRunningTime="2026-03-13 14:29:05.888580787 +0000 UTC m=+1444.788368476" watchObservedRunningTime="2026-03-13 14:29:05.921280076 +0000 UTC m=+1444.821067765" Mar 13 14:29:06 crc kubenswrapper[4907]: I0313 14:29:06.849283 4907 generic.go:334] "Generic (PLEG): container finished" podID="a99d655e-f6db-4c83-a950-b935588b7df1" containerID="c48900e9ab50388f765d2b50bcf7cb048feb1b78e5669c08c203f15509de3dac" exitCode=143 Mar 13 14:29:06 crc kubenswrapper[4907]: I0313 14:29:06.849998 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a99d655e-f6db-4c83-a950-b935588b7df1","Type":"ContainerDied","Data":"c48900e9ab50388f765d2b50bcf7cb048feb1b78e5669c08c203f15509de3dac"} Mar 13 14:29:09 crc kubenswrapper[4907]: I0313 14:29:09.879855 4907 generic.go:334] "Generic (PLEG): container finished" podID="1b701a42-f649-4657-9250-49188b10d3b9" containerID="dd9f647265d73fb65018c8a51bfa4b5a5aa86cde573e29956aaf3479042709d4" exitCode=0 Mar 13 14:29:09 crc kubenswrapper[4907]: I0313 14:29:09.879944 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-6g25t" event={"ID":"1b701a42-f649-4657-9250-49188b10d3b9","Type":"ContainerDied","Data":"dd9f647265d73fb65018c8a51bfa4b5a5aa86cde573e29956aaf3479042709d4"} Mar 13 14:29:09 crc kubenswrapper[4907]: I0313 14:29:09.911042 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=7.133873359 podStartE2EDuration="10.911026676s" podCreationTimestamp="2026-03-13 14:28:59 +0000 UTC" firstStartedPulling="2026-03-13 14:29:01.009811644 +0000 UTC m=+1439.909599323" lastFinishedPulling="2026-03-13 14:29:04.786964951 +0000 UTC m=+1443.686752640" observedRunningTime="2026-03-13 14:29:05.956229415 +0000 UTC m=+1444.856017104" watchObservedRunningTime="2026-03-13 14:29:09.911026676 +0000 UTC m=+1448.810814365" Mar 13 14:29:10 crc kubenswrapper[4907]: I0313 14:29:10.208827 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Mar 13 14:29:10 crc kubenswrapper[4907]: I0313 14:29:10.208892 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Mar 13 14:29:10 crc kubenswrapper[4907]: I0313 14:29:10.237324 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Mar 13 14:29:10 crc kubenswrapper[4907]: I0313 14:29:10.322922 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Mar 13 14:29:10 crc kubenswrapper[4907]: I0313 14:29:10.480872 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Mar 13 14:29:10 crc kubenswrapper[4907]: I0313 14:29:10.480943 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Mar 13 14:29:10 crc kubenswrapper[4907]: I0313 14:29:10.649252 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-69b4446475-ttj9k" Mar 13 14:29:10 crc kubenswrapper[4907]: I0313 14:29:10.710062 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-58b85ccffc-jt5lv"] Mar 13 14:29:10 crc kubenswrapper[4907]: I0313 14:29:10.710296 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-58b85ccffc-jt5lv" podUID="eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24" containerName="dnsmasq-dns" containerID="cri-o://6b06c4d709787d5a6f17229ea98351d8fef051feb5e0a2046eb23fd42751cf63" gracePeriod=10 Mar 13 14:29:10 crc kubenswrapper[4907]: I0313 14:29:10.894866 4907 generic.go:334] "Generic (PLEG): container finished" podID="65556420-dc7f-49c8-af59-abafebbb4674" containerID="28a2d9fc7d22dcae7fbbc67bc1cf31d4d10a281e4e795b4b2a25e7a449ecff2c" exitCode=0 Mar 13 14:29:10 crc kubenswrapper[4907]: I0313 14:29:10.894964 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-xw7n9" event={"ID":"65556420-dc7f-49c8-af59-abafebbb4674","Type":"ContainerDied","Data":"28a2d9fc7d22dcae7fbbc67bc1cf31d4d10a281e4e795b4b2a25e7a449ecff2c"} Mar 13 14:29:10 crc kubenswrapper[4907]: I0313 14:29:10.898967 4907 generic.go:334] "Generic (PLEG): container finished" podID="eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24" containerID="6b06c4d709787d5a6f17229ea98351d8fef051feb5e0a2046eb23fd42751cf63" exitCode=0 Mar 13 14:29:10 crc kubenswrapper[4907]: I0313 14:29:10.900157 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58b85ccffc-jt5lv" event={"ID":"eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24","Type":"ContainerDied","Data":"6b06c4d709787d5a6f17229ea98351d8fef051feb5e0a2046eb23fd42751cf63"} Mar 13 14:29:10 crc kubenswrapper[4907]: I0313 14:29:10.942175 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Mar 13 14:29:11 crc kubenswrapper[4907]: I0313 14:29:11.474709 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58b85ccffc-jt5lv" Mar 13 14:29:11 crc kubenswrapper[4907]: I0313 14:29:11.480743 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-6g25t" Mar 13 14:29:11 crc kubenswrapper[4907]: I0313 14:29:11.556103 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jrsdn\" (UniqueName: \"kubernetes.io/projected/1b701a42-f649-4657-9250-49188b10d3b9-kube-api-access-jrsdn\") pod \"1b701a42-f649-4657-9250-49188b10d3b9\" (UID: \"1b701a42-f649-4657-9250-49188b10d3b9\") " Mar 13 14:29:11 crc kubenswrapper[4907]: I0313 14:29:11.556253 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5zvgh\" (UniqueName: \"kubernetes.io/projected/eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24-kube-api-access-5zvgh\") pod \"eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24\" (UID: \"eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24\") " Mar 13 14:29:11 crc kubenswrapper[4907]: I0313 14:29:11.556305 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b701a42-f649-4657-9250-49188b10d3b9-config-data\") pod \"1b701a42-f649-4657-9250-49188b10d3b9\" (UID: \"1b701a42-f649-4657-9250-49188b10d3b9\") " Mar 13 14:29:11 crc kubenswrapper[4907]: I0313 14:29:11.556331 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24-ovsdbserver-sb\") pod \"eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24\" (UID: \"eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24\") " Mar 13 14:29:11 crc kubenswrapper[4907]: I0313 14:29:11.556347 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b701a42-f649-4657-9250-49188b10d3b9-combined-ca-bundle\") pod \"1b701a42-f649-4657-9250-49188b10d3b9\" (UID: \"1b701a42-f649-4657-9250-49188b10d3b9\") " Mar 13 14:29:11 crc kubenswrapper[4907]: I0313 14:29:11.556392 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24-ovsdbserver-nb\") pod \"eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24\" (UID: \"eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24\") " Mar 13 14:29:11 crc kubenswrapper[4907]: I0313 14:29:11.556450 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1b701a42-f649-4657-9250-49188b10d3b9-scripts\") pod \"1b701a42-f649-4657-9250-49188b10d3b9\" (UID: \"1b701a42-f649-4657-9250-49188b10d3b9\") " Mar 13 14:29:11 crc kubenswrapper[4907]: I0313 14:29:11.556490 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24-dns-svc\") pod \"eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24\" (UID: \"eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24\") " Mar 13 14:29:11 crc kubenswrapper[4907]: I0313 14:29:11.556515 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24-config\") pod \"eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24\" (UID: \"eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24\") " Mar 13 14:29:11 crc kubenswrapper[4907]: I0313 14:29:11.556594 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24-dns-swift-storage-0\") pod \"eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24\" (UID: \"eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24\") " Mar 13 14:29:11 crc kubenswrapper[4907]: I0313 14:29:11.566047 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="522d9ac2-33a0-4499-8575-b9e252e1c859" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.197:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 14:29:11 crc kubenswrapper[4907]: I0313 14:29:11.566180 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="522d9ac2-33a0-4499-8575-b9e252e1c859" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.197:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 14:29:11 crc kubenswrapper[4907]: I0313 14:29:11.575567 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b701a42-f649-4657-9250-49188b10d3b9-scripts" (OuterVolumeSpecName: "scripts") pod "1b701a42-f649-4657-9250-49188b10d3b9" (UID: "1b701a42-f649-4657-9250-49188b10d3b9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:29:11 crc kubenswrapper[4907]: I0313 14:29:11.575693 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24-kube-api-access-5zvgh" (OuterVolumeSpecName: "kube-api-access-5zvgh") pod "eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24" (UID: "eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24"). InnerVolumeSpecName "kube-api-access-5zvgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:29:11 crc kubenswrapper[4907]: I0313 14:29:11.582356 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1b701a42-f649-4657-9250-49188b10d3b9-kube-api-access-jrsdn" (OuterVolumeSpecName: "kube-api-access-jrsdn") pod "1b701a42-f649-4657-9250-49188b10d3b9" (UID: "1b701a42-f649-4657-9250-49188b10d3b9"). InnerVolumeSpecName "kube-api-access-jrsdn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:29:11 crc kubenswrapper[4907]: I0313 14:29:11.604437 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b701a42-f649-4657-9250-49188b10d3b9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1b701a42-f649-4657-9250-49188b10d3b9" (UID: "1b701a42-f649-4657-9250-49188b10d3b9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:29:11 crc kubenswrapper[4907]: I0313 14:29:11.640146 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b701a42-f649-4657-9250-49188b10d3b9-config-data" (OuterVolumeSpecName: "config-data") pod "1b701a42-f649-4657-9250-49188b10d3b9" (UID: "1b701a42-f649-4657-9250-49188b10d3b9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:29:11 crc kubenswrapper[4907]: I0313 14:29:11.641545 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24" (UID: "eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:29:11 crc kubenswrapper[4907]: I0313 14:29:11.656429 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24-config" (OuterVolumeSpecName: "config") pod "eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24" (UID: "eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:29:11 crc kubenswrapper[4907]: I0313 14:29:11.657695 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24" (UID: "eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:29:11 crc kubenswrapper[4907]: I0313 14:29:11.658734 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24" (UID: "eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:29:11 crc kubenswrapper[4907]: I0313 14:29:11.661923 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5zvgh\" (UniqueName: \"kubernetes.io/projected/eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24-kube-api-access-5zvgh\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:11 crc kubenswrapper[4907]: I0313 14:29:11.662153 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b701a42-f649-4657-9250-49188b10d3b9-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:11 crc kubenswrapper[4907]: I0313 14:29:11.662391 4907 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:11 crc kubenswrapper[4907]: I0313 14:29:11.662682 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b701a42-f649-4657-9250-49188b10d3b9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:11 crc kubenswrapper[4907]: I0313 14:29:11.662762 4907 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1b701a42-f649-4657-9250-49188b10d3b9-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:11 crc kubenswrapper[4907]: I0313 14:29:11.662833 4907 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:11 crc kubenswrapper[4907]: I0313 14:29:11.663270 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:11 crc kubenswrapper[4907]: I0313 14:29:11.663340 4907 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:11 crc kubenswrapper[4907]: I0313 14:29:11.663422 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jrsdn\" (UniqueName: \"kubernetes.io/projected/1b701a42-f649-4657-9250-49188b10d3b9-kube-api-access-jrsdn\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:11 crc kubenswrapper[4907]: I0313 14:29:11.673436 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24" (UID: "eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:29:11 crc kubenswrapper[4907]: I0313 14:29:11.765444 4907 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:11 crc kubenswrapper[4907]: I0313 14:29:11.910994 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58b85ccffc-jt5lv" Mar 13 14:29:11 crc kubenswrapper[4907]: I0313 14:29:11.910987 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58b85ccffc-jt5lv" event={"ID":"eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24","Type":"ContainerDied","Data":"a66bc3530699af62851033529fd1a73503f7565c8f14e919806b488f53d8de33"} Mar 13 14:29:11 crc kubenswrapper[4907]: I0313 14:29:11.911596 4907 scope.go:117] "RemoveContainer" containerID="6b06c4d709787d5a6f17229ea98351d8fef051feb5e0a2046eb23fd42751cf63" Mar 13 14:29:11 crc kubenswrapper[4907]: I0313 14:29:11.914547 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-6g25t" Mar 13 14:29:11 crc kubenswrapper[4907]: I0313 14:29:11.915105 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-6g25t" event={"ID":"1b701a42-f649-4657-9250-49188b10d3b9","Type":"ContainerDied","Data":"63f23d24127f696c6ac8a975f8b88fae3bcf7cd62b9d4f19bb8d47d4b8611d9a"} Mar 13 14:29:11 crc kubenswrapper[4907]: I0313 14:29:11.915148 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="63f23d24127f696c6ac8a975f8b88fae3bcf7cd62b9d4f19bb8d47d4b8611d9a" Mar 13 14:29:11 crc kubenswrapper[4907]: I0313 14:29:11.950625 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-58b85ccffc-jt5lv"] Mar 13 14:29:11 crc kubenswrapper[4907]: I0313 14:29:11.954846 4907 scope.go:117] "RemoveContainer" containerID="4f693ef5bd7b7f8e96bb97b86b9be101cd4aab9df400b462e0a79f5818a5b73b" Mar 13 14:29:11 crc kubenswrapper[4907]: I0313 14:29:11.974002 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-58b85ccffc-jt5lv"] Mar 13 14:29:12 crc kubenswrapper[4907]: I0313 14:29:12.069423 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 14:29:12 crc kubenswrapper[4907]: I0313 14:29:12.087517 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Mar 13 14:29:12 crc kubenswrapper[4907]: I0313 14:29:12.087978 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="522d9ac2-33a0-4499-8575-b9e252e1c859" containerName="nova-api-log" containerID="cri-o://7f405b45cecbda753272f1e3a9f4ad5fdb7025287b9661064da148b88958c60f" gracePeriod=30 Mar 13 14:29:12 crc kubenswrapper[4907]: I0313 14:29:12.088535 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="522d9ac2-33a0-4499-8575-b9e252e1c859" containerName="nova-api-api" containerID="cri-o://7dd20ac2524e3078e8bda1b7343498987597cfd7e20180905f0e9f71074bc2ec" gracePeriod=30 Mar 13 14:29:12 crc kubenswrapper[4907]: I0313 14:29:12.407742 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-xw7n9" Mar 13 14:29:12 crc kubenswrapper[4907]: I0313 14:29:12.491309 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zvml2\" (UniqueName: \"kubernetes.io/projected/65556420-dc7f-49c8-af59-abafebbb4674-kube-api-access-zvml2\") pod \"65556420-dc7f-49c8-af59-abafebbb4674\" (UID: \"65556420-dc7f-49c8-af59-abafebbb4674\") " Mar 13 14:29:12 crc kubenswrapper[4907]: I0313 14:29:12.491504 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/65556420-dc7f-49c8-af59-abafebbb4674-scripts\") pod \"65556420-dc7f-49c8-af59-abafebbb4674\" (UID: \"65556420-dc7f-49c8-af59-abafebbb4674\") " Mar 13 14:29:12 crc kubenswrapper[4907]: I0313 14:29:12.491621 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65556420-dc7f-49c8-af59-abafebbb4674-combined-ca-bundle\") pod \"65556420-dc7f-49c8-af59-abafebbb4674\" (UID: \"65556420-dc7f-49c8-af59-abafebbb4674\") " Mar 13 14:29:12 crc kubenswrapper[4907]: I0313 14:29:12.491666 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65556420-dc7f-49c8-af59-abafebbb4674-config-data\") pod \"65556420-dc7f-49c8-af59-abafebbb4674\" (UID: \"65556420-dc7f-49c8-af59-abafebbb4674\") " Mar 13 14:29:12 crc kubenswrapper[4907]: I0313 14:29:12.500758 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65556420-dc7f-49c8-af59-abafebbb4674-scripts" (OuterVolumeSpecName: "scripts") pod "65556420-dc7f-49c8-af59-abafebbb4674" (UID: "65556420-dc7f-49c8-af59-abafebbb4674"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:29:12 crc kubenswrapper[4907]: I0313 14:29:12.507196 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/65556420-dc7f-49c8-af59-abafebbb4674-kube-api-access-zvml2" (OuterVolumeSpecName: "kube-api-access-zvml2") pod "65556420-dc7f-49c8-af59-abafebbb4674" (UID: "65556420-dc7f-49c8-af59-abafebbb4674"). InnerVolumeSpecName "kube-api-access-zvml2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:29:12 crc kubenswrapper[4907]: I0313 14:29:12.527798 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65556420-dc7f-49c8-af59-abafebbb4674-config-data" (OuterVolumeSpecName: "config-data") pod "65556420-dc7f-49c8-af59-abafebbb4674" (UID: "65556420-dc7f-49c8-af59-abafebbb4674"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:29:12 crc kubenswrapper[4907]: I0313 14:29:12.538028 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65556420-dc7f-49c8-af59-abafebbb4674-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "65556420-dc7f-49c8-af59-abafebbb4674" (UID: "65556420-dc7f-49c8-af59-abafebbb4674"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:29:12 crc kubenswrapper[4907]: I0313 14:29:12.593297 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65556420-dc7f-49c8-af59-abafebbb4674-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:12 crc kubenswrapper[4907]: I0313 14:29:12.593321 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65556420-dc7f-49c8-af59-abafebbb4674-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:12 crc kubenswrapper[4907]: I0313 14:29:12.593332 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zvml2\" (UniqueName: \"kubernetes.io/projected/65556420-dc7f-49c8-af59-abafebbb4674-kube-api-access-zvml2\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:12 crc kubenswrapper[4907]: I0313 14:29:12.593341 4907 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/65556420-dc7f-49c8-af59-abafebbb4674-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:12 crc kubenswrapper[4907]: I0313 14:29:12.927605 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-xw7n9" Mar 13 14:29:12 crc kubenswrapper[4907]: I0313 14:29:12.927605 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-xw7n9" event={"ID":"65556420-dc7f-49c8-af59-abafebbb4674","Type":"ContainerDied","Data":"dc2a57efb2cfb970e90ab542c28a0ad4000fce19809d4298c8e7c3ed553951bb"} Mar 13 14:29:12 crc kubenswrapper[4907]: I0313 14:29:12.927981 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dc2a57efb2cfb970e90ab542c28a0ad4000fce19809d4298c8e7c3ed553951bb" Mar 13 14:29:12 crc kubenswrapper[4907]: I0313 14:29:12.928845 4907 generic.go:334] "Generic (PLEG): container finished" podID="522d9ac2-33a0-4499-8575-b9e252e1c859" containerID="7f405b45cecbda753272f1e3a9f4ad5fdb7025287b9661064da148b88958c60f" exitCode=143 Mar 13 14:29:12 crc kubenswrapper[4907]: I0313 14:29:12.928908 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"522d9ac2-33a0-4499-8575-b9e252e1c859","Type":"ContainerDied","Data":"7f405b45cecbda753272f1e3a9f4ad5fdb7025287b9661064da148b88958c60f"} Mar 13 14:29:12 crc kubenswrapper[4907]: I0313 14:29:12.930718 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="636c92f5-234f-4de8-b37b-09be481708e1" containerName="nova-scheduler-scheduler" containerID="cri-o://0307fce5fca271fe0872451ae17b038a302f7cddd1ff8ef3cc5a23fc73b4c7d5" gracePeriod=30 Mar 13 14:29:13 crc kubenswrapper[4907]: I0313 14:29:13.014726 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Mar 13 14:29:13 crc kubenswrapper[4907]: E0313 14:29:13.015438 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65556420-dc7f-49c8-af59-abafebbb4674" containerName="nova-cell1-conductor-db-sync" Mar 13 14:29:13 crc kubenswrapper[4907]: I0313 14:29:13.015465 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="65556420-dc7f-49c8-af59-abafebbb4674" containerName="nova-cell1-conductor-db-sync" Mar 13 14:29:13 crc kubenswrapper[4907]: E0313 14:29:13.015484 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b701a42-f649-4657-9250-49188b10d3b9" containerName="nova-manage" Mar 13 14:29:13 crc kubenswrapper[4907]: I0313 14:29:13.015492 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b701a42-f649-4657-9250-49188b10d3b9" containerName="nova-manage" Mar 13 14:29:13 crc kubenswrapper[4907]: E0313 14:29:13.015517 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24" containerName="dnsmasq-dns" Mar 13 14:29:13 crc kubenswrapper[4907]: I0313 14:29:13.015525 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24" containerName="dnsmasq-dns" Mar 13 14:29:13 crc kubenswrapper[4907]: E0313 14:29:13.015547 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24" containerName="init" Mar 13 14:29:13 crc kubenswrapper[4907]: I0313 14:29:13.015558 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24" containerName="init" Mar 13 14:29:13 crc kubenswrapper[4907]: I0313 14:29:13.015844 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b701a42-f649-4657-9250-49188b10d3b9" containerName="nova-manage" Mar 13 14:29:13 crc kubenswrapper[4907]: I0313 14:29:13.015868 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24" containerName="dnsmasq-dns" Mar 13 14:29:13 crc kubenswrapper[4907]: I0313 14:29:13.015896 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="65556420-dc7f-49c8-af59-abafebbb4674" containerName="nova-cell1-conductor-db-sync" Mar 13 14:29:13 crc kubenswrapper[4907]: I0313 14:29:13.016656 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Mar 13 14:29:13 crc kubenswrapper[4907]: I0313 14:29:13.018798 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Mar 13 14:29:13 crc kubenswrapper[4907]: I0313 14:29:13.040345 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Mar 13 14:29:13 crc kubenswrapper[4907]: I0313 14:29:13.103674 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25c1c7e6-a149-4a8e-aead-3ebdc7e56851-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"25c1c7e6-a149-4a8e-aead-3ebdc7e56851\") " pod="openstack/nova-cell1-conductor-0" Mar 13 14:29:13 crc kubenswrapper[4907]: I0313 14:29:13.103726 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gxd7v\" (UniqueName: \"kubernetes.io/projected/25c1c7e6-a149-4a8e-aead-3ebdc7e56851-kube-api-access-gxd7v\") pod \"nova-cell1-conductor-0\" (UID: \"25c1c7e6-a149-4a8e-aead-3ebdc7e56851\") " pod="openstack/nova-cell1-conductor-0" Mar 13 14:29:13 crc kubenswrapper[4907]: I0313 14:29:13.103975 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/25c1c7e6-a149-4a8e-aead-3ebdc7e56851-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"25c1c7e6-a149-4a8e-aead-3ebdc7e56851\") " pod="openstack/nova-cell1-conductor-0" Mar 13 14:29:13 crc kubenswrapper[4907]: I0313 14:29:13.205518 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/25c1c7e6-a149-4a8e-aead-3ebdc7e56851-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"25c1c7e6-a149-4a8e-aead-3ebdc7e56851\") " pod="openstack/nova-cell1-conductor-0" Mar 13 14:29:13 crc kubenswrapper[4907]: I0313 14:29:13.205976 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25c1c7e6-a149-4a8e-aead-3ebdc7e56851-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"25c1c7e6-a149-4a8e-aead-3ebdc7e56851\") " pod="openstack/nova-cell1-conductor-0" Mar 13 14:29:13 crc kubenswrapper[4907]: I0313 14:29:13.206086 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gxd7v\" (UniqueName: \"kubernetes.io/projected/25c1c7e6-a149-4a8e-aead-3ebdc7e56851-kube-api-access-gxd7v\") pod \"nova-cell1-conductor-0\" (UID: \"25c1c7e6-a149-4a8e-aead-3ebdc7e56851\") " pod="openstack/nova-cell1-conductor-0" Mar 13 14:29:13 crc kubenswrapper[4907]: I0313 14:29:13.210819 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25c1c7e6-a149-4a8e-aead-3ebdc7e56851-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"25c1c7e6-a149-4a8e-aead-3ebdc7e56851\") " pod="openstack/nova-cell1-conductor-0" Mar 13 14:29:13 crc kubenswrapper[4907]: I0313 14:29:13.217094 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/25c1c7e6-a149-4a8e-aead-3ebdc7e56851-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"25c1c7e6-a149-4a8e-aead-3ebdc7e56851\") " pod="openstack/nova-cell1-conductor-0" Mar 13 14:29:13 crc kubenswrapper[4907]: I0313 14:29:13.222430 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gxd7v\" (UniqueName: \"kubernetes.io/projected/25c1c7e6-a149-4a8e-aead-3ebdc7e56851-kube-api-access-gxd7v\") pod \"nova-cell1-conductor-0\" (UID: \"25c1c7e6-a149-4a8e-aead-3ebdc7e56851\") " pod="openstack/nova-cell1-conductor-0" Mar 13 14:29:13 crc kubenswrapper[4907]: I0313 14:29:13.335250 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Mar 13 14:29:13 crc kubenswrapper[4907]: I0313 14:29:13.793415 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24" path="/var/lib/kubelet/pods/eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24/volumes" Mar 13 14:29:13 crc kubenswrapper[4907]: I0313 14:29:13.836275 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Mar 13 14:29:13 crc kubenswrapper[4907]: I0313 14:29:13.947589 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"25c1c7e6-a149-4a8e-aead-3ebdc7e56851","Type":"ContainerStarted","Data":"b71e79c4c15864ccbc5c66db3afab587a50c68e66b3d7994a1fc84ed59fd938f"} Mar 13 14:29:14 crc kubenswrapper[4907]: I0313 14:29:14.957341 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"25c1c7e6-a149-4a8e-aead-3ebdc7e56851","Type":"ContainerStarted","Data":"77b52c0f64f55952d2255d5814b60083fd608735671f1f8507b09ea2e83db4d4"} Mar 13 14:29:14 crc kubenswrapper[4907]: I0313 14:29:14.957741 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Mar 13 14:29:14 crc kubenswrapper[4907]: I0313 14:29:14.977044 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.977027124 podStartE2EDuration="2.977027124s" podCreationTimestamp="2026-03-13 14:29:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:29:14.96947936 +0000 UTC m=+1453.869267069" watchObservedRunningTime="2026-03-13 14:29:14.977027124 +0000 UTC m=+1453.876814813" Mar 13 14:29:15 crc kubenswrapper[4907]: E0313 14:29:15.210082 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="0307fce5fca271fe0872451ae17b038a302f7cddd1ff8ef3cc5a23fc73b4c7d5" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Mar 13 14:29:15 crc kubenswrapper[4907]: E0313 14:29:15.211894 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="0307fce5fca271fe0872451ae17b038a302f7cddd1ff8ef3cc5a23fc73b4c7d5" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Mar 13 14:29:15 crc kubenswrapper[4907]: E0313 14:29:15.213305 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="0307fce5fca271fe0872451ae17b038a302f7cddd1ff8ef3cc5a23fc73b4c7d5" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Mar 13 14:29:15 crc kubenswrapper[4907]: E0313 14:29:15.213347 4907 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="636c92f5-234f-4de8-b37b-09be481708e1" containerName="nova-scheduler-scheduler" Mar 13 14:29:16 crc kubenswrapper[4907]: I0313 14:29:16.976258 4907 generic.go:334] "Generic (PLEG): container finished" podID="636c92f5-234f-4de8-b37b-09be481708e1" containerID="0307fce5fca271fe0872451ae17b038a302f7cddd1ff8ef3cc5a23fc73b4c7d5" exitCode=0 Mar 13 14:29:16 crc kubenswrapper[4907]: I0313 14:29:16.976329 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"636c92f5-234f-4de8-b37b-09be481708e1","Type":"ContainerDied","Data":"0307fce5fca271fe0872451ae17b038a302f7cddd1ff8ef3cc5a23fc73b4c7d5"} Mar 13 14:29:17 crc kubenswrapper[4907]: I0313 14:29:17.104394 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 13 14:29:17 crc kubenswrapper[4907]: I0313 14:29:17.189413 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/636c92f5-234f-4de8-b37b-09be481708e1-combined-ca-bundle\") pod \"636c92f5-234f-4de8-b37b-09be481708e1\" (UID: \"636c92f5-234f-4de8-b37b-09be481708e1\") " Mar 13 14:29:17 crc kubenswrapper[4907]: I0313 14:29:17.189607 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/636c92f5-234f-4de8-b37b-09be481708e1-config-data\") pod \"636c92f5-234f-4de8-b37b-09be481708e1\" (UID: \"636c92f5-234f-4de8-b37b-09be481708e1\") " Mar 13 14:29:17 crc kubenswrapper[4907]: I0313 14:29:17.189664 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6c8mc\" (UniqueName: \"kubernetes.io/projected/636c92f5-234f-4de8-b37b-09be481708e1-kube-api-access-6c8mc\") pod \"636c92f5-234f-4de8-b37b-09be481708e1\" (UID: \"636c92f5-234f-4de8-b37b-09be481708e1\") " Mar 13 14:29:17 crc kubenswrapper[4907]: I0313 14:29:17.195543 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/636c92f5-234f-4de8-b37b-09be481708e1-kube-api-access-6c8mc" (OuterVolumeSpecName: "kube-api-access-6c8mc") pod "636c92f5-234f-4de8-b37b-09be481708e1" (UID: "636c92f5-234f-4de8-b37b-09be481708e1"). InnerVolumeSpecName "kube-api-access-6c8mc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:29:17 crc kubenswrapper[4907]: I0313 14:29:17.221438 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/636c92f5-234f-4de8-b37b-09be481708e1-config-data" (OuterVolumeSpecName: "config-data") pod "636c92f5-234f-4de8-b37b-09be481708e1" (UID: "636c92f5-234f-4de8-b37b-09be481708e1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:29:17 crc kubenswrapper[4907]: I0313 14:29:17.222903 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/636c92f5-234f-4de8-b37b-09be481708e1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "636c92f5-234f-4de8-b37b-09be481708e1" (UID: "636c92f5-234f-4de8-b37b-09be481708e1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:29:17 crc kubenswrapper[4907]: I0313 14:29:17.292191 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/636c92f5-234f-4de8-b37b-09be481708e1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:17 crc kubenswrapper[4907]: I0313 14:29:17.292237 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/636c92f5-234f-4de8-b37b-09be481708e1-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:17 crc kubenswrapper[4907]: I0313 14:29:17.292251 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6c8mc\" (UniqueName: \"kubernetes.io/projected/636c92f5-234f-4de8-b37b-09be481708e1-kube-api-access-6c8mc\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:18 crc kubenswrapper[4907]: I0313 14:29:18.014539 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 13 14:29:18 crc kubenswrapper[4907]: I0313 14:29:18.014546 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"636c92f5-234f-4de8-b37b-09be481708e1","Type":"ContainerDied","Data":"bc36702cf525d044a0ffd3094a7d069c598da5cdc98d37c88eb61db1ff303142"} Mar 13 14:29:18 crc kubenswrapper[4907]: I0313 14:29:18.014974 4907 scope.go:117] "RemoveContainer" containerID="0307fce5fca271fe0872451ae17b038a302f7cddd1ff8ef3cc5a23fc73b4c7d5" Mar 13 14:29:18 crc kubenswrapper[4907]: I0313 14:29:18.021095 4907 generic.go:334] "Generic (PLEG): container finished" podID="522d9ac2-33a0-4499-8575-b9e252e1c859" containerID="7dd20ac2524e3078e8bda1b7343498987597cfd7e20180905f0e9f71074bc2ec" exitCode=0 Mar 13 14:29:18 crc kubenswrapper[4907]: I0313 14:29:18.021128 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"522d9ac2-33a0-4499-8575-b9e252e1c859","Type":"ContainerDied","Data":"7dd20ac2524e3078e8bda1b7343498987597cfd7e20180905f0e9f71074bc2ec"} Mar 13 14:29:18 crc kubenswrapper[4907]: I0313 14:29:18.102063 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 13 14:29:18 crc kubenswrapper[4907]: I0313 14:29:18.113934 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 14:29:18 crc kubenswrapper[4907]: I0313 14:29:18.132492 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 14:29:18 crc kubenswrapper[4907]: I0313 14:29:18.141406 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 14:29:18 crc kubenswrapper[4907]: E0313 14:29:18.141859 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="522d9ac2-33a0-4499-8575-b9e252e1c859" containerName="nova-api-api" Mar 13 14:29:18 crc kubenswrapper[4907]: I0313 14:29:18.141896 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="522d9ac2-33a0-4499-8575-b9e252e1c859" containerName="nova-api-api" Mar 13 14:29:18 crc kubenswrapper[4907]: E0313 14:29:18.141910 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="522d9ac2-33a0-4499-8575-b9e252e1c859" containerName="nova-api-log" Mar 13 14:29:18 crc kubenswrapper[4907]: I0313 14:29:18.141920 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="522d9ac2-33a0-4499-8575-b9e252e1c859" containerName="nova-api-log" Mar 13 14:29:18 crc kubenswrapper[4907]: E0313 14:29:18.141934 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="636c92f5-234f-4de8-b37b-09be481708e1" containerName="nova-scheduler-scheduler" Mar 13 14:29:18 crc kubenswrapper[4907]: I0313 14:29:18.141941 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="636c92f5-234f-4de8-b37b-09be481708e1" containerName="nova-scheduler-scheduler" Mar 13 14:29:18 crc kubenswrapper[4907]: I0313 14:29:18.142142 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="522d9ac2-33a0-4499-8575-b9e252e1c859" containerName="nova-api-api" Mar 13 14:29:18 crc kubenswrapper[4907]: I0313 14:29:18.142172 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="636c92f5-234f-4de8-b37b-09be481708e1" containerName="nova-scheduler-scheduler" Mar 13 14:29:18 crc kubenswrapper[4907]: I0313 14:29:18.142190 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="522d9ac2-33a0-4499-8575-b9e252e1c859" containerName="nova-api-log" Mar 13 14:29:18 crc kubenswrapper[4907]: I0313 14:29:18.142830 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 13 14:29:18 crc kubenswrapper[4907]: I0313 14:29:18.152323 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Mar 13 14:29:18 crc kubenswrapper[4907]: I0313 14:29:18.173417 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 14:29:18 crc kubenswrapper[4907]: I0313 14:29:18.204718 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/522d9ac2-33a0-4499-8575-b9e252e1c859-config-data\") pod \"522d9ac2-33a0-4499-8575-b9e252e1c859\" (UID: \"522d9ac2-33a0-4499-8575-b9e252e1c859\") " Mar 13 14:29:18 crc kubenswrapper[4907]: I0313 14:29:18.204853 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/522d9ac2-33a0-4499-8575-b9e252e1c859-combined-ca-bundle\") pod \"522d9ac2-33a0-4499-8575-b9e252e1c859\" (UID: \"522d9ac2-33a0-4499-8575-b9e252e1c859\") " Mar 13 14:29:18 crc kubenswrapper[4907]: I0313 14:29:18.204939 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2p88n\" (UniqueName: \"kubernetes.io/projected/522d9ac2-33a0-4499-8575-b9e252e1c859-kube-api-access-2p88n\") pod \"522d9ac2-33a0-4499-8575-b9e252e1c859\" (UID: \"522d9ac2-33a0-4499-8575-b9e252e1c859\") " Mar 13 14:29:18 crc kubenswrapper[4907]: I0313 14:29:18.205009 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/522d9ac2-33a0-4499-8575-b9e252e1c859-logs\") pod \"522d9ac2-33a0-4499-8575-b9e252e1c859\" (UID: \"522d9ac2-33a0-4499-8575-b9e252e1c859\") " Mar 13 14:29:18 crc kubenswrapper[4907]: I0313 14:29:18.206781 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/522d9ac2-33a0-4499-8575-b9e252e1c859-logs" (OuterVolumeSpecName: "logs") pod "522d9ac2-33a0-4499-8575-b9e252e1c859" (UID: "522d9ac2-33a0-4499-8575-b9e252e1c859"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:29:18 crc kubenswrapper[4907]: I0313 14:29:18.207601 4907 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/522d9ac2-33a0-4499-8575-b9e252e1c859-logs\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:18 crc kubenswrapper[4907]: I0313 14:29:18.229127 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/522d9ac2-33a0-4499-8575-b9e252e1c859-kube-api-access-2p88n" (OuterVolumeSpecName: "kube-api-access-2p88n") pod "522d9ac2-33a0-4499-8575-b9e252e1c859" (UID: "522d9ac2-33a0-4499-8575-b9e252e1c859"). InnerVolumeSpecName "kube-api-access-2p88n". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:29:18 crc kubenswrapper[4907]: E0313 14:29:18.232274 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/522d9ac2-33a0-4499-8575-b9e252e1c859-combined-ca-bundle podName:522d9ac2-33a0-4499-8575-b9e252e1c859 nodeName:}" failed. No retries permitted until 2026-03-13 14:29:18.732225797 +0000 UTC m=+1457.632013486 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "combined-ca-bundle" (UniqueName: "kubernetes.io/secret/522d9ac2-33a0-4499-8575-b9e252e1c859-combined-ca-bundle") pod "522d9ac2-33a0-4499-8575-b9e252e1c859" (UID: "522d9ac2-33a0-4499-8575-b9e252e1c859") : error deleting /var/lib/kubelet/pods/522d9ac2-33a0-4499-8575-b9e252e1c859/volume-subpaths: remove /var/lib/kubelet/pods/522d9ac2-33a0-4499-8575-b9e252e1c859/volume-subpaths: no such file or directory Mar 13 14:29:18 crc kubenswrapper[4907]: I0313 14:29:18.234509 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/522d9ac2-33a0-4499-8575-b9e252e1c859-config-data" (OuterVolumeSpecName: "config-data") pod "522d9ac2-33a0-4499-8575-b9e252e1c859" (UID: "522d9ac2-33a0-4499-8575-b9e252e1c859"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:29:18 crc kubenswrapper[4907]: I0313 14:29:18.308816 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0304716-7acd-4a2c-a018-73fc4bcbad51-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"c0304716-7acd-4a2c-a018-73fc4bcbad51\") " pod="openstack/nova-scheduler-0" Mar 13 14:29:18 crc kubenswrapper[4907]: I0313 14:29:18.308868 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xhvp6\" (UniqueName: \"kubernetes.io/projected/c0304716-7acd-4a2c-a018-73fc4bcbad51-kube-api-access-xhvp6\") pod \"nova-scheduler-0\" (UID: \"c0304716-7acd-4a2c-a018-73fc4bcbad51\") " pod="openstack/nova-scheduler-0" Mar 13 14:29:18 crc kubenswrapper[4907]: I0313 14:29:18.308959 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c0304716-7acd-4a2c-a018-73fc4bcbad51-config-data\") pod \"nova-scheduler-0\" (UID: \"c0304716-7acd-4a2c-a018-73fc4bcbad51\") " pod="openstack/nova-scheduler-0" Mar 13 14:29:18 crc kubenswrapper[4907]: I0313 14:29:18.309065 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/522d9ac2-33a0-4499-8575-b9e252e1c859-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:18 crc kubenswrapper[4907]: I0313 14:29:18.309077 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2p88n\" (UniqueName: \"kubernetes.io/projected/522d9ac2-33a0-4499-8575-b9e252e1c859-kube-api-access-2p88n\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:18 crc kubenswrapper[4907]: I0313 14:29:18.367485 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Mar 13 14:29:18 crc kubenswrapper[4907]: I0313 14:29:18.367540 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Mar 13 14:29:18 crc kubenswrapper[4907]: I0313 14:29:18.411288 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c0304716-7acd-4a2c-a018-73fc4bcbad51-config-data\") pod \"nova-scheduler-0\" (UID: \"c0304716-7acd-4a2c-a018-73fc4bcbad51\") " pod="openstack/nova-scheduler-0" Mar 13 14:29:18 crc kubenswrapper[4907]: I0313 14:29:18.411439 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0304716-7acd-4a2c-a018-73fc4bcbad51-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"c0304716-7acd-4a2c-a018-73fc4bcbad51\") " pod="openstack/nova-scheduler-0" Mar 13 14:29:18 crc kubenswrapper[4907]: I0313 14:29:18.411471 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xhvp6\" (UniqueName: \"kubernetes.io/projected/c0304716-7acd-4a2c-a018-73fc4bcbad51-kube-api-access-xhvp6\") pod \"nova-scheduler-0\" (UID: \"c0304716-7acd-4a2c-a018-73fc4bcbad51\") " pod="openstack/nova-scheduler-0" Mar 13 14:29:18 crc kubenswrapper[4907]: I0313 14:29:18.414623 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0304716-7acd-4a2c-a018-73fc4bcbad51-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"c0304716-7acd-4a2c-a018-73fc4bcbad51\") " pod="openstack/nova-scheduler-0" Mar 13 14:29:18 crc kubenswrapper[4907]: I0313 14:29:18.415367 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c0304716-7acd-4a2c-a018-73fc4bcbad51-config-data\") pod \"nova-scheduler-0\" (UID: \"c0304716-7acd-4a2c-a018-73fc4bcbad51\") " pod="openstack/nova-scheduler-0" Mar 13 14:29:18 crc kubenswrapper[4907]: I0313 14:29:18.432586 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xhvp6\" (UniqueName: \"kubernetes.io/projected/c0304716-7acd-4a2c-a018-73fc4bcbad51-kube-api-access-xhvp6\") pod \"nova-scheduler-0\" (UID: \"c0304716-7acd-4a2c-a018-73fc4bcbad51\") " pod="openstack/nova-scheduler-0" Mar 13 14:29:18 crc kubenswrapper[4907]: I0313 14:29:18.462957 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 13 14:29:18 crc kubenswrapper[4907]: I0313 14:29:18.818319 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/522d9ac2-33a0-4499-8575-b9e252e1c859-combined-ca-bundle\") pod \"522d9ac2-33a0-4499-8575-b9e252e1c859\" (UID: \"522d9ac2-33a0-4499-8575-b9e252e1c859\") " Mar 13 14:29:18 crc kubenswrapper[4907]: I0313 14:29:18.826378 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/522d9ac2-33a0-4499-8575-b9e252e1c859-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "522d9ac2-33a0-4499-8575-b9e252e1c859" (UID: "522d9ac2-33a0-4499-8575-b9e252e1c859"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:29:18 crc kubenswrapper[4907]: I0313 14:29:18.917855 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 14:29:18 crc kubenswrapper[4907]: I0313 14:29:18.920550 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/522d9ac2-33a0-4499-8575-b9e252e1c859-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:19 crc kubenswrapper[4907]: I0313 14:29:19.031096 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"c0304716-7acd-4a2c-a018-73fc4bcbad51","Type":"ContainerStarted","Data":"a765fc07ea8b46927f8043334da5e0885287349d5eefd55592990b1ae30cfba6"} Mar 13 14:29:19 crc kubenswrapper[4907]: I0313 14:29:19.034824 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"522d9ac2-33a0-4499-8575-b9e252e1c859","Type":"ContainerDied","Data":"62294b7ca355e8d04aaa630081ec7abb20dd7d9bb3d9f97d4610e23fae479816"} Mar 13 14:29:19 crc kubenswrapper[4907]: I0313 14:29:19.034857 4907 scope.go:117] "RemoveContainer" containerID="7dd20ac2524e3078e8bda1b7343498987597cfd7e20180905f0e9f71074bc2ec" Mar 13 14:29:19 crc kubenswrapper[4907]: I0313 14:29:19.034888 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 13 14:29:19 crc kubenswrapper[4907]: I0313 14:29:19.080849 4907 scope.go:117] "RemoveContainer" containerID="7f405b45cecbda753272f1e3a9f4ad5fdb7025287b9661064da148b88958c60f" Mar 13 14:29:19 crc kubenswrapper[4907]: I0313 14:29:19.106968 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Mar 13 14:29:19 crc kubenswrapper[4907]: I0313 14:29:19.122519 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Mar 13 14:29:19 crc kubenswrapper[4907]: I0313 14:29:19.136902 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Mar 13 14:29:19 crc kubenswrapper[4907]: I0313 14:29:19.138797 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 13 14:29:19 crc kubenswrapper[4907]: I0313 14:29:19.141677 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Mar 13 14:29:19 crc kubenswrapper[4907]: I0313 14:29:19.160511 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 13 14:29:19 crc kubenswrapper[4907]: I0313 14:29:19.326154 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41063592-cda5-443f-87e3-20c2c6ec5c73-config-data\") pod \"nova-api-0\" (UID: \"41063592-cda5-443f-87e3-20c2c6ec5c73\") " pod="openstack/nova-api-0" Mar 13 14:29:19 crc kubenswrapper[4907]: I0313 14:29:19.326384 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zhhbp\" (UniqueName: \"kubernetes.io/projected/41063592-cda5-443f-87e3-20c2c6ec5c73-kube-api-access-zhhbp\") pod \"nova-api-0\" (UID: \"41063592-cda5-443f-87e3-20c2c6ec5c73\") " pod="openstack/nova-api-0" Mar 13 14:29:19 crc kubenswrapper[4907]: I0313 14:29:19.326524 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41063592-cda5-443f-87e3-20c2c6ec5c73-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"41063592-cda5-443f-87e3-20c2c6ec5c73\") " pod="openstack/nova-api-0" Mar 13 14:29:19 crc kubenswrapper[4907]: I0313 14:29:19.326981 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/41063592-cda5-443f-87e3-20c2c6ec5c73-logs\") pod \"nova-api-0\" (UID: \"41063592-cda5-443f-87e3-20c2c6ec5c73\") " pod="openstack/nova-api-0" Mar 13 14:29:19 crc kubenswrapper[4907]: I0313 14:29:19.429054 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/41063592-cda5-443f-87e3-20c2c6ec5c73-logs\") pod \"nova-api-0\" (UID: \"41063592-cda5-443f-87e3-20c2c6ec5c73\") " pod="openstack/nova-api-0" Mar 13 14:29:19 crc kubenswrapper[4907]: I0313 14:29:19.429120 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41063592-cda5-443f-87e3-20c2c6ec5c73-config-data\") pod \"nova-api-0\" (UID: \"41063592-cda5-443f-87e3-20c2c6ec5c73\") " pod="openstack/nova-api-0" Mar 13 14:29:19 crc kubenswrapper[4907]: I0313 14:29:19.429161 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zhhbp\" (UniqueName: \"kubernetes.io/projected/41063592-cda5-443f-87e3-20c2c6ec5c73-kube-api-access-zhhbp\") pod \"nova-api-0\" (UID: \"41063592-cda5-443f-87e3-20c2c6ec5c73\") " pod="openstack/nova-api-0" Mar 13 14:29:19 crc kubenswrapper[4907]: I0313 14:29:19.429210 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41063592-cda5-443f-87e3-20c2c6ec5c73-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"41063592-cda5-443f-87e3-20c2c6ec5c73\") " pod="openstack/nova-api-0" Mar 13 14:29:19 crc kubenswrapper[4907]: I0313 14:29:19.429637 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/41063592-cda5-443f-87e3-20c2c6ec5c73-logs\") pod \"nova-api-0\" (UID: \"41063592-cda5-443f-87e3-20c2c6ec5c73\") " pod="openstack/nova-api-0" Mar 13 14:29:19 crc kubenswrapper[4907]: I0313 14:29:19.435274 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41063592-cda5-443f-87e3-20c2c6ec5c73-config-data\") pod \"nova-api-0\" (UID: \"41063592-cda5-443f-87e3-20c2c6ec5c73\") " pod="openstack/nova-api-0" Mar 13 14:29:19 crc kubenswrapper[4907]: I0313 14:29:19.435622 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41063592-cda5-443f-87e3-20c2c6ec5c73-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"41063592-cda5-443f-87e3-20c2c6ec5c73\") " pod="openstack/nova-api-0" Mar 13 14:29:19 crc kubenswrapper[4907]: I0313 14:29:19.450111 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zhhbp\" (UniqueName: \"kubernetes.io/projected/41063592-cda5-443f-87e3-20c2c6ec5c73-kube-api-access-zhhbp\") pod \"nova-api-0\" (UID: \"41063592-cda5-443f-87e3-20c2c6ec5c73\") " pod="openstack/nova-api-0" Mar 13 14:29:19 crc kubenswrapper[4907]: I0313 14:29:19.473831 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 13 14:29:19 crc kubenswrapper[4907]: I0313 14:29:19.816192 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="522d9ac2-33a0-4499-8575-b9e252e1c859" path="/var/lib/kubelet/pods/522d9ac2-33a0-4499-8575-b9e252e1c859/volumes" Mar 13 14:29:19 crc kubenswrapper[4907]: I0313 14:29:19.817967 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="636c92f5-234f-4de8-b37b-09be481708e1" path="/var/lib/kubelet/pods/636c92f5-234f-4de8-b37b-09be481708e1/volumes" Mar 13 14:29:19 crc kubenswrapper[4907]: I0313 14:29:19.936492 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 13 14:29:20 crc kubenswrapper[4907]: I0313 14:29:20.049458 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"c0304716-7acd-4a2c-a018-73fc4bcbad51","Type":"ContainerStarted","Data":"7dd9875251f25eba8556dcd6141b77fd52b2025f417f75c1f8e455693aab178b"} Mar 13 14:29:20 crc kubenswrapper[4907]: I0313 14:29:20.058301 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"41063592-cda5-443f-87e3-20c2c6ec5c73","Type":"ContainerStarted","Data":"11bb2f382efdf360af4e2164320ab74af741c059bdd8ac659bd82b1507cb52af"} Mar 13 14:29:20 crc kubenswrapper[4907]: I0313 14:29:20.082056 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.082029282 podStartE2EDuration="2.082029282s" podCreationTimestamp="2026-03-13 14:29:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:29:20.078835595 +0000 UTC m=+1458.978623284" watchObservedRunningTime="2026-03-13 14:29:20.082029282 +0000 UTC m=+1458.981816971" Mar 13 14:29:20 crc kubenswrapper[4907]: I0313 14:29:20.970451 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Mar 13 14:29:21 crc kubenswrapper[4907]: I0313 14:29:21.081850 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"41063592-cda5-443f-87e3-20c2c6ec5c73","Type":"ContainerStarted","Data":"bee51476e0503a2fc3a96ca20f1d42b6cb92e1822245b7b21033273b10f0e7a6"} Mar 13 14:29:21 crc kubenswrapper[4907]: I0313 14:29:21.082460 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"41063592-cda5-443f-87e3-20c2c6ec5c73","Type":"ContainerStarted","Data":"97742dae0519818b9c495c23cd7ee1fe57412023b1fc525ca01e7fc2236cef2b"} Mar 13 14:29:21 crc kubenswrapper[4907]: I0313 14:29:21.111187 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.111166431 podStartE2EDuration="2.111166431s" podCreationTimestamp="2026-03-13 14:29:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:29:21.105301411 +0000 UTC m=+1460.005089100" watchObservedRunningTime="2026-03-13 14:29:21.111166431 +0000 UTC m=+1460.010954120" Mar 13 14:29:23 crc kubenswrapper[4907]: I0313 14:29:23.380667 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Mar 13 14:29:23 crc kubenswrapper[4907]: I0313 14:29:23.463592 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Mar 13 14:29:25 crc kubenswrapper[4907]: I0313 14:29:25.274274 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Mar 13 14:29:25 crc kubenswrapper[4907]: I0313 14:29:25.274818 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="a7f6877c-1e9e-4e17-803d-90efa7d66469" containerName="kube-state-metrics" containerID="cri-o://f5792262fabc2fac4148b46830f3f0bdf740121dbf4200684e158ad3432a779c" gracePeriod=30 Mar 13 14:29:25 crc kubenswrapper[4907]: I0313 14:29:25.757683 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Mar 13 14:29:25 crc kubenswrapper[4907]: I0313 14:29:25.859438 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zg7qk\" (UniqueName: \"kubernetes.io/projected/a7f6877c-1e9e-4e17-803d-90efa7d66469-kube-api-access-zg7qk\") pod \"a7f6877c-1e9e-4e17-803d-90efa7d66469\" (UID: \"a7f6877c-1e9e-4e17-803d-90efa7d66469\") " Mar 13 14:29:25 crc kubenswrapper[4907]: I0313 14:29:25.886114 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a7f6877c-1e9e-4e17-803d-90efa7d66469-kube-api-access-zg7qk" (OuterVolumeSpecName: "kube-api-access-zg7qk") pod "a7f6877c-1e9e-4e17-803d-90efa7d66469" (UID: "a7f6877c-1e9e-4e17-803d-90efa7d66469"). InnerVolumeSpecName "kube-api-access-zg7qk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:29:25 crc kubenswrapper[4907]: I0313 14:29:25.962228 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zg7qk\" (UniqueName: \"kubernetes.io/projected/a7f6877c-1e9e-4e17-803d-90efa7d66469-kube-api-access-zg7qk\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:26 crc kubenswrapper[4907]: I0313 14:29:26.138802 4907 generic.go:334] "Generic (PLEG): container finished" podID="a7f6877c-1e9e-4e17-803d-90efa7d66469" containerID="f5792262fabc2fac4148b46830f3f0bdf740121dbf4200684e158ad3432a779c" exitCode=2 Mar 13 14:29:26 crc kubenswrapper[4907]: I0313 14:29:26.138868 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"a7f6877c-1e9e-4e17-803d-90efa7d66469","Type":"ContainerDied","Data":"f5792262fabc2fac4148b46830f3f0bdf740121dbf4200684e158ad3432a779c"} Mar 13 14:29:26 crc kubenswrapper[4907]: I0313 14:29:26.138927 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"a7f6877c-1e9e-4e17-803d-90efa7d66469","Type":"ContainerDied","Data":"3893671130f88340bc9f4b2c1c7ae2e64ebc3c35392832b178f2caf4f860f9fd"} Mar 13 14:29:26 crc kubenswrapper[4907]: I0313 14:29:26.138946 4907 scope.go:117] "RemoveContainer" containerID="f5792262fabc2fac4148b46830f3f0bdf740121dbf4200684e158ad3432a779c" Mar 13 14:29:26 crc kubenswrapper[4907]: I0313 14:29:26.138897 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Mar 13 14:29:26 crc kubenswrapper[4907]: I0313 14:29:26.172126 4907 scope.go:117] "RemoveContainer" containerID="f5792262fabc2fac4148b46830f3f0bdf740121dbf4200684e158ad3432a779c" Mar 13 14:29:26 crc kubenswrapper[4907]: E0313 14:29:26.172745 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f5792262fabc2fac4148b46830f3f0bdf740121dbf4200684e158ad3432a779c\": container with ID starting with f5792262fabc2fac4148b46830f3f0bdf740121dbf4200684e158ad3432a779c not found: ID does not exist" containerID="f5792262fabc2fac4148b46830f3f0bdf740121dbf4200684e158ad3432a779c" Mar 13 14:29:26 crc kubenswrapper[4907]: I0313 14:29:26.172777 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f5792262fabc2fac4148b46830f3f0bdf740121dbf4200684e158ad3432a779c"} err="failed to get container status \"f5792262fabc2fac4148b46830f3f0bdf740121dbf4200684e158ad3432a779c\": rpc error: code = NotFound desc = could not find container \"f5792262fabc2fac4148b46830f3f0bdf740121dbf4200684e158ad3432a779c\": container with ID starting with f5792262fabc2fac4148b46830f3f0bdf740121dbf4200684e158ad3432a779c not found: ID does not exist" Mar 13 14:29:26 crc kubenswrapper[4907]: I0313 14:29:26.182420 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Mar 13 14:29:26 crc kubenswrapper[4907]: I0313 14:29:26.194851 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Mar 13 14:29:26 crc kubenswrapper[4907]: I0313 14:29:26.201807 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Mar 13 14:29:26 crc kubenswrapper[4907]: E0313 14:29:26.202430 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7f6877c-1e9e-4e17-803d-90efa7d66469" containerName="kube-state-metrics" Mar 13 14:29:26 crc kubenswrapper[4907]: I0313 14:29:26.202459 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7f6877c-1e9e-4e17-803d-90efa7d66469" containerName="kube-state-metrics" Mar 13 14:29:26 crc kubenswrapper[4907]: I0313 14:29:26.202726 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7f6877c-1e9e-4e17-803d-90efa7d66469" containerName="kube-state-metrics" Mar 13 14:29:26 crc kubenswrapper[4907]: I0313 14:29:26.203601 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Mar 13 14:29:26 crc kubenswrapper[4907]: I0313 14:29:26.207117 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Mar 13 14:29:26 crc kubenswrapper[4907]: I0313 14:29:26.210450 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Mar 13 14:29:26 crc kubenswrapper[4907]: I0313 14:29:26.211945 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Mar 13 14:29:26 crc kubenswrapper[4907]: I0313 14:29:26.265574 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/c49286b1-2e9f-4de6-b7de-2e952f3ad607-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"c49286b1-2e9f-4de6-b7de-2e952f3ad607\") " pod="openstack/kube-state-metrics-0" Mar 13 14:29:26 crc kubenswrapper[4907]: I0313 14:29:26.265956 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-29ctv\" (UniqueName: \"kubernetes.io/projected/c49286b1-2e9f-4de6-b7de-2e952f3ad607-kube-api-access-29ctv\") pod \"kube-state-metrics-0\" (UID: \"c49286b1-2e9f-4de6-b7de-2e952f3ad607\") " pod="openstack/kube-state-metrics-0" Mar 13 14:29:26 crc kubenswrapper[4907]: I0313 14:29:26.266012 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c49286b1-2e9f-4de6-b7de-2e952f3ad607-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"c49286b1-2e9f-4de6-b7de-2e952f3ad607\") " pod="openstack/kube-state-metrics-0" Mar 13 14:29:26 crc kubenswrapper[4907]: I0313 14:29:26.266036 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/c49286b1-2e9f-4de6-b7de-2e952f3ad607-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"c49286b1-2e9f-4de6-b7de-2e952f3ad607\") " pod="openstack/kube-state-metrics-0" Mar 13 14:29:26 crc kubenswrapper[4907]: I0313 14:29:26.367028 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-29ctv\" (UniqueName: \"kubernetes.io/projected/c49286b1-2e9f-4de6-b7de-2e952f3ad607-kube-api-access-29ctv\") pod \"kube-state-metrics-0\" (UID: \"c49286b1-2e9f-4de6-b7de-2e952f3ad607\") " pod="openstack/kube-state-metrics-0" Mar 13 14:29:26 crc kubenswrapper[4907]: I0313 14:29:26.367112 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c49286b1-2e9f-4de6-b7de-2e952f3ad607-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"c49286b1-2e9f-4de6-b7de-2e952f3ad607\") " pod="openstack/kube-state-metrics-0" Mar 13 14:29:26 crc kubenswrapper[4907]: I0313 14:29:26.367153 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/c49286b1-2e9f-4de6-b7de-2e952f3ad607-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"c49286b1-2e9f-4de6-b7de-2e952f3ad607\") " pod="openstack/kube-state-metrics-0" Mar 13 14:29:26 crc kubenswrapper[4907]: I0313 14:29:26.367226 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/c49286b1-2e9f-4de6-b7de-2e952f3ad607-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"c49286b1-2e9f-4de6-b7de-2e952f3ad607\") " pod="openstack/kube-state-metrics-0" Mar 13 14:29:26 crc kubenswrapper[4907]: I0313 14:29:26.373226 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/c49286b1-2e9f-4de6-b7de-2e952f3ad607-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"c49286b1-2e9f-4de6-b7de-2e952f3ad607\") " pod="openstack/kube-state-metrics-0" Mar 13 14:29:26 crc kubenswrapper[4907]: I0313 14:29:26.373516 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c49286b1-2e9f-4de6-b7de-2e952f3ad607-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"c49286b1-2e9f-4de6-b7de-2e952f3ad607\") " pod="openstack/kube-state-metrics-0" Mar 13 14:29:26 crc kubenswrapper[4907]: I0313 14:29:26.376342 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/c49286b1-2e9f-4de6-b7de-2e952f3ad607-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"c49286b1-2e9f-4de6-b7de-2e952f3ad607\") " pod="openstack/kube-state-metrics-0" Mar 13 14:29:26 crc kubenswrapper[4907]: I0313 14:29:26.396338 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-29ctv\" (UniqueName: \"kubernetes.io/projected/c49286b1-2e9f-4de6-b7de-2e952f3ad607-kube-api-access-29ctv\") pod \"kube-state-metrics-0\" (UID: \"c49286b1-2e9f-4de6-b7de-2e952f3ad607\") " pod="openstack/kube-state-metrics-0" Mar 13 14:29:26 crc kubenswrapper[4907]: I0313 14:29:26.521653 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Mar 13 14:29:26 crc kubenswrapper[4907]: I0313 14:29:26.977306 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Mar 13 14:29:27 crc kubenswrapper[4907]: I0313 14:29:27.150057 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"c49286b1-2e9f-4de6-b7de-2e952f3ad607","Type":"ContainerStarted","Data":"50ab44d5af6ebe42498682725ea848a78fc975fa9daf535d39dda1ec402423d1"} Mar 13 14:29:27 crc kubenswrapper[4907]: I0313 14:29:27.249573 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 13 14:29:27 crc kubenswrapper[4907]: I0313 14:29:27.250086 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a85017ea-b984-471a-a9a8-22c02d94448b" containerName="ceilometer-central-agent" containerID="cri-o://62988d506f50133c1068883469efb51974597559ba58796a2aae59bf02bebb3a" gracePeriod=30 Mar 13 14:29:27 crc kubenswrapper[4907]: I0313 14:29:27.250152 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a85017ea-b984-471a-a9a8-22c02d94448b" containerName="ceilometer-notification-agent" containerID="cri-o://3b55543c8226645f0d4a440e41c724886657fe5c5d9924185cd3344bad9ddc01" gracePeriod=30 Mar 13 14:29:27 crc kubenswrapper[4907]: I0313 14:29:27.250164 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a85017ea-b984-471a-a9a8-22c02d94448b" containerName="sg-core" containerID="cri-o://f629b415ac0d40fdc7a871c081a5a0ceecff9578a74f62763cbbaae7d50396c7" gracePeriod=30 Mar 13 14:29:27 crc kubenswrapper[4907]: I0313 14:29:27.250127 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a85017ea-b984-471a-a9a8-22c02d94448b" containerName="proxy-httpd" containerID="cri-o://c2a32ce2fcf985bbc86c96669646554e0cd290d1eaa4396d46e02db24cec19b9" gracePeriod=30 Mar 13 14:29:27 crc kubenswrapper[4907]: I0313 14:29:27.793319 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a7f6877c-1e9e-4e17-803d-90efa7d66469" path="/var/lib/kubelet/pods/a7f6877c-1e9e-4e17-803d-90efa7d66469/volumes" Mar 13 14:29:28 crc kubenswrapper[4907]: I0313 14:29:28.165578 4907 generic.go:334] "Generic (PLEG): container finished" podID="a85017ea-b984-471a-a9a8-22c02d94448b" containerID="c2a32ce2fcf985bbc86c96669646554e0cd290d1eaa4396d46e02db24cec19b9" exitCode=0 Mar 13 14:29:28 crc kubenswrapper[4907]: I0313 14:29:28.165921 4907 generic.go:334] "Generic (PLEG): container finished" podID="a85017ea-b984-471a-a9a8-22c02d94448b" containerID="f629b415ac0d40fdc7a871c081a5a0ceecff9578a74f62763cbbaae7d50396c7" exitCode=2 Mar 13 14:29:28 crc kubenswrapper[4907]: I0313 14:29:28.165931 4907 generic.go:334] "Generic (PLEG): container finished" podID="a85017ea-b984-471a-a9a8-22c02d94448b" containerID="62988d506f50133c1068883469efb51974597559ba58796a2aae59bf02bebb3a" exitCode=0 Mar 13 14:29:28 crc kubenswrapper[4907]: I0313 14:29:28.165756 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a85017ea-b984-471a-a9a8-22c02d94448b","Type":"ContainerDied","Data":"c2a32ce2fcf985bbc86c96669646554e0cd290d1eaa4396d46e02db24cec19b9"} Mar 13 14:29:28 crc kubenswrapper[4907]: I0313 14:29:28.166004 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a85017ea-b984-471a-a9a8-22c02d94448b","Type":"ContainerDied","Data":"f629b415ac0d40fdc7a871c081a5a0ceecff9578a74f62763cbbaae7d50396c7"} Mar 13 14:29:28 crc kubenswrapper[4907]: I0313 14:29:28.166020 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a85017ea-b984-471a-a9a8-22c02d94448b","Type":"ContainerDied","Data":"62988d506f50133c1068883469efb51974597559ba58796a2aae59bf02bebb3a"} Mar 13 14:29:28 crc kubenswrapper[4907]: I0313 14:29:28.168374 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"c49286b1-2e9f-4de6-b7de-2e952f3ad607","Type":"ContainerStarted","Data":"0d242394ff0714a798d1ef385dde5709185206b531dfb7b51abcf0db5d059589"} Mar 13 14:29:28 crc kubenswrapper[4907]: I0313 14:29:28.169569 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Mar 13 14:29:28 crc kubenswrapper[4907]: I0313 14:29:28.190343 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=1.767998003 podStartE2EDuration="2.190325102s" podCreationTimestamp="2026-03-13 14:29:26 +0000 UTC" firstStartedPulling="2026-03-13 14:29:26.990380324 +0000 UTC m=+1465.890168013" lastFinishedPulling="2026-03-13 14:29:27.412707413 +0000 UTC m=+1466.312495112" observedRunningTime="2026-03-13 14:29:28.185208262 +0000 UTC m=+1467.084995951" watchObservedRunningTime="2026-03-13 14:29:28.190325102 +0000 UTC m=+1467.090112791" Mar 13 14:29:28 crc kubenswrapper[4907]: I0313 14:29:28.464108 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Mar 13 14:29:28 crc kubenswrapper[4907]: I0313 14:29:28.489097 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Mar 13 14:29:29 crc kubenswrapper[4907]: I0313 14:29:29.222146 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Mar 13 14:29:29 crc kubenswrapper[4907]: I0313 14:29:29.474900 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Mar 13 14:29:29 crc kubenswrapper[4907]: I0313 14:29:29.475203 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Mar 13 14:29:30 crc kubenswrapper[4907]: I0313 14:29:30.557082 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="41063592-cda5-443f-87e3-20c2c6ec5c73" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.202:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 14:29:30 crc kubenswrapper[4907]: I0313 14:29:30.557184 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="41063592-cda5-443f-87e3-20c2c6ec5c73" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.202:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 14:29:30 crc kubenswrapper[4907]: I0313 14:29:30.819692 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 14:29:30 crc kubenswrapper[4907]: I0313 14:29:30.956287 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a85017ea-b984-471a-a9a8-22c02d94448b-config-data\") pod \"a85017ea-b984-471a-a9a8-22c02d94448b\" (UID: \"a85017ea-b984-471a-a9a8-22c02d94448b\") " Mar 13 14:29:30 crc kubenswrapper[4907]: I0313 14:29:30.956380 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-462xb\" (UniqueName: \"kubernetes.io/projected/a85017ea-b984-471a-a9a8-22c02d94448b-kube-api-access-462xb\") pod \"a85017ea-b984-471a-a9a8-22c02d94448b\" (UID: \"a85017ea-b984-471a-a9a8-22c02d94448b\") " Mar 13 14:29:30 crc kubenswrapper[4907]: I0313 14:29:30.956399 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a85017ea-b984-471a-a9a8-22c02d94448b-sg-core-conf-yaml\") pod \"a85017ea-b984-471a-a9a8-22c02d94448b\" (UID: \"a85017ea-b984-471a-a9a8-22c02d94448b\") " Mar 13 14:29:30 crc kubenswrapper[4907]: I0313 14:29:30.956431 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a85017ea-b984-471a-a9a8-22c02d94448b-combined-ca-bundle\") pod \"a85017ea-b984-471a-a9a8-22c02d94448b\" (UID: \"a85017ea-b984-471a-a9a8-22c02d94448b\") " Mar 13 14:29:30 crc kubenswrapper[4907]: I0313 14:29:30.956463 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a85017ea-b984-471a-a9a8-22c02d94448b-log-httpd\") pod \"a85017ea-b984-471a-a9a8-22c02d94448b\" (UID: \"a85017ea-b984-471a-a9a8-22c02d94448b\") " Mar 13 14:29:30 crc kubenswrapper[4907]: I0313 14:29:30.956482 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a85017ea-b984-471a-a9a8-22c02d94448b-scripts\") pod \"a85017ea-b984-471a-a9a8-22c02d94448b\" (UID: \"a85017ea-b984-471a-a9a8-22c02d94448b\") " Mar 13 14:29:30 crc kubenswrapper[4907]: I0313 14:29:30.956616 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a85017ea-b984-471a-a9a8-22c02d94448b-run-httpd\") pod \"a85017ea-b984-471a-a9a8-22c02d94448b\" (UID: \"a85017ea-b984-471a-a9a8-22c02d94448b\") " Mar 13 14:29:30 crc kubenswrapper[4907]: I0313 14:29:30.958020 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a85017ea-b984-471a-a9a8-22c02d94448b-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "a85017ea-b984-471a-a9a8-22c02d94448b" (UID: "a85017ea-b984-471a-a9a8-22c02d94448b"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:29:30 crc kubenswrapper[4907]: I0313 14:29:30.958094 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a85017ea-b984-471a-a9a8-22c02d94448b-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "a85017ea-b984-471a-a9a8-22c02d94448b" (UID: "a85017ea-b984-471a-a9a8-22c02d94448b"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:29:30 crc kubenswrapper[4907]: I0313 14:29:30.968353 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a85017ea-b984-471a-a9a8-22c02d94448b-scripts" (OuterVolumeSpecName: "scripts") pod "a85017ea-b984-471a-a9a8-22c02d94448b" (UID: "a85017ea-b984-471a-a9a8-22c02d94448b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:29:30 crc kubenswrapper[4907]: I0313 14:29:30.968495 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a85017ea-b984-471a-a9a8-22c02d94448b-kube-api-access-462xb" (OuterVolumeSpecName: "kube-api-access-462xb") pod "a85017ea-b984-471a-a9a8-22c02d94448b" (UID: "a85017ea-b984-471a-a9a8-22c02d94448b"). InnerVolumeSpecName "kube-api-access-462xb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:29:30 crc kubenswrapper[4907]: I0313 14:29:30.991140 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a85017ea-b984-471a-a9a8-22c02d94448b-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "a85017ea-b984-471a-a9a8-22c02d94448b" (UID: "a85017ea-b984-471a-a9a8-22c02d94448b"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.056308 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a85017ea-b984-471a-a9a8-22c02d94448b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a85017ea-b984-471a-a9a8-22c02d94448b" (UID: "a85017ea-b984-471a-a9a8-22c02d94448b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.059229 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-462xb\" (UniqueName: \"kubernetes.io/projected/a85017ea-b984-471a-a9a8-22c02d94448b-kube-api-access-462xb\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.059256 4907 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a85017ea-b984-471a-a9a8-22c02d94448b-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.059268 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a85017ea-b984-471a-a9a8-22c02d94448b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.059281 4907 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a85017ea-b984-471a-a9a8-22c02d94448b-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.059290 4907 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a85017ea-b984-471a-a9a8-22c02d94448b-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.059301 4907 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a85017ea-b984-471a-a9a8-22c02d94448b-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.069374 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a85017ea-b984-471a-a9a8-22c02d94448b-config-data" (OuterVolumeSpecName: "config-data") pod "a85017ea-b984-471a-a9a8-22c02d94448b" (UID: "a85017ea-b984-471a-a9a8-22c02d94448b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.162640 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a85017ea-b984-471a-a9a8-22c02d94448b-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.210432 4907 generic.go:334] "Generic (PLEG): container finished" podID="a85017ea-b984-471a-a9a8-22c02d94448b" containerID="3b55543c8226645f0d4a440e41c724886657fe5c5d9924185cd3344bad9ddc01" exitCode=0 Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.210476 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a85017ea-b984-471a-a9a8-22c02d94448b","Type":"ContainerDied","Data":"3b55543c8226645f0d4a440e41c724886657fe5c5d9924185cd3344bad9ddc01"} Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.210508 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a85017ea-b984-471a-a9a8-22c02d94448b","Type":"ContainerDied","Data":"5a88cd9ef6cae1bcbb8c65cf464b690c68ec2dfeb0b4737e783c308ca4a8ffd6"} Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.210536 4907 scope.go:117] "RemoveContainer" containerID="c2a32ce2fcf985bbc86c96669646554e0cd290d1eaa4396d46e02db24cec19b9" Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.210688 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.253555 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.257530 4907 scope.go:117] "RemoveContainer" containerID="f629b415ac0d40fdc7a871c081a5a0ceecff9578a74f62763cbbaae7d50396c7" Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.274347 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.291500 4907 scope.go:117] "RemoveContainer" containerID="3b55543c8226645f0d4a440e41c724886657fe5c5d9924185cd3344bad9ddc01" Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.293008 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 13 14:29:31 crc kubenswrapper[4907]: E0313 14:29:31.293504 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a85017ea-b984-471a-a9a8-22c02d94448b" containerName="sg-core" Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.293527 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="a85017ea-b984-471a-a9a8-22c02d94448b" containerName="sg-core" Mar 13 14:29:31 crc kubenswrapper[4907]: E0313 14:29:31.293548 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a85017ea-b984-471a-a9a8-22c02d94448b" containerName="ceilometer-notification-agent" Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.293555 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="a85017ea-b984-471a-a9a8-22c02d94448b" containerName="ceilometer-notification-agent" Mar 13 14:29:31 crc kubenswrapper[4907]: E0313 14:29:31.293572 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a85017ea-b984-471a-a9a8-22c02d94448b" containerName="ceilometer-central-agent" Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.293579 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="a85017ea-b984-471a-a9a8-22c02d94448b" containerName="ceilometer-central-agent" Mar 13 14:29:31 crc kubenswrapper[4907]: E0313 14:29:31.293601 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a85017ea-b984-471a-a9a8-22c02d94448b" containerName="proxy-httpd" Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.293607 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="a85017ea-b984-471a-a9a8-22c02d94448b" containerName="proxy-httpd" Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.293796 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="a85017ea-b984-471a-a9a8-22c02d94448b" containerName="ceilometer-notification-agent" Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.293810 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="a85017ea-b984-471a-a9a8-22c02d94448b" containerName="sg-core" Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.293817 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="a85017ea-b984-471a-a9a8-22c02d94448b" containerName="ceilometer-central-agent" Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.293844 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="a85017ea-b984-471a-a9a8-22c02d94448b" containerName="proxy-httpd" Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.295819 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.298800 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.299089 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.299347 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.303051 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.331026 4907 scope.go:117] "RemoveContainer" containerID="62988d506f50133c1068883469efb51974597559ba58796a2aae59bf02bebb3a" Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.357243 4907 scope.go:117] "RemoveContainer" containerID="c2a32ce2fcf985bbc86c96669646554e0cd290d1eaa4396d46e02db24cec19b9" Mar 13 14:29:31 crc kubenswrapper[4907]: E0313 14:29:31.373018 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c2a32ce2fcf985bbc86c96669646554e0cd290d1eaa4396d46e02db24cec19b9\": container with ID starting with c2a32ce2fcf985bbc86c96669646554e0cd290d1eaa4396d46e02db24cec19b9 not found: ID does not exist" containerID="c2a32ce2fcf985bbc86c96669646554e0cd290d1eaa4396d46e02db24cec19b9" Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.373072 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c2a32ce2fcf985bbc86c96669646554e0cd290d1eaa4396d46e02db24cec19b9"} err="failed to get container status \"c2a32ce2fcf985bbc86c96669646554e0cd290d1eaa4396d46e02db24cec19b9\": rpc error: code = NotFound desc = could not find container \"c2a32ce2fcf985bbc86c96669646554e0cd290d1eaa4396d46e02db24cec19b9\": container with ID starting with c2a32ce2fcf985bbc86c96669646554e0cd290d1eaa4396d46e02db24cec19b9 not found: ID does not exist" Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.373096 4907 scope.go:117] "RemoveContainer" containerID="f629b415ac0d40fdc7a871c081a5a0ceecff9578a74f62763cbbaae7d50396c7" Mar 13 14:29:31 crc kubenswrapper[4907]: E0313 14:29:31.373997 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f629b415ac0d40fdc7a871c081a5a0ceecff9578a74f62763cbbaae7d50396c7\": container with ID starting with f629b415ac0d40fdc7a871c081a5a0ceecff9578a74f62763cbbaae7d50396c7 not found: ID does not exist" containerID="f629b415ac0d40fdc7a871c081a5a0ceecff9578a74f62763cbbaae7d50396c7" Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.374020 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f629b415ac0d40fdc7a871c081a5a0ceecff9578a74f62763cbbaae7d50396c7"} err="failed to get container status \"f629b415ac0d40fdc7a871c081a5a0ceecff9578a74f62763cbbaae7d50396c7\": rpc error: code = NotFound desc = could not find container \"f629b415ac0d40fdc7a871c081a5a0ceecff9578a74f62763cbbaae7d50396c7\": container with ID starting with f629b415ac0d40fdc7a871c081a5a0ceecff9578a74f62763cbbaae7d50396c7 not found: ID does not exist" Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.374035 4907 scope.go:117] "RemoveContainer" containerID="3b55543c8226645f0d4a440e41c724886657fe5c5d9924185cd3344bad9ddc01" Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.374335 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/41167fb8-73ed-4487-975c-365854b0d586-scripts\") pod \"ceilometer-0\" (UID: \"41167fb8-73ed-4487-975c-365854b0d586\") " pod="openstack/ceilometer-0" Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.374381 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/41167fb8-73ed-4487-975c-365854b0d586-run-httpd\") pod \"ceilometer-0\" (UID: \"41167fb8-73ed-4487-975c-365854b0d586\") " pod="openstack/ceilometer-0" Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.374431 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/41167fb8-73ed-4487-975c-365854b0d586-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"41167fb8-73ed-4487-975c-365854b0d586\") " pod="openstack/ceilometer-0" Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.374484 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/41167fb8-73ed-4487-975c-365854b0d586-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"41167fb8-73ed-4487-975c-365854b0d586\") " pod="openstack/ceilometer-0" Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.374508 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/41167fb8-73ed-4487-975c-365854b0d586-log-httpd\") pod \"ceilometer-0\" (UID: \"41167fb8-73ed-4487-975c-365854b0d586\") " pod="openstack/ceilometer-0" Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.374550 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w4ww5\" (UniqueName: \"kubernetes.io/projected/41167fb8-73ed-4487-975c-365854b0d586-kube-api-access-w4ww5\") pod \"ceilometer-0\" (UID: \"41167fb8-73ed-4487-975c-365854b0d586\") " pod="openstack/ceilometer-0" Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.374582 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41167fb8-73ed-4487-975c-365854b0d586-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"41167fb8-73ed-4487-975c-365854b0d586\") " pod="openstack/ceilometer-0" Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.374607 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41167fb8-73ed-4487-975c-365854b0d586-config-data\") pod \"ceilometer-0\" (UID: \"41167fb8-73ed-4487-975c-365854b0d586\") " pod="openstack/ceilometer-0" Mar 13 14:29:31 crc kubenswrapper[4907]: E0313 14:29:31.379316 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3b55543c8226645f0d4a440e41c724886657fe5c5d9924185cd3344bad9ddc01\": container with ID starting with 3b55543c8226645f0d4a440e41c724886657fe5c5d9924185cd3344bad9ddc01 not found: ID does not exist" containerID="3b55543c8226645f0d4a440e41c724886657fe5c5d9924185cd3344bad9ddc01" Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.379361 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3b55543c8226645f0d4a440e41c724886657fe5c5d9924185cd3344bad9ddc01"} err="failed to get container status \"3b55543c8226645f0d4a440e41c724886657fe5c5d9924185cd3344bad9ddc01\": rpc error: code = NotFound desc = could not find container \"3b55543c8226645f0d4a440e41c724886657fe5c5d9924185cd3344bad9ddc01\": container with ID starting with 3b55543c8226645f0d4a440e41c724886657fe5c5d9924185cd3344bad9ddc01 not found: ID does not exist" Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.379390 4907 scope.go:117] "RemoveContainer" containerID="62988d506f50133c1068883469efb51974597559ba58796a2aae59bf02bebb3a" Mar 13 14:29:31 crc kubenswrapper[4907]: E0313 14:29:31.379683 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"62988d506f50133c1068883469efb51974597559ba58796a2aae59bf02bebb3a\": container with ID starting with 62988d506f50133c1068883469efb51974597559ba58796a2aae59bf02bebb3a not found: ID does not exist" containerID="62988d506f50133c1068883469efb51974597559ba58796a2aae59bf02bebb3a" Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.379707 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"62988d506f50133c1068883469efb51974597559ba58796a2aae59bf02bebb3a"} err="failed to get container status \"62988d506f50133c1068883469efb51974597559ba58796a2aae59bf02bebb3a\": rpc error: code = NotFound desc = could not find container \"62988d506f50133c1068883469efb51974597559ba58796a2aae59bf02bebb3a\": container with ID starting with 62988d506f50133c1068883469efb51974597559ba58796a2aae59bf02bebb3a not found: ID does not exist" Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.475830 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/41167fb8-73ed-4487-975c-365854b0d586-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"41167fb8-73ed-4487-975c-365854b0d586\") " pod="openstack/ceilometer-0" Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.475908 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/41167fb8-73ed-4487-975c-365854b0d586-log-httpd\") pod \"ceilometer-0\" (UID: \"41167fb8-73ed-4487-975c-365854b0d586\") " pod="openstack/ceilometer-0" Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.475956 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w4ww5\" (UniqueName: \"kubernetes.io/projected/41167fb8-73ed-4487-975c-365854b0d586-kube-api-access-w4ww5\") pod \"ceilometer-0\" (UID: \"41167fb8-73ed-4487-975c-365854b0d586\") " pod="openstack/ceilometer-0" Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.475994 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41167fb8-73ed-4487-975c-365854b0d586-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"41167fb8-73ed-4487-975c-365854b0d586\") " pod="openstack/ceilometer-0" Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.476020 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41167fb8-73ed-4487-975c-365854b0d586-config-data\") pod \"ceilometer-0\" (UID: \"41167fb8-73ed-4487-975c-365854b0d586\") " pod="openstack/ceilometer-0" Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.476074 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/41167fb8-73ed-4487-975c-365854b0d586-scripts\") pod \"ceilometer-0\" (UID: \"41167fb8-73ed-4487-975c-365854b0d586\") " pod="openstack/ceilometer-0" Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.476109 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/41167fb8-73ed-4487-975c-365854b0d586-run-httpd\") pod \"ceilometer-0\" (UID: \"41167fb8-73ed-4487-975c-365854b0d586\") " pod="openstack/ceilometer-0" Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.476165 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/41167fb8-73ed-4487-975c-365854b0d586-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"41167fb8-73ed-4487-975c-365854b0d586\") " pod="openstack/ceilometer-0" Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.484849 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/41167fb8-73ed-4487-975c-365854b0d586-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"41167fb8-73ed-4487-975c-365854b0d586\") " pod="openstack/ceilometer-0" Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.485564 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41167fb8-73ed-4487-975c-365854b0d586-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"41167fb8-73ed-4487-975c-365854b0d586\") " pod="openstack/ceilometer-0" Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.493343 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/41167fb8-73ed-4487-975c-365854b0d586-log-httpd\") pod \"ceilometer-0\" (UID: \"41167fb8-73ed-4487-975c-365854b0d586\") " pod="openstack/ceilometer-0" Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.494058 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/41167fb8-73ed-4487-975c-365854b0d586-scripts\") pod \"ceilometer-0\" (UID: \"41167fb8-73ed-4487-975c-365854b0d586\") " pod="openstack/ceilometer-0" Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.494350 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41167fb8-73ed-4487-975c-365854b0d586-config-data\") pod \"ceilometer-0\" (UID: \"41167fb8-73ed-4487-975c-365854b0d586\") " pod="openstack/ceilometer-0" Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.494738 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/41167fb8-73ed-4487-975c-365854b0d586-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"41167fb8-73ed-4487-975c-365854b0d586\") " pod="openstack/ceilometer-0" Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.498163 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/41167fb8-73ed-4487-975c-365854b0d586-run-httpd\") pod \"ceilometer-0\" (UID: \"41167fb8-73ed-4487-975c-365854b0d586\") " pod="openstack/ceilometer-0" Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.501619 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w4ww5\" (UniqueName: \"kubernetes.io/projected/41167fb8-73ed-4487-975c-365854b0d586-kube-api-access-w4ww5\") pod \"ceilometer-0\" (UID: \"41167fb8-73ed-4487-975c-365854b0d586\") " pod="openstack/ceilometer-0" Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.632057 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.806847 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a85017ea-b984-471a-a9a8-22c02d94448b" path="/var/lib/kubelet/pods/a85017ea-b984-471a-a9a8-22c02d94448b/volumes" Mar 13 14:29:32 crc kubenswrapper[4907]: I0313 14:29:32.102527 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 13 14:29:32 crc kubenswrapper[4907]: I0313 14:29:32.220501 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"41167fb8-73ed-4487-975c-365854b0d586","Type":"ContainerStarted","Data":"a73a14261a590eee8da5263fe0ee191d82682588896d368eb8d1f46e7f9984ba"} Mar 13 14:29:33 crc kubenswrapper[4907]: I0313 14:29:33.234081 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"41167fb8-73ed-4487-975c-365854b0d586","Type":"ContainerStarted","Data":"9b5ab98c8fa734605e5e5d3d4bc91ad54650c1db1f1d0d56f2cbf2225fe14db6"} Mar 13 14:29:34 crc kubenswrapper[4907]: I0313 14:29:34.244640 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"41167fb8-73ed-4487-975c-365854b0d586","Type":"ContainerStarted","Data":"5939f8118a1b021358a2a1f3def6573f42d731f7807d9451cfe9c468306f6dbc"} Mar 13 14:29:34 crc kubenswrapper[4907]: I0313 14:29:34.245023 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"41167fb8-73ed-4487-975c-365854b0d586","Type":"ContainerStarted","Data":"5ecb51a13db91baac3a9eb0c7577b30ef982b63567b8d459d38ac04dfa9a4515"} Mar 13 14:29:36 crc kubenswrapper[4907]: I0313 14:29:36.220195 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Mar 13 14:29:36 crc kubenswrapper[4907]: I0313 14:29:36.263397 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"41167fb8-73ed-4487-975c-365854b0d586","Type":"ContainerStarted","Data":"8e591aebdcd4fb21580fb8235da93dec31fd3679a3943e9171c4f1541d57c538"} Mar 13 14:29:36 crc kubenswrapper[4907]: I0313 14:29:36.263812 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Mar 13 14:29:36 crc kubenswrapper[4907]: I0313 14:29:36.266845 4907 generic.go:334] "Generic (PLEG): container finished" podID="a99d655e-f6db-4c83-a950-b935588b7df1" containerID="ab262217f9b17cb1ba1a68c95414fc65431f7517928e9e7fc8f78902f26ff961" exitCode=137 Mar 13 14:29:36 crc kubenswrapper[4907]: I0313 14:29:36.266932 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a99d655e-f6db-4c83-a950-b935588b7df1","Type":"ContainerDied","Data":"ab262217f9b17cb1ba1a68c95414fc65431f7517928e9e7fc8f78902f26ff961"} Mar 13 14:29:36 crc kubenswrapper[4907]: I0313 14:29:36.269508 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-phktq\" (UniqueName: \"kubernetes.io/projected/1915a375-806d-43bb-8749-9524db071714-kube-api-access-phktq\") pod \"1915a375-806d-43bb-8749-9524db071714\" (UID: \"1915a375-806d-43bb-8749-9524db071714\") " Mar 13 14:29:36 crc kubenswrapper[4907]: I0313 14:29:36.269543 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1915a375-806d-43bb-8749-9524db071714-combined-ca-bundle\") pod \"1915a375-806d-43bb-8749-9524db071714\" (UID: \"1915a375-806d-43bb-8749-9524db071714\") " Mar 13 14:29:36 crc kubenswrapper[4907]: I0313 14:29:36.269587 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1915a375-806d-43bb-8749-9524db071714-config-data\") pod \"1915a375-806d-43bb-8749-9524db071714\" (UID: \"1915a375-806d-43bb-8749-9524db071714\") " Mar 13 14:29:36 crc kubenswrapper[4907]: I0313 14:29:36.271171 4907 generic.go:334] "Generic (PLEG): container finished" podID="1915a375-806d-43bb-8749-9524db071714" containerID="45b20ff335a1fe9f36ec78c46577e6fb80b04fdde4e9de1eb52326676228f285" exitCode=137 Mar 13 14:29:36 crc kubenswrapper[4907]: I0313 14:29:36.271378 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"1915a375-806d-43bb-8749-9524db071714","Type":"ContainerDied","Data":"45b20ff335a1fe9f36ec78c46577e6fb80b04fdde4e9de1eb52326676228f285"} Mar 13 14:29:36 crc kubenswrapper[4907]: I0313 14:29:36.271402 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"1915a375-806d-43bb-8749-9524db071714","Type":"ContainerDied","Data":"878ef11d68a7add33ce200311e92840ebc8d43565aea3bfc5d4ea32961e387a8"} Mar 13 14:29:36 crc kubenswrapper[4907]: I0313 14:29:36.271421 4907 scope.go:117] "RemoveContainer" containerID="45b20ff335a1fe9f36ec78c46577e6fb80b04fdde4e9de1eb52326676228f285" Mar 13 14:29:36 crc kubenswrapper[4907]: I0313 14:29:36.271581 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Mar 13 14:29:36 crc kubenswrapper[4907]: I0313 14:29:36.275973 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1915a375-806d-43bb-8749-9524db071714-kube-api-access-phktq" (OuterVolumeSpecName: "kube-api-access-phktq") pod "1915a375-806d-43bb-8749-9524db071714" (UID: "1915a375-806d-43bb-8749-9524db071714"). InnerVolumeSpecName "kube-api-access-phktq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:29:36 crc kubenswrapper[4907]: I0313 14:29:36.298381 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.084994597 podStartE2EDuration="5.298343132s" podCreationTimestamp="2026-03-13 14:29:31 +0000 UTC" firstStartedPulling="2026-03-13 14:29:32.118316015 +0000 UTC m=+1471.018103704" lastFinishedPulling="2026-03-13 14:29:35.33166455 +0000 UTC m=+1474.231452239" observedRunningTime="2026-03-13 14:29:36.281383231 +0000 UTC m=+1475.181170920" watchObservedRunningTime="2026-03-13 14:29:36.298343132 +0000 UTC m=+1475.198130821" Mar 13 14:29:36 crc kubenswrapper[4907]: I0313 14:29:36.305373 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1915a375-806d-43bb-8749-9524db071714-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1915a375-806d-43bb-8749-9524db071714" (UID: "1915a375-806d-43bb-8749-9524db071714"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:29:36 crc kubenswrapper[4907]: I0313 14:29:36.306269 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1915a375-806d-43bb-8749-9524db071714-config-data" (OuterVolumeSpecName: "config-data") pod "1915a375-806d-43bb-8749-9524db071714" (UID: "1915a375-806d-43bb-8749-9524db071714"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:29:36 crc kubenswrapper[4907]: I0313 14:29:36.314933 4907 scope.go:117] "RemoveContainer" containerID="45b20ff335a1fe9f36ec78c46577e6fb80b04fdde4e9de1eb52326676228f285" Mar 13 14:29:36 crc kubenswrapper[4907]: E0313 14:29:36.315726 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"45b20ff335a1fe9f36ec78c46577e6fb80b04fdde4e9de1eb52326676228f285\": container with ID starting with 45b20ff335a1fe9f36ec78c46577e6fb80b04fdde4e9de1eb52326676228f285 not found: ID does not exist" containerID="45b20ff335a1fe9f36ec78c46577e6fb80b04fdde4e9de1eb52326676228f285" Mar 13 14:29:36 crc kubenswrapper[4907]: I0313 14:29:36.315798 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"45b20ff335a1fe9f36ec78c46577e6fb80b04fdde4e9de1eb52326676228f285"} err="failed to get container status \"45b20ff335a1fe9f36ec78c46577e6fb80b04fdde4e9de1eb52326676228f285\": rpc error: code = NotFound desc = could not find container \"45b20ff335a1fe9f36ec78c46577e6fb80b04fdde4e9de1eb52326676228f285\": container with ID starting with 45b20ff335a1fe9f36ec78c46577e6fb80b04fdde4e9de1eb52326676228f285 not found: ID does not exist" Mar 13 14:29:36 crc kubenswrapper[4907]: I0313 14:29:36.373497 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-phktq\" (UniqueName: \"kubernetes.io/projected/1915a375-806d-43bb-8749-9524db071714-kube-api-access-phktq\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:36 crc kubenswrapper[4907]: I0313 14:29:36.373550 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1915a375-806d-43bb-8749-9524db071714-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:36 crc kubenswrapper[4907]: I0313 14:29:36.373569 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1915a375-806d-43bb-8749-9524db071714-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:36 crc kubenswrapper[4907]: I0313 14:29:36.540983 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Mar 13 14:29:36 crc kubenswrapper[4907]: I0313 14:29:36.652840 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 13 14:29:36 crc kubenswrapper[4907]: I0313 14:29:36.672518 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 13 14:29:36 crc kubenswrapper[4907]: I0313 14:29:36.682001 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 13 14:29:36 crc kubenswrapper[4907]: E0313 14:29:36.682319 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1915a375-806d-43bb-8749-9524db071714" containerName="nova-cell1-novncproxy-novncproxy" Mar 13 14:29:36 crc kubenswrapper[4907]: I0313 14:29:36.682330 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="1915a375-806d-43bb-8749-9524db071714" containerName="nova-cell1-novncproxy-novncproxy" Mar 13 14:29:36 crc kubenswrapper[4907]: I0313 14:29:36.682527 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="1915a375-806d-43bb-8749-9524db071714" containerName="nova-cell1-novncproxy-novncproxy" Mar 13 14:29:36 crc kubenswrapper[4907]: I0313 14:29:36.683148 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Mar 13 14:29:36 crc kubenswrapper[4907]: I0313 14:29:36.690703 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Mar 13 14:29:36 crc kubenswrapper[4907]: I0313 14:29:36.691920 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Mar 13 14:29:36 crc kubenswrapper[4907]: I0313 14:29:36.693866 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Mar 13 14:29:36 crc kubenswrapper[4907]: I0313 14:29:36.713347 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 13 14:29:36 crc kubenswrapper[4907]: I0313 14:29:36.728427 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 13 14:29:36 crc kubenswrapper[4907]: I0313 14:29:36.885279 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v4n9s\" (UniqueName: \"kubernetes.io/projected/a99d655e-f6db-4c83-a950-b935588b7df1-kube-api-access-v4n9s\") pod \"a99d655e-f6db-4c83-a950-b935588b7df1\" (UID: \"a99d655e-f6db-4c83-a950-b935588b7df1\") " Mar 13 14:29:36 crc kubenswrapper[4907]: I0313 14:29:36.885749 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a99d655e-f6db-4c83-a950-b935588b7df1-config-data\") pod \"a99d655e-f6db-4c83-a950-b935588b7df1\" (UID: \"a99d655e-f6db-4c83-a950-b935588b7df1\") " Mar 13 14:29:36 crc kubenswrapper[4907]: I0313 14:29:36.886773 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a99d655e-f6db-4c83-a950-b935588b7df1-logs\") pod \"a99d655e-f6db-4c83-a950-b935588b7df1\" (UID: \"a99d655e-f6db-4c83-a950-b935588b7df1\") " Mar 13 14:29:36 crc kubenswrapper[4907]: I0313 14:29:36.886932 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a99d655e-f6db-4c83-a950-b935588b7df1-combined-ca-bundle\") pod \"a99d655e-f6db-4c83-a950-b935588b7df1\" (UID: \"a99d655e-f6db-4c83-a950-b935588b7df1\") " Mar 13 14:29:36 crc kubenswrapper[4907]: I0313 14:29:36.887124 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a99d655e-f6db-4c83-a950-b935588b7df1-logs" (OuterVolumeSpecName: "logs") pod "a99d655e-f6db-4c83-a950-b935588b7df1" (UID: "a99d655e-f6db-4c83-a950-b935588b7df1"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:29:36 crc kubenswrapper[4907]: I0313 14:29:36.888036 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62983e04-f043-4ad3-9a46-96179d59671e-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"62983e04-f043-4ad3-9a46-96179d59671e\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 14:29:36 crc kubenswrapper[4907]: I0313 14:29:36.888139 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6kqzq\" (UniqueName: \"kubernetes.io/projected/62983e04-f043-4ad3-9a46-96179d59671e-kube-api-access-6kqzq\") pod \"nova-cell1-novncproxy-0\" (UID: \"62983e04-f043-4ad3-9a46-96179d59671e\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 14:29:36 crc kubenswrapper[4907]: I0313 14:29:36.888218 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62983e04-f043-4ad3-9a46-96179d59671e-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"62983e04-f043-4ad3-9a46-96179d59671e\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 14:29:36 crc kubenswrapper[4907]: I0313 14:29:36.888324 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/62983e04-f043-4ad3-9a46-96179d59671e-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"62983e04-f043-4ad3-9a46-96179d59671e\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 14:29:36 crc kubenswrapper[4907]: I0313 14:29:36.888353 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/62983e04-f043-4ad3-9a46-96179d59671e-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"62983e04-f043-4ad3-9a46-96179d59671e\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 14:29:36 crc kubenswrapper[4907]: I0313 14:29:36.888656 4907 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a99d655e-f6db-4c83-a950-b935588b7df1-logs\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:36 crc kubenswrapper[4907]: I0313 14:29:36.891760 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a99d655e-f6db-4c83-a950-b935588b7df1-kube-api-access-v4n9s" (OuterVolumeSpecName: "kube-api-access-v4n9s") pod "a99d655e-f6db-4c83-a950-b935588b7df1" (UID: "a99d655e-f6db-4c83-a950-b935588b7df1"). InnerVolumeSpecName "kube-api-access-v4n9s". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:29:36 crc kubenswrapper[4907]: I0313 14:29:36.922229 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a99d655e-f6db-4c83-a950-b935588b7df1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a99d655e-f6db-4c83-a950-b935588b7df1" (UID: "a99d655e-f6db-4c83-a950-b935588b7df1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:29:36 crc kubenswrapper[4907]: I0313 14:29:36.954135 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a99d655e-f6db-4c83-a950-b935588b7df1-config-data" (OuterVolumeSpecName: "config-data") pod "a99d655e-f6db-4c83-a950-b935588b7df1" (UID: "a99d655e-f6db-4c83-a950-b935588b7df1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:29:36 crc kubenswrapper[4907]: I0313 14:29:36.990161 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62983e04-f043-4ad3-9a46-96179d59671e-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"62983e04-f043-4ad3-9a46-96179d59671e\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 14:29:36 crc kubenswrapper[4907]: I0313 14:29:36.990251 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6kqzq\" (UniqueName: \"kubernetes.io/projected/62983e04-f043-4ad3-9a46-96179d59671e-kube-api-access-6kqzq\") pod \"nova-cell1-novncproxy-0\" (UID: \"62983e04-f043-4ad3-9a46-96179d59671e\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 14:29:36 crc kubenswrapper[4907]: I0313 14:29:36.990316 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62983e04-f043-4ad3-9a46-96179d59671e-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"62983e04-f043-4ad3-9a46-96179d59671e\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 14:29:36 crc kubenswrapper[4907]: I0313 14:29:36.990375 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/62983e04-f043-4ad3-9a46-96179d59671e-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"62983e04-f043-4ad3-9a46-96179d59671e\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 14:29:36 crc kubenswrapper[4907]: I0313 14:29:36.990396 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/62983e04-f043-4ad3-9a46-96179d59671e-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"62983e04-f043-4ad3-9a46-96179d59671e\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 14:29:36 crc kubenswrapper[4907]: I0313 14:29:36.990502 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a99d655e-f6db-4c83-a950-b935588b7df1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:36 crc kubenswrapper[4907]: I0313 14:29:36.990519 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v4n9s\" (UniqueName: \"kubernetes.io/projected/a99d655e-f6db-4c83-a950-b935588b7df1-kube-api-access-v4n9s\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:36 crc kubenswrapper[4907]: I0313 14:29:36.990532 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a99d655e-f6db-4c83-a950-b935588b7df1-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:36 crc kubenswrapper[4907]: I0313 14:29:36.993564 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/62983e04-f043-4ad3-9a46-96179d59671e-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"62983e04-f043-4ad3-9a46-96179d59671e\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 14:29:36 crc kubenswrapper[4907]: I0313 14:29:36.993599 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/62983e04-f043-4ad3-9a46-96179d59671e-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"62983e04-f043-4ad3-9a46-96179d59671e\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 14:29:36 crc kubenswrapper[4907]: I0313 14:29:36.994984 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62983e04-f043-4ad3-9a46-96179d59671e-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"62983e04-f043-4ad3-9a46-96179d59671e\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 14:29:37 crc kubenswrapper[4907]: I0313 14:29:37.001428 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62983e04-f043-4ad3-9a46-96179d59671e-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"62983e04-f043-4ad3-9a46-96179d59671e\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 14:29:37 crc kubenswrapper[4907]: I0313 14:29:37.007029 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6kqzq\" (UniqueName: \"kubernetes.io/projected/62983e04-f043-4ad3-9a46-96179d59671e-kube-api-access-6kqzq\") pod \"nova-cell1-novncproxy-0\" (UID: \"62983e04-f043-4ad3-9a46-96179d59671e\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 14:29:37 crc kubenswrapper[4907]: I0313 14:29:37.039486 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Mar 13 14:29:37 crc kubenswrapper[4907]: I0313 14:29:37.284309 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a99d655e-f6db-4c83-a950-b935588b7df1","Type":"ContainerDied","Data":"f5827c9400a454954bce2169d1bdd16e49d4ec2488819960057446785871a720"} Mar 13 14:29:37 crc kubenswrapper[4907]: I0313 14:29:37.284582 4907 scope.go:117] "RemoveContainer" containerID="ab262217f9b17cb1ba1a68c95414fc65431f7517928e9e7fc8f78902f26ff961" Mar 13 14:29:37 crc kubenswrapper[4907]: I0313 14:29:37.284711 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 13 14:29:37 crc kubenswrapper[4907]: I0313 14:29:37.328209 4907 scope.go:117] "RemoveContainer" containerID="c48900e9ab50388f765d2b50bcf7cb048feb1b78e5669c08c203f15509de3dac" Mar 13 14:29:37 crc kubenswrapper[4907]: I0313 14:29:37.332058 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Mar 13 14:29:37 crc kubenswrapper[4907]: I0313 14:29:37.347005 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Mar 13 14:29:37 crc kubenswrapper[4907]: I0313 14:29:37.358476 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Mar 13 14:29:37 crc kubenswrapper[4907]: E0313 14:29:37.358986 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a99d655e-f6db-4c83-a950-b935588b7df1" containerName="nova-metadata-log" Mar 13 14:29:37 crc kubenswrapper[4907]: I0313 14:29:37.359010 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="a99d655e-f6db-4c83-a950-b935588b7df1" containerName="nova-metadata-log" Mar 13 14:29:37 crc kubenswrapper[4907]: E0313 14:29:37.359055 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a99d655e-f6db-4c83-a950-b935588b7df1" containerName="nova-metadata-metadata" Mar 13 14:29:37 crc kubenswrapper[4907]: I0313 14:29:37.359063 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="a99d655e-f6db-4c83-a950-b935588b7df1" containerName="nova-metadata-metadata" Mar 13 14:29:37 crc kubenswrapper[4907]: I0313 14:29:37.359256 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="a99d655e-f6db-4c83-a950-b935588b7df1" containerName="nova-metadata-log" Mar 13 14:29:37 crc kubenswrapper[4907]: I0313 14:29:37.359287 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="a99d655e-f6db-4c83-a950-b935588b7df1" containerName="nova-metadata-metadata" Mar 13 14:29:37 crc kubenswrapper[4907]: I0313 14:29:37.360442 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 13 14:29:37 crc kubenswrapper[4907]: I0313 14:29:37.363327 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Mar 13 14:29:37 crc kubenswrapper[4907]: I0313 14:29:37.363526 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Mar 13 14:29:37 crc kubenswrapper[4907]: I0313 14:29:37.363769 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Mar 13 14:29:37 crc kubenswrapper[4907]: I0313 14:29:37.397792 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39488342-aa1a-4dba-bd5f-556da5081904-config-data\") pod \"nova-metadata-0\" (UID: \"39488342-aa1a-4dba-bd5f-556da5081904\") " pod="openstack/nova-metadata-0" Mar 13 14:29:37 crc kubenswrapper[4907]: I0313 14:29:37.397954 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/39488342-aa1a-4dba-bd5f-556da5081904-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"39488342-aa1a-4dba-bd5f-556da5081904\") " pod="openstack/nova-metadata-0" Mar 13 14:29:37 crc kubenswrapper[4907]: I0313 14:29:37.397981 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39488342-aa1a-4dba-bd5f-556da5081904-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"39488342-aa1a-4dba-bd5f-556da5081904\") " pod="openstack/nova-metadata-0" Mar 13 14:29:37 crc kubenswrapper[4907]: I0313 14:29:37.398067 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/39488342-aa1a-4dba-bd5f-556da5081904-logs\") pod \"nova-metadata-0\" (UID: \"39488342-aa1a-4dba-bd5f-556da5081904\") " pod="openstack/nova-metadata-0" Mar 13 14:29:37 crc kubenswrapper[4907]: I0313 14:29:37.398111 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-774w9\" (UniqueName: \"kubernetes.io/projected/39488342-aa1a-4dba-bd5f-556da5081904-kube-api-access-774w9\") pod \"nova-metadata-0\" (UID: \"39488342-aa1a-4dba-bd5f-556da5081904\") " pod="openstack/nova-metadata-0" Mar 13 14:29:37 crc kubenswrapper[4907]: I0313 14:29:37.474908 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Mar 13 14:29:37 crc kubenswrapper[4907]: I0313 14:29:37.475909 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Mar 13 14:29:37 crc kubenswrapper[4907]: I0313 14:29:37.481060 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 13 14:29:37 crc kubenswrapper[4907]: I0313 14:29:37.501161 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/39488342-aa1a-4dba-bd5f-556da5081904-logs\") pod \"nova-metadata-0\" (UID: \"39488342-aa1a-4dba-bd5f-556da5081904\") " pod="openstack/nova-metadata-0" Mar 13 14:29:37 crc kubenswrapper[4907]: I0313 14:29:37.501258 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-774w9\" (UniqueName: \"kubernetes.io/projected/39488342-aa1a-4dba-bd5f-556da5081904-kube-api-access-774w9\") pod \"nova-metadata-0\" (UID: \"39488342-aa1a-4dba-bd5f-556da5081904\") " pod="openstack/nova-metadata-0" Mar 13 14:29:37 crc kubenswrapper[4907]: I0313 14:29:37.501331 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39488342-aa1a-4dba-bd5f-556da5081904-config-data\") pod \"nova-metadata-0\" (UID: \"39488342-aa1a-4dba-bd5f-556da5081904\") " pod="openstack/nova-metadata-0" Mar 13 14:29:37 crc kubenswrapper[4907]: I0313 14:29:37.501407 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/39488342-aa1a-4dba-bd5f-556da5081904-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"39488342-aa1a-4dba-bd5f-556da5081904\") " pod="openstack/nova-metadata-0" Mar 13 14:29:37 crc kubenswrapper[4907]: I0313 14:29:37.501433 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39488342-aa1a-4dba-bd5f-556da5081904-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"39488342-aa1a-4dba-bd5f-556da5081904\") " pod="openstack/nova-metadata-0" Mar 13 14:29:37 crc kubenswrapper[4907]: I0313 14:29:37.501611 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/39488342-aa1a-4dba-bd5f-556da5081904-logs\") pod \"nova-metadata-0\" (UID: \"39488342-aa1a-4dba-bd5f-556da5081904\") " pod="openstack/nova-metadata-0" Mar 13 14:29:37 crc kubenswrapper[4907]: I0313 14:29:37.514862 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/39488342-aa1a-4dba-bd5f-556da5081904-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"39488342-aa1a-4dba-bd5f-556da5081904\") " pod="openstack/nova-metadata-0" Mar 13 14:29:37 crc kubenswrapper[4907]: I0313 14:29:37.514980 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39488342-aa1a-4dba-bd5f-556da5081904-config-data\") pod \"nova-metadata-0\" (UID: \"39488342-aa1a-4dba-bd5f-556da5081904\") " pod="openstack/nova-metadata-0" Mar 13 14:29:37 crc kubenswrapper[4907]: I0313 14:29:37.515000 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39488342-aa1a-4dba-bd5f-556da5081904-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"39488342-aa1a-4dba-bd5f-556da5081904\") " pod="openstack/nova-metadata-0" Mar 13 14:29:37 crc kubenswrapper[4907]: I0313 14:29:37.524051 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-774w9\" (UniqueName: \"kubernetes.io/projected/39488342-aa1a-4dba-bd5f-556da5081904-kube-api-access-774w9\") pod \"nova-metadata-0\" (UID: \"39488342-aa1a-4dba-bd5f-556da5081904\") " pod="openstack/nova-metadata-0" Mar 13 14:29:37 crc kubenswrapper[4907]: I0313 14:29:37.718732 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 13 14:29:37 crc kubenswrapper[4907]: I0313 14:29:37.795275 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1915a375-806d-43bb-8749-9524db071714" path="/var/lib/kubelet/pods/1915a375-806d-43bb-8749-9524db071714/volumes" Mar 13 14:29:37 crc kubenswrapper[4907]: I0313 14:29:37.795971 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a99d655e-f6db-4c83-a950-b935588b7df1" path="/var/lib/kubelet/pods/a99d655e-f6db-4c83-a950-b935588b7df1/volumes" Mar 13 14:29:38 crc kubenswrapper[4907]: I0313 14:29:38.184626 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Mar 13 14:29:38 crc kubenswrapper[4907]: W0313 14:29:38.186281 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod39488342_aa1a_4dba_bd5f_556da5081904.slice/crio-a7aca03c6cc93dfbda04afd1635961a9edebe57ed61e75bfcd08d8cd3550955a WatchSource:0}: Error finding container a7aca03c6cc93dfbda04afd1635961a9edebe57ed61e75bfcd08d8cd3550955a: Status 404 returned error can't find the container with id a7aca03c6cc93dfbda04afd1635961a9edebe57ed61e75bfcd08d8cd3550955a Mar 13 14:29:38 crc kubenswrapper[4907]: I0313 14:29:38.306240 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"62983e04-f043-4ad3-9a46-96179d59671e","Type":"ContainerStarted","Data":"9e2de4df42a807d1995b769c3c69187ddaaf7da1c1d09354b07088fb415f36d7"} Mar 13 14:29:38 crc kubenswrapper[4907]: I0313 14:29:38.306279 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"62983e04-f043-4ad3-9a46-96179d59671e","Type":"ContainerStarted","Data":"1cd31d9d43a5e6a87bb0836d84499cc5b9fab420b814ef48408f1a9549204fd0"} Mar 13 14:29:38 crc kubenswrapper[4907]: I0313 14:29:38.310485 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"39488342-aa1a-4dba-bd5f-556da5081904","Type":"ContainerStarted","Data":"a7aca03c6cc93dfbda04afd1635961a9edebe57ed61e75bfcd08d8cd3550955a"} Mar 13 14:29:39 crc kubenswrapper[4907]: I0313 14:29:39.324103 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"39488342-aa1a-4dba-bd5f-556da5081904","Type":"ContainerStarted","Data":"3f0638d25e0be962bf528749f86e7ed7d982188df63f11eef31f67ffdeb4e143"} Mar 13 14:29:39 crc kubenswrapper[4907]: I0313 14:29:39.324464 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"39488342-aa1a-4dba-bd5f-556da5081904","Type":"ContainerStarted","Data":"ee178515d58e37a284bcf6959726bc0672f424f0bddc3d61b496fcc0bf679507"} Mar 13 14:29:39 crc kubenswrapper[4907]: I0313 14:29:39.344710 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=3.344687712 podStartE2EDuration="3.344687712s" podCreationTimestamp="2026-03-13 14:29:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:29:38.330523681 +0000 UTC m=+1477.230311380" watchObservedRunningTime="2026-03-13 14:29:39.344687712 +0000 UTC m=+1478.244475401" Mar 13 14:29:39 crc kubenswrapper[4907]: I0313 14:29:39.478070 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Mar 13 14:29:39 crc kubenswrapper[4907]: I0313 14:29:39.483026 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Mar 13 14:29:39 crc kubenswrapper[4907]: I0313 14:29:39.484367 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Mar 13 14:29:39 crc kubenswrapper[4907]: I0313 14:29:39.508975 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.508954433 podStartE2EDuration="2.508954433s" podCreationTimestamp="2026-03-13 14:29:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:29:39.342893894 +0000 UTC m=+1478.242681593" watchObservedRunningTime="2026-03-13 14:29:39.508954433 +0000 UTC m=+1478.408742132" Mar 13 14:29:40 crc kubenswrapper[4907]: I0313 14:29:40.334228 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Mar 13 14:29:40 crc kubenswrapper[4907]: I0313 14:29:40.512691 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-fdb8f6449-nb4km"] Mar 13 14:29:40 crc kubenswrapper[4907]: I0313 14:29:40.514306 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-fdb8f6449-nb4km" Mar 13 14:29:40 crc kubenswrapper[4907]: I0313 14:29:40.541338 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-fdb8f6449-nb4km"] Mar 13 14:29:40 crc kubenswrapper[4907]: I0313 14:29:40.670188 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7da13722-732a-4b65-a894-fe4612f30d75-dns-svc\") pod \"dnsmasq-dns-fdb8f6449-nb4km\" (UID: \"7da13722-732a-4b65-a894-fe4612f30d75\") " pod="openstack/dnsmasq-dns-fdb8f6449-nb4km" Mar 13 14:29:40 crc kubenswrapper[4907]: I0313 14:29:40.670262 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7da13722-732a-4b65-a894-fe4612f30d75-ovsdbserver-sb\") pod \"dnsmasq-dns-fdb8f6449-nb4km\" (UID: \"7da13722-732a-4b65-a894-fe4612f30d75\") " pod="openstack/dnsmasq-dns-fdb8f6449-nb4km" Mar 13 14:29:40 crc kubenswrapper[4907]: I0313 14:29:40.670402 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g2kbb\" (UniqueName: \"kubernetes.io/projected/7da13722-732a-4b65-a894-fe4612f30d75-kube-api-access-g2kbb\") pod \"dnsmasq-dns-fdb8f6449-nb4km\" (UID: \"7da13722-732a-4b65-a894-fe4612f30d75\") " pod="openstack/dnsmasq-dns-fdb8f6449-nb4km" Mar 13 14:29:40 crc kubenswrapper[4907]: I0313 14:29:40.670500 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7da13722-732a-4b65-a894-fe4612f30d75-dns-swift-storage-0\") pod \"dnsmasq-dns-fdb8f6449-nb4km\" (UID: \"7da13722-732a-4b65-a894-fe4612f30d75\") " pod="openstack/dnsmasq-dns-fdb8f6449-nb4km" Mar 13 14:29:40 crc kubenswrapper[4907]: I0313 14:29:40.670597 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7da13722-732a-4b65-a894-fe4612f30d75-config\") pod \"dnsmasq-dns-fdb8f6449-nb4km\" (UID: \"7da13722-732a-4b65-a894-fe4612f30d75\") " pod="openstack/dnsmasq-dns-fdb8f6449-nb4km" Mar 13 14:29:40 crc kubenswrapper[4907]: I0313 14:29:40.670722 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7da13722-732a-4b65-a894-fe4612f30d75-ovsdbserver-nb\") pod \"dnsmasq-dns-fdb8f6449-nb4km\" (UID: \"7da13722-732a-4b65-a894-fe4612f30d75\") " pod="openstack/dnsmasq-dns-fdb8f6449-nb4km" Mar 13 14:29:40 crc kubenswrapper[4907]: I0313 14:29:40.772569 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7da13722-732a-4b65-a894-fe4612f30d75-ovsdbserver-nb\") pod \"dnsmasq-dns-fdb8f6449-nb4km\" (UID: \"7da13722-732a-4b65-a894-fe4612f30d75\") " pod="openstack/dnsmasq-dns-fdb8f6449-nb4km" Mar 13 14:29:40 crc kubenswrapper[4907]: I0313 14:29:40.772712 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7da13722-732a-4b65-a894-fe4612f30d75-dns-svc\") pod \"dnsmasq-dns-fdb8f6449-nb4km\" (UID: \"7da13722-732a-4b65-a894-fe4612f30d75\") " pod="openstack/dnsmasq-dns-fdb8f6449-nb4km" Mar 13 14:29:40 crc kubenswrapper[4907]: I0313 14:29:40.772760 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7da13722-732a-4b65-a894-fe4612f30d75-ovsdbserver-sb\") pod \"dnsmasq-dns-fdb8f6449-nb4km\" (UID: \"7da13722-732a-4b65-a894-fe4612f30d75\") " pod="openstack/dnsmasq-dns-fdb8f6449-nb4km" Mar 13 14:29:40 crc kubenswrapper[4907]: I0313 14:29:40.772805 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g2kbb\" (UniqueName: \"kubernetes.io/projected/7da13722-732a-4b65-a894-fe4612f30d75-kube-api-access-g2kbb\") pod \"dnsmasq-dns-fdb8f6449-nb4km\" (UID: \"7da13722-732a-4b65-a894-fe4612f30d75\") " pod="openstack/dnsmasq-dns-fdb8f6449-nb4km" Mar 13 14:29:40 crc kubenswrapper[4907]: I0313 14:29:40.772841 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7da13722-732a-4b65-a894-fe4612f30d75-dns-swift-storage-0\") pod \"dnsmasq-dns-fdb8f6449-nb4km\" (UID: \"7da13722-732a-4b65-a894-fe4612f30d75\") " pod="openstack/dnsmasq-dns-fdb8f6449-nb4km" Mar 13 14:29:40 crc kubenswrapper[4907]: I0313 14:29:40.772907 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7da13722-732a-4b65-a894-fe4612f30d75-config\") pod \"dnsmasq-dns-fdb8f6449-nb4km\" (UID: \"7da13722-732a-4b65-a894-fe4612f30d75\") " pod="openstack/dnsmasq-dns-fdb8f6449-nb4km" Mar 13 14:29:40 crc kubenswrapper[4907]: I0313 14:29:40.773873 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7da13722-732a-4b65-a894-fe4612f30d75-ovsdbserver-sb\") pod \"dnsmasq-dns-fdb8f6449-nb4km\" (UID: \"7da13722-732a-4b65-a894-fe4612f30d75\") " pod="openstack/dnsmasq-dns-fdb8f6449-nb4km" Mar 13 14:29:40 crc kubenswrapper[4907]: I0313 14:29:40.774029 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7da13722-732a-4b65-a894-fe4612f30d75-config\") pod \"dnsmasq-dns-fdb8f6449-nb4km\" (UID: \"7da13722-732a-4b65-a894-fe4612f30d75\") " pod="openstack/dnsmasq-dns-fdb8f6449-nb4km" Mar 13 14:29:40 crc kubenswrapper[4907]: I0313 14:29:40.774484 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7da13722-732a-4b65-a894-fe4612f30d75-ovsdbserver-nb\") pod \"dnsmasq-dns-fdb8f6449-nb4km\" (UID: \"7da13722-732a-4b65-a894-fe4612f30d75\") " pod="openstack/dnsmasq-dns-fdb8f6449-nb4km" Mar 13 14:29:40 crc kubenswrapper[4907]: I0313 14:29:40.775139 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7da13722-732a-4b65-a894-fe4612f30d75-dns-swift-storage-0\") pod \"dnsmasq-dns-fdb8f6449-nb4km\" (UID: \"7da13722-732a-4b65-a894-fe4612f30d75\") " pod="openstack/dnsmasq-dns-fdb8f6449-nb4km" Mar 13 14:29:40 crc kubenswrapper[4907]: I0313 14:29:40.775155 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7da13722-732a-4b65-a894-fe4612f30d75-dns-svc\") pod \"dnsmasq-dns-fdb8f6449-nb4km\" (UID: \"7da13722-732a-4b65-a894-fe4612f30d75\") " pod="openstack/dnsmasq-dns-fdb8f6449-nb4km" Mar 13 14:29:40 crc kubenswrapper[4907]: I0313 14:29:40.794743 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g2kbb\" (UniqueName: \"kubernetes.io/projected/7da13722-732a-4b65-a894-fe4612f30d75-kube-api-access-g2kbb\") pod \"dnsmasq-dns-fdb8f6449-nb4km\" (UID: \"7da13722-732a-4b65-a894-fe4612f30d75\") " pod="openstack/dnsmasq-dns-fdb8f6449-nb4km" Mar 13 14:29:40 crc kubenswrapper[4907]: I0313 14:29:40.839870 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-fdb8f6449-nb4km" Mar 13 14:29:41 crc kubenswrapper[4907]: W0313 14:29:41.998040 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7da13722_732a_4b65_a894_fe4612f30d75.slice/crio-fbaac4295e0e2ca7122793d1f083ea125d689fae676b037de247b0fa68fc7929 WatchSource:0}: Error finding container fbaac4295e0e2ca7122793d1f083ea125d689fae676b037de247b0fa68fc7929: Status 404 returned error can't find the container with id fbaac4295e0e2ca7122793d1f083ea125d689fae676b037de247b0fa68fc7929 Mar 13 14:29:41 crc kubenswrapper[4907]: I0313 14:29:41.999305 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-fdb8f6449-nb4km"] Mar 13 14:29:42 crc kubenswrapper[4907]: I0313 14:29:42.040206 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Mar 13 14:29:42 crc kubenswrapper[4907]: I0313 14:29:42.354629 4907 generic.go:334] "Generic (PLEG): container finished" podID="7da13722-732a-4b65-a894-fe4612f30d75" containerID="9c33ada0f7c03cde714537a63faadeb180f6ac07dd15e3f4ce8c7087574733bb" exitCode=0 Mar 13 14:29:42 crc kubenswrapper[4907]: I0313 14:29:42.354696 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-fdb8f6449-nb4km" event={"ID":"7da13722-732a-4b65-a894-fe4612f30d75","Type":"ContainerDied","Data":"9c33ada0f7c03cde714537a63faadeb180f6ac07dd15e3f4ce8c7087574733bb"} Mar 13 14:29:42 crc kubenswrapper[4907]: I0313 14:29:42.354774 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-fdb8f6449-nb4km" event={"ID":"7da13722-732a-4b65-a894-fe4612f30d75","Type":"ContainerStarted","Data":"fbaac4295e0e2ca7122793d1f083ea125d689fae676b037de247b0fa68fc7929"} Mar 13 14:29:43 crc kubenswrapper[4907]: I0313 14:29:43.196788 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 13 14:29:43 crc kubenswrapper[4907]: I0313 14:29:43.197337 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="41167fb8-73ed-4487-975c-365854b0d586" containerName="ceilometer-central-agent" containerID="cri-o://9b5ab98c8fa734605e5e5d3d4bc91ad54650c1db1f1d0d56f2cbf2225fe14db6" gracePeriod=30 Mar 13 14:29:43 crc kubenswrapper[4907]: I0313 14:29:43.197401 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="41167fb8-73ed-4487-975c-365854b0d586" containerName="proxy-httpd" containerID="cri-o://8e591aebdcd4fb21580fb8235da93dec31fd3679a3943e9171c4f1541d57c538" gracePeriod=30 Mar 13 14:29:43 crc kubenswrapper[4907]: I0313 14:29:43.197457 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="41167fb8-73ed-4487-975c-365854b0d586" containerName="sg-core" containerID="cri-o://5939f8118a1b021358a2a1f3def6573f42d731f7807d9451cfe9c468306f6dbc" gracePeriod=30 Mar 13 14:29:43 crc kubenswrapper[4907]: I0313 14:29:43.197496 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="41167fb8-73ed-4487-975c-365854b0d586" containerName="ceilometer-notification-agent" containerID="cri-o://5ecb51a13db91baac3a9eb0c7577b30ef982b63567b8d459d38ac04dfa9a4515" gracePeriod=30 Mar 13 14:29:43 crc kubenswrapper[4907]: I0313 14:29:43.367698 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-fdb8f6449-nb4km" event={"ID":"7da13722-732a-4b65-a894-fe4612f30d75","Type":"ContainerStarted","Data":"890693bd8cb16d526cbd4e7c94b7a9b8dcd67a1bc48dbb881c81f86b4e9727d1"} Mar 13 14:29:43 crc kubenswrapper[4907]: I0313 14:29:43.368544 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-fdb8f6449-nb4km" Mar 13 14:29:43 crc kubenswrapper[4907]: I0313 14:29:43.375954 4907 generic.go:334] "Generic (PLEG): container finished" podID="41167fb8-73ed-4487-975c-365854b0d586" containerID="8e591aebdcd4fb21580fb8235da93dec31fd3679a3943e9171c4f1541d57c538" exitCode=0 Mar 13 14:29:43 crc kubenswrapper[4907]: I0313 14:29:43.375987 4907 generic.go:334] "Generic (PLEG): container finished" podID="41167fb8-73ed-4487-975c-365854b0d586" containerID="5939f8118a1b021358a2a1f3def6573f42d731f7807d9451cfe9c468306f6dbc" exitCode=2 Mar 13 14:29:43 crc kubenswrapper[4907]: I0313 14:29:43.376010 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"41167fb8-73ed-4487-975c-365854b0d586","Type":"ContainerDied","Data":"8e591aebdcd4fb21580fb8235da93dec31fd3679a3943e9171c4f1541d57c538"} Mar 13 14:29:43 crc kubenswrapper[4907]: I0313 14:29:43.376033 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"41167fb8-73ed-4487-975c-365854b0d586","Type":"ContainerDied","Data":"5939f8118a1b021358a2a1f3def6573f42d731f7807d9451cfe9c468306f6dbc"} Mar 13 14:29:43 crc kubenswrapper[4907]: I0313 14:29:43.393580 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-fdb8f6449-nb4km" podStartSLOduration=3.393558399 podStartE2EDuration="3.393558399s" podCreationTimestamp="2026-03-13 14:29:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:29:43.386312332 +0000 UTC m=+1482.286100021" watchObservedRunningTime="2026-03-13 14:29:43.393558399 +0000 UTC m=+1482.293346088" Mar 13 14:29:43 crc kubenswrapper[4907]: I0313 14:29:43.443715 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Mar 13 14:29:43 crc kubenswrapper[4907]: I0313 14:29:43.444038 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="41063592-cda5-443f-87e3-20c2c6ec5c73" containerName="nova-api-log" containerID="cri-o://97742dae0519818b9c495c23cd7ee1fe57412023b1fc525ca01e7fc2236cef2b" gracePeriod=30 Mar 13 14:29:43 crc kubenswrapper[4907]: I0313 14:29:43.444122 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="41063592-cda5-443f-87e3-20c2c6ec5c73" containerName="nova-api-api" containerID="cri-o://bee51476e0503a2fc3a96ca20f1d42b6cb92e1822245b7b21033273b10f0e7a6" gracePeriod=30 Mar 13 14:29:44 crc kubenswrapper[4907]: I0313 14:29:44.391567 4907 generic.go:334] "Generic (PLEG): container finished" podID="41063592-cda5-443f-87e3-20c2c6ec5c73" containerID="97742dae0519818b9c495c23cd7ee1fe57412023b1fc525ca01e7fc2236cef2b" exitCode=143 Mar 13 14:29:44 crc kubenswrapper[4907]: I0313 14:29:44.391936 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"41063592-cda5-443f-87e3-20c2c6ec5c73","Type":"ContainerDied","Data":"97742dae0519818b9c495c23cd7ee1fe57412023b1fc525ca01e7fc2236cef2b"} Mar 13 14:29:44 crc kubenswrapper[4907]: I0313 14:29:44.394717 4907 generic.go:334] "Generic (PLEG): container finished" podID="41167fb8-73ed-4487-975c-365854b0d586" containerID="5ecb51a13db91baac3a9eb0c7577b30ef982b63567b8d459d38ac04dfa9a4515" exitCode=0 Mar 13 14:29:44 crc kubenswrapper[4907]: I0313 14:29:44.394741 4907 generic.go:334] "Generic (PLEG): container finished" podID="41167fb8-73ed-4487-975c-365854b0d586" containerID="9b5ab98c8fa734605e5e5d3d4bc91ad54650c1db1f1d0d56f2cbf2225fe14db6" exitCode=0 Mar 13 14:29:44 crc kubenswrapper[4907]: I0313 14:29:44.395676 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"41167fb8-73ed-4487-975c-365854b0d586","Type":"ContainerDied","Data":"5ecb51a13db91baac3a9eb0c7577b30ef982b63567b8d459d38ac04dfa9a4515"} Mar 13 14:29:44 crc kubenswrapper[4907]: I0313 14:29:44.395707 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"41167fb8-73ed-4487-975c-365854b0d586","Type":"ContainerDied","Data":"9b5ab98c8fa734605e5e5d3d4bc91ad54650c1db1f1d0d56f2cbf2225fe14db6"} Mar 13 14:29:44 crc kubenswrapper[4907]: I0313 14:29:44.632149 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 14:29:44 crc kubenswrapper[4907]: I0313 14:29:44.759017 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/41167fb8-73ed-4487-975c-365854b0d586-sg-core-conf-yaml\") pod \"41167fb8-73ed-4487-975c-365854b0d586\" (UID: \"41167fb8-73ed-4487-975c-365854b0d586\") " Mar 13 14:29:44 crc kubenswrapper[4907]: I0313 14:29:44.759129 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/41167fb8-73ed-4487-975c-365854b0d586-run-httpd\") pod \"41167fb8-73ed-4487-975c-365854b0d586\" (UID: \"41167fb8-73ed-4487-975c-365854b0d586\") " Mar 13 14:29:44 crc kubenswrapper[4907]: I0313 14:29:44.759150 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41167fb8-73ed-4487-975c-365854b0d586-combined-ca-bundle\") pod \"41167fb8-73ed-4487-975c-365854b0d586\" (UID: \"41167fb8-73ed-4487-975c-365854b0d586\") " Mar 13 14:29:44 crc kubenswrapper[4907]: I0313 14:29:44.759250 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4ww5\" (UniqueName: \"kubernetes.io/projected/41167fb8-73ed-4487-975c-365854b0d586-kube-api-access-w4ww5\") pod \"41167fb8-73ed-4487-975c-365854b0d586\" (UID: \"41167fb8-73ed-4487-975c-365854b0d586\") " Mar 13 14:29:44 crc kubenswrapper[4907]: I0313 14:29:44.759333 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/41167fb8-73ed-4487-975c-365854b0d586-ceilometer-tls-certs\") pod \"41167fb8-73ed-4487-975c-365854b0d586\" (UID: \"41167fb8-73ed-4487-975c-365854b0d586\") " Mar 13 14:29:44 crc kubenswrapper[4907]: I0313 14:29:44.759369 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/41167fb8-73ed-4487-975c-365854b0d586-log-httpd\") pod \"41167fb8-73ed-4487-975c-365854b0d586\" (UID: \"41167fb8-73ed-4487-975c-365854b0d586\") " Mar 13 14:29:44 crc kubenswrapper[4907]: I0313 14:29:44.759443 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/41167fb8-73ed-4487-975c-365854b0d586-scripts\") pod \"41167fb8-73ed-4487-975c-365854b0d586\" (UID: \"41167fb8-73ed-4487-975c-365854b0d586\") " Mar 13 14:29:44 crc kubenswrapper[4907]: I0313 14:29:44.759625 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/41167fb8-73ed-4487-975c-365854b0d586-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "41167fb8-73ed-4487-975c-365854b0d586" (UID: "41167fb8-73ed-4487-975c-365854b0d586"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:29:44 crc kubenswrapper[4907]: I0313 14:29:44.759845 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/41167fb8-73ed-4487-975c-365854b0d586-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "41167fb8-73ed-4487-975c-365854b0d586" (UID: "41167fb8-73ed-4487-975c-365854b0d586"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:29:44 crc kubenswrapper[4907]: I0313 14:29:44.759956 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41167fb8-73ed-4487-975c-365854b0d586-config-data\") pod \"41167fb8-73ed-4487-975c-365854b0d586\" (UID: \"41167fb8-73ed-4487-975c-365854b0d586\") " Mar 13 14:29:44 crc kubenswrapper[4907]: I0313 14:29:44.760740 4907 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/41167fb8-73ed-4487-975c-365854b0d586-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:44 crc kubenswrapper[4907]: I0313 14:29:44.760828 4907 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/41167fb8-73ed-4487-975c-365854b0d586-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:44 crc kubenswrapper[4907]: I0313 14:29:44.766119 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41167fb8-73ed-4487-975c-365854b0d586-scripts" (OuterVolumeSpecName: "scripts") pod "41167fb8-73ed-4487-975c-365854b0d586" (UID: "41167fb8-73ed-4487-975c-365854b0d586"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:29:44 crc kubenswrapper[4907]: I0313 14:29:44.766439 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/41167fb8-73ed-4487-975c-365854b0d586-kube-api-access-w4ww5" (OuterVolumeSpecName: "kube-api-access-w4ww5") pod "41167fb8-73ed-4487-975c-365854b0d586" (UID: "41167fb8-73ed-4487-975c-365854b0d586"). InnerVolumeSpecName "kube-api-access-w4ww5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:29:44 crc kubenswrapper[4907]: I0313 14:29:44.802360 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41167fb8-73ed-4487-975c-365854b0d586-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "41167fb8-73ed-4487-975c-365854b0d586" (UID: "41167fb8-73ed-4487-975c-365854b0d586"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:29:44 crc kubenswrapper[4907]: I0313 14:29:44.811400 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41167fb8-73ed-4487-975c-365854b0d586-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "41167fb8-73ed-4487-975c-365854b0d586" (UID: "41167fb8-73ed-4487-975c-365854b0d586"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:29:44 crc kubenswrapper[4907]: I0313 14:29:44.849076 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41167fb8-73ed-4487-975c-365854b0d586-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "41167fb8-73ed-4487-975c-365854b0d586" (UID: "41167fb8-73ed-4487-975c-365854b0d586"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:29:44 crc kubenswrapper[4907]: I0313 14:29:44.858942 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41167fb8-73ed-4487-975c-365854b0d586-config-data" (OuterVolumeSpecName: "config-data") pod "41167fb8-73ed-4487-975c-365854b0d586" (UID: "41167fb8-73ed-4487-975c-365854b0d586"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:29:44 crc kubenswrapper[4907]: I0313 14:29:44.862431 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41167fb8-73ed-4487-975c-365854b0d586-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:44 crc kubenswrapper[4907]: I0313 14:29:44.862477 4907 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/41167fb8-73ed-4487-975c-365854b0d586-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:44 crc kubenswrapper[4907]: I0313 14:29:44.862492 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41167fb8-73ed-4487-975c-365854b0d586-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:44 crc kubenswrapper[4907]: I0313 14:29:44.862506 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4ww5\" (UniqueName: \"kubernetes.io/projected/41167fb8-73ed-4487-975c-365854b0d586-kube-api-access-w4ww5\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:44 crc kubenswrapper[4907]: I0313 14:29:44.862518 4907 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/41167fb8-73ed-4487-975c-365854b0d586-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:44 crc kubenswrapper[4907]: I0313 14:29:44.862533 4907 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/41167fb8-73ed-4487-975c-365854b0d586-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:45 crc kubenswrapper[4907]: I0313 14:29:45.405830 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"41167fb8-73ed-4487-975c-365854b0d586","Type":"ContainerDied","Data":"a73a14261a590eee8da5263fe0ee191d82682588896d368eb8d1f46e7f9984ba"} Mar 13 14:29:45 crc kubenswrapper[4907]: I0313 14:29:45.405894 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 14:29:45 crc kubenswrapper[4907]: I0313 14:29:45.406189 4907 scope.go:117] "RemoveContainer" containerID="8e591aebdcd4fb21580fb8235da93dec31fd3679a3943e9171c4f1541d57c538" Mar 13 14:29:45 crc kubenswrapper[4907]: I0313 14:29:45.430432 4907 scope.go:117] "RemoveContainer" containerID="5939f8118a1b021358a2a1f3def6573f42d731f7807d9451cfe9c468306f6dbc" Mar 13 14:29:45 crc kubenswrapper[4907]: I0313 14:29:45.447120 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 13 14:29:45 crc kubenswrapper[4907]: I0313 14:29:45.456600 4907 scope.go:117] "RemoveContainer" containerID="5ecb51a13db91baac3a9eb0c7577b30ef982b63567b8d459d38ac04dfa9a4515" Mar 13 14:29:45 crc kubenswrapper[4907]: I0313 14:29:45.464032 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Mar 13 14:29:45 crc kubenswrapper[4907]: I0313 14:29:45.473794 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 13 14:29:45 crc kubenswrapper[4907]: E0313 14:29:45.474153 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41167fb8-73ed-4487-975c-365854b0d586" containerName="ceilometer-notification-agent" Mar 13 14:29:45 crc kubenswrapper[4907]: I0313 14:29:45.474169 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="41167fb8-73ed-4487-975c-365854b0d586" containerName="ceilometer-notification-agent" Mar 13 14:29:45 crc kubenswrapper[4907]: E0313 14:29:45.474190 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41167fb8-73ed-4487-975c-365854b0d586" containerName="sg-core" Mar 13 14:29:45 crc kubenswrapper[4907]: I0313 14:29:45.474198 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="41167fb8-73ed-4487-975c-365854b0d586" containerName="sg-core" Mar 13 14:29:45 crc kubenswrapper[4907]: E0313 14:29:45.474207 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41167fb8-73ed-4487-975c-365854b0d586" containerName="proxy-httpd" Mar 13 14:29:45 crc kubenswrapper[4907]: I0313 14:29:45.474213 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="41167fb8-73ed-4487-975c-365854b0d586" containerName="proxy-httpd" Mar 13 14:29:45 crc kubenswrapper[4907]: E0313 14:29:45.474237 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41167fb8-73ed-4487-975c-365854b0d586" containerName="ceilometer-central-agent" Mar 13 14:29:45 crc kubenswrapper[4907]: I0313 14:29:45.474243 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="41167fb8-73ed-4487-975c-365854b0d586" containerName="ceilometer-central-agent" Mar 13 14:29:45 crc kubenswrapper[4907]: I0313 14:29:45.474404 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="41167fb8-73ed-4487-975c-365854b0d586" containerName="proxy-httpd" Mar 13 14:29:45 crc kubenswrapper[4907]: I0313 14:29:45.474423 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="41167fb8-73ed-4487-975c-365854b0d586" containerName="ceilometer-central-agent" Mar 13 14:29:45 crc kubenswrapper[4907]: I0313 14:29:45.474445 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="41167fb8-73ed-4487-975c-365854b0d586" containerName="sg-core" Mar 13 14:29:45 crc kubenswrapper[4907]: I0313 14:29:45.474458 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="41167fb8-73ed-4487-975c-365854b0d586" containerName="ceilometer-notification-agent" Mar 13 14:29:45 crc kubenswrapper[4907]: I0313 14:29:45.476201 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 14:29:45 crc kubenswrapper[4907]: I0313 14:29:45.479564 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Mar 13 14:29:45 crc kubenswrapper[4907]: I0313 14:29:45.479744 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 13 14:29:45 crc kubenswrapper[4907]: I0313 14:29:45.479774 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 13 14:29:45 crc kubenswrapper[4907]: I0313 14:29:45.482091 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 13 14:29:45 crc kubenswrapper[4907]: I0313 14:29:45.482115 4907 scope.go:117] "RemoveContainer" containerID="9b5ab98c8fa734605e5e5d3d4bc91ad54650c1db1f1d0d56f2cbf2225fe14db6" Mar 13 14:29:45 crc kubenswrapper[4907]: I0313 14:29:45.584112 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/aa81fa2a-4c27-4e02-9891-5b39348feca9-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"aa81fa2a-4c27-4e02-9891-5b39348feca9\") " pod="openstack/ceilometer-0" Mar 13 14:29:45 crc kubenswrapper[4907]: I0313 14:29:45.584207 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dvn24\" (UniqueName: \"kubernetes.io/projected/aa81fa2a-4c27-4e02-9891-5b39348feca9-kube-api-access-dvn24\") pod \"ceilometer-0\" (UID: \"aa81fa2a-4c27-4e02-9891-5b39348feca9\") " pod="openstack/ceilometer-0" Mar 13 14:29:45 crc kubenswrapper[4907]: I0313 14:29:45.584254 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aa81fa2a-4c27-4e02-9891-5b39348feca9-scripts\") pod \"ceilometer-0\" (UID: \"aa81fa2a-4c27-4e02-9891-5b39348feca9\") " pod="openstack/ceilometer-0" Mar 13 14:29:45 crc kubenswrapper[4907]: I0313 14:29:45.584285 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa81fa2a-4c27-4e02-9891-5b39348feca9-config-data\") pod \"ceilometer-0\" (UID: \"aa81fa2a-4c27-4e02-9891-5b39348feca9\") " pod="openstack/ceilometer-0" Mar 13 14:29:45 crc kubenswrapper[4907]: I0313 14:29:45.584349 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aa81fa2a-4c27-4e02-9891-5b39348feca9-run-httpd\") pod \"ceilometer-0\" (UID: \"aa81fa2a-4c27-4e02-9891-5b39348feca9\") " pod="openstack/ceilometer-0" Mar 13 14:29:45 crc kubenswrapper[4907]: I0313 14:29:45.584380 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/aa81fa2a-4c27-4e02-9891-5b39348feca9-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"aa81fa2a-4c27-4e02-9891-5b39348feca9\") " pod="openstack/ceilometer-0" Mar 13 14:29:45 crc kubenswrapper[4907]: I0313 14:29:45.584430 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aa81fa2a-4c27-4e02-9891-5b39348feca9-log-httpd\") pod \"ceilometer-0\" (UID: \"aa81fa2a-4c27-4e02-9891-5b39348feca9\") " pod="openstack/ceilometer-0" Mar 13 14:29:45 crc kubenswrapper[4907]: I0313 14:29:45.584460 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa81fa2a-4c27-4e02-9891-5b39348feca9-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"aa81fa2a-4c27-4e02-9891-5b39348feca9\") " pod="openstack/ceilometer-0" Mar 13 14:29:45 crc kubenswrapper[4907]: I0313 14:29:45.686604 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aa81fa2a-4c27-4e02-9891-5b39348feca9-run-httpd\") pod \"ceilometer-0\" (UID: \"aa81fa2a-4c27-4e02-9891-5b39348feca9\") " pod="openstack/ceilometer-0" Mar 13 14:29:45 crc kubenswrapper[4907]: I0313 14:29:45.686665 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/aa81fa2a-4c27-4e02-9891-5b39348feca9-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"aa81fa2a-4c27-4e02-9891-5b39348feca9\") " pod="openstack/ceilometer-0" Mar 13 14:29:45 crc kubenswrapper[4907]: I0313 14:29:45.686692 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aa81fa2a-4c27-4e02-9891-5b39348feca9-log-httpd\") pod \"ceilometer-0\" (UID: \"aa81fa2a-4c27-4e02-9891-5b39348feca9\") " pod="openstack/ceilometer-0" Mar 13 14:29:45 crc kubenswrapper[4907]: I0313 14:29:45.686710 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa81fa2a-4c27-4e02-9891-5b39348feca9-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"aa81fa2a-4c27-4e02-9891-5b39348feca9\") " pod="openstack/ceilometer-0" Mar 13 14:29:45 crc kubenswrapper[4907]: I0313 14:29:45.686803 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/aa81fa2a-4c27-4e02-9891-5b39348feca9-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"aa81fa2a-4c27-4e02-9891-5b39348feca9\") " pod="openstack/ceilometer-0" Mar 13 14:29:45 crc kubenswrapper[4907]: I0313 14:29:45.686848 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dvn24\" (UniqueName: \"kubernetes.io/projected/aa81fa2a-4c27-4e02-9891-5b39348feca9-kube-api-access-dvn24\") pod \"ceilometer-0\" (UID: \"aa81fa2a-4c27-4e02-9891-5b39348feca9\") " pod="openstack/ceilometer-0" Mar 13 14:29:45 crc kubenswrapper[4907]: I0313 14:29:45.686894 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aa81fa2a-4c27-4e02-9891-5b39348feca9-scripts\") pod \"ceilometer-0\" (UID: \"aa81fa2a-4c27-4e02-9891-5b39348feca9\") " pod="openstack/ceilometer-0" Mar 13 14:29:45 crc kubenswrapper[4907]: I0313 14:29:45.686943 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa81fa2a-4c27-4e02-9891-5b39348feca9-config-data\") pod \"ceilometer-0\" (UID: \"aa81fa2a-4c27-4e02-9891-5b39348feca9\") " pod="openstack/ceilometer-0" Mar 13 14:29:45 crc kubenswrapper[4907]: I0313 14:29:45.687779 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aa81fa2a-4c27-4e02-9891-5b39348feca9-run-httpd\") pod \"ceilometer-0\" (UID: \"aa81fa2a-4c27-4e02-9891-5b39348feca9\") " pod="openstack/ceilometer-0" Mar 13 14:29:45 crc kubenswrapper[4907]: I0313 14:29:45.688165 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aa81fa2a-4c27-4e02-9891-5b39348feca9-log-httpd\") pod \"ceilometer-0\" (UID: \"aa81fa2a-4c27-4e02-9891-5b39348feca9\") " pod="openstack/ceilometer-0" Mar 13 14:29:45 crc kubenswrapper[4907]: I0313 14:29:45.692371 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/aa81fa2a-4c27-4e02-9891-5b39348feca9-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"aa81fa2a-4c27-4e02-9891-5b39348feca9\") " pod="openstack/ceilometer-0" Mar 13 14:29:45 crc kubenswrapper[4907]: I0313 14:29:45.694150 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/aa81fa2a-4c27-4e02-9891-5b39348feca9-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"aa81fa2a-4c27-4e02-9891-5b39348feca9\") " pod="openstack/ceilometer-0" Mar 13 14:29:45 crc kubenswrapper[4907]: I0313 14:29:45.695103 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa81fa2a-4c27-4e02-9891-5b39348feca9-config-data\") pod \"ceilometer-0\" (UID: \"aa81fa2a-4c27-4e02-9891-5b39348feca9\") " pod="openstack/ceilometer-0" Mar 13 14:29:45 crc kubenswrapper[4907]: I0313 14:29:45.696210 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aa81fa2a-4c27-4e02-9891-5b39348feca9-scripts\") pod \"ceilometer-0\" (UID: \"aa81fa2a-4c27-4e02-9891-5b39348feca9\") " pod="openstack/ceilometer-0" Mar 13 14:29:45 crc kubenswrapper[4907]: I0313 14:29:45.697536 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa81fa2a-4c27-4e02-9891-5b39348feca9-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"aa81fa2a-4c27-4e02-9891-5b39348feca9\") " pod="openstack/ceilometer-0" Mar 13 14:29:45 crc kubenswrapper[4907]: I0313 14:29:45.712202 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dvn24\" (UniqueName: \"kubernetes.io/projected/aa81fa2a-4c27-4e02-9891-5b39348feca9-kube-api-access-dvn24\") pod \"ceilometer-0\" (UID: \"aa81fa2a-4c27-4e02-9891-5b39348feca9\") " pod="openstack/ceilometer-0" Mar 13 14:29:45 crc kubenswrapper[4907]: I0313 14:29:45.793080 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="41167fb8-73ed-4487-975c-365854b0d586" path="/var/lib/kubelet/pods/41167fb8-73ed-4487-975c-365854b0d586/volumes" Mar 13 14:29:45 crc kubenswrapper[4907]: I0313 14:29:45.796723 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 14:29:46 crc kubenswrapper[4907]: I0313 14:29:46.587016 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.040542 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.062962 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.189592 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.320190 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41063592-cda5-443f-87e3-20c2c6ec5c73-combined-ca-bundle\") pod \"41063592-cda5-443f-87e3-20c2c6ec5c73\" (UID: \"41063592-cda5-443f-87e3-20c2c6ec5c73\") " Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.320601 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41063592-cda5-443f-87e3-20c2c6ec5c73-config-data\") pod \"41063592-cda5-443f-87e3-20c2c6ec5c73\" (UID: \"41063592-cda5-443f-87e3-20c2c6ec5c73\") " Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.320820 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/41063592-cda5-443f-87e3-20c2c6ec5c73-logs\") pod \"41063592-cda5-443f-87e3-20c2c6ec5c73\" (UID: \"41063592-cda5-443f-87e3-20c2c6ec5c73\") " Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.320859 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zhhbp\" (UniqueName: \"kubernetes.io/projected/41063592-cda5-443f-87e3-20c2c6ec5c73-kube-api-access-zhhbp\") pod \"41063592-cda5-443f-87e3-20c2c6ec5c73\" (UID: \"41063592-cda5-443f-87e3-20c2c6ec5c73\") " Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.321224 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/41063592-cda5-443f-87e3-20c2c6ec5c73-logs" (OuterVolumeSpecName: "logs") pod "41063592-cda5-443f-87e3-20c2c6ec5c73" (UID: "41063592-cda5-443f-87e3-20c2c6ec5c73"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.321530 4907 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/41063592-cda5-443f-87e3-20c2c6ec5c73-logs\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.326436 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/41063592-cda5-443f-87e3-20c2c6ec5c73-kube-api-access-zhhbp" (OuterVolumeSpecName: "kube-api-access-zhhbp") pod "41063592-cda5-443f-87e3-20c2c6ec5c73" (UID: "41063592-cda5-443f-87e3-20c2c6ec5c73"). InnerVolumeSpecName "kube-api-access-zhhbp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.355755 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41063592-cda5-443f-87e3-20c2c6ec5c73-config-data" (OuterVolumeSpecName: "config-data") pod "41063592-cda5-443f-87e3-20c2c6ec5c73" (UID: "41063592-cda5-443f-87e3-20c2c6ec5c73"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.376053 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41063592-cda5-443f-87e3-20c2c6ec5c73-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "41063592-cda5-443f-87e3-20c2c6ec5c73" (UID: "41063592-cda5-443f-87e3-20c2c6ec5c73"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.423232 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41063592-cda5-443f-87e3-20c2c6ec5c73-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.423261 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zhhbp\" (UniqueName: \"kubernetes.io/projected/41063592-cda5-443f-87e3-20c2c6ec5c73-kube-api-access-zhhbp\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.423270 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41063592-cda5-443f-87e3-20c2c6ec5c73-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.555129 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aa81fa2a-4c27-4e02-9891-5b39348feca9","Type":"ContainerStarted","Data":"285cdb1df855cbeeda75e9d50bffbe69e5c3f9a5485dd118f47a932fb1d36f19"} Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.555181 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aa81fa2a-4c27-4e02-9891-5b39348feca9","Type":"ContainerStarted","Data":"b4ef9b58ba405d6fb537a91498e7353e4bf4deae5a0e5f61d8548bb330be20cb"} Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.557606 4907 generic.go:334] "Generic (PLEG): container finished" podID="41063592-cda5-443f-87e3-20c2c6ec5c73" containerID="bee51476e0503a2fc3a96ca20f1d42b6cb92e1822245b7b21033273b10f0e7a6" exitCode=0 Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.557655 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.557678 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"41063592-cda5-443f-87e3-20c2c6ec5c73","Type":"ContainerDied","Data":"bee51476e0503a2fc3a96ca20f1d42b6cb92e1822245b7b21033273b10f0e7a6"} Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.557748 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"41063592-cda5-443f-87e3-20c2c6ec5c73","Type":"ContainerDied","Data":"11bb2f382efdf360af4e2164320ab74af741c059bdd8ac659bd82b1507cb52af"} Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.557771 4907 scope.go:117] "RemoveContainer" containerID="bee51476e0503a2fc3a96ca20f1d42b6cb92e1822245b7b21033273b10f0e7a6" Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.586297 4907 scope.go:117] "RemoveContainer" containerID="97742dae0519818b9c495c23cd7ee1fe57412023b1fc525ca01e7fc2236cef2b" Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.590171 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.672264 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.676575 4907 scope.go:117] "RemoveContainer" containerID="bee51476e0503a2fc3a96ca20f1d42b6cb92e1822245b7b21033273b10f0e7a6" Mar 13 14:29:47 crc kubenswrapper[4907]: E0313 14:29:47.678721 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bee51476e0503a2fc3a96ca20f1d42b6cb92e1822245b7b21033273b10f0e7a6\": container with ID starting with bee51476e0503a2fc3a96ca20f1d42b6cb92e1822245b7b21033273b10f0e7a6 not found: ID does not exist" containerID="bee51476e0503a2fc3a96ca20f1d42b6cb92e1822245b7b21033273b10f0e7a6" Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.678775 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bee51476e0503a2fc3a96ca20f1d42b6cb92e1822245b7b21033273b10f0e7a6"} err="failed to get container status \"bee51476e0503a2fc3a96ca20f1d42b6cb92e1822245b7b21033273b10f0e7a6\": rpc error: code = NotFound desc = could not find container \"bee51476e0503a2fc3a96ca20f1d42b6cb92e1822245b7b21033273b10f0e7a6\": container with ID starting with bee51476e0503a2fc3a96ca20f1d42b6cb92e1822245b7b21033273b10f0e7a6 not found: ID does not exist" Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.678807 4907 scope.go:117] "RemoveContainer" containerID="97742dae0519818b9c495c23cd7ee1fe57412023b1fc525ca01e7fc2236cef2b" Mar 13 14:29:47 crc kubenswrapper[4907]: E0313 14:29:47.680181 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"97742dae0519818b9c495c23cd7ee1fe57412023b1fc525ca01e7fc2236cef2b\": container with ID starting with 97742dae0519818b9c495c23cd7ee1fe57412023b1fc525ca01e7fc2236cef2b not found: ID does not exist" containerID="97742dae0519818b9c495c23cd7ee1fe57412023b1fc525ca01e7fc2236cef2b" Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.680226 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"97742dae0519818b9c495c23cd7ee1fe57412023b1fc525ca01e7fc2236cef2b"} err="failed to get container status \"97742dae0519818b9c495c23cd7ee1fe57412023b1fc525ca01e7fc2236cef2b\": rpc error: code = NotFound desc = could not find container \"97742dae0519818b9c495c23cd7ee1fe57412023b1fc525ca01e7fc2236cef2b\": container with ID starting with 97742dae0519818b9c495c23cd7ee1fe57412023b1fc525ca01e7fc2236cef2b not found: ID does not exist" Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.692955 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.707958 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Mar 13 14:29:47 crc kubenswrapper[4907]: E0313 14:29:47.708585 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41063592-cda5-443f-87e3-20c2c6ec5c73" containerName="nova-api-api" Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.708622 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="41063592-cda5-443f-87e3-20c2c6ec5c73" containerName="nova-api-api" Mar 13 14:29:47 crc kubenswrapper[4907]: E0313 14:29:47.708651 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41063592-cda5-443f-87e3-20c2c6ec5c73" containerName="nova-api-log" Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.708658 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="41063592-cda5-443f-87e3-20c2c6ec5c73" containerName="nova-api-log" Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.708983 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="41063592-cda5-443f-87e3-20c2c6ec5c73" containerName="nova-api-api" Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.709032 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="41063592-cda5-443f-87e3-20c2c6ec5c73" containerName="nova-api-log" Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.710424 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.713654 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.714151 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.714547 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.719449 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.719495 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.721645 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.811948 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="41063592-cda5-443f-87e3-20c2c6ec5c73" path="/var/lib/kubelet/pods/41063592-cda5-443f-87e3-20c2c6ec5c73/volumes" Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.832375 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q8tvk\" (UniqueName: \"kubernetes.io/projected/4851fd7f-a2e6-4ec1-96ea-7c0086edd03e-kube-api-access-q8tvk\") pod \"nova-api-0\" (UID: \"4851fd7f-a2e6-4ec1-96ea-7c0086edd03e\") " pod="openstack/nova-api-0" Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.832419 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4851fd7f-a2e6-4ec1-96ea-7c0086edd03e-config-data\") pod \"nova-api-0\" (UID: \"4851fd7f-a2e6-4ec1-96ea-7c0086edd03e\") " pod="openstack/nova-api-0" Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.832457 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4851fd7f-a2e6-4ec1-96ea-7c0086edd03e-internal-tls-certs\") pod \"nova-api-0\" (UID: \"4851fd7f-a2e6-4ec1-96ea-7c0086edd03e\") " pod="openstack/nova-api-0" Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.832496 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4851fd7f-a2e6-4ec1-96ea-7c0086edd03e-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"4851fd7f-a2e6-4ec1-96ea-7c0086edd03e\") " pod="openstack/nova-api-0" Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.832581 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4851fd7f-a2e6-4ec1-96ea-7c0086edd03e-logs\") pod \"nova-api-0\" (UID: \"4851fd7f-a2e6-4ec1-96ea-7c0086edd03e\") " pod="openstack/nova-api-0" Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.832600 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4851fd7f-a2e6-4ec1-96ea-7c0086edd03e-public-tls-certs\") pod \"nova-api-0\" (UID: \"4851fd7f-a2e6-4ec1-96ea-7c0086edd03e\") " pod="openstack/nova-api-0" Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.874472 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-2pvp9"] Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.893258 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-2pvp9" Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.897854 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.898122 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.933960 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q8tvk\" (UniqueName: \"kubernetes.io/projected/4851fd7f-a2e6-4ec1-96ea-7c0086edd03e-kube-api-access-q8tvk\") pod \"nova-api-0\" (UID: \"4851fd7f-a2e6-4ec1-96ea-7c0086edd03e\") " pod="openstack/nova-api-0" Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.934568 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4851fd7f-a2e6-4ec1-96ea-7c0086edd03e-config-data\") pod \"nova-api-0\" (UID: \"4851fd7f-a2e6-4ec1-96ea-7c0086edd03e\") " pod="openstack/nova-api-0" Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.934672 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4851fd7f-a2e6-4ec1-96ea-7c0086edd03e-internal-tls-certs\") pod \"nova-api-0\" (UID: \"4851fd7f-a2e6-4ec1-96ea-7c0086edd03e\") " pod="openstack/nova-api-0" Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.934791 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4851fd7f-a2e6-4ec1-96ea-7c0086edd03e-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"4851fd7f-a2e6-4ec1-96ea-7c0086edd03e\") " pod="openstack/nova-api-0" Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.934970 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4851fd7f-a2e6-4ec1-96ea-7c0086edd03e-logs\") pod \"nova-api-0\" (UID: \"4851fd7f-a2e6-4ec1-96ea-7c0086edd03e\") " pod="openstack/nova-api-0" Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.935067 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4851fd7f-a2e6-4ec1-96ea-7c0086edd03e-public-tls-certs\") pod \"nova-api-0\" (UID: \"4851fd7f-a2e6-4ec1-96ea-7c0086edd03e\") " pod="openstack/nova-api-0" Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.938711 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4851fd7f-a2e6-4ec1-96ea-7c0086edd03e-logs\") pod \"nova-api-0\" (UID: \"4851fd7f-a2e6-4ec1-96ea-7c0086edd03e\") " pod="openstack/nova-api-0" Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.942589 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-2pvp9"] Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.943348 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4851fd7f-a2e6-4ec1-96ea-7c0086edd03e-config-data\") pod \"nova-api-0\" (UID: \"4851fd7f-a2e6-4ec1-96ea-7c0086edd03e\") " pod="openstack/nova-api-0" Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.944088 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4851fd7f-a2e6-4ec1-96ea-7c0086edd03e-internal-tls-certs\") pod \"nova-api-0\" (UID: \"4851fd7f-a2e6-4ec1-96ea-7c0086edd03e\") " pod="openstack/nova-api-0" Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.952386 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4851fd7f-a2e6-4ec1-96ea-7c0086edd03e-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"4851fd7f-a2e6-4ec1-96ea-7c0086edd03e\") " pod="openstack/nova-api-0" Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.959556 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4851fd7f-a2e6-4ec1-96ea-7c0086edd03e-public-tls-certs\") pod \"nova-api-0\" (UID: \"4851fd7f-a2e6-4ec1-96ea-7c0086edd03e\") " pod="openstack/nova-api-0" Mar 13 14:29:48 crc kubenswrapper[4907]: I0313 14:29:47.989080 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q8tvk\" (UniqueName: \"kubernetes.io/projected/4851fd7f-a2e6-4ec1-96ea-7c0086edd03e-kube-api-access-q8tvk\") pod \"nova-api-0\" (UID: \"4851fd7f-a2e6-4ec1-96ea-7c0086edd03e\") " pod="openstack/nova-api-0" Mar 13 14:29:48 crc kubenswrapper[4907]: I0313 14:29:48.037022 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8901c771-b45a-4147-9f96-cef784165281-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-2pvp9\" (UID: \"8901c771-b45a-4147-9f96-cef784165281\") " pod="openstack/nova-cell1-cell-mapping-2pvp9" Mar 13 14:29:48 crc kubenswrapper[4907]: I0313 14:29:48.037074 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8901c771-b45a-4147-9f96-cef784165281-scripts\") pod \"nova-cell1-cell-mapping-2pvp9\" (UID: \"8901c771-b45a-4147-9f96-cef784165281\") " pod="openstack/nova-cell1-cell-mapping-2pvp9" Mar 13 14:29:48 crc kubenswrapper[4907]: I0313 14:29:48.037205 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8901c771-b45a-4147-9f96-cef784165281-config-data\") pod \"nova-cell1-cell-mapping-2pvp9\" (UID: \"8901c771-b45a-4147-9f96-cef784165281\") " pod="openstack/nova-cell1-cell-mapping-2pvp9" Mar 13 14:29:48 crc kubenswrapper[4907]: I0313 14:29:48.037243 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h9nqx\" (UniqueName: \"kubernetes.io/projected/8901c771-b45a-4147-9f96-cef784165281-kube-api-access-h9nqx\") pod \"nova-cell1-cell-mapping-2pvp9\" (UID: \"8901c771-b45a-4147-9f96-cef784165281\") " pod="openstack/nova-cell1-cell-mapping-2pvp9" Mar 13 14:29:48 crc kubenswrapper[4907]: I0313 14:29:48.065847 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 13 14:29:48 crc kubenswrapper[4907]: I0313 14:29:48.139509 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8901c771-b45a-4147-9f96-cef784165281-config-data\") pod \"nova-cell1-cell-mapping-2pvp9\" (UID: \"8901c771-b45a-4147-9f96-cef784165281\") " pod="openstack/nova-cell1-cell-mapping-2pvp9" Mar 13 14:29:48 crc kubenswrapper[4907]: I0313 14:29:48.139582 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h9nqx\" (UniqueName: \"kubernetes.io/projected/8901c771-b45a-4147-9f96-cef784165281-kube-api-access-h9nqx\") pod \"nova-cell1-cell-mapping-2pvp9\" (UID: \"8901c771-b45a-4147-9f96-cef784165281\") " pod="openstack/nova-cell1-cell-mapping-2pvp9" Mar 13 14:29:48 crc kubenswrapper[4907]: I0313 14:29:48.139630 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8901c771-b45a-4147-9f96-cef784165281-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-2pvp9\" (UID: \"8901c771-b45a-4147-9f96-cef784165281\") " pod="openstack/nova-cell1-cell-mapping-2pvp9" Mar 13 14:29:48 crc kubenswrapper[4907]: I0313 14:29:48.139648 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8901c771-b45a-4147-9f96-cef784165281-scripts\") pod \"nova-cell1-cell-mapping-2pvp9\" (UID: \"8901c771-b45a-4147-9f96-cef784165281\") " pod="openstack/nova-cell1-cell-mapping-2pvp9" Mar 13 14:29:48 crc kubenswrapper[4907]: I0313 14:29:48.149429 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8901c771-b45a-4147-9f96-cef784165281-scripts\") pod \"nova-cell1-cell-mapping-2pvp9\" (UID: \"8901c771-b45a-4147-9f96-cef784165281\") " pod="openstack/nova-cell1-cell-mapping-2pvp9" Mar 13 14:29:48 crc kubenswrapper[4907]: I0313 14:29:48.150435 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8901c771-b45a-4147-9f96-cef784165281-config-data\") pod \"nova-cell1-cell-mapping-2pvp9\" (UID: \"8901c771-b45a-4147-9f96-cef784165281\") " pod="openstack/nova-cell1-cell-mapping-2pvp9" Mar 13 14:29:48 crc kubenswrapper[4907]: I0313 14:29:48.152441 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8901c771-b45a-4147-9f96-cef784165281-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-2pvp9\" (UID: \"8901c771-b45a-4147-9f96-cef784165281\") " pod="openstack/nova-cell1-cell-mapping-2pvp9" Mar 13 14:29:48 crc kubenswrapper[4907]: I0313 14:29:48.159037 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h9nqx\" (UniqueName: \"kubernetes.io/projected/8901c771-b45a-4147-9f96-cef784165281-kube-api-access-h9nqx\") pod \"nova-cell1-cell-mapping-2pvp9\" (UID: \"8901c771-b45a-4147-9f96-cef784165281\") " pod="openstack/nova-cell1-cell-mapping-2pvp9" Mar 13 14:29:48 crc kubenswrapper[4907]: I0313 14:29:48.358183 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-2pvp9" Mar 13 14:29:48 crc kubenswrapper[4907]: I0313 14:29:48.578171 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aa81fa2a-4c27-4e02-9891-5b39348feca9","Type":"ContainerStarted","Data":"78891ea90d5eb95d642931ec25514d47603b8811d971d7d33d8fdc5eb0d3bb04"} Mar 13 14:29:48 crc kubenswrapper[4907]: I0313 14:29:48.635562 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 13 14:29:48 crc kubenswrapper[4907]: I0313 14:29:48.734113 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="39488342-aa1a-4dba-bd5f-556da5081904" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.206:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 13 14:29:48 crc kubenswrapper[4907]: I0313 14:29:48.734142 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="39488342-aa1a-4dba-bd5f-556da5081904" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.206:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 13 14:29:48 crc kubenswrapper[4907]: I0313 14:29:48.839785 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-2pvp9"] Mar 13 14:29:49 crc kubenswrapper[4907]: I0313 14:29:49.589174 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-2pvp9" event={"ID":"8901c771-b45a-4147-9f96-cef784165281","Type":"ContainerStarted","Data":"714551bd9afe88fbb37b8d1ba1d86847559c10eb4afc2d8c8cade9bb4be375ee"} Mar 13 14:29:49 crc kubenswrapper[4907]: I0313 14:29:49.589540 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-2pvp9" event={"ID":"8901c771-b45a-4147-9f96-cef784165281","Type":"ContainerStarted","Data":"9a7bf9979b09ab5543692c5fb89940c7553ed9a8ca26785e17126dc549ce6605"} Mar 13 14:29:49 crc kubenswrapper[4907]: I0313 14:29:49.591999 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4851fd7f-a2e6-4ec1-96ea-7c0086edd03e","Type":"ContainerStarted","Data":"8a9e8c8e4a79ece75fdb50c32528e4d6f87181c842fcd7d1b9802e179e72bd3b"} Mar 13 14:29:49 crc kubenswrapper[4907]: I0313 14:29:49.592115 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4851fd7f-a2e6-4ec1-96ea-7c0086edd03e","Type":"ContainerStarted","Data":"c4919269912183df77206bdb4ba3361c6048b9cadeaad50a6bcce3ec26e11c53"} Mar 13 14:29:49 crc kubenswrapper[4907]: I0313 14:29:49.592175 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4851fd7f-a2e6-4ec1-96ea-7c0086edd03e","Type":"ContainerStarted","Data":"2712a8612ce027f409c3dbd5c6adbabefb48875be5bd50702de7d61655bda2a8"} Mar 13 14:29:49 crc kubenswrapper[4907]: I0313 14:29:49.598397 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aa81fa2a-4c27-4e02-9891-5b39348feca9","Type":"ContainerStarted","Data":"16ef32bcd58d9ca78315416c39fe9eca0c5eee8682ac66cea46ec8f9242f8ee5"} Mar 13 14:29:49 crc kubenswrapper[4907]: I0313 14:29:49.627675 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-2pvp9" podStartSLOduration=2.627651369 podStartE2EDuration="2.627651369s" podCreationTimestamp="2026-03-13 14:29:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:29:49.6114689 +0000 UTC m=+1488.511256589" watchObservedRunningTime="2026-03-13 14:29:49.627651369 +0000 UTC m=+1488.527439068" Mar 13 14:29:49 crc kubenswrapper[4907]: I0313 14:29:49.640168 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.640149538 podStartE2EDuration="2.640149538s" podCreationTimestamp="2026-03-13 14:29:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:29:49.638139594 +0000 UTC m=+1488.537927293" watchObservedRunningTime="2026-03-13 14:29:49.640149538 +0000 UTC m=+1488.539937227" Mar 13 14:29:50 crc kubenswrapper[4907]: I0313 14:29:50.610648 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aa81fa2a-4c27-4e02-9891-5b39348feca9","Type":"ContainerStarted","Data":"79914c8007c9b10b224f5d4d1fcd270f17078ab7c002b89d3dab089884af5d6c"} Mar 13 14:29:50 crc kubenswrapper[4907]: I0313 14:29:50.631721 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.9883929930000002 podStartE2EDuration="5.631697206s" podCreationTimestamp="2026-03-13 14:29:45 +0000 UTC" firstStartedPulling="2026-03-13 14:29:46.582551852 +0000 UTC m=+1485.482339541" lastFinishedPulling="2026-03-13 14:29:50.225856065 +0000 UTC m=+1489.125643754" observedRunningTime="2026-03-13 14:29:50.62960675 +0000 UTC m=+1489.529394439" watchObservedRunningTime="2026-03-13 14:29:50.631697206 +0000 UTC m=+1489.531484895" Mar 13 14:29:50 crc kubenswrapper[4907]: I0313 14:29:50.841326 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-fdb8f6449-nb4km" Mar 13 14:29:50 crc kubenswrapper[4907]: I0313 14:29:50.905993 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-69b4446475-ttj9k"] Mar 13 14:29:50 crc kubenswrapper[4907]: I0313 14:29:50.907070 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-69b4446475-ttj9k" podUID="f56f7387-a1c9-4250-8abf-6c0a3830970e" containerName="dnsmasq-dns" containerID="cri-o://215938dfe070eab04984e52eb7b82e6d3f56a88ab88790cd28ad3b1a4bc7d77a" gracePeriod=10 Mar 13 14:29:51 crc kubenswrapper[4907]: I0313 14:29:51.441281 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-69b4446475-ttj9k" Mar 13 14:29:51 crc kubenswrapper[4907]: I0313 14:29:51.510828 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f56f7387-a1c9-4250-8abf-6c0a3830970e-ovsdbserver-sb\") pod \"f56f7387-a1c9-4250-8abf-6c0a3830970e\" (UID: \"f56f7387-a1c9-4250-8abf-6c0a3830970e\") " Mar 13 14:29:51 crc kubenswrapper[4907]: I0313 14:29:51.510988 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f56f7387-a1c9-4250-8abf-6c0a3830970e-dns-swift-storage-0\") pod \"f56f7387-a1c9-4250-8abf-6c0a3830970e\" (UID: \"f56f7387-a1c9-4250-8abf-6c0a3830970e\") " Mar 13 14:29:51 crc kubenswrapper[4907]: I0313 14:29:51.511111 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8c8lc\" (UniqueName: \"kubernetes.io/projected/f56f7387-a1c9-4250-8abf-6c0a3830970e-kube-api-access-8c8lc\") pod \"f56f7387-a1c9-4250-8abf-6c0a3830970e\" (UID: \"f56f7387-a1c9-4250-8abf-6c0a3830970e\") " Mar 13 14:29:51 crc kubenswrapper[4907]: I0313 14:29:51.511154 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f56f7387-a1c9-4250-8abf-6c0a3830970e-ovsdbserver-nb\") pod \"f56f7387-a1c9-4250-8abf-6c0a3830970e\" (UID: \"f56f7387-a1c9-4250-8abf-6c0a3830970e\") " Mar 13 14:29:51 crc kubenswrapper[4907]: I0313 14:29:51.511181 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f56f7387-a1c9-4250-8abf-6c0a3830970e-dns-svc\") pod \"f56f7387-a1c9-4250-8abf-6c0a3830970e\" (UID: \"f56f7387-a1c9-4250-8abf-6c0a3830970e\") " Mar 13 14:29:51 crc kubenswrapper[4907]: I0313 14:29:51.511221 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f56f7387-a1c9-4250-8abf-6c0a3830970e-config\") pod \"f56f7387-a1c9-4250-8abf-6c0a3830970e\" (UID: \"f56f7387-a1c9-4250-8abf-6c0a3830970e\") " Mar 13 14:29:51 crc kubenswrapper[4907]: I0313 14:29:51.518112 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f56f7387-a1c9-4250-8abf-6c0a3830970e-kube-api-access-8c8lc" (OuterVolumeSpecName: "kube-api-access-8c8lc") pod "f56f7387-a1c9-4250-8abf-6c0a3830970e" (UID: "f56f7387-a1c9-4250-8abf-6c0a3830970e"). InnerVolumeSpecName "kube-api-access-8c8lc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:29:51 crc kubenswrapper[4907]: I0313 14:29:51.582800 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f56f7387-a1c9-4250-8abf-6c0a3830970e-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "f56f7387-a1c9-4250-8abf-6c0a3830970e" (UID: "f56f7387-a1c9-4250-8abf-6c0a3830970e"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:29:51 crc kubenswrapper[4907]: I0313 14:29:51.589015 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f56f7387-a1c9-4250-8abf-6c0a3830970e-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "f56f7387-a1c9-4250-8abf-6c0a3830970e" (UID: "f56f7387-a1c9-4250-8abf-6c0a3830970e"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:29:51 crc kubenswrapper[4907]: I0313 14:29:51.595313 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f56f7387-a1c9-4250-8abf-6c0a3830970e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f56f7387-a1c9-4250-8abf-6c0a3830970e" (UID: "f56f7387-a1c9-4250-8abf-6c0a3830970e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:29:51 crc kubenswrapper[4907]: I0313 14:29:51.605332 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f56f7387-a1c9-4250-8abf-6c0a3830970e-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "f56f7387-a1c9-4250-8abf-6c0a3830970e" (UID: "f56f7387-a1c9-4250-8abf-6c0a3830970e"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:29:51 crc kubenswrapper[4907]: I0313 14:29:51.613040 4907 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f56f7387-a1c9-4250-8abf-6c0a3830970e-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:51 crc kubenswrapper[4907]: I0313 14:29:51.613077 4907 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f56f7387-a1c9-4250-8abf-6c0a3830970e-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:51 crc kubenswrapper[4907]: I0313 14:29:51.613095 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8c8lc\" (UniqueName: \"kubernetes.io/projected/f56f7387-a1c9-4250-8abf-6c0a3830970e-kube-api-access-8c8lc\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:51 crc kubenswrapper[4907]: I0313 14:29:51.613109 4907 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f56f7387-a1c9-4250-8abf-6c0a3830970e-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:51 crc kubenswrapper[4907]: I0313 14:29:51.613121 4907 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f56f7387-a1c9-4250-8abf-6c0a3830970e-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:51 crc kubenswrapper[4907]: I0313 14:29:51.614235 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f56f7387-a1c9-4250-8abf-6c0a3830970e-config" (OuterVolumeSpecName: "config") pod "f56f7387-a1c9-4250-8abf-6c0a3830970e" (UID: "f56f7387-a1c9-4250-8abf-6c0a3830970e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:29:51 crc kubenswrapper[4907]: I0313 14:29:51.621653 4907 generic.go:334] "Generic (PLEG): container finished" podID="f56f7387-a1c9-4250-8abf-6c0a3830970e" containerID="215938dfe070eab04984e52eb7b82e6d3f56a88ab88790cd28ad3b1a4bc7d77a" exitCode=0 Mar 13 14:29:51 crc kubenswrapper[4907]: I0313 14:29:51.623000 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-69b4446475-ttj9k" Mar 13 14:29:51 crc kubenswrapper[4907]: I0313 14:29:51.627949 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-69b4446475-ttj9k" event={"ID":"f56f7387-a1c9-4250-8abf-6c0a3830970e","Type":"ContainerDied","Data":"215938dfe070eab04984e52eb7b82e6d3f56a88ab88790cd28ad3b1a4bc7d77a"} Mar 13 14:29:51 crc kubenswrapper[4907]: I0313 14:29:51.628007 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Mar 13 14:29:51 crc kubenswrapper[4907]: I0313 14:29:51.628019 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-69b4446475-ttj9k" event={"ID":"f56f7387-a1c9-4250-8abf-6c0a3830970e","Type":"ContainerDied","Data":"c42518fd3cb49ffcc3d612c9faabeba7f93ee461bc56bb4d3b0a86fdeb96cbc5"} Mar 13 14:29:51 crc kubenswrapper[4907]: I0313 14:29:51.628041 4907 scope.go:117] "RemoveContainer" containerID="215938dfe070eab04984e52eb7b82e6d3f56a88ab88790cd28ad3b1a4bc7d77a" Mar 13 14:29:51 crc kubenswrapper[4907]: I0313 14:29:51.655742 4907 scope.go:117] "RemoveContainer" containerID="d71398ef9c9b9dc145195170a4e15d1e48bb6494190d13130bb61281addfc624" Mar 13 14:29:51 crc kubenswrapper[4907]: I0313 14:29:51.665920 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-69b4446475-ttj9k"] Mar 13 14:29:51 crc kubenswrapper[4907]: I0313 14:29:51.674891 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-69b4446475-ttj9k"] Mar 13 14:29:51 crc kubenswrapper[4907]: I0313 14:29:51.697750 4907 scope.go:117] "RemoveContainer" containerID="215938dfe070eab04984e52eb7b82e6d3f56a88ab88790cd28ad3b1a4bc7d77a" Mar 13 14:29:51 crc kubenswrapper[4907]: E0313 14:29:51.698162 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"215938dfe070eab04984e52eb7b82e6d3f56a88ab88790cd28ad3b1a4bc7d77a\": container with ID starting with 215938dfe070eab04984e52eb7b82e6d3f56a88ab88790cd28ad3b1a4bc7d77a not found: ID does not exist" containerID="215938dfe070eab04984e52eb7b82e6d3f56a88ab88790cd28ad3b1a4bc7d77a" Mar 13 14:29:51 crc kubenswrapper[4907]: I0313 14:29:51.698190 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"215938dfe070eab04984e52eb7b82e6d3f56a88ab88790cd28ad3b1a4bc7d77a"} err="failed to get container status \"215938dfe070eab04984e52eb7b82e6d3f56a88ab88790cd28ad3b1a4bc7d77a\": rpc error: code = NotFound desc = could not find container \"215938dfe070eab04984e52eb7b82e6d3f56a88ab88790cd28ad3b1a4bc7d77a\": container with ID starting with 215938dfe070eab04984e52eb7b82e6d3f56a88ab88790cd28ad3b1a4bc7d77a not found: ID does not exist" Mar 13 14:29:51 crc kubenswrapper[4907]: I0313 14:29:51.698213 4907 scope.go:117] "RemoveContainer" containerID="d71398ef9c9b9dc145195170a4e15d1e48bb6494190d13130bb61281addfc624" Mar 13 14:29:51 crc kubenswrapper[4907]: E0313 14:29:51.698401 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d71398ef9c9b9dc145195170a4e15d1e48bb6494190d13130bb61281addfc624\": container with ID starting with d71398ef9c9b9dc145195170a4e15d1e48bb6494190d13130bb61281addfc624 not found: ID does not exist" containerID="d71398ef9c9b9dc145195170a4e15d1e48bb6494190d13130bb61281addfc624" Mar 13 14:29:51 crc kubenswrapper[4907]: I0313 14:29:51.698418 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d71398ef9c9b9dc145195170a4e15d1e48bb6494190d13130bb61281addfc624"} err="failed to get container status \"d71398ef9c9b9dc145195170a4e15d1e48bb6494190d13130bb61281addfc624\": rpc error: code = NotFound desc = could not find container \"d71398ef9c9b9dc145195170a4e15d1e48bb6494190d13130bb61281addfc624\": container with ID starting with d71398ef9c9b9dc145195170a4e15d1e48bb6494190d13130bb61281addfc624 not found: ID does not exist" Mar 13 14:29:51 crc kubenswrapper[4907]: I0313 14:29:51.714983 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f56f7387-a1c9-4250-8abf-6c0a3830970e-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:51 crc kubenswrapper[4907]: I0313 14:29:51.805055 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f56f7387-a1c9-4250-8abf-6c0a3830970e" path="/var/lib/kubelet/pods/f56f7387-a1c9-4250-8abf-6c0a3830970e/volumes" Mar 13 14:29:54 crc kubenswrapper[4907]: I0313 14:29:54.650253 4907 generic.go:334] "Generic (PLEG): container finished" podID="8901c771-b45a-4147-9f96-cef784165281" containerID="714551bd9afe88fbb37b8d1ba1d86847559c10eb4afc2d8c8cade9bb4be375ee" exitCode=0 Mar 13 14:29:54 crc kubenswrapper[4907]: I0313 14:29:54.650344 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-2pvp9" event={"ID":"8901c771-b45a-4147-9f96-cef784165281","Type":"ContainerDied","Data":"714551bd9afe88fbb37b8d1ba1d86847559c10eb4afc2d8c8cade9bb4be375ee"} Mar 13 14:29:55 crc kubenswrapper[4907]: I0313 14:29:55.719293 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Mar 13 14:29:55 crc kubenswrapper[4907]: I0313 14:29:55.720625 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Mar 13 14:29:56 crc kubenswrapper[4907]: I0313 14:29:56.026410 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-2pvp9" Mar 13 14:29:56 crc kubenswrapper[4907]: I0313 14:29:56.099867 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8901c771-b45a-4147-9f96-cef784165281-scripts\") pod \"8901c771-b45a-4147-9f96-cef784165281\" (UID: \"8901c771-b45a-4147-9f96-cef784165281\") " Mar 13 14:29:56 crc kubenswrapper[4907]: I0313 14:29:56.100135 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8901c771-b45a-4147-9f96-cef784165281-config-data\") pod \"8901c771-b45a-4147-9f96-cef784165281\" (UID: \"8901c771-b45a-4147-9f96-cef784165281\") " Mar 13 14:29:56 crc kubenswrapper[4907]: I0313 14:29:56.100357 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h9nqx\" (UniqueName: \"kubernetes.io/projected/8901c771-b45a-4147-9f96-cef784165281-kube-api-access-h9nqx\") pod \"8901c771-b45a-4147-9f96-cef784165281\" (UID: \"8901c771-b45a-4147-9f96-cef784165281\") " Mar 13 14:29:56 crc kubenswrapper[4907]: I0313 14:29:56.100571 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8901c771-b45a-4147-9f96-cef784165281-combined-ca-bundle\") pod \"8901c771-b45a-4147-9f96-cef784165281\" (UID: \"8901c771-b45a-4147-9f96-cef784165281\") " Mar 13 14:29:56 crc kubenswrapper[4907]: I0313 14:29:56.105941 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8901c771-b45a-4147-9f96-cef784165281-scripts" (OuterVolumeSpecName: "scripts") pod "8901c771-b45a-4147-9f96-cef784165281" (UID: "8901c771-b45a-4147-9f96-cef784165281"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:29:56 crc kubenswrapper[4907]: I0313 14:29:56.106202 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8901c771-b45a-4147-9f96-cef784165281-kube-api-access-h9nqx" (OuterVolumeSpecName: "kube-api-access-h9nqx") pod "8901c771-b45a-4147-9f96-cef784165281" (UID: "8901c771-b45a-4147-9f96-cef784165281"). InnerVolumeSpecName "kube-api-access-h9nqx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:29:56 crc kubenswrapper[4907]: I0313 14:29:56.127369 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8901c771-b45a-4147-9f96-cef784165281-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8901c771-b45a-4147-9f96-cef784165281" (UID: "8901c771-b45a-4147-9f96-cef784165281"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:29:56 crc kubenswrapper[4907]: I0313 14:29:56.132698 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8901c771-b45a-4147-9f96-cef784165281-config-data" (OuterVolumeSpecName: "config-data") pod "8901c771-b45a-4147-9f96-cef784165281" (UID: "8901c771-b45a-4147-9f96-cef784165281"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:29:56 crc kubenswrapper[4907]: I0313 14:29:56.204970 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h9nqx\" (UniqueName: \"kubernetes.io/projected/8901c771-b45a-4147-9f96-cef784165281-kube-api-access-h9nqx\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:56 crc kubenswrapper[4907]: I0313 14:29:56.205204 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8901c771-b45a-4147-9f96-cef784165281-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:56 crc kubenswrapper[4907]: I0313 14:29:56.205260 4907 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8901c771-b45a-4147-9f96-cef784165281-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:56 crc kubenswrapper[4907]: I0313 14:29:56.205330 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8901c771-b45a-4147-9f96-cef784165281-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:56 crc kubenswrapper[4907]: I0313 14:29:56.669464 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-2pvp9" event={"ID":"8901c771-b45a-4147-9f96-cef784165281","Type":"ContainerDied","Data":"9a7bf9979b09ab5543692c5fb89940c7553ed9a8ca26785e17126dc549ce6605"} Mar 13 14:29:56 crc kubenswrapper[4907]: I0313 14:29:56.669819 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9a7bf9979b09ab5543692c5fb89940c7553ed9a8ca26785e17126dc549ce6605" Mar 13 14:29:56 crc kubenswrapper[4907]: I0313 14:29:56.669518 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-2pvp9" Mar 13 14:29:56 crc kubenswrapper[4907]: I0313 14:29:56.848549 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Mar 13 14:29:56 crc kubenswrapper[4907]: I0313 14:29:56.848897 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="4851fd7f-a2e6-4ec1-96ea-7c0086edd03e" containerName="nova-api-log" containerID="cri-o://c4919269912183df77206bdb4ba3361c6048b9cadeaad50a6bcce3ec26e11c53" gracePeriod=30 Mar 13 14:29:56 crc kubenswrapper[4907]: I0313 14:29:56.849054 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="4851fd7f-a2e6-4ec1-96ea-7c0086edd03e" containerName="nova-api-api" containerID="cri-o://8a9e8c8e4a79ece75fdb50c32528e4d6f87181c842fcd7d1b9802e179e72bd3b" gracePeriod=30 Mar 13 14:29:56 crc kubenswrapper[4907]: I0313 14:29:56.938900 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 14:29:56 crc kubenswrapper[4907]: I0313 14:29:56.939133 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="c0304716-7acd-4a2c-a018-73fc4bcbad51" containerName="nova-scheduler-scheduler" containerID="cri-o://7dd9875251f25eba8556dcd6141b77fd52b2025f417f75c1f8e455693aab178b" gracePeriod=30 Mar 13 14:29:56 crc kubenswrapper[4907]: I0313 14:29:56.954467 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.385403 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.528345 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4851fd7f-a2e6-4ec1-96ea-7c0086edd03e-internal-tls-certs\") pod \"4851fd7f-a2e6-4ec1-96ea-7c0086edd03e\" (UID: \"4851fd7f-a2e6-4ec1-96ea-7c0086edd03e\") " Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.528429 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4851fd7f-a2e6-4ec1-96ea-7c0086edd03e-config-data\") pod \"4851fd7f-a2e6-4ec1-96ea-7c0086edd03e\" (UID: \"4851fd7f-a2e6-4ec1-96ea-7c0086edd03e\") " Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.528510 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4851fd7f-a2e6-4ec1-96ea-7c0086edd03e-public-tls-certs\") pod \"4851fd7f-a2e6-4ec1-96ea-7c0086edd03e\" (UID: \"4851fd7f-a2e6-4ec1-96ea-7c0086edd03e\") " Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.528547 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q8tvk\" (UniqueName: \"kubernetes.io/projected/4851fd7f-a2e6-4ec1-96ea-7c0086edd03e-kube-api-access-q8tvk\") pod \"4851fd7f-a2e6-4ec1-96ea-7c0086edd03e\" (UID: \"4851fd7f-a2e6-4ec1-96ea-7c0086edd03e\") " Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.528633 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4851fd7f-a2e6-4ec1-96ea-7c0086edd03e-logs\") pod \"4851fd7f-a2e6-4ec1-96ea-7c0086edd03e\" (UID: \"4851fd7f-a2e6-4ec1-96ea-7c0086edd03e\") " Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.528722 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4851fd7f-a2e6-4ec1-96ea-7c0086edd03e-combined-ca-bundle\") pod \"4851fd7f-a2e6-4ec1-96ea-7c0086edd03e\" (UID: \"4851fd7f-a2e6-4ec1-96ea-7c0086edd03e\") " Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.533934 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4851fd7f-a2e6-4ec1-96ea-7c0086edd03e-kube-api-access-q8tvk" (OuterVolumeSpecName: "kube-api-access-q8tvk") pod "4851fd7f-a2e6-4ec1-96ea-7c0086edd03e" (UID: "4851fd7f-a2e6-4ec1-96ea-7c0086edd03e"). InnerVolumeSpecName "kube-api-access-q8tvk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.534305 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4851fd7f-a2e6-4ec1-96ea-7c0086edd03e-logs" (OuterVolumeSpecName: "logs") pod "4851fd7f-a2e6-4ec1-96ea-7c0086edd03e" (UID: "4851fd7f-a2e6-4ec1-96ea-7c0086edd03e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.561663 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4851fd7f-a2e6-4ec1-96ea-7c0086edd03e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4851fd7f-a2e6-4ec1-96ea-7c0086edd03e" (UID: "4851fd7f-a2e6-4ec1-96ea-7c0086edd03e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.561977 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4851fd7f-a2e6-4ec1-96ea-7c0086edd03e-config-data" (OuterVolumeSpecName: "config-data") pod "4851fd7f-a2e6-4ec1-96ea-7c0086edd03e" (UID: "4851fd7f-a2e6-4ec1-96ea-7c0086edd03e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.596527 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4851fd7f-a2e6-4ec1-96ea-7c0086edd03e-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "4851fd7f-a2e6-4ec1-96ea-7c0086edd03e" (UID: "4851fd7f-a2e6-4ec1-96ea-7c0086edd03e"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.598630 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4851fd7f-a2e6-4ec1-96ea-7c0086edd03e-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "4851fd7f-a2e6-4ec1-96ea-7c0086edd03e" (UID: "4851fd7f-a2e6-4ec1-96ea-7c0086edd03e"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.631188 4907 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4851fd7f-a2e6-4ec1-96ea-7c0086edd03e-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.631228 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4851fd7f-a2e6-4ec1-96ea-7c0086edd03e-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.631238 4907 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4851fd7f-a2e6-4ec1-96ea-7c0086edd03e-public-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.631251 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q8tvk\" (UniqueName: \"kubernetes.io/projected/4851fd7f-a2e6-4ec1-96ea-7c0086edd03e-kube-api-access-q8tvk\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.631263 4907 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4851fd7f-a2e6-4ec1-96ea-7c0086edd03e-logs\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.631272 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4851fd7f-a2e6-4ec1-96ea-7c0086edd03e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.678832 4907 generic.go:334] "Generic (PLEG): container finished" podID="4851fd7f-a2e6-4ec1-96ea-7c0086edd03e" containerID="8a9e8c8e4a79ece75fdb50c32528e4d6f87181c842fcd7d1b9802e179e72bd3b" exitCode=0 Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.678867 4907 generic.go:334] "Generic (PLEG): container finished" podID="4851fd7f-a2e6-4ec1-96ea-7c0086edd03e" containerID="c4919269912183df77206bdb4ba3361c6048b9cadeaad50a6bcce3ec26e11c53" exitCode=143 Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.678904 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.678903 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4851fd7f-a2e6-4ec1-96ea-7c0086edd03e","Type":"ContainerDied","Data":"8a9e8c8e4a79ece75fdb50c32528e4d6f87181c842fcd7d1b9802e179e72bd3b"} Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.678959 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4851fd7f-a2e6-4ec1-96ea-7c0086edd03e","Type":"ContainerDied","Data":"c4919269912183df77206bdb4ba3361c6048b9cadeaad50a6bcce3ec26e11c53"} Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.678974 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4851fd7f-a2e6-4ec1-96ea-7c0086edd03e","Type":"ContainerDied","Data":"2712a8612ce027f409c3dbd5c6adbabefb48875be5bd50702de7d61655bda2a8"} Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.678993 4907 scope.go:117] "RemoveContainer" containerID="8a9e8c8e4a79ece75fdb50c32528e4d6f87181c842fcd7d1b9802e179e72bd3b" Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.679512 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="39488342-aa1a-4dba-bd5f-556da5081904" containerName="nova-metadata-log" containerID="cri-o://ee178515d58e37a284bcf6959726bc0672f424f0bddc3d61b496fcc0bf679507" gracePeriod=30 Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.679633 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="39488342-aa1a-4dba-bd5f-556da5081904" containerName="nova-metadata-metadata" containerID="cri-o://3f0638d25e0be962bf528749f86e7ed7d982188df63f11eef31f67ffdeb4e143" gracePeriod=30 Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.705760 4907 scope.go:117] "RemoveContainer" containerID="c4919269912183df77206bdb4ba3361c6048b9cadeaad50a6bcce3ec26e11c53" Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.720301 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.729398 4907 scope.go:117] "RemoveContainer" containerID="8a9e8c8e4a79ece75fdb50c32528e4d6f87181c842fcd7d1b9802e179e72bd3b" Mar 13 14:29:57 crc kubenswrapper[4907]: E0313 14:29:57.730042 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8a9e8c8e4a79ece75fdb50c32528e4d6f87181c842fcd7d1b9802e179e72bd3b\": container with ID starting with 8a9e8c8e4a79ece75fdb50c32528e4d6f87181c842fcd7d1b9802e179e72bd3b not found: ID does not exist" containerID="8a9e8c8e4a79ece75fdb50c32528e4d6f87181c842fcd7d1b9802e179e72bd3b" Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.730096 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8a9e8c8e4a79ece75fdb50c32528e4d6f87181c842fcd7d1b9802e179e72bd3b"} err="failed to get container status \"8a9e8c8e4a79ece75fdb50c32528e4d6f87181c842fcd7d1b9802e179e72bd3b\": rpc error: code = NotFound desc = could not find container \"8a9e8c8e4a79ece75fdb50c32528e4d6f87181c842fcd7d1b9802e179e72bd3b\": container with ID starting with 8a9e8c8e4a79ece75fdb50c32528e4d6f87181c842fcd7d1b9802e179e72bd3b not found: ID does not exist" Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.730123 4907 scope.go:117] "RemoveContainer" containerID="c4919269912183df77206bdb4ba3361c6048b9cadeaad50a6bcce3ec26e11c53" Mar 13 14:29:57 crc kubenswrapper[4907]: E0313 14:29:57.730544 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c4919269912183df77206bdb4ba3361c6048b9cadeaad50a6bcce3ec26e11c53\": container with ID starting with c4919269912183df77206bdb4ba3361c6048b9cadeaad50a6bcce3ec26e11c53 not found: ID does not exist" containerID="c4919269912183df77206bdb4ba3361c6048b9cadeaad50a6bcce3ec26e11c53" Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.730581 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c4919269912183df77206bdb4ba3361c6048b9cadeaad50a6bcce3ec26e11c53"} err="failed to get container status \"c4919269912183df77206bdb4ba3361c6048b9cadeaad50a6bcce3ec26e11c53\": rpc error: code = NotFound desc = could not find container \"c4919269912183df77206bdb4ba3361c6048b9cadeaad50a6bcce3ec26e11c53\": container with ID starting with c4919269912183df77206bdb4ba3361c6048b9cadeaad50a6bcce3ec26e11c53 not found: ID does not exist" Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.730608 4907 scope.go:117] "RemoveContainer" containerID="8a9e8c8e4a79ece75fdb50c32528e4d6f87181c842fcd7d1b9802e179e72bd3b" Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.730925 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8a9e8c8e4a79ece75fdb50c32528e4d6f87181c842fcd7d1b9802e179e72bd3b"} err="failed to get container status \"8a9e8c8e4a79ece75fdb50c32528e4d6f87181c842fcd7d1b9802e179e72bd3b\": rpc error: code = NotFound desc = could not find container \"8a9e8c8e4a79ece75fdb50c32528e4d6f87181c842fcd7d1b9802e179e72bd3b\": container with ID starting with 8a9e8c8e4a79ece75fdb50c32528e4d6f87181c842fcd7d1b9802e179e72bd3b not found: ID does not exist" Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.730944 4907 scope.go:117] "RemoveContainer" containerID="c4919269912183df77206bdb4ba3361c6048b9cadeaad50a6bcce3ec26e11c53" Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.731279 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c4919269912183df77206bdb4ba3361c6048b9cadeaad50a6bcce3ec26e11c53"} err="failed to get container status \"c4919269912183df77206bdb4ba3361c6048b9cadeaad50a6bcce3ec26e11c53\": rpc error: code = NotFound desc = could not find container \"c4919269912183df77206bdb4ba3361c6048b9cadeaad50a6bcce3ec26e11c53\": container with ID starting with c4919269912183df77206bdb4ba3361c6048b9cadeaad50a6bcce3ec26e11c53 not found: ID does not exist" Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.732554 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.743698 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Mar 13 14:29:57 crc kubenswrapper[4907]: E0313 14:29:57.744168 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8901c771-b45a-4147-9f96-cef784165281" containerName="nova-manage" Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.744189 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="8901c771-b45a-4147-9f96-cef784165281" containerName="nova-manage" Mar 13 14:29:57 crc kubenswrapper[4907]: E0313 14:29:57.744203 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4851fd7f-a2e6-4ec1-96ea-7c0086edd03e" containerName="nova-api-log" Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.744220 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="4851fd7f-a2e6-4ec1-96ea-7c0086edd03e" containerName="nova-api-log" Mar 13 14:29:57 crc kubenswrapper[4907]: E0313 14:29:57.744233 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4851fd7f-a2e6-4ec1-96ea-7c0086edd03e" containerName="nova-api-api" Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.744242 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="4851fd7f-a2e6-4ec1-96ea-7c0086edd03e" containerName="nova-api-api" Mar 13 14:29:57 crc kubenswrapper[4907]: E0313 14:29:57.744258 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f56f7387-a1c9-4250-8abf-6c0a3830970e" containerName="dnsmasq-dns" Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.744266 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="f56f7387-a1c9-4250-8abf-6c0a3830970e" containerName="dnsmasq-dns" Mar 13 14:29:57 crc kubenswrapper[4907]: E0313 14:29:57.744292 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f56f7387-a1c9-4250-8abf-6c0a3830970e" containerName="init" Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.744301 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="f56f7387-a1c9-4250-8abf-6c0a3830970e" containerName="init" Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.744523 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="4851fd7f-a2e6-4ec1-96ea-7c0086edd03e" containerName="nova-api-log" Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.744553 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="f56f7387-a1c9-4250-8abf-6c0a3830970e" containerName="dnsmasq-dns" Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.744568 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="4851fd7f-a2e6-4ec1-96ea-7c0086edd03e" containerName="nova-api-api" Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.744585 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="8901c771-b45a-4147-9f96-cef784165281" containerName="nova-manage" Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.745653 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.749612 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.750640 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.750765 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.757937 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.811005 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4851fd7f-a2e6-4ec1-96ea-7c0086edd03e" path="/var/lib/kubelet/pods/4851fd7f-a2e6-4ec1-96ea-7c0086edd03e/volumes" Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.834195 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cc227d33-3bf1-4b5f-8368-8e6b8a512f7d-logs\") pod \"nova-api-0\" (UID: \"cc227d33-3bf1-4b5f-8368-8e6b8a512f7d\") " pod="openstack/nova-api-0" Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.834248 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cc227d33-3bf1-4b5f-8368-8e6b8a512f7d-public-tls-certs\") pod \"nova-api-0\" (UID: \"cc227d33-3bf1-4b5f-8368-8e6b8a512f7d\") " pod="openstack/nova-api-0" Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.834393 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-565tg\" (UniqueName: \"kubernetes.io/projected/cc227d33-3bf1-4b5f-8368-8e6b8a512f7d-kube-api-access-565tg\") pod \"nova-api-0\" (UID: \"cc227d33-3bf1-4b5f-8368-8e6b8a512f7d\") " pod="openstack/nova-api-0" Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.834639 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc227d33-3bf1-4b5f-8368-8e6b8a512f7d-config-data\") pod \"nova-api-0\" (UID: \"cc227d33-3bf1-4b5f-8368-8e6b8a512f7d\") " pod="openstack/nova-api-0" Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.834789 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc227d33-3bf1-4b5f-8368-8e6b8a512f7d-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"cc227d33-3bf1-4b5f-8368-8e6b8a512f7d\") " pod="openstack/nova-api-0" Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.834923 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cc227d33-3bf1-4b5f-8368-8e6b8a512f7d-internal-tls-certs\") pod \"nova-api-0\" (UID: \"cc227d33-3bf1-4b5f-8368-8e6b8a512f7d\") " pod="openstack/nova-api-0" Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.936438 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cc227d33-3bf1-4b5f-8368-8e6b8a512f7d-logs\") pod \"nova-api-0\" (UID: \"cc227d33-3bf1-4b5f-8368-8e6b8a512f7d\") " pod="openstack/nova-api-0" Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.936498 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cc227d33-3bf1-4b5f-8368-8e6b8a512f7d-public-tls-certs\") pod \"nova-api-0\" (UID: \"cc227d33-3bf1-4b5f-8368-8e6b8a512f7d\") " pod="openstack/nova-api-0" Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.936548 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-565tg\" (UniqueName: \"kubernetes.io/projected/cc227d33-3bf1-4b5f-8368-8e6b8a512f7d-kube-api-access-565tg\") pod \"nova-api-0\" (UID: \"cc227d33-3bf1-4b5f-8368-8e6b8a512f7d\") " pod="openstack/nova-api-0" Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.936618 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc227d33-3bf1-4b5f-8368-8e6b8a512f7d-config-data\") pod \"nova-api-0\" (UID: \"cc227d33-3bf1-4b5f-8368-8e6b8a512f7d\") " pod="openstack/nova-api-0" Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.936652 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc227d33-3bf1-4b5f-8368-8e6b8a512f7d-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"cc227d33-3bf1-4b5f-8368-8e6b8a512f7d\") " pod="openstack/nova-api-0" Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.936687 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cc227d33-3bf1-4b5f-8368-8e6b8a512f7d-internal-tls-certs\") pod \"nova-api-0\" (UID: \"cc227d33-3bf1-4b5f-8368-8e6b8a512f7d\") " pod="openstack/nova-api-0" Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.936925 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cc227d33-3bf1-4b5f-8368-8e6b8a512f7d-logs\") pod \"nova-api-0\" (UID: \"cc227d33-3bf1-4b5f-8368-8e6b8a512f7d\") " pod="openstack/nova-api-0" Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.941507 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cc227d33-3bf1-4b5f-8368-8e6b8a512f7d-internal-tls-certs\") pod \"nova-api-0\" (UID: \"cc227d33-3bf1-4b5f-8368-8e6b8a512f7d\") " pod="openstack/nova-api-0" Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.942100 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc227d33-3bf1-4b5f-8368-8e6b8a512f7d-config-data\") pod \"nova-api-0\" (UID: \"cc227d33-3bf1-4b5f-8368-8e6b8a512f7d\") " pod="openstack/nova-api-0" Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.942116 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cc227d33-3bf1-4b5f-8368-8e6b8a512f7d-public-tls-certs\") pod \"nova-api-0\" (UID: \"cc227d33-3bf1-4b5f-8368-8e6b8a512f7d\") " pod="openstack/nova-api-0" Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.942749 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc227d33-3bf1-4b5f-8368-8e6b8a512f7d-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"cc227d33-3bf1-4b5f-8368-8e6b8a512f7d\") " pod="openstack/nova-api-0" Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.958326 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-565tg\" (UniqueName: \"kubernetes.io/projected/cc227d33-3bf1-4b5f-8368-8e6b8a512f7d-kube-api-access-565tg\") pod \"nova-api-0\" (UID: \"cc227d33-3bf1-4b5f-8368-8e6b8a512f7d\") " pod="openstack/nova-api-0" Mar 13 14:29:58 crc kubenswrapper[4907]: I0313 14:29:58.062931 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 13 14:29:58 crc kubenswrapper[4907]: E0313 14:29:58.466464 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="7dd9875251f25eba8556dcd6141b77fd52b2025f417f75c1f8e455693aab178b" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Mar 13 14:29:58 crc kubenswrapper[4907]: E0313 14:29:58.468166 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="7dd9875251f25eba8556dcd6141b77fd52b2025f417f75c1f8e455693aab178b" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Mar 13 14:29:58 crc kubenswrapper[4907]: E0313 14:29:58.469676 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="7dd9875251f25eba8556dcd6141b77fd52b2025f417f75c1f8e455693aab178b" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Mar 13 14:29:58 crc kubenswrapper[4907]: E0313 14:29:58.469707 4907 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="c0304716-7acd-4a2c-a018-73fc4bcbad51" containerName="nova-scheduler-scheduler" Mar 13 14:29:58 crc kubenswrapper[4907]: I0313 14:29:58.514134 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 13 14:29:58 crc kubenswrapper[4907]: I0313 14:29:58.691989 4907 generic.go:334] "Generic (PLEG): container finished" podID="39488342-aa1a-4dba-bd5f-556da5081904" containerID="ee178515d58e37a284bcf6959726bc0672f424f0bddc3d61b496fcc0bf679507" exitCode=143 Mar 13 14:29:58 crc kubenswrapper[4907]: I0313 14:29:58.692069 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"39488342-aa1a-4dba-bd5f-556da5081904","Type":"ContainerDied","Data":"ee178515d58e37a284bcf6959726bc0672f424f0bddc3d61b496fcc0bf679507"} Mar 13 14:29:58 crc kubenswrapper[4907]: I0313 14:29:58.693786 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"cc227d33-3bf1-4b5f-8368-8e6b8a512f7d","Type":"ContainerStarted","Data":"ed58fd3bb4c7be310b9f67af6133d99b09325a11963ab01983393f4f0a20f895"} Mar 13 14:29:59 crc kubenswrapper[4907]: I0313 14:29:59.710368 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"cc227d33-3bf1-4b5f-8368-8e6b8a512f7d","Type":"ContainerStarted","Data":"58121d89217f995e7e932419445b98334a3ccb8b202e18c097026d64131a91fd"} Mar 13 14:29:59 crc kubenswrapper[4907]: I0313 14:29:59.710416 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"cc227d33-3bf1-4b5f-8368-8e6b8a512f7d","Type":"ContainerStarted","Data":"6c1fdac6667e489d30e980acdd57d1cf0fa750bc21c6f46f11adb992a74fe2bd"} Mar 13 14:30:00 crc kubenswrapper[4907]: I0313 14:30:00.134430 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.134406004 podStartE2EDuration="3.134406004s" podCreationTimestamp="2026-03-13 14:29:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:29:59.757490563 +0000 UTC m=+1498.657278252" watchObservedRunningTime="2026-03-13 14:30:00.134406004 +0000 UTC m=+1499.034193703" Mar 13 14:30:00 crc kubenswrapper[4907]: I0313 14:30:00.145122 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556870-8mlkz"] Mar 13 14:30:00 crc kubenswrapper[4907]: I0313 14:30:00.146565 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556870-8mlkz" Mar 13 14:30:00 crc kubenswrapper[4907]: I0313 14:30:00.153097 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 14:30:00 crc kubenswrapper[4907]: I0313 14:30:00.153227 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 14:30:00 crc kubenswrapper[4907]: I0313 14:30:00.153310 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 14:30:00 crc kubenswrapper[4907]: I0313 14:30:00.153795 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556870-rdhxg"] Mar 13 14:30:00 crc kubenswrapper[4907]: I0313 14:30:00.155292 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556870-rdhxg" Mar 13 14:30:00 crc kubenswrapper[4907]: I0313 14:30:00.160274 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 13 14:30:00 crc kubenswrapper[4907]: I0313 14:30:00.160538 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 13 14:30:00 crc kubenswrapper[4907]: I0313 14:30:00.164569 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556870-8mlkz"] Mar 13 14:30:00 crc kubenswrapper[4907]: I0313 14:30:00.184473 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556870-rdhxg"] Mar 13 14:30:00 crc kubenswrapper[4907]: I0313 14:30:00.284308 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/af9d29dd-6749-4fcd-82a8-ea528d832fcc-config-volume\") pod \"collect-profiles-29556870-rdhxg\" (UID: \"af9d29dd-6749-4fcd-82a8-ea528d832fcc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556870-rdhxg" Mar 13 14:30:00 crc kubenswrapper[4907]: I0313 14:30:00.284542 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/af9d29dd-6749-4fcd-82a8-ea528d832fcc-secret-volume\") pod \"collect-profiles-29556870-rdhxg\" (UID: \"af9d29dd-6749-4fcd-82a8-ea528d832fcc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556870-rdhxg" Mar 13 14:30:00 crc kubenswrapper[4907]: I0313 14:30:00.284794 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d6z4z\" (UniqueName: \"kubernetes.io/projected/af9d29dd-6749-4fcd-82a8-ea528d832fcc-kube-api-access-d6z4z\") pod \"collect-profiles-29556870-rdhxg\" (UID: \"af9d29dd-6749-4fcd-82a8-ea528d832fcc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556870-rdhxg" Mar 13 14:30:00 crc kubenswrapper[4907]: I0313 14:30:00.284986 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2kxlz\" (UniqueName: \"kubernetes.io/projected/e7993f31-1072-4c83-a3ac-16dca5eb04a3-kube-api-access-2kxlz\") pod \"auto-csr-approver-29556870-8mlkz\" (UID: \"e7993f31-1072-4c83-a3ac-16dca5eb04a3\") " pod="openshift-infra/auto-csr-approver-29556870-8mlkz" Mar 13 14:30:00 crc kubenswrapper[4907]: I0313 14:30:00.386703 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/af9d29dd-6749-4fcd-82a8-ea528d832fcc-secret-volume\") pod \"collect-profiles-29556870-rdhxg\" (UID: \"af9d29dd-6749-4fcd-82a8-ea528d832fcc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556870-rdhxg" Mar 13 14:30:00 crc kubenswrapper[4907]: I0313 14:30:00.386798 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d6z4z\" (UniqueName: \"kubernetes.io/projected/af9d29dd-6749-4fcd-82a8-ea528d832fcc-kube-api-access-d6z4z\") pod \"collect-profiles-29556870-rdhxg\" (UID: \"af9d29dd-6749-4fcd-82a8-ea528d832fcc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556870-rdhxg" Mar 13 14:30:00 crc kubenswrapper[4907]: I0313 14:30:00.386862 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2kxlz\" (UniqueName: \"kubernetes.io/projected/e7993f31-1072-4c83-a3ac-16dca5eb04a3-kube-api-access-2kxlz\") pod \"auto-csr-approver-29556870-8mlkz\" (UID: \"e7993f31-1072-4c83-a3ac-16dca5eb04a3\") " pod="openshift-infra/auto-csr-approver-29556870-8mlkz" Mar 13 14:30:00 crc kubenswrapper[4907]: I0313 14:30:00.386944 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/af9d29dd-6749-4fcd-82a8-ea528d832fcc-config-volume\") pod \"collect-profiles-29556870-rdhxg\" (UID: \"af9d29dd-6749-4fcd-82a8-ea528d832fcc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556870-rdhxg" Mar 13 14:30:00 crc kubenswrapper[4907]: I0313 14:30:00.388070 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/af9d29dd-6749-4fcd-82a8-ea528d832fcc-config-volume\") pod \"collect-profiles-29556870-rdhxg\" (UID: \"af9d29dd-6749-4fcd-82a8-ea528d832fcc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556870-rdhxg" Mar 13 14:30:00 crc kubenswrapper[4907]: I0313 14:30:00.395535 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/af9d29dd-6749-4fcd-82a8-ea528d832fcc-secret-volume\") pod \"collect-profiles-29556870-rdhxg\" (UID: \"af9d29dd-6749-4fcd-82a8-ea528d832fcc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556870-rdhxg" Mar 13 14:30:00 crc kubenswrapper[4907]: I0313 14:30:00.403982 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d6z4z\" (UniqueName: \"kubernetes.io/projected/af9d29dd-6749-4fcd-82a8-ea528d832fcc-kube-api-access-d6z4z\") pod \"collect-profiles-29556870-rdhxg\" (UID: \"af9d29dd-6749-4fcd-82a8-ea528d832fcc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556870-rdhxg" Mar 13 14:30:00 crc kubenswrapper[4907]: I0313 14:30:00.404670 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2kxlz\" (UniqueName: \"kubernetes.io/projected/e7993f31-1072-4c83-a3ac-16dca5eb04a3-kube-api-access-2kxlz\") pod \"auto-csr-approver-29556870-8mlkz\" (UID: \"e7993f31-1072-4c83-a3ac-16dca5eb04a3\") " pod="openshift-infra/auto-csr-approver-29556870-8mlkz" Mar 13 14:30:00 crc kubenswrapper[4907]: I0313 14:30:00.475902 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556870-8mlkz" Mar 13 14:30:00 crc kubenswrapper[4907]: I0313 14:30:00.495219 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556870-rdhxg" Mar 13 14:30:00 crc kubenswrapper[4907]: I0313 14:30:00.964274 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556870-8mlkz"] Mar 13 14:30:00 crc kubenswrapper[4907]: W0313 14:30:00.966816 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode7993f31_1072_4c83_a3ac_16dca5eb04a3.slice/crio-0fff58df59ba6aeb1bad1085410d3a35e667921192b7b80aca529cb3350038da WatchSource:0}: Error finding container 0fff58df59ba6aeb1bad1085410d3a35e667921192b7b80aca529cb3350038da: Status 404 returned error can't find the container with id 0fff58df59ba6aeb1bad1085410d3a35e667921192b7b80aca529cb3350038da Mar 13 14:30:00 crc kubenswrapper[4907]: I0313 14:30:00.972390 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556870-rdhxg"] Mar 13 14:30:00 crc kubenswrapper[4907]: W0313 14:30:00.983134 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaf9d29dd_6749_4fcd_82a8_ea528d832fcc.slice/crio-7c2a8f78d2f024990112bf064e09dd2f29128e2c064da8d891a6688df0ffe224 WatchSource:0}: Error finding container 7c2a8f78d2f024990112bf064e09dd2f29128e2c064da8d891a6688df0ffe224: Status 404 returned error can't find the container with id 7c2a8f78d2f024990112bf064e09dd2f29128e2c064da8d891a6688df0ffe224 Mar 13 14:30:01 crc kubenswrapper[4907]: I0313 14:30:01.227672 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 13 14:30:01 crc kubenswrapper[4907]: I0313 14:30:01.303237 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-774w9\" (UniqueName: \"kubernetes.io/projected/39488342-aa1a-4dba-bd5f-556da5081904-kube-api-access-774w9\") pod \"39488342-aa1a-4dba-bd5f-556da5081904\" (UID: \"39488342-aa1a-4dba-bd5f-556da5081904\") " Mar 13 14:30:01 crc kubenswrapper[4907]: I0313 14:30:01.303365 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39488342-aa1a-4dba-bd5f-556da5081904-config-data\") pod \"39488342-aa1a-4dba-bd5f-556da5081904\" (UID: \"39488342-aa1a-4dba-bd5f-556da5081904\") " Mar 13 14:30:01 crc kubenswrapper[4907]: I0313 14:30:01.303417 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/39488342-aa1a-4dba-bd5f-556da5081904-nova-metadata-tls-certs\") pod \"39488342-aa1a-4dba-bd5f-556da5081904\" (UID: \"39488342-aa1a-4dba-bd5f-556da5081904\") " Mar 13 14:30:01 crc kubenswrapper[4907]: I0313 14:30:01.303548 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/39488342-aa1a-4dba-bd5f-556da5081904-logs\") pod \"39488342-aa1a-4dba-bd5f-556da5081904\" (UID: \"39488342-aa1a-4dba-bd5f-556da5081904\") " Mar 13 14:30:01 crc kubenswrapper[4907]: I0313 14:30:01.303676 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39488342-aa1a-4dba-bd5f-556da5081904-combined-ca-bundle\") pod \"39488342-aa1a-4dba-bd5f-556da5081904\" (UID: \"39488342-aa1a-4dba-bd5f-556da5081904\") " Mar 13 14:30:01 crc kubenswrapper[4907]: I0313 14:30:01.304109 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/39488342-aa1a-4dba-bd5f-556da5081904-logs" (OuterVolumeSpecName: "logs") pod "39488342-aa1a-4dba-bd5f-556da5081904" (UID: "39488342-aa1a-4dba-bd5f-556da5081904"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:30:01 crc kubenswrapper[4907]: I0313 14:30:01.304378 4907 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/39488342-aa1a-4dba-bd5f-556da5081904-logs\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:01 crc kubenswrapper[4907]: I0313 14:30:01.322916 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/39488342-aa1a-4dba-bd5f-556da5081904-kube-api-access-774w9" (OuterVolumeSpecName: "kube-api-access-774w9") pod "39488342-aa1a-4dba-bd5f-556da5081904" (UID: "39488342-aa1a-4dba-bd5f-556da5081904"). InnerVolumeSpecName "kube-api-access-774w9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:30:01 crc kubenswrapper[4907]: I0313 14:30:01.343956 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/39488342-aa1a-4dba-bd5f-556da5081904-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "39488342-aa1a-4dba-bd5f-556da5081904" (UID: "39488342-aa1a-4dba-bd5f-556da5081904"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:01 crc kubenswrapper[4907]: I0313 14:30:01.347134 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/39488342-aa1a-4dba-bd5f-556da5081904-config-data" (OuterVolumeSpecName: "config-data") pod "39488342-aa1a-4dba-bd5f-556da5081904" (UID: "39488342-aa1a-4dba-bd5f-556da5081904"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:01 crc kubenswrapper[4907]: I0313 14:30:01.373162 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/39488342-aa1a-4dba-bd5f-556da5081904-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "39488342-aa1a-4dba-bd5f-556da5081904" (UID: "39488342-aa1a-4dba-bd5f-556da5081904"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:01 crc kubenswrapper[4907]: I0313 14:30:01.405663 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39488342-aa1a-4dba-bd5f-556da5081904-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:01 crc kubenswrapper[4907]: I0313 14:30:01.405702 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-774w9\" (UniqueName: \"kubernetes.io/projected/39488342-aa1a-4dba-bd5f-556da5081904-kube-api-access-774w9\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:01 crc kubenswrapper[4907]: I0313 14:30:01.405720 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39488342-aa1a-4dba-bd5f-556da5081904-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:01 crc kubenswrapper[4907]: I0313 14:30:01.405730 4907 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/39488342-aa1a-4dba-bd5f-556da5081904-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:01 crc kubenswrapper[4907]: I0313 14:30:01.732098 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556870-8mlkz" event={"ID":"e7993f31-1072-4c83-a3ac-16dca5eb04a3","Type":"ContainerStarted","Data":"0fff58df59ba6aeb1bad1085410d3a35e667921192b7b80aca529cb3350038da"} Mar 13 14:30:01 crc kubenswrapper[4907]: I0313 14:30:01.734782 4907 generic.go:334] "Generic (PLEG): container finished" podID="39488342-aa1a-4dba-bd5f-556da5081904" containerID="3f0638d25e0be962bf528749f86e7ed7d982188df63f11eef31f67ffdeb4e143" exitCode=0 Mar 13 14:30:01 crc kubenswrapper[4907]: I0313 14:30:01.734852 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"39488342-aa1a-4dba-bd5f-556da5081904","Type":"ContainerDied","Data":"3f0638d25e0be962bf528749f86e7ed7d982188df63f11eef31f67ffdeb4e143"} Mar 13 14:30:01 crc kubenswrapper[4907]: I0313 14:30:01.734904 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"39488342-aa1a-4dba-bd5f-556da5081904","Type":"ContainerDied","Data":"a7aca03c6cc93dfbda04afd1635961a9edebe57ed61e75bfcd08d8cd3550955a"} Mar 13 14:30:01 crc kubenswrapper[4907]: I0313 14:30:01.734916 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 13 14:30:01 crc kubenswrapper[4907]: I0313 14:30:01.734934 4907 scope.go:117] "RemoveContainer" containerID="3f0638d25e0be962bf528749f86e7ed7d982188df63f11eef31f67ffdeb4e143" Mar 13 14:30:01 crc kubenswrapper[4907]: I0313 14:30:01.739560 4907 generic.go:334] "Generic (PLEG): container finished" podID="af9d29dd-6749-4fcd-82a8-ea528d832fcc" containerID="46dfd46e5c29fa7764c024392038bd85a0e04092c9a231bbd08b67969852ba5c" exitCode=0 Mar 13 14:30:01 crc kubenswrapper[4907]: I0313 14:30:01.739609 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556870-rdhxg" event={"ID":"af9d29dd-6749-4fcd-82a8-ea528d832fcc","Type":"ContainerDied","Data":"46dfd46e5c29fa7764c024392038bd85a0e04092c9a231bbd08b67969852ba5c"} Mar 13 14:30:01 crc kubenswrapper[4907]: I0313 14:30:01.739636 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556870-rdhxg" event={"ID":"af9d29dd-6749-4fcd-82a8-ea528d832fcc","Type":"ContainerStarted","Data":"7c2a8f78d2f024990112bf064e09dd2f29128e2c064da8d891a6688df0ffe224"} Mar 13 14:30:01 crc kubenswrapper[4907]: I0313 14:30:01.767099 4907 scope.go:117] "RemoveContainer" containerID="ee178515d58e37a284bcf6959726bc0672f424f0bddc3d61b496fcc0bf679507" Mar 13 14:30:01 crc kubenswrapper[4907]: I0313 14:30:01.795772 4907 scope.go:117] "RemoveContainer" containerID="3f0638d25e0be962bf528749f86e7ed7d982188df63f11eef31f67ffdeb4e143" Mar 13 14:30:01 crc kubenswrapper[4907]: E0313 14:30:01.796235 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3f0638d25e0be962bf528749f86e7ed7d982188df63f11eef31f67ffdeb4e143\": container with ID starting with 3f0638d25e0be962bf528749f86e7ed7d982188df63f11eef31f67ffdeb4e143 not found: ID does not exist" containerID="3f0638d25e0be962bf528749f86e7ed7d982188df63f11eef31f67ffdeb4e143" Mar 13 14:30:01 crc kubenswrapper[4907]: I0313 14:30:01.796265 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3f0638d25e0be962bf528749f86e7ed7d982188df63f11eef31f67ffdeb4e143"} err="failed to get container status \"3f0638d25e0be962bf528749f86e7ed7d982188df63f11eef31f67ffdeb4e143\": rpc error: code = NotFound desc = could not find container \"3f0638d25e0be962bf528749f86e7ed7d982188df63f11eef31f67ffdeb4e143\": container with ID starting with 3f0638d25e0be962bf528749f86e7ed7d982188df63f11eef31f67ffdeb4e143 not found: ID does not exist" Mar 13 14:30:01 crc kubenswrapper[4907]: I0313 14:30:01.796283 4907 scope.go:117] "RemoveContainer" containerID="ee178515d58e37a284bcf6959726bc0672f424f0bddc3d61b496fcc0bf679507" Mar 13 14:30:01 crc kubenswrapper[4907]: E0313 14:30:01.796533 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ee178515d58e37a284bcf6959726bc0672f424f0bddc3d61b496fcc0bf679507\": container with ID starting with ee178515d58e37a284bcf6959726bc0672f424f0bddc3d61b496fcc0bf679507 not found: ID does not exist" containerID="ee178515d58e37a284bcf6959726bc0672f424f0bddc3d61b496fcc0bf679507" Mar 13 14:30:01 crc kubenswrapper[4907]: I0313 14:30:01.796567 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ee178515d58e37a284bcf6959726bc0672f424f0bddc3d61b496fcc0bf679507"} err="failed to get container status \"ee178515d58e37a284bcf6959726bc0672f424f0bddc3d61b496fcc0bf679507\": rpc error: code = NotFound desc = could not find container \"ee178515d58e37a284bcf6959726bc0672f424f0bddc3d61b496fcc0bf679507\": container with ID starting with ee178515d58e37a284bcf6959726bc0672f424f0bddc3d61b496fcc0bf679507 not found: ID does not exist" Mar 13 14:30:01 crc kubenswrapper[4907]: I0313 14:30:01.823827 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Mar 13 14:30:01 crc kubenswrapper[4907]: I0313 14:30:01.823911 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Mar 13 14:30:01 crc kubenswrapper[4907]: I0313 14:30:01.823931 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Mar 13 14:30:01 crc kubenswrapper[4907]: E0313 14:30:01.824559 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39488342-aa1a-4dba-bd5f-556da5081904" containerName="nova-metadata-metadata" Mar 13 14:30:01 crc kubenswrapper[4907]: I0313 14:30:01.824581 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="39488342-aa1a-4dba-bd5f-556da5081904" containerName="nova-metadata-metadata" Mar 13 14:30:01 crc kubenswrapper[4907]: E0313 14:30:01.824639 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39488342-aa1a-4dba-bd5f-556da5081904" containerName="nova-metadata-log" Mar 13 14:30:01 crc kubenswrapper[4907]: I0313 14:30:01.824649 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="39488342-aa1a-4dba-bd5f-556da5081904" containerName="nova-metadata-log" Mar 13 14:30:01 crc kubenswrapper[4907]: I0313 14:30:01.825017 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="39488342-aa1a-4dba-bd5f-556da5081904" containerName="nova-metadata-metadata" Mar 13 14:30:01 crc kubenswrapper[4907]: I0313 14:30:01.825067 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="39488342-aa1a-4dba-bd5f-556da5081904" containerName="nova-metadata-log" Mar 13 14:30:01 crc kubenswrapper[4907]: I0313 14:30:01.830259 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 13 14:30:01 crc kubenswrapper[4907]: I0313 14:30:01.832586 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Mar 13 14:30:01 crc kubenswrapper[4907]: I0313 14:30:01.833117 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Mar 13 14:30:01 crc kubenswrapper[4907]: I0313 14:30:01.846542 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Mar 13 14:30:01 crc kubenswrapper[4907]: I0313 14:30:01.923034 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8rvmk\" (UniqueName: \"kubernetes.io/projected/2e41272f-6e7b-47a7-825e-3571f6a1fd07-kube-api-access-8rvmk\") pod \"nova-metadata-0\" (UID: \"2e41272f-6e7b-47a7-825e-3571f6a1fd07\") " pod="openstack/nova-metadata-0" Mar 13 14:30:01 crc kubenswrapper[4907]: I0313 14:30:01.923329 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e41272f-6e7b-47a7-825e-3571f6a1fd07-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"2e41272f-6e7b-47a7-825e-3571f6a1fd07\") " pod="openstack/nova-metadata-0" Mar 13 14:30:01 crc kubenswrapper[4907]: I0313 14:30:01.923493 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e41272f-6e7b-47a7-825e-3571f6a1fd07-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"2e41272f-6e7b-47a7-825e-3571f6a1fd07\") " pod="openstack/nova-metadata-0" Mar 13 14:30:01 crc kubenswrapper[4907]: I0313 14:30:01.923576 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e41272f-6e7b-47a7-825e-3571f6a1fd07-config-data\") pod \"nova-metadata-0\" (UID: \"2e41272f-6e7b-47a7-825e-3571f6a1fd07\") " pod="openstack/nova-metadata-0" Mar 13 14:30:01 crc kubenswrapper[4907]: I0313 14:30:01.923729 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2e41272f-6e7b-47a7-825e-3571f6a1fd07-logs\") pod \"nova-metadata-0\" (UID: \"2e41272f-6e7b-47a7-825e-3571f6a1fd07\") " pod="openstack/nova-metadata-0" Mar 13 14:30:02 crc kubenswrapper[4907]: I0313 14:30:02.026662 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2e41272f-6e7b-47a7-825e-3571f6a1fd07-logs\") pod \"nova-metadata-0\" (UID: \"2e41272f-6e7b-47a7-825e-3571f6a1fd07\") " pod="openstack/nova-metadata-0" Mar 13 14:30:02 crc kubenswrapper[4907]: I0313 14:30:02.026801 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8rvmk\" (UniqueName: \"kubernetes.io/projected/2e41272f-6e7b-47a7-825e-3571f6a1fd07-kube-api-access-8rvmk\") pod \"nova-metadata-0\" (UID: \"2e41272f-6e7b-47a7-825e-3571f6a1fd07\") " pod="openstack/nova-metadata-0" Mar 13 14:30:02 crc kubenswrapper[4907]: I0313 14:30:02.026824 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e41272f-6e7b-47a7-825e-3571f6a1fd07-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"2e41272f-6e7b-47a7-825e-3571f6a1fd07\") " pod="openstack/nova-metadata-0" Mar 13 14:30:02 crc kubenswrapper[4907]: I0313 14:30:02.026848 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e41272f-6e7b-47a7-825e-3571f6a1fd07-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"2e41272f-6e7b-47a7-825e-3571f6a1fd07\") " pod="openstack/nova-metadata-0" Mar 13 14:30:02 crc kubenswrapper[4907]: I0313 14:30:02.026863 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e41272f-6e7b-47a7-825e-3571f6a1fd07-config-data\") pod \"nova-metadata-0\" (UID: \"2e41272f-6e7b-47a7-825e-3571f6a1fd07\") " pod="openstack/nova-metadata-0" Mar 13 14:30:02 crc kubenswrapper[4907]: I0313 14:30:02.030841 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2e41272f-6e7b-47a7-825e-3571f6a1fd07-logs\") pod \"nova-metadata-0\" (UID: \"2e41272f-6e7b-47a7-825e-3571f6a1fd07\") " pod="openstack/nova-metadata-0" Mar 13 14:30:02 crc kubenswrapper[4907]: I0313 14:30:02.044353 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e41272f-6e7b-47a7-825e-3571f6a1fd07-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"2e41272f-6e7b-47a7-825e-3571f6a1fd07\") " pod="openstack/nova-metadata-0" Mar 13 14:30:02 crc kubenswrapper[4907]: I0313 14:30:02.044414 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e41272f-6e7b-47a7-825e-3571f6a1fd07-config-data\") pod \"nova-metadata-0\" (UID: \"2e41272f-6e7b-47a7-825e-3571f6a1fd07\") " pod="openstack/nova-metadata-0" Mar 13 14:30:02 crc kubenswrapper[4907]: I0313 14:30:02.049113 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8rvmk\" (UniqueName: \"kubernetes.io/projected/2e41272f-6e7b-47a7-825e-3571f6a1fd07-kube-api-access-8rvmk\") pod \"nova-metadata-0\" (UID: \"2e41272f-6e7b-47a7-825e-3571f6a1fd07\") " pod="openstack/nova-metadata-0" Mar 13 14:30:02 crc kubenswrapper[4907]: I0313 14:30:02.049212 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e41272f-6e7b-47a7-825e-3571f6a1fd07-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"2e41272f-6e7b-47a7-825e-3571f6a1fd07\") " pod="openstack/nova-metadata-0" Mar 13 14:30:02 crc kubenswrapper[4907]: I0313 14:30:02.158894 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 13 14:30:02 crc kubenswrapper[4907]: I0313 14:30:02.622217 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Mar 13 14:30:02 crc kubenswrapper[4907]: W0313 14:30:02.630541 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2e41272f_6e7b_47a7_825e_3571f6a1fd07.slice/crio-68545429a61b864d7541170bf482dacf433239060ec793e8fb62a4bc2ee9a2f3 WatchSource:0}: Error finding container 68545429a61b864d7541170bf482dacf433239060ec793e8fb62a4bc2ee9a2f3: Status 404 returned error can't find the container with id 68545429a61b864d7541170bf482dacf433239060ec793e8fb62a4bc2ee9a2f3 Mar 13 14:30:02 crc kubenswrapper[4907]: I0313 14:30:02.751213 4907 generic.go:334] "Generic (PLEG): container finished" podID="e7993f31-1072-4c83-a3ac-16dca5eb04a3" containerID="93b983c69ed0cdc38b2c77a7998230bc3acd8a611b28bc2ce28fc6ae6eacb20d" exitCode=0 Mar 13 14:30:02 crc kubenswrapper[4907]: I0313 14:30:02.751653 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556870-8mlkz" event={"ID":"e7993f31-1072-4c83-a3ac-16dca5eb04a3","Type":"ContainerDied","Data":"93b983c69ed0cdc38b2c77a7998230bc3acd8a611b28bc2ce28fc6ae6eacb20d"} Mar 13 14:30:02 crc kubenswrapper[4907]: I0313 14:30:02.755242 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2e41272f-6e7b-47a7-825e-3571f6a1fd07","Type":"ContainerStarted","Data":"68545429a61b864d7541170bf482dacf433239060ec793e8fb62a4bc2ee9a2f3"} Mar 13 14:30:02 crc kubenswrapper[4907]: I0313 14:30:02.756779 4907 generic.go:334] "Generic (PLEG): container finished" podID="c0304716-7acd-4a2c-a018-73fc4bcbad51" containerID="7dd9875251f25eba8556dcd6141b77fd52b2025f417f75c1f8e455693aab178b" exitCode=0 Mar 13 14:30:02 crc kubenswrapper[4907]: I0313 14:30:02.756977 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"c0304716-7acd-4a2c-a018-73fc4bcbad51","Type":"ContainerDied","Data":"7dd9875251f25eba8556dcd6141b77fd52b2025f417f75c1f8e455693aab178b"} Mar 13 14:30:02 crc kubenswrapper[4907]: I0313 14:30:02.757009 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"c0304716-7acd-4a2c-a018-73fc4bcbad51","Type":"ContainerDied","Data":"a765fc07ea8b46927f8043334da5e0885287349d5eefd55592990b1ae30cfba6"} Mar 13 14:30:02 crc kubenswrapper[4907]: I0313 14:30:02.757024 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a765fc07ea8b46927f8043334da5e0885287349d5eefd55592990b1ae30cfba6" Mar 13 14:30:02 crc kubenswrapper[4907]: I0313 14:30:02.838928 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 13 14:30:02 crc kubenswrapper[4907]: I0313 14:30:02.948018 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0304716-7acd-4a2c-a018-73fc4bcbad51-combined-ca-bundle\") pod \"c0304716-7acd-4a2c-a018-73fc4bcbad51\" (UID: \"c0304716-7acd-4a2c-a018-73fc4bcbad51\") " Mar 13 14:30:02 crc kubenswrapper[4907]: I0313 14:30:02.948096 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xhvp6\" (UniqueName: \"kubernetes.io/projected/c0304716-7acd-4a2c-a018-73fc4bcbad51-kube-api-access-xhvp6\") pod \"c0304716-7acd-4a2c-a018-73fc4bcbad51\" (UID: \"c0304716-7acd-4a2c-a018-73fc4bcbad51\") " Mar 13 14:30:02 crc kubenswrapper[4907]: I0313 14:30:02.948226 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c0304716-7acd-4a2c-a018-73fc4bcbad51-config-data\") pod \"c0304716-7acd-4a2c-a018-73fc4bcbad51\" (UID: \"c0304716-7acd-4a2c-a018-73fc4bcbad51\") " Mar 13 14:30:02 crc kubenswrapper[4907]: I0313 14:30:02.952141 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c0304716-7acd-4a2c-a018-73fc4bcbad51-kube-api-access-xhvp6" (OuterVolumeSpecName: "kube-api-access-xhvp6") pod "c0304716-7acd-4a2c-a018-73fc4bcbad51" (UID: "c0304716-7acd-4a2c-a018-73fc4bcbad51"). InnerVolumeSpecName "kube-api-access-xhvp6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:30:02 crc kubenswrapper[4907]: I0313 14:30:02.976701 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c0304716-7acd-4a2c-a018-73fc4bcbad51-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c0304716-7acd-4a2c-a018-73fc4bcbad51" (UID: "c0304716-7acd-4a2c-a018-73fc4bcbad51"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:03 crc kubenswrapper[4907]: I0313 14:30:03.017142 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c0304716-7acd-4a2c-a018-73fc4bcbad51-config-data" (OuterVolumeSpecName: "config-data") pod "c0304716-7acd-4a2c-a018-73fc4bcbad51" (UID: "c0304716-7acd-4a2c-a018-73fc4bcbad51"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:03 crc kubenswrapper[4907]: I0313 14:30:03.051219 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0304716-7acd-4a2c-a018-73fc4bcbad51-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:03 crc kubenswrapper[4907]: I0313 14:30:03.051521 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xhvp6\" (UniqueName: \"kubernetes.io/projected/c0304716-7acd-4a2c-a018-73fc4bcbad51-kube-api-access-xhvp6\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:03 crc kubenswrapper[4907]: I0313 14:30:03.051537 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c0304716-7acd-4a2c-a018-73fc4bcbad51-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:03 crc kubenswrapper[4907]: I0313 14:30:03.078594 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556870-rdhxg" Mar 13 14:30:03 crc kubenswrapper[4907]: I0313 14:30:03.152638 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/af9d29dd-6749-4fcd-82a8-ea528d832fcc-config-volume\") pod \"af9d29dd-6749-4fcd-82a8-ea528d832fcc\" (UID: \"af9d29dd-6749-4fcd-82a8-ea528d832fcc\") " Mar 13 14:30:03 crc kubenswrapper[4907]: I0313 14:30:03.152867 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6z4z\" (UniqueName: \"kubernetes.io/projected/af9d29dd-6749-4fcd-82a8-ea528d832fcc-kube-api-access-d6z4z\") pod \"af9d29dd-6749-4fcd-82a8-ea528d832fcc\" (UID: \"af9d29dd-6749-4fcd-82a8-ea528d832fcc\") " Mar 13 14:30:03 crc kubenswrapper[4907]: I0313 14:30:03.152992 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/af9d29dd-6749-4fcd-82a8-ea528d832fcc-secret-volume\") pod \"af9d29dd-6749-4fcd-82a8-ea528d832fcc\" (UID: \"af9d29dd-6749-4fcd-82a8-ea528d832fcc\") " Mar 13 14:30:03 crc kubenswrapper[4907]: I0313 14:30:03.154603 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/af9d29dd-6749-4fcd-82a8-ea528d832fcc-config-volume" (OuterVolumeSpecName: "config-volume") pod "af9d29dd-6749-4fcd-82a8-ea528d832fcc" (UID: "af9d29dd-6749-4fcd-82a8-ea528d832fcc"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:30:03 crc kubenswrapper[4907]: I0313 14:30:03.159132 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af9d29dd-6749-4fcd-82a8-ea528d832fcc-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "af9d29dd-6749-4fcd-82a8-ea528d832fcc" (UID: "af9d29dd-6749-4fcd-82a8-ea528d832fcc"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:03 crc kubenswrapper[4907]: I0313 14:30:03.165539 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af9d29dd-6749-4fcd-82a8-ea528d832fcc-kube-api-access-d6z4z" (OuterVolumeSpecName: "kube-api-access-d6z4z") pod "af9d29dd-6749-4fcd-82a8-ea528d832fcc" (UID: "af9d29dd-6749-4fcd-82a8-ea528d832fcc"). InnerVolumeSpecName "kube-api-access-d6z4z". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:30:03 crc kubenswrapper[4907]: I0313 14:30:03.260621 4907 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/af9d29dd-6749-4fcd-82a8-ea528d832fcc-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:03 crc kubenswrapper[4907]: I0313 14:30:03.260683 4907 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/af9d29dd-6749-4fcd-82a8-ea528d832fcc-config-volume\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:03 crc kubenswrapper[4907]: I0313 14:30:03.260695 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6z4z\" (UniqueName: \"kubernetes.io/projected/af9d29dd-6749-4fcd-82a8-ea528d832fcc-kube-api-access-d6z4z\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:03 crc kubenswrapper[4907]: I0313 14:30:03.766771 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556870-rdhxg" event={"ID":"af9d29dd-6749-4fcd-82a8-ea528d832fcc","Type":"ContainerDied","Data":"7c2a8f78d2f024990112bf064e09dd2f29128e2c064da8d891a6688df0ffe224"} Mar 13 14:30:03 crc kubenswrapper[4907]: I0313 14:30:03.766815 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7c2a8f78d2f024990112bf064e09dd2f29128e2c064da8d891a6688df0ffe224" Mar 13 14:30:03 crc kubenswrapper[4907]: I0313 14:30:03.766828 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556870-rdhxg" Mar 13 14:30:03 crc kubenswrapper[4907]: I0313 14:30:03.769098 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2e41272f-6e7b-47a7-825e-3571f6a1fd07","Type":"ContainerStarted","Data":"ef8644527a89477971dc8c42d7fc7d2d08c1853a9324c6451f42196c1c50329e"} Mar 13 14:30:03 crc kubenswrapper[4907]: I0313 14:30:03.769145 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 13 14:30:03 crc kubenswrapper[4907]: I0313 14:30:03.769157 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2e41272f-6e7b-47a7-825e-3571f6a1fd07","Type":"ContainerStarted","Data":"c5ecad10cc9377aeda50af0d8c9213440ce021149e77309245ab84eb57179e9d"} Mar 13 14:30:03 crc kubenswrapper[4907]: I0313 14:30:03.826793 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.826771254 podStartE2EDuration="2.826771254s" podCreationTimestamp="2026-03-13 14:30:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:30:03.797093632 +0000 UTC m=+1502.696881331" watchObservedRunningTime="2026-03-13 14:30:03.826771254 +0000 UTC m=+1502.726558943" Mar 13 14:30:03 crc kubenswrapper[4907]: I0313 14:30:03.842692 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="39488342-aa1a-4dba-bd5f-556da5081904" path="/var/lib/kubelet/pods/39488342-aa1a-4dba-bd5f-556da5081904/volumes" Mar 13 14:30:03 crc kubenswrapper[4907]: I0313 14:30:03.869499 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 14:30:03 crc kubenswrapper[4907]: I0313 14:30:03.898210 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 14:30:03 crc kubenswrapper[4907]: I0313 14:30:03.909555 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 14:30:03 crc kubenswrapper[4907]: E0313 14:30:03.910102 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af9d29dd-6749-4fcd-82a8-ea528d832fcc" containerName="collect-profiles" Mar 13 14:30:03 crc kubenswrapper[4907]: I0313 14:30:03.910126 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="af9d29dd-6749-4fcd-82a8-ea528d832fcc" containerName="collect-profiles" Mar 13 14:30:03 crc kubenswrapper[4907]: E0313 14:30:03.910174 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c0304716-7acd-4a2c-a018-73fc4bcbad51" containerName="nova-scheduler-scheduler" Mar 13 14:30:03 crc kubenswrapper[4907]: I0313 14:30:03.910181 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="c0304716-7acd-4a2c-a018-73fc4bcbad51" containerName="nova-scheduler-scheduler" Mar 13 14:30:03 crc kubenswrapper[4907]: I0313 14:30:03.910367 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="c0304716-7acd-4a2c-a018-73fc4bcbad51" containerName="nova-scheduler-scheduler" Mar 13 14:30:03 crc kubenswrapper[4907]: I0313 14:30:03.910389 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="af9d29dd-6749-4fcd-82a8-ea528d832fcc" containerName="collect-profiles" Mar 13 14:30:03 crc kubenswrapper[4907]: I0313 14:30:03.911079 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 13 14:30:03 crc kubenswrapper[4907]: I0313 14:30:03.913194 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Mar 13 14:30:03 crc kubenswrapper[4907]: I0313 14:30:03.919640 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 14:30:03 crc kubenswrapper[4907]: I0313 14:30:03.973143 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4050d00b-0256-45c5-9dc4-0ab46956405d-config-data\") pod \"nova-scheduler-0\" (UID: \"4050d00b-0256-45c5-9dc4-0ab46956405d\") " pod="openstack/nova-scheduler-0" Mar 13 14:30:03 crc kubenswrapper[4907]: I0313 14:30:03.973355 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xdg29\" (UniqueName: \"kubernetes.io/projected/4050d00b-0256-45c5-9dc4-0ab46956405d-kube-api-access-xdg29\") pod \"nova-scheduler-0\" (UID: \"4050d00b-0256-45c5-9dc4-0ab46956405d\") " pod="openstack/nova-scheduler-0" Mar 13 14:30:03 crc kubenswrapper[4907]: I0313 14:30:03.973562 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4050d00b-0256-45c5-9dc4-0ab46956405d-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"4050d00b-0256-45c5-9dc4-0ab46956405d\") " pod="openstack/nova-scheduler-0" Mar 13 14:30:04 crc kubenswrapper[4907]: I0313 14:30:04.075803 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4050d00b-0256-45c5-9dc4-0ab46956405d-config-data\") pod \"nova-scheduler-0\" (UID: \"4050d00b-0256-45c5-9dc4-0ab46956405d\") " pod="openstack/nova-scheduler-0" Mar 13 14:30:04 crc kubenswrapper[4907]: I0313 14:30:04.075910 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xdg29\" (UniqueName: \"kubernetes.io/projected/4050d00b-0256-45c5-9dc4-0ab46956405d-kube-api-access-xdg29\") pod \"nova-scheduler-0\" (UID: \"4050d00b-0256-45c5-9dc4-0ab46956405d\") " pod="openstack/nova-scheduler-0" Mar 13 14:30:04 crc kubenswrapper[4907]: I0313 14:30:04.076000 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4050d00b-0256-45c5-9dc4-0ab46956405d-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"4050d00b-0256-45c5-9dc4-0ab46956405d\") " pod="openstack/nova-scheduler-0" Mar 13 14:30:04 crc kubenswrapper[4907]: I0313 14:30:04.081046 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4050d00b-0256-45c5-9dc4-0ab46956405d-config-data\") pod \"nova-scheduler-0\" (UID: \"4050d00b-0256-45c5-9dc4-0ab46956405d\") " pod="openstack/nova-scheduler-0" Mar 13 14:30:04 crc kubenswrapper[4907]: I0313 14:30:04.081417 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4050d00b-0256-45c5-9dc4-0ab46956405d-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"4050d00b-0256-45c5-9dc4-0ab46956405d\") " pod="openstack/nova-scheduler-0" Mar 13 14:30:04 crc kubenswrapper[4907]: I0313 14:30:04.096679 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xdg29\" (UniqueName: \"kubernetes.io/projected/4050d00b-0256-45c5-9dc4-0ab46956405d-kube-api-access-xdg29\") pod \"nova-scheduler-0\" (UID: \"4050d00b-0256-45c5-9dc4-0ab46956405d\") " pod="openstack/nova-scheduler-0" Mar 13 14:30:04 crc kubenswrapper[4907]: I0313 14:30:04.218945 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556870-8mlkz" Mar 13 14:30:04 crc kubenswrapper[4907]: I0313 14:30:04.229298 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 13 14:30:04 crc kubenswrapper[4907]: I0313 14:30:04.279340 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2kxlz\" (UniqueName: \"kubernetes.io/projected/e7993f31-1072-4c83-a3ac-16dca5eb04a3-kube-api-access-2kxlz\") pod \"e7993f31-1072-4c83-a3ac-16dca5eb04a3\" (UID: \"e7993f31-1072-4c83-a3ac-16dca5eb04a3\") " Mar 13 14:30:04 crc kubenswrapper[4907]: I0313 14:30:04.283927 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7993f31-1072-4c83-a3ac-16dca5eb04a3-kube-api-access-2kxlz" (OuterVolumeSpecName: "kube-api-access-2kxlz") pod "e7993f31-1072-4c83-a3ac-16dca5eb04a3" (UID: "e7993f31-1072-4c83-a3ac-16dca5eb04a3"). InnerVolumeSpecName "kube-api-access-2kxlz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:30:04 crc kubenswrapper[4907]: I0313 14:30:04.382252 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2kxlz\" (UniqueName: \"kubernetes.io/projected/e7993f31-1072-4c83-a3ac-16dca5eb04a3-kube-api-access-2kxlz\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:04 crc kubenswrapper[4907]: I0313 14:30:04.665828 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 14:30:04 crc kubenswrapper[4907]: W0313 14:30:04.669687 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4050d00b_0256_45c5_9dc4_0ab46956405d.slice/crio-c18abca4f12b62785c093b1834a9526c1fe011c7d23dc6808c045af32e0999dc WatchSource:0}: Error finding container c18abca4f12b62785c093b1834a9526c1fe011c7d23dc6808c045af32e0999dc: Status 404 returned error can't find the container with id c18abca4f12b62785c093b1834a9526c1fe011c7d23dc6808c045af32e0999dc Mar 13 14:30:04 crc kubenswrapper[4907]: I0313 14:30:04.782360 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556870-8mlkz" Mar 13 14:30:04 crc kubenswrapper[4907]: I0313 14:30:04.782359 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556870-8mlkz" event={"ID":"e7993f31-1072-4c83-a3ac-16dca5eb04a3","Type":"ContainerDied","Data":"0fff58df59ba6aeb1bad1085410d3a35e667921192b7b80aca529cb3350038da"} Mar 13 14:30:04 crc kubenswrapper[4907]: I0313 14:30:04.782527 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0fff58df59ba6aeb1bad1085410d3a35e667921192b7b80aca529cb3350038da" Mar 13 14:30:04 crc kubenswrapper[4907]: I0313 14:30:04.785221 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"4050d00b-0256-45c5-9dc4-0ab46956405d","Type":"ContainerStarted","Data":"c18abca4f12b62785c093b1834a9526c1fe011c7d23dc6808c045af32e0999dc"} Mar 13 14:30:05 crc kubenswrapper[4907]: I0313 14:30:05.309096 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556864-768cg"] Mar 13 14:30:05 crc kubenswrapper[4907]: I0313 14:30:05.317931 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556864-768cg"] Mar 13 14:30:05 crc kubenswrapper[4907]: I0313 14:30:05.796046 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="32f92b7e-159a-4824-93e7-ea6d49961db8" path="/var/lib/kubelet/pods/32f92b7e-159a-4824-93e7-ea6d49961db8/volumes" Mar 13 14:30:05 crc kubenswrapper[4907]: I0313 14:30:05.796827 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c0304716-7acd-4a2c-a018-73fc4bcbad51" path="/var/lib/kubelet/pods/c0304716-7acd-4a2c-a018-73fc4bcbad51/volumes" Mar 13 14:30:05 crc kubenswrapper[4907]: I0313 14:30:05.797348 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"4050d00b-0256-45c5-9dc4-0ab46956405d","Type":"ContainerStarted","Data":"8a59419fea27690bc54a6fc859fef55fbe86c693159ac3ff37906574c1369e56"} Mar 13 14:30:05 crc kubenswrapper[4907]: I0313 14:30:05.821533 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.821514209 podStartE2EDuration="2.821514209s" podCreationTimestamp="2026-03-13 14:30:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:30:05.813397225 +0000 UTC m=+1504.713184914" watchObservedRunningTime="2026-03-13 14:30:05.821514209 +0000 UTC m=+1504.721301898" Mar 13 14:30:08 crc kubenswrapper[4907]: I0313 14:30:08.064113 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Mar 13 14:30:08 crc kubenswrapper[4907]: I0313 14:30:08.064402 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Mar 13 14:30:09 crc kubenswrapper[4907]: I0313 14:30:09.078066 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="cc227d33-3bf1-4b5f-8368-8e6b8a512f7d" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.211:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 13 14:30:09 crc kubenswrapper[4907]: I0313 14:30:09.078154 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="cc227d33-3bf1-4b5f-8368-8e6b8a512f7d" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.211:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 13 14:30:09 crc kubenswrapper[4907]: I0313 14:30:09.230074 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Mar 13 14:30:12 crc kubenswrapper[4907]: I0313 14:30:12.159551 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Mar 13 14:30:12 crc kubenswrapper[4907]: I0313 14:30:12.160632 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Mar 13 14:30:13 crc kubenswrapper[4907]: I0313 14:30:13.174061 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="2e41272f-6e7b-47a7-825e-3571f6a1fd07" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.214:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 13 14:30:13 crc kubenswrapper[4907]: I0313 14:30:13.174100 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="2e41272f-6e7b-47a7-825e-3571f6a1fd07" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.214:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 13 14:30:14 crc kubenswrapper[4907]: I0313 14:30:14.229519 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Mar 13 14:30:14 crc kubenswrapper[4907]: I0313 14:30:14.255946 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Mar 13 14:30:14 crc kubenswrapper[4907]: I0313 14:30:14.916492 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Mar 13 14:30:15 crc kubenswrapper[4907]: I0313 14:30:15.809706 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Mar 13 14:30:16 crc kubenswrapper[4907]: I0313 14:30:16.063585 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Mar 13 14:30:16 crc kubenswrapper[4907]: I0313 14:30:16.063640 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Mar 13 14:30:18 crc kubenswrapper[4907]: I0313 14:30:18.073950 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Mar 13 14:30:18 crc kubenswrapper[4907]: I0313 14:30:18.075509 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Mar 13 14:30:18 crc kubenswrapper[4907]: I0313 14:30:18.098149 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Mar 13 14:30:18 crc kubenswrapper[4907]: I0313 14:30:18.927784 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Mar 13 14:30:20 crc kubenswrapper[4907]: I0313 14:30:20.159037 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Mar 13 14:30:20 crc kubenswrapper[4907]: I0313 14:30:20.159533 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Mar 13 14:30:21 crc kubenswrapper[4907]: I0313 14:30:21.726971 4907 scope.go:117] "RemoveContainer" containerID="a53552e35726d3cc22de750fc04d8a15ada7be07f868ed527291e30242034c7d" Mar 13 14:30:22 crc kubenswrapper[4907]: I0313 14:30:22.165168 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Mar 13 14:30:22 crc kubenswrapper[4907]: I0313 14:30:22.165277 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Mar 13 14:30:22 crc kubenswrapper[4907]: I0313 14:30:22.170567 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Mar 13 14:30:22 crc kubenswrapper[4907]: I0313 14:30:22.173425 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Mar 13 14:30:39 crc kubenswrapper[4907]: I0313 14:30:39.630126 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/root-account-create-update-r6vk2"] Mar 13 14:30:39 crc kubenswrapper[4907]: I0313 14:30:39.647704 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/root-account-create-update-r6vk2"] Mar 13 14:30:39 crc kubenswrapper[4907]: I0313 14:30:39.679011 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/root-account-create-update-glzbl"] Mar 13 14:30:39 crc kubenswrapper[4907]: E0313 14:30:39.679496 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7993f31-1072-4c83-a3ac-16dca5eb04a3" containerName="oc" Mar 13 14:30:39 crc kubenswrapper[4907]: I0313 14:30:39.679515 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7993f31-1072-4c83-a3ac-16dca5eb04a3" containerName="oc" Mar 13 14:30:39 crc kubenswrapper[4907]: I0313 14:30:39.679759 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="e7993f31-1072-4c83-a3ac-16dca5eb04a3" containerName="oc" Mar 13 14:30:39 crc kubenswrapper[4907]: I0313 14:30:39.680512 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-glzbl" Mar 13 14:30:39 crc kubenswrapper[4907]: I0313 14:30:39.682336 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-mariadb-root-db-secret" Mar 13 14:30:39 crc kubenswrapper[4907]: I0313 14:30:39.709267 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-glzbl"] Mar 13 14:30:39 crc kubenswrapper[4907]: I0313 14:30:39.751296 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Mar 13 14:30:39 crc kubenswrapper[4907]: I0313 14:30:39.751696 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstackclient" podUID="a055ce7f-0538-4e5e-938a-eaf6d0aa2dec" containerName="openstackclient" containerID="cri-o://3e056a670aa7307d9d211e2123e0cead9b6cd31175ff09e1feddb508c47f1a88" gracePeriod=2 Mar 13 14:30:39 crc kubenswrapper[4907]: I0313 14:30:39.771079 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Mar 13 14:30:39 crc kubenswrapper[4907]: I0313 14:30:39.778652 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/516d1499-3461-4af5-b426-9ae4a711a468-operator-scripts\") pod \"root-account-create-update-glzbl\" (UID: \"516d1499-3461-4af5-b426-9ae4a711a468\") " pod="openstack/root-account-create-update-glzbl" Mar 13 14:30:39 crc kubenswrapper[4907]: I0313 14:30:39.778790 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-df597\" (UniqueName: \"kubernetes.io/projected/516d1499-3461-4af5-b426-9ae4a711a468-kube-api-access-df597\") pod \"root-account-create-update-glzbl\" (UID: \"516d1499-3461-4af5-b426-9ae4a711a468\") " pod="openstack/root-account-create-update-glzbl" Mar 13 14:30:39 crc kubenswrapper[4907]: I0313 14:30:39.816340 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d1b4fd44-4a11-4521-bfd0-16c5436368fa" path="/var/lib/kubelet/pods/d1b4fd44-4a11-4521-bfd0-16c5436368fa/volumes" Mar 13 14:30:39 crc kubenswrapper[4907]: I0313 14:30:39.883173 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-df597\" (UniqueName: \"kubernetes.io/projected/516d1499-3461-4af5-b426-9ae4a711a468-kube-api-access-df597\") pod \"root-account-create-update-glzbl\" (UID: \"516d1499-3461-4af5-b426-9ae4a711a468\") " pod="openstack/root-account-create-update-glzbl" Mar 13 14:30:39 crc kubenswrapper[4907]: I0313 14:30:39.883346 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/516d1499-3461-4af5-b426-9ae4a711a468-operator-scripts\") pod \"root-account-create-update-glzbl\" (UID: \"516d1499-3461-4af5-b426-9ae4a711a468\") " pod="openstack/root-account-create-update-glzbl" Mar 13 14:30:39 crc kubenswrapper[4907]: I0313 14:30:39.884371 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/516d1499-3461-4af5-b426-9ae4a711a468-operator-scripts\") pod \"root-account-create-update-glzbl\" (UID: \"516d1499-3461-4af5-b426-9ae4a711a468\") " pod="openstack/root-account-create-update-glzbl" Mar 13 14:30:39 crc kubenswrapper[4907]: I0313 14:30:39.961040 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-9c64-account-create-update-v65tv"] Mar 13 14:30:39 crc kubenswrapper[4907]: I0313 14:30:39.965724 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-df597\" (UniqueName: \"kubernetes.io/projected/516d1499-3461-4af5-b426-9ae4a711a468-kube-api-access-df597\") pod \"root-account-create-update-glzbl\" (UID: \"516d1499-3461-4af5-b426-9ae4a711a468\") " pod="openstack/root-account-create-update-glzbl" Mar 13 14:30:40 crc kubenswrapper[4907]: I0313 14:30:40.009017 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-glzbl" Mar 13 14:30:40 crc kubenswrapper[4907]: I0313 14:30:40.012502 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-9c64-account-create-update-v65tv"] Mar 13 14:30:40 crc kubenswrapper[4907]: I0313 14:30:40.040379 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-sb-0"] Mar 13 14:30:40 crc kubenswrapper[4907]: I0313 14:30:40.041851 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-sb-0" podUID="c2938e39-f1d8-4cdc-a32c-5d57b8f2034f" containerName="openstack-network-exporter" containerID="cri-o://1ff45617937d399a62e2b1c7ef156ccf7052d51f57fba51d267daa3e49810ce6" gracePeriod=300 Mar 13 14:30:40 crc kubenswrapper[4907]: E0313 14:30:40.088204 4907 projected.go:263] Couldn't get secret openstack/swift-conf: secret "swift-conf" not found Mar 13 14:30:40 crc kubenswrapper[4907]: E0313 14:30:40.088269 4907 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: secret "swift-conf" not found Mar 13 14:30:40 crc kubenswrapper[4907]: E0313 14:30:40.088348 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/14f05f31-8185-43b0-be69-bcf8d5388ea5-etc-swift podName:14f05f31-8185-43b0-be69-bcf8d5388ea5 nodeName:}" failed. No retries permitted until 2026-03-13 14:30:40.588317102 +0000 UTC m=+1539.488104791 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/14f05f31-8185-43b0-be69-bcf8d5388ea5-etc-swift") pod "swift-storage-0" (UID: "14f05f31-8185-43b0-be69-bcf8d5388ea5") : secret "swift-conf" not found Mar 13 14:30:40 crc kubenswrapper[4907]: I0313 14:30:40.291960 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-9c64-account-create-update-5vvgg"] Mar 13 14:30:40 crc kubenswrapper[4907]: E0313 14:30:40.310937 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a055ce7f-0538-4e5e-938a-eaf6d0aa2dec" containerName="openstackclient" Mar 13 14:30:40 crc kubenswrapper[4907]: I0313 14:30:40.310979 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="a055ce7f-0538-4e5e-938a-eaf6d0aa2dec" containerName="openstackclient" Mar 13 14:30:40 crc kubenswrapper[4907]: I0313 14:30:40.311752 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="a055ce7f-0538-4e5e-938a-eaf6d0aa2dec" containerName="openstackclient" Mar 13 14:30:40 crc kubenswrapper[4907]: I0313 14:30:40.312611 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-9c64-account-create-update-5vvgg" Mar 13 14:30:40 crc kubenswrapper[4907]: I0313 14:30:40.366629 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4a4c64a3-7378-4d8d-9cf3-17583c6a90c0-operator-scripts\") pod \"placement-9c64-account-create-update-5vvgg\" (UID: \"4a4c64a3-7378-4d8d-9cf3-17583c6a90c0\") " pod="openstack/placement-9c64-account-create-update-5vvgg" Mar 13 14:30:40 crc kubenswrapper[4907]: I0313 14:30:40.366783 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wqnkx\" (UniqueName: \"kubernetes.io/projected/4a4c64a3-7378-4d8d-9cf3-17583c6a90c0-kube-api-access-wqnkx\") pod \"placement-9c64-account-create-update-5vvgg\" (UID: \"4a4c64a3-7378-4d8d-9cf3-17583c6a90c0\") " pod="openstack/placement-9c64-account-create-update-5vvgg" Mar 13 14:30:40 crc kubenswrapper[4907]: I0313 14:30:40.408447 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Mar 13 14:30:40 crc kubenswrapper[4907]: I0313 14:30:40.413035 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-47e2-account-create-update-wvjr7"] Mar 13 14:30:40 crc kubenswrapper[4907]: I0313 14:30:40.424146 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-47e2-account-create-update-wvjr7" Mar 13 14:30:40 crc kubenswrapper[4907]: I0313 14:30:40.445407 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Mar 13 14:30:40 crc kubenswrapper[4907]: I0313 14:30:40.474560 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wqnkx\" (UniqueName: \"kubernetes.io/projected/4a4c64a3-7378-4d8d-9cf3-17583c6a90c0-kube-api-access-wqnkx\") pod \"placement-9c64-account-create-update-5vvgg\" (UID: \"4a4c64a3-7378-4d8d-9cf3-17583c6a90c0\") " pod="openstack/placement-9c64-account-create-update-5vvgg" Mar 13 14:30:40 crc kubenswrapper[4907]: I0313 14:30:40.474634 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4a4c64a3-7378-4d8d-9cf3-17583c6a90c0-operator-scripts\") pod \"placement-9c64-account-create-update-5vvgg\" (UID: \"4a4c64a3-7378-4d8d-9cf3-17583c6a90c0\") " pod="openstack/placement-9c64-account-create-update-5vvgg" Mar 13 14:30:40 crc kubenswrapper[4907]: I0313 14:30:40.475393 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4a4c64a3-7378-4d8d-9cf3-17583c6a90c0-operator-scripts\") pod \"placement-9c64-account-create-update-5vvgg\" (UID: \"4a4c64a3-7378-4d8d-9cf3-17583c6a90c0\") " pod="openstack/placement-9c64-account-create-update-5vvgg" Mar 13 14:30:40 crc kubenswrapper[4907]: I0313 14:30:40.497814 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-9c64-account-create-update-5vvgg"] Mar 13 14:30:40 crc kubenswrapper[4907]: I0313 14:30:40.529226 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wqnkx\" (UniqueName: \"kubernetes.io/projected/4a4c64a3-7378-4d8d-9cf3-17583c6a90c0-kube-api-access-wqnkx\") pod \"placement-9c64-account-create-update-5vvgg\" (UID: \"4a4c64a3-7378-4d8d-9cf3-17583c6a90c0\") " pod="openstack/placement-9c64-account-create-update-5vvgg" Mar 13 14:30:40 crc kubenswrapper[4907]: I0313 14:30:40.550451 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-b69e-account-create-update-96wq7"] Mar 13 14:30:40 crc kubenswrapper[4907]: I0313 14:30:40.552086 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-b69e-account-create-update-96wq7" Mar 13 14:30:40 crc kubenswrapper[4907]: I0313 14:30:40.555374 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Mar 13 14:30:40 crc kubenswrapper[4907]: I0313 14:30:40.576780 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vlhjv\" (UniqueName: \"kubernetes.io/projected/82dec019-ba60-43f5-845b-322a6bcca225-kube-api-access-vlhjv\") pod \"cinder-47e2-account-create-update-wvjr7\" (UID: \"82dec019-ba60-43f5-845b-322a6bcca225\") " pod="openstack/cinder-47e2-account-create-update-wvjr7" Mar 13 14:30:40 crc kubenswrapper[4907]: I0313 14:30:40.576853 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/82dec019-ba60-43f5-845b-322a6bcca225-operator-scripts\") pod \"cinder-47e2-account-create-update-wvjr7\" (UID: \"82dec019-ba60-43f5-845b-322a6bcca225\") " pod="openstack/cinder-47e2-account-create-update-wvjr7" Mar 13 14:30:40 crc kubenswrapper[4907]: I0313 14:30:40.577379 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-b69e-account-create-update-96wq7"] Mar 13 14:30:40 crc kubenswrapper[4907]: I0313 14:30:40.585927 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-47e2-account-create-update-r4ccm"] Mar 13 14:30:40 crc kubenswrapper[4907]: I0313 14:30:40.596965 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-47e2-account-create-update-wvjr7"] Mar 13 14:30:40 crc kubenswrapper[4907]: I0313 14:30:40.631936 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-47e2-account-create-update-r4ccm"] Mar 13 14:30:40 crc kubenswrapper[4907]: I0313 14:30:40.664700 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 13 14:30:40 crc kubenswrapper[4907]: I0313 14:30:40.685588 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vlhjv\" (UniqueName: \"kubernetes.io/projected/82dec019-ba60-43f5-845b-322a6bcca225-kube-api-access-vlhjv\") pod \"cinder-47e2-account-create-update-wvjr7\" (UID: \"82dec019-ba60-43f5-845b-322a6bcca225\") " pod="openstack/cinder-47e2-account-create-update-wvjr7" Mar 13 14:30:40 crc kubenswrapper[4907]: I0313 14:30:40.685648 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/82dec019-ba60-43f5-845b-322a6bcca225-operator-scripts\") pod \"cinder-47e2-account-create-update-wvjr7\" (UID: \"82dec019-ba60-43f5-845b-322a6bcca225\") " pod="openstack/cinder-47e2-account-create-update-wvjr7" Mar 13 14:30:40 crc kubenswrapper[4907]: I0313 14:30:40.685689 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e8f6894c-b714-431e-912d-8b8de1a3914b-operator-scripts\") pod \"neutron-b69e-account-create-update-96wq7\" (UID: \"e8f6894c-b714-431e-912d-8b8de1a3914b\") " pod="openstack/neutron-b69e-account-create-update-96wq7" Mar 13 14:30:40 crc kubenswrapper[4907]: I0313 14:30:40.685744 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kfp65\" (UniqueName: \"kubernetes.io/projected/e8f6894c-b714-431e-912d-8b8de1a3914b-kube-api-access-kfp65\") pod \"neutron-b69e-account-create-update-96wq7\" (UID: \"e8f6894c-b714-431e-912d-8b8de1a3914b\") " pod="openstack/neutron-b69e-account-create-update-96wq7" Mar 13 14:30:40 crc kubenswrapper[4907]: I0313 14:30:40.704831 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/82dec019-ba60-43f5-845b-322a6bcca225-operator-scripts\") pod \"cinder-47e2-account-create-update-wvjr7\" (UID: \"82dec019-ba60-43f5-845b-322a6bcca225\") " pod="openstack/cinder-47e2-account-create-update-wvjr7" Mar 13 14:30:40 crc kubenswrapper[4907]: E0313 14:30:40.704926 4907 projected.go:263] Couldn't get secret openstack/swift-conf: secret "swift-conf" not found Mar 13 14:30:40 crc kubenswrapper[4907]: E0313 14:30:40.704941 4907 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Mar 13 14:30:40 crc kubenswrapper[4907]: E0313 14:30:40.704952 4907 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: [secret "swift-conf" not found, configmap "swift-ring-files" not found] Mar 13 14:30:40 crc kubenswrapper[4907]: E0313 14:30:40.704993 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/14f05f31-8185-43b0-be69-bcf8d5388ea5-etc-swift podName:14f05f31-8185-43b0-be69-bcf8d5388ea5 nodeName:}" failed. No retries permitted until 2026-03-13 14:30:41.704981383 +0000 UTC m=+1540.604769072 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/14f05f31-8185-43b0-be69-bcf8d5388ea5-etc-swift") pod "swift-storage-0" (UID: "14f05f31-8185-43b0-be69-bcf8d5388ea5") : [secret "swift-conf" not found, configmap "swift-ring-files" not found] Mar 13 14:30:40 crc kubenswrapper[4907]: I0313 14:30:40.709661 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-9c64-account-create-update-5vvgg" Mar 13 14:30:40 crc kubenswrapper[4907]: I0313 14:30:40.719716 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-sb-0" podUID="c2938e39-f1d8-4cdc-a32c-5d57b8f2034f" containerName="ovsdbserver-sb" containerID="cri-o://432aefb6d80f0522ac7d6d516626fd7ea34147a98d43c3ddba905f2a1bb074ad" gracePeriod=300 Mar 13 14:30:40 crc kubenswrapper[4907]: I0313 14:30:40.746312 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-b69e-account-create-update-ptpcx"] Mar 13 14:30:40 crc kubenswrapper[4907]: I0313 14:30:40.789395 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kfp65\" (UniqueName: \"kubernetes.io/projected/e8f6894c-b714-431e-912d-8b8de1a3914b-kube-api-access-kfp65\") pod \"neutron-b69e-account-create-update-96wq7\" (UID: \"e8f6894c-b714-431e-912d-8b8de1a3914b\") " pod="openstack/neutron-b69e-account-create-update-96wq7" Mar 13 14:30:40 crc kubenswrapper[4907]: I0313 14:30:40.789605 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e8f6894c-b714-431e-912d-8b8de1a3914b-operator-scripts\") pod \"neutron-b69e-account-create-update-96wq7\" (UID: \"e8f6894c-b714-431e-912d-8b8de1a3914b\") " pod="openstack/neutron-b69e-account-create-update-96wq7" Mar 13 14:30:40 crc kubenswrapper[4907]: E0313 14:30:40.791617 4907 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Mar 13 14:30:40 crc kubenswrapper[4907]: E0313 14:30:40.791664 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/286c82e8-b74a-49d7-a355-ac074aace10e-config-data podName:286c82e8-b74a-49d7-a355-ac074aace10e nodeName:}" failed. No retries permitted until 2026-03-13 14:30:41.291648284 +0000 UTC m=+1540.191435973 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/286c82e8-b74a-49d7-a355-ac074aace10e-config-data") pod "rabbitmq-cell1-server-0" (UID: "286c82e8-b74a-49d7-a355-ac074aace10e") : configmap "rabbitmq-cell1-config-data" not found Mar 13 14:30:40 crc kubenswrapper[4907]: I0313 14:30:40.792203 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e8f6894c-b714-431e-912d-8b8de1a3914b-operator-scripts\") pod \"neutron-b69e-account-create-update-96wq7\" (UID: \"e8f6894c-b714-431e-912d-8b8de1a3914b\") " pod="openstack/neutron-b69e-account-create-update-96wq7" Mar 13 14:30:40 crc kubenswrapper[4907]: I0313 14:30:40.807602 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vlhjv\" (UniqueName: \"kubernetes.io/projected/82dec019-ba60-43f5-845b-322a6bcca225-kube-api-access-vlhjv\") pod \"cinder-47e2-account-create-update-wvjr7\" (UID: \"82dec019-ba60-43f5-845b-322a6bcca225\") " pod="openstack/cinder-47e2-account-create-update-wvjr7" Mar 13 14:30:40 crc kubenswrapper[4907]: I0313 14:30:40.808097 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-17a8-account-create-update-nj9pw"] Mar 13 14:30:40 crc kubenswrapper[4907]: I0313 14:30:40.810438 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-17a8-account-create-update-nj9pw" Mar 13 14:30:40 crc kubenswrapper[4907]: I0313 14:30:40.824209 4907 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-galera-0" podUID="c4411bd2-c555-433a-9015-f623948b1401" containerName="galera" probeResult="failure" output="command timed out" Mar 13 14:30:40 crc kubenswrapper[4907]: I0313 14:30:40.824986 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Mar 13 14:30:40 crc kubenswrapper[4907]: I0313 14:30:40.828917 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/openstack-galera-0" podUID="c4411bd2-c555-433a-9015-f623948b1401" containerName="galera" probeResult="failure" output="command timed out" Mar 13 14:30:40 crc kubenswrapper[4907]: I0313 14:30:40.837574 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kfp65\" (UniqueName: \"kubernetes.io/projected/e8f6894c-b714-431e-912d-8b8de1a3914b-kube-api-access-kfp65\") pod \"neutron-b69e-account-create-update-96wq7\" (UID: \"e8f6894c-b714-431e-912d-8b8de1a3914b\") " pod="openstack/neutron-b69e-account-create-update-96wq7" Mar 13 14:30:40 crc kubenswrapper[4907]: I0313 14:30:40.849304 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-b69e-account-create-update-ptpcx"] Mar 13 14:30:40 crc kubenswrapper[4907]: I0313 14:30:40.871951 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-17a8-account-create-update-nj9pw"] Mar 13 14:30:40 crc kubenswrapper[4907]: I0313 14:30:40.892000 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bc15f080-f90d-4e97-a866-d5edfd662bb7-operator-scripts\") pod \"glance-17a8-account-create-update-nj9pw\" (UID: \"bc15f080-f90d-4e97-a866-d5edfd662bb7\") " pod="openstack/glance-17a8-account-create-update-nj9pw" Mar 13 14:30:40 crc kubenswrapper[4907]: I0313 14:30:40.892169 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jstl5\" (UniqueName: \"kubernetes.io/projected/bc15f080-f90d-4e97-a866-d5edfd662bb7-kube-api-access-jstl5\") pod \"glance-17a8-account-create-update-nj9pw\" (UID: \"bc15f080-f90d-4e97-a866-d5edfd662bb7\") " pod="openstack/glance-17a8-account-create-update-nj9pw" Mar 13 14:30:40 crc kubenswrapper[4907]: I0313 14:30:40.914711 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-b69e-account-create-update-96wq7" Mar 13 14:30:40 crc kubenswrapper[4907]: I0313 14:30:40.959118 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-nb-0"] Mar 13 14:30:40 crc kubenswrapper[4907]: I0313 14:30:40.959754 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-nb-0" podUID="056dd756-0d7b-471b-9929-f622d05ad606" containerName="openstack-network-exporter" containerID="cri-o://f294f1dfa23d177ca2e8f5511b6bd495bd3146c4baee6d3b95ccd81e1b4496fb" gracePeriod=300 Mar 13 14:30:40 crc kubenswrapper[4907]: I0313 14:30:40.988746 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-17a8-account-create-update-xv8sm"] Mar 13 14:30:40 crc kubenswrapper[4907]: I0313 14:30:40.995724 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bc15f080-f90d-4e97-a866-d5edfd662bb7-operator-scripts\") pod \"glance-17a8-account-create-update-nj9pw\" (UID: \"bc15f080-f90d-4e97-a866-d5edfd662bb7\") " pod="openstack/glance-17a8-account-create-update-nj9pw" Mar 13 14:30:40 crc kubenswrapper[4907]: I0313 14:30:40.996011 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jstl5\" (UniqueName: \"kubernetes.io/projected/bc15f080-f90d-4e97-a866-d5edfd662bb7-kube-api-access-jstl5\") pod \"glance-17a8-account-create-update-nj9pw\" (UID: \"bc15f080-f90d-4e97-a866-d5edfd662bb7\") " pod="openstack/glance-17a8-account-create-update-nj9pw" Mar 13 14:30:40 crc kubenswrapper[4907]: I0313 14:30:40.996985 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bc15f080-f90d-4e97-a866-d5edfd662bb7-operator-scripts\") pod \"glance-17a8-account-create-update-nj9pw\" (UID: \"bc15f080-f90d-4e97-a866-d5edfd662bb7\") " pod="openstack/glance-17a8-account-create-update-nj9pw" Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.030935 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-17a8-account-create-update-xv8sm"] Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.047283 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jstl5\" (UniqueName: \"kubernetes.io/projected/bc15f080-f90d-4e97-a866-d5edfd662bb7-kube-api-access-jstl5\") pod \"glance-17a8-account-create-update-nj9pw\" (UID: \"bc15f080-f90d-4e97-a866-d5edfd662bb7\") " pod="openstack/glance-17a8-account-create-update-nj9pw" Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.074070 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-northd-0"] Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.074293 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-northd-0" podUID="26827945-75f0-4867-ba04-31ff6428e06a" containerName="ovn-northd" containerID="cri-o://e23ef8c27779c528c50bff3fd731469de16d17d8b86a266ab0618151d6dd930b" gracePeriod=30 Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.074663 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-northd-0" podUID="26827945-75f0-4867-ba04-31ff6428e06a" containerName="openstack-network-exporter" containerID="cri-o://59dc65eade55429a5d17c956c987a8209b624ac898dd355bca1d93bdf510b508" gracePeriod=30 Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.102908 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-47e2-account-create-update-wvjr7" Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.124587 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-7q999"] Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.175401 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-17a8-account-create-update-nj9pw" Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.190782 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-7q999"] Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.223312 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-jt88g"] Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.253349 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-nb-0" podUID="056dd756-0d7b-471b-9929-f622d05ad606" containerName="ovsdbserver-nb" containerID="cri-o://46f9fcec5ce705120f46ecb4deec4665885423d12b23cabce7f929fa5941d7ab" gracePeriod=300 Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.270452 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-jt88g"] Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.308115 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-jjqr9"] Mar 13 14:30:41 crc kubenswrapper[4907]: E0313 14:30:41.315646 4907 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Mar 13 14:30:41 crc kubenswrapper[4907]: E0313 14:30:41.315770 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/286c82e8-b74a-49d7-a355-ac074aace10e-config-data podName:286c82e8-b74a-49d7-a355-ac074aace10e nodeName:}" failed. No retries permitted until 2026-03-13 14:30:42.3157553 +0000 UTC m=+1541.215542989 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/286c82e8-b74a-49d7-a355-ac074aace10e-config-data") pod "rabbitmq-cell1-server-0" (UID: "286c82e8-b74a-49d7-a355-ac074aace10e") : configmap "rabbitmq-cell1-config-data" not found Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.342826 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-jjqr9"] Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.377264 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-22bc-account-create-update-r969t"] Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.396939 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-22bc-account-create-update-r969t"] Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.398634 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_c2938e39-f1d8-4cdc-a32c-5d57b8f2034f/ovsdbserver-sb/0.log" Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.398674 4907 generic.go:334] "Generic (PLEG): container finished" podID="c2938e39-f1d8-4cdc-a32c-5d57b8f2034f" containerID="1ff45617937d399a62e2b1c7ef156ccf7052d51f57fba51d267daa3e49810ce6" exitCode=2 Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.398689 4907 generic.go:334] "Generic (PLEG): container finished" podID="c2938e39-f1d8-4cdc-a32c-5d57b8f2034f" containerID="432aefb6d80f0522ac7d6d516626fd7ea34147a98d43c3ddba905f2a1bb074ad" exitCode=143 Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.398935 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"c2938e39-f1d8-4cdc-a32c-5d57b8f2034f","Type":"ContainerDied","Data":"1ff45617937d399a62e2b1c7ef156ccf7052d51f57fba51d267daa3e49810ce6"} Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.399115 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"c2938e39-f1d8-4cdc-a32c-5d57b8f2034f","Type":"ContainerDied","Data":"432aefb6d80f0522ac7d6d516626fd7ea34147a98d43c3ddba905f2a1bb074ad"} Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.419134 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-rz6qh"] Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.420414 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_26827945-75f0-4867-ba04-31ff6428e06a/ovn-northd/0.log" Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.420445 4907 generic.go:334] "Generic (PLEG): container finished" podID="26827945-75f0-4867-ba04-31ff6428e06a" containerID="59dc65eade55429a5d17c956c987a8209b624ac898dd355bca1d93bdf510b508" exitCode=2 Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.420460 4907 generic.go:334] "Generic (PLEG): container finished" podID="26827945-75f0-4867-ba04-31ff6428e06a" containerID="e23ef8c27779c528c50bff3fd731469de16d17d8b86a266ab0618151d6dd930b" exitCode=143 Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.420515 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"26827945-75f0-4867-ba04-31ff6428e06a","Type":"ContainerDied","Data":"59dc65eade55429a5d17c956c987a8209b624ac898dd355bca1d93bdf510b508"} Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.420535 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"26827945-75f0-4867-ba04-31ff6428e06a","Type":"ContainerDied","Data":"e23ef8c27779c528c50bff3fd731469de16d17d8b86a266ab0618151d6dd930b"} Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.426597 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_056dd756-0d7b-471b-9929-f622d05ad606/ovsdbserver-nb/0.log" Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.426637 4907 generic.go:334] "Generic (PLEG): container finished" podID="056dd756-0d7b-471b-9929-f622d05ad606" containerID="f294f1dfa23d177ca2e8f5511b6bd495bd3146c4baee6d3b95ccd81e1b4496fb" exitCode=2 Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.426666 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"056dd756-0d7b-471b-9929-f622d05ad606","Type":"ContainerDied","Data":"f294f1dfa23d177ca2e8f5511b6bd495bd3146c4baee6d3b95ccd81e1b4496fb"} Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.453484 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-f148-account-create-update-wgw4b"] Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.464988 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-4v4ps"] Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.493788 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-rz6qh"] Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.548687 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-fdb8f6449-nb4km"] Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.549207 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-fdb8f6449-nb4km" podUID="7da13722-732a-4b65-a894-fe4612f30d75" containerName="dnsmasq-dns" containerID="cri-o://890693bd8cb16d526cbd4e7c94b7a9b8dcd67a1bc48dbb881c81f86b4e9727d1" gracePeriod=10 Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.621022 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-9fxfh"] Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.656032 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-metrics-nhnkq"] Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.656283 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-metrics-nhnkq" podUID="37978cc1-fcba-4032-a8b1-6632b61692ff" containerName="openstack-network-exporter" containerID="cri-o://62edff4734da4ab22bc7673f1d2c0e9a9db082a3ca22b12fadc41486071fe66a" gracePeriod=30 Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.686872 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-bw2xb"] Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.722163 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-ring-rebalance-9fxfh"] Mar 13 14:30:41 crc kubenswrapper[4907]: E0313 14:30:41.752786 4907 projected.go:288] Couldn't get configMap openstack/swift-storage-config-data: configmap "swift-storage-config-data" not found Mar 13 14:30:41 crc kubenswrapper[4907]: E0313 14:30:41.752819 4907 projected.go:263] Couldn't get secret openstack/swift-conf: secret "swift-conf" not found Mar 13 14:30:41 crc kubenswrapper[4907]: E0313 14:30:41.752828 4907 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Mar 13 14:30:41 crc kubenswrapper[4907]: E0313 14:30:41.752838 4907 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: [configmap "swift-storage-config-data" not found, secret "swift-conf" not found, configmap "swift-ring-files" not found] Mar 13 14:30:41 crc kubenswrapper[4907]: E0313 14:30:41.752900 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/14f05f31-8185-43b0-be69-bcf8d5388ea5-etc-swift podName:14f05f31-8185-43b0-be69-bcf8d5388ea5 nodeName:}" failed. No retries permitted until 2026-03-13 14:30:43.752872436 +0000 UTC m=+1542.652660125 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/14f05f31-8185-43b0-be69-bcf8d5388ea5-etc-swift") pod "swift-storage-0" (UID: "14f05f31-8185-43b0-be69-bcf8d5388ea5") : [configmap "swift-storage-config-data" not found, secret "swift-conf" not found, configmap "swift-ring-files" not found] Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.778498 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-f148-account-create-update-wgw4b"] Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.816472 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0983a1bd-a83c-4658-a405-09b8ab0d0002" path="/var/lib/kubelet/pods/0983a1bd-a83c-4658-a405-09b8ab0d0002/volumes" Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.817340 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="32b531c6-64c1-4137-b82e-3c10789325e6" path="/var/lib/kubelet/pods/32b531c6-64c1-4137-b82e-3c10789325e6/volumes" Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.817852 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3db3f0b6-9cae-40fa-b54d-3ed06c568c9c" path="/var/lib/kubelet/pods/3db3f0b6-9cae-40fa-b54d-3ed06c568c9c/volumes" Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.819203 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3fe8116b-9e68-41e2-aad2-cdd7d5e51c83" path="/var/lib/kubelet/pods/3fe8116b-9e68-41e2-aad2-cdd7d5e51c83/volumes" Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.820256 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4eeb2fb2-22fb-4ea8-a39b-5f959877cfe6" path="/var/lib/kubelet/pods/4eeb2fb2-22fb-4ea8-a39b-5f959877cfe6/volumes" Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.820797 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="594b61fa-12b7-47b6-8af3-17f024e0d54d" path="/var/lib/kubelet/pods/594b61fa-12b7-47b6-8af3-17f024e0d54d/volumes" Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.821606 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="85913727-bbcd-4343-9faf-a75f40b42dc8" path="/var/lib/kubelet/pods/85913727-bbcd-4343-9faf-a75f40b42dc8/volumes" Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.824417 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="870ec55c-fc6f-4b16-8c49-a3cda0d0d010" path="/var/lib/kubelet/pods/870ec55c-fc6f-4b16-8c49-a3cda0d0d010/volumes" Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.825364 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b144897c-8640-4c5c-afa1-68b44dfd4f98" path="/var/lib/kubelet/pods/b144897c-8640-4c5c-afa1-68b44dfd4f98/volumes" Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.826212 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cbb6c089-0112-43f4-8731-c2d68932795a" path="/var/lib/kubelet/pods/cbb6c089-0112-43f4-8731-c2d68932795a/volumes" Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.827774 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e2585497-6837-4d17-8d51-c3d7879fdb46" path="/var/lib/kubelet/pods/e2585497-6837-4d17-8d51-c3d7879fdb46/volumes" Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.828976 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-4v4ps"] Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.845018 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-ovs-vp6f5"] Mar 13 14:30:41 crc kubenswrapper[4907]: E0313 14:30:41.854267 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 46f9fcec5ce705120f46ecb4deec4665885423d12b23cabce7f929fa5941d7ab is running failed: container process not found" containerID="46f9fcec5ce705120f46ecb4deec4665885423d12b23cabce7f929fa5941d7ab" cmd=["/usr/bin/pidof","ovsdb-server"] Mar 13 14:30:41 crc kubenswrapper[4907]: E0313 14:30:41.863042 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 46f9fcec5ce705120f46ecb4deec4665885423d12b23cabce7f929fa5941d7ab is running failed: container process not found" containerID="46f9fcec5ce705120f46ecb4deec4665885423d12b23cabce7f929fa5941d7ab" cmd=["/usr/bin/pidof","ovsdb-server"] Mar 13 14:30:41 crc kubenswrapper[4907]: E0313 14:30:41.869780 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 46f9fcec5ce705120f46ecb4deec4665885423d12b23cabce7f929fa5941d7ab is running failed: container process not found" containerID="46f9fcec5ce705120f46ecb4deec4665885423d12b23cabce7f929fa5941d7ab" cmd=["/usr/bin/pidof","ovsdb-server"] Mar 13 14:30:41 crc kubenswrapper[4907]: E0313 14:30:41.869845 4907 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 46f9fcec5ce705120f46ecb4deec4665885423d12b23cabce7f929fa5941d7ab is running failed: container process not found" probeType="Readiness" pod="openstack/ovsdbserver-nb-0" podUID="056dd756-0d7b-471b-9929-f622d05ad606" containerName="ovsdbserver-nb" Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.874947 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.875174 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="3793f53a-ca36-4b27-8444-d6dbfd860424" containerName="cinder-scheduler" containerID="cri-o://d484e26e1510dc0830d22229bd3bbd9bfa0e7d5cab421e8e5b87043a8df1abeb" gracePeriod=30 Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.875516 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="3793f53a-ca36-4b27-8444-d6dbfd860424" containerName="probe" containerID="cri-o://94bc55f3afe5f98100548e7593f16440afc25ced9f4af83fb6cd9f896253e749" gracePeriod=30 Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.901306 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.918671 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-f6fe-account-create-update-56786"] Mar 13 14:30:41 crc kubenswrapper[4907]: E0313 14:30:41.949605 4907 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 13 14:30:41 crc kubenswrapper[4907]: container &Container{Name:mariadb-account-create-update,Image:quay.io/podified-antelope-centos9/openstack-mariadb@sha256:4caef2b55e01b9a7ee88a22bc69db1893521a91d95c7ad4c8e593f14f17a5f95,Command:[/bin/sh -c #!/bin/bash Mar 13 14:30:41 crc kubenswrapper[4907]: Mar 13 14:30:41 crc kubenswrapper[4907]: MYSQL_REMOTE_HOST="" source /var/lib/operator-scripts/mysql_root_auth.sh Mar 13 14:30:41 crc kubenswrapper[4907]: Mar 13 14:30:41 crc kubenswrapper[4907]: export DatabasePassword=${DatabasePassword:?"Please specify a DatabasePassword variable."} Mar 13 14:30:41 crc kubenswrapper[4907]: Mar 13 14:30:41 crc kubenswrapper[4907]: MYSQL_CMD="mysql -h -u root -P 3306" Mar 13 14:30:41 crc kubenswrapper[4907]: Mar 13 14:30:41 crc kubenswrapper[4907]: if [ -n "neutron" ]; then Mar 13 14:30:41 crc kubenswrapper[4907]: GRANT_DATABASE="neutron" Mar 13 14:30:41 crc kubenswrapper[4907]: else Mar 13 14:30:41 crc kubenswrapper[4907]: GRANT_DATABASE="*" Mar 13 14:30:41 crc kubenswrapper[4907]: fi Mar 13 14:30:41 crc kubenswrapper[4907]: Mar 13 14:30:41 crc kubenswrapper[4907]: # going for maximum compatibility here: Mar 13 14:30:41 crc kubenswrapper[4907]: # 1. MySQL 8 no longer allows implicit create user when GRANT is used Mar 13 14:30:41 crc kubenswrapper[4907]: # 2. MariaDB has "CREATE OR REPLACE", but MySQL does not Mar 13 14:30:41 crc kubenswrapper[4907]: # 3. create user with CREATE but then do all password and TLS with ALTER to Mar 13 14:30:41 crc kubenswrapper[4907]: # support updates Mar 13 14:30:41 crc kubenswrapper[4907]: Mar 13 14:30:41 crc kubenswrapper[4907]: $MYSQL_CMD < logger="UnhandledError" Mar 13 14:30:41 crc kubenswrapper[4907]: E0313 14:30:41.951751 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mariadb-account-create-update\" with CreateContainerConfigError: \"secret \\\"neutron-db-secret\\\" not found\"" pod="openstack/neutron-b69e-account-create-update-96wq7" podUID="e8f6894c-b714-431e-912d-8b8de1a3914b" Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.956919 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-7bf587b8d9-pvvbz"] Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.957718 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-7bf587b8d9-pvvbz" podUID="001078cc-c01e-4173-9740-4f9ff082c593" containerName="neutron-api" containerID="cri-o://8f41b5a31898b937ed3fab2872e25ea71023ac05d2c81d4552d7e5ae8d96095e" gracePeriod=30 Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.958442 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-7bf587b8d9-pvvbz" podUID="001078cc-c01e-4173-9740-4f9ff082c593" containerName="neutron-httpd" containerID="cri-o://3e3eb6f5fca9c3693edb3599f1d3dd3d706131518c37c580379168a8a1b7e5e0" gracePeriod=30 Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.978210 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-f6fe-account-create-update-56786"] Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.996294 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-85dc486cc4-fwrm6"] Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.996616 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement-85dc486cc4-fwrm6" podUID="e3ee069a-41fb-4cab-b650-9bb9c03ac271" containerName="placement-log" containerID="cri-o://c0c472fe526cce04d2912530f4de0e17fa4ad4be341dbd2a462b13899fc1ad89" gracePeriod=30 Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.999216 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement-85dc486cc4-fwrm6" podUID="e3ee069a-41fb-4cab-b650-9bb9c03ac271" containerName="placement-api" containerID="cri-o://6d08d480e1be3a68c3d20d884244fa154245d26dcb3fb269a2aeb6806364407e" gracePeriod=30 Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.020958 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.021472 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76" containerName="cinder-api-log" containerID="cri-o://2cbce92733c8c36aab8d17f31bd5085e394cb7fec21de278bf9301bd315d6bae" gracePeriod=30 Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.021985 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76" containerName="cinder-api" containerID="cri-o://a07507237e2303a227977bfc1acdae544607145ad994269ea3936668b9e9c650" gracePeriod=30 Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.042576 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-storage-0"] Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.043313 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerName="account-server" containerID="cri-o://8168297a51ad087d7375035164e54e4a0d24e6fd56cf85e922e11d80f7f3d091" gracePeriod=30 Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.043423 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerName="container-updater" containerID="cri-o://dbcb2b472a15fbcc053532418c2f7a654827914050e539781030bc0749d6218f" gracePeriod=30 Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.043461 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerName="container-auditor" containerID="cri-o://7a94b1f20e2cb964aaf2fe5c9f23f0cd193291a14edd5411717f7cad81ed9808" gracePeriod=30 Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.043503 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerName="container-replicator" containerID="cri-o://2bc2275a73f47c1a27db775d1cc1c3ae9c20c01f9ddad7a35bf652467b66ea80" gracePeriod=30 Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.043531 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerName="container-server" containerID="cri-o://b7f60cadabc1fd2d193073423cce7296e305ff6325001c7ec9823dd267a1787a" gracePeriod=30 Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.043560 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerName="account-reaper" containerID="cri-o://c7370f4c24ed7b82c59bc2c1e3ae152c65420b5dd6227a66eb64b5b83832e410" gracePeriod=30 Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.043589 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerName="account-auditor" containerID="cri-o://f8bc1f3dd2242db8254b664f00b3789940491fa59b38a7fc033c225b233b776e" gracePeriod=30 Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.043619 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerName="account-replicator" containerID="cri-o://9b6f16a5610e8f2c08a40a29bfba0c00e1ff883cb29b5290604713aad1dab566" gracePeriod=30 Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.043734 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerName="object-expirer" containerID="cri-o://8765bf4cde625366ddd57fa87118a09b18b3fbbf1d2ded31c12344e9c08dadb2" gracePeriod=30 Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.043418 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerName="object-server" containerID="cri-o://c81c73db839bc006e1b7fdce225dd05a9a50870ea60ca389ea146e1e69d195b7" gracePeriod=30 Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.043889 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerName="object-replicator" containerID="cri-o://5a6b392bf5f4cbd99ba2fc751327322f8960f4c73aba15cb9a501d4c9c3edd94" gracePeriod=30 Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.043832 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerName="rsync" containerID="cri-o://1e5ba68b21b276693f03089f51ca3955dd88adcaacae52727d03a999b543d411" gracePeriod=30 Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.043848 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerName="swift-recon-cron" containerID="cri-o://0696dc2db616f4616e43dd171ba757d11bce09169b35b516964d1af58bad39ea" gracePeriod=30 Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.043857 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerName="object-auditor" containerID="cri-o://c1b4fe0c395990393cfbace1600ea9872fde43535722ed738332ba0b26f178a4" gracePeriod=30 Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.043867 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerName="object-updater" containerID="cri-o://6f5cf167a5953dd1bdb0f436b238215d64975f16e188a1dacd546c8d2adfba14" gracePeriod=30 Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.057241 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-6g25t"] Mar 13 14:30:42 crc kubenswrapper[4907]: E0313 14:30:42.067819 4907 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Mar 13 14:30:42 crc kubenswrapper[4907]: E0313 14:30:42.071449 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/e3c34454-315e-4821-ab25-b0f331a0d521-config-data podName:e3c34454-315e-4821-ab25-b0f331a0d521 nodeName:}" failed. No retries permitted until 2026-03-13 14:30:42.57142518 +0000 UTC m=+1541.471212869 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/e3c34454-315e-4821-ab25-b0f331a0d521-config-data") pod "rabbitmq-server-0" (UID: "e3c34454-315e-4821-ab25-b0f331a0d521") : configmap "rabbitmq-config-data" not found Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.087988 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-6g25t"] Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.138788 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-2pvp9"] Mar 13 14:30:42 crc kubenswrapper[4907]: W0313 14:30:42.146245 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4a4c64a3_7378_4d8d_9cf3_17583c6a90c0.slice/crio-a22452114ba800cf0f41433cad0c073716fcc8d240adefbf18595bc1cdb4a098 WatchSource:0}: Error finding container a22452114ba800cf0f41433cad0c073716fcc8d240adefbf18595bc1cdb4a098: Status 404 returned error can't find the container with id a22452114ba800cf0f41433cad0c073716fcc8d240adefbf18595bc1cdb4a098 Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.158299 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-2pvp9"] Mar 13 14:30:42 crc kubenswrapper[4907]: E0313 14:30:42.176445 4907 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 13 14:30:42 crc kubenswrapper[4907]: container &Container{Name:mariadb-account-create-update,Image:quay.io/podified-antelope-centos9/openstack-mariadb@sha256:4caef2b55e01b9a7ee88a22bc69db1893521a91d95c7ad4c8e593f14f17a5f95,Command:[/bin/sh -c #!/bin/bash Mar 13 14:30:42 crc kubenswrapper[4907]: Mar 13 14:30:42 crc kubenswrapper[4907]: MYSQL_REMOTE_HOST="" source /var/lib/operator-scripts/mysql_root_auth.sh Mar 13 14:30:42 crc kubenswrapper[4907]: Mar 13 14:30:42 crc kubenswrapper[4907]: export DatabasePassword=${DatabasePassword:?"Please specify a DatabasePassword variable."} Mar 13 14:30:42 crc kubenswrapper[4907]: Mar 13 14:30:42 crc kubenswrapper[4907]: MYSQL_CMD="mysql -h -u root -P 3306" Mar 13 14:30:42 crc kubenswrapper[4907]: Mar 13 14:30:42 crc kubenswrapper[4907]: if [ -n "placement" ]; then Mar 13 14:30:42 crc kubenswrapper[4907]: GRANT_DATABASE="placement" Mar 13 14:30:42 crc kubenswrapper[4907]: else Mar 13 14:30:42 crc kubenswrapper[4907]: GRANT_DATABASE="*" Mar 13 14:30:42 crc kubenswrapper[4907]: fi Mar 13 14:30:42 crc kubenswrapper[4907]: Mar 13 14:30:42 crc kubenswrapper[4907]: # going for maximum compatibility here: Mar 13 14:30:42 crc kubenswrapper[4907]: # 1. MySQL 8 no longer allows implicit create user when GRANT is used Mar 13 14:30:42 crc kubenswrapper[4907]: # 2. MariaDB has "CREATE OR REPLACE", but MySQL does not Mar 13 14:30:42 crc kubenswrapper[4907]: # 3. create user with CREATE but then do all password and TLS with ALTER to Mar 13 14:30:42 crc kubenswrapper[4907]: # support updates Mar 13 14:30:42 crc kubenswrapper[4907]: Mar 13 14:30:42 crc kubenswrapper[4907]: $MYSQL_CMD < logger="UnhandledError" Mar 13 14:30:42 crc kubenswrapper[4907]: E0313 14:30:42.177762 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mariadb-account-create-update\" with CreateContainerConfigError: \"secret \\\"placement-db-secret\\\" not found\"" pod="openstack/placement-9c64-account-create-update-5vvgg" podUID="4a4c64a3-7378-4d8d-9cf3-17583c6a90c0" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.182868 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-glzbl"] Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.194275 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-proxy-7dfdbcdd5-dcb7n"] Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.194540 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-proxy-7dfdbcdd5-dcb7n" podUID="1a1494c5-e2d6-4d29-b161-97b720742d8d" containerName="proxy-httpd" containerID="cri-o://676efd80769fe771754cb647e65098f243a686b0c64467414fa0d6ccfbece2d4" gracePeriod=30 Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.194669 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-proxy-7dfdbcdd5-dcb7n" podUID="1a1494c5-e2d6-4d29-b161-97b720742d8d" containerName="proxy-server" containerID="cri-o://c6c4a6a54081de702756857f0341351ea26c5f9aaa725027df2c989c6b1a6e18" gracePeriod=30 Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.201131 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.201335 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="4251ae36-90ad-41ea-915e-862df60f5c07" containerName="glance-log" containerID="cri-o://70b1012f350567e4a184430af6b8324fc4a32cf4f648d2eb406bd3d58d2b10d8" gracePeriod=30 Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.201453 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="4251ae36-90ad-41ea-915e-862df60f5c07" containerName="glance-httpd" containerID="cri-o://479ee0fb693df48bcf5ef06770a45535f793e4153c556058631f42347318a544" gracePeriod=30 Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.208949 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-x9b2b"] Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.226621 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-x9b2b"] Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.259793 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.260023 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="0b2e62fa-af35-4f35-8852-362ea97c17e7" containerName="glance-log" containerID="cri-o://eeaee513015ed79d41df919baf6b1bec0513b97e239abe794d310237db6cde32" gracePeriod=30 Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.260456 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="0b2e62fa-af35-4f35-8852-362ea97c17e7" containerName="glance-httpd" containerID="cri-o://0d9941f62564b14687332e4dbc1f81a8df8a858860923a1b915fe67a744f75f5" gracePeriod=30 Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.280288 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-cnnxs"] Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.296975 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-b69e-account-create-update-96wq7"] Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.329507 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-cnnxs"] Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.377911 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-keystone-listener-7d7bc9f7d-fks76"] Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.378145 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-keystone-listener-7d7bc9f7d-fks76" podUID="65531461-cea2-4b2c-a9b6-8fd9e9bffb27" containerName="barbican-keystone-listener-log" containerID="cri-o://3ccd8eb5dd33f0cece989e6411df4852cb55e2cb0b138f16332e7a07ccf5c865" gracePeriod=30 Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.378565 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-keystone-listener-7d7bc9f7d-fks76" podUID="65531461-cea2-4b2c-a9b6-8fd9e9bffb27" containerName="barbican-keystone-listener" containerID="cri-o://7117af57f620e2d772f6e2ca6d92edfdc283b3eb640ab414d9483ba1758f36d4" gracePeriod=30 Mar 13 14:30:42 crc kubenswrapper[4907]: E0313 14:30:42.385360 4907 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Mar 13 14:30:42 crc kubenswrapper[4907]: E0313 14:30:42.385439 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/286c82e8-b74a-49d7-a355-ac074aace10e-config-data podName:286c82e8-b74a-49d7-a355-ac074aace10e nodeName:}" failed. No retries permitted until 2026-03-13 14:30:44.385419186 +0000 UTC m=+1543.285206875 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/286c82e8-b74a-49d7-a355-ac074aace10e-config-data") pod "rabbitmq-cell1-server-0" (UID: "286c82e8-b74a-49d7-a355-ac074aace10e") : configmap "rabbitmq-cell1-config-data" not found Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.389312 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-b69e-account-create-update-96wq7"] Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.401983 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-worker-76bbcdf7d9-j6hw5"] Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.402203 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-worker-76bbcdf7d9-j6hw5" podUID="25b5fafe-65c8-48ba-bc34-83442f8ace4c" containerName="barbican-worker-log" containerID="cri-o://b5356b2175cd794b4fc6692eb61775ae05e66fb791f7dbd913a5d208e5880c25" gracePeriod=30 Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.402561 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-worker-76bbcdf7d9-j6hw5" podUID="25b5fafe-65c8-48ba-bc34-83442f8ace4c" containerName="barbican-worker" containerID="cri-o://be9d4c930500ca917844d7687055220642421756a075ce078989248af8d81176" gracePeriod=30 Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.418165 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-9c64-account-create-update-5vvgg"] Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.497182 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-7fc86597fb-mt9pk"] Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.502674 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-7fc86597fb-mt9pk" podUID="1b2f62d4-0584-4cc3-81ed-e067d8db23b1" containerName="barbican-api-log" containerID="cri-o://f21714a9f3b46d6b404fb8600bc2b71af7d623654fb77b4f61140d6330569927" gracePeriod=30 Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.502807 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-7fc86597fb-mt9pk" podUID="1b2f62d4-0584-4cc3-81ed-e067d8db23b1" containerName="barbican-api" containerID="cri-o://22aa297a2f364f3884c53d61c34c94086546e7771e899b38ffebaf99570e4f07" gracePeriod=30 Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.513561 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_056dd756-0d7b-471b-9929-f622d05ad606/ovsdbserver-nb/0.log" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.513668 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.524308 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.524524 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="2e41272f-6e7b-47a7-825e-3571f6a1fd07" containerName="nova-metadata-log" containerID="cri-o://c5ecad10cc9377aeda50af0d8c9213440ce021149e77309245ab84eb57179e9d" gracePeriod=30 Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.524676 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="2e41272f-6e7b-47a7-825e-3571f6a1fd07" containerName="nova-metadata-metadata" containerID="cri-o://ef8644527a89477971dc8c42d7fc7d2d08c1853a9324c6451f42196c1c50329e" gracePeriod=30 Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.525019 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_26827945-75f0-4867-ba04-31ff6428e06a/ovn-northd/0.log" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.525089 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.536529 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-fdb8f6449-nb4km" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.538675 4907 generic.go:334] "Generic (PLEG): container finished" podID="4251ae36-90ad-41ea-915e-862df60f5c07" containerID="70b1012f350567e4a184430af6b8324fc4a32cf4f648d2eb406bd3d58d2b10d8" exitCode=143 Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.538732 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"4251ae36-90ad-41ea-915e-862df60f5c07","Type":"ContainerDied","Data":"70b1012f350567e4a184430af6b8324fc4a32cf4f648d2eb406bd3d58d2b10d8"} Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.550629 4907 generic.go:334] "Generic (PLEG): container finished" podID="e3ee069a-41fb-4cab-b650-9bb9c03ac271" containerID="c0c472fe526cce04d2912530f4de0e17fa4ad4be341dbd2a462b13899fc1ad89" exitCode=143 Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.550683 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-85dc486cc4-fwrm6" event={"ID":"e3ee069a-41fb-4cab-b650-9bb9c03ac271","Type":"ContainerDied","Data":"c0c472fe526cce04d2912530f4de0e17fa4ad4be341dbd2a462b13899fc1ad89"} Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.566555 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-g96mx"] Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.581156 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-nhnkq_37978cc1-fcba-4032-a8b1-6632b61692ff/openstack-network-exporter/0.log" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.583616 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-nhnkq" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.585184 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-9c64-account-create-update-5vvgg" event={"ID":"4a4c64a3-7378-4d8d-9cf3-17583c6a90c0","Type":"ContainerStarted","Data":"a22452114ba800cf0f41433cad0c073716fcc8d240adefbf18595bc1cdb4a098"} Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.592218 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-g96mx"] Mar 13 14:30:42 crc kubenswrapper[4907]: E0313 14:30:42.595244 4907 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 13 14:30:42 crc kubenswrapper[4907]: container &Container{Name:mariadb-account-create-update,Image:quay.io/podified-antelope-centos9/openstack-mariadb@sha256:4caef2b55e01b9a7ee88a22bc69db1893521a91d95c7ad4c8e593f14f17a5f95,Command:[/bin/sh -c #!/bin/bash Mar 13 14:30:42 crc kubenswrapper[4907]: Mar 13 14:30:42 crc kubenswrapper[4907]: MYSQL_REMOTE_HOST="" source /var/lib/operator-scripts/mysql_root_auth.sh Mar 13 14:30:42 crc kubenswrapper[4907]: Mar 13 14:30:42 crc kubenswrapper[4907]: export DatabasePassword=${DatabasePassword:?"Please specify a DatabasePassword variable."} Mar 13 14:30:42 crc kubenswrapper[4907]: Mar 13 14:30:42 crc kubenswrapper[4907]: MYSQL_CMD="mysql -h -u root -P 3306" Mar 13 14:30:42 crc kubenswrapper[4907]: Mar 13 14:30:42 crc kubenswrapper[4907]: if [ -n "placement" ]; then Mar 13 14:30:42 crc kubenswrapper[4907]: GRANT_DATABASE="placement" Mar 13 14:30:42 crc kubenswrapper[4907]: else Mar 13 14:30:42 crc kubenswrapper[4907]: GRANT_DATABASE="*" Mar 13 14:30:42 crc kubenswrapper[4907]: fi Mar 13 14:30:42 crc kubenswrapper[4907]: Mar 13 14:30:42 crc kubenswrapper[4907]: # going for maximum compatibility here: Mar 13 14:30:42 crc kubenswrapper[4907]: # 1. MySQL 8 no longer allows implicit create user when GRANT is used Mar 13 14:30:42 crc kubenswrapper[4907]: # 2. MariaDB has "CREATE OR REPLACE", but MySQL does not Mar 13 14:30:42 crc kubenswrapper[4907]: # 3. create user with CREATE but then do all password and TLS with ALTER to Mar 13 14:30:42 crc kubenswrapper[4907]: # support updates Mar 13 14:30:42 crc kubenswrapper[4907]: Mar 13 14:30:42 crc kubenswrapper[4907]: $MYSQL_CMD < logger="UnhandledError" Mar 13 14:30:42 crc kubenswrapper[4907]: E0313 14:30:42.598283 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mariadb-account-create-update\" with CreateContainerConfigError: \"secret \\\"placement-db-secret\\\" not found\"" pod="openstack/placement-9c64-account-create-update-5vvgg" podUID="4a4c64a3-7378-4d8d-9cf3-17583c6a90c0" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.603867 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-b69e-account-create-update-96wq7" event={"ID":"e8f6894c-b714-431e-912d-8b8de1a3914b","Type":"ContainerStarted","Data":"d38b680247762c5b2510a8776b9df1c80f0da04bd9de35e5d2db2452cf7446f0"} Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.607151 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_c2938e39-f1d8-4cdc-a32c-5d57b8f2034f/ovsdbserver-sb/0.log" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.607234 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.607946 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/26827945-75f0-4867-ba04-31ff6428e06a-config\") pod \"26827945-75f0-4867-ba04-31ff6428e06a\" (UID: \"26827945-75f0-4867-ba04-31ff6428e06a\") " Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.608024 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/26827945-75f0-4867-ba04-31ff6428e06a-metrics-certs-tls-certs\") pod \"26827945-75f0-4867-ba04-31ff6428e06a\" (UID: \"26827945-75f0-4867-ba04-31ff6428e06a\") " Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.608082 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ckz4p\" (UniqueName: \"kubernetes.io/projected/056dd756-0d7b-471b-9929-f622d05ad606-kube-api-access-ckz4p\") pod \"056dd756-0d7b-471b-9929-f622d05ad606\" (UID: \"056dd756-0d7b-471b-9929-f622d05ad606\") " Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.608164 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/056dd756-0d7b-471b-9929-f622d05ad606-ovsdbserver-nb-tls-certs\") pod \"056dd756-0d7b-471b-9929-f622d05ad606\" (UID: \"056dd756-0d7b-471b-9929-f622d05ad606\") " Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.608203 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/056dd756-0d7b-471b-9929-f622d05ad606-combined-ca-bundle\") pod \"056dd756-0d7b-471b-9929-f622d05ad606\" (UID: \"056dd756-0d7b-471b-9929-f622d05ad606\") " Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.608278 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndbcluster-nb-etc-ovn\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"056dd756-0d7b-471b-9929-f622d05ad606\" (UID: \"056dd756-0d7b-471b-9929-f622d05ad606\") " Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.608318 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/26827945-75f0-4867-ba04-31ff6428e06a-ovn-rundir\") pod \"26827945-75f0-4867-ba04-31ff6428e06a\" (UID: \"26827945-75f0-4867-ba04-31ff6428e06a\") " Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.608337 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/056dd756-0d7b-471b-9929-f622d05ad606-config\") pod \"056dd756-0d7b-471b-9929-f622d05ad606\" (UID: \"056dd756-0d7b-471b-9929-f622d05ad606\") " Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.608401 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lgj9j\" (UniqueName: \"kubernetes.io/projected/26827945-75f0-4867-ba04-31ff6428e06a-kube-api-access-lgj9j\") pod \"26827945-75f0-4867-ba04-31ff6428e06a\" (UID: \"26827945-75f0-4867-ba04-31ff6428e06a\") " Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.608418 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26827945-75f0-4867-ba04-31ff6428e06a-combined-ca-bundle\") pod \"26827945-75f0-4867-ba04-31ff6428e06a\" (UID: \"26827945-75f0-4867-ba04-31ff6428e06a\") " Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.608501 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/056dd756-0d7b-471b-9929-f622d05ad606-ovsdb-rundir\") pod \"056dd756-0d7b-471b-9929-f622d05ad606\" (UID: \"056dd756-0d7b-471b-9929-f622d05ad606\") " Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.608565 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/26827945-75f0-4867-ba04-31ff6428e06a-scripts\") pod \"26827945-75f0-4867-ba04-31ff6428e06a\" (UID: \"26827945-75f0-4867-ba04-31ff6428e06a\") " Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.608659 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/056dd756-0d7b-471b-9929-f622d05ad606-scripts\") pod \"056dd756-0d7b-471b-9929-f622d05ad606\" (UID: \"056dd756-0d7b-471b-9929-f622d05ad606\") " Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.608680 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/26827945-75f0-4867-ba04-31ff6428e06a-ovn-northd-tls-certs\") pod \"26827945-75f0-4867-ba04-31ff6428e06a\" (UID: \"26827945-75f0-4867-ba04-31ff6428e06a\") " Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.608713 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/056dd756-0d7b-471b-9929-f622d05ad606-metrics-certs-tls-certs\") pod \"056dd756-0d7b-471b-9929-f622d05ad606\" (UID: \"056dd756-0d7b-471b-9929-f622d05ad606\") " Mar 13 14:30:42 crc kubenswrapper[4907]: E0313 14:30:42.609181 4907 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Mar 13 14:30:42 crc kubenswrapper[4907]: E0313 14:30:42.609221 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/e3c34454-315e-4821-ab25-b0f331a0d521-config-data podName:e3c34454-315e-4821-ab25-b0f331a0d521 nodeName:}" failed. No retries permitted until 2026-03-13 14:30:43.609209046 +0000 UTC m=+1542.508996735 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/e3c34454-315e-4821-ab25-b0f331a0d521-config-data") pod "rabbitmq-server-0" (UID: "e3c34454-315e-4821-ab25-b0f331a0d521") : configmap "rabbitmq-config-data" not found Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.612368 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/26827945-75f0-4867-ba04-31ff6428e06a-config" (OuterVolumeSpecName: "config") pod "26827945-75f0-4867-ba04-31ff6428e06a" (UID: "26827945-75f0-4867-ba04-31ff6428e06a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.615194 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 13 14:30:42 crc kubenswrapper[4907]: E0313 14:30:42.616822 4907 handlers.go:78] "Exec lifecycle hook for Container in Pod failed" err=< Mar 13 14:30:42 crc kubenswrapper[4907]: command '/usr/local/bin/container-scripts/stop-ovsdb-server.sh' exited with 137: ++ dirname /usr/local/bin/container-scripts/stop-ovsdb-server.sh Mar 13 14:30:42 crc kubenswrapper[4907]: + source /usr/local/bin/container-scripts/functions Mar 13 14:30:42 crc kubenswrapper[4907]: ++ OVNBridge=br-int Mar 13 14:30:42 crc kubenswrapper[4907]: ++ OVNRemote=tcp:localhost:6642 Mar 13 14:30:42 crc kubenswrapper[4907]: ++ OVNEncapType=geneve Mar 13 14:30:42 crc kubenswrapper[4907]: ++ OVNAvailabilityZones= Mar 13 14:30:42 crc kubenswrapper[4907]: ++ EnableChassisAsGateway=true Mar 13 14:30:42 crc kubenswrapper[4907]: ++ PhysicalNetworks= Mar 13 14:30:42 crc kubenswrapper[4907]: ++ OVNHostName= Mar 13 14:30:42 crc kubenswrapper[4907]: ++ DB_FILE=/etc/openvswitch/conf.db Mar 13 14:30:42 crc kubenswrapper[4907]: ++ ovs_dir=/var/lib/openvswitch Mar 13 14:30:42 crc kubenswrapper[4907]: ++ FLOWS_RESTORE_SCRIPT=/var/lib/openvswitch/flows-script Mar 13 14:30:42 crc kubenswrapper[4907]: ++ FLOWS_RESTORE_DIR=/var/lib/openvswitch/saved-flows Mar 13 14:30:42 crc kubenswrapper[4907]: ++ SAFE_TO_STOP_OVSDB_SERVER_SEMAPHORE=/var/lib/openvswitch/is_safe_to_stop_ovsdb_server Mar 13 14:30:42 crc kubenswrapper[4907]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Mar 13 14:30:42 crc kubenswrapper[4907]: + sleep 0.5 Mar 13 14:30:42 crc kubenswrapper[4907]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Mar 13 14:30:42 crc kubenswrapper[4907]: + cleanup_ovsdb_server_semaphore Mar 13 14:30:42 crc kubenswrapper[4907]: + rm -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server Mar 13 14:30:42 crc kubenswrapper[4907]: + /usr/share/openvswitch/scripts/ovs-ctl stop --no-ovs-vswitchd Mar 13 14:30:42 crc kubenswrapper[4907]: > execCommand=["/usr/local/bin/container-scripts/stop-ovsdb-server.sh"] containerName="ovsdb-server" pod="openstack/ovn-controller-ovs-vp6f5" message=< Mar 13 14:30:42 crc kubenswrapper[4907]: Exiting ovsdb-server (5) ++ dirname /usr/local/bin/container-scripts/stop-ovsdb-server.sh Mar 13 14:30:42 crc kubenswrapper[4907]: + source /usr/local/bin/container-scripts/functions Mar 13 14:30:42 crc kubenswrapper[4907]: ++ OVNBridge=br-int Mar 13 14:30:42 crc kubenswrapper[4907]: ++ OVNRemote=tcp:localhost:6642 Mar 13 14:30:42 crc kubenswrapper[4907]: ++ OVNEncapType=geneve Mar 13 14:30:42 crc kubenswrapper[4907]: ++ OVNAvailabilityZones= Mar 13 14:30:42 crc kubenswrapper[4907]: ++ EnableChassisAsGateway=true Mar 13 14:30:42 crc kubenswrapper[4907]: ++ PhysicalNetworks= Mar 13 14:30:42 crc kubenswrapper[4907]: ++ OVNHostName= Mar 13 14:30:42 crc kubenswrapper[4907]: ++ DB_FILE=/etc/openvswitch/conf.db Mar 13 14:30:42 crc kubenswrapper[4907]: ++ ovs_dir=/var/lib/openvswitch Mar 13 14:30:42 crc kubenswrapper[4907]: ++ FLOWS_RESTORE_SCRIPT=/var/lib/openvswitch/flows-script Mar 13 14:30:42 crc kubenswrapper[4907]: ++ FLOWS_RESTORE_DIR=/var/lib/openvswitch/saved-flows Mar 13 14:30:42 crc kubenswrapper[4907]: ++ SAFE_TO_STOP_OVSDB_SERVER_SEMAPHORE=/var/lib/openvswitch/is_safe_to_stop_ovsdb_server Mar 13 14:30:42 crc kubenswrapper[4907]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Mar 13 14:30:42 crc kubenswrapper[4907]: + sleep 0.5 Mar 13 14:30:42 crc kubenswrapper[4907]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Mar 13 14:30:42 crc kubenswrapper[4907]: + cleanup_ovsdb_server_semaphore Mar 13 14:30:42 crc kubenswrapper[4907]: + rm -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server Mar 13 14:30:42 crc kubenswrapper[4907]: + /usr/share/openvswitch/scripts/ovs-ctl stop --no-ovs-vswitchd Mar 13 14:30:42 crc kubenswrapper[4907]: > Mar 13 14:30:42 crc kubenswrapper[4907]: E0313 14:30:42.616895 4907 kuberuntime_container.go:691] "PreStop hook failed" err=< Mar 13 14:30:42 crc kubenswrapper[4907]: command '/usr/local/bin/container-scripts/stop-ovsdb-server.sh' exited with 137: ++ dirname /usr/local/bin/container-scripts/stop-ovsdb-server.sh Mar 13 14:30:42 crc kubenswrapper[4907]: + source /usr/local/bin/container-scripts/functions Mar 13 14:30:42 crc kubenswrapper[4907]: ++ OVNBridge=br-int Mar 13 14:30:42 crc kubenswrapper[4907]: ++ OVNRemote=tcp:localhost:6642 Mar 13 14:30:42 crc kubenswrapper[4907]: ++ OVNEncapType=geneve Mar 13 14:30:42 crc kubenswrapper[4907]: ++ OVNAvailabilityZones= Mar 13 14:30:42 crc kubenswrapper[4907]: ++ EnableChassisAsGateway=true Mar 13 14:30:42 crc kubenswrapper[4907]: ++ PhysicalNetworks= Mar 13 14:30:42 crc kubenswrapper[4907]: ++ OVNHostName= Mar 13 14:30:42 crc kubenswrapper[4907]: ++ DB_FILE=/etc/openvswitch/conf.db Mar 13 14:30:42 crc kubenswrapper[4907]: ++ ovs_dir=/var/lib/openvswitch Mar 13 14:30:42 crc kubenswrapper[4907]: ++ FLOWS_RESTORE_SCRIPT=/var/lib/openvswitch/flows-script Mar 13 14:30:42 crc kubenswrapper[4907]: ++ FLOWS_RESTORE_DIR=/var/lib/openvswitch/saved-flows Mar 13 14:30:42 crc kubenswrapper[4907]: ++ SAFE_TO_STOP_OVSDB_SERVER_SEMAPHORE=/var/lib/openvswitch/is_safe_to_stop_ovsdb_server Mar 13 14:30:42 crc kubenswrapper[4907]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Mar 13 14:30:42 crc kubenswrapper[4907]: + sleep 0.5 Mar 13 14:30:42 crc kubenswrapper[4907]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Mar 13 14:30:42 crc kubenswrapper[4907]: + cleanup_ovsdb_server_semaphore Mar 13 14:30:42 crc kubenswrapper[4907]: + rm -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server Mar 13 14:30:42 crc kubenswrapper[4907]: + /usr/share/openvswitch/scripts/ovs-ctl stop --no-ovs-vswitchd Mar 13 14:30:42 crc kubenswrapper[4907]: > pod="openstack/ovn-controller-ovs-vp6f5" podUID="c1b70392-1240-40d9-8128-e7abe29c8398" containerName="ovsdb-server" containerID="cri-o://6b0b25079cc099dabe922ac89fa41ae76d2f01797aac9f273b042e242459c574" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.616938 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-ovs-vp6f5" podUID="c1b70392-1240-40d9-8128-e7abe29c8398" containerName="ovsdb-server" containerID="cri-o://6b0b25079cc099dabe922ac89fa41ae76d2f01797aac9f273b042e242459c574" gracePeriod=30 Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.617321 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-htpvb"] Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.621217 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/056dd756-0d7b-471b-9929-f622d05ad606-ovsdb-rundir" (OuterVolumeSpecName: "ovsdb-rundir") pod "056dd756-0d7b-471b-9929-f622d05ad606" (UID: "056dd756-0d7b-471b-9929-f622d05ad606"). InnerVolumeSpecName "ovsdb-rundir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.622003 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/26827945-75f0-4867-ba04-31ff6428e06a-scripts" (OuterVolumeSpecName: "scripts") pod "26827945-75f0-4867-ba04-31ff6428e06a" (UID: "26827945-75f0-4867-ba04-31ff6428e06a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.622927 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/056dd756-0d7b-471b-9929-f622d05ad606-scripts" (OuterVolumeSpecName: "scripts") pod "056dd756-0d7b-471b-9929-f622d05ad606" (UID: "056dd756-0d7b-471b-9929-f622d05ad606"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.636384 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/26827945-75f0-4867-ba04-31ff6428e06a-ovn-rundir" (OuterVolumeSpecName: "ovn-rundir") pod "26827945-75f0-4867-ba04-31ff6428e06a" (UID: "26827945-75f0-4867-ba04-31ff6428e06a"). InnerVolumeSpecName "ovn-rundir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.638020 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-htpvb"] Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.638176 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/056dd756-0d7b-471b-9929-f622d05ad606-config" (OuterVolumeSpecName: "config") pod "056dd756-0d7b-471b-9929-f622d05ad606" (UID: "056dd756-0d7b-471b-9929-f622d05ad606"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:30:42 crc kubenswrapper[4907]: E0313 14:30:42.649121 4907 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 13 14:30:42 crc kubenswrapper[4907]: container &Container{Name:mariadb-account-create-update,Image:quay.io/podified-antelope-centos9/openstack-mariadb@sha256:4caef2b55e01b9a7ee88a22bc69db1893521a91d95c7ad4c8e593f14f17a5f95,Command:[/bin/sh -c #!/bin/bash Mar 13 14:30:42 crc kubenswrapper[4907]: Mar 13 14:30:42 crc kubenswrapper[4907]: MYSQL_REMOTE_HOST="" source /var/lib/operator-scripts/mysql_root_auth.sh Mar 13 14:30:42 crc kubenswrapper[4907]: Mar 13 14:30:42 crc kubenswrapper[4907]: export DatabasePassword=${DatabasePassword:?"Please specify a DatabasePassword variable."} Mar 13 14:30:42 crc kubenswrapper[4907]: Mar 13 14:30:42 crc kubenswrapper[4907]: MYSQL_CMD="mysql -h -u root -P 3306" Mar 13 14:30:42 crc kubenswrapper[4907]: Mar 13 14:30:42 crc kubenswrapper[4907]: if [ -n "neutron" ]; then Mar 13 14:30:42 crc kubenswrapper[4907]: GRANT_DATABASE="neutron" Mar 13 14:30:42 crc kubenswrapper[4907]: else Mar 13 14:30:42 crc kubenswrapper[4907]: GRANT_DATABASE="*" Mar 13 14:30:42 crc kubenswrapper[4907]: fi Mar 13 14:30:42 crc kubenswrapper[4907]: Mar 13 14:30:42 crc kubenswrapper[4907]: # going for maximum compatibility here: Mar 13 14:30:42 crc kubenswrapper[4907]: # 1. MySQL 8 no longer allows implicit create user when GRANT is used Mar 13 14:30:42 crc kubenswrapper[4907]: # 2. MariaDB has "CREATE OR REPLACE", but MySQL does not Mar 13 14:30:42 crc kubenswrapper[4907]: # 3. create user with CREATE but then do all password and TLS with ALTER to Mar 13 14:30:42 crc kubenswrapper[4907]: # support updates Mar 13 14:30:42 crc kubenswrapper[4907]: Mar 13 14:30:42 crc kubenswrapper[4907]: $MYSQL_CMD < logger="UnhandledError" Mar 13 14:30:42 crc kubenswrapper[4907]: E0313 14:30:42.654152 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mariadb-account-create-update\" with CreateContainerConfigError: \"secret \\\"neutron-db-secret\\\" not found\"" pod="openstack/neutron-b69e-account-create-update-96wq7" podUID="e8f6894c-b714-431e-912d-8b8de1a3914b" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.655423 4907 generic.go:334] "Generic (PLEG): container finished" podID="41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76" containerID="2cbce92733c8c36aab8d17f31bd5085e394cb7fec21de278bf9301bd315d6bae" exitCode=143 Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.655542 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76","Type":"ContainerDied","Data":"2cbce92733c8c36aab8d17f31bd5085e394cb7fec21de278bf9301bd315d6bae"} Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.660278 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "ovndbcluster-nb-etc-ovn") pod "056dd756-0d7b-471b-9929-f622d05ad606" (UID: "056dd756-0d7b-471b-9929-f622d05ad606"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.662074 4907 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-ovs-vp6f5" podUID="c1b70392-1240-40d9-8128-e7abe29c8398" containerName="ovs-vswitchd" probeResult="failure" output=< Mar 13 14:30:42 crc kubenswrapper[4907]: 2026-03-13T14:30:42Z|00001|jsonrpc|WARN|unix:/var/run/openvswitch/ovs-vswitchd.11.ctl: receive error: Connection reset by peer Mar 13 14:30:42 crc kubenswrapper[4907]: 2026-03-13T14:30:42Z|00002|unixctl|WARN|error communicating with unix:/var/run/openvswitch/ovs-vswitchd.11.ctl: Connection reset by peer Mar 13 14:30:42 crc kubenswrapper[4907]: ovs-appctl: /var/run/openvswitch/ovs-vswitchd.11.ctl: transaction error (Connection reset by peer) Mar 13 14:30:42 crc kubenswrapper[4907]: ERROR - Failed retrieving ofproto/list from ovs-vswitchd Mar 13 14:30:42 crc kubenswrapper[4907]: > Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.667297 4907 generic.go:334] "Generic (PLEG): container finished" podID="516d1499-3461-4af5-b426-9ae4a711a468" containerID="783ee4ed5f573bc2e551fccdb256e8146b5d15e716dc18f3c1e347aa5d0fa4c4" exitCode=1 Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.668255 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-glzbl" event={"ID":"516d1499-3461-4af5-b426-9ae4a711a468","Type":"ContainerDied","Data":"783ee4ed5f573bc2e551fccdb256e8146b5d15e716dc18f3c1e347aa5d0fa4c4"} Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.669833 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-glzbl" event={"ID":"516d1499-3461-4af5-b426-9ae4a711a468","Type":"ContainerStarted","Data":"f6d2f722265ae72bf48903452fd81587c07136e043b3e098f648b2d79cb3fd59"} Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.669770 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/26827945-75f0-4867-ba04-31ff6428e06a-kube-api-access-lgj9j" (OuterVolumeSpecName: "kube-api-access-lgj9j") pod "26827945-75f0-4867-ba04-31ff6428e06a" (UID: "26827945-75f0-4867-ba04-31ff6428e06a"). InnerVolumeSpecName "kube-api-access-lgj9j". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.673147 4907 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openstack/root-account-create-update-glzbl" secret="" err="secret \"galera-openstack-cell1-dockercfg-szj6q\" not found" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.673277 4907 scope.go:117] "RemoveContainer" containerID="783ee4ed5f573bc2e551fccdb256e8146b5d15e716dc18f3c1e347aa5d0fa4c4" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.676934 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-cell1-galera-0"] Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.690303 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/056dd756-0d7b-471b-9929-f622d05ad606-kube-api-access-ckz4p" (OuterVolumeSpecName: "kube-api-access-ckz4p") pod "056dd756-0d7b-471b-9929-f622d05ad606" (UID: "056dd756-0d7b-471b-9929-f622d05ad606"). InnerVolumeSpecName "kube-api-access-ckz4p". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.691832 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-b501-account-create-update-dzlqr"] Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.713453 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7da13722-732a-4b65-a894-fe4612f30d75-config\") pod \"7da13722-732a-4b65-a894-fe4612f30d75\" (UID: \"7da13722-732a-4b65-a894-fe4612f30d75\") " Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.713490 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/37978cc1-fcba-4032-a8b1-6632b61692ff-ovs-rundir\") pod \"37978cc1-fcba-4032-a8b1-6632b61692ff\" (UID: \"37978cc1-fcba-4032-a8b1-6632b61692ff\") " Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.713519 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/37978cc1-fcba-4032-a8b1-6632b61692ff-metrics-certs-tls-certs\") pod \"37978cc1-fcba-4032-a8b1-6632b61692ff\" (UID: \"37978cc1-fcba-4032-a8b1-6632b61692ff\") " Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.713550 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pxn9s\" (UniqueName: \"kubernetes.io/projected/c2938e39-f1d8-4cdc-a32c-5d57b8f2034f-kube-api-access-pxn9s\") pod \"c2938e39-f1d8-4cdc-a32c-5d57b8f2034f\" (UID: \"c2938e39-f1d8-4cdc-a32c-5d57b8f2034f\") " Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.713583 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7da13722-732a-4b65-a894-fe4612f30d75-ovsdbserver-sb\") pod \"7da13722-732a-4b65-a894-fe4612f30d75\" (UID: \"7da13722-732a-4b65-a894-fe4612f30d75\") " Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.713607 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c2938e39-f1d8-4cdc-a32c-5d57b8f2034f-metrics-certs-tls-certs\") pod \"c2938e39-f1d8-4cdc-a32c-5d57b8f2034f\" (UID: \"c2938e39-f1d8-4cdc-a32c-5d57b8f2034f\") " Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.713643 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c2938e39-f1d8-4cdc-a32c-5d57b8f2034f-ovsdb-rundir\") pod \"c2938e39-f1d8-4cdc-a32c-5d57b8f2034f\" (UID: \"c2938e39-f1d8-4cdc-a32c-5d57b8f2034f\") " Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.713663 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c2938e39-f1d8-4cdc-a32c-5d57b8f2034f-scripts\") pod \"c2938e39-f1d8-4cdc-a32c-5d57b8f2034f\" (UID: \"c2938e39-f1d8-4cdc-a32c-5d57b8f2034f\") " Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.713680 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndbcluster-sb-etc-ovn\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"c2938e39-f1d8-4cdc-a32c-5d57b8f2034f\" (UID: \"c2938e39-f1d8-4cdc-a32c-5d57b8f2034f\") " Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.713735 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c2938e39-f1d8-4cdc-a32c-5d57b8f2034f-ovsdbserver-sb-tls-certs\") pod \"c2938e39-f1d8-4cdc-a32c-5d57b8f2034f\" (UID: \"c2938e39-f1d8-4cdc-a32c-5d57b8f2034f\") " Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.713762 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ndzf5\" (UniqueName: \"kubernetes.io/projected/37978cc1-fcba-4032-a8b1-6632b61692ff-kube-api-access-ndzf5\") pod \"37978cc1-fcba-4032-a8b1-6632b61692ff\" (UID: \"37978cc1-fcba-4032-a8b1-6632b61692ff\") " Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.713785 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7da13722-732a-4b65-a894-fe4612f30d75-dns-svc\") pod \"7da13722-732a-4b65-a894-fe4612f30d75\" (UID: \"7da13722-732a-4b65-a894-fe4612f30d75\") " Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.713804 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g2kbb\" (UniqueName: \"kubernetes.io/projected/7da13722-732a-4b65-a894-fe4612f30d75-kube-api-access-g2kbb\") pod \"7da13722-732a-4b65-a894-fe4612f30d75\" (UID: \"7da13722-732a-4b65-a894-fe4612f30d75\") " Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.713838 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7da13722-732a-4b65-a894-fe4612f30d75-dns-swift-storage-0\") pod \"7da13722-732a-4b65-a894-fe4612f30d75\" (UID: \"7da13722-732a-4b65-a894-fe4612f30d75\") " Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.713861 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7da13722-732a-4b65-a894-fe4612f30d75-ovsdbserver-nb\") pod \"7da13722-732a-4b65-a894-fe4612f30d75\" (UID: \"7da13722-732a-4b65-a894-fe4612f30d75\") " Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.713915 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c2938e39-f1d8-4cdc-a32c-5d57b8f2034f-config\") pod \"c2938e39-f1d8-4cdc-a32c-5d57b8f2034f\" (UID: \"c2938e39-f1d8-4cdc-a32c-5d57b8f2034f\") " Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.713937 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/37978cc1-fcba-4032-a8b1-6632b61692ff-ovn-rundir\") pod \"37978cc1-fcba-4032-a8b1-6632b61692ff\" (UID: \"37978cc1-fcba-4032-a8b1-6632b61692ff\") " Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.713956 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37978cc1-fcba-4032-a8b1-6632b61692ff-combined-ca-bundle\") pod \"37978cc1-fcba-4032-a8b1-6632b61692ff\" (UID: \"37978cc1-fcba-4032-a8b1-6632b61692ff\") " Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.714001 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/37978cc1-fcba-4032-a8b1-6632b61692ff-config\") pod \"37978cc1-fcba-4032-a8b1-6632b61692ff\" (UID: \"37978cc1-fcba-4032-a8b1-6632b61692ff\") " Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.714029 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2938e39-f1d8-4cdc-a32c-5d57b8f2034f-combined-ca-bundle\") pod \"c2938e39-f1d8-4cdc-a32c-5d57b8f2034f\" (UID: \"c2938e39-f1d8-4cdc-a32c-5d57b8f2034f\") " Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.714512 4907 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.715217 4907 reconciler_common.go:293] "Volume detached for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/26827945-75f0-4867-ba04-31ff6428e06a-ovn-rundir\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.725326 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/056dd756-0d7b-471b-9929-f622d05ad606-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.726469 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lgj9j\" (UniqueName: \"kubernetes.io/projected/26827945-75f0-4867-ba04-31ff6428e06a-kube-api-access-lgj9j\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.726537 4907 reconciler_common.go:293] "Volume detached for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/056dd756-0d7b-471b-9929-f622d05ad606-ovsdb-rundir\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.726609 4907 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/26827945-75f0-4867-ba04-31ff6428e06a-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.727407 4907 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/056dd756-0d7b-471b-9929-f622d05ad606-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.727472 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/26827945-75f0-4867-ba04-31ff6428e06a-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.727539 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ckz4p\" (UniqueName: \"kubernetes.io/projected/056dd756-0d7b-471b-9929-f622d05ad606-kube-api-access-ckz4p\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.721764 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/37978cc1-fcba-4032-a8b1-6632b61692ff-config" (OuterVolumeSpecName: "config") pod "37978cc1-fcba-4032-a8b1-6632b61692ff" (UID: "37978cc1-fcba-4032-a8b1-6632b61692ff"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.722341 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c2938e39-f1d8-4cdc-a32c-5d57b8f2034f-config" (OuterVolumeSpecName: "config") pod "c2938e39-f1d8-4cdc-a32c-5d57b8f2034f" (UID: "c2938e39-f1d8-4cdc-a32c-5d57b8f2034f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.723998 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/37978cc1-fcba-4032-a8b1-6632b61692ff-ovn-rundir" (OuterVolumeSpecName: "ovn-rundir") pod "37978cc1-fcba-4032-a8b1-6632b61692ff" (UID: "37978cc1-fcba-4032-a8b1-6632b61692ff"). InnerVolumeSpecName "ovn-rundir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.726197 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c2938e39-f1d8-4cdc-a32c-5d57b8f2034f-ovsdb-rundir" (OuterVolumeSpecName: "ovsdb-rundir") pod "c2938e39-f1d8-4cdc-a32c-5d57b8f2034f" (UID: "c2938e39-f1d8-4cdc-a32c-5d57b8f2034f"). InnerVolumeSpecName "ovsdb-rundir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.726225 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/37978cc1-fcba-4032-a8b1-6632b61692ff-ovs-rundir" (OuterVolumeSpecName: "ovs-rundir") pod "37978cc1-fcba-4032-a8b1-6632b61692ff" (UID: "37978cc1-fcba-4032-a8b1-6632b61692ff"). InnerVolumeSpecName "ovs-rundir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.716182 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="286c82e8-b74a-49d7-a355-ac074aace10e" containerName="rabbitmq" containerID="cri-o://d23153977130306af0e21b41c03dbbc96a217b011704059af7d65c0a4153cf43" gracePeriod=604800 Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.716360 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_26827945-75f0-4867-ba04-31ff6428e06a/ovn-northd/0.log" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.721036 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-b501-account-create-update-dzlqr"] Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.731207 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-sgpkl"] Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.736586 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"26827945-75f0-4867-ba04-31ff6428e06a","Type":"ContainerDied","Data":"59513110b8b488ed882794dedcca28bd1c64f5a6610bf84e8eaeafbea7ce6daa"} Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.736797 4907 scope.go:117] "RemoveContainer" containerID="59dc65eade55429a5d17c956c987a8209b624ac898dd355bca1d93bdf510b508" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.728608 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.730662 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/056dd756-0d7b-471b-9929-f622d05ad606-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "056dd756-0d7b-471b-9929-f622d05ad606" (UID: "056dd756-0d7b-471b-9929-f622d05ad606"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.740620 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c2938e39-f1d8-4cdc-a32c-5d57b8f2034f-scripts" (OuterVolumeSpecName: "scripts") pod "c2938e39-f1d8-4cdc-a32c-5d57b8f2034f" (UID: "c2938e39-f1d8-4cdc-a32c-5d57b8f2034f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.752595 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c2938e39-f1d8-4cdc-a32c-5d57b8f2034f-kube-api-access-pxn9s" (OuterVolumeSpecName: "kube-api-access-pxn9s") pod "c2938e39-f1d8-4cdc-a32c-5d57b8f2034f" (UID: "c2938e39-f1d8-4cdc-a32c-5d57b8f2034f"). InnerVolumeSpecName "kube-api-access-pxn9s". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.753580 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-ovs-vp6f5" podUID="c1b70392-1240-40d9-8128-e7abe29c8398" containerName="ovs-vswitchd" containerID="cri-o://265b6b6fe5f95195cdc86c65456998ad327fc7ff3a6641300cd6ac00798eff4f" gracePeriod=30 Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.754908 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7da13722-732a-4b65-a894-fe4612f30d75-kube-api-access-g2kbb" (OuterVolumeSpecName: "kube-api-access-g2kbb") pod "7da13722-732a-4b65-a894-fe4612f30d75" (UID: "7da13722-732a-4b65-a894-fe4612f30d75"). InnerVolumeSpecName "kube-api-access-g2kbb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.756359 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/37978cc1-fcba-4032-a8b1-6632b61692ff-kube-api-access-ndzf5" (OuterVolumeSpecName: "kube-api-access-ndzf5") pod "37978cc1-fcba-4032-a8b1-6632b61692ff" (UID: "37978cc1-fcba-4032-a8b1-6632b61692ff"). InnerVolumeSpecName "kube-api-access-ndzf5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.756677 4907 generic.go:334] "Generic (PLEG): container finished" podID="0b2e62fa-af35-4f35-8852-362ea97c17e7" containerID="eeaee513015ed79d41df919baf6b1bec0513b97e239abe794d310237db6cde32" exitCode=143 Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.756766 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"0b2e62fa-af35-4f35-8852-362ea97c17e7","Type":"ContainerDied","Data":"eeaee513015ed79d41df919baf6b1bec0513b97e239abe794d310237db6cde32"} Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.760381 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-sgpkl"] Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.768764 4907 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.779234 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-17a8-account-create-update-nj9pw"] Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.788032 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "ovndbcluster-sb-etc-ovn") pod "c2938e39-f1d8-4cdc-a32c-5d57b8f2034f" (UID: "c2938e39-f1d8-4cdc-a32c-5d57b8f2034f"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.795181 4907 generic.go:334] "Generic (PLEG): container finished" podID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerID="8765bf4cde625366ddd57fa87118a09b18b3fbbf1d2ded31c12344e9c08dadb2" exitCode=0 Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.795220 4907 generic.go:334] "Generic (PLEG): container finished" podID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerID="6f5cf167a5953dd1bdb0f436b238215d64975f16e188a1dacd546c8d2adfba14" exitCode=0 Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.795232 4907 generic.go:334] "Generic (PLEG): container finished" podID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerID="c1b4fe0c395990393cfbace1600ea9872fde43535722ed738332ba0b26f178a4" exitCode=0 Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.795240 4907 generic.go:334] "Generic (PLEG): container finished" podID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerID="5a6b392bf5f4cbd99ba2fc751327322f8960f4c73aba15cb9a501d4c9c3edd94" exitCode=0 Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.795248 4907 generic.go:334] "Generic (PLEG): container finished" podID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerID="dbcb2b472a15fbcc053532418c2f7a654827914050e539781030bc0749d6218f" exitCode=0 Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.795256 4907 generic.go:334] "Generic (PLEG): container finished" podID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerID="7a94b1f20e2cb964aaf2fe5c9f23f0cd193291a14edd5411717f7cad81ed9808" exitCode=0 Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.795264 4907 generic.go:334] "Generic (PLEG): container finished" podID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerID="2bc2275a73f47c1a27db775d1cc1c3ae9c20c01f9ddad7a35bf652467b66ea80" exitCode=0 Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.795273 4907 generic.go:334] "Generic (PLEG): container finished" podID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerID="c7370f4c24ed7b82c59bc2c1e3ae152c65420b5dd6227a66eb64b5b83832e410" exitCode=0 Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.795281 4907 generic.go:334] "Generic (PLEG): container finished" podID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerID="f8bc1f3dd2242db8254b664f00b3789940491fa59b38a7fc033c225b233b776e" exitCode=0 Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.795289 4907 generic.go:334] "Generic (PLEG): container finished" podID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerID="9b6f16a5610e8f2c08a40a29bfba0c00e1ff883cb29b5290604713aad1dab566" exitCode=0 Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.795336 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"14f05f31-8185-43b0-be69-bcf8d5388ea5","Type":"ContainerDied","Data":"8765bf4cde625366ddd57fa87118a09b18b3fbbf1d2ded31c12344e9c08dadb2"} Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.795368 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"14f05f31-8185-43b0-be69-bcf8d5388ea5","Type":"ContainerDied","Data":"6f5cf167a5953dd1bdb0f436b238215d64975f16e188a1dacd546c8d2adfba14"} Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.795381 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"14f05f31-8185-43b0-be69-bcf8d5388ea5","Type":"ContainerDied","Data":"c1b4fe0c395990393cfbace1600ea9872fde43535722ed738332ba0b26f178a4"} Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.795394 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"14f05f31-8185-43b0-be69-bcf8d5388ea5","Type":"ContainerDied","Data":"5a6b392bf5f4cbd99ba2fc751327322f8960f4c73aba15cb9a501d4c9c3edd94"} Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.795405 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"14f05f31-8185-43b0-be69-bcf8d5388ea5","Type":"ContainerDied","Data":"dbcb2b472a15fbcc053532418c2f7a654827914050e539781030bc0749d6218f"} Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.795418 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"14f05f31-8185-43b0-be69-bcf8d5388ea5","Type":"ContainerDied","Data":"7a94b1f20e2cb964aaf2fe5c9f23f0cd193291a14edd5411717f7cad81ed9808"} Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.795430 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"14f05f31-8185-43b0-be69-bcf8d5388ea5","Type":"ContainerDied","Data":"2bc2275a73f47c1a27db775d1cc1c3ae9c20c01f9ddad7a35bf652467b66ea80"} Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.795441 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"14f05f31-8185-43b0-be69-bcf8d5388ea5","Type":"ContainerDied","Data":"c7370f4c24ed7b82c59bc2c1e3ae152c65420b5dd6227a66eb64b5b83832e410"} Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.795453 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"14f05f31-8185-43b0-be69-bcf8d5388ea5","Type":"ContainerDied","Data":"f8bc1f3dd2242db8254b664f00b3789940491fa59b38a7fc033c225b233b776e"} Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.795461 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"14f05f31-8185-43b0-be69-bcf8d5388ea5","Type":"ContainerDied","Data":"9b6f16a5610e8f2c08a40a29bfba0c00e1ff883cb29b5290604713aad1dab566"} Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.795629 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.795849 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="cc227d33-3bf1-4b5f-8368-8e6b8a512f7d" containerName="nova-api-log" containerID="cri-o://6c1fdac6667e489d30e980acdd57d1cf0fa750bc21c6f46f11adb992a74fe2bd" gracePeriod=30 Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.796257 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="cc227d33-3bf1-4b5f-8368-8e6b8a512f7d" containerName="nova-api-api" containerID="cri-o://58121d89217f995e7e932419445b98334a3ccb8b202e18c097026d64131a91fd" gracePeriod=30 Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.796368 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/26827945-75f0-4867-ba04-31ff6428e06a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "26827945-75f0-4867-ba04-31ff6428e06a" (UID: "26827945-75f0-4867-ba04-31ff6428e06a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.813771 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_056dd756-0d7b-471b-9929-f622d05ad606/ovsdbserver-nb/0.log" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.813822 4907 generic.go:334] "Generic (PLEG): container finished" podID="056dd756-0d7b-471b-9929-f622d05ad606" containerID="46f9fcec5ce705120f46ecb4deec4665885423d12b23cabce7f929fa5941d7ab" exitCode=143 Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.813946 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.814582 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"056dd756-0d7b-471b-9929-f622d05ad606","Type":"ContainerDied","Data":"46f9fcec5ce705120f46ecb4deec4665885423d12b23cabce7f929fa5941d7ab"} Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.814639 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"056dd756-0d7b-471b-9929-f622d05ad606","Type":"ContainerDied","Data":"8f94280adbcbea9567f862d27736043270ac488194f6b0213cdb79c3aa5a9dee"} Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.829773 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26827945-75f0-4867-ba04-31ff6428e06a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.829801 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c2938e39-f1d8-4cdc-a32c-5d57b8f2034f-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.829810 4907 reconciler_common.go:293] "Volume detached for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/37978cc1-fcba-4032-a8b1-6632b61692ff-ovn-rundir\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.829819 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/37978cc1-fcba-4032-a8b1-6632b61692ff-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.829827 4907 reconciler_common.go:293] "Volume detached for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/37978cc1-fcba-4032-a8b1-6632b61692ff-ovs-rundir\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.829835 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pxn9s\" (UniqueName: \"kubernetes.io/projected/c2938e39-f1d8-4cdc-a32c-5d57b8f2034f-kube-api-access-pxn9s\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.829844 4907 reconciler_common.go:293] "Volume detached for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c2938e39-f1d8-4cdc-a32c-5d57b8f2034f-ovsdb-rundir\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.829853 4907 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c2938e39-f1d8-4cdc-a32c-5d57b8f2034f-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.829871 4907 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.829938 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/056dd756-0d7b-471b-9929-f622d05ad606-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.829949 4907 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.829958 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ndzf5\" (UniqueName: \"kubernetes.io/projected/37978cc1-fcba-4032-a8b1-6632b61692ff-kube-api-access-ndzf5\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.829967 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g2kbb\" (UniqueName: \"kubernetes.io/projected/7da13722-732a-4b65-a894-fe4612f30d75-kube-api-access-g2kbb\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:42 crc kubenswrapper[4907]: E0313 14:30:42.830473 4907 configmap.go:193] Couldn't get configMap openstack/openstack-cell1-scripts: configmap "openstack-cell1-scripts" not found Mar 13 14:30:42 crc kubenswrapper[4907]: E0313 14:30:42.830559 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/516d1499-3461-4af5-b426-9ae4a711a468-operator-scripts podName:516d1499-3461-4af5-b426-9ae4a711a468 nodeName:}" failed. No retries permitted until 2026-03-13 14:30:43.330540211 +0000 UTC m=+1542.230327900 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/516d1499-3461-4af5-b426-9ae4a711a468-operator-scripts") pod "root-account-create-update-glzbl" (UID: "516d1499-3461-4af5-b426-9ae4a711a468") : configmap "openstack-cell1-scripts" not found Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.843009 4907 scope.go:117] "RemoveContainer" containerID="e23ef8c27779c528c50bff3fd731469de16d17d8b86a266ab0618151d6dd930b" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.852429 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-nhnkq_37978cc1-fcba-4032-a8b1-6632b61692ff/openstack-network-exporter/0.log" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.852476 4907 generic.go:334] "Generic (PLEG): container finished" podID="37978cc1-fcba-4032-a8b1-6632b61692ff" containerID="62edff4734da4ab22bc7673f1d2c0e9a9db082a3ca22b12fadc41486071fe66a" exitCode=2 Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.852595 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-nhnkq" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.852910 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-nhnkq" event={"ID":"37978cc1-fcba-4032-a8b1-6632b61692ff","Type":"ContainerDied","Data":"62edff4734da4ab22bc7673f1d2c0e9a9db082a3ca22b12fadc41486071fe66a"} Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.855166 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-9c64-account-create-update-5vvgg"] Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.856710 4907 generic.go:334] "Generic (PLEG): container finished" podID="a055ce7f-0538-4e5e-938a-eaf6d0aa2dec" containerID="3e056a670aa7307d9d211e2123e0cead9b6cd31175ff09e1feddb508c47f1a88" exitCode=137 Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.859437 4907 generic.go:334] "Generic (PLEG): container finished" podID="7da13722-732a-4b65-a894-fe4612f30d75" containerID="890693bd8cb16d526cbd4e7c94b7a9b8dcd67a1bc48dbb881c81f86b4e9727d1" exitCode=0 Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.859466 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-fdb8f6449-nb4km" event={"ID":"7da13722-732a-4b65-a894-fe4612f30d75","Type":"ContainerDied","Data":"890693bd8cb16d526cbd4e7c94b7a9b8dcd67a1bc48dbb881c81f86b4e9727d1"} Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.859574 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-fdb8f6449-nb4km" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.862413 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-vlp6r"] Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.869661 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-vlp6r"] Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.874198 4907 scope.go:117] "RemoveContainer" containerID="f294f1dfa23d177ca2e8f5511b6bd495bd3146c4baee6d3b95ccd81e1b4496fb" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.877963 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-ldbt9"] Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.890107 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-47e2-account-create-update-wvjr7"] Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.904746 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-z92h5"] Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.916688 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-vbcql"] Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.923424 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-z92h5"] Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.927051 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/056dd756-0d7b-471b-9929-f622d05ad606-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "056dd756-0d7b-471b-9929-f622d05ad606" (UID: "056dd756-0d7b-471b-9929-f622d05ad606"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.928661 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-ldbt9"] Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.933893 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7da13722-732a-4b65-a894-fe4612f30d75-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "7da13722-732a-4b65-a894-fe4612f30d75" (UID: "7da13722-732a-4b65-a894-fe4612f30d75"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.933958 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7da13722-732a-4b65-a894-fe4612f30d75-ovsdbserver-nb\") pod \"7da13722-732a-4b65-a894-fe4612f30d75\" (UID: \"7da13722-732a-4b65-a894-fe4612f30d75\") " Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.934496 4907 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/056dd756-0d7b-471b-9929-f622d05ad606-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:42 crc kubenswrapper[4907]: W0313 14:30:42.935387 4907 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/7da13722-732a-4b65-a894-fe4612f30d75/volumes/kubernetes.io~configmap/ovsdbserver-nb Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.935409 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7da13722-732a-4b65-a894-fe4612f30d75-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "7da13722-732a-4b65-a894-fe4612f30d75" (UID: "7da13722-732a-4b65-a894-fe4612f30d75"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.936679 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-vbcql"] Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.943201 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.943387 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell0-conductor-0" podUID="364a702e-e6c6-42bc-8cee-4ba19ea2f1ec" containerName="nova-cell0-conductor-conductor" containerID="cri-o://5ab7cf17ecb1804f362340163cfceb48e6f2c8900108799d01d1025c485c1e8f" gracePeriod=30 Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.949717 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.949870 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-conductor-0" podUID="25c1c7e6-a149-4a8e-aead-3ebdc7e56851" containerName="nova-cell1-conductor-conductor" containerID="cri-o://77b52c0f64f55952d2255d5814b60083fd608735671f1f8507b09ea2e83db4d4" gracePeriod=30 Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.956774 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-xw7n9"] Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.968362 4907 scope.go:117] "RemoveContainer" containerID="46f9fcec5ce705120f46ecb4deec4665885423d12b23cabce7f929fa5941d7ab" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.972705 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-xw7n9"] Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.981246 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.981646 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="62983e04-f043-4ad3-9a46-96179d59671e" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://9e2de4df42a807d1995b769c3c69187ddaaf7da1c1d09354b07088fb415f36d7" gracePeriod=30 Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:42.993609 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/root-account-create-update-glzbl"] Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:42.995234 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/056dd756-0d7b-471b-9929-f622d05ad606-ovsdbserver-nb-tls-certs" (OuterVolumeSpecName: "ovsdbserver-nb-tls-certs") pod "056dd756-0d7b-471b-9929-f622d05ad606" (UID: "056dd756-0d7b-471b-9929-f622d05ad606"). InnerVolumeSpecName "ovsdbserver-nb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.000864 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.001587 4907 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.039164 4907 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7da13722-732a-4b65-a894-fe4612f30d75-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.039428 4907 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.039438 4907 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/056dd756-0d7b-471b-9929-f622d05ad606-ovsdbserver-nb-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.045190 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2938e39-f1d8-4cdc-a32c-5d57b8f2034f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c2938e39-f1d8-4cdc-a32c-5d57b8f2034f" (UID: "c2938e39-f1d8-4cdc-a32c-5d57b8f2034f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:43 crc kubenswrapper[4907]: E0313 14:30:43.046040 4907 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 13 14:30:43 crc kubenswrapper[4907]: container &Container{Name:mariadb-account-create-update,Image:quay.io/podified-antelope-centos9/openstack-mariadb@sha256:4caef2b55e01b9a7ee88a22bc69db1893521a91d95c7ad4c8e593f14f17a5f95,Command:[/bin/sh -c #!/bin/bash Mar 13 14:30:43 crc kubenswrapper[4907]: Mar 13 14:30:43 crc kubenswrapper[4907]: MYSQL_REMOTE_HOST="" source /var/lib/operator-scripts/mysql_root_auth.sh Mar 13 14:30:43 crc kubenswrapper[4907]: Mar 13 14:30:43 crc kubenswrapper[4907]: export DatabasePassword=${DatabasePassword:?"Please specify a DatabasePassword variable."} Mar 13 14:30:43 crc kubenswrapper[4907]: Mar 13 14:30:43 crc kubenswrapper[4907]: MYSQL_CMD="mysql -h -u root -P 3306" Mar 13 14:30:43 crc kubenswrapper[4907]: Mar 13 14:30:43 crc kubenswrapper[4907]: if [ -n "cinder" ]; then Mar 13 14:30:43 crc kubenswrapper[4907]: GRANT_DATABASE="cinder" Mar 13 14:30:43 crc kubenswrapper[4907]: else Mar 13 14:30:43 crc kubenswrapper[4907]: GRANT_DATABASE="*" Mar 13 14:30:43 crc kubenswrapper[4907]: fi Mar 13 14:30:43 crc kubenswrapper[4907]: Mar 13 14:30:43 crc kubenswrapper[4907]: # going for maximum compatibility here: Mar 13 14:30:43 crc kubenswrapper[4907]: # 1. MySQL 8 no longer allows implicit create user when GRANT is used Mar 13 14:30:43 crc kubenswrapper[4907]: # 2. MariaDB has "CREATE OR REPLACE", but MySQL does not Mar 13 14:30:43 crc kubenswrapper[4907]: # 3. create user with CREATE but then do all password and TLS with ALTER to Mar 13 14:30:43 crc kubenswrapper[4907]: # support updates Mar 13 14:30:43 crc kubenswrapper[4907]: Mar 13 14:30:43 crc kubenswrapper[4907]: $MYSQL_CMD < logger="UnhandledError" Mar 13 14:30:43 crc kubenswrapper[4907]: E0313 14:30:43.048426 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mariadb-account-create-update\" with CreateContainerConfigError: \"secret \\\"cinder-db-secret\\\" not found\"" pod="openstack/cinder-47e2-account-create-update-wvjr7" podUID="82dec019-ba60-43f5-845b-322a6bcca225" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.105781 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/26827945-75f0-4867-ba04-31ff6428e06a-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "26827945-75f0-4867-ba04-31ff6428e06a" (UID: "26827945-75f0-4867-ba04-31ff6428e06a"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.140918 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2938e39-f1d8-4cdc-a32c-5d57b8f2034f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.140960 4907 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/26827945-75f0-4867-ba04-31ff6428e06a-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.143941 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstack-cell1-galera-0" podUID="58ab9689-3ab2-42f2-ad56-beb22e29f8da" containerName="galera" containerID="cri-o://af7b1c6225d2e94f5a2f9827f7b02dc18651ab20c1d53b4a73e6d032e463897a" gracePeriod=30 Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.159400 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7da13722-732a-4b65-a894-fe4612f30d75-config" (OuterVolumeSpecName: "config") pod "7da13722-732a-4b65-a894-fe4612f30d75" (UID: "7da13722-732a-4b65-a894-fe4612f30d75"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.163169 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7da13722-732a-4b65-a894-fe4612f30d75-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "7da13722-732a-4b65-a894-fe4612f30d75" (UID: "7da13722-732a-4b65-a894-fe4612f30d75"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.174249 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.174502 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="4050d00b-0256-45c5-9dc4-0ab46956405d" containerName="nova-scheduler-scheduler" containerID="cri-o://8a59419fea27690bc54a6fc859fef55fbe86c693159ac3ff37906574c1369e56" gracePeriod=30 Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.202302 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-47e2-account-create-update-wvjr7"] Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.202794 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="e3c34454-315e-4821-ab25-b0f331a0d521" containerName="rabbitmq" containerID="cri-o://c5d00dd71aba57d230e51645212e11968b2acd8a912b082abd6373002fb907b4" gracePeriod=604800 Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.213613 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37978cc1-fcba-4032-a8b1-6632b61692ff-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "37978cc1-fcba-4032-a8b1-6632b61692ff" (UID: "37978cc1-fcba-4032-a8b1-6632b61692ff"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.220922 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37978cc1-fcba-4032-a8b1-6632b61692ff-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "37978cc1-fcba-4032-a8b1-6632b61692ff" (UID: "37978cc1-fcba-4032-a8b1-6632b61692ff"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.226327 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7da13722-732a-4b65-a894-fe4612f30d75-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7da13722-732a-4b65-a894-fe4612f30d75" (UID: "7da13722-732a-4b65-a894-fe4612f30d75"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.246207 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7da13722-732a-4b65-a894-fe4612f30d75-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.246242 4907 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/37978cc1-fcba-4032-a8b1-6632b61692ff-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.246251 4907 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7da13722-732a-4b65-a894-fe4612f30d75-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.246260 4907 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7da13722-732a-4b65-a894-fe4612f30d75-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.246268 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37978cc1-fcba-4032-a8b1-6632b61692ff-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.254639 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7da13722-732a-4b65-a894-fe4612f30d75-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "7da13722-732a-4b65-a894-fe4612f30d75" (UID: "7da13722-732a-4b65-a894-fe4612f30d75"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.266724 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2938e39-f1d8-4cdc-a32c-5d57b8f2034f-ovsdbserver-sb-tls-certs" (OuterVolumeSpecName: "ovsdbserver-sb-tls-certs") pod "c2938e39-f1d8-4cdc-a32c-5d57b8f2034f" (UID: "c2938e39-f1d8-4cdc-a32c-5d57b8f2034f"). InnerVolumeSpecName "ovsdbserver-sb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.268091 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/26827945-75f0-4867-ba04-31ff6428e06a-ovn-northd-tls-certs" (OuterVolumeSpecName: "ovn-northd-tls-certs") pod "26827945-75f0-4867-ba04-31ff6428e06a" (UID: "26827945-75f0-4867-ba04-31ff6428e06a"). InnerVolumeSpecName "ovn-northd-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.286093 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2938e39-f1d8-4cdc-a32c-5d57b8f2034f-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "c2938e39-f1d8-4cdc-a32c-5d57b8f2034f" (UID: "c2938e39-f1d8-4cdc-a32c-5d57b8f2034f"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:43 crc kubenswrapper[4907]: E0313 14:30:43.338254 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="77b52c0f64f55952d2255d5814b60083fd608735671f1f8507b09ea2e83db4d4" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Mar 13 14:30:43 crc kubenswrapper[4907]: E0313 14:30:43.340219 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="77b52c0f64f55952d2255d5814b60083fd608735671f1f8507b09ea2e83db4d4" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.348110 4907 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7da13722-732a-4b65-a894-fe4612f30d75-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.348145 4907 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c2938e39-f1d8-4cdc-a32c-5d57b8f2034f-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.348159 4907 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c2938e39-f1d8-4cdc-a32c-5d57b8f2034f-ovsdbserver-sb-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.348197 4907 reconciler_common.go:293] "Volume detached for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/26827945-75f0-4867-ba04-31ff6428e06a-ovn-northd-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:43 crc kubenswrapper[4907]: E0313 14:30:43.348283 4907 configmap.go:193] Couldn't get configMap openstack/openstack-cell1-scripts: configmap "openstack-cell1-scripts" not found Mar 13 14:30:43 crc kubenswrapper[4907]: E0313 14:30:43.348342 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/516d1499-3461-4af5-b426-9ae4a711a468-operator-scripts podName:516d1499-3461-4af5-b426-9ae4a711a468 nodeName:}" failed. No retries permitted until 2026-03-13 14:30:44.348325611 +0000 UTC m=+1543.248113300 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/516d1499-3461-4af5-b426-9ae4a711a468-operator-scripts") pod "root-account-create-update-glzbl" (UID: "516d1499-3461-4af5-b426-9ae4a711a468") : configmap "openstack-cell1-scripts" not found Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.357042 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-17a8-account-create-update-nj9pw"] Mar 13 14:30:43 crc kubenswrapper[4907]: E0313 14:30:43.358635 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="77b52c0f64f55952d2255d5814b60083fd608735671f1f8507b09ea2e83db4d4" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Mar 13 14:30:43 crc kubenswrapper[4907]: E0313 14:30:43.358705 4907 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell1-conductor-0" podUID="25c1c7e6-a149-4a8e-aead-3ebdc7e56851" containerName="nova-cell1-conductor-conductor" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.419851 4907 scope.go:117] "RemoveContainer" containerID="f294f1dfa23d177ca2e8f5511b6bd495bd3146c4baee6d3b95ccd81e1b4496fb" Mar 13 14:30:43 crc kubenswrapper[4907]: E0313 14:30:43.420565 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f294f1dfa23d177ca2e8f5511b6bd495bd3146c4baee6d3b95ccd81e1b4496fb\": container with ID starting with f294f1dfa23d177ca2e8f5511b6bd495bd3146c4baee6d3b95ccd81e1b4496fb not found: ID does not exist" containerID="f294f1dfa23d177ca2e8f5511b6bd495bd3146c4baee6d3b95ccd81e1b4496fb" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.420617 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f294f1dfa23d177ca2e8f5511b6bd495bd3146c4baee6d3b95ccd81e1b4496fb"} err="failed to get container status \"f294f1dfa23d177ca2e8f5511b6bd495bd3146c4baee6d3b95ccd81e1b4496fb\": rpc error: code = NotFound desc = could not find container \"f294f1dfa23d177ca2e8f5511b6bd495bd3146c4baee6d3b95ccd81e1b4496fb\": container with ID starting with f294f1dfa23d177ca2e8f5511b6bd495bd3146c4baee6d3b95ccd81e1b4496fb not found: ID does not exist" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.420650 4907 scope.go:117] "RemoveContainer" containerID="46f9fcec5ce705120f46ecb4deec4665885423d12b23cabce7f929fa5941d7ab" Mar 13 14:30:43 crc kubenswrapper[4907]: E0313 14:30:43.422089 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"46f9fcec5ce705120f46ecb4deec4665885423d12b23cabce7f929fa5941d7ab\": container with ID starting with 46f9fcec5ce705120f46ecb4deec4665885423d12b23cabce7f929fa5941d7ab not found: ID does not exist" containerID="46f9fcec5ce705120f46ecb4deec4665885423d12b23cabce7f929fa5941d7ab" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.422108 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"46f9fcec5ce705120f46ecb4deec4665885423d12b23cabce7f929fa5941d7ab"} err="failed to get container status \"46f9fcec5ce705120f46ecb4deec4665885423d12b23cabce7f929fa5941d7ab\": rpc error: code = NotFound desc = could not find container \"46f9fcec5ce705120f46ecb4deec4665885423d12b23cabce7f929fa5941d7ab\": container with ID starting with 46f9fcec5ce705120f46ecb4deec4665885423d12b23cabce7f929fa5941d7ab not found: ID does not exist" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.422124 4907 scope.go:117] "RemoveContainer" containerID="62edff4734da4ab22bc7673f1d2c0e9a9db082a3ca22b12fadc41486071fe66a" Mar 13 14:30:43 crc kubenswrapper[4907]: W0313 14:30:43.426227 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbc15f080_f90d_4e97_a866_d5edfd662bb7.slice/crio-bc3dc3b0fd90acb1ff4424b36cbf0bd4f5dd344f9b23a791c924f552a3241a63 WatchSource:0}: Error finding container bc3dc3b0fd90acb1ff4424b36cbf0bd4f5dd344f9b23a791c924f552a3241a63: Status 404 returned error can't find the container with id bc3dc3b0fd90acb1ff4424b36cbf0bd4f5dd344f9b23a791c924f552a3241a63 Mar 13 14:30:43 crc kubenswrapper[4907]: E0313 14:30:43.428678 4907 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 13 14:30:43 crc kubenswrapper[4907]: container &Container{Name:mariadb-account-create-update,Image:quay.io/podified-antelope-centos9/openstack-mariadb@sha256:4caef2b55e01b9a7ee88a22bc69db1893521a91d95c7ad4c8e593f14f17a5f95,Command:[/bin/sh -c #!/bin/bash Mar 13 14:30:43 crc kubenswrapper[4907]: Mar 13 14:30:43 crc kubenswrapper[4907]: MYSQL_REMOTE_HOST="" source /var/lib/operator-scripts/mysql_root_auth.sh Mar 13 14:30:43 crc kubenswrapper[4907]: Mar 13 14:30:43 crc kubenswrapper[4907]: export DatabasePassword=${DatabasePassword:?"Please specify a DatabasePassword variable."} Mar 13 14:30:43 crc kubenswrapper[4907]: Mar 13 14:30:43 crc kubenswrapper[4907]: MYSQL_CMD="mysql -h -u root -P 3306" Mar 13 14:30:43 crc kubenswrapper[4907]: Mar 13 14:30:43 crc kubenswrapper[4907]: if [ -n "glance" ]; then Mar 13 14:30:43 crc kubenswrapper[4907]: GRANT_DATABASE="glance" Mar 13 14:30:43 crc kubenswrapper[4907]: else Mar 13 14:30:43 crc kubenswrapper[4907]: GRANT_DATABASE="*" Mar 13 14:30:43 crc kubenswrapper[4907]: fi Mar 13 14:30:43 crc kubenswrapper[4907]: Mar 13 14:30:43 crc kubenswrapper[4907]: # going for maximum compatibility here: Mar 13 14:30:43 crc kubenswrapper[4907]: # 1. MySQL 8 no longer allows implicit create user when GRANT is used Mar 13 14:30:43 crc kubenswrapper[4907]: # 2. MariaDB has "CREATE OR REPLACE", but MySQL does not Mar 13 14:30:43 crc kubenswrapper[4907]: # 3. create user with CREATE but then do all password and TLS with ALTER to Mar 13 14:30:43 crc kubenswrapper[4907]: # support updates Mar 13 14:30:43 crc kubenswrapper[4907]: Mar 13 14:30:43 crc kubenswrapper[4907]: $MYSQL_CMD < logger="UnhandledError" Mar 13 14:30:43 crc kubenswrapper[4907]: E0313 14:30:43.438529 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mariadb-account-create-update\" with CreateContainerConfigError: \"secret \\\"glance-db-secret\\\" not found\"" pod="openstack/glance-17a8-account-create-update-nj9pw" podUID="bc15f080-f90d-4e97-a866-d5edfd662bb7" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.450499 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-nb-0"] Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.459374 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovsdbserver-nb-0"] Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.459982 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.476210 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-northd-0"] Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.483858 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-northd-0"] Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.486550 4907 scope.go:117] "RemoveContainer" containerID="890693bd8cb16d526cbd4e7c94b7a9b8dcd67a1bc48dbb881c81f86b4e9727d1" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.491738 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-7dfdbcdd5-dcb7n" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.511276 4907 scope.go:117] "RemoveContainer" containerID="9c33ada0f7c03cde714537a63faadeb180f6ac07dd15e3f4ce8c7087574733bb" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.572109 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-metrics-nhnkq"] Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.584571 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-metrics-nhnkq"] Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.591534 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-fdb8f6449-nb4km"] Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.601067 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-fdb8f6449-nb4km"] Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.671635 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1a1494c5-e2d6-4d29-b161-97b720742d8d-run-httpd\") pod \"1a1494c5-e2d6-4d29-b161-97b720742d8d\" (UID: \"1a1494c5-e2d6-4d29-b161-97b720742d8d\") " Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.671678 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7gbp6\" (UniqueName: \"kubernetes.io/projected/1a1494c5-e2d6-4d29-b161-97b720742d8d-kube-api-access-7gbp6\") pod \"1a1494c5-e2d6-4d29-b161-97b720742d8d\" (UID: \"1a1494c5-e2d6-4d29-b161-97b720742d8d\") " Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.671700 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/a055ce7f-0538-4e5e-938a-eaf6d0aa2dec-openstack-config\") pod \"a055ce7f-0538-4e5e-938a-eaf6d0aa2dec\" (UID: \"a055ce7f-0538-4e5e-938a-eaf6d0aa2dec\") " Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.671754 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/1a1494c5-e2d6-4d29-b161-97b720742d8d-etc-swift\") pod \"1a1494c5-e2d6-4d29-b161-97b720742d8d\" (UID: \"1a1494c5-e2d6-4d29-b161-97b720742d8d\") " Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.671804 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a055ce7f-0538-4e5e-938a-eaf6d0aa2dec-combined-ca-bundle\") pod \"a055ce7f-0538-4e5e-938a-eaf6d0aa2dec\" (UID: \"a055ce7f-0538-4e5e-938a-eaf6d0aa2dec\") " Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.671854 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1a1494c5-e2d6-4d29-b161-97b720742d8d-internal-tls-certs\") pod \"1a1494c5-e2d6-4d29-b161-97b720742d8d\" (UID: \"1a1494c5-e2d6-4d29-b161-97b720742d8d\") " Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.671916 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a1494c5-e2d6-4d29-b161-97b720742d8d-combined-ca-bundle\") pod \"1a1494c5-e2d6-4d29-b161-97b720742d8d\" (UID: \"1a1494c5-e2d6-4d29-b161-97b720742d8d\") " Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.671946 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/a055ce7f-0538-4e5e-938a-eaf6d0aa2dec-openstack-config-secret\") pod \"a055ce7f-0538-4e5e-938a-eaf6d0aa2dec\" (UID: \"a055ce7f-0538-4e5e-938a-eaf6d0aa2dec\") " Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.671972 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a1494c5-e2d6-4d29-b161-97b720742d8d-config-data\") pod \"1a1494c5-e2d6-4d29-b161-97b720742d8d\" (UID: \"1a1494c5-e2d6-4d29-b161-97b720742d8d\") " Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.672000 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fk82k\" (UniqueName: \"kubernetes.io/projected/a055ce7f-0538-4e5e-938a-eaf6d0aa2dec-kube-api-access-fk82k\") pod \"a055ce7f-0538-4e5e-938a-eaf6d0aa2dec\" (UID: \"a055ce7f-0538-4e5e-938a-eaf6d0aa2dec\") " Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.672061 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1a1494c5-e2d6-4d29-b161-97b720742d8d-log-httpd\") pod \"1a1494c5-e2d6-4d29-b161-97b720742d8d\" (UID: \"1a1494c5-e2d6-4d29-b161-97b720742d8d\") " Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.672083 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1a1494c5-e2d6-4d29-b161-97b720742d8d-public-tls-certs\") pod \"1a1494c5-e2d6-4d29-b161-97b720742d8d\" (UID: \"1a1494c5-e2d6-4d29-b161-97b720742d8d\") " Mar 13 14:30:43 crc kubenswrapper[4907]: E0313 14:30:43.672535 4907 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Mar 13 14:30:43 crc kubenswrapper[4907]: E0313 14:30:43.672591 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/e3c34454-315e-4821-ab25-b0f331a0d521-config-data podName:e3c34454-315e-4821-ab25-b0f331a0d521 nodeName:}" failed. No retries permitted until 2026-03-13 14:30:45.672576966 +0000 UTC m=+1544.572364655 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/e3c34454-315e-4821-ab25-b0f331a0d521-config-data") pod "rabbitmq-server-0" (UID: "e3c34454-315e-4821-ab25-b0f331a0d521") : configmap "rabbitmq-config-data" not found Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.675311 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1a1494c5-e2d6-4d29-b161-97b720742d8d-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "1a1494c5-e2d6-4d29-b161-97b720742d8d" (UID: "1a1494c5-e2d6-4d29-b161-97b720742d8d"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.676612 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a055ce7f-0538-4e5e-938a-eaf6d0aa2dec-kube-api-access-fk82k" (OuterVolumeSpecName: "kube-api-access-fk82k") pod "a055ce7f-0538-4e5e-938a-eaf6d0aa2dec" (UID: "a055ce7f-0538-4e5e-938a-eaf6d0aa2dec"). InnerVolumeSpecName "kube-api-access-fk82k". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.677385 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1a1494c5-e2d6-4d29-b161-97b720742d8d-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "1a1494c5-e2d6-4d29-b161-97b720742d8d" (UID: "1a1494c5-e2d6-4d29-b161-97b720742d8d"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.682180 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a1494c5-e2d6-4d29-b161-97b720742d8d-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "1a1494c5-e2d6-4d29-b161-97b720742d8d" (UID: "1a1494c5-e2d6-4d29-b161-97b720742d8d"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.682545 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a1494c5-e2d6-4d29-b161-97b720742d8d-kube-api-access-7gbp6" (OuterVolumeSpecName: "kube-api-access-7gbp6") pod "1a1494c5-e2d6-4d29-b161-97b720742d8d" (UID: "1a1494c5-e2d6-4d29-b161-97b720742d8d"). InnerVolumeSpecName "kube-api-access-7gbp6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.717023 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.728365 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a055ce7f-0538-4e5e-938a-eaf6d0aa2dec-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "a055ce7f-0538-4e5e-938a-eaf6d0aa2dec" (UID: "a055ce7f-0538-4e5e-938a-eaf6d0aa2dec"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.756635 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a055ce7f-0538-4e5e-938a-eaf6d0aa2dec-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a055ce7f-0538-4e5e-938a-eaf6d0aa2dec" (UID: "a055ce7f-0538-4e5e-938a-eaf6d0aa2dec"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:43 crc kubenswrapper[4907]: E0313 14:30:43.774389 4907 projected.go:288] Couldn't get configMap openstack/swift-storage-config-data: configmap "swift-storage-config-data" not found Mar 13 14:30:43 crc kubenswrapper[4907]: E0313 14:30:43.774418 4907 projected.go:263] Couldn't get secret openstack/swift-conf: secret "swift-conf" not found Mar 13 14:30:43 crc kubenswrapper[4907]: E0313 14:30:43.774465 4907 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Mar 13 14:30:43 crc kubenswrapper[4907]: E0313 14:30:43.774479 4907 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: [configmap "swift-storage-config-data" not found, secret "swift-conf" not found, configmap "swift-ring-files" not found] Mar 13 14:30:43 crc kubenswrapper[4907]: E0313 14:30:43.774533 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/14f05f31-8185-43b0-be69-bcf8d5388ea5-etc-swift podName:14f05f31-8185-43b0-be69-bcf8d5388ea5 nodeName:}" failed. No retries permitted until 2026-03-13 14:30:47.774510309 +0000 UTC m=+1546.674297998 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/14f05f31-8185-43b0-be69-bcf8d5388ea5-etc-swift") pod "swift-storage-0" (UID: "14f05f31-8185-43b0-be69-bcf8d5388ea5") : [configmap "swift-storage-config-data" not found, secret "swift-conf" not found, configmap "swift-ring-files" not found] Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.775008 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fk82k\" (UniqueName: \"kubernetes.io/projected/a055ce7f-0538-4e5e-938a-eaf6d0aa2dec-kube-api-access-fk82k\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.775027 4907 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1a1494c5-e2d6-4d29-b161-97b720742d8d-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.775038 4907 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1a1494c5-e2d6-4d29-b161-97b720742d8d-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.775048 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7gbp6\" (UniqueName: \"kubernetes.io/projected/1a1494c5-e2d6-4d29-b161-97b720742d8d-kube-api-access-7gbp6\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.775291 4907 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/a055ce7f-0538-4e5e-938a-eaf6d0aa2dec-openstack-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.775305 4907 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/1a1494c5-e2d6-4d29-b161-97b720742d8d-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.775313 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a055ce7f-0538-4e5e-938a-eaf6d0aa2dec-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.786667 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a1494c5-e2d6-4d29-b161-97b720742d8d-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "1a1494c5-e2d6-4d29-b161-97b720742d8d" (UID: "1a1494c5-e2d6-4d29-b161-97b720742d8d"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.790831 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a1494c5-e2d6-4d29-b161-97b720742d8d-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "1a1494c5-e2d6-4d29-b161-97b720742d8d" (UID: "1a1494c5-e2d6-4d29-b161-97b720742d8d"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.797262 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a1494c5-e2d6-4d29-b161-97b720742d8d-config-data" (OuterVolumeSpecName: "config-data") pod "1a1494c5-e2d6-4d29-b161-97b720742d8d" (UID: "1a1494c5-e2d6-4d29-b161-97b720742d8d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.798540 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a1494c5-e2d6-4d29-b161-97b720742d8d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1a1494c5-e2d6-4d29-b161-97b720742d8d" (UID: "1a1494c5-e2d6-4d29-b161-97b720742d8d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.802611 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="02262347-eaec-462e-a3ef-58c670885d99" path="/var/lib/kubelet/pods/02262347-eaec-462e-a3ef-58c670885d99/volumes" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.803438 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="056dd756-0d7b-471b-9929-f622d05ad606" path="/var/lib/kubelet/pods/056dd756-0d7b-471b-9929-f622d05ad606/volumes" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.804145 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1b701a42-f649-4657-9250-49188b10d3b9" path="/var/lib/kubelet/pods/1b701a42-f649-4657-9250-49188b10d3b9/volumes" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.805091 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="251ab76e-b9df-406f-9b07-c7fa6e227feb" path="/var/lib/kubelet/pods/251ab76e-b9df-406f-9b07-c7fa6e227feb/volumes" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.805644 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="26827945-75f0-4867-ba04-31ff6428e06a" path="/var/lib/kubelet/pods/26827945-75f0-4867-ba04-31ff6428e06a/volumes" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.806264 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2bab6ed6-5335-40cc-bdff-3d7461876fb8" path="/var/lib/kubelet/pods/2bab6ed6-5335-40cc-bdff-3d7461876fb8/volumes" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.807272 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="37978cc1-fcba-4032-a8b1-6632b61692ff" path="/var/lib/kubelet/pods/37978cc1-fcba-4032-a8b1-6632b61692ff/volumes" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.807846 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3f2b7163-b5d4-4575-a72c-e6ad1b3137df" path="/var/lib/kubelet/pods/3f2b7163-b5d4-4575-a72c-e6ad1b3137df/volumes" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.808463 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4d7d2639-4229-4286-be5c-d8b15ed91d17" path="/var/lib/kubelet/pods/4d7d2639-4229-4286-be5c-d8b15ed91d17/volumes" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.809365 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4d7ea7a2-ab25-4cb4-8749-10bc129cbd22" path="/var/lib/kubelet/pods/4d7ea7a2-ab25-4cb4-8749-10bc129cbd22/volumes" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.810613 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5dcaf7c6-b4dd-4312-b68b-a833a827a2c2" path="/var/lib/kubelet/pods/5dcaf7c6-b4dd-4312-b68b-a833a827a2c2/volumes" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.811186 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="65556420-dc7f-49c8-af59-abafebbb4674" path="/var/lib/kubelet/pods/65556420-dc7f-49c8-af59-abafebbb4674/volumes" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.811641 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7af00fde-8a92-4530-8fa8-9df7ff787c93" path="/var/lib/kubelet/pods/7af00fde-8a92-4530-8fa8-9df7ff787c93/volumes" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.813942 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7da13722-732a-4b65-a894-fe4612f30d75" path="/var/lib/kubelet/pods/7da13722-732a-4b65-a894-fe4612f30d75/volumes" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.814461 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8901c771-b45a-4147-9f96-cef784165281" path="/var/lib/kubelet/pods/8901c771-b45a-4147-9f96-cef784165281/volumes" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.815232 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9118baf0-df6a-41e5-98e5-3e4d8edfbaaa" path="/var/lib/kubelet/pods/9118baf0-df6a-41e5-98e5-3e4d8edfbaaa/volumes" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.816444 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="98748d4d-f934-49ed-91d9-7a9e467b2753" path="/var/lib/kubelet/pods/98748d4d-f934-49ed-91d9-7a9e467b2753/volumes" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.817005 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a89525da-db62-49c2-b548-bab4c974dfdd" path="/var/lib/kubelet/pods/a89525da-db62-49c2-b548-bab4c974dfdd/volumes" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.817517 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e1068e89-10df-4d4c-a213-d9e9643fab38" path="/var/lib/kubelet/pods/e1068e89-10df-4d4c-a213-d9e9643fab38/volumes" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.828808 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a055ce7f-0538-4e5e-938a-eaf6d0aa2dec-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "a055ce7f-0538-4e5e-938a-eaf6d0aa2dec" (UID: "a055ce7f-0538-4e5e-938a-eaf6d0aa2dec"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.871723 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-17a8-account-create-update-nj9pw" event={"ID":"bc15f080-f90d-4e97-a866-d5edfd662bb7","Type":"ContainerStarted","Data":"bc3dc3b0fd90acb1ff4424b36cbf0bd4f5dd344f9b23a791c924f552a3241a63"} Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.875601 4907 generic.go:334] "Generic (PLEG): container finished" podID="c1b70392-1240-40d9-8128-e7abe29c8398" containerID="6b0b25079cc099dabe922ac89fa41ae76d2f01797aac9f273b042e242459c574" exitCode=0 Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.875723 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-vp6f5" event={"ID":"c1b70392-1240-40d9-8128-e7abe29c8398","Type":"ContainerDied","Data":"6b0b25079cc099dabe922ac89fa41ae76d2f01797aac9f273b042e242459c574"} Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.876357 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3793f53a-ca36-4b27-8444-d6dbfd860424-etc-machine-id\") pod \"3793f53a-ca36-4b27-8444-d6dbfd860424\" (UID: \"3793f53a-ca36-4b27-8444-d6dbfd860424\") " Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.876422 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3793f53a-ca36-4b27-8444-d6dbfd860424-config-data-custom\") pod \"3793f53a-ca36-4b27-8444-d6dbfd860424\" (UID: \"3793f53a-ca36-4b27-8444-d6dbfd860424\") " Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.876456 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3793f53a-ca36-4b27-8444-d6dbfd860424-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "3793f53a-ca36-4b27-8444-d6dbfd860424" (UID: "3793f53a-ca36-4b27-8444-d6dbfd860424"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.876474 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3793f53a-ca36-4b27-8444-d6dbfd860424-config-data\") pod \"3793f53a-ca36-4b27-8444-d6dbfd860424\" (UID: \"3793f53a-ca36-4b27-8444-d6dbfd860424\") " Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.877025 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8gkl4\" (UniqueName: \"kubernetes.io/projected/3793f53a-ca36-4b27-8444-d6dbfd860424-kube-api-access-8gkl4\") pod \"3793f53a-ca36-4b27-8444-d6dbfd860424\" (UID: \"3793f53a-ca36-4b27-8444-d6dbfd860424\") " Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.877100 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3793f53a-ca36-4b27-8444-d6dbfd860424-combined-ca-bundle\") pod \"3793f53a-ca36-4b27-8444-d6dbfd860424\" (UID: \"3793f53a-ca36-4b27-8444-d6dbfd860424\") " Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.877144 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3793f53a-ca36-4b27-8444-d6dbfd860424-scripts\") pod \"3793f53a-ca36-4b27-8444-d6dbfd860424\" (UID: \"3793f53a-ca36-4b27-8444-d6dbfd860424\") " Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.877964 4907 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1a1494c5-e2d6-4d29-b161-97b720742d8d-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.878000 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a1494c5-e2d6-4d29-b161-97b720742d8d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.878010 4907 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/a055ce7f-0538-4e5e-938a-eaf6d0aa2dec-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.878021 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a1494c5-e2d6-4d29-b161-97b720742d8d-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.878030 4907 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3793f53a-ca36-4b27-8444-d6dbfd860424-etc-machine-id\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.878038 4907 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1a1494c5-e2d6-4d29-b161-97b720742d8d-public-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.882049 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3793f53a-ca36-4b27-8444-d6dbfd860424-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "3793f53a-ca36-4b27-8444-d6dbfd860424" (UID: "3793f53a-ca36-4b27-8444-d6dbfd860424"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.882290 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3793f53a-ca36-4b27-8444-d6dbfd860424-scripts" (OuterVolumeSpecName: "scripts") pod "3793f53a-ca36-4b27-8444-d6dbfd860424" (UID: "3793f53a-ca36-4b27-8444-d6dbfd860424"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.884471 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.884520 4907 scope.go:117] "RemoveContainer" containerID="3e056a670aa7307d9d211e2123e0cead9b6cd31175ff09e1feddb508c47f1a88" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.888227 4907 generic.go:334] "Generic (PLEG): container finished" podID="62983e04-f043-4ad3-9a46-96179d59671e" containerID="9e2de4df42a807d1995b769c3c69187ddaaf7da1c1d09354b07088fb415f36d7" exitCode=0 Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.888706 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"62983e04-f043-4ad3-9a46-96179d59671e","Type":"ContainerDied","Data":"9e2de4df42a807d1995b769c3c69187ddaaf7da1c1d09354b07088fb415f36d7"} Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.900369 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3793f53a-ca36-4b27-8444-d6dbfd860424-kube-api-access-8gkl4" (OuterVolumeSpecName: "kube-api-access-8gkl4") pod "3793f53a-ca36-4b27-8444-d6dbfd860424" (UID: "3793f53a-ca36-4b27-8444-d6dbfd860424"). InnerVolumeSpecName "kube-api-access-8gkl4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.900743 4907 generic.go:334] "Generic (PLEG): container finished" podID="2e41272f-6e7b-47a7-825e-3571f6a1fd07" containerID="c5ecad10cc9377aeda50af0d8c9213440ce021149e77309245ab84eb57179e9d" exitCode=143 Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.900916 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2e41272f-6e7b-47a7-825e-3571f6a1fd07","Type":"ContainerDied","Data":"c5ecad10cc9377aeda50af0d8c9213440ce021149e77309245ab84eb57179e9d"} Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.923126 4907 generic.go:334] "Generic (PLEG): container finished" podID="516d1499-3461-4af5-b426-9ae4a711a468" containerID="73f68d1aced419107233520623fb182eb32be93775f62b9a147d1f6ec4671d78" exitCode=1 Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.923187 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-glzbl" event={"ID":"516d1499-3461-4af5-b426-9ae4a711a468","Type":"ContainerDied","Data":"73f68d1aced419107233520623fb182eb32be93775f62b9a147d1f6ec4671d78"} Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.923216 4907 scope.go:117] "RemoveContainer" containerID="783ee4ed5f573bc2e551fccdb256e8146b5d15e716dc18f3c1e347aa5d0fa4c4" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.942741 4907 generic.go:334] "Generic (PLEG): container finished" podID="25b5fafe-65c8-48ba-bc34-83442f8ace4c" containerID="b5356b2175cd794b4fc6692eb61775ae05e66fb791f7dbd913a5d208e5880c25" exitCode=143 Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.943000 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-76bbcdf7d9-j6hw5" event={"ID":"25b5fafe-65c8-48ba-bc34-83442f8ace4c","Type":"ContainerDied","Data":"b5356b2175cd794b4fc6692eb61775ae05e66fb791f7dbd913a5d208e5880c25"} Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.962664 4907 generic.go:334] "Generic (PLEG): container finished" podID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerID="1e5ba68b21b276693f03089f51ca3955dd88adcaacae52727d03a999b543d411" exitCode=0 Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.963007 4907 generic.go:334] "Generic (PLEG): container finished" podID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerID="c81c73db839bc006e1b7fdce225dd05a9a50870ea60ca389ea146e1e69d195b7" exitCode=0 Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.964738 4907 generic.go:334] "Generic (PLEG): container finished" podID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerID="b7f60cadabc1fd2d193073423cce7296e305ff6325001c7ec9823dd267a1787a" exitCode=0 Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.964864 4907 generic.go:334] "Generic (PLEG): container finished" podID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerID="8168297a51ad087d7375035164e54e4a0d24e6fd56cf85e922e11d80f7f3d091" exitCode=0 Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.963138 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"14f05f31-8185-43b0-be69-bcf8d5388ea5","Type":"ContainerDied","Data":"1e5ba68b21b276693f03089f51ca3955dd88adcaacae52727d03a999b543d411"} Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.965074 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"14f05f31-8185-43b0-be69-bcf8d5388ea5","Type":"ContainerDied","Data":"c81c73db839bc006e1b7fdce225dd05a9a50870ea60ca389ea146e1e69d195b7"} Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.965110 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"14f05f31-8185-43b0-be69-bcf8d5388ea5","Type":"ContainerDied","Data":"b7f60cadabc1fd2d193073423cce7296e305ff6325001c7ec9823dd267a1787a"} Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.965123 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"14f05f31-8185-43b0-be69-bcf8d5388ea5","Type":"ContainerDied","Data":"8168297a51ad087d7375035164e54e4a0d24e6fd56cf85e922e11d80f7f3d091"} Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.980105 4907 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3793f53a-ca36-4b27-8444-d6dbfd860424-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.980446 4907 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3793f53a-ca36-4b27-8444-d6dbfd860424-config-data-custom\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.980456 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8gkl4\" (UniqueName: \"kubernetes.io/projected/3793f53a-ca36-4b27-8444-d6dbfd860424-kube-api-access-8gkl4\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.984544 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3793f53a-ca36-4b27-8444-d6dbfd860424-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3793f53a-ca36-4b27-8444-d6dbfd860424" (UID: "3793f53a-ca36-4b27-8444-d6dbfd860424"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.998272 4907 generic.go:334] "Generic (PLEG): container finished" podID="1a1494c5-e2d6-4d29-b161-97b720742d8d" containerID="c6c4a6a54081de702756857f0341351ea26c5f9aaa725027df2c989c6b1a6e18" exitCode=0 Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.998637 4907 generic.go:334] "Generic (PLEG): container finished" podID="1a1494c5-e2d6-4d29-b161-97b720742d8d" containerID="676efd80769fe771754cb647e65098f243a686b0c64467414fa0d6ccfbece2d4" exitCode=0 Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.998594 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-7dfdbcdd5-dcb7n" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.998618 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-7dfdbcdd5-dcb7n" event={"ID":"1a1494c5-e2d6-4d29-b161-97b720742d8d","Type":"ContainerDied","Data":"c6c4a6a54081de702756857f0341351ea26c5f9aaa725027df2c989c6b1a6e18"} Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.999801 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-7dfdbcdd5-dcb7n" event={"ID":"1a1494c5-e2d6-4d29-b161-97b720742d8d","Type":"ContainerDied","Data":"676efd80769fe771754cb647e65098f243a686b0c64467414fa0d6ccfbece2d4"} Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.999833 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-7dfdbcdd5-dcb7n" event={"ID":"1a1494c5-e2d6-4d29-b161-97b720742d8d","Type":"ContainerDied","Data":"3472978398de27f6988019a31485f47afeca847038681d998902a9798bd82c2a"} Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.016034 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3793f53a-ca36-4b27-8444-d6dbfd860424-config-data" (OuterVolumeSpecName: "config-data") pod "3793f53a-ca36-4b27-8444-d6dbfd860424" (UID: "3793f53a-ca36-4b27-8444-d6dbfd860424"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.038242 4907 generic.go:334] "Generic (PLEG): container finished" podID="3793f53a-ca36-4b27-8444-d6dbfd860424" containerID="94bc55f3afe5f98100548e7593f16440afc25ced9f4af83fb6cd9f896253e749" exitCode=0 Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.038277 4907 generic.go:334] "Generic (PLEG): container finished" podID="3793f53a-ca36-4b27-8444-d6dbfd860424" containerID="d484e26e1510dc0830d22229bd3bbd9bfa0e7d5cab421e8e5b87043a8df1abeb" exitCode=0 Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.038327 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"3793f53a-ca36-4b27-8444-d6dbfd860424","Type":"ContainerDied","Data":"94bc55f3afe5f98100548e7593f16440afc25ced9f4af83fb6cd9f896253e749"} Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.038357 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"3793f53a-ca36-4b27-8444-d6dbfd860424","Type":"ContainerDied","Data":"d484e26e1510dc0830d22229bd3bbd9bfa0e7d5cab421e8e5b87043a8df1abeb"} Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.038368 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"3793f53a-ca36-4b27-8444-d6dbfd860424","Type":"ContainerDied","Data":"8158569f0778590e4a7e4122294d9844f9402dad6a901826d7dd5af6e7fab659"} Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.038433 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.047074 4907 generic.go:334] "Generic (PLEG): container finished" podID="cc227d33-3bf1-4b5f-8368-8e6b8a512f7d" containerID="6c1fdac6667e489d30e980acdd57d1cf0fa750bc21c6f46f11adb992a74fe2bd" exitCode=143 Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.047180 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"cc227d33-3bf1-4b5f-8368-8e6b8a512f7d","Type":"ContainerDied","Data":"6c1fdac6667e489d30e980acdd57d1cf0fa750bc21c6f46f11adb992a74fe2bd"} Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.048706 4907 generic.go:334] "Generic (PLEG): container finished" podID="1b2f62d4-0584-4cc3-81ed-e067d8db23b1" containerID="f21714a9f3b46d6b404fb8600bc2b71af7d623654fb77b4f61140d6330569927" exitCode=143 Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.048768 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7fc86597fb-mt9pk" event={"ID":"1b2f62d4-0584-4cc3-81ed-e067d8db23b1","Type":"ContainerDied","Data":"f21714a9f3b46d6b404fb8600bc2b71af7d623654fb77b4f61140d6330569927"} Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.049902 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-47e2-account-create-update-wvjr7" event={"ID":"82dec019-ba60-43f5-845b-322a6bcca225","Type":"ContainerStarted","Data":"cbbe85502bda2ac6ef5bc690e92165a5b92af3bcd3162f655a1bd43c95395317"} Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.056243 4907 generic.go:334] "Generic (PLEG): container finished" podID="001078cc-c01e-4173-9740-4f9ff082c593" containerID="3e3eb6f5fca9c3693edb3599f1d3dd3d706131518c37c580379168a8a1b7e5e0" exitCode=0 Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.056337 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7bf587b8d9-pvvbz" event={"ID":"001078cc-c01e-4173-9740-4f9ff082c593","Type":"ContainerDied","Data":"3e3eb6f5fca9c3693edb3599f1d3dd3d706131518c37c580379168a8a1b7e5e0"} Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.059772 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_c2938e39-f1d8-4cdc-a32c-5d57b8f2034f/ovsdbserver-sb/0.log" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.059875 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"c2938e39-f1d8-4cdc-a32c-5d57b8f2034f","Type":"ContainerDied","Data":"c836a111eed36848627fde431a8ac117cdf68714d43b9b328bad3f2315e96e0e"} Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.059990 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Mar 13 14:30:44 crc kubenswrapper[4907]: E0313 14:30:44.067406 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5ab7cf17ecb1804f362340163cfceb48e6f2c8900108799d01d1025c485c1e8f" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.072123 4907 generic.go:334] "Generic (PLEG): container finished" podID="65531461-cea2-4b2c-a9b6-8fd9e9bffb27" containerID="3ccd8eb5dd33f0cece989e6411df4852cb55e2cb0b138f16332e7a07ccf5c865" exitCode=143 Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.072278 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7d7bc9f7d-fks76" event={"ID":"65531461-cea2-4b2c-a9b6-8fd9e9bffb27","Type":"ContainerDied","Data":"3ccd8eb5dd33f0cece989e6411df4852cb55e2cb0b138f16332e7a07ccf5c865"} Mar 13 14:30:44 crc kubenswrapper[4907]: E0313 14:30:44.072628 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5ab7cf17ecb1804f362340163cfceb48e6f2c8900108799d01d1025c485c1e8f" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Mar 13 14:30:44 crc kubenswrapper[4907]: E0313 14:30:44.079706 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5ab7cf17ecb1804f362340163cfceb48e6f2c8900108799d01d1025c485c1e8f" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Mar 13 14:30:44 crc kubenswrapper[4907]: E0313 14:30:44.079761 4907 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell0-conductor-0" podUID="364a702e-e6c6-42bc-8cee-4ba19ea2f1ec" containerName="nova-cell0-conductor-conductor" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.081817 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3793f53a-ca36-4b27-8444-d6dbfd860424-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.081917 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3793f53a-ca36-4b27-8444-d6dbfd860424-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.103406 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.109099 4907 scope.go:117] "RemoveContainer" containerID="c6c4a6a54081de702756857f0341351ea26c5f9aaa725027df2c989c6b1a6e18" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.145333 4907 scope.go:117] "RemoveContainer" containerID="676efd80769fe771754cb647e65098f243a686b0c64467414fa0d6ccfbece2d4" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.181700 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-proxy-7dfdbcdd5-dcb7n"] Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.187957 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-proxy-7dfdbcdd5-dcb7n"] Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.202933 4907 scope.go:117] "RemoveContainer" containerID="c6c4a6a54081de702756857f0341351ea26c5f9aaa725027df2c989c6b1a6e18" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.204807 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Mar 13 14:30:44 crc kubenswrapper[4907]: E0313 14:30:44.208347 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c6c4a6a54081de702756857f0341351ea26c5f9aaa725027df2c989c6b1a6e18\": container with ID starting with c6c4a6a54081de702756857f0341351ea26c5f9aaa725027df2c989c6b1a6e18 not found: ID does not exist" containerID="c6c4a6a54081de702756857f0341351ea26c5f9aaa725027df2c989c6b1a6e18" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.208380 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c6c4a6a54081de702756857f0341351ea26c5f9aaa725027df2c989c6b1a6e18"} err="failed to get container status \"c6c4a6a54081de702756857f0341351ea26c5f9aaa725027df2c989c6b1a6e18\": rpc error: code = NotFound desc = could not find container \"c6c4a6a54081de702756857f0341351ea26c5f9aaa725027df2c989c6b1a6e18\": container with ID starting with c6c4a6a54081de702756857f0341351ea26c5f9aaa725027df2c989c6b1a6e18 not found: ID does not exist" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.208413 4907 scope.go:117] "RemoveContainer" containerID="676efd80769fe771754cb647e65098f243a686b0c64467414fa0d6ccfbece2d4" Mar 13 14:30:44 crc kubenswrapper[4907]: E0313 14:30:44.209073 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"676efd80769fe771754cb647e65098f243a686b0c64467414fa0d6ccfbece2d4\": container with ID starting with 676efd80769fe771754cb647e65098f243a686b0c64467414fa0d6ccfbece2d4 not found: ID does not exist" containerID="676efd80769fe771754cb647e65098f243a686b0c64467414fa0d6ccfbece2d4" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.209358 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"676efd80769fe771754cb647e65098f243a686b0c64467414fa0d6ccfbece2d4"} err="failed to get container status \"676efd80769fe771754cb647e65098f243a686b0c64467414fa0d6ccfbece2d4\": rpc error: code = NotFound desc = could not find container \"676efd80769fe771754cb647e65098f243a686b0c64467414fa0d6ccfbece2d4\": container with ID starting with 676efd80769fe771754cb647e65098f243a686b0c64467414fa0d6ccfbece2d4 not found: ID does not exist" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.209379 4907 scope.go:117] "RemoveContainer" containerID="c6c4a6a54081de702756857f0341351ea26c5f9aaa725027df2c989c6b1a6e18" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.209585 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c6c4a6a54081de702756857f0341351ea26c5f9aaa725027df2c989c6b1a6e18"} err="failed to get container status \"c6c4a6a54081de702756857f0341351ea26c5f9aaa725027df2c989c6b1a6e18\": rpc error: code = NotFound desc = could not find container \"c6c4a6a54081de702756857f0341351ea26c5f9aaa725027df2c989c6b1a6e18\": container with ID starting with c6c4a6a54081de702756857f0341351ea26c5f9aaa725027df2c989c6b1a6e18 not found: ID does not exist" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.209603 4907 scope.go:117] "RemoveContainer" containerID="676efd80769fe771754cb647e65098f243a686b0c64467414fa0d6ccfbece2d4" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.212349 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"676efd80769fe771754cb647e65098f243a686b0c64467414fa0d6ccfbece2d4"} err="failed to get container status \"676efd80769fe771754cb647e65098f243a686b0c64467414fa0d6ccfbece2d4\": rpc error: code = NotFound desc = could not find container \"676efd80769fe771754cb647e65098f243a686b0c64467414fa0d6ccfbece2d4\": container with ID starting with 676efd80769fe771754cb647e65098f243a686b0c64467414fa0d6ccfbece2d4 not found: ID does not exist" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.212394 4907 scope.go:117] "RemoveContainer" containerID="94bc55f3afe5f98100548e7593f16440afc25ced9f4af83fb6cd9f896253e749" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.213124 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.221908 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-sb-0"] Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.230022 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovsdbserver-sb-0"] Mar 13 14:30:44 crc kubenswrapper[4907]: E0313 14:30:44.231307 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="8a59419fea27690bc54a6fc859fef55fbe86c693159ac3ff37906574c1369e56" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Mar 13 14:30:44 crc kubenswrapper[4907]: E0313 14:30:44.232506 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="8a59419fea27690bc54a6fc859fef55fbe86c693159ac3ff37906574c1369e56" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Mar 13 14:30:44 crc kubenswrapper[4907]: E0313 14:30:44.233473 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="8a59419fea27690bc54a6fc859fef55fbe86c693159ac3ff37906574c1369e56" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Mar 13 14:30:44 crc kubenswrapper[4907]: E0313 14:30:44.233500 4907 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="4050d00b-0256-45c5-9dc4-0ab46956405d" containerName="nova-scheduler-scheduler" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.234676 4907 scope.go:117] "RemoveContainer" containerID="d484e26e1510dc0830d22229bd3bbd9bfa0e7d5cab421e8e5b87043a8df1abeb" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.262182 4907 scope.go:117] "RemoveContainer" containerID="94bc55f3afe5f98100548e7593f16440afc25ced9f4af83fb6cd9f896253e749" Mar 13 14:30:44 crc kubenswrapper[4907]: E0313 14:30:44.272630 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"94bc55f3afe5f98100548e7593f16440afc25ced9f4af83fb6cd9f896253e749\": container with ID starting with 94bc55f3afe5f98100548e7593f16440afc25ced9f4af83fb6cd9f896253e749 not found: ID does not exist" containerID="94bc55f3afe5f98100548e7593f16440afc25ced9f4af83fb6cd9f896253e749" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.272670 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"94bc55f3afe5f98100548e7593f16440afc25ced9f4af83fb6cd9f896253e749"} err="failed to get container status \"94bc55f3afe5f98100548e7593f16440afc25ced9f4af83fb6cd9f896253e749\": rpc error: code = NotFound desc = could not find container \"94bc55f3afe5f98100548e7593f16440afc25ced9f4af83fb6cd9f896253e749\": container with ID starting with 94bc55f3afe5f98100548e7593f16440afc25ced9f4af83fb6cd9f896253e749 not found: ID does not exist" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.272700 4907 scope.go:117] "RemoveContainer" containerID="d484e26e1510dc0830d22229bd3bbd9bfa0e7d5cab421e8e5b87043a8df1abeb" Mar 13 14:30:44 crc kubenswrapper[4907]: E0313 14:30:44.280145 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d484e26e1510dc0830d22229bd3bbd9bfa0e7d5cab421e8e5b87043a8df1abeb\": container with ID starting with d484e26e1510dc0830d22229bd3bbd9bfa0e7d5cab421e8e5b87043a8df1abeb not found: ID does not exist" containerID="d484e26e1510dc0830d22229bd3bbd9bfa0e7d5cab421e8e5b87043a8df1abeb" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.280192 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d484e26e1510dc0830d22229bd3bbd9bfa0e7d5cab421e8e5b87043a8df1abeb"} err="failed to get container status \"d484e26e1510dc0830d22229bd3bbd9bfa0e7d5cab421e8e5b87043a8df1abeb\": rpc error: code = NotFound desc = could not find container \"d484e26e1510dc0830d22229bd3bbd9bfa0e7d5cab421e8e5b87043a8df1abeb\": container with ID starting with d484e26e1510dc0830d22229bd3bbd9bfa0e7d5cab421e8e5b87043a8df1abeb not found: ID does not exist" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.280219 4907 scope.go:117] "RemoveContainer" containerID="94bc55f3afe5f98100548e7593f16440afc25ced9f4af83fb6cd9f896253e749" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.282761 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"94bc55f3afe5f98100548e7593f16440afc25ced9f4af83fb6cd9f896253e749"} err="failed to get container status \"94bc55f3afe5f98100548e7593f16440afc25ced9f4af83fb6cd9f896253e749\": rpc error: code = NotFound desc = could not find container \"94bc55f3afe5f98100548e7593f16440afc25ced9f4af83fb6cd9f896253e749\": container with ID starting with 94bc55f3afe5f98100548e7593f16440afc25ced9f4af83fb6cd9f896253e749 not found: ID does not exist" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.282807 4907 scope.go:117] "RemoveContainer" containerID="d484e26e1510dc0830d22229bd3bbd9bfa0e7d5cab421e8e5b87043a8df1abeb" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.285389 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.285672 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="aa81fa2a-4c27-4e02-9891-5b39348feca9" containerName="ceilometer-central-agent" containerID="cri-o://285cdb1df855cbeeda75e9d50bffbe69e5c3f9a5485dd118f47a932fb1d36f19" gracePeriod=30 Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.286135 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="aa81fa2a-4c27-4e02-9891-5b39348feca9" containerName="sg-core" containerID="cri-o://16ef32bcd58d9ca78315416c39fe9eca0c5eee8682ac66cea46ec8f9242f8ee5" gracePeriod=30 Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.286190 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="aa81fa2a-4c27-4e02-9891-5b39348feca9" containerName="ceilometer-notification-agent" containerID="cri-o://78891ea90d5eb95d642931ec25514d47603b8811d971d7d33d8fdc5eb0d3bb04" gracePeriod=30 Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.286188 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="aa81fa2a-4c27-4e02-9891-5b39348feca9" containerName="proxy-httpd" containerID="cri-o://79914c8007c9b10b224f5d4d1fcd270f17078ab7c002b89d3dab089884af5d6c" gracePeriod=30 Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.287349 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d484e26e1510dc0830d22229bd3bbd9bfa0e7d5cab421e8e5b87043a8df1abeb"} err="failed to get container status \"d484e26e1510dc0830d22229bd3bbd9bfa0e7d5cab421e8e5b87043a8df1abeb\": rpc error: code = NotFound desc = could not find container \"d484e26e1510dc0830d22229bd3bbd9bfa0e7d5cab421e8e5b87043a8df1abeb\": container with ID starting with d484e26e1510dc0830d22229bd3bbd9bfa0e7d5cab421e8e5b87043a8df1abeb not found: ID does not exist" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.287403 4907 scope.go:117] "RemoveContainer" containerID="1ff45617937d399a62e2b1c7ef156ccf7052d51f57fba51d267daa3e49810ce6" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.295272 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62983e04-f043-4ad3-9a46-96179d59671e-config-data\") pod \"62983e04-f043-4ad3-9a46-96179d59671e\" (UID: \"62983e04-f043-4ad3-9a46-96179d59671e\") " Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.295354 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/62983e04-f043-4ad3-9a46-96179d59671e-nova-novncproxy-tls-certs\") pod \"62983e04-f043-4ad3-9a46-96179d59671e\" (UID: \"62983e04-f043-4ad3-9a46-96179d59671e\") " Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.295382 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62983e04-f043-4ad3-9a46-96179d59671e-combined-ca-bundle\") pod \"62983e04-f043-4ad3-9a46-96179d59671e\" (UID: \"62983e04-f043-4ad3-9a46-96179d59671e\") " Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.295432 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/62983e04-f043-4ad3-9a46-96179d59671e-vencrypt-tls-certs\") pod \"62983e04-f043-4ad3-9a46-96179d59671e\" (UID: \"62983e04-f043-4ad3-9a46-96179d59671e\") " Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.295488 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6kqzq\" (UniqueName: \"kubernetes.io/projected/62983e04-f043-4ad3-9a46-96179d59671e-kube-api-access-6kqzq\") pod \"62983e04-f043-4ad3-9a46-96179d59671e\" (UID: \"62983e04-f043-4ad3-9a46-96179d59671e\") " Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.319616 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/62983e04-f043-4ad3-9a46-96179d59671e-kube-api-access-6kqzq" (OuterVolumeSpecName: "kube-api-access-6kqzq") pod "62983e04-f043-4ad3-9a46-96179d59671e" (UID: "62983e04-f043-4ad3-9a46-96179d59671e"). InnerVolumeSpecName "kube-api-access-6kqzq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.339463 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62983e04-f043-4ad3-9a46-96179d59671e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "62983e04-f043-4ad3-9a46-96179d59671e" (UID: "62983e04-f043-4ad3-9a46-96179d59671e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.383822 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.384131 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="c49286b1-2e9f-4de6-b7de-2e952f3ad607" containerName="kube-state-metrics" containerID="cri-o://0d242394ff0714a798d1ef385dde5709185206b531dfb7b51abcf0db5d059589" gracePeriod=30 Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.410113 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6kqzq\" (UniqueName: \"kubernetes.io/projected/62983e04-f043-4ad3-9a46-96179d59671e-kube-api-access-6kqzq\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.410168 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62983e04-f043-4ad3-9a46-96179d59671e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:44 crc kubenswrapper[4907]: E0313 14:30:44.410271 4907 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Mar 13 14:30:44 crc kubenswrapper[4907]: E0313 14:30:44.410333 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/286c82e8-b74a-49d7-a355-ac074aace10e-config-data podName:286c82e8-b74a-49d7-a355-ac074aace10e nodeName:}" failed. No retries permitted until 2026-03-13 14:30:48.410318135 +0000 UTC m=+1547.310105824 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/286c82e8-b74a-49d7-a355-ac074aace10e-config-data") pod "rabbitmq-cell1-server-0" (UID: "286c82e8-b74a-49d7-a355-ac074aace10e") : configmap "rabbitmq-cell1-config-data" not found Mar 13 14:30:44 crc kubenswrapper[4907]: E0313 14:30:44.410778 4907 configmap.go:193] Couldn't get configMap openstack/openstack-cell1-scripts: configmap "openstack-cell1-scripts" not found Mar 13 14:30:44 crc kubenswrapper[4907]: E0313 14:30:44.410818 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/516d1499-3461-4af5-b426-9ae4a711a468-operator-scripts podName:516d1499-3461-4af5-b426-9ae4a711a468 nodeName:}" failed. No retries permitted until 2026-03-13 14:30:46.410805297 +0000 UTC m=+1545.310592986 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/516d1499-3461-4af5-b426-9ae4a711a468-operator-scripts") pod "root-account-create-update-glzbl" (UID: "516d1499-3461-4af5-b426-9ae4a711a468") : configmap "openstack-cell1-scripts" not found Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.426754 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-17a8-account-create-update-nj9pw" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.440868 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62983e04-f043-4ad3-9a46-96179d59671e-vencrypt-tls-certs" (OuterVolumeSpecName: "vencrypt-tls-certs") pod "62983e04-f043-4ad3-9a46-96179d59671e" (UID: "62983e04-f043-4ad3-9a46-96179d59671e"). InnerVolumeSpecName "vencrypt-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.441579 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62983e04-f043-4ad3-9a46-96179d59671e-nova-novncproxy-tls-certs" (OuterVolumeSpecName: "nova-novncproxy-tls-certs") pod "62983e04-f043-4ad3-9a46-96179d59671e" (UID: "62983e04-f043-4ad3-9a46-96179d59671e"). InnerVolumeSpecName "nova-novncproxy-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.463211 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62983e04-f043-4ad3-9a46-96179d59671e-config-data" (OuterVolumeSpecName: "config-data") pod "62983e04-f043-4ad3-9a46-96179d59671e" (UID: "62983e04-f043-4ad3-9a46-96179d59671e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.468553 4907 scope.go:117] "RemoveContainer" containerID="432aefb6d80f0522ac7d6d516626fd7ea34147a98d43c3ddba905f2a1bb074ad" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.475053 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/memcached-0"] Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.475299 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/memcached-0" podUID="cacb281b-ee3b-42b6-8754-d07e6ea6aa4b" containerName="memcached" containerID="cri-o://65475a4b4006963b7fec1fa34278a44b7bf08cbcf9c088a764d7d29c870c630c" gracePeriod=30 Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.495962 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-04a8-account-create-update-ws5ht"] Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.511737 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bc15f080-f90d-4e97-a866-d5edfd662bb7-operator-scripts\") pod \"bc15f080-f90d-4e97-a866-d5edfd662bb7\" (UID: \"bc15f080-f90d-4e97-a866-d5edfd662bb7\") " Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.511900 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jstl5\" (UniqueName: \"kubernetes.io/projected/bc15f080-f90d-4e97-a866-d5edfd662bb7-kube-api-access-jstl5\") pod \"bc15f080-f90d-4e97-a866-d5edfd662bb7\" (UID: \"bc15f080-f90d-4e97-a866-d5edfd662bb7\") " Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.512412 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62983e04-f043-4ad3-9a46-96179d59671e-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.512427 4907 reconciler_common.go:293] "Volume detached for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/62983e04-f043-4ad3-9a46-96179d59671e-nova-novncproxy-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.512438 4907 reconciler_common.go:293] "Volume detached for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/62983e04-f043-4ad3-9a46-96179d59671e-vencrypt-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.517056 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bc15f080-f90d-4e97-a866-d5edfd662bb7-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "bc15f080-f90d-4e97-a866-d5edfd662bb7" (UID: "bc15f080-f90d-4e97-a866-d5edfd662bb7"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.519374 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-04a8-account-create-update-ws5ht"] Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.528456 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc15f080-f90d-4e97-a866-d5edfd662bb7-kube-api-access-jstl5" (OuterVolumeSpecName: "kube-api-access-jstl5") pod "bc15f080-f90d-4e97-a866-d5edfd662bb7" (UID: "bc15f080-f90d-4e97-a866-d5edfd662bb7"). InnerVolumeSpecName "kube-api-access-jstl5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.528514 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-04a8-account-create-update-kntb5"] Mar 13 14:30:44 crc kubenswrapper[4907]: E0313 14:30:44.528870 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2938e39-f1d8-4cdc-a32c-5d57b8f2034f" containerName="openstack-network-exporter" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.528898 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2938e39-f1d8-4cdc-a32c-5d57b8f2034f" containerName="openstack-network-exporter" Mar 13 14:30:44 crc kubenswrapper[4907]: E0313 14:30:44.528911 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3793f53a-ca36-4b27-8444-d6dbfd860424" containerName="probe" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.528917 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="3793f53a-ca36-4b27-8444-d6dbfd860424" containerName="probe" Mar 13 14:30:44 crc kubenswrapper[4907]: E0313 14:30:44.528935 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a1494c5-e2d6-4d29-b161-97b720742d8d" containerName="proxy-httpd" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.528941 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a1494c5-e2d6-4d29-b161-97b720742d8d" containerName="proxy-httpd" Mar 13 14:30:44 crc kubenswrapper[4907]: E0313 14:30:44.528953 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62983e04-f043-4ad3-9a46-96179d59671e" containerName="nova-cell1-novncproxy-novncproxy" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.528959 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="62983e04-f043-4ad3-9a46-96179d59671e" containerName="nova-cell1-novncproxy-novncproxy" Mar 13 14:30:44 crc kubenswrapper[4907]: E0313 14:30:44.528968 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26827945-75f0-4867-ba04-31ff6428e06a" containerName="openstack-network-exporter" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.528974 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="26827945-75f0-4867-ba04-31ff6428e06a" containerName="openstack-network-exporter" Mar 13 14:30:44 crc kubenswrapper[4907]: E0313 14:30:44.528981 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3793f53a-ca36-4b27-8444-d6dbfd860424" containerName="cinder-scheduler" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.528987 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="3793f53a-ca36-4b27-8444-d6dbfd860424" containerName="cinder-scheduler" Mar 13 14:30:44 crc kubenswrapper[4907]: E0313 14:30:44.528999 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="056dd756-0d7b-471b-9929-f622d05ad606" containerName="openstack-network-exporter" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.529005 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="056dd756-0d7b-471b-9929-f622d05ad606" containerName="openstack-network-exporter" Mar 13 14:30:44 crc kubenswrapper[4907]: E0313 14:30:44.529012 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a1494c5-e2d6-4d29-b161-97b720742d8d" containerName="proxy-server" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.529017 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a1494c5-e2d6-4d29-b161-97b720742d8d" containerName="proxy-server" Mar 13 14:30:44 crc kubenswrapper[4907]: E0313 14:30:44.529027 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26827945-75f0-4867-ba04-31ff6428e06a" containerName="ovn-northd" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.529033 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="26827945-75f0-4867-ba04-31ff6428e06a" containerName="ovn-northd" Mar 13 14:30:44 crc kubenswrapper[4907]: E0313 14:30:44.529043 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37978cc1-fcba-4032-a8b1-6632b61692ff" containerName="openstack-network-exporter" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.529049 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="37978cc1-fcba-4032-a8b1-6632b61692ff" containerName="openstack-network-exporter" Mar 13 14:30:44 crc kubenswrapper[4907]: E0313 14:30:44.529056 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7da13722-732a-4b65-a894-fe4612f30d75" containerName="init" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.529061 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="7da13722-732a-4b65-a894-fe4612f30d75" containerName="init" Mar 13 14:30:44 crc kubenswrapper[4907]: E0313 14:30:44.529068 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2938e39-f1d8-4cdc-a32c-5d57b8f2034f" containerName="ovsdbserver-sb" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.529075 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2938e39-f1d8-4cdc-a32c-5d57b8f2034f" containerName="ovsdbserver-sb" Mar 13 14:30:44 crc kubenswrapper[4907]: E0313 14:30:44.529086 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="056dd756-0d7b-471b-9929-f622d05ad606" containerName="ovsdbserver-nb" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.529092 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="056dd756-0d7b-471b-9929-f622d05ad606" containerName="ovsdbserver-nb" Mar 13 14:30:44 crc kubenswrapper[4907]: E0313 14:30:44.529102 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7da13722-732a-4b65-a894-fe4612f30d75" containerName="dnsmasq-dns" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.529109 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="7da13722-732a-4b65-a894-fe4612f30d75" containerName="dnsmasq-dns" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.529270 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="37978cc1-fcba-4032-a8b1-6632b61692ff" containerName="openstack-network-exporter" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.529282 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="056dd756-0d7b-471b-9929-f622d05ad606" containerName="ovsdbserver-nb" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.529291 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="7da13722-732a-4b65-a894-fe4612f30d75" containerName="dnsmasq-dns" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.529300 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a1494c5-e2d6-4d29-b161-97b720742d8d" containerName="proxy-server" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.529308 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="26827945-75f0-4867-ba04-31ff6428e06a" containerName="openstack-network-exporter" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.529318 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="c2938e39-f1d8-4cdc-a32c-5d57b8f2034f" containerName="openstack-network-exporter" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.529326 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="c2938e39-f1d8-4cdc-a32c-5d57b8f2034f" containerName="ovsdbserver-sb" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.529336 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a1494c5-e2d6-4d29-b161-97b720742d8d" containerName="proxy-httpd" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.529345 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="62983e04-f043-4ad3-9a46-96179d59671e" containerName="nova-cell1-novncproxy-novncproxy" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.529353 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="056dd756-0d7b-471b-9929-f622d05ad606" containerName="openstack-network-exporter" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.529362 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="3793f53a-ca36-4b27-8444-d6dbfd860424" containerName="probe" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.529375 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="3793f53a-ca36-4b27-8444-d6dbfd860424" containerName="cinder-scheduler" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.529383 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="26827945-75f0-4867-ba04-31ff6428e06a" containerName="ovn-northd" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.530046 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-04a8-account-create-update-kntb5" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.535558 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.549685 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-04a8-account-create-update-kntb5"] Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.580593 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-crrww"] Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.597970 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-kcjn8"] Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.613975 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-kcjn8"] Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.615481 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-48ngf\" (UniqueName: \"kubernetes.io/projected/4685651c-2690-48f8-97d2-8817fedaff56-kube-api-access-48ngf\") pod \"keystone-04a8-account-create-update-kntb5\" (UID: \"4685651c-2690-48f8-97d2-8817fedaff56\") " pod="openstack/keystone-04a8-account-create-update-kntb5" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.615695 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4685651c-2690-48f8-97d2-8817fedaff56-operator-scripts\") pod \"keystone-04a8-account-create-update-kntb5\" (UID: \"4685651c-2690-48f8-97d2-8817fedaff56\") " pod="openstack/keystone-04a8-account-create-update-kntb5" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.616367 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jstl5\" (UniqueName: \"kubernetes.io/projected/bc15f080-f90d-4e97-a866-d5edfd662bb7-kube-api-access-jstl5\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.616494 4907 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bc15f080-f90d-4e97-a866-d5edfd662bb7-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.628647 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-crrww"] Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.665320 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-85b9d64669-dxnf6"] Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.665550 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/keystone-85b9d64669-dxnf6" podUID="8a738a66-54bc-4f5e-86d5-6e3004d8f265" containerName="keystone-api" containerID="cri-o://e6a1933783ac7223012561b093ce055e734ac2c3c802390f6f0bf14354694b72" gracePeriod=30 Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.679206 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-galera-0"] Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.684533 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-47e2-account-create-update-wvjr7" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.711233 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-hg688"] Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.720677 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/82dec019-ba60-43f5-845b-322a6bcca225-operator-scripts\") pod \"82dec019-ba60-43f5-845b-322a6bcca225\" (UID: \"82dec019-ba60-43f5-845b-322a6bcca225\") " Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.720840 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vlhjv\" (UniqueName: \"kubernetes.io/projected/82dec019-ba60-43f5-845b-322a6bcca225-kube-api-access-vlhjv\") pod \"82dec019-ba60-43f5-845b-322a6bcca225\" (UID: \"82dec019-ba60-43f5-845b-322a6bcca225\") " Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.721303 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-48ngf\" (UniqueName: \"kubernetes.io/projected/4685651c-2690-48f8-97d2-8817fedaff56-kube-api-access-48ngf\") pod \"keystone-04a8-account-create-update-kntb5\" (UID: \"4685651c-2690-48f8-97d2-8817fedaff56\") " pod="openstack/keystone-04a8-account-create-update-kntb5" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.721348 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4685651c-2690-48f8-97d2-8817fedaff56-operator-scripts\") pod \"keystone-04a8-account-create-update-kntb5\" (UID: \"4685651c-2690-48f8-97d2-8817fedaff56\") " pod="openstack/keystone-04a8-account-create-update-kntb5" Mar 13 14:30:44 crc kubenswrapper[4907]: E0313 14:30:44.721553 4907 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Mar 13 14:30:44 crc kubenswrapper[4907]: E0313 14:30:44.721605 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/4685651c-2690-48f8-97d2-8817fedaff56-operator-scripts podName:4685651c-2690-48f8-97d2-8817fedaff56 nodeName:}" failed. No retries permitted until 2026-03-13 14:30:45.221588078 +0000 UTC m=+1544.121375767 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/4685651c-2690-48f8-97d2-8817fedaff56-operator-scripts") pod "keystone-04a8-account-create-update-kntb5" (UID: "4685651c-2690-48f8-97d2-8817fedaff56") : configmap "openstack-scripts" not found Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.722278 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/82dec019-ba60-43f5-845b-322a6bcca225-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "82dec019-ba60-43f5-845b-322a6bcca225" (UID: "82dec019-ba60-43f5-845b-322a6bcca225"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.726536 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/82dec019-ba60-43f5-845b-322a6bcca225-kube-api-access-vlhjv" (OuterVolumeSpecName: "kube-api-access-vlhjv") pod "82dec019-ba60-43f5-845b-322a6bcca225" (UID: "82dec019-ba60-43f5-845b-322a6bcca225"). InnerVolumeSpecName "kube-api-access-vlhjv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.726731 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-glzbl" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.733136 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-04a8-account-create-update-kntb5"] Mar 13 14:30:44 crc kubenswrapper[4907]: E0313 14:30:44.733577 4907 projected.go:194] Error preparing data for projected volume kube-api-access-48ngf for pod openstack/keystone-04a8-account-create-update-kntb5: failed to fetch token: serviceaccounts "galera-openstack" not found Mar 13 14:30:44 crc kubenswrapper[4907]: E0313 14:30:44.733639 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/4685651c-2690-48f8-97d2-8817fedaff56-kube-api-access-48ngf podName:4685651c-2690-48f8-97d2-8817fedaff56 nodeName:}" failed. No retries permitted until 2026-03-13 14:30:45.233621365 +0000 UTC m=+1544.133409064 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-48ngf" (UniqueName: "kubernetes.io/projected/4685651c-2690-48f8-97d2-8817fedaff56-kube-api-access-48ngf") pod "keystone-04a8-account-create-update-kntb5" (UID: "4685651c-2690-48f8-97d2-8817fedaff56") : failed to fetch token: serviceaccounts "galera-openstack" not found Mar 13 14:30:44 crc kubenswrapper[4907]: E0313 14:30:44.733910 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[kube-api-access-48ngf operator-scripts], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/keystone-04a8-account-create-update-kntb5" podUID="4685651c-2690-48f8-97d2-8817fedaff56" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.749181 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-hg688"] Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.750640 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.822498 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58ab9689-3ab2-42f2-ad56-beb22e29f8da-combined-ca-bundle\") pod \"58ab9689-3ab2-42f2-ad56-beb22e29f8da\" (UID: \"58ab9689-3ab2-42f2-ad56-beb22e29f8da\") " Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.822590 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kvwz4\" (UniqueName: \"kubernetes.io/projected/58ab9689-3ab2-42f2-ad56-beb22e29f8da-kube-api-access-kvwz4\") pod \"58ab9689-3ab2-42f2-ad56-beb22e29f8da\" (UID: \"58ab9689-3ab2-42f2-ad56-beb22e29f8da\") " Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.822628 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/58ab9689-3ab2-42f2-ad56-beb22e29f8da-config-data-default\") pod \"58ab9689-3ab2-42f2-ad56-beb22e29f8da\" (UID: \"58ab9689-3ab2-42f2-ad56-beb22e29f8da\") " Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.822730 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/58ab9689-3ab2-42f2-ad56-beb22e29f8da-galera-tls-certs\") pod \"58ab9689-3ab2-42f2-ad56-beb22e29f8da\" (UID: \"58ab9689-3ab2-42f2-ad56-beb22e29f8da\") " Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.822841 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/516d1499-3461-4af5-b426-9ae4a711a468-operator-scripts\") pod \"516d1499-3461-4af5-b426-9ae4a711a468\" (UID: \"516d1499-3461-4af5-b426-9ae4a711a468\") " Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.822914 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-df597\" (UniqueName: \"kubernetes.io/projected/516d1499-3461-4af5-b426-9ae4a711a468-kube-api-access-df597\") pod \"516d1499-3461-4af5-b426-9ae4a711a468\" (UID: \"516d1499-3461-4af5-b426-9ae4a711a468\") " Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.822948 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mysql-db\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"58ab9689-3ab2-42f2-ad56-beb22e29f8da\" (UID: \"58ab9689-3ab2-42f2-ad56-beb22e29f8da\") " Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.822975 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/58ab9689-3ab2-42f2-ad56-beb22e29f8da-operator-scripts\") pod \"58ab9689-3ab2-42f2-ad56-beb22e29f8da\" (UID: \"58ab9689-3ab2-42f2-ad56-beb22e29f8da\") " Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.823003 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/58ab9689-3ab2-42f2-ad56-beb22e29f8da-kolla-config\") pod \"58ab9689-3ab2-42f2-ad56-beb22e29f8da\" (UID: \"58ab9689-3ab2-42f2-ad56-beb22e29f8da\") " Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.823042 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/58ab9689-3ab2-42f2-ad56-beb22e29f8da-config-data-generated\") pod \"58ab9689-3ab2-42f2-ad56-beb22e29f8da\" (UID: \"58ab9689-3ab2-42f2-ad56-beb22e29f8da\") " Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.823822 4907 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/82dec019-ba60-43f5-845b-322a6bcca225-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.823842 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vlhjv\" (UniqueName: \"kubernetes.io/projected/82dec019-ba60-43f5-845b-322a6bcca225-kube-api-access-vlhjv\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.827211 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/58ab9689-3ab2-42f2-ad56-beb22e29f8da-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "58ab9689-3ab2-42f2-ad56-beb22e29f8da" (UID: "58ab9689-3ab2-42f2-ad56-beb22e29f8da"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.827673 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/58ab9689-3ab2-42f2-ad56-beb22e29f8da-config-data-generated" (OuterVolumeSpecName: "config-data-generated") pod "58ab9689-3ab2-42f2-ad56-beb22e29f8da" (UID: "58ab9689-3ab2-42f2-ad56-beb22e29f8da"). InnerVolumeSpecName "config-data-generated". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.827687 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/58ab9689-3ab2-42f2-ad56-beb22e29f8da-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "58ab9689-3ab2-42f2-ad56-beb22e29f8da" (UID: "58ab9689-3ab2-42f2-ad56-beb22e29f8da"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.828347 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/516d1499-3461-4af5-b426-9ae4a711a468-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "516d1499-3461-4af5-b426-9ae4a711a468" (UID: "516d1499-3461-4af5-b426-9ae4a711a468"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.828479 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/58ab9689-3ab2-42f2-ad56-beb22e29f8da-config-data-default" (OuterVolumeSpecName: "config-data-default") pod "58ab9689-3ab2-42f2-ad56-beb22e29f8da" (UID: "58ab9689-3ab2-42f2-ad56-beb22e29f8da"). InnerVolumeSpecName "config-data-default". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.834396 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/58ab9689-3ab2-42f2-ad56-beb22e29f8da-kube-api-access-kvwz4" (OuterVolumeSpecName: "kube-api-access-kvwz4") pod "58ab9689-3ab2-42f2-ad56-beb22e29f8da" (UID: "58ab9689-3ab2-42f2-ad56-beb22e29f8da"). InnerVolumeSpecName "kube-api-access-kvwz4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.834623 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/516d1499-3461-4af5-b426-9ae4a711a468-kube-api-access-df597" (OuterVolumeSpecName: "kube-api-access-df597") pod "516d1499-3461-4af5-b426-9ae4a711a468" (UID: "516d1499-3461-4af5-b426-9ae4a711a468"). InnerVolumeSpecName "kube-api-access-df597". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.866751 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "mysql-db") pod "58ab9689-3ab2-42f2-ad56-beb22e29f8da" (UID: "58ab9689-3ab2-42f2-ad56-beb22e29f8da"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.877263 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58ab9689-3ab2-42f2-ad56-beb22e29f8da-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "58ab9689-3ab2-42f2-ad56-beb22e29f8da" (UID: "58ab9689-3ab2-42f2-ad56-beb22e29f8da"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.919561 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58ab9689-3ab2-42f2-ad56-beb22e29f8da-galera-tls-certs" (OuterVolumeSpecName: "galera-tls-certs") pod "58ab9689-3ab2-42f2-ad56-beb22e29f8da" (UID: "58ab9689-3ab2-42f2-ad56-beb22e29f8da"). InnerVolumeSpecName "galera-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.932465 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstack-galera-0" podUID="c4411bd2-c555-433a-9015-f623948b1401" containerName="galera" containerID="cri-o://b5e68fc91f35838327b362dd7ecd6d2f5ee1173bc7ae4fafcb541ac0c96b44fa" gracePeriod=30 Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.933774 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-9c64-account-create-update-5vvgg" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.934616 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58ab9689-3ab2-42f2-ad56-beb22e29f8da-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.934641 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kvwz4\" (UniqueName: \"kubernetes.io/projected/58ab9689-3ab2-42f2-ad56-beb22e29f8da-kube-api-access-kvwz4\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.934651 4907 reconciler_common.go:293] "Volume detached for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/58ab9689-3ab2-42f2-ad56-beb22e29f8da-config-data-default\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.934661 4907 reconciler_common.go:293] "Volume detached for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/58ab9689-3ab2-42f2-ad56-beb22e29f8da-galera-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.934669 4907 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/516d1499-3461-4af5-b426-9ae4a711a468-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.934677 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-df597\" (UniqueName: \"kubernetes.io/projected/516d1499-3461-4af5-b426-9ae4a711a468-kube-api-access-df597\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.934699 4907 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.934708 4907 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/58ab9689-3ab2-42f2-ad56-beb22e29f8da-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.934731 4907 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/58ab9689-3ab2-42f2-ad56-beb22e29f8da-kolla-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.934742 4907 reconciler_common.go:293] "Volume detached for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/58ab9689-3ab2-42f2-ad56-beb22e29f8da-config-data-generated\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.947132 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-b69e-account-create-update-96wq7" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.969395 4907 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.032145 4907 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-bw2xb" podUID="6f8119b2-e38a-494d-967f-5198b83512c7" containerName="ovn-controller" probeResult="failure" output="" Mar 13 14:30:45 crc kubenswrapper[4907]: E0313 14:30:45.035967 4907 handlers.go:78] "Exec lifecycle hook for Container in Pod failed" err=< Mar 13 14:30:45 crc kubenswrapper[4907]: command '/usr/share/ovn/scripts/ovn-ctl stop_controller' exited with 137: 2026-03-13T14:30:42Z|00001|fatal_signal|WARN|terminating with signal 14 (Alarm clock) Mar 13 14:30:45 crc kubenswrapper[4907]: /etc/init.d/functions: line 589: 400 Alarm clock "$@" Mar 13 14:30:45 crc kubenswrapper[4907]: > execCommand=["/usr/share/ovn/scripts/ovn-ctl","stop_controller"] containerName="ovn-controller" pod="openstack/ovn-controller-bw2xb" message=< Mar 13 14:30:45 crc kubenswrapper[4907]: Exiting ovn-controller (1) [FAILED] Mar 13 14:30:45 crc kubenswrapper[4907]: Killing ovn-controller (1) [ OK ] Mar 13 14:30:45 crc kubenswrapper[4907]: 2026-03-13T14:30:42Z|00001|fatal_signal|WARN|terminating with signal 14 (Alarm clock) Mar 13 14:30:45 crc kubenswrapper[4907]: /etc/init.d/functions: line 589: 400 Alarm clock "$@" Mar 13 14:30:45 crc kubenswrapper[4907]: > Mar 13 14:30:45 crc kubenswrapper[4907]: E0313 14:30:45.036010 4907 kuberuntime_container.go:691] "PreStop hook failed" err=< Mar 13 14:30:45 crc kubenswrapper[4907]: command '/usr/share/ovn/scripts/ovn-ctl stop_controller' exited with 137: 2026-03-13T14:30:42Z|00001|fatal_signal|WARN|terminating with signal 14 (Alarm clock) Mar 13 14:30:45 crc kubenswrapper[4907]: /etc/init.d/functions: line 589: 400 Alarm clock "$@" Mar 13 14:30:45 crc kubenswrapper[4907]: > pod="openstack/ovn-controller-bw2xb" podUID="6f8119b2-e38a-494d-967f-5198b83512c7" containerName="ovn-controller" containerID="cri-o://df7f17b832ba26f166592b0696a663ded9a2e03678601f86699f09b76e050e58" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.036043 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-bw2xb" podUID="6f8119b2-e38a-494d-967f-5198b83512c7" containerName="ovn-controller" containerID="cri-o://df7f17b832ba26f166592b0696a663ded9a2e03678601f86699f09b76e050e58" gracePeriod=27 Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.036737 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4a4c64a3-7378-4d8d-9cf3-17583c6a90c0-operator-scripts\") pod \"4a4c64a3-7378-4d8d-9cf3-17583c6a90c0\" (UID: \"4a4c64a3-7378-4d8d-9cf3-17583c6a90c0\") " Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.036806 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfp65\" (UniqueName: \"kubernetes.io/projected/e8f6894c-b714-431e-912d-8b8de1a3914b-kube-api-access-kfp65\") pod \"e8f6894c-b714-431e-912d-8b8de1a3914b\" (UID: \"e8f6894c-b714-431e-912d-8b8de1a3914b\") " Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.036875 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e8f6894c-b714-431e-912d-8b8de1a3914b-operator-scripts\") pod \"e8f6894c-b714-431e-912d-8b8de1a3914b\" (UID: \"e8f6894c-b714-431e-912d-8b8de1a3914b\") " Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.036941 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wqnkx\" (UniqueName: \"kubernetes.io/projected/4a4c64a3-7378-4d8d-9cf3-17583c6a90c0-kube-api-access-wqnkx\") pod \"4a4c64a3-7378-4d8d-9cf3-17583c6a90c0\" (UID: \"4a4c64a3-7378-4d8d-9cf3-17583c6a90c0\") " Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.037233 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4a4c64a3-7378-4d8d-9cf3-17583c6a90c0-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "4a4c64a3-7378-4d8d-9cf3-17583c6a90c0" (UID: "4a4c64a3-7378-4d8d-9cf3-17583c6a90c0"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.037417 4907 reconciler_common.go:293] "Volume detached for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.037432 4907 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4a4c64a3-7378-4d8d-9cf3-17583c6a90c0-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.038332 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e8f6894c-b714-431e-912d-8b8de1a3914b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e8f6894c-b714-431e-912d-8b8de1a3914b" (UID: "e8f6894c-b714-431e-912d-8b8de1a3914b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.041498 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e8f6894c-b714-431e-912d-8b8de1a3914b-kube-api-access-kfp65" (OuterVolumeSpecName: "kube-api-access-kfp65") pod "e8f6894c-b714-431e-912d-8b8de1a3914b" (UID: "e8f6894c-b714-431e-912d-8b8de1a3914b"). InnerVolumeSpecName "kube-api-access-kfp65". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.047066 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4a4c64a3-7378-4d8d-9cf3-17583c6a90c0-kube-api-access-wqnkx" (OuterVolumeSpecName: "kube-api-access-wqnkx") pod "4a4c64a3-7378-4d8d-9cf3-17583c6a90c0" (UID: "4a4c64a3-7378-4d8d-9cf3-17583c6a90c0"). InnerVolumeSpecName "kube-api-access-wqnkx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.098800 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"62983e04-f043-4ad3-9a46-96179d59671e","Type":"ContainerDied","Data":"1cd31d9d43a5e6a87bb0836d84499cc5b9fab420b814ef48408f1a9549204fd0"} Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.098854 4907 scope.go:117] "RemoveContainer" containerID="9e2de4df42a807d1995b769c3c69187ddaaf7da1c1d09354b07088fb415f36d7" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.098993 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.107056 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-17a8-account-create-update-nj9pw" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.107711 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-17a8-account-create-update-nj9pw" event={"ID":"bc15f080-f90d-4e97-a866-d5edfd662bb7","Type":"ContainerDied","Data":"bc3dc3b0fd90acb1ff4424b36cbf0bd4f5dd344f9b23a791c924f552a3241a63"} Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.125857 4907 generic.go:334] "Generic (PLEG): container finished" podID="58ab9689-3ab2-42f2-ad56-beb22e29f8da" containerID="af7b1c6225d2e94f5a2f9827f7b02dc18651ab20c1d53b4a73e6d032e463897a" exitCode=0 Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.125934 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"58ab9689-3ab2-42f2-ad56-beb22e29f8da","Type":"ContainerDied","Data":"af7b1c6225d2e94f5a2f9827f7b02dc18651ab20c1d53b4a73e6d032e463897a"} Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.125961 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"58ab9689-3ab2-42f2-ad56-beb22e29f8da","Type":"ContainerDied","Data":"be11df2bf36396e7ed2d074eed97b3e79df7e43e0e8a1a8dec1ef8739586a724"} Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.126019 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.138519 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfp65\" (UniqueName: \"kubernetes.io/projected/e8f6894c-b714-431e-912d-8b8de1a3914b-kube-api-access-kfp65\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.138547 4907 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e8f6894c-b714-431e-912d-8b8de1a3914b-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.138556 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wqnkx\" (UniqueName: \"kubernetes.io/projected/4a4c64a3-7378-4d8d-9cf3-17583c6a90c0-kube-api-access-wqnkx\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.139561 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-glzbl" event={"ID":"516d1499-3461-4af5-b426-9ae4a711a468","Type":"ContainerDied","Data":"f6d2f722265ae72bf48903452fd81587c07136e043b3e098f648b2d79cb3fd59"} Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.139631 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-glzbl" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.147154 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.149424 4907 generic.go:334] "Generic (PLEG): container finished" podID="aa81fa2a-4c27-4e02-9891-5b39348feca9" containerID="79914c8007c9b10b224f5d4d1fcd270f17078ab7c002b89d3dab089884af5d6c" exitCode=0 Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.149446 4907 generic.go:334] "Generic (PLEG): container finished" podID="aa81fa2a-4c27-4e02-9891-5b39348feca9" containerID="16ef32bcd58d9ca78315416c39fe9eca0c5eee8682ac66cea46ec8f9242f8ee5" exitCode=2 Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.149454 4907 generic.go:334] "Generic (PLEG): container finished" podID="aa81fa2a-4c27-4e02-9891-5b39348feca9" containerID="285cdb1df855cbeeda75e9d50bffbe69e5c3f9a5485dd118f47a932fb1d36f19" exitCode=0 Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.149771 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aa81fa2a-4c27-4e02-9891-5b39348feca9","Type":"ContainerDied","Data":"79914c8007c9b10b224f5d4d1fcd270f17078ab7c002b89d3dab089884af5d6c"} Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.149802 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aa81fa2a-4c27-4e02-9891-5b39348feca9","Type":"ContainerDied","Data":"16ef32bcd58d9ca78315416c39fe9eca0c5eee8682ac66cea46ec8f9242f8ee5"} Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.149815 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aa81fa2a-4c27-4e02-9891-5b39348feca9","Type":"ContainerDied","Data":"285cdb1df855cbeeda75e9d50bffbe69e5c3f9a5485dd118f47a932fb1d36f19"} Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.152815 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-9c64-account-create-update-5vvgg" event={"ID":"4a4c64a3-7378-4d8d-9cf3-17583c6a90c0","Type":"ContainerDied","Data":"a22452114ba800cf0f41433cad0c073716fcc8d240adefbf18595bc1cdb4a098"} Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.152865 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-9c64-account-create-update-5vvgg" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.158153 4907 generic.go:334] "Generic (PLEG): container finished" podID="c49286b1-2e9f-4de6-b7de-2e952f3ad607" containerID="0d242394ff0714a798d1ef385dde5709185206b531dfb7b51abcf0db5d059589" exitCode=2 Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.158195 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"c49286b1-2e9f-4de6-b7de-2e952f3ad607","Type":"ContainerDied","Data":"0d242394ff0714a798d1ef385dde5709185206b531dfb7b51abcf0db5d059589"} Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.158212 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"c49286b1-2e9f-4de6-b7de-2e952f3ad607","Type":"ContainerDied","Data":"50ab44d5af6ebe42498682725ea848a78fc975fa9daf535d39dda1ec402423d1"} Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.158265 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.159931 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-b69e-account-create-update-96wq7" event={"ID":"e8f6894c-b714-431e-912d-8b8de1a3914b","Type":"ContainerDied","Data":"d38b680247762c5b2510a8776b9df1c80f0da04bd9de35e5d2db2452cf7446f0"} Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.160037 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-b69e-account-create-update-96wq7" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.161357 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-bw2xb_6f8119b2-e38a-494d-967f-5198b83512c7/ovn-controller/0.log" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.161386 4907 generic.go:334] "Generic (PLEG): container finished" podID="6f8119b2-e38a-494d-967f-5198b83512c7" containerID="df7f17b832ba26f166592b0696a663ded9a2e03678601f86699f09b76e050e58" exitCode=143 Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.161424 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-bw2xb" event={"ID":"6f8119b2-e38a-494d-967f-5198b83512c7","Type":"ContainerDied","Data":"df7f17b832ba26f166592b0696a663ded9a2e03678601f86699f09b76e050e58"} Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.171632 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-47e2-account-create-update-wvjr7" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.172061 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-47e2-account-create-update-wvjr7" event={"ID":"82dec019-ba60-43f5-845b-322a6bcca225","Type":"ContainerDied","Data":"cbbe85502bda2ac6ef5bc690e92165a5b92af3bcd3162f655a1bd43c95395317"} Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.172568 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-04a8-account-create-update-kntb5" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.187315 4907 scope.go:117] "RemoveContainer" containerID="af7b1c6225d2e94f5a2f9827f7b02dc18651ab20c1d53b4a73e6d032e463897a" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.199732 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-04a8-account-create-update-kntb5" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.201159 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-cell1-galera-0"] Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.224923 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstack-cell1-galera-0"] Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.239709 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/c49286b1-2e9f-4de6-b7de-2e952f3ad607-kube-state-metrics-tls-config\") pod \"c49286b1-2e9f-4de6-b7de-2e952f3ad607\" (UID: \"c49286b1-2e9f-4de6-b7de-2e952f3ad607\") " Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.239761 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/c49286b1-2e9f-4de6-b7de-2e952f3ad607-kube-state-metrics-tls-certs\") pod \"c49286b1-2e9f-4de6-b7de-2e952f3ad607\" (UID: \"c49286b1-2e9f-4de6-b7de-2e952f3ad607\") " Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.239833 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c49286b1-2e9f-4de6-b7de-2e952f3ad607-combined-ca-bundle\") pod \"c49286b1-2e9f-4de6-b7de-2e952f3ad607\" (UID: \"c49286b1-2e9f-4de6-b7de-2e952f3ad607\") " Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.239967 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-29ctv\" (UniqueName: \"kubernetes.io/projected/c49286b1-2e9f-4de6-b7de-2e952f3ad607-kube-api-access-29ctv\") pod \"c49286b1-2e9f-4de6-b7de-2e952f3ad607\" (UID: \"c49286b1-2e9f-4de6-b7de-2e952f3ad607\") " Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.240216 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4685651c-2690-48f8-97d2-8817fedaff56-operator-scripts\") pod \"keystone-04a8-account-create-update-kntb5\" (UID: \"4685651c-2690-48f8-97d2-8817fedaff56\") " pod="openstack/keystone-04a8-account-create-update-kntb5" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.240405 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-48ngf\" (UniqueName: \"kubernetes.io/projected/4685651c-2690-48f8-97d2-8817fedaff56-kube-api-access-48ngf\") pod \"keystone-04a8-account-create-update-kntb5\" (UID: \"4685651c-2690-48f8-97d2-8817fedaff56\") " pod="openstack/keystone-04a8-account-create-update-kntb5" Mar 13 14:30:45 crc kubenswrapper[4907]: E0313 14:30:45.240407 4907 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Mar 13 14:30:45 crc kubenswrapper[4907]: E0313 14:30:45.240493 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/4685651c-2690-48f8-97d2-8817fedaff56-operator-scripts podName:4685651c-2690-48f8-97d2-8817fedaff56 nodeName:}" failed. No retries permitted until 2026-03-13 14:30:46.240470876 +0000 UTC m=+1545.140258635 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/4685651c-2690-48f8-97d2-8817fedaff56-operator-scripts") pod "keystone-04a8-account-create-update-kntb5" (UID: "4685651c-2690-48f8-97d2-8817fedaff56") : configmap "openstack-scripts" not found Mar 13 14:30:45 crc kubenswrapper[4907]: E0313 14:30:45.250783 4907 projected.go:194] Error preparing data for projected volume kube-api-access-48ngf for pod openstack/keystone-04a8-account-create-update-kntb5: failed to fetch token: serviceaccounts "galera-openstack" not found Mar 13 14:30:45 crc kubenswrapper[4907]: E0313 14:30:45.250865 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/4685651c-2690-48f8-97d2-8817fedaff56-kube-api-access-48ngf podName:4685651c-2690-48f8-97d2-8817fedaff56 nodeName:}" failed. No retries permitted until 2026-03-13 14:30:46.250843999 +0000 UTC m=+1545.150631698 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-48ngf" (UniqueName: "kubernetes.io/projected/4685651c-2690-48f8-97d2-8817fedaff56-kube-api-access-48ngf") pod "keystone-04a8-account-create-update-kntb5" (UID: "4685651c-2690-48f8-97d2-8817fedaff56") : failed to fetch token: serviceaccounts "galera-openstack" not found Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.259585 4907 scope.go:117] "RemoveContainer" containerID="62c42ba7e98c9bfe072975553c668c323179c088fa7a92a9812e075126aebb86" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.278111 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-17a8-account-create-update-nj9pw"] Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.280899 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c49286b1-2e9f-4de6-b7de-2e952f3ad607-kube-api-access-29ctv" (OuterVolumeSpecName: "kube-api-access-29ctv") pod "c49286b1-2e9f-4de6-b7de-2e952f3ad607" (UID: "c49286b1-2e9f-4de6-b7de-2e952f3ad607"). InnerVolumeSpecName "kube-api-access-29ctv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.286621 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-17a8-account-create-update-nj9pw"] Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.307017 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.342149 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-29ctv\" (UniqueName: \"kubernetes.io/projected/c49286b1-2e9f-4de6-b7de-2e952f3ad607-kube-api-access-29ctv\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.348252 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.356788 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c49286b1-2e9f-4de6-b7de-2e952f3ad607-kube-state-metrics-tls-certs" (OuterVolumeSpecName: "kube-state-metrics-tls-certs") pod "c49286b1-2e9f-4de6-b7de-2e952f3ad607" (UID: "c49286b1-2e9f-4de6-b7de-2e952f3ad607"). InnerVolumeSpecName "kube-state-metrics-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.380251 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c49286b1-2e9f-4de6-b7de-2e952f3ad607-kube-state-metrics-tls-config" (OuterVolumeSpecName: "kube-state-metrics-tls-config") pod "c49286b1-2e9f-4de6-b7de-2e952f3ad607" (UID: "c49286b1-2e9f-4de6-b7de-2e952f3ad607"). InnerVolumeSpecName "kube-state-metrics-tls-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.384230 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-47e2-account-create-update-wvjr7"] Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.392142 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c49286b1-2e9f-4de6-b7de-2e952f3ad607-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c49286b1-2e9f-4de6-b7de-2e952f3ad607" (UID: "c49286b1-2e9f-4de6-b7de-2e952f3ad607"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.392215 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-47e2-account-create-update-wvjr7"] Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.406002 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-bw2xb_6f8119b2-e38a-494d-967f-5198b83512c7/ovn-controller/0.log" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.406086 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-bw2xb" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.413330 4907 scope.go:117] "RemoveContainer" containerID="af7b1c6225d2e94f5a2f9827f7b02dc18651ab20c1d53b4a73e6d032e463897a" Mar 13 14:30:45 crc kubenswrapper[4907]: E0313 14:30:45.414314 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"af7b1c6225d2e94f5a2f9827f7b02dc18651ab20c1d53b4a73e6d032e463897a\": container with ID starting with af7b1c6225d2e94f5a2f9827f7b02dc18651ab20c1d53b4a73e6d032e463897a not found: ID does not exist" containerID="af7b1c6225d2e94f5a2f9827f7b02dc18651ab20c1d53b4a73e6d032e463897a" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.414359 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af7b1c6225d2e94f5a2f9827f7b02dc18651ab20c1d53b4a73e6d032e463897a"} err="failed to get container status \"af7b1c6225d2e94f5a2f9827f7b02dc18651ab20c1d53b4a73e6d032e463897a\": rpc error: code = NotFound desc = could not find container \"af7b1c6225d2e94f5a2f9827f7b02dc18651ab20c1d53b4a73e6d032e463897a\": container with ID starting with af7b1c6225d2e94f5a2f9827f7b02dc18651ab20c1d53b4a73e6d032e463897a not found: ID does not exist" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.414385 4907 scope.go:117] "RemoveContainer" containerID="62c42ba7e98c9bfe072975553c668c323179c088fa7a92a9812e075126aebb86" Mar 13 14:30:45 crc kubenswrapper[4907]: E0313 14:30:45.415231 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"62c42ba7e98c9bfe072975553c668c323179c088fa7a92a9812e075126aebb86\": container with ID starting with 62c42ba7e98c9bfe072975553c668c323179c088fa7a92a9812e075126aebb86 not found: ID does not exist" containerID="62c42ba7e98c9bfe072975553c668c323179c088fa7a92a9812e075126aebb86" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.415350 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"62c42ba7e98c9bfe072975553c668c323179c088fa7a92a9812e075126aebb86"} err="failed to get container status \"62c42ba7e98c9bfe072975553c668c323179c088fa7a92a9812e075126aebb86\": rpc error: code = NotFound desc = could not find container \"62c42ba7e98c9bfe072975553c668c323179c088fa7a92a9812e075126aebb86\": container with ID starting with 62c42ba7e98c9bfe072975553c668c323179c088fa7a92a9812e075126aebb86 not found: ID does not exist" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.415471 4907 scope.go:117] "RemoveContainer" containerID="73f68d1aced419107233520623fb182eb32be93775f62b9a147d1f6ec4671d78" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.421149 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/root-account-create-update-glzbl"] Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.439464 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/root-account-create-update-glzbl"] Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.444128 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/6f8119b2-e38a-494d-967f-5198b83512c7-var-run-ovn\") pod \"6f8119b2-e38a-494d-967f-5198b83512c7\" (UID: \"6f8119b2-e38a-494d-967f-5198b83512c7\") " Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.444197 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6f8119b2-e38a-494d-967f-5198b83512c7-var-run\") pod \"6f8119b2-e38a-494d-967f-5198b83512c7\" (UID: \"6f8119b2-e38a-494d-967f-5198b83512c7\") " Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.444264 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v2qq5\" (UniqueName: \"kubernetes.io/projected/6f8119b2-e38a-494d-967f-5198b83512c7-kube-api-access-v2qq5\") pod \"6f8119b2-e38a-494d-967f-5198b83512c7\" (UID: \"6f8119b2-e38a-494d-967f-5198b83512c7\") " Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.444293 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/6f8119b2-e38a-494d-967f-5198b83512c7-ovn-controller-tls-certs\") pod \"6f8119b2-e38a-494d-967f-5198b83512c7\" (UID: \"6f8119b2-e38a-494d-967f-5198b83512c7\") " Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.444386 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6f8119b2-e38a-494d-967f-5198b83512c7-var-run" (OuterVolumeSpecName: "var-run") pod "6f8119b2-e38a-494d-967f-5198b83512c7" (UID: "6f8119b2-e38a-494d-967f-5198b83512c7"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.444405 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f8119b2-e38a-494d-967f-5198b83512c7-combined-ca-bundle\") pod \"6f8119b2-e38a-494d-967f-5198b83512c7\" (UID: \"6f8119b2-e38a-494d-967f-5198b83512c7\") " Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.444475 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6f8119b2-e38a-494d-967f-5198b83512c7-scripts\") pod \"6f8119b2-e38a-494d-967f-5198b83512c7\" (UID: \"6f8119b2-e38a-494d-967f-5198b83512c7\") " Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.444543 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/6f8119b2-e38a-494d-967f-5198b83512c7-var-log-ovn\") pod \"6f8119b2-e38a-494d-967f-5198b83512c7\" (UID: \"6f8119b2-e38a-494d-967f-5198b83512c7\") " Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.445304 4907 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6f8119b2-e38a-494d-967f-5198b83512c7-var-run\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.445323 4907 reconciler_common.go:293] "Volume detached for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/c49286b1-2e9f-4de6-b7de-2e952f3ad607-kube-state-metrics-tls-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.445334 4907 reconciler_common.go:293] "Volume detached for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/c49286b1-2e9f-4de6-b7de-2e952f3ad607-kube-state-metrics-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.445343 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c49286b1-2e9f-4de6-b7de-2e952f3ad607-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.445532 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6f8119b2-e38a-494d-967f-5198b83512c7-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "6f8119b2-e38a-494d-967f-5198b83512c7" (UID: "6f8119b2-e38a-494d-967f-5198b83512c7"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.445927 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6f8119b2-e38a-494d-967f-5198b83512c7-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "6f8119b2-e38a-494d-967f-5198b83512c7" (UID: "6f8119b2-e38a-494d-967f-5198b83512c7"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.447602 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6f8119b2-e38a-494d-967f-5198b83512c7-scripts" (OuterVolumeSpecName: "scripts") pod "6f8119b2-e38a-494d-967f-5198b83512c7" (UID: "6f8119b2-e38a-494d-967f-5198b83512c7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.454475 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6f8119b2-e38a-494d-967f-5198b83512c7-kube-api-access-v2qq5" (OuterVolumeSpecName: "kube-api-access-v2qq5") pod "6f8119b2-e38a-494d-967f-5198b83512c7" (UID: "6f8119b2-e38a-494d-967f-5198b83512c7"). InnerVolumeSpecName "kube-api-access-v2qq5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.457010 4907 scope.go:117] "RemoveContainer" containerID="0d242394ff0714a798d1ef385dde5709185206b531dfb7b51abcf0db5d059589" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.481826 4907 scope.go:117] "RemoveContainer" containerID="0d242394ff0714a798d1ef385dde5709185206b531dfb7b51abcf0db5d059589" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.483125 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6f8119b2-e38a-494d-967f-5198b83512c7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6f8119b2-e38a-494d-967f-5198b83512c7" (UID: "6f8119b2-e38a-494d-967f-5198b83512c7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:45 crc kubenswrapper[4907]: E0313 14:30:45.483211 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0d242394ff0714a798d1ef385dde5709185206b531dfb7b51abcf0db5d059589\": container with ID starting with 0d242394ff0714a798d1ef385dde5709185206b531dfb7b51abcf0db5d059589 not found: ID does not exist" containerID="0d242394ff0714a798d1ef385dde5709185206b531dfb7b51abcf0db5d059589" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.483279 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0d242394ff0714a798d1ef385dde5709185206b531dfb7b51abcf0db5d059589"} err="failed to get container status \"0d242394ff0714a798d1ef385dde5709185206b531dfb7b51abcf0db5d059589\": rpc error: code = NotFound desc = could not find container \"0d242394ff0714a798d1ef385dde5709185206b531dfb7b51abcf0db5d059589\": container with ID starting with 0d242394ff0714a798d1ef385dde5709185206b531dfb7b51abcf0db5d059589 not found: ID does not exist" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.508423 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-9c64-account-create-update-5vvgg"] Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.518180 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-9c64-account-create-update-5vvgg"] Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.528757 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6f8119b2-e38a-494d-967f-5198b83512c7-ovn-controller-tls-certs" (OuterVolumeSpecName: "ovn-controller-tls-certs") pod "6f8119b2-e38a-494d-967f-5198b83512c7" (UID: "6f8119b2-e38a-494d-967f-5198b83512c7"). InnerVolumeSpecName "ovn-controller-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.534263 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-b69e-account-create-update-96wq7"] Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.543451 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-b69e-account-create-update-96wq7"] Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.547509 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f8119b2-e38a-494d-967f-5198b83512c7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.547539 4907 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6f8119b2-e38a-494d-967f-5198b83512c7-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.547550 4907 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/6f8119b2-e38a-494d-967f-5198b83512c7-var-log-ovn\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.547560 4907 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/6f8119b2-e38a-494d-967f-5198b83512c7-var-run-ovn\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.547569 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v2qq5\" (UniqueName: \"kubernetes.io/projected/6f8119b2-e38a-494d-967f-5198b83512c7-kube-api-access-v2qq5\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.547577 4907 reconciler_common.go:293] "Volume detached for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/6f8119b2-e38a-494d-967f-5198b83512c7-ovn-controller-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.555480 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.560449 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.599503 4907 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cinder-api-0" podUID="41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76" containerName="cinder-api" probeResult="failure" output="Get \"https://10.217.0.172:8776/healthcheck\": read tcp 10.217.0.2:39704->10.217.0.172:8776: read: connection reset by peer" Mar 13 14:30:45 crc kubenswrapper[4907]: E0313 14:30:45.752226 4907 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Mar 13 14:30:45 crc kubenswrapper[4907]: E0313 14:30:45.752296 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/e3c34454-315e-4821-ab25-b0f331a0d521-config-data podName:e3c34454-315e-4821-ab25-b0f331a0d521 nodeName:}" failed. No retries permitted until 2026-03-13 14:30:49.752279757 +0000 UTC m=+1548.652067436 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/e3c34454-315e-4821-ab25-b0f331a0d521-config-data") pod "rabbitmq-server-0" (UID: "e3c34454-315e-4821-ab25-b0f331a0d521") : configmap "rabbitmq-config-data" not found Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.799830 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1a1494c5-e2d6-4d29-b161-97b720742d8d" path="/var/lib/kubelet/pods/1a1494c5-e2d6-4d29-b161-97b720742d8d/volumes" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.800854 4907 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="aa81fa2a-4c27-4e02-9891-5b39348feca9" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.208:3000/\": dial tcp 10.217.0.208:3000: connect: connection refused" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.806518 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="34a0ea40-bdaa-4eef-8a28-263d8db74f10" path="/var/lib/kubelet/pods/34a0ea40-bdaa-4eef-8a28-263d8db74f10/volumes" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.807551 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3793f53a-ca36-4b27-8444-d6dbfd860424" path="/var/lib/kubelet/pods/3793f53a-ca36-4b27-8444-d6dbfd860424/volumes" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.824556 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="37ba2d41-5de4-48cb-aad6-59df13307ed1" path="/var/lib/kubelet/pods/37ba2d41-5de4-48cb-aad6-59df13307ed1/volumes" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.826520 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="480d23fe-2980-46b3-9e86-e74f2d051357" path="/var/lib/kubelet/pods/480d23fe-2980-46b3-9e86-e74f2d051357/volumes" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.827207 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4a4c64a3-7378-4d8d-9cf3-17583c6a90c0" path="/var/lib/kubelet/pods/4a4c64a3-7378-4d8d-9cf3-17583c6a90c0/volumes" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.827790 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="516d1499-3461-4af5-b426-9ae4a711a468" path="/var/lib/kubelet/pods/516d1499-3461-4af5-b426-9ae4a711a468/volumes" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.836529 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="58ab9689-3ab2-42f2-ad56-beb22e29f8da" path="/var/lib/kubelet/pods/58ab9689-3ab2-42f2-ad56-beb22e29f8da/volumes" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.838766 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="62983e04-f043-4ad3-9a46-96179d59671e" path="/var/lib/kubelet/pods/62983e04-f043-4ad3-9a46-96179d59671e/volumes" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.840932 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="82dec019-ba60-43f5-845b-322a6bcca225" path="/var/lib/kubelet/pods/82dec019-ba60-43f5-845b-322a6bcca225/volumes" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.841542 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9c985ff9-3545-4544-acc8-ca9eb2bfad6a" path="/var/lib/kubelet/pods/9c985ff9-3545-4544-acc8-ca9eb2bfad6a/volumes" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.846517 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a055ce7f-0538-4e5e-938a-eaf6d0aa2dec" path="/var/lib/kubelet/pods/a055ce7f-0538-4e5e-938a-eaf6d0aa2dec/volumes" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.848832 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc15f080-f90d-4e97-a866-d5edfd662bb7" path="/var/lib/kubelet/pods/bc15f080-f90d-4e97-a866-d5edfd662bb7/volumes" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.852983 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c2938e39-f1d8-4cdc-a32c-5d57b8f2034f" path="/var/lib/kubelet/pods/c2938e39-f1d8-4cdc-a32c-5d57b8f2034f/volumes" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.854522 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c49286b1-2e9f-4de6-b7de-2e952f3ad607" path="/var/lib/kubelet/pods/c49286b1-2e9f-4de6-b7de-2e952f3ad607/volumes" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.867724 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e8f6894c-b714-431e-912d-8b8de1a3914b" path="/var/lib/kubelet/pods/e8f6894c-b714-431e-912d-8b8de1a3914b/volumes" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.946435 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-85dc486cc4-fwrm6" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.058375 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3ee069a-41fb-4cab-b650-9bb9c03ac271-config-data\") pod \"e3ee069a-41fb-4cab-b650-9bb9c03ac271\" (UID: \"e3ee069a-41fb-4cab-b650-9bb9c03ac271\") " Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.058440 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3ee069a-41fb-4cab-b650-9bb9c03ac271-combined-ca-bundle\") pod \"e3ee069a-41fb-4cab-b650-9bb9c03ac271\" (UID: \"e3ee069a-41fb-4cab-b650-9bb9c03ac271\") " Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.058526 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gds2t\" (UniqueName: \"kubernetes.io/projected/e3ee069a-41fb-4cab-b650-9bb9c03ac271-kube-api-access-gds2t\") pod \"e3ee069a-41fb-4cab-b650-9bb9c03ac271\" (UID: \"e3ee069a-41fb-4cab-b650-9bb9c03ac271\") " Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.058586 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e3ee069a-41fb-4cab-b650-9bb9c03ac271-scripts\") pod \"e3ee069a-41fb-4cab-b650-9bb9c03ac271\" (UID: \"e3ee069a-41fb-4cab-b650-9bb9c03ac271\") " Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.058606 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e3ee069a-41fb-4cab-b650-9bb9c03ac271-internal-tls-certs\") pod \"e3ee069a-41fb-4cab-b650-9bb9c03ac271\" (UID: \"e3ee069a-41fb-4cab-b650-9bb9c03ac271\") " Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.058649 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e3ee069a-41fb-4cab-b650-9bb9c03ac271-public-tls-certs\") pod \"e3ee069a-41fb-4cab-b650-9bb9c03ac271\" (UID: \"e3ee069a-41fb-4cab-b650-9bb9c03ac271\") " Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.058695 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e3ee069a-41fb-4cab-b650-9bb9c03ac271-logs\") pod \"e3ee069a-41fb-4cab-b650-9bb9c03ac271\" (UID: \"e3ee069a-41fb-4cab-b650-9bb9c03ac271\") " Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.059698 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e3ee069a-41fb-4cab-b650-9bb9c03ac271-logs" (OuterVolumeSpecName: "logs") pod "e3ee069a-41fb-4cab-b650-9bb9c03ac271" (UID: "e3ee069a-41fb-4cab-b650-9bb9c03ac271"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.067437 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e3ee069a-41fb-4cab-b650-9bb9c03ac271-kube-api-access-gds2t" (OuterVolumeSpecName: "kube-api-access-gds2t") pod "e3ee069a-41fb-4cab-b650-9bb9c03ac271" (UID: "e3ee069a-41fb-4cab-b650-9bb9c03ac271"). InnerVolumeSpecName "kube-api-access-gds2t". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.069648 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3ee069a-41fb-4cab-b650-9bb9c03ac271-scripts" (OuterVolumeSpecName: "scripts") pod "e3ee069a-41fb-4cab-b650-9bb9c03ac271" (UID: "e3ee069a-41fb-4cab-b650-9bb9c03ac271"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.126919 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3ee069a-41fb-4cab-b650-9bb9c03ac271-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e3ee069a-41fb-4cab-b650-9bb9c03ac271" (UID: "e3ee069a-41fb-4cab-b650-9bb9c03ac271"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.160076 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3ee069a-41fb-4cab-b650-9bb9c03ac271-config-data" (OuterVolumeSpecName: "config-data") pod "e3ee069a-41fb-4cab-b650-9bb9c03ac271" (UID: "e3ee069a-41fb-4cab-b650-9bb9c03ac271"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.160447 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3ee069a-41fb-4cab-b650-9bb9c03ac271-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.160464 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3ee069a-41fb-4cab-b650-9bb9c03ac271-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.160474 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gds2t\" (UniqueName: \"kubernetes.io/projected/e3ee069a-41fb-4cab-b650-9bb9c03ac271-kube-api-access-gds2t\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.160484 4907 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e3ee069a-41fb-4cab-b650-9bb9c03ac271-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.160495 4907 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e3ee069a-41fb-4cab-b650-9bb9c03ac271-logs\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.164415 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3ee069a-41fb-4cab-b650-9bb9c03ac271-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "e3ee069a-41fb-4cab-b650-9bb9c03ac271" (UID: "e3ee069a-41fb-4cab-b650-9bb9c03ac271"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.181227 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3ee069a-41fb-4cab-b650-9bb9c03ac271-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "e3ee069a-41fb-4cab-b650-9bb9c03ac271" (UID: "e3ee069a-41fb-4cab-b650-9bb9c03ac271"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.184454 4907 generic.go:334] "Generic (PLEG): container finished" podID="4251ae36-90ad-41ea-915e-862df60f5c07" containerID="479ee0fb693df48bcf5ef06770a45535f793e4153c556058631f42347318a544" exitCode=0 Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.184551 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"4251ae36-90ad-41ea-915e-862df60f5c07","Type":"ContainerDied","Data":"479ee0fb693df48bcf5ef06770a45535f793e4153c556058631f42347318a544"} Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.191489 4907 generic.go:334] "Generic (PLEG): container finished" podID="cacb281b-ee3b-42b6-8754-d07e6ea6aa4b" containerID="65475a4b4006963b7fec1fa34278a44b7bf08cbcf9c088a764d7d29c870c630c" exitCode=0 Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.191573 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"cacb281b-ee3b-42b6-8754-d07e6ea6aa4b","Type":"ContainerDied","Data":"65475a4b4006963b7fec1fa34278a44b7bf08cbcf9c088a764d7d29c870c630c"} Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.191620 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"cacb281b-ee3b-42b6-8754-d07e6ea6aa4b","Type":"ContainerDied","Data":"c174ea292d3346503480cf55d66ad04b5cc26c89fac549263ee565359ea5a081"} Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.191632 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c174ea292d3346503480cf55d66ad04b5cc26c89fac549263ee565359ea5a081" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.194385 4907 generic.go:334] "Generic (PLEG): container finished" podID="2e41272f-6e7b-47a7-825e-3571f6a1fd07" containerID="ef8644527a89477971dc8c42d7fc7d2d08c1853a9324c6451f42196c1c50329e" exitCode=0 Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.194449 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2e41272f-6e7b-47a7-825e-3571f6a1fd07","Type":"ContainerDied","Data":"ef8644527a89477971dc8c42d7fc7d2d08c1853a9324c6451f42196c1c50329e"} Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.196217 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-bw2xb_6f8119b2-e38a-494d-967f-5198b83512c7/ovn-controller/0.log" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.196346 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-bw2xb" event={"ID":"6f8119b2-e38a-494d-967f-5198b83512c7","Type":"ContainerDied","Data":"4f06fa5373719502d5e6526674b7a5bdf213bd13582c27f9af62c582539db3b6"} Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.196436 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-bw2xb" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.196445 4907 scope.go:117] "RemoveContainer" containerID="df7f17b832ba26f166592b0696a663ded9a2e03678601f86699f09b76e050e58" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.201099 4907 generic.go:334] "Generic (PLEG): container finished" podID="0b2e62fa-af35-4f35-8852-362ea97c17e7" containerID="0d9941f62564b14687332e4dbc1f81a8df8a858860923a1b915fe67a744f75f5" exitCode=0 Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.201174 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"0b2e62fa-af35-4f35-8852-362ea97c17e7","Type":"ContainerDied","Data":"0d9941f62564b14687332e4dbc1f81a8df8a858860923a1b915fe67a744f75f5"} Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.203595 4907 generic.go:334] "Generic (PLEG): container finished" podID="41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76" containerID="a07507237e2303a227977bfc1acdae544607145ad994269ea3936668b9e9c650" exitCode=0 Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.203676 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76","Type":"ContainerDied","Data":"a07507237e2303a227977bfc1acdae544607145ad994269ea3936668b9e9c650"} Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.203730 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76","Type":"ContainerDied","Data":"10db10f108be05500948726431cb9db777e566b5230954b5ba74892f495e9236"} Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.203742 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="10db10f108be05500948726431cb9db777e566b5230954b5ba74892f495e9236" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.206044 4907 generic.go:334] "Generic (PLEG): container finished" podID="e3ee069a-41fb-4cab-b650-9bb9c03ac271" containerID="6d08d480e1be3a68c3d20d884244fa154245d26dcb3fb269a2aeb6806364407e" exitCode=0 Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.206099 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-85dc486cc4-fwrm6" event={"ID":"e3ee069a-41fb-4cab-b650-9bb9c03ac271","Type":"ContainerDied","Data":"6d08d480e1be3a68c3d20d884244fa154245d26dcb3fb269a2aeb6806364407e"} Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.206115 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-85dc486cc4-fwrm6" event={"ID":"e3ee069a-41fb-4cab-b650-9bb9c03ac271","Type":"ContainerDied","Data":"80317f4f67489584d7a194b5c3eced56f03150f3c1e4e1db7ac581e0ad05f710"} Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.206141 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-85dc486cc4-fwrm6" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.210520 4907 generic.go:334] "Generic (PLEG): container finished" podID="1b2f62d4-0584-4cc3-81ed-e067d8db23b1" containerID="22aa297a2f364f3884c53d61c34c94086546e7771e899b38ffebaf99570e4f07" exitCode=0 Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.210614 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-04a8-account-create-update-kntb5" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.210925 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7fc86597fb-mt9pk" event={"ID":"1b2f62d4-0584-4cc3-81ed-e067d8db23b1","Type":"ContainerDied","Data":"22aa297a2f364f3884c53d61c34c94086546e7771e899b38ffebaf99570e4f07"} Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.262858 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-48ngf\" (UniqueName: \"kubernetes.io/projected/4685651c-2690-48f8-97d2-8817fedaff56-kube-api-access-48ngf\") pod \"keystone-04a8-account-create-update-kntb5\" (UID: \"4685651c-2690-48f8-97d2-8817fedaff56\") " pod="openstack/keystone-04a8-account-create-update-kntb5" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.262936 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4685651c-2690-48f8-97d2-8817fedaff56-operator-scripts\") pod \"keystone-04a8-account-create-update-kntb5\" (UID: \"4685651c-2690-48f8-97d2-8817fedaff56\") " pod="openstack/keystone-04a8-account-create-update-kntb5" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.263164 4907 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e3ee069a-41fb-4cab-b650-9bb9c03ac271-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.263187 4907 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e3ee069a-41fb-4cab-b650-9bb9c03ac271-public-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:46 crc kubenswrapper[4907]: E0313 14:30:46.263251 4907 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Mar 13 14:30:46 crc kubenswrapper[4907]: E0313 14:30:46.263316 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/4685651c-2690-48f8-97d2-8817fedaff56-operator-scripts podName:4685651c-2690-48f8-97d2-8817fedaff56 nodeName:}" failed. No retries permitted until 2026-03-13 14:30:48.263300808 +0000 UTC m=+1547.163088497 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/4685651c-2690-48f8-97d2-8817fedaff56-operator-scripts") pod "keystone-04a8-account-create-update-kntb5" (UID: "4685651c-2690-48f8-97d2-8817fedaff56") : configmap "openstack-scripts" not found Mar 13 14:30:46 crc kubenswrapper[4907]: E0313 14:30:46.268574 4907 projected.go:194] Error preparing data for projected volume kube-api-access-48ngf for pod openstack/keystone-04a8-account-create-update-kntb5: failed to fetch token: serviceaccounts "galera-openstack" not found Mar 13 14:30:46 crc kubenswrapper[4907]: E0313 14:30:46.268690 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/4685651c-2690-48f8-97d2-8817fedaff56-kube-api-access-48ngf podName:4685651c-2690-48f8-97d2-8817fedaff56 nodeName:}" failed. No retries permitted until 2026-03-13 14:30:48.26867153 +0000 UTC m=+1547.168459219 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-48ngf" (UniqueName: "kubernetes.io/projected/4685651c-2690-48f8-97d2-8817fedaff56-kube-api-access-48ngf") pod "keystone-04a8-account-create-update-kntb5" (UID: "4685651c-2690-48f8-97d2-8817fedaff56") : failed to fetch token: serviceaccounts "galera-openstack" not found Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.345952 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.351577 4907 scope.go:117] "RemoveContainer" containerID="6d08d480e1be3a68c3d20d884244fa154245d26dcb3fb269a2aeb6806364407e" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.376036 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.398967 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-04a8-account-create-update-kntb5"] Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.400587 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.405361 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-04a8-account-create-update-kntb5"] Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.406025 4907 scope.go:117] "RemoveContainer" containerID="c0c472fe526cce04d2912530f4de0e17fa4ad4be341dbd2a462b13899fc1ad89" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.427722 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-bw2xb"] Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.436224 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.439213 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.444105 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-bw2xb"] Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.447187 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7fc86597fb-mt9pk" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.456971 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-85dc486cc4-fwrm6"] Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.459838 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-85dc486cc4-fwrm6"] Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.470522 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/cacb281b-ee3b-42b6-8754-d07e6ea6aa4b-kolla-config\") pod \"cacb281b-ee3b-42b6-8754-d07e6ea6aa4b\" (UID: \"cacb281b-ee3b-42b6-8754-d07e6ea6aa4b\") " Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.470581 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vdj7f\" (UniqueName: \"kubernetes.io/projected/cacb281b-ee3b-42b6-8754-d07e6ea6aa4b-kube-api-access-vdj7f\") pod \"cacb281b-ee3b-42b6-8754-d07e6ea6aa4b\" (UID: \"cacb281b-ee3b-42b6-8754-d07e6ea6aa4b\") " Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.470620 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76-public-tls-certs\") pod \"41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76\" (UID: \"41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76\") " Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.470650 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e41272f-6e7b-47a7-825e-3571f6a1fd07-combined-ca-bundle\") pod \"2e41272f-6e7b-47a7-825e-3571f6a1fd07\" (UID: \"2e41272f-6e7b-47a7-825e-3571f6a1fd07\") " Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.470678 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4251ae36-90ad-41ea-915e-862df60f5c07-httpd-run\") pod \"4251ae36-90ad-41ea-915e-862df60f5c07\" (UID: \"4251ae36-90ad-41ea-915e-862df60f5c07\") " Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.470733 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/cacb281b-ee3b-42b6-8754-d07e6ea6aa4b-memcached-tls-certs\") pod \"cacb281b-ee3b-42b6-8754-d07e6ea6aa4b\" (UID: \"cacb281b-ee3b-42b6-8754-d07e6ea6aa4b\") " Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.470773 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76-logs\") pod \"41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76\" (UID: \"41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76\") " Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.470807 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0b2e62fa-af35-4f35-8852-362ea97c17e7-public-tls-certs\") pod \"0b2e62fa-af35-4f35-8852-362ea97c17e7\" (UID: \"0b2e62fa-af35-4f35-8852-362ea97c17e7\") " Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.470829 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4251ae36-90ad-41ea-915e-862df60f5c07-logs\") pod \"4251ae36-90ad-41ea-915e-862df60f5c07\" (UID: \"4251ae36-90ad-41ea-915e-862df60f5c07\") " Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.470867 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2jswj\" (UniqueName: \"kubernetes.io/projected/0b2e62fa-af35-4f35-8852-362ea97c17e7-kube-api-access-2jswj\") pod \"0b2e62fa-af35-4f35-8852-362ea97c17e7\" (UID: \"0b2e62fa-af35-4f35-8852-362ea97c17e7\") " Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.470927 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4251ae36-90ad-41ea-915e-862df60f5c07-combined-ca-bundle\") pod \"4251ae36-90ad-41ea-915e-862df60f5c07\" (UID: \"4251ae36-90ad-41ea-915e-862df60f5c07\") " Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.470956 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76-combined-ca-bundle\") pod \"41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76\" (UID: \"41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76\") " Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.470996 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dnjx6\" (UniqueName: \"kubernetes.io/projected/41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76-kube-api-access-dnjx6\") pod \"41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76\" (UID: \"41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76\") " Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.471021 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76-etc-machine-id\") pod \"41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76\" (UID: \"41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76\") " Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.471044 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cacb281b-ee3b-42b6-8754-d07e6ea6aa4b-config-data\") pod \"cacb281b-ee3b-42b6-8754-d07e6ea6aa4b\" (UID: \"cacb281b-ee3b-42b6-8754-d07e6ea6aa4b\") " Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.471072 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0b2e62fa-af35-4f35-8852-362ea97c17e7-logs\") pod \"0b2e62fa-af35-4f35-8852-362ea97c17e7\" (UID: \"0b2e62fa-af35-4f35-8852-362ea97c17e7\") " Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.471096 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"0b2e62fa-af35-4f35-8852-362ea97c17e7\" (UID: \"0b2e62fa-af35-4f35-8852-362ea97c17e7\") " Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.471119 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b2e62fa-af35-4f35-8852-362ea97c17e7-config-data\") pod \"0b2e62fa-af35-4f35-8852-362ea97c17e7\" (UID: \"0b2e62fa-af35-4f35-8852-362ea97c17e7\") " Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.471152 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cacb281b-ee3b-42b6-8754-d07e6ea6aa4b-combined-ca-bundle\") pod \"cacb281b-ee3b-42b6-8754-d07e6ea6aa4b\" (UID: \"cacb281b-ee3b-42b6-8754-d07e6ea6aa4b\") " Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.471178 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76-config-data\") pod \"41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76\" (UID: \"41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76\") " Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.471210 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b2e62fa-af35-4f35-8852-362ea97c17e7-combined-ca-bundle\") pod \"0b2e62fa-af35-4f35-8852-362ea97c17e7\" (UID: \"0b2e62fa-af35-4f35-8852-362ea97c17e7\") " Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.471245 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4251ae36-90ad-41ea-915e-862df60f5c07-scripts\") pod \"4251ae36-90ad-41ea-915e-862df60f5c07\" (UID: \"4251ae36-90ad-41ea-915e-862df60f5c07\") " Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.471267 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76-config-data-custom\") pod \"41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76\" (UID: \"41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76\") " Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.471302 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h2bph\" (UniqueName: \"kubernetes.io/projected/4251ae36-90ad-41ea-915e-862df60f5c07-kube-api-access-h2bph\") pod \"4251ae36-90ad-41ea-915e-862df60f5c07\" (UID: \"4251ae36-90ad-41ea-915e-862df60f5c07\") " Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.471331 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4251ae36-90ad-41ea-915e-862df60f5c07-config-data\") pod \"4251ae36-90ad-41ea-915e-862df60f5c07\" (UID: \"4251ae36-90ad-41ea-915e-862df60f5c07\") " Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.471360 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e41272f-6e7b-47a7-825e-3571f6a1fd07-config-data\") pod \"2e41272f-6e7b-47a7-825e-3571f6a1fd07\" (UID: \"2e41272f-6e7b-47a7-825e-3571f6a1fd07\") " Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.471391 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4251ae36-90ad-41ea-915e-862df60f5c07-internal-tls-certs\") pod \"4251ae36-90ad-41ea-915e-862df60f5c07\" (UID: \"4251ae36-90ad-41ea-915e-862df60f5c07\") " Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.471440 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2e41272f-6e7b-47a7-825e-3571f6a1fd07-logs\") pod \"2e41272f-6e7b-47a7-825e-3571f6a1fd07\" (UID: \"2e41272f-6e7b-47a7-825e-3571f6a1fd07\") " Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.471461 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76-scripts\") pod \"41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76\" (UID: \"41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76\") " Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.471489 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76-internal-tls-certs\") pod \"41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76\" (UID: \"41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76\") " Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.471531 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8rvmk\" (UniqueName: \"kubernetes.io/projected/2e41272f-6e7b-47a7-825e-3571f6a1fd07-kube-api-access-8rvmk\") pod \"2e41272f-6e7b-47a7-825e-3571f6a1fd07\" (UID: \"2e41272f-6e7b-47a7-825e-3571f6a1fd07\") " Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.471570 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0b2e62fa-af35-4f35-8852-362ea97c17e7-scripts\") pod \"0b2e62fa-af35-4f35-8852-362ea97c17e7\" (UID: \"0b2e62fa-af35-4f35-8852-362ea97c17e7\") " Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.471618 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e41272f-6e7b-47a7-825e-3571f6a1fd07-nova-metadata-tls-certs\") pod \"2e41272f-6e7b-47a7-825e-3571f6a1fd07\" (UID: \"2e41272f-6e7b-47a7-825e-3571f6a1fd07\") " Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.471642 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"4251ae36-90ad-41ea-915e-862df60f5c07\" (UID: \"4251ae36-90ad-41ea-915e-862df60f5c07\") " Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.471663 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0b2e62fa-af35-4f35-8852-362ea97c17e7-httpd-run\") pod \"0b2e62fa-af35-4f35-8852-362ea97c17e7\" (UID: \"0b2e62fa-af35-4f35-8852-362ea97c17e7\") " Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.472048 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-48ngf\" (UniqueName: \"kubernetes.io/projected/4685651c-2690-48f8-97d2-8817fedaff56-kube-api-access-48ngf\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.472076 4907 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4685651c-2690-48f8-97d2-8817fedaff56-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.473016 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0b2e62fa-af35-4f35-8852-362ea97c17e7-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "0b2e62fa-af35-4f35-8852-362ea97c17e7" (UID: "0b2e62fa-af35-4f35-8852-362ea97c17e7"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.476140 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2e41272f-6e7b-47a7-825e-3571f6a1fd07-logs" (OuterVolumeSpecName: "logs") pod "2e41272f-6e7b-47a7-825e-3571f6a1fd07" (UID: "2e41272f-6e7b-47a7-825e-3571f6a1fd07"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.483504 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2e41272f-6e7b-47a7-825e-3571f6a1fd07-kube-api-access-8rvmk" (OuterVolumeSpecName: "kube-api-access-8rvmk") pod "2e41272f-6e7b-47a7-825e-3571f6a1fd07" (UID: "2e41272f-6e7b-47a7-825e-3571f6a1fd07"). InnerVolumeSpecName "kube-api-access-8rvmk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.488848 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76" (UID: "41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.493930 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4251ae36-90ad-41ea-915e-862df60f5c07-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "4251ae36-90ad-41ea-915e-862df60f5c07" (UID: "4251ae36-90ad-41ea-915e-862df60f5c07"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.499347 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cacb281b-ee3b-42b6-8754-d07e6ea6aa4b-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "cacb281b-ee3b-42b6-8754-d07e6ea6aa4b" (UID: "cacb281b-ee3b-42b6-8754-d07e6ea6aa4b"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.499466 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cacb281b-ee3b-42b6-8754-d07e6ea6aa4b-kube-api-access-vdj7f" (OuterVolumeSpecName: "kube-api-access-vdj7f") pod "cacb281b-ee3b-42b6-8754-d07e6ea6aa4b" (UID: "cacb281b-ee3b-42b6-8754-d07e6ea6aa4b"). InnerVolumeSpecName "kube-api-access-vdj7f". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.501391 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76-logs" (OuterVolumeSpecName: "logs") pod "41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76" (UID: "41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.505135 4907 scope.go:117] "RemoveContainer" containerID="6d08d480e1be3a68c3d20d884244fa154245d26dcb3fb269a2aeb6806364407e" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.507101 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4251ae36-90ad-41ea-915e-862df60f5c07-scripts" (OuterVolumeSpecName: "scripts") pod "4251ae36-90ad-41ea-915e-862df60f5c07" (UID: "4251ae36-90ad-41ea-915e-862df60f5c07"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.508478 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cacb281b-ee3b-42b6-8754-d07e6ea6aa4b-config-data" (OuterVolumeSpecName: "config-data") pod "cacb281b-ee3b-42b6-8754-d07e6ea6aa4b" (UID: "cacb281b-ee3b-42b6-8754-d07e6ea6aa4b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.508519 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76" (UID: "41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.510131 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4251ae36-90ad-41ea-915e-862df60f5c07-logs" (OuterVolumeSpecName: "logs") pod "4251ae36-90ad-41ea-915e-862df60f5c07" (UID: "4251ae36-90ad-41ea-915e-862df60f5c07"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.510233 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "glance") pod "0b2e62fa-af35-4f35-8852-362ea97c17e7" (UID: "0b2e62fa-af35-4f35-8852-362ea97c17e7"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.511596 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0b2e62fa-af35-4f35-8852-362ea97c17e7-logs" (OuterVolumeSpecName: "logs") pod "0b2e62fa-af35-4f35-8852-362ea97c17e7" (UID: "0b2e62fa-af35-4f35-8852-362ea97c17e7"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.511721 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76-scripts" (OuterVolumeSpecName: "scripts") pod "41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76" (UID: "41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:46 crc kubenswrapper[4907]: E0313 14:30:46.511922 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6d08d480e1be3a68c3d20d884244fa154245d26dcb3fb269a2aeb6806364407e\": container with ID starting with 6d08d480e1be3a68c3d20d884244fa154245d26dcb3fb269a2aeb6806364407e not found: ID does not exist" containerID="6d08d480e1be3a68c3d20d884244fa154245d26dcb3fb269a2aeb6806364407e" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.511994 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6d08d480e1be3a68c3d20d884244fa154245d26dcb3fb269a2aeb6806364407e"} err="failed to get container status \"6d08d480e1be3a68c3d20d884244fa154245d26dcb3fb269a2aeb6806364407e\": rpc error: code = NotFound desc = could not find container \"6d08d480e1be3a68c3d20d884244fa154245d26dcb3fb269a2aeb6806364407e\": container with ID starting with 6d08d480e1be3a68c3d20d884244fa154245d26dcb3fb269a2aeb6806364407e not found: ID does not exist" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.512033 4907 scope.go:117] "RemoveContainer" containerID="c0c472fe526cce04d2912530f4de0e17fa4ad4be341dbd2a462b13899fc1ad89" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.513218 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b2e62fa-af35-4f35-8852-362ea97c17e7-kube-api-access-2jswj" (OuterVolumeSpecName: "kube-api-access-2jswj") pod "0b2e62fa-af35-4f35-8852-362ea97c17e7" (UID: "0b2e62fa-af35-4f35-8852-362ea97c17e7"). InnerVolumeSpecName "kube-api-access-2jswj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.514100 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4251ae36-90ad-41ea-915e-862df60f5c07-kube-api-access-h2bph" (OuterVolumeSpecName: "kube-api-access-h2bph") pod "4251ae36-90ad-41ea-915e-862df60f5c07" (UID: "4251ae36-90ad-41ea-915e-862df60f5c07"). InnerVolumeSpecName "kube-api-access-h2bph". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.514107 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b2e62fa-af35-4f35-8852-362ea97c17e7-scripts" (OuterVolumeSpecName: "scripts") pod "0b2e62fa-af35-4f35-8852-362ea97c17e7" (UID: "0b2e62fa-af35-4f35-8852-362ea97c17e7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.514182 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "glance") pod "4251ae36-90ad-41ea-915e-862df60f5c07" (UID: "4251ae36-90ad-41ea-915e-862df60f5c07"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Mar 13 14:30:46 crc kubenswrapper[4907]: E0313 14:30:46.516266 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c0c472fe526cce04d2912530f4de0e17fa4ad4be341dbd2a462b13899fc1ad89\": container with ID starting with c0c472fe526cce04d2912530f4de0e17fa4ad4be341dbd2a462b13899fc1ad89 not found: ID does not exist" containerID="c0c472fe526cce04d2912530f4de0e17fa4ad4be341dbd2a462b13899fc1ad89" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.516302 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c0c472fe526cce04d2912530f4de0e17fa4ad4be341dbd2a462b13899fc1ad89"} err="failed to get container status \"c0c472fe526cce04d2912530f4de0e17fa4ad4be341dbd2a462b13899fc1ad89\": rpc error: code = NotFound desc = could not find container \"c0c472fe526cce04d2912530f4de0e17fa4ad4be341dbd2a462b13899fc1ad89\": container with ID starting with c0c472fe526cce04d2912530f4de0e17fa4ad4be341dbd2a462b13899fc1ad89 not found: ID does not exist" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.518593 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76-kube-api-access-dnjx6" (OuterVolumeSpecName: "kube-api-access-dnjx6") pod "41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76" (UID: "41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76"). InnerVolumeSpecName "kube-api-access-dnjx6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.554794 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b2e62fa-af35-4f35-8852-362ea97c17e7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0b2e62fa-af35-4f35-8852-362ea97c17e7" (UID: "0b2e62fa-af35-4f35-8852-362ea97c17e7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.565004 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cacb281b-ee3b-42b6-8754-d07e6ea6aa4b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cacb281b-ee3b-42b6-8754-d07e6ea6aa4b" (UID: "cacb281b-ee3b-42b6-8754-d07e6ea6aa4b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.575442 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1b2f62d4-0584-4cc3-81ed-e067d8db23b1-logs\") pod \"1b2f62d4-0584-4cc3-81ed-e067d8db23b1\" (UID: \"1b2f62d4-0584-4cc3-81ed-e067d8db23b1\") " Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.575484 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b2f62d4-0584-4cc3-81ed-e067d8db23b1-config-data\") pod \"1b2f62d4-0584-4cc3-81ed-e067d8db23b1\" (UID: \"1b2f62d4-0584-4cc3-81ed-e067d8db23b1\") " Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.575532 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1b2f62d4-0584-4cc3-81ed-e067d8db23b1-internal-tls-certs\") pod \"1b2f62d4-0584-4cc3-81ed-e067d8db23b1\" (UID: \"1b2f62d4-0584-4cc3-81ed-e067d8db23b1\") " Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.575631 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1b2f62d4-0584-4cc3-81ed-e067d8db23b1-public-tls-certs\") pod \"1b2f62d4-0584-4cc3-81ed-e067d8db23b1\" (UID: \"1b2f62d4-0584-4cc3-81ed-e067d8db23b1\") " Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.575719 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ktf7c\" (UniqueName: \"kubernetes.io/projected/1b2f62d4-0584-4cc3-81ed-e067d8db23b1-kube-api-access-ktf7c\") pod \"1b2f62d4-0584-4cc3-81ed-e067d8db23b1\" (UID: \"1b2f62d4-0584-4cc3-81ed-e067d8db23b1\") " Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.575737 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1b2f62d4-0584-4cc3-81ed-e067d8db23b1-config-data-custom\") pod \"1b2f62d4-0584-4cc3-81ed-e067d8db23b1\" (UID: \"1b2f62d4-0584-4cc3-81ed-e067d8db23b1\") " Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.575760 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b2f62d4-0584-4cc3-81ed-e067d8db23b1-combined-ca-bundle\") pod \"1b2f62d4-0584-4cc3-81ed-e067d8db23b1\" (UID: \"1b2f62d4-0584-4cc3-81ed-e067d8db23b1\") " Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.576140 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b2e62fa-af35-4f35-8852-362ea97c17e7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.576157 4907 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4251ae36-90ad-41ea-915e-862df60f5c07-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.576166 4907 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76-config-data-custom\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.576176 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h2bph\" (UniqueName: \"kubernetes.io/projected/4251ae36-90ad-41ea-915e-862df60f5c07-kube-api-access-h2bph\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.576188 4907 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2e41272f-6e7b-47a7-825e-3571f6a1fd07-logs\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.576197 4907 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.576208 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8rvmk\" (UniqueName: \"kubernetes.io/projected/2e41272f-6e7b-47a7-825e-3571f6a1fd07-kube-api-access-8rvmk\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.576217 4907 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0b2e62fa-af35-4f35-8852-362ea97c17e7-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.576239 4907 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.576250 4907 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0b2e62fa-af35-4f35-8852-362ea97c17e7-httpd-run\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.576260 4907 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/cacb281b-ee3b-42b6-8754-d07e6ea6aa4b-kolla-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.576270 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vdj7f\" (UniqueName: \"kubernetes.io/projected/cacb281b-ee3b-42b6-8754-d07e6ea6aa4b-kube-api-access-vdj7f\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.576278 4907 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4251ae36-90ad-41ea-915e-862df60f5c07-httpd-run\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.576286 4907 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76-logs\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.576296 4907 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4251ae36-90ad-41ea-915e-862df60f5c07-logs\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.576304 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2jswj\" (UniqueName: \"kubernetes.io/projected/0b2e62fa-af35-4f35-8852-362ea97c17e7-kube-api-access-2jswj\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.576312 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dnjx6\" (UniqueName: \"kubernetes.io/projected/41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76-kube-api-access-dnjx6\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.576321 4907 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76-etc-machine-id\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.576329 4907 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0b2e62fa-af35-4f35-8852-362ea97c17e7-logs\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.576337 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cacb281b-ee3b-42b6-8754-d07e6ea6aa4b-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.576350 4907 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.576358 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cacb281b-ee3b-42b6-8754-d07e6ea6aa4b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.581999 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1b2f62d4-0584-4cc3-81ed-e067d8db23b1-logs" (OuterVolumeSpecName: "logs") pod "1b2f62d4-0584-4cc3-81ed-e067d8db23b1" (UID: "1b2f62d4-0584-4cc3-81ed-e067d8db23b1"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.587694 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1b2f62d4-0584-4cc3-81ed-e067d8db23b1-kube-api-access-ktf7c" (OuterVolumeSpecName: "kube-api-access-ktf7c") pod "1b2f62d4-0584-4cc3-81ed-e067d8db23b1" (UID: "1b2f62d4-0584-4cc3-81ed-e067d8db23b1"). InnerVolumeSpecName "kube-api-access-ktf7c". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.605307 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b2f62d4-0584-4cc3-81ed-e067d8db23b1-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "1b2f62d4-0584-4cc3-81ed-e067d8db23b1" (UID: "1b2f62d4-0584-4cc3-81ed-e067d8db23b1"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.614702 4907 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.639397 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76" (UID: "41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.640970 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e41272f-6e7b-47a7-825e-3571f6a1fd07-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2e41272f-6e7b-47a7-825e-3571f6a1fd07" (UID: "2e41272f-6e7b-47a7-825e-3571f6a1fd07"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.651105 4907 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.678666 4907 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.678970 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e41272f-6e7b-47a7-825e-3571f6a1fd07-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.687008 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.687201 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ktf7c\" (UniqueName: \"kubernetes.io/projected/1b2f62d4-0584-4cc3-81ed-e067d8db23b1-kube-api-access-ktf7c\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.687299 4907 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1b2f62d4-0584-4cc3-81ed-e067d8db23b1-config-data-custom\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.687365 4907 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.687422 4907 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1b2f62d4-0584-4cc3-81ed-e067d8db23b1-logs\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.705124 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e41272f-6e7b-47a7-825e-3571f6a1fd07-config-data" (OuterVolumeSpecName: "config-data") pod "2e41272f-6e7b-47a7-825e-3571f6a1fd07" (UID: "2e41272f-6e7b-47a7-825e-3571f6a1fd07"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.710317 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76" (UID: "41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.714418 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.725891 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4251ae36-90ad-41ea-915e-862df60f5c07-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "4251ae36-90ad-41ea-915e-862df60f5c07" (UID: "4251ae36-90ad-41ea-915e-862df60f5c07"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.754776 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b2f62d4-0584-4cc3-81ed-e067d8db23b1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1b2f62d4-0584-4cc3-81ed-e067d8db23b1" (UID: "1b2f62d4-0584-4cc3-81ed-e067d8db23b1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.758338 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b2e62fa-af35-4f35-8852-362ea97c17e7-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "0b2e62fa-af35-4f35-8852-362ea97c17e7" (UID: "0b2e62fa-af35-4f35-8852-362ea97c17e7"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.759129 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b2e62fa-af35-4f35-8852-362ea97c17e7-config-data" (OuterVolumeSpecName: "config-data") pod "0b2e62fa-af35-4f35-8852-362ea97c17e7" (UID: "0b2e62fa-af35-4f35-8852-362ea97c17e7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.768960 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b2f62d4-0584-4cc3-81ed-e067d8db23b1-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "1b2f62d4-0584-4cc3-81ed-e067d8db23b1" (UID: "1b2f62d4-0584-4cc3-81ed-e067d8db23b1"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.773487 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e41272f-6e7b-47a7-825e-3571f6a1fd07-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "2e41272f-6e7b-47a7-825e-3571f6a1fd07" (UID: "2e41272f-6e7b-47a7-825e-3571f6a1fd07"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.777040 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4251ae36-90ad-41ea-915e-862df60f5c07-config-data" (OuterVolumeSpecName: "config-data") pod "4251ae36-90ad-41ea-915e-862df60f5c07" (UID: "4251ae36-90ad-41ea-915e-862df60f5c07"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.788049 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4251ae36-90ad-41ea-915e-862df60f5c07-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4251ae36-90ad-41ea-915e-862df60f5c07" (UID: "4251ae36-90ad-41ea-915e-862df60f5c07"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.788616 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cc227d33-3bf1-4b5f-8368-8e6b8a512f7d-public-tls-certs\") pod \"cc227d33-3bf1-4b5f-8368-8e6b8a512f7d\" (UID: \"cc227d33-3bf1-4b5f-8368-8e6b8a512f7d\") " Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.788655 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc227d33-3bf1-4b5f-8368-8e6b8a512f7d-config-data\") pod \"cc227d33-3bf1-4b5f-8368-8e6b8a512f7d\" (UID: \"cc227d33-3bf1-4b5f-8368-8e6b8a512f7d\") " Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.790059 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76" (UID: "41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.791092 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cc227d33-3bf1-4b5f-8368-8e6b8a512f7d-internal-tls-certs\") pod \"cc227d33-3bf1-4b5f-8368-8e6b8a512f7d\" (UID: \"cc227d33-3bf1-4b5f-8368-8e6b8a512f7d\") " Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.791153 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-565tg\" (UniqueName: \"kubernetes.io/projected/cc227d33-3bf1-4b5f-8368-8e6b8a512f7d-kube-api-access-565tg\") pod \"cc227d33-3bf1-4b5f-8368-8e6b8a512f7d\" (UID: \"cc227d33-3bf1-4b5f-8368-8e6b8a512f7d\") " Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.791176 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76-public-tls-certs\") pod \"41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76\" (UID: \"41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76\") " Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.791208 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cc227d33-3bf1-4b5f-8368-8e6b8a512f7d-logs\") pod \"cc227d33-3bf1-4b5f-8368-8e6b8a512f7d\" (UID: \"cc227d33-3bf1-4b5f-8368-8e6b8a512f7d\") " Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.791275 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc227d33-3bf1-4b5f-8368-8e6b8a512f7d-combined-ca-bundle\") pod \"cc227d33-3bf1-4b5f-8368-8e6b8a512f7d\" (UID: \"cc227d33-3bf1-4b5f-8368-8e6b8a512f7d\") " Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.791970 4907 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0b2e62fa-af35-4f35-8852-362ea97c17e7-public-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.791994 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4251ae36-90ad-41ea-915e-862df60f5c07-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.792007 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b2e62fa-af35-4f35-8852-362ea97c17e7-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.792019 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b2f62d4-0584-4cc3-81ed-e067d8db23b1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.792030 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4251ae36-90ad-41ea-915e-862df60f5c07-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.792041 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e41272f-6e7b-47a7-825e-3571f6a1fd07-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.792051 4907 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4251ae36-90ad-41ea-915e-862df60f5c07-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.792062 4907 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.792076 4907 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1b2f62d4-0584-4cc3-81ed-e067d8db23b1-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.792087 4907 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e41272f-6e7b-47a7-825e-3571f6a1fd07-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.792659 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76-config-data" (OuterVolumeSpecName: "config-data") pod "41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76" (UID: "41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:46 crc kubenswrapper[4907]: W0313 14:30:46.794066 4907 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76/volumes/kubernetes.io~secret/public-tls-certs Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.794110 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76" (UID: "41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.794299 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cc227d33-3bf1-4b5f-8368-8e6b8a512f7d-logs" (OuterVolumeSpecName: "logs") pod "cc227d33-3bf1-4b5f-8368-8e6b8a512f7d" (UID: "cc227d33-3bf1-4b5f-8368-8e6b8a512f7d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.796078 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc227d33-3bf1-4b5f-8368-8e6b8a512f7d-kube-api-access-565tg" (OuterVolumeSpecName: "kube-api-access-565tg") pod "cc227d33-3bf1-4b5f-8368-8e6b8a512f7d" (UID: "cc227d33-3bf1-4b5f-8368-8e6b8a512f7d"). InnerVolumeSpecName "kube-api-access-565tg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.802608 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b2f62d4-0584-4cc3-81ed-e067d8db23b1-config-data" (OuterVolumeSpecName: "config-data") pod "1b2f62d4-0584-4cc3-81ed-e067d8db23b1" (UID: "1b2f62d4-0584-4cc3-81ed-e067d8db23b1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.805642 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b2f62d4-0584-4cc3-81ed-e067d8db23b1-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "1b2f62d4-0584-4cc3-81ed-e067d8db23b1" (UID: "1b2f62d4-0584-4cc3-81ed-e067d8db23b1"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.813646 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cacb281b-ee3b-42b6-8754-d07e6ea6aa4b-memcached-tls-certs" (OuterVolumeSpecName: "memcached-tls-certs") pod "cacb281b-ee3b-42b6-8754-d07e6ea6aa4b" (UID: "cacb281b-ee3b-42b6-8754-d07e6ea6aa4b"). InnerVolumeSpecName "memcached-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.816606 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc227d33-3bf1-4b5f-8368-8e6b8a512f7d-config-data" (OuterVolumeSpecName: "config-data") pod "cc227d33-3bf1-4b5f-8368-8e6b8a512f7d" (UID: "cc227d33-3bf1-4b5f-8368-8e6b8a512f7d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.818755 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc227d33-3bf1-4b5f-8368-8e6b8a512f7d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cc227d33-3bf1-4b5f-8368-8e6b8a512f7d" (UID: "cc227d33-3bf1-4b5f-8368-8e6b8a512f7d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.838714 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc227d33-3bf1-4b5f-8368-8e6b8a512f7d-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "cc227d33-3bf1-4b5f-8368-8e6b8a512f7d" (UID: "cc227d33-3bf1-4b5f-8368-8e6b8a512f7d"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.877919 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc227d33-3bf1-4b5f-8368-8e6b8a512f7d-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "cc227d33-3bf1-4b5f-8368-8e6b8a512f7d" (UID: "cc227d33-3bf1-4b5f-8368-8e6b8a512f7d"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.893906 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.893930 4907 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cc227d33-3bf1-4b5f-8368-8e6b8a512f7d-public-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.893942 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc227d33-3bf1-4b5f-8368-8e6b8a512f7d-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.893952 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b2f62d4-0584-4cc3-81ed-e067d8db23b1-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.893960 4907 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cc227d33-3bf1-4b5f-8368-8e6b8a512f7d-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.893968 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-565tg\" (UniqueName: \"kubernetes.io/projected/cc227d33-3bf1-4b5f-8368-8e6b8a512f7d-kube-api-access-565tg\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.893977 4907 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76-public-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.893985 4907 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cc227d33-3bf1-4b5f-8368-8e6b8a512f7d-logs\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.893993 4907 reconciler_common.go:293] "Volume detached for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/cacb281b-ee3b-42b6-8754-d07e6ea6aa4b-memcached-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.894001 4907 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1b2f62d4-0584-4cc3-81ed-e067d8db23b1-public-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.894009 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc227d33-3bf1-4b5f-8368-8e6b8a512f7d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.165209 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-7d7bc9f7d-fks76" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.233028 4907 generic.go:334] "Generic (PLEG): container finished" podID="65531461-cea2-4b2c-a9b6-8fd9e9bffb27" containerID="7117af57f620e2d772f6e2ca6d92edfdc283b3eb640ab414d9483ba1758f36d4" exitCode=0 Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.233100 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7d7bc9f7d-fks76" event={"ID":"65531461-cea2-4b2c-a9b6-8fd9e9bffb27","Type":"ContainerDied","Data":"7117af57f620e2d772f6e2ca6d92edfdc283b3eb640ab414d9483ba1758f36d4"} Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.233128 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7d7bc9f7d-fks76" event={"ID":"65531461-cea2-4b2c-a9b6-8fd9e9bffb27","Type":"ContainerDied","Data":"d83bd2c5278d976f4863dc597e16a40e27298e4624048379804b2a946c5145cc"} Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.233146 4907 scope.go:117] "RemoveContainer" containerID="7117af57f620e2d772f6e2ca6d92edfdc283b3eb640ab414d9483ba1758f36d4" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.233261 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-7d7bc9f7d-fks76" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.255721 4907 generic.go:334] "Generic (PLEG): container finished" podID="c4411bd2-c555-433a-9015-f623948b1401" containerID="b5e68fc91f35838327b362dd7ecd6d2f5ee1173bc7ae4fafcb541ac0c96b44fa" exitCode=0 Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.255786 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"c4411bd2-c555-433a-9015-f623948b1401","Type":"ContainerDied","Data":"b5e68fc91f35838327b362dd7ecd6d2f5ee1173bc7ae4fafcb541ac0c96b44fa"} Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.255831 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"c4411bd2-c555-433a-9015-f623948b1401","Type":"ContainerDied","Data":"da35a91fc58ce0787798d501f92eec9bbafae6711281a9caca5018e3a65c78c8"} Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.255846 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="da35a91fc58ce0787798d501f92eec9bbafae6711281a9caca5018e3a65c78c8" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.266265 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2e41272f-6e7b-47a7-825e-3571f6a1fd07","Type":"ContainerDied","Data":"68545429a61b864d7541170bf482dacf433239060ec793e8fb62a4bc2ee9a2f3"} Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.266376 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.302660 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz7hb\" (UniqueName: \"kubernetes.io/projected/65531461-cea2-4b2c-a9b6-8fd9e9bffb27-kube-api-access-lz7hb\") pod \"65531461-cea2-4b2c-a9b6-8fd9e9bffb27\" (UID: \"65531461-cea2-4b2c-a9b6-8fd9e9bffb27\") " Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.302732 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/65531461-cea2-4b2c-a9b6-8fd9e9bffb27-config-data-custom\") pod \"65531461-cea2-4b2c-a9b6-8fd9e9bffb27\" (UID: \"65531461-cea2-4b2c-a9b6-8fd9e9bffb27\") " Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.302798 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65531461-cea2-4b2c-a9b6-8fd9e9bffb27-combined-ca-bundle\") pod \"65531461-cea2-4b2c-a9b6-8fd9e9bffb27\" (UID: \"65531461-cea2-4b2c-a9b6-8fd9e9bffb27\") " Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.302938 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65531461-cea2-4b2c-a9b6-8fd9e9bffb27-config-data\") pod \"65531461-cea2-4b2c-a9b6-8fd9e9bffb27\" (UID: \"65531461-cea2-4b2c-a9b6-8fd9e9bffb27\") " Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.302959 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/65531461-cea2-4b2c-a9b6-8fd9e9bffb27-logs\") pod \"65531461-cea2-4b2c-a9b6-8fd9e9bffb27\" (UID: \"65531461-cea2-4b2c-a9b6-8fd9e9bffb27\") " Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.303566 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/65531461-cea2-4b2c-a9b6-8fd9e9bffb27-logs" (OuterVolumeSpecName: "logs") pod "65531461-cea2-4b2c-a9b6-8fd9e9bffb27" (UID: "65531461-cea2-4b2c-a9b6-8fd9e9bffb27"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.331245 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/65531461-cea2-4b2c-a9b6-8fd9e9bffb27-kube-api-access-lz7hb" (OuterVolumeSpecName: "kube-api-access-lz7hb") pod "65531461-cea2-4b2c-a9b6-8fd9e9bffb27" (UID: "65531461-cea2-4b2c-a9b6-8fd9e9bffb27"). InnerVolumeSpecName "kube-api-access-lz7hb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.331793 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"0b2e62fa-af35-4f35-8852-362ea97c17e7","Type":"ContainerDied","Data":"0ede32ea598e7a7a739c54b190ba597842e8cec1fea7546f52eefbb2b1fa0f83"} Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.331968 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.359939 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65531461-cea2-4b2c-a9b6-8fd9e9bffb27-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "65531461-cea2-4b2c-a9b6-8fd9e9bffb27" (UID: "65531461-cea2-4b2c-a9b6-8fd9e9bffb27"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.360519 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.364088 4907 generic.go:334] "Generic (PLEG): container finished" podID="cc227d33-3bf1-4b5f-8368-8e6b8a512f7d" containerID="58121d89217f995e7e932419445b98334a3ccb8b202e18c097026d64131a91fd" exitCode=0 Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.364170 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"cc227d33-3bf1-4b5f-8368-8e6b8a512f7d","Type":"ContainerDied","Data":"58121d89217f995e7e932419445b98334a3ccb8b202e18c097026d64131a91fd"} Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.364196 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"cc227d33-3bf1-4b5f-8368-8e6b8a512f7d","Type":"ContainerDied","Data":"ed58fd3bb4c7be310b9f67af6133d99b09325a11963ab01983393f4f0a20f895"} Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.364260 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.367188 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65531461-cea2-4b2c-a9b6-8fd9e9bffb27-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "65531461-cea2-4b2c-a9b6-8fd9e9bffb27" (UID: "65531461-cea2-4b2c-a9b6-8fd9e9bffb27"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.377966 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"4251ae36-90ad-41ea-915e-862df60f5c07","Type":"ContainerDied","Data":"d3f2a65d466df300e932d28e3a4a5fb0f68b272bc80ae37edb5a080aa7144cae"} Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.378077 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.383200 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.383268 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7fc86597fb-mt9pk" event={"ID":"1b2f62d4-0584-4cc3-81ed-e067d8db23b1","Type":"ContainerDied","Data":"97dce4dda51605af3970178d1030468b7ba5970d94173d175f92a65d74507fff"} Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.383392 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7fc86597fb-mt9pk" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.383895 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.385466 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.403871 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/c4411bd2-c555-433a-9015-f623948b1401-galera-tls-certs\") pod \"c4411bd2-c555-433a-9015-f623948b1401\" (UID: \"c4411bd2-c555-433a-9015-f623948b1401\") " Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.404031 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.404044 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/c4411bd2-c555-433a-9015-f623948b1401-config-data-default\") pod \"c4411bd2-c555-433a-9015-f623948b1401\" (UID: \"c4411bd2-c555-433a-9015-f623948b1401\") " Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.404147 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c4411bd2-c555-433a-9015-f623948b1401-operator-scripts\") pod \"c4411bd2-c555-433a-9015-f623948b1401\" (UID: \"c4411bd2-c555-433a-9015-f623948b1401\") " Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.404203 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mysql-db\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"c4411bd2-c555-433a-9015-f623948b1401\" (UID: \"c4411bd2-c555-433a-9015-f623948b1401\") " Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.404688 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c4411bd2-c555-433a-9015-f623948b1401-config-data-default" (OuterVolumeSpecName: "config-data-default") pod "c4411bd2-c555-433a-9015-f623948b1401" (UID: "c4411bd2-c555-433a-9015-f623948b1401"). InnerVolumeSpecName "config-data-default". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.405329 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c4411bd2-c555-433a-9015-f623948b1401-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c4411bd2-c555-433a-9015-f623948b1401" (UID: "c4411bd2-c555-433a-9015-f623948b1401"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.411602 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65531461-cea2-4b2c-a9b6-8fd9e9bffb27-config-data" (OuterVolumeSpecName: "config-data") pod "65531461-cea2-4b2c-a9b6-8fd9e9bffb27" (UID: "65531461-cea2-4b2c-a9b6-8fd9e9bffb27"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.412217 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcwkz\" (UniqueName: \"kubernetes.io/projected/c4411bd2-c555-433a-9015-f623948b1401-kube-api-access-pcwkz\") pod \"c4411bd2-c555-433a-9015-f623948b1401\" (UID: \"c4411bd2-c555-433a-9015-f623948b1401\") " Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.412245 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4411bd2-c555-433a-9015-f623948b1401-combined-ca-bundle\") pod \"c4411bd2-c555-433a-9015-f623948b1401\" (UID: \"c4411bd2-c555-433a-9015-f623948b1401\") " Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.412275 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/c4411bd2-c555-433a-9015-f623948b1401-kolla-config\") pod \"c4411bd2-c555-433a-9015-f623948b1401\" (UID: \"c4411bd2-c555-433a-9015-f623948b1401\") " Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.412294 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/c4411bd2-c555-433a-9015-f623948b1401-config-data-generated\") pod \"c4411bd2-c555-433a-9015-f623948b1401\" (UID: \"c4411bd2-c555-433a-9015-f623948b1401\") " Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.412912 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c4411bd2-c555-433a-9015-f623948b1401-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "c4411bd2-c555-433a-9015-f623948b1401" (UID: "c4411bd2-c555-433a-9015-f623948b1401"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.413832 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c4411bd2-c555-433a-9015-f623948b1401-config-data-generated" (OuterVolumeSpecName: "config-data-generated") pod "c4411bd2-c555-433a-9015-f623948b1401" (UID: "c4411bd2-c555-433a-9015-f623948b1401"). InnerVolumeSpecName "config-data-generated". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.414324 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65531461-cea2-4b2c-a9b6-8fd9e9bffb27-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.414353 4907 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/65531461-cea2-4b2c-a9b6-8fd9e9bffb27-logs\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.414365 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz7hb\" (UniqueName: \"kubernetes.io/projected/65531461-cea2-4b2c-a9b6-8fd9e9bffb27-kube-api-access-lz7hb\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.414380 4907 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/65531461-cea2-4b2c-a9b6-8fd9e9bffb27-config-data-custom\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.414390 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65531461-cea2-4b2c-a9b6-8fd9e9bffb27-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.414399 4907 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/c4411bd2-c555-433a-9015-f623948b1401-kolla-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.414408 4907 reconciler_common.go:293] "Volume detached for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/c4411bd2-c555-433a-9015-f623948b1401-config-data-generated\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.414417 4907 reconciler_common.go:293] "Volume detached for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/c4411bd2-c555-433a-9015-f623948b1401-config-data-default\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.414425 4907 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c4411bd2-c555-433a-9015-f623948b1401-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.421627 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c4411bd2-c555-433a-9015-f623948b1401-kube-api-access-pcwkz" (OuterVolumeSpecName: "kube-api-access-pcwkz") pod "c4411bd2-c555-433a-9015-f623948b1401" (UID: "c4411bd2-c555-433a-9015-f623948b1401"). InnerVolumeSpecName "kube-api-access-pcwkz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.421834 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "mysql-db") pod "c4411bd2-c555-433a-9015-f623948b1401" (UID: "c4411bd2-c555-433a-9015-f623948b1401"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Mar 13 14:30:47 crc kubenswrapper[4907]: E0313 14:30:47.439255 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="265b6b6fe5f95195cdc86c65456998ad327fc7ff3a6641300cd6ac00798eff4f" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Mar 13 14:30:47 crc kubenswrapper[4907]: E0313 14:30:47.439399 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 6b0b25079cc099dabe922ac89fa41ae76d2f01797aac9f273b042e242459c574 is running failed: container process not found" containerID="6b0b25079cc099dabe922ac89fa41ae76d2f01797aac9f273b042e242459c574" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Mar 13 14:30:47 crc kubenswrapper[4907]: E0313 14:30:47.445084 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 6b0b25079cc099dabe922ac89fa41ae76d2f01797aac9f273b042e242459c574 is running failed: container process not found" containerID="6b0b25079cc099dabe922ac89fa41ae76d2f01797aac9f273b042e242459c574" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Mar 13 14:30:47 crc kubenswrapper[4907]: E0313 14:30:47.445479 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 6b0b25079cc099dabe922ac89fa41ae76d2f01797aac9f273b042e242459c574 is running failed: container process not found" containerID="6b0b25079cc099dabe922ac89fa41ae76d2f01797aac9f273b042e242459c574" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Mar 13 14:30:47 crc kubenswrapper[4907]: E0313 14:30:47.445531 4907 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 6b0b25079cc099dabe922ac89fa41ae76d2f01797aac9f273b042e242459c574 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-vp6f5" podUID="c1b70392-1240-40d9-8128-e7abe29c8398" containerName="ovsdb-server" Mar 13 14:30:47 crc kubenswrapper[4907]: E0313 14:30:47.445941 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="265b6b6fe5f95195cdc86c65456998ad327fc7ff3a6641300cd6ac00798eff4f" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Mar 13 14:30:47 crc kubenswrapper[4907]: E0313 14:30:47.447047 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="265b6b6fe5f95195cdc86c65456998ad327fc7ff3a6641300cd6ac00798eff4f" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Mar 13 14:30:47 crc kubenswrapper[4907]: E0313 14:30:47.447124 4907 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-vp6f5" podUID="c1b70392-1240-40d9-8128-e7abe29c8398" containerName="ovs-vswitchd" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.451002 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.457724 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.470264 4907 scope.go:117] "RemoveContainer" containerID="3ccd8eb5dd33f0cece989e6411df4852cb55e2cb0b138f16332e7a07ccf5c865" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.486096 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c4411bd2-c555-433a-9015-f623948b1401-galera-tls-certs" (OuterVolumeSpecName: "galera-tls-certs") pod "c4411bd2-c555-433a-9015-f623948b1401" (UID: "c4411bd2-c555-433a-9015-f623948b1401"). InnerVolumeSpecName "galera-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.498001 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.517920 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcwkz\" (UniqueName: \"kubernetes.io/projected/c4411bd2-c555-433a-9015-f623948b1401-kube-api-access-pcwkz\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.517957 4907 reconciler_common.go:293] "Volume detached for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/c4411bd2-c555-433a-9015-f623948b1401-galera-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.517993 4907 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.523279 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c4411bd2-c555-433a-9015-f623948b1401-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c4411bd2-c555-433a-9015-f623948b1401" (UID: "c4411bd2-c555-433a-9015-f623948b1401"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.526585 4907 scope.go:117] "RemoveContainer" containerID="7117af57f620e2d772f6e2ca6d92edfdc283b3eb640ab414d9483ba1758f36d4" Mar 13 14:30:47 crc kubenswrapper[4907]: E0313 14:30:47.527016 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7117af57f620e2d772f6e2ca6d92edfdc283b3eb640ab414d9483ba1758f36d4\": container with ID starting with 7117af57f620e2d772f6e2ca6d92edfdc283b3eb640ab414d9483ba1758f36d4 not found: ID does not exist" containerID="7117af57f620e2d772f6e2ca6d92edfdc283b3eb640ab414d9483ba1758f36d4" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.527040 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7117af57f620e2d772f6e2ca6d92edfdc283b3eb640ab414d9483ba1758f36d4"} err="failed to get container status \"7117af57f620e2d772f6e2ca6d92edfdc283b3eb640ab414d9483ba1758f36d4\": rpc error: code = NotFound desc = could not find container \"7117af57f620e2d772f6e2ca6d92edfdc283b3eb640ab414d9483ba1758f36d4\": container with ID starting with 7117af57f620e2d772f6e2ca6d92edfdc283b3eb640ab414d9483ba1758f36d4 not found: ID does not exist" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.527056 4907 scope.go:117] "RemoveContainer" containerID="3ccd8eb5dd33f0cece989e6411df4852cb55e2cb0b138f16332e7a07ccf5c865" Mar 13 14:30:47 crc kubenswrapper[4907]: E0313 14:30:47.527613 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3ccd8eb5dd33f0cece989e6411df4852cb55e2cb0b138f16332e7a07ccf5c865\": container with ID starting with 3ccd8eb5dd33f0cece989e6411df4852cb55e2cb0b138f16332e7a07ccf5c865 not found: ID does not exist" containerID="3ccd8eb5dd33f0cece989e6411df4852cb55e2cb0b138f16332e7a07ccf5c865" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.527633 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3ccd8eb5dd33f0cece989e6411df4852cb55e2cb0b138f16332e7a07ccf5c865"} err="failed to get container status \"3ccd8eb5dd33f0cece989e6411df4852cb55e2cb0b138f16332e7a07ccf5c865\": rpc error: code = NotFound desc = could not find container \"3ccd8eb5dd33f0cece989e6411df4852cb55e2cb0b138f16332e7a07ccf5c865\": container with ID starting with 3ccd8eb5dd33f0cece989e6411df4852cb55e2cb0b138f16332e7a07ccf5c865 not found: ID does not exist" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.527647 4907 scope.go:117] "RemoveContainer" containerID="ef8644527a89477971dc8c42d7fc7d2d08c1853a9324c6451f42196c1c50329e" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.533965 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.539212 4907 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.547750 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.553300 4907 scope.go:117] "RemoveContainer" containerID="c5ecad10cc9377aeda50af0d8c9213440ce021149e77309245ab84eb57179e9d" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.565416 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.587835 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.593284 4907 scope.go:117] "RemoveContainer" containerID="0d9941f62564b14687332e4dbc1f81a8df8a858860923a1b915fe67a744f75f5" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.594003 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.605832 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-7fc86597fb-mt9pk"] Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.608129 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-7fc86597fb-mt9pk"] Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.613471 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/memcached-0"] Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.618777 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/memcached-0"] Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.619001 4907 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.619026 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4411bd2-c555-433a-9015-f623948b1401-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.623267 4907 scope.go:117] "RemoveContainer" containerID="eeaee513015ed79d41df919baf6b1bec0513b97e239abe794d310237db6cde32" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.629985 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-keystone-listener-7d7bc9f7d-fks76"] Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.636777 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-keystone-listener-7d7bc9f7d-fks76"] Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.664197 4907 scope.go:117] "RemoveContainer" containerID="58121d89217f995e7e932419445b98334a3ccb8b202e18c097026d64131a91fd" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.687089 4907 scope.go:117] "RemoveContainer" containerID="6c1fdac6667e489d30e980acdd57d1cf0fa750bc21c6f46f11adb992a74fe2bd" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.707192 4907 scope.go:117] "RemoveContainer" containerID="58121d89217f995e7e932419445b98334a3ccb8b202e18c097026d64131a91fd" Mar 13 14:30:47 crc kubenswrapper[4907]: E0313 14:30:47.711832 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"58121d89217f995e7e932419445b98334a3ccb8b202e18c097026d64131a91fd\": container with ID starting with 58121d89217f995e7e932419445b98334a3ccb8b202e18c097026d64131a91fd not found: ID does not exist" containerID="58121d89217f995e7e932419445b98334a3ccb8b202e18c097026d64131a91fd" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.711896 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"58121d89217f995e7e932419445b98334a3ccb8b202e18c097026d64131a91fd"} err="failed to get container status \"58121d89217f995e7e932419445b98334a3ccb8b202e18c097026d64131a91fd\": rpc error: code = NotFound desc = could not find container \"58121d89217f995e7e932419445b98334a3ccb8b202e18c097026d64131a91fd\": container with ID starting with 58121d89217f995e7e932419445b98334a3ccb8b202e18c097026d64131a91fd not found: ID does not exist" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.711929 4907 scope.go:117] "RemoveContainer" containerID="6c1fdac6667e489d30e980acdd57d1cf0fa750bc21c6f46f11adb992a74fe2bd" Mar 13 14:30:47 crc kubenswrapper[4907]: E0313 14:30:47.712336 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6c1fdac6667e489d30e980acdd57d1cf0fa750bc21c6f46f11adb992a74fe2bd\": container with ID starting with 6c1fdac6667e489d30e980acdd57d1cf0fa750bc21c6f46f11adb992a74fe2bd not found: ID does not exist" containerID="6c1fdac6667e489d30e980acdd57d1cf0fa750bc21c6f46f11adb992a74fe2bd" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.712360 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6c1fdac6667e489d30e980acdd57d1cf0fa750bc21c6f46f11adb992a74fe2bd"} err="failed to get container status \"6c1fdac6667e489d30e980acdd57d1cf0fa750bc21c6f46f11adb992a74fe2bd\": rpc error: code = NotFound desc = could not find container \"6c1fdac6667e489d30e980acdd57d1cf0fa750bc21c6f46f11adb992a74fe2bd\": container with ID starting with 6c1fdac6667e489d30e980acdd57d1cf0fa750bc21c6f46f11adb992a74fe2bd not found: ID does not exist" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.712376 4907 scope.go:117] "RemoveContainer" containerID="479ee0fb693df48bcf5ef06770a45535f793e4153c556058631f42347318a544" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.740861 4907 scope.go:117] "RemoveContainer" containerID="70b1012f350567e4a184430af6b8324fc4a32cf4f648d2eb406bd3d58d2b10d8" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.762870 4907 scope.go:117] "RemoveContainer" containerID="22aa297a2f364f3884c53d61c34c94086546e7771e899b38ffebaf99570e4f07" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.784507 4907 scope.go:117] "RemoveContainer" containerID="f21714a9f3b46d6b404fb8600bc2b71af7d623654fb77b4f61140d6330569927" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.793916 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b2e62fa-af35-4f35-8852-362ea97c17e7" path="/var/lib/kubelet/pods/0b2e62fa-af35-4f35-8852-362ea97c17e7/volumes" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.794597 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1b2f62d4-0584-4cc3-81ed-e067d8db23b1" path="/var/lib/kubelet/pods/1b2f62d4-0584-4cc3-81ed-e067d8db23b1/volumes" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.795234 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2e41272f-6e7b-47a7-825e-3571f6a1fd07" path="/var/lib/kubelet/pods/2e41272f-6e7b-47a7-825e-3571f6a1fd07/volumes" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.796725 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76" path="/var/lib/kubelet/pods/41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76/volumes" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.797822 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4251ae36-90ad-41ea-915e-862df60f5c07" path="/var/lib/kubelet/pods/4251ae36-90ad-41ea-915e-862df60f5c07/volumes" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.798741 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4685651c-2690-48f8-97d2-8817fedaff56" path="/var/lib/kubelet/pods/4685651c-2690-48f8-97d2-8817fedaff56/volumes" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.801278 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="65531461-cea2-4b2c-a9b6-8fd9e9bffb27" path="/var/lib/kubelet/pods/65531461-cea2-4b2c-a9b6-8fd9e9bffb27/volumes" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.802097 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6f8119b2-e38a-494d-967f-5198b83512c7" path="/var/lib/kubelet/pods/6f8119b2-e38a-494d-967f-5198b83512c7/volumes" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.803012 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cacb281b-ee3b-42b6-8754-d07e6ea6aa4b" path="/var/lib/kubelet/pods/cacb281b-ee3b-42b6-8754-d07e6ea6aa4b/volumes" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.804197 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cc227d33-3bf1-4b5f-8368-8e6b8a512f7d" path="/var/lib/kubelet/pods/cc227d33-3bf1-4b5f-8368-8e6b8a512f7d/volumes" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.804960 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e3ee069a-41fb-4cab-b650-9bb9c03ac271" path="/var/lib/kubelet/pods/e3ee069a-41fb-4cab-b650-9bb9c03ac271/volumes" Mar 13 14:30:47 crc kubenswrapper[4907]: E0313 14:30:47.828551 4907 projected.go:288] Couldn't get configMap openstack/swift-storage-config-data: configmap "swift-storage-config-data" not found Mar 13 14:30:47 crc kubenswrapper[4907]: E0313 14:30:47.828587 4907 projected.go:263] Couldn't get secret openstack/swift-conf: secret "swift-conf" not found Mar 13 14:30:47 crc kubenswrapper[4907]: E0313 14:30:47.828598 4907 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Mar 13 14:30:47 crc kubenswrapper[4907]: E0313 14:30:47.828609 4907 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: [configmap "swift-storage-config-data" not found, secret "swift-conf" not found, configmap "swift-ring-files" not found] Mar 13 14:30:47 crc kubenswrapper[4907]: E0313 14:30:47.828665 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/14f05f31-8185-43b0-be69-bcf8d5388ea5-etc-swift podName:14f05f31-8185-43b0-be69-bcf8d5388ea5 nodeName:}" failed. No retries permitted until 2026-03-13 14:30:55.828646281 +0000 UTC m=+1554.728433970 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/14f05f31-8185-43b0-be69-bcf8d5388ea5-etc-swift") pod "swift-storage-0" (UID: "14f05f31-8185-43b0-be69-bcf8d5388ea5") : [configmap "swift-storage-config-data" not found, secret "swift-conf" not found, configmap "swift-ring-files" not found] Mar 13 14:30:48 crc kubenswrapper[4907]: I0313 14:30:48.215239 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-85b9d64669-dxnf6" Mar 13 14:30:48 crc kubenswrapper[4907]: I0313 14:30:48.276765 4907 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="286c82e8-b74a-49d7-a355-ac074aace10e" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.104:5671: connect: connection refused" Mar 13 14:30:48 crc kubenswrapper[4907]: I0313 14:30:48.337520 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a738a66-54bc-4f5e-86d5-6e3004d8f265-scripts\") pod \"8a738a66-54bc-4f5e-86d5-6e3004d8f265\" (UID: \"8a738a66-54bc-4f5e-86d5-6e3004d8f265\") " Mar 13 14:30:48 crc kubenswrapper[4907]: I0313 14:30:48.337566 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pvwdb\" (UniqueName: \"kubernetes.io/projected/8a738a66-54bc-4f5e-86d5-6e3004d8f265-kube-api-access-pvwdb\") pod \"8a738a66-54bc-4f5e-86d5-6e3004d8f265\" (UID: \"8a738a66-54bc-4f5e-86d5-6e3004d8f265\") " Mar 13 14:30:48 crc kubenswrapper[4907]: I0313 14:30:48.337604 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8a738a66-54bc-4f5e-86d5-6e3004d8f265-fernet-keys\") pod \"8a738a66-54bc-4f5e-86d5-6e3004d8f265\" (UID: \"8a738a66-54bc-4f5e-86d5-6e3004d8f265\") " Mar 13 14:30:48 crc kubenswrapper[4907]: I0313 14:30:48.337692 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a738a66-54bc-4f5e-86d5-6e3004d8f265-internal-tls-certs\") pod \"8a738a66-54bc-4f5e-86d5-6e3004d8f265\" (UID: \"8a738a66-54bc-4f5e-86d5-6e3004d8f265\") " Mar 13 14:30:48 crc kubenswrapper[4907]: I0313 14:30:48.337712 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a738a66-54bc-4f5e-86d5-6e3004d8f265-config-data\") pod \"8a738a66-54bc-4f5e-86d5-6e3004d8f265\" (UID: \"8a738a66-54bc-4f5e-86d5-6e3004d8f265\") " Mar 13 14:30:48 crc kubenswrapper[4907]: I0313 14:30:48.337770 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a738a66-54bc-4f5e-86d5-6e3004d8f265-combined-ca-bundle\") pod \"8a738a66-54bc-4f5e-86d5-6e3004d8f265\" (UID: \"8a738a66-54bc-4f5e-86d5-6e3004d8f265\") " Mar 13 14:30:48 crc kubenswrapper[4907]: I0313 14:30:48.337843 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a738a66-54bc-4f5e-86d5-6e3004d8f265-public-tls-certs\") pod \"8a738a66-54bc-4f5e-86d5-6e3004d8f265\" (UID: \"8a738a66-54bc-4f5e-86d5-6e3004d8f265\") " Mar 13 14:30:48 crc kubenswrapper[4907]: I0313 14:30:48.337896 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/8a738a66-54bc-4f5e-86d5-6e3004d8f265-credential-keys\") pod \"8a738a66-54bc-4f5e-86d5-6e3004d8f265\" (UID: \"8a738a66-54bc-4f5e-86d5-6e3004d8f265\") " Mar 13 14:30:48 crc kubenswrapper[4907]: E0313 14:30:48.339655 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="77b52c0f64f55952d2255d5814b60083fd608735671f1f8507b09ea2e83db4d4" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Mar 13 14:30:48 crc kubenswrapper[4907]: E0313 14:30:48.343098 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="77b52c0f64f55952d2255d5814b60083fd608735671f1f8507b09ea2e83db4d4" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Mar 13 14:30:48 crc kubenswrapper[4907]: I0313 14:30:48.343319 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a738a66-54bc-4f5e-86d5-6e3004d8f265-kube-api-access-pvwdb" (OuterVolumeSpecName: "kube-api-access-pvwdb") pod "8a738a66-54bc-4f5e-86d5-6e3004d8f265" (UID: "8a738a66-54bc-4f5e-86d5-6e3004d8f265"). InnerVolumeSpecName "kube-api-access-pvwdb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:30:48 crc kubenswrapper[4907]: I0313 14:30:48.346245 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a738a66-54bc-4f5e-86d5-6e3004d8f265-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "8a738a66-54bc-4f5e-86d5-6e3004d8f265" (UID: "8a738a66-54bc-4f5e-86d5-6e3004d8f265"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:48 crc kubenswrapper[4907]: I0313 14:30:48.349281 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a738a66-54bc-4f5e-86d5-6e3004d8f265-scripts" (OuterVolumeSpecName: "scripts") pod "8a738a66-54bc-4f5e-86d5-6e3004d8f265" (UID: "8a738a66-54bc-4f5e-86d5-6e3004d8f265"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:48 crc kubenswrapper[4907]: I0313 14:30:48.351308 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a738a66-54bc-4f5e-86d5-6e3004d8f265-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "8a738a66-54bc-4f5e-86d5-6e3004d8f265" (UID: "8a738a66-54bc-4f5e-86d5-6e3004d8f265"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:48 crc kubenswrapper[4907]: E0313 14:30:48.355417 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="77b52c0f64f55952d2255d5814b60083fd608735671f1f8507b09ea2e83db4d4" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Mar 13 14:30:48 crc kubenswrapper[4907]: E0313 14:30:48.355569 4907 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell1-conductor-0" podUID="25c1c7e6-a149-4a8e-aead-3ebdc7e56851" containerName="nova-cell1-conductor-conductor" Mar 13 14:30:48 crc kubenswrapper[4907]: I0313 14:30:48.374378 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a738a66-54bc-4f5e-86d5-6e3004d8f265-config-data" (OuterVolumeSpecName: "config-data") pod "8a738a66-54bc-4f5e-86d5-6e3004d8f265" (UID: "8a738a66-54bc-4f5e-86d5-6e3004d8f265"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:48 crc kubenswrapper[4907]: I0313 14:30:48.376020 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a738a66-54bc-4f5e-86d5-6e3004d8f265-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8a738a66-54bc-4f5e-86d5-6e3004d8f265" (UID: "8a738a66-54bc-4f5e-86d5-6e3004d8f265"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:48 crc kubenswrapper[4907]: I0313 14:30:48.424317 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a738a66-54bc-4f5e-86d5-6e3004d8f265-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "8a738a66-54bc-4f5e-86d5-6e3004d8f265" (UID: "8a738a66-54bc-4f5e-86d5-6e3004d8f265"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:48 crc kubenswrapper[4907]: I0313 14:30:48.427156 4907 generic.go:334] "Generic (PLEG): container finished" podID="8a738a66-54bc-4f5e-86d5-6e3004d8f265" containerID="e6a1933783ac7223012561b093ce055e734ac2c3c802390f6f0bf14354694b72" exitCode=0 Mar 13 14:30:48 crc kubenswrapper[4907]: I0313 14:30:48.427222 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-85b9d64669-dxnf6" event={"ID":"8a738a66-54bc-4f5e-86d5-6e3004d8f265","Type":"ContainerDied","Data":"e6a1933783ac7223012561b093ce055e734ac2c3c802390f6f0bf14354694b72"} Mar 13 14:30:48 crc kubenswrapper[4907]: I0313 14:30:48.427254 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-85b9d64669-dxnf6" event={"ID":"8a738a66-54bc-4f5e-86d5-6e3004d8f265","Type":"ContainerDied","Data":"50cd2791b5f7c237a1259945e1c40427e424019970480fa895cfa345fab258b4"} Mar 13 14:30:48 crc kubenswrapper[4907]: I0313 14:30:48.427273 4907 scope.go:117] "RemoveContainer" containerID="e6a1933783ac7223012561b093ce055e734ac2c3c802390f6f0bf14354694b72" Mar 13 14:30:48 crc kubenswrapper[4907]: I0313 14:30:48.427394 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-85b9d64669-dxnf6" Mar 13 14:30:48 crc kubenswrapper[4907]: I0313 14:30:48.428697 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a738a66-54bc-4f5e-86d5-6e3004d8f265-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "8a738a66-54bc-4f5e-86d5-6e3004d8f265" (UID: "8a738a66-54bc-4f5e-86d5-6e3004d8f265"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:48 crc kubenswrapper[4907]: I0313 14:30:48.429632 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Mar 13 14:30:48 crc kubenswrapper[4907]: I0313 14:30:48.452293 4907 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8a738a66-54bc-4f5e-86d5-6e3004d8f265-fernet-keys\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:48 crc kubenswrapper[4907]: I0313 14:30:48.452325 4907 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a738a66-54bc-4f5e-86d5-6e3004d8f265-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:48 crc kubenswrapper[4907]: I0313 14:30:48.452338 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a738a66-54bc-4f5e-86d5-6e3004d8f265-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:48 crc kubenswrapper[4907]: I0313 14:30:48.452350 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a738a66-54bc-4f5e-86d5-6e3004d8f265-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:48 crc kubenswrapper[4907]: I0313 14:30:48.452361 4907 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a738a66-54bc-4f5e-86d5-6e3004d8f265-public-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:48 crc kubenswrapper[4907]: E0313 14:30:48.452359 4907 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Mar 13 14:30:48 crc kubenswrapper[4907]: I0313 14:30:48.452376 4907 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/8a738a66-54bc-4f5e-86d5-6e3004d8f265-credential-keys\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:48 crc kubenswrapper[4907]: E0313 14:30:48.452434 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/286c82e8-b74a-49d7-a355-ac074aace10e-config-data podName:286c82e8-b74a-49d7-a355-ac074aace10e nodeName:}" failed. No retries permitted until 2026-03-13 14:30:56.45241674 +0000 UTC m=+1555.352204429 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/286c82e8-b74a-49d7-a355-ac074aace10e-config-data") pod "rabbitmq-cell1-server-0" (UID: "286c82e8-b74a-49d7-a355-ac074aace10e") : configmap "rabbitmq-cell1-config-data" not found Mar 13 14:30:48 crc kubenswrapper[4907]: I0313 14:30:48.452459 4907 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a738a66-54bc-4f5e-86d5-6e3004d8f265-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:48 crc kubenswrapper[4907]: I0313 14:30:48.452475 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pvwdb\" (UniqueName: \"kubernetes.io/projected/8a738a66-54bc-4f5e-86d5-6e3004d8f265-kube-api-access-pvwdb\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:48 crc kubenswrapper[4907]: I0313 14:30:48.456557 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-galera-0"] Mar 13 14:30:48 crc kubenswrapper[4907]: I0313 14:30:48.461344 4907 scope.go:117] "RemoveContainer" containerID="e6a1933783ac7223012561b093ce055e734ac2c3c802390f6f0bf14354694b72" Mar 13 14:30:48 crc kubenswrapper[4907]: E0313 14:30:48.461726 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e6a1933783ac7223012561b093ce055e734ac2c3c802390f6f0bf14354694b72\": container with ID starting with e6a1933783ac7223012561b093ce055e734ac2c3c802390f6f0bf14354694b72 not found: ID does not exist" containerID="e6a1933783ac7223012561b093ce055e734ac2c3c802390f6f0bf14354694b72" Mar 13 14:30:48 crc kubenswrapper[4907]: I0313 14:30:48.461784 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e6a1933783ac7223012561b093ce055e734ac2c3c802390f6f0bf14354694b72"} err="failed to get container status \"e6a1933783ac7223012561b093ce055e734ac2c3c802390f6f0bf14354694b72\": rpc error: code = NotFound desc = could not find container \"e6a1933783ac7223012561b093ce055e734ac2c3c802390f6f0bf14354694b72\": container with ID starting with e6a1933783ac7223012561b093ce055e734ac2c3c802390f6f0bf14354694b72 not found: ID does not exist" Mar 13 14:30:48 crc kubenswrapper[4907]: I0313 14:30:48.463246 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstack-galera-0"] Mar 13 14:30:48 crc kubenswrapper[4907]: I0313 14:30:48.696802 4907 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="e3c34454-315e-4821-ab25-b0f331a0d521" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.105:5671: connect: connection refused" Mar 13 14:30:48 crc kubenswrapper[4907]: I0313 14:30:48.761836 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-85b9d64669-dxnf6"] Mar 13 14:30:48 crc kubenswrapper[4907]: I0313 14:30:48.766666 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-85b9d64669-dxnf6"] Mar 13 14:30:49 crc kubenswrapper[4907]: E0313 14:30:49.054318 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5ab7cf17ecb1804f362340163cfceb48e6f2c8900108799d01d1025c485c1e8f" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Mar 13 14:30:49 crc kubenswrapper[4907]: E0313 14:30:49.055943 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5ab7cf17ecb1804f362340163cfceb48e6f2c8900108799d01d1025c485c1e8f" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Mar 13 14:30:49 crc kubenswrapper[4907]: E0313 14:30:49.057781 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5ab7cf17ecb1804f362340163cfceb48e6f2c8900108799d01d1025c485c1e8f" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Mar 13 14:30:49 crc kubenswrapper[4907]: E0313 14:30:49.057821 4907 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell0-conductor-0" podUID="364a702e-e6c6-42bc-8cee-4ba19ea2f1ec" containerName="nova-cell0-conductor-conductor" Mar 13 14:30:49 crc kubenswrapper[4907]: E0313 14:30:49.231491 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="8a59419fea27690bc54a6fc859fef55fbe86c693159ac3ff37906574c1369e56" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Mar 13 14:30:49 crc kubenswrapper[4907]: E0313 14:30:49.236199 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="8a59419fea27690bc54a6fc859fef55fbe86c693159ac3ff37906574c1369e56" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Mar 13 14:30:49 crc kubenswrapper[4907]: E0313 14:30:49.240650 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="8a59419fea27690bc54a6fc859fef55fbe86c693159ac3ff37906574c1369e56" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Mar 13 14:30:49 crc kubenswrapper[4907]: E0313 14:30:49.240692 4907 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="4050d00b-0256-45c5-9dc4-0ab46956405d" containerName="nova-scheduler-scheduler" Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.334732 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.443631 4907 generic.go:334] "Generic (PLEG): container finished" podID="e3c34454-315e-4821-ab25-b0f331a0d521" containerID="c5d00dd71aba57d230e51645212e11968b2acd8a912b082abd6373002fb907b4" exitCode=0 Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.443678 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"e3c34454-315e-4821-ab25-b0f331a0d521","Type":"ContainerDied","Data":"c5d00dd71aba57d230e51645212e11968b2acd8a912b082abd6373002fb907b4"} Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.446044 4907 generic.go:334] "Generic (PLEG): container finished" podID="286c82e8-b74a-49d7-a355-ac074aace10e" containerID="d23153977130306af0e21b41c03dbbc96a217b011704059af7d65c0a4153cf43" exitCode=0 Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.446070 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"286c82e8-b74a-49d7-a355-ac074aace10e","Type":"ContainerDied","Data":"d23153977130306af0e21b41c03dbbc96a217b011704059af7d65c0a4153cf43"} Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.446097 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"286c82e8-b74a-49d7-a355-ac074aace10e","Type":"ContainerDied","Data":"84aa3724ad3336159381b40fcc1f3bf130caa7609c9affeddec48f79002eb26c"} Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.446114 4907 scope.go:117] "RemoveContainer" containerID="d23153977130306af0e21b41c03dbbc96a217b011704059af7d65c0a4153cf43" Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.446216 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.471236 4907 scope.go:117] "RemoveContainer" containerID="275b76780d99e94e2c958194ce29198632d04fd0fe0f0668ff5025ad387fc2f7" Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.475723 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/286c82e8-b74a-49d7-a355-ac074aace10e-config-data\") pod \"286c82e8-b74a-49d7-a355-ac074aace10e\" (UID: \"286c82e8-b74a-49d7-a355-ac074aace10e\") " Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.475773 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/286c82e8-b74a-49d7-a355-ac074aace10e-pod-info\") pod \"286c82e8-b74a-49d7-a355-ac074aace10e\" (UID: \"286c82e8-b74a-49d7-a355-ac074aace10e\") " Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.475819 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/286c82e8-b74a-49d7-a355-ac074aace10e-erlang-cookie-secret\") pod \"286c82e8-b74a-49d7-a355-ac074aace10e\" (UID: \"286c82e8-b74a-49d7-a355-ac074aace10e\") " Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.475859 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"286c82e8-b74a-49d7-a355-ac074aace10e\" (UID: \"286c82e8-b74a-49d7-a355-ac074aace10e\") " Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.475931 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/286c82e8-b74a-49d7-a355-ac074aace10e-rabbitmq-confd\") pod \"286c82e8-b74a-49d7-a355-ac074aace10e\" (UID: \"286c82e8-b74a-49d7-a355-ac074aace10e\") " Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.475971 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/286c82e8-b74a-49d7-a355-ac074aace10e-rabbitmq-erlang-cookie\") pod \"286c82e8-b74a-49d7-a355-ac074aace10e\" (UID: \"286c82e8-b74a-49d7-a355-ac074aace10e\") " Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.475996 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w8snq\" (UniqueName: \"kubernetes.io/projected/286c82e8-b74a-49d7-a355-ac074aace10e-kube-api-access-w8snq\") pod \"286c82e8-b74a-49d7-a355-ac074aace10e\" (UID: \"286c82e8-b74a-49d7-a355-ac074aace10e\") " Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.476030 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/286c82e8-b74a-49d7-a355-ac074aace10e-plugins-conf\") pod \"286c82e8-b74a-49d7-a355-ac074aace10e\" (UID: \"286c82e8-b74a-49d7-a355-ac074aace10e\") " Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.476050 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/286c82e8-b74a-49d7-a355-ac074aace10e-rabbitmq-plugins\") pod \"286c82e8-b74a-49d7-a355-ac074aace10e\" (UID: \"286c82e8-b74a-49d7-a355-ac074aace10e\") " Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.476068 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/286c82e8-b74a-49d7-a355-ac074aace10e-server-conf\") pod \"286c82e8-b74a-49d7-a355-ac074aace10e\" (UID: \"286c82e8-b74a-49d7-a355-ac074aace10e\") " Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.476096 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/286c82e8-b74a-49d7-a355-ac074aace10e-rabbitmq-tls\") pod \"286c82e8-b74a-49d7-a355-ac074aace10e\" (UID: \"286c82e8-b74a-49d7-a355-ac074aace10e\") " Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.477227 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/286c82e8-b74a-49d7-a355-ac074aace10e-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "286c82e8-b74a-49d7-a355-ac074aace10e" (UID: "286c82e8-b74a-49d7-a355-ac074aace10e"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.477243 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/286c82e8-b74a-49d7-a355-ac074aace10e-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "286c82e8-b74a-49d7-a355-ac074aace10e" (UID: "286c82e8-b74a-49d7-a355-ac074aace10e"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.477232 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/286c82e8-b74a-49d7-a355-ac074aace10e-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "286c82e8-b74a-49d7-a355-ac074aace10e" (UID: "286c82e8-b74a-49d7-a355-ac074aace10e"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.481211 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "persistence") pod "286c82e8-b74a-49d7-a355-ac074aace10e" (UID: "286c82e8-b74a-49d7-a355-ac074aace10e"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.481554 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/286c82e8-b74a-49d7-a355-ac074aace10e-pod-info" (OuterVolumeSpecName: "pod-info") pod "286c82e8-b74a-49d7-a355-ac074aace10e" (UID: "286c82e8-b74a-49d7-a355-ac074aace10e"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.481603 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/286c82e8-b74a-49d7-a355-ac074aace10e-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "286c82e8-b74a-49d7-a355-ac074aace10e" (UID: "286c82e8-b74a-49d7-a355-ac074aace10e"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.494674 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/286c82e8-b74a-49d7-a355-ac074aace10e-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "286c82e8-b74a-49d7-a355-ac074aace10e" (UID: "286c82e8-b74a-49d7-a355-ac074aace10e"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.497495 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/286c82e8-b74a-49d7-a355-ac074aace10e-kube-api-access-w8snq" (OuterVolumeSpecName: "kube-api-access-w8snq") pod "286c82e8-b74a-49d7-a355-ac074aace10e" (UID: "286c82e8-b74a-49d7-a355-ac074aace10e"). InnerVolumeSpecName "kube-api-access-w8snq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.508467 4907 scope.go:117] "RemoveContainer" containerID="d23153977130306af0e21b41c03dbbc96a217b011704059af7d65c0a4153cf43" Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.508562 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/286c82e8-b74a-49d7-a355-ac074aace10e-config-data" (OuterVolumeSpecName: "config-data") pod "286c82e8-b74a-49d7-a355-ac074aace10e" (UID: "286c82e8-b74a-49d7-a355-ac074aace10e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:30:49 crc kubenswrapper[4907]: E0313 14:30:49.508898 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d23153977130306af0e21b41c03dbbc96a217b011704059af7d65c0a4153cf43\": container with ID starting with d23153977130306af0e21b41c03dbbc96a217b011704059af7d65c0a4153cf43 not found: ID does not exist" containerID="d23153977130306af0e21b41c03dbbc96a217b011704059af7d65c0a4153cf43" Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.508931 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d23153977130306af0e21b41c03dbbc96a217b011704059af7d65c0a4153cf43"} err="failed to get container status \"d23153977130306af0e21b41c03dbbc96a217b011704059af7d65c0a4153cf43\": rpc error: code = NotFound desc = could not find container \"d23153977130306af0e21b41c03dbbc96a217b011704059af7d65c0a4153cf43\": container with ID starting with d23153977130306af0e21b41c03dbbc96a217b011704059af7d65c0a4153cf43 not found: ID does not exist" Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.508955 4907 scope.go:117] "RemoveContainer" containerID="275b76780d99e94e2c958194ce29198632d04fd0fe0f0668ff5025ad387fc2f7" Mar 13 14:30:49 crc kubenswrapper[4907]: E0313 14:30:49.509395 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"275b76780d99e94e2c958194ce29198632d04fd0fe0f0668ff5025ad387fc2f7\": container with ID starting with 275b76780d99e94e2c958194ce29198632d04fd0fe0f0668ff5025ad387fc2f7 not found: ID does not exist" containerID="275b76780d99e94e2c958194ce29198632d04fd0fe0f0668ff5025ad387fc2f7" Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.509469 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"275b76780d99e94e2c958194ce29198632d04fd0fe0f0668ff5025ad387fc2f7"} err="failed to get container status \"275b76780d99e94e2c958194ce29198632d04fd0fe0f0668ff5025ad387fc2f7\": rpc error: code = NotFound desc = could not find container \"275b76780d99e94e2c958194ce29198632d04fd0fe0f0668ff5025ad387fc2f7\": container with ID starting with 275b76780d99e94e2c958194ce29198632d04fd0fe0f0668ff5025ad387fc2f7 not found: ID does not exist" Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.526429 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/286c82e8-b74a-49d7-a355-ac074aace10e-server-conf" (OuterVolumeSpecName: "server-conf") pod "286c82e8-b74a-49d7-a355-ac074aace10e" (UID: "286c82e8-b74a-49d7-a355-ac074aace10e"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.578058 4907 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/286c82e8-b74a-49d7-a355-ac074aace10e-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.578107 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/286c82e8-b74a-49d7-a355-ac074aace10e-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.578116 4907 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/286c82e8-b74a-49d7-a355-ac074aace10e-pod-info\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.578126 4907 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/286c82e8-b74a-49d7-a355-ac074aace10e-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.578155 4907 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.578166 4907 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/286c82e8-b74a-49d7-a355-ac074aace10e-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.578178 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w8snq\" (UniqueName: \"kubernetes.io/projected/286c82e8-b74a-49d7-a355-ac074aace10e-kube-api-access-w8snq\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.578187 4907 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/286c82e8-b74a-49d7-a355-ac074aace10e-plugins-conf\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.578203 4907 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/286c82e8-b74a-49d7-a355-ac074aace10e-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.578212 4907 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/286c82e8-b74a-49d7-a355-ac074aace10e-server-conf\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.594093 4907 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.597099 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/286c82e8-b74a-49d7-a355-ac074aace10e-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "286c82e8-b74a-49d7-a355-ac074aace10e" (UID: "286c82e8-b74a-49d7-a355-ac074aace10e"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.680417 4907 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.680444 4907 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/286c82e8-b74a-49d7-a355-ac074aace10e-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.753046 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Mar 13 14:30:49 crc kubenswrapper[4907]: E0313 14:30:49.782203 4907 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Mar 13 14:30:49 crc kubenswrapper[4907]: E0313 14:30:49.782284 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/e3c34454-315e-4821-ab25-b0f331a0d521-config-data podName:e3c34454-315e-4821-ab25-b0f331a0d521 nodeName:}" failed. No retries permitted until 2026-03-13 14:30:57.782266645 +0000 UTC m=+1556.682054334 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/e3c34454-315e-4821-ab25-b0f331a0d521-config-data") pod "rabbitmq-server-0" (UID: "e3c34454-315e-4821-ab25-b0f331a0d521") : configmap "rabbitmq-config-data" not found Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.805841 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8a738a66-54bc-4f5e-86d5-6e3004d8f265" path="/var/lib/kubelet/pods/8a738a66-54bc-4f5e-86d5-6e3004d8f265/volumes" Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.806466 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c4411bd2-c555-433a-9015-f623948b1401" path="/var/lib/kubelet/pods/c4411bd2-c555-433a-9015-f623948b1401/volumes" Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.809587 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.814366 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.883665 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/e3c34454-315e-4821-ab25-b0f331a0d521-rabbitmq-erlang-cookie\") pod \"e3c34454-315e-4821-ab25-b0f331a0d521\" (UID: \"e3c34454-315e-4821-ab25-b0f331a0d521\") " Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.883778 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/e3c34454-315e-4821-ab25-b0f331a0d521-erlang-cookie-secret\") pod \"e3c34454-315e-4821-ab25-b0f331a0d521\" (UID: \"e3c34454-315e-4821-ab25-b0f331a0d521\") " Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.883869 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/e3c34454-315e-4821-ab25-b0f331a0d521-pod-info\") pod \"e3c34454-315e-4821-ab25-b0f331a0d521\" (UID: \"e3c34454-315e-4821-ab25-b0f331a0d521\") " Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.883986 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e3c34454-315e-4821-ab25-b0f331a0d521-config-data\") pod \"e3c34454-315e-4821-ab25-b0f331a0d521\" (UID: \"e3c34454-315e-4821-ab25-b0f331a0d521\") " Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.884083 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"e3c34454-315e-4821-ab25-b0f331a0d521\" (UID: \"e3c34454-315e-4821-ab25-b0f331a0d521\") " Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.884159 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/e3c34454-315e-4821-ab25-b0f331a0d521-server-conf\") pod \"e3c34454-315e-4821-ab25-b0f331a0d521\" (UID: \"e3c34454-315e-4821-ab25-b0f331a0d521\") " Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.884231 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnz7l\" (UniqueName: \"kubernetes.io/projected/e3c34454-315e-4821-ab25-b0f331a0d521-kube-api-access-mnz7l\") pod \"e3c34454-315e-4821-ab25-b0f331a0d521\" (UID: \"e3c34454-315e-4821-ab25-b0f331a0d521\") " Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.884263 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e3c34454-315e-4821-ab25-b0f331a0d521-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "e3c34454-315e-4821-ab25-b0f331a0d521" (UID: "e3c34454-315e-4821-ab25-b0f331a0d521"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.884279 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/e3c34454-315e-4821-ab25-b0f331a0d521-rabbitmq-plugins\") pod \"e3c34454-315e-4821-ab25-b0f331a0d521\" (UID: \"e3c34454-315e-4821-ab25-b0f331a0d521\") " Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.884375 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/e3c34454-315e-4821-ab25-b0f331a0d521-rabbitmq-confd\") pod \"e3c34454-315e-4821-ab25-b0f331a0d521\" (UID: \"e3c34454-315e-4821-ab25-b0f331a0d521\") " Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.884419 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/e3c34454-315e-4821-ab25-b0f331a0d521-plugins-conf\") pod \"e3c34454-315e-4821-ab25-b0f331a0d521\" (UID: \"e3c34454-315e-4821-ab25-b0f331a0d521\") " Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.884444 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/e3c34454-315e-4821-ab25-b0f331a0d521-rabbitmq-tls\") pod \"e3c34454-315e-4821-ab25-b0f331a0d521\" (UID: \"e3c34454-315e-4821-ab25-b0f331a0d521\") " Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.884764 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e3c34454-315e-4821-ab25-b0f331a0d521-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "e3c34454-315e-4821-ab25-b0f331a0d521" (UID: "e3c34454-315e-4821-ab25-b0f331a0d521"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.885266 4907 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/e3c34454-315e-4821-ab25-b0f331a0d521-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.885292 4907 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/e3c34454-315e-4821-ab25-b0f331a0d521-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.887177 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3c34454-315e-4821-ab25-b0f331a0d521-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "e3c34454-315e-4821-ab25-b0f331a0d521" (UID: "e3c34454-315e-4821-ab25-b0f331a0d521"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.887393 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "persistence") pod "e3c34454-315e-4821-ab25-b0f331a0d521" (UID: "e3c34454-315e-4821-ab25-b0f331a0d521"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.887408 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e3c34454-315e-4821-ab25-b0f331a0d521-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "e3c34454-315e-4821-ab25-b0f331a0d521" (UID: "e3c34454-315e-4821-ab25-b0f331a0d521"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.889449 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e3c34454-315e-4821-ab25-b0f331a0d521-kube-api-access-mnz7l" (OuterVolumeSpecName: "kube-api-access-mnz7l") pod "e3c34454-315e-4821-ab25-b0f331a0d521" (UID: "e3c34454-315e-4821-ab25-b0f331a0d521"). InnerVolumeSpecName "kube-api-access-mnz7l". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.890168 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e3c34454-315e-4821-ab25-b0f331a0d521-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "e3c34454-315e-4821-ab25-b0f331a0d521" (UID: "e3c34454-315e-4821-ab25-b0f331a0d521"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.890630 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/e3c34454-315e-4821-ab25-b0f331a0d521-pod-info" (OuterVolumeSpecName: "pod-info") pod "e3c34454-315e-4821-ab25-b0f331a0d521" (UID: "e3c34454-315e-4821-ab25-b0f331a0d521"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.903545 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e3c34454-315e-4821-ab25-b0f331a0d521-config-data" (OuterVolumeSpecName: "config-data") pod "e3c34454-315e-4821-ab25-b0f331a0d521" (UID: "e3c34454-315e-4821-ab25-b0f331a0d521"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.924096 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e3c34454-315e-4821-ab25-b0f331a0d521-server-conf" (OuterVolumeSpecName: "server-conf") pod "e3c34454-315e-4821-ab25-b0f331a0d521" (UID: "e3c34454-315e-4821-ab25-b0f331a0d521"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.968481 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e3c34454-315e-4821-ab25-b0f331a0d521-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "e3c34454-315e-4821-ab25-b0f331a0d521" (UID: "e3c34454-315e-4821-ab25-b0f331a0d521"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.988774 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e3c34454-315e-4821-ab25-b0f331a0d521-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.988829 4907 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.988839 4907 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/e3c34454-315e-4821-ab25-b0f331a0d521-server-conf\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.988849 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnz7l\" (UniqueName: \"kubernetes.io/projected/e3c34454-315e-4821-ab25-b0f331a0d521-kube-api-access-mnz7l\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.988858 4907 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/e3c34454-315e-4821-ab25-b0f331a0d521-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.988867 4907 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/e3c34454-315e-4821-ab25-b0f331a0d521-plugins-conf\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.988874 4907 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/e3c34454-315e-4821-ab25-b0f331a0d521-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.988897 4907 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/e3c34454-315e-4821-ab25-b0f331a0d521-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.988905 4907 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/e3c34454-315e-4821-ab25-b0f331a0d521-pod-info\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:50 crc kubenswrapper[4907]: I0313 14:30:50.008873 4907 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Mar 13 14:30:50 crc kubenswrapper[4907]: I0313 14:30:50.090602 4907 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:50 crc kubenswrapper[4907]: I0313 14:30:50.468160 4907 generic.go:334] "Generic (PLEG): container finished" podID="25c1c7e6-a149-4a8e-aead-3ebdc7e56851" containerID="77b52c0f64f55952d2255d5814b60083fd608735671f1f8507b09ea2e83db4d4" exitCode=0 Mar 13 14:30:50 crc kubenswrapper[4907]: I0313 14:30:50.468225 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"25c1c7e6-a149-4a8e-aead-3ebdc7e56851","Type":"ContainerDied","Data":"77b52c0f64f55952d2255d5814b60083fd608735671f1f8507b09ea2e83db4d4"} Mar 13 14:30:50 crc kubenswrapper[4907]: I0313 14:30:50.472034 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"e3c34454-315e-4821-ab25-b0f331a0d521","Type":"ContainerDied","Data":"0dddf7e485bfeda5a825f940bee1a34df372878429700a5d41310e0d0412399b"} Mar 13 14:30:50 crc kubenswrapper[4907]: I0313 14:30:50.472084 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Mar 13 14:30:50 crc kubenswrapper[4907]: I0313 14:30:50.472102 4907 scope.go:117] "RemoveContainer" containerID="c5d00dd71aba57d230e51645212e11968b2acd8a912b082abd6373002fb907b4" Mar 13 14:30:50 crc kubenswrapper[4907]: I0313 14:30:50.506431 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Mar 13 14:30:50 crc kubenswrapper[4907]: I0313 14:30:50.518557 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Mar 13 14:30:50 crc kubenswrapper[4907]: I0313 14:30:50.522449 4907 scope.go:117] "RemoveContainer" containerID="987e5ce0ff0b4c3691f9df5efc31a14e40c988230f0cf14b4b06e00cd7129b38" Mar 13 14:30:50 crc kubenswrapper[4907]: I0313 14:30:50.883329 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.014670 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25c1c7e6-a149-4a8e-aead-3ebdc7e56851-combined-ca-bundle\") pod \"25c1c7e6-a149-4a8e-aead-3ebdc7e56851\" (UID: \"25c1c7e6-a149-4a8e-aead-3ebdc7e56851\") " Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.014746 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gxd7v\" (UniqueName: \"kubernetes.io/projected/25c1c7e6-a149-4a8e-aead-3ebdc7e56851-kube-api-access-gxd7v\") pod \"25c1c7e6-a149-4a8e-aead-3ebdc7e56851\" (UID: \"25c1c7e6-a149-4a8e-aead-3ebdc7e56851\") " Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.014781 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/25c1c7e6-a149-4a8e-aead-3ebdc7e56851-config-data\") pod \"25c1c7e6-a149-4a8e-aead-3ebdc7e56851\" (UID: \"25c1c7e6-a149-4a8e-aead-3ebdc7e56851\") " Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.027100 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25c1c7e6-a149-4a8e-aead-3ebdc7e56851-kube-api-access-gxd7v" (OuterVolumeSpecName: "kube-api-access-gxd7v") pod "25c1c7e6-a149-4a8e-aead-3ebdc7e56851" (UID: "25c1c7e6-a149-4a8e-aead-3ebdc7e56851"). InnerVolumeSpecName "kube-api-access-gxd7v". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.048329 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25c1c7e6-a149-4a8e-aead-3ebdc7e56851-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "25c1c7e6-a149-4a8e-aead-3ebdc7e56851" (UID: "25c1c7e6-a149-4a8e-aead-3ebdc7e56851"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.059329 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25c1c7e6-a149-4a8e-aead-3ebdc7e56851-config-data" (OuterVolumeSpecName: "config-data") pod "25c1c7e6-a149-4a8e-aead-3ebdc7e56851" (UID: "25c1c7e6-a149-4a8e-aead-3ebdc7e56851"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.116360 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gxd7v\" (UniqueName: \"kubernetes.io/projected/25c1c7e6-a149-4a8e-aead-3ebdc7e56851-kube-api-access-gxd7v\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.116396 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/25c1c7e6-a149-4a8e-aead-3ebdc7e56851-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.116406 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25c1c7e6-a149-4a8e-aead-3ebdc7e56851-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.145228 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.155700 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-76bbcdf7d9-j6hw5" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.162002 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.162076 4907 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-7fc86597fb-mt9pk" podUID="1b2f62d4-0584-4cc3-81ed-e067d8db23b1" containerName="barbican-api" probeResult="failure" output="Get \"https://10.217.0.167:9311/healthcheck\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.162285 4907 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-7fc86597fb-mt9pk" podUID="1b2f62d4-0584-4cc3-81ed-e067d8db23b1" containerName="barbican-api-log" probeResult="failure" output="Get \"https://10.217.0.167:9311/healthcheck\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.256513 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.275574 4907 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/memcached-0" podUID="cacb281b-ee3b-42b6-8754-d07e6ea6aa4b" containerName="memcached" probeResult="failure" output="dial tcp 10.217.0.108:11211: i/o timeout" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.319147 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/aa81fa2a-4c27-4e02-9891-5b39348feca9-sg-core-conf-yaml\") pod \"aa81fa2a-4c27-4e02-9891-5b39348feca9\" (UID: \"aa81fa2a-4c27-4e02-9891-5b39348feca9\") " Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.319217 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xdg29\" (UniqueName: \"kubernetes.io/projected/4050d00b-0256-45c5-9dc4-0ab46956405d-kube-api-access-xdg29\") pod \"4050d00b-0256-45c5-9dc4-0ab46956405d\" (UID: \"4050d00b-0256-45c5-9dc4-0ab46956405d\") " Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.319245 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4050d00b-0256-45c5-9dc4-0ab46956405d-combined-ca-bundle\") pod \"4050d00b-0256-45c5-9dc4-0ab46956405d\" (UID: \"4050d00b-0256-45c5-9dc4-0ab46956405d\") " Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.319294 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aa81fa2a-4c27-4e02-9891-5b39348feca9-log-httpd\") pod \"aa81fa2a-4c27-4e02-9891-5b39348feca9\" (UID: \"aa81fa2a-4c27-4e02-9891-5b39348feca9\") " Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.319320 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa81fa2a-4c27-4e02-9891-5b39348feca9-combined-ca-bundle\") pod \"aa81fa2a-4c27-4e02-9891-5b39348feca9\" (UID: \"aa81fa2a-4c27-4e02-9891-5b39348feca9\") " Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.319342 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4050d00b-0256-45c5-9dc4-0ab46956405d-config-data\") pod \"4050d00b-0256-45c5-9dc4-0ab46956405d\" (UID: \"4050d00b-0256-45c5-9dc4-0ab46956405d\") " Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.319365 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/aa81fa2a-4c27-4e02-9891-5b39348feca9-ceilometer-tls-certs\") pod \"aa81fa2a-4c27-4e02-9891-5b39348feca9\" (UID: \"aa81fa2a-4c27-4e02-9891-5b39348feca9\") " Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.319419 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/25b5fafe-65c8-48ba-bc34-83442f8ace4c-logs\") pod \"25b5fafe-65c8-48ba-bc34-83442f8ace4c\" (UID: \"25b5fafe-65c8-48ba-bc34-83442f8ace4c\") " Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.319449 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dvn24\" (UniqueName: \"kubernetes.io/projected/aa81fa2a-4c27-4e02-9891-5b39348feca9-kube-api-access-dvn24\") pod \"aa81fa2a-4c27-4e02-9891-5b39348feca9\" (UID: \"aa81fa2a-4c27-4e02-9891-5b39348feca9\") " Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.319476 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/25b5fafe-65c8-48ba-bc34-83442f8ace4c-config-data-custom\") pod \"25b5fafe-65c8-48ba-bc34-83442f8ace4c\" (UID: \"25b5fafe-65c8-48ba-bc34-83442f8ace4c\") " Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.319495 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/25b5fafe-65c8-48ba-bc34-83442f8ace4c-config-data\") pod \"25b5fafe-65c8-48ba-bc34-83442f8ace4c\" (UID: \"25b5fafe-65c8-48ba-bc34-83442f8ace4c\") " Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.319518 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-99bgf\" (UniqueName: \"kubernetes.io/projected/25b5fafe-65c8-48ba-bc34-83442f8ace4c-kube-api-access-99bgf\") pod \"25b5fafe-65c8-48ba-bc34-83442f8ace4c\" (UID: \"25b5fafe-65c8-48ba-bc34-83442f8ace4c\") " Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.319532 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aa81fa2a-4c27-4e02-9891-5b39348feca9-run-httpd\") pod \"aa81fa2a-4c27-4e02-9891-5b39348feca9\" (UID: \"aa81fa2a-4c27-4e02-9891-5b39348feca9\") " Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.319553 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aa81fa2a-4c27-4e02-9891-5b39348feca9-scripts\") pod \"aa81fa2a-4c27-4e02-9891-5b39348feca9\" (UID: \"aa81fa2a-4c27-4e02-9891-5b39348feca9\") " Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.319590 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa81fa2a-4c27-4e02-9891-5b39348feca9-config-data\") pod \"aa81fa2a-4c27-4e02-9891-5b39348feca9\" (UID: \"aa81fa2a-4c27-4e02-9891-5b39348feca9\") " Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.319613 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25b5fafe-65c8-48ba-bc34-83442f8ace4c-combined-ca-bundle\") pod \"25b5fafe-65c8-48ba-bc34-83442f8ace4c\" (UID: \"25b5fafe-65c8-48ba-bc34-83442f8ace4c\") " Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.320247 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/25b5fafe-65c8-48ba-bc34-83442f8ace4c-logs" (OuterVolumeSpecName: "logs") pod "25b5fafe-65c8-48ba-bc34-83442f8ace4c" (UID: "25b5fafe-65c8-48ba-bc34-83442f8ace4c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.321106 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aa81fa2a-4c27-4e02-9891-5b39348feca9-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "aa81fa2a-4c27-4e02-9891-5b39348feca9" (UID: "aa81fa2a-4c27-4e02-9891-5b39348feca9"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.321970 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aa81fa2a-4c27-4e02-9891-5b39348feca9-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "aa81fa2a-4c27-4e02-9891-5b39348feca9" (UID: "aa81fa2a-4c27-4e02-9891-5b39348feca9"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.323446 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25b5fafe-65c8-48ba-bc34-83442f8ace4c-kube-api-access-99bgf" (OuterVolumeSpecName: "kube-api-access-99bgf") pod "25b5fafe-65c8-48ba-bc34-83442f8ace4c" (UID: "25b5fafe-65c8-48ba-bc34-83442f8ace4c"). InnerVolumeSpecName "kube-api-access-99bgf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.326265 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aa81fa2a-4c27-4e02-9891-5b39348feca9-kube-api-access-dvn24" (OuterVolumeSpecName: "kube-api-access-dvn24") pod "aa81fa2a-4c27-4e02-9891-5b39348feca9" (UID: "aa81fa2a-4c27-4e02-9891-5b39348feca9"). InnerVolumeSpecName "kube-api-access-dvn24". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.326376 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa81fa2a-4c27-4e02-9891-5b39348feca9-scripts" (OuterVolumeSpecName: "scripts") pod "aa81fa2a-4c27-4e02-9891-5b39348feca9" (UID: "aa81fa2a-4c27-4e02-9891-5b39348feca9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.326458 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25b5fafe-65c8-48ba-bc34-83442f8ace4c-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "25b5fafe-65c8-48ba-bc34-83442f8ace4c" (UID: "25b5fafe-65c8-48ba-bc34-83442f8ace4c"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.326713 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4050d00b-0256-45c5-9dc4-0ab46956405d-kube-api-access-xdg29" (OuterVolumeSpecName: "kube-api-access-xdg29") pod "4050d00b-0256-45c5-9dc4-0ab46956405d" (UID: "4050d00b-0256-45c5-9dc4-0ab46956405d"). InnerVolumeSpecName "kube-api-access-xdg29". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.346541 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4050d00b-0256-45c5-9dc4-0ab46956405d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4050d00b-0256-45c5-9dc4-0ab46956405d" (UID: "4050d00b-0256-45c5-9dc4-0ab46956405d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.347166 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa81fa2a-4c27-4e02-9891-5b39348feca9-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "aa81fa2a-4c27-4e02-9891-5b39348feca9" (UID: "aa81fa2a-4c27-4e02-9891-5b39348feca9"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.357062 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25b5fafe-65c8-48ba-bc34-83442f8ace4c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "25b5fafe-65c8-48ba-bc34-83442f8ace4c" (UID: "25b5fafe-65c8-48ba-bc34-83442f8ace4c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.357967 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4050d00b-0256-45c5-9dc4-0ab46956405d-config-data" (OuterVolumeSpecName: "config-data") pod "4050d00b-0256-45c5-9dc4-0ab46956405d" (UID: "4050d00b-0256-45c5-9dc4-0ab46956405d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.366467 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa81fa2a-4c27-4e02-9891-5b39348feca9-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "aa81fa2a-4c27-4e02-9891-5b39348feca9" (UID: "aa81fa2a-4c27-4e02-9891-5b39348feca9"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.372498 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25b5fafe-65c8-48ba-bc34-83442f8ace4c-config-data" (OuterVolumeSpecName: "config-data") pod "25b5fafe-65c8-48ba-bc34-83442f8ace4c" (UID: "25b5fafe-65c8-48ba-bc34-83442f8ace4c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.379682 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa81fa2a-4c27-4e02-9891-5b39348feca9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "aa81fa2a-4c27-4e02-9891-5b39348feca9" (UID: "aa81fa2a-4c27-4e02-9891-5b39348feca9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.399465 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa81fa2a-4c27-4e02-9891-5b39348feca9-config-data" (OuterVolumeSpecName: "config-data") pod "aa81fa2a-4c27-4e02-9891-5b39348feca9" (UID: "aa81fa2a-4c27-4e02-9891-5b39348feca9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.420978 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/364a702e-e6c6-42bc-8cee-4ba19ea2f1ec-config-data\") pod \"364a702e-e6c6-42bc-8cee-4ba19ea2f1ec\" (UID: \"364a702e-e6c6-42bc-8cee-4ba19ea2f1ec\") " Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.421033 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/364a702e-e6c6-42bc-8cee-4ba19ea2f1ec-combined-ca-bundle\") pod \"364a702e-e6c6-42bc-8cee-4ba19ea2f1ec\" (UID: \"364a702e-e6c6-42bc-8cee-4ba19ea2f1ec\") " Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.421074 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cdxcj\" (UniqueName: \"kubernetes.io/projected/364a702e-e6c6-42bc-8cee-4ba19ea2f1ec-kube-api-access-cdxcj\") pod \"364a702e-e6c6-42bc-8cee-4ba19ea2f1ec\" (UID: \"364a702e-e6c6-42bc-8cee-4ba19ea2f1ec\") " Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.421351 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/25b5fafe-65c8-48ba-bc34-83442f8ace4c-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.421367 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-99bgf\" (UniqueName: \"kubernetes.io/projected/25b5fafe-65c8-48ba-bc34-83442f8ace4c-kube-api-access-99bgf\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.421377 4907 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aa81fa2a-4c27-4e02-9891-5b39348feca9-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.421385 4907 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aa81fa2a-4c27-4e02-9891-5b39348feca9-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.421397 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa81fa2a-4c27-4e02-9891-5b39348feca9-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.421404 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25b5fafe-65c8-48ba-bc34-83442f8ace4c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.421412 4907 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/aa81fa2a-4c27-4e02-9891-5b39348feca9-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.421422 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xdg29\" (UniqueName: \"kubernetes.io/projected/4050d00b-0256-45c5-9dc4-0ab46956405d-kube-api-access-xdg29\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.421433 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4050d00b-0256-45c5-9dc4-0ab46956405d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.421444 4907 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aa81fa2a-4c27-4e02-9891-5b39348feca9-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.421454 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa81fa2a-4c27-4e02-9891-5b39348feca9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.421462 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4050d00b-0256-45c5-9dc4-0ab46956405d-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.421470 4907 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/aa81fa2a-4c27-4e02-9891-5b39348feca9-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.421477 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dvn24\" (UniqueName: \"kubernetes.io/projected/aa81fa2a-4c27-4e02-9891-5b39348feca9-kube-api-access-dvn24\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.421485 4907 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/25b5fafe-65c8-48ba-bc34-83442f8ace4c-logs\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.421492 4907 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/25b5fafe-65c8-48ba-bc34-83442f8ace4c-config-data-custom\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.424072 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/364a702e-e6c6-42bc-8cee-4ba19ea2f1ec-kube-api-access-cdxcj" (OuterVolumeSpecName: "kube-api-access-cdxcj") pod "364a702e-e6c6-42bc-8cee-4ba19ea2f1ec" (UID: "364a702e-e6c6-42bc-8cee-4ba19ea2f1ec"). InnerVolumeSpecName "kube-api-access-cdxcj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.442518 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/364a702e-e6c6-42bc-8cee-4ba19ea2f1ec-config-data" (OuterVolumeSpecName: "config-data") pod "364a702e-e6c6-42bc-8cee-4ba19ea2f1ec" (UID: "364a702e-e6c6-42bc-8cee-4ba19ea2f1ec"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.453732 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/364a702e-e6c6-42bc-8cee-4ba19ea2f1ec-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "364a702e-e6c6-42bc-8cee-4ba19ea2f1ec" (UID: "364a702e-e6c6-42bc-8cee-4ba19ea2f1ec"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.488399 4907 generic.go:334] "Generic (PLEG): container finished" podID="aa81fa2a-4c27-4e02-9891-5b39348feca9" containerID="78891ea90d5eb95d642931ec25514d47603b8811d971d7d33d8fdc5eb0d3bb04" exitCode=0 Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.488495 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aa81fa2a-4c27-4e02-9891-5b39348feca9","Type":"ContainerDied","Data":"78891ea90d5eb95d642931ec25514d47603b8811d971d7d33d8fdc5eb0d3bb04"} Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.488510 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.488539 4907 scope.go:117] "RemoveContainer" containerID="79914c8007c9b10b224f5d4d1fcd270f17078ab7c002b89d3dab089884af5d6c" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.488527 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aa81fa2a-4c27-4e02-9891-5b39348feca9","Type":"ContainerDied","Data":"b4ef9b58ba405d6fb537a91498e7353e4bf4deae5a0e5f61d8548bb330be20cb"} Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.491559 4907 generic.go:334] "Generic (PLEG): container finished" podID="364a702e-e6c6-42bc-8cee-4ba19ea2f1ec" containerID="5ab7cf17ecb1804f362340163cfceb48e6f2c8900108799d01d1025c485c1e8f" exitCode=0 Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.491631 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"364a702e-e6c6-42bc-8cee-4ba19ea2f1ec","Type":"ContainerDied","Data":"5ab7cf17ecb1804f362340163cfceb48e6f2c8900108799d01d1025c485c1e8f"} Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.491651 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"364a702e-e6c6-42bc-8cee-4ba19ea2f1ec","Type":"ContainerDied","Data":"416b8c99953bb29e1da48539894292de008485b785183dda880949342468afcc"} Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.492002 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.493156 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.494353 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"25c1c7e6-a149-4a8e-aead-3ebdc7e56851","Type":"ContainerDied","Data":"b71e79c4c15864ccbc5c66db3afab587a50c68e66b3d7994a1fc84ed59fd938f"} Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.515294 4907 generic.go:334] "Generic (PLEG): container finished" podID="25b5fafe-65c8-48ba-bc34-83442f8ace4c" containerID="be9d4c930500ca917844d7687055220642421756a075ce078989248af8d81176" exitCode=0 Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.515329 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-76bbcdf7d9-j6hw5" event={"ID":"25b5fafe-65c8-48ba-bc34-83442f8ace4c","Type":"ContainerDied","Data":"be9d4c930500ca917844d7687055220642421756a075ce078989248af8d81176"} Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.515365 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-76bbcdf7d9-j6hw5" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.515377 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-76bbcdf7d9-j6hw5" event={"ID":"25b5fafe-65c8-48ba-bc34-83442f8ace4c","Type":"ContainerDied","Data":"57bab78a65adf0bc133213358e1f66cc1c97a4b0aefeb088fc087562e2ca2999"} Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.517435 4907 generic.go:334] "Generic (PLEG): container finished" podID="4050d00b-0256-45c5-9dc4-0ab46956405d" containerID="8a59419fea27690bc54a6fc859fef55fbe86c693159ac3ff37906574c1369e56" exitCode=0 Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.517461 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"4050d00b-0256-45c5-9dc4-0ab46956405d","Type":"ContainerDied","Data":"8a59419fea27690bc54a6fc859fef55fbe86c693159ac3ff37906574c1369e56"} Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.517479 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"4050d00b-0256-45c5-9dc4-0ab46956405d","Type":"ContainerDied","Data":"c18abca4f12b62785c093b1834a9526c1fe011c7d23dc6808c045af32e0999dc"} Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.517540 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.520995 4907 scope.go:117] "RemoveContainer" containerID="16ef32bcd58d9ca78315416c39fe9eca0c5eee8682ac66cea46ec8f9242f8ee5" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.528462 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/364a702e-e6c6-42bc-8cee-4ba19ea2f1ec-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.528494 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/364a702e-e6c6-42bc-8cee-4ba19ea2f1ec-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.528510 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cdxcj\" (UniqueName: \"kubernetes.io/projected/364a702e-e6c6-42bc-8cee-4ba19ea2f1ec-kube-api-access-cdxcj\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.534665 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.539629 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.551213 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.554228 4907 scope.go:117] "RemoveContainer" containerID="78891ea90d5eb95d642931ec25514d47603b8811d971d7d33d8fdc5eb0d3bb04" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.558636 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-0"] Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.568561 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.580495 4907 scope.go:117] "RemoveContainer" containerID="285cdb1df855cbeeda75e9d50bffbe69e5c3f9a5485dd118f47a932fb1d36f19" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.582336 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-0"] Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.593470 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.600935 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.602274 4907 scope.go:117] "RemoveContainer" containerID="79914c8007c9b10b224f5d4d1fcd270f17078ab7c002b89d3dab089884af5d6c" Mar 13 14:30:51 crc kubenswrapper[4907]: E0313 14:30:51.602756 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"79914c8007c9b10b224f5d4d1fcd270f17078ab7c002b89d3dab089884af5d6c\": container with ID starting with 79914c8007c9b10b224f5d4d1fcd270f17078ab7c002b89d3dab089884af5d6c not found: ID does not exist" containerID="79914c8007c9b10b224f5d4d1fcd270f17078ab7c002b89d3dab089884af5d6c" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.602819 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"79914c8007c9b10b224f5d4d1fcd270f17078ab7c002b89d3dab089884af5d6c"} err="failed to get container status \"79914c8007c9b10b224f5d4d1fcd270f17078ab7c002b89d3dab089884af5d6c\": rpc error: code = NotFound desc = could not find container \"79914c8007c9b10b224f5d4d1fcd270f17078ab7c002b89d3dab089884af5d6c\": container with ID starting with 79914c8007c9b10b224f5d4d1fcd270f17078ab7c002b89d3dab089884af5d6c not found: ID does not exist" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.602846 4907 scope.go:117] "RemoveContainer" containerID="16ef32bcd58d9ca78315416c39fe9eca0c5eee8682ac66cea46ec8f9242f8ee5" Mar 13 14:30:51 crc kubenswrapper[4907]: E0313 14:30:51.603242 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"16ef32bcd58d9ca78315416c39fe9eca0c5eee8682ac66cea46ec8f9242f8ee5\": container with ID starting with 16ef32bcd58d9ca78315416c39fe9eca0c5eee8682ac66cea46ec8f9242f8ee5 not found: ID does not exist" containerID="16ef32bcd58d9ca78315416c39fe9eca0c5eee8682ac66cea46ec8f9242f8ee5" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.603267 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"16ef32bcd58d9ca78315416c39fe9eca0c5eee8682ac66cea46ec8f9242f8ee5"} err="failed to get container status \"16ef32bcd58d9ca78315416c39fe9eca0c5eee8682ac66cea46ec8f9242f8ee5\": rpc error: code = NotFound desc = could not find container \"16ef32bcd58d9ca78315416c39fe9eca0c5eee8682ac66cea46ec8f9242f8ee5\": container with ID starting with 16ef32bcd58d9ca78315416c39fe9eca0c5eee8682ac66cea46ec8f9242f8ee5 not found: ID does not exist" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.603279 4907 scope.go:117] "RemoveContainer" containerID="78891ea90d5eb95d642931ec25514d47603b8811d971d7d33d8fdc5eb0d3bb04" Mar 13 14:30:51 crc kubenswrapper[4907]: E0313 14:30:51.603517 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"78891ea90d5eb95d642931ec25514d47603b8811d971d7d33d8fdc5eb0d3bb04\": container with ID starting with 78891ea90d5eb95d642931ec25514d47603b8811d971d7d33d8fdc5eb0d3bb04 not found: ID does not exist" containerID="78891ea90d5eb95d642931ec25514d47603b8811d971d7d33d8fdc5eb0d3bb04" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.603551 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"78891ea90d5eb95d642931ec25514d47603b8811d971d7d33d8fdc5eb0d3bb04"} err="failed to get container status \"78891ea90d5eb95d642931ec25514d47603b8811d971d7d33d8fdc5eb0d3bb04\": rpc error: code = NotFound desc = could not find container \"78891ea90d5eb95d642931ec25514d47603b8811d971d7d33d8fdc5eb0d3bb04\": container with ID starting with 78891ea90d5eb95d642931ec25514d47603b8811d971d7d33d8fdc5eb0d3bb04 not found: ID does not exist" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.603578 4907 scope.go:117] "RemoveContainer" containerID="285cdb1df855cbeeda75e9d50bffbe69e5c3f9a5485dd118f47a932fb1d36f19" Mar 13 14:30:51 crc kubenswrapper[4907]: E0313 14:30:51.603948 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"285cdb1df855cbeeda75e9d50bffbe69e5c3f9a5485dd118f47a932fb1d36f19\": container with ID starting with 285cdb1df855cbeeda75e9d50bffbe69e5c3f9a5485dd118f47a932fb1d36f19 not found: ID does not exist" containerID="285cdb1df855cbeeda75e9d50bffbe69e5c3f9a5485dd118f47a932fb1d36f19" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.604001 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"285cdb1df855cbeeda75e9d50bffbe69e5c3f9a5485dd118f47a932fb1d36f19"} err="failed to get container status \"285cdb1df855cbeeda75e9d50bffbe69e5c3f9a5485dd118f47a932fb1d36f19\": rpc error: code = NotFound desc = could not find container \"285cdb1df855cbeeda75e9d50bffbe69e5c3f9a5485dd118f47a932fb1d36f19\": container with ID starting with 285cdb1df855cbeeda75e9d50bffbe69e5c3f9a5485dd118f47a932fb1d36f19 not found: ID does not exist" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.604019 4907 scope.go:117] "RemoveContainer" containerID="5ab7cf17ecb1804f362340163cfceb48e6f2c8900108799d01d1025c485c1e8f" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.607221 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-worker-76bbcdf7d9-j6hw5"] Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.612402 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-worker-76bbcdf7d9-j6hw5"] Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.619654 4907 scope.go:117] "RemoveContainer" containerID="5ab7cf17ecb1804f362340163cfceb48e6f2c8900108799d01d1025c485c1e8f" Mar 13 14:30:51 crc kubenswrapper[4907]: E0313 14:30:51.620127 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5ab7cf17ecb1804f362340163cfceb48e6f2c8900108799d01d1025c485c1e8f\": container with ID starting with 5ab7cf17ecb1804f362340163cfceb48e6f2c8900108799d01d1025c485c1e8f not found: ID does not exist" containerID="5ab7cf17ecb1804f362340163cfceb48e6f2c8900108799d01d1025c485c1e8f" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.620158 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5ab7cf17ecb1804f362340163cfceb48e6f2c8900108799d01d1025c485c1e8f"} err="failed to get container status \"5ab7cf17ecb1804f362340163cfceb48e6f2c8900108799d01d1025c485c1e8f\": rpc error: code = NotFound desc = could not find container \"5ab7cf17ecb1804f362340163cfceb48e6f2c8900108799d01d1025c485c1e8f\": container with ID starting with 5ab7cf17ecb1804f362340163cfceb48e6f2c8900108799d01d1025c485c1e8f not found: ID does not exist" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.620177 4907 scope.go:117] "RemoveContainer" containerID="77b52c0f64f55952d2255d5814b60083fd608735671f1f8507b09ea2e83db4d4" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.635203 4907 scope.go:117] "RemoveContainer" containerID="be9d4c930500ca917844d7687055220642421756a075ce078989248af8d81176" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.652627 4907 scope.go:117] "RemoveContainer" containerID="b5356b2175cd794b4fc6692eb61775ae05e66fb791f7dbd913a5d208e5880c25" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.669200 4907 scope.go:117] "RemoveContainer" containerID="be9d4c930500ca917844d7687055220642421756a075ce078989248af8d81176" Mar 13 14:30:51 crc kubenswrapper[4907]: E0313 14:30:51.669615 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"be9d4c930500ca917844d7687055220642421756a075ce078989248af8d81176\": container with ID starting with be9d4c930500ca917844d7687055220642421756a075ce078989248af8d81176 not found: ID does not exist" containerID="be9d4c930500ca917844d7687055220642421756a075ce078989248af8d81176" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.669644 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"be9d4c930500ca917844d7687055220642421756a075ce078989248af8d81176"} err="failed to get container status \"be9d4c930500ca917844d7687055220642421756a075ce078989248af8d81176\": rpc error: code = NotFound desc = could not find container \"be9d4c930500ca917844d7687055220642421756a075ce078989248af8d81176\": container with ID starting with be9d4c930500ca917844d7687055220642421756a075ce078989248af8d81176 not found: ID does not exist" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.669665 4907 scope.go:117] "RemoveContainer" containerID="b5356b2175cd794b4fc6692eb61775ae05e66fb791f7dbd913a5d208e5880c25" Mar 13 14:30:51 crc kubenswrapper[4907]: E0313 14:30:51.669996 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b5356b2175cd794b4fc6692eb61775ae05e66fb791f7dbd913a5d208e5880c25\": container with ID starting with b5356b2175cd794b4fc6692eb61775ae05e66fb791f7dbd913a5d208e5880c25 not found: ID does not exist" containerID="b5356b2175cd794b4fc6692eb61775ae05e66fb791f7dbd913a5d208e5880c25" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.670018 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b5356b2175cd794b4fc6692eb61775ae05e66fb791f7dbd913a5d208e5880c25"} err="failed to get container status \"b5356b2175cd794b4fc6692eb61775ae05e66fb791f7dbd913a5d208e5880c25\": rpc error: code = NotFound desc = could not find container \"b5356b2175cd794b4fc6692eb61775ae05e66fb791f7dbd913a5d208e5880c25\": container with ID starting with b5356b2175cd794b4fc6692eb61775ae05e66fb791f7dbd913a5d208e5880c25 not found: ID does not exist" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.670030 4907 scope.go:117] "RemoveContainer" containerID="8a59419fea27690bc54a6fc859fef55fbe86c693159ac3ff37906574c1369e56" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.689829 4907 scope.go:117] "RemoveContainer" containerID="8a59419fea27690bc54a6fc859fef55fbe86c693159ac3ff37906574c1369e56" Mar 13 14:30:51 crc kubenswrapper[4907]: E0313 14:30:51.690319 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8a59419fea27690bc54a6fc859fef55fbe86c693159ac3ff37906574c1369e56\": container with ID starting with 8a59419fea27690bc54a6fc859fef55fbe86c693159ac3ff37906574c1369e56 not found: ID does not exist" containerID="8a59419fea27690bc54a6fc859fef55fbe86c693159ac3ff37906574c1369e56" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.690357 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8a59419fea27690bc54a6fc859fef55fbe86c693159ac3ff37906574c1369e56"} err="failed to get container status \"8a59419fea27690bc54a6fc859fef55fbe86c693159ac3ff37906574c1369e56\": rpc error: code = NotFound desc = could not find container \"8a59419fea27690bc54a6fc859fef55fbe86c693159ac3ff37906574c1369e56\": container with ID starting with 8a59419fea27690bc54a6fc859fef55fbe86c693159ac3ff37906574c1369e56 not found: ID does not exist" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.805978 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25b5fafe-65c8-48ba-bc34-83442f8ace4c" path="/var/lib/kubelet/pods/25b5fafe-65c8-48ba-bc34-83442f8ace4c/volumes" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.807239 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25c1c7e6-a149-4a8e-aead-3ebdc7e56851" path="/var/lib/kubelet/pods/25c1c7e6-a149-4a8e-aead-3ebdc7e56851/volumes" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.809162 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="286c82e8-b74a-49d7-a355-ac074aace10e" path="/var/lib/kubelet/pods/286c82e8-b74a-49d7-a355-ac074aace10e/volumes" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.811838 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="364a702e-e6c6-42bc-8cee-4ba19ea2f1ec" path="/var/lib/kubelet/pods/364a702e-e6c6-42bc-8cee-4ba19ea2f1ec/volumes" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.813136 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4050d00b-0256-45c5-9dc4-0ab46956405d" path="/var/lib/kubelet/pods/4050d00b-0256-45c5-9dc4-0ab46956405d/volumes" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.813828 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aa81fa2a-4c27-4e02-9891-5b39348feca9" path="/var/lib/kubelet/pods/aa81fa2a-4c27-4e02-9891-5b39348feca9/volumes" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.815687 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e3c34454-315e-4821-ab25-b0f331a0d521" path="/var/lib/kubelet/pods/e3c34454-315e-4821-ab25-b0f331a0d521/volumes" Mar 13 14:30:52 crc kubenswrapper[4907]: E0313 14:30:52.434931 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 6b0b25079cc099dabe922ac89fa41ae76d2f01797aac9f273b042e242459c574 is running failed: container process not found" containerID="6b0b25079cc099dabe922ac89fa41ae76d2f01797aac9f273b042e242459c574" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Mar 13 14:30:52 crc kubenswrapper[4907]: E0313 14:30:52.435498 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 6b0b25079cc099dabe922ac89fa41ae76d2f01797aac9f273b042e242459c574 is running failed: container process not found" containerID="6b0b25079cc099dabe922ac89fa41ae76d2f01797aac9f273b042e242459c574" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Mar 13 14:30:52 crc kubenswrapper[4907]: E0313 14:30:52.435899 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 6b0b25079cc099dabe922ac89fa41ae76d2f01797aac9f273b042e242459c574 is running failed: container process not found" containerID="6b0b25079cc099dabe922ac89fa41ae76d2f01797aac9f273b042e242459c574" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Mar 13 14:30:52 crc kubenswrapper[4907]: E0313 14:30:52.435928 4907 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 6b0b25079cc099dabe922ac89fa41ae76d2f01797aac9f273b042e242459c574 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-vp6f5" podUID="c1b70392-1240-40d9-8128-e7abe29c8398" containerName="ovsdb-server" Mar 13 14:30:52 crc kubenswrapper[4907]: E0313 14:30:52.438855 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="265b6b6fe5f95195cdc86c65456998ad327fc7ff3a6641300cd6ac00798eff4f" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Mar 13 14:30:52 crc kubenswrapper[4907]: E0313 14:30:52.440327 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="265b6b6fe5f95195cdc86c65456998ad327fc7ff3a6641300cd6ac00798eff4f" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Mar 13 14:30:52 crc kubenswrapper[4907]: E0313 14:30:52.442082 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="265b6b6fe5f95195cdc86c65456998ad327fc7ff3a6641300cd6ac00798eff4f" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Mar 13 14:30:52 crc kubenswrapper[4907]: E0313 14:30:52.442118 4907 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-vp6f5" podUID="c1b70392-1240-40d9-8128-e7abe29c8398" containerName="ovs-vswitchd" Mar 13 14:30:53 crc kubenswrapper[4907]: I0313 14:30:53.551029 4907 generic.go:334] "Generic (PLEG): container finished" podID="001078cc-c01e-4173-9740-4f9ff082c593" containerID="8f41b5a31898b937ed3fab2872e25ea71023ac05d2c81d4552d7e5ae8d96095e" exitCode=0 Mar 13 14:30:53 crc kubenswrapper[4907]: I0313 14:30:53.551207 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7bf587b8d9-pvvbz" event={"ID":"001078cc-c01e-4173-9740-4f9ff082c593","Type":"ContainerDied","Data":"8f41b5a31898b937ed3fab2872e25ea71023ac05d2c81d4552d7e5ae8d96095e"} Mar 13 14:30:53 crc kubenswrapper[4907]: I0313 14:30:53.921246 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7bf587b8d9-pvvbz" Mar 13 14:30:54 crc kubenswrapper[4907]: I0313 14:30:54.064419 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/001078cc-c01e-4173-9740-4f9ff082c593-ovndb-tls-certs\") pod \"001078cc-c01e-4173-9740-4f9ff082c593\" (UID: \"001078cc-c01e-4173-9740-4f9ff082c593\") " Mar 13 14:30:54 crc kubenswrapper[4907]: I0313 14:30:54.064716 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/001078cc-c01e-4173-9740-4f9ff082c593-httpd-config\") pod \"001078cc-c01e-4173-9740-4f9ff082c593\" (UID: \"001078cc-c01e-4173-9740-4f9ff082c593\") " Mar 13 14:30:54 crc kubenswrapper[4907]: I0313 14:30:54.064850 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/001078cc-c01e-4173-9740-4f9ff082c593-public-tls-certs\") pod \"001078cc-c01e-4173-9740-4f9ff082c593\" (UID: \"001078cc-c01e-4173-9740-4f9ff082c593\") " Mar 13 14:30:54 crc kubenswrapper[4907]: I0313 14:30:54.064989 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/001078cc-c01e-4173-9740-4f9ff082c593-internal-tls-certs\") pod \"001078cc-c01e-4173-9740-4f9ff082c593\" (UID: \"001078cc-c01e-4173-9740-4f9ff082c593\") " Mar 13 14:30:54 crc kubenswrapper[4907]: I0313 14:30:54.065109 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/001078cc-c01e-4173-9740-4f9ff082c593-combined-ca-bundle\") pod \"001078cc-c01e-4173-9740-4f9ff082c593\" (UID: \"001078cc-c01e-4173-9740-4f9ff082c593\") " Mar 13 14:30:54 crc kubenswrapper[4907]: I0313 14:30:54.065247 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/001078cc-c01e-4173-9740-4f9ff082c593-config\") pod \"001078cc-c01e-4173-9740-4f9ff082c593\" (UID: \"001078cc-c01e-4173-9740-4f9ff082c593\") " Mar 13 14:30:54 crc kubenswrapper[4907]: I0313 14:30:54.065361 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-67wmh\" (UniqueName: \"kubernetes.io/projected/001078cc-c01e-4173-9740-4f9ff082c593-kube-api-access-67wmh\") pod \"001078cc-c01e-4173-9740-4f9ff082c593\" (UID: \"001078cc-c01e-4173-9740-4f9ff082c593\") " Mar 13 14:30:54 crc kubenswrapper[4907]: I0313 14:30:54.071486 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/001078cc-c01e-4173-9740-4f9ff082c593-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "001078cc-c01e-4173-9740-4f9ff082c593" (UID: "001078cc-c01e-4173-9740-4f9ff082c593"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:54 crc kubenswrapper[4907]: I0313 14:30:54.085541 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/001078cc-c01e-4173-9740-4f9ff082c593-kube-api-access-67wmh" (OuterVolumeSpecName: "kube-api-access-67wmh") pod "001078cc-c01e-4173-9740-4f9ff082c593" (UID: "001078cc-c01e-4173-9740-4f9ff082c593"). InnerVolumeSpecName "kube-api-access-67wmh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:30:54 crc kubenswrapper[4907]: I0313 14:30:54.109048 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/001078cc-c01e-4173-9740-4f9ff082c593-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "001078cc-c01e-4173-9740-4f9ff082c593" (UID: "001078cc-c01e-4173-9740-4f9ff082c593"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:54 crc kubenswrapper[4907]: I0313 14:30:54.109989 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/001078cc-c01e-4173-9740-4f9ff082c593-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "001078cc-c01e-4173-9740-4f9ff082c593" (UID: "001078cc-c01e-4173-9740-4f9ff082c593"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:54 crc kubenswrapper[4907]: I0313 14:30:54.117033 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/001078cc-c01e-4173-9740-4f9ff082c593-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "001078cc-c01e-4173-9740-4f9ff082c593" (UID: "001078cc-c01e-4173-9740-4f9ff082c593"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:54 crc kubenswrapper[4907]: I0313 14:30:54.122270 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/001078cc-c01e-4173-9740-4f9ff082c593-config" (OuterVolumeSpecName: "config") pod "001078cc-c01e-4173-9740-4f9ff082c593" (UID: "001078cc-c01e-4173-9740-4f9ff082c593"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:54 crc kubenswrapper[4907]: I0313 14:30:54.131629 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/001078cc-c01e-4173-9740-4f9ff082c593-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "001078cc-c01e-4173-9740-4f9ff082c593" (UID: "001078cc-c01e-4173-9740-4f9ff082c593"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:54 crc kubenswrapper[4907]: I0313 14:30:54.167260 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/001078cc-c01e-4173-9740-4f9ff082c593-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:54 crc kubenswrapper[4907]: I0313 14:30:54.167304 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-67wmh\" (UniqueName: \"kubernetes.io/projected/001078cc-c01e-4173-9740-4f9ff082c593-kube-api-access-67wmh\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:54 crc kubenswrapper[4907]: I0313 14:30:54.167316 4907 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/001078cc-c01e-4173-9740-4f9ff082c593-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:54 crc kubenswrapper[4907]: I0313 14:30:54.167325 4907 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/001078cc-c01e-4173-9740-4f9ff082c593-httpd-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:54 crc kubenswrapper[4907]: I0313 14:30:54.167335 4907 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/001078cc-c01e-4173-9740-4f9ff082c593-public-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:54 crc kubenswrapper[4907]: I0313 14:30:54.167346 4907 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/001078cc-c01e-4173-9740-4f9ff082c593-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:54 crc kubenswrapper[4907]: I0313 14:30:54.167355 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/001078cc-c01e-4173-9740-4f9ff082c593-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:54 crc kubenswrapper[4907]: I0313 14:30:54.565595 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7bf587b8d9-pvvbz" event={"ID":"001078cc-c01e-4173-9740-4f9ff082c593","Type":"ContainerDied","Data":"502cbc857cf878844cea8a77f9b3d530353b13301737ac4a045e3f57f1cad50e"} Mar 13 14:30:54 crc kubenswrapper[4907]: I0313 14:30:54.565690 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7bf587b8d9-pvvbz" Mar 13 14:30:54 crc kubenswrapper[4907]: I0313 14:30:54.565922 4907 scope.go:117] "RemoveContainer" containerID="3e3eb6f5fca9c3693edb3599f1d3dd3d706131518c37c580379168a8a1b7e5e0" Mar 13 14:30:54 crc kubenswrapper[4907]: I0313 14:30:54.609148 4907 scope.go:117] "RemoveContainer" containerID="8f41b5a31898b937ed3fab2872e25ea71023ac05d2c81d4552d7e5ae8d96095e" Mar 13 14:30:54 crc kubenswrapper[4907]: I0313 14:30:54.610911 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-7bf587b8d9-pvvbz"] Mar 13 14:30:54 crc kubenswrapper[4907]: I0313 14:30:54.617393 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-7bf587b8d9-pvvbz"] Mar 13 14:30:55 crc kubenswrapper[4907]: I0313 14:30:55.792336 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="001078cc-c01e-4173-9740-4f9ff082c593" path="/var/lib/kubelet/pods/001078cc-c01e-4173-9740-4f9ff082c593/volumes" Mar 13 14:30:55 crc kubenswrapper[4907]: E0313 14:30:55.893634 4907 projected.go:288] Couldn't get configMap openstack/swift-storage-config-data: configmap "swift-storage-config-data" not found Mar 13 14:30:55 crc kubenswrapper[4907]: E0313 14:30:55.893676 4907 projected.go:263] Couldn't get secret openstack/swift-conf: secret "swift-conf" not found Mar 13 14:30:55 crc kubenswrapper[4907]: E0313 14:30:55.893686 4907 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Mar 13 14:30:55 crc kubenswrapper[4907]: E0313 14:30:55.893698 4907 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: [configmap "swift-storage-config-data" not found, secret "swift-conf" not found, configmap "swift-ring-files" not found] Mar 13 14:30:55 crc kubenswrapper[4907]: E0313 14:30:55.893758 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/14f05f31-8185-43b0-be69-bcf8d5388ea5-etc-swift podName:14f05f31-8185-43b0-be69-bcf8d5388ea5 nodeName:}" failed. No retries permitted until 2026-03-13 14:31:11.893740019 +0000 UTC m=+1570.793527708 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/14f05f31-8185-43b0-be69-bcf8d5388ea5-etc-swift") pod "swift-storage-0" (UID: "14f05f31-8185-43b0-be69-bcf8d5388ea5") : [configmap "swift-storage-config-data" not found, secret "swift-conf" not found, configmap "swift-ring-files" not found] Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.253471 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-5przm"] Mar 13 14:30:56 crc kubenswrapper[4907]: E0313 14:30:56.253747 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58ab9689-3ab2-42f2-ad56-beb22e29f8da" containerName="mysql-bootstrap" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.253758 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="58ab9689-3ab2-42f2-ad56-beb22e29f8da" containerName="mysql-bootstrap" Mar 13 14:30:56 crc kubenswrapper[4907]: E0313 14:30:56.253766 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="516d1499-3461-4af5-b426-9ae4a711a468" containerName="mariadb-account-create-update" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.253772 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="516d1499-3461-4af5-b426-9ae4a711a468" containerName="mariadb-account-create-update" Mar 13 14:30:56 crc kubenswrapper[4907]: E0313 14:30:56.253780 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4411bd2-c555-433a-9015-f623948b1401" containerName="galera" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.253786 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4411bd2-c555-433a-9015-f623948b1401" containerName="galera" Mar 13 14:30:56 crc kubenswrapper[4907]: E0313 14:30:56.253791 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4251ae36-90ad-41ea-915e-862df60f5c07" containerName="glance-httpd" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.253797 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="4251ae36-90ad-41ea-915e-862df60f5c07" containerName="glance-httpd" Mar 13 14:30:56 crc kubenswrapper[4907]: E0313 14:30:56.253805 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa81fa2a-4c27-4e02-9891-5b39348feca9" containerName="sg-core" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.253811 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa81fa2a-4c27-4e02-9891-5b39348feca9" containerName="sg-core" Mar 13 14:30:56 crc kubenswrapper[4907]: E0313 14:30:56.253820 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa81fa2a-4c27-4e02-9891-5b39348feca9" containerName="proxy-httpd" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.253827 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa81fa2a-4c27-4e02-9891-5b39348feca9" containerName="proxy-httpd" Mar 13 14:30:56 crc kubenswrapper[4907]: E0313 14:30:56.253835 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e41272f-6e7b-47a7-825e-3571f6a1fd07" containerName="nova-metadata-log" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.253841 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e41272f-6e7b-47a7-825e-3571f6a1fd07" containerName="nova-metadata-log" Mar 13 14:30:56 crc kubenswrapper[4907]: E0313 14:30:56.253854 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25b5fafe-65c8-48ba-bc34-83442f8ace4c" containerName="barbican-worker-log" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.253859 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="25b5fafe-65c8-48ba-bc34-83442f8ace4c" containerName="barbican-worker-log" Mar 13 14:30:56 crc kubenswrapper[4907]: E0313 14:30:56.253871 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b2e62fa-af35-4f35-8852-362ea97c17e7" containerName="glance-log" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.253876 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b2e62fa-af35-4f35-8852-362ea97c17e7" containerName="glance-log" Mar 13 14:30:56 crc kubenswrapper[4907]: E0313 14:30:56.253909 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3c34454-315e-4821-ab25-b0f331a0d521" containerName="rabbitmq" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.253915 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3c34454-315e-4821-ab25-b0f331a0d521" containerName="rabbitmq" Mar 13 14:30:56 crc kubenswrapper[4907]: E0313 14:30:56.253922 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="286c82e8-b74a-49d7-a355-ac074aace10e" containerName="rabbitmq" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.253927 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="286c82e8-b74a-49d7-a355-ac074aace10e" containerName="rabbitmq" Mar 13 14:30:56 crc kubenswrapper[4907]: E0313 14:30:56.253938 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cacb281b-ee3b-42b6-8754-d07e6ea6aa4b" containerName="memcached" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.253944 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="cacb281b-ee3b-42b6-8754-d07e6ea6aa4b" containerName="memcached" Mar 13 14:30:56 crc kubenswrapper[4907]: E0313 14:30:56.253955 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b2e62fa-af35-4f35-8852-362ea97c17e7" containerName="glance-httpd" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.253960 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b2e62fa-af35-4f35-8852-362ea97c17e7" containerName="glance-httpd" Mar 13 14:30:56 crc kubenswrapper[4907]: E0313 14:30:56.253970 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3c34454-315e-4821-ab25-b0f331a0d521" containerName="setup-container" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.253976 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3c34454-315e-4821-ab25-b0f331a0d521" containerName="setup-container" Mar 13 14:30:56 crc kubenswrapper[4907]: E0313 14:30:56.253985 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4411bd2-c555-433a-9015-f623948b1401" containerName="mysql-bootstrap" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.253991 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4411bd2-c555-433a-9015-f623948b1401" containerName="mysql-bootstrap" Mar 13 14:30:56 crc kubenswrapper[4907]: E0313 14:30:56.253998 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="286c82e8-b74a-49d7-a355-ac074aace10e" containerName="setup-container" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.254003 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="286c82e8-b74a-49d7-a355-ac074aace10e" containerName="setup-container" Mar 13 14:30:56 crc kubenswrapper[4907]: E0313 14:30:56.254010 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58ab9689-3ab2-42f2-ad56-beb22e29f8da" containerName="galera" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.254016 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="58ab9689-3ab2-42f2-ad56-beb22e29f8da" containerName="galera" Mar 13 14:30:56 crc kubenswrapper[4907]: E0313 14:30:56.254026 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc227d33-3bf1-4b5f-8368-8e6b8a512f7d" containerName="nova-api-log" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.254032 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc227d33-3bf1-4b5f-8368-8e6b8a512f7d" containerName="nova-api-log" Mar 13 14:30:56 crc kubenswrapper[4907]: E0313 14:30:56.254042 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a738a66-54bc-4f5e-86d5-6e3004d8f265" containerName="keystone-api" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.254047 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a738a66-54bc-4f5e-86d5-6e3004d8f265" containerName="keystone-api" Mar 13 14:30:56 crc kubenswrapper[4907]: E0313 14:30:56.254055 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4251ae36-90ad-41ea-915e-862df60f5c07" containerName="glance-log" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.254061 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="4251ae36-90ad-41ea-915e-862df60f5c07" containerName="glance-log" Mar 13 14:30:56 crc kubenswrapper[4907]: E0313 14:30:56.254072 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa81fa2a-4c27-4e02-9891-5b39348feca9" containerName="ceilometer-notification-agent" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.254077 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa81fa2a-4c27-4e02-9891-5b39348feca9" containerName="ceilometer-notification-agent" Mar 13 14:30:56 crc kubenswrapper[4907]: E0313 14:30:56.254085 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65531461-cea2-4b2c-a9b6-8fd9e9bffb27" containerName="barbican-keystone-listener" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.254091 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="65531461-cea2-4b2c-a9b6-8fd9e9bffb27" containerName="barbican-keystone-listener" Mar 13 14:30:56 crc kubenswrapper[4907]: E0313 14:30:56.254102 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b2f62d4-0584-4cc3-81ed-e067d8db23b1" containerName="barbican-api" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.254107 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b2f62d4-0584-4cc3-81ed-e067d8db23b1" containerName="barbican-api" Mar 13 14:30:56 crc kubenswrapper[4907]: E0313 14:30:56.254116 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="001078cc-c01e-4173-9740-4f9ff082c593" containerName="neutron-api" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.254121 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="001078cc-c01e-4173-9740-4f9ff082c593" containerName="neutron-api" Mar 13 14:30:56 crc kubenswrapper[4907]: E0313 14:30:56.254129 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f8119b2-e38a-494d-967f-5198b83512c7" containerName="ovn-controller" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.254135 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f8119b2-e38a-494d-967f-5198b83512c7" containerName="ovn-controller" Mar 13 14:30:56 crc kubenswrapper[4907]: E0313 14:30:56.254142 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76" containerName="cinder-api-log" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.254147 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76" containerName="cinder-api-log" Mar 13 14:30:56 crc kubenswrapper[4907]: E0313 14:30:56.254157 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b2f62d4-0584-4cc3-81ed-e067d8db23b1" containerName="barbican-api-log" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.254163 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b2f62d4-0584-4cc3-81ed-e067d8db23b1" containerName="barbican-api-log" Mar 13 14:30:56 crc kubenswrapper[4907]: E0313 14:30:56.254173 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25c1c7e6-a149-4a8e-aead-3ebdc7e56851" containerName="nova-cell1-conductor-conductor" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.254179 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="25c1c7e6-a149-4a8e-aead-3ebdc7e56851" containerName="nova-cell1-conductor-conductor" Mar 13 14:30:56 crc kubenswrapper[4907]: E0313 14:30:56.254187 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e41272f-6e7b-47a7-825e-3571f6a1fd07" containerName="nova-metadata-metadata" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.254192 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e41272f-6e7b-47a7-825e-3571f6a1fd07" containerName="nova-metadata-metadata" Mar 13 14:30:56 crc kubenswrapper[4907]: E0313 14:30:56.254200 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="001078cc-c01e-4173-9740-4f9ff082c593" containerName="neutron-httpd" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.254206 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="001078cc-c01e-4173-9740-4f9ff082c593" containerName="neutron-httpd" Mar 13 14:30:56 crc kubenswrapper[4907]: E0313 14:30:56.254213 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3ee069a-41fb-4cab-b650-9bb9c03ac271" containerName="placement-log" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.254218 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3ee069a-41fb-4cab-b650-9bb9c03ac271" containerName="placement-log" Mar 13 14:30:56 crc kubenswrapper[4907]: E0313 14:30:56.254228 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65531461-cea2-4b2c-a9b6-8fd9e9bffb27" containerName="barbican-keystone-listener-log" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.254233 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="65531461-cea2-4b2c-a9b6-8fd9e9bffb27" containerName="barbican-keystone-listener-log" Mar 13 14:30:56 crc kubenswrapper[4907]: E0313 14:30:56.254241 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76" containerName="cinder-api" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.254247 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76" containerName="cinder-api" Mar 13 14:30:56 crc kubenswrapper[4907]: E0313 14:30:56.254253 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3ee069a-41fb-4cab-b650-9bb9c03ac271" containerName="placement-api" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.254259 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3ee069a-41fb-4cab-b650-9bb9c03ac271" containerName="placement-api" Mar 13 14:30:56 crc kubenswrapper[4907]: E0313 14:30:56.254265 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4050d00b-0256-45c5-9dc4-0ab46956405d" containerName="nova-scheduler-scheduler" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.254270 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="4050d00b-0256-45c5-9dc4-0ab46956405d" containerName="nova-scheduler-scheduler" Mar 13 14:30:56 crc kubenswrapper[4907]: E0313 14:30:56.254278 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25b5fafe-65c8-48ba-bc34-83442f8ace4c" containerName="barbican-worker" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.254284 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="25b5fafe-65c8-48ba-bc34-83442f8ace4c" containerName="barbican-worker" Mar 13 14:30:56 crc kubenswrapper[4907]: E0313 14:30:56.254293 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c49286b1-2e9f-4de6-b7de-2e952f3ad607" containerName="kube-state-metrics" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.254299 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="c49286b1-2e9f-4de6-b7de-2e952f3ad607" containerName="kube-state-metrics" Mar 13 14:30:56 crc kubenswrapper[4907]: E0313 14:30:56.254307 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="364a702e-e6c6-42bc-8cee-4ba19ea2f1ec" containerName="nova-cell0-conductor-conductor" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.254313 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="364a702e-e6c6-42bc-8cee-4ba19ea2f1ec" containerName="nova-cell0-conductor-conductor" Mar 13 14:30:56 crc kubenswrapper[4907]: E0313 14:30:56.254323 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa81fa2a-4c27-4e02-9891-5b39348feca9" containerName="ceilometer-central-agent" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.254328 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa81fa2a-4c27-4e02-9891-5b39348feca9" containerName="ceilometer-central-agent" Mar 13 14:30:56 crc kubenswrapper[4907]: E0313 14:30:56.254336 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc227d33-3bf1-4b5f-8368-8e6b8a512f7d" containerName="nova-api-api" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.254342 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc227d33-3bf1-4b5f-8368-8e6b8a512f7d" containerName="nova-api-api" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.254457 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="001078cc-c01e-4173-9740-4f9ff082c593" containerName="neutron-api" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.254467 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="0b2e62fa-af35-4f35-8852-362ea97c17e7" containerName="glance-httpd" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.254474 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e41272f-6e7b-47a7-825e-3571f6a1fd07" containerName="nova-metadata-metadata" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.254482 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="25c1c7e6-a149-4a8e-aead-3ebdc7e56851" containerName="nova-cell1-conductor-conductor" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.254492 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="cacb281b-ee3b-42b6-8754-d07e6ea6aa4b" containerName="memcached" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.254500 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3ee069a-41fb-4cab-b650-9bb9c03ac271" containerName="placement-log" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.254507 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc227d33-3bf1-4b5f-8368-8e6b8a512f7d" containerName="nova-api-log" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.254514 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b2f62d4-0584-4cc3-81ed-e067d8db23b1" containerName="barbican-api-log" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.254523 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="001078cc-c01e-4173-9740-4f9ff082c593" containerName="neutron-httpd" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.254533 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="65531461-cea2-4b2c-a9b6-8fd9e9bffb27" containerName="barbican-keystone-listener" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.254540 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="6f8119b2-e38a-494d-967f-5198b83512c7" containerName="ovn-controller" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.254548 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa81fa2a-4c27-4e02-9891-5b39348feca9" containerName="sg-core" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.254556 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc227d33-3bf1-4b5f-8368-8e6b8a512f7d" containerName="nova-api-api" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.254563 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="4251ae36-90ad-41ea-915e-862df60f5c07" containerName="glance-log" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.254570 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="364a702e-e6c6-42bc-8cee-4ba19ea2f1ec" containerName="nova-cell0-conductor-conductor" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.254579 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="4050d00b-0256-45c5-9dc4-0ab46956405d" containerName="nova-scheduler-scheduler" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.254586 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76" containerName="cinder-api-log" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.254594 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="c49286b1-2e9f-4de6-b7de-2e952f3ad607" containerName="kube-state-metrics" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.254603 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="0b2e62fa-af35-4f35-8852-362ea97c17e7" containerName="glance-log" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.254611 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="58ab9689-3ab2-42f2-ad56-beb22e29f8da" containerName="galera" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.254620 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa81fa2a-4c27-4e02-9891-5b39348feca9" containerName="ceilometer-notification-agent" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.254627 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="516d1499-3461-4af5-b426-9ae4a711a468" containerName="mariadb-account-create-update" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.254634 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa81fa2a-4c27-4e02-9891-5b39348feca9" containerName="ceilometer-central-agent" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.254644 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="c4411bd2-c555-433a-9015-f623948b1401" containerName="galera" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.254653 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="25b5fafe-65c8-48ba-bc34-83442f8ace4c" containerName="barbican-worker" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.254661 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="4251ae36-90ad-41ea-915e-862df60f5c07" containerName="glance-httpd" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.254669 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="25b5fafe-65c8-48ba-bc34-83442f8ace4c" containerName="barbican-worker-log" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.254677 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b2f62d4-0584-4cc3-81ed-e067d8db23b1" containerName="barbican-api" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.254685 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a738a66-54bc-4f5e-86d5-6e3004d8f265" containerName="keystone-api" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.254694 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="65531461-cea2-4b2c-a9b6-8fd9e9bffb27" containerName="barbican-keystone-listener-log" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.254702 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3c34454-315e-4821-ab25-b0f331a0d521" containerName="rabbitmq" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.254712 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76" containerName="cinder-api" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.254721 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3ee069a-41fb-4cab-b650-9bb9c03ac271" containerName="placement-api" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.254728 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa81fa2a-4c27-4e02-9891-5b39348feca9" containerName="proxy-httpd" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.254736 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e41272f-6e7b-47a7-825e-3571f6a1fd07" containerName="nova-metadata-log" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.254744 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="286c82e8-b74a-49d7-a355-ac074aace10e" containerName="rabbitmq" Mar 13 14:30:56 crc kubenswrapper[4907]: E0313 14:30:56.254867 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="516d1499-3461-4af5-b426-9ae4a711a468" containerName="mariadb-account-create-update" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.254874 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="516d1499-3461-4af5-b426-9ae4a711a468" containerName="mariadb-account-create-update" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.255006 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="516d1499-3461-4af5-b426-9ae4a711a468" containerName="mariadb-account-create-update" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.255650 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5przm" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.274928 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5przm"] Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.400961 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1e43e4fe-9084-48bb-a931-33e2fc5ecaaa-catalog-content\") pod \"redhat-operators-5przm\" (UID: \"1e43e4fe-9084-48bb-a931-33e2fc5ecaaa\") " pod="openshift-marketplace/redhat-operators-5przm" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.401031 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g8dvd\" (UniqueName: \"kubernetes.io/projected/1e43e4fe-9084-48bb-a931-33e2fc5ecaaa-kube-api-access-g8dvd\") pod \"redhat-operators-5przm\" (UID: \"1e43e4fe-9084-48bb-a931-33e2fc5ecaaa\") " pod="openshift-marketplace/redhat-operators-5przm" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.401189 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1e43e4fe-9084-48bb-a931-33e2fc5ecaaa-utilities\") pod \"redhat-operators-5przm\" (UID: \"1e43e4fe-9084-48bb-a931-33e2fc5ecaaa\") " pod="openshift-marketplace/redhat-operators-5przm" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.503025 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1e43e4fe-9084-48bb-a931-33e2fc5ecaaa-utilities\") pod \"redhat-operators-5przm\" (UID: \"1e43e4fe-9084-48bb-a931-33e2fc5ecaaa\") " pod="openshift-marketplace/redhat-operators-5przm" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.503152 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1e43e4fe-9084-48bb-a931-33e2fc5ecaaa-catalog-content\") pod \"redhat-operators-5przm\" (UID: \"1e43e4fe-9084-48bb-a931-33e2fc5ecaaa\") " pod="openshift-marketplace/redhat-operators-5przm" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.503224 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g8dvd\" (UniqueName: \"kubernetes.io/projected/1e43e4fe-9084-48bb-a931-33e2fc5ecaaa-kube-api-access-g8dvd\") pod \"redhat-operators-5przm\" (UID: \"1e43e4fe-9084-48bb-a931-33e2fc5ecaaa\") " pod="openshift-marketplace/redhat-operators-5przm" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.503567 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1e43e4fe-9084-48bb-a931-33e2fc5ecaaa-utilities\") pod \"redhat-operators-5przm\" (UID: \"1e43e4fe-9084-48bb-a931-33e2fc5ecaaa\") " pod="openshift-marketplace/redhat-operators-5przm" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.503700 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1e43e4fe-9084-48bb-a931-33e2fc5ecaaa-catalog-content\") pod \"redhat-operators-5przm\" (UID: \"1e43e4fe-9084-48bb-a931-33e2fc5ecaaa\") " pod="openshift-marketplace/redhat-operators-5przm" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.528768 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g8dvd\" (UniqueName: \"kubernetes.io/projected/1e43e4fe-9084-48bb-a931-33e2fc5ecaaa-kube-api-access-g8dvd\") pod \"redhat-operators-5przm\" (UID: \"1e43e4fe-9084-48bb-a931-33e2fc5ecaaa\") " pod="openshift-marketplace/redhat-operators-5przm" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.576319 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5przm" Mar 13 14:30:57 crc kubenswrapper[4907]: I0313 14:30:57.071761 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5przm"] Mar 13 14:30:57 crc kubenswrapper[4907]: E0313 14:30:57.434502 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 6b0b25079cc099dabe922ac89fa41ae76d2f01797aac9f273b042e242459c574 is running failed: container process not found" containerID="6b0b25079cc099dabe922ac89fa41ae76d2f01797aac9f273b042e242459c574" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Mar 13 14:30:57 crc kubenswrapper[4907]: E0313 14:30:57.434964 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 6b0b25079cc099dabe922ac89fa41ae76d2f01797aac9f273b042e242459c574 is running failed: container process not found" containerID="6b0b25079cc099dabe922ac89fa41ae76d2f01797aac9f273b042e242459c574" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Mar 13 14:30:57 crc kubenswrapper[4907]: E0313 14:30:57.435334 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 6b0b25079cc099dabe922ac89fa41ae76d2f01797aac9f273b042e242459c574 is running failed: container process not found" containerID="6b0b25079cc099dabe922ac89fa41ae76d2f01797aac9f273b042e242459c574" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Mar 13 14:30:57 crc kubenswrapper[4907]: E0313 14:30:57.435406 4907 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 6b0b25079cc099dabe922ac89fa41ae76d2f01797aac9f273b042e242459c574 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-vp6f5" podUID="c1b70392-1240-40d9-8128-e7abe29c8398" containerName="ovsdb-server" Mar 13 14:30:57 crc kubenswrapper[4907]: E0313 14:30:57.435633 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="265b6b6fe5f95195cdc86c65456998ad327fc7ff3a6641300cd6ac00798eff4f" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Mar 13 14:30:57 crc kubenswrapper[4907]: E0313 14:30:57.437290 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="265b6b6fe5f95195cdc86c65456998ad327fc7ff3a6641300cd6ac00798eff4f" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Mar 13 14:30:57 crc kubenswrapper[4907]: E0313 14:30:57.438597 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="265b6b6fe5f95195cdc86c65456998ad327fc7ff3a6641300cd6ac00798eff4f" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Mar 13 14:30:57 crc kubenswrapper[4907]: E0313 14:30:57.438670 4907 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-vp6f5" podUID="c1b70392-1240-40d9-8128-e7abe29c8398" containerName="ovs-vswitchd" Mar 13 14:30:57 crc kubenswrapper[4907]: I0313 14:30:57.593644 4907 generic.go:334] "Generic (PLEG): container finished" podID="1e43e4fe-9084-48bb-a931-33e2fc5ecaaa" containerID="4b6734fa58a17ab6193b32c4603463a610dfd0ffb321199337a7a16643dced00" exitCode=0 Mar 13 14:30:57 crc kubenswrapper[4907]: I0313 14:30:57.593714 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5przm" event={"ID":"1e43e4fe-9084-48bb-a931-33e2fc5ecaaa","Type":"ContainerDied","Data":"4b6734fa58a17ab6193b32c4603463a610dfd0ffb321199337a7a16643dced00"} Mar 13 14:30:57 crc kubenswrapper[4907]: I0313 14:30:57.593758 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5przm" event={"ID":"1e43e4fe-9084-48bb-a931-33e2fc5ecaaa","Type":"ContainerStarted","Data":"420aae8884cc744f6583795347d67462ea3274678bbdefc76fe718c1b6545b81"} Mar 13 14:30:58 crc kubenswrapper[4907]: I0313 14:30:58.608059 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5przm" event={"ID":"1e43e4fe-9084-48bb-a931-33e2fc5ecaaa","Type":"ContainerStarted","Data":"057b7d13b67e10cac7b4063457b964d60358da6d919d6ef053cd4aaf19089d04"} Mar 13 14:30:59 crc kubenswrapper[4907]: I0313 14:30:59.619952 4907 generic.go:334] "Generic (PLEG): container finished" podID="1e43e4fe-9084-48bb-a931-33e2fc5ecaaa" containerID="057b7d13b67e10cac7b4063457b964d60358da6d919d6ef053cd4aaf19089d04" exitCode=0 Mar 13 14:30:59 crc kubenswrapper[4907]: I0313 14:30:59.620004 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5przm" event={"ID":"1e43e4fe-9084-48bb-a931-33e2fc5ecaaa","Type":"ContainerDied","Data":"057b7d13b67e10cac7b4063457b964d60358da6d919d6ef053cd4aaf19089d04"} Mar 13 14:31:00 crc kubenswrapper[4907]: I0313 14:31:00.632764 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5przm" event={"ID":"1e43e4fe-9084-48bb-a931-33e2fc5ecaaa","Type":"ContainerStarted","Data":"b4fc4a9c98855f93f56a77779ad655ec33841f4a3e7acba65768c38b4a3718bf"} Mar 13 14:31:00 crc kubenswrapper[4907]: I0313 14:31:00.656609 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-5przm" podStartSLOduration=2.051220609 podStartE2EDuration="4.656585792s" podCreationTimestamp="2026-03-13 14:30:56 +0000 UTC" firstStartedPulling="2026-03-13 14:30:57.5980306 +0000 UTC m=+1556.497818289" lastFinishedPulling="2026-03-13 14:31:00.203395743 +0000 UTC m=+1559.103183472" observedRunningTime="2026-03-13 14:31:00.652823671 +0000 UTC m=+1559.552611370" watchObservedRunningTime="2026-03-13 14:31:00.656585792 +0000 UTC m=+1559.556373491" Mar 13 14:31:02 crc kubenswrapper[4907]: E0313 14:31:02.433422 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 6b0b25079cc099dabe922ac89fa41ae76d2f01797aac9f273b042e242459c574 is running failed: container process not found" containerID="6b0b25079cc099dabe922ac89fa41ae76d2f01797aac9f273b042e242459c574" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Mar 13 14:31:02 crc kubenswrapper[4907]: E0313 14:31:02.434184 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 6b0b25079cc099dabe922ac89fa41ae76d2f01797aac9f273b042e242459c574 is running failed: container process not found" containerID="6b0b25079cc099dabe922ac89fa41ae76d2f01797aac9f273b042e242459c574" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Mar 13 14:31:02 crc kubenswrapper[4907]: E0313 14:31:02.434741 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 6b0b25079cc099dabe922ac89fa41ae76d2f01797aac9f273b042e242459c574 is running failed: container process not found" containerID="6b0b25079cc099dabe922ac89fa41ae76d2f01797aac9f273b042e242459c574" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Mar 13 14:31:02 crc kubenswrapper[4907]: E0313 14:31:02.434773 4907 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 6b0b25079cc099dabe922ac89fa41ae76d2f01797aac9f273b042e242459c574 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-vp6f5" podUID="c1b70392-1240-40d9-8128-e7abe29c8398" containerName="ovsdb-server" Mar 13 14:31:02 crc kubenswrapper[4907]: E0313 14:31:02.436529 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="265b6b6fe5f95195cdc86c65456998ad327fc7ff3a6641300cd6ac00798eff4f" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Mar 13 14:31:02 crc kubenswrapper[4907]: E0313 14:31:02.437959 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="265b6b6fe5f95195cdc86c65456998ad327fc7ff3a6641300cd6ac00798eff4f" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Mar 13 14:31:02 crc kubenswrapper[4907]: E0313 14:31:02.439419 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="265b6b6fe5f95195cdc86c65456998ad327fc7ff3a6641300cd6ac00798eff4f" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Mar 13 14:31:02 crc kubenswrapper[4907]: E0313 14:31:02.439473 4907 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-vp6f5" podUID="c1b70392-1240-40d9-8128-e7abe29c8398" containerName="ovs-vswitchd" Mar 13 14:31:06 crc kubenswrapper[4907]: I0313 14:31:06.576792 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-5przm" Mar 13 14:31:06 crc kubenswrapper[4907]: I0313 14:31:06.577154 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-5przm" Mar 13 14:31:07 crc kubenswrapper[4907]: E0313 14:31:07.434220 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 6b0b25079cc099dabe922ac89fa41ae76d2f01797aac9f273b042e242459c574 is running failed: container process not found" containerID="6b0b25079cc099dabe922ac89fa41ae76d2f01797aac9f273b042e242459c574" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Mar 13 14:31:07 crc kubenswrapper[4907]: E0313 14:31:07.435462 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 6b0b25079cc099dabe922ac89fa41ae76d2f01797aac9f273b042e242459c574 is running failed: container process not found" containerID="6b0b25079cc099dabe922ac89fa41ae76d2f01797aac9f273b042e242459c574" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Mar 13 14:31:07 crc kubenswrapper[4907]: E0313 14:31:07.435828 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="265b6b6fe5f95195cdc86c65456998ad327fc7ff3a6641300cd6ac00798eff4f" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Mar 13 14:31:07 crc kubenswrapper[4907]: E0313 14:31:07.435970 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 6b0b25079cc099dabe922ac89fa41ae76d2f01797aac9f273b042e242459c574 is running failed: container process not found" containerID="6b0b25079cc099dabe922ac89fa41ae76d2f01797aac9f273b042e242459c574" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Mar 13 14:31:07 crc kubenswrapper[4907]: E0313 14:31:07.436064 4907 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 6b0b25079cc099dabe922ac89fa41ae76d2f01797aac9f273b042e242459c574 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-vp6f5" podUID="c1b70392-1240-40d9-8128-e7abe29c8398" containerName="ovsdb-server" Mar 13 14:31:07 crc kubenswrapper[4907]: E0313 14:31:07.438070 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="265b6b6fe5f95195cdc86c65456998ad327fc7ff3a6641300cd6ac00798eff4f" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Mar 13 14:31:07 crc kubenswrapper[4907]: E0313 14:31:07.440463 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="265b6b6fe5f95195cdc86c65456998ad327fc7ff3a6641300cd6ac00798eff4f" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Mar 13 14:31:07 crc kubenswrapper[4907]: E0313 14:31:07.440525 4907 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-vp6f5" podUID="c1b70392-1240-40d9-8128-e7abe29c8398" containerName="ovs-vswitchd" Mar 13 14:31:07 crc kubenswrapper[4907]: I0313 14:31:07.621342 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-5przm" podUID="1e43e4fe-9084-48bb-a931-33e2fc5ecaaa" containerName="registry-server" probeResult="failure" output=< Mar 13 14:31:07 crc kubenswrapper[4907]: timeout: failed to connect service ":50051" within 1s Mar 13 14:31:07 crc kubenswrapper[4907]: > Mar 13 14:31:11 crc kubenswrapper[4907]: E0313 14:31:11.939458 4907 projected.go:288] Couldn't get configMap openstack/swift-storage-config-data: configmap "swift-storage-config-data" not found Mar 13 14:31:11 crc kubenswrapper[4907]: E0313 14:31:11.940087 4907 projected.go:263] Couldn't get secret openstack/swift-conf: secret "swift-conf" not found Mar 13 14:31:11 crc kubenswrapper[4907]: E0313 14:31:11.940101 4907 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Mar 13 14:31:11 crc kubenswrapper[4907]: E0313 14:31:11.940112 4907 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: [configmap "swift-storage-config-data" not found, secret "swift-conf" not found, configmap "swift-ring-files" not found] Mar 13 14:31:11 crc kubenswrapper[4907]: E0313 14:31:11.940223 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/14f05f31-8185-43b0-be69-bcf8d5388ea5-etc-swift podName:14f05f31-8185-43b0-be69-bcf8d5388ea5 nodeName:}" failed. No retries permitted until 2026-03-13 14:31:43.940204977 +0000 UTC m=+1602.839992666 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/14f05f31-8185-43b0-be69-bcf8d5388ea5-etc-swift") pod "swift-storage-0" (UID: "14f05f31-8185-43b0-be69-bcf8d5388ea5") : [configmap "swift-storage-config-data" not found, secret "swift-conf" not found, configmap "swift-ring-files" not found] Mar 13 14:31:12 crc kubenswrapper[4907]: E0313 14:31:12.435433 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 6b0b25079cc099dabe922ac89fa41ae76d2f01797aac9f273b042e242459c574 is running failed: container process not found" containerID="6b0b25079cc099dabe922ac89fa41ae76d2f01797aac9f273b042e242459c574" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Mar 13 14:31:12 crc kubenswrapper[4907]: E0313 14:31:12.436583 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="265b6b6fe5f95195cdc86c65456998ad327fc7ff3a6641300cd6ac00798eff4f" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Mar 13 14:31:12 crc kubenswrapper[4907]: E0313 14:31:12.436627 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 6b0b25079cc099dabe922ac89fa41ae76d2f01797aac9f273b042e242459c574 is running failed: container process not found" containerID="6b0b25079cc099dabe922ac89fa41ae76d2f01797aac9f273b042e242459c574" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Mar 13 14:31:12 crc kubenswrapper[4907]: E0313 14:31:12.437131 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 6b0b25079cc099dabe922ac89fa41ae76d2f01797aac9f273b042e242459c574 is running failed: container process not found" containerID="6b0b25079cc099dabe922ac89fa41ae76d2f01797aac9f273b042e242459c574" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Mar 13 14:31:12 crc kubenswrapper[4907]: E0313 14:31:12.437165 4907 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 6b0b25079cc099dabe922ac89fa41ae76d2f01797aac9f273b042e242459c574 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-vp6f5" podUID="c1b70392-1240-40d9-8128-e7abe29c8398" containerName="ovsdb-server" Mar 13 14:31:12 crc kubenswrapper[4907]: E0313 14:31:12.438263 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="265b6b6fe5f95195cdc86c65456998ad327fc7ff3a6641300cd6ac00798eff4f" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Mar 13 14:31:12 crc kubenswrapper[4907]: E0313 14:31:12.440493 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="265b6b6fe5f95195cdc86c65456998ad327fc7ff3a6641300cd6ac00798eff4f" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Mar 13 14:31:12 crc kubenswrapper[4907]: E0313 14:31:12.440528 4907 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-vp6f5" podUID="c1b70392-1240-40d9-8128-e7abe29c8398" containerName="ovs-vswitchd" Mar 13 14:31:12 crc kubenswrapper[4907]: I0313 14:31:12.540830 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Mar 13 14:31:12 crc kubenswrapper[4907]: I0313 14:31:12.650746 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14f05f31-8185-43b0-be69-bcf8d5388ea5-combined-ca-bundle\") pod \"14f05f31-8185-43b0-be69-bcf8d5388ea5\" (UID: \"14f05f31-8185-43b0-be69-bcf8d5388ea5\") " Mar 13 14:31:12 crc kubenswrapper[4907]: I0313 14:31:12.650800 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/14f05f31-8185-43b0-be69-bcf8d5388ea5-etc-swift\") pod \"14f05f31-8185-43b0-be69-bcf8d5388ea5\" (UID: \"14f05f31-8185-43b0-be69-bcf8d5388ea5\") " Mar 13 14:31:12 crc kubenswrapper[4907]: I0313 14:31:12.650831 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/14f05f31-8185-43b0-be69-bcf8d5388ea5-lock\") pod \"14f05f31-8185-43b0-be69-bcf8d5388ea5\" (UID: \"14f05f31-8185-43b0-be69-bcf8d5388ea5\") " Mar 13 14:31:12 crc kubenswrapper[4907]: I0313 14:31:12.650897 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swift\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"14f05f31-8185-43b0-be69-bcf8d5388ea5\" (UID: \"14f05f31-8185-43b0-be69-bcf8d5388ea5\") " Mar 13 14:31:12 crc kubenswrapper[4907]: I0313 14:31:12.650978 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/14f05f31-8185-43b0-be69-bcf8d5388ea5-cache\") pod \"14f05f31-8185-43b0-be69-bcf8d5388ea5\" (UID: \"14f05f31-8185-43b0-be69-bcf8d5388ea5\") " Mar 13 14:31:12 crc kubenswrapper[4907]: I0313 14:31:12.651000 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7ps87\" (UniqueName: \"kubernetes.io/projected/14f05f31-8185-43b0-be69-bcf8d5388ea5-kube-api-access-7ps87\") pod \"14f05f31-8185-43b0-be69-bcf8d5388ea5\" (UID: \"14f05f31-8185-43b0-be69-bcf8d5388ea5\") " Mar 13 14:31:12 crc kubenswrapper[4907]: I0313 14:31:12.651547 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/14f05f31-8185-43b0-be69-bcf8d5388ea5-lock" (OuterVolumeSpecName: "lock") pod "14f05f31-8185-43b0-be69-bcf8d5388ea5" (UID: "14f05f31-8185-43b0-be69-bcf8d5388ea5"). InnerVolumeSpecName "lock". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:31:12 crc kubenswrapper[4907]: I0313 14:31:12.652103 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/14f05f31-8185-43b0-be69-bcf8d5388ea5-cache" (OuterVolumeSpecName: "cache") pod "14f05f31-8185-43b0-be69-bcf8d5388ea5" (UID: "14f05f31-8185-43b0-be69-bcf8d5388ea5"). InnerVolumeSpecName "cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:31:12 crc kubenswrapper[4907]: I0313 14:31:12.656306 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/14f05f31-8185-43b0-be69-bcf8d5388ea5-kube-api-access-7ps87" (OuterVolumeSpecName: "kube-api-access-7ps87") pod "14f05f31-8185-43b0-be69-bcf8d5388ea5" (UID: "14f05f31-8185-43b0-be69-bcf8d5388ea5"). InnerVolumeSpecName "kube-api-access-7ps87". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:31:12 crc kubenswrapper[4907]: I0313 14:31:12.656489 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "swift") pod "14f05f31-8185-43b0-be69-bcf8d5388ea5" (UID: "14f05f31-8185-43b0-be69-bcf8d5388ea5"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Mar 13 14:31:12 crc kubenswrapper[4907]: I0313 14:31:12.656995 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/14f05f31-8185-43b0-be69-bcf8d5388ea5-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "14f05f31-8185-43b0-be69-bcf8d5388ea5" (UID: "14f05f31-8185-43b0-be69-bcf8d5388ea5"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:31:12 crc kubenswrapper[4907]: I0313 14:31:12.753300 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7ps87\" (UniqueName: \"kubernetes.io/projected/14f05f31-8185-43b0-be69-bcf8d5388ea5-kube-api-access-7ps87\") on node \"crc\" DevicePath \"\"" Mar 13 14:31:12 crc kubenswrapper[4907]: I0313 14:31:12.753409 4907 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/14f05f31-8185-43b0-be69-bcf8d5388ea5-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 13 14:31:12 crc kubenswrapper[4907]: I0313 14:31:12.753430 4907 reconciler_common.go:293] "Volume detached for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/14f05f31-8185-43b0-be69-bcf8d5388ea5-lock\") on node \"crc\" DevicePath \"\"" Mar 13 14:31:12 crc kubenswrapper[4907]: I0313 14:31:12.753503 4907 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Mar 13 14:31:12 crc kubenswrapper[4907]: I0313 14:31:12.753523 4907 reconciler_common.go:293] "Volume detached for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/14f05f31-8185-43b0-be69-bcf8d5388ea5-cache\") on node \"crc\" DevicePath \"\"" Mar 13 14:31:12 crc kubenswrapper[4907]: I0313 14:31:12.769483 4907 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Mar 13 14:31:12 crc kubenswrapper[4907]: I0313 14:31:12.770624 4907 generic.go:334] "Generic (PLEG): container finished" podID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerID="0696dc2db616f4616e43dd171ba757d11bce09169b35b516964d1af58bad39ea" exitCode=137 Mar 13 14:31:12 crc kubenswrapper[4907]: I0313 14:31:12.770662 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"14f05f31-8185-43b0-be69-bcf8d5388ea5","Type":"ContainerDied","Data":"0696dc2db616f4616e43dd171ba757d11bce09169b35b516964d1af58bad39ea"} Mar 13 14:31:12 crc kubenswrapper[4907]: I0313 14:31:12.770689 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"14f05f31-8185-43b0-be69-bcf8d5388ea5","Type":"ContainerDied","Data":"8ab726bec0bb062d6ac5d70e873cfd78800b9921e7163b9cb0ca74467cc3c7f4"} Mar 13 14:31:12 crc kubenswrapper[4907]: I0313 14:31:12.770708 4907 scope.go:117] "RemoveContainer" containerID="0696dc2db616f4616e43dd171ba757d11bce09169b35b516964d1af58bad39ea" Mar 13 14:31:12 crc kubenswrapper[4907]: I0313 14:31:12.770757 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Mar 13 14:31:12 crc kubenswrapper[4907]: I0313 14:31:12.854381 4907 scope.go:117] "RemoveContainer" containerID="1e5ba68b21b276693f03089f51ca3955dd88adcaacae52727d03a999b543d411" Mar 13 14:31:12 crc kubenswrapper[4907]: I0313 14:31:12.855368 4907 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Mar 13 14:31:12 crc kubenswrapper[4907]: I0313 14:31:12.879977 4907 scope.go:117] "RemoveContainer" containerID="8765bf4cde625366ddd57fa87118a09b18b3fbbf1d2ded31c12344e9c08dadb2" Mar 13 14:31:12 crc kubenswrapper[4907]: I0313 14:31:12.913744 4907 scope.go:117] "RemoveContainer" containerID="6f5cf167a5953dd1bdb0f436b238215d64975f16e188a1dacd546c8d2adfba14" Mar 13 14:31:12 crc kubenswrapper[4907]: I0313 14:31:12.943539 4907 scope.go:117] "RemoveContainer" containerID="c1b4fe0c395990393cfbace1600ea9872fde43535722ed738332ba0b26f178a4" Mar 13 14:31:12 crc kubenswrapper[4907]: I0313 14:31:12.972200 4907 scope.go:117] "RemoveContainer" containerID="5a6b392bf5f4cbd99ba2fc751327322f8960f4c73aba15cb9a501d4c9c3edd94" Mar 13 14:31:12 crc kubenswrapper[4907]: I0313 14:31:12.974841 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14f05f31-8185-43b0-be69-bcf8d5388ea5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "14f05f31-8185-43b0-be69-bcf8d5388ea5" (UID: "14f05f31-8185-43b0-be69-bcf8d5388ea5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:31:12 crc kubenswrapper[4907]: I0313 14:31:12.997393 4907 scope.go:117] "RemoveContainer" containerID="c81c73db839bc006e1b7fdce225dd05a9a50870ea60ca389ea146e1e69d195b7" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.021719 4907 scope.go:117] "RemoveContainer" containerID="dbcb2b472a15fbcc053532418c2f7a654827914050e539781030bc0749d6218f" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.048177 4907 scope.go:117] "RemoveContainer" containerID="7a94b1f20e2cb964aaf2fe5c9f23f0cd193291a14edd5411717f7cad81ed9808" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.061529 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14f05f31-8185-43b0-be69-bcf8d5388ea5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.068156 4907 scope.go:117] "RemoveContainer" containerID="2bc2275a73f47c1a27db775d1cc1c3ae9c20c01f9ddad7a35bf652467b66ea80" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.104748 4907 scope.go:117] "RemoveContainer" containerID="b7f60cadabc1fd2d193073423cce7296e305ff6325001c7ec9823dd267a1787a" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.138490 4907 scope.go:117] "RemoveContainer" containerID="c7370f4c24ed7b82c59bc2c1e3ae152c65420b5dd6227a66eb64b5b83832e410" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.145524 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-storage-0"] Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.152819 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-storage-0"] Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.193453 4907 scope.go:117] "RemoveContainer" containerID="f8bc1f3dd2242db8254b664f00b3789940491fa59b38a7fc033c225b233b776e" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.219025 4907 scope.go:117] "RemoveContainer" containerID="9b6f16a5610e8f2c08a40a29bfba0c00e1ff883cb29b5290604713aad1dab566" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.274519 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-vp6f5_c1b70392-1240-40d9-8128-e7abe29c8398/ovs-vswitchd/0.log" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.282419 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-vp6f5" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.287127 4907 scope.go:117] "RemoveContainer" containerID="8168297a51ad087d7375035164e54e4a0d24e6fd56cf85e922e11d80f7f3d091" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.308506 4907 scope.go:117] "RemoveContainer" containerID="0696dc2db616f4616e43dd171ba757d11bce09169b35b516964d1af58bad39ea" Mar 13 14:31:13 crc kubenswrapper[4907]: E0313 14:31:13.309048 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0696dc2db616f4616e43dd171ba757d11bce09169b35b516964d1af58bad39ea\": container with ID starting with 0696dc2db616f4616e43dd171ba757d11bce09169b35b516964d1af58bad39ea not found: ID does not exist" containerID="0696dc2db616f4616e43dd171ba757d11bce09169b35b516964d1af58bad39ea" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.309090 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0696dc2db616f4616e43dd171ba757d11bce09169b35b516964d1af58bad39ea"} err="failed to get container status \"0696dc2db616f4616e43dd171ba757d11bce09169b35b516964d1af58bad39ea\": rpc error: code = NotFound desc = could not find container \"0696dc2db616f4616e43dd171ba757d11bce09169b35b516964d1af58bad39ea\": container with ID starting with 0696dc2db616f4616e43dd171ba757d11bce09169b35b516964d1af58bad39ea not found: ID does not exist" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.309116 4907 scope.go:117] "RemoveContainer" containerID="1e5ba68b21b276693f03089f51ca3955dd88adcaacae52727d03a999b543d411" Mar 13 14:31:13 crc kubenswrapper[4907]: E0313 14:31:13.309374 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1e5ba68b21b276693f03089f51ca3955dd88adcaacae52727d03a999b543d411\": container with ID starting with 1e5ba68b21b276693f03089f51ca3955dd88adcaacae52727d03a999b543d411 not found: ID does not exist" containerID="1e5ba68b21b276693f03089f51ca3955dd88adcaacae52727d03a999b543d411" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.309394 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1e5ba68b21b276693f03089f51ca3955dd88adcaacae52727d03a999b543d411"} err="failed to get container status \"1e5ba68b21b276693f03089f51ca3955dd88adcaacae52727d03a999b543d411\": rpc error: code = NotFound desc = could not find container \"1e5ba68b21b276693f03089f51ca3955dd88adcaacae52727d03a999b543d411\": container with ID starting with 1e5ba68b21b276693f03089f51ca3955dd88adcaacae52727d03a999b543d411 not found: ID does not exist" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.309406 4907 scope.go:117] "RemoveContainer" containerID="8765bf4cde625366ddd57fa87118a09b18b3fbbf1d2ded31c12344e9c08dadb2" Mar 13 14:31:13 crc kubenswrapper[4907]: E0313 14:31:13.309615 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8765bf4cde625366ddd57fa87118a09b18b3fbbf1d2ded31c12344e9c08dadb2\": container with ID starting with 8765bf4cde625366ddd57fa87118a09b18b3fbbf1d2ded31c12344e9c08dadb2 not found: ID does not exist" containerID="8765bf4cde625366ddd57fa87118a09b18b3fbbf1d2ded31c12344e9c08dadb2" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.309655 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8765bf4cde625366ddd57fa87118a09b18b3fbbf1d2ded31c12344e9c08dadb2"} err="failed to get container status \"8765bf4cde625366ddd57fa87118a09b18b3fbbf1d2ded31c12344e9c08dadb2\": rpc error: code = NotFound desc = could not find container \"8765bf4cde625366ddd57fa87118a09b18b3fbbf1d2ded31c12344e9c08dadb2\": container with ID starting with 8765bf4cde625366ddd57fa87118a09b18b3fbbf1d2ded31c12344e9c08dadb2 not found: ID does not exist" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.309668 4907 scope.go:117] "RemoveContainer" containerID="6f5cf167a5953dd1bdb0f436b238215d64975f16e188a1dacd546c8d2adfba14" Mar 13 14:31:13 crc kubenswrapper[4907]: E0313 14:31:13.309817 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6f5cf167a5953dd1bdb0f436b238215d64975f16e188a1dacd546c8d2adfba14\": container with ID starting with 6f5cf167a5953dd1bdb0f436b238215d64975f16e188a1dacd546c8d2adfba14 not found: ID does not exist" containerID="6f5cf167a5953dd1bdb0f436b238215d64975f16e188a1dacd546c8d2adfba14" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.309835 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6f5cf167a5953dd1bdb0f436b238215d64975f16e188a1dacd546c8d2adfba14"} err="failed to get container status \"6f5cf167a5953dd1bdb0f436b238215d64975f16e188a1dacd546c8d2adfba14\": rpc error: code = NotFound desc = could not find container \"6f5cf167a5953dd1bdb0f436b238215d64975f16e188a1dacd546c8d2adfba14\": container with ID starting with 6f5cf167a5953dd1bdb0f436b238215d64975f16e188a1dacd546c8d2adfba14 not found: ID does not exist" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.309847 4907 scope.go:117] "RemoveContainer" containerID="c1b4fe0c395990393cfbace1600ea9872fde43535722ed738332ba0b26f178a4" Mar 13 14:31:13 crc kubenswrapper[4907]: E0313 14:31:13.310048 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c1b4fe0c395990393cfbace1600ea9872fde43535722ed738332ba0b26f178a4\": container with ID starting with c1b4fe0c395990393cfbace1600ea9872fde43535722ed738332ba0b26f178a4 not found: ID does not exist" containerID="c1b4fe0c395990393cfbace1600ea9872fde43535722ed738332ba0b26f178a4" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.310068 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c1b4fe0c395990393cfbace1600ea9872fde43535722ed738332ba0b26f178a4"} err="failed to get container status \"c1b4fe0c395990393cfbace1600ea9872fde43535722ed738332ba0b26f178a4\": rpc error: code = NotFound desc = could not find container \"c1b4fe0c395990393cfbace1600ea9872fde43535722ed738332ba0b26f178a4\": container with ID starting with c1b4fe0c395990393cfbace1600ea9872fde43535722ed738332ba0b26f178a4 not found: ID does not exist" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.310078 4907 scope.go:117] "RemoveContainer" containerID="5a6b392bf5f4cbd99ba2fc751327322f8960f4c73aba15cb9a501d4c9c3edd94" Mar 13 14:31:13 crc kubenswrapper[4907]: E0313 14:31:13.310210 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5a6b392bf5f4cbd99ba2fc751327322f8960f4c73aba15cb9a501d4c9c3edd94\": container with ID starting with 5a6b392bf5f4cbd99ba2fc751327322f8960f4c73aba15cb9a501d4c9c3edd94 not found: ID does not exist" containerID="5a6b392bf5f4cbd99ba2fc751327322f8960f4c73aba15cb9a501d4c9c3edd94" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.310228 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5a6b392bf5f4cbd99ba2fc751327322f8960f4c73aba15cb9a501d4c9c3edd94"} err="failed to get container status \"5a6b392bf5f4cbd99ba2fc751327322f8960f4c73aba15cb9a501d4c9c3edd94\": rpc error: code = NotFound desc = could not find container \"5a6b392bf5f4cbd99ba2fc751327322f8960f4c73aba15cb9a501d4c9c3edd94\": container with ID starting with 5a6b392bf5f4cbd99ba2fc751327322f8960f4c73aba15cb9a501d4c9c3edd94 not found: ID does not exist" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.310238 4907 scope.go:117] "RemoveContainer" containerID="c81c73db839bc006e1b7fdce225dd05a9a50870ea60ca389ea146e1e69d195b7" Mar 13 14:31:13 crc kubenswrapper[4907]: E0313 14:31:13.310402 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c81c73db839bc006e1b7fdce225dd05a9a50870ea60ca389ea146e1e69d195b7\": container with ID starting with c81c73db839bc006e1b7fdce225dd05a9a50870ea60ca389ea146e1e69d195b7 not found: ID does not exist" containerID="c81c73db839bc006e1b7fdce225dd05a9a50870ea60ca389ea146e1e69d195b7" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.310421 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c81c73db839bc006e1b7fdce225dd05a9a50870ea60ca389ea146e1e69d195b7"} err="failed to get container status \"c81c73db839bc006e1b7fdce225dd05a9a50870ea60ca389ea146e1e69d195b7\": rpc error: code = NotFound desc = could not find container \"c81c73db839bc006e1b7fdce225dd05a9a50870ea60ca389ea146e1e69d195b7\": container with ID starting with c81c73db839bc006e1b7fdce225dd05a9a50870ea60ca389ea146e1e69d195b7 not found: ID does not exist" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.310432 4907 scope.go:117] "RemoveContainer" containerID="dbcb2b472a15fbcc053532418c2f7a654827914050e539781030bc0749d6218f" Mar 13 14:31:13 crc kubenswrapper[4907]: E0313 14:31:13.310585 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dbcb2b472a15fbcc053532418c2f7a654827914050e539781030bc0749d6218f\": container with ID starting with dbcb2b472a15fbcc053532418c2f7a654827914050e539781030bc0749d6218f not found: ID does not exist" containerID="dbcb2b472a15fbcc053532418c2f7a654827914050e539781030bc0749d6218f" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.310602 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dbcb2b472a15fbcc053532418c2f7a654827914050e539781030bc0749d6218f"} err="failed to get container status \"dbcb2b472a15fbcc053532418c2f7a654827914050e539781030bc0749d6218f\": rpc error: code = NotFound desc = could not find container \"dbcb2b472a15fbcc053532418c2f7a654827914050e539781030bc0749d6218f\": container with ID starting with dbcb2b472a15fbcc053532418c2f7a654827914050e539781030bc0749d6218f not found: ID does not exist" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.310614 4907 scope.go:117] "RemoveContainer" containerID="7a94b1f20e2cb964aaf2fe5c9f23f0cd193291a14edd5411717f7cad81ed9808" Mar 13 14:31:13 crc kubenswrapper[4907]: E0313 14:31:13.310754 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7a94b1f20e2cb964aaf2fe5c9f23f0cd193291a14edd5411717f7cad81ed9808\": container with ID starting with 7a94b1f20e2cb964aaf2fe5c9f23f0cd193291a14edd5411717f7cad81ed9808 not found: ID does not exist" containerID="7a94b1f20e2cb964aaf2fe5c9f23f0cd193291a14edd5411717f7cad81ed9808" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.310774 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7a94b1f20e2cb964aaf2fe5c9f23f0cd193291a14edd5411717f7cad81ed9808"} err="failed to get container status \"7a94b1f20e2cb964aaf2fe5c9f23f0cd193291a14edd5411717f7cad81ed9808\": rpc error: code = NotFound desc = could not find container \"7a94b1f20e2cb964aaf2fe5c9f23f0cd193291a14edd5411717f7cad81ed9808\": container with ID starting with 7a94b1f20e2cb964aaf2fe5c9f23f0cd193291a14edd5411717f7cad81ed9808 not found: ID does not exist" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.310787 4907 scope.go:117] "RemoveContainer" containerID="2bc2275a73f47c1a27db775d1cc1c3ae9c20c01f9ddad7a35bf652467b66ea80" Mar 13 14:31:13 crc kubenswrapper[4907]: E0313 14:31:13.310942 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2bc2275a73f47c1a27db775d1cc1c3ae9c20c01f9ddad7a35bf652467b66ea80\": container with ID starting with 2bc2275a73f47c1a27db775d1cc1c3ae9c20c01f9ddad7a35bf652467b66ea80 not found: ID does not exist" containerID="2bc2275a73f47c1a27db775d1cc1c3ae9c20c01f9ddad7a35bf652467b66ea80" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.310966 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2bc2275a73f47c1a27db775d1cc1c3ae9c20c01f9ddad7a35bf652467b66ea80"} err="failed to get container status \"2bc2275a73f47c1a27db775d1cc1c3ae9c20c01f9ddad7a35bf652467b66ea80\": rpc error: code = NotFound desc = could not find container \"2bc2275a73f47c1a27db775d1cc1c3ae9c20c01f9ddad7a35bf652467b66ea80\": container with ID starting with 2bc2275a73f47c1a27db775d1cc1c3ae9c20c01f9ddad7a35bf652467b66ea80 not found: ID does not exist" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.310978 4907 scope.go:117] "RemoveContainer" containerID="b7f60cadabc1fd2d193073423cce7296e305ff6325001c7ec9823dd267a1787a" Mar 13 14:31:13 crc kubenswrapper[4907]: E0313 14:31:13.311155 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b7f60cadabc1fd2d193073423cce7296e305ff6325001c7ec9823dd267a1787a\": container with ID starting with b7f60cadabc1fd2d193073423cce7296e305ff6325001c7ec9823dd267a1787a not found: ID does not exist" containerID="b7f60cadabc1fd2d193073423cce7296e305ff6325001c7ec9823dd267a1787a" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.311172 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b7f60cadabc1fd2d193073423cce7296e305ff6325001c7ec9823dd267a1787a"} err="failed to get container status \"b7f60cadabc1fd2d193073423cce7296e305ff6325001c7ec9823dd267a1787a\": rpc error: code = NotFound desc = could not find container \"b7f60cadabc1fd2d193073423cce7296e305ff6325001c7ec9823dd267a1787a\": container with ID starting with b7f60cadabc1fd2d193073423cce7296e305ff6325001c7ec9823dd267a1787a not found: ID does not exist" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.311184 4907 scope.go:117] "RemoveContainer" containerID="c7370f4c24ed7b82c59bc2c1e3ae152c65420b5dd6227a66eb64b5b83832e410" Mar 13 14:31:13 crc kubenswrapper[4907]: E0313 14:31:13.311321 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c7370f4c24ed7b82c59bc2c1e3ae152c65420b5dd6227a66eb64b5b83832e410\": container with ID starting with c7370f4c24ed7b82c59bc2c1e3ae152c65420b5dd6227a66eb64b5b83832e410 not found: ID does not exist" containerID="c7370f4c24ed7b82c59bc2c1e3ae152c65420b5dd6227a66eb64b5b83832e410" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.311339 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c7370f4c24ed7b82c59bc2c1e3ae152c65420b5dd6227a66eb64b5b83832e410"} err="failed to get container status \"c7370f4c24ed7b82c59bc2c1e3ae152c65420b5dd6227a66eb64b5b83832e410\": rpc error: code = NotFound desc = could not find container \"c7370f4c24ed7b82c59bc2c1e3ae152c65420b5dd6227a66eb64b5b83832e410\": container with ID starting with c7370f4c24ed7b82c59bc2c1e3ae152c65420b5dd6227a66eb64b5b83832e410 not found: ID does not exist" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.311350 4907 scope.go:117] "RemoveContainer" containerID="f8bc1f3dd2242db8254b664f00b3789940491fa59b38a7fc033c225b233b776e" Mar 13 14:31:13 crc kubenswrapper[4907]: E0313 14:31:13.311519 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f8bc1f3dd2242db8254b664f00b3789940491fa59b38a7fc033c225b233b776e\": container with ID starting with f8bc1f3dd2242db8254b664f00b3789940491fa59b38a7fc033c225b233b776e not found: ID does not exist" containerID="f8bc1f3dd2242db8254b664f00b3789940491fa59b38a7fc033c225b233b776e" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.311537 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f8bc1f3dd2242db8254b664f00b3789940491fa59b38a7fc033c225b233b776e"} err="failed to get container status \"f8bc1f3dd2242db8254b664f00b3789940491fa59b38a7fc033c225b233b776e\": rpc error: code = NotFound desc = could not find container \"f8bc1f3dd2242db8254b664f00b3789940491fa59b38a7fc033c225b233b776e\": container with ID starting with f8bc1f3dd2242db8254b664f00b3789940491fa59b38a7fc033c225b233b776e not found: ID does not exist" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.311551 4907 scope.go:117] "RemoveContainer" containerID="9b6f16a5610e8f2c08a40a29bfba0c00e1ff883cb29b5290604713aad1dab566" Mar 13 14:31:13 crc kubenswrapper[4907]: E0313 14:31:13.311754 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9b6f16a5610e8f2c08a40a29bfba0c00e1ff883cb29b5290604713aad1dab566\": container with ID starting with 9b6f16a5610e8f2c08a40a29bfba0c00e1ff883cb29b5290604713aad1dab566 not found: ID does not exist" containerID="9b6f16a5610e8f2c08a40a29bfba0c00e1ff883cb29b5290604713aad1dab566" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.311863 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9b6f16a5610e8f2c08a40a29bfba0c00e1ff883cb29b5290604713aad1dab566"} err="failed to get container status \"9b6f16a5610e8f2c08a40a29bfba0c00e1ff883cb29b5290604713aad1dab566\": rpc error: code = NotFound desc = could not find container \"9b6f16a5610e8f2c08a40a29bfba0c00e1ff883cb29b5290604713aad1dab566\": container with ID starting with 9b6f16a5610e8f2c08a40a29bfba0c00e1ff883cb29b5290604713aad1dab566 not found: ID does not exist" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.311889 4907 scope.go:117] "RemoveContainer" containerID="8168297a51ad087d7375035164e54e4a0d24e6fd56cf85e922e11d80f7f3d091" Mar 13 14:31:13 crc kubenswrapper[4907]: E0313 14:31:13.312219 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8168297a51ad087d7375035164e54e4a0d24e6fd56cf85e922e11d80f7f3d091\": container with ID starting with 8168297a51ad087d7375035164e54e4a0d24e6fd56cf85e922e11d80f7f3d091 not found: ID does not exist" containerID="8168297a51ad087d7375035164e54e4a0d24e6fd56cf85e922e11d80f7f3d091" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.312269 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8168297a51ad087d7375035164e54e4a0d24e6fd56cf85e922e11d80f7f3d091"} err="failed to get container status \"8168297a51ad087d7375035164e54e4a0d24e6fd56cf85e922e11d80f7f3d091\": rpc error: code = NotFound desc = could not find container \"8168297a51ad087d7375035164e54e4a0d24e6fd56cf85e922e11d80f7f3d091\": container with ID starting with 8168297a51ad087d7375035164e54e4a0d24e6fd56cf85e922e11d80f7f3d091 not found: ID does not exist" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.365599 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/c1b70392-1240-40d9-8128-e7abe29c8398-var-log\") pod \"c1b70392-1240-40d9-8128-e7abe29c8398\" (UID: \"c1b70392-1240-40d9-8128-e7abe29c8398\") " Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.365737 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/c1b70392-1240-40d9-8128-e7abe29c8398-var-lib\") pod \"c1b70392-1240-40d9-8128-e7abe29c8398\" (UID: \"c1b70392-1240-40d9-8128-e7abe29c8398\") " Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.365734 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c1b70392-1240-40d9-8128-e7abe29c8398-var-log" (OuterVolumeSpecName: "var-log") pod "c1b70392-1240-40d9-8128-e7abe29c8398" (UID: "c1b70392-1240-40d9-8128-e7abe29c8398"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.365775 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4qvcj\" (UniqueName: \"kubernetes.io/projected/c1b70392-1240-40d9-8128-e7abe29c8398-kube-api-access-4qvcj\") pod \"c1b70392-1240-40d9-8128-e7abe29c8398\" (UID: \"c1b70392-1240-40d9-8128-e7abe29c8398\") " Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.365792 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c1b70392-1240-40d9-8128-e7abe29c8398-var-lib" (OuterVolumeSpecName: "var-lib") pod "c1b70392-1240-40d9-8128-e7abe29c8398" (UID: "c1b70392-1240-40d9-8128-e7abe29c8398"). InnerVolumeSpecName "var-lib". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.365824 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/c1b70392-1240-40d9-8128-e7abe29c8398-var-run\") pod \"c1b70392-1240-40d9-8128-e7abe29c8398\" (UID: \"c1b70392-1240-40d9-8128-e7abe29c8398\") " Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.365866 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c1b70392-1240-40d9-8128-e7abe29c8398-scripts\") pod \"c1b70392-1240-40d9-8128-e7abe29c8398\" (UID: \"c1b70392-1240-40d9-8128-e7abe29c8398\") " Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.365900 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/c1b70392-1240-40d9-8128-e7abe29c8398-etc-ovs\") pod \"c1b70392-1240-40d9-8128-e7abe29c8398\" (UID: \"c1b70392-1240-40d9-8128-e7abe29c8398\") " Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.366054 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c1b70392-1240-40d9-8128-e7abe29c8398-etc-ovs" (OuterVolumeSpecName: "etc-ovs") pod "c1b70392-1240-40d9-8128-e7abe29c8398" (UID: "c1b70392-1240-40d9-8128-e7abe29c8398"). InnerVolumeSpecName "etc-ovs". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.366017 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c1b70392-1240-40d9-8128-e7abe29c8398-var-run" (OuterVolumeSpecName: "var-run") pod "c1b70392-1240-40d9-8128-e7abe29c8398" (UID: "c1b70392-1240-40d9-8128-e7abe29c8398"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.366269 4907 reconciler_common.go:293] "Volume detached for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/c1b70392-1240-40d9-8128-e7abe29c8398-var-lib\") on node \"crc\" DevicePath \"\"" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.366285 4907 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/c1b70392-1240-40d9-8128-e7abe29c8398-var-run\") on node \"crc\" DevicePath \"\"" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.366297 4907 reconciler_common.go:293] "Volume detached for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/c1b70392-1240-40d9-8128-e7abe29c8398-etc-ovs\") on node \"crc\" DevicePath \"\"" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.366309 4907 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/c1b70392-1240-40d9-8128-e7abe29c8398-var-log\") on node \"crc\" DevicePath \"\"" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.367198 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c1b70392-1240-40d9-8128-e7abe29c8398-scripts" (OuterVolumeSpecName: "scripts") pod "c1b70392-1240-40d9-8128-e7abe29c8398" (UID: "c1b70392-1240-40d9-8128-e7abe29c8398"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.373463 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c1b70392-1240-40d9-8128-e7abe29c8398-kube-api-access-4qvcj" (OuterVolumeSpecName: "kube-api-access-4qvcj") pod "c1b70392-1240-40d9-8128-e7abe29c8398" (UID: "c1b70392-1240-40d9-8128-e7abe29c8398"). InnerVolumeSpecName "kube-api-access-4qvcj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.467575 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4qvcj\" (UniqueName: \"kubernetes.io/projected/c1b70392-1240-40d9-8128-e7abe29c8398-kube-api-access-4qvcj\") on node \"crc\" DevicePath \"\"" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.467615 4907 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c1b70392-1240-40d9-8128-e7abe29c8398-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.784971 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-vp6f5_c1b70392-1240-40d9-8128-e7abe29c8398/ovs-vswitchd/0.log" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.786120 4907 generic.go:334] "Generic (PLEG): container finished" podID="c1b70392-1240-40d9-8128-e7abe29c8398" containerID="265b6b6fe5f95195cdc86c65456998ad327fc7ff3a6641300cd6ac00798eff4f" exitCode=137 Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.786228 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-vp6f5" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.789979 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="14f05f31-8185-43b0-be69-bcf8d5388ea5" path="/var/lib/kubelet/pods/14f05f31-8185-43b0-be69-bcf8d5388ea5/volumes" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.792520 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-vp6f5" event={"ID":"c1b70392-1240-40d9-8128-e7abe29c8398","Type":"ContainerDied","Data":"265b6b6fe5f95195cdc86c65456998ad327fc7ff3a6641300cd6ac00798eff4f"} Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.792581 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-vp6f5" event={"ID":"c1b70392-1240-40d9-8128-e7abe29c8398","Type":"ContainerDied","Data":"99abe9d38052154162343b52b5e60bd6560b49e824682c932c7a2a64adbb5d61"} Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.792607 4907 scope.go:117] "RemoveContainer" containerID="265b6b6fe5f95195cdc86c65456998ad327fc7ff3a6641300cd6ac00798eff4f" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.810845 4907 scope.go:117] "RemoveContainer" containerID="6b0b25079cc099dabe922ac89fa41ae76d2f01797aac9f273b042e242459c574" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.833058 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-ovs-vp6f5"] Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.841067 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-ovs-vp6f5"] Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.843091 4907 scope.go:117] "RemoveContainer" containerID="4cf539e8b8cde03127504ad0dcb18c074f6f0c3dab53805c59e991c594c32a31" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.873460 4907 scope.go:117] "RemoveContainer" containerID="265b6b6fe5f95195cdc86c65456998ad327fc7ff3a6641300cd6ac00798eff4f" Mar 13 14:31:13 crc kubenswrapper[4907]: E0313 14:31:13.874307 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"265b6b6fe5f95195cdc86c65456998ad327fc7ff3a6641300cd6ac00798eff4f\": container with ID starting with 265b6b6fe5f95195cdc86c65456998ad327fc7ff3a6641300cd6ac00798eff4f not found: ID does not exist" containerID="265b6b6fe5f95195cdc86c65456998ad327fc7ff3a6641300cd6ac00798eff4f" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.874350 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"265b6b6fe5f95195cdc86c65456998ad327fc7ff3a6641300cd6ac00798eff4f"} err="failed to get container status \"265b6b6fe5f95195cdc86c65456998ad327fc7ff3a6641300cd6ac00798eff4f\": rpc error: code = NotFound desc = could not find container \"265b6b6fe5f95195cdc86c65456998ad327fc7ff3a6641300cd6ac00798eff4f\": container with ID starting with 265b6b6fe5f95195cdc86c65456998ad327fc7ff3a6641300cd6ac00798eff4f not found: ID does not exist" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.874379 4907 scope.go:117] "RemoveContainer" containerID="6b0b25079cc099dabe922ac89fa41ae76d2f01797aac9f273b042e242459c574" Mar 13 14:31:13 crc kubenswrapper[4907]: E0313 14:31:13.874706 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6b0b25079cc099dabe922ac89fa41ae76d2f01797aac9f273b042e242459c574\": container with ID starting with 6b0b25079cc099dabe922ac89fa41ae76d2f01797aac9f273b042e242459c574 not found: ID does not exist" containerID="6b0b25079cc099dabe922ac89fa41ae76d2f01797aac9f273b042e242459c574" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.874729 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b0b25079cc099dabe922ac89fa41ae76d2f01797aac9f273b042e242459c574"} err="failed to get container status \"6b0b25079cc099dabe922ac89fa41ae76d2f01797aac9f273b042e242459c574\": rpc error: code = NotFound desc = could not find container \"6b0b25079cc099dabe922ac89fa41ae76d2f01797aac9f273b042e242459c574\": container with ID starting with 6b0b25079cc099dabe922ac89fa41ae76d2f01797aac9f273b042e242459c574 not found: ID does not exist" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.874747 4907 scope.go:117] "RemoveContainer" containerID="4cf539e8b8cde03127504ad0dcb18c074f6f0c3dab53805c59e991c594c32a31" Mar 13 14:31:13 crc kubenswrapper[4907]: E0313 14:31:13.875229 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4cf539e8b8cde03127504ad0dcb18c074f6f0c3dab53805c59e991c594c32a31\": container with ID starting with 4cf539e8b8cde03127504ad0dcb18c074f6f0c3dab53805c59e991c594c32a31 not found: ID does not exist" containerID="4cf539e8b8cde03127504ad0dcb18c074f6f0c3dab53805c59e991c594c32a31" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.875263 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4cf539e8b8cde03127504ad0dcb18c074f6f0c3dab53805c59e991c594c32a31"} err="failed to get container status \"4cf539e8b8cde03127504ad0dcb18c074f6f0c3dab53805c59e991c594c32a31\": rpc error: code = NotFound desc = could not find container \"4cf539e8b8cde03127504ad0dcb18c074f6f0c3dab53805c59e991c594c32a31\": container with ID starting with 4cf539e8b8cde03127504ad0dcb18c074f6f0c3dab53805c59e991c594c32a31 not found: ID does not exist" Mar 13 14:31:15 crc kubenswrapper[4907]: I0313 14:31:15.791758 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c1b70392-1240-40d9-8128-e7abe29c8398" path="/var/lib/kubelet/pods/c1b70392-1240-40d9-8128-e7abe29c8398/volumes" Mar 13 14:31:16 crc kubenswrapper[4907]: I0313 14:31:16.616847 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-5przm" Mar 13 14:31:16 crc kubenswrapper[4907]: I0313 14:31:16.660987 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-5przm" Mar 13 14:31:16 crc kubenswrapper[4907]: I0313 14:31:16.845009 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5przm"] Mar 13 14:31:17 crc kubenswrapper[4907]: I0313 14:31:17.824407 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-5przm" podUID="1e43e4fe-9084-48bb-a931-33e2fc5ecaaa" containerName="registry-server" containerID="cri-o://b4fc4a9c98855f93f56a77779ad655ec33841f4a3e7acba65768c38b4a3718bf" gracePeriod=2 Mar 13 14:31:18 crc kubenswrapper[4907]: I0313 14:31:18.042311 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 14:31:18 crc kubenswrapper[4907]: I0313 14:31:18.042867 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 14:31:18 crc kubenswrapper[4907]: I0313 14:31:18.321682 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5przm" Mar 13 14:31:18 crc kubenswrapper[4907]: I0313 14:31:18.432956 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g8dvd\" (UniqueName: \"kubernetes.io/projected/1e43e4fe-9084-48bb-a931-33e2fc5ecaaa-kube-api-access-g8dvd\") pod \"1e43e4fe-9084-48bb-a931-33e2fc5ecaaa\" (UID: \"1e43e4fe-9084-48bb-a931-33e2fc5ecaaa\") " Mar 13 14:31:18 crc kubenswrapper[4907]: I0313 14:31:18.433009 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1e43e4fe-9084-48bb-a931-33e2fc5ecaaa-utilities\") pod \"1e43e4fe-9084-48bb-a931-33e2fc5ecaaa\" (UID: \"1e43e4fe-9084-48bb-a931-33e2fc5ecaaa\") " Mar 13 14:31:18 crc kubenswrapper[4907]: I0313 14:31:18.433086 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1e43e4fe-9084-48bb-a931-33e2fc5ecaaa-catalog-content\") pod \"1e43e4fe-9084-48bb-a931-33e2fc5ecaaa\" (UID: \"1e43e4fe-9084-48bb-a931-33e2fc5ecaaa\") " Mar 13 14:31:18 crc kubenswrapper[4907]: I0313 14:31:18.434384 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1e43e4fe-9084-48bb-a931-33e2fc5ecaaa-utilities" (OuterVolumeSpecName: "utilities") pod "1e43e4fe-9084-48bb-a931-33e2fc5ecaaa" (UID: "1e43e4fe-9084-48bb-a931-33e2fc5ecaaa"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:31:18 crc kubenswrapper[4907]: I0313 14:31:18.443375 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1e43e4fe-9084-48bb-a931-33e2fc5ecaaa-kube-api-access-g8dvd" (OuterVolumeSpecName: "kube-api-access-g8dvd") pod "1e43e4fe-9084-48bb-a931-33e2fc5ecaaa" (UID: "1e43e4fe-9084-48bb-a931-33e2fc5ecaaa"). InnerVolumeSpecName "kube-api-access-g8dvd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:31:18 crc kubenswrapper[4907]: I0313 14:31:18.534679 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1e43e4fe-9084-48bb-a931-33e2fc5ecaaa-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 14:31:18 crc kubenswrapper[4907]: I0313 14:31:18.535000 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g8dvd\" (UniqueName: \"kubernetes.io/projected/1e43e4fe-9084-48bb-a931-33e2fc5ecaaa-kube-api-access-g8dvd\") on node \"crc\" DevicePath \"\"" Mar 13 14:31:18 crc kubenswrapper[4907]: I0313 14:31:18.583033 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1e43e4fe-9084-48bb-a931-33e2fc5ecaaa-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1e43e4fe-9084-48bb-a931-33e2fc5ecaaa" (UID: "1e43e4fe-9084-48bb-a931-33e2fc5ecaaa"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:31:18 crc kubenswrapper[4907]: I0313 14:31:18.636515 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1e43e4fe-9084-48bb-a931-33e2fc5ecaaa-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 14:31:18 crc kubenswrapper[4907]: I0313 14:31:18.840640 4907 generic.go:334] "Generic (PLEG): container finished" podID="1e43e4fe-9084-48bb-a931-33e2fc5ecaaa" containerID="b4fc4a9c98855f93f56a77779ad655ec33841f4a3e7acba65768c38b4a3718bf" exitCode=0 Mar 13 14:31:18 crc kubenswrapper[4907]: I0313 14:31:18.840701 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5przm" event={"ID":"1e43e4fe-9084-48bb-a931-33e2fc5ecaaa","Type":"ContainerDied","Data":"b4fc4a9c98855f93f56a77779ad655ec33841f4a3e7acba65768c38b4a3718bf"} Mar 13 14:31:18 crc kubenswrapper[4907]: I0313 14:31:18.840743 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5przm" event={"ID":"1e43e4fe-9084-48bb-a931-33e2fc5ecaaa","Type":"ContainerDied","Data":"420aae8884cc744f6583795347d67462ea3274678bbdefc76fe718c1b6545b81"} Mar 13 14:31:18 crc kubenswrapper[4907]: I0313 14:31:18.840742 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5przm" Mar 13 14:31:18 crc kubenswrapper[4907]: I0313 14:31:18.840771 4907 scope.go:117] "RemoveContainer" containerID="b4fc4a9c98855f93f56a77779ad655ec33841f4a3e7acba65768c38b4a3718bf" Mar 13 14:31:18 crc kubenswrapper[4907]: I0313 14:31:18.867334 4907 scope.go:117] "RemoveContainer" containerID="057b7d13b67e10cac7b4063457b964d60358da6d919d6ef053cd4aaf19089d04" Mar 13 14:31:18 crc kubenswrapper[4907]: I0313 14:31:18.891946 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5przm"] Mar 13 14:31:18 crc kubenswrapper[4907]: I0313 14:31:18.900009 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-5przm"] Mar 13 14:31:18 crc kubenswrapper[4907]: I0313 14:31:18.919539 4907 scope.go:117] "RemoveContainer" containerID="4b6734fa58a17ab6193b32c4603463a610dfd0ffb321199337a7a16643dced00" Mar 13 14:31:18 crc kubenswrapper[4907]: I0313 14:31:18.945823 4907 scope.go:117] "RemoveContainer" containerID="b4fc4a9c98855f93f56a77779ad655ec33841f4a3e7acba65768c38b4a3718bf" Mar 13 14:31:18 crc kubenswrapper[4907]: E0313 14:31:18.948366 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b4fc4a9c98855f93f56a77779ad655ec33841f4a3e7acba65768c38b4a3718bf\": container with ID starting with b4fc4a9c98855f93f56a77779ad655ec33841f4a3e7acba65768c38b4a3718bf not found: ID does not exist" containerID="b4fc4a9c98855f93f56a77779ad655ec33841f4a3e7acba65768c38b4a3718bf" Mar 13 14:31:18 crc kubenswrapper[4907]: I0313 14:31:18.948460 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b4fc4a9c98855f93f56a77779ad655ec33841f4a3e7acba65768c38b4a3718bf"} err="failed to get container status \"b4fc4a9c98855f93f56a77779ad655ec33841f4a3e7acba65768c38b4a3718bf\": rpc error: code = NotFound desc = could not find container \"b4fc4a9c98855f93f56a77779ad655ec33841f4a3e7acba65768c38b4a3718bf\": container with ID starting with b4fc4a9c98855f93f56a77779ad655ec33841f4a3e7acba65768c38b4a3718bf not found: ID does not exist" Mar 13 14:31:18 crc kubenswrapper[4907]: I0313 14:31:18.948562 4907 scope.go:117] "RemoveContainer" containerID="057b7d13b67e10cac7b4063457b964d60358da6d919d6ef053cd4aaf19089d04" Mar 13 14:31:18 crc kubenswrapper[4907]: E0313 14:31:18.951127 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"057b7d13b67e10cac7b4063457b964d60358da6d919d6ef053cd4aaf19089d04\": container with ID starting with 057b7d13b67e10cac7b4063457b964d60358da6d919d6ef053cd4aaf19089d04 not found: ID does not exist" containerID="057b7d13b67e10cac7b4063457b964d60358da6d919d6ef053cd4aaf19089d04" Mar 13 14:31:18 crc kubenswrapper[4907]: I0313 14:31:18.951250 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"057b7d13b67e10cac7b4063457b964d60358da6d919d6ef053cd4aaf19089d04"} err="failed to get container status \"057b7d13b67e10cac7b4063457b964d60358da6d919d6ef053cd4aaf19089d04\": rpc error: code = NotFound desc = could not find container \"057b7d13b67e10cac7b4063457b964d60358da6d919d6ef053cd4aaf19089d04\": container with ID starting with 057b7d13b67e10cac7b4063457b964d60358da6d919d6ef053cd4aaf19089d04 not found: ID does not exist" Mar 13 14:31:18 crc kubenswrapper[4907]: I0313 14:31:18.951295 4907 scope.go:117] "RemoveContainer" containerID="4b6734fa58a17ab6193b32c4603463a610dfd0ffb321199337a7a16643dced00" Mar 13 14:31:18 crc kubenswrapper[4907]: E0313 14:31:18.957300 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4b6734fa58a17ab6193b32c4603463a610dfd0ffb321199337a7a16643dced00\": container with ID starting with 4b6734fa58a17ab6193b32c4603463a610dfd0ffb321199337a7a16643dced00 not found: ID does not exist" containerID="4b6734fa58a17ab6193b32c4603463a610dfd0ffb321199337a7a16643dced00" Mar 13 14:31:18 crc kubenswrapper[4907]: I0313 14:31:18.957359 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4b6734fa58a17ab6193b32c4603463a610dfd0ffb321199337a7a16643dced00"} err="failed to get container status \"4b6734fa58a17ab6193b32c4603463a610dfd0ffb321199337a7a16643dced00\": rpc error: code = NotFound desc = could not find container \"4b6734fa58a17ab6193b32c4603463a610dfd0ffb321199337a7a16643dced00\": container with ID starting with 4b6734fa58a17ab6193b32c4603463a610dfd0ffb321199337a7a16643dced00 not found: ID does not exist" Mar 13 14:31:19 crc kubenswrapper[4907]: I0313 14:31:19.793769 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1e43e4fe-9084-48bb-a931-33e2fc5ecaaa" path="/var/lib/kubelet/pods/1e43e4fe-9084-48bb-a931-33e2fc5ecaaa/volumes" Mar 13 14:31:30 crc kubenswrapper[4907]: I0313 14:31:30.528160 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-2dpx8"] Mar 13 14:31:30 crc kubenswrapper[4907]: E0313 14:31:30.529468 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerName="object-expirer" Mar 13 14:31:30 crc kubenswrapper[4907]: I0313 14:31:30.529487 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerName="object-expirer" Mar 13 14:31:30 crc kubenswrapper[4907]: E0313 14:31:30.529668 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerName="account-reaper" Mar 13 14:31:30 crc kubenswrapper[4907]: I0313 14:31:30.529682 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerName="account-reaper" Mar 13 14:31:30 crc kubenswrapper[4907]: E0313 14:31:30.529690 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e43e4fe-9084-48bb-a931-33e2fc5ecaaa" containerName="extract-content" Mar 13 14:31:30 crc kubenswrapper[4907]: I0313 14:31:30.529699 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e43e4fe-9084-48bb-a931-33e2fc5ecaaa" containerName="extract-content" Mar 13 14:31:30 crc kubenswrapper[4907]: E0313 14:31:30.529717 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerName="object-auditor" Mar 13 14:31:30 crc kubenswrapper[4907]: I0313 14:31:30.529725 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerName="object-auditor" Mar 13 14:31:30 crc kubenswrapper[4907]: E0313 14:31:30.529742 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1b70392-1240-40d9-8128-e7abe29c8398" containerName="ovsdb-server" Mar 13 14:31:30 crc kubenswrapper[4907]: I0313 14:31:30.529749 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1b70392-1240-40d9-8128-e7abe29c8398" containerName="ovsdb-server" Mar 13 14:31:30 crc kubenswrapper[4907]: E0313 14:31:30.529759 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerName="account-auditor" Mar 13 14:31:30 crc kubenswrapper[4907]: I0313 14:31:30.529767 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerName="account-auditor" Mar 13 14:31:30 crc kubenswrapper[4907]: E0313 14:31:30.529781 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerName="container-replicator" Mar 13 14:31:30 crc kubenswrapper[4907]: I0313 14:31:30.529789 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerName="container-replicator" Mar 13 14:31:30 crc kubenswrapper[4907]: E0313 14:31:30.529798 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerName="account-server" Mar 13 14:31:30 crc kubenswrapper[4907]: I0313 14:31:30.529806 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerName="account-server" Mar 13 14:31:30 crc kubenswrapper[4907]: E0313 14:31:30.529820 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerName="container-auditor" Mar 13 14:31:30 crc kubenswrapper[4907]: I0313 14:31:30.529830 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerName="container-auditor" Mar 13 14:31:30 crc kubenswrapper[4907]: E0313 14:31:30.529841 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerName="container-updater" Mar 13 14:31:30 crc kubenswrapper[4907]: I0313 14:31:30.529848 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerName="container-updater" Mar 13 14:31:30 crc kubenswrapper[4907]: E0313 14:31:30.529858 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerName="object-server" Mar 13 14:31:30 crc kubenswrapper[4907]: I0313 14:31:30.529866 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerName="object-server" Mar 13 14:31:30 crc kubenswrapper[4907]: E0313 14:31:30.529893 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerName="rsync" Mar 13 14:31:30 crc kubenswrapper[4907]: I0313 14:31:30.529901 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerName="rsync" Mar 13 14:31:30 crc kubenswrapper[4907]: E0313 14:31:30.529912 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerName="container-server" Mar 13 14:31:30 crc kubenswrapper[4907]: I0313 14:31:30.529920 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerName="container-server" Mar 13 14:31:30 crc kubenswrapper[4907]: E0313 14:31:30.529933 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerName="swift-recon-cron" Mar 13 14:31:30 crc kubenswrapper[4907]: I0313 14:31:30.529941 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerName="swift-recon-cron" Mar 13 14:31:30 crc kubenswrapper[4907]: E0313 14:31:30.529955 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1b70392-1240-40d9-8128-e7abe29c8398" containerName="ovsdb-server-init" Mar 13 14:31:30 crc kubenswrapper[4907]: I0313 14:31:30.529962 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1b70392-1240-40d9-8128-e7abe29c8398" containerName="ovsdb-server-init" Mar 13 14:31:30 crc kubenswrapper[4907]: E0313 14:31:30.529973 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1b70392-1240-40d9-8128-e7abe29c8398" containerName="ovs-vswitchd" Mar 13 14:31:30 crc kubenswrapper[4907]: I0313 14:31:30.529980 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1b70392-1240-40d9-8128-e7abe29c8398" containerName="ovs-vswitchd" Mar 13 14:31:30 crc kubenswrapper[4907]: E0313 14:31:30.529990 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerName="account-replicator" Mar 13 14:31:30 crc kubenswrapper[4907]: I0313 14:31:30.529998 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerName="account-replicator" Mar 13 14:31:30 crc kubenswrapper[4907]: E0313 14:31:30.530006 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerName="object-replicator" Mar 13 14:31:30 crc kubenswrapper[4907]: I0313 14:31:30.530013 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerName="object-replicator" Mar 13 14:31:30 crc kubenswrapper[4907]: E0313 14:31:30.530024 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerName="object-updater" Mar 13 14:31:30 crc kubenswrapper[4907]: I0313 14:31:30.530033 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerName="object-updater" Mar 13 14:31:30 crc kubenswrapper[4907]: E0313 14:31:30.530047 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e43e4fe-9084-48bb-a931-33e2fc5ecaaa" containerName="extract-utilities" Mar 13 14:31:30 crc kubenswrapper[4907]: I0313 14:31:30.530055 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e43e4fe-9084-48bb-a931-33e2fc5ecaaa" containerName="extract-utilities" Mar 13 14:31:30 crc kubenswrapper[4907]: E0313 14:31:30.530064 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e43e4fe-9084-48bb-a931-33e2fc5ecaaa" containerName="registry-server" Mar 13 14:31:30 crc kubenswrapper[4907]: I0313 14:31:30.530070 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e43e4fe-9084-48bb-a931-33e2fc5ecaaa" containerName="registry-server" Mar 13 14:31:30 crc kubenswrapper[4907]: I0313 14:31:30.530248 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerName="account-server" Mar 13 14:31:30 crc kubenswrapper[4907]: I0313 14:31:30.530265 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerName="object-expirer" Mar 13 14:31:30 crc kubenswrapper[4907]: I0313 14:31:30.530277 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerName="object-updater" Mar 13 14:31:30 crc kubenswrapper[4907]: I0313 14:31:30.530291 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerName="swift-recon-cron" Mar 13 14:31:30 crc kubenswrapper[4907]: I0313 14:31:30.530303 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerName="account-auditor" Mar 13 14:31:30 crc kubenswrapper[4907]: I0313 14:31:30.530311 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerName="account-reaper" Mar 13 14:31:30 crc kubenswrapper[4907]: I0313 14:31:30.530320 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerName="container-replicator" Mar 13 14:31:30 crc kubenswrapper[4907]: I0313 14:31:30.530328 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerName="container-updater" Mar 13 14:31:30 crc kubenswrapper[4907]: I0313 14:31:30.530341 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerName="rsync" Mar 13 14:31:30 crc kubenswrapper[4907]: I0313 14:31:30.530350 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="c1b70392-1240-40d9-8128-e7abe29c8398" containerName="ovsdb-server" Mar 13 14:31:30 crc kubenswrapper[4907]: I0313 14:31:30.530361 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerName="object-auditor" Mar 13 14:31:30 crc kubenswrapper[4907]: I0313 14:31:30.530371 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerName="object-replicator" Mar 13 14:31:30 crc kubenswrapper[4907]: I0313 14:31:30.530381 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e43e4fe-9084-48bb-a931-33e2fc5ecaaa" containerName="registry-server" Mar 13 14:31:30 crc kubenswrapper[4907]: I0313 14:31:30.530389 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerName="account-replicator" Mar 13 14:31:30 crc kubenswrapper[4907]: I0313 14:31:30.530401 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerName="object-server" Mar 13 14:31:30 crc kubenswrapper[4907]: I0313 14:31:30.530412 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerName="container-server" Mar 13 14:31:30 crc kubenswrapper[4907]: I0313 14:31:30.530424 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="c1b70392-1240-40d9-8128-e7abe29c8398" containerName="ovs-vswitchd" Mar 13 14:31:30 crc kubenswrapper[4907]: I0313 14:31:30.530433 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerName="container-auditor" Mar 13 14:31:30 crc kubenswrapper[4907]: I0313 14:31:30.531597 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2dpx8" Mar 13 14:31:30 crc kubenswrapper[4907]: I0313 14:31:30.545362 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2dpx8"] Mar 13 14:31:30 crc kubenswrapper[4907]: I0313 14:31:30.702171 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d9c047f6-207b-451e-acb0-e909774dc44e-catalog-content\") pod \"certified-operators-2dpx8\" (UID: \"d9c047f6-207b-451e-acb0-e909774dc44e\") " pod="openshift-marketplace/certified-operators-2dpx8" Mar 13 14:31:30 crc kubenswrapper[4907]: I0313 14:31:30.702231 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ttd6l\" (UniqueName: \"kubernetes.io/projected/d9c047f6-207b-451e-acb0-e909774dc44e-kube-api-access-ttd6l\") pod \"certified-operators-2dpx8\" (UID: \"d9c047f6-207b-451e-acb0-e909774dc44e\") " pod="openshift-marketplace/certified-operators-2dpx8" Mar 13 14:31:30 crc kubenswrapper[4907]: I0313 14:31:30.702494 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d9c047f6-207b-451e-acb0-e909774dc44e-utilities\") pod \"certified-operators-2dpx8\" (UID: \"d9c047f6-207b-451e-acb0-e909774dc44e\") " pod="openshift-marketplace/certified-operators-2dpx8" Mar 13 14:31:30 crc kubenswrapper[4907]: I0313 14:31:30.804392 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d9c047f6-207b-451e-acb0-e909774dc44e-utilities\") pod \"certified-operators-2dpx8\" (UID: \"d9c047f6-207b-451e-acb0-e909774dc44e\") " pod="openshift-marketplace/certified-operators-2dpx8" Mar 13 14:31:30 crc kubenswrapper[4907]: I0313 14:31:30.804527 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d9c047f6-207b-451e-acb0-e909774dc44e-catalog-content\") pod \"certified-operators-2dpx8\" (UID: \"d9c047f6-207b-451e-acb0-e909774dc44e\") " pod="openshift-marketplace/certified-operators-2dpx8" Mar 13 14:31:30 crc kubenswrapper[4907]: I0313 14:31:30.804585 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ttd6l\" (UniqueName: \"kubernetes.io/projected/d9c047f6-207b-451e-acb0-e909774dc44e-kube-api-access-ttd6l\") pod \"certified-operators-2dpx8\" (UID: \"d9c047f6-207b-451e-acb0-e909774dc44e\") " pod="openshift-marketplace/certified-operators-2dpx8" Mar 13 14:31:30 crc kubenswrapper[4907]: I0313 14:31:30.804867 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d9c047f6-207b-451e-acb0-e909774dc44e-utilities\") pod \"certified-operators-2dpx8\" (UID: \"d9c047f6-207b-451e-acb0-e909774dc44e\") " pod="openshift-marketplace/certified-operators-2dpx8" Mar 13 14:31:30 crc kubenswrapper[4907]: I0313 14:31:30.805289 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d9c047f6-207b-451e-acb0-e909774dc44e-catalog-content\") pod \"certified-operators-2dpx8\" (UID: \"d9c047f6-207b-451e-acb0-e909774dc44e\") " pod="openshift-marketplace/certified-operators-2dpx8" Mar 13 14:31:30 crc kubenswrapper[4907]: I0313 14:31:30.824677 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ttd6l\" (UniqueName: \"kubernetes.io/projected/d9c047f6-207b-451e-acb0-e909774dc44e-kube-api-access-ttd6l\") pod \"certified-operators-2dpx8\" (UID: \"d9c047f6-207b-451e-acb0-e909774dc44e\") " pod="openshift-marketplace/certified-operators-2dpx8" Mar 13 14:31:30 crc kubenswrapper[4907]: I0313 14:31:30.855614 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2dpx8" Mar 13 14:31:31 crc kubenswrapper[4907]: I0313 14:31:31.349444 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2dpx8"] Mar 13 14:31:31 crc kubenswrapper[4907]: I0313 14:31:31.967708 4907 generic.go:334] "Generic (PLEG): container finished" podID="d9c047f6-207b-451e-acb0-e909774dc44e" containerID="764acaf87f75bed17bfb66456c3fd4c1cb85fa1c820674eb12e9306b1a048961" exitCode=0 Mar 13 14:31:31 crc kubenswrapper[4907]: I0313 14:31:31.967802 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2dpx8" event={"ID":"d9c047f6-207b-451e-acb0-e909774dc44e","Type":"ContainerDied","Data":"764acaf87f75bed17bfb66456c3fd4c1cb85fa1c820674eb12e9306b1a048961"} Mar 13 14:31:31 crc kubenswrapper[4907]: I0313 14:31:31.968076 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2dpx8" event={"ID":"d9c047f6-207b-451e-acb0-e909774dc44e","Type":"ContainerStarted","Data":"86437821014ef87426f69aad718b9aa1d504d40345fbe79e3129c90fbf55b0d2"} Mar 13 14:31:31 crc kubenswrapper[4907]: I0313 14:31:31.969986 4907 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 13 14:31:32 crc kubenswrapper[4907]: I0313 14:31:32.979175 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2dpx8" event={"ID":"d9c047f6-207b-451e-acb0-e909774dc44e","Type":"ContainerStarted","Data":"a97db1b0d0bfe03ca4dcb913a70fc3028d28f841560e547d602b7d23a4710e8e"} Mar 13 14:31:33 crc kubenswrapper[4907]: I0313 14:31:33.990978 4907 generic.go:334] "Generic (PLEG): container finished" podID="d9c047f6-207b-451e-acb0-e909774dc44e" containerID="a97db1b0d0bfe03ca4dcb913a70fc3028d28f841560e547d602b7d23a4710e8e" exitCode=0 Mar 13 14:31:33 crc kubenswrapper[4907]: I0313 14:31:33.991032 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2dpx8" event={"ID":"d9c047f6-207b-451e-acb0-e909774dc44e","Type":"ContainerDied","Data":"a97db1b0d0bfe03ca4dcb913a70fc3028d28f841560e547d602b7d23a4710e8e"} Mar 13 14:31:35 crc kubenswrapper[4907]: I0313 14:31:35.002265 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2dpx8" event={"ID":"d9c047f6-207b-451e-acb0-e909774dc44e","Type":"ContainerStarted","Data":"e745fb2cd1882130e47435a6e6cfa910c9eb1a8493a9ef4cf1add5adbf625f8c"} Mar 13 14:31:35 crc kubenswrapper[4907]: I0313 14:31:35.020458 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-2dpx8" podStartSLOduration=2.458874806 podStartE2EDuration="5.020437561s" podCreationTimestamp="2026-03-13 14:31:30 +0000 UTC" firstStartedPulling="2026-03-13 14:31:31.969588264 +0000 UTC m=+1590.869375983" lastFinishedPulling="2026-03-13 14:31:34.531151049 +0000 UTC m=+1593.430938738" observedRunningTime="2026-03-13 14:31:35.019114285 +0000 UTC m=+1593.918901984" watchObservedRunningTime="2026-03-13 14:31:35.020437561 +0000 UTC m=+1593.920225250" Mar 13 14:31:40 crc kubenswrapper[4907]: I0313 14:31:40.855982 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-2dpx8" Mar 13 14:31:40 crc kubenswrapper[4907]: I0313 14:31:40.856308 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-2dpx8" Mar 13 14:31:40 crc kubenswrapper[4907]: I0313 14:31:40.901049 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-2dpx8" Mar 13 14:31:41 crc kubenswrapper[4907]: I0313 14:31:41.104754 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-2dpx8" Mar 13 14:31:41 crc kubenswrapper[4907]: I0313 14:31:41.161658 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-2dpx8"] Mar 13 14:31:43 crc kubenswrapper[4907]: I0313 14:31:43.076987 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-2dpx8" podUID="d9c047f6-207b-451e-acb0-e909774dc44e" containerName="registry-server" containerID="cri-o://e745fb2cd1882130e47435a6e6cfa910c9eb1a8493a9ef4cf1add5adbf625f8c" gracePeriod=2 Mar 13 14:31:43 crc kubenswrapper[4907]: I0313 14:31:43.437747 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2dpx8" Mar 13 14:31:43 crc kubenswrapper[4907]: I0313 14:31:43.589975 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d9c047f6-207b-451e-acb0-e909774dc44e-catalog-content\") pod \"d9c047f6-207b-451e-acb0-e909774dc44e\" (UID: \"d9c047f6-207b-451e-acb0-e909774dc44e\") " Mar 13 14:31:43 crc kubenswrapper[4907]: I0313 14:31:43.590026 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ttd6l\" (UniqueName: \"kubernetes.io/projected/d9c047f6-207b-451e-acb0-e909774dc44e-kube-api-access-ttd6l\") pod \"d9c047f6-207b-451e-acb0-e909774dc44e\" (UID: \"d9c047f6-207b-451e-acb0-e909774dc44e\") " Mar 13 14:31:43 crc kubenswrapper[4907]: I0313 14:31:43.590050 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d9c047f6-207b-451e-acb0-e909774dc44e-utilities\") pod \"d9c047f6-207b-451e-acb0-e909774dc44e\" (UID: \"d9c047f6-207b-451e-acb0-e909774dc44e\") " Mar 13 14:31:43 crc kubenswrapper[4907]: I0313 14:31:43.591143 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d9c047f6-207b-451e-acb0-e909774dc44e-utilities" (OuterVolumeSpecName: "utilities") pod "d9c047f6-207b-451e-acb0-e909774dc44e" (UID: "d9c047f6-207b-451e-acb0-e909774dc44e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:31:43 crc kubenswrapper[4907]: I0313 14:31:43.595763 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d9c047f6-207b-451e-acb0-e909774dc44e-kube-api-access-ttd6l" (OuterVolumeSpecName: "kube-api-access-ttd6l") pod "d9c047f6-207b-451e-acb0-e909774dc44e" (UID: "d9c047f6-207b-451e-acb0-e909774dc44e"). InnerVolumeSpecName "kube-api-access-ttd6l". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:31:43 crc kubenswrapper[4907]: I0313 14:31:43.692036 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ttd6l\" (UniqueName: \"kubernetes.io/projected/d9c047f6-207b-451e-acb0-e909774dc44e-kube-api-access-ttd6l\") on node \"crc\" DevicePath \"\"" Mar 13 14:31:43 crc kubenswrapper[4907]: I0313 14:31:43.692068 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d9c047f6-207b-451e-acb0-e909774dc44e-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 14:31:44 crc kubenswrapper[4907]: I0313 14:31:44.088978 4907 generic.go:334] "Generic (PLEG): container finished" podID="d9c047f6-207b-451e-acb0-e909774dc44e" containerID="e745fb2cd1882130e47435a6e6cfa910c9eb1a8493a9ef4cf1add5adbf625f8c" exitCode=0 Mar 13 14:31:44 crc kubenswrapper[4907]: I0313 14:31:44.089027 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2dpx8" event={"ID":"d9c047f6-207b-451e-acb0-e909774dc44e","Type":"ContainerDied","Data":"e745fb2cd1882130e47435a6e6cfa910c9eb1a8493a9ef4cf1add5adbf625f8c"} Mar 13 14:31:44 crc kubenswrapper[4907]: I0313 14:31:44.089110 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2dpx8" event={"ID":"d9c047f6-207b-451e-acb0-e909774dc44e","Type":"ContainerDied","Data":"86437821014ef87426f69aad718b9aa1d504d40345fbe79e3129c90fbf55b0d2"} Mar 13 14:31:44 crc kubenswrapper[4907]: I0313 14:31:44.089138 4907 scope.go:117] "RemoveContainer" containerID="e745fb2cd1882130e47435a6e6cfa910c9eb1a8493a9ef4cf1add5adbf625f8c" Mar 13 14:31:44 crc kubenswrapper[4907]: I0313 14:31:44.089044 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2dpx8" Mar 13 14:31:44 crc kubenswrapper[4907]: I0313 14:31:44.109195 4907 scope.go:117] "RemoveContainer" containerID="a97db1b0d0bfe03ca4dcb913a70fc3028d28f841560e547d602b7d23a4710e8e" Mar 13 14:31:44 crc kubenswrapper[4907]: I0313 14:31:44.130990 4907 scope.go:117] "RemoveContainer" containerID="764acaf87f75bed17bfb66456c3fd4c1cb85fa1c820674eb12e9306b1a048961" Mar 13 14:31:44 crc kubenswrapper[4907]: I0313 14:31:44.172987 4907 scope.go:117] "RemoveContainer" containerID="e745fb2cd1882130e47435a6e6cfa910c9eb1a8493a9ef4cf1add5adbf625f8c" Mar 13 14:31:44 crc kubenswrapper[4907]: E0313 14:31:44.173543 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e745fb2cd1882130e47435a6e6cfa910c9eb1a8493a9ef4cf1add5adbf625f8c\": container with ID starting with e745fb2cd1882130e47435a6e6cfa910c9eb1a8493a9ef4cf1add5adbf625f8c not found: ID does not exist" containerID="e745fb2cd1882130e47435a6e6cfa910c9eb1a8493a9ef4cf1add5adbf625f8c" Mar 13 14:31:44 crc kubenswrapper[4907]: I0313 14:31:44.173582 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e745fb2cd1882130e47435a6e6cfa910c9eb1a8493a9ef4cf1add5adbf625f8c"} err="failed to get container status \"e745fb2cd1882130e47435a6e6cfa910c9eb1a8493a9ef4cf1add5adbf625f8c\": rpc error: code = NotFound desc = could not find container \"e745fb2cd1882130e47435a6e6cfa910c9eb1a8493a9ef4cf1add5adbf625f8c\": container with ID starting with e745fb2cd1882130e47435a6e6cfa910c9eb1a8493a9ef4cf1add5adbf625f8c not found: ID does not exist" Mar 13 14:31:44 crc kubenswrapper[4907]: I0313 14:31:44.173605 4907 scope.go:117] "RemoveContainer" containerID="a97db1b0d0bfe03ca4dcb913a70fc3028d28f841560e547d602b7d23a4710e8e" Mar 13 14:31:44 crc kubenswrapper[4907]: E0313 14:31:44.173894 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a97db1b0d0bfe03ca4dcb913a70fc3028d28f841560e547d602b7d23a4710e8e\": container with ID starting with a97db1b0d0bfe03ca4dcb913a70fc3028d28f841560e547d602b7d23a4710e8e not found: ID does not exist" containerID="a97db1b0d0bfe03ca4dcb913a70fc3028d28f841560e547d602b7d23a4710e8e" Mar 13 14:31:44 crc kubenswrapper[4907]: I0313 14:31:44.173927 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a97db1b0d0bfe03ca4dcb913a70fc3028d28f841560e547d602b7d23a4710e8e"} err="failed to get container status \"a97db1b0d0bfe03ca4dcb913a70fc3028d28f841560e547d602b7d23a4710e8e\": rpc error: code = NotFound desc = could not find container \"a97db1b0d0bfe03ca4dcb913a70fc3028d28f841560e547d602b7d23a4710e8e\": container with ID starting with a97db1b0d0bfe03ca4dcb913a70fc3028d28f841560e547d602b7d23a4710e8e not found: ID does not exist" Mar 13 14:31:44 crc kubenswrapper[4907]: I0313 14:31:44.173947 4907 scope.go:117] "RemoveContainer" containerID="764acaf87f75bed17bfb66456c3fd4c1cb85fa1c820674eb12e9306b1a048961" Mar 13 14:31:44 crc kubenswrapper[4907]: E0313 14:31:44.174228 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"764acaf87f75bed17bfb66456c3fd4c1cb85fa1c820674eb12e9306b1a048961\": container with ID starting with 764acaf87f75bed17bfb66456c3fd4c1cb85fa1c820674eb12e9306b1a048961 not found: ID does not exist" containerID="764acaf87f75bed17bfb66456c3fd4c1cb85fa1c820674eb12e9306b1a048961" Mar 13 14:31:44 crc kubenswrapper[4907]: I0313 14:31:44.174250 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"764acaf87f75bed17bfb66456c3fd4c1cb85fa1c820674eb12e9306b1a048961"} err="failed to get container status \"764acaf87f75bed17bfb66456c3fd4c1cb85fa1c820674eb12e9306b1a048961\": rpc error: code = NotFound desc = could not find container \"764acaf87f75bed17bfb66456c3fd4c1cb85fa1c820674eb12e9306b1a048961\": container with ID starting with 764acaf87f75bed17bfb66456c3fd4c1cb85fa1c820674eb12e9306b1a048961 not found: ID does not exist" Mar 13 14:31:44 crc kubenswrapper[4907]: I0313 14:31:44.270264 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d9c047f6-207b-451e-acb0-e909774dc44e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d9c047f6-207b-451e-acb0-e909774dc44e" (UID: "d9c047f6-207b-451e-acb0-e909774dc44e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:31:44 crc kubenswrapper[4907]: I0313 14:31:44.301509 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d9c047f6-207b-451e-acb0-e909774dc44e-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 14:31:44 crc kubenswrapper[4907]: I0313 14:31:44.423525 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-2dpx8"] Mar 13 14:31:44 crc kubenswrapper[4907]: I0313 14:31:44.431048 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-2dpx8"] Mar 13 14:31:45 crc kubenswrapper[4907]: I0313 14:31:45.792844 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d9c047f6-207b-451e-acb0-e909774dc44e" path="/var/lib/kubelet/pods/d9c047f6-207b-451e-acb0-e909774dc44e/volumes" Mar 13 14:31:48 crc kubenswrapper[4907]: I0313 14:31:48.041869 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 14:31:48 crc kubenswrapper[4907]: I0313 14:31:48.042746 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 14:32:00 crc kubenswrapper[4907]: I0313 14:32:00.145460 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556872-srgh9"] Mar 13 14:32:00 crc kubenswrapper[4907]: E0313 14:32:00.146416 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9c047f6-207b-451e-acb0-e909774dc44e" containerName="registry-server" Mar 13 14:32:00 crc kubenswrapper[4907]: I0313 14:32:00.146434 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9c047f6-207b-451e-acb0-e909774dc44e" containerName="registry-server" Mar 13 14:32:00 crc kubenswrapper[4907]: E0313 14:32:00.146457 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9c047f6-207b-451e-acb0-e909774dc44e" containerName="extract-content" Mar 13 14:32:00 crc kubenswrapper[4907]: I0313 14:32:00.146465 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9c047f6-207b-451e-acb0-e909774dc44e" containerName="extract-content" Mar 13 14:32:00 crc kubenswrapper[4907]: E0313 14:32:00.146478 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9c047f6-207b-451e-acb0-e909774dc44e" containerName="extract-utilities" Mar 13 14:32:00 crc kubenswrapper[4907]: I0313 14:32:00.146486 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9c047f6-207b-451e-acb0-e909774dc44e" containerName="extract-utilities" Mar 13 14:32:00 crc kubenswrapper[4907]: I0313 14:32:00.146638 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="d9c047f6-207b-451e-acb0-e909774dc44e" containerName="registry-server" Mar 13 14:32:00 crc kubenswrapper[4907]: I0313 14:32:00.147235 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556872-srgh9" Mar 13 14:32:00 crc kubenswrapper[4907]: I0313 14:32:00.152124 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 14:32:00 crc kubenswrapper[4907]: I0313 14:32:00.152683 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 14:32:00 crc kubenswrapper[4907]: I0313 14:32:00.155632 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 14:32:00 crc kubenswrapper[4907]: I0313 14:32:00.158646 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556872-srgh9"] Mar 13 14:32:00 crc kubenswrapper[4907]: I0313 14:32:00.312793 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wfpwt\" (UniqueName: \"kubernetes.io/projected/32297c61-6461-4357-832f-9de703d88000-kube-api-access-wfpwt\") pod \"auto-csr-approver-29556872-srgh9\" (UID: \"32297c61-6461-4357-832f-9de703d88000\") " pod="openshift-infra/auto-csr-approver-29556872-srgh9" Mar 13 14:32:00 crc kubenswrapper[4907]: I0313 14:32:00.413726 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wfpwt\" (UniqueName: \"kubernetes.io/projected/32297c61-6461-4357-832f-9de703d88000-kube-api-access-wfpwt\") pod \"auto-csr-approver-29556872-srgh9\" (UID: \"32297c61-6461-4357-832f-9de703d88000\") " pod="openshift-infra/auto-csr-approver-29556872-srgh9" Mar 13 14:32:00 crc kubenswrapper[4907]: I0313 14:32:00.435165 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wfpwt\" (UniqueName: \"kubernetes.io/projected/32297c61-6461-4357-832f-9de703d88000-kube-api-access-wfpwt\") pod \"auto-csr-approver-29556872-srgh9\" (UID: \"32297c61-6461-4357-832f-9de703d88000\") " pod="openshift-infra/auto-csr-approver-29556872-srgh9" Mar 13 14:32:00 crc kubenswrapper[4907]: I0313 14:32:00.468560 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556872-srgh9" Mar 13 14:32:00 crc kubenswrapper[4907]: I0313 14:32:00.885270 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556872-srgh9"] Mar 13 14:32:00 crc kubenswrapper[4907]: I0313 14:32:00.992891 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556872-srgh9" event={"ID":"32297c61-6461-4357-832f-9de703d88000","Type":"ContainerStarted","Data":"4b360592a61dc96e94c5095ba7d5510824692c1ccb09aa93abcdd78a8df14367"} Mar 13 14:32:03 crc kubenswrapper[4907]: I0313 14:32:03.011514 4907 generic.go:334] "Generic (PLEG): container finished" podID="32297c61-6461-4357-832f-9de703d88000" containerID="8eea35f2339f97a9cd81df041bdf950f47b0d8ec3a69dea570886bdde6ad88a7" exitCode=0 Mar 13 14:32:03 crc kubenswrapper[4907]: I0313 14:32:03.011622 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556872-srgh9" event={"ID":"32297c61-6461-4357-832f-9de703d88000","Type":"ContainerDied","Data":"8eea35f2339f97a9cd81df041bdf950f47b0d8ec3a69dea570886bdde6ad88a7"} Mar 13 14:32:04 crc kubenswrapper[4907]: I0313 14:32:04.324175 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556872-srgh9" Mar 13 14:32:04 crc kubenswrapper[4907]: I0313 14:32:04.475253 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wfpwt\" (UniqueName: \"kubernetes.io/projected/32297c61-6461-4357-832f-9de703d88000-kube-api-access-wfpwt\") pod \"32297c61-6461-4357-832f-9de703d88000\" (UID: \"32297c61-6461-4357-832f-9de703d88000\") " Mar 13 14:32:04 crc kubenswrapper[4907]: I0313 14:32:04.481553 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/32297c61-6461-4357-832f-9de703d88000-kube-api-access-wfpwt" (OuterVolumeSpecName: "kube-api-access-wfpwt") pod "32297c61-6461-4357-832f-9de703d88000" (UID: "32297c61-6461-4357-832f-9de703d88000"). InnerVolumeSpecName "kube-api-access-wfpwt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:32:04 crc kubenswrapper[4907]: I0313 14:32:04.577556 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wfpwt\" (UniqueName: \"kubernetes.io/projected/32297c61-6461-4357-832f-9de703d88000-kube-api-access-wfpwt\") on node \"crc\" DevicePath \"\"" Mar 13 14:32:05 crc kubenswrapper[4907]: I0313 14:32:05.026602 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556872-srgh9" event={"ID":"32297c61-6461-4357-832f-9de703d88000","Type":"ContainerDied","Data":"4b360592a61dc96e94c5095ba7d5510824692c1ccb09aa93abcdd78a8df14367"} Mar 13 14:32:05 crc kubenswrapper[4907]: I0313 14:32:05.026642 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4b360592a61dc96e94c5095ba7d5510824692c1ccb09aa93abcdd78a8df14367" Mar 13 14:32:05 crc kubenswrapper[4907]: I0313 14:32:05.026672 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556872-srgh9" Mar 13 14:32:05 crc kubenswrapper[4907]: I0313 14:32:05.393785 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556866-fwg48"] Mar 13 14:32:05 crc kubenswrapper[4907]: I0313 14:32:05.400003 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556866-fwg48"] Mar 13 14:32:05 crc kubenswrapper[4907]: I0313 14:32:05.793311 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7d930b98-c1af-4e47-a663-86afa9484856" path="/var/lib/kubelet/pods/7d930b98-c1af-4e47-a663-86afa9484856/volumes" Mar 13 14:32:18 crc kubenswrapper[4907]: I0313 14:32:18.041437 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 14:32:18 crc kubenswrapper[4907]: I0313 14:32:18.042034 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 14:32:18 crc kubenswrapper[4907]: I0313 14:32:18.042077 4907 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" Mar 13 14:32:18 crc kubenswrapper[4907]: I0313 14:32:18.042715 4907 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a09a3bec467b7e275495edb5827150af756bcc256911e7e06cbae5765922b5e2"} pod="openshift-machine-config-operator/machine-config-daemon-hm56j" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 14:32:18 crc kubenswrapper[4907]: I0313 14:32:18.042770 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" containerID="cri-o://a09a3bec467b7e275495edb5827150af756bcc256911e7e06cbae5765922b5e2" gracePeriod=600 Mar 13 14:32:18 crc kubenswrapper[4907]: E0313 14:32:18.171104 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:32:19 crc kubenswrapper[4907]: I0313 14:32:19.142250 4907 generic.go:334] "Generic (PLEG): container finished" podID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerID="a09a3bec467b7e275495edb5827150af756bcc256911e7e06cbae5765922b5e2" exitCode=0 Mar 13 14:32:19 crc kubenswrapper[4907]: I0313 14:32:19.142295 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerDied","Data":"a09a3bec467b7e275495edb5827150af756bcc256911e7e06cbae5765922b5e2"} Mar 13 14:32:19 crc kubenswrapper[4907]: I0313 14:32:19.142332 4907 scope.go:117] "RemoveContainer" containerID="1f79a88cd4431d2b6772ef1dc1fb18ce20d965a3a936a7157e03bc28ca239c57" Mar 13 14:32:19 crc kubenswrapper[4907]: I0313 14:32:19.143122 4907 scope.go:117] "RemoveContainer" containerID="a09a3bec467b7e275495edb5827150af756bcc256911e7e06cbae5765922b5e2" Mar 13 14:32:19 crc kubenswrapper[4907]: E0313 14:32:19.143512 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:32:22 crc kubenswrapper[4907]: I0313 14:32:22.539643 4907 scope.go:117] "RemoveContainer" containerID="93ab187126d08591f8a6823b5ca150c5522a7db6aea7e1ca8cc868ee0d426800" Mar 13 14:32:22 crc kubenswrapper[4907]: I0313 14:32:22.562639 4907 scope.go:117] "RemoveContainer" containerID="37d0b4abf880f9988e205a5d09debf0e634cb45e2049ee542b43a256eccbec25" Mar 13 14:32:22 crc kubenswrapper[4907]: I0313 14:32:22.624463 4907 scope.go:117] "RemoveContainer" containerID="65475a4b4006963b7fec1fa34278a44b7bf08cbcf9c088a764d7d29c870c630c" Mar 13 14:32:22 crc kubenswrapper[4907]: I0313 14:32:22.648198 4907 scope.go:117] "RemoveContainer" containerID="b2e917cc4c559e06be2ed2a503af651b2cead394bb06409f6e1c888788df817a" Mar 13 14:32:22 crc kubenswrapper[4907]: I0313 14:32:22.666477 4907 scope.go:117] "RemoveContainer" containerID="d5c1cbc11101f46dd814717af56c35f8ffa0c1cd1640b4a7f3a2d4236359a48e" Mar 13 14:32:22 crc kubenswrapper[4907]: I0313 14:32:22.691329 4907 scope.go:117] "RemoveContainer" containerID="ae6341ae2f61b11a2ad0c4f86bf51fc609fce628a9af3e20ce77aa5f1d917a69" Mar 13 14:32:22 crc kubenswrapper[4907]: I0313 14:32:22.712484 4907 scope.go:117] "RemoveContainer" containerID="1953abb8d3f9054081c314893bde9d173fee7430bf61ccd6ae11b908a7002729" Mar 13 14:32:22 crc kubenswrapper[4907]: I0313 14:32:22.743136 4907 scope.go:117] "RemoveContainer" containerID="e6d88705fdc67e86fd0214f4c6bbf1c37e140065fdd712f700392f915f879795" Mar 13 14:32:22 crc kubenswrapper[4907]: I0313 14:32:22.762120 4907 scope.go:117] "RemoveContainer" containerID="3d70e191441a97faf79353905c86d9af8e9fd167c63431a9f984ad9d2a36445e" Mar 13 14:32:22 crc kubenswrapper[4907]: I0313 14:32:22.778796 4907 scope.go:117] "RemoveContainer" containerID="b5e68fc91f35838327b362dd7ecd6d2f5ee1173bc7ae4fafcb541ac0c96b44fa" Mar 13 14:32:22 crc kubenswrapper[4907]: I0313 14:32:22.799167 4907 scope.go:117] "RemoveContainer" containerID="c688bfe6d6ea48e683001f8125ee585ba89819e96749b44bbe5f071b99df1226" Mar 13 14:32:22 crc kubenswrapper[4907]: I0313 14:32:22.821586 4907 scope.go:117] "RemoveContainer" containerID="1a454f90b2465a04739b60faa1ea54a4c62cae0b7a6742d83f6b477594d8a8a9" Mar 13 14:32:22 crc kubenswrapper[4907]: I0313 14:32:22.843779 4907 scope.go:117] "RemoveContainer" containerID="3c90645adf755c84a8db83dd2e03403f66a250afbd541346b9261208e83c56ff" Mar 13 14:32:29 crc kubenswrapper[4907]: I0313 14:32:29.883195 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-4rn6k"] Mar 13 14:32:29 crc kubenswrapper[4907]: E0313 14:32:29.885382 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32297c61-6461-4357-832f-9de703d88000" containerName="oc" Mar 13 14:32:29 crc kubenswrapper[4907]: I0313 14:32:29.885405 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="32297c61-6461-4357-832f-9de703d88000" containerName="oc" Mar 13 14:32:29 crc kubenswrapper[4907]: I0313 14:32:29.885607 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="32297c61-6461-4357-832f-9de703d88000" containerName="oc" Mar 13 14:32:29 crc kubenswrapper[4907]: I0313 14:32:29.886870 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4rn6k" Mar 13 14:32:29 crc kubenswrapper[4907]: I0313 14:32:29.898074 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4rn6k"] Mar 13 14:32:29 crc kubenswrapper[4907]: I0313 14:32:29.926453 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cxlz4\" (UniqueName: \"kubernetes.io/projected/bf79c546-2d9a-4eba-bf86-f78d196bc557-kube-api-access-cxlz4\") pod \"community-operators-4rn6k\" (UID: \"bf79c546-2d9a-4eba-bf86-f78d196bc557\") " pod="openshift-marketplace/community-operators-4rn6k" Mar 13 14:32:29 crc kubenswrapper[4907]: I0313 14:32:29.926527 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf79c546-2d9a-4eba-bf86-f78d196bc557-catalog-content\") pod \"community-operators-4rn6k\" (UID: \"bf79c546-2d9a-4eba-bf86-f78d196bc557\") " pod="openshift-marketplace/community-operators-4rn6k" Mar 13 14:32:29 crc kubenswrapper[4907]: I0313 14:32:29.926544 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf79c546-2d9a-4eba-bf86-f78d196bc557-utilities\") pod \"community-operators-4rn6k\" (UID: \"bf79c546-2d9a-4eba-bf86-f78d196bc557\") " pod="openshift-marketplace/community-operators-4rn6k" Mar 13 14:32:30 crc kubenswrapper[4907]: I0313 14:32:30.027747 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf79c546-2d9a-4eba-bf86-f78d196bc557-catalog-content\") pod \"community-operators-4rn6k\" (UID: \"bf79c546-2d9a-4eba-bf86-f78d196bc557\") " pod="openshift-marketplace/community-operators-4rn6k" Mar 13 14:32:30 crc kubenswrapper[4907]: I0313 14:32:30.027790 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf79c546-2d9a-4eba-bf86-f78d196bc557-utilities\") pod \"community-operators-4rn6k\" (UID: \"bf79c546-2d9a-4eba-bf86-f78d196bc557\") " pod="openshift-marketplace/community-operators-4rn6k" Mar 13 14:32:30 crc kubenswrapper[4907]: I0313 14:32:30.027855 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cxlz4\" (UniqueName: \"kubernetes.io/projected/bf79c546-2d9a-4eba-bf86-f78d196bc557-kube-api-access-cxlz4\") pod \"community-operators-4rn6k\" (UID: \"bf79c546-2d9a-4eba-bf86-f78d196bc557\") " pod="openshift-marketplace/community-operators-4rn6k" Mar 13 14:32:30 crc kubenswrapper[4907]: I0313 14:32:30.028354 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf79c546-2d9a-4eba-bf86-f78d196bc557-catalog-content\") pod \"community-operators-4rn6k\" (UID: \"bf79c546-2d9a-4eba-bf86-f78d196bc557\") " pod="openshift-marketplace/community-operators-4rn6k" Mar 13 14:32:30 crc kubenswrapper[4907]: I0313 14:32:30.028444 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf79c546-2d9a-4eba-bf86-f78d196bc557-utilities\") pod \"community-operators-4rn6k\" (UID: \"bf79c546-2d9a-4eba-bf86-f78d196bc557\") " pod="openshift-marketplace/community-operators-4rn6k" Mar 13 14:32:30 crc kubenswrapper[4907]: I0313 14:32:30.049006 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cxlz4\" (UniqueName: \"kubernetes.io/projected/bf79c546-2d9a-4eba-bf86-f78d196bc557-kube-api-access-cxlz4\") pod \"community-operators-4rn6k\" (UID: \"bf79c546-2d9a-4eba-bf86-f78d196bc557\") " pod="openshift-marketplace/community-operators-4rn6k" Mar 13 14:32:30 crc kubenswrapper[4907]: I0313 14:32:30.212268 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4rn6k" Mar 13 14:32:30 crc kubenswrapper[4907]: I0313 14:32:30.690100 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4rn6k"] Mar 13 14:32:31 crc kubenswrapper[4907]: I0313 14:32:31.262994 4907 generic.go:334] "Generic (PLEG): container finished" podID="bf79c546-2d9a-4eba-bf86-f78d196bc557" containerID="4bf9c94aaa9cc7c5c0ba8069c717611a0d87cc031d7ffb209791060330b75f67" exitCode=0 Mar 13 14:32:31 crc kubenswrapper[4907]: I0313 14:32:31.263178 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4rn6k" event={"ID":"bf79c546-2d9a-4eba-bf86-f78d196bc557","Type":"ContainerDied","Data":"4bf9c94aaa9cc7c5c0ba8069c717611a0d87cc031d7ffb209791060330b75f67"} Mar 13 14:32:31 crc kubenswrapper[4907]: I0313 14:32:31.263397 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4rn6k" event={"ID":"bf79c546-2d9a-4eba-bf86-f78d196bc557","Type":"ContainerStarted","Data":"13c2f49681df8442c517f0c438f3ee426387fdb04f4ee6868a9b1082ae36005d"} Mar 13 14:32:32 crc kubenswrapper[4907]: I0313 14:32:32.276161 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4rn6k" event={"ID":"bf79c546-2d9a-4eba-bf86-f78d196bc557","Type":"ContainerStarted","Data":"b81eb78ae61aa093de15332c1ee04faf01a340ba43eebb6b2242c49cc2235cbd"} Mar 13 14:32:32 crc kubenswrapper[4907]: I0313 14:32:32.782898 4907 scope.go:117] "RemoveContainer" containerID="a09a3bec467b7e275495edb5827150af756bcc256911e7e06cbae5765922b5e2" Mar 13 14:32:32 crc kubenswrapper[4907]: E0313 14:32:32.783191 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:32:33 crc kubenswrapper[4907]: I0313 14:32:33.287685 4907 generic.go:334] "Generic (PLEG): container finished" podID="bf79c546-2d9a-4eba-bf86-f78d196bc557" containerID="b81eb78ae61aa093de15332c1ee04faf01a340ba43eebb6b2242c49cc2235cbd" exitCode=0 Mar 13 14:32:33 crc kubenswrapper[4907]: I0313 14:32:33.287779 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4rn6k" event={"ID":"bf79c546-2d9a-4eba-bf86-f78d196bc557","Type":"ContainerDied","Data":"b81eb78ae61aa093de15332c1ee04faf01a340ba43eebb6b2242c49cc2235cbd"} Mar 13 14:32:34 crc kubenswrapper[4907]: I0313 14:32:34.299122 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4rn6k" event={"ID":"bf79c546-2d9a-4eba-bf86-f78d196bc557","Type":"ContainerStarted","Data":"aaee29c2d401955a8cb458cf5d817e6a9250580d773d1a23710ccd6b61e1d612"} Mar 13 14:32:34 crc kubenswrapper[4907]: I0313 14:32:34.319554 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-4rn6k" podStartSLOduration=2.747760435 podStartE2EDuration="5.319532045s" podCreationTimestamp="2026-03-13 14:32:29 +0000 UTC" firstStartedPulling="2026-03-13 14:32:31.26784225 +0000 UTC m=+1650.167629949" lastFinishedPulling="2026-03-13 14:32:33.83961387 +0000 UTC m=+1652.739401559" observedRunningTime="2026-03-13 14:32:34.318675883 +0000 UTC m=+1653.218463572" watchObservedRunningTime="2026-03-13 14:32:34.319532045 +0000 UTC m=+1653.219319734" Mar 13 14:32:40 crc kubenswrapper[4907]: I0313 14:32:40.212820 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-4rn6k" Mar 13 14:32:40 crc kubenswrapper[4907]: I0313 14:32:40.213400 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-4rn6k" Mar 13 14:32:40 crc kubenswrapper[4907]: I0313 14:32:40.271814 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-4rn6k" Mar 13 14:32:40 crc kubenswrapper[4907]: I0313 14:32:40.423174 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-4rn6k" Mar 13 14:32:40 crc kubenswrapper[4907]: I0313 14:32:40.512311 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-4rn6k"] Mar 13 14:32:42 crc kubenswrapper[4907]: I0313 14:32:42.373864 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-4rn6k" podUID="bf79c546-2d9a-4eba-bf86-f78d196bc557" containerName="registry-server" containerID="cri-o://aaee29c2d401955a8cb458cf5d817e6a9250580d773d1a23710ccd6b61e1d612" gracePeriod=2 Mar 13 14:32:43 crc kubenswrapper[4907]: I0313 14:32:43.336413 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4rn6k" Mar 13 14:32:43 crc kubenswrapper[4907]: I0313 14:32:43.386649 4907 generic.go:334] "Generic (PLEG): container finished" podID="bf79c546-2d9a-4eba-bf86-f78d196bc557" containerID="aaee29c2d401955a8cb458cf5d817e6a9250580d773d1a23710ccd6b61e1d612" exitCode=0 Mar 13 14:32:43 crc kubenswrapper[4907]: I0313 14:32:43.386696 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4rn6k" event={"ID":"bf79c546-2d9a-4eba-bf86-f78d196bc557","Type":"ContainerDied","Data":"aaee29c2d401955a8cb458cf5d817e6a9250580d773d1a23710ccd6b61e1d612"} Mar 13 14:32:43 crc kubenswrapper[4907]: I0313 14:32:43.386725 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4rn6k" event={"ID":"bf79c546-2d9a-4eba-bf86-f78d196bc557","Type":"ContainerDied","Data":"13c2f49681df8442c517f0c438f3ee426387fdb04f4ee6868a9b1082ae36005d"} Mar 13 14:32:43 crc kubenswrapper[4907]: I0313 14:32:43.386744 4907 scope.go:117] "RemoveContainer" containerID="aaee29c2d401955a8cb458cf5d817e6a9250580d773d1a23710ccd6b61e1d612" Mar 13 14:32:43 crc kubenswrapper[4907]: I0313 14:32:43.386921 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4rn6k" Mar 13 14:32:43 crc kubenswrapper[4907]: I0313 14:32:43.420367 4907 scope.go:117] "RemoveContainer" containerID="b81eb78ae61aa093de15332c1ee04faf01a340ba43eebb6b2242c49cc2235cbd" Mar 13 14:32:43 crc kubenswrapper[4907]: I0313 14:32:43.428431 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cxlz4\" (UniqueName: \"kubernetes.io/projected/bf79c546-2d9a-4eba-bf86-f78d196bc557-kube-api-access-cxlz4\") pod \"bf79c546-2d9a-4eba-bf86-f78d196bc557\" (UID: \"bf79c546-2d9a-4eba-bf86-f78d196bc557\") " Mar 13 14:32:43 crc kubenswrapper[4907]: I0313 14:32:43.428548 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf79c546-2d9a-4eba-bf86-f78d196bc557-utilities\") pod \"bf79c546-2d9a-4eba-bf86-f78d196bc557\" (UID: \"bf79c546-2d9a-4eba-bf86-f78d196bc557\") " Mar 13 14:32:43 crc kubenswrapper[4907]: I0313 14:32:43.428579 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf79c546-2d9a-4eba-bf86-f78d196bc557-catalog-content\") pod \"bf79c546-2d9a-4eba-bf86-f78d196bc557\" (UID: \"bf79c546-2d9a-4eba-bf86-f78d196bc557\") " Mar 13 14:32:43 crc kubenswrapper[4907]: I0313 14:32:43.430258 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bf79c546-2d9a-4eba-bf86-f78d196bc557-utilities" (OuterVolumeSpecName: "utilities") pod "bf79c546-2d9a-4eba-bf86-f78d196bc557" (UID: "bf79c546-2d9a-4eba-bf86-f78d196bc557"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:32:43 crc kubenswrapper[4907]: I0313 14:32:43.438217 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf79c546-2d9a-4eba-bf86-f78d196bc557-kube-api-access-cxlz4" (OuterVolumeSpecName: "kube-api-access-cxlz4") pod "bf79c546-2d9a-4eba-bf86-f78d196bc557" (UID: "bf79c546-2d9a-4eba-bf86-f78d196bc557"). InnerVolumeSpecName "kube-api-access-cxlz4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:32:43 crc kubenswrapper[4907]: I0313 14:32:43.498196 4907 scope.go:117] "RemoveContainer" containerID="4bf9c94aaa9cc7c5c0ba8069c717611a0d87cc031d7ffb209791060330b75f67" Mar 13 14:32:43 crc kubenswrapper[4907]: I0313 14:32:43.519244 4907 scope.go:117] "RemoveContainer" containerID="aaee29c2d401955a8cb458cf5d817e6a9250580d773d1a23710ccd6b61e1d612" Mar 13 14:32:43 crc kubenswrapper[4907]: E0313 14:32:43.519642 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aaee29c2d401955a8cb458cf5d817e6a9250580d773d1a23710ccd6b61e1d612\": container with ID starting with aaee29c2d401955a8cb458cf5d817e6a9250580d773d1a23710ccd6b61e1d612 not found: ID does not exist" containerID="aaee29c2d401955a8cb458cf5d817e6a9250580d773d1a23710ccd6b61e1d612" Mar 13 14:32:43 crc kubenswrapper[4907]: I0313 14:32:43.519674 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aaee29c2d401955a8cb458cf5d817e6a9250580d773d1a23710ccd6b61e1d612"} err="failed to get container status \"aaee29c2d401955a8cb458cf5d817e6a9250580d773d1a23710ccd6b61e1d612\": rpc error: code = NotFound desc = could not find container \"aaee29c2d401955a8cb458cf5d817e6a9250580d773d1a23710ccd6b61e1d612\": container with ID starting with aaee29c2d401955a8cb458cf5d817e6a9250580d773d1a23710ccd6b61e1d612 not found: ID does not exist" Mar 13 14:32:43 crc kubenswrapper[4907]: I0313 14:32:43.519693 4907 scope.go:117] "RemoveContainer" containerID="b81eb78ae61aa093de15332c1ee04faf01a340ba43eebb6b2242c49cc2235cbd" Mar 13 14:32:43 crc kubenswrapper[4907]: E0313 14:32:43.520096 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b81eb78ae61aa093de15332c1ee04faf01a340ba43eebb6b2242c49cc2235cbd\": container with ID starting with b81eb78ae61aa093de15332c1ee04faf01a340ba43eebb6b2242c49cc2235cbd not found: ID does not exist" containerID="b81eb78ae61aa093de15332c1ee04faf01a340ba43eebb6b2242c49cc2235cbd" Mar 13 14:32:43 crc kubenswrapper[4907]: I0313 14:32:43.520119 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b81eb78ae61aa093de15332c1ee04faf01a340ba43eebb6b2242c49cc2235cbd"} err="failed to get container status \"b81eb78ae61aa093de15332c1ee04faf01a340ba43eebb6b2242c49cc2235cbd\": rpc error: code = NotFound desc = could not find container \"b81eb78ae61aa093de15332c1ee04faf01a340ba43eebb6b2242c49cc2235cbd\": container with ID starting with b81eb78ae61aa093de15332c1ee04faf01a340ba43eebb6b2242c49cc2235cbd not found: ID does not exist" Mar 13 14:32:43 crc kubenswrapper[4907]: I0313 14:32:43.520133 4907 scope.go:117] "RemoveContainer" containerID="4bf9c94aaa9cc7c5c0ba8069c717611a0d87cc031d7ffb209791060330b75f67" Mar 13 14:32:43 crc kubenswrapper[4907]: E0313 14:32:43.520418 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4bf9c94aaa9cc7c5c0ba8069c717611a0d87cc031d7ffb209791060330b75f67\": container with ID starting with 4bf9c94aaa9cc7c5c0ba8069c717611a0d87cc031d7ffb209791060330b75f67 not found: ID does not exist" containerID="4bf9c94aaa9cc7c5c0ba8069c717611a0d87cc031d7ffb209791060330b75f67" Mar 13 14:32:43 crc kubenswrapper[4907]: I0313 14:32:43.520436 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4bf9c94aaa9cc7c5c0ba8069c717611a0d87cc031d7ffb209791060330b75f67"} err="failed to get container status \"4bf9c94aaa9cc7c5c0ba8069c717611a0d87cc031d7ffb209791060330b75f67\": rpc error: code = NotFound desc = could not find container \"4bf9c94aaa9cc7c5c0ba8069c717611a0d87cc031d7ffb209791060330b75f67\": container with ID starting with 4bf9c94aaa9cc7c5c0ba8069c717611a0d87cc031d7ffb209791060330b75f67 not found: ID does not exist" Mar 13 14:32:43 crc kubenswrapper[4907]: I0313 14:32:43.529706 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cxlz4\" (UniqueName: \"kubernetes.io/projected/bf79c546-2d9a-4eba-bf86-f78d196bc557-kube-api-access-cxlz4\") on node \"crc\" DevicePath \"\"" Mar 13 14:32:43 crc kubenswrapper[4907]: I0313 14:32:43.529749 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf79c546-2d9a-4eba-bf86-f78d196bc557-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 14:32:43 crc kubenswrapper[4907]: I0313 14:32:43.531950 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bf79c546-2d9a-4eba-bf86-f78d196bc557-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bf79c546-2d9a-4eba-bf86-f78d196bc557" (UID: "bf79c546-2d9a-4eba-bf86-f78d196bc557"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:32:43 crc kubenswrapper[4907]: I0313 14:32:43.631013 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf79c546-2d9a-4eba-bf86-f78d196bc557-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 14:32:43 crc kubenswrapper[4907]: I0313 14:32:43.715986 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-4rn6k"] Mar 13 14:32:43 crc kubenswrapper[4907]: I0313 14:32:43.723073 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-4rn6k"] Mar 13 14:32:43 crc kubenswrapper[4907]: I0313 14:32:43.797264 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf79c546-2d9a-4eba-bf86-f78d196bc557" path="/var/lib/kubelet/pods/bf79c546-2d9a-4eba-bf86-f78d196bc557/volumes" Mar 13 14:32:45 crc kubenswrapper[4907]: I0313 14:32:45.783495 4907 scope.go:117] "RemoveContainer" containerID="a09a3bec467b7e275495edb5827150af756bcc256911e7e06cbae5765922b5e2" Mar 13 14:32:45 crc kubenswrapper[4907]: E0313 14:32:45.783791 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:33:00 crc kubenswrapper[4907]: I0313 14:33:00.782149 4907 scope.go:117] "RemoveContainer" containerID="a09a3bec467b7e275495edb5827150af756bcc256911e7e06cbae5765922b5e2" Mar 13 14:33:00 crc kubenswrapper[4907]: E0313 14:33:00.783200 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:33:13 crc kubenswrapper[4907]: I0313 14:33:13.782873 4907 scope.go:117] "RemoveContainer" containerID="a09a3bec467b7e275495edb5827150af756bcc256911e7e06cbae5765922b5e2" Mar 13 14:33:13 crc kubenswrapper[4907]: E0313 14:33:13.784904 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:33:23 crc kubenswrapper[4907]: I0313 14:33:23.091140 4907 scope.go:117] "RemoveContainer" containerID="8a34f94ac114c7c6f4deeb646d486da44df7c2de5f8d5ce22726005779d1271f" Mar 13 14:33:23 crc kubenswrapper[4907]: I0313 14:33:23.128715 4907 scope.go:117] "RemoveContainer" containerID="8f4cb6476a276775b25f72fa46f11e3f550f531236e5cf19234f8b274637da17" Mar 13 14:33:23 crc kubenswrapper[4907]: I0313 14:33:23.147224 4907 scope.go:117] "RemoveContainer" containerID="4e3ebec78f62227a1eec084cd7d40d9bb86b61261221bf385169185e602cfad5" Mar 13 14:33:23 crc kubenswrapper[4907]: I0313 14:33:23.175111 4907 scope.go:117] "RemoveContainer" containerID="fad76ed01fe4f9787942e4f47ff8ffafe18d1d8c62a2533bd080f0d2d246c933" Mar 13 14:33:23 crc kubenswrapper[4907]: I0313 14:33:23.210033 4907 scope.go:117] "RemoveContainer" containerID="3e1e04ee01cb4c593a590da900f137153aed8f7c58309d1a24f19dd94c9af998" Mar 13 14:33:23 crc kubenswrapper[4907]: I0313 14:33:23.231657 4907 scope.go:117] "RemoveContainer" containerID="5b6a2afb9b4ad5f576a34ca212d94796e12f98db7c7c3a2d4bb1600fc3d60034" Mar 13 14:33:23 crc kubenswrapper[4907]: I0313 14:33:23.260445 4907 scope.go:117] "RemoveContainer" containerID="70df9585a66270db2bec365cac2492109be5b468d9fc1bab776dec0ef7304eaa" Mar 13 14:33:23 crc kubenswrapper[4907]: I0313 14:33:23.297041 4907 scope.go:117] "RemoveContainer" containerID="c90c150f12af47e1b50f63c9ddceacb4d1551893cbeff831172fa51757c596cb" Mar 13 14:33:23 crc kubenswrapper[4907]: I0313 14:33:23.332855 4907 scope.go:117] "RemoveContainer" containerID="cdd8592044b8b42302c000b60e593bcd7aa7352aa0bfd95a951e13ddb4d1cef1" Mar 13 14:33:23 crc kubenswrapper[4907]: I0313 14:33:23.350976 4907 scope.go:117] "RemoveContainer" containerID="f407f7b1e1671cab73872a73a57d80595693b20473de90bda4d6de97abc60cd5" Mar 13 14:33:23 crc kubenswrapper[4907]: I0313 14:33:23.379347 4907 scope.go:117] "RemoveContainer" containerID="0a3f47d3069fe894c10b414d1b636502b45151a11f2acdaa45abe3e56ee5a7a1" Mar 13 14:33:23 crc kubenswrapper[4907]: I0313 14:33:23.402739 4907 scope.go:117] "RemoveContainer" containerID="46c2ff1e9e66c6fc449295a888dc400f857863d7df33a386b4b1e1cf42f6a3f4" Mar 13 14:33:23 crc kubenswrapper[4907]: I0313 14:33:23.427217 4907 scope.go:117] "RemoveContainer" containerID="2ad298467d04021b9b1a343e12a26a79c8a6605114bfe6ee4b9cfd24470e49c4" Mar 13 14:33:23 crc kubenswrapper[4907]: I0313 14:33:23.468019 4907 scope.go:117] "RemoveContainer" containerID="71b8e65c6bc18f48a724ed12b9dcf478104abf3f0beba637548e1962f43f12bb" Mar 13 14:33:23 crc kubenswrapper[4907]: I0313 14:33:23.485560 4907 scope.go:117] "RemoveContainer" containerID="766d1e6536dea7ad49da9b30346790d8d4eb902b9e7e7b3c7bf460cfd415e154" Mar 13 14:33:23 crc kubenswrapper[4907]: I0313 14:33:23.506910 4907 scope.go:117] "RemoveContainer" containerID="e3a0771ba1e8115d911c740ed70ee5fca37f11ff18ec41bfab54c46258690b83" Mar 13 14:33:27 crc kubenswrapper[4907]: I0313 14:33:27.782215 4907 scope.go:117] "RemoveContainer" containerID="a09a3bec467b7e275495edb5827150af756bcc256911e7e06cbae5765922b5e2" Mar 13 14:33:27 crc kubenswrapper[4907]: E0313 14:33:27.782865 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:33:39 crc kubenswrapper[4907]: I0313 14:33:39.782423 4907 scope.go:117] "RemoveContainer" containerID="a09a3bec467b7e275495edb5827150af756bcc256911e7e06cbae5765922b5e2" Mar 13 14:33:39 crc kubenswrapper[4907]: E0313 14:33:39.783000 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:33:51 crc kubenswrapper[4907]: I0313 14:33:51.787385 4907 scope.go:117] "RemoveContainer" containerID="a09a3bec467b7e275495edb5827150af756bcc256911e7e06cbae5765922b5e2" Mar 13 14:33:51 crc kubenswrapper[4907]: E0313 14:33:51.788234 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:34:00 crc kubenswrapper[4907]: I0313 14:34:00.139806 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556874-chrll"] Mar 13 14:34:00 crc kubenswrapper[4907]: E0313 14:34:00.140720 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf79c546-2d9a-4eba-bf86-f78d196bc557" containerName="extract-utilities" Mar 13 14:34:00 crc kubenswrapper[4907]: I0313 14:34:00.140735 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf79c546-2d9a-4eba-bf86-f78d196bc557" containerName="extract-utilities" Mar 13 14:34:00 crc kubenswrapper[4907]: E0313 14:34:00.140752 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf79c546-2d9a-4eba-bf86-f78d196bc557" containerName="extract-content" Mar 13 14:34:00 crc kubenswrapper[4907]: I0313 14:34:00.140758 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf79c546-2d9a-4eba-bf86-f78d196bc557" containerName="extract-content" Mar 13 14:34:00 crc kubenswrapper[4907]: E0313 14:34:00.140786 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf79c546-2d9a-4eba-bf86-f78d196bc557" containerName="registry-server" Mar 13 14:34:00 crc kubenswrapper[4907]: I0313 14:34:00.140793 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf79c546-2d9a-4eba-bf86-f78d196bc557" containerName="registry-server" Mar 13 14:34:00 crc kubenswrapper[4907]: I0313 14:34:00.140997 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf79c546-2d9a-4eba-bf86-f78d196bc557" containerName="registry-server" Mar 13 14:34:00 crc kubenswrapper[4907]: I0313 14:34:00.141551 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556874-chrll" Mar 13 14:34:00 crc kubenswrapper[4907]: I0313 14:34:00.143956 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 14:34:00 crc kubenswrapper[4907]: I0313 14:34:00.143970 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 14:34:00 crc kubenswrapper[4907]: I0313 14:34:00.143975 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 14:34:00 crc kubenswrapper[4907]: I0313 14:34:00.150111 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556874-chrll"] Mar 13 14:34:00 crc kubenswrapper[4907]: I0313 14:34:00.329457 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gdsq6\" (UniqueName: \"kubernetes.io/projected/bb80c3ca-ddb4-49fe-8b24-5dc43269ab1e-kube-api-access-gdsq6\") pod \"auto-csr-approver-29556874-chrll\" (UID: \"bb80c3ca-ddb4-49fe-8b24-5dc43269ab1e\") " pod="openshift-infra/auto-csr-approver-29556874-chrll" Mar 13 14:34:00 crc kubenswrapper[4907]: I0313 14:34:00.430459 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gdsq6\" (UniqueName: \"kubernetes.io/projected/bb80c3ca-ddb4-49fe-8b24-5dc43269ab1e-kube-api-access-gdsq6\") pod \"auto-csr-approver-29556874-chrll\" (UID: \"bb80c3ca-ddb4-49fe-8b24-5dc43269ab1e\") " pod="openshift-infra/auto-csr-approver-29556874-chrll" Mar 13 14:34:00 crc kubenswrapper[4907]: I0313 14:34:00.450959 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gdsq6\" (UniqueName: \"kubernetes.io/projected/bb80c3ca-ddb4-49fe-8b24-5dc43269ab1e-kube-api-access-gdsq6\") pod \"auto-csr-approver-29556874-chrll\" (UID: \"bb80c3ca-ddb4-49fe-8b24-5dc43269ab1e\") " pod="openshift-infra/auto-csr-approver-29556874-chrll" Mar 13 14:34:00 crc kubenswrapper[4907]: I0313 14:34:00.477136 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556874-chrll" Mar 13 14:34:00 crc kubenswrapper[4907]: I0313 14:34:00.917631 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556874-chrll"] Mar 13 14:34:00 crc kubenswrapper[4907]: I0313 14:34:00.967153 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556874-chrll" event={"ID":"bb80c3ca-ddb4-49fe-8b24-5dc43269ab1e","Type":"ContainerStarted","Data":"ed8867f760fa43ecca52e81252b98798addeb34b9728f71dd81a483c15c6d605"} Mar 13 14:34:03 crc kubenswrapper[4907]: I0313 14:34:03.994989 4907 generic.go:334] "Generic (PLEG): container finished" podID="bb80c3ca-ddb4-49fe-8b24-5dc43269ab1e" containerID="6a420a24c23cba094e67bdf8ae1b0ab87504764d74798dee0bc7e8921659b3b8" exitCode=0 Mar 13 14:34:03 crc kubenswrapper[4907]: I0313 14:34:03.995363 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556874-chrll" event={"ID":"bb80c3ca-ddb4-49fe-8b24-5dc43269ab1e","Type":"ContainerDied","Data":"6a420a24c23cba094e67bdf8ae1b0ab87504764d74798dee0bc7e8921659b3b8"} Mar 13 14:34:05 crc kubenswrapper[4907]: I0313 14:34:05.293452 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556874-chrll" Mar 13 14:34:05 crc kubenswrapper[4907]: I0313 14:34:05.398416 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gdsq6\" (UniqueName: \"kubernetes.io/projected/bb80c3ca-ddb4-49fe-8b24-5dc43269ab1e-kube-api-access-gdsq6\") pod \"bb80c3ca-ddb4-49fe-8b24-5dc43269ab1e\" (UID: \"bb80c3ca-ddb4-49fe-8b24-5dc43269ab1e\") " Mar 13 14:34:05 crc kubenswrapper[4907]: I0313 14:34:05.403111 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb80c3ca-ddb4-49fe-8b24-5dc43269ab1e-kube-api-access-gdsq6" (OuterVolumeSpecName: "kube-api-access-gdsq6") pod "bb80c3ca-ddb4-49fe-8b24-5dc43269ab1e" (UID: "bb80c3ca-ddb4-49fe-8b24-5dc43269ab1e"). InnerVolumeSpecName "kube-api-access-gdsq6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:34:05 crc kubenswrapper[4907]: I0313 14:34:05.499585 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gdsq6\" (UniqueName: \"kubernetes.io/projected/bb80c3ca-ddb4-49fe-8b24-5dc43269ab1e-kube-api-access-gdsq6\") on node \"crc\" DevicePath \"\"" Mar 13 14:34:06 crc kubenswrapper[4907]: I0313 14:34:06.007953 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556874-chrll" event={"ID":"bb80c3ca-ddb4-49fe-8b24-5dc43269ab1e","Type":"ContainerDied","Data":"ed8867f760fa43ecca52e81252b98798addeb34b9728f71dd81a483c15c6d605"} Mar 13 14:34:06 crc kubenswrapper[4907]: I0313 14:34:06.008002 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ed8867f760fa43ecca52e81252b98798addeb34b9728f71dd81a483c15c6d605" Mar 13 14:34:06 crc kubenswrapper[4907]: I0313 14:34:06.008019 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556874-chrll" Mar 13 14:34:06 crc kubenswrapper[4907]: I0313 14:34:06.351793 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556868-pzml6"] Mar 13 14:34:06 crc kubenswrapper[4907]: I0313 14:34:06.357363 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556868-pzml6"] Mar 13 14:34:06 crc kubenswrapper[4907]: I0313 14:34:06.782300 4907 scope.go:117] "RemoveContainer" containerID="a09a3bec467b7e275495edb5827150af756bcc256911e7e06cbae5765922b5e2" Mar 13 14:34:06 crc kubenswrapper[4907]: E0313 14:34:06.782990 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:34:07 crc kubenswrapper[4907]: I0313 14:34:07.791440 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6301088c-1715-4d57-a9b8-1ea9f7128560" path="/var/lib/kubelet/pods/6301088c-1715-4d57-a9b8-1ea9f7128560/volumes" Mar 13 14:34:17 crc kubenswrapper[4907]: I0313 14:34:17.782471 4907 scope.go:117] "RemoveContainer" containerID="a09a3bec467b7e275495edb5827150af756bcc256911e7e06cbae5765922b5e2" Mar 13 14:34:17 crc kubenswrapper[4907]: E0313 14:34:17.783253 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:34:23 crc kubenswrapper[4907]: I0313 14:34:23.760748 4907 scope.go:117] "RemoveContainer" containerID="57e7ca150948c3586cf56019f626e80fe0c0e49f021cce99a8bbf1bcfa081624" Mar 13 14:34:23 crc kubenswrapper[4907]: I0313 14:34:23.781016 4907 scope.go:117] "RemoveContainer" containerID="2ef6cd9eb88308a93f1d78fb9954045701e6f121a3c20317549b8db40b280bd4" Mar 13 14:34:23 crc kubenswrapper[4907]: I0313 14:34:23.804295 4907 scope.go:117] "RemoveContainer" containerID="3f5c4dcc986bdd72e19fe37fafad59be240102706caf33d75e722b2f347bd3e3" Mar 13 14:34:23 crc kubenswrapper[4907]: I0313 14:34:23.837099 4907 scope.go:117] "RemoveContainer" containerID="bb7f9aa2bb828fff3a17bc00a23c15ddb0d4b9b7ae4d6e61a19fc6386d089dbb" Mar 13 14:34:23 crc kubenswrapper[4907]: I0313 14:34:23.853817 4907 scope.go:117] "RemoveContainer" containerID="6ad9081a27708afd67b77cedc2aa19d22a6afb40f65f24b3a9da7ca3a519bf38" Mar 13 14:34:23 crc kubenswrapper[4907]: I0313 14:34:23.869093 4907 scope.go:117] "RemoveContainer" containerID="5feda018c451c4b446f95e35e5721969c1c913a7887bdf916e5ea2193df476ec" Mar 13 14:34:23 crc kubenswrapper[4907]: I0313 14:34:23.909344 4907 scope.go:117] "RemoveContainer" containerID="2cbce92733c8c36aab8d17f31bd5085e394cb7fec21de278bf9301bd315d6bae" Mar 13 14:34:23 crc kubenswrapper[4907]: I0313 14:34:23.926025 4907 scope.go:117] "RemoveContainer" containerID="52f2356a35bf1ec5bd881f141be785ea7a9d11be6b6e8da92a08cbad79935d8d" Mar 13 14:34:23 crc kubenswrapper[4907]: I0313 14:34:23.979008 4907 scope.go:117] "RemoveContainer" containerID="9fe21d916b51d235fb7ba28d7a6934f6f8ffaa95ccce23ab15efee2b6e69acd5" Mar 13 14:34:23 crc kubenswrapper[4907]: I0313 14:34:23.997081 4907 scope.go:117] "RemoveContainer" containerID="a07507237e2303a227977bfc1acdae544607145ad994269ea3936668b9e9c650" Mar 13 14:34:24 crc kubenswrapper[4907]: I0313 14:34:24.011121 4907 scope.go:117] "RemoveContainer" containerID="598b6fbc5376df701a762ba6746b37137dc421681ebf472e4298098628662035" Mar 13 14:34:24 crc kubenswrapper[4907]: I0313 14:34:24.026644 4907 scope.go:117] "RemoveContainer" containerID="8ce806c1853f002c1ca5aee8d1ea69c43b2c38c4c666d7df49a773673406c573" Mar 13 14:34:24 crc kubenswrapper[4907]: I0313 14:34:24.068673 4907 scope.go:117] "RemoveContainer" containerID="107f7252eaea9f43079c72f27a3a08aa9bdfea92d244d423f8695cb45030deb6" Mar 13 14:34:24 crc kubenswrapper[4907]: I0313 14:34:24.084365 4907 scope.go:117] "RemoveContainer" containerID="ed14d79ba88da2a61d40c8ed633b8377f32a153aac0ee43765b80eb04fb89188" Mar 13 14:34:24 crc kubenswrapper[4907]: I0313 14:34:24.103684 4907 scope.go:117] "RemoveContainer" containerID="13e164891acf22f2bb718fb70e94321cdc023003038bdb8dbc28a6f2b84707b6" Mar 13 14:34:24 crc kubenswrapper[4907]: I0313 14:34:24.122629 4907 scope.go:117] "RemoveContainer" containerID="e156e90b4d1cbea6f3a55f14df88e09d8aeafa1d0449cf8de460be879c418cf0" Mar 13 14:34:28 crc kubenswrapper[4907]: I0313 14:34:28.783542 4907 scope.go:117] "RemoveContainer" containerID="a09a3bec467b7e275495edb5827150af756bcc256911e7e06cbae5765922b5e2" Mar 13 14:34:28 crc kubenswrapper[4907]: E0313 14:34:28.784494 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:34:40 crc kubenswrapper[4907]: I0313 14:34:40.782646 4907 scope.go:117] "RemoveContainer" containerID="a09a3bec467b7e275495edb5827150af756bcc256911e7e06cbae5765922b5e2" Mar 13 14:34:40 crc kubenswrapper[4907]: E0313 14:34:40.784334 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:34:55 crc kubenswrapper[4907]: I0313 14:34:55.783336 4907 scope.go:117] "RemoveContainer" containerID="a09a3bec467b7e275495edb5827150af756bcc256911e7e06cbae5765922b5e2" Mar 13 14:34:55 crc kubenswrapper[4907]: E0313 14:34:55.784936 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:35:10 crc kubenswrapper[4907]: I0313 14:35:10.783220 4907 scope.go:117] "RemoveContainer" containerID="a09a3bec467b7e275495edb5827150af756bcc256911e7e06cbae5765922b5e2" Mar 13 14:35:10 crc kubenswrapper[4907]: E0313 14:35:10.784430 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:35:21 crc kubenswrapper[4907]: I0313 14:35:21.782649 4907 scope.go:117] "RemoveContainer" containerID="a09a3bec467b7e275495edb5827150af756bcc256911e7e06cbae5765922b5e2" Mar 13 14:35:21 crc kubenswrapper[4907]: E0313 14:35:21.784381 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:35:24 crc kubenswrapper[4907]: I0313 14:35:24.376614 4907 scope.go:117] "RemoveContainer" containerID="7dd9875251f25eba8556dcd6141b77fd52b2025f417f75c1f8e455693aab178b" Mar 13 14:35:24 crc kubenswrapper[4907]: I0313 14:35:24.397176 4907 scope.go:117] "RemoveContainer" containerID="dd9f647265d73fb65018c8a51bfa4b5a5aa86cde573e29956aaf3479042709d4" Mar 13 14:35:24 crc kubenswrapper[4907]: I0313 14:35:24.441004 4907 scope.go:117] "RemoveContainer" containerID="d935db5c32cf1af5583418c37d51248a6d17b0514af1c3fa70469ca109747682" Mar 13 14:35:24 crc kubenswrapper[4907]: I0313 14:35:24.477063 4907 scope.go:117] "RemoveContainer" containerID="28a2d9fc7d22dcae7fbbc67bc1cf31d4d10a281e4e795b4b2a25e7a449ecff2c" Mar 13 14:35:34 crc kubenswrapper[4907]: I0313 14:35:34.783534 4907 scope.go:117] "RemoveContainer" containerID="a09a3bec467b7e275495edb5827150af756bcc256911e7e06cbae5765922b5e2" Mar 13 14:35:34 crc kubenswrapper[4907]: E0313 14:35:34.784770 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:35:49 crc kubenswrapper[4907]: I0313 14:35:49.782667 4907 scope.go:117] "RemoveContainer" containerID="a09a3bec467b7e275495edb5827150af756bcc256911e7e06cbae5765922b5e2" Mar 13 14:35:49 crc kubenswrapper[4907]: E0313 14:35:49.784784 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:36:00 crc kubenswrapper[4907]: I0313 14:36:00.150533 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556876-dwnhg"] Mar 13 14:36:00 crc kubenswrapper[4907]: E0313 14:36:00.151505 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb80c3ca-ddb4-49fe-8b24-5dc43269ab1e" containerName="oc" Mar 13 14:36:00 crc kubenswrapper[4907]: I0313 14:36:00.151523 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb80c3ca-ddb4-49fe-8b24-5dc43269ab1e" containerName="oc" Mar 13 14:36:00 crc kubenswrapper[4907]: I0313 14:36:00.151698 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb80c3ca-ddb4-49fe-8b24-5dc43269ab1e" containerName="oc" Mar 13 14:36:00 crc kubenswrapper[4907]: I0313 14:36:00.152328 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556876-dwnhg" Mar 13 14:36:00 crc kubenswrapper[4907]: I0313 14:36:00.156119 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 14:36:00 crc kubenswrapper[4907]: I0313 14:36:00.156492 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 14:36:00 crc kubenswrapper[4907]: I0313 14:36:00.157705 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556876-dwnhg"] Mar 13 14:36:00 crc kubenswrapper[4907]: I0313 14:36:00.161481 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 14:36:00 crc kubenswrapper[4907]: I0313 14:36:00.216567 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-65b49\" (UniqueName: \"kubernetes.io/projected/d32570e6-4671-4b95-b312-f4af6c4a1478-kube-api-access-65b49\") pod \"auto-csr-approver-29556876-dwnhg\" (UID: \"d32570e6-4671-4b95-b312-f4af6c4a1478\") " pod="openshift-infra/auto-csr-approver-29556876-dwnhg" Mar 13 14:36:00 crc kubenswrapper[4907]: I0313 14:36:00.318189 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-65b49\" (UniqueName: \"kubernetes.io/projected/d32570e6-4671-4b95-b312-f4af6c4a1478-kube-api-access-65b49\") pod \"auto-csr-approver-29556876-dwnhg\" (UID: \"d32570e6-4671-4b95-b312-f4af6c4a1478\") " pod="openshift-infra/auto-csr-approver-29556876-dwnhg" Mar 13 14:36:00 crc kubenswrapper[4907]: I0313 14:36:00.345110 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-65b49\" (UniqueName: \"kubernetes.io/projected/d32570e6-4671-4b95-b312-f4af6c4a1478-kube-api-access-65b49\") pod \"auto-csr-approver-29556876-dwnhg\" (UID: \"d32570e6-4671-4b95-b312-f4af6c4a1478\") " pod="openshift-infra/auto-csr-approver-29556876-dwnhg" Mar 13 14:36:00 crc kubenswrapper[4907]: I0313 14:36:00.469525 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556876-dwnhg" Mar 13 14:36:00 crc kubenswrapper[4907]: I0313 14:36:00.943867 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556876-dwnhg"] Mar 13 14:36:01 crc kubenswrapper[4907]: I0313 14:36:01.786607 4907 scope.go:117] "RemoveContainer" containerID="a09a3bec467b7e275495edb5827150af756bcc256911e7e06cbae5765922b5e2" Mar 13 14:36:01 crc kubenswrapper[4907]: E0313 14:36:01.787195 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:36:01 crc kubenswrapper[4907]: I0313 14:36:01.953568 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556876-dwnhg" event={"ID":"d32570e6-4671-4b95-b312-f4af6c4a1478","Type":"ContainerStarted","Data":"46040338384cd3899901e9d87c65468c58bf61984fc713bee92d18141877dbab"} Mar 13 14:36:02 crc kubenswrapper[4907]: I0313 14:36:02.962702 4907 generic.go:334] "Generic (PLEG): container finished" podID="d32570e6-4671-4b95-b312-f4af6c4a1478" containerID="ec22972508d423bdab0d6a15e5b2be7962b2967f440897919fe9018a840dad0f" exitCode=0 Mar 13 14:36:02 crc kubenswrapper[4907]: I0313 14:36:02.963027 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556876-dwnhg" event={"ID":"d32570e6-4671-4b95-b312-f4af6c4a1478","Type":"ContainerDied","Data":"ec22972508d423bdab0d6a15e5b2be7962b2967f440897919fe9018a840dad0f"} Mar 13 14:36:04 crc kubenswrapper[4907]: I0313 14:36:04.399821 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556876-dwnhg" Mar 13 14:36:04 crc kubenswrapper[4907]: I0313 14:36:04.573527 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-65b49\" (UniqueName: \"kubernetes.io/projected/d32570e6-4671-4b95-b312-f4af6c4a1478-kube-api-access-65b49\") pod \"d32570e6-4671-4b95-b312-f4af6c4a1478\" (UID: \"d32570e6-4671-4b95-b312-f4af6c4a1478\") " Mar 13 14:36:04 crc kubenswrapper[4907]: I0313 14:36:04.578943 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d32570e6-4671-4b95-b312-f4af6c4a1478-kube-api-access-65b49" (OuterVolumeSpecName: "kube-api-access-65b49") pod "d32570e6-4671-4b95-b312-f4af6c4a1478" (UID: "d32570e6-4671-4b95-b312-f4af6c4a1478"). InnerVolumeSpecName "kube-api-access-65b49". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:36:04 crc kubenswrapper[4907]: I0313 14:36:04.675173 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-65b49\" (UniqueName: \"kubernetes.io/projected/d32570e6-4671-4b95-b312-f4af6c4a1478-kube-api-access-65b49\") on node \"crc\" DevicePath \"\"" Mar 13 14:36:04 crc kubenswrapper[4907]: I0313 14:36:04.986928 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556876-dwnhg" event={"ID":"d32570e6-4671-4b95-b312-f4af6c4a1478","Type":"ContainerDied","Data":"46040338384cd3899901e9d87c65468c58bf61984fc713bee92d18141877dbab"} Mar 13 14:36:04 crc kubenswrapper[4907]: I0313 14:36:04.987000 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="46040338384cd3899901e9d87c65468c58bf61984fc713bee92d18141877dbab" Mar 13 14:36:04 crc kubenswrapper[4907]: I0313 14:36:04.986953 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556876-dwnhg" Mar 13 14:36:05 crc kubenswrapper[4907]: I0313 14:36:05.461419 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556870-8mlkz"] Mar 13 14:36:05 crc kubenswrapper[4907]: I0313 14:36:05.466949 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556870-8mlkz"] Mar 13 14:36:05 crc kubenswrapper[4907]: I0313 14:36:05.793568 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7993f31-1072-4c83-a3ac-16dca5eb04a3" path="/var/lib/kubelet/pods/e7993f31-1072-4c83-a3ac-16dca5eb04a3/volumes" Mar 13 14:36:16 crc kubenswrapper[4907]: I0313 14:36:16.783478 4907 scope.go:117] "RemoveContainer" containerID="a09a3bec467b7e275495edb5827150af756bcc256911e7e06cbae5765922b5e2" Mar 13 14:36:16 crc kubenswrapper[4907]: E0313 14:36:16.784823 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:36:24 crc kubenswrapper[4907]: I0313 14:36:24.558450 4907 scope.go:117] "RemoveContainer" containerID="714551bd9afe88fbb37b8d1ba1d86847559c10eb4afc2d8c8cade9bb4be375ee" Mar 13 14:36:24 crc kubenswrapper[4907]: I0313 14:36:24.612190 4907 scope.go:117] "RemoveContainer" containerID="93b983c69ed0cdc38b2c77a7998230bc3acd8a611b28bc2ce28fc6ae6eacb20d" Mar 13 14:36:29 crc kubenswrapper[4907]: I0313 14:36:29.782465 4907 scope.go:117] "RemoveContainer" containerID="a09a3bec467b7e275495edb5827150af756bcc256911e7e06cbae5765922b5e2" Mar 13 14:36:29 crc kubenswrapper[4907]: E0313 14:36:29.783317 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:36:43 crc kubenswrapper[4907]: I0313 14:36:43.782839 4907 scope.go:117] "RemoveContainer" containerID="a09a3bec467b7e275495edb5827150af756bcc256911e7e06cbae5765922b5e2" Mar 13 14:36:43 crc kubenswrapper[4907]: E0313 14:36:43.783579 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:36:56 crc kubenswrapper[4907]: I0313 14:36:56.782532 4907 scope.go:117] "RemoveContainer" containerID="a09a3bec467b7e275495edb5827150af756bcc256911e7e06cbae5765922b5e2" Mar 13 14:36:56 crc kubenswrapper[4907]: E0313 14:36:56.783491 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:37:10 crc kubenswrapper[4907]: I0313 14:37:10.782439 4907 scope.go:117] "RemoveContainer" containerID="a09a3bec467b7e275495edb5827150af756bcc256911e7e06cbae5765922b5e2" Mar 13 14:37:10 crc kubenswrapper[4907]: E0313 14:37:10.783294 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:37:25 crc kubenswrapper[4907]: I0313 14:37:25.783258 4907 scope.go:117] "RemoveContainer" containerID="a09a3bec467b7e275495edb5827150af756bcc256911e7e06cbae5765922b5e2" Mar 13 14:37:26 crc kubenswrapper[4907]: I0313 14:37:26.676273 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerStarted","Data":"3c1970952dcd249074d61677266bef44d3d78378f555ffc20b89b00a033811dd"} Mar 13 14:38:00 crc kubenswrapper[4907]: I0313 14:38:00.150870 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556878-fqsf6"] Mar 13 14:38:00 crc kubenswrapper[4907]: E0313 14:38:00.151941 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d32570e6-4671-4b95-b312-f4af6c4a1478" containerName="oc" Mar 13 14:38:00 crc kubenswrapper[4907]: I0313 14:38:00.151964 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="d32570e6-4671-4b95-b312-f4af6c4a1478" containerName="oc" Mar 13 14:38:00 crc kubenswrapper[4907]: I0313 14:38:00.152201 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="d32570e6-4671-4b95-b312-f4af6c4a1478" containerName="oc" Mar 13 14:38:00 crc kubenswrapper[4907]: I0313 14:38:00.152764 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556878-fqsf6" Mar 13 14:38:00 crc kubenswrapper[4907]: I0313 14:38:00.157471 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 14:38:00 crc kubenswrapper[4907]: I0313 14:38:00.157676 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 14:38:00 crc kubenswrapper[4907]: I0313 14:38:00.158201 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 14:38:00 crc kubenswrapper[4907]: I0313 14:38:00.163001 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556878-fqsf6"] Mar 13 14:38:00 crc kubenswrapper[4907]: I0313 14:38:00.206799 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nkf9c\" (UniqueName: \"kubernetes.io/projected/ec5b140c-a271-4a8e-a9af-8e7c5bddcad3-kube-api-access-nkf9c\") pod \"auto-csr-approver-29556878-fqsf6\" (UID: \"ec5b140c-a271-4a8e-a9af-8e7c5bddcad3\") " pod="openshift-infra/auto-csr-approver-29556878-fqsf6" Mar 13 14:38:00 crc kubenswrapper[4907]: I0313 14:38:00.307672 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nkf9c\" (UniqueName: \"kubernetes.io/projected/ec5b140c-a271-4a8e-a9af-8e7c5bddcad3-kube-api-access-nkf9c\") pod \"auto-csr-approver-29556878-fqsf6\" (UID: \"ec5b140c-a271-4a8e-a9af-8e7c5bddcad3\") " pod="openshift-infra/auto-csr-approver-29556878-fqsf6" Mar 13 14:38:00 crc kubenswrapper[4907]: I0313 14:38:00.327282 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nkf9c\" (UniqueName: \"kubernetes.io/projected/ec5b140c-a271-4a8e-a9af-8e7c5bddcad3-kube-api-access-nkf9c\") pod \"auto-csr-approver-29556878-fqsf6\" (UID: \"ec5b140c-a271-4a8e-a9af-8e7c5bddcad3\") " pod="openshift-infra/auto-csr-approver-29556878-fqsf6" Mar 13 14:38:00 crc kubenswrapper[4907]: I0313 14:38:00.478100 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556878-fqsf6" Mar 13 14:38:00 crc kubenswrapper[4907]: I0313 14:38:00.763652 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556878-fqsf6"] Mar 13 14:38:00 crc kubenswrapper[4907]: I0313 14:38:00.765951 4907 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 13 14:38:00 crc kubenswrapper[4907]: I0313 14:38:00.943167 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556878-fqsf6" event={"ID":"ec5b140c-a271-4a8e-a9af-8e7c5bddcad3","Type":"ContainerStarted","Data":"efa9657ca20cfe1e7c6d4d50a0574daaf6c8bf76cb2e8a66510138fcb99666c3"} Mar 13 14:38:02 crc kubenswrapper[4907]: I0313 14:38:02.963779 4907 generic.go:334] "Generic (PLEG): container finished" podID="ec5b140c-a271-4a8e-a9af-8e7c5bddcad3" containerID="3d950127b4213b5d0d64f24ed2cfdf8b4f35b40aa5e7072728ac45679d7fc695" exitCode=0 Mar 13 14:38:02 crc kubenswrapper[4907]: I0313 14:38:02.963838 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556878-fqsf6" event={"ID":"ec5b140c-a271-4a8e-a9af-8e7c5bddcad3","Type":"ContainerDied","Data":"3d950127b4213b5d0d64f24ed2cfdf8b4f35b40aa5e7072728ac45679d7fc695"} Mar 13 14:38:04 crc kubenswrapper[4907]: I0313 14:38:04.288496 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556878-fqsf6" Mar 13 14:38:04 crc kubenswrapper[4907]: I0313 14:38:04.369705 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nkf9c\" (UniqueName: \"kubernetes.io/projected/ec5b140c-a271-4a8e-a9af-8e7c5bddcad3-kube-api-access-nkf9c\") pod \"ec5b140c-a271-4a8e-a9af-8e7c5bddcad3\" (UID: \"ec5b140c-a271-4a8e-a9af-8e7c5bddcad3\") " Mar 13 14:38:04 crc kubenswrapper[4907]: I0313 14:38:04.378265 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ec5b140c-a271-4a8e-a9af-8e7c5bddcad3-kube-api-access-nkf9c" (OuterVolumeSpecName: "kube-api-access-nkf9c") pod "ec5b140c-a271-4a8e-a9af-8e7c5bddcad3" (UID: "ec5b140c-a271-4a8e-a9af-8e7c5bddcad3"). InnerVolumeSpecName "kube-api-access-nkf9c". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:38:04 crc kubenswrapper[4907]: I0313 14:38:04.471433 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nkf9c\" (UniqueName: \"kubernetes.io/projected/ec5b140c-a271-4a8e-a9af-8e7c5bddcad3-kube-api-access-nkf9c\") on node \"crc\" DevicePath \"\"" Mar 13 14:38:04 crc kubenswrapper[4907]: I0313 14:38:04.980944 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556878-fqsf6" event={"ID":"ec5b140c-a271-4a8e-a9af-8e7c5bddcad3","Type":"ContainerDied","Data":"efa9657ca20cfe1e7c6d4d50a0574daaf6c8bf76cb2e8a66510138fcb99666c3"} Mar 13 14:38:04 crc kubenswrapper[4907]: I0313 14:38:04.981123 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="efa9657ca20cfe1e7c6d4d50a0574daaf6c8bf76cb2e8a66510138fcb99666c3" Mar 13 14:38:04 crc kubenswrapper[4907]: I0313 14:38:04.981179 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556878-fqsf6" Mar 13 14:38:05 crc kubenswrapper[4907]: I0313 14:38:05.345971 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556872-srgh9"] Mar 13 14:38:05 crc kubenswrapper[4907]: I0313 14:38:05.351766 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556872-srgh9"] Mar 13 14:38:05 crc kubenswrapper[4907]: I0313 14:38:05.799002 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="32297c61-6461-4357-832f-9de703d88000" path="/var/lib/kubelet/pods/32297c61-6461-4357-832f-9de703d88000/volumes" Mar 13 14:38:24 crc kubenswrapper[4907]: I0313 14:38:24.725010 4907 scope.go:117] "RemoveContainer" containerID="8eea35f2339f97a9cd81df041bdf950f47b0d8ec3a69dea570886bdde6ad88a7" Mar 13 14:39:48 crc kubenswrapper[4907]: I0313 14:39:48.042016 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 14:39:48 crc kubenswrapper[4907]: I0313 14:39:48.042652 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 14:39:48 crc kubenswrapper[4907]: I0313 14:39:48.432216 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-mqhct"] Mar 13 14:39:48 crc kubenswrapper[4907]: E0313 14:39:48.433088 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec5b140c-a271-4a8e-a9af-8e7c5bddcad3" containerName="oc" Mar 13 14:39:48 crc kubenswrapper[4907]: I0313 14:39:48.433329 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec5b140c-a271-4a8e-a9af-8e7c5bddcad3" containerName="oc" Mar 13 14:39:48 crc kubenswrapper[4907]: I0313 14:39:48.433867 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="ec5b140c-a271-4a8e-a9af-8e7c5bddcad3" containerName="oc" Mar 13 14:39:48 crc kubenswrapper[4907]: I0313 14:39:48.436327 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mqhct" Mar 13 14:39:48 crc kubenswrapper[4907]: I0313 14:39:48.442009 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mqhct"] Mar 13 14:39:48 crc kubenswrapper[4907]: I0313 14:39:48.572215 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/631842f8-7a11-4ec0-bab3-5793a949806d-catalog-content\") pod \"redhat-marketplace-mqhct\" (UID: \"631842f8-7a11-4ec0-bab3-5793a949806d\") " pod="openshift-marketplace/redhat-marketplace-mqhct" Mar 13 14:39:48 crc kubenswrapper[4907]: I0313 14:39:48.572344 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9ffj9\" (UniqueName: \"kubernetes.io/projected/631842f8-7a11-4ec0-bab3-5793a949806d-kube-api-access-9ffj9\") pod \"redhat-marketplace-mqhct\" (UID: \"631842f8-7a11-4ec0-bab3-5793a949806d\") " pod="openshift-marketplace/redhat-marketplace-mqhct" Mar 13 14:39:48 crc kubenswrapper[4907]: I0313 14:39:48.572394 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/631842f8-7a11-4ec0-bab3-5793a949806d-utilities\") pod \"redhat-marketplace-mqhct\" (UID: \"631842f8-7a11-4ec0-bab3-5793a949806d\") " pod="openshift-marketplace/redhat-marketplace-mqhct" Mar 13 14:39:48 crc kubenswrapper[4907]: I0313 14:39:48.674244 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/631842f8-7a11-4ec0-bab3-5793a949806d-catalog-content\") pod \"redhat-marketplace-mqhct\" (UID: \"631842f8-7a11-4ec0-bab3-5793a949806d\") " pod="openshift-marketplace/redhat-marketplace-mqhct" Mar 13 14:39:48 crc kubenswrapper[4907]: I0313 14:39:48.674757 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/631842f8-7a11-4ec0-bab3-5793a949806d-catalog-content\") pod \"redhat-marketplace-mqhct\" (UID: \"631842f8-7a11-4ec0-bab3-5793a949806d\") " pod="openshift-marketplace/redhat-marketplace-mqhct" Mar 13 14:39:48 crc kubenswrapper[4907]: I0313 14:39:48.675055 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9ffj9\" (UniqueName: \"kubernetes.io/projected/631842f8-7a11-4ec0-bab3-5793a949806d-kube-api-access-9ffj9\") pod \"redhat-marketplace-mqhct\" (UID: \"631842f8-7a11-4ec0-bab3-5793a949806d\") " pod="openshift-marketplace/redhat-marketplace-mqhct" Mar 13 14:39:48 crc kubenswrapper[4907]: I0313 14:39:48.675174 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/631842f8-7a11-4ec0-bab3-5793a949806d-utilities\") pod \"redhat-marketplace-mqhct\" (UID: \"631842f8-7a11-4ec0-bab3-5793a949806d\") " pod="openshift-marketplace/redhat-marketplace-mqhct" Mar 13 14:39:48 crc kubenswrapper[4907]: I0313 14:39:48.675820 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/631842f8-7a11-4ec0-bab3-5793a949806d-utilities\") pod \"redhat-marketplace-mqhct\" (UID: \"631842f8-7a11-4ec0-bab3-5793a949806d\") " pod="openshift-marketplace/redhat-marketplace-mqhct" Mar 13 14:39:48 crc kubenswrapper[4907]: I0313 14:39:48.713137 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9ffj9\" (UniqueName: \"kubernetes.io/projected/631842f8-7a11-4ec0-bab3-5793a949806d-kube-api-access-9ffj9\") pod \"redhat-marketplace-mqhct\" (UID: \"631842f8-7a11-4ec0-bab3-5793a949806d\") " pod="openshift-marketplace/redhat-marketplace-mqhct" Mar 13 14:39:48 crc kubenswrapper[4907]: I0313 14:39:48.782120 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mqhct" Mar 13 14:39:49 crc kubenswrapper[4907]: I0313 14:39:49.196744 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mqhct"] Mar 13 14:39:49 crc kubenswrapper[4907]: I0313 14:39:49.922780 4907 generic.go:334] "Generic (PLEG): container finished" podID="631842f8-7a11-4ec0-bab3-5793a949806d" containerID="8b6db5a70191fecc20fe10365c72f6b6153346de6e602aa9933891080440586f" exitCode=0 Mar 13 14:39:49 crc kubenswrapper[4907]: I0313 14:39:49.922841 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mqhct" event={"ID":"631842f8-7a11-4ec0-bab3-5793a949806d","Type":"ContainerDied","Data":"8b6db5a70191fecc20fe10365c72f6b6153346de6e602aa9933891080440586f"} Mar 13 14:39:49 crc kubenswrapper[4907]: I0313 14:39:49.922950 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mqhct" event={"ID":"631842f8-7a11-4ec0-bab3-5793a949806d","Type":"ContainerStarted","Data":"3849a706a8af92bea878d778fade40002b62f87b6f1a978975072a05bbe7b2c8"} Mar 13 14:39:50 crc kubenswrapper[4907]: I0313 14:39:50.932853 4907 generic.go:334] "Generic (PLEG): container finished" podID="631842f8-7a11-4ec0-bab3-5793a949806d" containerID="fe74f9a5411a632ebac06894970db3a7519b223f6ba533c92f921f1566e8a759" exitCode=0 Mar 13 14:39:50 crc kubenswrapper[4907]: I0313 14:39:50.932915 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mqhct" event={"ID":"631842f8-7a11-4ec0-bab3-5793a949806d","Type":"ContainerDied","Data":"fe74f9a5411a632ebac06894970db3a7519b223f6ba533c92f921f1566e8a759"} Mar 13 14:39:52 crc kubenswrapper[4907]: I0313 14:39:52.953253 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mqhct" event={"ID":"631842f8-7a11-4ec0-bab3-5793a949806d","Type":"ContainerStarted","Data":"acb00c615a16c4afc4f4512ed580341d6c225c7726089602e71b32f317ded8e2"} Mar 13 14:39:52 crc kubenswrapper[4907]: I0313 14:39:52.979294 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-mqhct" podStartSLOduration=2.545098287 podStartE2EDuration="4.97927148s" podCreationTimestamp="2026-03-13 14:39:48 +0000 UTC" firstStartedPulling="2026-03-13 14:39:49.925841359 +0000 UTC m=+2088.825629088" lastFinishedPulling="2026-03-13 14:39:52.360014562 +0000 UTC m=+2091.259802281" observedRunningTime="2026-03-13 14:39:52.976401087 +0000 UTC m=+2091.876188816" watchObservedRunningTime="2026-03-13 14:39:52.97927148 +0000 UTC m=+2091.879059179" Mar 13 14:39:58 crc kubenswrapper[4907]: I0313 14:39:58.782313 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-mqhct" Mar 13 14:39:58 crc kubenswrapper[4907]: I0313 14:39:58.782719 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-mqhct" Mar 13 14:39:58 crc kubenswrapper[4907]: I0313 14:39:58.838397 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-mqhct" Mar 13 14:39:59 crc kubenswrapper[4907]: I0313 14:39:59.060928 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-mqhct" Mar 13 14:39:59 crc kubenswrapper[4907]: I0313 14:39:59.116742 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-mqhct"] Mar 13 14:40:00 crc kubenswrapper[4907]: I0313 14:40:00.148674 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556880-wxdzr"] Mar 13 14:40:00 crc kubenswrapper[4907]: I0313 14:40:00.149864 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556880-wxdzr" Mar 13 14:40:00 crc kubenswrapper[4907]: I0313 14:40:00.155835 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 14:40:00 crc kubenswrapper[4907]: I0313 14:40:00.156124 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 14:40:00 crc kubenswrapper[4907]: I0313 14:40:00.156235 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 14:40:00 crc kubenswrapper[4907]: I0313 14:40:00.185781 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556880-wxdzr"] Mar 13 14:40:00 crc kubenswrapper[4907]: I0313 14:40:00.261647 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nj4vk\" (UniqueName: \"kubernetes.io/projected/20dce4c7-7cbf-4e49-bd0e-93ee53d72551-kube-api-access-nj4vk\") pod \"auto-csr-approver-29556880-wxdzr\" (UID: \"20dce4c7-7cbf-4e49-bd0e-93ee53d72551\") " pod="openshift-infra/auto-csr-approver-29556880-wxdzr" Mar 13 14:40:00 crc kubenswrapper[4907]: I0313 14:40:00.362765 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nj4vk\" (UniqueName: \"kubernetes.io/projected/20dce4c7-7cbf-4e49-bd0e-93ee53d72551-kube-api-access-nj4vk\") pod \"auto-csr-approver-29556880-wxdzr\" (UID: \"20dce4c7-7cbf-4e49-bd0e-93ee53d72551\") " pod="openshift-infra/auto-csr-approver-29556880-wxdzr" Mar 13 14:40:00 crc kubenswrapper[4907]: I0313 14:40:00.387777 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nj4vk\" (UniqueName: \"kubernetes.io/projected/20dce4c7-7cbf-4e49-bd0e-93ee53d72551-kube-api-access-nj4vk\") pod \"auto-csr-approver-29556880-wxdzr\" (UID: \"20dce4c7-7cbf-4e49-bd0e-93ee53d72551\") " pod="openshift-infra/auto-csr-approver-29556880-wxdzr" Mar 13 14:40:00 crc kubenswrapper[4907]: I0313 14:40:00.481079 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556880-wxdzr" Mar 13 14:40:00 crc kubenswrapper[4907]: I0313 14:40:00.956459 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556880-wxdzr"] Mar 13 14:40:00 crc kubenswrapper[4907]: W0313 14:40:00.967528 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod20dce4c7_7cbf_4e49_bd0e_93ee53d72551.slice/crio-879e5062915a12a2855140a77d7a884e87419cf083062608de3e6cfa55590d80 WatchSource:0}: Error finding container 879e5062915a12a2855140a77d7a884e87419cf083062608de3e6cfa55590d80: Status 404 returned error can't find the container with id 879e5062915a12a2855140a77d7a884e87419cf083062608de3e6cfa55590d80 Mar 13 14:40:01 crc kubenswrapper[4907]: I0313 14:40:01.019479 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556880-wxdzr" event={"ID":"20dce4c7-7cbf-4e49-bd0e-93ee53d72551","Type":"ContainerStarted","Data":"879e5062915a12a2855140a77d7a884e87419cf083062608de3e6cfa55590d80"} Mar 13 14:40:01 crc kubenswrapper[4907]: I0313 14:40:01.019604 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-mqhct" podUID="631842f8-7a11-4ec0-bab3-5793a949806d" containerName="registry-server" containerID="cri-o://acb00c615a16c4afc4f4512ed580341d6c225c7726089602e71b32f317ded8e2" gracePeriod=2 Mar 13 14:40:01 crc kubenswrapper[4907]: I0313 14:40:01.462560 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mqhct" Mar 13 14:40:01 crc kubenswrapper[4907]: I0313 14:40:01.586836 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/631842f8-7a11-4ec0-bab3-5793a949806d-catalog-content\") pod \"631842f8-7a11-4ec0-bab3-5793a949806d\" (UID: \"631842f8-7a11-4ec0-bab3-5793a949806d\") " Mar 13 14:40:01 crc kubenswrapper[4907]: I0313 14:40:01.587089 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/631842f8-7a11-4ec0-bab3-5793a949806d-utilities\") pod \"631842f8-7a11-4ec0-bab3-5793a949806d\" (UID: \"631842f8-7a11-4ec0-bab3-5793a949806d\") " Mar 13 14:40:01 crc kubenswrapper[4907]: I0313 14:40:01.587290 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9ffj9\" (UniqueName: \"kubernetes.io/projected/631842f8-7a11-4ec0-bab3-5793a949806d-kube-api-access-9ffj9\") pod \"631842f8-7a11-4ec0-bab3-5793a949806d\" (UID: \"631842f8-7a11-4ec0-bab3-5793a949806d\") " Mar 13 14:40:01 crc kubenswrapper[4907]: I0313 14:40:01.588655 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/631842f8-7a11-4ec0-bab3-5793a949806d-utilities" (OuterVolumeSpecName: "utilities") pod "631842f8-7a11-4ec0-bab3-5793a949806d" (UID: "631842f8-7a11-4ec0-bab3-5793a949806d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:40:01 crc kubenswrapper[4907]: I0313 14:40:01.594649 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/631842f8-7a11-4ec0-bab3-5793a949806d-kube-api-access-9ffj9" (OuterVolumeSpecName: "kube-api-access-9ffj9") pod "631842f8-7a11-4ec0-bab3-5793a949806d" (UID: "631842f8-7a11-4ec0-bab3-5793a949806d"). InnerVolumeSpecName "kube-api-access-9ffj9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:40:01 crc kubenswrapper[4907]: I0313 14:40:01.625458 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/631842f8-7a11-4ec0-bab3-5793a949806d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "631842f8-7a11-4ec0-bab3-5793a949806d" (UID: "631842f8-7a11-4ec0-bab3-5793a949806d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:40:01 crc kubenswrapper[4907]: I0313 14:40:01.689608 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9ffj9\" (UniqueName: \"kubernetes.io/projected/631842f8-7a11-4ec0-bab3-5793a949806d-kube-api-access-9ffj9\") on node \"crc\" DevicePath \"\"" Mar 13 14:40:01 crc kubenswrapper[4907]: I0313 14:40:01.689670 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/631842f8-7a11-4ec0-bab3-5793a949806d-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 14:40:01 crc kubenswrapper[4907]: I0313 14:40:01.689690 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/631842f8-7a11-4ec0-bab3-5793a949806d-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 14:40:02 crc kubenswrapper[4907]: I0313 14:40:02.029051 4907 generic.go:334] "Generic (PLEG): container finished" podID="631842f8-7a11-4ec0-bab3-5793a949806d" containerID="acb00c615a16c4afc4f4512ed580341d6c225c7726089602e71b32f317ded8e2" exitCode=0 Mar 13 14:40:02 crc kubenswrapper[4907]: I0313 14:40:02.029109 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mqhct" Mar 13 14:40:02 crc kubenswrapper[4907]: I0313 14:40:02.029127 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mqhct" event={"ID":"631842f8-7a11-4ec0-bab3-5793a949806d","Type":"ContainerDied","Data":"acb00c615a16c4afc4f4512ed580341d6c225c7726089602e71b32f317ded8e2"} Mar 13 14:40:02 crc kubenswrapper[4907]: I0313 14:40:02.029465 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mqhct" event={"ID":"631842f8-7a11-4ec0-bab3-5793a949806d","Type":"ContainerDied","Data":"3849a706a8af92bea878d778fade40002b62f87b6f1a978975072a05bbe7b2c8"} Mar 13 14:40:02 crc kubenswrapper[4907]: I0313 14:40:02.029488 4907 scope.go:117] "RemoveContainer" containerID="acb00c615a16c4afc4f4512ed580341d6c225c7726089602e71b32f317ded8e2" Mar 13 14:40:02 crc kubenswrapper[4907]: I0313 14:40:02.052785 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-mqhct"] Mar 13 14:40:02 crc kubenswrapper[4907]: I0313 14:40:02.058730 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-mqhct"] Mar 13 14:40:02 crc kubenswrapper[4907]: I0313 14:40:02.066092 4907 scope.go:117] "RemoveContainer" containerID="fe74f9a5411a632ebac06894970db3a7519b223f6ba533c92f921f1566e8a759" Mar 13 14:40:02 crc kubenswrapper[4907]: I0313 14:40:02.083670 4907 scope.go:117] "RemoveContainer" containerID="8b6db5a70191fecc20fe10365c72f6b6153346de6e602aa9933891080440586f" Mar 13 14:40:02 crc kubenswrapper[4907]: I0313 14:40:02.110689 4907 scope.go:117] "RemoveContainer" containerID="acb00c615a16c4afc4f4512ed580341d6c225c7726089602e71b32f317ded8e2" Mar 13 14:40:02 crc kubenswrapper[4907]: E0313 14:40:02.111259 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"acb00c615a16c4afc4f4512ed580341d6c225c7726089602e71b32f317ded8e2\": container with ID starting with acb00c615a16c4afc4f4512ed580341d6c225c7726089602e71b32f317ded8e2 not found: ID does not exist" containerID="acb00c615a16c4afc4f4512ed580341d6c225c7726089602e71b32f317ded8e2" Mar 13 14:40:02 crc kubenswrapper[4907]: I0313 14:40:02.111359 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"acb00c615a16c4afc4f4512ed580341d6c225c7726089602e71b32f317ded8e2"} err="failed to get container status \"acb00c615a16c4afc4f4512ed580341d6c225c7726089602e71b32f317ded8e2\": rpc error: code = NotFound desc = could not find container \"acb00c615a16c4afc4f4512ed580341d6c225c7726089602e71b32f317ded8e2\": container with ID starting with acb00c615a16c4afc4f4512ed580341d6c225c7726089602e71b32f317ded8e2 not found: ID does not exist" Mar 13 14:40:02 crc kubenswrapper[4907]: I0313 14:40:02.111439 4907 scope.go:117] "RemoveContainer" containerID="fe74f9a5411a632ebac06894970db3a7519b223f6ba533c92f921f1566e8a759" Mar 13 14:40:02 crc kubenswrapper[4907]: E0313 14:40:02.111944 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fe74f9a5411a632ebac06894970db3a7519b223f6ba533c92f921f1566e8a759\": container with ID starting with fe74f9a5411a632ebac06894970db3a7519b223f6ba533c92f921f1566e8a759 not found: ID does not exist" containerID="fe74f9a5411a632ebac06894970db3a7519b223f6ba533c92f921f1566e8a759" Mar 13 14:40:02 crc kubenswrapper[4907]: I0313 14:40:02.111972 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fe74f9a5411a632ebac06894970db3a7519b223f6ba533c92f921f1566e8a759"} err="failed to get container status \"fe74f9a5411a632ebac06894970db3a7519b223f6ba533c92f921f1566e8a759\": rpc error: code = NotFound desc = could not find container \"fe74f9a5411a632ebac06894970db3a7519b223f6ba533c92f921f1566e8a759\": container with ID starting with fe74f9a5411a632ebac06894970db3a7519b223f6ba533c92f921f1566e8a759 not found: ID does not exist" Mar 13 14:40:02 crc kubenswrapper[4907]: I0313 14:40:02.111990 4907 scope.go:117] "RemoveContainer" containerID="8b6db5a70191fecc20fe10365c72f6b6153346de6e602aa9933891080440586f" Mar 13 14:40:02 crc kubenswrapper[4907]: E0313 14:40:02.112349 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8b6db5a70191fecc20fe10365c72f6b6153346de6e602aa9933891080440586f\": container with ID starting with 8b6db5a70191fecc20fe10365c72f6b6153346de6e602aa9933891080440586f not found: ID does not exist" containerID="8b6db5a70191fecc20fe10365c72f6b6153346de6e602aa9933891080440586f" Mar 13 14:40:02 crc kubenswrapper[4907]: I0313 14:40:02.112374 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8b6db5a70191fecc20fe10365c72f6b6153346de6e602aa9933891080440586f"} err="failed to get container status \"8b6db5a70191fecc20fe10365c72f6b6153346de6e602aa9933891080440586f\": rpc error: code = NotFound desc = could not find container \"8b6db5a70191fecc20fe10365c72f6b6153346de6e602aa9933891080440586f\": container with ID starting with 8b6db5a70191fecc20fe10365c72f6b6153346de6e602aa9933891080440586f not found: ID does not exist" Mar 13 14:40:03 crc kubenswrapper[4907]: I0313 14:40:03.043786 4907 generic.go:334] "Generic (PLEG): container finished" podID="20dce4c7-7cbf-4e49-bd0e-93ee53d72551" containerID="fc46a881237631d2b078045b48ba1243ef9d84385c8036fc8ea4400ece996277" exitCode=0 Mar 13 14:40:03 crc kubenswrapper[4907]: I0313 14:40:03.043946 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556880-wxdzr" event={"ID":"20dce4c7-7cbf-4e49-bd0e-93ee53d72551","Type":"ContainerDied","Data":"fc46a881237631d2b078045b48ba1243ef9d84385c8036fc8ea4400ece996277"} Mar 13 14:40:03 crc kubenswrapper[4907]: I0313 14:40:03.807851 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="631842f8-7a11-4ec0-bab3-5793a949806d" path="/var/lib/kubelet/pods/631842f8-7a11-4ec0-bab3-5793a949806d/volumes" Mar 13 14:40:04 crc kubenswrapper[4907]: I0313 14:40:04.349327 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556880-wxdzr" Mar 13 14:40:04 crc kubenswrapper[4907]: I0313 14:40:04.434141 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nj4vk\" (UniqueName: \"kubernetes.io/projected/20dce4c7-7cbf-4e49-bd0e-93ee53d72551-kube-api-access-nj4vk\") pod \"20dce4c7-7cbf-4e49-bd0e-93ee53d72551\" (UID: \"20dce4c7-7cbf-4e49-bd0e-93ee53d72551\") " Mar 13 14:40:04 crc kubenswrapper[4907]: I0313 14:40:04.442543 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20dce4c7-7cbf-4e49-bd0e-93ee53d72551-kube-api-access-nj4vk" (OuterVolumeSpecName: "kube-api-access-nj4vk") pod "20dce4c7-7cbf-4e49-bd0e-93ee53d72551" (UID: "20dce4c7-7cbf-4e49-bd0e-93ee53d72551"). InnerVolumeSpecName "kube-api-access-nj4vk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:40:04 crc kubenswrapper[4907]: I0313 14:40:04.536116 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nj4vk\" (UniqueName: \"kubernetes.io/projected/20dce4c7-7cbf-4e49-bd0e-93ee53d72551-kube-api-access-nj4vk\") on node \"crc\" DevicePath \"\"" Mar 13 14:40:05 crc kubenswrapper[4907]: I0313 14:40:05.065038 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556880-wxdzr" event={"ID":"20dce4c7-7cbf-4e49-bd0e-93ee53d72551","Type":"ContainerDied","Data":"879e5062915a12a2855140a77d7a884e87419cf083062608de3e6cfa55590d80"} Mar 13 14:40:05 crc kubenswrapper[4907]: I0313 14:40:05.065085 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556880-wxdzr" Mar 13 14:40:05 crc kubenswrapper[4907]: I0313 14:40:05.065092 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="879e5062915a12a2855140a77d7a884e87419cf083062608de3e6cfa55590d80" Mar 13 14:40:05 crc kubenswrapper[4907]: I0313 14:40:05.415336 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556874-chrll"] Mar 13 14:40:05 crc kubenswrapper[4907]: I0313 14:40:05.421123 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556874-chrll"] Mar 13 14:40:05 crc kubenswrapper[4907]: I0313 14:40:05.792980 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bb80c3ca-ddb4-49fe-8b24-5dc43269ab1e" path="/var/lib/kubelet/pods/bb80c3ca-ddb4-49fe-8b24-5dc43269ab1e/volumes" Mar 13 14:40:08 crc kubenswrapper[4907]: E0313 14:40:08.225242 4907 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod631842f8_7a11_4ec0_bab3_5793a949806d.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod631842f8_7a11_4ec0_bab3_5793a949806d.slice/crio-3849a706a8af92bea878d778fade40002b62f87b6f1a978975072a05bbe7b2c8\": RecentStats: unable to find data in memory cache]" Mar 13 14:40:18 crc kubenswrapper[4907]: I0313 14:40:18.041961 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 14:40:18 crc kubenswrapper[4907]: I0313 14:40:18.042710 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 14:40:18 crc kubenswrapper[4907]: E0313 14:40:18.432648 4907 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod631842f8_7a11_4ec0_bab3_5793a949806d.slice/crio-3849a706a8af92bea878d778fade40002b62f87b6f1a978975072a05bbe7b2c8\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod631842f8_7a11_4ec0_bab3_5793a949806d.slice\": RecentStats: unable to find data in memory cache]" Mar 13 14:40:24 crc kubenswrapper[4907]: I0313 14:40:24.805206 4907 scope.go:117] "RemoveContainer" containerID="6a420a24c23cba094e67bdf8ae1b0ab87504764d74798dee0bc7e8921659b3b8" Mar 13 14:40:28 crc kubenswrapper[4907]: E0313 14:40:28.626060 4907 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod631842f8_7a11_4ec0_bab3_5793a949806d.slice/crio-3849a706a8af92bea878d778fade40002b62f87b6f1a978975072a05bbe7b2c8\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod631842f8_7a11_4ec0_bab3_5793a949806d.slice\": RecentStats: unable to find data in memory cache]" Mar 13 14:40:38 crc kubenswrapper[4907]: E0313 14:40:38.805560 4907 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod631842f8_7a11_4ec0_bab3_5793a949806d.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod631842f8_7a11_4ec0_bab3_5793a949806d.slice/crio-3849a706a8af92bea878d778fade40002b62f87b6f1a978975072a05bbe7b2c8\": RecentStats: unable to find data in memory cache]" Mar 13 14:40:48 crc kubenswrapper[4907]: I0313 14:40:48.042219 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 14:40:48 crc kubenswrapper[4907]: I0313 14:40:48.042637 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 14:40:48 crc kubenswrapper[4907]: I0313 14:40:48.042678 4907 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" Mar 13 14:40:48 crc kubenswrapper[4907]: I0313 14:40:48.043108 4907 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3c1970952dcd249074d61677266bef44d3d78378f555ffc20b89b00a033811dd"} pod="openshift-machine-config-operator/machine-config-daemon-hm56j" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 14:40:48 crc kubenswrapper[4907]: I0313 14:40:48.043149 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" containerID="cri-o://3c1970952dcd249074d61677266bef44d3d78378f555ffc20b89b00a033811dd" gracePeriod=600 Mar 13 14:40:48 crc kubenswrapper[4907]: I0313 14:40:48.421640 4907 generic.go:334] "Generic (PLEG): container finished" podID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerID="3c1970952dcd249074d61677266bef44d3d78378f555ffc20b89b00a033811dd" exitCode=0 Mar 13 14:40:48 crc kubenswrapper[4907]: I0313 14:40:48.421685 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerDied","Data":"3c1970952dcd249074d61677266bef44d3d78378f555ffc20b89b00a033811dd"} Mar 13 14:40:48 crc kubenswrapper[4907]: I0313 14:40:48.421921 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerStarted","Data":"43810caaacd8c4e6d986aec4d7996e802ba4907ce25cbd96e3f26a101537186b"} Mar 13 14:40:48 crc kubenswrapper[4907]: I0313 14:40:48.421958 4907 scope.go:117] "RemoveContainer" containerID="a09a3bec467b7e275495edb5827150af756bcc256911e7e06cbae5765922b5e2" Mar 13 14:40:49 crc kubenswrapper[4907]: E0313 14:40:49.011314 4907 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod631842f8_7a11_4ec0_bab3_5793a949806d.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod631842f8_7a11_4ec0_bab3_5793a949806d.slice/crio-3849a706a8af92bea878d778fade40002b62f87b6f1a978975072a05bbe7b2c8\": RecentStats: unable to find data in memory cache]" Mar 13 14:40:59 crc kubenswrapper[4907]: E0313 14:40:59.191465 4907 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod631842f8_7a11_4ec0_bab3_5793a949806d.slice/crio-3849a706a8af92bea878d778fade40002b62f87b6f1a978975072a05bbe7b2c8\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod631842f8_7a11_4ec0_bab3_5793a949806d.slice\": RecentStats: unable to find data in memory cache]" Mar 13 14:41:31 crc kubenswrapper[4907]: I0313 14:41:31.494127 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-gg9hb"] Mar 13 14:41:31 crc kubenswrapper[4907]: E0313 14:41:31.495515 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20dce4c7-7cbf-4e49-bd0e-93ee53d72551" containerName="oc" Mar 13 14:41:31 crc kubenswrapper[4907]: I0313 14:41:31.495543 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="20dce4c7-7cbf-4e49-bd0e-93ee53d72551" containerName="oc" Mar 13 14:41:31 crc kubenswrapper[4907]: E0313 14:41:31.495565 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="631842f8-7a11-4ec0-bab3-5793a949806d" containerName="extract-utilities" Mar 13 14:41:31 crc kubenswrapper[4907]: I0313 14:41:31.495574 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="631842f8-7a11-4ec0-bab3-5793a949806d" containerName="extract-utilities" Mar 13 14:41:31 crc kubenswrapper[4907]: E0313 14:41:31.495595 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="631842f8-7a11-4ec0-bab3-5793a949806d" containerName="extract-content" Mar 13 14:41:31 crc kubenswrapper[4907]: I0313 14:41:31.495604 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="631842f8-7a11-4ec0-bab3-5793a949806d" containerName="extract-content" Mar 13 14:41:31 crc kubenswrapper[4907]: E0313 14:41:31.495636 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="631842f8-7a11-4ec0-bab3-5793a949806d" containerName="registry-server" Mar 13 14:41:31 crc kubenswrapper[4907]: I0313 14:41:31.495646 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="631842f8-7a11-4ec0-bab3-5793a949806d" containerName="registry-server" Mar 13 14:41:31 crc kubenswrapper[4907]: I0313 14:41:31.495831 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="20dce4c7-7cbf-4e49-bd0e-93ee53d72551" containerName="oc" Mar 13 14:41:31 crc kubenswrapper[4907]: I0313 14:41:31.495901 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="631842f8-7a11-4ec0-bab3-5793a949806d" containerName="registry-server" Mar 13 14:41:31 crc kubenswrapper[4907]: I0313 14:41:31.498571 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gg9hb" Mar 13 14:41:31 crc kubenswrapper[4907]: I0313 14:41:31.503932 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-gg9hb"] Mar 13 14:41:31 crc kubenswrapper[4907]: I0313 14:41:31.511322 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ff81c42-6310-47a8-a691-5a7193a1f806-catalog-content\") pod \"redhat-operators-gg9hb\" (UID: \"7ff81c42-6310-47a8-a691-5a7193a1f806\") " pod="openshift-marketplace/redhat-operators-gg9hb" Mar 13 14:41:31 crc kubenswrapper[4907]: I0313 14:41:31.511381 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ff81c42-6310-47a8-a691-5a7193a1f806-utilities\") pod \"redhat-operators-gg9hb\" (UID: \"7ff81c42-6310-47a8-a691-5a7193a1f806\") " pod="openshift-marketplace/redhat-operators-gg9hb" Mar 13 14:41:31 crc kubenswrapper[4907]: I0313 14:41:31.511416 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2c5n9\" (UniqueName: \"kubernetes.io/projected/7ff81c42-6310-47a8-a691-5a7193a1f806-kube-api-access-2c5n9\") pod \"redhat-operators-gg9hb\" (UID: \"7ff81c42-6310-47a8-a691-5a7193a1f806\") " pod="openshift-marketplace/redhat-operators-gg9hb" Mar 13 14:41:31 crc kubenswrapper[4907]: I0313 14:41:31.613119 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ff81c42-6310-47a8-a691-5a7193a1f806-utilities\") pod \"redhat-operators-gg9hb\" (UID: \"7ff81c42-6310-47a8-a691-5a7193a1f806\") " pod="openshift-marketplace/redhat-operators-gg9hb" Mar 13 14:41:31 crc kubenswrapper[4907]: I0313 14:41:31.613188 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2c5n9\" (UniqueName: \"kubernetes.io/projected/7ff81c42-6310-47a8-a691-5a7193a1f806-kube-api-access-2c5n9\") pod \"redhat-operators-gg9hb\" (UID: \"7ff81c42-6310-47a8-a691-5a7193a1f806\") " pod="openshift-marketplace/redhat-operators-gg9hb" Mar 13 14:41:31 crc kubenswrapper[4907]: I0313 14:41:31.613454 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ff81c42-6310-47a8-a691-5a7193a1f806-catalog-content\") pod \"redhat-operators-gg9hb\" (UID: \"7ff81c42-6310-47a8-a691-5a7193a1f806\") " pod="openshift-marketplace/redhat-operators-gg9hb" Mar 13 14:41:31 crc kubenswrapper[4907]: I0313 14:41:31.613800 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ff81c42-6310-47a8-a691-5a7193a1f806-utilities\") pod \"redhat-operators-gg9hb\" (UID: \"7ff81c42-6310-47a8-a691-5a7193a1f806\") " pod="openshift-marketplace/redhat-operators-gg9hb" Mar 13 14:41:31 crc kubenswrapper[4907]: I0313 14:41:31.613800 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ff81c42-6310-47a8-a691-5a7193a1f806-catalog-content\") pod \"redhat-operators-gg9hb\" (UID: \"7ff81c42-6310-47a8-a691-5a7193a1f806\") " pod="openshift-marketplace/redhat-operators-gg9hb" Mar 13 14:41:31 crc kubenswrapper[4907]: I0313 14:41:31.636349 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2c5n9\" (UniqueName: \"kubernetes.io/projected/7ff81c42-6310-47a8-a691-5a7193a1f806-kube-api-access-2c5n9\") pod \"redhat-operators-gg9hb\" (UID: \"7ff81c42-6310-47a8-a691-5a7193a1f806\") " pod="openshift-marketplace/redhat-operators-gg9hb" Mar 13 14:41:31 crc kubenswrapper[4907]: I0313 14:41:31.816952 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gg9hb" Mar 13 14:41:32 crc kubenswrapper[4907]: I0313 14:41:32.276645 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-gg9hb"] Mar 13 14:41:32 crc kubenswrapper[4907]: I0313 14:41:32.757840 4907 generic.go:334] "Generic (PLEG): container finished" podID="7ff81c42-6310-47a8-a691-5a7193a1f806" containerID="04c538f4410703a67c9a1f6f70fb480854f64cb603b98dc4fb30a4da10e1cfe1" exitCode=0 Mar 13 14:41:32 crc kubenswrapper[4907]: I0313 14:41:32.758036 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gg9hb" event={"ID":"7ff81c42-6310-47a8-a691-5a7193a1f806","Type":"ContainerDied","Data":"04c538f4410703a67c9a1f6f70fb480854f64cb603b98dc4fb30a4da10e1cfe1"} Mar 13 14:41:32 crc kubenswrapper[4907]: I0313 14:41:32.759196 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gg9hb" event={"ID":"7ff81c42-6310-47a8-a691-5a7193a1f806","Type":"ContainerStarted","Data":"7e29cff77ed5e8b8c0c9fb8223d475e0bf231b38d49b85e0d18a0eff7bad0daf"} Mar 13 14:41:33 crc kubenswrapper[4907]: I0313 14:41:33.767139 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gg9hb" event={"ID":"7ff81c42-6310-47a8-a691-5a7193a1f806","Type":"ContainerStarted","Data":"b5a0413a722d5577be71421092aa45b7b0b6b423078bf9472fdcb79739eef422"} Mar 13 14:41:34 crc kubenswrapper[4907]: I0313 14:41:34.780074 4907 generic.go:334] "Generic (PLEG): container finished" podID="7ff81c42-6310-47a8-a691-5a7193a1f806" containerID="b5a0413a722d5577be71421092aa45b7b0b6b423078bf9472fdcb79739eef422" exitCode=0 Mar 13 14:41:34 crc kubenswrapper[4907]: I0313 14:41:34.780127 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gg9hb" event={"ID":"7ff81c42-6310-47a8-a691-5a7193a1f806","Type":"ContainerDied","Data":"b5a0413a722d5577be71421092aa45b7b0b6b423078bf9472fdcb79739eef422"} Mar 13 14:41:35 crc kubenswrapper[4907]: I0313 14:41:35.790182 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gg9hb" event={"ID":"7ff81c42-6310-47a8-a691-5a7193a1f806","Type":"ContainerStarted","Data":"1554edf4b3f861c840f593511dedd2375f020bd21204d76082b3b72355050269"} Mar 13 14:41:35 crc kubenswrapper[4907]: I0313 14:41:35.809133 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-gg9hb" podStartSLOduration=2.364032716 podStartE2EDuration="4.809114429s" podCreationTimestamp="2026-03-13 14:41:31 +0000 UTC" firstStartedPulling="2026-03-13 14:41:32.760064205 +0000 UTC m=+2191.659851894" lastFinishedPulling="2026-03-13 14:41:35.205145918 +0000 UTC m=+2194.104933607" observedRunningTime="2026-03-13 14:41:35.805193203 +0000 UTC m=+2194.704980902" watchObservedRunningTime="2026-03-13 14:41:35.809114429 +0000 UTC m=+2194.708902128" Mar 13 14:41:41 crc kubenswrapper[4907]: I0313 14:41:41.817559 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-gg9hb" Mar 13 14:41:41 crc kubenswrapper[4907]: I0313 14:41:41.819043 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-gg9hb" Mar 13 14:41:41 crc kubenswrapper[4907]: I0313 14:41:41.867142 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-gg9hb" Mar 13 14:41:41 crc kubenswrapper[4907]: I0313 14:41:41.910426 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-gg9hb" Mar 13 14:41:42 crc kubenswrapper[4907]: I0313 14:41:42.099508 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-gg9hb"] Mar 13 14:41:43 crc kubenswrapper[4907]: I0313 14:41:43.856188 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-gg9hb" podUID="7ff81c42-6310-47a8-a691-5a7193a1f806" containerName="registry-server" containerID="cri-o://1554edf4b3f861c840f593511dedd2375f020bd21204d76082b3b72355050269" gracePeriod=2 Mar 13 14:41:44 crc kubenswrapper[4907]: I0313 14:41:44.220640 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gg9hb" Mar 13 14:41:44 crc kubenswrapper[4907]: I0313 14:41:44.390861 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ff81c42-6310-47a8-a691-5a7193a1f806-catalog-content\") pod \"7ff81c42-6310-47a8-a691-5a7193a1f806\" (UID: \"7ff81c42-6310-47a8-a691-5a7193a1f806\") " Mar 13 14:41:44 crc kubenswrapper[4907]: I0313 14:41:44.390987 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2c5n9\" (UniqueName: \"kubernetes.io/projected/7ff81c42-6310-47a8-a691-5a7193a1f806-kube-api-access-2c5n9\") pod \"7ff81c42-6310-47a8-a691-5a7193a1f806\" (UID: \"7ff81c42-6310-47a8-a691-5a7193a1f806\") " Mar 13 14:41:44 crc kubenswrapper[4907]: I0313 14:41:44.392098 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ff81c42-6310-47a8-a691-5a7193a1f806-utilities\") pod \"7ff81c42-6310-47a8-a691-5a7193a1f806\" (UID: \"7ff81c42-6310-47a8-a691-5a7193a1f806\") " Mar 13 14:41:44 crc kubenswrapper[4907]: I0313 14:41:44.392810 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7ff81c42-6310-47a8-a691-5a7193a1f806-utilities" (OuterVolumeSpecName: "utilities") pod "7ff81c42-6310-47a8-a691-5a7193a1f806" (UID: "7ff81c42-6310-47a8-a691-5a7193a1f806"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:41:44 crc kubenswrapper[4907]: I0313 14:41:44.401547 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ff81c42-6310-47a8-a691-5a7193a1f806-kube-api-access-2c5n9" (OuterVolumeSpecName: "kube-api-access-2c5n9") pod "7ff81c42-6310-47a8-a691-5a7193a1f806" (UID: "7ff81c42-6310-47a8-a691-5a7193a1f806"). InnerVolumeSpecName "kube-api-access-2c5n9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:41:44 crc kubenswrapper[4907]: I0313 14:41:44.494004 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2c5n9\" (UniqueName: \"kubernetes.io/projected/7ff81c42-6310-47a8-a691-5a7193a1f806-kube-api-access-2c5n9\") on node \"crc\" DevicePath \"\"" Mar 13 14:41:44 crc kubenswrapper[4907]: I0313 14:41:44.494040 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ff81c42-6310-47a8-a691-5a7193a1f806-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 14:41:44 crc kubenswrapper[4907]: I0313 14:41:44.865713 4907 generic.go:334] "Generic (PLEG): container finished" podID="7ff81c42-6310-47a8-a691-5a7193a1f806" containerID="1554edf4b3f861c840f593511dedd2375f020bd21204d76082b3b72355050269" exitCode=0 Mar 13 14:41:44 crc kubenswrapper[4907]: I0313 14:41:44.865760 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gg9hb" event={"ID":"7ff81c42-6310-47a8-a691-5a7193a1f806","Type":"ContainerDied","Data":"1554edf4b3f861c840f593511dedd2375f020bd21204d76082b3b72355050269"} Mar 13 14:41:44 crc kubenswrapper[4907]: I0313 14:41:44.865787 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gg9hb" event={"ID":"7ff81c42-6310-47a8-a691-5a7193a1f806","Type":"ContainerDied","Data":"7e29cff77ed5e8b8c0c9fb8223d475e0bf231b38d49b85e0d18a0eff7bad0daf"} Mar 13 14:41:44 crc kubenswrapper[4907]: I0313 14:41:44.865806 4907 scope.go:117] "RemoveContainer" containerID="1554edf4b3f861c840f593511dedd2375f020bd21204d76082b3b72355050269" Mar 13 14:41:44 crc kubenswrapper[4907]: I0313 14:41:44.867013 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gg9hb" Mar 13 14:41:44 crc kubenswrapper[4907]: I0313 14:41:44.884661 4907 scope.go:117] "RemoveContainer" containerID="b5a0413a722d5577be71421092aa45b7b0b6b423078bf9472fdcb79739eef422" Mar 13 14:41:44 crc kubenswrapper[4907]: I0313 14:41:44.899871 4907 scope.go:117] "RemoveContainer" containerID="04c538f4410703a67c9a1f6f70fb480854f64cb603b98dc4fb30a4da10e1cfe1" Mar 13 14:41:44 crc kubenswrapper[4907]: I0313 14:41:44.922303 4907 scope.go:117] "RemoveContainer" containerID="1554edf4b3f861c840f593511dedd2375f020bd21204d76082b3b72355050269" Mar 13 14:41:44 crc kubenswrapper[4907]: E0313 14:41:44.922754 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1554edf4b3f861c840f593511dedd2375f020bd21204d76082b3b72355050269\": container with ID starting with 1554edf4b3f861c840f593511dedd2375f020bd21204d76082b3b72355050269 not found: ID does not exist" containerID="1554edf4b3f861c840f593511dedd2375f020bd21204d76082b3b72355050269" Mar 13 14:41:44 crc kubenswrapper[4907]: I0313 14:41:44.922814 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1554edf4b3f861c840f593511dedd2375f020bd21204d76082b3b72355050269"} err="failed to get container status \"1554edf4b3f861c840f593511dedd2375f020bd21204d76082b3b72355050269\": rpc error: code = NotFound desc = could not find container \"1554edf4b3f861c840f593511dedd2375f020bd21204d76082b3b72355050269\": container with ID starting with 1554edf4b3f861c840f593511dedd2375f020bd21204d76082b3b72355050269 not found: ID does not exist" Mar 13 14:41:44 crc kubenswrapper[4907]: I0313 14:41:44.922855 4907 scope.go:117] "RemoveContainer" containerID="b5a0413a722d5577be71421092aa45b7b0b6b423078bf9472fdcb79739eef422" Mar 13 14:41:44 crc kubenswrapper[4907]: E0313 14:41:44.923284 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b5a0413a722d5577be71421092aa45b7b0b6b423078bf9472fdcb79739eef422\": container with ID starting with b5a0413a722d5577be71421092aa45b7b0b6b423078bf9472fdcb79739eef422 not found: ID does not exist" containerID="b5a0413a722d5577be71421092aa45b7b0b6b423078bf9472fdcb79739eef422" Mar 13 14:41:44 crc kubenswrapper[4907]: I0313 14:41:44.923319 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b5a0413a722d5577be71421092aa45b7b0b6b423078bf9472fdcb79739eef422"} err="failed to get container status \"b5a0413a722d5577be71421092aa45b7b0b6b423078bf9472fdcb79739eef422\": rpc error: code = NotFound desc = could not find container \"b5a0413a722d5577be71421092aa45b7b0b6b423078bf9472fdcb79739eef422\": container with ID starting with b5a0413a722d5577be71421092aa45b7b0b6b423078bf9472fdcb79739eef422 not found: ID does not exist" Mar 13 14:41:44 crc kubenswrapper[4907]: I0313 14:41:44.923341 4907 scope.go:117] "RemoveContainer" containerID="04c538f4410703a67c9a1f6f70fb480854f64cb603b98dc4fb30a4da10e1cfe1" Mar 13 14:41:44 crc kubenswrapper[4907]: E0313 14:41:44.923630 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"04c538f4410703a67c9a1f6f70fb480854f64cb603b98dc4fb30a4da10e1cfe1\": container with ID starting with 04c538f4410703a67c9a1f6f70fb480854f64cb603b98dc4fb30a4da10e1cfe1 not found: ID does not exist" containerID="04c538f4410703a67c9a1f6f70fb480854f64cb603b98dc4fb30a4da10e1cfe1" Mar 13 14:41:44 crc kubenswrapper[4907]: I0313 14:41:44.923666 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"04c538f4410703a67c9a1f6f70fb480854f64cb603b98dc4fb30a4da10e1cfe1"} err="failed to get container status \"04c538f4410703a67c9a1f6f70fb480854f64cb603b98dc4fb30a4da10e1cfe1\": rpc error: code = NotFound desc = could not find container \"04c538f4410703a67c9a1f6f70fb480854f64cb603b98dc4fb30a4da10e1cfe1\": container with ID starting with 04c538f4410703a67c9a1f6f70fb480854f64cb603b98dc4fb30a4da10e1cfe1 not found: ID does not exist" Mar 13 14:41:45 crc kubenswrapper[4907]: I0313 14:41:45.556679 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7ff81c42-6310-47a8-a691-5a7193a1f806-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7ff81c42-6310-47a8-a691-5a7193a1f806" (UID: "7ff81c42-6310-47a8-a691-5a7193a1f806"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:41:45 crc kubenswrapper[4907]: I0313 14:41:45.610996 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ff81c42-6310-47a8-a691-5a7193a1f806-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 14:41:45 crc kubenswrapper[4907]: I0313 14:41:45.804243 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-gg9hb"] Mar 13 14:41:45 crc kubenswrapper[4907]: I0313 14:41:45.811168 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-gg9hb"] Mar 13 14:41:47 crc kubenswrapper[4907]: I0313 14:41:47.792617 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7ff81c42-6310-47a8-a691-5a7193a1f806" path="/var/lib/kubelet/pods/7ff81c42-6310-47a8-a691-5a7193a1f806/volumes" Mar 13 14:41:49 crc kubenswrapper[4907]: I0313 14:41:49.318970 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-kh8q2"] Mar 13 14:41:49 crc kubenswrapper[4907]: E0313 14:41:49.319947 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ff81c42-6310-47a8-a691-5a7193a1f806" containerName="extract-content" Mar 13 14:41:49 crc kubenswrapper[4907]: I0313 14:41:49.320251 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ff81c42-6310-47a8-a691-5a7193a1f806" containerName="extract-content" Mar 13 14:41:49 crc kubenswrapper[4907]: E0313 14:41:49.320342 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ff81c42-6310-47a8-a691-5a7193a1f806" containerName="extract-utilities" Mar 13 14:41:49 crc kubenswrapper[4907]: I0313 14:41:49.320417 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ff81c42-6310-47a8-a691-5a7193a1f806" containerName="extract-utilities" Mar 13 14:41:49 crc kubenswrapper[4907]: E0313 14:41:49.320494 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ff81c42-6310-47a8-a691-5a7193a1f806" containerName="registry-server" Mar 13 14:41:49 crc kubenswrapper[4907]: I0313 14:41:49.320576 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ff81c42-6310-47a8-a691-5a7193a1f806" containerName="registry-server" Mar 13 14:41:49 crc kubenswrapper[4907]: I0313 14:41:49.320800 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ff81c42-6310-47a8-a691-5a7193a1f806" containerName="registry-server" Mar 13 14:41:49 crc kubenswrapper[4907]: I0313 14:41:49.322360 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kh8q2" Mar 13 14:41:49 crc kubenswrapper[4907]: I0313 14:41:49.336633 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kh8q2"] Mar 13 14:41:49 crc kubenswrapper[4907]: I0313 14:41:49.461609 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bbd9d05a-e055-4974-b018-f1d8fdd10178-catalog-content\") pod \"certified-operators-kh8q2\" (UID: \"bbd9d05a-e055-4974-b018-f1d8fdd10178\") " pod="openshift-marketplace/certified-operators-kh8q2" Mar 13 14:41:49 crc kubenswrapper[4907]: I0313 14:41:49.461655 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rmldf\" (UniqueName: \"kubernetes.io/projected/bbd9d05a-e055-4974-b018-f1d8fdd10178-kube-api-access-rmldf\") pod \"certified-operators-kh8q2\" (UID: \"bbd9d05a-e055-4974-b018-f1d8fdd10178\") " pod="openshift-marketplace/certified-operators-kh8q2" Mar 13 14:41:49 crc kubenswrapper[4907]: I0313 14:41:49.461675 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bbd9d05a-e055-4974-b018-f1d8fdd10178-utilities\") pod \"certified-operators-kh8q2\" (UID: \"bbd9d05a-e055-4974-b018-f1d8fdd10178\") " pod="openshift-marketplace/certified-operators-kh8q2" Mar 13 14:41:49 crc kubenswrapper[4907]: I0313 14:41:49.562838 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bbd9d05a-e055-4974-b018-f1d8fdd10178-catalog-content\") pod \"certified-operators-kh8q2\" (UID: \"bbd9d05a-e055-4974-b018-f1d8fdd10178\") " pod="openshift-marketplace/certified-operators-kh8q2" Mar 13 14:41:49 crc kubenswrapper[4907]: I0313 14:41:49.562932 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rmldf\" (UniqueName: \"kubernetes.io/projected/bbd9d05a-e055-4974-b018-f1d8fdd10178-kube-api-access-rmldf\") pod \"certified-operators-kh8q2\" (UID: \"bbd9d05a-e055-4974-b018-f1d8fdd10178\") " pod="openshift-marketplace/certified-operators-kh8q2" Mar 13 14:41:49 crc kubenswrapper[4907]: I0313 14:41:49.562975 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bbd9d05a-e055-4974-b018-f1d8fdd10178-utilities\") pod \"certified-operators-kh8q2\" (UID: \"bbd9d05a-e055-4974-b018-f1d8fdd10178\") " pod="openshift-marketplace/certified-operators-kh8q2" Mar 13 14:41:49 crc kubenswrapper[4907]: I0313 14:41:49.563397 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bbd9d05a-e055-4974-b018-f1d8fdd10178-catalog-content\") pod \"certified-operators-kh8q2\" (UID: \"bbd9d05a-e055-4974-b018-f1d8fdd10178\") " pod="openshift-marketplace/certified-operators-kh8q2" Mar 13 14:41:49 crc kubenswrapper[4907]: I0313 14:41:49.563597 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bbd9d05a-e055-4974-b018-f1d8fdd10178-utilities\") pod \"certified-operators-kh8q2\" (UID: \"bbd9d05a-e055-4974-b018-f1d8fdd10178\") " pod="openshift-marketplace/certified-operators-kh8q2" Mar 13 14:41:49 crc kubenswrapper[4907]: I0313 14:41:49.618029 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rmldf\" (UniqueName: \"kubernetes.io/projected/bbd9d05a-e055-4974-b018-f1d8fdd10178-kube-api-access-rmldf\") pod \"certified-operators-kh8q2\" (UID: \"bbd9d05a-e055-4974-b018-f1d8fdd10178\") " pod="openshift-marketplace/certified-operators-kh8q2" Mar 13 14:41:49 crc kubenswrapper[4907]: I0313 14:41:49.642004 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kh8q2" Mar 13 14:41:50 crc kubenswrapper[4907]: I0313 14:41:50.150076 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kh8q2"] Mar 13 14:41:50 crc kubenswrapper[4907]: I0313 14:41:50.905586 4907 generic.go:334] "Generic (PLEG): container finished" podID="bbd9d05a-e055-4974-b018-f1d8fdd10178" containerID="c29f21e67b78bfa6321f01394b48fae88a9c4f77cce83e14401fce3ca805ce44" exitCode=0 Mar 13 14:41:50 crc kubenswrapper[4907]: I0313 14:41:50.905661 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kh8q2" event={"ID":"bbd9d05a-e055-4974-b018-f1d8fdd10178","Type":"ContainerDied","Data":"c29f21e67b78bfa6321f01394b48fae88a9c4f77cce83e14401fce3ca805ce44"} Mar 13 14:41:50 crc kubenswrapper[4907]: I0313 14:41:50.905932 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kh8q2" event={"ID":"bbd9d05a-e055-4974-b018-f1d8fdd10178","Type":"ContainerStarted","Data":"410b8cd335bd7034fb5a1e0759b6ec51be56bf7900b40557e8a393482878522f"} Mar 13 14:41:51 crc kubenswrapper[4907]: I0313 14:41:51.914307 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kh8q2" event={"ID":"bbd9d05a-e055-4974-b018-f1d8fdd10178","Type":"ContainerStarted","Data":"da85c297a9d3589a9b32a44fd9795a8682a003114947247f16b543e60836dd31"} Mar 13 14:41:52 crc kubenswrapper[4907]: I0313 14:41:52.921807 4907 generic.go:334] "Generic (PLEG): container finished" podID="bbd9d05a-e055-4974-b018-f1d8fdd10178" containerID="da85c297a9d3589a9b32a44fd9795a8682a003114947247f16b543e60836dd31" exitCode=0 Mar 13 14:41:52 crc kubenswrapper[4907]: I0313 14:41:52.921863 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kh8q2" event={"ID":"bbd9d05a-e055-4974-b018-f1d8fdd10178","Type":"ContainerDied","Data":"da85c297a9d3589a9b32a44fd9795a8682a003114947247f16b543e60836dd31"} Mar 13 14:41:53 crc kubenswrapper[4907]: I0313 14:41:53.934936 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kh8q2" event={"ID":"bbd9d05a-e055-4974-b018-f1d8fdd10178","Type":"ContainerStarted","Data":"838f1e0eb2d0fa893d5a1cfcea314424ab3346a5d58f11c35c05c50224a6bf64"} Mar 13 14:41:53 crc kubenswrapper[4907]: I0313 14:41:53.961325 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-kh8q2" podStartSLOduration=2.124273589 podStartE2EDuration="4.961295885s" podCreationTimestamp="2026-03-13 14:41:49 +0000 UTC" firstStartedPulling="2026-03-13 14:41:50.907080918 +0000 UTC m=+2209.806868607" lastFinishedPulling="2026-03-13 14:41:53.744103214 +0000 UTC m=+2212.643890903" observedRunningTime="2026-03-13 14:41:53.960161253 +0000 UTC m=+2212.859948942" watchObservedRunningTime="2026-03-13 14:41:53.961295885 +0000 UTC m=+2212.861083614" Mar 13 14:41:59 crc kubenswrapper[4907]: I0313 14:41:59.649119 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-kh8q2" Mar 13 14:41:59 crc kubenswrapper[4907]: I0313 14:41:59.649805 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-kh8q2" Mar 13 14:41:59 crc kubenswrapper[4907]: I0313 14:41:59.692127 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-kh8q2" Mar 13 14:42:00 crc kubenswrapper[4907]: I0313 14:42:00.020515 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-kh8q2" Mar 13 14:42:00 crc kubenswrapper[4907]: I0313 14:42:00.065903 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kh8q2"] Mar 13 14:42:00 crc kubenswrapper[4907]: I0313 14:42:00.143402 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556882-c7dc8"] Mar 13 14:42:00 crc kubenswrapper[4907]: I0313 14:42:00.144258 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556882-c7dc8" Mar 13 14:42:00 crc kubenswrapper[4907]: I0313 14:42:00.146932 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 14:42:00 crc kubenswrapper[4907]: I0313 14:42:00.147426 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 14:42:00 crc kubenswrapper[4907]: I0313 14:42:00.147433 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 14:42:00 crc kubenswrapper[4907]: I0313 14:42:00.158459 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556882-c7dc8"] Mar 13 14:42:00 crc kubenswrapper[4907]: I0313 14:42:00.212997 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xkkqt\" (UniqueName: \"kubernetes.io/projected/81733636-fc5e-4687-9ceb-0dd2e18fa15a-kube-api-access-xkkqt\") pod \"auto-csr-approver-29556882-c7dc8\" (UID: \"81733636-fc5e-4687-9ceb-0dd2e18fa15a\") " pod="openshift-infra/auto-csr-approver-29556882-c7dc8" Mar 13 14:42:00 crc kubenswrapper[4907]: I0313 14:42:00.315795 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xkkqt\" (UniqueName: \"kubernetes.io/projected/81733636-fc5e-4687-9ceb-0dd2e18fa15a-kube-api-access-xkkqt\") pod \"auto-csr-approver-29556882-c7dc8\" (UID: \"81733636-fc5e-4687-9ceb-0dd2e18fa15a\") " pod="openshift-infra/auto-csr-approver-29556882-c7dc8" Mar 13 14:42:00 crc kubenswrapper[4907]: I0313 14:42:00.336935 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xkkqt\" (UniqueName: \"kubernetes.io/projected/81733636-fc5e-4687-9ceb-0dd2e18fa15a-kube-api-access-xkkqt\") pod \"auto-csr-approver-29556882-c7dc8\" (UID: \"81733636-fc5e-4687-9ceb-0dd2e18fa15a\") " pod="openshift-infra/auto-csr-approver-29556882-c7dc8" Mar 13 14:42:00 crc kubenswrapper[4907]: I0313 14:42:00.467025 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556882-c7dc8" Mar 13 14:42:00 crc kubenswrapper[4907]: I0313 14:42:00.897340 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556882-c7dc8"] Mar 13 14:42:00 crc kubenswrapper[4907]: W0313 14:42:00.904952 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod81733636_fc5e_4687_9ceb_0dd2e18fa15a.slice/crio-47339f6bc7f00baa2ec0d1f9fb7f56f3dda07e47348aedc197d070f454b4bb76 WatchSource:0}: Error finding container 47339f6bc7f00baa2ec0d1f9fb7f56f3dda07e47348aedc197d070f454b4bb76: Status 404 returned error can't find the container with id 47339f6bc7f00baa2ec0d1f9fb7f56f3dda07e47348aedc197d070f454b4bb76 Mar 13 14:42:00 crc kubenswrapper[4907]: I0313 14:42:00.985594 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556882-c7dc8" event={"ID":"81733636-fc5e-4687-9ceb-0dd2e18fa15a","Type":"ContainerStarted","Data":"47339f6bc7f00baa2ec0d1f9fb7f56f3dda07e47348aedc197d070f454b4bb76"} Mar 13 14:42:01 crc kubenswrapper[4907]: I0313 14:42:01.994427 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-kh8q2" podUID="bbd9d05a-e055-4974-b018-f1d8fdd10178" containerName="registry-server" containerID="cri-o://838f1e0eb2d0fa893d5a1cfcea314424ab3346a5d58f11c35c05c50224a6bf64" gracePeriod=2 Mar 13 14:42:02 crc kubenswrapper[4907]: I0313 14:42:02.926270 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kh8q2" Mar 13 14:42:03 crc kubenswrapper[4907]: I0313 14:42:03.006358 4907 generic.go:334] "Generic (PLEG): container finished" podID="81733636-fc5e-4687-9ceb-0dd2e18fa15a" containerID="888132585d34376894f6b13900aa0293c6191b881a23e9abf82e3ca297b39ec9" exitCode=0 Mar 13 14:42:03 crc kubenswrapper[4907]: I0313 14:42:03.006411 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556882-c7dc8" event={"ID":"81733636-fc5e-4687-9ceb-0dd2e18fa15a","Type":"ContainerDied","Data":"888132585d34376894f6b13900aa0293c6191b881a23e9abf82e3ca297b39ec9"} Mar 13 14:42:03 crc kubenswrapper[4907]: I0313 14:42:03.009730 4907 generic.go:334] "Generic (PLEG): container finished" podID="bbd9d05a-e055-4974-b018-f1d8fdd10178" containerID="838f1e0eb2d0fa893d5a1cfcea314424ab3346a5d58f11c35c05c50224a6bf64" exitCode=0 Mar 13 14:42:03 crc kubenswrapper[4907]: I0313 14:42:03.009770 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kh8q2" event={"ID":"bbd9d05a-e055-4974-b018-f1d8fdd10178","Type":"ContainerDied","Data":"838f1e0eb2d0fa893d5a1cfcea314424ab3346a5d58f11c35c05c50224a6bf64"} Mar 13 14:42:03 crc kubenswrapper[4907]: I0313 14:42:03.009792 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kh8q2" Mar 13 14:42:03 crc kubenswrapper[4907]: I0313 14:42:03.009830 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kh8q2" event={"ID":"bbd9d05a-e055-4974-b018-f1d8fdd10178","Type":"ContainerDied","Data":"410b8cd335bd7034fb5a1e0759b6ec51be56bf7900b40557e8a393482878522f"} Mar 13 14:42:03 crc kubenswrapper[4907]: I0313 14:42:03.009861 4907 scope.go:117] "RemoveContainer" containerID="838f1e0eb2d0fa893d5a1cfcea314424ab3346a5d58f11c35c05c50224a6bf64" Mar 13 14:42:03 crc kubenswrapper[4907]: I0313 14:42:03.034816 4907 scope.go:117] "RemoveContainer" containerID="da85c297a9d3589a9b32a44fd9795a8682a003114947247f16b543e60836dd31" Mar 13 14:42:03 crc kubenswrapper[4907]: I0313 14:42:03.050324 4907 scope.go:117] "RemoveContainer" containerID="c29f21e67b78bfa6321f01394b48fae88a9c4f77cce83e14401fce3ca805ce44" Mar 13 14:42:03 crc kubenswrapper[4907]: I0313 14:42:03.062492 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rmldf\" (UniqueName: \"kubernetes.io/projected/bbd9d05a-e055-4974-b018-f1d8fdd10178-kube-api-access-rmldf\") pod \"bbd9d05a-e055-4974-b018-f1d8fdd10178\" (UID: \"bbd9d05a-e055-4974-b018-f1d8fdd10178\") " Mar 13 14:42:03 crc kubenswrapper[4907]: I0313 14:42:03.062577 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bbd9d05a-e055-4974-b018-f1d8fdd10178-utilities\") pod \"bbd9d05a-e055-4974-b018-f1d8fdd10178\" (UID: \"bbd9d05a-e055-4974-b018-f1d8fdd10178\") " Mar 13 14:42:03 crc kubenswrapper[4907]: I0313 14:42:03.062663 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bbd9d05a-e055-4974-b018-f1d8fdd10178-catalog-content\") pod \"bbd9d05a-e055-4974-b018-f1d8fdd10178\" (UID: \"bbd9d05a-e055-4974-b018-f1d8fdd10178\") " Mar 13 14:42:03 crc kubenswrapper[4907]: I0313 14:42:03.063366 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bbd9d05a-e055-4974-b018-f1d8fdd10178-utilities" (OuterVolumeSpecName: "utilities") pod "bbd9d05a-e055-4974-b018-f1d8fdd10178" (UID: "bbd9d05a-e055-4974-b018-f1d8fdd10178"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:42:03 crc kubenswrapper[4907]: I0313 14:42:03.068944 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bbd9d05a-e055-4974-b018-f1d8fdd10178-kube-api-access-rmldf" (OuterVolumeSpecName: "kube-api-access-rmldf") pod "bbd9d05a-e055-4974-b018-f1d8fdd10178" (UID: "bbd9d05a-e055-4974-b018-f1d8fdd10178"). InnerVolumeSpecName "kube-api-access-rmldf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:42:03 crc kubenswrapper[4907]: I0313 14:42:03.076985 4907 scope.go:117] "RemoveContainer" containerID="838f1e0eb2d0fa893d5a1cfcea314424ab3346a5d58f11c35c05c50224a6bf64" Mar 13 14:42:03 crc kubenswrapper[4907]: E0313 14:42:03.077463 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"838f1e0eb2d0fa893d5a1cfcea314424ab3346a5d58f11c35c05c50224a6bf64\": container with ID starting with 838f1e0eb2d0fa893d5a1cfcea314424ab3346a5d58f11c35c05c50224a6bf64 not found: ID does not exist" containerID="838f1e0eb2d0fa893d5a1cfcea314424ab3346a5d58f11c35c05c50224a6bf64" Mar 13 14:42:03 crc kubenswrapper[4907]: I0313 14:42:03.077500 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"838f1e0eb2d0fa893d5a1cfcea314424ab3346a5d58f11c35c05c50224a6bf64"} err="failed to get container status \"838f1e0eb2d0fa893d5a1cfcea314424ab3346a5d58f11c35c05c50224a6bf64\": rpc error: code = NotFound desc = could not find container \"838f1e0eb2d0fa893d5a1cfcea314424ab3346a5d58f11c35c05c50224a6bf64\": container with ID starting with 838f1e0eb2d0fa893d5a1cfcea314424ab3346a5d58f11c35c05c50224a6bf64 not found: ID does not exist" Mar 13 14:42:03 crc kubenswrapper[4907]: I0313 14:42:03.077542 4907 scope.go:117] "RemoveContainer" containerID="da85c297a9d3589a9b32a44fd9795a8682a003114947247f16b543e60836dd31" Mar 13 14:42:03 crc kubenswrapper[4907]: E0313 14:42:03.078105 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"da85c297a9d3589a9b32a44fd9795a8682a003114947247f16b543e60836dd31\": container with ID starting with da85c297a9d3589a9b32a44fd9795a8682a003114947247f16b543e60836dd31 not found: ID does not exist" containerID="da85c297a9d3589a9b32a44fd9795a8682a003114947247f16b543e60836dd31" Mar 13 14:42:03 crc kubenswrapper[4907]: I0313 14:42:03.078129 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da85c297a9d3589a9b32a44fd9795a8682a003114947247f16b543e60836dd31"} err="failed to get container status \"da85c297a9d3589a9b32a44fd9795a8682a003114947247f16b543e60836dd31\": rpc error: code = NotFound desc = could not find container \"da85c297a9d3589a9b32a44fd9795a8682a003114947247f16b543e60836dd31\": container with ID starting with da85c297a9d3589a9b32a44fd9795a8682a003114947247f16b543e60836dd31 not found: ID does not exist" Mar 13 14:42:03 crc kubenswrapper[4907]: I0313 14:42:03.078162 4907 scope.go:117] "RemoveContainer" containerID="c29f21e67b78bfa6321f01394b48fae88a9c4f77cce83e14401fce3ca805ce44" Mar 13 14:42:03 crc kubenswrapper[4907]: E0313 14:42:03.078418 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c29f21e67b78bfa6321f01394b48fae88a9c4f77cce83e14401fce3ca805ce44\": container with ID starting with c29f21e67b78bfa6321f01394b48fae88a9c4f77cce83e14401fce3ca805ce44 not found: ID does not exist" containerID="c29f21e67b78bfa6321f01394b48fae88a9c4f77cce83e14401fce3ca805ce44" Mar 13 14:42:03 crc kubenswrapper[4907]: I0313 14:42:03.078458 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c29f21e67b78bfa6321f01394b48fae88a9c4f77cce83e14401fce3ca805ce44"} err="failed to get container status \"c29f21e67b78bfa6321f01394b48fae88a9c4f77cce83e14401fce3ca805ce44\": rpc error: code = NotFound desc = could not find container \"c29f21e67b78bfa6321f01394b48fae88a9c4f77cce83e14401fce3ca805ce44\": container with ID starting with c29f21e67b78bfa6321f01394b48fae88a9c4f77cce83e14401fce3ca805ce44 not found: ID does not exist" Mar 13 14:42:03 crc kubenswrapper[4907]: I0313 14:42:03.130341 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bbd9d05a-e055-4974-b018-f1d8fdd10178-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bbd9d05a-e055-4974-b018-f1d8fdd10178" (UID: "bbd9d05a-e055-4974-b018-f1d8fdd10178"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:42:03 crc kubenswrapper[4907]: I0313 14:42:03.164581 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rmldf\" (UniqueName: \"kubernetes.io/projected/bbd9d05a-e055-4974-b018-f1d8fdd10178-kube-api-access-rmldf\") on node \"crc\" DevicePath \"\"" Mar 13 14:42:03 crc kubenswrapper[4907]: I0313 14:42:03.164609 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bbd9d05a-e055-4974-b018-f1d8fdd10178-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 14:42:03 crc kubenswrapper[4907]: I0313 14:42:03.164621 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bbd9d05a-e055-4974-b018-f1d8fdd10178-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 14:42:03 crc kubenswrapper[4907]: I0313 14:42:03.361396 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kh8q2"] Mar 13 14:42:03 crc kubenswrapper[4907]: I0313 14:42:03.375308 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-kh8q2"] Mar 13 14:42:03 crc kubenswrapper[4907]: I0313 14:42:03.790008 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bbd9d05a-e055-4974-b018-f1d8fdd10178" path="/var/lib/kubelet/pods/bbd9d05a-e055-4974-b018-f1d8fdd10178/volumes" Mar 13 14:42:04 crc kubenswrapper[4907]: I0313 14:42:04.345776 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556882-c7dc8" Mar 13 14:42:04 crc kubenswrapper[4907]: I0313 14:42:04.486099 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xkkqt\" (UniqueName: \"kubernetes.io/projected/81733636-fc5e-4687-9ceb-0dd2e18fa15a-kube-api-access-xkkqt\") pod \"81733636-fc5e-4687-9ceb-0dd2e18fa15a\" (UID: \"81733636-fc5e-4687-9ceb-0dd2e18fa15a\") " Mar 13 14:42:04 crc kubenswrapper[4907]: I0313 14:42:04.493493 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/81733636-fc5e-4687-9ceb-0dd2e18fa15a-kube-api-access-xkkqt" (OuterVolumeSpecName: "kube-api-access-xkkqt") pod "81733636-fc5e-4687-9ceb-0dd2e18fa15a" (UID: "81733636-fc5e-4687-9ceb-0dd2e18fa15a"). InnerVolumeSpecName "kube-api-access-xkkqt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:42:04 crc kubenswrapper[4907]: I0313 14:42:04.588638 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xkkqt\" (UniqueName: \"kubernetes.io/projected/81733636-fc5e-4687-9ceb-0dd2e18fa15a-kube-api-access-xkkqt\") on node \"crc\" DevicePath \"\"" Mar 13 14:42:05 crc kubenswrapper[4907]: I0313 14:42:05.031845 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556882-c7dc8" event={"ID":"81733636-fc5e-4687-9ceb-0dd2e18fa15a","Type":"ContainerDied","Data":"47339f6bc7f00baa2ec0d1f9fb7f56f3dda07e47348aedc197d070f454b4bb76"} Mar 13 14:42:05 crc kubenswrapper[4907]: I0313 14:42:05.031899 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="47339f6bc7f00baa2ec0d1f9fb7f56f3dda07e47348aedc197d070f454b4bb76" Mar 13 14:42:05 crc kubenswrapper[4907]: I0313 14:42:05.031911 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556882-c7dc8" Mar 13 14:42:05 crc kubenswrapper[4907]: I0313 14:42:05.418741 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556876-dwnhg"] Mar 13 14:42:05 crc kubenswrapper[4907]: I0313 14:42:05.424185 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556876-dwnhg"] Mar 13 14:42:05 crc kubenswrapper[4907]: I0313 14:42:05.794667 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d32570e6-4671-4b95-b312-f4af6c4a1478" path="/var/lib/kubelet/pods/d32570e6-4671-4b95-b312-f4af6c4a1478/volumes" Mar 13 14:42:24 crc kubenswrapper[4907]: I0313 14:42:24.923607 4907 scope.go:117] "RemoveContainer" containerID="ec22972508d423bdab0d6a15e5b2be7962b2967f440897919fe9018a840dad0f" Mar 13 14:42:34 crc kubenswrapper[4907]: I0313 14:42:34.185426 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-bwznz"] Mar 13 14:42:34 crc kubenswrapper[4907]: E0313 14:42:34.186673 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bbd9d05a-e055-4974-b018-f1d8fdd10178" containerName="registry-server" Mar 13 14:42:34 crc kubenswrapper[4907]: I0313 14:42:34.186707 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="bbd9d05a-e055-4974-b018-f1d8fdd10178" containerName="registry-server" Mar 13 14:42:34 crc kubenswrapper[4907]: E0313 14:42:34.186724 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bbd9d05a-e055-4974-b018-f1d8fdd10178" containerName="extract-content" Mar 13 14:42:34 crc kubenswrapper[4907]: I0313 14:42:34.186733 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="bbd9d05a-e055-4974-b018-f1d8fdd10178" containerName="extract-content" Mar 13 14:42:34 crc kubenswrapper[4907]: E0313 14:42:34.186750 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81733636-fc5e-4687-9ceb-0dd2e18fa15a" containerName="oc" Mar 13 14:42:34 crc kubenswrapper[4907]: I0313 14:42:34.186758 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="81733636-fc5e-4687-9ceb-0dd2e18fa15a" containerName="oc" Mar 13 14:42:34 crc kubenswrapper[4907]: E0313 14:42:34.186782 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bbd9d05a-e055-4974-b018-f1d8fdd10178" containerName="extract-utilities" Mar 13 14:42:34 crc kubenswrapper[4907]: I0313 14:42:34.186790 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="bbd9d05a-e055-4974-b018-f1d8fdd10178" containerName="extract-utilities" Mar 13 14:42:34 crc kubenswrapper[4907]: I0313 14:42:34.186987 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="81733636-fc5e-4687-9ceb-0dd2e18fa15a" containerName="oc" Mar 13 14:42:34 crc kubenswrapper[4907]: I0313 14:42:34.187018 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="bbd9d05a-e055-4974-b018-f1d8fdd10178" containerName="registry-server" Mar 13 14:42:34 crc kubenswrapper[4907]: I0313 14:42:34.188061 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bwznz" Mar 13 14:42:34 crc kubenswrapper[4907]: I0313 14:42:34.201745 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-bwznz"] Mar 13 14:42:34 crc kubenswrapper[4907]: I0313 14:42:34.259812 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rxdf7\" (UniqueName: \"kubernetes.io/projected/e3a3b609-3625-4670-b669-d7fd07386be6-kube-api-access-rxdf7\") pod \"community-operators-bwznz\" (UID: \"e3a3b609-3625-4670-b669-d7fd07386be6\") " pod="openshift-marketplace/community-operators-bwznz" Mar 13 14:42:34 crc kubenswrapper[4907]: I0313 14:42:34.259930 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e3a3b609-3625-4670-b669-d7fd07386be6-utilities\") pod \"community-operators-bwznz\" (UID: \"e3a3b609-3625-4670-b669-d7fd07386be6\") " pod="openshift-marketplace/community-operators-bwznz" Mar 13 14:42:34 crc kubenswrapper[4907]: I0313 14:42:34.259965 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e3a3b609-3625-4670-b669-d7fd07386be6-catalog-content\") pod \"community-operators-bwznz\" (UID: \"e3a3b609-3625-4670-b669-d7fd07386be6\") " pod="openshift-marketplace/community-operators-bwznz" Mar 13 14:42:34 crc kubenswrapper[4907]: I0313 14:42:34.361345 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rxdf7\" (UniqueName: \"kubernetes.io/projected/e3a3b609-3625-4670-b669-d7fd07386be6-kube-api-access-rxdf7\") pod \"community-operators-bwznz\" (UID: \"e3a3b609-3625-4670-b669-d7fd07386be6\") " pod="openshift-marketplace/community-operators-bwznz" Mar 13 14:42:34 crc kubenswrapper[4907]: I0313 14:42:34.361455 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e3a3b609-3625-4670-b669-d7fd07386be6-utilities\") pod \"community-operators-bwznz\" (UID: \"e3a3b609-3625-4670-b669-d7fd07386be6\") " pod="openshift-marketplace/community-operators-bwznz" Mar 13 14:42:34 crc kubenswrapper[4907]: I0313 14:42:34.361484 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e3a3b609-3625-4670-b669-d7fd07386be6-catalog-content\") pod \"community-operators-bwznz\" (UID: \"e3a3b609-3625-4670-b669-d7fd07386be6\") " pod="openshift-marketplace/community-operators-bwznz" Mar 13 14:42:34 crc kubenswrapper[4907]: I0313 14:42:34.362119 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e3a3b609-3625-4670-b669-d7fd07386be6-catalog-content\") pod \"community-operators-bwznz\" (UID: \"e3a3b609-3625-4670-b669-d7fd07386be6\") " pod="openshift-marketplace/community-operators-bwznz" Mar 13 14:42:34 crc kubenswrapper[4907]: I0313 14:42:34.362244 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e3a3b609-3625-4670-b669-d7fd07386be6-utilities\") pod \"community-operators-bwznz\" (UID: \"e3a3b609-3625-4670-b669-d7fd07386be6\") " pod="openshift-marketplace/community-operators-bwznz" Mar 13 14:42:34 crc kubenswrapper[4907]: I0313 14:42:34.384745 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rxdf7\" (UniqueName: \"kubernetes.io/projected/e3a3b609-3625-4670-b669-d7fd07386be6-kube-api-access-rxdf7\") pod \"community-operators-bwznz\" (UID: \"e3a3b609-3625-4670-b669-d7fd07386be6\") " pod="openshift-marketplace/community-operators-bwznz" Mar 13 14:42:34 crc kubenswrapper[4907]: I0313 14:42:34.526109 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bwznz" Mar 13 14:42:35 crc kubenswrapper[4907]: I0313 14:42:35.009946 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-bwznz"] Mar 13 14:42:35 crc kubenswrapper[4907]: I0313 14:42:35.267729 4907 generic.go:334] "Generic (PLEG): container finished" podID="e3a3b609-3625-4670-b669-d7fd07386be6" containerID="d1ed90bc0d96aa6d2be8063d734ff2e5ee2f08e2f5f3cf16acdb090807cbe152" exitCode=0 Mar 13 14:42:35 crc kubenswrapper[4907]: I0313 14:42:35.267798 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bwznz" event={"ID":"e3a3b609-3625-4670-b669-d7fd07386be6","Type":"ContainerDied","Data":"d1ed90bc0d96aa6d2be8063d734ff2e5ee2f08e2f5f3cf16acdb090807cbe152"} Mar 13 14:42:35 crc kubenswrapper[4907]: I0313 14:42:35.268031 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bwznz" event={"ID":"e3a3b609-3625-4670-b669-d7fd07386be6","Type":"ContainerStarted","Data":"b12c47624fc9c120d0c9e10ec66f5d5bf123a990a2f6d121f6b346f5dcde3df9"} Mar 13 14:42:39 crc kubenswrapper[4907]: I0313 14:42:39.292992 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bwznz" event={"ID":"e3a3b609-3625-4670-b669-d7fd07386be6","Type":"ContainerStarted","Data":"e9afb226fd8c064f22333c56811e1ee9b967bd4ba350cccbeba6a963b90b37da"} Mar 13 14:42:40 crc kubenswrapper[4907]: I0313 14:42:40.301007 4907 generic.go:334] "Generic (PLEG): container finished" podID="e3a3b609-3625-4670-b669-d7fd07386be6" containerID="e9afb226fd8c064f22333c56811e1ee9b967bd4ba350cccbeba6a963b90b37da" exitCode=0 Mar 13 14:42:40 crc kubenswrapper[4907]: I0313 14:42:40.301047 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bwznz" event={"ID":"e3a3b609-3625-4670-b669-d7fd07386be6","Type":"ContainerDied","Data":"e9afb226fd8c064f22333c56811e1ee9b967bd4ba350cccbeba6a963b90b37da"} Mar 13 14:42:41 crc kubenswrapper[4907]: I0313 14:42:41.309417 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bwznz" event={"ID":"e3a3b609-3625-4670-b669-d7fd07386be6","Type":"ContainerStarted","Data":"0c1015ce501c5ee89732127ca1e7eceae8fce9d0539ede2b5801447664bc10f4"} Mar 13 14:42:41 crc kubenswrapper[4907]: I0313 14:42:41.327842 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-bwznz" podStartSLOduration=1.858644712 podStartE2EDuration="7.327826578s" podCreationTimestamp="2026-03-13 14:42:34 +0000 UTC" firstStartedPulling="2026-03-13 14:42:35.269381382 +0000 UTC m=+2254.169169071" lastFinishedPulling="2026-03-13 14:42:40.738563248 +0000 UTC m=+2259.638350937" observedRunningTime="2026-03-13 14:42:41.327536479 +0000 UTC m=+2260.227324178" watchObservedRunningTime="2026-03-13 14:42:41.327826578 +0000 UTC m=+2260.227614267" Mar 13 14:42:44 crc kubenswrapper[4907]: I0313 14:42:44.527232 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-bwznz" Mar 13 14:42:44 crc kubenswrapper[4907]: I0313 14:42:44.527521 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-bwznz" Mar 13 14:42:44 crc kubenswrapper[4907]: I0313 14:42:44.584706 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-bwznz" Mar 13 14:42:45 crc kubenswrapper[4907]: I0313 14:42:45.387671 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-bwznz" Mar 13 14:42:45 crc kubenswrapper[4907]: I0313 14:42:45.476956 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-bwznz"] Mar 13 14:42:45 crc kubenswrapper[4907]: I0313 14:42:45.504076 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-bkqkq"] Mar 13 14:42:45 crc kubenswrapper[4907]: I0313 14:42:45.507982 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-bkqkq" podUID="8f55641f-96d1-4abe-b29e-35c03423c86f" containerName="registry-server" containerID="cri-o://dd30544066f21dab70ce804b909280851ba1f1bd4d4cb32479861077957a9178" gracePeriod=2 Mar 13 14:42:45 crc kubenswrapper[4907]: I0313 14:42:45.910760 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bkqkq" Mar 13 14:42:45 crc kubenswrapper[4907]: I0313 14:42:45.918353 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f55641f-96d1-4abe-b29e-35c03423c86f-utilities\") pod \"8f55641f-96d1-4abe-b29e-35c03423c86f\" (UID: \"8f55641f-96d1-4abe-b29e-35c03423c86f\") " Mar 13 14:42:45 crc kubenswrapper[4907]: I0313 14:42:45.918422 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f55641f-96d1-4abe-b29e-35c03423c86f-catalog-content\") pod \"8f55641f-96d1-4abe-b29e-35c03423c86f\" (UID: \"8f55641f-96d1-4abe-b29e-35c03423c86f\") " Mar 13 14:42:45 crc kubenswrapper[4907]: I0313 14:42:45.918601 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-blsqh\" (UniqueName: \"kubernetes.io/projected/8f55641f-96d1-4abe-b29e-35c03423c86f-kube-api-access-blsqh\") pod \"8f55641f-96d1-4abe-b29e-35c03423c86f\" (UID: \"8f55641f-96d1-4abe-b29e-35c03423c86f\") " Mar 13 14:42:45 crc kubenswrapper[4907]: I0313 14:42:45.920543 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f55641f-96d1-4abe-b29e-35c03423c86f-utilities" (OuterVolumeSpecName: "utilities") pod "8f55641f-96d1-4abe-b29e-35c03423c86f" (UID: "8f55641f-96d1-4abe-b29e-35c03423c86f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:42:45 crc kubenswrapper[4907]: I0313 14:42:45.931841 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f55641f-96d1-4abe-b29e-35c03423c86f-kube-api-access-blsqh" (OuterVolumeSpecName: "kube-api-access-blsqh") pod "8f55641f-96d1-4abe-b29e-35c03423c86f" (UID: "8f55641f-96d1-4abe-b29e-35c03423c86f"). InnerVolumeSpecName "kube-api-access-blsqh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:42:45 crc kubenswrapper[4907]: I0313 14:42:45.965725 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f55641f-96d1-4abe-b29e-35c03423c86f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8f55641f-96d1-4abe-b29e-35c03423c86f" (UID: "8f55641f-96d1-4abe-b29e-35c03423c86f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:42:46 crc kubenswrapper[4907]: I0313 14:42:46.019796 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-blsqh\" (UniqueName: \"kubernetes.io/projected/8f55641f-96d1-4abe-b29e-35c03423c86f-kube-api-access-blsqh\") on node \"crc\" DevicePath \"\"" Mar 13 14:42:46 crc kubenswrapper[4907]: I0313 14:42:46.019956 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f55641f-96d1-4abe-b29e-35c03423c86f-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 14:42:46 crc kubenswrapper[4907]: I0313 14:42:46.020021 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f55641f-96d1-4abe-b29e-35c03423c86f-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 14:42:46 crc kubenswrapper[4907]: I0313 14:42:46.344764 4907 generic.go:334] "Generic (PLEG): container finished" podID="8f55641f-96d1-4abe-b29e-35c03423c86f" containerID="dd30544066f21dab70ce804b909280851ba1f1bd4d4cb32479861077957a9178" exitCode=0 Mar 13 14:42:46 crc kubenswrapper[4907]: I0313 14:42:46.344836 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bkqkq" Mar 13 14:42:46 crc kubenswrapper[4907]: I0313 14:42:46.344868 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bkqkq" event={"ID":"8f55641f-96d1-4abe-b29e-35c03423c86f","Type":"ContainerDied","Data":"dd30544066f21dab70ce804b909280851ba1f1bd4d4cb32479861077957a9178"} Mar 13 14:42:46 crc kubenswrapper[4907]: I0313 14:42:46.344928 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bkqkq" event={"ID":"8f55641f-96d1-4abe-b29e-35c03423c86f","Type":"ContainerDied","Data":"51741992329c7cbf452d56e06ea4b68fc6d3bd2347af300a47210ccc94035b08"} Mar 13 14:42:46 crc kubenswrapper[4907]: I0313 14:42:46.344949 4907 scope.go:117] "RemoveContainer" containerID="dd30544066f21dab70ce804b909280851ba1f1bd4d4cb32479861077957a9178" Mar 13 14:42:46 crc kubenswrapper[4907]: I0313 14:42:46.372832 4907 scope.go:117] "RemoveContainer" containerID="7fb98f669ca95b8966acee50a69bfe3ca16749a7b12337396bf39c2c672c9654" Mar 13 14:42:46 crc kubenswrapper[4907]: I0313 14:42:46.384304 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-bkqkq"] Mar 13 14:42:46 crc kubenswrapper[4907]: I0313 14:42:46.389389 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-bkqkq"] Mar 13 14:42:46 crc kubenswrapper[4907]: I0313 14:42:46.398252 4907 scope.go:117] "RemoveContainer" containerID="4a29b590b6d8237b3616d05e54dcdd32085b2501383cd2a01ceaf05da4c45946" Mar 13 14:42:46 crc kubenswrapper[4907]: I0313 14:42:46.428739 4907 scope.go:117] "RemoveContainer" containerID="dd30544066f21dab70ce804b909280851ba1f1bd4d4cb32479861077957a9178" Mar 13 14:42:46 crc kubenswrapper[4907]: E0313 14:42:46.429097 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dd30544066f21dab70ce804b909280851ba1f1bd4d4cb32479861077957a9178\": container with ID starting with dd30544066f21dab70ce804b909280851ba1f1bd4d4cb32479861077957a9178 not found: ID does not exist" containerID="dd30544066f21dab70ce804b909280851ba1f1bd4d4cb32479861077957a9178" Mar 13 14:42:46 crc kubenswrapper[4907]: I0313 14:42:46.429127 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd30544066f21dab70ce804b909280851ba1f1bd4d4cb32479861077957a9178"} err="failed to get container status \"dd30544066f21dab70ce804b909280851ba1f1bd4d4cb32479861077957a9178\": rpc error: code = NotFound desc = could not find container \"dd30544066f21dab70ce804b909280851ba1f1bd4d4cb32479861077957a9178\": container with ID starting with dd30544066f21dab70ce804b909280851ba1f1bd4d4cb32479861077957a9178 not found: ID does not exist" Mar 13 14:42:46 crc kubenswrapper[4907]: I0313 14:42:46.429148 4907 scope.go:117] "RemoveContainer" containerID="7fb98f669ca95b8966acee50a69bfe3ca16749a7b12337396bf39c2c672c9654" Mar 13 14:42:46 crc kubenswrapper[4907]: E0313 14:42:46.429343 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7fb98f669ca95b8966acee50a69bfe3ca16749a7b12337396bf39c2c672c9654\": container with ID starting with 7fb98f669ca95b8966acee50a69bfe3ca16749a7b12337396bf39c2c672c9654 not found: ID does not exist" containerID="7fb98f669ca95b8966acee50a69bfe3ca16749a7b12337396bf39c2c672c9654" Mar 13 14:42:46 crc kubenswrapper[4907]: I0313 14:42:46.429363 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7fb98f669ca95b8966acee50a69bfe3ca16749a7b12337396bf39c2c672c9654"} err="failed to get container status \"7fb98f669ca95b8966acee50a69bfe3ca16749a7b12337396bf39c2c672c9654\": rpc error: code = NotFound desc = could not find container \"7fb98f669ca95b8966acee50a69bfe3ca16749a7b12337396bf39c2c672c9654\": container with ID starting with 7fb98f669ca95b8966acee50a69bfe3ca16749a7b12337396bf39c2c672c9654 not found: ID does not exist" Mar 13 14:42:46 crc kubenswrapper[4907]: I0313 14:42:46.429394 4907 scope.go:117] "RemoveContainer" containerID="4a29b590b6d8237b3616d05e54dcdd32085b2501383cd2a01ceaf05da4c45946" Mar 13 14:42:46 crc kubenswrapper[4907]: E0313 14:42:46.429604 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4a29b590b6d8237b3616d05e54dcdd32085b2501383cd2a01ceaf05da4c45946\": container with ID starting with 4a29b590b6d8237b3616d05e54dcdd32085b2501383cd2a01ceaf05da4c45946 not found: ID does not exist" containerID="4a29b590b6d8237b3616d05e54dcdd32085b2501383cd2a01ceaf05da4c45946" Mar 13 14:42:46 crc kubenswrapper[4907]: I0313 14:42:46.429625 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a29b590b6d8237b3616d05e54dcdd32085b2501383cd2a01ceaf05da4c45946"} err="failed to get container status \"4a29b590b6d8237b3616d05e54dcdd32085b2501383cd2a01ceaf05da4c45946\": rpc error: code = NotFound desc = could not find container \"4a29b590b6d8237b3616d05e54dcdd32085b2501383cd2a01ceaf05da4c45946\": container with ID starting with 4a29b590b6d8237b3616d05e54dcdd32085b2501383cd2a01ceaf05da4c45946 not found: ID does not exist" Mar 13 14:42:47 crc kubenswrapper[4907]: I0313 14:42:47.792463 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f55641f-96d1-4abe-b29e-35c03423c86f" path="/var/lib/kubelet/pods/8f55641f-96d1-4abe-b29e-35c03423c86f/volumes" Mar 13 14:42:48 crc kubenswrapper[4907]: I0313 14:42:48.042404 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 14:42:48 crc kubenswrapper[4907]: I0313 14:42:48.042485 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 14:43:18 crc kubenswrapper[4907]: I0313 14:43:18.041364 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 14:43:18 crc kubenswrapper[4907]: I0313 14:43:18.042951 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 14:43:48 crc kubenswrapper[4907]: I0313 14:43:48.042035 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 14:43:48 crc kubenswrapper[4907]: I0313 14:43:48.044054 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 14:43:48 crc kubenswrapper[4907]: I0313 14:43:48.044142 4907 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" Mar 13 14:43:48 crc kubenswrapper[4907]: I0313 14:43:48.045197 4907 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"43810caaacd8c4e6d986aec4d7996e802ba4907ce25cbd96e3f26a101537186b"} pod="openshift-machine-config-operator/machine-config-daemon-hm56j" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 14:43:48 crc kubenswrapper[4907]: I0313 14:43:48.045288 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" containerID="cri-o://43810caaacd8c4e6d986aec4d7996e802ba4907ce25cbd96e3f26a101537186b" gracePeriod=600 Mar 13 14:43:48 crc kubenswrapper[4907]: E0313 14:43:48.184178 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:43:48 crc kubenswrapper[4907]: I0313 14:43:48.857141 4907 generic.go:334] "Generic (PLEG): container finished" podID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerID="43810caaacd8c4e6d986aec4d7996e802ba4907ce25cbd96e3f26a101537186b" exitCode=0 Mar 13 14:43:48 crc kubenswrapper[4907]: I0313 14:43:48.857194 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerDied","Data":"43810caaacd8c4e6d986aec4d7996e802ba4907ce25cbd96e3f26a101537186b"} Mar 13 14:43:48 crc kubenswrapper[4907]: I0313 14:43:48.857241 4907 scope.go:117] "RemoveContainer" containerID="3c1970952dcd249074d61677266bef44d3d78378f555ffc20b89b00a033811dd" Mar 13 14:43:48 crc kubenswrapper[4907]: I0313 14:43:48.857782 4907 scope.go:117] "RemoveContainer" containerID="43810caaacd8c4e6d986aec4d7996e802ba4907ce25cbd96e3f26a101537186b" Mar 13 14:43:48 crc kubenswrapper[4907]: E0313 14:43:48.858076 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:44:00 crc kubenswrapper[4907]: I0313 14:44:00.135507 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556884-zzm6t"] Mar 13 14:44:00 crc kubenswrapper[4907]: E0313 14:44:00.138904 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f55641f-96d1-4abe-b29e-35c03423c86f" containerName="extract-utilities" Mar 13 14:44:00 crc kubenswrapper[4907]: I0313 14:44:00.138986 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f55641f-96d1-4abe-b29e-35c03423c86f" containerName="extract-utilities" Mar 13 14:44:00 crc kubenswrapper[4907]: E0313 14:44:00.139014 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f55641f-96d1-4abe-b29e-35c03423c86f" containerName="extract-content" Mar 13 14:44:00 crc kubenswrapper[4907]: I0313 14:44:00.139037 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f55641f-96d1-4abe-b29e-35c03423c86f" containerName="extract-content" Mar 13 14:44:00 crc kubenswrapper[4907]: E0313 14:44:00.139050 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f55641f-96d1-4abe-b29e-35c03423c86f" containerName="registry-server" Mar 13 14:44:00 crc kubenswrapper[4907]: I0313 14:44:00.139059 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f55641f-96d1-4abe-b29e-35c03423c86f" containerName="registry-server" Mar 13 14:44:00 crc kubenswrapper[4907]: I0313 14:44:00.139316 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f55641f-96d1-4abe-b29e-35c03423c86f" containerName="registry-server" Mar 13 14:44:00 crc kubenswrapper[4907]: I0313 14:44:00.140507 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556884-zzm6t" Mar 13 14:44:00 crc kubenswrapper[4907]: I0313 14:44:00.146683 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 14:44:00 crc kubenswrapper[4907]: I0313 14:44:00.146732 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 14:44:00 crc kubenswrapper[4907]: I0313 14:44:00.146761 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 14:44:00 crc kubenswrapper[4907]: I0313 14:44:00.147726 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556884-zzm6t"] Mar 13 14:44:00 crc kubenswrapper[4907]: I0313 14:44:00.243562 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b92lk\" (UniqueName: \"kubernetes.io/projected/bd8a84e4-33ca-4236-8c28-bf80cf64b6bc-kube-api-access-b92lk\") pod \"auto-csr-approver-29556884-zzm6t\" (UID: \"bd8a84e4-33ca-4236-8c28-bf80cf64b6bc\") " pod="openshift-infra/auto-csr-approver-29556884-zzm6t" Mar 13 14:44:00 crc kubenswrapper[4907]: I0313 14:44:00.345008 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b92lk\" (UniqueName: \"kubernetes.io/projected/bd8a84e4-33ca-4236-8c28-bf80cf64b6bc-kube-api-access-b92lk\") pod \"auto-csr-approver-29556884-zzm6t\" (UID: \"bd8a84e4-33ca-4236-8c28-bf80cf64b6bc\") " pod="openshift-infra/auto-csr-approver-29556884-zzm6t" Mar 13 14:44:00 crc kubenswrapper[4907]: I0313 14:44:00.362170 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b92lk\" (UniqueName: \"kubernetes.io/projected/bd8a84e4-33ca-4236-8c28-bf80cf64b6bc-kube-api-access-b92lk\") pod \"auto-csr-approver-29556884-zzm6t\" (UID: \"bd8a84e4-33ca-4236-8c28-bf80cf64b6bc\") " pod="openshift-infra/auto-csr-approver-29556884-zzm6t" Mar 13 14:44:00 crc kubenswrapper[4907]: I0313 14:44:00.462598 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556884-zzm6t" Mar 13 14:44:00 crc kubenswrapper[4907]: I0313 14:44:00.738264 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556884-zzm6t"] Mar 13 14:44:00 crc kubenswrapper[4907]: I0313 14:44:00.749714 4907 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 13 14:44:00 crc kubenswrapper[4907]: I0313 14:44:00.782129 4907 scope.go:117] "RemoveContainer" containerID="43810caaacd8c4e6d986aec4d7996e802ba4907ce25cbd96e3f26a101537186b" Mar 13 14:44:00 crc kubenswrapper[4907]: E0313 14:44:00.782352 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:44:00 crc kubenswrapper[4907]: I0313 14:44:00.963149 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556884-zzm6t" event={"ID":"bd8a84e4-33ca-4236-8c28-bf80cf64b6bc","Type":"ContainerStarted","Data":"2083bf0f478da887209b5b2d38e198b775cc56d66e0058fae942f158d67a7dda"} Mar 13 14:44:02 crc kubenswrapper[4907]: I0313 14:44:02.981476 4907 generic.go:334] "Generic (PLEG): container finished" podID="bd8a84e4-33ca-4236-8c28-bf80cf64b6bc" containerID="1d99f9504e5b5c3d1b261df414bc1841776b381d4074946ac4e7f45b3d386557" exitCode=0 Mar 13 14:44:02 crc kubenswrapper[4907]: I0313 14:44:02.981556 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556884-zzm6t" event={"ID":"bd8a84e4-33ca-4236-8c28-bf80cf64b6bc","Type":"ContainerDied","Data":"1d99f9504e5b5c3d1b261df414bc1841776b381d4074946ac4e7f45b3d386557"} Mar 13 14:44:04 crc kubenswrapper[4907]: I0313 14:44:04.349765 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556884-zzm6t" Mar 13 14:44:04 crc kubenswrapper[4907]: I0313 14:44:04.512123 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b92lk\" (UniqueName: \"kubernetes.io/projected/bd8a84e4-33ca-4236-8c28-bf80cf64b6bc-kube-api-access-b92lk\") pod \"bd8a84e4-33ca-4236-8c28-bf80cf64b6bc\" (UID: \"bd8a84e4-33ca-4236-8c28-bf80cf64b6bc\") " Mar 13 14:44:04 crc kubenswrapper[4907]: I0313 14:44:04.518505 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd8a84e4-33ca-4236-8c28-bf80cf64b6bc-kube-api-access-b92lk" (OuterVolumeSpecName: "kube-api-access-b92lk") pod "bd8a84e4-33ca-4236-8c28-bf80cf64b6bc" (UID: "bd8a84e4-33ca-4236-8c28-bf80cf64b6bc"). InnerVolumeSpecName "kube-api-access-b92lk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:44:04 crc kubenswrapper[4907]: I0313 14:44:04.613958 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b92lk\" (UniqueName: \"kubernetes.io/projected/bd8a84e4-33ca-4236-8c28-bf80cf64b6bc-kube-api-access-b92lk\") on node \"crc\" DevicePath \"\"" Mar 13 14:44:04 crc kubenswrapper[4907]: I0313 14:44:04.996951 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556884-zzm6t" event={"ID":"bd8a84e4-33ca-4236-8c28-bf80cf64b6bc","Type":"ContainerDied","Data":"2083bf0f478da887209b5b2d38e198b775cc56d66e0058fae942f158d67a7dda"} Mar 13 14:44:04 crc kubenswrapper[4907]: I0313 14:44:04.997301 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2083bf0f478da887209b5b2d38e198b775cc56d66e0058fae942f158d67a7dda" Mar 13 14:44:04 crc kubenswrapper[4907]: I0313 14:44:04.997450 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556884-zzm6t" Mar 13 14:44:05 crc kubenswrapper[4907]: I0313 14:44:05.423366 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556878-fqsf6"] Mar 13 14:44:05 crc kubenswrapper[4907]: I0313 14:44:05.431138 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556878-fqsf6"] Mar 13 14:44:05 crc kubenswrapper[4907]: I0313 14:44:05.796478 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ec5b140c-a271-4a8e-a9af-8e7c5bddcad3" path="/var/lib/kubelet/pods/ec5b140c-a271-4a8e-a9af-8e7c5bddcad3/volumes" Mar 13 14:44:15 crc kubenswrapper[4907]: I0313 14:44:15.783134 4907 scope.go:117] "RemoveContainer" containerID="43810caaacd8c4e6d986aec4d7996e802ba4907ce25cbd96e3f26a101537186b" Mar 13 14:44:15 crc kubenswrapper[4907]: E0313 14:44:15.784138 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:44:25 crc kubenswrapper[4907]: I0313 14:44:25.061723 4907 scope.go:117] "RemoveContainer" containerID="3d950127b4213b5d0d64f24ed2cfdf8b4f35b40aa5e7072728ac45679d7fc695" Mar 13 14:44:29 crc kubenswrapper[4907]: I0313 14:44:29.782381 4907 scope.go:117] "RemoveContainer" containerID="43810caaacd8c4e6d986aec4d7996e802ba4907ce25cbd96e3f26a101537186b" Mar 13 14:44:29 crc kubenswrapper[4907]: E0313 14:44:29.783118 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:44:43 crc kubenswrapper[4907]: I0313 14:44:43.782717 4907 scope.go:117] "RemoveContainer" containerID="43810caaacd8c4e6d986aec4d7996e802ba4907ce25cbd96e3f26a101537186b" Mar 13 14:44:43 crc kubenswrapper[4907]: E0313 14:44:43.783564 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:44:54 crc kubenswrapper[4907]: I0313 14:44:54.783183 4907 scope.go:117] "RemoveContainer" containerID="43810caaacd8c4e6d986aec4d7996e802ba4907ce25cbd96e3f26a101537186b" Mar 13 14:44:54 crc kubenswrapper[4907]: E0313 14:44:54.784353 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:45:00 crc kubenswrapper[4907]: I0313 14:45:00.160436 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556885-9xsw6"] Mar 13 14:45:00 crc kubenswrapper[4907]: E0313 14:45:00.161010 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd8a84e4-33ca-4236-8c28-bf80cf64b6bc" containerName="oc" Mar 13 14:45:00 crc kubenswrapper[4907]: I0313 14:45:00.161022 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd8a84e4-33ca-4236-8c28-bf80cf64b6bc" containerName="oc" Mar 13 14:45:00 crc kubenswrapper[4907]: I0313 14:45:00.161183 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd8a84e4-33ca-4236-8c28-bf80cf64b6bc" containerName="oc" Mar 13 14:45:00 crc kubenswrapper[4907]: I0313 14:45:00.161642 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556885-9xsw6" Mar 13 14:45:00 crc kubenswrapper[4907]: I0313 14:45:00.163581 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 13 14:45:00 crc kubenswrapper[4907]: I0313 14:45:00.167538 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 13 14:45:00 crc kubenswrapper[4907]: I0313 14:45:00.175911 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556885-9xsw6"] Mar 13 14:45:00 crc kubenswrapper[4907]: I0313 14:45:00.238320 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d79a5eae-26dc-4f69-aa63-9bd6c5788a03-secret-volume\") pod \"collect-profiles-29556885-9xsw6\" (UID: \"d79a5eae-26dc-4f69-aa63-9bd6c5788a03\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556885-9xsw6" Mar 13 14:45:00 crc kubenswrapper[4907]: I0313 14:45:00.238407 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d79a5eae-26dc-4f69-aa63-9bd6c5788a03-config-volume\") pod \"collect-profiles-29556885-9xsw6\" (UID: \"d79a5eae-26dc-4f69-aa63-9bd6c5788a03\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556885-9xsw6" Mar 13 14:45:00 crc kubenswrapper[4907]: I0313 14:45:00.238497 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-456sb\" (UniqueName: \"kubernetes.io/projected/d79a5eae-26dc-4f69-aa63-9bd6c5788a03-kube-api-access-456sb\") pod \"collect-profiles-29556885-9xsw6\" (UID: \"d79a5eae-26dc-4f69-aa63-9bd6c5788a03\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556885-9xsw6" Mar 13 14:45:00 crc kubenswrapper[4907]: I0313 14:45:00.340095 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d79a5eae-26dc-4f69-aa63-9bd6c5788a03-secret-volume\") pod \"collect-profiles-29556885-9xsw6\" (UID: \"d79a5eae-26dc-4f69-aa63-9bd6c5788a03\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556885-9xsw6" Mar 13 14:45:00 crc kubenswrapper[4907]: I0313 14:45:00.340168 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d79a5eae-26dc-4f69-aa63-9bd6c5788a03-config-volume\") pod \"collect-profiles-29556885-9xsw6\" (UID: \"d79a5eae-26dc-4f69-aa63-9bd6c5788a03\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556885-9xsw6" Mar 13 14:45:00 crc kubenswrapper[4907]: I0313 14:45:00.340205 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-456sb\" (UniqueName: \"kubernetes.io/projected/d79a5eae-26dc-4f69-aa63-9bd6c5788a03-kube-api-access-456sb\") pod \"collect-profiles-29556885-9xsw6\" (UID: \"d79a5eae-26dc-4f69-aa63-9bd6c5788a03\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556885-9xsw6" Mar 13 14:45:00 crc kubenswrapper[4907]: I0313 14:45:00.341215 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d79a5eae-26dc-4f69-aa63-9bd6c5788a03-config-volume\") pod \"collect-profiles-29556885-9xsw6\" (UID: \"d79a5eae-26dc-4f69-aa63-9bd6c5788a03\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556885-9xsw6" Mar 13 14:45:00 crc kubenswrapper[4907]: I0313 14:45:00.348768 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d79a5eae-26dc-4f69-aa63-9bd6c5788a03-secret-volume\") pod \"collect-profiles-29556885-9xsw6\" (UID: \"d79a5eae-26dc-4f69-aa63-9bd6c5788a03\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556885-9xsw6" Mar 13 14:45:00 crc kubenswrapper[4907]: I0313 14:45:00.359314 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-456sb\" (UniqueName: \"kubernetes.io/projected/d79a5eae-26dc-4f69-aa63-9bd6c5788a03-kube-api-access-456sb\") pod \"collect-profiles-29556885-9xsw6\" (UID: \"d79a5eae-26dc-4f69-aa63-9bd6c5788a03\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556885-9xsw6" Mar 13 14:45:00 crc kubenswrapper[4907]: I0313 14:45:00.479035 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556885-9xsw6" Mar 13 14:45:00 crc kubenswrapper[4907]: I0313 14:45:00.912955 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556885-9xsw6"] Mar 13 14:45:01 crc kubenswrapper[4907]: I0313 14:45:01.451951 4907 generic.go:334] "Generic (PLEG): container finished" podID="d79a5eae-26dc-4f69-aa63-9bd6c5788a03" containerID="34346eaf9fc976e698b78600d581550e67e1b840574df19f7129d1f617293a7d" exitCode=0 Mar 13 14:45:01 crc kubenswrapper[4907]: I0313 14:45:01.451993 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556885-9xsw6" event={"ID":"d79a5eae-26dc-4f69-aa63-9bd6c5788a03","Type":"ContainerDied","Data":"34346eaf9fc976e698b78600d581550e67e1b840574df19f7129d1f617293a7d"} Mar 13 14:45:01 crc kubenswrapper[4907]: I0313 14:45:01.452017 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556885-9xsw6" event={"ID":"d79a5eae-26dc-4f69-aa63-9bd6c5788a03","Type":"ContainerStarted","Data":"500440f870c9906ee7451693b9eaf24fd71405810b18783bcc0bedd8966b5ea2"} Mar 13 14:45:02 crc kubenswrapper[4907]: I0313 14:45:02.828507 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556885-9xsw6" Mar 13 14:45:02 crc kubenswrapper[4907]: I0313 14:45:02.878387 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-456sb\" (UniqueName: \"kubernetes.io/projected/d79a5eae-26dc-4f69-aa63-9bd6c5788a03-kube-api-access-456sb\") pod \"d79a5eae-26dc-4f69-aa63-9bd6c5788a03\" (UID: \"d79a5eae-26dc-4f69-aa63-9bd6c5788a03\") " Mar 13 14:45:02 crc kubenswrapper[4907]: I0313 14:45:02.878437 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d79a5eae-26dc-4f69-aa63-9bd6c5788a03-config-volume\") pod \"d79a5eae-26dc-4f69-aa63-9bd6c5788a03\" (UID: \"d79a5eae-26dc-4f69-aa63-9bd6c5788a03\") " Mar 13 14:45:02 crc kubenswrapper[4907]: I0313 14:45:02.878507 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d79a5eae-26dc-4f69-aa63-9bd6c5788a03-secret-volume\") pod \"d79a5eae-26dc-4f69-aa63-9bd6c5788a03\" (UID: \"d79a5eae-26dc-4f69-aa63-9bd6c5788a03\") " Mar 13 14:45:02 crc kubenswrapper[4907]: I0313 14:45:02.879565 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d79a5eae-26dc-4f69-aa63-9bd6c5788a03-config-volume" (OuterVolumeSpecName: "config-volume") pod "d79a5eae-26dc-4f69-aa63-9bd6c5788a03" (UID: "d79a5eae-26dc-4f69-aa63-9bd6c5788a03"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:45:02 crc kubenswrapper[4907]: I0313 14:45:02.890237 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d79a5eae-26dc-4f69-aa63-9bd6c5788a03-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "d79a5eae-26dc-4f69-aa63-9bd6c5788a03" (UID: "d79a5eae-26dc-4f69-aa63-9bd6c5788a03"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:45:02 crc kubenswrapper[4907]: I0313 14:45:02.890591 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d79a5eae-26dc-4f69-aa63-9bd6c5788a03-kube-api-access-456sb" (OuterVolumeSpecName: "kube-api-access-456sb") pod "d79a5eae-26dc-4f69-aa63-9bd6c5788a03" (UID: "d79a5eae-26dc-4f69-aa63-9bd6c5788a03"). InnerVolumeSpecName "kube-api-access-456sb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:45:02 crc kubenswrapper[4907]: I0313 14:45:02.980661 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-456sb\" (UniqueName: \"kubernetes.io/projected/d79a5eae-26dc-4f69-aa63-9bd6c5788a03-kube-api-access-456sb\") on node \"crc\" DevicePath \"\"" Mar 13 14:45:02 crc kubenswrapper[4907]: I0313 14:45:02.980693 4907 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d79a5eae-26dc-4f69-aa63-9bd6c5788a03-config-volume\") on node \"crc\" DevicePath \"\"" Mar 13 14:45:02 crc kubenswrapper[4907]: I0313 14:45:02.980702 4907 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d79a5eae-26dc-4f69-aa63-9bd6c5788a03-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 13 14:45:03 crc kubenswrapper[4907]: I0313 14:45:03.471728 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556885-9xsw6" event={"ID":"d79a5eae-26dc-4f69-aa63-9bd6c5788a03","Type":"ContainerDied","Data":"500440f870c9906ee7451693b9eaf24fd71405810b18783bcc0bedd8966b5ea2"} Mar 13 14:45:03 crc kubenswrapper[4907]: I0313 14:45:03.471787 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="500440f870c9906ee7451693b9eaf24fd71405810b18783bcc0bedd8966b5ea2" Mar 13 14:45:03 crc kubenswrapper[4907]: I0313 14:45:03.471847 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556885-9xsw6" Mar 13 14:45:03 crc kubenswrapper[4907]: I0313 14:45:03.914705 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556840-pgmwj"] Mar 13 14:45:03 crc kubenswrapper[4907]: I0313 14:45:03.922269 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556840-pgmwj"] Mar 13 14:45:05 crc kubenswrapper[4907]: I0313 14:45:05.793098 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="19edf37b-a1d3-4c7c-945e-5d0b8bb601bf" path="/var/lib/kubelet/pods/19edf37b-a1d3-4c7c-945e-5d0b8bb601bf/volumes" Mar 13 14:45:07 crc kubenswrapper[4907]: I0313 14:45:07.783410 4907 scope.go:117] "RemoveContainer" containerID="43810caaacd8c4e6d986aec4d7996e802ba4907ce25cbd96e3f26a101537186b" Mar 13 14:45:07 crc kubenswrapper[4907]: E0313 14:45:07.784489 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:45:19 crc kubenswrapper[4907]: I0313 14:45:19.782353 4907 scope.go:117] "RemoveContainer" containerID="43810caaacd8c4e6d986aec4d7996e802ba4907ce25cbd96e3f26a101537186b" Mar 13 14:45:19 crc kubenswrapper[4907]: E0313 14:45:19.783056 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:45:25 crc kubenswrapper[4907]: I0313 14:45:25.136823 4907 scope.go:117] "RemoveContainer" containerID="d2284227e021b37d9e8d2812f8621bebf16efd8eeb5019b51cfa28773e7a5aa9" Mar 13 14:45:31 crc kubenswrapper[4907]: I0313 14:45:31.786706 4907 scope.go:117] "RemoveContainer" containerID="43810caaacd8c4e6d986aec4d7996e802ba4907ce25cbd96e3f26a101537186b" Mar 13 14:45:31 crc kubenswrapper[4907]: E0313 14:45:31.787460 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:45:44 crc kubenswrapper[4907]: I0313 14:45:44.782639 4907 scope.go:117] "RemoveContainer" containerID="43810caaacd8c4e6d986aec4d7996e802ba4907ce25cbd96e3f26a101537186b" Mar 13 14:45:44 crc kubenswrapper[4907]: E0313 14:45:44.783580 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:45:58 crc kubenswrapper[4907]: I0313 14:45:58.782384 4907 scope.go:117] "RemoveContainer" containerID="43810caaacd8c4e6d986aec4d7996e802ba4907ce25cbd96e3f26a101537186b" Mar 13 14:45:58 crc kubenswrapper[4907]: E0313 14:45:58.783092 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:46:00 crc kubenswrapper[4907]: I0313 14:46:00.151783 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556886-tqfp7"] Mar 13 14:46:00 crc kubenswrapper[4907]: E0313 14:46:00.153507 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d79a5eae-26dc-4f69-aa63-9bd6c5788a03" containerName="collect-profiles" Mar 13 14:46:00 crc kubenswrapper[4907]: I0313 14:46:00.153625 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="d79a5eae-26dc-4f69-aa63-9bd6c5788a03" containerName="collect-profiles" Mar 13 14:46:00 crc kubenswrapper[4907]: I0313 14:46:00.153926 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="d79a5eae-26dc-4f69-aa63-9bd6c5788a03" containerName="collect-profiles" Mar 13 14:46:00 crc kubenswrapper[4907]: I0313 14:46:00.154589 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556886-tqfp7" Mar 13 14:46:00 crc kubenswrapper[4907]: I0313 14:46:00.158302 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 14:46:00 crc kubenswrapper[4907]: I0313 14:46:00.159689 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 14:46:00 crc kubenswrapper[4907]: I0313 14:46:00.160484 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 14:46:00 crc kubenswrapper[4907]: I0313 14:46:00.162585 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556886-tqfp7"] Mar 13 14:46:00 crc kubenswrapper[4907]: I0313 14:46:00.223127 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9skvj\" (UniqueName: \"kubernetes.io/projected/7e68a473-ca00-4a39-bdca-1a122010d02f-kube-api-access-9skvj\") pod \"auto-csr-approver-29556886-tqfp7\" (UID: \"7e68a473-ca00-4a39-bdca-1a122010d02f\") " pod="openshift-infra/auto-csr-approver-29556886-tqfp7" Mar 13 14:46:00 crc kubenswrapper[4907]: I0313 14:46:00.324175 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9skvj\" (UniqueName: \"kubernetes.io/projected/7e68a473-ca00-4a39-bdca-1a122010d02f-kube-api-access-9skvj\") pod \"auto-csr-approver-29556886-tqfp7\" (UID: \"7e68a473-ca00-4a39-bdca-1a122010d02f\") " pod="openshift-infra/auto-csr-approver-29556886-tqfp7" Mar 13 14:46:00 crc kubenswrapper[4907]: I0313 14:46:00.342164 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9skvj\" (UniqueName: \"kubernetes.io/projected/7e68a473-ca00-4a39-bdca-1a122010d02f-kube-api-access-9skvj\") pod \"auto-csr-approver-29556886-tqfp7\" (UID: \"7e68a473-ca00-4a39-bdca-1a122010d02f\") " pod="openshift-infra/auto-csr-approver-29556886-tqfp7" Mar 13 14:46:00 crc kubenswrapper[4907]: I0313 14:46:00.477541 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556886-tqfp7" Mar 13 14:46:00 crc kubenswrapper[4907]: I0313 14:46:00.916290 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556886-tqfp7"] Mar 13 14:46:00 crc kubenswrapper[4907]: I0313 14:46:00.948589 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556886-tqfp7" event={"ID":"7e68a473-ca00-4a39-bdca-1a122010d02f","Type":"ContainerStarted","Data":"e8d00dfed9d39506586f930b123b132eb8946a22941081dff69f3cd7065de584"} Mar 13 14:46:02 crc kubenswrapper[4907]: I0313 14:46:02.963012 4907 generic.go:334] "Generic (PLEG): container finished" podID="7e68a473-ca00-4a39-bdca-1a122010d02f" containerID="07d959b411798c04dbd7ce18d3d50f7352525fb8aa854268d1f201958d491e3c" exitCode=0 Mar 13 14:46:02 crc kubenswrapper[4907]: I0313 14:46:02.963276 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556886-tqfp7" event={"ID":"7e68a473-ca00-4a39-bdca-1a122010d02f","Type":"ContainerDied","Data":"07d959b411798c04dbd7ce18d3d50f7352525fb8aa854268d1f201958d491e3c"} Mar 13 14:46:04 crc kubenswrapper[4907]: I0313 14:46:04.227117 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556886-tqfp7" Mar 13 14:46:04 crc kubenswrapper[4907]: I0313 14:46:04.379303 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9skvj\" (UniqueName: \"kubernetes.io/projected/7e68a473-ca00-4a39-bdca-1a122010d02f-kube-api-access-9skvj\") pod \"7e68a473-ca00-4a39-bdca-1a122010d02f\" (UID: \"7e68a473-ca00-4a39-bdca-1a122010d02f\") " Mar 13 14:46:04 crc kubenswrapper[4907]: I0313 14:46:04.385627 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7e68a473-ca00-4a39-bdca-1a122010d02f-kube-api-access-9skvj" (OuterVolumeSpecName: "kube-api-access-9skvj") pod "7e68a473-ca00-4a39-bdca-1a122010d02f" (UID: "7e68a473-ca00-4a39-bdca-1a122010d02f"). InnerVolumeSpecName "kube-api-access-9skvj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:46:04 crc kubenswrapper[4907]: I0313 14:46:04.481999 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9skvj\" (UniqueName: \"kubernetes.io/projected/7e68a473-ca00-4a39-bdca-1a122010d02f-kube-api-access-9skvj\") on node \"crc\" DevicePath \"\"" Mar 13 14:46:04 crc kubenswrapper[4907]: I0313 14:46:04.978019 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556886-tqfp7" event={"ID":"7e68a473-ca00-4a39-bdca-1a122010d02f","Type":"ContainerDied","Data":"e8d00dfed9d39506586f930b123b132eb8946a22941081dff69f3cd7065de584"} Mar 13 14:46:04 crc kubenswrapper[4907]: I0313 14:46:04.978070 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e8d00dfed9d39506586f930b123b132eb8946a22941081dff69f3cd7065de584" Mar 13 14:46:04 crc kubenswrapper[4907]: I0313 14:46:04.978175 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556886-tqfp7" Mar 13 14:46:05 crc kubenswrapper[4907]: I0313 14:46:05.291527 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556880-wxdzr"] Mar 13 14:46:05 crc kubenswrapper[4907]: I0313 14:46:05.298216 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556880-wxdzr"] Mar 13 14:46:05 crc kubenswrapper[4907]: I0313 14:46:05.795046 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20dce4c7-7cbf-4e49-bd0e-93ee53d72551" path="/var/lib/kubelet/pods/20dce4c7-7cbf-4e49-bd0e-93ee53d72551/volumes" Mar 13 14:46:11 crc kubenswrapper[4907]: I0313 14:46:11.788621 4907 scope.go:117] "RemoveContainer" containerID="43810caaacd8c4e6d986aec4d7996e802ba4907ce25cbd96e3f26a101537186b" Mar 13 14:46:11 crc kubenswrapper[4907]: E0313 14:46:11.789287 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:46:25 crc kubenswrapper[4907]: I0313 14:46:25.185072 4907 scope.go:117] "RemoveContainer" containerID="fc46a881237631d2b078045b48ba1243ef9d84385c8036fc8ea4400ece996277" Mar 13 14:46:26 crc kubenswrapper[4907]: I0313 14:46:26.783275 4907 scope.go:117] "RemoveContainer" containerID="43810caaacd8c4e6d986aec4d7996e802ba4907ce25cbd96e3f26a101537186b" Mar 13 14:46:26 crc kubenswrapper[4907]: E0313 14:46:26.783787 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:46:39 crc kubenswrapper[4907]: I0313 14:46:39.782099 4907 scope.go:117] "RemoveContainer" containerID="43810caaacd8c4e6d986aec4d7996e802ba4907ce25cbd96e3f26a101537186b" Mar 13 14:46:39 crc kubenswrapper[4907]: E0313 14:46:39.782747 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:46:52 crc kubenswrapper[4907]: I0313 14:46:52.782982 4907 scope.go:117] "RemoveContainer" containerID="43810caaacd8c4e6d986aec4d7996e802ba4907ce25cbd96e3f26a101537186b" Mar 13 14:46:52 crc kubenswrapper[4907]: E0313 14:46:52.783860 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:47:07 crc kubenswrapper[4907]: I0313 14:47:07.782550 4907 scope.go:117] "RemoveContainer" containerID="43810caaacd8c4e6d986aec4d7996e802ba4907ce25cbd96e3f26a101537186b" Mar 13 14:47:07 crc kubenswrapper[4907]: E0313 14:47:07.783337 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:47:20 crc kubenswrapper[4907]: I0313 14:47:20.783424 4907 scope.go:117] "RemoveContainer" containerID="43810caaacd8c4e6d986aec4d7996e802ba4907ce25cbd96e3f26a101537186b" Mar 13 14:47:20 crc kubenswrapper[4907]: E0313 14:47:20.784704 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:47:35 crc kubenswrapper[4907]: I0313 14:47:35.782377 4907 scope.go:117] "RemoveContainer" containerID="43810caaacd8c4e6d986aec4d7996e802ba4907ce25cbd96e3f26a101537186b" Mar 13 14:47:35 crc kubenswrapper[4907]: E0313 14:47:35.783214 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:47:49 crc kubenswrapper[4907]: I0313 14:47:49.782185 4907 scope.go:117] "RemoveContainer" containerID="43810caaacd8c4e6d986aec4d7996e802ba4907ce25cbd96e3f26a101537186b" Mar 13 14:47:49 crc kubenswrapper[4907]: E0313 14:47:49.782819 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:48:00 crc kubenswrapper[4907]: I0313 14:48:00.168256 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556888-6dcv4"] Mar 13 14:48:00 crc kubenswrapper[4907]: E0313 14:48:00.169490 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e68a473-ca00-4a39-bdca-1a122010d02f" containerName="oc" Mar 13 14:48:00 crc kubenswrapper[4907]: I0313 14:48:00.169516 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e68a473-ca00-4a39-bdca-1a122010d02f" containerName="oc" Mar 13 14:48:00 crc kubenswrapper[4907]: I0313 14:48:00.169754 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="7e68a473-ca00-4a39-bdca-1a122010d02f" containerName="oc" Mar 13 14:48:00 crc kubenswrapper[4907]: I0313 14:48:00.170491 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556888-6dcv4" Mar 13 14:48:00 crc kubenswrapper[4907]: I0313 14:48:00.173197 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 14:48:00 crc kubenswrapper[4907]: I0313 14:48:00.173374 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 14:48:00 crc kubenswrapper[4907]: I0313 14:48:00.173809 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 14:48:00 crc kubenswrapper[4907]: I0313 14:48:00.184000 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556888-6dcv4"] Mar 13 14:48:00 crc kubenswrapper[4907]: I0313 14:48:00.334113 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n2nsj\" (UniqueName: \"kubernetes.io/projected/4b255770-f854-44fe-93fb-7ae5082081be-kube-api-access-n2nsj\") pod \"auto-csr-approver-29556888-6dcv4\" (UID: \"4b255770-f854-44fe-93fb-7ae5082081be\") " pod="openshift-infra/auto-csr-approver-29556888-6dcv4" Mar 13 14:48:00 crc kubenswrapper[4907]: I0313 14:48:00.436150 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n2nsj\" (UniqueName: \"kubernetes.io/projected/4b255770-f854-44fe-93fb-7ae5082081be-kube-api-access-n2nsj\") pod \"auto-csr-approver-29556888-6dcv4\" (UID: \"4b255770-f854-44fe-93fb-7ae5082081be\") " pod="openshift-infra/auto-csr-approver-29556888-6dcv4" Mar 13 14:48:00 crc kubenswrapper[4907]: I0313 14:48:00.455014 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n2nsj\" (UniqueName: \"kubernetes.io/projected/4b255770-f854-44fe-93fb-7ae5082081be-kube-api-access-n2nsj\") pod \"auto-csr-approver-29556888-6dcv4\" (UID: \"4b255770-f854-44fe-93fb-7ae5082081be\") " pod="openshift-infra/auto-csr-approver-29556888-6dcv4" Mar 13 14:48:00 crc kubenswrapper[4907]: I0313 14:48:00.498239 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556888-6dcv4" Mar 13 14:48:00 crc kubenswrapper[4907]: I0313 14:48:00.938822 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556888-6dcv4"] Mar 13 14:48:01 crc kubenswrapper[4907]: I0313 14:48:01.849551 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556888-6dcv4" event={"ID":"4b255770-f854-44fe-93fb-7ae5082081be","Type":"ContainerStarted","Data":"fdf6f7dd09a291140d72d172172a962dbe55255551fbcb81eaab9d7b5b3eb8a5"} Mar 13 14:48:02 crc kubenswrapper[4907]: I0313 14:48:02.859787 4907 generic.go:334] "Generic (PLEG): container finished" podID="4b255770-f854-44fe-93fb-7ae5082081be" containerID="33e7c832124ee1aa4196cb0bd674abcb6224ac6318050e8ed5dbb3afadecce6f" exitCode=0 Mar 13 14:48:02 crc kubenswrapper[4907]: I0313 14:48:02.859843 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556888-6dcv4" event={"ID":"4b255770-f854-44fe-93fb-7ae5082081be","Type":"ContainerDied","Data":"33e7c832124ee1aa4196cb0bd674abcb6224ac6318050e8ed5dbb3afadecce6f"} Mar 13 14:48:03 crc kubenswrapper[4907]: I0313 14:48:03.783383 4907 scope.go:117] "RemoveContainer" containerID="43810caaacd8c4e6d986aec4d7996e802ba4907ce25cbd96e3f26a101537186b" Mar 13 14:48:03 crc kubenswrapper[4907]: E0313 14:48:03.784009 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:48:04 crc kubenswrapper[4907]: I0313 14:48:04.135486 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556888-6dcv4" Mar 13 14:48:04 crc kubenswrapper[4907]: I0313 14:48:04.292582 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n2nsj\" (UniqueName: \"kubernetes.io/projected/4b255770-f854-44fe-93fb-7ae5082081be-kube-api-access-n2nsj\") pod \"4b255770-f854-44fe-93fb-7ae5082081be\" (UID: \"4b255770-f854-44fe-93fb-7ae5082081be\") " Mar 13 14:48:04 crc kubenswrapper[4907]: I0313 14:48:04.297608 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4b255770-f854-44fe-93fb-7ae5082081be-kube-api-access-n2nsj" (OuterVolumeSpecName: "kube-api-access-n2nsj") pod "4b255770-f854-44fe-93fb-7ae5082081be" (UID: "4b255770-f854-44fe-93fb-7ae5082081be"). InnerVolumeSpecName "kube-api-access-n2nsj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:48:04 crc kubenswrapper[4907]: I0313 14:48:04.394575 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n2nsj\" (UniqueName: \"kubernetes.io/projected/4b255770-f854-44fe-93fb-7ae5082081be-kube-api-access-n2nsj\") on node \"crc\" DevicePath \"\"" Mar 13 14:48:04 crc kubenswrapper[4907]: I0313 14:48:04.881079 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556888-6dcv4" event={"ID":"4b255770-f854-44fe-93fb-7ae5082081be","Type":"ContainerDied","Data":"fdf6f7dd09a291140d72d172172a962dbe55255551fbcb81eaab9d7b5b3eb8a5"} Mar 13 14:48:04 crc kubenswrapper[4907]: I0313 14:48:04.881156 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fdf6f7dd09a291140d72d172172a962dbe55255551fbcb81eaab9d7b5b3eb8a5" Mar 13 14:48:04 crc kubenswrapper[4907]: I0313 14:48:04.881162 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556888-6dcv4" Mar 13 14:48:05 crc kubenswrapper[4907]: I0313 14:48:05.208014 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556882-c7dc8"] Mar 13 14:48:05 crc kubenswrapper[4907]: I0313 14:48:05.215687 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556882-c7dc8"] Mar 13 14:48:05 crc kubenswrapper[4907]: I0313 14:48:05.798993 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="81733636-fc5e-4687-9ceb-0dd2e18fa15a" path="/var/lib/kubelet/pods/81733636-fc5e-4687-9ceb-0dd2e18fa15a/volumes" Mar 13 14:48:17 crc kubenswrapper[4907]: I0313 14:48:17.783576 4907 scope.go:117] "RemoveContainer" containerID="43810caaacd8c4e6d986aec4d7996e802ba4907ce25cbd96e3f26a101537186b" Mar 13 14:48:17 crc kubenswrapper[4907]: E0313 14:48:17.784245 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:48:25 crc kubenswrapper[4907]: I0313 14:48:25.278320 4907 scope.go:117] "RemoveContainer" containerID="888132585d34376894f6b13900aa0293c6191b881a23e9abf82e3ca297b39ec9" Mar 13 14:48:32 crc kubenswrapper[4907]: I0313 14:48:32.782579 4907 scope.go:117] "RemoveContainer" containerID="43810caaacd8c4e6d986aec4d7996e802ba4907ce25cbd96e3f26a101537186b" Mar 13 14:48:32 crc kubenswrapper[4907]: E0313 14:48:32.783280 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:48:44 crc kubenswrapper[4907]: I0313 14:48:44.782256 4907 scope.go:117] "RemoveContainer" containerID="43810caaacd8c4e6d986aec4d7996e802ba4907ce25cbd96e3f26a101537186b" Mar 13 14:48:44 crc kubenswrapper[4907]: E0313 14:48:44.782869 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:48:58 crc kubenswrapper[4907]: I0313 14:48:58.782772 4907 scope.go:117] "RemoveContainer" containerID="43810caaacd8c4e6d986aec4d7996e802ba4907ce25cbd96e3f26a101537186b" Mar 13 14:48:59 crc kubenswrapper[4907]: I0313 14:48:59.324808 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerStarted","Data":"5b16948da306cc82d0f15e426031c3bf4ecc434619902cfc2cfc82bb26ad6352"} Mar 13 14:50:00 crc kubenswrapper[4907]: I0313 14:50:00.145000 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556890-5kp94"] Mar 13 14:50:00 crc kubenswrapper[4907]: E0313 14:50:00.146633 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b255770-f854-44fe-93fb-7ae5082081be" containerName="oc" Mar 13 14:50:00 crc kubenswrapper[4907]: I0313 14:50:00.146658 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b255770-f854-44fe-93fb-7ae5082081be" containerName="oc" Mar 13 14:50:00 crc kubenswrapper[4907]: I0313 14:50:00.146913 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b255770-f854-44fe-93fb-7ae5082081be" containerName="oc" Mar 13 14:50:00 crc kubenswrapper[4907]: I0313 14:50:00.147861 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556890-5kp94" Mar 13 14:50:00 crc kubenswrapper[4907]: I0313 14:50:00.149944 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 14:50:00 crc kubenswrapper[4907]: I0313 14:50:00.152362 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 14:50:00 crc kubenswrapper[4907]: I0313 14:50:00.152581 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 14:50:00 crc kubenswrapper[4907]: I0313 14:50:00.156187 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556890-5kp94"] Mar 13 14:50:00 crc kubenswrapper[4907]: I0313 14:50:00.228847 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xn2vf\" (UniqueName: \"kubernetes.io/projected/d63f291e-b12e-4b0a-b4c3-e7b9c2cc3c40-kube-api-access-xn2vf\") pod \"auto-csr-approver-29556890-5kp94\" (UID: \"d63f291e-b12e-4b0a-b4c3-e7b9c2cc3c40\") " pod="openshift-infra/auto-csr-approver-29556890-5kp94" Mar 13 14:50:00 crc kubenswrapper[4907]: I0313 14:50:00.331031 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xn2vf\" (UniqueName: \"kubernetes.io/projected/d63f291e-b12e-4b0a-b4c3-e7b9c2cc3c40-kube-api-access-xn2vf\") pod \"auto-csr-approver-29556890-5kp94\" (UID: \"d63f291e-b12e-4b0a-b4c3-e7b9c2cc3c40\") " pod="openshift-infra/auto-csr-approver-29556890-5kp94" Mar 13 14:50:00 crc kubenswrapper[4907]: I0313 14:50:00.352163 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xn2vf\" (UniqueName: \"kubernetes.io/projected/d63f291e-b12e-4b0a-b4c3-e7b9c2cc3c40-kube-api-access-xn2vf\") pod \"auto-csr-approver-29556890-5kp94\" (UID: \"d63f291e-b12e-4b0a-b4c3-e7b9c2cc3c40\") " pod="openshift-infra/auto-csr-approver-29556890-5kp94" Mar 13 14:50:00 crc kubenswrapper[4907]: I0313 14:50:00.471982 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556890-5kp94" Mar 13 14:50:00 crc kubenswrapper[4907]: I0313 14:50:00.896038 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556890-5kp94"] Mar 13 14:50:00 crc kubenswrapper[4907]: I0313 14:50:00.904753 4907 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 13 14:50:01 crc kubenswrapper[4907]: I0313 14:50:01.819020 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556890-5kp94" event={"ID":"d63f291e-b12e-4b0a-b4c3-e7b9c2cc3c40","Type":"ContainerStarted","Data":"9ae7c3fb56c210bef178a4a1f6238ce90b42264a05c6b1a96392a1fe36274965"} Mar 13 14:50:03 crc kubenswrapper[4907]: I0313 14:50:03.840546 4907 generic.go:334] "Generic (PLEG): container finished" podID="d63f291e-b12e-4b0a-b4c3-e7b9c2cc3c40" containerID="f4cc0414786125661ff7aeac8cccd8627fe8bdb42c54577e769f1e1b046242f5" exitCode=0 Mar 13 14:50:03 crc kubenswrapper[4907]: I0313 14:50:03.840623 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556890-5kp94" event={"ID":"d63f291e-b12e-4b0a-b4c3-e7b9c2cc3c40","Type":"ContainerDied","Data":"f4cc0414786125661ff7aeac8cccd8627fe8bdb42c54577e769f1e1b046242f5"} Mar 13 14:50:05 crc kubenswrapper[4907]: I0313 14:50:05.225913 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556890-5kp94" Mar 13 14:50:05 crc kubenswrapper[4907]: I0313 14:50:05.309963 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xn2vf\" (UniqueName: \"kubernetes.io/projected/d63f291e-b12e-4b0a-b4c3-e7b9c2cc3c40-kube-api-access-xn2vf\") pod \"d63f291e-b12e-4b0a-b4c3-e7b9c2cc3c40\" (UID: \"d63f291e-b12e-4b0a-b4c3-e7b9c2cc3c40\") " Mar 13 14:50:05 crc kubenswrapper[4907]: I0313 14:50:05.315718 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d63f291e-b12e-4b0a-b4c3-e7b9c2cc3c40-kube-api-access-xn2vf" (OuterVolumeSpecName: "kube-api-access-xn2vf") pod "d63f291e-b12e-4b0a-b4c3-e7b9c2cc3c40" (UID: "d63f291e-b12e-4b0a-b4c3-e7b9c2cc3c40"). InnerVolumeSpecName "kube-api-access-xn2vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:50:05 crc kubenswrapper[4907]: I0313 14:50:05.411969 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xn2vf\" (UniqueName: \"kubernetes.io/projected/d63f291e-b12e-4b0a-b4c3-e7b9c2cc3c40-kube-api-access-xn2vf\") on node \"crc\" DevicePath \"\"" Mar 13 14:50:05 crc kubenswrapper[4907]: I0313 14:50:05.856443 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556890-5kp94" event={"ID":"d63f291e-b12e-4b0a-b4c3-e7b9c2cc3c40","Type":"ContainerDied","Data":"9ae7c3fb56c210bef178a4a1f6238ce90b42264a05c6b1a96392a1fe36274965"} Mar 13 14:50:05 crc kubenswrapper[4907]: I0313 14:50:05.856751 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9ae7c3fb56c210bef178a4a1f6238ce90b42264a05c6b1a96392a1fe36274965" Mar 13 14:50:05 crc kubenswrapper[4907]: I0313 14:50:05.856497 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556890-5kp94" Mar 13 14:50:06 crc kubenswrapper[4907]: I0313 14:50:06.302348 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556884-zzm6t"] Mar 13 14:50:06 crc kubenswrapper[4907]: I0313 14:50:06.307037 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556884-zzm6t"] Mar 13 14:50:07 crc kubenswrapper[4907]: I0313 14:50:07.793921 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd8a84e4-33ca-4236-8c28-bf80cf64b6bc" path="/var/lib/kubelet/pods/bd8a84e4-33ca-4236-8c28-bf80cf64b6bc/volumes" Mar 13 14:50:15 crc kubenswrapper[4907]: I0313 14:50:15.185876 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-tbj2h"] Mar 13 14:50:15 crc kubenswrapper[4907]: E0313 14:50:15.186671 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d63f291e-b12e-4b0a-b4c3-e7b9c2cc3c40" containerName="oc" Mar 13 14:50:15 crc kubenswrapper[4907]: I0313 14:50:15.186682 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="d63f291e-b12e-4b0a-b4c3-e7b9c2cc3c40" containerName="oc" Mar 13 14:50:15 crc kubenswrapper[4907]: I0313 14:50:15.186813 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="d63f291e-b12e-4b0a-b4c3-e7b9c2cc3c40" containerName="oc" Mar 13 14:50:15 crc kubenswrapper[4907]: I0313 14:50:15.187806 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tbj2h" Mar 13 14:50:15 crc kubenswrapper[4907]: I0313 14:50:15.192506 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-tbj2h"] Mar 13 14:50:15 crc kubenswrapper[4907]: I0313 14:50:15.266804 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b7f6fd9e-d07b-4339-97aa-50903ea7b614-utilities\") pod \"redhat-marketplace-tbj2h\" (UID: \"b7f6fd9e-d07b-4339-97aa-50903ea7b614\") " pod="openshift-marketplace/redhat-marketplace-tbj2h" Mar 13 14:50:15 crc kubenswrapper[4907]: I0313 14:50:15.267138 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kgqmz\" (UniqueName: \"kubernetes.io/projected/b7f6fd9e-d07b-4339-97aa-50903ea7b614-kube-api-access-kgqmz\") pod \"redhat-marketplace-tbj2h\" (UID: \"b7f6fd9e-d07b-4339-97aa-50903ea7b614\") " pod="openshift-marketplace/redhat-marketplace-tbj2h" Mar 13 14:50:15 crc kubenswrapper[4907]: I0313 14:50:15.267242 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b7f6fd9e-d07b-4339-97aa-50903ea7b614-catalog-content\") pod \"redhat-marketplace-tbj2h\" (UID: \"b7f6fd9e-d07b-4339-97aa-50903ea7b614\") " pod="openshift-marketplace/redhat-marketplace-tbj2h" Mar 13 14:50:15 crc kubenswrapper[4907]: I0313 14:50:15.368247 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b7f6fd9e-d07b-4339-97aa-50903ea7b614-utilities\") pod \"redhat-marketplace-tbj2h\" (UID: \"b7f6fd9e-d07b-4339-97aa-50903ea7b614\") " pod="openshift-marketplace/redhat-marketplace-tbj2h" Mar 13 14:50:15 crc kubenswrapper[4907]: I0313 14:50:15.368813 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b7f6fd9e-d07b-4339-97aa-50903ea7b614-utilities\") pod \"redhat-marketplace-tbj2h\" (UID: \"b7f6fd9e-d07b-4339-97aa-50903ea7b614\") " pod="openshift-marketplace/redhat-marketplace-tbj2h" Mar 13 14:50:15 crc kubenswrapper[4907]: I0313 14:50:15.369512 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kgqmz\" (UniqueName: \"kubernetes.io/projected/b7f6fd9e-d07b-4339-97aa-50903ea7b614-kube-api-access-kgqmz\") pod \"redhat-marketplace-tbj2h\" (UID: \"b7f6fd9e-d07b-4339-97aa-50903ea7b614\") " pod="openshift-marketplace/redhat-marketplace-tbj2h" Mar 13 14:50:15 crc kubenswrapper[4907]: I0313 14:50:15.370159 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b7f6fd9e-d07b-4339-97aa-50903ea7b614-catalog-content\") pod \"redhat-marketplace-tbj2h\" (UID: \"b7f6fd9e-d07b-4339-97aa-50903ea7b614\") " pod="openshift-marketplace/redhat-marketplace-tbj2h" Mar 13 14:50:15 crc kubenswrapper[4907]: I0313 14:50:15.370192 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b7f6fd9e-d07b-4339-97aa-50903ea7b614-catalog-content\") pod \"redhat-marketplace-tbj2h\" (UID: \"b7f6fd9e-d07b-4339-97aa-50903ea7b614\") " pod="openshift-marketplace/redhat-marketplace-tbj2h" Mar 13 14:50:15 crc kubenswrapper[4907]: I0313 14:50:15.389529 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kgqmz\" (UniqueName: \"kubernetes.io/projected/b7f6fd9e-d07b-4339-97aa-50903ea7b614-kube-api-access-kgqmz\") pod \"redhat-marketplace-tbj2h\" (UID: \"b7f6fd9e-d07b-4339-97aa-50903ea7b614\") " pod="openshift-marketplace/redhat-marketplace-tbj2h" Mar 13 14:50:15 crc kubenswrapper[4907]: I0313 14:50:15.502927 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tbj2h" Mar 13 14:50:15 crc kubenswrapper[4907]: I0313 14:50:15.916488 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-tbj2h"] Mar 13 14:50:15 crc kubenswrapper[4907]: I0313 14:50:15.938774 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tbj2h" event={"ID":"b7f6fd9e-d07b-4339-97aa-50903ea7b614","Type":"ContainerStarted","Data":"a49953d7498cf6c74e49e1a76f262c8cf7a2a5083aa06bb1bb21fd359058fb8f"} Mar 13 14:50:16 crc kubenswrapper[4907]: I0313 14:50:16.949643 4907 generic.go:334] "Generic (PLEG): container finished" podID="b7f6fd9e-d07b-4339-97aa-50903ea7b614" containerID="be8896282b4c21bb9015035503186642512b3e9f125e62a6c4d4cf624108a9f1" exitCode=0 Mar 13 14:50:16 crc kubenswrapper[4907]: I0313 14:50:16.949708 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tbj2h" event={"ID":"b7f6fd9e-d07b-4339-97aa-50903ea7b614","Type":"ContainerDied","Data":"be8896282b4c21bb9015035503186642512b3e9f125e62a6c4d4cf624108a9f1"} Mar 13 14:50:17 crc kubenswrapper[4907]: I0313 14:50:17.957237 4907 generic.go:334] "Generic (PLEG): container finished" podID="b7f6fd9e-d07b-4339-97aa-50903ea7b614" containerID="7e5377b54b517be40da3dfb2c0604762de65b6f2cdabc2cfb353510ffcd39ab0" exitCode=0 Mar 13 14:50:17 crc kubenswrapper[4907]: I0313 14:50:17.957277 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tbj2h" event={"ID":"b7f6fd9e-d07b-4339-97aa-50903ea7b614","Type":"ContainerDied","Data":"7e5377b54b517be40da3dfb2c0604762de65b6f2cdabc2cfb353510ffcd39ab0"} Mar 13 14:50:18 crc kubenswrapper[4907]: I0313 14:50:18.965619 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tbj2h" event={"ID":"b7f6fd9e-d07b-4339-97aa-50903ea7b614","Type":"ContainerStarted","Data":"88decf377cd617e46acde155df67a278293b44791d1045d2c952e59fa72f3b12"} Mar 13 14:50:18 crc kubenswrapper[4907]: I0313 14:50:18.986661 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-tbj2h" podStartSLOduration=2.576377226 podStartE2EDuration="3.986644117s" podCreationTimestamp="2026-03-13 14:50:15 +0000 UTC" firstStartedPulling="2026-03-13 14:50:16.953796667 +0000 UTC m=+2715.853584356" lastFinishedPulling="2026-03-13 14:50:18.364063518 +0000 UTC m=+2717.263851247" observedRunningTime="2026-03-13 14:50:18.982912665 +0000 UTC m=+2717.882700364" watchObservedRunningTime="2026-03-13 14:50:18.986644117 +0000 UTC m=+2717.886431806" Mar 13 14:50:25 crc kubenswrapper[4907]: I0313 14:50:25.388808 4907 scope.go:117] "RemoveContainer" containerID="1d99f9504e5b5c3d1b261df414bc1841776b381d4074946ac4e7f45b3d386557" Mar 13 14:50:25 crc kubenswrapper[4907]: I0313 14:50:25.503945 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-tbj2h" Mar 13 14:50:25 crc kubenswrapper[4907]: I0313 14:50:25.504557 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-tbj2h" Mar 13 14:50:25 crc kubenswrapper[4907]: I0313 14:50:25.545011 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-tbj2h" Mar 13 14:50:26 crc kubenswrapper[4907]: I0313 14:50:26.054306 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-tbj2h" Mar 13 14:50:26 crc kubenswrapper[4907]: I0313 14:50:26.096250 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-tbj2h"] Mar 13 14:50:28 crc kubenswrapper[4907]: I0313 14:50:28.036466 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-tbj2h" podUID="b7f6fd9e-d07b-4339-97aa-50903ea7b614" containerName="registry-server" containerID="cri-o://88decf377cd617e46acde155df67a278293b44791d1045d2c952e59fa72f3b12" gracePeriod=2 Mar 13 14:50:28 crc kubenswrapper[4907]: I0313 14:50:28.406285 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tbj2h" Mar 13 14:50:28 crc kubenswrapper[4907]: I0313 14:50:28.535449 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b7f6fd9e-d07b-4339-97aa-50903ea7b614-catalog-content\") pod \"b7f6fd9e-d07b-4339-97aa-50903ea7b614\" (UID: \"b7f6fd9e-d07b-4339-97aa-50903ea7b614\") " Mar 13 14:50:28 crc kubenswrapper[4907]: I0313 14:50:28.535595 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kgqmz\" (UniqueName: \"kubernetes.io/projected/b7f6fd9e-d07b-4339-97aa-50903ea7b614-kube-api-access-kgqmz\") pod \"b7f6fd9e-d07b-4339-97aa-50903ea7b614\" (UID: \"b7f6fd9e-d07b-4339-97aa-50903ea7b614\") " Mar 13 14:50:28 crc kubenswrapper[4907]: I0313 14:50:28.535677 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b7f6fd9e-d07b-4339-97aa-50903ea7b614-utilities\") pod \"b7f6fd9e-d07b-4339-97aa-50903ea7b614\" (UID: \"b7f6fd9e-d07b-4339-97aa-50903ea7b614\") " Mar 13 14:50:28 crc kubenswrapper[4907]: I0313 14:50:28.536969 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b7f6fd9e-d07b-4339-97aa-50903ea7b614-utilities" (OuterVolumeSpecName: "utilities") pod "b7f6fd9e-d07b-4339-97aa-50903ea7b614" (UID: "b7f6fd9e-d07b-4339-97aa-50903ea7b614"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:50:28 crc kubenswrapper[4907]: I0313 14:50:28.542729 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b7f6fd9e-d07b-4339-97aa-50903ea7b614-kube-api-access-kgqmz" (OuterVolumeSpecName: "kube-api-access-kgqmz") pod "b7f6fd9e-d07b-4339-97aa-50903ea7b614" (UID: "b7f6fd9e-d07b-4339-97aa-50903ea7b614"). InnerVolumeSpecName "kube-api-access-kgqmz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:50:28 crc kubenswrapper[4907]: I0313 14:50:28.571427 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b7f6fd9e-d07b-4339-97aa-50903ea7b614-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b7f6fd9e-d07b-4339-97aa-50903ea7b614" (UID: "b7f6fd9e-d07b-4339-97aa-50903ea7b614"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:50:28 crc kubenswrapper[4907]: I0313 14:50:28.637701 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kgqmz\" (UniqueName: \"kubernetes.io/projected/b7f6fd9e-d07b-4339-97aa-50903ea7b614-kube-api-access-kgqmz\") on node \"crc\" DevicePath \"\"" Mar 13 14:50:28 crc kubenswrapper[4907]: I0313 14:50:28.637732 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b7f6fd9e-d07b-4339-97aa-50903ea7b614-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 14:50:28 crc kubenswrapper[4907]: I0313 14:50:28.637742 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b7f6fd9e-d07b-4339-97aa-50903ea7b614-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 14:50:29 crc kubenswrapper[4907]: I0313 14:50:29.047061 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tbj2h" Mar 13 14:50:29 crc kubenswrapper[4907]: I0313 14:50:29.047102 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tbj2h" event={"ID":"b7f6fd9e-d07b-4339-97aa-50903ea7b614","Type":"ContainerDied","Data":"88decf377cd617e46acde155df67a278293b44791d1045d2c952e59fa72f3b12"} Mar 13 14:50:29 crc kubenswrapper[4907]: I0313 14:50:29.047231 4907 scope.go:117] "RemoveContainer" containerID="88decf377cd617e46acde155df67a278293b44791d1045d2c952e59fa72f3b12" Mar 13 14:50:29 crc kubenswrapper[4907]: I0313 14:50:29.046845 4907 generic.go:334] "Generic (PLEG): container finished" podID="b7f6fd9e-d07b-4339-97aa-50903ea7b614" containerID="88decf377cd617e46acde155df67a278293b44791d1045d2c952e59fa72f3b12" exitCode=0 Mar 13 14:50:29 crc kubenswrapper[4907]: I0313 14:50:29.056103 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tbj2h" event={"ID":"b7f6fd9e-d07b-4339-97aa-50903ea7b614","Type":"ContainerDied","Data":"a49953d7498cf6c74e49e1a76f262c8cf7a2a5083aa06bb1bb21fd359058fb8f"} Mar 13 14:50:29 crc kubenswrapper[4907]: I0313 14:50:29.084007 4907 scope.go:117] "RemoveContainer" containerID="7e5377b54b517be40da3dfb2c0604762de65b6f2cdabc2cfb353510ffcd39ab0" Mar 13 14:50:29 crc kubenswrapper[4907]: I0313 14:50:29.090534 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-tbj2h"] Mar 13 14:50:29 crc kubenswrapper[4907]: I0313 14:50:29.102775 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-tbj2h"] Mar 13 14:50:29 crc kubenswrapper[4907]: I0313 14:50:29.112047 4907 scope.go:117] "RemoveContainer" containerID="be8896282b4c21bb9015035503186642512b3e9f125e62a6c4d4cf624108a9f1" Mar 13 14:50:29 crc kubenswrapper[4907]: I0313 14:50:29.136613 4907 scope.go:117] "RemoveContainer" containerID="88decf377cd617e46acde155df67a278293b44791d1045d2c952e59fa72f3b12" Mar 13 14:50:29 crc kubenswrapper[4907]: E0313 14:50:29.137196 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"88decf377cd617e46acde155df67a278293b44791d1045d2c952e59fa72f3b12\": container with ID starting with 88decf377cd617e46acde155df67a278293b44791d1045d2c952e59fa72f3b12 not found: ID does not exist" containerID="88decf377cd617e46acde155df67a278293b44791d1045d2c952e59fa72f3b12" Mar 13 14:50:29 crc kubenswrapper[4907]: I0313 14:50:29.137240 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"88decf377cd617e46acde155df67a278293b44791d1045d2c952e59fa72f3b12"} err="failed to get container status \"88decf377cd617e46acde155df67a278293b44791d1045d2c952e59fa72f3b12\": rpc error: code = NotFound desc = could not find container \"88decf377cd617e46acde155df67a278293b44791d1045d2c952e59fa72f3b12\": container with ID starting with 88decf377cd617e46acde155df67a278293b44791d1045d2c952e59fa72f3b12 not found: ID does not exist" Mar 13 14:50:29 crc kubenswrapper[4907]: I0313 14:50:29.137268 4907 scope.go:117] "RemoveContainer" containerID="7e5377b54b517be40da3dfb2c0604762de65b6f2cdabc2cfb353510ffcd39ab0" Mar 13 14:50:29 crc kubenswrapper[4907]: E0313 14:50:29.138231 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7e5377b54b517be40da3dfb2c0604762de65b6f2cdabc2cfb353510ffcd39ab0\": container with ID starting with 7e5377b54b517be40da3dfb2c0604762de65b6f2cdabc2cfb353510ffcd39ab0 not found: ID does not exist" containerID="7e5377b54b517be40da3dfb2c0604762de65b6f2cdabc2cfb353510ffcd39ab0" Mar 13 14:50:29 crc kubenswrapper[4907]: I0313 14:50:29.138256 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7e5377b54b517be40da3dfb2c0604762de65b6f2cdabc2cfb353510ffcd39ab0"} err="failed to get container status \"7e5377b54b517be40da3dfb2c0604762de65b6f2cdabc2cfb353510ffcd39ab0\": rpc error: code = NotFound desc = could not find container \"7e5377b54b517be40da3dfb2c0604762de65b6f2cdabc2cfb353510ffcd39ab0\": container with ID starting with 7e5377b54b517be40da3dfb2c0604762de65b6f2cdabc2cfb353510ffcd39ab0 not found: ID does not exist" Mar 13 14:50:29 crc kubenswrapper[4907]: I0313 14:50:29.138270 4907 scope.go:117] "RemoveContainer" containerID="be8896282b4c21bb9015035503186642512b3e9f125e62a6c4d4cf624108a9f1" Mar 13 14:50:29 crc kubenswrapper[4907]: E0313 14:50:29.138706 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"be8896282b4c21bb9015035503186642512b3e9f125e62a6c4d4cf624108a9f1\": container with ID starting with be8896282b4c21bb9015035503186642512b3e9f125e62a6c4d4cf624108a9f1 not found: ID does not exist" containerID="be8896282b4c21bb9015035503186642512b3e9f125e62a6c4d4cf624108a9f1" Mar 13 14:50:29 crc kubenswrapper[4907]: I0313 14:50:29.138744 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"be8896282b4c21bb9015035503186642512b3e9f125e62a6c4d4cf624108a9f1"} err="failed to get container status \"be8896282b4c21bb9015035503186642512b3e9f125e62a6c4d4cf624108a9f1\": rpc error: code = NotFound desc = could not find container \"be8896282b4c21bb9015035503186642512b3e9f125e62a6c4d4cf624108a9f1\": container with ID starting with be8896282b4c21bb9015035503186642512b3e9f125e62a6c4d4cf624108a9f1 not found: ID does not exist" Mar 13 14:50:29 crc kubenswrapper[4907]: I0313 14:50:29.793653 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b7f6fd9e-d07b-4339-97aa-50903ea7b614" path="/var/lib/kubelet/pods/b7f6fd9e-d07b-4339-97aa-50903ea7b614/volumes" Mar 13 14:51:18 crc kubenswrapper[4907]: I0313 14:51:18.041500 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 14:51:18 crc kubenswrapper[4907]: I0313 14:51:18.042096 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 14:51:48 crc kubenswrapper[4907]: I0313 14:51:48.041255 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 14:51:48 crc kubenswrapper[4907]: I0313 14:51:48.041875 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 14:52:00 crc kubenswrapper[4907]: I0313 14:52:00.151860 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-hglxf"] Mar 13 14:52:00 crc kubenswrapper[4907]: E0313 14:52:00.152723 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7f6fd9e-d07b-4339-97aa-50903ea7b614" containerName="registry-server" Mar 13 14:52:00 crc kubenswrapper[4907]: I0313 14:52:00.152736 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7f6fd9e-d07b-4339-97aa-50903ea7b614" containerName="registry-server" Mar 13 14:52:00 crc kubenswrapper[4907]: E0313 14:52:00.152766 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7f6fd9e-d07b-4339-97aa-50903ea7b614" containerName="extract-utilities" Mar 13 14:52:00 crc kubenswrapper[4907]: I0313 14:52:00.152773 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7f6fd9e-d07b-4339-97aa-50903ea7b614" containerName="extract-utilities" Mar 13 14:52:00 crc kubenswrapper[4907]: E0313 14:52:00.152786 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7f6fd9e-d07b-4339-97aa-50903ea7b614" containerName="extract-content" Mar 13 14:52:00 crc kubenswrapper[4907]: I0313 14:52:00.152792 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7f6fd9e-d07b-4339-97aa-50903ea7b614" containerName="extract-content" Mar 13 14:52:00 crc kubenswrapper[4907]: I0313 14:52:00.152940 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="b7f6fd9e-d07b-4339-97aa-50903ea7b614" containerName="registry-server" Mar 13 14:52:00 crc kubenswrapper[4907]: I0313 14:52:00.153870 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hglxf" Mar 13 14:52:00 crc kubenswrapper[4907]: I0313 14:52:00.167549 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hglxf"] Mar 13 14:52:00 crc kubenswrapper[4907]: I0313 14:52:00.244372 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73e2ff11-5a7d-4fd9-9908-b57c774d4793-catalog-content\") pod \"certified-operators-hglxf\" (UID: \"73e2ff11-5a7d-4fd9-9908-b57c774d4793\") " pod="openshift-marketplace/certified-operators-hglxf" Mar 13 14:52:00 crc kubenswrapper[4907]: I0313 14:52:00.244458 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jhsvf\" (UniqueName: \"kubernetes.io/projected/73e2ff11-5a7d-4fd9-9908-b57c774d4793-kube-api-access-jhsvf\") pod \"certified-operators-hglxf\" (UID: \"73e2ff11-5a7d-4fd9-9908-b57c774d4793\") " pod="openshift-marketplace/certified-operators-hglxf" Mar 13 14:52:00 crc kubenswrapper[4907]: I0313 14:52:00.244801 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73e2ff11-5a7d-4fd9-9908-b57c774d4793-utilities\") pod \"certified-operators-hglxf\" (UID: \"73e2ff11-5a7d-4fd9-9908-b57c774d4793\") " pod="openshift-marketplace/certified-operators-hglxf" Mar 13 14:52:00 crc kubenswrapper[4907]: I0313 14:52:00.251463 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556892-gq5qv"] Mar 13 14:52:00 crc kubenswrapper[4907]: I0313 14:52:00.258943 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556892-gq5qv" Mar 13 14:52:00 crc kubenswrapper[4907]: I0313 14:52:00.260567 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556892-gq5qv"] Mar 13 14:52:00 crc kubenswrapper[4907]: I0313 14:52:00.261793 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 14:52:00 crc kubenswrapper[4907]: I0313 14:52:00.263166 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 14:52:00 crc kubenswrapper[4907]: I0313 14:52:00.264241 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 14:52:00 crc kubenswrapper[4907]: I0313 14:52:00.346828 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73e2ff11-5a7d-4fd9-9908-b57c774d4793-utilities\") pod \"certified-operators-hglxf\" (UID: \"73e2ff11-5a7d-4fd9-9908-b57c774d4793\") " pod="openshift-marketplace/certified-operators-hglxf" Mar 13 14:52:00 crc kubenswrapper[4907]: I0313 14:52:00.346961 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k8zbg\" (UniqueName: \"kubernetes.io/projected/cbb8b9ec-85bd-4617-952d-57ce0bdb4553-kube-api-access-k8zbg\") pod \"auto-csr-approver-29556892-gq5qv\" (UID: \"cbb8b9ec-85bd-4617-952d-57ce0bdb4553\") " pod="openshift-infra/auto-csr-approver-29556892-gq5qv" Mar 13 14:52:00 crc kubenswrapper[4907]: I0313 14:52:00.346997 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73e2ff11-5a7d-4fd9-9908-b57c774d4793-catalog-content\") pod \"certified-operators-hglxf\" (UID: \"73e2ff11-5a7d-4fd9-9908-b57c774d4793\") " pod="openshift-marketplace/certified-operators-hglxf" Mar 13 14:52:00 crc kubenswrapper[4907]: I0313 14:52:00.347023 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jhsvf\" (UniqueName: \"kubernetes.io/projected/73e2ff11-5a7d-4fd9-9908-b57c774d4793-kube-api-access-jhsvf\") pod \"certified-operators-hglxf\" (UID: \"73e2ff11-5a7d-4fd9-9908-b57c774d4793\") " pod="openshift-marketplace/certified-operators-hglxf" Mar 13 14:52:00 crc kubenswrapper[4907]: I0313 14:52:00.347495 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73e2ff11-5a7d-4fd9-9908-b57c774d4793-utilities\") pod \"certified-operators-hglxf\" (UID: \"73e2ff11-5a7d-4fd9-9908-b57c774d4793\") " pod="openshift-marketplace/certified-operators-hglxf" Mar 13 14:52:00 crc kubenswrapper[4907]: I0313 14:52:00.347547 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73e2ff11-5a7d-4fd9-9908-b57c774d4793-catalog-content\") pod \"certified-operators-hglxf\" (UID: \"73e2ff11-5a7d-4fd9-9908-b57c774d4793\") " pod="openshift-marketplace/certified-operators-hglxf" Mar 13 14:52:00 crc kubenswrapper[4907]: I0313 14:52:00.367841 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jhsvf\" (UniqueName: \"kubernetes.io/projected/73e2ff11-5a7d-4fd9-9908-b57c774d4793-kube-api-access-jhsvf\") pod \"certified-operators-hglxf\" (UID: \"73e2ff11-5a7d-4fd9-9908-b57c774d4793\") " pod="openshift-marketplace/certified-operators-hglxf" Mar 13 14:52:00 crc kubenswrapper[4907]: I0313 14:52:00.448856 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k8zbg\" (UniqueName: \"kubernetes.io/projected/cbb8b9ec-85bd-4617-952d-57ce0bdb4553-kube-api-access-k8zbg\") pod \"auto-csr-approver-29556892-gq5qv\" (UID: \"cbb8b9ec-85bd-4617-952d-57ce0bdb4553\") " pod="openshift-infra/auto-csr-approver-29556892-gq5qv" Mar 13 14:52:00 crc kubenswrapper[4907]: I0313 14:52:00.466994 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k8zbg\" (UniqueName: \"kubernetes.io/projected/cbb8b9ec-85bd-4617-952d-57ce0bdb4553-kube-api-access-k8zbg\") pod \"auto-csr-approver-29556892-gq5qv\" (UID: \"cbb8b9ec-85bd-4617-952d-57ce0bdb4553\") " pod="openshift-infra/auto-csr-approver-29556892-gq5qv" Mar 13 14:52:00 crc kubenswrapper[4907]: I0313 14:52:00.474925 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hglxf" Mar 13 14:52:00 crc kubenswrapper[4907]: I0313 14:52:00.575306 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556892-gq5qv" Mar 13 14:52:00 crc kubenswrapper[4907]: I0313 14:52:00.939988 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hglxf"] Mar 13 14:52:01 crc kubenswrapper[4907]: I0313 14:52:01.069426 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556892-gq5qv"] Mar 13 14:52:01 crc kubenswrapper[4907]: I0313 14:52:01.072290 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hglxf" event={"ID":"73e2ff11-5a7d-4fd9-9908-b57c774d4793","Type":"ContainerStarted","Data":"76918a7b1d9b781c78870a05cada882f98b668b545bb9ef8a4888d25a3d213a8"} Mar 13 14:52:01 crc kubenswrapper[4907]: W0313 14:52:01.084217 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcbb8b9ec_85bd_4617_952d_57ce0bdb4553.slice/crio-6540b5f3ac658f051ef622c91f789b6c6c1726d92e07753b794ac5b4496497aa WatchSource:0}: Error finding container 6540b5f3ac658f051ef622c91f789b6c6c1726d92e07753b794ac5b4496497aa: Status 404 returned error can't find the container with id 6540b5f3ac658f051ef622c91f789b6c6c1726d92e07753b794ac5b4496497aa Mar 13 14:52:02 crc kubenswrapper[4907]: I0313 14:52:02.080694 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556892-gq5qv" event={"ID":"cbb8b9ec-85bd-4617-952d-57ce0bdb4553","Type":"ContainerStarted","Data":"6540b5f3ac658f051ef622c91f789b6c6c1726d92e07753b794ac5b4496497aa"} Mar 13 14:52:02 crc kubenswrapper[4907]: I0313 14:52:02.082504 4907 generic.go:334] "Generic (PLEG): container finished" podID="73e2ff11-5a7d-4fd9-9908-b57c774d4793" containerID="1039a932888e0112ca2f48289a4503fb0f292131eefcd895d4539232ce58909c" exitCode=0 Mar 13 14:52:02 crc kubenswrapper[4907]: I0313 14:52:02.082859 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hglxf" event={"ID":"73e2ff11-5a7d-4fd9-9908-b57c774d4793","Type":"ContainerDied","Data":"1039a932888e0112ca2f48289a4503fb0f292131eefcd895d4539232ce58909c"} Mar 13 14:52:03 crc kubenswrapper[4907]: I0313 14:52:03.090237 4907 generic.go:334] "Generic (PLEG): container finished" podID="cbb8b9ec-85bd-4617-952d-57ce0bdb4553" containerID="2c73e4b83765c168348350266e6ee7d5565c27f33d1acbef663a50d336ac5e75" exitCode=0 Mar 13 14:52:03 crc kubenswrapper[4907]: I0313 14:52:03.090469 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556892-gq5qv" event={"ID":"cbb8b9ec-85bd-4617-952d-57ce0bdb4553","Type":"ContainerDied","Data":"2c73e4b83765c168348350266e6ee7d5565c27f33d1acbef663a50d336ac5e75"} Mar 13 14:52:03 crc kubenswrapper[4907]: I0313 14:52:03.093271 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hglxf" event={"ID":"73e2ff11-5a7d-4fd9-9908-b57c774d4793","Type":"ContainerStarted","Data":"69543302314ed4849971909b44e1e718b46042e5aea85023f93ed2a8109086e8"} Mar 13 14:52:04 crc kubenswrapper[4907]: I0313 14:52:04.101856 4907 generic.go:334] "Generic (PLEG): container finished" podID="73e2ff11-5a7d-4fd9-9908-b57c774d4793" containerID="69543302314ed4849971909b44e1e718b46042e5aea85023f93ed2a8109086e8" exitCode=0 Mar 13 14:52:04 crc kubenswrapper[4907]: I0313 14:52:04.101912 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hglxf" event={"ID":"73e2ff11-5a7d-4fd9-9908-b57c774d4793","Type":"ContainerDied","Data":"69543302314ed4849971909b44e1e718b46042e5aea85023f93ed2a8109086e8"} Mar 13 14:52:04 crc kubenswrapper[4907]: I0313 14:52:04.405837 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556892-gq5qv" Mar 13 14:52:04 crc kubenswrapper[4907]: I0313 14:52:04.504952 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k8zbg\" (UniqueName: \"kubernetes.io/projected/cbb8b9ec-85bd-4617-952d-57ce0bdb4553-kube-api-access-k8zbg\") pod \"cbb8b9ec-85bd-4617-952d-57ce0bdb4553\" (UID: \"cbb8b9ec-85bd-4617-952d-57ce0bdb4553\") " Mar 13 14:52:04 crc kubenswrapper[4907]: I0313 14:52:04.511675 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cbb8b9ec-85bd-4617-952d-57ce0bdb4553-kube-api-access-k8zbg" (OuterVolumeSpecName: "kube-api-access-k8zbg") pod "cbb8b9ec-85bd-4617-952d-57ce0bdb4553" (UID: "cbb8b9ec-85bd-4617-952d-57ce0bdb4553"). InnerVolumeSpecName "kube-api-access-k8zbg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:52:04 crc kubenswrapper[4907]: I0313 14:52:04.606456 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k8zbg\" (UniqueName: \"kubernetes.io/projected/cbb8b9ec-85bd-4617-952d-57ce0bdb4553-kube-api-access-k8zbg\") on node \"crc\" DevicePath \"\"" Mar 13 14:52:05 crc kubenswrapper[4907]: I0313 14:52:05.110320 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556892-gq5qv" Mar 13 14:52:05 crc kubenswrapper[4907]: I0313 14:52:05.110323 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556892-gq5qv" event={"ID":"cbb8b9ec-85bd-4617-952d-57ce0bdb4553","Type":"ContainerDied","Data":"6540b5f3ac658f051ef622c91f789b6c6c1726d92e07753b794ac5b4496497aa"} Mar 13 14:52:05 crc kubenswrapper[4907]: I0313 14:52:05.110693 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6540b5f3ac658f051ef622c91f789b6c6c1726d92e07753b794ac5b4496497aa" Mar 13 14:52:05 crc kubenswrapper[4907]: I0313 14:52:05.112219 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hglxf" event={"ID":"73e2ff11-5a7d-4fd9-9908-b57c774d4793","Type":"ContainerStarted","Data":"08451a948422d2c35b846b2a876db014a9cf925cdae25fd1738e4b8845d5de4e"} Mar 13 14:52:05 crc kubenswrapper[4907]: I0313 14:52:05.136316 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-hglxf" podStartSLOduration=2.495860144 podStartE2EDuration="5.136298253s" podCreationTimestamp="2026-03-13 14:52:00 +0000 UTC" firstStartedPulling="2026-03-13 14:52:02.084588612 +0000 UTC m=+2820.984376301" lastFinishedPulling="2026-03-13 14:52:04.725026721 +0000 UTC m=+2823.624814410" observedRunningTime="2026-03-13 14:52:05.132686735 +0000 UTC m=+2824.032474424" watchObservedRunningTime="2026-03-13 14:52:05.136298253 +0000 UTC m=+2824.036085942" Mar 13 14:52:05 crc kubenswrapper[4907]: I0313 14:52:05.462356 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556886-tqfp7"] Mar 13 14:52:05 crc kubenswrapper[4907]: I0313 14:52:05.468462 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556886-tqfp7"] Mar 13 14:52:05 crc kubenswrapper[4907]: I0313 14:52:05.790146 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7e68a473-ca00-4a39-bdca-1a122010d02f" path="/var/lib/kubelet/pods/7e68a473-ca00-4a39-bdca-1a122010d02f/volumes" Mar 13 14:52:10 crc kubenswrapper[4907]: I0313 14:52:10.475553 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-hglxf" Mar 13 14:52:10 crc kubenswrapper[4907]: I0313 14:52:10.476272 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-hglxf" Mar 13 14:52:10 crc kubenswrapper[4907]: I0313 14:52:10.521268 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-hglxf" Mar 13 14:52:11 crc kubenswrapper[4907]: I0313 14:52:11.200294 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-hglxf" Mar 13 14:52:11 crc kubenswrapper[4907]: I0313 14:52:11.250863 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-hglxf"] Mar 13 14:52:13 crc kubenswrapper[4907]: I0313 14:52:13.173327 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-hglxf" podUID="73e2ff11-5a7d-4fd9-9908-b57c774d4793" containerName="registry-server" containerID="cri-o://08451a948422d2c35b846b2a876db014a9cf925cdae25fd1738e4b8845d5de4e" gracePeriod=2 Mar 13 14:52:13 crc kubenswrapper[4907]: I0313 14:52:13.691548 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hglxf" Mar 13 14:52:13 crc kubenswrapper[4907]: I0313 14:52:13.736752 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhsvf\" (UniqueName: \"kubernetes.io/projected/73e2ff11-5a7d-4fd9-9908-b57c774d4793-kube-api-access-jhsvf\") pod \"73e2ff11-5a7d-4fd9-9908-b57c774d4793\" (UID: \"73e2ff11-5a7d-4fd9-9908-b57c774d4793\") " Mar 13 14:52:13 crc kubenswrapper[4907]: I0313 14:52:13.736862 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73e2ff11-5a7d-4fd9-9908-b57c774d4793-utilities\") pod \"73e2ff11-5a7d-4fd9-9908-b57c774d4793\" (UID: \"73e2ff11-5a7d-4fd9-9908-b57c774d4793\") " Mar 13 14:52:13 crc kubenswrapper[4907]: I0313 14:52:13.736990 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73e2ff11-5a7d-4fd9-9908-b57c774d4793-catalog-content\") pod \"73e2ff11-5a7d-4fd9-9908-b57c774d4793\" (UID: \"73e2ff11-5a7d-4fd9-9908-b57c774d4793\") " Mar 13 14:52:13 crc kubenswrapper[4907]: I0313 14:52:13.738115 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/73e2ff11-5a7d-4fd9-9908-b57c774d4793-utilities" (OuterVolumeSpecName: "utilities") pod "73e2ff11-5a7d-4fd9-9908-b57c774d4793" (UID: "73e2ff11-5a7d-4fd9-9908-b57c774d4793"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:52:13 crc kubenswrapper[4907]: I0313 14:52:13.744196 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/73e2ff11-5a7d-4fd9-9908-b57c774d4793-kube-api-access-jhsvf" (OuterVolumeSpecName: "kube-api-access-jhsvf") pod "73e2ff11-5a7d-4fd9-9908-b57c774d4793" (UID: "73e2ff11-5a7d-4fd9-9908-b57c774d4793"). InnerVolumeSpecName "kube-api-access-jhsvf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:52:13 crc kubenswrapper[4907]: I0313 14:52:13.809342 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/73e2ff11-5a7d-4fd9-9908-b57c774d4793-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "73e2ff11-5a7d-4fd9-9908-b57c774d4793" (UID: "73e2ff11-5a7d-4fd9-9908-b57c774d4793"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:52:13 crc kubenswrapper[4907]: I0313 14:52:13.838970 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73e2ff11-5a7d-4fd9-9908-b57c774d4793-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 14:52:13 crc kubenswrapper[4907]: I0313 14:52:13.839004 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhsvf\" (UniqueName: \"kubernetes.io/projected/73e2ff11-5a7d-4fd9-9908-b57c774d4793-kube-api-access-jhsvf\") on node \"crc\" DevicePath \"\"" Mar 13 14:52:13 crc kubenswrapper[4907]: I0313 14:52:13.839017 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73e2ff11-5a7d-4fd9-9908-b57c774d4793-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 14:52:14 crc kubenswrapper[4907]: I0313 14:52:14.185695 4907 generic.go:334] "Generic (PLEG): container finished" podID="73e2ff11-5a7d-4fd9-9908-b57c774d4793" containerID="08451a948422d2c35b846b2a876db014a9cf925cdae25fd1738e4b8845d5de4e" exitCode=0 Mar 13 14:52:14 crc kubenswrapper[4907]: I0313 14:52:14.185783 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hglxf" event={"ID":"73e2ff11-5a7d-4fd9-9908-b57c774d4793","Type":"ContainerDied","Data":"08451a948422d2c35b846b2a876db014a9cf925cdae25fd1738e4b8845d5de4e"} Mar 13 14:52:14 crc kubenswrapper[4907]: I0313 14:52:14.185839 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hglxf" event={"ID":"73e2ff11-5a7d-4fd9-9908-b57c774d4793","Type":"ContainerDied","Data":"76918a7b1d9b781c78870a05cada882f98b668b545bb9ef8a4888d25a3d213a8"} Mar 13 14:52:14 crc kubenswrapper[4907]: I0313 14:52:14.185923 4907 scope.go:117] "RemoveContainer" containerID="08451a948422d2c35b846b2a876db014a9cf925cdae25fd1738e4b8845d5de4e" Mar 13 14:52:14 crc kubenswrapper[4907]: I0313 14:52:14.186171 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hglxf" Mar 13 14:52:14 crc kubenswrapper[4907]: I0313 14:52:14.217835 4907 scope.go:117] "RemoveContainer" containerID="69543302314ed4849971909b44e1e718b46042e5aea85023f93ed2a8109086e8" Mar 13 14:52:14 crc kubenswrapper[4907]: I0313 14:52:14.231201 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-hglxf"] Mar 13 14:52:14 crc kubenswrapper[4907]: I0313 14:52:14.244090 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-hglxf"] Mar 13 14:52:14 crc kubenswrapper[4907]: I0313 14:52:14.248147 4907 scope.go:117] "RemoveContainer" containerID="1039a932888e0112ca2f48289a4503fb0f292131eefcd895d4539232ce58909c" Mar 13 14:52:14 crc kubenswrapper[4907]: I0313 14:52:14.269504 4907 scope.go:117] "RemoveContainer" containerID="08451a948422d2c35b846b2a876db014a9cf925cdae25fd1738e4b8845d5de4e" Mar 13 14:52:14 crc kubenswrapper[4907]: E0313 14:52:14.270009 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"08451a948422d2c35b846b2a876db014a9cf925cdae25fd1738e4b8845d5de4e\": container with ID starting with 08451a948422d2c35b846b2a876db014a9cf925cdae25fd1738e4b8845d5de4e not found: ID does not exist" containerID="08451a948422d2c35b846b2a876db014a9cf925cdae25fd1738e4b8845d5de4e" Mar 13 14:52:14 crc kubenswrapper[4907]: I0313 14:52:14.270059 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"08451a948422d2c35b846b2a876db014a9cf925cdae25fd1738e4b8845d5de4e"} err="failed to get container status \"08451a948422d2c35b846b2a876db014a9cf925cdae25fd1738e4b8845d5de4e\": rpc error: code = NotFound desc = could not find container \"08451a948422d2c35b846b2a876db014a9cf925cdae25fd1738e4b8845d5de4e\": container with ID starting with 08451a948422d2c35b846b2a876db014a9cf925cdae25fd1738e4b8845d5de4e not found: ID does not exist" Mar 13 14:52:14 crc kubenswrapper[4907]: I0313 14:52:14.270092 4907 scope.go:117] "RemoveContainer" containerID="69543302314ed4849971909b44e1e718b46042e5aea85023f93ed2a8109086e8" Mar 13 14:52:14 crc kubenswrapper[4907]: E0313 14:52:14.270531 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"69543302314ed4849971909b44e1e718b46042e5aea85023f93ed2a8109086e8\": container with ID starting with 69543302314ed4849971909b44e1e718b46042e5aea85023f93ed2a8109086e8 not found: ID does not exist" containerID="69543302314ed4849971909b44e1e718b46042e5aea85023f93ed2a8109086e8" Mar 13 14:52:14 crc kubenswrapper[4907]: I0313 14:52:14.270843 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"69543302314ed4849971909b44e1e718b46042e5aea85023f93ed2a8109086e8"} err="failed to get container status \"69543302314ed4849971909b44e1e718b46042e5aea85023f93ed2a8109086e8\": rpc error: code = NotFound desc = could not find container \"69543302314ed4849971909b44e1e718b46042e5aea85023f93ed2a8109086e8\": container with ID starting with 69543302314ed4849971909b44e1e718b46042e5aea85023f93ed2a8109086e8 not found: ID does not exist" Mar 13 14:52:14 crc kubenswrapper[4907]: I0313 14:52:14.270910 4907 scope.go:117] "RemoveContainer" containerID="1039a932888e0112ca2f48289a4503fb0f292131eefcd895d4539232ce58909c" Mar 13 14:52:14 crc kubenswrapper[4907]: E0313 14:52:14.271251 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1039a932888e0112ca2f48289a4503fb0f292131eefcd895d4539232ce58909c\": container with ID starting with 1039a932888e0112ca2f48289a4503fb0f292131eefcd895d4539232ce58909c not found: ID does not exist" containerID="1039a932888e0112ca2f48289a4503fb0f292131eefcd895d4539232ce58909c" Mar 13 14:52:14 crc kubenswrapper[4907]: I0313 14:52:14.271289 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1039a932888e0112ca2f48289a4503fb0f292131eefcd895d4539232ce58909c"} err="failed to get container status \"1039a932888e0112ca2f48289a4503fb0f292131eefcd895d4539232ce58909c\": rpc error: code = NotFound desc = could not find container \"1039a932888e0112ca2f48289a4503fb0f292131eefcd895d4539232ce58909c\": container with ID starting with 1039a932888e0112ca2f48289a4503fb0f292131eefcd895d4539232ce58909c not found: ID does not exist" Mar 13 14:52:15 crc kubenswrapper[4907]: I0313 14:52:15.811089 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="73e2ff11-5a7d-4fd9-9908-b57c774d4793" path="/var/lib/kubelet/pods/73e2ff11-5a7d-4fd9-9908-b57c774d4793/volumes" Mar 13 14:52:18 crc kubenswrapper[4907]: I0313 14:52:18.041728 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 14:52:18 crc kubenswrapper[4907]: I0313 14:52:18.042085 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 14:52:18 crc kubenswrapper[4907]: I0313 14:52:18.042167 4907 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" Mar 13 14:52:18 crc kubenswrapper[4907]: I0313 14:52:18.042865 4907 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5b16948da306cc82d0f15e426031c3bf4ecc434619902cfc2cfc82bb26ad6352"} pod="openshift-machine-config-operator/machine-config-daemon-hm56j" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 14:52:18 crc kubenswrapper[4907]: I0313 14:52:18.042958 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" containerID="cri-o://5b16948da306cc82d0f15e426031c3bf4ecc434619902cfc2cfc82bb26ad6352" gracePeriod=600 Mar 13 14:52:18 crc kubenswrapper[4907]: I0313 14:52:18.228785 4907 generic.go:334] "Generic (PLEG): container finished" podID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerID="5b16948da306cc82d0f15e426031c3bf4ecc434619902cfc2cfc82bb26ad6352" exitCode=0 Mar 13 14:52:18 crc kubenswrapper[4907]: I0313 14:52:18.228866 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerDied","Data":"5b16948da306cc82d0f15e426031c3bf4ecc434619902cfc2cfc82bb26ad6352"} Mar 13 14:52:18 crc kubenswrapper[4907]: I0313 14:52:18.228992 4907 scope.go:117] "RemoveContainer" containerID="43810caaacd8c4e6d986aec4d7996e802ba4907ce25cbd96e3f26a101537186b" Mar 13 14:52:19 crc kubenswrapper[4907]: I0313 14:52:19.240103 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerStarted","Data":"a44259e1e0a7edc8acdd625ba3f7415d70aa78b760d8eeb81254942722d3acb3"} Mar 13 14:52:25 crc kubenswrapper[4907]: I0313 14:52:25.472983 4907 scope.go:117] "RemoveContainer" containerID="07d959b411798c04dbd7ce18d3d50f7352525fb8aa854268d1f201958d491e3c" Mar 13 14:52:31 crc kubenswrapper[4907]: I0313 14:52:31.334772 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-s8srt"] Mar 13 14:52:31 crc kubenswrapper[4907]: E0313 14:52:31.335667 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cbb8b9ec-85bd-4617-952d-57ce0bdb4553" containerName="oc" Mar 13 14:52:31 crc kubenswrapper[4907]: I0313 14:52:31.335683 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="cbb8b9ec-85bd-4617-952d-57ce0bdb4553" containerName="oc" Mar 13 14:52:31 crc kubenswrapper[4907]: E0313 14:52:31.335697 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73e2ff11-5a7d-4fd9-9908-b57c774d4793" containerName="extract-content" Mar 13 14:52:31 crc kubenswrapper[4907]: I0313 14:52:31.335718 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="73e2ff11-5a7d-4fd9-9908-b57c774d4793" containerName="extract-content" Mar 13 14:52:31 crc kubenswrapper[4907]: E0313 14:52:31.335739 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73e2ff11-5a7d-4fd9-9908-b57c774d4793" containerName="extract-utilities" Mar 13 14:52:31 crc kubenswrapper[4907]: I0313 14:52:31.335748 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="73e2ff11-5a7d-4fd9-9908-b57c774d4793" containerName="extract-utilities" Mar 13 14:52:31 crc kubenswrapper[4907]: E0313 14:52:31.335759 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73e2ff11-5a7d-4fd9-9908-b57c774d4793" containerName="registry-server" Mar 13 14:52:31 crc kubenswrapper[4907]: I0313 14:52:31.335767 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="73e2ff11-5a7d-4fd9-9908-b57c774d4793" containerName="registry-server" Mar 13 14:52:31 crc kubenswrapper[4907]: I0313 14:52:31.335944 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="cbb8b9ec-85bd-4617-952d-57ce0bdb4553" containerName="oc" Mar 13 14:52:31 crc kubenswrapper[4907]: I0313 14:52:31.335978 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="73e2ff11-5a7d-4fd9-9908-b57c774d4793" containerName="registry-server" Mar 13 14:52:31 crc kubenswrapper[4907]: I0313 14:52:31.337148 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-s8srt" Mar 13 14:52:31 crc kubenswrapper[4907]: I0313 14:52:31.349540 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-s8srt"] Mar 13 14:52:31 crc kubenswrapper[4907]: I0313 14:52:31.478245 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a53a97c4-ae1d-4e67-b816-c1e4852b08e5-utilities\") pod \"redhat-operators-s8srt\" (UID: \"a53a97c4-ae1d-4e67-b816-c1e4852b08e5\") " pod="openshift-marketplace/redhat-operators-s8srt" Mar 13 14:52:31 crc kubenswrapper[4907]: I0313 14:52:31.478592 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-892qz\" (UniqueName: \"kubernetes.io/projected/a53a97c4-ae1d-4e67-b816-c1e4852b08e5-kube-api-access-892qz\") pod \"redhat-operators-s8srt\" (UID: \"a53a97c4-ae1d-4e67-b816-c1e4852b08e5\") " pod="openshift-marketplace/redhat-operators-s8srt" Mar 13 14:52:31 crc kubenswrapper[4907]: I0313 14:52:31.478804 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a53a97c4-ae1d-4e67-b816-c1e4852b08e5-catalog-content\") pod \"redhat-operators-s8srt\" (UID: \"a53a97c4-ae1d-4e67-b816-c1e4852b08e5\") " pod="openshift-marketplace/redhat-operators-s8srt" Mar 13 14:52:31 crc kubenswrapper[4907]: I0313 14:52:31.579693 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a53a97c4-ae1d-4e67-b816-c1e4852b08e5-utilities\") pod \"redhat-operators-s8srt\" (UID: \"a53a97c4-ae1d-4e67-b816-c1e4852b08e5\") " pod="openshift-marketplace/redhat-operators-s8srt" Mar 13 14:52:31 crc kubenswrapper[4907]: I0313 14:52:31.579750 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-892qz\" (UniqueName: \"kubernetes.io/projected/a53a97c4-ae1d-4e67-b816-c1e4852b08e5-kube-api-access-892qz\") pod \"redhat-operators-s8srt\" (UID: \"a53a97c4-ae1d-4e67-b816-c1e4852b08e5\") " pod="openshift-marketplace/redhat-operators-s8srt" Mar 13 14:52:31 crc kubenswrapper[4907]: I0313 14:52:31.579779 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a53a97c4-ae1d-4e67-b816-c1e4852b08e5-catalog-content\") pod \"redhat-operators-s8srt\" (UID: \"a53a97c4-ae1d-4e67-b816-c1e4852b08e5\") " pod="openshift-marketplace/redhat-operators-s8srt" Mar 13 14:52:31 crc kubenswrapper[4907]: I0313 14:52:31.580266 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a53a97c4-ae1d-4e67-b816-c1e4852b08e5-utilities\") pod \"redhat-operators-s8srt\" (UID: \"a53a97c4-ae1d-4e67-b816-c1e4852b08e5\") " pod="openshift-marketplace/redhat-operators-s8srt" Mar 13 14:52:31 crc kubenswrapper[4907]: I0313 14:52:31.580352 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a53a97c4-ae1d-4e67-b816-c1e4852b08e5-catalog-content\") pod \"redhat-operators-s8srt\" (UID: \"a53a97c4-ae1d-4e67-b816-c1e4852b08e5\") " pod="openshift-marketplace/redhat-operators-s8srt" Mar 13 14:52:31 crc kubenswrapper[4907]: I0313 14:52:31.609338 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-892qz\" (UniqueName: \"kubernetes.io/projected/a53a97c4-ae1d-4e67-b816-c1e4852b08e5-kube-api-access-892qz\") pod \"redhat-operators-s8srt\" (UID: \"a53a97c4-ae1d-4e67-b816-c1e4852b08e5\") " pod="openshift-marketplace/redhat-operators-s8srt" Mar 13 14:52:31 crc kubenswrapper[4907]: I0313 14:52:31.657689 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-s8srt" Mar 13 14:52:32 crc kubenswrapper[4907]: I0313 14:52:32.088895 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-s8srt"] Mar 13 14:52:32 crc kubenswrapper[4907]: W0313 14:52:32.100538 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda53a97c4_ae1d_4e67_b816_c1e4852b08e5.slice/crio-0fd20c97ede4d0b2cf316bf01e48438f62ec92b557262bb0310559909098c715 WatchSource:0}: Error finding container 0fd20c97ede4d0b2cf316bf01e48438f62ec92b557262bb0310559909098c715: Status 404 returned error can't find the container with id 0fd20c97ede4d0b2cf316bf01e48438f62ec92b557262bb0310559909098c715 Mar 13 14:52:32 crc kubenswrapper[4907]: I0313 14:52:32.331160 4907 generic.go:334] "Generic (PLEG): container finished" podID="a53a97c4-ae1d-4e67-b816-c1e4852b08e5" containerID="0b057665a0d00f3a48117344966f922b2928ae4d7a2f92b272a3107b565359be" exitCode=0 Mar 13 14:52:32 crc kubenswrapper[4907]: I0313 14:52:32.331245 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-s8srt" event={"ID":"a53a97c4-ae1d-4e67-b816-c1e4852b08e5","Type":"ContainerDied","Data":"0b057665a0d00f3a48117344966f922b2928ae4d7a2f92b272a3107b565359be"} Mar 13 14:52:32 crc kubenswrapper[4907]: I0313 14:52:32.331855 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-s8srt" event={"ID":"a53a97c4-ae1d-4e67-b816-c1e4852b08e5","Type":"ContainerStarted","Data":"0fd20c97ede4d0b2cf316bf01e48438f62ec92b557262bb0310559909098c715"} Mar 13 14:52:33 crc kubenswrapper[4907]: I0313 14:52:33.340476 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-s8srt" event={"ID":"a53a97c4-ae1d-4e67-b816-c1e4852b08e5","Type":"ContainerStarted","Data":"5f1bfe1233a28e84068fa3690e2155ac3883d0d7f36e3aac56ab833ab2614920"} Mar 13 14:52:34 crc kubenswrapper[4907]: I0313 14:52:34.348772 4907 generic.go:334] "Generic (PLEG): container finished" podID="a53a97c4-ae1d-4e67-b816-c1e4852b08e5" containerID="5f1bfe1233a28e84068fa3690e2155ac3883d0d7f36e3aac56ab833ab2614920" exitCode=0 Mar 13 14:52:34 crc kubenswrapper[4907]: I0313 14:52:34.348829 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-s8srt" event={"ID":"a53a97c4-ae1d-4e67-b816-c1e4852b08e5","Type":"ContainerDied","Data":"5f1bfe1233a28e84068fa3690e2155ac3883d0d7f36e3aac56ab833ab2614920"} Mar 13 14:52:36 crc kubenswrapper[4907]: I0313 14:52:36.364793 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-s8srt" event={"ID":"a53a97c4-ae1d-4e67-b816-c1e4852b08e5","Type":"ContainerStarted","Data":"a0f45585bd9bea343e63d05743c65bf9a3e6218aa8b37cbe283fae944e8b7e8d"} Mar 13 14:52:41 crc kubenswrapper[4907]: I0313 14:52:41.658611 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-s8srt" Mar 13 14:52:41 crc kubenswrapper[4907]: I0313 14:52:41.658979 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-s8srt" Mar 13 14:52:41 crc kubenswrapper[4907]: I0313 14:52:41.700549 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-s8srt" Mar 13 14:52:41 crc kubenswrapper[4907]: I0313 14:52:41.721655 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-s8srt" podStartSLOduration=7.649174469 podStartE2EDuration="10.721639457s" podCreationTimestamp="2026-03-13 14:52:31 +0000 UTC" firstStartedPulling="2026-03-13 14:52:32.332464934 +0000 UTC m=+2851.232252623" lastFinishedPulling="2026-03-13 14:52:35.404929922 +0000 UTC m=+2854.304717611" observedRunningTime="2026-03-13 14:52:36.384110369 +0000 UTC m=+2855.283898078" watchObservedRunningTime="2026-03-13 14:52:41.721639457 +0000 UTC m=+2860.621427146" Mar 13 14:52:42 crc kubenswrapper[4907]: I0313 14:52:42.441830 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-s8srt" Mar 13 14:52:42 crc kubenswrapper[4907]: I0313 14:52:42.494762 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-s8srt"] Mar 13 14:52:44 crc kubenswrapper[4907]: I0313 14:52:44.414463 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-s8srt" podUID="a53a97c4-ae1d-4e67-b816-c1e4852b08e5" containerName="registry-server" containerID="cri-o://a0f45585bd9bea343e63d05743c65bf9a3e6218aa8b37cbe283fae944e8b7e8d" gracePeriod=2 Mar 13 14:52:45 crc kubenswrapper[4907]: I0313 14:52:45.422717 4907 generic.go:334] "Generic (PLEG): container finished" podID="a53a97c4-ae1d-4e67-b816-c1e4852b08e5" containerID="a0f45585bd9bea343e63d05743c65bf9a3e6218aa8b37cbe283fae944e8b7e8d" exitCode=0 Mar 13 14:52:45 crc kubenswrapper[4907]: I0313 14:52:45.422787 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-s8srt" event={"ID":"a53a97c4-ae1d-4e67-b816-c1e4852b08e5","Type":"ContainerDied","Data":"a0f45585bd9bea343e63d05743c65bf9a3e6218aa8b37cbe283fae944e8b7e8d"} Mar 13 14:52:45 crc kubenswrapper[4907]: I0313 14:52:45.927559 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-s8srt" Mar 13 14:52:45 crc kubenswrapper[4907]: I0313 14:52:45.976270 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-892qz\" (UniqueName: \"kubernetes.io/projected/a53a97c4-ae1d-4e67-b816-c1e4852b08e5-kube-api-access-892qz\") pod \"a53a97c4-ae1d-4e67-b816-c1e4852b08e5\" (UID: \"a53a97c4-ae1d-4e67-b816-c1e4852b08e5\") " Mar 13 14:52:45 crc kubenswrapper[4907]: I0313 14:52:45.982087 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a53a97c4-ae1d-4e67-b816-c1e4852b08e5-kube-api-access-892qz" (OuterVolumeSpecName: "kube-api-access-892qz") pod "a53a97c4-ae1d-4e67-b816-c1e4852b08e5" (UID: "a53a97c4-ae1d-4e67-b816-c1e4852b08e5"). InnerVolumeSpecName "kube-api-access-892qz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:52:46 crc kubenswrapper[4907]: I0313 14:52:46.077829 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a53a97c4-ae1d-4e67-b816-c1e4852b08e5-catalog-content\") pod \"a53a97c4-ae1d-4e67-b816-c1e4852b08e5\" (UID: \"a53a97c4-ae1d-4e67-b816-c1e4852b08e5\") " Mar 13 14:52:46 crc kubenswrapper[4907]: I0313 14:52:46.077938 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a53a97c4-ae1d-4e67-b816-c1e4852b08e5-utilities\") pod \"a53a97c4-ae1d-4e67-b816-c1e4852b08e5\" (UID: \"a53a97c4-ae1d-4e67-b816-c1e4852b08e5\") " Mar 13 14:52:46 crc kubenswrapper[4907]: I0313 14:52:46.078297 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-892qz\" (UniqueName: \"kubernetes.io/projected/a53a97c4-ae1d-4e67-b816-c1e4852b08e5-kube-api-access-892qz\") on node \"crc\" DevicePath \"\"" Mar 13 14:52:46 crc kubenswrapper[4907]: I0313 14:52:46.079105 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a53a97c4-ae1d-4e67-b816-c1e4852b08e5-utilities" (OuterVolumeSpecName: "utilities") pod "a53a97c4-ae1d-4e67-b816-c1e4852b08e5" (UID: "a53a97c4-ae1d-4e67-b816-c1e4852b08e5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:52:46 crc kubenswrapper[4907]: I0313 14:52:46.179526 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a53a97c4-ae1d-4e67-b816-c1e4852b08e5-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 14:52:46 crc kubenswrapper[4907]: I0313 14:52:46.210780 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a53a97c4-ae1d-4e67-b816-c1e4852b08e5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a53a97c4-ae1d-4e67-b816-c1e4852b08e5" (UID: "a53a97c4-ae1d-4e67-b816-c1e4852b08e5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:52:46 crc kubenswrapper[4907]: I0313 14:52:46.281546 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a53a97c4-ae1d-4e67-b816-c1e4852b08e5-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 14:52:46 crc kubenswrapper[4907]: I0313 14:52:46.435133 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-s8srt" event={"ID":"a53a97c4-ae1d-4e67-b816-c1e4852b08e5","Type":"ContainerDied","Data":"0fd20c97ede4d0b2cf316bf01e48438f62ec92b557262bb0310559909098c715"} Mar 13 14:52:46 crc kubenswrapper[4907]: I0313 14:52:46.435207 4907 scope.go:117] "RemoveContainer" containerID="a0f45585bd9bea343e63d05743c65bf9a3e6218aa8b37cbe283fae944e8b7e8d" Mar 13 14:52:46 crc kubenswrapper[4907]: I0313 14:52:46.435280 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-s8srt" Mar 13 14:52:46 crc kubenswrapper[4907]: I0313 14:52:46.457101 4907 scope.go:117] "RemoveContainer" containerID="5f1bfe1233a28e84068fa3690e2155ac3883d0d7f36e3aac56ab833ab2614920" Mar 13 14:52:46 crc kubenswrapper[4907]: I0313 14:52:46.481158 4907 scope.go:117] "RemoveContainer" containerID="0b057665a0d00f3a48117344966f922b2928ae4d7a2f92b272a3107b565359be" Mar 13 14:52:46 crc kubenswrapper[4907]: I0313 14:52:46.497168 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-s8srt"] Mar 13 14:52:46 crc kubenswrapper[4907]: I0313 14:52:46.504347 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-s8srt"] Mar 13 14:52:47 crc kubenswrapper[4907]: I0313 14:52:47.790600 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a53a97c4-ae1d-4e67-b816-c1e4852b08e5" path="/var/lib/kubelet/pods/a53a97c4-ae1d-4e67-b816-c1e4852b08e5/volumes" Mar 13 14:53:40 crc kubenswrapper[4907]: I0313 14:53:40.557327 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-zpk2j"] Mar 13 14:53:40 crc kubenswrapper[4907]: E0313 14:53:40.558105 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a53a97c4-ae1d-4e67-b816-c1e4852b08e5" containerName="registry-server" Mar 13 14:53:40 crc kubenswrapper[4907]: I0313 14:53:40.558117 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="a53a97c4-ae1d-4e67-b816-c1e4852b08e5" containerName="registry-server" Mar 13 14:53:40 crc kubenswrapper[4907]: E0313 14:53:40.558137 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a53a97c4-ae1d-4e67-b816-c1e4852b08e5" containerName="extract-content" Mar 13 14:53:40 crc kubenswrapper[4907]: I0313 14:53:40.558142 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="a53a97c4-ae1d-4e67-b816-c1e4852b08e5" containerName="extract-content" Mar 13 14:53:40 crc kubenswrapper[4907]: E0313 14:53:40.558165 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a53a97c4-ae1d-4e67-b816-c1e4852b08e5" containerName="extract-utilities" Mar 13 14:53:40 crc kubenswrapper[4907]: I0313 14:53:40.558172 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="a53a97c4-ae1d-4e67-b816-c1e4852b08e5" containerName="extract-utilities" Mar 13 14:53:40 crc kubenswrapper[4907]: I0313 14:53:40.558298 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="a53a97c4-ae1d-4e67-b816-c1e4852b08e5" containerName="registry-server" Mar 13 14:53:40 crc kubenswrapper[4907]: I0313 14:53:40.559230 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zpk2j" Mar 13 14:53:40 crc kubenswrapper[4907]: I0313 14:53:40.575687 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zpk2j"] Mar 13 14:53:40 crc kubenswrapper[4907]: I0313 14:53:40.579965 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/40a01f10-ea81-4843-bee8-cc7e5851b1e0-catalog-content\") pod \"community-operators-zpk2j\" (UID: \"40a01f10-ea81-4843-bee8-cc7e5851b1e0\") " pod="openshift-marketplace/community-operators-zpk2j" Mar 13 14:53:40 crc kubenswrapper[4907]: I0313 14:53:40.580061 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/40a01f10-ea81-4843-bee8-cc7e5851b1e0-utilities\") pod \"community-operators-zpk2j\" (UID: \"40a01f10-ea81-4843-bee8-cc7e5851b1e0\") " pod="openshift-marketplace/community-operators-zpk2j" Mar 13 14:53:40 crc kubenswrapper[4907]: I0313 14:53:40.580124 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-blcs2\" (UniqueName: \"kubernetes.io/projected/40a01f10-ea81-4843-bee8-cc7e5851b1e0-kube-api-access-blcs2\") pod \"community-operators-zpk2j\" (UID: \"40a01f10-ea81-4843-bee8-cc7e5851b1e0\") " pod="openshift-marketplace/community-operators-zpk2j" Mar 13 14:53:40 crc kubenswrapper[4907]: I0313 14:53:40.681246 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-blcs2\" (UniqueName: \"kubernetes.io/projected/40a01f10-ea81-4843-bee8-cc7e5851b1e0-kube-api-access-blcs2\") pod \"community-operators-zpk2j\" (UID: \"40a01f10-ea81-4843-bee8-cc7e5851b1e0\") " pod="openshift-marketplace/community-operators-zpk2j" Mar 13 14:53:40 crc kubenswrapper[4907]: I0313 14:53:40.681351 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/40a01f10-ea81-4843-bee8-cc7e5851b1e0-catalog-content\") pod \"community-operators-zpk2j\" (UID: \"40a01f10-ea81-4843-bee8-cc7e5851b1e0\") " pod="openshift-marketplace/community-operators-zpk2j" Mar 13 14:53:40 crc kubenswrapper[4907]: I0313 14:53:40.681389 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/40a01f10-ea81-4843-bee8-cc7e5851b1e0-utilities\") pod \"community-operators-zpk2j\" (UID: \"40a01f10-ea81-4843-bee8-cc7e5851b1e0\") " pod="openshift-marketplace/community-operators-zpk2j" Mar 13 14:53:40 crc kubenswrapper[4907]: I0313 14:53:40.681812 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/40a01f10-ea81-4843-bee8-cc7e5851b1e0-catalog-content\") pod \"community-operators-zpk2j\" (UID: \"40a01f10-ea81-4843-bee8-cc7e5851b1e0\") " pod="openshift-marketplace/community-operators-zpk2j" Mar 13 14:53:40 crc kubenswrapper[4907]: I0313 14:53:40.681860 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/40a01f10-ea81-4843-bee8-cc7e5851b1e0-utilities\") pod \"community-operators-zpk2j\" (UID: \"40a01f10-ea81-4843-bee8-cc7e5851b1e0\") " pod="openshift-marketplace/community-operators-zpk2j" Mar 13 14:53:40 crc kubenswrapper[4907]: I0313 14:53:40.711667 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-blcs2\" (UniqueName: \"kubernetes.io/projected/40a01f10-ea81-4843-bee8-cc7e5851b1e0-kube-api-access-blcs2\") pod \"community-operators-zpk2j\" (UID: \"40a01f10-ea81-4843-bee8-cc7e5851b1e0\") " pod="openshift-marketplace/community-operators-zpk2j" Mar 13 14:53:40 crc kubenswrapper[4907]: I0313 14:53:40.877731 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zpk2j" Mar 13 14:53:41 crc kubenswrapper[4907]: I0313 14:53:41.346861 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zpk2j"] Mar 13 14:53:41 crc kubenswrapper[4907]: I0313 14:53:41.832227 4907 generic.go:334] "Generic (PLEG): container finished" podID="40a01f10-ea81-4843-bee8-cc7e5851b1e0" containerID="90a60a553cdd587785b5c5a14302529824eeffe1f263fc6e5179be6c780af850" exitCode=0 Mar 13 14:53:41 crc kubenswrapper[4907]: I0313 14:53:41.832295 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zpk2j" event={"ID":"40a01f10-ea81-4843-bee8-cc7e5851b1e0","Type":"ContainerDied","Data":"90a60a553cdd587785b5c5a14302529824eeffe1f263fc6e5179be6c780af850"} Mar 13 14:53:41 crc kubenswrapper[4907]: I0313 14:53:41.832471 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zpk2j" event={"ID":"40a01f10-ea81-4843-bee8-cc7e5851b1e0","Type":"ContainerStarted","Data":"507df3aaee8a6dbfeb22b5e97bee364fc1ad89fcaba6062646ae95dc721d6f17"} Mar 13 14:53:43 crc kubenswrapper[4907]: I0313 14:53:43.847154 4907 generic.go:334] "Generic (PLEG): container finished" podID="40a01f10-ea81-4843-bee8-cc7e5851b1e0" containerID="401a1d6750944d46aa81b47dac6e65af576075e988be53e4b38e87b67ecebd67" exitCode=0 Mar 13 14:53:43 crc kubenswrapper[4907]: I0313 14:53:43.847215 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zpk2j" event={"ID":"40a01f10-ea81-4843-bee8-cc7e5851b1e0","Type":"ContainerDied","Data":"401a1d6750944d46aa81b47dac6e65af576075e988be53e4b38e87b67ecebd67"} Mar 13 14:53:44 crc kubenswrapper[4907]: I0313 14:53:44.857066 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zpk2j" event={"ID":"40a01f10-ea81-4843-bee8-cc7e5851b1e0","Type":"ContainerStarted","Data":"0d00e6e05c6219b1e79b7b2d7b334c8b87d4b7aeb54d70b84bcf93cbd75db2d8"} Mar 13 14:53:44 crc kubenswrapper[4907]: I0313 14:53:44.875546 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-zpk2j" podStartSLOduration=2.430917728 podStartE2EDuration="4.875529439s" podCreationTimestamp="2026-03-13 14:53:40 +0000 UTC" firstStartedPulling="2026-03-13 14:53:41.834489641 +0000 UTC m=+2920.734277330" lastFinishedPulling="2026-03-13 14:53:44.279101312 +0000 UTC m=+2923.178889041" observedRunningTime="2026-03-13 14:53:44.872452284 +0000 UTC m=+2923.772239973" watchObservedRunningTime="2026-03-13 14:53:44.875529439 +0000 UTC m=+2923.775317128" Mar 13 14:53:50 crc kubenswrapper[4907]: I0313 14:53:50.878767 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-zpk2j" Mar 13 14:53:50 crc kubenswrapper[4907]: I0313 14:53:50.880085 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-zpk2j" Mar 13 14:53:50 crc kubenswrapper[4907]: I0313 14:53:50.926631 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-zpk2j" Mar 13 14:53:50 crc kubenswrapper[4907]: I0313 14:53:50.969747 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-zpk2j" Mar 13 14:53:51 crc kubenswrapper[4907]: I0313 14:53:51.508944 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zpk2j"] Mar 13 14:53:52 crc kubenswrapper[4907]: I0313 14:53:52.913814 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-zpk2j" podUID="40a01f10-ea81-4843-bee8-cc7e5851b1e0" containerName="registry-server" containerID="cri-o://0d00e6e05c6219b1e79b7b2d7b334c8b87d4b7aeb54d70b84bcf93cbd75db2d8" gracePeriod=2 Mar 13 14:53:53 crc kubenswrapper[4907]: I0313 14:53:53.270557 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zpk2j" Mar 13 14:53:53 crc kubenswrapper[4907]: I0313 14:53:53.464669 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/40a01f10-ea81-4843-bee8-cc7e5851b1e0-catalog-content\") pod \"40a01f10-ea81-4843-bee8-cc7e5851b1e0\" (UID: \"40a01f10-ea81-4843-bee8-cc7e5851b1e0\") " Mar 13 14:53:53 crc kubenswrapper[4907]: I0313 14:53:53.464789 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/40a01f10-ea81-4843-bee8-cc7e5851b1e0-utilities\") pod \"40a01f10-ea81-4843-bee8-cc7e5851b1e0\" (UID: \"40a01f10-ea81-4843-bee8-cc7e5851b1e0\") " Mar 13 14:53:53 crc kubenswrapper[4907]: I0313 14:53:53.464818 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-blcs2\" (UniqueName: \"kubernetes.io/projected/40a01f10-ea81-4843-bee8-cc7e5851b1e0-kube-api-access-blcs2\") pod \"40a01f10-ea81-4843-bee8-cc7e5851b1e0\" (UID: \"40a01f10-ea81-4843-bee8-cc7e5851b1e0\") " Mar 13 14:53:53 crc kubenswrapper[4907]: I0313 14:53:53.466221 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/40a01f10-ea81-4843-bee8-cc7e5851b1e0-utilities" (OuterVolumeSpecName: "utilities") pod "40a01f10-ea81-4843-bee8-cc7e5851b1e0" (UID: "40a01f10-ea81-4843-bee8-cc7e5851b1e0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:53:53 crc kubenswrapper[4907]: I0313 14:53:53.471257 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/40a01f10-ea81-4843-bee8-cc7e5851b1e0-kube-api-access-blcs2" (OuterVolumeSpecName: "kube-api-access-blcs2") pod "40a01f10-ea81-4843-bee8-cc7e5851b1e0" (UID: "40a01f10-ea81-4843-bee8-cc7e5851b1e0"). InnerVolumeSpecName "kube-api-access-blcs2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:53:53 crc kubenswrapper[4907]: I0313 14:53:53.523409 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/40a01f10-ea81-4843-bee8-cc7e5851b1e0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "40a01f10-ea81-4843-bee8-cc7e5851b1e0" (UID: "40a01f10-ea81-4843-bee8-cc7e5851b1e0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:53:53 crc kubenswrapper[4907]: I0313 14:53:53.572358 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/40a01f10-ea81-4843-bee8-cc7e5851b1e0-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 14:53:53 crc kubenswrapper[4907]: I0313 14:53:53.572405 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/40a01f10-ea81-4843-bee8-cc7e5851b1e0-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 14:53:53 crc kubenswrapper[4907]: I0313 14:53:53.572420 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-blcs2\" (UniqueName: \"kubernetes.io/projected/40a01f10-ea81-4843-bee8-cc7e5851b1e0-kube-api-access-blcs2\") on node \"crc\" DevicePath \"\"" Mar 13 14:53:53 crc kubenswrapper[4907]: I0313 14:53:53.922300 4907 generic.go:334] "Generic (PLEG): container finished" podID="40a01f10-ea81-4843-bee8-cc7e5851b1e0" containerID="0d00e6e05c6219b1e79b7b2d7b334c8b87d4b7aeb54d70b84bcf93cbd75db2d8" exitCode=0 Mar 13 14:53:53 crc kubenswrapper[4907]: I0313 14:53:53.922359 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zpk2j" event={"ID":"40a01f10-ea81-4843-bee8-cc7e5851b1e0","Type":"ContainerDied","Data":"0d00e6e05c6219b1e79b7b2d7b334c8b87d4b7aeb54d70b84bcf93cbd75db2d8"} Mar 13 14:53:53 crc kubenswrapper[4907]: I0313 14:53:53.922416 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zpk2j" event={"ID":"40a01f10-ea81-4843-bee8-cc7e5851b1e0","Type":"ContainerDied","Data":"507df3aaee8a6dbfeb22b5e97bee364fc1ad89fcaba6062646ae95dc721d6f17"} Mar 13 14:53:53 crc kubenswrapper[4907]: I0313 14:53:53.922364 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zpk2j" Mar 13 14:53:53 crc kubenswrapper[4907]: I0313 14:53:53.922442 4907 scope.go:117] "RemoveContainer" containerID="0d00e6e05c6219b1e79b7b2d7b334c8b87d4b7aeb54d70b84bcf93cbd75db2d8" Mar 13 14:53:53 crc kubenswrapper[4907]: I0313 14:53:53.945316 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zpk2j"] Mar 13 14:53:53 crc kubenswrapper[4907]: I0313 14:53:53.946565 4907 scope.go:117] "RemoveContainer" containerID="401a1d6750944d46aa81b47dac6e65af576075e988be53e4b38e87b67ecebd67" Mar 13 14:53:53 crc kubenswrapper[4907]: I0313 14:53:53.950636 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-zpk2j"] Mar 13 14:53:53 crc kubenswrapper[4907]: I0313 14:53:53.965708 4907 scope.go:117] "RemoveContainer" containerID="90a60a553cdd587785b5c5a14302529824eeffe1f263fc6e5179be6c780af850" Mar 13 14:53:53 crc kubenswrapper[4907]: I0313 14:53:53.981576 4907 scope.go:117] "RemoveContainer" containerID="0d00e6e05c6219b1e79b7b2d7b334c8b87d4b7aeb54d70b84bcf93cbd75db2d8" Mar 13 14:53:53 crc kubenswrapper[4907]: E0313 14:53:53.981928 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0d00e6e05c6219b1e79b7b2d7b334c8b87d4b7aeb54d70b84bcf93cbd75db2d8\": container with ID starting with 0d00e6e05c6219b1e79b7b2d7b334c8b87d4b7aeb54d70b84bcf93cbd75db2d8 not found: ID does not exist" containerID="0d00e6e05c6219b1e79b7b2d7b334c8b87d4b7aeb54d70b84bcf93cbd75db2d8" Mar 13 14:53:53 crc kubenswrapper[4907]: I0313 14:53:53.981965 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0d00e6e05c6219b1e79b7b2d7b334c8b87d4b7aeb54d70b84bcf93cbd75db2d8"} err="failed to get container status \"0d00e6e05c6219b1e79b7b2d7b334c8b87d4b7aeb54d70b84bcf93cbd75db2d8\": rpc error: code = NotFound desc = could not find container \"0d00e6e05c6219b1e79b7b2d7b334c8b87d4b7aeb54d70b84bcf93cbd75db2d8\": container with ID starting with 0d00e6e05c6219b1e79b7b2d7b334c8b87d4b7aeb54d70b84bcf93cbd75db2d8 not found: ID does not exist" Mar 13 14:53:53 crc kubenswrapper[4907]: I0313 14:53:53.982006 4907 scope.go:117] "RemoveContainer" containerID="401a1d6750944d46aa81b47dac6e65af576075e988be53e4b38e87b67ecebd67" Mar 13 14:53:53 crc kubenswrapper[4907]: E0313 14:53:53.982226 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"401a1d6750944d46aa81b47dac6e65af576075e988be53e4b38e87b67ecebd67\": container with ID starting with 401a1d6750944d46aa81b47dac6e65af576075e988be53e4b38e87b67ecebd67 not found: ID does not exist" containerID="401a1d6750944d46aa81b47dac6e65af576075e988be53e4b38e87b67ecebd67" Mar 13 14:53:53 crc kubenswrapper[4907]: I0313 14:53:53.982252 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"401a1d6750944d46aa81b47dac6e65af576075e988be53e4b38e87b67ecebd67"} err="failed to get container status \"401a1d6750944d46aa81b47dac6e65af576075e988be53e4b38e87b67ecebd67\": rpc error: code = NotFound desc = could not find container \"401a1d6750944d46aa81b47dac6e65af576075e988be53e4b38e87b67ecebd67\": container with ID starting with 401a1d6750944d46aa81b47dac6e65af576075e988be53e4b38e87b67ecebd67 not found: ID does not exist" Mar 13 14:53:53 crc kubenswrapper[4907]: I0313 14:53:53.982266 4907 scope.go:117] "RemoveContainer" containerID="90a60a553cdd587785b5c5a14302529824eeffe1f263fc6e5179be6c780af850" Mar 13 14:53:53 crc kubenswrapper[4907]: E0313 14:53:53.982570 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"90a60a553cdd587785b5c5a14302529824eeffe1f263fc6e5179be6c780af850\": container with ID starting with 90a60a553cdd587785b5c5a14302529824eeffe1f263fc6e5179be6c780af850 not found: ID does not exist" containerID="90a60a553cdd587785b5c5a14302529824eeffe1f263fc6e5179be6c780af850" Mar 13 14:53:53 crc kubenswrapper[4907]: I0313 14:53:53.982590 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"90a60a553cdd587785b5c5a14302529824eeffe1f263fc6e5179be6c780af850"} err="failed to get container status \"90a60a553cdd587785b5c5a14302529824eeffe1f263fc6e5179be6c780af850\": rpc error: code = NotFound desc = could not find container \"90a60a553cdd587785b5c5a14302529824eeffe1f263fc6e5179be6c780af850\": container with ID starting with 90a60a553cdd587785b5c5a14302529824eeffe1f263fc6e5179be6c780af850 not found: ID does not exist" Mar 13 14:53:55 crc kubenswrapper[4907]: I0313 14:53:55.792635 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="40a01f10-ea81-4843-bee8-cc7e5851b1e0" path="/var/lib/kubelet/pods/40a01f10-ea81-4843-bee8-cc7e5851b1e0/volumes" Mar 13 14:54:00 crc kubenswrapper[4907]: I0313 14:54:00.140385 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556894-8mk4g"] Mar 13 14:54:00 crc kubenswrapper[4907]: E0313 14:54:00.141250 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40a01f10-ea81-4843-bee8-cc7e5851b1e0" containerName="extract-content" Mar 13 14:54:00 crc kubenswrapper[4907]: I0313 14:54:00.141266 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="40a01f10-ea81-4843-bee8-cc7e5851b1e0" containerName="extract-content" Mar 13 14:54:00 crc kubenswrapper[4907]: E0313 14:54:00.141283 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40a01f10-ea81-4843-bee8-cc7e5851b1e0" containerName="registry-server" Mar 13 14:54:00 crc kubenswrapper[4907]: I0313 14:54:00.141293 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="40a01f10-ea81-4843-bee8-cc7e5851b1e0" containerName="registry-server" Mar 13 14:54:00 crc kubenswrapper[4907]: E0313 14:54:00.141310 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40a01f10-ea81-4843-bee8-cc7e5851b1e0" containerName="extract-utilities" Mar 13 14:54:00 crc kubenswrapper[4907]: I0313 14:54:00.141319 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="40a01f10-ea81-4843-bee8-cc7e5851b1e0" containerName="extract-utilities" Mar 13 14:54:00 crc kubenswrapper[4907]: I0313 14:54:00.141509 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="40a01f10-ea81-4843-bee8-cc7e5851b1e0" containerName="registry-server" Mar 13 14:54:00 crc kubenswrapper[4907]: I0313 14:54:00.142120 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556894-8mk4g" Mar 13 14:54:00 crc kubenswrapper[4907]: I0313 14:54:00.144540 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 14:54:00 crc kubenswrapper[4907]: I0313 14:54:00.145744 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 14:54:00 crc kubenswrapper[4907]: I0313 14:54:00.150996 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556894-8mk4g"] Mar 13 14:54:00 crc kubenswrapper[4907]: I0313 14:54:00.153208 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 14:54:00 crc kubenswrapper[4907]: I0313 14:54:00.161570 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dkvf8\" (UniqueName: \"kubernetes.io/projected/14af3ced-cea4-46f4-82b2-bdcc55744a4f-kube-api-access-dkvf8\") pod \"auto-csr-approver-29556894-8mk4g\" (UID: \"14af3ced-cea4-46f4-82b2-bdcc55744a4f\") " pod="openshift-infra/auto-csr-approver-29556894-8mk4g" Mar 13 14:54:00 crc kubenswrapper[4907]: I0313 14:54:00.262680 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dkvf8\" (UniqueName: \"kubernetes.io/projected/14af3ced-cea4-46f4-82b2-bdcc55744a4f-kube-api-access-dkvf8\") pod \"auto-csr-approver-29556894-8mk4g\" (UID: \"14af3ced-cea4-46f4-82b2-bdcc55744a4f\") " pod="openshift-infra/auto-csr-approver-29556894-8mk4g" Mar 13 14:54:00 crc kubenswrapper[4907]: I0313 14:54:00.281011 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dkvf8\" (UniqueName: \"kubernetes.io/projected/14af3ced-cea4-46f4-82b2-bdcc55744a4f-kube-api-access-dkvf8\") pod \"auto-csr-approver-29556894-8mk4g\" (UID: \"14af3ced-cea4-46f4-82b2-bdcc55744a4f\") " pod="openshift-infra/auto-csr-approver-29556894-8mk4g" Mar 13 14:54:00 crc kubenswrapper[4907]: I0313 14:54:00.458983 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556894-8mk4g" Mar 13 14:54:00 crc kubenswrapper[4907]: I0313 14:54:00.884993 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556894-8mk4g"] Mar 13 14:54:00 crc kubenswrapper[4907]: I0313 14:54:00.979589 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556894-8mk4g" event={"ID":"14af3ced-cea4-46f4-82b2-bdcc55744a4f","Type":"ContainerStarted","Data":"6c0e1e2490d9f0a95c8081e3db1bb9df146ef5a4b0b734e629bd40b26691c155"} Mar 13 14:54:03 crc kubenswrapper[4907]: I0313 14:54:03.000040 4907 generic.go:334] "Generic (PLEG): container finished" podID="14af3ced-cea4-46f4-82b2-bdcc55744a4f" containerID="1f3fec4f332e7900eb7076ad444f074ae92511e972577dd77d05eebc0931b626" exitCode=0 Mar 13 14:54:03 crc kubenswrapper[4907]: I0313 14:54:03.000123 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556894-8mk4g" event={"ID":"14af3ced-cea4-46f4-82b2-bdcc55744a4f","Type":"ContainerDied","Data":"1f3fec4f332e7900eb7076ad444f074ae92511e972577dd77d05eebc0931b626"} Mar 13 14:54:04 crc kubenswrapper[4907]: I0313 14:54:04.434050 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556894-8mk4g" Mar 13 14:54:04 crc kubenswrapper[4907]: I0313 14:54:04.629861 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dkvf8\" (UniqueName: \"kubernetes.io/projected/14af3ced-cea4-46f4-82b2-bdcc55744a4f-kube-api-access-dkvf8\") pod \"14af3ced-cea4-46f4-82b2-bdcc55744a4f\" (UID: \"14af3ced-cea4-46f4-82b2-bdcc55744a4f\") " Mar 13 14:54:04 crc kubenswrapper[4907]: I0313 14:54:04.634581 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/14af3ced-cea4-46f4-82b2-bdcc55744a4f-kube-api-access-dkvf8" (OuterVolumeSpecName: "kube-api-access-dkvf8") pod "14af3ced-cea4-46f4-82b2-bdcc55744a4f" (UID: "14af3ced-cea4-46f4-82b2-bdcc55744a4f"). InnerVolumeSpecName "kube-api-access-dkvf8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:54:04 crc kubenswrapper[4907]: I0313 14:54:04.732001 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dkvf8\" (UniqueName: \"kubernetes.io/projected/14af3ced-cea4-46f4-82b2-bdcc55744a4f-kube-api-access-dkvf8\") on node \"crc\" DevicePath \"\"" Mar 13 14:54:05 crc kubenswrapper[4907]: I0313 14:54:05.016189 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556894-8mk4g" event={"ID":"14af3ced-cea4-46f4-82b2-bdcc55744a4f","Type":"ContainerDied","Data":"6c0e1e2490d9f0a95c8081e3db1bb9df146ef5a4b0b734e629bd40b26691c155"} Mar 13 14:54:05 crc kubenswrapper[4907]: I0313 14:54:05.016239 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6c0e1e2490d9f0a95c8081e3db1bb9df146ef5a4b0b734e629bd40b26691c155" Mar 13 14:54:05 crc kubenswrapper[4907]: I0313 14:54:05.016244 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556894-8mk4g" Mar 13 14:54:05 crc kubenswrapper[4907]: I0313 14:54:05.500023 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556888-6dcv4"] Mar 13 14:54:05 crc kubenswrapper[4907]: I0313 14:54:05.508711 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556888-6dcv4"] Mar 13 14:54:05 crc kubenswrapper[4907]: I0313 14:54:05.795166 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4b255770-f854-44fe-93fb-7ae5082081be" path="/var/lib/kubelet/pods/4b255770-f854-44fe-93fb-7ae5082081be/volumes" Mar 13 14:54:18 crc kubenswrapper[4907]: I0313 14:54:18.041849 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 14:54:18 crc kubenswrapper[4907]: I0313 14:54:18.042337 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 14:54:25 crc kubenswrapper[4907]: I0313 14:54:25.574570 4907 scope.go:117] "RemoveContainer" containerID="33e7c832124ee1aa4196cb0bd674abcb6224ac6318050e8ed5dbb3afadecce6f" Mar 13 14:54:48 crc kubenswrapper[4907]: I0313 14:54:48.042033 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 14:54:48 crc kubenswrapper[4907]: I0313 14:54:48.042606 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 14:55:18 crc kubenswrapper[4907]: I0313 14:55:18.041927 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 14:55:18 crc kubenswrapper[4907]: I0313 14:55:18.042512 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 14:55:18 crc kubenswrapper[4907]: I0313 14:55:18.042566 4907 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" Mar 13 14:55:18 crc kubenswrapper[4907]: I0313 14:55:18.043268 4907 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a44259e1e0a7edc8acdd625ba3f7415d70aa78b760d8eeb81254942722d3acb3"} pod="openshift-machine-config-operator/machine-config-daemon-hm56j" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 14:55:18 crc kubenswrapper[4907]: I0313 14:55:18.043344 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" containerID="cri-o://a44259e1e0a7edc8acdd625ba3f7415d70aa78b760d8eeb81254942722d3acb3" gracePeriod=600 Mar 13 14:55:18 crc kubenswrapper[4907]: E0313 14:55:18.167225 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:55:18 crc kubenswrapper[4907]: I0313 14:55:18.562674 4907 generic.go:334] "Generic (PLEG): container finished" podID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerID="a44259e1e0a7edc8acdd625ba3f7415d70aa78b760d8eeb81254942722d3acb3" exitCode=0 Mar 13 14:55:18 crc kubenswrapper[4907]: I0313 14:55:18.562738 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerDied","Data":"a44259e1e0a7edc8acdd625ba3f7415d70aa78b760d8eeb81254942722d3acb3"} Mar 13 14:55:18 crc kubenswrapper[4907]: I0313 14:55:18.562788 4907 scope.go:117] "RemoveContainer" containerID="5b16948da306cc82d0f15e426031c3bf4ecc434619902cfc2cfc82bb26ad6352" Mar 13 14:55:18 crc kubenswrapper[4907]: I0313 14:55:18.563647 4907 scope.go:117] "RemoveContainer" containerID="a44259e1e0a7edc8acdd625ba3f7415d70aa78b760d8eeb81254942722d3acb3" Mar 13 14:55:18 crc kubenswrapper[4907]: E0313 14:55:18.564081 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:55:31 crc kubenswrapper[4907]: I0313 14:55:31.790212 4907 scope.go:117] "RemoveContainer" containerID="a44259e1e0a7edc8acdd625ba3f7415d70aa78b760d8eeb81254942722d3acb3" Mar 13 14:55:31 crc kubenswrapper[4907]: E0313 14:55:31.791402 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:55:42 crc kubenswrapper[4907]: I0313 14:55:42.782798 4907 scope.go:117] "RemoveContainer" containerID="a44259e1e0a7edc8acdd625ba3f7415d70aa78b760d8eeb81254942722d3acb3" Mar 13 14:55:42 crc kubenswrapper[4907]: E0313 14:55:42.783592 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:55:53 crc kubenswrapper[4907]: I0313 14:55:53.782759 4907 scope.go:117] "RemoveContainer" containerID="a44259e1e0a7edc8acdd625ba3f7415d70aa78b760d8eeb81254942722d3acb3" Mar 13 14:55:53 crc kubenswrapper[4907]: E0313 14:55:53.784223 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:56:00 crc kubenswrapper[4907]: I0313 14:56:00.144802 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556896-7dlx2"] Mar 13 14:56:00 crc kubenswrapper[4907]: E0313 14:56:00.145804 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14af3ced-cea4-46f4-82b2-bdcc55744a4f" containerName="oc" Mar 13 14:56:00 crc kubenswrapper[4907]: I0313 14:56:00.145826 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="14af3ced-cea4-46f4-82b2-bdcc55744a4f" containerName="oc" Mar 13 14:56:00 crc kubenswrapper[4907]: I0313 14:56:00.146117 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="14af3ced-cea4-46f4-82b2-bdcc55744a4f" containerName="oc" Mar 13 14:56:00 crc kubenswrapper[4907]: I0313 14:56:00.146789 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556896-7dlx2" Mar 13 14:56:00 crc kubenswrapper[4907]: I0313 14:56:00.151150 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 14:56:00 crc kubenswrapper[4907]: I0313 14:56:00.151483 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 14:56:00 crc kubenswrapper[4907]: I0313 14:56:00.151650 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 14:56:00 crc kubenswrapper[4907]: I0313 14:56:00.159437 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556896-7dlx2"] Mar 13 14:56:00 crc kubenswrapper[4907]: I0313 14:56:00.316399 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j9wkv\" (UniqueName: \"kubernetes.io/projected/e193f9f5-a033-4f42-9932-62f7b4edd0b2-kube-api-access-j9wkv\") pod \"auto-csr-approver-29556896-7dlx2\" (UID: \"e193f9f5-a033-4f42-9932-62f7b4edd0b2\") " pod="openshift-infra/auto-csr-approver-29556896-7dlx2" Mar 13 14:56:00 crc kubenswrapper[4907]: I0313 14:56:00.417726 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j9wkv\" (UniqueName: \"kubernetes.io/projected/e193f9f5-a033-4f42-9932-62f7b4edd0b2-kube-api-access-j9wkv\") pod \"auto-csr-approver-29556896-7dlx2\" (UID: \"e193f9f5-a033-4f42-9932-62f7b4edd0b2\") " pod="openshift-infra/auto-csr-approver-29556896-7dlx2" Mar 13 14:56:00 crc kubenswrapper[4907]: I0313 14:56:00.448853 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j9wkv\" (UniqueName: \"kubernetes.io/projected/e193f9f5-a033-4f42-9932-62f7b4edd0b2-kube-api-access-j9wkv\") pod \"auto-csr-approver-29556896-7dlx2\" (UID: \"e193f9f5-a033-4f42-9932-62f7b4edd0b2\") " pod="openshift-infra/auto-csr-approver-29556896-7dlx2" Mar 13 14:56:00 crc kubenswrapper[4907]: I0313 14:56:00.471979 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556896-7dlx2" Mar 13 14:56:00 crc kubenswrapper[4907]: I0313 14:56:00.911827 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556896-7dlx2"] Mar 13 14:56:00 crc kubenswrapper[4907]: I0313 14:56:00.923212 4907 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 13 14:56:00 crc kubenswrapper[4907]: I0313 14:56:00.936952 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556896-7dlx2" event={"ID":"e193f9f5-a033-4f42-9932-62f7b4edd0b2","Type":"ContainerStarted","Data":"fe1ac4052638fbf8e2ed7b02b6cc93e68c5a8f741a2e3a0784fd85d47094c7ff"} Mar 13 14:56:02 crc kubenswrapper[4907]: I0313 14:56:02.954006 4907 generic.go:334] "Generic (PLEG): container finished" podID="e193f9f5-a033-4f42-9932-62f7b4edd0b2" containerID="19c525d3167fb54f80c181f5117cf495aa5c0602e81951a3395f3a05b6fbc6c9" exitCode=0 Mar 13 14:56:02 crc kubenswrapper[4907]: I0313 14:56:02.954140 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556896-7dlx2" event={"ID":"e193f9f5-a033-4f42-9932-62f7b4edd0b2","Type":"ContainerDied","Data":"19c525d3167fb54f80c181f5117cf495aa5c0602e81951a3395f3a05b6fbc6c9"} Mar 13 14:56:04 crc kubenswrapper[4907]: I0313 14:56:04.264171 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556896-7dlx2" Mar 13 14:56:04 crc kubenswrapper[4907]: I0313 14:56:04.283110 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j9wkv\" (UniqueName: \"kubernetes.io/projected/e193f9f5-a033-4f42-9932-62f7b4edd0b2-kube-api-access-j9wkv\") pod \"e193f9f5-a033-4f42-9932-62f7b4edd0b2\" (UID: \"e193f9f5-a033-4f42-9932-62f7b4edd0b2\") " Mar 13 14:56:04 crc kubenswrapper[4907]: I0313 14:56:04.288556 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e193f9f5-a033-4f42-9932-62f7b4edd0b2-kube-api-access-j9wkv" (OuterVolumeSpecName: "kube-api-access-j9wkv") pod "e193f9f5-a033-4f42-9932-62f7b4edd0b2" (UID: "e193f9f5-a033-4f42-9932-62f7b4edd0b2"). InnerVolumeSpecName "kube-api-access-j9wkv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:56:04 crc kubenswrapper[4907]: I0313 14:56:04.384497 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j9wkv\" (UniqueName: \"kubernetes.io/projected/e193f9f5-a033-4f42-9932-62f7b4edd0b2-kube-api-access-j9wkv\") on node \"crc\" DevicePath \"\"" Mar 13 14:56:04 crc kubenswrapper[4907]: I0313 14:56:04.971768 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556896-7dlx2" event={"ID":"e193f9f5-a033-4f42-9932-62f7b4edd0b2","Type":"ContainerDied","Data":"fe1ac4052638fbf8e2ed7b02b6cc93e68c5a8f741a2e3a0784fd85d47094c7ff"} Mar 13 14:56:04 crc kubenswrapper[4907]: I0313 14:56:04.971808 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fe1ac4052638fbf8e2ed7b02b6cc93e68c5a8f741a2e3a0784fd85d47094c7ff" Mar 13 14:56:04 crc kubenswrapper[4907]: I0313 14:56:04.971856 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556896-7dlx2" Mar 13 14:56:05 crc kubenswrapper[4907]: I0313 14:56:05.329744 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556890-5kp94"] Mar 13 14:56:05 crc kubenswrapper[4907]: I0313 14:56:05.335771 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556890-5kp94"] Mar 13 14:56:05 crc kubenswrapper[4907]: I0313 14:56:05.790828 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d63f291e-b12e-4b0a-b4c3-e7b9c2cc3c40" path="/var/lib/kubelet/pods/d63f291e-b12e-4b0a-b4c3-e7b9c2cc3c40/volumes" Mar 13 14:56:07 crc kubenswrapper[4907]: I0313 14:56:07.782526 4907 scope.go:117] "RemoveContainer" containerID="a44259e1e0a7edc8acdd625ba3f7415d70aa78b760d8eeb81254942722d3acb3" Mar 13 14:56:07 crc kubenswrapper[4907]: E0313 14:56:07.783101 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:56:20 crc kubenswrapper[4907]: I0313 14:56:20.782968 4907 scope.go:117] "RemoveContainer" containerID="a44259e1e0a7edc8acdd625ba3f7415d70aa78b760d8eeb81254942722d3acb3" Mar 13 14:56:20 crc kubenswrapper[4907]: E0313 14:56:20.783796 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:56:25 crc kubenswrapper[4907]: I0313 14:56:25.714329 4907 scope.go:117] "RemoveContainer" containerID="f4cc0414786125661ff7aeac8cccd8627fe8bdb42c54577e769f1e1b046242f5" Mar 13 14:56:34 crc kubenswrapper[4907]: I0313 14:56:34.782733 4907 scope.go:117] "RemoveContainer" containerID="a44259e1e0a7edc8acdd625ba3f7415d70aa78b760d8eeb81254942722d3acb3" Mar 13 14:56:34 crc kubenswrapper[4907]: E0313 14:56:34.783515 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:56:47 crc kubenswrapper[4907]: I0313 14:56:47.950910 4907 scope.go:117] "RemoveContainer" containerID="a44259e1e0a7edc8acdd625ba3f7415d70aa78b760d8eeb81254942722d3acb3" Mar 13 14:56:47 crc kubenswrapper[4907]: E0313 14:56:47.951630 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:57:01 crc kubenswrapper[4907]: I0313 14:57:01.787123 4907 scope.go:117] "RemoveContainer" containerID="a44259e1e0a7edc8acdd625ba3f7415d70aa78b760d8eeb81254942722d3acb3" Mar 13 14:57:01 crc kubenswrapper[4907]: E0313 14:57:01.788053 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:57:13 crc kubenswrapper[4907]: I0313 14:57:13.783488 4907 scope.go:117] "RemoveContainer" containerID="a44259e1e0a7edc8acdd625ba3f7415d70aa78b760d8eeb81254942722d3acb3" Mar 13 14:57:13 crc kubenswrapper[4907]: E0313 14:57:13.784646 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:57:24 crc kubenswrapper[4907]: I0313 14:57:24.783355 4907 scope.go:117] "RemoveContainer" containerID="a44259e1e0a7edc8acdd625ba3f7415d70aa78b760d8eeb81254942722d3acb3" Mar 13 14:57:24 crc kubenswrapper[4907]: E0313 14:57:24.784512 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:57:36 crc kubenswrapper[4907]: I0313 14:57:36.782498 4907 scope.go:117] "RemoveContainer" containerID="a44259e1e0a7edc8acdd625ba3f7415d70aa78b760d8eeb81254942722d3acb3" Mar 13 14:57:36 crc kubenswrapper[4907]: E0313 14:57:36.784535 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:57:48 crc kubenswrapper[4907]: I0313 14:57:48.783019 4907 scope.go:117] "RemoveContainer" containerID="a44259e1e0a7edc8acdd625ba3f7415d70aa78b760d8eeb81254942722d3acb3" Mar 13 14:57:48 crc kubenswrapper[4907]: E0313 14:57:48.783684 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:58:00 crc kubenswrapper[4907]: I0313 14:58:00.140246 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556898-55zxj"] Mar 13 14:58:00 crc kubenswrapper[4907]: E0313 14:58:00.141141 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e193f9f5-a033-4f42-9932-62f7b4edd0b2" containerName="oc" Mar 13 14:58:00 crc kubenswrapper[4907]: I0313 14:58:00.141156 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="e193f9f5-a033-4f42-9932-62f7b4edd0b2" containerName="oc" Mar 13 14:58:00 crc kubenswrapper[4907]: I0313 14:58:00.141340 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="e193f9f5-a033-4f42-9932-62f7b4edd0b2" containerName="oc" Mar 13 14:58:00 crc kubenswrapper[4907]: I0313 14:58:00.141871 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556898-55zxj" Mar 13 14:58:00 crc kubenswrapper[4907]: I0313 14:58:00.144152 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 14:58:00 crc kubenswrapper[4907]: I0313 14:58:00.144941 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 14:58:00 crc kubenswrapper[4907]: I0313 14:58:00.148774 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556898-55zxj"] Mar 13 14:58:00 crc kubenswrapper[4907]: I0313 14:58:00.150908 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 14:58:00 crc kubenswrapper[4907]: I0313 14:58:00.195098 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rjqcf\" (UniqueName: \"kubernetes.io/projected/4efb5dca-e973-443a-bcc2-3b8799fb680c-kube-api-access-rjqcf\") pod \"auto-csr-approver-29556898-55zxj\" (UID: \"4efb5dca-e973-443a-bcc2-3b8799fb680c\") " pod="openshift-infra/auto-csr-approver-29556898-55zxj" Mar 13 14:58:00 crc kubenswrapper[4907]: I0313 14:58:00.296587 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rjqcf\" (UniqueName: \"kubernetes.io/projected/4efb5dca-e973-443a-bcc2-3b8799fb680c-kube-api-access-rjqcf\") pod \"auto-csr-approver-29556898-55zxj\" (UID: \"4efb5dca-e973-443a-bcc2-3b8799fb680c\") " pod="openshift-infra/auto-csr-approver-29556898-55zxj" Mar 13 14:58:00 crc kubenswrapper[4907]: I0313 14:58:00.324422 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rjqcf\" (UniqueName: \"kubernetes.io/projected/4efb5dca-e973-443a-bcc2-3b8799fb680c-kube-api-access-rjqcf\") pod \"auto-csr-approver-29556898-55zxj\" (UID: \"4efb5dca-e973-443a-bcc2-3b8799fb680c\") " pod="openshift-infra/auto-csr-approver-29556898-55zxj" Mar 13 14:58:00 crc kubenswrapper[4907]: I0313 14:58:00.461985 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556898-55zxj" Mar 13 14:58:00 crc kubenswrapper[4907]: I0313 14:58:00.874638 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556898-55zxj"] Mar 13 14:58:01 crc kubenswrapper[4907]: I0313 14:58:01.480476 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556898-55zxj" event={"ID":"4efb5dca-e973-443a-bcc2-3b8799fb680c","Type":"ContainerStarted","Data":"f047cdaa087410c849606e22ef215db03f44d03ec14d4cb7f91672c24bbbade7"} Mar 13 14:58:01 crc kubenswrapper[4907]: I0313 14:58:01.786517 4907 scope.go:117] "RemoveContainer" containerID="a44259e1e0a7edc8acdd625ba3f7415d70aa78b760d8eeb81254942722d3acb3" Mar 13 14:58:01 crc kubenswrapper[4907]: E0313 14:58:01.786730 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:58:02 crc kubenswrapper[4907]: I0313 14:58:02.490096 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556898-55zxj" event={"ID":"4efb5dca-e973-443a-bcc2-3b8799fb680c","Type":"ContainerStarted","Data":"770adaf80533ac339976a419aecf995f45ec4e4221e3c81f0b0f3f79b4736ef0"} Mar 13 14:58:02 crc kubenswrapper[4907]: I0313 14:58:02.514527 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29556898-55zxj" podStartSLOduration=1.240730842 podStartE2EDuration="2.51450287s" podCreationTimestamp="2026-03-13 14:58:00 +0000 UTC" firstStartedPulling="2026-03-13 14:58:00.8823956 +0000 UTC m=+3179.782183289" lastFinishedPulling="2026-03-13 14:58:02.156167588 +0000 UTC m=+3181.055955317" observedRunningTime="2026-03-13 14:58:02.511556289 +0000 UTC m=+3181.411343978" watchObservedRunningTime="2026-03-13 14:58:02.51450287 +0000 UTC m=+3181.414290569" Mar 13 14:58:03 crc kubenswrapper[4907]: I0313 14:58:03.503181 4907 generic.go:334] "Generic (PLEG): container finished" podID="4efb5dca-e973-443a-bcc2-3b8799fb680c" containerID="770adaf80533ac339976a419aecf995f45ec4e4221e3c81f0b0f3f79b4736ef0" exitCode=0 Mar 13 14:58:03 crc kubenswrapper[4907]: I0313 14:58:03.503325 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556898-55zxj" event={"ID":"4efb5dca-e973-443a-bcc2-3b8799fb680c","Type":"ContainerDied","Data":"770adaf80533ac339976a419aecf995f45ec4e4221e3c81f0b0f3f79b4736ef0"} Mar 13 14:58:04 crc kubenswrapper[4907]: I0313 14:58:04.767370 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556898-55zxj" Mar 13 14:58:04 crc kubenswrapper[4907]: I0313 14:58:04.861225 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556892-gq5qv"] Mar 13 14:58:04 crc kubenswrapper[4907]: I0313 14:58:04.867842 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rjqcf\" (UniqueName: \"kubernetes.io/projected/4efb5dca-e973-443a-bcc2-3b8799fb680c-kube-api-access-rjqcf\") pod \"4efb5dca-e973-443a-bcc2-3b8799fb680c\" (UID: \"4efb5dca-e973-443a-bcc2-3b8799fb680c\") " Mar 13 14:58:04 crc kubenswrapper[4907]: I0313 14:58:04.873949 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556892-gq5qv"] Mar 13 14:58:04 crc kubenswrapper[4907]: I0313 14:58:04.879314 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4efb5dca-e973-443a-bcc2-3b8799fb680c-kube-api-access-rjqcf" (OuterVolumeSpecName: "kube-api-access-rjqcf") pod "4efb5dca-e973-443a-bcc2-3b8799fb680c" (UID: "4efb5dca-e973-443a-bcc2-3b8799fb680c"). InnerVolumeSpecName "kube-api-access-rjqcf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:58:04 crc kubenswrapper[4907]: I0313 14:58:04.970760 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rjqcf\" (UniqueName: \"kubernetes.io/projected/4efb5dca-e973-443a-bcc2-3b8799fb680c-kube-api-access-rjqcf\") on node \"crc\" DevicePath \"\"" Mar 13 14:58:05 crc kubenswrapper[4907]: I0313 14:58:05.521429 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556898-55zxj" event={"ID":"4efb5dca-e973-443a-bcc2-3b8799fb680c","Type":"ContainerDied","Data":"f047cdaa087410c849606e22ef215db03f44d03ec14d4cb7f91672c24bbbade7"} Mar 13 14:58:05 crc kubenswrapper[4907]: I0313 14:58:05.521911 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f047cdaa087410c849606e22ef215db03f44d03ec14d4cb7f91672c24bbbade7" Mar 13 14:58:05 crc kubenswrapper[4907]: I0313 14:58:05.521549 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556898-55zxj" Mar 13 14:58:05 crc kubenswrapper[4907]: I0313 14:58:05.793990 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cbb8b9ec-85bd-4617-952d-57ce0bdb4553" path="/var/lib/kubelet/pods/cbb8b9ec-85bd-4617-952d-57ce0bdb4553/volumes" Mar 13 14:58:13 crc kubenswrapper[4907]: I0313 14:58:13.782830 4907 scope.go:117] "RemoveContainer" containerID="a44259e1e0a7edc8acdd625ba3f7415d70aa78b760d8eeb81254942722d3acb3" Mar 13 14:58:13 crc kubenswrapper[4907]: E0313 14:58:13.783416 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:58:25 crc kubenswrapper[4907]: I0313 14:58:25.798765 4907 scope.go:117] "RemoveContainer" containerID="2c73e4b83765c168348350266e6ee7d5565c27f33d1acbef663a50d336ac5e75" Mar 13 14:58:26 crc kubenswrapper[4907]: I0313 14:58:26.786407 4907 scope.go:117] "RemoveContainer" containerID="a44259e1e0a7edc8acdd625ba3f7415d70aa78b760d8eeb81254942722d3acb3" Mar 13 14:58:26 crc kubenswrapper[4907]: E0313 14:58:26.786678 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:58:40 crc kubenswrapper[4907]: I0313 14:58:40.782938 4907 scope.go:117] "RemoveContainer" containerID="a44259e1e0a7edc8acdd625ba3f7415d70aa78b760d8eeb81254942722d3acb3" Mar 13 14:58:40 crc kubenswrapper[4907]: E0313 14:58:40.783705 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:58:52 crc kubenswrapper[4907]: I0313 14:58:52.782994 4907 scope.go:117] "RemoveContainer" containerID="a44259e1e0a7edc8acdd625ba3f7415d70aa78b760d8eeb81254942722d3acb3" Mar 13 14:58:52 crc kubenswrapper[4907]: E0313 14:58:52.784946 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:59:07 crc kubenswrapper[4907]: I0313 14:59:07.783352 4907 scope.go:117] "RemoveContainer" containerID="a44259e1e0a7edc8acdd625ba3f7415d70aa78b760d8eeb81254942722d3acb3" Mar 13 14:59:07 crc kubenswrapper[4907]: E0313 14:59:07.784162 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:59:19 crc kubenswrapper[4907]: I0313 14:59:19.782731 4907 scope.go:117] "RemoveContainer" containerID="a44259e1e0a7edc8acdd625ba3f7415d70aa78b760d8eeb81254942722d3acb3" Mar 13 14:59:19 crc kubenswrapper[4907]: E0313 14:59:19.783405 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:59:33 crc kubenswrapper[4907]: I0313 14:59:33.782412 4907 scope.go:117] "RemoveContainer" containerID="a44259e1e0a7edc8acdd625ba3f7415d70aa78b760d8eeb81254942722d3acb3" Mar 13 14:59:33 crc kubenswrapper[4907]: E0313 14:59:33.783164 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:59:47 crc kubenswrapper[4907]: I0313 14:59:47.782328 4907 scope.go:117] "RemoveContainer" containerID="a44259e1e0a7edc8acdd625ba3f7415d70aa78b760d8eeb81254942722d3acb3" Mar 13 14:59:47 crc kubenswrapper[4907]: E0313 14:59:47.784597 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:00:00 crc kubenswrapper[4907]: I0313 15:00:00.185013 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556900-thmpg"] Mar 13 15:00:00 crc kubenswrapper[4907]: E0313 15:00:00.185744 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4efb5dca-e973-443a-bcc2-3b8799fb680c" containerName="oc" Mar 13 15:00:00 crc kubenswrapper[4907]: I0313 15:00:00.185761 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="4efb5dca-e973-443a-bcc2-3b8799fb680c" containerName="oc" Mar 13 15:00:00 crc kubenswrapper[4907]: I0313 15:00:00.185969 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="4efb5dca-e973-443a-bcc2-3b8799fb680c" containerName="oc" Mar 13 15:00:00 crc kubenswrapper[4907]: I0313 15:00:00.186670 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556900-thmpg" Mar 13 15:00:00 crc kubenswrapper[4907]: I0313 15:00:00.190895 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 15:00:00 crc kubenswrapper[4907]: I0313 15:00:00.191259 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 15:00:00 crc kubenswrapper[4907]: I0313 15:00:00.192132 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 15:00:00 crc kubenswrapper[4907]: I0313 15:00:00.192369 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556900-cxbzm"] Mar 13 15:00:00 crc kubenswrapper[4907]: I0313 15:00:00.193343 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556900-cxbzm" Mar 13 15:00:00 crc kubenswrapper[4907]: I0313 15:00:00.194470 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 13 15:00:00 crc kubenswrapper[4907]: I0313 15:00:00.202030 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556900-thmpg"] Mar 13 15:00:00 crc kubenswrapper[4907]: I0313 15:00:00.204309 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 13 15:00:00 crc kubenswrapper[4907]: I0313 15:00:00.210925 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556900-cxbzm"] Mar 13 15:00:00 crc kubenswrapper[4907]: I0313 15:00:00.282874 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-92pc9\" (UniqueName: \"kubernetes.io/projected/48ae7008-a336-47d8-a3da-fd198db91cba-kube-api-access-92pc9\") pod \"collect-profiles-29556900-cxbzm\" (UID: \"48ae7008-a336-47d8-a3da-fd198db91cba\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556900-cxbzm" Mar 13 15:00:00 crc kubenswrapper[4907]: I0313 15:00:00.282990 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fktlp\" (UniqueName: \"kubernetes.io/projected/aafe442b-c955-456c-ac6d-a31322a5a27e-kube-api-access-fktlp\") pod \"auto-csr-approver-29556900-thmpg\" (UID: \"aafe442b-c955-456c-ac6d-a31322a5a27e\") " pod="openshift-infra/auto-csr-approver-29556900-thmpg" Mar 13 15:00:00 crc kubenswrapper[4907]: I0313 15:00:00.283040 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/48ae7008-a336-47d8-a3da-fd198db91cba-secret-volume\") pod \"collect-profiles-29556900-cxbzm\" (UID: \"48ae7008-a336-47d8-a3da-fd198db91cba\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556900-cxbzm" Mar 13 15:00:00 crc kubenswrapper[4907]: I0313 15:00:00.283072 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/48ae7008-a336-47d8-a3da-fd198db91cba-config-volume\") pod \"collect-profiles-29556900-cxbzm\" (UID: \"48ae7008-a336-47d8-a3da-fd198db91cba\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556900-cxbzm" Mar 13 15:00:00 crc kubenswrapper[4907]: I0313 15:00:00.384342 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fktlp\" (UniqueName: \"kubernetes.io/projected/aafe442b-c955-456c-ac6d-a31322a5a27e-kube-api-access-fktlp\") pod \"auto-csr-approver-29556900-thmpg\" (UID: \"aafe442b-c955-456c-ac6d-a31322a5a27e\") " pod="openshift-infra/auto-csr-approver-29556900-thmpg" Mar 13 15:00:00 crc kubenswrapper[4907]: I0313 15:00:00.384404 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/48ae7008-a336-47d8-a3da-fd198db91cba-secret-volume\") pod \"collect-profiles-29556900-cxbzm\" (UID: \"48ae7008-a336-47d8-a3da-fd198db91cba\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556900-cxbzm" Mar 13 15:00:00 crc kubenswrapper[4907]: I0313 15:00:00.384438 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/48ae7008-a336-47d8-a3da-fd198db91cba-config-volume\") pod \"collect-profiles-29556900-cxbzm\" (UID: \"48ae7008-a336-47d8-a3da-fd198db91cba\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556900-cxbzm" Mar 13 15:00:00 crc kubenswrapper[4907]: I0313 15:00:00.384502 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-92pc9\" (UniqueName: \"kubernetes.io/projected/48ae7008-a336-47d8-a3da-fd198db91cba-kube-api-access-92pc9\") pod \"collect-profiles-29556900-cxbzm\" (UID: \"48ae7008-a336-47d8-a3da-fd198db91cba\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556900-cxbzm" Mar 13 15:00:00 crc kubenswrapper[4907]: I0313 15:00:00.386037 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/48ae7008-a336-47d8-a3da-fd198db91cba-config-volume\") pod \"collect-profiles-29556900-cxbzm\" (UID: \"48ae7008-a336-47d8-a3da-fd198db91cba\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556900-cxbzm" Mar 13 15:00:00 crc kubenswrapper[4907]: I0313 15:00:00.396666 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/48ae7008-a336-47d8-a3da-fd198db91cba-secret-volume\") pod \"collect-profiles-29556900-cxbzm\" (UID: \"48ae7008-a336-47d8-a3da-fd198db91cba\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556900-cxbzm" Mar 13 15:00:00 crc kubenswrapper[4907]: I0313 15:00:00.400860 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fktlp\" (UniqueName: \"kubernetes.io/projected/aafe442b-c955-456c-ac6d-a31322a5a27e-kube-api-access-fktlp\") pod \"auto-csr-approver-29556900-thmpg\" (UID: \"aafe442b-c955-456c-ac6d-a31322a5a27e\") " pod="openshift-infra/auto-csr-approver-29556900-thmpg" Mar 13 15:00:00 crc kubenswrapper[4907]: I0313 15:00:00.400923 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-92pc9\" (UniqueName: \"kubernetes.io/projected/48ae7008-a336-47d8-a3da-fd198db91cba-kube-api-access-92pc9\") pod \"collect-profiles-29556900-cxbzm\" (UID: \"48ae7008-a336-47d8-a3da-fd198db91cba\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556900-cxbzm" Mar 13 15:00:00 crc kubenswrapper[4907]: I0313 15:00:00.517742 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556900-thmpg" Mar 13 15:00:00 crc kubenswrapper[4907]: I0313 15:00:00.523676 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556900-cxbzm" Mar 13 15:00:00 crc kubenswrapper[4907]: I0313 15:00:00.941768 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556900-thmpg"] Mar 13 15:00:01 crc kubenswrapper[4907]: I0313 15:00:01.005177 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556900-cxbzm"] Mar 13 15:00:01 crc kubenswrapper[4907]: W0313 15:00:01.012117 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod48ae7008_a336_47d8_a3da_fd198db91cba.slice/crio-f74ce56bde217c7c03564c264b33895175d6afdd3867b429f1f77de4792d29c8 WatchSource:0}: Error finding container f74ce56bde217c7c03564c264b33895175d6afdd3867b429f1f77de4792d29c8: Status 404 returned error can't find the container with id f74ce56bde217c7c03564c264b33895175d6afdd3867b429f1f77de4792d29c8 Mar 13 15:00:01 crc kubenswrapper[4907]: I0313 15:00:01.301681 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556900-thmpg" event={"ID":"aafe442b-c955-456c-ac6d-a31322a5a27e","Type":"ContainerStarted","Data":"500cd8b2f75b1b643859a926fad44ef0385176e754d864d21bafc545c9b9ead7"} Mar 13 15:00:01 crc kubenswrapper[4907]: I0313 15:00:01.304012 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556900-cxbzm" event={"ID":"48ae7008-a336-47d8-a3da-fd198db91cba","Type":"ContainerStarted","Data":"189d9f0966abd854ff0ae3d577047f2f411f08ff9ec8a2c05b10583fd1185431"} Mar 13 15:00:01 crc kubenswrapper[4907]: I0313 15:00:01.304056 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556900-cxbzm" event={"ID":"48ae7008-a336-47d8-a3da-fd198db91cba","Type":"ContainerStarted","Data":"f74ce56bde217c7c03564c264b33895175d6afdd3867b429f1f77de4792d29c8"} Mar 13 15:00:01 crc kubenswrapper[4907]: I0313 15:00:01.324971 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29556900-cxbzm" podStartSLOduration=1.324953346 podStartE2EDuration="1.324953346s" podCreationTimestamp="2026-03-13 15:00:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:00:01.32002831 +0000 UTC m=+3300.219816019" watchObservedRunningTime="2026-03-13 15:00:01.324953346 +0000 UTC m=+3300.224741035" Mar 13 15:00:01 crc kubenswrapper[4907]: I0313 15:00:01.785655 4907 scope.go:117] "RemoveContainer" containerID="a44259e1e0a7edc8acdd625ba3f7415d70aa78b760d8eeb81254942722d3acb3" Mar 13 15:00:01 crc kubenswrapper[4907]: E0313 15:00:01.786264 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:00:02 crc kubenswrapper[4907]: I0313 15:00:02.311156 4907 generic.go:334] "Generic (PLEG): container finished" podID="48ae7008-a336-47d8-a3da-fd198db91cba" containerID="189d9f0966abd854ff0ae3d577047f2f411f08ff9ec8a2c05b10583fd1185431" exitCode=0 Mar 13 15:00:02 crc kubenswrapper[4907]: I0313 15:00:02.311224 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556900-cxbzm" event={"ID":"48ae7008-a336-47d8-a3da-fd198db91cba","Type":"ContainerDied","Data":"189d9f0966abd854ff0ae3d577047f2f411f08ff9ec8a2c05b10583fd1185431"} Mar 13 15:00:03 crc kubenswrapper[4907]: I0313 15:00:03.602114 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556900-cxbzm" Mar 13 15:00:03 crc kubenswrapper[4907]: I0313 15:00:03.763422 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-92pc9\" (UniqueName: \"kubernetes.io/projected/48ae7008-a336-47d8-a3da-fd198db91cba-kube-api-access-92pc9\") pod \"48ae7008-a336-47d8-a3da-fd198db91cba\" (UID: \"48ae7008-a336-47d8-a3da-fd198db91cba\") " Mar 13 15:00:03 crc kubenswrapper[4907]: I0313 15:00:03.763553 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/48ae7008-a336-47d8-a3da-fd198db91cba-config-volume\") pod \"48ae7008-a336-47d8-a3da-fd198db91cba\" (UID: \"48ae7008-a336-47d8-a3da-fd198db91cba\") " Mar 13 15:00:03 crc kubenswrapper[4907]: I0313 15:00:03.763587 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/48ae7008-a336-47d8-a3da-fd198db91cba-secret-volume\") pod \"48ae7008-a336-47d8-a3da-fd198db91cba\" (UID: \"48ae7008-a336-47d8-a3da-fd198db91cba\") " Mar 13 15:00:03 crc kubenswrapper[4907]: I0313 15:00:03.764625 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/48ae7008-a336-47d8-a3da-fd198db91cba-config-volume" (OuterVolumeSpecName: "config-volume") pod "48ae7008-a336-47d8-a3da-fd198db91cba" (UID: "48ae7008-a336-47d8-a3da-fd198db91cba"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:00:03 crc kubenswrapper[4907]: I0313 15:00:03.768930 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/48ae7008-a336-47d8-a3da-fd198db91cba-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "48ae7008-a336-47d8-a3da-fd198db91cba" (UID: "48ae7008-a336-47d8-a3da-fd198db91cba"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:00:03 crc kubenswrapper[4907]: I0313 15:00:03.769310 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/48ae7008-a336-47d8-a3da-fd198db91cba-kube-api-access-92pc9" (OuterVolumeSpecName: "kube-api-access-92pc9") pod "48ae7008-a336-47d8-a3da-fd198db91cba" (UID: "48ae7008-a336-47d8-a3da-fd198db91cba"). InnerVolumeSpecName "kube-api-access-92pc9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:00:03 crc kubenswrapper[4907]: I0313 15:00:03.865537 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-92pc9\" (UniqueName: \"kubernetes.io/projected/48ae7008-a336-47d8-a3da-fd198db91cba-kube-api-access-92pc9\") on node \"crc\" DevicePath \"\"" Mar 13 15:00:03 crc kubenswrapper[4907]: I0313 15:00:03.865663 4907 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/48ae7008-a336-47d8-a3da-fd198db91cba-config-volume\") on node \"crc\" DevicePath \"\"" Mar 13 15:00:03 crc kubenswrapper[4907]: I0313 15:00:03.865698 4907 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/48ae7008-a336-47d8-a3da-fd198db91cba-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 13 15:00:04 crc kubenswrapper[4907]: I0313 15:00:04.323116 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556900-cxbzm" event={"ID":"48ae7008-a336-47d8-a3da-fd198db91cba","Type":"ContainerDied","Data":"f74ce56bde217c7c03564c264b33895175d6afdd3867b429f1f77de4792d29c8"} Mar 13 15:00:04 crc kubenswrapper[4907]: I0313 15:00:04.323159 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f74ce56bde217c7c03564c264b33895175d6afdd3867b429f1f77de4792d29c8" Mar 13 15:00:04 crc kubenswrapper[4907]: I0313 15:00:04.323209 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556900-cxbzm" Mar 13 15:00:04 crc kubenswrapper[4907]: I0313 15:00:04.393777 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556855-t8w6z"] Mar 13 15:00:04 crc kubenswrapper[4907]: I0313 15:00:04.398802 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556855-t8w6z"] Mar 13 15:00:05 crc kubenswrapper[4907]: I0313 15:00:05.333207 4907 generic.go:334] "Generic (PLEG): container finished" podID="aafe442b-c955-456c-ac6d-a31322a5a27e" containerID="5fa3c1b84916d90898ac1395ff1cb1a7f88260f563edf9c1fdc38cde419411f5" exitCode=0 Mar 13 15:00:05 crc kubenswrapper[4907]: I0313 15:00:05.333299 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556900-thmpg" event={"ID":"aafe442b-c955-456c-ac6d-a31322a5a27e","Type":"ContainerDied","Data":"5fa3c1b84916d90898ac1395ff1cb1a7f88260f563edf9c1fdc38cde419411f5"} Mar 13 15:00:05 crc kubenswrapper[4907]: I0313 15:00:05.791110 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="80282185-9b05-4101-b906-e069f57e9bae" path="/var/lib/kubelet/pods/80282185-9b05-4101-b906-e069f57e9bae/volumes" Mar 13 15:00:06 crc kubenswrapper[4907]: I0313 15:00:06.588539 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556900-thmpg" Mar 13 15:00:06 crc kubenswrapper[4907]: I0313 15:00:06.706872 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fktlp\" (UniqueName: \"kubernetes.io/projected/aafe442b-c955-456c-ac6d-a31322a5a27e-kube-api-access-fktlp\") pod \"aafe442b-c955-456c-ac6d-a31322a5a27e\" (UID: \"aafe442b-c955-456c-ac6d-a31322a5a27e\") " Mar 13 15:00:06 crc kubenswrapper[4907]: I0313 15:00:06.713035 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aafe442b-c955-456c-ac6d-a31322a5a27e-kube-api-access-fktlp" (OuterVolumeSpecName: "kube-api-access-fktlp") pod "aafe442b-c955-456c-ac6d-a31322a5a27e" (UID: "aafe442b-c955-456c-ac6d-a31322a5a27e"). InnerVolumeSpecName "kube-api-access-fktlp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:00:06 crc kubenswrapper[4907]: I0313 15:00:06.809121 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fktlp\" (UniqueName: \"kubernetes.io/projected/aafe442b-c955-456c-ac6d-a31322a5a27e-kube-api-access-fktlp\") on node \"crc\" DevicePath \"\"" Mar 13 15:00:07 crc kubenswrapper[4907]: I0313 15:00:07.350739 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556900-thmpg" event={"ID":"aafe442b-c955-456c-ac6d-a31322a5a27e","Type":"ContainerDied","Data":"500cd8b2f75b1b643859a926fad44ef0385176e754d864d21bafc545c9b9ead7"} Mar 13 15:00:07 crc kubenswrapper[4907]: I0313 15:00:07.351066 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="500cd8b2f75b1b643859a926fad44ef0385176e754d864d21bafc545c9b9ead7" Mar 13 15:00:07 crc kubenswrapper[4907]: I0313 15:00:07.350818 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556900-thmpg" Mar 13 15:00:07 crc kubenswrapper[4907]: I0313 15:00:07.648544 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556894-8mk4g"] Mar 13 15:00:07 crc kubenswrapper[4907]: I0313 15:00:07.654564 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556894-8mk4g"] Mar 13 15:00:07 crc kubenswrapper[4907]: I0313 15:00:07.794844 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="14af3ced-cea4-46f4-82b2-bdcc55744a4f" path="/var/lib/kubelet/pods/14af3ced-cea4-46f4-82b2-bdcc55744a4f/volumes" Mar 13 15:00:13 crc kubenswrapper[4907]: I0313 15:00:13.783613 4907 scope.go:117] "RemoveContainer" containerID="a44259e1e0a7edc8acdd625ba3f7415d70aa78b760d8eeb81254942722d3acb3" Mar 13 15:00:13 crc kubenswrapper[4907]: E0313 15:00:13.784398 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:00:24 crc kubenswrapper[4907]: I0313 15:00:24.783251 4907 scope.go:117] "RemoveContainer" containerID="a44259e1e0a7edc8acdd625ba3f7415d70aa78b760d8eeb81254942722d3acb3" Mar 13 15:00:25 crc kubenswrapper[4907]: I0313 15:00:25.488096 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerStarted","Data":"62e3208576ab64ce53d2a23023d86e9e8cd79b456c2ef95c5adcb1299b258391"} Mar 13 15:00:25 crc kubenswrapper[4907]: I0313 15:00:25.872785 4907 scope.go:117] "RemoveContainer" containerID="1113c32af3f3a8484a7765dfa4c4a4904cf0b0cad6e1387975333f03729886ff" Mar 13 15:00:25 crc kubenswrapper[4907]: I0313 15:00:25.901168 4907 scope.go:117] "RemoveContainer" containerID="1f3fec4f332e7900eb7076ad444f074ae92511e972577dd77d05eebc0931b626" Mar 13 15:00:49 crc kubenswrapper[4907]: I0313 15:00:49.385426 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-m2b5z"] Mar 13 15:00:49 crc kubenswrapper[4907]: E0313 15:00:49.386293 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aafe442b-c955-456c-ac6d-a31322a5a27e" containerName="oc" Mar 13 15:00:49 crc kubenswrapper[4907]: I0313 15:00:49.386305 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="aafe442b-c955-456c-ac6d-a31322a5a27e" containerName="oc" Mar 13 15:00:49 crc kubenswrapper[4907]: E0313 15:00:49.386328 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48ae7008-a336-47d8-a3da-fd198db91cba" containerName="collect-profiles" Mar 13 15:00:49 crc kubenswrapper[4907]: I0313 15:00:49.386335 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="48ae7008-a336-47d8-a3da-fd198db91cba" containerName="collect-profiles" Mar 13 15:00:49 crc kubenswrapper[4907]: I0313 15:00:49.386529 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="48ae7008-a336-47d8-a3da-fd198db91cba" containerName="collect-profiles" Mar 13 15:00:49 crc kubenswrapper[4907]: I0313 15:00:49.386546 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="aafe442b-c955-456c-ac6d-a31322a5a27e" containerName="oc" Mar 13 15:00:49 crc kubenswrapper[4907]: I0313 15:00:49.387643 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-m2b5z" Mar 13 15:00:49 crc kubenswrapper[4907]: I0313 15:00:49.396199 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-m2b5z"] Mar 13 15:00:49 crc kubenswrapper[4907]: I0313 15:00:49.542504 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98b7807e-6fa0-4bde-8f09-8289ed316462-catalog-content\") pod \"redhat-marketplace-m2b5z\" (UID: \"98b7807e-6fa0-4bde-8f09-8289ed316462\") " pod="openshift-marketplace/redhat-marketplace-m2b5z" Mar 13 15:00:49 crc kubenswrapper[4907]: I0313 15:00:49.542571 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n9wr4\" (UniqueName: \"kubernetes.io/projected/98b7807e-6fa0-4bde-8f09-8289ed316462-kube-api-access-n9wr4\") pod \"redhat-marketplace-m2b5z\" (UID: \"98b7807e-6fa0-4bde-8f09-8289ed316462\") " pod="openshift-marketplace/redhat-marketplace-m2b5z" Mar 13 15:00:49 crc kubenswrapper[4907]: I0313 15:00:49.542658 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98b7807e-6fa0-4bde-8f09-8289ed316462-utilities\") pod \"redhat-marketplace-m2b5z\" (UID: \"98b7807e-6fa0-4bde-8f09-8289ed316462\") " pod="openshift-marketplace/redhat-marketplace-m2b5z" Mar 13 15:00:49 crc kubenswrapper[4907]: I0313 15:00:49.643355 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98b7807e-6fa0-4bde-8f09-8289ed316462-utilities\") pod \"redhat-marketplace-m2b5z\" (UID: \"98b7807e-6fa0-4bde-8f09-8289ed316462\") " pod="openshift-marketplace/redhat-marketplace-m2b5z" Mar 13 15:00:49 crc kubenswrapper[4907]: I0313 15:00:49.643429 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98b7807e-6fa0-4bde-8f09-8289ed316462-catalog-content\") pod \"redhat-marketplace-m2b5z\" (UID: \"98b7807e-6fa0-4bde-8f09-8289ed316462\") " pod="openshift-marketplace/redhat-marketplace-m2b5z" Mar 13 15:00:49 crc kubenswrapper[4907]: I0313 15:00:49.643492 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n9wr4\" (UniqueName: \"kubernetes.io/projected/98b7807e-6fa0-4bde-8f09-8289ed316462-kube-api-access-n9wr4\") pod \"redhat-marketplace-m2b5z\" (UID: \"98b7807e-6fa0-4bde-8f09-8289ed316462\") " pod="openshift-marketplace/redhat-marketplace-m2b5z" Mar 13 15:00:49 crc kubenswrapper[4907]: I0313 15:00:49.643859 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98b7807e-6fa0-4bde-8f09-8289ed316462-utilities\") pod \"redhat-marketplace-m2b5z\" (UID: \"98b7807e-6fa0-4bde-8f09-8289ed316462\") " pod="openshift-marketplace/redhat-marketplace-m2b5z" Mar 13 15:00:49 crc kubenswrapper[4907]: I0313 15:00:49.644024 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98b7807e-6fa0-4bde-8f09-8289ed316462-catalog-content\") pod \"redhat-marketplace-m2b5z\" (UID: \"98b7807e-6fa0-4bde-8f09-8289ed316462\") " pod="openshift-marketplace/redhat-marketplace-m2b5z" Mar 13 15:00:49 crc kubenswrapper[4907]: I0313 15:00:49.663783 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n9wr4\" (UniqueName: \"kubernetes.io/projected/98b7807e-6fa0-4bde-8f09-8289ed316462-kube-api-access-n9wr4\") pod \"redhat-marketplace-m2b5z\" (UID: \"98b7807e-6fa0-4bde-8f09-8289ed316462\") " pod="openshift-marketplace/redhat-marketplace-m2b5z" Mar 13 15:00:49 crc kubenswrapper[4907]: I0313 15:00:49.707908 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-m2b5z" Mar 13 15:00:50 crc kubenswrapper[4907]: I0313 15:00:50.125960 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-m2b5z"] Mar 13 15:00:50 crc kubenswrapper[4907]: I0313 15:00:50.689926 4907 generic.go:334] "Generic (PLEG): container finished" podID="98b7807e-6fa0-4bde-8f09-8289ed316462" containerID="f6c782f8c5b76bea23f8182eb1e7f3a7d38bf37fbaa586bf1d7d22a54504041e" exitCode=0 Mar 13 15:00:50 crc kubenswrapper[4907]: I0313 15:00:50.690045 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m2b5z" event={"ID":"98b7807e-6fa0-4bde-8f09-8289ed316462","Type":"ContainerDied","Data":"f6c782f8c5b76bea23f8182eb1e7f3a7d38bf37fbaa586bf1d7d22a54504041e"} Mar 13 15:00:50 crc kubenswrapper[4907]: I0313 15:00:50.690332 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m2b5z" event={"ID":"98b7807e-6fa0-4bde-8f09-8289ed316462","Type":"ContainerStarted","Data":"6430f9fc5bca0ed430ebc1bfa36b769f8542a5c922626b992fd79be104079725"} Mar 13 15:00:52 crc kubenswrapper[4907]: I0313 15:00:52.706223 4907 generic.go:334] "Generic (PLEG): container finished" podID="98b7807e-6fa0-4bde-8f09-8289ed316462" containerID="8087bdb8f2e06dc3cb2d038da4ecf2ebe93fd908793fddc37698e9e8f496792b" exitCode=0 Mar 13 15:00:52 crc kubenswrapper[4907]: I0313 15:00:52.706341 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m2b5z" event={"ID":"98b7807e-6fa0-4bde-8f09-8289ed316462","Type":"ContainerDied","Data":"8087bdb8f2e06dc3cb2d038da4ecf2ebe93fd908793fddc37698e9e8f496792b"} Mar 13 15:00:53 crc kubenswrapper[4907]: I0313 15:00:53.715837 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m2b5z" event={"ID":"98b7807e-6fa0-4bde-8f09-8289ed316462","Type":"ContainerStarted","Data":"1b9d8269123996da45f8c0e1ef845fe700956a76fbab44efafea4a51c69437b5"} Mar 13 15:00:53 crc kubenswrapper[4907]: I0313 15:00:53.736936 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-m2b5z" podStartSLOduration=2.302601292 podStartE2EDuration="4.736919185s" podCreationTimestamp="2026-03-13 15:00:49 +0000 UTC" firstStartedPulling="2026-03-13 15:00:50.692777377 +0000 UTC m=+3349.592565066" lastFinishedPulling="2026-03-13 15:00:53.12709527 +0000 UTC m=+3352.026882959" observedRunningTime="2026-03-13 15:00:53.731445845 +0000 UTC m=+3352.631233554" watchObservedRunningTime="2026-03-13 15:00:53.736919185 +0000 UTC m=+3352.636706884" Mar 13 15:00:59 crc kubenswrapper[4907]: I0313 15:00:59.707997 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-m2b5z" Mar 13 15:00:59 crc kubenswrapper[4907]: I0313 15:00:59.708291 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-m2b5z" Mar 13 15:00:59 crc kubenswrapper[4907]: I0313 15:00:59.753764 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-m2b5z" Mar 13 15:00:59 crc kubenswrapper[4907]: I0313 15:00:59.839336 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-m2b5z" Mar 13 15:01:00 crc kubenswrapper[4907]: I0313 15:01:00.023862 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-m2b5z"] Mar 13 15:01:01 crc kubenswrapper[4907]: I0313 15:01:01.790654 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-m2b5z" podUID="98b7807e-6fa0-4bde-8f09-8289ed316462" containerName="registry-server" containerID="cri-o://1b9d8269123996da45f8c0e1ef845fe700956a76fbab44efafea4a51c69437b5" gracePeriod=2 Mar 13 15:01:02 crc kubenswrapper[4907]: I0313 15:01:02.194284 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-m2b5z" Mar 13 15:01:02 crc kubenswrapper[4907]: I0313 15:01:02.320838 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98b7807e-6fa0-4bde-8f09-8289ed316462-utilities\") pod \"98b7807e-6fa0-4bde-8f09-8289ed316462\" (UID: \"98b7807e-6fa0-4bde-8f09-8289ed316462\") " Mar 13 15:01:02 crc kubenswrapper[4907]: I0313 15:01:02.321334 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98b7807e-6fa0-4bde-8f09-8289ed316462-catalog-content\") pod \"98b7807e-6fa0-4bde-8f09-8289ed316462\" (UID: \"98b7807e-6fa0-4bde-8f09-8289ed316462\") " Mar 13 15:01:02 crc kubenswrapper[4907]: I0313 15:01:02.321459 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n9wr4\" (UniqueName: \"kubernetes.io/projected/98b7807e-6fa0-4bde-8f09-8289ed316462-kube-api-access-n9wr4\") pod \"98b7807e-6fa0-4bde-8f09-8289ed316462\" (UID: \"98b7807e-6fa0-4bde-8f09-8289ed316462\") " Mar 13 15:01:02 crc kubenswrapper[4907]: I0313 15:01:02.322359 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/98b7807e-6fa0-4bde-8f09-8289ed316462-utilities" (OuterVolumeSpecName: "utilities") pod "98b7807e-6fa0-4bde-8f09-8289ed316462" (UID: "98b7807e-6fa0-4bde-8f09-8289ed316462"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:01:02 crc kubenswrapper[4907]: I0313 15:01:02.328095 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/98b7807e-6fa0-4bde-8f09-8289ed316462-kube-api-access-n9wr4" (OuterVolumeSpecName: "kube-api-access-n9wr4") pod "98b7807e-6fa0-4bde-8f09-8289ed316462" (UID: "98b7807e-6fa0-4bde-8f09-8289ed316462"). InnerVolumeSpecName "kube-api-access-n9wr4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:01:02 crc kubenswrapper[4907]: I0313 15:01:02.347634 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/98b7807e-6fa0-4bde-8f09-8289ed316462-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "98b7807e-6fa0-4bde-8f09-8289ed316462" (UID: "98b7807e-6fa0-4bde-8f09-8289ed316462"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:01:02 crc kubenswrapper[4907]: I0313 15:01:02.422947 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n9wr4\" (UniqueName: \"kubernetes.io/projected/98b7807e-6fa0-4bde-8f09-8289ed316462-kube-api-access-n9wr4\") on node \"crc\" DevicePath \"\"" Mar 13 15:01:02 crc kubenswrapper[4907]: I0313 15:01:02.422984 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98b7807e-6fa0-4bde-8f09-8289ed316462-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 15:01:02 crc kubenswrapper[4907]: I0313 15:01:02.422993 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98b7807e-6fa0-4bde-8f09-8289ed316462-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 15:01:02 crc kubenswrapper[4907]: I0313 15:01:02.797795 4907 generic.go:334] "Generic (PLEG): container finished" podID="98b7807e-6fa0-4bde-8f09-8289ed316462" containerID="1b9d8269123996da45f8c0e1ef845fe700956a76fbab44efafea4a51c69437b5" exitCode=0 Mar 13 15:01:02 crc kubenswrapper[4907]: I0313 15:01:02.797835 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m2b5z" event={"ID":"98b7807e-6fa0-4bde-8f09-8289ed316462","Type":"ContainerDied","Data":"1b9d8269123996da45f8c0e1ef845fe700956a76fbab44efafea4a51c69437b5"} Mar 13 15:01:02 crc kubenswrapper[4907]: I0313 15:01:02.797856 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-m2b5z" Mar 13 15:01:02 crc kubenswrapper[4907]: I0313 15:01:02.797874 4907 scope.go:117] "RemoveContainer" containerID="1b9d8269123996da45f8c0e1ef845fe700956a76fbab44efafea4a51c69437b5" Mar 13 15:01:02 crc kubenswrapper[4907]: I0313 15:01:02.797863 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m2b5z" event={"ID":"98b7807e-6fa0-4bde-8f09-8289ed316462","Type":"ContainerDied","Data":"6430f9fc5bca0ed430ebc1bfa36b769f8542a5c922626b992fd79be104079725"} Mar 13 15:01:02 crc kubenswrapper[4907]: I0313 15:01:02.815650 4907 scope.go:117] "RemoveContainer" containerID="8087bdb8f2e06dc3cb2d038da4ecf2ebe93fd908793fddc37698e9e8f496792b" Mar 13 15:01:02 crc kubenswrapper[4907]: I0313 15:01:02.833845 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-m2b5z"] Mar 13 15:01:02 crc kubenswrapper[4907]: I0313 15:01:02.839170 4907 scope.go:117] "RemoveContainer" containerID="f6c782f8c5b76bea23f8182eb1e7f3a7d38bf37fbaa586bf1d7d22a54504041e" Mar 13 15:01:02 crc kubenswrapper[4907]: I0313 15:01:02.839558 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-m2b5z"] Mar 13 15:01:02 crc kubenswrapper[4907]: I0313 15:01:02.861442 4907 scope.go:117] "RemoveContainer" containerID="1b9d8269123996da45f8c0e1ef845fe700956a76fbab44efafea4a51c69437b5" Mar 13 15:01:02 crc kubenswrapper[4907]: E0313 15:01:02.861793 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1b9d8269123996da45f8c0e1ef845fe700956a76fbab44efafea4a51c69437b5\": container with ID starting with 1b9d8269123996da45f8c0e1ef845fe700956a76fbab44efafea4a51c69437b5 not found: ID does not exist" containerID="1b9d8269123996da45f8c0e1ef845fe700956a76fbab44efafea4a51c69437b5" Mar 13 15:01:02 crc kubenswrapper[4907]: I0313 15:01:02.861830 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1b9d8269123996da45f8c0e1ef845fe700956a76fbab44efafea4a51c69437b5"} err="failed to get container status \"1b9d8269123996da45f8c0e1ef845fe700956a76fbab44efafea4a51c69437b5\": rpc error: code = NotFound desc = could not find container \"1b9d8269123996da45f8c0e1ef845fe700956a76fbab44efafea4a51c69437b5\": container with ID starting with 1b9d8269123996da45f8c0e1ef845fe700956a76fbab44efafea4a51c69437b5 not found: ID does not exist" Mar 13 15:01:02 crc kubenswrapper[4907]: I0313 15:01:02.861860 4907 scope.go:117] "RemoveContainer" containerID="8087bdb8f2e06dc3cb2d038da4ecf2ebe93fd908793fddc37698e9e8f496792b" Mar 13 15:01:02 crc kubenswrapper[4907]: E0313 15:01:02.862379 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8087bdb8f2e06dc3cb2d038da4ecf2ebe93fd908793fddc37698e9e8f496792b\": container with ID starting with 8087bdb8f2e06dc3cb2d038da4ecf2ebe93fd908793fddc37698e9e8f496792b not found: ID does not exist" containerID="8087bdb8f2e06dc3cb2d038da4ecf2ebe93fd908793fddc37698e9e8f496792b" Mar 13 15:01:02 crc kubenswrapper[4907]: I0313 15:01:02.862448 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8087bdb8f2e06dc3cb2d038da4ecf2ebe93fd908793fddc37698e9e8f496792b"} err="failed to get container status \"8087bdb8f2e06dc3cb2d038da4ecf2ebe93fd908793fddc37698e9e8f496792b\": rpc error: code = NotFound desc = could not find container \"8087bdb8f2e06dc3cb2d038da4ecf2ebe93fd908793fddc37698e9e8f496792b\": container with ID starting with 8087bdb8f2e06dc3cb2d038da4ecf2ebe93fd908793fddc37698e9e8f496792b not found: ID does not exist" Mar 13 15:01:02 crc kubenswrapper[4907]: I0313 15:01:02.862471 4907 scope.go:117] "RemoveContainer" containerID="f6c782f8c5b76bea23f8182eb1e7f3a7d38bf37fbaa586bf1d7d22a54504041e" Mar 13 15:01:02 crc kubenswrapper[4907]: E0313 15:01:02.862726 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f6c782f8c5b76bea23f8182eb1e7f3a7d38bf37fbaa586bf1d7d22a54504041e\": container with ID starting with f6c782f8c5b76bea23f8182eb1e7f3a7d38bf37fbaa586bf1d7d22a54504041e not found: ID does not exist" containerID="f6c782f8c5b76bea23f8182eb1e7f3a7d38bf37fbaa586bf1d7d22a54504041e" Mar 13 15:01:02 crc kubenswrapper[4907]: I0313 15:01:02.862749 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f6c782f8c5b76bea23f8182eb1e7f3a7d38bf37fbaa586bf1d7d22a54504041e"} err="failed to get container status \"f6c782f8c5b76bea23f8182eb1e7f3a7d38bf37fbaa586bf1d7d22a54504041e\": rpc error: code = NotFound desc = could not find container \"f6c782f8c5b76bea23f8182eb1e7f3a7d38bf37fbaa586bf1d7d22a54504041e\": container with ID starting with f6c782f8c5b76bea23f8182eb1e7f3a7d38bf37fbaa586bf1d7d22a54504041e not found: ID does not exist" Mar 13 15:01:03 crc kubenswrapper[4907]: I0313 15:01:03.793842 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="98b7807e-6fa0-4bde-8f09-8289ed316462" path="/var/lib/kubelet/pods/98b7807e-6fa0-4bde-8f09-8289ed316462/volumes" Mar 13 15:01:10 crc kubenswrapper[4907]: E0313 15:01:10.990951 4907 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod98b7807e_6fa0_4bde_8f09_8289ed316462.slice/crio-conmon-1b9d8269123996da45f8c0e1ef845fe700956a76fbab44efafea4a51c69437b5.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod98b7807e_6fa0_4bde_8f09_8289ed316462.slice/crio-1b9d8269123996da45f8c0e1ef845fe700956a76fbab44efafea4a51c69437b5.scope\": RecentStats: unable to find data in memory cache]" Mar 13 15:01:21 crc kubenswrapper[4907]: E0313 15:01:21.175499 4907 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod98b7807e_6fa0_4bde_8f09_8289ed316462.slice/crio-conmon-1b9d8269123996da45f8c0e1ef845fe700956a76fbab44efafea4a51c69437b5.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod98b7807e_6fa0_4bde_8f09_8289ed316462.slice/crio-1b9d8269123996da45f8c0e1ef845fe700956a76fbab44efafea4a51c69437b5.scope\": RecentStats: unable to find data in memory cache]" Mar 13 15:01:31 crc kubenswrapper[4907]: E0313 15:01:31.332831 4907 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod98b7807e_6fa0_4bde_8f09_8289ed316462.slice/crio-1b9d8269123996da45f8c0e1ef845fe700956a76fbab44efafea4a51c69437b5.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod98b7807e_6fa0_4bde_8f09_8289ed316462.slice/crio-conmon-1b9d8269123996da45f8c0e1ef845fe700956a76fbab44efafea4a51c69437b5.scope\": RecentStats: unable to find data in memory cache]" Mar 13 15:01:41 crc kubenswrapper[4907]: E0313 15:01:41.483739 4907 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod98b7807e_6fa0_4bde_8f09_8289ed316462.slice/crio-1b9d8269123996da45f8c0e1ef845fe700956a76fbab44efafea4a51c69437b5.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod98b7807e_6fa0_4bde_8f09_8289ed316462.slice/crio-conmon-1b9d8269123996da45f8c0e1ef845fe700956a76fbab44efafea4a51c69437b5.scope\": RecentStats: unable to find data in memory cache]" Mar 13 15:01:51 crc kubenswrapper[4907]: E0313 15:01:51.670101 4907 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod98b7807e_6fa0_4bde_8f09_8289ed316462.slice/crio-1b9d8269123996da45f8c0e1ef845fe700956a76fbab44efafea4a51c69437b5.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod98b7807e_6fa0_4bde_8f09_8289ed316462.slice/crio-conmon-1b9d8269123996da45f8c0e1ef845fe700956a76fbab44efafea4a51c69437b5.scope\": RecentStats: unable to find data in memory cache]" Mar 13 15:02:00 crc kubenswrapper[4907]: I0313 15:02:00.139651 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556902-92pzz"] Mar 13 15:02:00 crc kubenswrapper[4907]: E0313 15:02:00.142065 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98b7807e-6fa0-4bde-8f09-8289ed316462" containerName="extract-content" Mar 13 15:02:00 crc kubenswrapper[4907]: I0313 15:02:00.142165 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="98b7807e-6fa0-4bde-8f09-8289ed316462" containerName="extract-content" Mar 13 15:02:00 crc kubenswrapper[4907]: E0313 15:02:00.142258 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98b7807e-6fa0-4bde-8f09-8289ed316462" containerName="registry-server" Mar 13 15:02:00 crc kubenswrapper[4907]: I0313 15:02:00.142317 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="98b7807e-6fa0-4bde-8f09-8289ed316462" containerName="registry-server" Mar 13 15:02:00 crc kubenswrapper[4907]: E0313 15:02:00.142483 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98b7807e-6fa0-4bde-8f09-8289ed316462" containerName="extract-utilities" Mar 13 15:02:00 crc kubenswrapper[4907]: I0313 15:02:00.142560 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="98b7807e-6fa0-4bde-8f09-8289ed316462" containerName="extract-utilities" Mar 13 15:02:00 crc kubenswrapper[4907]: I0313 15:02:00.142745 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="98b7807e-6fa0-4bde-8f09-8289ed316462" containerName="registry-server" Mar 13 15:02:00 crc kubenswrapper[4907]: I0313 15:02:00.143377 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556902-92pzz" Mar 13 15:02:00 crc kubenswrapper[4907]: I0313 15:02:00.146928 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 15:02:00 crc kubenswrapper[4907]: I0313 15:02:00.146986 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 15:02:00 crc kubenswrapper[4907]: I0313 15:02:00.148620 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 15:02:00 crc kubenswrapper[4907]: I0313 15:02:00.149054 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556902-92pzz"] Mar 13 15:02:00 crc kubenswrapper[4907]: I0313 15:02:00.257325 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qxf2h\" (UniqueName: \"kubernetes.io/projected/076c2b87-8e1f-4dba-9b60-eb24660aa1f1-kube-api-access-qxf2h\") pod \"auto-csr-approver-29556902-92pzz\" (UID: \"076c2b87-8e1f-4dba-9b60-eb24660aa1f1\") " pod="openshift-infra/auto-csr-approver-29556902-92pzz" Mar 13 15:02:00 crc kubenswrapper[4907]: I0313 15:02:00.358634 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qxf2h\" (UniqueName: \"kubernetes.io/projected/076c2b87-8e1f-4dba-9b60-eb24660aa1f1-kube-api-access-qxf2h\") pod \"auto-csr-approver-29556902-92pzz\" (UID: \"076c2b87-8e1f-4dba-9b60-eb24660aa1f1\") " pod="openshift-infra/auto-csr-approver-29556902-92pzz" Mar 13 15:02:00 crc kubenswrapper[4907]: I0313 15:02:00.376121 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qxf2h\" (UniqueName: \"kubernetes.io/projected/076c2b87-8e1f-4dba-9b60-eb24660aa1f1-kube-api-access-qxf2h\") pod \"auto-csr-approver-29556902-92pzz\" (UID: \"076c2b87-8e1f-4dba-9b60-eb24660aa1f1\") " pod="openshift-infra/auto-csr-approver-29556902-92pzz" Mar 13 15:02:00 crc kubenswrapper[4907]: I0313 15:02:00.469351 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556902-92pzz" Mar 13 15:02:00 crc kubenswrapper[4907]: I0313 15:02:00.924397 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556902-92pzz"] Mar 13 15:02:00 crc kubenswrapper[4907]: I0313 15:02:00.935203 4907 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 13 15:02:01 crc kubenswrapper[4907]: I0313 15:02:01.257369 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556902-92pzz" event={"ID":"076c2b87-8e1f-4dba-9b60-eb24660aa1f1","Type":"ContainerStarted","Data":"bd9a11fb2fb4bd3d8600ccfeb75929b4157aca3412cb52bb36e9d792add6c662"} Mar 13 15:02:01 crc kubenswrapper[4907]: E0313 15:02:01.853672 4907 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod98b7807e_6fa0_4bde_8f09_8289ed316462.slice/crio-conmon-1b9d8269123996da45f8c0e1ef845fe700956a76fbab44efafea4a51c69437b5.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod98b7807e_6fa0_4bde_8f09_8289ed316462.slice/crio-1b9d8269123996da45f8c0e1ef845fe700956a76fbab44efafea4a51c69437b5.scope\": RecentStats: unable to find data in memory cache]" Mar 13 15:02:02 crc kubenswrapper[4907]: I0313 15:02:02.268777 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556902-92pzz" event={"ID":"076c2b87-8e1f-4dba-9b60-eb24660aa1f1","Type":"ContainerStarted","Data":"be729cf40227f8298124611ddf4a79928c79ddf05e8dab4417abc40fd127f33b"} Mar 13 15:02:02 crc kubenswrapper[4907]: I0313 15:02:02.284871 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29556902-92pzz" podStartSLOduration=1.3925067819999999 podStartE2EDuration="2.284841431s" podCreationTimestamp="2026-03-13 15:02:00 +0000 UTC" firstStartedPulling="2026-03-13 15:02:00.934531009 +0000 UTC m=+3419.834318688" lastFinishedPulling="2026-03-13 15:02:01.826865648 +0000 UTC m=+3420.726653337" observedRunningTime="2026-03-13 15:02:02.280559423 +0000 UTC m=+3421.180347112" watchObservedRunningTime="2026-03-13 15:02:02.284841431 +0000 UTC m=+3421.184629120" Mar 13 15:02:03 crc kubenswrapper[4907]: I0313 15:02:03.278676 4907 generic.go:334] "Generic (PLEG): container finished" podID="076c2b87-8e1f-4dba-9b60-eb24660aa1f1" containerID="be729cf40227f8298124611ddf4a79928c79ddf05e8dab4417abc40fd127f33b" exitCode=0 Mar 13 15:02:03 crc kubenswrapper[4907]: I0313 15:02:03.278771 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556902-92pzz" event={"ID":"076c2b87-8e1f-4dba-9b60-eb24660aa1f1","Type":"ContainerDied","Data":"be729cf40227f8298124611ddf4a79928c79ddf05e8dab4417abc40fd127f33b"} Mar 13 15:02:05 crc kubenswrapper[4907]: I0313 15:02:04.586305 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556902-92pzz" Mar 13 15:02:05 crc kubenswrapper[4907]: I0313 15:02:04.724658 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qxf2h\" (UniqueName: \"kubernetes.io/projected/076c2b87-8e1f-4dba-9b60-eb24660aa1f1-kube-api-access-qxf2h\") pod \"076c2b87-8e1f-4dba-9b60-eb24660aa1f1\" (UID: \"076c2b87-8e1f-4dba-9b60-eb24660aa1f1\") " Mar 13 15:02:05 crc kubenswrapper[4907]: I0313 15:02:04.730294 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/076c2b87-8e1f-4dba-9b60-eb24660aa1f1-kube-api-access-qxf2h" (OuterVolumeSpecName: "kube-api-access-qxf2h") pod "076c2b87-8e1f-4dba-9b60-eb24660aa1f1" (UID: "076c2b87-8e1f-4dba-9b60-eb24660aa1f1"). InnerVolumeSpecName "kube-api-access-qxf2h". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:02:05 crc kubenswrapper[4907]: I0313 15:02:04.830654 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qxf2h\" (UniqueName: \"kubernetes.io/projected/076c2b87-8e1f-4dba-9b60-eb24660aa1f1-kube-api-access-qxf2h\") on node \"crc\" DevicePath \"\"" Mar 13 15:02:05 crc kubenswrapper[4907]: I0313 15:02:04.872573 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556896-7dlx2"] Mar 13 15:02:05 crc kubenswrapper[4907]: I0313 15:02:04.880988 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556896-7dlx2"] Mar 13 15:02:05 crc kubenswrapper[4907]: I0313 15:02:05.292605 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556902-92pzz" event={"ID":"076c2b87-8e1f-4dba-9b60-eb24660aa1f1","Type":"ContainerDied","Data":"bd9a11fb2fb4bd3d8600ccfeb75929b4157aca3412cb52bb36e9d792add6c662"} Mar 13 15:02:05 crc kubenswrapper[4907]: I0313 15:02:05.292669 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bd9a11fb2fb4bd3d8600ccfeb75929b4157aca3412cb52bb36e9d792add6c662" Mar 13 15:02:05 crc kubenswrapper[4907]: I0313 15:02:05.292669 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556902-92pzz" Mar 13 15:02:05 crc kubenswrapper[4907]: I0313 15:02:05.790448 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e193f9f5-a033-4f42-9932-62f7b4edd0b2" path="/var/lib/kubelet/pods/e193f9f5-a033-4f42-9932-62f7b4edd0b2/volumes" Mar 13 15:02:25 crc kubenswrapper[4907]: I0313 15:02:25.996314 4907 scope.go:117] "RemoveContainer" containerID="19c525d3167fb54f80c181f5117cf495aa5c0602e81951a3395f3a05b6fbc6c9" Mar 13 15:02:48 crc kubenswrapper[4907]: I0313 15:02:48.041668 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 15:02:48 crc kubenswrapper[4907]: I0313 15:02:48.042261 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 15:03:18 crc kubenswrapper[4907]: I0313 15:03:18.042455 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 15:03:18 crc kubenswrapper[4907]: I0313 15:03:18.043077 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 15:03:27 crc kubenswrapper[4907]: I0313 15:03:27.772415 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-66vvh"] Mar 13 15:03:27 crc kubenswrapper[4907]: E0313 15:03:27.774189 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="076c2b87-8e1f-4dba-9b60-eb24660aa1f1" containerName="oc" Mar 13 15:03:27 crc kubenswrapper[4907]: I0313 15:03:27.774275 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="076c2b87-8e1f-4dba-9b60-eb24660aa1f1" containerName="oc" Mar 13 15:03:27 crc kubenswrapper[4907]: I0313 15:03:27.774550 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="076c2b87-8e1f-4dba-9b60-eb24660aa1f1" containerName="oc" Mar 13 15:03:27 crc kubenswrapper[4907]: I0313 15:03:27.775977 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-66vvh" Mar 13 15:03:27 crc kubenswrapper[4907]: I0313 15:03:27.800243 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-66vvh"] Mar 13 15:03:27 crc kubenswrapper[4907]: I0313 15:03:27.887784 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7a96a53-ed40-4b76-b1f3-58edd7dec59c-utilities\") pod \"redhat-operators-66vvh\" (UID: \"e7a96a53-ed40-4b76-b1f3-58edd7dec59c\") " pod="openshift-marketplace/redhat-operators-66vvh" Mar 13 15:03:27 crc kubenswrapper[4907]: I0313 15:03:27.888359 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j5wrg\" (UniqueName: \"kubernetes.io/projected/e7a96a53-ed40-4b76-b1f3-58edd7dec59c-kube-api-access-j5wrg\") pod \"redhat-operators-66vvh\" (UID: \"e7a96a53-ed40-4b76-b1f3-58edd7dec59c\") " pod="openshift-marketplace/redhat-operators-66vvh" Mar 13 15:03:27 crc kubenswrapper[4907]: I0313 15:03:27.888512 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7a96a53-ed40-4b76-b1f3-58edd7dec59c-catalog-content\") pod \"redhat-operators-66vvh\" (UID: \"e7a96a53-ed40-4b76-b1f3-58edd7dec59c\") " pod="openshift-marketplace/redhat-operators-66vvh" Mar 13 15:03:27 crc kubenswrapper[4907]: I0313 15:03:27.989513 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j5wrg\" (UniqueName: \"kubernetes.io/projected/e7a96a53-ed40-4b76-b1f3-58edd7dec59c-kube-api-access-j5wrg\") pod \"redhat-operators-66vvh\" (UID: \"e7a96a53-ed40-4b76-b1f3-58edd7dec59c\") " pod="openshift-marketplace/redhat-operators-66vvh" Mar 13 15:03:27 crc kubenswrapper[4907]: I0313 15:03:27.989565 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7a96a53-ed40-4b76-b1f3-58edd7dec59c-catalog-content\") pod \"redhat-operators-66vvh\" (UID: \"e7a96a53-ed40-4b76-b1f3-58edd7dec59c\") " pod="openshift-marketplace/redhat-operators-66vvh" Mar 13 15:03:27 crc kubenswrapper[4907]: I0313 15:03:27.989618 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7a96a53-ed40-4b76-b1f3-58edd7dec59c-utilities\") pod \"redhat-operators-66vvh\" (UID: \"e7a96a53-ed40-4b76-b1f3-58edd7dec59c\") " pod="openshift-marketplace/redhat-operators-66vvh" Mar 13 15:03:27 crc kubenswrapper[4907]: I0313 15:03:27.990164 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7a96a53-ed40-4b76-b1f3-58edd7dec59c-utilities\") pod \"redhat-operators-66vvh\" (UID: \"e7a96a53-ed40-4b76-b1f3-58edd7dec59c\") " pod="openshift-marketplace/redhat-operators-66vvh" Mar 13 15:03:27 crc kubenswrapper[4907]: I0313 15:03:27.990279 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7a96a53-ed40-4b76-b1f3-58edd7dec59c-catalog-content\") pod \"redhat-operators-66vvh\" (UID: \"e7a96a53-ed40-4b76-b1f3-58edd7dec59c\") " pod="openshift-marketplace/redhat-operators-66vvh" Mar 13 15:03:28 crc kubenswrapper[4907]: I0313 15:03:28.011233 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j5wrg\" (UniqueName: \"kubernetes.io/projected/e7a96a53-ed40-4b76-b1f3-58edd7dec59c-kube-api-access-j5wrg\") pod \"redhat-operators-66vvh\" (UID: \"e7a96a53-ed40-4b76-b1f3-58edd7dec59c\") " pod="openshift-marketplace/redhat-operators-66vvh" Mar 13 15:03:28 crc kubenswrapper[4907]: I0313 15:03:28.123541 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-66vvh" Mar 13 15:03:28 crc kubenswrapper[4907]: I0313 15:03:28.548091 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-66vvh"] Mar 13 15:03:28 crc kubenswrapper[4907]: I0313 15:03:28.910878 4907 generic.go:334] "Generic (PLEG): container finished" podID="e7a96a53-ed40-4b76-b1f3-58edd7dec59c" containerID="9b879515804206ce1b2b94caf0c492d6383320b9348237ccf02d879b6d4dc8ff" exitCode=0 Mar 13 15:03:28 crc kubenswrapper[4907]: I0313 15:03:28.910968 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-66vvh" event={"ID":"e7a96a53-ed40-4b76-b1f3-58edd7dec59c","Type":"ContainerDied","Data":"9b879515804206ce1b2b94caf0c492d6383320b9348237ccf02d879b6d4dc8ff"} Mar 13 15:03:28 crc kubenswrapper[4907]: I0313 15:03:28.911126 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-66vvh" event={"ID":"e7a96a53-ed40-4b76-b1f3-58edd7dec59c","Type":"ContainerStarted","Data":"ded7a4a7c9ba3e3a3c449453acd07df91c27c365325bee0b5ef808e168610d02"} Mar 13 15:03:30 crc kubenswrapper[4907]: I0313 15:03:30.172465 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-gghnk"] Mar 13 15:03:30 crc kubenswrapper[4907]: I0313 15:03:30.174842 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gghnk" Mar 13 15:03:30 crc kubenswrapper[4907]: I0313 15:03:30.192040 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-gghnk"] Mar 13 15:03:30 crc kubenswrapper[4907]: I0313 15:03:30.318208 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4m69g\" (UniqueName: \"kubernetes.io/projected/58109682-9dbe-4b9f-ac68-436adf1e5943-kube-api-access-4m69g\") pod \"certified-operators-gghnk\" (UID: \"58109682-9dbe-4b9f-ac68-436adf1e5943\") " pod="openshift-marketplace/certified-operators-gghnk" Mar 13 15:03:30 crc kubenswrapper[4907]: I0313 15:03:30.318262 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/58109682-9dbe-4b9f-ac68-436adf1e5943-utilities\") pod \"certified-operators-gghnk\" (UID: \"58109682-9dbe-4b9f-ac68-436adf1e5943\") " pod="openshift-marketplace/certified-operators-gghnk" Mar 13 15:03:30 crc kubenswrapper[4907]: I0313 15:03:30.318866 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/58109682-9dbe-4b9f-ac68-436adf1e5943-catalog-content\") pod \"certified-operators-gghnk\" (UID: \"58109682-9dbe-4b9f-ac68-436adf1e5943\") " pod="openshift-marketplace/certified-operators-gghnk" Mar 13 15:03:30 crc kubenswrapper[4907]: I0313 15:03:30.420156 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4m69g\" (UniqueName: \"kubernetes.io/projected/58109682-9dbe-4b9f-ac68-436adf1e5943-kube-api-access-4m69g\") pod \"certified-operators-gghnk\" (UID: \"58109682-9dbe-4b9f-ac68-436adf1e5943\") " pod="openshift-marketplace/certified-operators-gghnk" Mar 13 15:03:30 crc kubenswrapper[4907]: I0313 15:03:30.420222 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/58109682-9dbe-4b9f-ac68-436adf1e5943-utilities\") pod \"certified-operators-gghnk\" (UID: \"58109682-9dbe-4b9f-ac68-436adf1e5943\") " pod="openshift-marketplace/certified-operators-gghnk" Mar 13 15:03:30 crc kubenswrapper[4907]: I0313 15:03:30.420301 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/58109682-9dbe-4b9f-ac68-436adf1e5943-catalog-content\") pod \"certified-operators-gghnk\" (UID: \"58109682-9dbe-4b9f-ac68-436adf1e5943\") " pod="openshift-marketplace/certified-operators-gghnk" Mar 13 15:03:30 crc kubenswrapper[4907]: I0313 15:03:30.420796 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/58109682-9dbe-4b9f-ac68-436adf1e5943-catalog-content\") pod \"certified-operators-gghnk\" (UID: \"58109682-9dbe-4b9f-ac68-436adf1e5943\") " pod="openshift-marketplace/certified-operators-gghnk" Mar 13 15:03:30 crc kubenswrapper[4907]: I0313 15:03:30.420896 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/58109682-9dbe-4b9f-ac68-436adf1e5943-utilities\") pod \"certified-operators-gghnk\" (UID: \"58109682-9dbe-4b9f-ac68-436adf1e5943\") " pod="openshift-marketplace/certified-operators-gghnk" Mar 13 15:03:30 crc kubenswrapper[4907]: I0313 15:03:30.440905 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4m69g\" (UniqueName: \"kubernetes.io/projected/58109682-9dbe-4b9f-ac68-436adf1e5943-kube-api-access-4m69g\") pod \"certified-operators-gghnk\" (UID: \"58109682-9dbe-4b9f-ac68-436adf1e5943\") " pod="openshift-marketplace/certified-operators-gghnk" Mar 13 15:03:30 crc kubenswrapper[4907]: I0313 15:03:30.493688 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gghnk" Mar 13 15:03:30 crc kubenswrapper[4907]: I0313 15:03:30.926577 4907 generic.go:334] "Generic (PLEG): container finished" podID="e7a96a53-ed40-4b76-b1f3-58edd7dec59c" containerID="0ebf89d94f759be904e95ad5288965a4b932133de012c3259bfbd8638ad7acdb" exitCode=0 Mar 13 15:03:30 crc kubenswrapper[4907]: I0313 15:03:30.926683 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-66vvh" event={"ID":"e7a96a53-ed40-4b76-b1f3-58edd7dec59c","Type":"ContainerDied","Data":"0ebf89d94f759be904e95ad5288965a4b932133de012c3259bfbd8638ad7acdb"} Mar 13 15:03:30 crc kubenswrapper[4907]: W0313 15:03:30.982138 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod58109682_9dbe_4b9f_ac68_436adf1e5943.slice/crio-a6de591612bd7f598975b473588315c1b21aeee80aebf09d8fc818c82ba74d05 WatchSource:0}: Error finding container a6de591612bd7f598975b473588315c1b21aeee80aebf09d8fc818c82ba74d05: Status 404 returned error can't find the container with id a6de591612bd7f598975b473588315c1b21aeee80aebf09d8fc818c82ba74d05 Mar 13 15:03:30 crc kubenswrapper[4907]: I0313 15:03:30.986459 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-gghnk"] Mar 13 15:03:31 crc kubenswrapper[4907]: I0313 15:03:31.935029 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-66vvh" event={"ID":"e7a96a53-ed40-4b76-b1f3-58edd7dec59c","Type":"ContainerStarted","Data":"554246066c0b4d2e5bcab179363cf1a580411eeb84336e62f65f4c2957f75cfb"} Mar 13 15:03:31 crc kubenswrapper[4907]: I0313 15:03:31.936237 4907 generic.go:334] "Generic (PLEG): container finished" podID="58109682-9dbe-4b9f-ac68-436adf1e5943" containerID="b15bbdd7e88ef6a86f31df0036fa2f21235ebe45494d497ad718bac5b50fb58e" exitCode=0 Mar 13 15:03:31 crc kubenswrapper[4907]: I0313 15:03:31.936268 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gghnk" event={"ID":"58109682-9dbe-4b9f-ac68-436adf1e5943","Type":"ContainerDied","Data":"b15bbdd7e88ef6a86f31df0036fa2f21235ebe45494d497ad718bac5b50fb58e"} Mar 13 15:03:31 crc kubenswrapper[4907]: I0313 15:03:31.936285 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gghnk" event={"ID":"58109682-9dbe-4b9f-ac68-436adf1e5943","Type":"ContainerStarted","Data":"a6de591612bd7f598975b473588315c1b21aeee80aebf09d8fc818c82ba74d05"} Mar 13 15:03:31 crc kubenswrapper[4907]: I0313 15:03:31.952556 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-66vvh" podStartSLOduration=2.34704898 podStartE2EDuration="4.952532762s" podCreationTimestamp="2026-03-13 15:03:27 +0000 UTC" firstStartedPulling="2026-03-13 15:03:28.912746037 +0000 UTC m=+3507.812533726" lastFinishedPulling="2026-03-13 15:03:31.518229819 +0000 UTC m=+3510.418017508" observedRunningTime="2026-03-13 15:03:31.949317465 +0000 UTC m=+3510.849105154" watchObservedRunningTime="2026-03-13 15:03:31.952532762 +0000 UTC m=+3510.852320451" Mar 13 15:03:37 crc kubenswrapper[4907]: I0313 15:03:37.973612 4907 generic.go:334] "Generic (PLEG): container finished" podID="58109682-9dbe-4b9f-ac68-436adf1e5943" containerID="b74e9e26c2ace1b371fd700c4c00135c61ddd6fae9f88bedfe17d9937c863902" exitCode=0 Mar 13 15:03:37 crc kubenswrapper[4907]: I0313 15:03:37.973674 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gghnk" event={"ID":"58109682-9dbe-4b9f-ac68-436adf1e5943","Type":"ContainerDied","Data":"b74e9e26c2ace1b371fd700c4c00135c61ddd6fae9f88bedfe17d9937c863902"} Mar 13 15:03:38 crc kubenswrapper[4907]: I0313 15:03:38.124394 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-66vvh" Mar 13 15:03:38 crc kubenswrapper[4907]: I0313 15:03:38.124448 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-66vvh" Mar 13 15:03:38 crc kubenswrapper[4907]: I0313 15:03:38.161426 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-66vvh" Mar 13 15:03:38 crc kubenswrapper[4907]: I0313 15:03:38.983236 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gghnk" event={"ID":"58109682-9dbe-4b9f-ac68-436adf1e5943","Type":"ContainerStarted","Data":"24c5f8f053b603164ac21c58d2935e095541de88fc62f30e30f087b6e3342310"} Mar 13 15:03:39 crc kubenswrapper[4907]: I0313 15:03:39.004832 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-gghnk" podStartSLOduration=2.481883791 podStartE2EDuration="9.004811144s" podCreationTimestamp="2026-03-13 15:03:30 +0000 UTC" firstStartedPulling="2026-03-13 15:03:31.938010715 +0000 UTC m=+3510.837798404" lastFinishedPulling="2026-03-13 15:03:38.460938068 +0000 UTC m=+3517.360725757" observedRunningTime="2026-03-13 15:03:39.004811834 +0000 UTC m=+3517.904599563" watchObservedRunningTime="2026-03-13 15:03:39.004811144 +0000 UTC m=+3517.904598833" Mar 13 15:03:39 crc kubenswrapper[4907]: I0313 15:03:39.028545 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-66vvh" Mar 13 15:03:40 crc kubenswrapper[4907]: I0313 15:03:40.210021 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-66vvh"] Mar 13 15:03:40 crc kubenswrapper[4907]: I0313 15:03:40.494924 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-gghnk" Mar 13 15:03:40 crc kubenswrapper[4907]: I0313 15:03:40.494973 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-gghnk" Mar 13 15:03:40 crc kubenswrapper[4907]: I0313 15:03:40.579409 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-gghnk" Mar 13 15:03:40 crc kubenswrapper[4907]: I0313 15:03:40.995621 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-66vvh" podUID="e7a96a53-ed40-4b76-b1f3-58edd7dec59c" containerName="registry-server" containerID="cri-o://554246066c0b4d2e5bcab179363cf1a580411eeb84336e62f65f4c2957f75cfb" gracePeriod=2 Mar 13 15:03:42 crc kubenswrapper[4907]: I0313 15:03:42.004246 4907 generic.go:334] "Generic (PLEG): container finished" podID="e7a96a53-ed40-4b76-b1f3-58edd7dec59c" containerID="554246066c0b4d2e5bcab179363cf1a580411eeb84336e62f65f4c2957f75cfb" exitCode=0 Mar 13 15:03:42 crc kubenswrapper[4907]: I0313 15:03:42.004327 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-66vvh" event={"ID":"e7a96a53-ed40-4b76-b1f3-58edd7dec59c","Type":"ContainerDied","Data":"554246066c0b4d2e5bcab179363cf1a580411eeb84336e62f65f4c2957f75cfb"} Mar 13 15:03:42 crc kubenswrapper[4907]: I0313 15:03:42.477267 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-66vvh" Mar 13 15:03:42 crc kubenswrapper[4907]: I0313 15:03:42.589683 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7a96a53-ed40-4b76-b1f3-58edd7dec59c-catalog-content\") pod \"e7a96a53-ed40-4b76-b1f3-58edd7dec59c\" (UID: \"e7a96a53-ed40-4b76-b1f3-58edd7dec59c\") " Mar 13 15:03:42 crc kubenswrapper[4907]: I0313 15:03:42.589861 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j5wrg\" (UniqueName: \"kubernetes.io/projected/e7a96a53-ed40-4b76-b1f3-58edd7dec59c-kube-api-access-j5wrg\") pod \"e7a96a53-ed40-4b76-b1f3-58edd7dec59c\" (UID: \"e7a96a53-ed40-4b76-b1f3-58edd7dec59c\") " Mar 13 15:03:42 crc kubenswrapper[4907]: I0313 15:03:42.589931 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7a96a53-ed40-4b76-b1f3-58edd7dec59c-utilities\") pod \"e7a96a53-ed40-4b76-b1f3-58edd7dec59c\" (UID: \"e7a96a53-ed40-4b76-b1f3-58edd7dec59c\") " Mar 13 15:03:42 crc kubenswrapper[4907]: I0313 15:03:42.591174 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e7a96a53-ed40-4b76-b1f3-58edd7dec59c-utilities" (OuterVolumeSpecName: "utilities") pod "e7a96a53-ed40-4b76-b1f3-58edd7dec59c" (UID: "e7a96a53-ed40-4b76-b1f3-58edd7dec59c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:03:42 crc kubenswrapper[4907]: I0313 15:03:42.591477 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7a96a53-ed40-4b76-b1f3-58edd7dec59c-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 15:03:42 crc kubenswrapper[4907]: I0313 15:03:42.596967 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7a96a53-ed40-4b76-b1f3-58edd7dec59c-kube-api-access-j5wrg" (OuterVolumeSpecName: "kube-api-access-j5wrg") pod "e7a96a53-ed40-4b76-b1f3-58edd7dec59c" (UID: "e7a96a53-ed40-4b76-b1f3-58edd7dec59c"). InnerVolumeSpecName "kube-api-access-j5wrg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:03:42 crc kubenswrapper[4907]: I0313 15:03:42.702488 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j5wrg\" (UniqueName: \"kubernetes.io/projected/e7a96a53-ed40-4b76-b1f3-58edd7dec59c-kube-api-access-j5wrg\") on node \"crc\" DevicePath \"\"" Mar 13 15:03:42 crc kubenswrapper[4907]: I0313 15:03:42.744175 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e7a96a53-ed40-4b76-b1f3-58edd7dec59c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e7a96a53-ed40-4b76-b1f3-58edd7dec59c" (UID: "e7a96a53-ed40-4b76-b1f3-58edd7dec59c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:03:42 crc kubenswrapper[4907]: I0313 15:03:42.804025 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7a96a53-ed40-4b76-b1f3-58edd7dec59c-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 15:03:43 crc kubenswrapper[4907]: I0313 15:03:43.016009 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-66vvh" event={"ID":"e7a96a53-ed40-4b76-b1f3-58edd7dec59c","Type":"ContainerDied","Data":"ded7a4a7c9ba3e3a3c449453acd07df91c27c365325bee0b5ef808e168610d02"} Mar 13 15:03:43 crc kubenswrapper[4907]: I0313 15:03:43.016063 4907 scope.go:117] "RemoveContainer" containerID="554246066c0b4d2e5bcab179363cf1a580411eeb84336e62f65f4c2957f75cfb" Mar 13 15:03:43 crc kubenswrapper[4907]: I0313 15:03:43.016219 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-66vvh" Mar 13 15:03:43 crc kubenswrapper[4907]: I0313 15:03:43.049453 4907 scope.go:117] "RemoveContainer" containerID="0ebf89d94f759be904e95ad5288965a4b932133de012c3259bfbd8638ad7acdb" Mar 13 15:03:43 crc kubenswrapper[4907]: I0313 15:03:43.052787 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-66vvh"] Mar 13 15:03:43 crc kubenswrapper[4907]: I0313 15:03:43.058063 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-66vvh"] Mar 13 15:03:43 crc kubenswrapper[4907]: I0313 15:03:43.079567 4907 scope.go:117] "RemoveContainer" containerID="9b879515804206ce1b2b94caf0c492d6383320b9348237ccf02d879b6d4dc8ff" Mar 13 15:03:43 crc kubenswrapper[4907]: I0313 15:03:43.791397 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7a96a53-ed40-4b76-b1f3-58edd7dec59c" path="/var/lib/kubelet/pods/e7a96a53-ed40-4b76-b1f3-58edd7dec59c/volumes" Mar 13 15:03:48 crc kubenswrapper[4907]: I0313 15:03:48.041948 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 15:03:48 crc kubenswrapper[4907]: I0313 15:03:48.042540 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 15:03:48 crc kubenswrapper[4907]: I0313 15:03:48.042590 4907 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" Mar 13 15:03:48 crc kubenswrapper[4907]: I0313 15:03:48.043201 4907 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"62e3208576ab64ce53d2a23023d86e9e8cd79b456c2ef95c5adcb1299b258391"} pod="openshift-machine-config-operator/machine-config-daemon-hm56j" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 15:03:48 crc kubenswrapper[4907]: I0313 15:03:48.043258 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" containerID="cri-o://62e3208576ab64ce53d2a23023d86e9e8cd79b456c2ef95c5adcb1299b258391" gracePeriod=600 Mar 13 15:03:49 crc kubenswrapper[4907]: I0313 15:03:49.062815 4907 generic.go:334] "Generic (PLEG): container finished" podID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerID="62e3208576ab64ce53d2a23023d86e9e8cd79b456c2ef95c5adcb1299b258391" exitCode=0 Mar 13 15:03:49 crc kubenswrapper[4907]: I0313 15:03:49.062920 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerDied","Data":"62e3208576ab64ce53d2a23023d86e9e8cd79b456c2ef95c5adcb1299b258391"} Mar 13 15:03:49 crc kubenswrapper[4907]: I0313 15:03:49.063186 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerStarted","Data":"caeb29efd9b0ec75cd14d4d1c0fbb54dce8073ba7639ceb605cc152e79561f8a"} Mar 13 15:03:49 crc kubenswrapper[4907]: I0313 15:03:49.063229 4907 scope.go:117] "RemoveContainer" containerID="a44259e1e0a7edc8acdd625ba3f7415d70aa78b760d8eeb81254942722d3acb3" Mar 13 15:03:50 crc kubenswrapper[4907]: I0313 15:03:50.539109 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-gghnk" Mar 13 15:03:50 crc kubenswrapper[4907]: I0313 15:03:50.587400 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-gghnk"] Mar 13 15:03:51 crc kubenswrapper[4907]: I0313 15:03:51.082074 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-gghnk" podUID="58109682-9dbe-4b9f-ac68-436adf1e5943" containerName="registry-server" containerID="cri-o://24c5f8f053b603164ac21c58d2935e095541de88fc62f30e30f087b6e3342310" gracePeriod=2 Mar 13 15:03:51 crc kubenswrapper[4907]: I0313 15:03:51.476302 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gghnk" Mar 13 15:03:51 crc kubenswrapper[4907]: I0313 15:03:51.631397 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4m69g\" (UniqueName: \"kubernetes.io/projected/58109682-9dbe-4b9f-ac68-436adf1e5943-kube-api-access-4m69g\") pod \"58109682-9dbe-4b9f-ac68-436adf1e5943\" (UID: \"58109682-9dbe-4b9f-ac68-436adf1e5943\") " Mar 13 15:03:51 crc kubenswrapper[4907]: I0313 15:03:51.631479 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/58109682-9dbe-4b9f-ac68-436adf1e5943-utilities\") pod \"58109682-9dbe-4b9f-ac68-436adf1e5943\" (UID: \"58109682-9dbe-4b9f-ac68-436adf1e5943\") " Mar 13 15:03:51 crc kubenswrapper[4907]: I0313 15:03:51.631551 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/58109682-9dbe-4b9f-ac68-436adf1e5943-catalog-content\") pod \"58109682-9dbe-4b9f-ac68-436adf1e5943\" (UID: \"58109682-9dbe-4b9f-ac68-436adf1e5943\") " Mar 13 15:03:51 crc kubenswrapper[4907]: I0313 15:03:51.632516 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/58109682-9dbe-4b9f-ac68-436adf1e5943-utilities" (OuterVolumeSpecName: "utilities") pod "58109682-9dbe-4b9f-ac68-436adf1e5943" (UID: "58109682-9dbe-4b9f-ac68-436adf1e5943"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:03:51 crc kubenswrapper[4907]: I0313 15:03:51.642065 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/58109682-9dbe-4b9f-ac68-436adf1e5943-kube-api-access-4m69g" (OuterVolumeSpecName: "kube-api-access-4m69g") pod "58109682-9dbe-4b9f-ac68-436adf1e5943" (UID: "58109682-9dbe-4b9f-ac68-436adf1e5943"). InnerVolumeSpecName "kube-api-access-4m69g". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:03:51 crc kubenswrapper[4907]: I0313 15:03:51.690500 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/58109682-9dbe-4b9f-ac68-436adf1e5943-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "58109682-9dbe-4b9f-ac68-436adf1e5943" (UID: "58109682-9dbe-4b9f-ac68-436adf1e5943"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:03:51 crc kubenswrapper[4907]: I0313 15:03:51.733712 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/58109682-9dbe-4b9f-ac68-436adf1e5943-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 15:03:51 crc kubenswrapper[4907]: I0313 15:03:51.733762 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4m69g\" (UniqueName: \"kubernetes.io/projected/58109682-9dbe-4b9f-ac68-436adf1e5943-kube-api-access-4m69g\") on node \"crc\" DevicePath \"\"" Mar 13 15:03:51 crc kubenswrapper[4907]: I0313 15:03:51.733781 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/58109682-9dbe-4b9f-ac68-436adf1e5943-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 15:03:52 crc kubenswrapper[4907]: I0313 15:03:52.091216 4907 generic.go:334] "Generic (PLEG): container finished" podID="58109682-9dbe-4b9f-ac68-436adf1e5943" containerID="24c5f8f053b603164ac21c58d2935e095541de88fc62f30e30f087b6e3342310" exitCode=0 Mar 13 15:03:52 crc kubenswrapper[4907]: I0313 15:03:52.091263 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gghnk" event={"ID":"58109682-9dbe-4b9f-ac68-436adf1e5943","Type":"ContainerDied","Data":"24c5f8f053b603164ac21c58d2935e095541de88fc62f30e30f087b6e3342310"} Mar 13 15:03:52 crc kubenswrapper[4907]: I0313 15:03:52.091310 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gghnk" event={"ID":"58109682-9dbe-4b9f-ac68-436adf1e5943","Type":"ContainerDied","Data":"a6de591612bd7f598975b473588315c1b21aeee80aebf09d8fc818c82ba74d05"} Mar 13 15:03:52 crc kubenswrapper[4907]: I0313 15:03:52.091315 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gghnk" Mar 13 15:03:52 crc kubenswrapper[4907]: I0313 15:03:52.091336 4907 scope.go:117] "RemoveContainer" containerID="24c5f8f053b603164ac21c58d2935e095541de88fc62f30e30f087b6e3342310" Mar 13 15:03:52 crc kubenswrapper[4907]: I0313 15:03:52.112965 4907 scope.go:117] "RemoveContainer" containerID="b74e9e26c2ace1b371fd700c4c00135c61ddd6fae9f88bedfe17d9937c863902" Mar 13 15:03:52 crc kubenswrapper[4907]: I0313 15:03:52.114907 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-gghnk"] Mar 13 15:03:52 crc kubenswrapper[4907]: I0313 15:03:52.122439 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-gghnk"] Mar 13 15:03:52 crc kubenswrapper[4907]: I0313 15:03:52.146786 4907 scope.go:117] "RemoveContainer" containerID="b15bbdd7e88ef6a86f31df0036fa2f21235ebe45494d497ad718bac5b50fb58e" Mar 13 15:03:52 crc kubenswrapper[4907]: I0313 15:03:52.167188 4907 scope.go:117] "RemoveContainer" containerID="24c5f8f053b603164ac21c58d2935e095541de88fc62f30e30f087b6e3342310" Mar 13 15:03:52 crc kubenswrapper[4907]: E0313 15:03:52.167588 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"24c5f8f053b603164ac21c58d2935e095541de88fc62f30e30f087b6e3342310\": container with ID starting with 24c5f8f053b603164ac21c58d2935e095541de88fc62f30e30f087b6e3342310 not found: ID does not exist" containerID="24c5f8f053b603164ac21c58d2935e095541de88fc62f30e30f087b6e3342310" Mar 13 15:03:52 crc kubenswrapper[4907]: I0313 15:03:52.167723 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"24c5f8f053b603164ac21c58d2935e095541de88fc62f30e30f087b6e3342310"} err="failed to get container status \"24c5f8f053b603164ac21c58d2935e095541de88fc62f30e30f087b6e3342310\": rpc error: code = NotFound desc = could not find container \"24c5f8f053b603164ac21c58d2935e095541de88fc62f30e30f087b6e3342310\": container with ID starting with 24c5f8f053b603164ac21c58d2935e095541de88fc62f30e30f087b6e3342310 not found: ID does not exist" Mar 13 15:03:52 crc kubenswrapper[4907]: I0313 15:03:52.167799 4907 scope.go:117] "RemoveContainer" containerID="b74e9e26c2ace1b371fd700c4c00135c61ddd6fae9f88bedfe17d9937c863902" Mar 13 15:03:52 crc kubenswrapper[4907]: E0313 15:03:52.168303 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b74e9e26c2ace1b371fd700c4c00135c61ddd6fae9f88bedfe17d9937c863902\": container with ID starting with b74e9e26c2ace1b371fd700c4c00135c61ddd6fae9f88bedfe17d9937c863902 not found: ID does not exist" containerID="b74e9e26c2ace1b371fd700c4c00135c61ddd6fae9f88bedfe17d9937c863902" Mar 13 15:03:52 crc kubenswrapper[4907]: I0313 15:03:52.168384 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b74e9e26c2ace1b371fd700c4c00135c61ddd6fae9f88bedfe17d9937c863902"} err="failed to get container status \"b74e9e26c2ace1b371fd700c4c00135c61ddd6fae9f88bedfe17d9937c863902\": rpc error: code = NotFound desc = could not find container \"b74e9e26c2ace1b371fd700c4c00135c61ddd6fae9f88bedfe17d9937c863902\": container with ID starting with b74e9e26c2ace1b371fd700c4c00135c61ddd6fae9f88bedfe17d9937c863902 not found: ID does not exist" Mar 13 15:03:52 crc kubenswrapper[4907]: I0313 15:03:52.168446 4907 scope.go:117] "RemoveContainer" containerID="b15bbdd7e88ef6a86f31df0036fa2f21235ebe45494d497ad718bac5b50fb58e" Mar 13 15:03:52 crc kubenswrapper[4907]: E0313 15:03:52.168804 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b15bbdd7e88ef6a86f31df0036fa2f21235ebe45494d497ad718bac5b50fb58e\": container with ID starting with b15bbdd7e88ef6a86f31df0036fa2f21235ebe45494d497ad718bac5b50fb58e not found: ID does not exist" containerID="b15bbdd7e88ef6a86f31df0036fa2f21235ebe45494d497ad718bac5b50fb58e" Mar 13 15:03:52 crc kubenswrapper[4907]: I0313 15:03:52.168914 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b15bbdd7e88ef6a86f31df0036fa2f21235ebe45494d497ad718bac5b50fb58e"} err="failed to get container status \"b15bbdd7e88ef6a86f31df0036fa2f21235ebe45494d497ad718bac5b50fb58e\": rpc error: code = NotFound desc = could not find container \"b15bbdd7e88ef6a86f31df0036fa2f21235ebe45494d497ad718bac5b50fb58e\": container with ID starting with b15bbdd7e88ef6a86f31df0036fa2f21235ebe45494d497ad718bac5b50fb58e not found: ID does not exist" Mar 13 15:03:53 crc kubenswrapper[4907]: I0313 15:03:53.791131 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="58109682-9dbe-4b9f-ac68-436adf1e5943" path="/var/lib/kubelet/pods/58109682-9dbe-4b9f-ac68-436adf1e5943/volumes" Mar 13 15:03:55 crc kubenswrapper[4907]: I0313 15:03:55.383716 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-zx5wb"] Mar 13 15:03:55 crc kubenswrapper[4907]: E0313 15:03:55.385235 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58109682-9dbe-4b9f-ac68-436adf1e5943" containerName="extract-content" Mar 13 15:03:55 crc kubenswrapper[4907]: I0313 15:03:55.385253 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="58109682-9dbe-4b9f-ac68-436adf1e5943" containerName="extract-content" Mar 13 15:03:55 crc kubenswrapper[4907]: E0313 15:03:55.385271 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7a96a53-ed40-4b76-b1f3-58edd7dec59c" containerName="extract-utilities" Mar 13 15:03:55 crc kubenswrapper[4907]: I0313 15:03:55.385278 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7a96a53-ed40-4b76-b1f3-58edd7dec59c" containerName="extract-utilities" Mar 13 15:03:55 crc kubenswrapper[4907]: E0313 15:03:55.385286 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7a96a53-ed40-4b76-b1f3-58edd7dec59c" containerName="extract-content" Mar 13 15:03:55 crc kubenswrapper[4907]: I0313 15:03:55.385292 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7a96a53-ed40-4b76-b1f3-58edd7dec59c" containerName="extract-content" Mar 13 15:03:55 crc kubenswrapper[4907]: E0313 15:03:55.385301 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58109682-9dbe-4b9f-ac68-436adf1e5943" containerName="extract-utilities" Mar 13 15:03:55 crc kubenswrapper[4907]: I0313 15:03:55.385307 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="58109682-9dbe-4b9f-ac68-436adf1e5943" containerName="extract-utilities" Mar 13 15:03:55 crc kubenswrapper[4907]: E0313 15:03:55.385324 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7a96a53-ed40-4b76-b1f3-58edd7dec59c" containerName="registry-server" Mar 13 15:03:55 crc kubenswrapper[4907]: I0313 15:03:55.385329 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7a96a53-ed40-4b76-b1f3-58edd7dec59c" containerName="registry-server" Mar 13 15:03:55 crc kubenswrapper[4907]: E0313 15:03:55.385345 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58109682-9dbe-4b9f-ac68-436adf1e5943" containerName="registry-server" Mar 13 15:03:55 crc kubenswrapper[4907]: I0313 15:03:55.385351 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="58109682-9dbe-4b9f-ac68-436adf1e5943" containerName="registry-server" Mar 13 15:03:55 crc kubenswrapper[4907]: I0313 15:03:55.385492 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="e7a96a53-ed40-4b76-b1f3-58edd7dec59c" containerName="registry-server" Mar 13 15:03:55 crc kubenswrapper[4907]: I0313 15:03:55.385510 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="58109682-9dbe-4b9f-ac68-436adf1e5943" containerName="registry-server" Mar 13 15:03:55 crc kubenswrapper[4907]: I0313 15:03:55.386497 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zx5wb" Mar 13 15:03:55 crc kubenswrapper[4907]: I0313 15:03:55.438743 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zx5wb"] Mar 13 15:03:55 crc kubenswrapper[4907]: I0313 15:03:55.495290 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bdee4689-2396-4e32-ae39-6ffc60033f38-catalog-content\") pod \"community-operators-zx5wb\" (UID: \"bdee4689-2396-4e32-ae39-6ffc60033f38\") " pod="openshift-marketplace/community-operators-zx5wb" Mar 13 15:03:55 crc kubenswrapper[4907]: I0313 15:03:55.495382 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q7dql\" (UniqueName: \"kubernetes.io/projected/bdee4689-2396-4e32-ae39-6ffc60033f38-kube-api-access-q7dql\") pod \"community-operators-zx5wb\" (UID: \"bdee4689-2396-4e32-ae39-6ffc60033f38\") " pod="openshift-marketplace/community-operators-zx5wb" Mar 13 15:03:55 crc kubenswrapper[4907]: I0313 15:03:55.495432 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bdee4689-2396-4e32-ae39-6ffc60033f38-utilities\") pod \"community-operators-zx5wb\" (UID: \"bdee4689-2396-4e32-ae39-6ffc60033f38\") " pod="openshift-marketplace/community-operators-zx5wb" Mar 13 15:03:55 crc kubenswrapper[4907]: I0313 15:03:55.596503 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bdee4689-2396-4e32-ae39-6ffc60033f38-catalog-content\") pod \"community-operators-zx5wb\" (UID: \"bdee4689-2396-4e32-ae39-6ffc60033f38\") " pod="openshift-marketplace/community-operators-zx5wb" Mar 13 15:03:55 crc kubenswrapper[4907]: I0313 15:03:55.596588 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q7dql\" (UniqueName: \"kubernetes.io/projected/bdee4689-2396-4e32-ae39-6ffc60033f38-kube-api-access-q7dql\") pod \"community-operators-zx5wb\" (UID: \"bdee4689-2396-4e32-ae39-6ffc60033f38\") " pod="openshift-marketplace/community-operators-zx5wb" Mar 13 15:03:55 crc kubenswrapper[4907]: I0313 15:03:55.596642 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bdee4689-2396-4e32-ae39-6ffc60033f38-utilities\") pod \"community-operators-zx5wb\" (UID: \"bdee4689-2396-4e32-ae39-6ffc60033f38\") " pod="openshift-marketplace/community-operators-zx5wb" Mar 13 15:03:55 crc kubenswrapper[4907]: I0313 15:03:55.597127 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bdee4689-2396-4e32-ae39-6ffc60033f38-catalog-content\") pod \"community-operators-zx5wb\" (UID: \"bdee4689-2396-4e32-ae39-6ffc60033f38\") " pod="openshift-marketplace/community-operators-zx5wb" Mar 13 15:03:55 crc kubenswrapper[4907]: I0313 15:03:55.597179 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bdee4689-2396-4e32-ae39-6ffc60033f38-utilities\") pod \"community-operators-zx5wb\" (UID: \"bdee4689-2396-4e32-ae39-6ffc60033f38\") " pod="openshift-marketplace/community-operators-zx5wb" Mar 13 15:03:55 crc kubenswrapper[4907]: I0313 15:03:55.624106 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q7dql\" (UniqueName: \"kubernetes.io/projected/bdee4689-2396-4e32-ae39-6ffc60033f38-kube-api-access-q7dql\") pod \"community-operators-zx5wb\" (UID: \"bdee4689-2396-4e32-ae39-6ffc60033f38\") " pod="openshift-marketplace/community-operators-zx5wb" Mar 13 15:03:55 crc kubenswrapper[4907]: I0313 15:03:55.710858 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zx5wb" Mar 13 15:03:56 crc kubenswrapper[4907]: I0313 15:03:56.016274 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zx5wb"] Mar 13 15:03:56 crc kubenswrapper[4907]: I0313 15:03:56.118973 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zx5wb" event={"ID":"bdee4689-2396-4e32-ae39-6ffc60033f38","Type":"ContainerStarted","Data":"0bc7f96bf4dee2dfb710654f87828dd414be19d802599fada94514556b97115a"} Mar 13 15:03:57 crc kubenswrapper[4907]: I0313 15:03:57.126293 4907 generic.go:334] "Generic (PLEG): container finished" podID="bdee4689-2396-4e32-ae39-6ffc60033f38" containerID="6568d87c6bad18c68d28d286c5ed63b49b3f4a5ac95f46d9d41dc5bd03f93775" exitCode=0 Mar 13 15:03:57 crc kubenswrapper[4907]: I0313 15:03:57.126413 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zx5wb" event={"ID":"bdee4689-2396-4e32-ae39-6ffc60033f38","Type":"ContainerDied","Data":"6568d87c6bad18c68d28d286c5ed63b49b3f4a5ac95f46d9d41dc5bd03f93775"} Mar 13 15:03:58 crc kubenswrapper[4907]: I0313 15:03:58.135405 4907 generic.go:334] "Generic (PLEG): container finished" podID="bdee4689-2396-4e32-ae39-6ffc60033f38" containerID="84f8ced411df60770107c3973266fb82290633ef4138573f621b9eb0d07085e5" exitCode=0 Mar 13 15:03:58 crc kubenswrapper[4907]: I0313 15:03:58.135466 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zx5wb" event={"ID":"bdee4689-2396-4e32-ae39-6ffc60033f38","Type":"ContainerDied","Data":"84f8ced411df60770107c3973266fb82290633ef4138573f621b9eb0d07085e5"} Mar 13 15:03:59 crc kubenswrapper[4907]: I0313 15:03:59.144849 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zx5wb" event={"ID":"bdee4689-2396-4e32-ae39-6ffc60033f38","Type":"ContainerStarted","Data":"83fa90112c03ad8ce15458ee0ff9f3e2e51c4749967eee63928ff69aae2e9b88"} Mar 13 15:03:59 crc kubenswrapper[4907]: I0313 15:03:59.170625 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-zx5wb" podStartSLOduration=2.5354631789999997 podStartE2EDuration="4.170608146s" podCreationTimestamp="2026-03-13 15:03:55 +0000 UTC" firstStartedPulling="2026-03-13 15:03:57.128076673 +0000 UTC m=+3536.027864362" lastFinishedPulling="2026-03-13 15:03:58.76322164 +0000 UTC m=+3537.663009329" observedRunningTime="2026-03-13 15:03:59.16054434 +0000 UTC m=+3538.060332029" watchObservedRunningTime="2026-03-13 15:03:59.170608146 +0000 UTC m=+3538.070395835" Mar 13 15:04:00 crc kubenswrapper[4907]: I0313 15:04:00.139036 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556904-4qqhq"] Mar 13 15:04:00 crc kubenswrapper[4907]: I0313 15:04:00.140736 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556904-4qqhq" Mar 13 15:04:00 crc kubenswrapper[4907]: I0313 15:04:00.147762 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 15:04:00 crc kubenswrapper[4907]: I0313 15:04:00.148838 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 15:04:00 crc kubenswrapper[4907]: I0313 15:04:00.153044 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 15:04:00 crc kubenswrapper[4907]: I0313 15:04:00.167349 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556904-4qqhq"] Mar 13 15:04:00 crc kubenswrapper[4907]: I0313 15:04:00.270119 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c67s5\" (UniqueName: \"kubernetes.io/projected/856cd90a-f3c2-4b46-b5eb-77e85004dbf6-kube-api-access-c67s5\") pod \"auto-csr-approver-29556904-4qqhq\" (UID: \"856cd90a-f3c2-4b46-b5eb-77e85004dbf6\") " pod="openshift-infra/auto-csr-approver-29556904-4qqhq" Mar 13 15:04:00 crc kubenswrapper[4907]: I0313 15:04:00.372150 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c67s5\" (UniqueName: \"kubernetes.io/projected/856cd90a-f3c2-4b46-b5eb-77e85004dbf6-kube-api-access-c67s5\") pod \"auto-csr-approver-29556904-4qqhq\" (UID: \"856cd90a-f3c2-4b46-b5eb-77e85004dbf6\") " pod="openshift-infra/auto-csr-approver-29556904-4qqhq" Mar 13 15:04:00 crc kubenswrapper[4907]: I0313 15:04:00.401845 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c67s5\" (UniqueName: \"kubernetes.io/projected/856cd90a-f3c2-4b46-b5eb-77e85004dbf6-kube-api-access-c67s5\") pod \"auto-csr-approver-29556904-4qqhq\" (UID: \"856cd90a-f3c2-4b46-b5eb-77e85004dbf6\") " pod="openshift-infra/auto-csr-approver-29556904-4qqhq" Mar 13 15:04:00 crc kubenswrapper[4907]: I0313 15:04:00.474539 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556904-4qqhq" Mar 13 15:04:00 crc kubenswrapper[4907]: I0313 15:04:00.900061 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556904-4qqhq"] Mar 13 15:04:00 crc kubenswrapper[4907]: W0313 15:04:00.906279 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod856cd90a_f3c2_4b46_b5eb_77e85004dbf6.slice/crio-51d2b9d00ee0e8fc8c6c0d1f2930961379e1f2ed812db0fb202c46d98f7acbbe WatchSource:0}: Error finding container 51d2b9d00ee0e8fc8c6c0d1f2930961379e1f2ed812db0fb202c46d98f7acbbe: Status 404 returned error can't find the container with id 51d2b9d00ee0e8fc8c6c0d1f2930961379e1f2ed812db0fb202c46d98f7acbbe Mar 13 15:04:01 crc kubenswrapper[4907]: I0313 15:04:01.158432 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556904-4qqhq" event={"ID":"856cd90a-f3c2-4b46-b5eb-77e85004dbf6","Type":"ContainerStarted","Data":"51d2b9d00ee0e8fc8c6c0d1f2930961379e1f2ed812db0fb202c46d98f7acbbe"} Mar 13 15:04:02 crc kubenswrapper[4907]: I0313 15:04:02.167342 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556904-4qqhq" event={"ID":"856cd90a-f3c2-4b46-b5eb-77e85004dbf6","Type":"ContainerStarted","Data":"6b49bfb9250df806b9fbdd020bba3661e9f6dd127456008ebc47a00dd7745594"} Mar 13 15:04:02 crc kubenswrapper[4907]: I0313 15:04:02.191616 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29556904-4qqhq" podStartSLOduration=1.378680985 podStartE2EDuration="2.191587665s" podCreationTimestamp="2026-03-13 15:04:00 +0000 UTC" firstStartedPulling="2026-03-13 15:04:00.90820107 +0000 UTC m=+3539.807988759" lastFinishedPulling="2026-03-13 15:04:01.72110772 +0000 UTC m=+3540.620895439" observedRunningTime="2026-03-13 15:04:02.189366795 +0000 UTC m=+3541.089154494" watchObservedRunningTime="2026-03-13 15:04:02.191587665 +0000 UTC m=+3541.091375384" Mar 13 15:04:03 crc kubenswrapper[4907]: I0313 15:04:03.175593 4907 generic.go:334] "Generic (PLEG): container finished" podID="856cd90a-f3c2-4b46-b5eb-77e85004dbf6" containerID="6b49bfb9250df806b9fbdd020bba3661e9f6dd127456008ebc47a00dd7745594" exitCode=0 Mar 13 15:04:03 crc kubenswrapper[4907]: I0313 15:04:03.175651 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556904-4qqhq" event={"ID":"856cd90a-f3c2-4b46-b5eb-77e85004dbf6","Type":"ContainerDied","Data":"6b49bfb9250df806b9fbdd020bba3661e9f6dd127456008ebc47a00dd7745594"} Mar 13 15:04:04 crc kubenswrapper[4907]: I0313 15:04:04.439809 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556904-4qqhq" Mar 13 15:04:04 crc kubenswrapper[4907]: I0313 15:04:04.628922 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c67s5\" (UniqueName: \"kubernetes.io/projected/856cd90a-f3c2-4b46-b5eb-77e85004dbf6-kube-api-access-c67s5\") pod \"856cd90a-f3c2-4b46-b5eb-77e85004dbf6\" (UID: \"856cd90a-f3c2-4b46-b5eb-77e85004dbf6\") " Mar 13 15:04:04 crc kubenswrapper[4907]: I0313 15:04:04.634290 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/856cd90a-f3c2-4b46-b5eb-77e85004dbf6-kube-api-access-c67s5" (OuterVolumeSpecName: "kube-api-access-c67s5") pod "856cd90a-f3c2-4b46-b5eb-77e85004dbf6" (UID: "856cd90a-f3c2-4b46-b5eb-77e85004dbf6"). InnerVolumeSpecName "kube-api-access-c67s5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:04:04 crc kubenswrapper[4907]: I0313 15:04:04.730475 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c67s5\" (UniqueName: \"kubernetes.io/projected/856cd90a-f3c2-4b46-b5eb-77e85004dbf6-kube-api-access-c67s5\") on node \"crc\" DevicePath \"\"" Mar 13 15:04:04 crc kubenswrapper[4907]: I0313 15:04:04.869366 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556898-55zxj"] Mar 13 15:04:04 crc kubenswrapper[4907]: I0313 15:04:04.875031 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556898-55zxj"] Mar 13 15:04:05 crc kubenswrapper[4907]: I0313 15:04:05.196724 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556904-4qqhq" event={"ID":"856cd90a-f3c2-4b46-b5eb-77e85004dbf6","Type":"ContainerDied","Data":"51d2b9d00ee0e8fc8c6c0d1f2930961379e1f2ed812db0fb202c46d98f7acbbe"} Mar 13 15:04:05 crc kubenswrapper[4907]: I0313 15:04:05.196766 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556904-4qqhq" Mar 13 15:04:05 crc kubenswrapper[4907]: I0313 15:04:05.196773 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="51d2b9d00ee0e8fc8c6c0d1f2930961379e1f2ed812db0fb202c46d98f7acbbe" Mar 13 15:04:05 crc kubenswrapper[4907]: I0313 15:04:05.711921 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-zx5wb" Mar 13 15:04:05 crc kubenswrapper[4907]: I0313 15:04:05.712190 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-zx5wb" Mar 13 15:04:05 crc kubenswrapper[4907]: I0313 15:04:05.757609 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-zx5wb" Mar 13 15:04:05 crc kubenswrapper[4907]: I0313 15:04:05.790737 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4efb5dca-e973-443a-bcc2-3b8799fb680c" path="/var/lib/kubelet/pods/4efb5dca-e973-443a-bcc2-3b8799fb680c/volumes" Mar 13 15:04:06 crc kubenswrapper[4907]: I0313 15:04:06.254038 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-zx5wb" Mar 13 15:04:06 crc kubenswrapper[4907]: I0313 15:04:06.302658 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zx5wb"] Mar 13 15:04:08 crc kubenswrapper[4907]: I0313 15:04:08.215613 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-zx5wb" podUID="bdee4689-2396-4e32-ae39-6ffc60033f38" containerName="registry-server" containerID="cri-o://83fa90112c03ad8ce15458ee0ff9f3e2e51c4749967eee63928ff69aae2e9b88" gracePeriod=2 Mar 13 15:04:09 crc kubenswrapper[4907]: I0313 15:04:09.108929 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zx5wb" Mar 13 15:04:09 crc kubenswrapper[4907]: I0313 15:04:09.223120 4907 generic.go:334] "Generic (PLEG): container finished" podID="bdee4689-2396-4e32-ae39-6ffc60033f38" containerID="83fa90112c03ad8ce15458ee0ff9f3e2e51c4749967eee63928ff69aae2e9b88" exitCode=0 Mar 13 15:04:09 crc kubenswrapper[4907]: I0313 15:04:09.223173 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zx5wb" event={"ID":"bdee4689-2396-4e32-ae39-6ffc60033f38","Type":"ContainerDied","Data":"83fa90112c03ad8ce15458ee0ff9f3e2e51c4749967eee63928ff69aae2e9b88"} Mar 13 15:04:09 crc kubenswrapper[4907]: I0313 15:04:09.223184 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zx5wb" Mar 13 15:04:09 crc kubenswrapper[4907]: I0313 15:04:09.223204 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zx5wb" event={"ID":"bdee4689-2396-4e32-ae39-6ffc60033f38","Type":"ContainerDied","Data":"0bc7f96bf4dee2dfb710654f87828dd414be19d802599fada94514556b97115a"} Mar 13 15:04:09 crc kubenswrapper[4907]: I0313 15:04:09.223224 4907 scope.go:117] "RemoveContainer" containerID="83fa90112c03ad8ce15458ee0ff9f3e2e51c4749967eee63928ff69aae2e9b88" Mar 13 15:04:09 crc kubenswrapper[4907]: I0313 15:04:09.240464 4907 scope.go:117] "RemoveContainer" containerID="84f8ced411df60770107c3973266fb82290633ef4138573f621b9eb0d07085e5" Mar 13 15:04:09 crc kubenswrapper[4907]: I0313 15:04:09.257224 4907 scope.go:117] "RemoveContainer" containerID="6568d87c6bad18c68d28d286c5ed63b49b3f4a5ac95f46d9d41dc5bd03f93775" Mar 13 15:04:09 crc kubenswrapper[4907]: I0313 15:04:09.284054 4907 scope.go:117] "RemoveContainer" containerID="83fa90112c03ad8ce15458ee0ff9f3e2e51c4749967eee63928ff69aae2e9b88" Mar 13 15:04:09 crc kubenswrapper[4907]: E0313 15:04:09.286606 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"83fa90112c03ad8ce15458ee0ff9f3e2e51c4749967eee63928ff69aae2e9b88\": container with ID starting with 83fa90112c03ad8ce15458ee0ff9f3e2e51c4749967eee63928ff69aae2e9b88 not found: ID does not exist" containerID="83fa90112c03ad8ce15458ee0ff9f3e2e51c4749967eee63928ff69aae2e9b88" Mar 13 15:04:09 crc kubenswrapper[4907]: I0313 15:04:09.286652 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"83fa90112c03ad8ce15458ee0ff9f3e2e51c4749967eee63928ff69aae2e9b88"} err="failed to get container status \"83fa90112c03ad8ce15458ee0ff9f3e2e51c4749967eee63928ff69aae2e9b88\": rpc error: code = NotFound desc = could not find container \"83fa90112c03ad8ce15458ee0ff9f3e2e51c4749967eee63928ff69aae2e9b88\": container with ID starting with 83fa90112c03ad8ce15458ee0ff9f3e2e51c4749967eee63928ff69aae2e9b88 not found: ID does not exist" Mar 13 15:04:09 crc kubenswrapper[4907]: I0313 15:04:09.286676 4907 scope.go:117] "RemoveContainer" containerID="84f8ced411df60770107c3973266fb82290633ef4138573f621b9eb0d07085e5" Mar 13 15:04:09 crc kubenswrapper[4907]: E0313 15:04:09.287223 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"84f8ced411df60770107c3973266fb82290633ef4138573f621b9eb0d07085e5\": container with ID starting with 84f8ced411df60770107c3973266fb82290633ef4138573f621b9eb0d07085e5 not found: ID does not exist" containerID="84f8ced411df60770107c3973266fb82290633ef4138573f621b9eb0d07085e5" Mar 13 15:04:09 crc kubenswrapper[4907]: I0313 15:04:09.287253 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"84f8ced411df60770107c3973266fb82290633ef4138573f621b9eb0d07085e5"} err="failed to get container status \"84f8ced411df60770107c3973266fb82290633ef4138573f621b9eb0d07085e5\": rpc error: code = NotFound desc = could not find container \"84f8ced411df60770107c3973266fb82290633ef4138573f621b9eb0d07085e5\": container with ID starting with 84f8ced411df60770107c3973266fb82290633ef4138573f621b9eb0d07085e5 not found: ID does not exist" Mar 13 15:04:09 crc kubenswrapper[4907]: I0313 15:04:09.287267 4907 scope.go:117] "RemoveContainer" containerID="6568d87c6bad18c68d28d286c5ed63b49b3f4a5ac95f46d9d41dc5bd03f93775" Mar 13 15:04:09 crc kubenswrapper[4907]: E0313 15:04:09.288131 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6568d87c6bad18c68d28d286c5ed63b49b3f4a5ac95f46d9d41dc5bd03f93775\": container with ID starting with 6568d87c6bad18c68d28d286c5ed63b49b3f4a5ac95f46d9d41dc5bd03f93775 not found: ID does not exist" containerID="6568d87c6bad18c68d28d286c5ed63b49b3f4a5ac95f46d9d41dc5bd03f93775" Mar 13 15:04:09 crc kubenswrapper[4907]: I0313 15:04:09.288160 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6568d87c6bad18c68d28d286c5ed63b49b3f4a5ac95f46d9d41dc5bd03f93775"} err="failed to get container status \"6568d87c6bad18c68d28d286c5ed63b49b3f4a5ac95f46d9d41dc5bd03f93775\": rpc error: code = NotFound desc = could not find container \"6568d87c6bad18c68d28d286c5ed63b49b3f4a5ac95f46d9d41dc5bd03f93775\": container with ID starting with 6568d87c6bad18c68d28d286c5ed63b49b3f4a5ac95f46d9d41dc5bd03f93775 not found: ID does not exist" Mar 13 15:04:09 crc kubenswrapper[4907]: I0313 15:04:09.290623 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q7dql\" (UniqueName: \"kubernetes.io/projected/bdee4689-2396-4e32-ae39-6ffc60033f38-kube-api-access-q7dql\") pod \"bdee4689-2396-4e32-ae39-6ffc60033f38\" (UID: \"bdee4689-2396-4e32-ae39-6ffc60033f38\") " Mar 13 15:04:09 crc kubenswrapper[4907]: I0313 15:04:09.290713 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bdee4689-2396-4e32-ae39-6ffc60033f38-utilities\") pod \"bdee4689-2396-4e32-ae39-6ffc60033f38\" (UID: \"bdee4689-2396-4e32-ae39-6ffc60033f38\") " Mar 13 15:04:09 crc kubenswrapper[4907]: I0313 15:04:09.290826 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bdee4689-2396-4e32-ae39-6ffc60033f38-catalog-content\") pod \"bdee4689-2396-4e32-ae39-6ffc60033f38\" (UID: \"bdee4689-2396-4e32-ae39-6ffc60033f38\") " Mar 13 15:04:09 crc kubenswrapper[4907]: I0313 15:04:09.291790 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bdee4689-2396-4e32-ae39-6ffc60033f38-utilities" (OuterVolumeSpecName: "utilities") pod "bdee4689-2396-4e32-ae39-6ffc60033f38" (UID: "bdee4689-2396-4e32-ae39-6ffc60033f38"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:04:09 crc kubenswrapper[4907]: I0313 15:04:09.295475 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bdee4689-2396-4e32-ae39-6ffc60033f38-kube-api-access-q7dql" (OuterVolumeSpecName: "kube-api-access-q7dql") pod "bdee4689-2396-4e32-ae39-6ffc60033f38" (UID: "bdee4689-2396-4e32-ae39-6ffc60033f38"). InnerVolumeSpecName "kube-api-access-q7dql". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:04:09 crc kubenswrapper[4907]: I0313 15:04:09.345566 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bdee4689-2396-4e32-ae39-6ffc60033f38-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bdee4689-2396-4e32-ae39-6ffc60033f38" (UID: "bdee4689-2396-4e32-ae39-6ffc60033f38"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:04:09 crc kubenswrapper[4907]: I0313 15:04:09.392169 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bdee4689-2396-4e32-ae39-6ffc60033f38-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 15:04:09 crc kubenswrapper[4907]: I0313 15:04:09.392203 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q7dql\" (UniqueName: \"kubernetes.io/projected/bdee4689-2396-4e32-ae39-6ffc60033f38-kube-api-access-q7dql\") on node \"crc\" DevicePath \"\"" Mar 13 15:04:09 crc kubenswrapper[4907]: I0313 15:04:09.392215 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bdee4689-2396-4e32-ae39-6ffc60033f38-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 15:04:09 crc kubenswrapper[4907]: I0313 15:04:09.553556 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zx5wb"] Mar 13 15:04:09 crc kubenswrapper[4907]: I0313 15:04:09.563115 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-zx5wb"] Mar 13 15:04:09 crc kubenswrapper[4907]: I0313 15:04:09.790218 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bdee4689-2396-4e32-ae39-6ffc60033f38" path="/var/lib/kubelet/pods/bdee4689-2396-4e32-ae39-6ffc60033f38/volumes" Mar 13 15:04:26 crc kubenswrapper[4907]: I0313 15:04:26.075695 4907 scope.go:117] "RemoveContainer" containerID="770adaf80533ac339976a419aecf995f45ec4e4221e3c81f0b0f3f79b4736ef0" Mar 13 15:05:48 crc kubenswrapper[4907]: I0313 15:05:48.041325 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 15:05:48 crc kubenswrapper[4907]: I0313 15:05:48.041833 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 15:06:00 crc kubenswrapper[4907]: I0313 15:06:00.148927 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556906-kqncj"] Mar 13 15:06:00 crc kubenswrapper[4907]: E0313 15:06:00.151977 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bdee4689-2396-4e32-ae39-6ffc60033f38" containerName="extract-utilities" Mar 13 15:06:00 crc kubenswrapper[4907]: I0313 15:06:00.152137 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="bdee4689-2396-4e32-ae39-6ffc60033f38" containerName="extract-utilities" Mar 13 15:06:00 crc kubenswrapper[4907]: E0313 15:06:00.152262 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bdee4689-2396-4e32-ae39-6ffc60033f38" containerName="registry-server" Mar 13 15:06:00 crc kubenswrapper[4907]: I0313 15:06:00.152378 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="bdee4689-2396-4e32-ae39-6ffc60033f38" containerName="registry-server" Mar 13 15:06:00 crc kubenswrapper[4907]: E0313 15:06:00.152495 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="856cd90a-f3c2-4b46-b5eb-77e85004dbf6" containerName="oc" Mar 13 15:06:00 crc kubenswrapper[4907]: I0313 15:06:00.152612 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="856cd90a-f3c2-4b46-b5eb-77e85004dbf6" containerName="oc" Mar 13 15:06:00 crc kubenswrapper[4907]: E0313 15:06:00.152729 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bdee4689-2396-4e32-ae39-6ffc60033f38" containerName="extract-content" Mar 13 15:06:00 crc kubenswrapper[4907]: I0313 15:06:00.152905 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="bdee4689-2396-4e32-ae39-6ffc60033f38" containerName="extract-content" Mar 13 15:06:00 crc kubenswrapper[4907]: I0313 15:06:00.153323 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="bdee4689-2396-4e32-ae39-6ffc60033f38" containerName="registry-server" Mar 13 15:06:00 crc kubenswrapper[4907]: I0313 15:06:00.153536 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="856cd90a-f3c2-4b46-b5eb-77e85004dbf6" containerName="oc" Mar 13 15:06:00 crc kubenswrapper[4907]: I0313 15:06:00.154409 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556906-kqncj" Mar 13 15:06:00 crc kubenswrapper[4907]: I0313 15:06:00.157930 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 15:06:00 crc kubenswrapper[4907]: I0313 15:06:00.158549 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 15:06:00 crc kubenswrapper[4907]: I0313 15:06:00.158658 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 15:06:00 crc kubenswrapper[4907]: I0313 15:06:00.161239 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556906-kqncj"] Mar 13 15:06:00 crc kubenswrapper[4907]: I0313 15:06:00.309705 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ph8jm\" (UniqueName: \"kubernetes.io/projected/7a501d2e-0774-414e-8afb-39a9a9d8b731-kube-api-access-ph8jm\") pod \"auto-csr-approver-29556906-kqncj\" (UID: \"7a501d2e-0774-414e-8afb-39a9a9d8b731\") " pod="openshift-infra/auto-csr-approver-29556906-kqncj" Mar 13 15:06:00 crc kubenswrapper[4907]: I0313 15:06:00.411677 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ph8jm\" (UniqueName: \"kubernetes.io/projected/7a501d2e-0774-414e-8afb-39a9a9d8b731-kube-api-access-ph8jm\") pod \"auto-csr-approver-29556906-kqncj\" (UID: \"7a501d2e-0774-414e-8afb-39a9a9d8b731\") " pod="openshift-infra/auto-csr-approver-29556906-kqncj" Mar 13 15:06:00 crc kubenswrapper[4907]: I0313 15:06:00.450838 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ph8jm\" (UniqueName: \"kubernetes.io/projected/7a501d2e-0774-414e-8afb-39a9a9d8b731-kube-api-access-ph8jm\") pod \"auto-csr-approver-29556906-kqncj\" (UID: \"7a501d2e-0774-414e-8afb-39a9a9d8b731\") " pod="openshift-infra/auto-csr-approver-29556906-kqncj" Mar 13 15:06:00 crc kubenswrapper[4907]: I0313 15:06:00.473206 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556906-kqncj" Mar 13 15:06:00 crc kubenswrapper[4907]: I0313 15:06:00.911187 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556906-kqncj"] Mar 13 15:06:01 crc kubenswrapper[4907]: I0313 15:06:01.439667 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556906-kqncj" event={"ID":"7a501d2e-0774-414e-8afb-39a9a9d8b731","Type":"ContainerStarted","Data":"a11a599b13906f7f8bcbe895f454a7ae524d8105c5bce3da9b1f7d5287818788"} Mar 13 15:06:02 crc kubenswrapper[4907]: I0313 15:06:02.448973 4907 generic.go:334] "Generic (PLEG): container finished" podID="7a501d2e-0774-414e-8afb-39a9a9d8b731" containerID="c0f40228828ee5195fb41b365b21febab7b0feb3cf02e777e21a472c7ba50dbb" exitCode=0 Mar 13 15:06:02 crc kubenswrapper[4907]: I0313 15:06:02.449087 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556906-kqncj" event={"ID":"7a501d2e-0774-414e-8afb-39a9a9d8b731","Type":"ContainerDied","Data":"c0f40228828ee5195fb41b365b21febab7b0feb3cf02e777e21a472c7ba50dbb"} Mar 13 15:06:03 crc kubenswrapper[4907]: I0313 15:06:03.721515 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556906-kqncj" Mar 13 15:06:03 crc kubenswrapper[4907]: I0313 15:06:03.859085 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ph8jm\" (UniqueName: \"kubernetes.io/projected/7a501d2e-0774-414e-8afb-39a9a9d8b731-kube-api-access-ph8jm\") pod \"7a501d2e-0774-414e-8afb-39a9a9d8b731\" (UID: \"7a501d2e-0774-414e-8afb-39a9a9d8b731\") " Mar 13 15:06:03 crc kubenswrapper[4907]: I0313 15:06:03.864415 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7a501d2e-0774-414e-8afb-39a9a9d8b731-kube-api-access-ph8jm" (OuterVolumeSpecName: "kube-api-access-ph8jm") pod "7a501d2e-0774-414e-8afb-39a9a9d8b731" (UID: "7a501d2e-0774-414e-8afb-39a9a9d8b731"). InnerVolumeSpecName "kube-api-access-ph8jm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:06:03 crc kubenswrapper[4907]: I0313 15:06:03.960969 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ph8jm\" (UniqueName: \"kubernetes.io/projected/7a501d2e-0774-414e-8afb-39a9a9d8b731-kube-api-access-ph8jm\") on node \"crc\" DevicePath \"\"" Mar 13 15:06:04 crc kubenswrapper[4907]: I0313 15:06:04.463187 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556906-kqncj" event={"ID":"7a501d2e-0774-414e-8afb-39a9a9d8b731","Type":"ContainerDied","Data":"a11a599b13906f7f8bcbe895f454a7ae524d8105c5bce3da9b1f7d5287818788"} Mar 13 15:06:04 crc kubenswrapper[4907]: I0313 15:06:04.463232 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a11a599b13906f7f8bcbe895f454a7ae524d8105c5bce3da9b1f7d5287818788" Mar 13 15:06:04 crc kubenswrapper[4907]: I0313 15:06:04.463249 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556906-kqncj" Mar 13 15:06:04 crc kubenswrapper[4907]: I0313 15:06:04.782140 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556900-thmpg"] Mar 13 15:06:04 crc kubenswrapper[4907]: I0313 15:06:04.788721 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556900-thmpg"] Mar 13 15:06:05 crc kubenswrapper[4907]: I0313 15:06:05.799943 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aafe442b-c955-456c-ac6d-a31322a5a27e" path="/var/lib/kubelet/pods/aafe442b-c955-456c-ac6d-a31322a5a27e/volumes" Mar 13 15:06:18 crc kubenswrapper[4907]: I0313 15:06:18.041600 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 15:06:18 crc kubenswrapper[4907]: I0313 15:06:18.042306 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 15:06:26 crc kubenswrapper[4907]: I0313 15:06:26.189604 4907 scope.go:117] "RemoveContainer" containerID="5fa3c1b84916d90898ac1395ff1cb1a7f88260f563edf9c1fdc38cde419411f5" Mar 13 15:06:48 crc kubenswrapper[4907]: I0313 15:06:48.042263 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 15:06:48 crc kubenswrapper[4907]: I0313 15:06:48.042794 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 15:06:48 crc kubenswrapper[4907]: I0313 15:06:48.042848 4907 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" Mar 13 15:06:48 crc kubenswrapper[4907]: I0313 15:06:48.043501 4907 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"caeb29efd9b0ec75cd14d4d1c0fbb54dce8073ba7639ceb605cc152e79561f8a"} pod="openshift-machine-config-operator/machine-config-daemon-hm56j" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 15:06:48 crc kubenswrapper[4907]: I0313 15:06:48.043553 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" containerID="cri-o://caeb29efd9b0ec75cd14d4d1c0fbb54dce8073ba7639ceb605cc152e79561f8a" gracePeriod=600 Mar 13 15:06:48 crc kubenswrapper[4907]: E0313 15:06:48.164973 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:06:48 crc kubenswrapper[4907]: I0313 15:06:48.788688 4907 generic.go:334] "Generic (PLEG): container finished" podID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerID="caeb29efd9b0ec75cd14d4d1c0fbb54dce8073ba7639ceb605cc152e79561f8a" exitCode=0 Mar 13 15:06:48 crc kubenswrapper[4907]: I0313 15:06:48.788745 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerDied","Data":"caeb29efd9b0ec75cd14d4d1c0fbb54dce8073ba7639ceb605cc152e79561f8a"} Mar 13 15:06:48 crc kubenswrapper[4907]: I0313 15:06:48.789116 4907 scope.go:117] "RemoveContainer" containerID="62e3208576ab64ce53d2a23023d86e9e8cd79b456c2ef95c5adcb1299b258391" Mar 13 15:06:48 crc kubenswrapper[4907]: I0313 15:06:48.790980 4907 scope.go:117] "RemoveContainer" containerID="caeb29efd9b0ec75cd14d4d1c0fbb54dce8073ba7639ceb605cc152e79561f8a" Mar 13 15:06:48 crc kubenswrapper[4907]: E0313 15:06:48.791577 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:07:03 crc kubenswrapper[4907]: I0313 15:07:03.782013 4907 scope.go:117] "RemoveContainer" containerID="caeb29efd9b0ec75cd14d4d1c0fbb54dce8073ba7639ceb605cc152e79561f8a" Mar 13 15:07:03 crc kubenswrapper[4907]: E0313 15:07:03.782780 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:07:17 crc kubenswrapper[4907]: I0313 15:07:17.783019 4907 scope.go:117] "RemoveContainer" containerID="caeb29efd9b0ec75cd14d4d1c0fbb54dce8073ba7639ceb605cc152e79561f8a" Mar 13 15:07:17 crc kubenswrapper[4907]: E0313 15:07:17.783697 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:07:30 crc kubenswrapper[4907]: I0313 15:07:30.782645 4907 scope.go:117] "RemoveContainer" containerID="caeb29efd9b0ec75cd14d4d1c0fbb54dce8073ba7639ceb605cc152e79561f8a" Mar 13 15:07:30 crc kubenswrapper[4907]: E0313 15:07:30.783472 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:07:44 crc kubenswrapper[4907]: I0313 15:07:44.782945 4907 scope.go:117] "RemoveContainer" containerID="caeb29efd9b0ec75cd14d4d1c0fbb54dce8073ba7639ceb605cc152e79561f8a" Mar 13 15:07:44 crc kubenswrapper[4907]: E0313 15:07:44.783571 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:07:57 crc kubenswrapper[4907]: I0313 15:07:57.783515 4907 scope.go:117] "RemoveContainer" containerID="caeb29efd9b0ec75cd14d4d1c0fbb54dce8073ba7639ceb605cc152e79561f8a" Mar 13 15:07:57 crc kubenswrapper[4907]: E0313 15:07:57.784279 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:08:00 crc kubenswrapper[4907]: I0313 15:08:00.143746 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556908-lsqcs"] Mar 13 15:08:00 crc kubenswrapper[4907]: E0313 15:08:00.144184 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a501d2e-0774-414e-8afb-39a9a9d8b731" containerName="oc" Mar 13 15:08:00 crc kubenswrapper[4907]: I0313 15:08:00.144201 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a501d2e-0774-414e-8afb-39a9a9d8b731" containerName="oc" Mar 13 15:08:00 crc kubenswrapper[4907]: I0313 15:08:00.144371 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a501d2e-0774-414e-8afb-39a9a9d8b731" containerName="oc" Mar 13 15:08:00 crc kubenswrapper[4907]: I0313 15:08:00.144928 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556908-lsqcs" Mar 13 15:08:00 crc kubenswrapper[4907]: I0313 15:08:00.146993 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 15:08:00 crc kubenswrapper[4907]: I0313 15:08:00.147242 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 15:08:00 crc kubenswrapper[4907]: I0313 15:08:00.147772 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 15:08:00 crc kubenswrapper[4907]: I0313 15:08:00.155552 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556908-lsqcs"] Mar 13 15:08:00 crc kubenswrapper[4907]: I0313 15:08:00.180710 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cl9zt\" (UniqueName: \"kubernetes.io/projected/555f4e30-b850-428b-a0fc-e15af18a58be-kube-api-access-cl9zt\") pod \"auto-csr-approver-29556908-lsqcs\" (UID: \"555f4e30-b850-428b-a0fc-e15af18a58be\") " pod="openshift-infra/auto-csr-approver-29556908-lsqcs" Mar 13 15:08:00 crc kubenswrapper[4907]: I0313 15:08:00.282690 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cl9zt\" (UniqueName: \"kubernetes.io/projected/555f4e30-b850-428b-a0fc-e15af18a58be-kube-api-access-cl9zt\") pod \"auto-csr-approver-29556908-lsqcs\" (UID: \"555f4e30-b850-428b-a0fc-e15af18a58be\") " pod="openshift-infra/auto-csr-approver-29556908-lsqcs" Mar 13 15:08:00 crc kubenswrapper[4907]: I0313 15:08:00.299972 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cl9zt\" (UniqueName: \"kubernetes.io/projected/555f4e30-b850-428b-a0fc-e15af18a58be-kube-api-access-cl9zt\") pod \"auto-csr-approver-29556908-lsqcs\" (UID: \"555f4e30-b850-428b-a0fc-e15af18a58be\") " pod="openshift-infra/auto-csr-approver-29556908-lsqcs" Mar 13 15:08:00 crc kubenswrapper[4907]: I0313 15:08:00.468703 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556908-lsqcs" Mar 13 15:08:00 crc kubenswrapper[4907]: I0313 15:08:00.871357 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556908-lsqcs"] Mar 13 15:08:00 crc kubenswrapper[4907]: I0313 15:08:00.877257 4907 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 13 15:08:01 crc kubenswrapper[4907]: I0313 15:08:01.325595 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556908-lsqcs" event={"ID":"555f4e30-b850-428b-a0fc-e15af18a58be","Type":"ContainerStarted","Data":"a174107a87cee85d700bd1c92ce4cce88bc356e63d464a68d165ca89b478e124"} Mar 13 15:08:03 crc kubenswrapper[4907]: I0313 15:08:03.342683 4907 generic.go:334] "Generic (PLEG): container finished" podID="555f4e30-b850-428b-a0fc-e15af18a58be" containerID="b199bdf3c496349ef8f75e1af0202d9bd0ccc18b9b13016f199b8123d6045267" exitCode=0 Mar 13 15:08:03 crc kubenswrapper[4907]: I0313 15:08:03.342740 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556908-lsqcs" event={"ID":"555f4e30-b850-428b-a0fc-e15af18a58be","Type":"ContainerDied","Data":"b199bdf3c496349ef8f75e1af0202d9bd0ccc18b9b13016f199b8123d6045267"} Mar 13 15:08:04 crc kubenswrapper[4907]: I0313 15:08:04.610642 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556908-lsqcs" Mar 13 15:08:04 crc kubenswrapper[4907]: I0313 15:08:04.649392 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cl9zt\" (UniqueName: \"kubernetes.io/projected/555f4e30-b850-428b-a0fc-e15af18a58be-kube-api-access-cl9zt\") pod \"555f4e30-b850-428b-a0fc-e15af18a58be\" (UID: \"555f4e30-b850-428b-a0fc-e15af18a58be\") " Mar 13 15:08:04 crc kubenswrapper[4907]: I0313 15:08:04.654544 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/555f4e30-b850-428b-a0fc-e15af18a58be-kube-api-access-cl9zt" (OuterVolumeSpecName: "kube-api-access-cl9zt") pod "555f4e30-b850-428b-a0fc-e15af18a58be" (UID: "555f4e30-b850-428b-a0fc-e15af18a58be"). InnerVolumeSpecName "kube-api-access-cl9zt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:08:04 crc kubenswrapper[4907]: I0313 15:08:04.750722 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cl9zt\" (UniqueName: \"kubernetes.io/projected/555f4e30-b850-428b-a0fc-e15af18a58be-kube-api-access-cl9zt\") on node \"crc\" DevicePath \"\"" Mar 13 15:08:05 crc kubenswrapper[4907]: I0313 15:08:05.356852 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556908-lsqcs" event={"ID":"555f4e30-b850-428b-a0fc-e15af18a58be","Type":"ContainerDied","Data":"a174107a87cee85d700bd1c92ce4cce88bc356e63d464a68d165ca89b478e124"} Mar 13 15:08:05 crc kubenswrapper[4907]: I0313 15:08:05.357098 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a174107a87cee85d700bd1c92ce4cce88bc356e63d464a68d165ca89b478e124" Mar 13 15:08:05 crc kubenswrapper[4907]: I0313 15:08:05.356949 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556908-lsqcs" Mar 13 15:08:05 crc kubenswrapper[4907]: I0313 15:08:05.675222 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556902-92pzz"] Mar 13 15:08:05 crc kubenswrapper[4907]: I0313 15:08:05.681109 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556902-92pzz"] Mar 13 15:08:05 crc kubenswrapper[4907]: I0313 15:08:05.792757 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="076c2b87-8e1f-4dba-9b60-eb24660aa1f1" path="/var/lib/kubelet/pods/076c2b87-8e1f-4dba-9b60-eb24660aa1f1/volumes" Mar 13 15:08:12 crc kubenswrapper[4907]: I0313 15:08:12.782190 4907 scope.go:117] "RemoveContainer" containerID="caeb29efd9b0ec75cd14d4d1c0fbb54dce8073ba7639ceb605cc152e79561f8a" Mar 13 15:08:12 crc kubenswrapper[4907]: E0313 15:08:12.782997 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:08:26 crc kubenswrapper[4907]: I0313 15:08:26.270916 4907 scope.go:117] "RemoveContainer" containerID="be729cf40227f8298124611ddf4a79928c79ddf05e8dab4417abc40fd127f33b" Mar 13 15:08:26 crc kubenswrapper[4907]: I0313 15:08:26.782469 4907 scope.go:117] "RemoveContainer" containerID="caeb29efd9b0ec75cd14d4d1c0fbb54dce8073ba7639ceb605cc152e79561f8a" Mar 13 15:08:26 crc kubenswrapper[4907]: E0313 15:08:26.783017 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:08:41 crc kubenswrapper[4907]: I0313 15:08:41.787201 4907 scope.go:117] "RemoveContainer" containerID="caeb29efd9b0ec75cd14d4d1c0fbb54dce8073ba7639ceb605cc152e79561f8a" Mar 13 15:08:41 crc kubenswrapper[4907]: E0313 15:08:41.788000 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:08:54 crc kubenswrapper[4907]: I0313 15:08:54.783203 4907 scope.go:117] "RemoveContainer" containerID="caeb29efd9b0ec75cd14d4d1c0fbb54dce8073ba7639ceb605cc152e79561f8a" Mar 13 15:08:54 crc kubenswrapper[4907]: E0313 15:08:54.784306 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:09:07 crc kubenswrapper[4907]: I0313 15:09:07.782047 4907 scope.go:117] "RemoveContainer" containerID="caeb29efd9b0ec75cd14d4d1c0fbb54dce8073ba7639ceb605cc152e79561f8a" Mar 13 15:09:07 crc kubenswrapper[4907]: E0313 15:09:07.782838 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:09:18 crc kubenswrapper[4907]: I0313 15:09:18.782540 4907 scope.go:117] "RemoveContainer" containerID="caeb29efd9b0ec75cd14d4d1c0fbb54dce8073ba7639ceb605cc152e79561f8a" Mar 13 15:09:18 crc kubenswrapper[4907]: E0313 15:09:18.783140 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:09:31 crc kubenswrapper[4907]: I0313 15:09:31.790525 4907 scope.go:117] "RemoveContainer" containerID="caeb29efd9b0ec75cd14d4d1c0fbb54dce8073ba7639ceb605cc152e79561f8a" Mar 13 15:09:31 crc kubenswrapper[4907]: E0313 15:09:31.791208 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:09:42 crc kubenswrapper[4907]: I0313 15:09:42.782732 4907 scope.go:117] "RemoveContainer" containerID="caeb29efd9b0ec75cd14d4d1c0fbb54dce8073ba7639ceb605cc152e79561f8a" Mar 13 15:09:42 crc kubenswrapper[4907]: E0313 15:09:42.783629 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:09:57 crc kubenswrapper[4907]: I0313 15:09:57.782752 4907 scope.go:117] "RemoveContainer" containerID="caeb29efd9b0ec75cd14d4d1c0fbb54dce8073ba7639ceb605cc152e79561f8a" Mar 13 15:09:57 crc kubenswrapper[4907]: E0313 15:09:57.784155 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:10:00 crc kubenswrapper[4907]: I0313 15:10:00.148097 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556910-5pmwx"] Mar 13 15:10:00 crc kubenswrapper[4907]: E0313 15:10:00.148547 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="555f4e30-b850-428b-a0fc-e15af18a58be" containerName="oc" Mar 13 15:10:00 crc kubenswrapper[4907]: I0313 15:10:00.148565 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="555f4e30-b850-428b-a0fc-e15af18a58be" containerName="oc" Mar 13 15:10:00 crc kubenswrapper[4907]: I0313 15:10:00.148814 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="555f4e30-b850-428b-a0fc-e15af18a58be" containerName="oc" Mar 13 15:10:00 crc kubenswrapper[4907]: I0313 15:10:00.149509 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556910-5pmwx" Mar 13 15:10:00 crc kubenswrapper[4907]: I0313 15:10:00.151380 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 15:10:00 crc kubenswrapper[4907]: I0313 15:10:00.152026 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 15:10:00 crc kubenswrapper[4907]: I0313 15:10:00.152359 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 15:10:00 crc kubenswrapper[4907]: I0313 15:10:00.164512 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556910-5pmwx"] Mar 13 15:10:00 crc kubenswrapper[4907]: I0313 15:10:00.285597 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8kv6l\" (UniqueName: \"kubernetes.io/projected/e74e8820-9ca4-4010-9070-ef32010a85c8-kube-api-access-8kv6l\") pod \"auto-csr-approver-29556910-5pmwx\" (UID: \"e74e8820-9ca4-4010-9070-ef32010a85c8\") " pod="openshift-infra/auto-csr-approver-29556910-5pmwx" Mar 13 15:10:00 crc kubenswrapper[4907]: I0313 15:10:00.387073 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8kv6l\" (UniqueName: \"kubernetes.io/projected/e74e8820-9ca4-4010-9070-ef32010a85c8-kube-api-access-8kv6l\") pod \"auto-csr-approver-29556910-5pmwx\" (UID: \"e74e8820-9ca4-4010-9070-ef32010a85c8\") " pod="openshift-infra/auto-csr-approver-29556910-5pmwx" Mar 13 15:10:00 crc kubenswrapper[4907]: I0313 15:10:00.425134 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8kv6l\" (UniqueName: \"kubernetes.io/projected/e74e8820-9ca4-4010-9070-ef32010a85c8-kube-api-access-8kv6l\") pod \"auto-csr-approver-29556910-5pmwx\" (UID: \"e74e8820-9ca4-4010-9070-ef32010a85c8\") " pod="openshift-infra/auto-csr-approver-29556910-5pmwx" Mar 13 15:10:00 crc kubenswrapper[4907]: I0313 15:10:00.468050 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556910-5pmwx" Mar 13 15:10:00 crc kubenswrapper[4907]: I0313 15:10:00.712852 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556910-5pmwx"] Mar 13 15:10:00 crc kubenswrapper[4907]: W0313 15:10:00.717251 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode74e8820_9ca4_4010_9070_ef32010a85c8.slice/crio-db3872fd4c5e6ef8fbd9653013661910abc918fc75aec32d6e1e26db23ebf10d WatchSource:0}: Error finding container db3872fd4c5e6ef8fbd9653013661910abc918fc75aec32d6e1e26db23ebf10d: Status 404 returned error can't find the container with id db3872fd4c5e6ef8fbd9653013661910abc918fc75aec32d6e1e26db23ebf10d Mar 13 15:10:01 crc kubenswrapper[4907]: I0313 15:10:01.233503 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556910-5pmwx" event={"ID":"e74e8820-9ca4-4010-9070-ef32010a85c8","Type":"ContainerStarted","Data":"db3872fd4c5e6ef8fbd9653013661910abc918fc75aec32d6e1e26db23ebf10d"} Mar 13 15:10:04 crc kubenswrapper[4907]: I0313 15:10:04.254523 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556910-5pmwx" event={"ID":"e74e8820-9ca4-4010-9070-ef32010a85c8","Type":"ContainerStarted","Data":"6156d45b4121389ab4793e9ba61feb497d3ccb0338571d8b44490cfdec3820ce"} Mar 13 15:10:04 crc kubenswrapper[4907]: I0313 15:10:04.270839 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29556910-5pmwx" podStartSLOduration=1.299812045 podStartE2EDuration="4.270808645s" podCreationTimestamp="2026-03-13 15:10:00 +0000 UTC" firstStartedPulling="2026-03-13 15:10:00.720067765 +0000 UTC m=+3899.619855464" lastFinishedPulling="2026-03-13 15:10:03.691064335 +0000 UTC m=+3902.590852064" observedRunningTime="2026-03-13 15:10:04.267683409 +0000 UTC m=+3903.167471098" watchObservedRunningTime="2026-03-13 15:10:04.270808645 +0000 UTC m=+3903.170596334" Mar 13 15:10:05 crc kubenswrapper[4907]: I0313 15:10:05.263337 4907 generic.go:334] "Generic (PLEG): container finished" podID="e74e8820-9ca4-4010-9070-ef32010a85c8" containerID="6156d45b4121389ab4793e9ba61feb497d3ccb0338571d8b44490cfdec3820ce" exitCode=0 Mar 13 15:10:05 crc kubenswrapper[4907]: I0313 15:10:05.263439 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556910-5pmwx" event={"ID":"e74e8820-9ca4-4010-9070-ef32010a85c8","Type":"ContainerDied","Data":"6156d45b4121389ab4793e9ba61feb497d3ccb0338571d8b44490cfdec3820ce"} Mar 13 15:10:06 crc kubenswrapper[4907]: I0313 15:10:06.536432 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556910-5pmwx" Mar 13 15:10:06 crc kubenswrapper[4907]: I0313 15:10:06.682067 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8kv6l\" (UniqueName: \"kubernetes.io/projected/e74e8820-9ca4-4010-9070-ef32010a85c8-kube-api-access-8kv6l\") pod \"e74e8820-9ca4-4010-9070-ef32010a85c8\" (UID: \"e74e8820-9ca4-4010-9070-ef32010a85c8\") " Mar 13 15:10:06 crc kubenswrapper[4907]: I0313 15:10:06.688296 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e74e8820-9ca4-4010-9070-ef32010a85c8-kube-api-access-8kv6l" (OuterVolumeSpecName: "kube-api-access-8kv6l") pod "e74e8820-9ca4-4010-9070-ef32010a85c8" (UID: "e74e8820-9ca4-4010-9070-ef32010a85c8"). InnerVolumeSpecName "kube-api-access-8kv6l". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:10:06 crc kubenswrapper[4907]: I0313 15:10:06.783983 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8kv6l\" (UniqueName: \"kubernetes.io/projected/e74e8820-9ca4-4010-9070-ef32010a85c8-kube-api-access-8kv6l\") on node \"crc\" DevicePath \"\"" Mar 13 15:10:07 crc kubenswrapper[4907]: I0313 15:10:07.282800 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556910-5pmwx" event={"ID":"e74e8820-9ca4-4010-9070-ef32010a85c8","Type":"ContainerDied","Data":"db3872fd4c5e6ef8fbd9653013661910abc918fc75aec32d6e1e26db23ebf10d"} Mar 13 15:10:07 crc kubenswrapper[4907]: I0313 15:10:07.282861 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="db3872fd4c5e6ef8fbd9653013661910abc918fc75aec32d6e1e26db23ebf10d" Mar 13 15:10:07 crc kubenswrapper[4907]: I0313 15:10:07.282872 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556910-5pmwx" Mar 13 15:10:07 crc kubenswrapper[4907]: I0313 15:10:07.350184 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556904-4qqhq"] Mar 13 15:10:07 crc kubenswrapper[4907]: I0313 15:10:07.357302 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556904-4qqhq"] Mar 13 15:10:07 crc kubenswrapper[4907]: I0313 15:10:07.794632 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="856cd90a-f3c2-4b46-b5eb-77e85004dbf6" path="/var/lib/kubelet/pods/856cd90a-f3c2-4b46-b5eb-77e85004dbf6/volumes" Mar 13 15:10:10 crc kubenswrapper[4907]: I0313 15:10:10.782925 4907 scope.go:117] "RemoveContainer" containerID="caeb29efd9b0ec75cd14d4d1c0fbb54dce8073ba7639ceb605cc152e79561f8a" Mar 13 15:10:10 crc kubenswrapper[4907]: E0313 15:10:10.783997 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:10:25 crc kubenswrapper[4907]: I0313 15:10:25.782753 4907 scope.go:117] "RemoveContainer" containerID="caeb29efd9b0ec75cd14d4d1c0fbb54dce8073ba7639ceb605cc152e79561f8a" Mar 13 15:10:25 crc kubenswrapper[4907]: E0313 15:10:25.783392 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:10:26 crc kubenswrapper[4907]: I0313 15:10:26.345522 4907 scope.go:117] "RemoveContainer" containerID="6b49bfb9250df806b9fbdd020bba3661e9f6dd127456008ebc47a00dd7745594" Mar 13 15:10:40 crc kubenswrapper[4907]: I0313 15:10:40.782777 4907 scope.go:117] "RemoveContainer" containerID="caeb29efd9b0ec75cd14d4d1c0fbb54dce8073ba7639ceb605cc152e79561f8a" Mar 13 15:10:40 crc kubenswrapper[4907]: E0313 15:10:40.783542 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:10:55 crc kubenswrapper[4907]: I0313 15:10:55.782783 4907 scope.go:117] "RemoveContainer" containerID="caeb29efd9b0ec75cd14d4d1c0fbb54dce8073ba7639ceb605cc152e79561f8a" Mar 13 15:10:55 crc kubenswrapper[4907]: E0313 15:10:55.785115 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:11:10 crc kubenswrapper[4907]: I0313 15:11:10.782957 4907 scope.go:117] "RemoveContainer" containerID="caeb29efd9b0ec75cd14d4d1c0fbb54dce8073ba7639ceb605cc152e79561f8a" Mar 13 15:11:10 crc kubenswrapper[4907]: E0313 15:11:10.783788 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:11:20 crc kubenswrapper[4907]: I0313 15:11:20.357840 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-cxj9h"] Mar 13 15:11:20 crc kubenswrapper[4907]: E0313 15:11:20.358715 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e74e8820-9ca4-4010-9070-ef32010a85c8" containerName="oc" Mar 13 15:11:20 crc kubenswrapper[4907]: I0313 15:11:20.358730 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="e74e8820-9ca4-4010-9070-ef32010a85c8" containerName="oc" Mar 13 15:11:20 crc kubenswrapper[4907]: I0313 15:11:20.358871 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="e74e8820-9ca4-4010-9070-ef32010a85c8" containerName="oc" Mar 13 15:11:20 crc kubenswrapper[4907]: I0313 15:11:20.359796 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cxj9h" Mar 13 15:11:20 crc kubenswrapper[4907]: I0313 15:11:20.371657 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-cxj9h"] Mar 13 15:11:20 crc kubenswrapper[4907]: I0313 15:11:20.557569 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dd7cb0e4-ac20-4ff7-8b04-d40718c41d40-utilities\") pod \"redhat-marketplace-cxj9h\" (UID: \"dd7cb0e4-ac20-4ff7-8b04-d40718c41d40\") " pod="openshift-marketplace/redhat-marketplace-cxj9h" Mar 13 15:11:20 crc kubenswrapper[4907]: I0313 15:11:20.557643 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nzn6b\" (UniqueName: \"kubernetes.io/projected/dd7cb0e4-ac20-4ff7-8b04-d40718c41d40-kube-api-access-nzn6b\") pod \"redhat-marketplace-cxj9h\" (UID: \"dd7cb0e4-ac20-4ff7-8b04-d40718c41d40\") " pod="openshift-marketplace/redhat-marketplace-cxj9h" Mar 13 15:11:20 crc kubenswrapper[4907]: I0313 15:11:20.557766 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dd7cb0e4-ac20-4ff7-8b04-d40718c41d40-catalog-content\") pod \"redhat-marketplace-cxj9h\" (UID: \"dd7cb0e4-ac20-4ff7-8b04-d40718c41d40\") " pod="openshift-marketplace/redhat-marketplace-cxj9h" Mar 13 15:11:20 crc kubenswrapper[4907]: I0313 15:11:20.659471 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dd7cb0e4-ac20-4ff7-8b04-d40718c41d40-catalog-content\") pod \"redhat-marketplace-cxj9h\" (UID: \"dd7cb0e4-ac20-4ff7-8b04-d40718c41d40\") " pod="openshift-marketplace/redhat-marketplace-cxj9h" Mar 13 15:11:20 crc kubenswrapper[4907]: I0313 15:11:20.659596 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dd7cb0e4-ac20-4ff7-8b04-d40718c41d40-utilities\") pod \"redhat-marketplace-cxj9h\" (UID: \"dd7cb0e4-ac20-4ff7-8b04-d40718c41d40\") " pod="openshift-marketplace/redhat-marketplace-cxj9h" Mar 13 15:11:20 crc kubenswrapper[4907]: I0313 15:11:20.659635 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nzn6b\" (UniqueName: \"kubernetes.io/projected/dd7cb0e4-ac20-4ff7-8b04-d40718c41d40-kube-api-access-nzn6b\") pod \"redhat-marketplace-cxj9h\" (UID: \"dd7cb0e4-ac20-4ff7-8b04-d40718c41d40\") " pod="openshift-marketplace/redhat-marketplace-cxj9h" Mar 13 15:11:20 crc kubenswrapper[4907]: I0313 15:11:20.660085 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dd7cb0e4-ac20-4ff7-8b04-d40718c41d40-catalog-content\") pod \"redhat-marketplace-cxj9h\" (UID: \"dd7cb0e4-ac20-4ff7-8b04-d40718c41d40\") " pod="openshift-marketplace/redhat-marketplace-cxj9h" Mar 13 15:11:20 crc kubenswrapper[4907]: I0313 15:11:20.660140 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dd7cb0e4-ac20-4ff7-8b04-d40718c41d40-utilities\") pod \"redhat-marketplace-cxj9h\" (UID: \"dd7cb0e4-ac20-4ff7-8b04-d40718c41d40\") " pod="openshift-marketplace/redhat-marketplace-cxj9h" Mar 13 15:11:20 crc kubenswrapper[4907]: I0313 15:11:20.707130 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nzn6b\" (UniqueName: \"kubernetes.io/projected/dd7cb0e4-ac20-4ff7-8b04-d40718c41d40-kube-api-access-nzn6b\") pod \"redhat-marketplace-cxj9h\" (UID: \"dd7cb0e4-ac20-4ff7-8b04-d40718c41d40\") " pod="openshift-marketplace/redhat-marketplace-cxj9h" Mar 13 15:11:20 crc kubenswrapper[4907]: I0313 15:11:20.980245 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cxj9h" Mar 13 15:11:21 crc kubenswrapper[4907]: I0313 15:11:21.537960 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-cxj9h"] Mar 13 15:11:21 crc kubenswrapper[4907]: I0313 15:11:21.786390 4907 scope.go:117] "RemoveContainer" containerID="caeb29efd9b0ec75cd14d4d1c0fbb54dce8073ba7639ceb605cc152e79561f8a" Mar 13 15:11:21 crc kubenswrapper[4907]: E0313 15:11:21.786923 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:11:21 crc kubenswrapper[4907]: I0313 15:11:21.940628 4907 generic.go:334] "Generic (PLEG): container finished" podID="dd7cb0e4-ac20-4ff7-8b04-d40718c41d40" containerID="7d419ed4bd94d03f181c4d03b47d6360d106e83f89d0f80631b8c98942f0811d" exitCode=0 Mar 13 15:11:21 crc kubenswrapper[4907]: I0313 15:11:21.940671 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cxj9h" event={"ID":"dd7cb0e4-ac20-4ff7-8b04-d40718c41d40","Type":"ContainerDied","Data":"7d419ed4bd94d03f181c4d03b47d6360d106e83f89d0f80631b8c98942f0811d"} Mar 13 15:11:21 crc kubenswrapper[4907]: I0313 15:11:21.940717 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cxj9h" event={"ID":"dd7cb0e4-ac20-4ff7-8b04-d40718c41d40","Type":"ContainerStarted","Data":"5464248f8a1cc4999ddd5ec071572b52cc1accb86479d7ff5b0ea37426d14fde"} Mar 13 15:11:22 crc kubenswrapper[4907]: I0313 15:11:22.948931 4907 generic.go:334] "Generic (PLEG): container finished" podID="dd7cb0e4-ac20-4ff7-8b04-d40718c41d40" containerID="59fe9c3ca129487a9017d04f24aff47275d2cc5ba27e2667acd811f849f3d50c" exitCode=0 Mar 13 15:11:22 crc kubenswrapper[4907]: I0313 15:11:22.948983 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cxj9h" event={"ID":"dd7cb0e4-ac20-4ff7-8b04-d40718c41d40","Type":"ContainerDied","Data":"59fe9c3ca129487a9017d04f24aff47275d2cc5ba27e2667acd811f849f3d50c"} Mar 13 15:11:23 crc kubenswrapper[4907]: I0313 15:11:23.958861 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cxj9h" event={"ID":"dd7cb0e4-ac20-4ff7-8b04-d40718c41d40","Type":"ContainerStarted","Data":"78745dd5b00ec2ec5e49e3d60ba52f4f1eb321ccddaa4efe0e88a95ab80f8364"} Mar 13 15:11:23 crc kubenswrapper[4907]: I0313 15:11:23.982721 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-cxj9h" podStartSLOduration=2.495653829 podStartE2EDuration="3.982700466s" podCreationTimestamp="2026-03-13 15:11:20 +0000 UTC" firstStartedPulling="2026-03-13 15:11:21.942042095 +0000 UTC m=+3980.841829784" lastFinishedPulling="2026-03-13 15:11:23.429088722 +0000 UTC m=+3982.328876421" observedRunningTime="2026-03-13 15:11:23.978558653 +0000 UTC m=+3982.878346342" watchObservedRunningTime="2026-03-13 15:11:23.982700466 +0000 UTC m=+3982.882488155" Mar 13 15:11:30 crc kubenswrapper[4907]: I0313 15:11:30.980817 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-cxj9h" Mar 13 15:11:30 crc kubenswrapper[4907]: I0313 15:11:30.981369 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-cxj9h" Mar 13 15:11:31 crc kubenswrapper[4907]: I0313 15:11:31.044453 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-cxj9h" Mar 13 15:11:31 crc kubenswrapper[4907]: I0313 15:11:31.105050 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-cxj9h" Mar 13 15:11:31 crc kubenswrapper[4907]: I0313 15:11:31.287183 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-cxj9h"] Mar 13 15:11:32 crc kubenswrapper[4907]: I0313 15:11:32.782682 4907 scope.go:117] "RemoveContainer" containerID="caeb29efd9b0ec75cd14d4d1c0fbb54dce8073ba7639ceb605cc152e79561f8a" Mar 13 15:11:32 crc kubenswrapper[4907]: E0313 15:11:32.782944 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:11:33 crc kubenswrapper[4907]: I0313 15:11:33.020042 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-cxj9h" podUID="dd7cb0e4-ac20-4ff7-8b04-d40718c41d40" containerName="registry-server" containerID="cri-o://78745dd5b00ec2ec5e49e3d60ba52f4f1eb321ccddaa4efe0e88a95ab80f8364" gracePeriod=2 Mar 13 15:11:33 crc kubenswrapper[4907]: I0313 15:11:33.979858 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cxj9h" Mar 13 15:11:34 crc kubenswrapper[4907]: I0313 15:11:34.030897 4907 generic.go:334] "Generic (PLEG): container finished" podID="dd7cb0e4-ac20-4ff7-8b04-d40718c41d40" containerID="78745dd5b00ec2ec5e49e3d60ba52f4f1eb321ccddaa4efe0e88a95ab80f8364" exitCode=0 Mar 13 15:11:34 crc kubenswrapper[4907]: I0313 15:11:34.030942 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cxj9h" event={"ID":"dd7cb0e4-ac20-4ff7-8b04-d40718c41d40","Type":"ContainerDied","Data":"78745dd5b00ec2ec5e49e3d60ba52f4f1eb321ccddaa4efe0e88a95ab80f8364"} Mar 13 15:11:34 crc kubenswrapper[4907]: I0313 15:11:34.030972 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cxj9h" Mar 13 15:11:34 crc kubenswrapper[4907]: I0313 15:11:34.030988 4907 scope.go:117] "RemoveContainer" containerID="78745dd5b00ec2ec5e49e3d60ba52f4f1eb321ccddaa4efe0e88a95ab80f8364" Mar 13 15:11:34 crc kubenswrapper[4907]: I0313 15:11:34.030974 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cxj9h" event={"ID":"dd7cb0e4-ac20-4ff7-8b04-d40718c41d40","Type":"ContainerDied","Data":"5464248f8a1cc4999ddd5ec071572b52cc1accb86479d7ff5b0ea37426d14fde"} Mar 13 15:11:34 crc kubenswrapper[4907]: I0313 15:11:34.048937 4907 scope.go:117] "RemoveContainer" containerID="59fe9c3ca129487a9017d04f24aff47275d2cc5ba27e2667acd811f849f3d50c" Mar 13 15:11:34 crc kubenswrapper[4907]: I0313 15:11:34.069504 4907 scope.go:117] "RemoveContainer" containerID="7d419ed4bd94d03f181c4d03b47d6360d106e83f89d0f80631b8c98942f0811d" Mar 13 15:11:34 crc kubenswrapper[4907]: I0313 15:11:34.089958 4907 scope.go:117] "RemoveContainer" containerID="78745dd5b00ec2ec5e49e3d60ba52f4f1eb321ccddaa4efe0e88a95ab80f8364" Mar 13 15:11:34 crc kubenswrapper[4907]: E0313 15:11:34.090461 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"78745dd5b00ec2ec5e49e3d60ba52f4f1eb321ccddaa4efe0e88a95ab80f8364\": container with ID starting with 78745dd5b00ec2ec5e49e3d60ba52f4f1eb321ccddaa4efe0e88a95ab80f8364 not found: ID does not exist" containerID="78745dd5b00ec2ec5e49e3d60ba52f4f1eb321ccddaa4efe0e88a95ab80f8364" Mar 13 15:11:34 crc kubenswrapper[4907]: I0313 15:11:34.090490 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"78745dd5b00ec2ec5e49e3d60ba52f4f1eb321ccddaa4efe0e88a95ab80f8364"} err="failed to get container status \"78745dd5b00ec2ec5e49e3d60ba52f4f1eb321ccddaa4efe0e88a95ab80f8364\": rpc error: code = NotFound desc = could not find container \"78745dd5b00ec2ec5e49e3d60ba52f4f1eb321ccddaa4efe0e88a95ab80f8364\": container with ID starting with 78745dd5b00ec2ec5e49e3d60ba52f4f1eb321ccddaa4efe0e88a95ab80f8364 not found: ID does not exist" Mar 13 15:11:34 crc kubenswrapper[4907]: I0313 15:11:34.090515 4907 scope.go:117] "RemoveContainer" containerID="59fe9c3ca129487a9017d04f24aff47275d2cc5ba27e2667acd811f849f3d50c" Mar 13 15:11:34 crc kubenswrapper[4907]: E0313 15:11:34.091008 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"59fe9c3ca129487a9017d04f24aff47275d2cc5ba27e2667acd811f849f3d50c\": container with ID starting with 59fe9c3ca129487a9017d04f24aff47275d2cc5ba27e2667acd811f849f3d50c not found: ID does not exist" containerID="59fe9c3ca129487a9017d04f24aff47275d2cc5ba27e2667acd811f849f3d50c" Mar 13 15:11:34 crc kubenswrapper[4907]: I0313 15:11:34.091042 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"59fe9c3ca129487a9017d04f24aff47275d2cc5ba27e2667acd811f849f3d50c"} err="failed to get container status \"59fe9c3ca129487a9017d04f24aff47275d2cc5ba27e2667acd811f849f3d50c\": rpc error: code = NotFound desc = could not find container \"59fe9c3ca129487a9017d04f24aff47275d2cc5ba27e2667acd811f849f3d50c\": container with ID starting with 59fe9c3ca129487a9017d04f24aff47275d2cc5ba27e2667acd811f849f3d50c not found: ID does not exist" Mar 13 15:11:34 crc kubenswrapper[4907]: I0313 15:11:34.091062 4907 scope.go:117] "RemoveContainer" containerID="7d419ed4bd94d03f181c4d03b47d6360d106e83f89d0f80631b8c98942f0811d" Mar 13 15:11:34 crc kubenswrapper[4907]: E0313 15:11:34.091358 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7d419ed4bd94d03f181c4d03b47d6360d106e83f89d0f80631b8c98942f0811d\": container with ID starting with 7d419ed4bd94d03f181c4d03b47d6360d106e83f89d0f80631b8c98942f0811d not found: ID does not exist" containerID="7d419ed4bd94d03f181c4d03b47d6360d106e83f89d0f80631b8c98942f0811d" Mar 13 15:11:34 crc kubenswrapper[4907]: I0313 15:11:34.091385 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d419ed4bd94d03f181c4d03b47d6360d106e83f89d0f80631b8c98942f0811d"} err="failed to get container status \"7d419ed4bd94d03f181c4d03b47d6360d106e83f89d0f80631b8c98942f0811d\": rpc error: code = NotFound desc = could not find container \"7d419ed4bd94d03f181c4d03b47d6360d106e83f89d0f80631b8c98942f0811d\": container with ID starting with 7d419ed4bd94d03f181c4d03b47d6360d106e83f89d0f80631b8c98942f0811d not found: ID does not exist" Mar 13 15:11:34 crc kubenswrapper[4907]: I0313 15:11:34.145776 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dd7cb0e4-ac20-4ff7-8b04-d40718c41d40-utilities\") pod \"dd7cb0e4-ac20-4ff7-8b04-d40718c41d40\" (UID: \"dd7cb0e4-ac20-4ff7-8b04-d40718c41d40\") " Mar 13 15:11:34 crc kubenswrapper[4907]: I0313 15:11:34.145927 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzn6b\" (UniqueName: \"kubernetes.io/projected/dd7cb0e4-ac20-4ff7-8b04-d40718c41d40-kube-api-access-nzn6b\") pod \"dd7cb0e4-ac20-4ff7-8b04-d40718c41d40\" (UID: \"dd7cb0e4-ac20-4ff7-8b04-d40718c41d40\") " Mar 13 15:11:34 crc kubenswrapper[4907]: I0313 15:11:34.145953 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dd7cb0e4-ac20-4ff7-8b04-d40718c41d40-catalog-content\") pod \"dd7cb0e4-ac20-4ff7-8b04-d40718c41d40\" (UID: \"dd7cb0e4-ac20-4ff7-8b04-d40718c41d40\") " Mar 13 15:11:34 crc kubenswrapper[4907]: I0313 15:11:34.147743 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dd7cb0e4-ac20-4ff7-8b04-d40718c41d40-utilities" (OuterVolumeSpecName: "utilities") pod "dd7cb0e4-ac20-4ff7-8b04-d40718c41d40" (UID: "dd7cb0e4-ac20-4ff7-8b04-d40718c41d40"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:11:34 crc kubenswrapper[4907]: I0313 15:11:34.151406 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dd7cb0e4-ac20-4ff7-8b04-d40718c41d40-kube-api-access-nzn6b" (OuterVolumeSpecName: "kube-api-access-nzn6b") pod "dd7cb0e4-ac20-4ff7-8b04-d40718c41d40" (UID: "dd7cb0e4-ac20-4ff7-8b04-d40718c41d40"). InnerVolumeSpecName "kube-api-access-nzn6b". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:11:34 crc kubenswrapper[4907]: I0313 15:11:34.178154 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dd7cb0e4-ac20-4ff7-8b04-d40718c41d40-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "dd7cb0e4-ac20-4ff7-8b04-d40718c41d40" (UID: "dd7cb0e4-ac20-4ff7-8b04-d40718c41d40"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:11:34 crc kubenswrapper[4907]: I0313 15:11:34.248130 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dd7cb0e4-ac20-4ff7-8b04-d40718c41d40-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 15:11:34 crc kubenswrapper[4907]: I0313 15:11:34.248175 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzn6b\" (UniqueName: \"kubernetes.io/projected/dd7cb0e4-ac20-4ff7-8b04-d40718c41d40-kube-api-access-nzn6b\") on node \"crc\" DevicePath \"\"" Mar 13 15:11:34 crc kubenswrapper[4907]: I0313 15:11:34.248187 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dd7cb0e4-ac20-4ff7-8b04-d40718c41d40-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 15:11:34 crc kubenswrapper[4907]: I0313 15:11:34.368976 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-cxj9h"] Mar 13 15:11:34 crc kubenswrapper[4907]: I0313 15:11:34.374786 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-cxj9h"] Mar 13 15:11:35 crc kubenswrapper[4907]: I0313 15:11:35.795525 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dd7cb0e4-ac20-4ff7-8b04-d40718c41d40" path="/var/lib/kubelet/pods/dd7cb0e4-ac20-4ff7-8b04-d40718c41d40/volumes" Mar 13 15:11:46 crc kubenswrapper[4907]: I0313 15:11:46.782051 4907 scope.go:117] "RemoveContainer" containerID="caeb29efd9b0ec75cd14d4d1c0fbb54dce8073ba7639ceb605cc152e79561f8a" Mar 13 15:11:46 crc kubenswrapper[4907]: E0313 15:11:46.782827 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:12:00 crc kubenswrapper[4907]: I0313 15:12:00.164466 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556912-smqnn"] Mar 13 15:12:00 crc kubenswrapper[4907]: E0313 15:12:00.165563 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd7cb0e4-ac20-4ff7-8b04-d40718c41d40" containerName="extract-content" Mar 13 15:12:00 crc kubenswrapper[4907]: I0313 15:12:00.165582 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd7cb0e4-ac20-4ff7-8b04-d40718c41d40" containerName="extract-content" Mar 13 15:12:00 crc kubenswrapper[4907]: E0313 15:12:00.165599 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd7cb0e4-ac20-4ff7-8b04-d40718c41d40" containerName="registry-server" Mar 13 15:12:00 crc kubenswrapper[4907]: I0313 15:12:00.165608 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd7cb0e4-ac20-4ff7-8b04-d40718c41d40" containerName="registry-server" Mar 13 15:12:00 crc kubenswrapper[4907]: E0313 15:12:00.165620 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd7cb0e4-ac20-4ff7-8b04-d40718c41d40" containerName="extract-utilities" Mar 13 15:12:00 crc kubenswrapper[4907]: I0313 15:12:00.165629 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd7cb0e4-ac20-4ff7-8b04-d40718c41d40" containerName="extract-utilities" Mar 13 15:12:00 crc kubenswrapper[4907]: I0313 15:12:00.165846 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="dd7cb0e4-ac20-4ff7-8b04-d40718c41d40" containerName="registry-server" Mar 13 15:12:00 crc kubenswrapper[4907]: I0313 15:12:00.166642 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556912-smqnn" Mar 13 15:12:00 crc kubenswrapper[4907]: I0313 15:12:00.169168 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 15:12:00 crc kubenswrapper[4907]: I0313 15:12:00.169203 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 15:12:00 crc kubenswrapper[4907]: I0313 15:12:00.170428 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 15:12:00 crc kubenswrapper[4907]: I0313 15:12:00.178822 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556912-smqnn"] Mar 13 15:12:00 crc kubenswrapper[4907]: I0313 15:12:00.240577 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wg9qs\" (UniqueName: \"kubernetes.io/projected/449e0ca3-4756-4ec2-8ea2-b2b6fbe52ca9-kube-api-access-wg9qs\") pod \"auto-csr-approver-29556912-smqnn\" (UID: \"449e0ca3-4756-4ec2-8ea2-b2b6fbe52ca9\") " pod="openshift-infra/auto-csr-approver-29556912-smqnn" Mar 13 15:12:00 crc kubenswrapper[4907]: I0313 15:12:00.342473 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wg9qs\" (UniqueName: \"kubernetes.io/projected/449e0ca3-4756-4ec2-8ea2-b2b6fbe52ca9-kube-api-access-wg9qs\") pod \"auto-csr-approver-29556912-smqnn\" (UID: \"449e0ca3-4756-4ec2-8ea2-b2b6fbe52ca9\") " pod="openshift-infra/auto-csr-approver-29556912-smqnn" Mar 13 15:12:00 crc kubenswrapper[4907]: I0313 15:12:00.368143 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wg9qs\" (UniqueName: \"kubernetes.io/projected/449e0ca3-4756-4ec2-8ea2-b2b6fbe52ca9-kube-api-access-wg9qs\") pod \"auto-csr-approver-29556912-smqnn\" (UID: \"449e0ca3-4756-4ec2-8ea2-b2b6fbe52ca9\") " pod="openshift-infra/auto-csr-approver-29556912-smqnn" Mar 13 15:12:00 crc kubenswrapper[4907]: I0313 15:12:00.485613 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556912-smqnn" Mar 13 15:12:00 crc kubenswrapper[4907]: I0313 15:12:00.941230 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556912-smqnn"] Mar 13 15:12:01 crc kubenswrapper[4907]: I0313 15:12:01.216992 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556912-smqnn" event={"ID":"449e0ca3-4756-4ec2-8ea2-b2b6fbe52ca9","Type":"ContainerStarted","Data":"8b02e5696714ca6118ebedfe0bc75d417d937d6760d4c0c0b82ce55f9a51fdba"} Mar 13 15:12:01 crc kubenswrapper[4907]: I0313 15:12:01.792419 4907 scope.go:117] "RemoveContainer" containerID="caeb29efd9b0ec75cd14d4d1c0fbb54dce8073ba7639ceb605cc152e79561f8a" Mar 13 15:12:02 crc kubenswrapper[4907]: I0313 15:12:02.223300 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerStarted","Data":"36ef7b96f3457426d961c14798528949d4c51116631e09ea3be0c17a7c07b91d"} Mar 13 15:12:03 crc kubenswrapper[4907]: I0313 15:12:03.230478 4907 generic.go:334] "Generic (PLEG): container finished" podID="449e0ca3-4756-4ec2-8ea2-b2b6fbe52ca9" containerID="eb18542ee2a2c90ab854737b99c04f55fb12626f0e4778cc4e2869a3c3a6cd0f" exitCode=0 Mar 13 15:12:03 crc kubenswrapper[4907]: I0313 15:12:03.230529 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556912-smqnn" event={"ID":"449e0ca3-4756-4ec2-8ea2-b2b6fbe52ca9","Type":"ContainerDied","Data":"eb18542ee2a2c90ab854737b99c04f55fb12626f0e4778cc4e2869a3c3a6cd0f"} Mar 13 15:12:04 crc kubenswrapper[4907]: I0313 15:12:04.567963 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556912-smqnn" Mar 13 15:12:04 crc kubenswrapper[4907]: I0313 15:12:04.616561 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wg9qs\" (UniqueName: \"kubernetes.io/projected/449e0ca3-4756-4ec2-8ea2-b2b6fbe52ca9-kube-api-access-wg9qs\") pod \"449e0ca3-4756-4ec2-8ea2-b2b6fbe52ca9\" (UID: \"449e0ca3-4756-4ec2-8ea2-b2b6fbe52ca9\") " Mar 13 15:12:04 crc kubenswrapper[4907]: I0313 15:12:04.623193 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/449e0ca3-4756-4ec2-8ea2-b2b6fbe52ca9-kube-api-access-wg9qs" (OuterVolumeSpecName: "kube-api-access-wg9qs") pod "449e0ca3-4756-4ec2-8ea2-b2b6fbe52ca9" (UID: "449e0ca3-4756-4ec2-8ea2-b2b6fbe52ca9"). InnerVolumeSpecName "kube-api-access-wg9qs". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:12:04 crc kubenswrapper[4907]: I0313 15:12:04.718299 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wg9qs\" (UniqueName: \"kubernetes.io/projected/449e0ca3-4756-4ec2-8ea2-b2b6fbe52ca9-kube-api-access-wg9qs\") on node \"crc\" DevicePath \"\"" Mar 13 15:12:05 crc kubenswrapper[4907]: I0313 15:12:05.249947 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556912-smqnn" event={"ID":"449e0ca3-4756-4ec2-8ea2-b2b6fbe52ca9","Type":"ContainerDied","Data":"8b02e5696714ca6118ebedfe0bc75d417d937d6760d4c0c0b82ce55f9a51fdba"} Mar 13 15:12:05 crc kubenswrapper[4907]: I0313 15:12:05.249987 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8b02e5696714ca6118ebedfe0bc75d417d937d6760d4c0c0b82ce55f9a51fdba" Mar 13 15:12:05 crc kubenswrapper[4907]: I0313 15:12:05.249992 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556912-smqnn" Mar 13 15:12:05 crc kubenswrapper[4907]: I0313 15:12:05.637935 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556906-kqncj"] Mar 13 15:12:05 crc kubenswrapper[4907]: I0313 15:12:05.647847 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556906-kqncj"] Mar 13 15:12:05 crc kubenswrapper[4907]: I0313 15:12:05.792499 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7a501d2e-0774-414e-8afb-39a9a9d8b731" path="/var/lib/kubelet/pods/7a501d2e-0774-414e-8afb-39a9a9d8b731/volumes" Mar 13 15:12:26 crc kubenswrapper[4907]: I0313 15:12:26.477165 4907 scope.go:117] "RemoveContainer" containerID="c0f40228828ee5195fb41b365b21febab7b0feb3cf02e777e21a472c7ba50dbb" Mar 13 15:14:00 crc kubenswrapper[4907]: I0313 15:14:00.154281 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556914-98b9m"] Mar 13 15:14:00 crc kubenswrapper[4907]: E0313 15:14:00.156858 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="449e0ca3-4756-4ec2-8ea2-b2b6fbe52ca9" containerName="oc" Mar 13 15:14:00 crc kubenswrapper[4907]: I0313 15:14:00.156992 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="449e0ca3-4756-4ec2-8ea2-b2b6fbe52ca9" containerName="oc" Mar 13 15:14:00 crc kubenswrapper[4907]: I0313 15:14:00.157251 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="449e0ca3-4756-4ec2-8ea2-b2b6fbe52ca9" containerName="oc" Mar 13 15:14:00 crc kubenswrapper[4907]: I0313 15:14:00.157938 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556914-98b9m" Mar 13 15:14:00 crc kubenswrapper[4907]: I0313 15:14:00.161545 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556914-98b9m"] Mar 13 15:14:00 crc kubenswrapper[4907]: I0313 15:14:00.162270 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 15:14:00 crc kubenswrapper[4907]: I0313 15:14:00.162790 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 15:14:00 crc kubenswrapper[4907]: I0313 15:14:00.163293 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 15:14:00 crc kubenswrapper[4907]: I0313 15:14:00.331227 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j7vqk\" (UniqueName: \"kubernetes.io/projected/561da6ce-3fe7-4e35-97f0-6a39582d3d4b-kube-api-access-j7vqk\") pod \"auto-csr-approver-29556914-98b9m\" (UID: \"561da6ce-3fe7-4e35-97f0-6a39582d3d4b\") " pod="openshift-infra/auto-csr-approver-29556914-98b9m" Mar 13 15:14:00 crc kubenswrapper[4907]: I0313 15:14:00.433218 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j7vqk\" (UniqueName: \"kubernetes.io/projected/561da6ce-3fe7-4e35-97f0-6a39582d3d4b-kube-api-access-j7vqk\") pod \"auto-csr-approver-29556914-98b9m\" (UID: \"561da6ce-3fe7-4e35-97f0-6a39582d3d4b\") " pod="openshift-infra/auto-csr-approver-29556914-98b9m" Mar 13 15:14:00 crc kubenswrapper[4907]: I0313 15:14:00.453816 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j7vqk\" (UniqueName: \"kubernetes.io/projected/561da6ce-3fe7-4e35-97f0-6a39582d3d4b-kube-api-access-j7vqk\") pod \"auto-csr-approver-29556914-98b9m\" (UID: \"561da6ce-3fe7-4e35-97f0-6a39582d3d4b\") " pod="openshift-infra/auto-csr-approver-29556914-98b9m" Mar 13 15:14:00 crc kubenswrapper[4907]: I0313 15:14:00.498461 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556914-98b9m" Mar 13 15:14:00 crc kubenswrapper[4907]: I0313 15:14:00.913090 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556914-98b9m"] Mar 13 15:14:00 crc kubenswrapper[4907]: I0313 15:14:00.922726 4907 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 13 15:14:01 crc kubenswrapper[4907]: I0313 15:14:01.157204 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556914-98b9m" event={"ID":"561da6ce-3fe7-4e35-97f0-6a39582d3d4b","Type":"ContainerStarted","Data":"01eb835bc62db5d86ae4c0f5cf722d7dc8f945ba3bb95e0ff8b3af8bb49d05e1"} Mar 13 15:14:03 crc kubenswrapper[4907]: I0313 15:14:03.171976 4907 generic.go:334] "Generic (PLEG): container finished" podID="561da6ce-3fe7-4e35-97f0-6a39582d3d4b" containerID="692bff8b244c6ed0093ba50083b6fcff2e9a78403c7f3e3cc6626818a66971bb" exitCode=0 Mar 13 15:14:03 crc kubenswrapper[4907]: I0313 15:14:03.172172 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556914-98b9m" event={"ID":"561da6ce-3fe7-4e35-97f0-6a39582d3d4b","Type":"ContainerDied","Data":"692bff8b244c6ed0093ba50083b6fcff2e9a78403c7f3e3cc6626818a66971bb"} Mar 13 15:14:04 crc kubenswrapper[4907]: I0313 15:14:04.459985 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556914-98b9m" Mar 13 15:14:04 crc kubenswrapper[4907]: I0313 15:14:04.593206 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j7vqk\" (UniqueName: \"kubernetes.io/projected/561da6ce-3fe7-4e35-97f0-6a39582d3d4b-kube-api-access-j7vqk\") pod \"561da6ce-3fe7-4e35-97f0-6a39582d3d4b\" (UID: \"561da6ce-3fe7-4e35-97f0-6a39582d3d4b\") " Mar 13 15:14:04 crc kubenswrapper[4907]: I0313 15:14:04.598963 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/561da6ce-3fe7-4e35-97f0-6a39582d3d4b-kube-api-access-j7vqk" (OuterVolumeSpecName: "kube-api-access-j7vqk") pod "561da6ce-3fe7-4e35-97f0-6a39582d3d4b" (UID: "561da6ce-3fe7-4e35-97f0-6a39582d3d4b"). InnerVolumeSpecName "kube-api-access-j7vqk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:14:04 crc kubenswrapper[4907]: I0313 15:14:04.694637 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j7vqk\" (UniqueName: \"kubernetes.io/projected/561da6ce-3fe7-4e35-97f0-6a39582d3d4b-kube-api-access-j7vqk\") on node \"crc\" DevicePath \"\"" Mar 13 15:14:05 crc kubenswrapper[4907]: I0313 15:14:05.185101 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556914-98b9m" event={"ID":"561da6ce-3fe7-4e35-97f0-6a39582d3d4b","Type":"ContainerDied","Data":"01eb835bc62db5d86ae4c0f5cf722d7dc8f945ba3bb95e0ff8b3af8bb49d05e1"} Mar 13 15:14:05 crc kubenswrapper[4907]: I0313 15:14:05.185497 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="01eb835bc62db5d86ae4c0f5cf722d7dc8f945ba3bb95e0ff8b3af8bb49d05e1" Mar 13 15:14:05 crc kubenswrapper[4907]: I0313 15:14:05.185157 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556914-98b9m" Mar 13 15:14:05 crc kubenswrapper[4907]: I0313 15:14:05.526841 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556908-lsqcs"] Mar 13 15:14:05 crc kubenswrapper[4907]: I0313 15:14:05.531609 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556908-lsqcs"] Mar 13 15:14:05 crc kubenswrapper[4907]: I0313 15:14:05.792557 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="555f4e30-b850-428b-a0fc-e15af18a58be" path="/var/lib/kubelet/pods/555f4e30-b850-428b-a0fc-e15af18a58be/volumes" Mar 13 15:14:18 crc kubenswrapper[4907]: I0313 15:14:18.041630 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 15:14:18 crc kubenswrapper[4907]: I0313 15:14:18.043261 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 15:14:26 crc kubenswrapper[4907]: I0313 15:14:26.576481 4907 scope.go:117] "RemoveContainer" containerID="b199bdf3c496349ef8f75e1af0202d9bd0ccc18b9b13016f199b8123d6045267" Mar 13 15:14:48 crc kubenswrapper[4907]: I0313 15:14:48.042189 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 15:14:48 crc kubenswrapper[4907]: I0313 15:14:48.042831 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 15:14:58 crc kubenswrapper[4907]: I0313 15:14:58.029957 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-jpjgk"] Mar 13 15:14:58 crc kubenswrapper[4907]: E0313 15:14:58.035058 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="561da6ce-3fe7-4e35-97f0-6a39582d3d4b" containerName="oc" Mar 13 15:14:58 crc kubenswrapper[4907]: I0313 15:14:58.035105 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="561da6ce-3fe7-4e35-97f0-6a39582d3d4b" containerName="oc" Mar 13 15:14:58 crc kubenswrapper[4907]: I0313 15:14:58.035467 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="561da6ce-3fe7-4e35-97f0-6a39582d3d4b" containerName="oc" Mar 13 15:14:58 crc kubenswrapper[4907]: I0313 15:14:58.036779 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jpjgk" Mar 13 15:14:58 crc kubenswrapper[4907]: I0313 15:14:58.056796 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jpjgk"] Mar 13 15:14:58 crc kubenswrapper[4907]: I0313 15:14:58.159790 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bkfd5\" (UniqueName: \"kubernetes.io/projected/db74eac8-6eb9-4abc-8bd4-05a7c4de409d-kube-api-access-bkfd5\") pod \"community-operators-jpjgk\" (UID: \"db74eac8-6eb9-4abc-8bd4-05a7c4de409d\") " pod="openshift-marketplace/community-operators-jpjgk" Mar 13 15:14:58 crc kubenswrapper[4907]: I0313 15:14:58.159927 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db74eac8-6eb9-4abc-8bd4-05a7c4de409d-utilities\") pod \"community-operators-jpjgk\" (UID: \"db74eac8-6eb9-4abc-8bd4-05a7c4de409d\") " pod="openshift-marketplace/community-operators-jpjgk" Mar 13 15:14:58 crc kubenswrapper[4907]: I0313 15:14:58.160097 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db74eac8-6eb9-4abc-8bd4-05a7c4de409d-catalog-content\") pod \"community-operators-jpjgk\" (UID: \"db74eac8-6eb9-4abc-8bd4-05a7c4de409d\") " pod="openshift-marketplace/community-operators-jpjgk" Mar 13 15:14:58 crc kubenswrapper[4907]: I0313 15:14:58.261178 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db74eac8-6eb9-4abc-8bd4-05a7c4de409d-catalog-content\") pod \"community-operators-jpjgk\" (UID: \"db74eac8-6eb9-4abc-8bd4-05a7c4de409d\") " pod="openshift-marketplace/community-operators-jpjgk" Mar 13 15:14:58 crc kubenswrapper[4907]: I0313 15:14:58.261295 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bkfd5\" (UniqueName: \"kubernetes.io/projected/db74eac8-6eb9-4abc-8bd4-05a7c4de409d-kube-api-access-bkfd5\") pod \"community-operators-jpjgk\" (UID: \"db74eac8-6eb9-4abc-8bd4-05a7c4de409d\") " pod="openshift-marketplace/community-operators-jpjgk" Mar 13 15:14:58 crc kubenswrapper[4907]: I0313 15:14:58.261340 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db74eac8-6eb9-4abc-8bd4-05a7c4de409d-utilities\") pod \"community-operators-jpjgk\" (UID: \"db74eac8-6eb9-4abc-8bd4-05a7c4de409d\") " pod="openshift-marketplace/community-operators-jpjgk" Mar 13 15:14:58 crc kubenswrapper[4907]: I0313 15:14:58.261710 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db74eac8-6eb9-4abc-8bd4-05a7c4de409d-catalog-content\") pod \"community-operators-jpjgk\" (UID: \"db74eac8-6eb9-4abc-8bd4-05a7c4de409d\") " pod="openshift-marketplace/community-operators-jpjgk" Mar 13 15:14:58 crc kubenswrapper[4907]: I0313 15:14:58.261844 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db74eac8-6eb9-4abc-8bd4-05a7c4de409d-utilities\") pod \"community-operators-jpjgk\" (UID: \"db74eac8-6eb9-4abc-8bd4-05a7c4de409d\") " pod="openshift-marketplace/community-operators-jpjgk" Mar 13 15:14:58 crc kubenswrapper[4907]: I0313 15:14:58.281454 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bkfd5\" (UniqueName: \"kubernetes.io/projected/db74eac8-6eb9-4abc-8bd4-05a7c4de409d-kube-api-access-bkfd5\") pod \"community-operators-jpjgk\" (UID: \"db74eac8-6eb9-4abc-8bd4-05a7c4de409d\") " pod="openshift-marketplace/community-operators-jpjgk" Mar 13 15:14:58 crc kubenswrapper[4907]: I0313 15:14:58.369338 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jpjgk" Mar 13 15:14:58 crc kubenswrapper[4907]: I0313 15:14:58.652082 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jpjgk"] Mar 13 15:14:59 crc kubenswrapper[4907]: I0313 15:14:59.577463 4907 generic.go:334] "Generic (PLEG): container finished" podID="db74eac8-6eb9-4abc-8bd4-05a7c4de409d" containerID="40cf43ff58a1d7d18d1fe371ca4cdf1f9b55f77c9ec021d33f6565bbdaadc497" exitCode=0 Mar 13 15:14:59 crc kubenswrapper[4907]: I0313 15:14:59.577520 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jpjgk" event={"ID":"db74eac8-6eb9-4abc-8bd4-05a7c4de409d","Type":"ContainerDied","Data":"40cf43ff58a1d7d18d1fe371ca4cdf1f9b55f77c9ec021d33f6565bbdaadc497"} Mar 13 15:14:59 crc kubenswrapper[4907]: I0313 15:14:59.577551 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jpjgk" event={"ID":"db74eac8-6eb9-4abc-8bd4-05a7c4de409d","Type":"ContainerStarted","Data":"1e8f5904cbc6648ae855379c24f68a418662b1fb626449b7dd05f4af7ab877dc"} Mar 13 15:15:00 crc kubenswrapper[4907]: I0313 15:15:00.151615 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556915-jtzgq"] Mar 13 15:15:00 crc kubenswrapper[4907]: I0313 15:15:00.153066 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556915-jtzgq" Mar 13 15:15:00 crc kubenswrapper[4907]: I0313 15:15:00.155363 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 13 15:15:00 crc kubenswrapper[4907]: I0313 15:15:00.155496 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 13 15:15:00 crc kubenswrapper[4907]: I0313 15:15:00.159658 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556915-jtzgq"] Mar 13 15:15:00 crc kubenswrapper[4907]: I0313 15:15:00.189473 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1fb6a3c2-e8df-47a4-9ae7-2ee38842051a-secret-volume\") pod \"collect-profiles-29556915-jtzgq\" (UID: \"1fb6a3c2-e8df-47a4-9ae7-2ee38842051a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556915-jtzgq" Mar 13 15:15:00 crc kubenswrapper[4907]: I0313 15:15:00.190267 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fxh86\" (UniqueName: \"kubernetes.io/projected/1fb6a3c2-e8df-47a4-9ae7-2ee38842051a-kube-api-access-fxh86\") pod \"collect-profiles-29556915-jtzgq\" (UID: \"1fb6a3c2-e8df-47a4-9ae7-2ee38842051a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556915-jtzgq" Mar 13 15:15:00 crc kubenswrapper[4907]: I0313 15:15:00.190360 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1fb6a3c2-e8df-47a4-9ae7-2ee38842051a-config-volume\") pod \"collect-profiles-29556915-jtzgq\" (UID: \"1fb6a3c2-e8df-47a4-9ae7-2ee38842051a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556915-jtzgq" Mar 13 15:15:00 crc kubenswrapper[4907]: I0313 15:15:00.291041 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1fb6a3c2-e8df-47a4-9ae7-2ee38842051a-secret-volume\") pod \"collect-profiles-29556915-jtzgq\" (UID: \"1fb6a3c2-e8df-47a4-9ae7-2ee38842051a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556915-jtzgq" Mar 13 15:15:00 crc kubenswrapper[4907]: I0313 15:15:00.291108 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fxh86\" (UniqueName: \"kubernetes.io/projected/1fb6a3c2-e8df-47a4-9ae7-2ee38842051a-kube-api-access-fxh86\") pod \"collect-profiles-29556915-jtzgq\" (UID: \"1fb6a3c2-e8df-47a4-9ae7-2ee38842051a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556915-jtzgq" Mar 13 15:15:00 crc kubenswrapper[4907]: I0313 15:15:00.291171 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1fb6a3c2-e8df-47a4-9ae7-2ee38842051a-config-volume\") pod \"collect-profiles-29556915-jtzgq\" (UID: \"1fb6a3c2-e8df-47a4-9ae7-2ee38842051a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556915-jtzgq" Mar 13 15:15:00 crc kubenswrapper[4907]: I0313 15:15:00.292256 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1fb6a3c2-e8df-47a4-9ae7-2ee38842051a-config-volume\") pod \"collect-profiles-29556915-jtzgq\" (UID: \"1fb6a3c2-e8df-47a4-9ae7-2ee38842051a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556915-jtzgq" Mar 13 15:15:00 crc kubenswrapper[4907]: I0313 15:15:00.304625 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1fb6a3c2-e8df-47a4-9ae7-2ee38842051a-secret-volume\") pod \"collect-profiles-29556915-jtzgq\" (UID: \"1fb6a3c2-e8df-47a4-9ae7-2ee38842051a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556915-jtzgq" Mar 13 15:15:00 crc kubenswrapper[4907]: I0313 15:15:00.307974 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fxh86\" (UniqueName: \"kubernetes.io/projected/1fb6a3c2-e8df-47a4-9ae7-2ee38842051a-kube-api-access-fxh86\") pod \"collect-profiles-29556915-jtzgq\" (UID: \"1fb6a3c2-e8df-47a4-9ae7-2ee38842051a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556915-jtzgq" Mar 13 15:15:00 crc kubenswrapper[4907]: I0313 15:15:00.476986 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556915-jtzgq" Mar 13 15:15:00 crc kubenswrapper[4907]: I0313 15:15:00.596960 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jpjgk" event={"ID":"db74eac8-6eb9-4abc-8bd4-05a7c4de409d","Type":"ContainerStarted","Data":"43c417bf05f6d13ba2647567c1378035b8d2b1457c7230a97bd627f1015c7e9b"} Mar 13 15:15:00 crc kubenswrapper[4907]: I0313 15:15:00.898129 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556915-jtzgq"] Mar 13 15:15:00 crc kubenswrapper[4907]: W0313 15:15:00.908571 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1fb6a3c2_e8df_47a4_9ae7_2ee38842051a.slice/crio-8763e76268e6fdebf8c7827617d0774f78f5d7e8ce50c07417aa72eaaa7a2d83 WatchSource:0}: Error finding container 8763e76268e6fdebf8c7827617d0774f78f5d7e8ce50c07417aa72eaaa7a2d83: Status 404 returned error can't find the container with id 8763e76268e6fdebf8c7827617d0774f78f5d7e8ce50c07417aa72eaaa7a2d83 Mar 13 15:15:01 crc kubenswrapper[4907]: I0313 15:15:01.608666 4907 generic.go:334] "Generic (PLEG): container finished" podID="db74eac8-6eb9-4abc-8bd4-05a7c4de409d" containerID="43c417bf05f6d13ba2647567c1378035b8d2b1457c7230a97bd627f1015c7e9b" exitCode=0 Mar 13 15:15:01 crc kubenswrapper[4907]: I0313 15:15:01.608776 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jpjgk" event={"ID":"db74eac8-6eb9-4abc-8bd4-05a7c4de409d","Type":"ContainerDied","Data":"43c417bf05f6d13ba2647567c1378035b8d2b1457c7230a97bd627f1015c7e9b"} Mar 13 15:15:01 crc kubenswrapper[4907]: I0313 15:15:01.612313 4907 generic.go:334] "Generic (PLEG): container finished" podID="1fb6a3c2-e8df-47a4-9ae7-2ee38842051a" containerID="f8d1778bda06ccfd24371ed557383017005b23e6277f212ccf01bbaf98e64b94" exitCode=0 Mar 13 15:15:01 crc kubenswrapper[4907]: I0313 15:15:01.612368 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556915-jtzgq" event={"ID":"1fb6a3c2-e8df-47a4-9ae7-2ee38842051a","Type":"ContainerDied","Data":"f8d1778bda06ccfd24371ed557383017005b23e6277f212ccf01bbaf98e64b94"} Mar 13 15:15:01 crc kubenswrapper[4907]: I0313 15:15:01.612403 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556915-jtzgq" event={"ID":"1fb6a3c2-e8df-47a4-9ae7-2ee38842051a","Type":"ContainerStarted","Data":"8763e76268e6fdebf8c7827617d0774f78f5d7e8ce50c07417aa72eaaa7a2d83"} Mar 13 15:15:02 crc kubenswrapper[4907]: I0313 15:15:02.621257 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jpjgk" event={"ID":"db74eac8-6eb9-4abc-8bd4-05a7c4de409d","Type":"ContainerStarted","Data":"2832afb1ca9f1fcdd05a62540f3537314948bd51bef218b6399c500e2dfcca30"} Mar 13 15:15:02 crc kubenswrapper[4907]: I0313 15:15:02.643180 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-jpjgk" podStartSLOduration=2.119638662 podStartE2EDuration="4.643163727s" podCreationTimestamp="2026-03-13 15:14:58 +0000 UTC" firstStartedPulling="2026-03-13 15:14:59.580836934 +0000 UTC m=+4198.480624663" lastFinishedPulling="2026-03-13 15:15:02.104362039 +0000 UTC m=+4201.004149728" observedRunningTime="2026-03-13 15:15:02.638380766 +0000 UTC m=+4201.538168475" watchObservedRunningTime="2026-03-13 15:15:02.643163727 +0000 UTC m=+4201.542951416" Mar 13 15:15:02 crc kubenswrapper[4907]: I0313 15:15:02.882035 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556915-jtzgq" Mar 13 15:15:03 crc kubenswrapper[4907]: I0313 15:15:03.041270 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1fb6a3c2-e8df-47a4-9ae7-2ee38842051a-config-volume\") pod \"1fb6a3c2-e8df-47a4-9ae7-2ee38842051a\" (UID: \"1fb6a3c2-e8df-47a4-9ae7-2ee38842051a\") " Mar 13 15:15:03 crc kubenswrapper[4907]: I0313 15:15:03.041379 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fxh86\" (UniqueName: \"kubernetes.io/projected/1fb6a3c2-e8df-47a4-9ae7-2ee38842051a-kube-api-access-fxh86\") pod \"1fb6a3c2-e8df-47a4-9ae7-2ee38842051a\" (UID: \"1fb6a3c2-e8df-47a4-9ae7-2ee38842051a\") " Mar 13 15:15:03 crc kubenswrapper[4907]: I0313 15:15:03.041399 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1fb6a3c2-e8df-47a4-9ae7-2ee38842051a-secret-volume\") pod \"1fb6a3c2-e8df-47a4-9ae7-2ee38842051a\" (UID: \"1fb6a3c2-e8df-47a4-9ae7-2ee38842051a\") " Mar 13 15:15:03 crc kubenswrapper[4907]: I0313 15:15:03.042468 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1fb6a3c2-e8df-47a4-9ae7-2ee38842051a-config-volume" (OuterVolumeSpecName: "config-volume") pod "1fb6a3c2-e8df-47a4-9ae7-2ee38842051a" (UID: "1fb6a3c2-e8df-47a4-9ae7-2ee38842051a"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:15:03 crc kubenswrapper[4907]: I0313 15:15:03.046715 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1fb6a3c2-e8df-47a4-9ae7-2ee38842051a-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "1fb6a3c2-e8df-47a4-9ae7-2ee38842051a" (UID: "1fb6a3c2-e8df-47a4-9ae7-2ee38842051a"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:15:03 crc kubenswrapper[4907]: I0313 15:15:03.053702 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1fb6a3c2-e8df-47a4-9ae7-2ee38842051a-kube-api-access-fxh86" (OuterVolumeSpecName: "kube-api-access-fxh86") pod "1fb6a3c2-e8df-47a4-9ae7-2ee38842051a" (UID: "1fb6a3c2-e8df-47a4-9ae7-2ee38842051a"). InnerVolumeSpecName "kube-api-access-fxh86". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:15:03 crc kubenswrapper[4907]: I0313 15:15:03.143196 4907 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1fb6a3c2-e8df-47a4-9ae7-2ee38842051a-config-volume\") on node \"crc\" DevicePath \"\"" Mar 13 15:15:03 crc kubenswrapper[4907]: I0313 15:15:03.143255 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fxh86\" (UniqueName: \"kubernetes.io/projected/1fb6a3c2-e8df-47a4-9ae7-2ee38842051a-kube-api-access-fxh86\") on node \"crc\" DevicePath \"\"" Mar 13 15:15:03 crc kubenswrapper[4907]: I0313 15:15:03.143273 4907 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1fb6a3c2-e8df-47a4-9ae7-2ee38842051a-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 13 15:15:03 crc kubenswrapper[4907]: I0313 15:15:03.628056 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556915-jtzgq" Mar 13 15:15:03 crc kubenswrapper[4907]: I0313 15:15:03.628077 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556915-jtzgq" event={"ID":"1fb6a3c2-e8df-47a4-9ae7-2ee38842051a","Type":"ContainerDied","Data":"8763e76268e6fdebf8c7827617d0774f78f5d7e8ce50c07417aa72eaaa7a2d83"} Mar 13 15:15:03 crc kubenswrapper[4907]: I0313 15:15:03.628130 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8763e76268e6fdebf8c7827617d0774f78f5d7e8ce50c07417aa72eaaa7a2d83" Mar 13 15:15:03 crc kubenswrapper[4907]: I0313 15:15:03.967837 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556870-rdhxg"] Mar 13 15:15:03 crc kubenswrapper[4907]: I0313 15:15:03.972849 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556870-rdhxg"] Mar 13 15:15:05 crc kubenswrapper[4907]: I0313 15:15:05.790693 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="af9d29dd-6749-4fcd-82a8-ea528d832fcc" path="/var/lib/kubelet/pods/af9d29dd-6749-4fcd-82a8-ea528d832fcc/volumes" Mar 13 15:15:08 crc kubenswrapper[4907]: I0313 15:15:08.370171 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-jpjgk" Mar 13 15:15:08 crc kubenswrapper[4907]: I0313 15:15:08.370539 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-jpjgk" Mar 13 15:15:08 crc kubenswrapper[4907]: I0313 15:15:08.432807 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-jpjgk" Mar 13 15:15:08 crc kubenswrapper[4907]: I0313 15:15:08.702004 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-jpjgk" Mar 13 15:15:08 crc kubenswrapper[4907]: I0313 15:15:08.747163 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jpjgk"] Mar 13 15:15:10 crc kubenswrapper[4907]: I0313 15:15:10.681513 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-jpjgk" podUID="db74eac8-6eb9-4abc-8bd4-05a7c4de409d" containerName="registry-server" containerID="cri-o://2832afb1ca9f1fcdd05a62540f3537314948bd51bef218b6399c500e2dfcca30" gracePeriod=2 Mar 13 15:15:11 crc kubenswrapper[4907]: I0313 15:15:11.113669 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jpjgk" Mar 13 15:15:11 crc kubenswrapper[4907]: I0313 15:15:11.257397 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db74eac8-6eb9-4abc-8bd4-05a7c4de409d-utilities\") pod \"db74eac8-6eb9-4abc-8bd4-05a7c4de409d\" (UID: \"db74eac8-6eb9-4abc-8bd4-05a7c4de409d\") " Mar 13 15:15:11 crc kubenswrapper[4907]: I0313 15:15:11.257602 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db74eac8-6eb9-4abc-8bd4-05a7c4de409d-catalog-content\") pod \"db74eac8-6eb9-4abc-8bd4-05a7c4de409d\" (UID: \"db74eac8-6eb9-4abc-8bd4-05a7c4de409d\") " Mar 13 15:15:11 crc kubenswrapper[4907]: I0313 15:15:11.257710 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bkfd5\" (UniqueName: \"kubernetes.io/projected/db74eac8-6eb9-4abc-8bd4-05a7c4de409d-kube-api-access-bkfd5\") pod \"db74eac8-6eb9-4abc-8bd4-05a7c4de409d\" (UID: \"db74eac8-6eb9-4abc-8bd4-05a7c4de409d\") " Mar 13 15:15:11 crc kubenswrapper[4907]: I0313 15:15:11.258996 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/db74eac8-6eb9-4abc-8bd4-05a7c4de409d-utilities" (OuterVolumeSpecName: "utilities") pod "db74eac8-6eb9-4abc-8bd4-05a7c4de409d" (UID: "db74eac8-6eb9-4abc-8bd4-05a7c4de409d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:15:11 crc kubenswrapper[4907]: I0313 15:15:11.359532 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db74eac8-6eb9-4abc-8bd4-05a7c4de409d-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 15:15:11 crc kubenswrapper[4907]: I0313 15:15:11.412309 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db74eac8-6eb9-4abc-8bd4-05a7c4de409d-kube-api-access-bkfd5" (OuterVolumeSpecName: "kube-api-access-bkfd5") pod "db74eac8-6eb9-4abc-8bd4-05a7c4de409d" (UID: "db74eac8-6eb9-4abc-8bd4-05a7c4de409d"). InnerVolumeSpecName "kube-api-access-bkfd5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:15:11 crc kubenswrapper[4907]: I0313 15:15:11.460704 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bkfd5\" (UniqueName: \"kubernetes.io/projected/db74eac8-6eb9-4abc-8bd4-05a7c4de409d-kube-api-access-bkfd5\") on node \"crc\" DevicePath \"\"" Mar 13 15:15:11 crc kubenswrapper[4907]: I0313 15:15:11.692843 4907 generic.go:334] "Generic (PLEG): container finished" podID="db74eac8-6eb9-4abc-8bd4-05a7c4de409d" containerID="2832afb1ca9f1fcdd05a62540f3537314948bd51bef218b6399c500e2dfcca30" exitCode=0 Mar 13 15:15:11 crc kubenswrapper[4907]: I0313 15:15:11.692906 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jpjgk" event={"ID":"db74eac8-6eb9-4abc-8bd4-05a7c4de409d","Type":"ContainerDied","Data":"2832afb1ca9f1fcdd05a62540f3537314948bd51bef218b6399c500e2dfcca30"} Mar 13 15:15:11 crc kubenswrapper[4907]: I0313 15:15:11.692957 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jpjgk" event={"ID":"db74eac8-6eb9-4abc-8bd4-05a7c4de409d","Type":"ContainerDied","Data":"1e8f5904cbc6648ae855379c24f68a418662b1fb626449b7dd05f4af7ab877dc"} Mar 13 15:15:11 crc kubenswrapper[4907]: I0313 15:15:11.692983 4907 scope.go:117] "RemoveContainer" containerID="2832afb1ca9f1fcdd05a62540f3537314948bd51bef218b6399c500e2dfcca30" Mar 13 15:15:11 crc kubenswrapper[4907]: I0313 15:15:11.693128 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jpjgk" Mar 13 15:15:11 crc kubenswrapper[4907]: I0313 15:15:11.721917 4907 scope.go:117] "RemoveContainer" containerID="43c417bf05f6d13ba2647567c1378035b8d2b1457c7230a97bd627f1015c7e9b" Mar 13 15:15:11 crc kubenswrapper[4907]: I0313 15:15:11.742676 4907 scope.go:117] "RemoveContainer" containerID="40cf43ff58a1d7d18d1fe371ca4cdf1f9b55f77c9ec021d33f6565bbdaadc497" Mar 13 15:15:11 crc kubenswrapper[4907]: I0313 15:15:11.778174 4907 scope.go:117] "RemoveContainer" containerID="2832afb1ca9f1fcdd05a62540f3537314948bd51bef218b6399c500e2dfcca30" Mar 13 15:15:11 crc kubenswrapper[4907]: E0313 15:15:11.778723 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2832afb1ca9f1fcdd05a62540f3537314948bd51bef218b6399c500e2dfcca30\": container with ID starting with 2832afb1ca9f1fcdd05a62540f3537314948bd51bef218b6399c500e2dfcca30 not found: ID does not exist" containerID="2832afb1ca9f1fcdd05a62540f3537314948bd51bef218b6399c500e2dfcca30" Mar 13 15:15:11 crc kubenswrapper[4907]: I0313 15:15:11.778762 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2832afb1ca9f1fcdd05a62540f3537314948bd51bef218b6399c500e2dfcca30"} err="failed to get container status \"2832afb1ca9f1fcdd05a62540f3537314948bd51bef218b6399c500e2dfcca30\": rpc error: code = NotFound desc = could not find container \"2832afb1ca9f1fcdd05a62540f3537314948bd51bef218b6399c500e2dfcca30\": container with ID starting with 2832afb1ca9f1fcdd05a62540f3537314948bd51bef218b6399c500e2dfcca30 not found: ID does not exist" Mar 13 15:15:11 crc kubenswrapper[4907]: I0313 15:15:11.778789 4907 scope.go:117] "RemoveContainer" containerID="43c417bf05f6d13ba2647567c1378035b8d2b1457c7230a97bd627f1015c7e9b" Mar 13 15:15:11 crc kubenswrapper[4907]: E0313 15:15:11.779224 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"43c417bf05f6d13ba2647567c1378035b8d2b1457c7230a97bd627f1015c7e9b\": container with ID starting with 43c417bf05f6d13ba2647567c1378035b8d2b1457c7230a97bd627f1015c7e9b not found: ID does not exist" containerID="43c417bf05f6d13ba2647567c1378035b8d2b1457c7230a97bd627f1015c7e9b" Mar 13 15:15:11 crc kubenswrapper[4907]: I0313 15:15:11.779277 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"43c417bf05f6d13ba2647567c1378035b8d2b1457c7230a97bd627f1015c7e9b"} err="failed to get container status \"43c417bf05f6d13ba2647567c1378035b8d2b1457c7230a97bd627f1015c7e9b\": rpc error: code = NotFound desc = could not find container \"43c417bf05f6d13ba2647567c1378035b8d2b1457c7230a97bd627f1015c7e9b\": container with ID starting with 43c417bf05f6d13ba2647567c1378035b8d2b1457c7230a97bd627f1015c7e9b not found: ID does not exist" Mar 13 15:15:11 crc kubenswrapper[4907]: I0313 15:15:11.779310 4907 scope.go:117] "RemoveContainer" containerID="40cf43ff58a1d7d18d1fe371ca4cdf1f9b55f77c9ec021d33f6565bbdaadc497" Mar 13 15:15:11 crc kubenswrapper[4907]: E0313 15:15:11.779669 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"40cf43ff58a1d7d18d1fe371ca4cdf1f9b55f77c9ec021d33f6565bbdaadc497\": container with ID starting with 40cf43ff58a1d7d18d1fe371ca4cdf1f9b55f77c9ec021d33f6565bbdaadc497 not found: ID does not exist" containerID="40cf43ff58a1d7d18d1fe371ca4cdf1f9b55f77c9ec021d33f6565bbdaadc497" Mar 13 15:15:11 crc kubenswrapper[4907]: I0313 15:15:11.779709 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"40cf43ff58a1d7d18d1fe371ca4cdf1f9b55f77c9ec021d33f6565bbdaadc497"} err="failed to get container status \"40cf43ff58a1d7d18d1fe371ca4cdf1f9b55f77c9ec021d33f6565bbdaadc497\": rpc error: code = NotFound desc = could not find container \"40cf43ff58a1d7d18d1fe371ca4cdf1f9b55f77c9ec021d33f6565bbdaadc497\": container with ID starting with 40cf43ff58a1d7d18d1fe371ca4cdf1f9b55f77c9ec021d33f6565bbdaadc497 not found: ID does not exist" Mar 13 15:15:11 crc kubenswrapper[4907]: I0313 15:15:11.801640 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/db74eac8-6eb9-4abc-8bd4-05a7c4de409d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "db74eac8-6eb9-4abc-8bd4-05a7c4de409d" (UID: "db74eac8-6eb9-4abc-8bd4-05a7c4de409d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:15:11 crc kubenswrapper[4907]: I0313 15:15:11.866704 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db74eac8-6eb9-4abc-8bd4-05a7c4de409d-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 15:15:12 crc kubenswrapper[4907]: I0313 15:15:12.030580 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jpjgk"] Mar 13 15:15:12 crc kubenswrapper[4907]: I0313 15:15:12.037742 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-jpjgk"] Mar 13 15:15:13 crc kubenswrapper[4907]: I0313 15:15:13.795183 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="db74eac8-6eb9-4abc-8bd4-05a7c4de409d" path="/var/lib/kubelet/pods/db74eac8-6eb9-4abc-8bd4-05a7c4de409d/volumes" Mar 13 15:15:18 crc kubenswrapper[4907]: I0313 15:15:18.041731 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 15:15:18 crc kubenswrapper[4907]: I0313 15:15:18.042123 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 15:15:18 crc kubenswrapper[4907]: I0313 15:15:18.042184 4907 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" Mar 13 15:15:18 crc kubenswrapper[4907]: I0313 15:15:18.043153 4907 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"36ef7b96f3457426d961c14798528949d4c51116631e09ea3be0c17a7c07b91d"} pod="openshift-machine-config-operator/machine-config-daemon-hm56j" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 15:15:18 crc kubenswrapper[4907]: I0313 15:15:18.043278 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" containerID="cri-o://36ef7b96f3457426d961c14798528949d4c51116631e09ea3be0c17a7c07b91d" gracePeriod=600 Mar 13 15:15:18 crc kubenswrapper[4907]: I0313 15:15:18.740675 4907 generic.go:334] "Generic (PLEG): container finished" podID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerID="36ef7b96f3457426d961c14798528949d4c51116631e09ea3be0c17a7c07b91d" exitCode=0 Mar 13 15:15:18 crc kubenswrapper[4907]: I0313 15:15:18.740769 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerDied","Data":"36ef7b96f3457426d961c14798528949d4c51116631e09ea3be0c17a7c07b91d"} Mar 13 15:15:18 crc kubenswrapper[4907]: I0313 15:15:18.741437 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerStarted","Data":"dd934d7c5fc7e95362f79c2301ff40920ce7e3112722cbc11a053a56c085e05c"} Mar 13 15:15:18 crc kubenswrapper[4907]: I0313 15:15:18.741485 4907 scope.go:117] "RemoveContainer" containerID="caeb29efd9b0ec75cd14d4d1c0fbb54dce8073ba7639ceb605cc152e79561f8a" Mar 13 15:15:26 crc kubenswrapper[4907]: I0313 15:15:26.633480 4907 scope.go:117] "RemoveContainer" containerID="46dfd46e5c29fa7764c024392038bd85a0e04092c9a231bbd08b67969852ba5c" Mar 13 15:16:00 crc kubenswrapper[4907]: I0313 15:16:00.145626 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556916-gpp98"] Mar 13 15:16:00 crc kubenswrapper[4907]: E0313 15:16:00.146767 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db74eac8-6eb9-4abc-8bd4-05a7c4de409d" containerName="extract-content" Mar 13 15:16:00 crc kubenswrapper[4907]: I0313 15:16:00.146786 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="db74eac8-6eb9-4abc-8bd4-05a7c4de409d" containerName="extract-content" Mar 13 15:16:00 crc kubenswrapper[4907]: E0313 15:16:00.146814 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db74eac8-6eb9-4abc-8bd4-05a7c4de409d" containerName="registry-server" Mar 13 15:16:00 crc kubenswrapper[4907]: I0313 15:16:00.146827 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="db74eac8-6eb9-4abc-8bd4-05a7c4de409d" containerName="registry-server" Mar 13 15:16:00 crc kubenswrapper[4907]: E0313 15:16:00.146856 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db74eac8-6eb9-4abc-8bd4-05a7c4de409d" containerName="extract-utilities" Mar 13 15:16:00 crc kubenswrapper[4907]: I0313 15:16:00.146870 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="db74eac8-6eb9-4abc-8bd4-05a7c4de409d" containerName="extract-utilities" Mar 13 15:16:00 crc kubenswrapper[4907]: E0313 15:16:00.147010 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1fb6a3c2-e8df-47a4-9ae7-2ee38842051a" containerName="collect-profiles" Mar 13 15:16:00 crc kubenswrapper[4907]: I0313 15:16:00.147024 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="1fb6a3c2-e8df-47a4-9ae7-2ee38842051a" containerName="collect-profiles" Mar 13 15:16:00 crc kubenswrapper[4907]: I0313 15:16:00.147253 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="db74eac8-6eb9-4abc-8bd4-05a7c4de409d" containerName="registry-server" Mar 13 15:16:00 crc kubenswrapper[4907]: I0313 15:16:00.147281 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="1fb6a3c2-e8df-47a4-9ae7-2ee38842051a" containerName="collect-profiles" Mar 13 15:16:00 crc kubenswrapper[4907]: I0313 15:16:00.147987 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556916-gpp98" Mar 13 15:16:00 crc kubenswrapper[4907]: I0313 15:16:00.153343 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 15:16:00 crc kubenswrapper[4907]: I0313 15:16:00.154008 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 15:16:00 crc kubenswrapper[4907]: I0313 15:16:00.154211 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 15:16:00 crc kubenswrapper[4907]: I0313 15:16:00.164417 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556916-gpp98"] Mar 13 15:16:00 crc kubenswrapper[4907]: I0313 15:16:00.281779 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sxmtz\" (UniqueName: \"kubernetes.io/projected/c5c630da-de73-4f29-89f0-43c5497dd89a-kube-api-access-sxmtz\") pod \"auto-csr-approver-29556916-gpp98\" (UID: \"c5c630da-de73-4f29-89f0-43c5497dd89a\") " pod="openshift-infra/auto-csr-approver-29556916-gpp98" Mar 13 15:16:00 crc kubenswrapper[4907]: I0313 15:16:00.383771 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sxmtz\" (UniqueName: \"kubernetes.io/projected/c5c630da-de73-4f29-89f0-43c5497dd89a-kube-api-access-sxmtz\") pod \"auto-csr-approver-29556916-gpp98\" (UID: \"c5c630da-de73-4f29-89f0-43c5497dd89a\") " pod="openshift-infra/auto-csr-approver-29556916-gpp98" Mar 13 15:16:00 crc kubenswrapper[4907]: I0313 15:16:00.417159 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sxmtz\" (UniqueName: \"kubernetes.io/projected/c5c630da-de73-4f29-89f0-43c5497dd89a-kube-api-access-sxmtz\") pod \"auto-csr-approver-29556916-gpp98\" (UID: \"c5c630da-de73-4f29-89f0-43c5497dd89a\") " pod="openshift-infra/auto-csr-approver-29556916-gpp98" Mar 13 15:16:00 crc kubenswrapper[4907]: I0313 15:16:00.478120 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556916-gpp98" Mar 13 15:16:00 crc kubenswrapper[4907]: I0313 15:16:00.907920 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556916-gpp98"] Mar 13 15:16:01 crc kubenswrapper[4907]: I0313 15:16:01.046708 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556916-gpp98" event={"ID":"c5c630da-de73-4f29-89f0-43c5497dd89a","Type":"ContainerStarted","Data":"93ed1baee1884167038c99f469dc1ebb5a70d1aba4262c7b4e9fae45a3e3a33c"} Mar 13 15:16:03 crc kubenswrapper[4907]: I0313 15:16:03.062105 4907 generic.go:334] "Generic (PLEG): container finished" podID="c5c630da-de73-4f29-89f0-43c5497dd89a" containerID="348c87a146da948df0e4139184e97c7692cc71aad3ec14fe9f015dc10660b3b8" exitCode=0 Mar 13 15:16:03 crc kubenswrapper[4907]: I0313 15:16:03.062209 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556916-gpp98" event={"ID":"c5c630da-de73-4f29-89f0-43c5497dd89a","Type":"ContainerDied","Data":"348c87a146da948df0e4139184e97c7692cc71aad3ec14fe9f015dc10660b3b8"} Mar 13 15:16:04 crc kubenswrapper[4907]: I0313 15:16:04.358190 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556916-gpp98" Mar 13 15:16:04 crc kubenswrapper[4907]: I0313 15:16:04.548403 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sxmtz\" (UniqueName: \"kubernetes.io/projected/c5c630da-de73-4f29-89f0-43c5497dd89a-kube-api-access-sxmtz\") pod \"c5c630da-de73-4f29-89f0-43c5497dd89a\" (UID: \"c5c630da-de73-4f29-89f0-43c5497dd89a\") " Mar 13 15:16:04 crc kubenswrapper[4907]: I0313 15:16:04.555706 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c5c630da-de73-4f29-89f0-43c5497dd89a-kube-api-access-sxmtz" (OuterVolumeSpecName: "kube-api-access-sxmtz") pod "c5c630da-de73-4f29-89f0-43c5497dd89a" (UID: "c5c630da-de73-4f29-89f0-43c5497dd89a"). InnerVolumeSpecName "kube-api-access-sxmtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:16:04 crc kubenswrapper[4907]: I0313 15:16:04.649761 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sxmtz\" (UniqueName: \"kubernetes.io/projected/c5c630da-de73-4f29-89f0-43c5497dd89a-kube-api-access-sxmtz\") on node \"crc\" DevicePath \"\"" Mar 13 15:16:05 crc kubenswrapper[4907]: I0313 15:16:05.078625 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556916-gpp98" event={"ID":"c5c630da-de73-4f29-89f0-43c5497dd89a","Type":"ContainerDied","Data":"93ed1baee1884167038c99f469dc1ebb5a70d1aba4262c7b4e9fae45a3e3a33c"} Mar 13 15:16:05 crc kubenswrapper[4907]: I0313 15:16:05.078674 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="93ed1baee1884167038c99f469dc1ebb5a70d1aba4262c7b4e9fae45a3e3a33c" Mar 13 15:16:05 crc kubenswrapper[4907]: I0313 15:16:05.078733 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556916-gpp98" Mar 13 15:16:05 crc kubenswrapper[4907]: I0313 15:16:05.425082 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556910-5pmwx"] Mar 13 15:16:05 crc kubenswrapper[4907]: I0313 15:16:05.433631 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556910-5pmwx"] Mar 13 15:16:05 crc kubenswrapper[4907]: I0313 15:16:05.791223 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e74e8820-9ca4-4010-9070-ef32010a85c8" path="/var/lib/kubelet/pods/e74e8820-9ca4-4010-9070-ef32010a85c8/volumes" Mar 13 15:16:26 crc kubenswrapper[4907]: I0313 15:16:26.704012 4907 scope.go:117] "RemoveContainer" containerID="6156d45b4121389ab4793e9ba61feb497d3ccb0338571d8b44490cfdec3820ce" Mar 13 15:17:18 crc kubenswrapper[4907]: I0313 15:17:18.042201 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 15:17:18 crc kubenswrapper[4907]: I0313 15:17:18.042988 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 15:17:32 crc kubenswrapper[4907]: I0313 15:17:32.675569 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-k675m"] Mar 13 15:17:32 crc kubenswrapper[4907]: E0313 15:17:32.676593 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5c630da-de73-4f29-89f0-43c5497dd89a" containerName="oc" Mar 13 15:17:32 crc kubenswrapper[4907]: I0313 15:17:32.676611 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5c630da-de73-4f29-89f0-43c5497dd89a" containerName="oc" Mar 13 15:17:32 crc kubenswrapper[4907]: I0313 15:17:32.676841 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="c5c630da-de73-4f29-89f0-43c5497dd89a" containerName="oc" Mar 13 15:17:32 crc kubenswrapper[4907]: I0313 15:17:32.678163 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-k675m" Mar 13 15:17:32 crc kubenswrapper[4907]: I0313 15:17:32.683471 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-k675m"] Mar 13 15:17:32 crc kubenswrapper[4907]: I0313 15:17:32.859556 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ng4nl\" (UniqueName: \"kubernetes.io/projected/7a6ef1dc-77f0-4a25-9465-f9b881023f83-kube-api-access-ng4nl\") pod \"certified-operators-k675m\" (UID: \"7a6ef1dc-77f0-4a25-9465-f9b881023f83\") " pod="openshift-marketplace/certified-operators-k675m" Mar 13 15:17:32 crc kubenswrapper[4907]: I0313 15:17:32.859613 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a6ef1dc-77f0-4a25-9465-f9b881023f83-catalog-content\") pod \"certified-operators-k675m\" (UID: \"7a6ef1dc-77f0-4a25-9465-f9b881023f83\") " pod="openshift-marketplace/certified-operators-k675m" Mar 13 15:17:32 crc kubenswrapper[4907]: I0313 15:17:32.859638 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a6ef1dc-77f0-4a25-9465-f9b881023f83-utilities\") pod \"certified-operators-k675m\" (UID: \"7a6ef1dc-77f0-4a25-9465-f9b881023f83\") " pod="openshift-marketplace/certified-operators-k675m" Mar 13 15:17:32 crc kubenswrapper[4907]: I0313 15:17:32.960776 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a6ef1dc-77f0-4a25-9465-f9b881023f83-catalog-content\") pod \"certified-operators-k675m\" (UID: \"7a6ef1dc-77f0-4a25-9465-f9b881023f83\") " pod="openshift-marketplace/certified-operators-k675m" Mar 13 15:17:32 crc kubenswrapper[4907]: I0313 15:17:32.960836 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a6ef1dc-77f0-4a25-9465-f9b881023f83-utilities\") pod \"certified-operators-k675m\" (UID: \"7a6ef1dc-77f0-4a25-9465-f9b881023f83\") " pod="openshift-marketplace/certified-operators-k675m" Mar 13 15:17:32 crc kubenswrapper[4907]: I0313 15:17:32.960971 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ng4nl\" (UniqueName: \"kubernetes.io/projected/7a6ef1dc-77f0-4a25-9465-f9b881023f83-kube-api-access-ng4nl\") pod \"certified-operators-k675m\" (UID: \"7a6ef1dc-77f0-4a25-9465-f9b881023f83\") " pod="openshift-marketplace/certified-operators-k675m" Mar 13 15:17:32 crc kubenswrapper[4907]: I0313 15:17:32.961324 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a6ef1dc-77f0-4a25-9465-f9b881023f83-catalog-content\") pod \"certified-operators-k675m\" (UID: \"7a6ef1dc-77f0-4a25-9465-f9b881023f83\") " pod="openshift-marketplace/certified-operators-k675m" Mar 13 15:17:32 crc kubenswrapper[4907]: I0313 15:17:32.961410 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a6ef1dc-77f0-4a25-9465-f9b881023f83-utilities\") pod \"certified-operators-k675m\" (UID: \"7a6ef1dc-77f0-4a25-9465-f9b881023f83\") " pod="openshift-marketplace/certified-operators-k675m" Mar 13 15:17:32 crc kubenswrapper[4907]: I0313 15:17:32.992446 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ng4nl\" (UniqueName: \"kubernetes.io/projected/7a6ef1dc-77f0-4a25-9465-f9b881023f83-kube-api-access-ng4nl\") pod \"certified-operators-k675m\" (UID: \"7a6ef1dc-77f0-4a25-9465-f9b881023f83\") " pod="openshift-marketplace/certified-operators-k675m" Mar 13 15:17:33 crc kubenswrapper[4907]: I0313 15:17:33.002324 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-k675m" Mar 13 15:17:33 crc kubenswrapper[4907]: I0313 15:17:33.526603 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-k675m"] Mar 13 15:17:33 crc kubenswrapper[4907]: I0313 15:17:33.741560 4907 generic.go:334] "Generic (PLEG): container finished" podID="7a6ef1dc-77f0-4a25-9465-f9b881023f83" containerID="f64382383e10d8bc77380b8411efc74df96b9cacb9ce301d06c4749e0ccc97b6" exitCode=0 Mar 13 15:17:33 crc kubenswrapper[4907]: I0313 15:17:33.741656 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k675m" event={"ID":"7a6ef1dc-77f0-4a25-9465-f9b881023f83","Type":"ContainerDied","Data":"f64382383e10d8bc77380b8411efc74df96b9cacb9ce301d06c4749e0ccc97b6"} Mar 13 15:17:33 crc kubenswrapper[4907]: I0313 15:17:33.741832 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k675m" event={"ID":"7a6ef1dc-77f0-4a25-9465-f9b881023f83","Type":"ContainerStarted","Data":"4e3fd1a5082ad9dee2e764ae30a95c9a0adcb1f77f2fd1c6a9389907f1dd1779"} Mar 13 15:17:34 crc kubenswrapper[4907]: I0313 15:17:34.749255 4907 generic.go:334] "Generic (PLEG): container finished" podID="7a6ef1dc-77f0-4a25-9465-f9b881023f83" containerID="2cbc3b0e8cb89dede628cdba2545a3b54133f559dd5f2d0fcaf7f50b62b92188" exitCode=0 Mar 13 15:17:34 crc kubenswrapper[4907]: I0313 15:17:34.749422 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k675m" event={"ID":"7a6ef1dc-77f0-4a25-9465-f9b881023f83","Type":"ContainerDied","Data":"2cbc3b0e8cb89dede628cdba2545a3b54133f559dd5f2d0fcaf7f50b62b92188"} Mar 13 15:17:35 crc kubenswrapper[4907]: I0313 15:17:35.757284 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k675m" event={"ID":"7a6ef1dc-77f0-4a25-9465-f9b881023f83","Type":"ContainerStarted","Data":"de1a6aa0e4c46c2959d6be7168e2b145807f3a5c7a00ecfda7923f20d6b96d58"} Mar 13 15:17:35 crc kubenswrapper[4907]: I0313 15:17:35.776028 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-k675m" podStartSLOduration=2.372191954 podStartE2EDuration="3.776011351s" podCreationTimestamp="2026-03-13 15:17:32 +0000 UTC" firstStartedPulling="2026-03-13 15:17:33.74337868 +0000 UTC m=+4352.643166369" lastFinishedPulling="2026-03-13 15:17:35.147198087 +0000 UTC m=+4354.046985766" observedRunningTime="2026-03-13 15:17:35.774464169 +0000 UTC m=+4354.674251878" watchObservedRunningTime="2026-03-13 15:17:35.776011351 +0000 UTC m=+4354.675799040" Mar 13 15:17:35 crc kubenswrapper[4907]: I0313 15:17:35.867149 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-xhrm7"] Mar 13 15:17:35 crc kubenswrapper[4907]: I0313 15:17:35.868910 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xhrm7" Mar 13 15:17:35 crc kubenswrapper[4907]: I0313 15:17:35.877033 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xhrm7"] Mar 13 15:17:36 crc kubenswrapper[4907]: I0313 15:17:36.000286 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m4gbx\" (UniqueName: \"kubernetes.io/projected/78255f5a-d6ed-45c3-ade0-4ff6f01c7a0f-kube-api-access-m4gbx\") pod \"redhat-operators-xhrm7\" (UID: \"78255f5a-d6ed-45c3-ade0-4ff6f01c7a0f\") " pod="openshift-marketplace/redhat-operators-xhrm7" Mar 13 15:17:36 crc kubenswrapper[4907]: I0313 15:17:36.000344 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/78255f5a-d6ed-45c3-ade0-4ff6f01c7a0f-catalog-content\") pod \"redhat-operators-xhrm7\" (UID: \"78255f5a-d6ed-45c3-ade0-4ff6f01c7a0f\") " pod="openshift-marketplace/redhat-operators-xhrm7" Mar 13 15:17:36 crc kubenswrapper[4907]: I0313 15:17:36.000377 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/78255f5a-d6ed-45c3-ade0-4ff6f01c7a0f-utilities\") pod \"redhat-operators-xhrm7\" (UID: \"78255f5a-d6ed-45c3-ade0-4ff6f01c7a0f\") " pod="openshift-marketplace/redhat-operators-xhrm7" Mar 13 15:17:36 crc kubenswrapper[4907]: I0313 15:17:36.102934 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m4gbx\" (UniqueName: \"kubernetes.io/projected/78255f5a-d6ed-45c3-ade0-4ff6f01c7a0f-kube-api-access-m4gbx\") pod \"redhat-operators-xhrm7\" (UID: \"78255f5a-d6ed-45c3-ade0-4ff6f01c7a0f\") " pod="openshift-marketplace/redhat-operators-xhrm7" Mar 13 15:17:36 crc kubenswrapper[4907]: I0313 15:17:36.103217 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/78255f5a-d6ed-45c3-ade0-4ff6f01c7a0f-catalog-content\") pod \"redhat-operators-xhrm7\" (UID: \"78255f5a-d6ed-45c3-ade0-4ff6f01c7a0f\") " pod="openshift-marketplace/redhat-operators-xhrm7" Mar 13 15:17:36 crc kubenswrapper[4907]: I0313 15:17:36.103338 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/78255f5a-d6ed-45c3-ade0-4ff6f01c7a0f-utilities\") pod \"redhat-operators-xhrm7\" (UID: \"78255f5a-d6ed-45c3-ade0-4ff6f01c7a0f\") " pod="openshift-marketplace/redhat-operators-xhrm7" Mar 13 15:17:36 crc kubenswrapper[4907]: I0313 15:17:36.103804 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/78255f5a-d6ed-45c3-ade0-4ff6f01c7a0f-catalog-content\") pod \"redhat-operators-xhrm7\" (UID: \"78255f5a-d6ed-45c3-ade0-4ff6f01c7a0f\") " pod="openshift-marketplace/redhat-operators-xhrm7" Mar 13 15:17:36 crc kubenswrapper[4907]: I0313 15:17:36.103910 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/78255f5a-d6ed-45c3-ade0-4ff6f01c7a0f-utilities\") pod \"redhat-operators-xhrm7\" (UID: \"78255f5a-d6ed-45c3-ade0-4ff6f01c7a0f\") " pod="openshift-marketplace/redhat-operators-xhrm7" Mar 13 15:17:36 crc kubenswrapper[4907]: I0313 15:17:36.121366 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m4gbx\" (UniqueName: \"kubernetes.io/projected/78255f5a-d6ed-45c3-ade0-4ff6f01c7a0f-kube-api-access-m4gbx\") pod \"redhat-operators-xhrm7\" (UID: \"78255f5a-d6ed-45c3-ade0-4ff6f01c7a0f\") " pod="openshift-marketplace/redhat-operators-xhrm7" Mar 13 15:17:36 crc kubenswrapper[4907]: I0313 15:17:36.191336 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xhrm7" Mar 13 15:17:36 crc kubenswrapper[4907]: I0313 15:17:36.647592 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xhrm7"] Mar 13 15:17:36 crc kubenswrapper[4907]: I0313 15:17:36.764382 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xhrm7" event={"ID":"78255f5a-d6ed-45c3-ade0-4ff6f01c7a0f","Type":"ContainerStarted","Data":"e710e5af60dad227fb28f3f3f8736be015c34c3cbf424e8cb7767ac5bea1fdd9"} Mar 13 15:17:37 crc kubenswrapper[4907]: I0313 15:17:37.774118 4907 generic.go:334] "Generic (PLEG): container finished" podID="78255f5a-d6ed-45c3-ade0-4ff6f01c7a0f" containerID="cdd1adc8cc682b604b6e997800c49a6fbb65e62d6431a882b09277e919d44d4d" exitCode=0 Mar 13 15:17:37 crc kubenswrapper[4907]: I0313 15:17:37.774166 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xhrm7" event={"ID":"78255f5a-d6ed-45c3-ade0-4ff6f01c7a0f","Type":"ContainerDied","Data":"cdd1adc8cc682b604b6e997800c49a6fbb65e62d6431a882b09277e919d44d4d"} Mar 13 15:17:39 crc kubenswrapper[4907]: I0313 15:17:39.792068 4907 generic.go:334] "Generic (PLEG): container finished" podID="78255f5a-d6ed-45c3-ade0-4ff6f01c7a0f" containerID="21f6ddb10036b57ffc641c1c43e81015c667a267b0b5f356fdd1eeded7b6a958" exitCode=0 Mar 13 15:17:39 crc kubenswrapper[4907]: I0313 15:17:39.792131 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xhrm7" event={"ID":"78255f5a-d6ed-45c3-ade0-4ff6f01c7a0f","Type":"ContainerDied","Data":"21f6ddb10036b57ffc641c1c43e81015c667a267b0b5f356fdd1eeded7b6a958"} Mar 13 15:17:40 crc kubenswrapper[4907]: I0313 15:17:40.800497 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xhrm7" event={"ID":"78255f5a-d6ed-45c3-ade0-4ff6f01c7a0f","Type":"ContainerStarted","Data":"91ebf771b78bb7a2fe3a8359fd7a3263f5fdd1c2032beba161125b81cdf81d3c"} Mar 13 15:17:40 crc kubenswrapper[4907]: I0313 15:17:40.818841 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-xhrm7" podStartSLOduration=3.399974499 podStartE2EDuration="5.818818746s" podCreationTimestamp="2026-03-13 15:17:35 +0000 UTC" firstStartedPulling="2026-03-13 15:17:37.775720649 +0000 UTC m=+4356.675508338" lastFinishedPulling="2026-03-13 15:17:40.194564896 +0000 UTC m=+4359.094352585" observedRunningTime="2026-03-13 15:17:40.815851685 +0000 UTC m=+4359.715639374" watchObservedRunningTime="2026-03-13 15:17:40.818818746 +0000 UTC m=+4359.718606435" Mar 13 15:17:43 crc kubenswrapper[4907]: I0313 15:17:43.003703 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-k675m" Mar 13 15:17:43 crc kubenswrapper[4907]: I0313 15:17:43.004235 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-k675m" Mar 13 15:17:43 crc kubenswrapper[4907]: I0313 15:17:43.070688 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-k675m" Mar 13 15:17:43 crc kubenswrapper[4907]: I0313 15:17:43.864296 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-k675m" Mar 13 15:17:46 crc kubenswrapper[4907]: I0313 15:17:46.390324 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-xhrm7" Mar 13 15:17:46 crc kubenswrapper[4907]: I0313 15:17:46.391115 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-xhrm7" Mar 13 15:17:47 crc kubenswrapper[4907]: I0313 15:17:47.432516 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-xhrm7" podUID="78255f5a-d6ed-45c3-ade0-4ff6f01c7a0f" containerName="registry-server" probeResult="failure" output=< Mar 13 15:17:47 crc kubenswrapper[4907]: timeout: failed to connect service ":50051" within 1s Mar 13 15:17:47 crc kubenswrapper[4907]: > Mar 13 15:17:47 crc kubenswrapper[4907]: I0313 15:17:47.859561 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-k675m"] Mar 13 15:17:47 crc kubenswrapper[4907]: I0313 15:17:47.859857 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-k675m" podUID="7a6ef1dc-77f0-4a25-9465-f9b881023f83" containerName="registry-server" containerID="cri-o://de1a6aa0e4c46c2959d6be7168e2b145807f3a5c7a00ecfda7923f20d6b96d58" gracePeriod=2 Mar 13 15:17:48 crc kubenswrapper[4907]: I0313 15:17:48.042282 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 15:17:48 crc kubenswrapper[4907]: I0313 15:17:48.042739 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 15:17:48 crc kubenswrapper[4907]: I0313 15:17:48.226330 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-k675m" Mar 13 15:17:48 crc kubenswrapper[4907]: I0313 15:17:48.366602 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ng4nl\" (UniqueName: \"kubernetes.io/projected/7a6ef1dc-77f0-4a25-9465-f9b881023f83-kube-api-access-ng4nl\") pod \"7a6ef1dc-77f0-4a25-9465-f9b881023f83\" (UID: \"7a6ef1dc-77f0-4a25-9465-f9b881023f83\") " Mar 13 15:17:48 crc kubenswrapper[4907]: I0313 15:17:48.366702 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a6ef1dc-77f0-4a25-9465-f9b881023f83-catalog-content\") pod \"7a6ef1dc-77f0-4a25-9465-f9b881023f83\" (UID: \"7a6ef1dc-77f0-4a25-9465-f9b881023f83\") " Mar 13 15:17:48 crc kubenswrapper[4907]: I0313 15:17:48.366728 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a6ef1dc-77f0-4a25-9465-f9b881023f83-utilities\") pod \"7a6ef1dc-77f0-4a25-9465-f9b881023f83\" (UID: \"7a6ef1dc-77f0-4a25-9465-f9b881023f83\") " Mar 13 15:17:48 crc kubenswrapper[4907]: I0313 15:17:48.367731 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7a6ef1dc-77f0-4a25-9465-f9b881023f83-utilities" (OuterVolumeSpecName: "utilities") pod "7a6ef1dc-77f0-4a25-9465-f9b881023f83" (UID: "7a6ef1dc-77f0-4a25-9465-f9b881023f83"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:17:48 crc kubenswrapper[4907]: I0313 15:17:48.368073 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a6ef1dc-77f0-4a25-9465-f9b881023f83-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 15:17:48 crc kubenswrapper[4907]: I0313 15:17:48.373988 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7a6ef1dc-77f0-4a25-9465-f9b881023f83-kube-api-access-ng4nl" (OuterVolumeSpecName: "kube-api-access-ng4nl") pod "7a6ef1dc-77f0-4a25-9465-f9b881023f83" (UID: "7a6ef1dc-77f0-4a25-9465-f9b881023f83"). InnerVolumeSpecName "kube-api-access-ng4nl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:17:48 crc kubenswrapper[4907]: I0313 15:17:48.428510 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7a6ef1dc-77f0-4a25-9465-f9b881023f83-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7a6ef1dc-77f0-4a25-9465-f9b881023f83" (UID: "7a6ef1dc-77f0-4a25-9465-f9b881023f83"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:17:48 crc kubenswrapper[4907]: I0313 15:17:48.469225 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a6ef1dc-77f0-4a25-9465-f9b881023f83-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 15:17:48 crc kubenswrapper[4907]: I0313 15:17:48.469252 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ng4nl\" (UniqueName: \"kubernetes.io/projected/7a6ef1dc-77f0-4a25-9465-f9b881023f83-kube-api-access-ng4nl\") on node \"crc\" DevicePath \"\"" Mar 13 15:17:48 crc kubenswrapper[4907]: I0313 15:17:48.858700 4907 generic.go:334] "Generic (PLEG): container finished" podID="7a6ef1dc-77f0-4a25-9465-f9b881023f83" containerID="de1a6aa0e4c46c2959d6be7168e2b145807f3a5c7a00ecfda7923f20d6b96d58" exitCode=0 Mar 13 15:17:48 crc kubenswrapper[4907]: I0313 15:17:48.858780 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-k675m" Mar 13 15:17:48 crc kubenswrapper[4907]: I0313 15:17:48.858789 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k675m" event={"ID":"7a6ef1dc-77f0-4a25-9465-f9b881023f83","Type":"ContainerDied","Data":"de1a6aa0e4c46c2959d6be7168e2b145807f3a5c7a00ecfda7923f20d6b96d58"} Mar 13 15:17:48 crc kubenswrapper[4907]: I0313 15:17:48.859459 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k675m" event={"ID":"7a6ef1dc-77f0-4a25-9465-f9b881023f83","Type":"ContainerDied","Data":"4e3fd1a5082ad9dee2e764ae30a95c9a0adcb1f77f2fd1c6a9389907f1dd1779"} Mar 13 15:17:48 crc kubenswrapper[4907]: I0313 15:17:48.859517 4907 scope.go:117] "RemoveContainer" containerID="de1a6aa0e4c46c2959d6be7168e2b145807f3a5c7a00ecfda7923f20d6b96d58" Mar 13 15:17:48 crc kubenswrapper[4907]: I0313 15:17:48.899762 4907 scope.go:117] "RemoveContainer" containerID="2cbc3b0e8cb89dede628cdba2545a3b54133f559dd5f2d0fcaf7f50b62b92188" Mar 13 15:17:48 crc kubenswrapper[4907]: I0313 15:17:48.903143 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-k675m"] Mar 13 15:17:48 crc kubenswrapper[4907]: I0313 15:17:48.924257 4907 scope.go:117] "RemoveContainer" containerID="f64382383e10d8bc77380b8411efc74df96b9cacb9ce301d06c4749e0ccc97b6" Mar 13 15:17:48 crc kubenswrapper[4907]: I0313 15:17:48.928928 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-k675m"] Mar 13 15:17:48 crc kubenswrapper[4907]: I0313 15:17:48.948807 4907 scope.go:117] "RemoveContainer" containerID="de1a6aa0e4c46c2959d6be7168e2b145807f3a5c7a00ecfda7923f20d6b96d58" Mar 13 15:17:48 crc kubenswrapper[4907]: E0313 15:17:48.949324 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"de1a6aa0e4c46c2959d6be7168e2b145807f3a5c7a00ecfda7923f20d6b96d58\": container with ID starting with de1a6aa0e4c46c2959d6be7168e2b145807f3a5c7a00ecfda7923f20d6b96d58 not found: ID does not exist" containerID="de1a6aa0e4c46c2959d6be7168e2b145807f3a5c7a00ecfda7923f20d6b96d58" Mar 13 15:17:48 crc kubenswrapper[4907]: I0313 15:17:48.949395 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de1a6aa0e4c46c2959d6be7168e2b145807f3a5c7a00ecfda7923f20d6b96d58"} err="failed to get container status \"de1a6aa0e4c46c2959d6be7168e2b145807f3a5c7a00ecfda7923f20d6b96d58\": rpc error: code = NotFound desc = could not find container \"de1a6aa0e4c46c2959d6be7168e2b145807f3a5c7a00ecfda7923f20d6b96d58\": container with ID starting with de1a6aa0e4c46c2959d6be7168e2b145807f3a5c7a00ecfda7923f20d6b96d58 not found: ID does not exist" Mar 13 15:17:48 crc kubenswrapper[4907]: I0313 15:17:48.949434 4907 scope.go:117] "RemoveContainer" containerID="2cbc3b0e8cb89dede628cdba2545a3b54133f559dd5f2d0fcaf7f50b62b92188" Mar 13 15:17:48 crc kubenswrapper[4907]: E0313 15:17:48.949943 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2cbc3b0e8cb89dede628cdba2545a3b54133f559dd5f2d0fcaf7f50b62b92188\": container with ID starting with 2cbc3b0e8cb89dede628cdba2545a3b54133f559dd5f2d0fcaf7f50b62b92188 not found: ID does not exist" containerID="2cbc3b0e8cb89dede628cdba2545a3b54133f559dd5f2d0fcaf7f50b62b92188" Mar 13 15:17:48 crc kubenswrapper[4907]: I0313 15:17:48.950012 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2cbc3b0e8cb89dede628cdba2545a3b54133f559dd5f2d0fcaf7f50b62b92188"} err="failed to get container status \"2cbc3b0e8cb89dede628cdba2545a3b54133f559dd5f2d0fcaf7f50b62b92188\": rpc error: code = NotFound desc = could not find container \"2cbc3b0e8cb89dede628cdba2545a3b54133f559dd5f2d0fcaf7f50b62b92188\": container with ID starting with 2cbc3b0e8cb89dede628cdba2545a3b54133f559dd5f2d0fcaf7f50b62b92188 not found: ID does not exist" Mar 13 15:17:48 crc kubenswrapper[4907]: I0313 15:17:48.950047 4907 scope.go:117] "RemoveContainer" containerID="f64382383e10d8bc77380b8411efc74df96b9cacb9ce301d06c4749e0ccc97b6" Mar 13 15:17:48 crc kubenswrapper[4907]: E0313 15:17:48.950359 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f64382383e10d8bc77380b8411efc74df96b9cacb9ce301d06c4749e0ccc97b6\": container with ID starting with f64382383e10d8bc77380b8411efc74df96b9cacb9ce301d06c4749e0ccc97b6 not found: ID does not exist" containerID="f64382383e10d8bc77380b8411efc74df96b9cacb9ce301d06c4749e0ccc97b6" Mar 13 15:17:48 crc kubenswrapper[4907]: I0313 15:17:48.950394 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f64382383e10d8bc77380b8411efc74df96b9cacb9ce301d06c4749e0ccc97b6"} err="failed to get container status \"f64382383e10d8bc77380b8411efc74df96b9cacb9ce301d06c4749e0ccc97b6\": rpc error: code = NotFound desc = could not find container \"f64382383e10d8bc77380b8411efc74df96b9cacb9ce301d06c4749e0ccc97b6\": container with ID starting with f64382383e10d8bc77380b8411efc74df96b9cacb9ce301d06c4749e0ccc97b6 not found: ID does not exist" Mar 13 15:17:49 crc kubenswrapper[4907]: E0313 15:17:49.070000 4907 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7a6ef1dc_77f0_4a25_9465_f9b881023f83.slice/crio-4e3fd1a5082ad9dee2e764ae30a95c9a0adcb1f77f2fd1c6a9389907f1dd1779\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7a6ef1dc_77f0_4a25_9465_f9b881023f83.slice\": RecentStats: unable to find data in memory cache]" Mar 13 15:17:49 crc kubenswrapper[4907]: I0313 15:17:49.791764 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7a6ef1dc-77f0-4a25-9465-f9b881023f83" path="/var/lib/kubelet/pods/7a6ef1dc-77f0-4a25-9465-f9b881023f83/volumes" Mar 13 15:17:56 crc kubenswrapper[4907]: I0313 15:17:56.234351 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-xhrm7" Mar 13 15:17:56 crc kubenswrapper[4907]: I0313 15:17:56.276760 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-xhrm7" Mar 13 15:17:56 crc kubenswrapper[4907]: I0313 15:17:56.467630 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-xhrm7"] Mar 13 15:17:57 crc kubenswrapper[4907]: I0313 15:17:57.927638 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-xhrm7" podUID="78255f5a-d6ed-45c3-ade0-4ff6f01c7a0f" containerName="registry-server" containerID="cri-o://91ebf771b78bb7a2fe3a8359fd7a3263f5fdd1c2032beba161125b81cdf81d3c" gracePeriod=2 Mar 13 15:17:58 crc kubenswrapper[4907]: I0313 15:17:58.321717 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xhrm7" Mar 13 15:17:58 crc kubenswrapper[4907]: I0313 15:17:58.412955 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m4gbx\" (UniqueName: \"kubernetes.io/projected/78255f5a-d6ed-45c3-ade0-4ff6f01c7a0f-kube-api-access-m4gbx\") pod \"78255f5a-d6ed-45c3-ade0-4ff6f01c7a0f\" (UID: \"78255f5a-d6ed-45c3-ade0-4ff6f01c7a0f\") " Mar 13 15:17:58 crc kubenswrapper[4907]: I0313 15:17:58.413022 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/78255f5a-d6ed-45c3-ade0-4ff6f01c7a0f-utilities\") pod \"78255f5a-d6ed-45c3-ade0-4ff6f01c7a0f\" (UID: \"78255f5a-d6ed-45c3-ade0-4ff6f01c7a0f\") " Mar 13 15:17:58 crc kubenswrapper[4907]: I0313 15:17:58.413099 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/78255f5a-d6ed-45c3-ade0-4ff6f01c7a0f-catalog-content\") pod \"78255f5a-d6ed-45c3-ade0-4ff6f01c7a0f\" (UID: \"78255f5a-d6ed-45c3-ade0-4ff6f01c7a0f\") " Mar 13 15:17:58 crc kubenswrapper[4907]: I0313 15:17:58.414387 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/78255f5a-d6ed-45c3-ade0-4ff6f01c7a0f-utilities" (OuterVolumeSpecName: "utilities") pod "78255f5a-d6ed-45c3-ade0-4ff6f01c7a0f" (UID: "78255f5a-d6ed-45c3-ade0-4ff6f01c7a0f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:17:58 crc kubenswrapper[4907]: I0313 15:17:58.424189 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/78255f5a-d6ed-45c3-ade0-4ff6f01c7a0f-kube-api-access-m4gbx" (OuterVolumeSpecName: "kube-api-access-m4gbx") pod "78255f5a-d6ed-45c3-ade0-4ff6f01c7a0f" (UID: "78255f5a-d6ed-45c3-ade0-4ff6f01c7a0f"). InnerVolumeSpecName "kube-api-access-m4gbx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:17:58 crc kubenswrapper[4907]: I0313 15:17:58.515143 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m4gbx\" (UniqueName: \"kubernetes.io/projected/78255f5a-d6ed-45c3-ade0-4ff6f01c7a0f-kube-api-access-m4gbx\") on node \"crc\" DevicePath \"\"" Mar 13 15:17:58 crc kubenswrapper[4907]: I0313 15:17:58.515194 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/78255f5a-d6ed-45c3-ade0-4ff6f01c7a0f-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 15:17:58 crc kubenswrapper[4907]: I0313 15:17:58.563484 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/78255f5a-d6ed-45c3-ade0-4ff6f01c7a0f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "78255f5a-d6ed-45c3-ade0-4ff6f01c7a0f" (UID: "78255f5a-d6ed-45c3-ade0-4ff6f01c7a0f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:17:58 crc kubenswrapper[4907]: I0313 15:17:58.616728 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/78255f5a-d6ed-45c3-ade0-4ff6f01c7a0f-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 15:17:58 crc kubenswrapper[4907]: I0313 15:17:58.937451 4907 generic.go:334] "Generic (PLEG): container finished" podID="78255f5a-d6ed-45c3-ade0-4ff6f01c7a0f" containerID="91ebf771b78bb7a2fe3a8359fd7a3263f5fdd1c2032beba161125b81cdf81d3c" exitCode=0 Mar 13 15:17:58 crc kubenswrapper[4907]: I0313 15:17:58.937533 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xhrm7" event={"ID":"78255f5a-d6ed-45c3-ade0-4ff6f01c7a0f","Type":"ContainerDied","Data":"91ebf771b78bb7a2fe3a8359fd7a3263f5fdd1c2032beba161125b81cdf81d3c"} Mar 13 15:17:58 crc kubenswrapper[4907]: I0313 15:17:58.937574 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xhrm7" Mar 13 15:17:58 crc kubenswrapper[4907]: I0313 15:17:58.937616 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xhrm7" event={"ID":"78255f5a-d6ed-45c3-ade0-4ff6f01c7a0f","Type":"ContainerDied","Data":"e710e5af60dad227fb28f3f3f8736be015c34c3cbf424e8cb7767ac5bea1fdd9"} Mar 13 15:17:58 crc kubenswrapper[4907]: I0313 15:17:58.937647 4907 scope.go:117] "RemoveContainer" containerID="91ebf771b78bb7a2fe3a8359fd7a3263f5fdd1c2032beba161125b81cdf81d3c" Mar 13 15:17:58 crc kubenswrapper[4907]: I0313 15:17:58.973660 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-xhrm7"] Mar 13 15:17:58 crc kubenswrapper[4907]: I0313 15:17:58.973811 4907 scope.go:117] "RemoveContainer" containerID="21f6ddb10036b57ffc641c1c43e81015c667a267b0b5f356fdd1eeded7b6a958" Mar 13 15:17:58 crc kubenswrapper[4907]: I0313 15:17:58.988225 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-xhrm7"] Mar 13 15:17:59 crc kubenswrapper[4907]: I0313 15:17:59.011280 4907 scope.go:117] "RemoveContainer" containerID="cdd1adc8cc682b604b6e997800c49a6fbb65e62d6431a882b09277e919d44d4d" Mar 13 15:17:59 crc kubenswrapper[4907]: I0313 15:17:59.040273 4907 scope.go:117] "RemoveContainer" containerID="91ebf771b78bb7a2fe3a8359fd7a3263f5fdd1c2032beba161125b81cdf81d3c" Mar 13 15:17:59 crc kubenswrapper[4907]: E0313 15:17:59.041504 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"91ebf771b78bb7a2fe3a8359fd7a3263f5fdd1c2032beba161125b81cdf81d3c\": container with ID starting with 91ebf771b78bb7a2fe3a8359fd7a3263f5fdd1c2032beba161125b81cdf81d3c not found: ID does not exist" containerID="91ebf771b78bb7a2fe3a8359fd7a3263f5fdd1c2032beba161125b81cdf81d3c" Mar 13 15:17:59 crc kubenswrapper[4907]: I0313 15:17:59.041546 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"91ebf771b78bb7a2fe3a8359fd7a3263f5fdd1c2032beba161125b81cdf81d3c"} err="failed to get container status \"91ebf771b78bb7a2fe3a8359fd7a3263f5fdd1c2032beba161125b81cdf81d3c\": rpc error: code = NotFound desc = could not find container \"91ebf771b78bb7a2fe3a8359fd7a3263f5fdd1c2032beba161125b81cdf81d3c\": container with ID starting with 91ebf771b78bb7a2fe3a8359fd7a3263f5fdd1c2032beba161125b81cdf81d3c not found: ID does not exist" Mar 13 15:17:59 crc kubenswrapper[4907]: I0313 15:17:59.041573 4907 scope.go:117] "RemoveContainer" containerID="21f6ddb10036b57ffc641c1c43e81015c667a267b0b5f356fdd1eeded7b6a958" Mar 13 15:17:59 crc kubenswrapper[4907]: E0313 15:17:59.042070 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"21f6ddb10036b57ffc641c1c43e81015c667a267b0b5f356fdd1eeded7b6a958\": container with ID starting with 21f6ddb10036b57ffc641c1c43e81015c667a267b0b5f356fdd1eeded7b6a958 not found: ID does not exist" containerID="21f6ddb10036b57ffc641c1c43e81015c667a267b0b5f356fdd1eeded7b6a958" Mar 13 15:17:59 crc kubenswrapper[4907]: I0313 15:17:59.042112 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"21f6ddb10036b57ffc641c1c43e81015c667a267b0b5f356fdd1eeded7b6a958"} err="failed to get container status \"21f6ddb10036b57ffc641c1c43e81015c667a267b0b5f356fdd1eeded7b6a958\": rpc error: code = NotFound desc = could not find container \"21f6ddb10036b57ffc641c1c43e81015c667a267b0b5f356fdd1eeded7b6a958\": container with ID starting with 21f6ddb10036b57ffc641c1c43e81015c667a267b0b5f356fdd1eeded7b6a958 not found: ID does not exist" Mar 13 15:17:59 crc kubenswrapper[4907]: I0313 15:17:59.042140 4907 scope.go:117] "RemoveContainer" containerID="cdd1adc8cc682b604b6e997800c49a6fbb65e62d6431a882b09277e919d44d4d" Mar 13 15:17:59 crc kubenswrapper[4907]: E0313 15:17:59.042520 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cdd1adc8cc682b604b6e997800c49a6fbb65e62d6431a882b09277e919d44d4d\": container with ID starting with cdd1adc8cc682b604b6e997800c49a6fbb65e62d6431a882b09277e919d44d4d not found: ID does not exist" containerID="cdd1adc8cc682b604b6e997800c49a6fbb65e62d6431a882b09277e919d44d4d" Mar 13 15:17:59 crc kubenswrapper[4907]: I0313 15:17:59.042573 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cdd1adc8cc682b604b6e997800c49a6fbb65e62d6431a882b09277e919d44d4d"} err="failed to get container status \"cdd1adc8cc682b604b6e997800c49a6fbb65e62d6431a882b09277e919d44d4d\": rpc error: code = NotFound desc = could not find container \"cdd1adc8cc682b604b6e997800c49a6fbb65e62d6431a882b09277e919d44d4d\": container with ID starting with cdd1adc8cc682b604b6e997800c49a6fbb65e62d6431a882b09277e919d44d4d not found: ID does not exist" Mar 13 15:17:59 crc kubenswrapper[4907]: I0313 15:17:59.794493 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="78255f5a-d6ed-45c3-ade0-4ff6f01c7a0f" path="/var/lib/kubelet/pods/78255f5a-d6ed-45c3-ade0-4ff6f01c7a0f/volumes" Mar 13 15:18:00 crc kubenswrapper[4907]: I0313 15:18:00.154960 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556918-cs4dh"] Mar 13 15:18:00 crc kubenswrapper[4907]: E0313 15:18:00.155832 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a6ef1dc-77f0-4a25-9465-f9b881023f83" containerName="extract-content" Mar 13 15:18:00 crc kubenswrapper[4907]: I0313 15:18:00.155855 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a6ef1dc-77f0-4a25-9465-f9b881023f83" containerName="extract-content" Mar 13 15:18:00 crc kubenswrapper[4907]: E0313 15:18:00.155910 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78255f5a-d6ed-45c3-ade0-4ff6f01c7a0f" containerName="extract-utilities" Mar 13 15:18:00 crc kubenswrapper[4907]: I0313 15:18:00.155926 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="78255f5a-d6ed-45c3-ade0-4ff6f01c7a0f" containerName="extract-utilities" Mar 13 15:18:00 crc kubenswrapper[4907]: E0313 15:18:00.155954 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a6ef1dc-77f0-4a25-9465-f9b881023f83" containerName="registry-server" Mar 13 15:18:00 crc kubenswrapper[4907]: I0313 15:18:00.155967 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a6ef1dc-77f0-4a25-9465-f9b881023f83" containerName="registry-server" Mar 13 15:18:00 crc kubenswrapper[4907]: E0313 15:18:00.155985 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a6ef1dc-77f0-4a25-9465-f9b881023f83" containerName="extract-utilities" Mar 13 15:18:00 crc kubenswrapper[4907]: I0313 15:18:00.155998 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a6ef1dc-77f0-4a25-9465-f9b881023f83" containerName="extract-utilities" Mar 13 15:18:00 crc kubenswrapper[4907]: E0313 15:18:00.156018 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78255f5a-d6ed-45c3-ade0-4ff6f01c7a0f" containerName="registry-server" Mar 13 15:18:00 crc kubenswrapper[4907]: I0313 15:18:00.156030 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="78255f5a-d6ed-45c3-ade0-4ff6f01c7a0f" containerName="registry-server" Mar 13 15:18:00 crc kubenswrapper[4907]: E0313 15:18:00.156057 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78255f5a-d6ed-45c3-ade0-4ff6f01c7a0f" containerName="extract-content" Mar 13 15:18:00 crc kubenswrapper[4907]: I0313 15:18:00.156069 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="78255f5a-d6ed-45c3-ade0-4ff6f01c7a0f" containerName="extract-content" Mar 13 15:18:00 crc kubenswrapper[4907]: I0313 15:18:00.156309 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="78255f5a-d6ed-45c3-ade0-4ff6f01c7a0f" containerName="registry-server" Mar 13 15:18:00 crc kubenswrapper[4907]: I0313 15:18:00.156346 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a6ef1dc-77f0-4a25-9465-f9b881023f83" containerName="registry-server" Mar 13 15:18:00 crc kubenswrapper[4907]: I0313 15:18:00.157167 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556918-cs4dh" Mar 13 15:18:00 crc kubenswrapper[4907]: I0313 15:18:00.160631 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 15:18:00 crc kubenswrapper[4907]: I0313 15:18:00.160832 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 15:18:00 crc kubenswrapper[4907]: I0313 15:18:00.160993 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 15:18:00 crc kubenswrapper[4907]: I0313 15:18:00.192942 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556918-cs4dh"] Mar 13 15:18:00 crc kubenswrapper[4907]: I0313 15:18:00.237709 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6m4sb\" (UniqueName: \"kubernetes.io/projected/fdd030a7-3ebb-4fbf-91f2-2ab96e526252-kube-api-access-6m4sb\") pod \"auto-csr-approver-29556918-cs4dh\" (UID: \"fdd030a7-3ebb-4fbf-91f2-2ab96e526252\") " pod="openshift-infra/auto-csr-approver-29556918-cs4dh" Mar 13 15:18:00 crc kubenswrapper[4907]: I0313 15:18:00.338844 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6m4sb\" (UniqueName: \"kubernetes.io/projected/fdd030a7-3ebb-4fbf-91f2-2ab96e526252-kube-api-access-6m4sb\") pod \"auto-csr-approver-29556918-cs4dh\" (UID: \"fdd030a7-3ebb-4fbf-91f2-2ab96e526252\") " pod="openshift-infra/auto-csr-approver-29556918-cs4dh" Mar 13 15:18:00 crc kubenswrapper[4907]: I0313 15:18:00.356234 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6m4sb\" (UniqueName: \"kubernetes.io/projected/fdd030a7-3ebb-4fbf-91f2-2ab96e526252-kube-api-access-6m4sb\") pod \"auto-csr-approver-29556918-cs4dh\" (UID: \"fdd030a7-3ebb-4fbf-91f2-2ab96e526252\") " pod="openshift-infra/auto-csr-approver-29556918-cs4dh" Mar 13 15:18:00 crc kubenswrapper[4907]: I0313 15:18:00.501374 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556918-cs4dh" Mar 13 15:18:00 crc kubenswrapper[4907]: I0313 15:18:00.902714 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556918-cs4dh"] Mar 13 15:18:00 crc kubenswrapper[4907]: I0313 15:18:00.960608 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556918-cs4dh" event={"ID":"fdd030a7-3ebb-4fbf-91f2-2ab96e526252","Type":"ContainerStarted","Data":"51efcbdac2654ebcc5d956fb9add4884a6d192c679448f44a80b43591fcf224b"} Mar 13 15:18:02 crc kubenswrapper[4907]: I0313 15:18:02.985500 4907 generic.go:334] "Generic (PLEG): container finished" podID="fdd030a7-3ebb-4fbf-91f2-2ab96e526252" containerID="e984813306df1ab82de9a9dd04affc9e989fc2092eb575a81e8fc5794b033f1a" exitCode=0 Mar 13 15:18:02 crc kubenswrapper[4907]: I0313 15:18:02.985664 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556918-cs4dh" event={"ID":"fdd030a7-3ebb-4fbf-91f2-2ab96e526252","Type":"ContainerDied","Data":"e984813306df1ab82de9a9dd04affc9e989fc2092eb575a81e8fc5794b033f1a"} Mar 13 15:18:04 crc kubenswrapper[4907]: I0313 15:18:04.285676 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556918-cs4dh" Mar 13 15:18:04 crc kubenswrapper[4907]: I0313 15:18:04.395834 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6m4sb\" (UniqueName: \"kubernetes.io/projected/fdd030a7-3ebb-4fbf-91f2-2ab96e526252-kube-api-access-6m4sb\") pod \"fdd030a7-3ebb-4fbf-91f2-2ab96e526252\" (UID: \"fdd030a7-3ebb-4fbf-91f2-2ab96e526252\") " Mar 13 15:18:04 crc kubenswrapper[4907]: I0313 15:18:04.400849 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fdd030a7-3ebb-4fbf-91f2-2ab96e526252-kube-api-access-6m4sb" (OuterVolumeSpecName: "kube-api-access-6m4sb") pod "fdd030a7-3ebb-4fbf-91f2-2ab96e526252" (UID: "fdd030a7-3ebb-4fbf-91f2-2ab96e526252"). InnerVolumeSpecName "kube-api-access-6m4sb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:18:04 crc kubenswrapper[4907]: I0313 15:18:04.497475 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6m4sb\" (UniqueName: \"kubernetes.io/projected/fdd030a7-3ebb-4fbf-91f2-2ab96e526252-kube-api-access-6m4sb\") on node \"crc\" DevicePath \"\"" Mar 13 15:18:05 crc kubenswrapper[4907]: I0313 15:18:05.002147 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556918-cs4dh" event={"ID":"fdd030a7-3ebb-4fbf-91f2-2ab96e526252","Type":"ContainerDied","Data":"51efcbdac2654ebcc5d956fb9add4884a6d192c679448f44a80b43591fcf224b"} Mar 13 15:18:05 crc kubenswrapper[4907]: I0313 15:18:05.002202 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="51efcbdac2654ebcc5d956fb9add4884a6d192c679448f44a80b43591fcf224b" Mar 13 15:18:05 crc kubenswrapper[4907]: I0313 15:18:05.002266 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556918-cs4dh" Mar 13 15:18:05 crc kubenswrapper[4907]: I0313 15:18:05.368274 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556912-smqnn"] Mar 13 15:18:05 crc kubenswrapper[4907]: I0313 15:18:05.376970 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556912-smqnn"] Mar 13 15:18:05 crc kubenswrapper[4907]: I0313 15:18:05.790337 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="449e0ca3-4756-4ec2-8ea2-b2b6fbe52ca9" path="/var/lib/kubelet/pods/449e0ca3-4756-4ec2-8ea2-b2b6fbe52ca9/volumes" Mar 13 15:18:18 crc kubenswrapper[4907]: I0313 15:18:18.041665 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 15:18:18 crc kubenswrapper[4907]: I0313 15:18:18.042293 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 15:18:18 crc kubenswrapper[4907]: I0313 15:18:18.042355 4907 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" Mar 13 15:18:18 crc kubenswrapper[4907]: I0313 15:18:18.043144 4907 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"dd934d7c5fc7e95362f79c2301ff40920ce7e3112722cbc11a053a56c085e05c"} pod="openshift-machine-config-operator/machine-config-daemon-hm56j" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 15:18:18 crc kubenswrapper[4907]: I0313 15:18:18.043245 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" containerID="cri-o://dd934d7c5fc7e95362f79c2301ff40920ce7e3112722cbc11a053a56c085e05c" gracePeriod=600 Mar 13 15:18:18 crc kubenswrapper[4907]: E0313 15:18:18.178096 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:18:19 crc kubenswrapper[4907]: I0313 15:18:19.120998 4907 generic.go:334] "Generic (PLEG): container finished" podID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerID="dd934d7c5fc7e95362f79c2301ff40920ce7e3112722cbc11a053a56c085e05c" exitCode=0 Mar 13 15:18:19 crc kubenswrapper[4907]: I0313 15:18:19.121046 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerDied","Data":"dd934d7c5fc7e95362f79c2301ff40920ce7e3112722cbc11a053a56c085e05c"} Mar 13 15:18:19 crc kubenswrapper[4907]: I0313 15:18:19.121088 4907 scope.go:117] "RemoveContainer" containerID="36ef7b96f3457426d961c14798528949d4c51116631e09ea3be0c17a7c07b91d" Mar 13 15:18:19 crc kubenswrapper[4907]: I0313 15:18:19.121841 4907 scope.go:117] "RemoveContainer" containerID="dd934d7c5fc7e95362f79c2301ff40920ce7e3112722cbc11a053a56c085e05c" Mar 13 15:18:19 crc kubenswrapper[4907]: E0313 15:18:19.122108 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:18:26 crc kubenswrapper[4907]: I0313 15:18:26.807193 4907 scope.go:117] "RemoveContainer" containerID="eb18542ee2a2c90ab854737b99c04f55fb12626f0e4778cc4e2869a3c3a6cd0f" Mar 13 15:18:30 crc kubenswrapper[4907]: I0313 15:18:30.782751 4907 scope.go:117] "RemoveContainer" containerID="dd934d7c5fc7e95362f79c2301ff40920ce7e3112722cbc11a053a56c085e05c" Mar 13 15:18:30 crc kubenswrapper[4907]: E0313 15:18:30.783980 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:18:45 crc kubenswrapper[4907]: I0313 15:18:45.782847 4907 scope.go:117] "RemoveContainer" containerID="dd934d7c5fc7e95362f79c2301ff40920ce7e3112722cbc11a053a56c085e05c" Mar 13 15:18:45 crc kubenswrapper[4907]: E0313 15:18:45.783718 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:18:58 crc kubenswrapper[4907]: I0313 15:18:58.783088 4907 scope.go:117] "RemoveContainer" containerID="dd934d7c5fc7e95362f79c2301ff40920ce7e3112722cbc11a053a56c085e05c" Mar 13 15:18:58 crc kubenswrapper[4907]: E0313 15:18:58.784170 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:19:13 crc kubenswrapper[4907]: I0313 15:19:13.782153 4907 scope.go:117] "RemoveContainer" containerID="dd934d7c5fc7e95362f79c2301ff40920ce7e3112722cbc11a053a56c085e05c" Mar 13 15:19:13 crc kubenswrapper[4907]: E0313 15:19:13.782938 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:19:27 crc kubenswrapper[4907]: I0313 15:19:27.783013 4907 scope.go:117] "RemoveContainer" containerID="dd934d7c5fc7e95362f79c2301ff40920ce7e3112722cbc11a053a56c085e05c" Mar 13 15:19:27 crc kubenswrapper[4907]: E0313 15:19:27.783863 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:19:39 crc kubenswrapper[4907]: I0313 15:19:39.782958 4907 scope.go:117] "RemoveContainer" containerID="dd934d7c5fc7e95362f79c2301ff40920ce7e3112722cbc11a053a56c085e05c" Mar 13 15:19:39 crc kubenswrapper[4907]: E0313 15:19:39.783827 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:19:52 crc kubenswrapper[4907]: I0313 15:19:52.782783 4907 scope.go:117] "RemoveContainer" containerID="dd934d7c5fc7e95362f79c2301ff40920ce7e3112722cbc11a053a56c085e05c" Mar 13 15:19:52 crc kubenswrapper[4907]: E0313 15:19:52.783497 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:20:00 crc kubenswrapper[4907]: I0313 15:20:00.146158 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556920-lrh8k"] Mar 13 15:20:00 crc kubenswrapper[4907]: E0313 15:20:00.147080 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fdd030a7-3ebb-4fbf-91f2-2ab96e526252" containerName="oc" Mar 13 15:20:00 crc kubenswrapper[4907]: I0313 15:20:00.147096 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="fdd030a7-3ebb-4fbf-91f2-2ab96e526252" containerName="oc" Mar 13 15:20:00 crc kubenswrapper[4907]: I0313 15:20:00.147262 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="fdd030a7-3ebb-4fbf-91f2-2ab96e526252" containerName="oc" Mar 13 15:20:00 crc kubenswrapper[4907]: I0313 15:20:00.147826 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556920-lrh8k" Mar 13 15:20:00 crc kubenswrapper[4907]: I0313 15:20:00.150925 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 15:20:00 crc kubenswrapper[4907]: I0313 15:20:00.151489 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 15:20:00 crc kubenswrapper[4907]: I0313 15:20:00.151582 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 15:20:00 crc kubenswrapper[4907]: I0313 15:20:00.151924 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556920-lrh8k"] Mar 13 15:20:00 crc kubenswrapper[4907]: I0313 15:20:00.260669 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ltdsd\" (UniqueName: \"kubernetes.io/projected/925bf3a9-9978-49a9-8ac3-7c92a928adb7-kube-api-access-ltdsd\") pod \"auto-csr-approver-29556920-lrh8k\" (UID: \"925bf3a9-9978-49a9-8ac3-7c92a928adb7\") " pod="openshift-infra/auto-csr-approver-29556920-lrh8k" Mar 13 15:20:00 crc kubenswrapper[4907]: I0313 15:20:00.362730 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ltdsd\" (UniqueName: \"kubernetes.io/projected/925bf3a9-9978-49a9-8ac3-7c92a928adb7-kube-api-access-ltdsd\") pod \"auto-csr-approver-29556920-lrh8k\" (UID: \"925bf3a9-9978-49a9-8ac3-7c92a928adb7\") " pod="openshift-infra/auto-csr-approver-29556920-lrh8k" Mar 13 15:20:00 crc kubenswrapper[4907]: I0313 15:20:00.381426 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ltdsd\" (UniqueName: \"kubernetes.io/projected/925bf3a9-9978-49a9-8ac3-7c92a928adb7-kube-api-access-ltdsd\") pod \"auto-csr-approver-29556920-lrh8k\" (UID: \"925bf3a9-9978-49a9-8ac3-7c92a928adb7\") " pod="openshift-infra/auto-csr-approver-29556920-lrh8k" Mar 13 15:20:00 crc kubenswrapper[4907]: I0313 15:20:00.469447 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556920-lrh8k" Mar 13 15:20:00 crc kubenswrapper[4907]: I0313 15:20:00.681123 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556920-lrh8k"] Mar 13 15:20:00 crc kubenswrapper[4907]: I0313 15:20:00.695127 4907 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 13 15:20:01 crc kubenswrapper[4907]: I0313 15:20:01.241714 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556920-lrh8k" event={"ID":"925bf3a9-9978-49a9-8ac3-7c92a928adb7","Type":"ContainerStarted","Data":"9f0ae93c37a3701544ad75512f3f6ea72f66a0bf17bc5937952fb61e324abb61"} Mar 13 15:20:02 crc kubenswrapper[4907]: I0313 15:20:02.251218 4907 generic.go:334] "Generic (PLEG): container finished" podID="925bf3a9-9978-49a9-8ac3-7c92a928adb7" containerID="e5083f5e7cb4434ec8bff6d3c9df8bdd433195c0fc9dbe495454864580276a9b" exitCode=0 Mar 13 15:20:02 crc kubenswrapper[4907]: I0313 15:20:02.251308 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556920-lrh8k" event={"ID":"925bf3a9-9978-49a9-8ac3-7c92a928adb7","Type":"ContainerDied","Data":"e5083f5e7cb4434ec8bff6d3c9df8bdd433195c0fc9dbe495454864580276a9b"} Mar 13 15:20:03 crc kubenswrapper[4907]: I0313 15:20:03.563074 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556920-lrh8k" Mar 13 15:20:03 crc kubenswrapper[4907]: I0313 15:20:03.612918 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ltdsd\" (UniqueName: \"kubernetes.io/projected/925bf3a9-9978-49a9-8ac3-7c92a928adb7-kube-api-access-ltdsd\") pod \"925bf3a9-9978-49a9-8ac3-7c92a928adb7\" (UID: \"925bf3a9-9978-49a9-8ac3-7c92a928adb7\") " Mar 13 15:20:03 crc kubenswrapper[4907]: I0313 15:20:03.705122 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925bf3a9-9978-49a9-8ac3-7c92a928adb7-kube-api-access-ltdsd" (OuterVolumeSpecName: "kube-api-access-ltdsd") pod "925bf3a9-9978-49a9-8ac3-7c92a928adb7" (UID: "925bf3a9-9978-49a9-8ac3-7c92a928adb7"). InnerVolumeSpecName "kube-api-access-ltdsd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:20:03 crc kubenswrapper[4907]: I0313 15:20:03.714661 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ltdsd\" (UniqueName: \"kubernetes.io/projected/925bf3a9-9978-49a9-8ac3-7c92a928adb7-kube-api-access-ltdsd\") on node \"crc\" DevicePath \"\"" Mar 13 15:20:04 crc kubenswrapper[4907]: I0313 15:20:04.265199 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556920-lrh8k" event={"ID":"925bf3a9-9978-49a9-8ac3-7c92a928adb7","Type":"ContainerDied","Data":"9f0ae93c37a3701544ad75512f3f6ea72f66a0bf17bc5937952fb61e324abb61"} Mar 13 15:20:04 crc kubenswrapper[4907]: I0313 15:20:04.265479 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9f0ae93c37a3701544ad75512f3f6ea72f66a0bf17bc5937952fb61e324abb61" Mar 13 15:20:04 crc kubenswrapper[4907]: I0313 15:20:04.265292 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556920-lrh8k" Mar 13 15:20:04 crc kubenswrapper[4907]: I0313 15:20:04.624039 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556914-98b9m"] Mar 13 15:20:04 crc kubenswrapper[4907]: I0313 15:20:04.629318 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556914-98b9m"] Mar 13 15:20:05 crc kubenswrapper[4907]: I0313 15:20:05.782160 4907 scope.go:117] "RemoveContainer" containerID="dd934d7c5fc7e95362f79c2301ff40920ce7e3112722cbc11a053a56c085e05c" Mar 13 15:20:05 crc kubenswrapper[4907]: E0313 15:20:05.782378 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:20:05 crc kubenswrapper[4907]: I0313 15:20:05.795424 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="561da6ce-3fe7-4e35-97f0-6a39582d3d4b" path="/var/lib/kubelet/pods/561da6ce-3fe7-4e35-97f0-6a39582d3d4b/volumes" Mar 13 15:20:18 crc kubenswrapper[4907]: I0313 15:20:18.782996 4907 scope.go:117] "RemoveContainer" containerID="dd934d7c5fc7e95362f79c2301ff40920ce7e3112722cbc11a053a56c085e05c" Mar 13 15:20:18 crc kubenswrapper[4907]: E0313 15:20:18.784350 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:20:27 crc kubenswrapper[4907]: I0313 15:20:27.003448 4907 scope.go:117] "RemoveContainer" containerID="692bff8b244c6ed0093ba50083b6fcff2e9a78403c7f3e3cc6626818a66971bb" Mar 13 15:20:32 crc kubenswrapper[4907]: I0313 15:20:32.782391 4907 scope.go:117] "RemoveContainer" containerID="dd934d7c5fc7e95362f79c2301ff40920ce7e3112722cbc11a053a56c085e05c" Mar 13 15:20:32 crc kubenswrapper[4907]: E0313 15:20:32.783127 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:20:43 crc kubenswrapper[4907]: I0313 15:20:43.782324 4907 scope.go:117] "RemoveContainer" containerID="dd934d7c5fc7e95362f79c2301ff40920ce7e3112722cbc11a053a56c085e05c" Mar 13 15:20:43 crc kubenswrapper[4907]: E0313 15:20:43.784271 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:20:57 crc kubenswrapper[4907]: I0313 15:20:57.782508 4907 scope.go:117] "RemoveContainer" containerID="dd934d7c5fc7e95362f79c2301ff40920ce7e3112722cbc11a053a56c085e05c" Mar 13 15:20:57 crc kubenswrapper[4907]: E0313 15:20:57.783246 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:21:11 crc kubenswrapper[4907]: I0313 15:21:11.786666 4907 scope.go:117] "RemoveContainer" containerID="dd934d7c5fc7e95362f79c2301ff40920ce7e3112722cbc11a053a56c085e05c" Mar 13 15:21:11 crc kubenswrapper[4907]: E0313 15:21:11.787455 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:21:22 crc kubenswrapper[4907]: I0313 15:21:22.781721 4907 scope.go:117] "RemoveContainer" containerID="dd934d7c5fc7e95362f79c2301ff40920ce7e3112722cbc11a053a56c085e05c" Mar 13 15:21:22 crc kubenswrapper[4907]: E0313 15:21:22.782562 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:21:36 crc kubenswrapper[4907]: I0313 15:21:36.783553 4907 scope.go:117] "RemoveContainer" containerID="dd934d7c5fc7e95362f79c2301ff40920ce7e3112722cbc11a053a56c085e05c" Mar 13 15:21:36 crc kubenswrapper[4907]: E0313 15:21:36.784507 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:21:48 crc kubenswrapper[4907]: I0313 15:21:48.743213 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-qtts8"] Mar 13 15:21:48 crc kubenswrapper[4907]: E0313 15:21:48.744139 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="925bf3a9-9978-49a9-8ac3-7c92a928adb7" containerName="oc" Mar 13 15:21:48 crc kubenswrapper[4907]: I0313 15:21:48.744155 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="925bf3a9-9978-49a9-8ac3-7c92a928adb7" containerName="oc" Mar 13 15:21:48 crc kubenswrapper[4907]: I0313 15:21:48.744324 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="925bf3a9-9978-49a9-8ac3-7c92a928adb7" containerName="oc" Mar 13 15:21:48 crc kubenswrapper[4907]: I0313 15:21:48.745485 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qtts8" Mar 13 15:21:48 crc kubenswrapper[4907]: I0313 15:21:48.762253 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qtts8"] Mar 13 15:21:48 crc kubenswrapper[4907]: I0313 15:21:48.781863 4907 scope.go:117] "RemoveContainer" containerID="dd934d7c5fc7e95362f79c2301ff40920ce7e3112722cbc11a053a56c085e05c" Mar 13 15:21:48 crc kubenswrapper[4907]: E0313 15:21:48.782232 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:21:48 crc kubenswrapper[4907]: I0313 15:21:48.871100 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ce4bcd2-1af1-43b6-8e59-d470e9089686-utilities\") pod \"redhat-marketplace-qtts8\" (UID: \"8ce4bcd2-1af1-43b6-8e59-d470e9089686\") " pod="openshift-marketplace/redhat-marketplace-qtts8" Mar 13 15:21:48 crc kubenswrapper[4907]: I0313 15:21:48.871277 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-56pfq\" (UniqueName: \"kubernetes.io/projected/8ce4bcd2-1af1-43b6-8e59-d470e9089686-kube-api-access-56pfq\") pod \"redhat-marketplace-qtts8\" (UID: \"8ce4bcd2-1af1-43b6-8e59-d470e9089686\") " pod="openshift-marketplace/redhat-marketplace-qtts8" Mar 13 15:21:48 crc kubenswrapper[4907]: I0313 15:21:48.871343 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ce4bcd2-1af1-43b6-8e59-d470e9089686-catalog-content\") pod \"redhat-marketplace-qtts8\" (UID: \"8ce4bcd2-1af1-43b6-8e59-d470e9089686\") " pod="openshift-marketplace/redhat-marketplace-qtts8" Mar 13 15:21:48 crc kubenswrapper[4907]: I0313 15:21:48.972360 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ce4bcd2-1af1-43b6-8e59-d470e9089686-utilities\") pod \"redhat-marketplace-qtts8\" (UID: \"8ce4bcd2-1af1-43b6-8e59-d470e9089686\") " pod="openshift-marketplace/redhat-marketplace-qtts8" Mar 13 15:21:48 crc kubenswrapper[4907]: I0313 15:21:48.972435 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-56pfq\" (UniqueName: \"kubernetes.io/projected/8ce4bcd2-1af1-43b6-8e59-d470e9089686-kube-api-access-56pfq\") pod \"redhat-marketplace-qtts8\" (UID: \"8ce4bcd2-1af1-43b6-8e59-d470e9089686\") " pod="openshift-marketplace/redhat-marketplace-qtts8" Mar 13 15:21:48 crc kubenswrapper[4907]: I0313 15:21:48.972466 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ce4bcd2-1af1-43b6-8e59-d470e9089686-catalog-content\") pod \"redhat-marketplace-qtts8\" (UID: \"8ce4bcd2-1af1-43b6-8e59-d470e9089686\") " pod="openshift-marketplace/redhat-marketplace-qtts8" Mar 13 15:21:48 crc kubenswrapper[4907]: I0313 15:21:48.972921 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ce4bcd2-1af1-43b6-8e59-d470e9089686-utilities\") pod \"redhat-marketplace-qtts8\" (UID: \"8ce4bcd2-1af1-43b6-8e59-d470e9089686\") " pod="openshift-marketplace/redhat-marketplace-qtts8" Mar 13 15:21:48 crc kubenswrapper[4907]: I0313 15:21:48.972970 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ce4bcd2-1af1-43b6-8e59-d470e9089686-catalog-content\") pod \"redhat-marketplace-qtts8\" (UID: \"8ce4bcd2-1af1-43b6-8e59-d470e9089686\") " pod="openshift-marketplace/redhat-marketplace-qtts8" Mar 13 15:21:48 crc kubenswrapper[4907]: I0313 15:21:48.994940 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-56pfq\" (UniqueName: \"kubernetes.io/projected/8ce4bcd2-1af1-43b6-8e59-d470e9089686-kube-api-access-56pfq\") pod \"redhat-marketplace-qtts8\" (UID: \"8ce4bcd2-1af1-43b6-8e59-d470e9089686\") " pod="openshift-marketplace/redhat-marketplace-qtts8" Mar 13 15:21:49 crc kubenswrapper[4907]: I0313 15:21:49.064182 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qtts8" Mar 13 15:21:49 crc kubenswrapper[4907]: I0313 15:21:49.562767 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qtts8"] Mar 13 15:21:50 crc kubenswrapper[4907]: I0313 15:21:50.040227 4907 generic.go:334] "Generic (PLEG): container finished" podID="8ce4bcd2-1af1-43b6-8e59-d470e9089686" containerID="52eff7ddefc909e7c1ea25af560e2b9a36da156db25cc2701e4a6f34d753a948" exitCode=0 Mar 13 15:21:50 crc kubenswrapper[4907]: I0313 15:21:50.040284 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qtts8" event={"ID":"8ce4bcd2-1af1-43b6-8e59-d470e9089686","Type":"ContainerDied","Data":"52eff7ddefc909e7c1ea25af560e2b9a36da156db25cc2701e4a6f34d753a948"} Mar 13 15:21:50 crc kubenswrapper[4907]: I0313 15:21:50.040320 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qtts8" event={"ID":"8ce4bcd2-1af1-43b6-8e59-d470e9089686","Type":"ContainerStarted","Data":"6986ac91f0a1405d74e10dd104ecf7a3a66af918e31112a29fb3477814a7cb12"} Mar 13 15:21:51 crc kubenswrapper[4907]: I0313 15:21:51.050141 4907 generic.go:334] "Generic (PLEG): container finished" podID="8ce4bcd2-1af1-43b6-8e59-d470e9089686" containerID="7f4c959d1547bf070df1ec07daaac52181e0605870a8c576d5c25db49214bf5a" exitCode=0 Mar 13 15:21:51 crc kubenswrapper[4907]: I0313 15:21:51.050297 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qtts8" event={"ID":"8ce4bcd2-1af1-43b6-8e59-d470e9089686","Type":"ContainerDied","Data":"7f4c959d1547bf070df1ec07daaac52181e0605870a8c576d5c25db49214bf5a"} Mar 13 15:21:52 crc kubenswrapper[4907]: I0313 15:21:52.058388 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qtts8" event={"ID":"8ce4bcd2-1af1-43b6-8e59-d470e9089686","Type":"ContainerStarted","Data":"011e7e094e32e738642f126d721a01c4dd2932844c6b73664a0cb1d95cb88f04"} Mar 13 15:21:52 crc kubenswrapper[4907]: I0313 15:21:52.081105 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-qtts8" podStartSLOduration=2.659200975 podStartE2EDuration="4.081085389s" podCreationTimestamp="2026-03-13 15:21:48 +0000 UTC" firstStartedPulling="2026-03-13 15:21:50.042803523 +0000 UTC m=+4608.942591212" lastFinishedPulling="2026-03-13 15:21:51.464687937 +0000 UTC m=+4610.364475626" observedRunningTime="2026-03-13 15:21:52.075749643 +0000 UTC m=+4610.975537342" watchObservedRunningTime="2026-03-13 15:21:52.081085389 +0000 UTC m=+4610.980873078" Mar 13 15:21:59 crc kubenswrapper[4907]: I0313 15:21:59.064801 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-qtts8" Mar 13 15:21:59 crc kubenswrapper[4907]: I0313 15:21:59.065302 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-qtts8" Mar 13 15:21:59 crc kubenswrapper[4907]: I0313 15:21:59.108649 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-qtts8" Mar 13 15:21:59 crc kubenswrapper[4907]: I0313 15:21:59.152116 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-qtts8" Mar 13 15:21:59 crc kubenswrapper[4907]: I0313 15:21:59.343276 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-qtts8"] Mar 13 15:22:00 crc kubenswrapper[4907]: I0313 15:22:00.136830 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556922-b7qtk"] Mar 13 15:22:00 crc kubenswrapper[4907]: I0313 15:22:00.138130 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556922-b7qtk" Mar 13 15:22:00 crc kubenswrapper[4907]: I0313 15:22:00.143512 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 15:22:00 crc kubenswrapper[4907]: I0313 15:22:00.143555 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 15:22:00 crc kubenswrapper[4907]: I0313 15:22:00.143571 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 15:22:00 crc kubenswrapper[4907]: I0313 15:22:00.145162 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556922-b7qtk"] Mar 13 15:22:00 crc kubenswrapper[4907]: I0313 15:22:00.161264 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9kjk8\" (UniqueName: \"kubernetes.io/projected/31cb9010-2aab-475d-aab3-9aa35a538581-kube-api-access-9kjk8\") pod \"auto-csr-approver-29556922-b7qtk\" (UID: \"31cb9010-2aab-475d-aab3-9aa35a538581\") " pod="openshift-infra/auto-csr-approver-29556922-b7qtk" Mar 13 15:22:00 crc kubenswrapper[4907]: I0313 15:22:00.263088 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9kjk8\" (UniqueName: \"kubernetes.io/projected/31cb9010-2aab-475d-aab3-9aa35a538581-kube-api-access-9kjk8\") pod \"auto-csr-approver-29556922-b7qtk\" (UID: \"31cb9010-2aab-475d-aab3-9aa35a538581\") " pod="openshift-infra/auto-csr-approver-29556922-b7qtk" Mar 13 15:22:00 crc kubenswrapper[4907]: I0313 15:22:00.283936 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9kjk8\" (UniqueName: \"kubernetes.io/projected/31cb9010-2aab-475d-aab3-9aa35a538581-kube-api-access-9kjk8\") pod \"auto-csr-approver-29556922-b7qtk\" (UID: \"31cb9010-2aab-475d-aab3-9aa35a538581\") " pod="openshift-infra/auto-csr-approver-29556922-b7qtk" Mar 13 15:22:00 crc kubenswrapper[4907]: I0313 15:22:00.458839 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556922-b7qtk" Mar 13 15:22:00 crc kubenswrapper[4907]: I0313 15:22:00.783180 4907 scope.go:117] "RemoveContainer" containerID="dd934d7c5fc7e95362f79c2301ff40920ce7e3112722cbc11a053a56c085e05c" Mar 13 15:22:00 crc kubenswrapper[4907]: E0313 15:22:00.783777 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:22:00 crc kubenswrapper[4907]: I0313 15:22:00.879917 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556922-b7qtk"] Mar 13 15:22:01 crc kubenswrapper[4907]: I0313 15:22:01.121222 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556922-b7qtk" event={"ID":"31cb9010-2aab-475d-aab3-9aa35a538581","Type":"ContainerStarted","Data":"0e6b3e4cee3836f04db13b82a09d26c6bd36e7d4ef4e9ebe75b7a51fdd3a4b97"} Mar 13 15:22:01 crc kubenswrapper[4907]: I0313 15:22:01.121388 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-qtts8" podUID="8ce4bcd2-1af1-43b6-8e59-d470e9089686" containerName="registry-server" containerID="cri-o://011e7e094e32e738642f126d721a01c4dd2932844c6b73664a0cb1d95cb88f04" gracePeriod=2 Mar 13 15:22:01 crc kubenswrapper[4907]: I0313 15:22:01.659362 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qtts8" Mar 13 15:22:01 crc kubenswrapper[4907]: I0313 15:22:01.684827 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ce4bcd2-1af1-43b6-8e59-d470e9089686-catalog-content\") pod \"8ce4bcd2-1af1-43b6-8e59-d470e9089686\" (UID: \"8ce4bcd2-1af1-43b6-8e59-d470e9089686\") " Mar 13 15:22:01 crc kubenswrapper[4907]: I0313 15:22:01.684885 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-56pfq\" (UniqueName: \"kubernetes.io/projected/8ce4bcd2-1af1-43b6-8e59-d470e9089686-kube-api-access-56pfq\") pod \"8ce4bcd2-1af1-43b6-8e59-d470e9089686\" (UID: \"8ce4bcd2-1af1-43b6-8e59-d470e9089686\") " Mar 13 15:22:01 crc kubenswrapper[4907]: I0313 15:22:01.685022 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ce4bcd2-1af1-43b6-8e59-d470e9089686-utilities\") pod \"8ce4bcd2-1af1-43b6-8e59-d470e9089686\" (UID: \"8ce4bcd2-1af1-43b6-8e59-d470e9089686\") " Mar 13 15:22:01 crc kubenswrapper[4907]: I0313 15:22:01.685982 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8ce4bcd2-1af1-43b6-8e59-d470e9089686-utilities" (OuterVolumeSpecName: "utilities") pod "8ce4bcd2-1af1-43b6-8e59-d470e9089686" (UID: "8ce4bcd2-1af1-43b6-8e59-d470e9089686"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:22:01 crc kubenswrapper[4907]: I0313 15:22:01.691592 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8ce4bcd2-1af1-43b6-8e59-d470e9089686-kube-api-access-56pfq" (OuterVolumeSpecName: "kube-api-access-56pfq") pod "8ce4bcd2-1af1-43b6-8e59-d470e9089686" (UID: "8ce4bcd2-1af1-43b6-8e59-d470e9089686"). InnerVolumeSpecName "kube-api-access-56pfq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:22:01 crc kubenswrapper[4907]: I0313 15:22:01.714617 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8ce4bcd2-1af1-43b6-8e59-d470e9089686-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8ce4bcd2-1af1-43b6-8e59-d470e9089686" (UID: "8ce4bcd2-1af1-43b6-8e59-d470e9089686"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:22:01 crc kubenswrapper[4907]: I0313 15:22:01.786118 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ce4bcd2-1af1-43b6-8e59-d470e9089686-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 15:22:01 crc kubenswrapper[4907]: I0313 15:22:01.786160 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ce4bcd2-1af1-43b6-8e59-d470e9089686-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 15:22:01 crc kubenswrapper[4907]: I0313 15:22:01.786178 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-56pfq\" (UniqueName: \"kubernetes.io/projected/8ce4bcd2-1af1-43b6-8e59-d470e9089686-kube-api-access-56pfq\") on node \"crc\" DevicePath \"\"" Mar 13 15:22:02 crc kubenswrapper[4907]: I0313 15:22:02.135141 4907 generic.go:334] "Generic (PLEG): container finished" podID="8ce4bcd2-1af1-43b6-8e59-d470e9089686" containerID="011e7e094e32e738642f126d721a01c4dd2932844c6b73664a0cb1d95cb88f04" exitCode=0 Mar 13 15:22:02 crc kubenswrapper[4907]: I0313 15:22:02.135220 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qtts8" event={"ID":"8ce4bcd2-1af1-43b6-8e59-d470e9089686","Type":"ContainerDied","Data":"011e7e094e32e738642f126d721a01c4dd2932844c6b73664a0cb1d95cb88f04"} Mar 13 15:22:02 crc kubenswrapper[4907]: I0313 15:22:02.135703 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qtts8" event={"ID":"8ce4bcd2-1af1-43b6-8e59-d470e9089686","Type":"ContainerDied","Data":"6986ac91f0a1405d74e10dd104ecf7a3a66af918e31112a29fb3477814a7cb12"} Mar 13 15:22:02 crc kubenswrapper[4907]: I0313 15:22:02.135728 4907 scope.go:117] "RemoveContainer" containerID="011e7e094e32e738642f126d721a01c4dd2932844c6b73664a0cb1d95cb88f04" Mar 13 15:22:02 crc kubenswrapper[4907]: I0313 15:22:02.135245 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qtts8" Mar 13 15:22:02 crc kubenswrapper[4907]: I0313 15:22:02.141946 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556922-b7qtk" event={"ID":"31cb9010-2aab-475d-aab3-9aa35a538581","Type":"ContainerStarted","Data":"2cd2416cb09c17cb001811639f27834ed34890771029fd10f4c0a96794fa7519"} Mar 13 15:22:02 crc kubenswrapper[4907]: I0313 15:22:02.156784 4907 scope.go:117] "RemoveContainer" containerID="7f4c959d1547bf070df1ec07daaac52181e0605870a8c576d5c25db49214bf5a" Mar 13 15:22:02 crc kubenswrapper[4907]: I0313 15:22:02.158680 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-qtts8"] Mar 13 15:22:02 crc kubenswrapper[4907]: I0313 15:22:02.164769 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-qtts8"] Mar 13 15:22:02 crc kubenswrapper[4907]: I0313 15:22:02.174177 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29556922-b7qtk" podStartSLOduration=1.223280165 podStartE2EDuration="2.174157671s" podCreationTimestamp="2026-03-13 15:22:00 +0000 UTC" firstStartedPulling="2026-03-13 15:22:00.903750218 +0000 UTC m=+4619.803537907" lastFinishedPulling="2026-03-13 15:22:01.854627724 +0000 UTC m=+4620.754415413" observedRunningTime="2026-03-13 15:22:02.169751791 +0000 UTC m=+4621.069539480" watchObservedRunningTime="2026-03-13 15:22:02.174157671 +0000 UTC m=+4621.073945360" Mar 13 15:22:02 crc kubenswrapper[4907]: I0313 15:22:02.182978 4907 scope.go:117] "RemoveContainer" containerID="52eff7ddefc909e7c1ea25af560e2b9a36da156db25cc2701e4a6f34d753a948" Mar 13 15:22:02 crc kubenswrapper[4907]: I0313 15:22:02.206719 4907 scope.go:117] "RemoveContainer" containerID="011e7e094e32e738642f126d721a01c4dd2932844c6b73664a0cb1d95cb88f04" Mar 13 15:22:02 crc kubenswrapper[4907]: E0313 15:22:02.207090 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"011e7e094e32e738642f126d721a01c4dd2932844c6b73664a0cb1d95cb88f04\": container with ID starting with 011e7e094e32e738642f126d721a01c4dd2932844c6b73664a0cb1d95cb88f04 not found: ID does not exist" containerID="011e7e094e32e738642f126d721a01c4dd2932844c6b73664a0cb1d95cb88f04" Mar 13 15:22:02 crc kubenswrapper[4907]: I0313 15:22:02.207117 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"011e7e094e32e738642f126d721a01c4dd2932844c6b73664a0cb1d95cb88f04"} err="failed to get container status \"011e7e094e32e738642f126d721a01c4dd2932844c6b73664a0cb1d95cb88f04\": rpc error: code = NotFound desc = could not find container \"011e7e094e32e738642f126d721a01c4dd2932844c6b73664a0cb1d95cb88f04\": container with ID starting with 011e7e094e32e738642f126d721a01c4dd2932844c6b73664a0cb1d95cb88f04 not found: ID does not exist" Mar 13 15:22:02 crc kubenswrapper[4907]: I0313 15:22:02.207136 4907 scope.go:117] "RemoveContainer" containerID="7f4c959d1547bf070df1ec07daaac52181e0605870a8c576d5c25db49214bf5a" Mar 13 15:22:02 crc kubenswrapper[4907]: E0313 15:22:02.207309 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7f4c959d1547bf070df1ec07daaac52181e0605870a8c576d5c25db49214bf5a\": container with ID starting with 7f4c959d1547bf070df1ec07daaac52181e0605870a8c576d5c25db49214bf5a not found: ID does not exist" containerID="7f4c959d1547bf070df1ec07daaac52181e0605870a8c576d5c25db49214bf5a" Mar 13 15:22:02 crc kubenswrapper[4907]: I0313 15:22:02.207329 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7f4c959d1547bf070df1ec07daaac52181e0605870a8c576d5c25db49214bf5a"} err="failed to get container status \"7f4c959d1547bf070df1ec07daaac52181e0605870a8c576d5c25db49214bf5a\": rpc error: code = NotFound desc = could not find container \"7f4c959d1547bf070df1ec07daaac52181e0605870a8c576d5c25db49214bf5a\": container with ID starting with 7f4c959d1547bf070df1ec07daaac52181e0605870a8c576d5c25db49214bf5a not found: ID does not exist" Mar 13 15:22:02 crc kubenswrapper[4907]: I0313 15:22:02.207357 4907 scope.go:117] "RemoveContainer" containerID="52eff7ddefc909e7c1ea25af560e2b9a36da156db25cc2701e4a6f34d753a948" Mar 13 15:22:02 crc kubenswrapper[4907]: E0313 15:22:02.207570 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"52eff7ddefc909e7c1ea25af560e2b9a36da156db25cc2701e4a6f34d753a948\": container with ID starting with 52eff7ddefc909e7c1ea25af560e2b9a36da156db25cc2701e4a6f34d753a948 not found: ID does not exist" containerID="52eff7ddefc909e7c1ea25af560e2b9a36da156db25cc2701e4a6f34d753a948" Mar 13 15:22:02 crc kubenswrapper[4907]: I0313 15:22:02.207588 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"52eff7ddefc909e7c1ea25af560e2b9a36da156db25cc2701e4a6f34d753a948"} err="failed to get container status \"52eff7ddefc909e7c1ea25af560e2b9a36da156db25cc2701e4a6f34d753a948\": rpc error: code = NotFound desc = could not find container \"52eff7ddefc909e7c1ea25af560e2b9a36da156db25cc2701e4a6f34d753a948\": container with ID starting with 52eff7ddefc909e7c1ea25af560e2b9a36da156db25cc2701e4a6f34d753a948 not found: ID does not exist" Mar 13 15:22:03 crc kubenswrapper[4907]: I0313 15:22:03.151064 4907 generic.go:334] "Generic (PLEG): container finished" podID="31cb9010-2aab-475d-aab3-9aa35a538581" containerID="2cd2416cb09c17cb001811639f27834ed34890771029fd10f4c0a96794fa7519" exitCode=0 Mar 13 15:22:03 crc kubenswrapper[4907]: I0313 15:22:03.151192 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556922-b7qtk" event={"ID":"31cb9010-2aab-475d-aab3-9aa35a538581","Type":"ContainerDied","Data":"2cd2416cb09c17cb001811639f27834ed34890771029fd10f4c0a96794fa7519"} Mar 13 15:22:03 crc kubenswrapper[4907]: E0313 15:22:03.599595 4907 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8ce4bcd2_1af1_43b6_8e59_d470e9089686.slice/crio-6986ac91f0a1405d74e10dd104ecf7a3a66af918e31112a29fb3477814a7cb12\": RecentStats: unable to find data in memory cache]" Mar 13 15:22:03 crc kubenswrapper[4907]: I0313 15:22:03.791056 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8ce4bcd2-1af1-43b6-8e59-d470e9089686" path="/var/lib/kubelet/pods/8ce4bcd2-1af1-43b6-8e59-d470e9089686/volumes" Mar 13 15:22:04 crc kubenswrapper[4907]: I0313 15:22:04.466382 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556922-b7qtk" Mar 13 15:22:04 crc kubenswrapper[4907]: I0313 15:22:04.629354 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9kjk8\" (UniqueName: \"kubernetes.io/projected/31cb9010-2aab-475d-aab3-9aa35a538581-kube-api-access-9kjk8\") pod \"31cb9010-2aab-475d-aab3-9aa35a538581\" (UID: \"31cb9010-2aab-475d-aab3-9aa35a538581\") " Mar 13 15:22:04 crc kubenswrapper[4907]: I0313 15:22:04.635076 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31cb9010-2aab-475d-aab3-9aa35a538581-kube-api-access-9kjk8" (OuterVolumeSpecName: "kube-api-access-9kjk8") pod "31cb9010-2aab-475d-aab3-9aa35a538581" (UID: "31cb9010-2aab-475d-aab3-9aa35a538581"). InnerVolumeSpecName "kube-api-access-9kjk8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:22:04 crc kubenswrapper[4907]: I0313 15:22:04.731422 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9kjk8\" (UniqueName: \"kubernetes.io/projected/31cb9010-2aab-475d-aab3-9aa35a538581-kube-api-access-9kjk8\") on node \"crc\" DevicePath \"\"" Mar 13 15:22:04 crc kubenswrapper[4907]: I0313 15:22:04.878241 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556916-gpp98"] Mar 13 15:22:04 crc kubenswrapper[4907]: I0313 15:22:04.886806 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556916-gpp98"] Mar 13 15:22:05 crc kubenswrapper[4907]: I0313 15:22:05.166114 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556922-b7qtk" event={"ID":"31cb9010-2aab-475d-aab3-9aa35a538581","Type":"ContainerDied","Data":"0e6b3e4cee3836f04db13b82a09d26c6bd36e7d4ef4e9ebe75b7a51fdd3a4b97"} Mar 13 15:22:05 crc kubenswrapper[4907]: I0313 15:22:05.166153 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0e6b3e4cee3836f04db13b82a09d26c6bd36e7d4ef4e9ebe75b7a51fdd3a4b97" Mar 13 15:22:05 crc kubenswrapper[4907]: I0313 15:22:05.166186 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556922-b7qtk" Mar 13 15:22:05 crc kubenswrapper[4907]: I0313 15:22:05.791392 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c5c630da-de73-4f29-89f0-43c5497dd89a" path="/var/lib/kubelet/pods/c5c630da-de73-4f29-89f0-43c5497dd89a/volumes" Mar 13 15:22:13 crc kubenswrapper[4907]: E0313 15:22:13.761836 4907 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8ce4bcd2_1af1_43b6_8e59_d470e9089686.slice/crio-6986ac91f0a1405d74e10dd104ecf7a3a66af918e31112a29fb3477814a7cb12\": RecentStats: unable to find data in memory cache]" Mar 13 15:22:14 crc kubenswrapper[4907]: I0313 15:22:14.782962 4907 scope.go:117] "RemoveContainer" containerID="dd934d7c5fc7e95362f79c2301ff40920ce7e3112722cbc11a053a56c085e05c" Mar 13 15:22:14 crc kubenswrapper[4907]: E0313 15:22:14.783476 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:22:21 crc kubenswrapper[4907]: I0313 15:22:21.672191 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["crc-storage/crc-storage-crc-wds2z"] Mar 13 15:22:21 crc kubenswrapper[4907]: I0313 15:22:21.680209 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["crc-storage/crc-storage-crc-wds2z"] Mar 13 15:22:21 crc kubenswrapper[4907]: I0313 15:22:21.792116 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b501aea0-66d1-4655-b67f-c6a3bacb7cdb" path="/var/lib/kubelet/pods/b501aea0-66d1-4655-b67f-c6a3bacb7cdb/volumes" Mar 13 15:22:21 crc kubenswrapper[4907]: I0313 15:22:21.821743 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["crc-storage/crc-storage-crc-g69ht"] Mar 13 15:22:21 crc kubenswrapper[4907]: E0313 15:22:21.822081 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31cb9010-2aab-475d-aab3-9aa35a538581" containerName="oc" Mar 13 15:22:21 crc kubenswrapper[4907]: I0313 15:22:21.822096 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="31cb9010-2aab-475d-aab3-9aa35a538581" containerName="oc" Mar 13 15:22:21 crc kubenswrapper[4907]: E0313 15:22:21.822111 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ce4bcd2-1af1-43b6-8e59-d470e9089686" containerName="registry-server" Mar 13 15:22:21 crc kubenswrapper[4907]: I0313 15:22:21.822120 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ce4bcd2-1af1-43b6-8e59-d470e9089686" containerName="registry-server" Mar 13 15:22:21 crc kubenswrapper[4907]: E0313 15:22:21.822134 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ce4bcd2-1af1-43b6-8e59-d470e9089686" containerName="extract-content" Mar 13 15:22:21 crc kubenswrapper[4907]: I0313 15:22:21.822141 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ce4bcd2-1af1-43b6-8e59-d470e9089686" containerName="extract-content" Mar 13 15:22:21 crc kubenswrapper[4907]: E0313 15:22:21.822149 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ce4bcd2-1af1-43b6-8e59-d470e9089686" containerName="extract-utilities" Mar 13 15:22:21 crc kubenswrapper[4907]: I0313 15:22:21.822157 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ce4bcd2-1af1-43b6-8e59-d470e9089686" containerName="extract-utilities" Mar 13 15:22:21 crc kubenswrapper[4907]: I0313 15:22:21.822339 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="31cb9010-2aab-475d-aab3-9aa35a538581" containerName="oc" Mar 13 15:22:21 crc kubenswrapper[4907]: I0313 15:22:21.822353 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ce4bcd2-1af1-43b6-8e59-d470e9089686" containerName="registry-server" Mar 13 15:22:21 crc kubenswrapper[4907]: I0313 15:22:21.822896 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-g69ht" Mar 13 15:22:21 crc kubenswrapper[4907]: I0313 15:22:21.827796 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"openshift-service-ca.crt" Mar 13 15:22:21 crc kubenswrapper[4907]: I0313 15:22:21.827931 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"kube-root-ca.crt" Mar 13 15:22:21 crc kubenswrapper[4907]: I0313 15:22:21.828135 4907 reflector.go:368] Caches populated for *v1.Secret from object-"crc-storage"/"crc-storage-dockercfg-2d28j" Mar 13 15:22:21 crc kubenswrapper[4907]: I0313 15:22:21.828753 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"crc-storage" Mar 13 15:22:21 crc kubenswrapper[4907]: I0313 15:22:21.832173 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-g69ht"] Mar 13 15:22:21 crc kubenswrapper[4907]: I0313 15:22:21.981754 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/9dbd65f5-4738-4d22-a8da-5be0af6211fa-crc-storage\") pod \"crc-storage-crc-g69ht\" (UID: \"9dbd65f5-4738-4d22-a8da-5be0af6211fa\") " pod="crc-storage/crc-storage-crc-g69ht" Mar 13 15:22:21 crc kubenswrapper[4907]: I0313 15:22:21.981805 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/9dbd65f5-4738-4d22-a8da-5be0af6211fa-node-mnt\") pod \"crc-storage-crc-g69ht\" (UID: \"9dbd65f5-4738-4d22-a8da-5be0af6211fa\") " pod="crc-storage/crc-storage-crc-g69ht" Mar 13 15:22:21 crc kubenswrapper[4907]: I0313 15:22:21.981853 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-km7dg\" (UniqueName: \"kubernetes.io/projected/9dbd65f5-4738-4d22-a8da-5be0af6211fa-kube-api-access-km7dg\") pod \"crc-storage-crc-g69ht\" (UID: \"9dbd65f5-4738-4d22-a8da-5be0af6211fa\") " pod="crc-storage/crc-storage-crc-g69ht" Mar 13 15:22:22 crc kubenswrapper[4907]: I0313 15:22:22.083437 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/9dbd65f5-4738-4d22-a8da-5be0af6211fa-crc-storage\") pod \"crc-storage-crc-g69ht\" (UID: \"9dbd65f5-4738-4d22-a8da-5be0af6211fa\") " pod="crc-storage/crc-storage-crc-g69ht" Mar 13 15:22:22 crc kubenswrapper[4907]: I0313 15:22:22.083495 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/9dbd65f5-4738-4d22-a8da-5be0af6211fa-node-mnt\") pod \"crc-storage-crc-g69ht\" (UID: \"9dbd65f5-4738-4d22-a8da-5be0af6211fa\") " pod="crc-storage/crc-storage-crc-g69ht" Mar 13 15:22:22 crc kubenswrapper[4907]: I0313 15:22:22.083541 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-km7dg\" (UniqueName: \"kubernetes.io/projected/9dbd65f5-4738-4d22-a8da-5be0af6211fa-kube-api-access-km7dg\") pod \"crc-storage-crc-g69ht\" (UID: \"9dbd65f5-4738-4d22-a8da-5be0af6211fa\") " pod="crc-storage/crc-storage-crc-g69ht" Mar 13 15:22:22 crc kubenswrapper[4907]: I0313 15:22:22.084624 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/9dbd65f5-4738-4d22-a8da-5be0af6211fa-crc-storage\") pod \"crc-storage-crc-g69ht\" (UID: \"9dbd65f5-4738-4d22-a8da-5be0af6211fa\") " pod="crc-storage/crc-storage-crc-g69ht" Mar 13 15:22:22 crc kubenswrapper[4907]: I0313 15:22:22.084800 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/9dbd65f5-4738-4d22-a8da-5be0af6211fa-node-mnt\") pod \"crc-storage-crc-g69ht\" (UID: \"9dbd65f5-4738-4d22-a8da-5be0af6211fa\") " pod="crc-storage/crc-storage-crc-g69ht" Mar 13 15:22:22 crc kubenswrapper[4907]: I0313 15:22:22.103191 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-km7dg\" (UniqueName: \"kubernetes.io/projected/9dbd65f5-4738-4d22-a8da-5be0af6211fa-kube-api-access-km7dg\") pod \"crc-storage-crc-g69ht\" (UID: \"9dbd65f5-4738-4d22-a8da-5be0af6211fa\") " pod="crc-storage/crc-storage-crc-g69ht" Mar 13 15:22:22 crc kubenswrapper[4907]: I0313 15:22:22.143569 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-g69ht" Mar 13 15:22:22 crc kubenswrapper[4907]: I0313 15:22:22.624869 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-g69ht"] Mar 13 15:22:23 crc kubenswrapper[4907]: I0313 15:22:23.289756 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-g69ht" event={"ID":"9dbd65f5-4738-4d22-a8da-5be0af6211fa","Type":"ContainerStarted","Data":"71e8f6f74869e988544440e1e55a81639417cc910f590802c295e1f053265c1e"} Mar 13 15:22:23 crc kubenswrapper[4907]: E0313 15:22:23.940611 4907 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8ce4bcd2_1af1_43b6_8e59_d470e9089686.slice/crio-6986ac91f0a1405d74e10dd104ecf7a3a66af918e31112a29fb3477814a7cb12\": RecentStats: unable to find data in memory cache]" Mar 13 15:22:24 crc kubenswrapper[4907]: I0313 15:22:24.303224 4907 generic.go:334] "Generic (PLEG): container finished" podID="9dbd65f5-4738-4d22-a8da-5be0af6211fa" containerID="d45e59416766d20c4f95b995286178dd988b4b6257768b1fef8c82dcc281cd6c" exitCode=0 Mar 13 15:22:24 crc kubenswrapper[4907]: I0313 15:22:24.303300 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-g69ht" event={"ID":"9dbd65f5-4738-4d22-a8da-5be0af6211fa","Type":"ContainerDied","Data":"d45e59416766d20c4f95b995286178dd988b4b6257768b1fef8c82dcc281cd6c"} Mar 13 15:22:25 crc kubenswrapper[4907]: I0313 15:22:25.650746 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-g69ht" Mar 13 15:22:25 crc kubenswrapper[4907]: I0313 15:22:25.838268 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-km7dg\" (UniqueName: \"kubernetes.io/projected/9dbd65f5-4738-4d22-a8da-5be0af6211fa-kube-api-access-km7dg\") pod \"9dbd65f5-4738-4d22-a8da-5be0af6211fa\" (UID: \"9dbd65f5-4738-4d22-a8da-5be0af6211fa\") " Mar 13 15:22:25 crc kubenswrapper[4907]: I0313 15:22:25.838459 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/9dbd65f5-4738-4d22-a8da-5be0af6211fa-node-mnt\") pod \"9dbd65f5-4738-4d22-a8da-5be0af6211fa\" (UID: \"9dbd65f5-4738-4d22-a8da-5be0af6211fa\") " Mar 13 15:22:25 crc kubenswrapper[4907]: I0313 15:22:25.838489 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/9dbd65f5-4738-4d22-a8da-5be0af6211fa-crc-storage\") pod \"9dbd65f5-4738-4d22-a8da-5be0af6211fa\" (UID: \"9dbd65f5-4738-4d22-a8da-5be0af6211fa\") " Mar 13 15:22:25 crc kubenswrapper[4907]: I0313 15:22:25.838819 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9dbd65f5-4738-4d22-a8da-5be0af6211fa-node-mnt" (OuterVolumeSpecName: "node-mnt") pod "9dbd65f5-4738-4d22-a8da-5be0af6211fa" (UID: "9dbd65f5-4738-4d22-a8da-5be0af6211fa"). InnerVolumeSpecName "node-mnt". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 15:22:25 crc kubenswrapper[4907]: I0313 15:22:25.844732 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9dbd65f5-4738-4d22-a8da-5be0af6211fa-kube-api-access-km7dg" (OuterVolumeSpecName: "kube-api-access-km7dg") pod "9dbd65f5-4738-4d22-a8da-5be0af6211fa" (UID: "9dbd65f5-4738-4d22-a8da-5be0af6211fa"). InnerVolumeSpecName "kube-api-access-km7dg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:22:25 crc kubenswrapper[4907]: I0313 15:22:25.877671 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9dbd65f5-4738-4d22-a8da-5be0af6211fa-crc-storage" (OuterVolumeSpecName: "crc-storage") pod "9dbd65f5-4738-4d22-a8da-5be0af6211fa" (UID: "9dbd65f5-4738-4d22-a8da-5be0af6211fa"). InnerVolumeSpecName "crc-storage". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:22:25 crc kubenswrapper[4907]: I0313 15:22:25.939896 4907 reconciler_common.go:293] "Volume detached for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/9dbd65f5-4738-4d22-a8da-5be0af6211fa-node-mnt\") on node \"crc\" DevicePath \"\"" Mar 13 15:22:25 crc kubenswrapper[4907]: I0313 15:22:25.939926 4907 reconciler_common.go:293] "Volume detached for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/9dbd65f5-4738-4d22-a8da-5be0af6211fa-crc-storage\") on node \"crc\" DevicePath \"\"" Mar 13 15:22:25 crc kubenswrapper[4907]: I0313 15:22:25.939936 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-km7dg\" (UniqueName: \"kubernetes.io/projected/9dbd65f5-4738-4d22-a8da-5be0af6211fa-kube-api-access-km7dg\") on node \"crc\" DevicePath \"\"" Mar 13 15:22:26 crc kubenswrapper[4907]: I0313 15:22:26.318541 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-g69ht" event={"ID":"9dbd65f5-4738-4d22-a8da-5be0af6211fa","Type":"ContainerDied","Data":"71e8f6f74869e988544440e1e55a81639417cc910f590802c295e1f053265c1e"} Mar 13 15:22:26 crc kubenswrapper[4907]: I0313 15:22:26.318582 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="71e8f6f74869e988544440e1e55a81639417cc910f590802c295e1f053265c1e" Mar 13 15:22:26 crc kubenswrapper[4907]: I0313 15:22:26.318608 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-g69ht" Mar 13 15:22:27 crc kubenswrapper[4907]: I0313 15:22:27.083213 4907 scope.go:117] "RemoveContainer" containerID="311a031e1ff4ef3b74d10161b82c176889b7b82821fc88f2a4edeca5e5657ee5" Mar 13 15:22:27 crc kubenswrapper[4907]: I0313 15:22:27.106503 4907 scope.go:117] "RemoveContainer" containerID="348c87a146da948df0e4139184e97c7692cc71aad3ec14fe9f015dc10660b3b8" Mar 13 15:22:27 crc kubenswrapper[4907]: I0313 15:22:27.782081 4907 scope.go:117] "RemoveContainer" containerID="dd934d7c5fc7e95362f79c2301ff40920ce7e3112722cbc11a053a56c085e05c" Mar 13 15:22:27 crc kubenswrapper[4907]: E0313 15:22:27.782435 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:22:27 crc kubenswrapper[4907]: I0313 15:22:27.986206 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["crc-storage/crc-storage-crc-g69ht"] Mar 13 15:22:27 crc kubenswrapper[4907]: I0313 15:22:27.993156 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["crc-storage/crc-storage-crc-g69ht"] Mar 13 15:22:28 crc kubenswrapper[4907]: I0313 15:22:28.140830 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["crc-storage/crc-storage-crc-vl976"] Mar 13 15:22:28 crc kubenswrapper[4907]: E0313 15:22:28.141225 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9dbd65f5-4738-4d22-a8da-5be0af6211fa" containerName="storage" Mar 13 15:22:28 crc kubenswrapper[4907]: I0313 15:22:28.141242 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="9dbd65f5-4738-4d22-a8da-5be0af6211fa" containerName="storage" Mar 13 15:22:28 crc kubenswrapper[4907]: I0313 15:22:28.141405 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="9dbd65f5-4738-4d22-a8da-5be0af6211fa" containerName="storage" Mar 13 15:22:28 crc kubenswrapper[4907]: I0313 15:22:28.141971 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-vl976" Mar 13 15:22:28 crc kubenswrapper[4907]: I0313 15:22:28.145785 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"openshift-service-ca.crt" Mar 13 15:22:28 crc kubenswrapper[4907]: I0313 15:22:28.146096 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"crc-storage" Mar 13 15:22:28 crc kubenswrapper[4907]: I0313 15:22:28.147073 4907 reflector.go:368] Caches populated for *v1.Secret from object-"crc-storage"/"crc-storage-dockercfg-2d28j" Mar 13 15:22:28 crc kubenswrapper[4907]: I0313 15:22:28.147134 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"kube-root-ca.crt" Mar 13 15:22:28 crc kubenswrapper[4907]: I0313 15:22:28.152641 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-vl976"] Mar 13 15:22:28 crc kubenswrapper[4907]: I0313 15:22:28.274137 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/b7d369a5-227b-42bb-9e19-e2cc88a42a24-crc-storage\") pod \"crc-storage-crc-vl976\" (UID: \"b7d369a5-227b-42bb-9e19-e2cc88a42a24\") " pod="crc-storage/crc-storage-crc-vl976" Mar 13 15:22:28 crc kubenswrapper[4907]: I0313 15:22:28.274212 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/b7d369a5-227b-42bb-9e19-e2cc88a42a24-node-mnt\") pod \"crc-storage-crc-vl976\" (UID: \"b7d369a5-227b-42bb-9e19-e2cc88a42a24\") " pod="crc-storage/crc-storage-crc-vl976" Mar 13 15:22:28 crc kubenswrapper[4907]: I0313 15:22:28.274248 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zl6d9\" (UniqueName: \"kubernetes.io/projected/b7d369a5-227b-42bb-9e19-e2cc88a42a24-kube-api-access-zl6d9\") pod \"crc-storage-crc-vl976\" (UID: \"b7d369a5-227b-42bb-9e19-e2cc88a42a24\") " pod="crc-storage/crc-storage-crc-vl976" Mar 13 15:22:28 crc kubenswrapper[4907]: I0313 15:22:28.375972 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/b7d369a5-227b-42bb-9e19-e2cc88a42a24-crc-storage\") pod \"crc-storage-crc-vl976\" (UID: \"b7d369a5-227b-42bb-9e19-e2cc88a42a24\") " pod="crc-storage/crc-storage-crc-vl976" Mar 13 15:22:28 crc kubenswrapper[4907]: I0313 15:22:28.376044 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/b7d369a5-227b-42bb-9e19-e2cc88a42a24-node-mnt\") pod \"crc-storage-crc-vl976\" (UID: \"b7d369a5-227b-42bb-9e19-e2cc88a42a24\") " pod="crc-storage/crc-storage-crc-vl976" Mar 13 15:22:28 crc kubenswrapper[4907]: I0313 15:22:28.376074 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zl6d9\" (UniqueName: \"kubernetes.io/projected/b7d369a5-227b-42bb-9e19-e2cc88a42a24-kube-api-access-zl6d9\") pod \"crc-storage-crc-vl976\" (UID: \"b7d369a5-227b-42bb-9e19-e2cc88a42a24\") " pod="crc-storage/crc-storage-crc-vl976" Mar 13 15:22:28 crc kubenswrapper[4907]: I0313 15:22:28.376631 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/b7d369a5-227b-42bb-9e19-e2cc88a42a24-node-mnt\") pod \"crc-storage-crc-vl976\" (UID: \"b7d369a5-227b-42bb-9e19-e2cc88a42a24\") " pod="crc-storage/crc-storage-crc-vl976" Mar 13 15:22:28 crc kubenswrapper[4907]: I0313 15:22:28.379752 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/b7d369a5-227b-42bb-9e19-e2cc88a42a24-crc-storage\") pod \"crc-storage-crc-vl976\" (UID: \"b7d369a5-227b-42bb-9e19-e2cc88a42a24\") " pod="crc-storage/crc-storage-crc-vl976" Mar 13 15:22:28 crc kubenswrapper[4907]: I0313 15:22:28.397932 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zl6d9\" (UniqueName: \"kubernetes.io/projected/b7d369a5-227b-42bb-9e19-e2cc88a42a24-kube-api-access-zl6d9\") pod \"crc-storage-crc-vl976\" (UID: \"b7d369a5-227b-42bb-9e19-e2cc88a42a24\") " pod="crc-storage/crc-storage-crc-vl976" Mar 13 15:22:28 crc kubenswrapper[4907]: I0313 15:22:28.459864 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-vl976" Mar 13 15:22:28 crc kubenswrapper[4907]: I0313 15:22:28.879788 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-vl976"] Mar 13 15:22:29 crc kubenswrapper[4907]: I0313 15:22:29.342128 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-vl976" event={"ID":"b7d369a5-227b-42bb-9e19-e2cc88a42a24","Type":"ContainerStarted","Data":"762a2cc05f7d23fe18e4297120fcdfa9ad522d3b313575d0337e171ffdddeab6"} Mar 13 15:22:29 crc kubenswrapper[4907]: I0313 15:22:29.791964 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9dbd65f5-4738-4d22-a8da-5be0af6211fa" path="/var/lib/kubelet/pods/9dbd65f5-4738-4d22-a8da-5be0af6211fa/volumes" Mar 13 15:22:30 crc kubenswrapper[4907]: E0313 15:22:30.281405 4907 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = parsing image configuration: Get \"https://cdn01.quay.io/quayio-production-s3/sha256/a6/a6f5c002ec832d6e6a25c08e60f361308d0e752c78eebf8774d041416f5e0be9?X-Amz-Algorithm=AWS4-HMAC-SHA256&X-Amz-Credential=AKIATAAF2YHTGR23ZTE6%2F20260313%2Fus-east-1%2Fs3%2Faws4_request&X-Amz-Date=20260313T152229Z&X-Amz-Expires=600&X-Amz-SignedHeaders=host&X-Amz-Signature=a401e230d4c2c642de4064ed340803135dac3ca9bbe99019dcef287f4718d3f1®ion=us-east-1&namespace=openstack-k8s-operators&username=openshift-release-dev+ocm_access_1b89217552bc42d1be3fb06a1aed001a&repo_name=bash&akamai_signature=exp=1773416249~hmac=84d0e01e00b0a4a6dec744898931c695f38a70686f6235dc6c63b84df4e4ef32\": remote error: tls: internal error" image="quay.io/openstack-k8s-operators/bash:latest" Mar 13 15:22:30 crc kubenswrapper[4907]: E0313 15:22:30.281600 4907 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:storage,Image:quay.io/openstack-k8s-operators/bash:latest,Command:[bash],Args:[/usr/local/bin/crc-storage.sh],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:PV_NUM,Value:12,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:crc-storage,ReadOnly:true,MountPath:/usr/local/bin/crc-storage.sh,SubPath:create-storage.sh,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:node-mnt,ReadOnly:false,MountPath:/mnt/nodeMnt,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-zl6d9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod crc-storage-crc-vl976_crc-storage(b7d369a5-227b-42bb-9e19-e2cc88a42a24): ErrImagePull: parsing image configuration: Get \"https://cdn01.quay.io/quayio-production-s3/sha256/a6/a6f5c002ec832d6e6a25c08e60f361308d0e752c78eebf8774d041416f5e0be9?X-Amz-Algorithm=AWS4-HMAC-SHA256&X-Amz-Credential=AKIATAAF2YHTGR23ZTE6%2F20260313%2Fus-east-1%2Fs3%2Faws4_request&X-Amz-Date=20260313T152229Z&X-Amz-Expires=600&X-Amz-SignedHeaders=host&X-Amz-Signature=a401e230d4c2c642de4064ed340803135dac3ca9bbe99019dcef287f4718d3f1®ion=us-east-1&namespace=openstack-k8s-operators&username=openshift-release-dev+ocm_access_1b89217552bc42d1be3fb06a1aed001a&repo_name=bash&akamai_signature=exp=1773416249~hmac=84d0e01e00b0a4a6dec744898931c695f38a70686f6235dc6c63b84df4e4ef32\": remote error: tls: internal error" logger="UnhandledError" Mar 13 15:22:30 crc kubenswrapper[4907]: E0313 15:22:30.283042 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"storage\" with ErrImagePull: \"parsing image configuration: Get \\\"https://cdn01.quay.io/quayio-production-s3/sha256/a6/a6f5c002ec832d6e6a25c08e60f361308d0e752c78eebf8774d041416f5e0be9?X-Amz-Algorithm=AWS4-HMAC-SHA256&X-Amz-Credential=AKIATAAF2YHTGR23ZTE6%2F20260313%2Fus-east-1%2Fs3%2Faws4_request&X-Amz-Date=20260313T152229Z&X-Amz-Expires=600&X-Amz-SignedHeaders=host&X-Amz-Signature=a401e230d4c2c642de4064ed340803135dac3ca9bbe99019dcef287f4718d3f1®ion=us-east-1&namespace=openstack-k8s-operators&username=openshift-release-dev+ocm_access_1b89217552bc42d1be3fb06a1aed001a&repo_name=bash&akamai_signature=exp=1773416249~hmac=84d0e01e00b0a4a6dec744898931c695f38a70686f6235dc6c63b84df4e4ef32\\\": remote error: tls: internal error\"" pod="crc-storage/crc-storage-crc-vl976" podUID="b7d369a5-227b-42bb-9e19-e2cc88a42a24" Mar 13 15:22:30 crc kubenswrapper[4907]: E0313 15:22:30.351807 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"storage\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/bash:latest\\\"\"" pod="crc-storage/crc-storage-crc-vl976" podUID="b7d369a5-227b-42bb-9e19-e2cc88a42a24" Mar 13 15:22:34 crc kubenswrapper[4907]: E0313 15:22:34.112033 4907 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8ce4bcd2_1af1_43b6_8e59_d470e9089686.slice/crio-6986ac91f0a1405d74e10dd104ecf7a3a66af918e31112a29fb3477814a7cb12\": RecentStats: unable to find data in memory cache]" Mar 13 15:22:39 crc kubenswrapper[4907]: I0313 15:22:39.782842 4907 scope.go:117] "RemoveContainer" containerID="dd934d7c5fc7e95362f79c2301ff40920ce7e3112722cbc11a053a56c085e05c" Mar 13 15:22:39 crc kubenswrapper[4907]: E0313 15:22:39.783574 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:22:44 crc kubenswrapper[4907]: E0313 15:22:44.304861 4907 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8ce4bcd2_1af1_43b6_8e59_d470e9089686.slice/crio-6986ac91f0a1405d74e10dd104ecf7a3a66af918e31112a29fb3477814a7cb12\": RecentStats: unable to find data in memory cache]" Mar 13 15:22:46 crc kubenswrapper[4907]: I0313 15:22:46.454331 4907 generic.go:334] "Generic (PLEG): container finished" podID="b7d369a5-227b-42bb-9e19-e2cc88a42a24" containerID="720dafffc2f4de439a0af5f54cec3d94c4aabe1ca444e2ead9104c7711dcd4fe" exitCode=0 Mar 13 15:22:46 crc kubenswrapper[4907]: I0313 15:22:46.454422 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-vl976" event={"ID":"b7d369a5-227b-42bb-9e19-e2cc88a42a24","Type":"ContainerDied","Data":"720dafffc2f4de439a0af5f54cec3d94c4aabe1ca444e2ead9104c7711dcd4fe"} Mar 13 15:22:47 crc kubenswrapper[4907]: I0313 15:22:47.814312 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-vl976" Mar 13 15:22:47 crc kubenswrapper[4907]: I0313 15:22:47.952304 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/b7d369a5-227b-42bb-9e19-e2cc88a42a24-crc-storage\") pod \"b7d369a5-227b-42bb-9e19-e2cc88a42a24\" (UID: \"b7d369a5-227b-42bb-9e19-e2cc88a42a24\") " Mar 13 15:22:47 crc kubenswrapper[4907]: I0313 15:22:47.952390 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zl6d9\" (UniqueName: \"kubernetes.io/projected/b7d369a5-227b-42bb-9e19-e2cc88a42a24-kube-api-access-zl6d9\") pod \"b7d369a5-227b-42bb-9e19-e2cc88a42a24\" (UID: \"b7d369a5-227b-42bb-9e19-e2cc88a42a24\") " Mar 13 15:22:47 crc kubenswrapper[4907]: I0313 15:22:47.952422 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/b7d369a5-227b-42bb-9e19-e2cc88a42a24-node-mnt\") pod \"b7d369a5-227b-42bb-9e19-e2cc88a42a24\" (UID: \"b7d369a5-227b-42bb-9e19-e2cc88a42a24\") " Mar 13 15:22:47 crc kubenswrapper[4907]: I0313 15:22:47.952649 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b7d369a5-227b-42bb-9e19-e2cc88a42a24-node-mnt" (OuterVolumeSpecName: "node-mnt") pod "b7d369a5-227b-42bb-9e19-e2cc88a42a24" (UID: "b7d369a5-227b-42bb-9e19-e2cc88a42a24"). InnerVolumeSpecName "node-mnt". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 15:22:47 crc kubenswrapper[4907]: I0313 15:22:47.957641 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b7d369a5-227b-42bb-9e19-e2cc88a42a24-kube-api-access-zl6d9" (OuterVolumeSpecName: "kube-api-access-zl6d9") pod "b7d369a5-227b-42bb-9e19-e2cc88a42a24" (UID: "b7d369a5-227b-42bb-9e19-e2cc88a42a24"). InnerVolumeSpecName "kube-api-access-zl6d9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:22:47 crc kubenswrapper[4907]: I0313 15:22:47.969902 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b7d369a5-227b-42bb-9e19-e2cc88a42a24-crc-storage" (OuterVolumeSpecName: "crc-storage") pod "b7d369a5-227b-42bb-9e19-e2cc88a42a24" (UID: "b7d369a5-227b-42bb-9e19-e2cc88a42a24"). InnerVolumeSpecName "crc-storage". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:22:48 crc kubenswrapper[4907]: I0313 15:22:48.053785 4907 reconciler_common.go:293] "Volume detached for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/b7d369a5-227b-42bb-9e19-e2cc88a42a24-crc-storage\") on node \"crc\" DevicePath \"\"" Mar 13 15:22:48 crc kubenswrapper[4907]: I0313 15:22:48.053832 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zl6d9\" (UniqueName: \"kubernetes.io/projected/b7d369a5-227b-42bb-9e19-e2cc88a42a24-kube-api-access-zl6d9\") on node \"crc\" DevicePath \"\"" Mar 13 15:22:48 crc kubenswrapper[4907]: I0313 15:22:48.053848 4907 reconciler_common.go:293] "Volume detached for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/b7d369a5-227b-42bb-9e19-e2cc88a42a24-node-mnt\") on node \"crc\" DevicePath \"\"" Mar 13 15:22:48 crc kubenswrapper[4907]: I0313 15:22:48.473108 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-vl976" event={"ID":"b7d369a5-227b-42bb-9e19-e2cc88a42a24","Type":"ContainerDied","Data":"762a2cc05f7d23fe18e4297120fcdfa9ad522d3b313575d0337e171ffdddeab6"} Mar 13 15:22:48 crc kubenswrapper[4907]: I0313 15:22:48.473146 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-vl976" Mar 13 15:22:48 crc kubenswrapper[4907]: I0313 15:22:48.473156 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="762a2cc05f7d23fe18e4297120fcdfa9ad522d3b313575d0337e171ffdddeab6" Mar 13 15:22:53 crc kubenswrapper[4907]: I0313 15:22:53.782916 4907 scope.go:117] "RemoveContainer" containerID="dd934d7c5fc7e95362f79c2301ff40920ce7e3112722cbc11a053a56c085e05c" Mar 13 15:22:53 crc kubenswrapper[4907]: E0313 15:22:53.783943 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:22:54 crc kubenswrapper[4907]: E0313 15:22:54.463821 4907 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8ce4bcd2_1af1_43b6_8e59_d470e9089686.slice/crio-6986ac91f0a1405d74e10dd104ecf7a3a66af918e31112a29fb3477814a7cb12\": RecentStats: unable to find data in memory cache]" Mar 13 15:23:01 crc kubenswrapper[4907]: E0313 15:23:01.811926 4907 fsHandler.go:119] failed to collect filesystem stats - rootDiskErr: could not stat "/var/lib/containers/storage/overlay/25607d4d87b5b5a88585c2cd1d2ea551307e53f784e3d597bd98d5b6a061095c/diff" to get inode usage: stat /var/lib/containers/storage/overlay/25607d4d87b5b5a88585c2cd1d2ea551307e53f784e3d597bd98d5b6a061095c/diff: no such file or directory, extraDiskErr: Mar 13 15:23:06 crc kubenswrapper[4907]: I0313 15:23:06.782670 4907 scope.go:117] "RemoveContainer" containerID="dd934d7c5fc7e95362f79c2301ff40920ce7e3112722cbc11a053a56c085e05c" Mar 13 15:23:06 crc kubenswrapper[4907]: E0313 15:23:06.783761 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:23:21 crc kubenswrapper[4907]: I0313 15:23:21.791721 4907 scope.go:117] "RemoveContainer" containerID="dd934d7c5fc7e95362f79c2301ff40920ce7e3112722cbc11a053a56c085e05c" Mar 13 15:23:22 crc kubenswrapper[4907]: I0313 15:23:22.719756 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerStarted","Data":"b5bb3cf2979988d507615c0a856eebc9a02c4d2c3275ceb2cee45dcc6b625cfd"} Mar 13 15:24:00 crc kubenswrapper[4907]: I0313 15:24:00.141937 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556924-8z5rs"] Mar 13 15:24:00 crc kubenswrapper[4907]: E0313 15:24:00.142854 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7d369a5-227b-42bb-9e19-e2cc88a42a24" containerName="storage" Mar 13 15:24:00 crc kubenswrapper[4907]: I0313 15:24:00.142871 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7d369a5-227b-42bb-9e19-e2cc88a42a24" containerName="storage" Mar 13 15:24:00 crc kubenswrapper[4907]: I0313 15:24:00.143089 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="b7d369a5-227b-42bb-9e19-e2cc88a42a24" containerName="storage" Mar 13 15:24:00 crc kubenswrapper[4907]: I0313 15:24:00.143662 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556924-8z5rs" Mar 13 15:24:00 crc kubenswrapper[4907]: I0313 15:24:00.145447 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 15:24:00 crc kubenswrapper[4907]: I0313 15:24:00.145663 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 15:24:00 crc kubenswrapper[4907]: I0313 15:24:00.149487 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 15:24:00 crc kubenswrapper[4907]: I0313 15:24:00.162517 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556924-8z5rs"] Mar 13 15:24:00 crc kubenswrapper[4907]: I0313 15:24:00.291294 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p2pm7\" (UniqueName: \"kubernetes.io/projected/11a50278-3da1-4a63-ac7e-1fcb4ae77f52-kube-api-access-p2pm7\") pod \"auto-csr-approver-29556924-8z5rs\" (UID: \"11a50278-3da1-4a63-ac7e-1fcb4ae77f52\") " pod="openshift-infra/auto-csr-approver-29556924-8z5rs" Mar 13 15:24:00 crc kubenswrapper[4907]: I0313 15:24:00.393062 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p2pm7\" (UniqueName: \"kubernetes.io/projected/11a50278-3da1-4a63-ac7e-1fcb4ae77f52-kube-api-access-p2pm7\") pod \"auto-csr-approver-29556924-8z5rs\" (UID: \"11a50278-3da1-4a63-ac7e-1fcb4ae77f52\") " pod="openshift-infra/auto-csr-approver-29556924-8z5rs" Mar 13 15:24:00 crc kubenswrapper[4907]: I0313 15:24:00.410541 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p2pm7\" (UniqueName: \"kubernetes.io/projected/11a50278-3da1-4a63-ac7e-1fcb4ae77f52-kube-api-access-p2pm7\") pod \"auto-csr-approver-29556924-8z5rs\" (UID: \"11a50278-3da1-4a63-ac7e-1fcb4ae77f52\") " pod="openshift-infra/auto-csr-approver-29556924-8z5rs" Mar 13 15:24:00 crc kubenswrapper[4907]: I0313 15:24:00.467164 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556924-8z5rs" Mar 13 15:24:00 crc kubenswrapper[4907]: I0313 15:24:00.874336 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556924-8z5rs"] Mar 13 15:24:00 crc kubenswrapper[4907]: I0313 15:24:00.970822 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556924-8z5rs" event={"ID":"11a50278-3da1-4a63-ac7e-1fcb4ae77f52","Type":"ContainerStarted","Data":"1ce5fdadd767eafa59d4051e216d3c016e27cf39dffbc0f38e628d186e7e2aef"} Mar 13 15:24:02 crc kubenswrapper[4907]: I0313 15:24:02.988404 4907 generic.go:334] "Generic (PLEG): container finished" podID="11a50278-3da1-4a63-ac7e-1fcb4ae77f52" containerID="16a638d449eaf9553071c678dcf92f038cd2ddde8f659099ade22f5e78a0fcae" exitCode=0 Mar 13 15:24:02 crc kubenswrapper[4907]: I0313 15:24:02.988475 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556924-8z5rs" event={"ID":"11a50278-3da1-4a63-ac7e-1fcb4ae77f52","Type":"ContainerDied","Data":"16a638d449eaf9553071c678dcf92f038cd2ddde8f659099ade22f5e78a0fcae"} Mar 13 15:24:04 crc kubenswrapper[4907]: I0313 15:24:04.350423 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556924-8z5rs" Mar 13 15:24:04 crc kubenswrapper[4907]: I0313 15:24:04.449384 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p2pm7\" (UniqueName: \"kubernetes.io/projected/11a50278-3da1-4a63-ac7e-1fcb4ae77f52-kube-api-access-p2pm7\") pod \"11a50278-3da1-4a63-ac7e-1fcb4ae77f52\" (UID: \"11a50278-3da1-4a63-ac7e-1fcb4ae77f52\") " Mar 13 15:24:04 crc kubenswrapper[4907]: I0313 15:24:04.454947 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/11a50278-3da1-4a63-ac7e-1fcb4ae77f52-kube-api-access-p2pm7" (OuterVolumeSpecName: "kube-api-access-p2pm7") pod "11a50278-3da1-4a63-ac7e-1fcb4ae77f52" (UID: "11a50278-3da1-4a63-ac7e-1fcb4ae77f52"). InnerVolumeSpecName "kube-api-access-p2pm7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:24:04 crc kubenswrapper[4907]: I0313 15:24:04.550406 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p2pm7\" (UniqueName: \"kubernetes.io/projected/11a50278-3da1-4a63-ac7e-1fcb4ae77f52-kube-api-access-p2pm7\") on node \"crc\" DevicePath \"\"" Mar 13 15:24:05 crc kubenswrapper[4907]: I0313 15:24:05.004558 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556924-8z5rs" event={"ID":"11a50278-3da1-4a63-ac7e-1fcb4ae77f52","Type":"ContainerDied","Data":"1ce5fdadd767eafa59d4051e216d3c016e27cf39dffbc0f38e628d186e7e2aef"} Mar 13 15:24:05 crc kubenswrapper[4907]: I0313 15:24:05.005006 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1ce5fdadd767eafa59d4051e216d3c016e27cf39dffbc0f38e628d186e7e2aef" Mar 13 15:24:05 crc kubenswrapper[4907]: I0313 15:24:05.004602 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556924-8z5rs" Mar 13 15:24:05 crc kubenswrapper[4907]: I0313 15:24:05.416348 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556918-cs4dh"] Mar 13 15:24:05 crc kubenswrapper[4907]: I0313 15:24:05.422081 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556918-cs4dh"] Mar 13 15:24:05 crc kubenswrapper[4907]: I0313 15:24:05.791334 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fdd030a7-3ebb-4fbf-91f2-2ab96e526252" path="/var/lib/kubelet/pods/fdd030a7-3ebb-4fbf-91f2-2ab96e526252/volumes" Mar 13 15:24:27 crc kubenswrapper[4907]: I0313 15:24:27.207304 4907 scope.go:117] "RemoveContainer" containerID="e984813306df1ab82de9a9dd04affc9e989fc2092eb575a81e8fc5794b033f1a" Mar 13 15:25:36 crc kubenswrapper[4907]: I0313 15:25:36.610047 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-skd22"] Mar 13 15:25:36 crc kubenswrapper[4907]: E0313 15:25:36.611011 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11a50278-3da1-4a63-ac7e-1fcb4ae77f52" containerName="oc" Mar 13 15:25:36 crc kubenswrapper[4907]: I0313 15:25:36.611028 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="11a50278-3da1-4a63-ac7e-1fcb4ae77f52" containerName="oc" Mar 13 15:25:36 crc kubenswrapper[4907]: I0313 15:25:36.611186 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="11a50278-3da1-4a63-ac7e-1fcb4ae77f52" containerName="oc" Mar 13 15:25:36 crc kubenswrapper[4907]: I0313 15:25:36.612392 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-skd22" Mar 13 15:25:36 crc kubenswrapper[4907]: I0313 15:25:36.622806 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-skd22"] Mar 13 15:25:36 crc kubenswrapper[4907]: I0313 15:25:36.766485 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z4r6m\" (UniqueName: \"kubernetes.io/projected/6454399b-51ab-41f1-b4db-bbe12b4b7c95-kube-api-access-z4r6m\") pod \"community-operators-skd22\" (UID: \"6454399b-51ab-41f1-b4db-bbe12b4b7c95\") " pod="openshift-marketplace/community-operators-skd22" Mar 13 15:25:36 crc kubenswrapper[4907]: I0313 15:25:36.766728 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6454399b-51ab-41f1-b4db-bbe12b4b7c95-utilities\") pod \"community-operators-skd22\" (UID: \"6454399b-51ab-41f1-b4db-bbe12b4b7c95\") " pod="openshift-marketplace/community-operators-skd22" Mar 13 15:25:36 crc kubenswrapper[4907]: I0313 15:25:36.766853 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6454399b-51ab-41f1-b4db-bbe12b4b7c95-catalog-content\") pod \"community-operators-skd22\" (UID: \"6454399b-51ab-41f1-b4db-bbe12b4b7c95\") " pod="openshift-marketplace/community-operators-skd22" Mar 13 15:25:36 crc kubenswrapper[4907]: I0313 15:25:36.867952 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6454399b-51ab-41f1-b4db-bbe12b4b7c95-catalog-content\") pod \"community-operators-skd22\" (UID: \"6454399b-51ab-41f1-b4db-bbe12b4b7c95\") " pod="openshift-marketplace/community-operators-skd22" Mar 13 15:25:36 crc kubenswrapper[4907]: I0313 15:25:36.868044 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z4r6m\" (UniqueName: \"kubernetes.io/projected/6454399b-51ab-41f1-b4db-bbe12b4b7c95-kube-api-access-z4r6m\") pod \"community-operators-skd22\" (UID: \"6454399b-51ab-41f1-b4db-bbe12b4b7c95\") " pod="openshift-marketplace/community-operators-skd22" Mar 13 15:25:36 crc kubenswrapper[4907]: I0313 15:25:36.868105 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6454399b-51ab-41f1-b4db-bbe12b4b7c95-utilities\") pod \"community-operators-skd22\" (UID: \"6454399b-51ab-41f1-b4db-bbe12b4b7c95\") " pod="openshift-marketplace/community-operators-skd22" Mar 13 15:25:36 crc kubenswrapper[4907]: I0313 15:25:36.868461 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6454399b-51ab-41f1-b4db-bbe12b4b7c95-catalog-content\") pod \"community-operators-skd22\" (UID: \"6454399b-51ab-41f1-b4db-bbe12b4b7c95\") " pod="openshift-marketplace/community-operators-skd22" Mar 13 15:25:36 crc kubenswrapper[4907]: I0313 15:25:36.868901 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6454399b-51ab-41f1-b4db-bbe12b4b7c95-utilities\") pod \"community-operators-skd22\" (UID: \"6454399b-51ab-41f1-b4db-bbe12b4b7c95\") " pod="openshift-marketplace/community-operators-skd22" Mar 13 15:25:36 crc kubenswrapper[4907]: I0313 15:25:36.891012 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z4r6m\" (UniqueName: \"kubernetes.io/projected/6454399b-51ab-41f1-b4db-bbe12b4b7c95-kube-api-access-z4r6m\") pod \"community-operators-skd22\" (UID: \"6454399b-51ab-41f1-b4db-bbe12b4b7c95\") " pod="openshift-marketplace/community-operators-skd22" Mar 13 15:25:36 crc kubenswrapper[4907]: I0313 15:25:36.933457 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-skd22" Mar 13 15:25:37 crc kubenswrapper[4907]: I0313 15:25:37.449297 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-skd22"] Mar 13 15:25:37 crc kubenswrapper[4907]: I0313 15:25:37.656586 4907 generic.go:334] "Generic (PLEG): container finished" podID="6454399b-51ab-41f1-b4db-bbe12b4b7c95" containerID="c1c98d52ad0aba6e93f5f3a8f2f40fa1e06cc70b71f70d32c0e3077df5f68e0d" exitCode=0 Mar 13 15:25:37 crc kubenswrapper[4907]: I0313 15:25:37.656629 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-skd22" event={"ID":"6454399b-51ab-41f1-b4db-bbe12b4b7c95","Type":"ContainerDied","Data":"c1c98d52ad0aba6e93f5f3a8f2f40fa1e06cc70b71f70d32c0e3077df5f68e0d"} Mar 13 15:25:37 crc kubenswrapper[4907]: I0313 15:25:37.656654 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-skd22" event={"ID":"6454399b-51ab-41f1-b4db-bbe12b4b7c95","Type":"ContainerStarted","Data":"4ff9f5b55970d7e99e658759b98269c42d65a135c1963e0a4852afb8d683eb2f"} Mar 13 15:25:37 crc kubenswrapper[4907]: I0313 15:25:37.659711 4907 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 13 15:25:39 crc kubenswrapper[4907]: I0313 15:25:39.677794 4907 generic.go:334] "Generic (PLEG): container finished" podID="6454399b-51ab-41f1-b4db-bbe12b4b7c95" containerID="40a57078c576bd33006fc990f6f567d9963ea1da66645391e9fcf035de5e33fa" exitCode=0 Mar 13 15:25:39 crc kubenswrapper[4907]: I0313 15:25:39.677850 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-skd22" event={"ID":"6454399b-51ab-41f1-b4db-bbe12b4b7c95","Type":"ContainerDied","Data":"40a57078c576bd33006fc990f6f567d9963ea1da66645391e9fcf035de5e33fa"} Mar 13 15:25:40 crc kubenswrapper[4907]: I0313 15:25:40.688502 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-skd22" event={"ID":"6454399b-51ab-41f1-b4db-bbe12b4b7c95","Type":"ContainerStarted","Data":"61ac5444abd6e1ac4c4de07ff227d8deea674a73498eeb60cb027d561df5df1f"} Mar 13 15:25:46 crc kubenswrapper[4907]: I0313 15:25:46.933951 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-skd22" Mar 13 15:25:46 crc kubenswrapper[4907]: I0313 15:25:46.934524 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-skd22" Mar 13 15:25:46 crc kubenswrapper[4907]: I0313 15:25:46.985258 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-skd22" Mar 13 15:25:47 crc kubenswrapper[4907]: I0313 15:25:47.008404 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-skd22" podStartSLOduration=8.609087121 podStartE2EDuration="11.008377126s" podCreationTimestamp="2026-03-13 15:25:36 +0000 UTC" firstStartedPulling="2026-03-13 15:25:37.659441848 +0000 UTC m=+4836.559229537" lastFinishedPulling="2026-03-13 15:25:40.058731853 +0000 UTC m=+4838.958519542" observedRunningTime="2026-03-13 15:25:40.718072509 +0000 UTC m=+4839.617860208" watchObservedRunningTime="2026-03-13 15:25:47.008377126 +0000 UTC m=+4845.908164825" Mar 13 15:25:48 crc kubenswrapper[4907]: I0313 15:25:48.041124 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-skd22" Mar 13 15:25:48 crc kubenswrapper[4907]: I0313 15:25:48.041323 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 15:25:48 crc kubenswrapper[4907]: I0313 15:25:48.041375 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 15:25:48 crc kubenswrapper[4907]: I0313 15:25:48.085672 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-skd22"] Mar 13 15:25:49 crc kubenswrapper[4907]: I0313 15:25:49.743860 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-skd22" podUID="6454399b-51ab-41f1-b4db-bbe12b4b7c95" containerName="registry-server" containerID="cri-o://61ac5444abd6e1ac4c4de07ff227d8deea674a73498eeb60cb027d561df5df1f" gracePeriod=2 Mar 13 15:25:50 crc kubenswrapper[4907]: I0313 15:25:50.182223 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-skd22" Mar 13 15:25:50 crc kubenswrapper[4907]: I0313 15:25:50.372439 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6454399b-51ab-41f1-b4db-bbe12b4b7c95-utilities\") pod \"6454399b-51ab-41f1-b4db-bbe12b4b7c95\" (UID: \"6454399b-51ab-41f1-b4db-bbe12b4b7c95\") " Mar 13 15:25:50 crc kubenswrapper[4907]: I0313 15:25:50.372606 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z4r6m\" (UniqueName: \"kubernetes.io/projected/6454399b-51ab-41f1-b4db-bbe12b4b7c95-kube-api-access-z4r6m\") pod \"6454399b-51ab-41f1-b4db-bbe12b4b7c95\" (UID: \"6454399b-51ab-41f1-b4db-bbe12b4b7c95\") " Mar 13 15:25:50 crc kubenswrapper[4907]: I0313 15:25:50.372638 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6454399b-51ab-41f1-b4db-bbe12b4b7c95-catalog-content\") pod \"6454399b-51ab-41f1-b4db-bbe12b4b7c95\" (UID: \"6454399b-51ab-41f1-b4db-bbe12b4b7c95\") " Mar 13 15:25:50 crc kubenswrapper[4907]: I0313 15:25:50.407133 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6454399b-51ab-41f1-b4db-bbe12b4b7c95-kube-api-access-z4r6m" (OuterVolumeSpecName: "kube-api-access-z4r6m") pod "6454399b-51ab-41f1-b4db-bbe12b4b7c95" (UID: "6454399b-51ab-41f1-b4db-bbe12b4b7c95"). InnerVolumeSpecName "kube-api-access-z4r6m". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:25:50 crc kubenswrapper[4907]: I0313 15:25:50.411209 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6454399b-51ab-41f1-b4db-bbe12b4b7c95-utilities" (OuterVolumeSpecName: "utilities") pod "6454399b-51ab-41f1-b4db-bbe12b4b7c95" (UID: "6454399b-51ab-41f1-b4db-bbe12b4b7c95"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:25:50 crc kubenswrapper[4907]: I0313 15:25:50.451150 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6454399b-51ab-41f1-b4db-bbe12b4b7c95-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6454399b-51ab-41f1-b4db-bbe12b4b7c95" (UID: "6454399b-51ab-41f1-b4db-bbe12b4b7c95"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:25:50 crc kubenswrapper[4907]: I0313 15:25:50.473913 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z4r6m\" (UniqueName: \"kubernetes.io/projected/6454399b-51ab-41f1-b4db-bbe12b4b7c95-kube-api-access-z4r6m\") on node \"crc\" DevicePath \"\"" Mar 13 15:25:50 crc kubenswrapper[4907]: I0313 15:25:50.473946 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6454399b-51ab-41f1-b4db-bbe12b4b7c95-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 15:25:50 crc kubenswrapper[4907]: I0313 15:25:50.473961 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6454399b-51ab-41f1-b4db-bbe12b4b7c95-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 15:25:50 crc kubenswrapper[4907]: I0313 15:25:50.753245 4907 generic.go:334] "Generic (PLEG): container finished" podID="6454399b-51ab-41f1-b4db-bbe12b4b7c95" containerID="61ac5444abd6e1ac4c4de07ff227d8deea674a73498eeb60cb027d561df5df1f" exitCode=0 Mar 13 15:25:50 crc kubenswrapper[4907]: I0313 15:25:50.753295 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-skd22" event={"ID":"6454399b-51ab-41f1-b4db-bbe12b4b7c95","Type":"ContainerDied","Data":"61ac5444abd6e1ac4c4de07ff227d8deea674a73498eeb60cb027d561df5df1f"} Mar 13 15:25:50 crc kubenswrapper[4907]: I0313 15:25:50.753319 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-skd22" Mar 13 15:25:50 crc kubenswrapper[4907]: I0313 15:25:50.753348 4907 scope.go:117] "RemoveContainer" containerID="61ac5444abd6e1ac4c4de07ff227d8deea674a73498eeb60cb027d561df5df1f" Mar 13 15:25:50 crc kubenswrapper[4907]: I0313 15:25:50.753334 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-skd22" event={"ID":"6454399b-51ab-41f1-b4db-bbe12b4b7c95","Type":"ContainerDied","Data":"4ff9f5b55970d7e99e658759b98269c42d65a135c1963e0a4852afb8d683eb2f"} Mar 13 15:25:50 crc kubenswrapper[4907]: I0313 15:25:50.773336 4907 scope.go:117] "RemoveContainer" containerID="40a57078c576bd33006fc990f6f567d9963ea1da66645391e9fcf035de5e33fa" Mar 13 15:25:50 crc kubenswrapper[4907]: I0313 15:25:50.794079 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-skd22"] Mar 13 15:25:50 crc kubenswrapper[4907]: I0313 15:25:50.799416 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-skd22"] Mar 13 15:25:50 crc kubenswrapper[4907]: I0313 15:25:50.808230 4907 scope.go:117] "RemoveContainer" containerID="c1c98d52ad0aba6e93f5f3a8f2f40fa1e06cc70b71f70d32c0e3077df5f68e0d" Mar 13 15:25:50 crc kubenswrapper[4907]: I0313 15:25:50.824672 4907 scope.go:117] "RemoveContainer" containerID="61ac5444abd6e1ac4c4de07ff227d8deea674a73498eeb60cb027d561df5df1f" Mar 13 15:25:50 crc kubenswrapper[4907]: E0313 15:25:50.825264 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"61ac5444abd6e1ac4c4de07ff227d8deea674a73498eeb60cb027d561df5df1f\": container with ID starting with 61ac5444abd6e1ac4c4de07ff227d8deea674a73498eeb60cb027d561df5df1f not found: ID does not exist" containerID="61ac5444abd6e1ac4c4de07ff227d8deea674a73498eeb60cb027d561df5df1f" Mar 13 15:25:50 crc kubenswrapper[4907]: I0313 15:25:50.825297 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"61ac5444abd6e1ac4c4de07ff227d8deea674a73498eeb60cb027d561df5df1f"} err="failed to get container status \"61ac5444abd6e1ac4c4de07ff227d8deea674a73498eeb60cb027d561df5df1f\": rpc error: code = NotFound desc = could not find container \"61ac5444abd6e1ac4c4de07ff227d8deea674a73498eeb60cb027d561df5df1f\": container with ID starting with 61ac5444abd6e1ac4c4de07ff227d8deea674a73498eeb60cb027d561df5df1f not found: ID does not exist" Mar 13 15:25:50 crc kubenswrapper[4907]: I0313 15:25:50.825326 4907 scope.go:117] "RemoveContainer" containerID="40a57078c576bd33006fc990f6f567d9963ea1da66645391e9fcf035de5e33fa" Mar 13 15:25:50 crc kubenswrapper[4907]: E0313 15:25:50.825593 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"40a57078c576bd33006fc990f6f567d9963ea1da66645391e9fcf035de5e33fa\": container with ID starting with 40a57078c576bd33006fc990f6f567d9963ea1da66645391e9fcf035de5e33fa not found: ID does not exist" containerID="40a57078c576bd33006fc990f6f567d9963ea1da66645391e9fcf035de5e33fa" Mar 13 15:25:50 crc kubenswrapper[4907]: I0313 15:25:50.825643 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"40a57078c576bd33006fc990f6f567d9963ea1da66645391e9fcf035de5e33fa"} err="failed to get container status \"40a57078c576bd33006fc990f6f567d9963ea1da66645391e9fcf035de5e33fa\": rpc error: code = NotFound desc = could not find container \"40a57078c576bd33006fc990f6f567d9963ea1da66645391e9fcf035de5e33fa\": container with ID starting with 40a57078c576bd33006fc990f6f567d9963ea1da66645391e9fcf035de5e33fa not found: ID does not exist" Mar 13 15:25:50 crc kubenswrapper[4907]: I0313 15:25:50.825671 4907 scope.go:117] "RemoveContainer" containerID="c1c98d52ad0aba6e93f5f3a8f2f40fa1e06cc70b71f70d32c0e3077df5f68e0d" Mar 13 15:25:50 crc kubenswrapper[4907]: E0313 15:25:50.826114 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c1c98d52ad0aba6e93f5f3a8f2f40fa1e06cc70b71f70d32c0e3077df5f68e0d\": container with ID starting with c1c98d52ad0aba6e93f5f3a8f2f40fa1e06cc70b71f70d32c0e3077df5f68e0d not found: ID does not exist" containerID="c1c98d52ad0aba6e93f5f3a8f2f40fa1e06cc70b71f70d32c0e3077df5f68e0d" Mar 13 15:25:50 crc kubenswrapper[4907]: I0313 15:25:50.826135 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c1c98d52ad0aba6e93f5f3a8f2f40fa1e06cc70b71f70d32c0e3077df5f68e0d"} err="failed to get container status \"c1c98d52ad0aba6e93f5f3a8f2f40fa1e06cc70b71f70d32c0e3077df5f68e0d\": rpc error: code = NotFound desc = could not find container \"c1c98d52ad0aba6e93f5f3a8f2f40fa1e06cc70b71f70d32c0e3077df5f68e0d\": container with ID starting with c1c98d52ad0aba6e93f5f3a8f2f40fa1e06cc70b71f70d32c0e3077df5f68e0d not found: ID does not exist" Mar 13 15:25:51 crc kubenswrapper[4907]: I0313 15:25:51.791638 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6454399b-51ab-41f1-b4db-bbe12b4b7c95" path="/var/lib/kubelet/pods/6454399b-51ab-41f1-b4db-bbe12b4b7c95/volumes" Mar 13 15:26:00 crc kubenswrapper[4907]: I0313 15:26:00.139144 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556926-r9tzc"] Mar 13 15:26:00 crc kubenswrapper[4907]: E0313 15:26:00.139993 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6454399b-51ab-41f1-b4db-bbe12b4b7c95" containerName="extract-content" Mar 13 15:26:00 crc kubenswrapper[4907]: I0313 15:26:00.140014 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="6454399b-51ab-41f1-b4db-bbe12b4b7c95" containerName="extract-content" Mar 13 15:26:00 crc kubenswrapper[4907]: E0313 15:26:00.140030 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6454399b-51ab-41f1-b4db-bbe12b4b7c95" containerName="registry-server" Mar 13 15:26:00 crc kubenswrapper[4907]: I0313 15:26:00.140038 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="6454399b-51ab-41f1-b4db-bbe12b4b7c95" containerName="registry-server" Mar 13 15:26:00 crc kubenswrapper[4907]: E0313 15:26:00.140058 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6454399b-51ab-41f1-b4db-bbe12b4b7c95" containerName="extract-utilities" Mar 13 15:26:00 crc kubenswrapper[4907]: I0313 15:26:00.140069 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="6454399b-51ab-41f1-b4db-bbe12b4b7c95" containerName="extract-utilities" Mar 13 15:26:00 crc kubenswrapper[4907]: I0313 15:26:00.140321 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="6454399b-51ab-41f1-b4db-bbe12b4b7c95" containerName="registry-server" Mar 13 15:26:00 crc kubenswrapper[4907]: I0313 15:26:00.143304 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556926-r9tzc" Mar 13 15:26:00 crc kubenswrapper[4907]: I0313 15:26:00.145216 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 15:26:00 crc kubenswrapper[4907]: I0313 15:26:00.145242 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 15:26:00 crc kubenswrapper[4907]: I0313 15:26:00.146422 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 15:26:00 crc kubenswrapper[4907]: I0313 15:26:00.154905 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556926-r9tzc"] Mar 13 15:26:00 crc kubenswrapper[4907]: I0313 15:26:00.307527 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nntns\" (UniqueName: \"kubernetes.io/projected/81a286dc-1af0-43f3-ad06-0e7d42141a0a-kube-api-access-nntns\") pod \"auto-csr-approver-29556926-r9tzc\" (UID: \"81a286dc-1af0-43f3-ad06-0e7d42141a0a\") " pod="openshift-infra/auto-csr-approver-29556926-r9tzc" Mar 13 15:26:00 crc kubenswrapper[4907]: I0313 15:26:00.408629 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nntns\" (UniqueName: \"kubernetes.io/projected/81a286dc-1af0-43f3-ad06-0e7d42141a0a-kube-api-access-nntns\") pod \"auto-csr-approver-29556926-r9tzc\" (UID: \"81a286dc-1af0-43f3-ad06-0e7d42141a0a\") " pod="openshift-infra/auto-csr-approver-29556926-r9tzc" Mar 13 15:26:00 crc kubenswrapper[4907]: I0313 15:26:00.430326 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nntns\" (UniqueName: \"kubernetes.io/projected/81a286dc-1af0-43f3-ad06-0e7d42141a0a-kube-api-access-nntns\") pod \"auto-csr-approver-29556926-r9tzc\" (UID: \"81a286dc-1af0-43f3-ad06-0e7d42141a0a\") " pod="openshift-infra/auto-csr-approver-29556926-r9tzc" Mar 13 15:26:00 crc kubenswrapper[4907]: I0313 15:26:00.469411 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556926-r9tzc" Mar 13 15:26:00 crc kubenswrapper[4907]: I0313 15:26:00.876336 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556926-r9tzc"] Mar 13 15:26:01 crc kubenswrapper[4907]: I0313 15:26:01.824852 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556926-r9tzc" event={"ID":"81a286dc-1af0-43f3-ad06-0e7d42141a0a","Type":"ContainerStarted","Data":"1dee45477813eb103f3973c857556e21148b420ea81f710f1f3bbf6e830f1279"} Mar 13 15:26:02 crc kubenswrapper[4907]: I0313 15:26:02.833301 4907 generic.go:334] "Generic (PLEG): container finished" podID="81a286dc-1af0-43f3-ad06-0e7d42141a0a" containerID="fdb72e54eff786808e8d908105afbd91c2b9b22c84cb13b6b15ad7f081283322" exitCode=0 Mar 13 15:26:02 crc kubenswrapper[4907]: I0313 15:26:02.833387 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556926-r9tzc" event={"ID":"81a286dc-1af0-43f3-ad06-0e7d42141a0a","Type":"ContainerDied","Data":"fdb72e54eff786808e8d908105afbd91c2b9b22c84cb13b6b15ad7f081283322"} Mar 13 15:26:04 crc kubenswrapper[4907]: I0313 15:26:04.083242 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556926-r9tzc" Mar 13 15:26:04 crc kubenswrapper[4907]: I0313 15:26:04.160217 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nntns\" (UniqueName: \"kubernetes.io/projected/81a286dc-1af0-43f3-ad06-0e7d42141a0a-kube-api-access-nntns\") pod \"81a286dc-1af0-43f3-ad06-0e7d42141a0a\" (UID: \"81a286dc-1af0-43f3-ad06-0e7d42141a0a\") " Mar 13 15:26:04 crc kubenswrapper[4907]: I0313 15:26:04.168123 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/81a286dc-1af0-43f3-ad06-0e7d42141a0a-kube-api-access-nntns" (OuterVolumeSpecName: "kube-api-access-nntns") pod "81a286dc-1af0-43f3-ad06-0e7d42141a0a" (UID: "81a286dc-1af0-43f3-ad06-0e7d42141a0a"). InnerVolumeSpecName "kube-api-access-nntns". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:26:04 crc kubenswrapper[4907]: I0313 15:26:04.261755 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nntns\" (UniqueName: \"kubernetes.io/projected/81a286dc-1af0-43f3-ad06-0e7d42141a0a-kube-api-access-nntns\") on node \"crc\" DevicePath \"\"" Mar 13 15:26:04 crc kubenswrapper[4907]: I0313 15:26:04.847333 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556926-r9tzc" event={"ID":"81a286dc-1af0-43f3-ad06-0e7d42141a0a","Type":"ContainerDied","Data":"1dee45477813eb103f3973c857556e21148b420ea81f710f1f3bbf6e830f1279"} Mar 13 15:26:04 crc kubenswrapper[4907]: I0313 15:26:04.847715 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1dee45477813eb103f3973c857556e21148b420ea81f710f1f3bbf6e830f1279" Mar 13 15:26:04 crc kubenswrapper[4907]: I0313 15:26:04.847372 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556926-r9tzc" Mar 13 15:26:05 crc kubenswrapper[4907]: I0313 15:26:05.154011 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556920-lrh8k"] Mar 13 15:26:05 crc kubenswrapper[4907]: I0313 15:26:05.160540 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556920-lrh8k"] Mar 13 15:26:05 crc kubenswrapper[4907]: I0313 15:26:05.791252 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925bf3a9-9978-49a9-8ac3-7c92a928adb7" path="/var/lib/kubelet/pods/925bf3a9-9978-49a9-8ac3-7c92a928adb7/volumes" Mar 13 15:26:12 crc kubenswrapper[4907]: I0313 15:26:12.923936 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7fc7c884dc-98jxk"] Mar 13 15:26:12 crc kubenswrapper[4907]: E0313 15:26:12.924605 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81a286dc-1af0-43f3-ad06-0e7d42141a0a" containerName="oc" Mar 13 15:26:12 crc kubenswrapper[4907]: I0313 15:26:12.924618 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="81a286dc-1af0-43f3-ad06-0e7d42141a0a" containerName="oc" Mar 13 15:26:12 crc kubenswrapper[4907]: I0313 15:26:12.924733 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="81a286dc-1af0-43f3-ad06-0e7d42141a0a" containerName="oc" Mar 13 15:26:12 crc kubenswrapper[4907]: I0313 15:26:12.925418 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fc7c884dc-98jxk" Mar 13 15:26:12 crc kubenswrapper[4907]: I0313 15:26:12.928520 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Mar 13 15:26:12 crc kubenswrapper[4907]: I0313 15:26:12.928700 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-5b4k5" Mar 13 15:26:12 crc kubenswrapper[4907]: I0313 15:26:12.928820 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Mar 13 15:26:12 crc kubenswrapper[4907]: I0313 15:26:12.929426 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Mar 13 15:26:12 crc kubenswrapper[4907]: I0313 15:26:12.929515 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Mar 13 15:26:12 crc kubenswrapper[4907]: I0313 15:26:12.942648 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7fc7c884dc-98jxk"] Mar 13 15:26:12 crc kubenswrapper[4907]: I0313 15:26:12.972869 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2cdac990-bf65-48a3-8892-b375611f1e70-config\") pod \"dnsmasq-dns-7fc7c884dc-98jxk\" (UID: \"2cdac990-bf65-48a3-8892-b375611f1e70\") " pod="openstack/dnsmasq-dns-7fc7c884dc-98jxk" Mar 13 15:26:12 crc kubenswrapper[4907]: I0313 15:26:12.972941 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2cdac990-bf65-48a3-8892-b375611f1e70-dns-svc\") pod \"dnsmasq-dns-7fc7c884dc-98jxk\" (UID: \"2cdac990-bf65-48a3-8892-b375611f1e70\") " pod="openstack/dnsmasq-dns-7fc7c884dc-98jxk" Mar 13 15:26:12 crc kubenswrapper[4907]: I0313 15:26:12.972964 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6c4r9\" (UniqueName: \"kubernetes.io/projected/2cdac990-bf65-48a3-8892-b375611f1e70-kube-api-access-6c4r9\") pod \"dnsmasq-dns-7fc7c884dc-98jxk\" (UID: \"2cdac990-bf65-48a3-8892-b375611f1e70\") " pod="openstack/dnsmasq-dns-7fc7c884dc-98jxk" Mar 13 15:26:13 crc kubenswrapper[4907]: I0313 15:26:13.074210 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2cdac990-bf65-48a3-8892-b375611f1e70-config\") pod \"dnsmasq-dns-7fc7c884dc-98jxk\" (UID: \"2cdac990-bf65-48a3-8892-b375611f1e70\") " pod="openstack/dnsmasq-dns-7fc7c884dc-98jxk" Mar 13 15:26:13 crc kubenswrapper[4907]: I0313 15:26:13.074260 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2cdac990-bf65-48a3-8892-b375611f1e70-dns-svc\") pod \"dnsmasq-dns-7fc7c884dc-98jxk\" (UID: \"2cdac990-bf65-48a3-8892-b375611f1e70\") " pod="openstack/dnsmasq-dns-7fc7c884dc-98jxk" Mar 13 15:26:13 crc kubenswrapper[4907]: I0313 15:26:13.074287 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6c4r9\" (UniqueName: \"kubernetes.io/projected/2cdac990-bf65-48a3-8892-b375611f1e70-kube-api-access-6c4r9\") pod \"dnsmasq-dns-7fc7c884dc-98jxk\" (UID: \"2cdac990-bf65-48a3-8892-b375611f1e70\") " pod="openstack/dnsmasq-dns-7fc7c884dc-98jxk" Mar 13 15:26:13 crc kubenswrapper[4907]: I0313 15:26:13.075314 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2cdac990-bf65-48a3-8892-b375611f1e70-config\") pod \"dnsmasq-dns-7fc7c884dc-98jxk\" (UID: \"2cdac990-bf65-48a3-8892-b375611f1e70\") " pod="openstack/dnsmasq-dns-7fc7c884dc-98jxk" Mar 13 15:26:13 crc kubenswrapper[4907]: I0313 15:26:13.075519 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2cdac990-bf65-48a3-8892-b375611f1e70-dns-svc\") pod \"dnsmasq-dns-7fc7c884dc-98jxk\" (UID: \"2cdac990-bf65-48a3-8892-b375611f1e70\") " pod="openstack/dnsmasq-dns-7fc7c884dc-98jxk" Mar 13 15:26:13 crc kubenswrapper[4907]: I0313 15:26:13.092212 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6c4r9\" (UniqueName: \"kubernetes.io/projected/2cdac990-bf65-48a3-8892-b375611f1e70-kube-api-access-6c4r9\") pod \"dnsmasq-dns-7fc7c884dc-98jxk\" (UID: \"2cdac990-bf65-48a3-8892-b375611f1e70\") " pod="openstack/dnsmasq-dns-7fc7c884dc-98jxk" Mar 13 15:26:13 crc kubenswrapper[4907]: I0313 15:26:13.203027 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7c95686bd5-8kb7m"] Mar 13 15:26:13 crc kubenswrapper[4907]: I0313 15:26:13.204173 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c95686bd5-8kb7m" Mar 13 15:26:13 crc kubenswrapper[4907]: I0313 15:26:13.214476 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7c95686bd5-8kb7m"] Mar 13 15:26:13 crc kubenswrapper[4907]: I0313 15:26:13.242381 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fc7c884dc-98jxk" Mar 13 15:26:13 crc kubenswrapper[4907]: I0313 15:26:13.377868 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d1b7f262-ac9a-4a7e-ae94-6ce13d4105a2-dns-svc\") pod \"dnsmasq-dns-7c95686bd5-8kb7m\" (UID: \"d1b7f262-ac9a-4a7e-ae94-6ce13d4105a2\") " pod="openstack/dnsmasq-dns-7c95686bd5-8kb7m" Mar 13 15:26:13 crc kubenswrapper[4907]: I0313 15:26:13.378145 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d1b7f262-ac9a-4a7e-ae94-6ce13d4105a2-config\") pod \"dnsmasq-dns-7c95686bd5-8kb7m\" (UID: \"d1b7f262-ac9a-4a7e-ae94-6ce13d4105a2\") " pod="openstack/dnsmasq-dns-7c95686bd5-8kb7m" Mar 13 15:26:13 crc kubenswrapper[4907]: I0313 15:26:13.378230 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r9x6s\" (UniqueName: \"kubernetes.io/projected/d1b7f262-ac9a-4a7e-ae94-6ce13d4105a2-kube-api-access-r9x6s\") pod \"dnsmasq-dns-7c95686bd5-8kb7m\" (UID: \"d1b7f262-ac9a-4a7e-ae94-6ce13d4105a2\") " pod="openstack/dnsmasq-dns-7c95686bd5-8kb7m" Mar 13 15:26:13 crc kubenswrapper[4907]: I0313 15:26:13.479650 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r9x6s\" (UniqueName: \"kubernetes.io/projected/d1b7f262-ac9a-4a7e-ae94-6ce13d4105a2-kube-api-access-r9x6s\") pod \"dnsmasq-dns-7c95686bd5-8kb7m\" (UID: \"d1b7f262-ac9a-4a7e-ae94-6ce13d4105a2\") " pod="openstack/dnsmasq-dns-7c95686bd5-8kb7m" Mar 13 15:26:13 crc kubenswrapper[4907]: I0313 15:26:13.479842 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d1b7f262-ac9a-4a7e-ae94-6ce13d4105a2-dns-svc\") pod \"dnsmasq-dns-7c95686bd5-8kb7m\" (UID: \"d1b7f262-ac9a-4a7e-ae94-6ce13d4105a2\") " pod="openstack/dnsmasq-dns-7c95686bd5-8kb7m" Mar 13 15:26:13 crc kubenswrapper[4907]: I0313 15:26:13.479963 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d1b7f262-ac9a-4a7e-ae94-6ce13d4105a2-config\") pod \"dnsmasq-dns-7c95686bd5-8kb7m\" (UID: \"d1b7f262-ac9a-4a7e-ae94-6ce13d4105a2\") " pod="openstack/dnsmasq-dns-7c95686bd5-8kb7m" Mar 13 15:26:13 crc kubenswrapper[4907]: I0313 15:26:13.480933 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d1b7f262-ac9a-4a7e-ae94-6ce13d4105a2-dns-svc\") pod \"dnsmasq-dns-7c95686bd5-8kb7m\" (UID: \"d1b7f262-ac9a-4a7e-ae94-6ce13d4105a2\") " pod="openstack/dnsmasq-dns-7c95686bd5-8kb7m" Mar 13 15:26:13 crc kubenswrapper[4907]: I0313 15:26:13.480929 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d1b7f262-ac9a-4a7e-ae94-6ce13d4105a2-config\") pod \"dnsmasq-dns-7c95686bd5-8kb7m\" (UID: \"d1b7f262-ac9a-4a7e-ae94-6ce13d4105a2\") " pod="openstack/dnsmasq-dns-7c95686bd5-8kb7m" Mar 13 15:26:13 crc kubenswrapper[4907]: I0313 15:26:13.500323 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r9x6s\" (UniqueName: \"kubernetes.io/projected/d1b7f262-ac9a-4a7e-ae94-6ce13d4105a2-kube-api-access-r9x6s\") pod \"dnsmasq-dns-7c95686bd5-8kb7m\" (UID: \"d1b7f262-ac9a-4a7e-ae94-6ce13d4105a2\") " pod="openstack/dnsmasq-dns-7c95686bd5-8kb7m" Mar 13 15:26:13 crc kubenswrapper[4907]: I0313 15:26:13.534094 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c95686bd5-8kb7m" Mar 13 15:26:13 crc kubenswrapper[4907]: I0313 15:26:13.692732 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7fc7c884dc-98jxk"] Mar 13 15:26:13 crc kubenswrapper[4907]: W0313 15:26:13.698305 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2cdac990_bf65_48a3_8892_b375611f1e70.slice/crio-77a29a36e6edcd7453504c99bd8d6185ca0a664cb336729564e5ea2647a574ad WatchSource:0}: Error finding container 77a29a36e6edcd7453504c99bd8d6185ca0a664cb336729564e5ea2647a574ad: Status 404 returned error can't find the container with id 77a29a36e6edcd7453504c99bd8d6185ca0a664cb336729564e5ea2647a574ad Mar 13 15:26:13 crc kubenswrapper[4907]: I0313 15:26:13.909615 4907 generic.go:334] "Generic (PLEG): container finished" podID="2cdac990-bf65-48a3-8892-b375611f1e70" containerID="380d0c10f65356a76638952335ee06865a37a23f94d9e43041a4522fb0545d73" exitCode=0 Mar 13 15:26:13 crc kubenswrapper[4907]: I0313 15:26:13.909665 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fc7c884dc-98jxk" event={"ID":"2cdac990-bf65-48a3-8892-b375611f1e70","Type":"ContainerDied","Data":"380d0c10f65356a76638952335ee06865a37a23f94d9e43041a4522fb0545d73"} Mar 13 15:26:13 crc kubenswrapper[4907]: I0313 15:26:13.909729 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fc7c884dc-98jxk" event={"ID":"2cdac990-bf65-48a3-8892-b375611f1e70","Type":"ContainerStarted","Data":"77a29a36e6edcd7453504c99bd8d6185ca0a664cb336729564e5ea2647a574ad"} Mar 13 15:26:13 crc kubenswrapper[4907]: I0313 15:26:13.995719 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7c95686bd5-8kb7m"] Mar 13 15:26:14 crc kubenswrapper[4907]: W0313 15:26:14.002098 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd1b7f262_ac9a_4a7e_ae94_6ce13d4105a2.slice/crio-18cfff04ebd6c2ab270b3fa1a352afaef9dc310f2e13e5df39289e1f0f968058 WatchSource:0}: Error finding container 18cfff04ebd6c2ab270b3fa1a352afaef9dc310f2e13e5df39289e1f0f968058: Status 404 returned error can't find the container with id 18cfff04ebd6c2ab270b3fa1a352afaef9dc310f2e13e5df39289e1f0f968058 Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.080864 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.081937 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.084138 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.084488 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.085373 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-vlpfs" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.085613 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.085705 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.095298 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.196790 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d-server-conf\") pod \"rabbitmq-server-0\" (UID: \"7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d\") " pod="openstack/rabbitmq-server-0" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.196869 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d\") " pod="openstack/rabbitmq-server-0" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.196925 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d\") " pod="openstack/rabbitmq-server-0" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.197450 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d\") " pod="openstack/rabbitmq-server-0" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.197582 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d\") " pod="openstack/rabbitmq-server-0" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.197625 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-9187efe6-871d-4bbe-9376-bff5d8a151f2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-9187efe6-871d-4bbe-9376-bff5d8a151f2\") pod \"rabbitmq-server-0\" (UID: \"7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d\") " pod="openstack/rabbitmq-server-0" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.197675 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vsj5v\" (UniqueName: \"kubernetes.io/projected/7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d-kube-api-access-vsj5v\") pod \"rabbitmq-server-0\" (UID: \"7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d\") " pod="openstack/rabbitmq-server-0" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.197802 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d-pod-info\") pod \"rabbitmq-server-0\" (UID: \"7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d\") " pod="openstack/rabbitmq-server-0" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.197850 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d\") " pod="openstack/rabbitmq-server-0" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.299483 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d\") " pod="openstack/rabbitmq-server-0" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.299551 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d\") " pod="openstack/rabbitmq-server-0" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.299590 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d\") " pod="openstack/rabbitmq-server-0" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.299644 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d\") " pod="openstack/rabbitmq-server-0" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.299670 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-9187efe6-871d-4bbe-9376-bff5d8a151f2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-9187efe6-871d-4bbe-9376-bff5d8a151f2\") pod \"rabbitmq-server-0\" (UID: \"7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d\") " pod="openstack/rabbitmq-server-0" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.299703 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vsj5v\" (UniqueName: \"kubernetes.io/projected/7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d-kube-api-access-vsj5v\") pod \"rabbitmq-server-0\" (UID: \"7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d\") " pod="openstack/rabbitmq-server-0" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.299739 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d-pod-info\") pod \"rabbitmq-server-0\" (UID: \"7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d\") " pod="openstack/rabbitmq-server-0" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.299766 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d\") " pod="openstack/rabbitmq-server-0" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.299801 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d-server-conf\") pod \"rabbitmq-server-0\" (UID: \"7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d\") " pod="openstack/rabbitmq-server-0" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.300126 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d\") " pod="openstack/rabbitmq-server-0" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.300155 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d\") " pod="openstack/rabbitmq-server-0" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.301012 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d\") " pod="openstack/rabbitmq-server-0" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.301538 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d-server-conf\") pod \"rabbitmq-server-0\" (UID: \"7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d\") " pod="openstack/rabbitmq-server-0" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.303453 4907 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.303484 4907 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-9187efe6-871d-4bbe-9376-bff5d8a151f2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-9187efe6-871d-4bbe-9376-bff5d8a151f2\") pod \"rabbitmq-server-0\" (UID: \"7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/c926015d5702d92351724596ba892321797e624e5f7a01b00404b7d1c8a17788/globalmount\"" pod="openstack/rabbitmq-server-0" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.303648 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d\") " pod="openstack/rabbitmq-server-0" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.303923 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d-pod-info\") pod \"rabbitmq-server-0\" (UID: \"7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d\") " pod="openstack/rabbitmq-server-0" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.304230 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d\") " pod="openstack/rabbitmq-server-0" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.326904 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vsj5v\" (UniqueName: \"kubernetes.io/projected/7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d-kube-api-access-vsj5v\") pod \"rabbitmq-server-0\" (UID: \"7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d\") " pod="openstack/rabbitmq-server-0" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.337205 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-9187efe6-871d-4bbe-9376-bff5d8a151f2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-9187efe6-871d-4bbe-9376-bff5d8a151f2\") pod \"rabbitmq-server-0\" (UID: \"7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d\") " pod="openstack/rabbitmq-server-0" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.399367 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.400851 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.402814 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.403146 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.403276 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.403269 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-grjtp" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.403439 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.417047 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.469241 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.502808 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.502906 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.502938 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.502966 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vhr7j\" (UniqueName: \"kubernetes.io/projected/75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd-kube-api-access-vhr7j\") pod \"rabbitmq-cell1-server-0\" (UID: \"75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.503014 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-7c39547e-1168-404d-865b-13cfd1526f6f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7c39547e-1168-404d-865b-13cfd1526f6f\") pod \"rabbitmq-cell1-server-0\" (UID: \"75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.503052 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.503089 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.503137 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.503190 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.604953 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.605001 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.605037 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.605061 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.605087 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.605104 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.605123 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vhr7j\" (UniqueName: \"kubernetes.io/projected/75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd-kube-api-access-vhr7j\") pod \"rabbitmq-cell1-server-0\" (UID: \"75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.605154 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-7c39547e-1168-404d-865b-13cfd1526f6f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7c39547e-1168-404d-865b-13cfd1526f6f\") pod \"rabbitmq-cell1-server-0\" (UID: \"75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.605178 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.606272 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.606524 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.612012 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.612025 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.612618 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.613620 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.615020 4907 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.615047 4907 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-7c39547e-1168-404d-865b-13cfd1526f6f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7c39547e-1168-404d-865b-13cfd1526f6f\") pod \"rabbitmq-cell1-server-0\" (UID: \"75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/d84a704eae1ca42fe329ab2f9d170d2a410be00de88c5e0a709d7b62965c26cb/globalmount\"" pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.615107 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.630829 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vhr7j\" (UniqueName: \"kubernetes.io/projected/75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd-kube-api-access-vhr7j\") pod \"rabbitmq-cell1-server-0\" (UID: \"75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.643475 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-7c39547e-1168-404d-865b-13cfd1526f6f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7c39547e-1168-404d-865b-13cfd1526f6f\") pod \"rabbitmq-cell1-server-0\" (UID: \"75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.720940 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.893616 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.918143 4907 generic.go:334] "Generic (PLEG): container finished" podID="d1b7f262-ac9a-4a7e-ae94-6ce13d4105a2" containerID="8f7fad973a21b1188eda2310b2c3e66fb866528b33ce2f78d5f9e1986ac222d3" exitCode=0 Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.918214 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c95686bd5-8kb7m" event={"ID":"d1b7f262-ac9a-4a7e-ae94-6ce13d4105a2","Type":"ContainerDied","Data":"8f7fad973a21b1188eda2310b2c3e66fb866528b33ce2f78d5f9e1986ac222d3"} Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.918245 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c95686bd5-8kb7m" event={"ID":"d1b7f262-ac9a-4a7e-ae94-6ce13d4105a2","Type":"ContainerStarted","Data":"18cfff04ebd6c2ab270b3fa1a352afaef9dc310f2e13e5df39289e1f0f968058"} Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.921749 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d","Type":"ContainerStarted","Data":"7fa7486d25e0b1f49302f99040c36181735134e64166f989efe3f3ae26956d78"} Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.926316 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fc7c884dc-98jxk" event={"ID":"2cdac990-bf65-48a3-8892-b375611f1e70","Type":"ContainerStarted","Data":"4d2e0d92c2b302461c3a8b6d60213e7b81411578e2bc548ac86409412c045bf8"} Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.928635 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7fc7c884dc-98jxk" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.965228 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7fc7c884dc-98jxk" podStartSLOduration=2.9652065050000003 podStartE2EDuration="2.965206505s" podCreationTimestamp="2026-03-13 15:26:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:26:14.963069347 +0000 UTC m=+4873.862857036" watchObservedRunningTime="2026-03-13 15:26:14.965206505 +0000 UTC m=+4873.864994194" Mar 13 15:26:15 crc kubenswrapper[4907]: I0313 15:26:15.124863 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 13 15:26:15 crc kubenswrapper[4907]: W0313 15:26:15.129411 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod75f775f2_b77c_4ff4_80e4_12f6d3e2d0fd.slice/crio-24618a7621fc23ef2173f3056fdb4f82fc6d9f9738f1b8f67a17813327203788 WatchSource:0}: Error finding container 24618a7621fc23ef2173f3056fdb4f82fc6d9f9738f1b8f67a17813327203788: Status 404 returned error can't find the container with id 24618a7621fc23ef2173f3056fdb4f82fc6d9f9738f1b8f67a17813327203788 Mar 13 15:26:15 crc kubenswrapper[4907]: I0313 15:26:15.484549 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Mar 13 15:26:15 crc kubenswrapper[4907]: I0313 15:26:15.490706 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Mar 13 15:26:15 crc kubenswrapper[4907]: I0313 15:26:15.498409 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Mar 13 15:26:15 crc kubenswrapper[4907]: I0313 15:26:15.498762 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-xl4lf" Mar 13 15:26:15 crc kubenswrapper[4907]: I0313 15:26:15.499079 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Mar 13 15:26:15 crc kubenswrapper[4907]: I0313 15:26:15.499316 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Mar 13 15:26:15 crc kubenswrapper[4907]: I0313 15:26:15.502801 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Mar 13 15:26:15 crc kubenswrapper[4907]: I0313 15:26:15.508849 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Mar 13 15:26:15 crc kubenswrapper[4907]: I0313 15:26:15.618656 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/585e3be7-1e17-443c-b430-0c03727167cb-operator-scripts\") pod \"openstack-galera-0\" (UID: \"585e3be7-1e17-443c-b430-0c03727167cb\") " pod="openstack/openstack-galera-0" Mar 13 15:26:15 crc kubenswrapper[4907]: I0313 15:26:15.618705 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/585e3be7-1e17-443c-b430-0c03727167cb-config-data-default\") pod \"openstack-galera-0\" (UID: \"585e3be7-1e17-443c-b430-0c03727167cb\") " pod="openstack/openstack-galera-0" Mar 13 15:26:15 crc kubenswrapper[4907]: I0313 15:26:15.618737 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/585e3be7-1e17-443c-b430-0c03727167cb-config-data-generated\") pod \"openstack-galera-0\" (UID: \"585e3be7-1e17-443c-b430-0c03727167cb\") " pod="openstack/openstack-galera-0" Mar 13 15:26:15 crc kubenswrapper[4907]: I0313 15:26:15.618802 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rvvpz\" (UniqueName: \"kubernetes.io/projected/585e3be7-1e17-443c-b430-0c03727167cb-kube-api-access-rvvpz\") pod \"openstack-galera-0\" (UID: \"585e3be7-1e17-443c-b430-0c03727167cb\") " pod="openstack/openstack-galera-0" Mar 13 15:26:15 crc kubenswrapper[4907]: I0313 15:26:15.618844 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/585e3be7-1e17-443c-b430-0c03727167cb-kolla-config\") pod \"openstack-galera-0\" (UID: \"585e3be7-1e17-443c-b430-0c03727167cb\") " pod="openstack/openstack-galera-0" Mar 13 15:26:15 crc kubenswrapper[4907]: I0313 15:26:15.618897 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-2e0cba70-4127-4af3-a5ff-dca1a2fa2510\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2e0cba70-4127-4af3-a5ff-dca1a2fa2510\") pod \"openstack-galera-0\" (UID: \"585e3be7-1e17-443c-b430-0c03727167cb\") " pod="openstack/openstack-galera-0" Mar 13 15:26:15 crc kubenswrapper[4907]: I0313 15:26:15.618931 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/585e3be7-1e17-443c-b430-0c03727167cb-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"585e3be7-1e17-443c-b430-0c03727167cb\") " pod="openstack/openstack-galera-0" Mar 13 15:26:15 crc kubenswrapper[4907]: I0313 15:26:15.618963 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/585e3be7-1e17-443c-b430-0c03727167cb-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"585e3be7-1e17-443c-b430-0c03727167cb\") " pod="openstack/openstack-galera-0" Mar 13 15:26:15 crc kubenswrapper[4907]: I0313 15:26:15.720651 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/585e3be7-1e17-443c-b430-0c03727167cb-operator-scripts\") pod \"openstack-galera-0\" (UID: \"585e3be7-1e17-443c-b430-0c03727167cb\") " pod="openstack/openstack-galera-0" Mar 13 15:26:15 crc kubenswrapper[4907]: I0313 15:26:15.720734 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/585e3be7-1e17-443c-b430-0c03727167cb-config-data-default\") pod \"openstack-galera-0\" (UID: \"585e3be7-1e17-443c-b430-0c03727167cb\") " pod="openstack/openstack-galera-0" Mar 13 15:26:15 crc kubenswrapper[4907]: I0313 15:26:15.720775 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/585e3be7-1e17-443c-b430-0c03727167cb-config-data-generated\") pod \"openstack-galera-0\" (UID: \"585e3be7-1e17-443c-b430-0c03727167cb\") " pod="openstack/openstack-galera-0" Mar 13 15:26:15 crc kubenswrapper[4907]: I0313 15:26:15.720801 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rvvpz\" (UniqueName: \"kubernetes.io/projected/585e3be7-1e17-443c-b430-0c03727167cb-kube-api-access-rvvpz\") pod \"openstack-galera-0\" (UID: \"585e3be7-1e17-443c-b430-0c03727167cb\") " pod="openstack/openstack-galera-0" Mar 13 15:26:15 crc kubenswrapper[4907]: I0313 15:26:15.720827 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/585e3be7-1e17-443c-b430-0c03727167cb-kolla-config\") pod \"openstack-galera-0\" (UID: \"585e3be7-1e17-443c-b430-0c03727167cb\") " pod="openstack/openstack-galera-0" Mar 13 15:26:15 crc kubenswrapper[4907]: I0313 15:26:15.720868 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-2e0cba70-4127-4af3-a5ff-dca1a2fa2510\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2e0cba70-4127-4af3-a5ff-dca1a2fa2510\") pod \"openstack-galera-0\" (UID: \"585e3be7-1e17-443c-b430-0c03727167cb\") " pod="openstack/openstack-galera-0" Mar 13 15:26:15 crc kubenswrapper[4907]: I0313 15:26:15.720918 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/585e3be7-1e17-443c-b430-0c03727167cb-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"585e3be7-1e17-443c-b430-0c03727167cb\") " pod="openstack/openstack-galera-0" Mar 13 15:26:15 crc kubenswrapper[4907]: I0313 15:26:15.720964 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/585e3be7-1e17-443c-b430-0c03727167cb-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"585e3be7-1e17-443c-b430-0c03727167cb\") " pod="openstack/openstack-galera-0" Mar 13 15:26:15 crc kubenswrapper[4907]: I0313 15:26:15.722000 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/585e3be7-1e17-443c-b430-0c03727167cb-config-data-generated\") pod \"openstack-galera-0\" (UID: \"585e3be7-1e17-443c-b430-0c03727167cb\") " pod="openstack/openstack-galera-0" Mar 13 15:26:15 crc kubenswrapper[4907]: I0313 15:26:15.722364 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/585e3be7-1e17-443c-b430-0c03727167cb-operator-scripts\") pod \"openstack-galera-0\" (UID: \"585e3be7-1e17-443c-b430-0c03727167cb\") " pod="openstack/openstack-galera-0" Mar 13 15:26:15 crc kubenswrapper[4907]: I0313 15:26:15.722665 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/585e3be7-1e17-443c-b430-0c03727167cb-kolla-config\") pod \"openstack-galera-0\" (UID: \"585e3be7-1e17-443c-b430-0c03727167cb\") " pod="openstack/openstack-galera-0" Mar 13 15:26:15 crc kubenswrapper[4907]: I0313 15:26:15.723071 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/585e3be7-1e17-443c-b430-0c03727167cb-config-data-default\") pod \"openstack-galera-0\" (UID: \"585e3be7-1e17-443c-b430-0c03727167cb\") " pod="openstack/openstack-galera-0" Mar 13 15:26:15 crc kubenswrapper[4907]: I0313 15:26:15.726183 4907 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 13 15:26:15 crc kubenswrapper[4907]: I0313 15:26:15.726225 4907 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-2e0cba70-4127-4af3-a5ff-dca1a2fa2510\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2e0cba70-4127-4af3-a5ff-dca1a2fa2510\") pod \"openstack-galera-0\" (UID: \"585e3be7-1e17-443c-b430-0c03727167cb\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/2835f978c80613ada9b2606f7e7b901545616ef9a453011deca8ef169d3eb6f9/globalmount\"" pod="openstack/openstack-galera-0" Mar 13 15:26:15 crc kubenswrapper[4907]: I0313 15:26:15.806861 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/585e3be7-1e17-443c-b430-0c03727167cb-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"585e3be7-1e17-443c-b430-0c03727167cb\") " pod="openstack/openstack-galera-0" Mar 13 15:26:15 crc kubenswrapper[4907]: I0313 15:26:15.807371 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/585e3be7-1e17-443c-b430-0c03727167cb-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"585e3be7-1e17-443c-b430-0c03727167cb\") " pod="openstack/openstack-galera-0" Mar 13 15:26:15 crc kubenswrapper[4907]: I0313 15:26:15.808279 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rvvpz\" (UniqueName: \"kubernetes.io/projected/585e3be7-1e17-443c-b430-0c03727167cb-kube-api-access-rvvpz\") pod \"openstack-galera-0\" (UID: \"585e3be7-1e17-443c-b430-0c03727167cb\") " pod="openstack/openstack-galera-0" Mar 13 15:26:15 crc kubenswrapper[4907]: I0313 15:26:15.933415 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Mar 13 15:26:15 crc kubenswrapper[4907]: I0313 15:26:15.934456 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Mar 13 15:26:15 crc kubenswrapper[4907]: I0313 15:26:15.937418 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Mar 13 15:26:15 crc kubenswrapper[4907]: I0313 15:26:15.937707 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c95686bd5-8kb7m" event={"ID":"d1b7f262-ac9a-4a7e-ae94-6ce13d4105a2","Type":"ContainerStarted","Data":"7dc8560cb6dc85650a7b469d68aaea22e66e3e209dca62e699bcb1b0807b2fea"} Mar 13 15:26:15 crc kubenswrapper[4907]: I0313 15:26:15.937754 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7c95686bd5-8kb7m" Mar 13 15:26:15 crc kubenswrapper[4907]: I0313 15:26:15.939165 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd","Type":"ContainerStarted","Data":"24618a7621fc23ef2173f3056fdb4f82fc6d9f9738f1b8f67a17813327203788"} Mar 13 15:26:15 crc kubenswrapper[4907]: I0313 15:26:15.943751 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-s66hx" Mar 13 15:26:15 crc kubenswrapper[4907]: I0313 15:26:15.946119 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Mar 13 15:26:15 crc kubenswrapper[4907]: I0313 15:26:15.979712 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7c95686bd5-8kb7m" podStartSLOduration=2.979689441 podStartE2EDuration="2.979689441s" podCreationTimestamp="2026-03-13 15:26:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:26:15.973397779 +0000 UTC m=+4874.873185468" watchObservedRunningTime="2026-03-13 15:26:15.979689441 +0000 UTC m=+4874.879477140" Mar 13 15:26:16 crc kubenswrapper[4907]: I0313 15:26:16.025041 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8c712e4f-4999-476b-970f-eae3189b4e59-config-data\") pod \"memcached-0\" (UID: \"8c712e4f-4999-476b-970f-eae3189b4e59\") " pod="openstack/memcached-0" Mar 13 15:26:16 crc kubenswrapper[4907]: I0313 15:26:16.025355 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qpmrh\" (UniqueName: \"kubernetes.io/projected/8c712e4f-4999-476b-970f-eae3189b4e59-kube-api-access-qpmrh\") pod \"memcached-0\" (UID: \"8c712e4f-4999-476b-970f-eae3189b4e59\") " pod="openstack/memcached-0" Mar 13 15:26:16 crc kubenswrapper[4907]: I0313 15:26:16.025418 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/8c712e4f-4999-476b-970f-eae3189b4e59-kolla-config\") pod \"memcached-0\" (UID: \"8c712e4f-4999-476b-970f-eae3189b4e59\") " pod="openstack/memcached-0" Mar 13 15:26:16 crc kubenswrapper[4907]: I0313 15:26:16.121522 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-2e0cba70-4127-4af3-a5ff-dca1a2fa2510\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2e0cba70-4127-4af3-a5ff-dca1a2fa2510\") pod \"openstack-galera-0\" (UID: \"585e3be7-1e17-443c-b430-0c03727167cb\") " pod="openstack/openstack-galera-0" Mar 13 15:26:16 crc kubenswrapper[4907]: I0313 15:26:16.126429 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8c712e4f-4999-476b-970f-eae3189b4e59-config-data\") pod \"memcached-0\" (UID: \"8c712e4f-4999-476b-970f-eae3189b4e59\") " pod="openstack/memcached-0" Mar 13 15:26:16 crc kubenswrapper[4907]: I0313 15:26:16.126508 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qpmrh\" (UniqueName: \"kubernetes.io/projected/8c712e4f-4999-476b-970f-eae3189b4e59-kube-api-access-qpmrh\") pod \"memcached-0\" (UID: \"8c712e4f-4999-476b-970f-eae3189b4e59\") " pod="openstack/memcached-0" Mar 13 15:26:16 crc kubenswrapper[4907]: I0313 15:26:16.126565 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/8c712e4f-4999-476b-970f-eae3189b4e59-kolla-config\") pod \"memcached-0\" (UID: \"8c712e4f-4999-476b-970f-eae3189b4e59\") " pod="openstack/memcached-0" Mar 13 15:26:16 crc kubenswrapper[4907]: I0313 15:26:16.127407 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/8c712e4f-4999-476b-970f-eae3189b4e59-kolla-config\") pod \"memcached-0\" (UID: \"8c712e4f-4999-476b-970f-eae3189b4e59\") " pod="openstack/memcached-0" Mar 13 15:26:16 crc kubenswrapper[4907]: I0313 15:26:16.127438 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8c712e4f-4999-476b-970f-eae3189b4e59-config-data\") pod \"memcached-0\" (UID: \"8c712e4f-4999-476b-970f-eae3189b4e59\") " pod="openstack/memcached-0" Mar 13 15:26:16 crc kubenswrapper[4907]: I0313 15:26:16.161525 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qpmrh\" (UniqueName: \"kubernetes.io/projected/8c712e4f-4999-476b-970f-eae3189b4e59-kube-api-access-qpmrh\") pod \"memcached-0\" (UID: \"8c712e4f-4999-476b-970f-eae3189b4e59\") " pod="openstack/memcached-0" Mar 13 15:26:16 crc kubenswrapper[4907]: I0313 15:26:16.250360 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Mar 13 15:26:16 crc kubenswrapper[4907]: I0313 15:26:16.412261 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Mar 13 15:26:16 crc kubenswrapper[4907]: I0313 15:26:16.724546 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Mar 13 15:26:16 crc kubenswrapper[4907]: I0313 15:26:16.840975 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Mar 13 15:26:16 crc kubenswrapper[4907]: W0313 15:26:16.913177 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8c712e4f_4999_476b_970f_eae3189b4e59.slice/crio-a0f9a382e21caec5c6f87f1cc77b784feb6313013846ee54439ff8d42725be02 WatchSource:0}: Error finding container a0f9a382e21caec5c6f87f1cc77b784feb6313013846ee54439ff8d42725be02: Status 404 returned error can't find the container with id a0f9a382e21caec5c6f87f1cc77b784feb6313013846ee54439ff8d42725be02 Mar 13 15:26:16 crc kubenswrapper[4907]: I0313 15:26:16.948391 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d","Type":"ContainerStarted","Data":"bf96e8e63801badfacccb754ff56342cdc5ba8417a4f20f022b8a5f76d1c9272"} Mar 13 15:26:16 crc kubenswrapper[4907]: I0313 15:26:16.950047 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"8c712e4f-4999-476b-970f-eae3189b4e59","Type":"ContainerStarted","Data":"a0f9a382e21caec5c6f87f1cc77b784feb6313013846ee54439ff8d42725be02"} Mar 13 15:26:16 crc kubenswrapper[4907]: I0313 15:26:16.951068 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"585e3be7-1e17-443c-b430-0c03727167cb","Type":"ContainerStarted","Data":"5c54f8dc7a3d2ec4cd9dcf35e8b0c95c41f32bfe48a3f1eab285bc0ae0bac5eb"} Mar 13 15:26:16 crc kubenswrapper[4907]: I0313 15:26:16.952775 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd","Type":"ContainerStarted","Data":"0b4af08a62b26495d3442675a1ce03aa549cddc0c7a28e8f6cb98a7deb792d53"} Mar 13 15:26:17 crc kubenswrapper[4907]: I0313 15:26:17.101445 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Mar 13 15:26:17 crc kubenswrapper[4907]: I0313 15:26:17.102937 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Mar 13 15:26:17 crc kubenswrapper[4907]: I0313 15:26:17.106258 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Mar 13 15:26:17 crc kubenswrapper[4907]: I0313 15:26:17.106595 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Mar 13 15:26:17 crc kubenswrapper[4907]: I0313 15:26:17.109695 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-stnw8" Mar 13 15:26:17 crc kubenswrapper[4907]: I0313 15:26:17.115066 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Mar 13 15:26:17 crc kubenswrapper[4907]: I0313 15:26:17.117461 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Mar 13 15:26:17 crc kubenswrapper[4907]: I0313 15:26:17.143340 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-3bdc965c-6ae3-457e-9054-e9bd61be219a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3bdc965c-6ae3-457e-9054-e9bd61be219a\") pod \"openstack-cell1-galera-0\" (UID: \"edb521bb-211e-48c0-b644-8dc44b235feb\") " pod="openstack/openstack-cell1-galera-0" Mar 13 15:26:17 crc kubenswrapper[4907]: I0313 15:26:17.143480 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edb521bb-211e-48c0-b644-8dc44b235feb-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"edb521bb-211e-48c0-b644-8dc44b235feb\") " pod="openstack/openstack-cell1-galera-0" Mar 13 15:26:17 crc kubenswrapper[4907]: I0313 15:26:17.143570 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/edb521bb-211e-48c0-b644-8dc44b235feb-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"edb521bb-211e-48c0-b644-8dc44b235feb\") " pod="openstack/openstack-cell1-galera-0" Mar 13 15:26:17 crc kubenswrapper[4907]: I0313 15:26:17.143688 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/edb521bb-211e-48c0-b644-8dc44b235feb-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"edb521bb-211e-48c0-b644-8dc44b235feb\") " pod="openstack/openstack-cell1-galera-0" Mar 13 15:26:17 crc kubenswrapper[4907]: I0313 15:26:17.143822 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pjp8c\" (UniqueName: \"kubernetes.io/projected/edb521bb-211e-48c0-b644-8dc44b235feb-kube-api-access-pjp8c\") pod \"openstack-cell1-galera-0\" (UID: \"edb521bb-211e-48c0-b644-8dc44b235feb\") " pod="openstack/openstack-cell1-galera-0" Mar 13 15:26:17 crc kubenswrapper[4907]: I0313 15:26:17.143915 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/edb521bb-211e-48c0-b644-8dc44b235feb-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"edb521bb-211e-48c0-b644-8dc44b235feb\") " pod="openstack/openstack-cell1-galera-0" Mar 13 15:26:17 crc kubenswrapper[4907]: I0313 15:26:17.143973 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/edb521bb-211e-48c0-b644-8dc44b235feb-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"edb521bb-211e-48c0-b644-8dc44b235feb\") " pod="openstack/openstack-cell1-galera-0" Mar 13 15:26:17 crc kubenswrapper[4907]: I0313 15:26:17.144047 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/edb521bb-211e-48c0-b644-8dc44b235feb-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"edb521bb-211e-48c0-b644-8dc44b235feb\") " pod="openstack/openstack-cell1-galera-0" Mar 13 15:26:17 crc kubenswrapper[4907]: I0313 15:26:17.245545 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/edb521bb-211e-48c0-b644-8dc44b235feb-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"edb521bb-211e-48c0-b644-8dc44b235feb\") " pod="openstack/openstack-cell1-galera-0" Mar 13 15:26:17 crc kubenswrapper[4907]: I0313 15:26:17.245597 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/edb521bb-211e-48c0-b644-8dc44b235feb-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"edb521bb-211e-48c0-b644-8dc44b235feb\") " pod="openstack/openstack-cell1-galera-0" Mar 13 15:26:17 crc kubenswrapper[4907]: I0313 15:26:17.245638 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pjp8c\" (UniqueName: \"kubernetes.io/projected/edb521bb-211e-48c0-b644-8dc44b235feb-kube-api-access-pjp8c\") pod \"openstack-cell1-galera-0\" (UID: \"edb521bb-211e-48c0-b644-8dc44b235feb\") " pod="openstack/openstack-cell1-galera-0" Mar 13 15:26:17 crc kubenswrapper[4907]: I0313 15:26:17.245656 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/edb521bb-211e-48c0-b644-8dc44b235feb-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"edb521bb-211e-48c0-b644-8dc44b235feb\") " pod="openstack/openstack-cell1-galera-0" Mar 13 15:26:17 crc kubenswrapper[4907]: I0313 15:26:17.245675 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/edb521bb-211e-48c0-b644-8dc44b235feb-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"edb521bb-211e-48c0-b644-8dc44b235feb\") " pod="openstack/openstack-cell1-galera-0" Mar 13 15:26:17 crc kubenswrapper[4907]: I0313 15:26:17.245699 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/edb521bb-211e-48c0-b644-8dc44b235feb-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"edb521bb-211e-48c0-b644-8dc44b235feb\") " pod="openstack/openstack-cell1-galera-0" Mar 13 15:26:17 crc kubenswrapper[4907]: I0313 15:26:17.245731 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/edb521bb-211e-48c0-b644-8dc44b235feb-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"edb521bb-211e-48c0-b644-8dc44b235feb\") " pod="openstack/openstack-cell1-galera-0" Mar 13 15:26:17 crc kubenswrapper[4907]: I0313 15:26:17.245804 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-3bdc965c-6ae3-457e-9054-e9bd61be219a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3bdc965c-6ae3-457e-9054-e9bd61be219a\") pod \"openstack-cell1-galera-0\" (UID: \"edb521bb-211e-48c0-b644-8dc44b235feb\") " pod="openstack/openstack-cell1-galera-0" Mar 13 15:26:17 crc kubenswrapper[4907]: I0313 15:26:17.245836 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edb521bb-211e-48c0-b644-8dc44b235feb-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"edb521bb-211e-48c0-b644-8dc44b235feb\") " pod="openstack/openstack-cell1-galera-0" Mar 13 15:26:17 crc kubenswrapper[4907]: I0313 15:26:17.246086 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/edb521bb-211e-48c0-b644-8dc44b235feb-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"edb521bb-211e-48c0-b644-8dc44b235feb\") " pod="openstack/openstack-cell1-galera-0" Mar 13 15:26:17 crc kubenswrapper[4907]: I0313 15:26:17.246304 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/edb521bb-211e-48c0-b644-8dc44b235feb-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"edb521bb-211e-48c0-b644-8dc44b235feb\") " pod="openstack/openstack-cell1-galera-0" Mar 13 15:26:17 crc kubenswrapper[4907]: I0313 15:26:17.247109 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/edb521bb-211e-48c0-b644-8dc44b235feb-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"edb521bb-211e-48c0-b644-8dc44b235feb\") " pod="openstack/openstack-cell1-galera-0" Mar 13 15:26:17 crc kubenswrapper[4907]: I0313 15:26:17.249825 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edb521bb-211e-48c0-b644-8dc44b235feb-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"edb521bb-211e-48c0-b644-8dc44b235feb\") " pod="openstack/openstack-cell1-galera-0" Mar 13 15:26:17 crc kubenswrapper[4907]: I0313 15:26:17.249934 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/edb521bb-211e-48c0-b644-8dc44b235feb-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"edb521bb-211e-48c0-b644-8dc44b235feb\") " pod="openstack/openstack-cell1-galera-0" Mar 13 15:26:17 crc kubenswrapper[4907]: I0313 15:26:17.251507 4907 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 13 15:26:17 crc kubenswrapper[4907]: I0313 15:26:17.251537 4907 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-3bdc965c-6ae3-457e-9054-e9bd61be219a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3bdc965c-6ae3-457e-9054-e9bd61be219a\") pod \"openstack-cell1-galera-0\" (UID: \"edb521bb-211e-48c0-b644-8dc44b235feb\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/bc929da5e54fcbc45e8b64e6c7cf70de64445330602309f6aca5ea888f5267cb/globalmount\"" pod="openstack/openstack-cell1-galera-0" Mar 13 15:26:17 crc kubenswrapper[4907]: I0313 15:26:17.263652 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pjp8c\" (UniqueName: \"kubernetes.io/projected/edb521bb-211e-48c0-b644-8dc44b235feb-kube-api-access-pjp8c\") pod \"openstack-cell1-galera-0\" (UID: \"edb521bb-211e-48c0-b644-8dc44b235feb\") " pod="openstack/openstack-cell1-galera-0" Mar 13 15:26:17 crc kubenswrapper[4907]: I0313 15:26:17.276480 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-3bdc965c-6ae3-457e-9054-e9bd61be219a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3bdc965c-6ae3-457e-9054-e9bd61be219a\") pod \"openstack-cell1-galera-0\" (UID: \"edb521bb-211e-48c0-b644-8dc44b235feb\") " pod="openstack/openstack-cell1-galera-0" Mar 13 15:26:17 crc kubenswrapper[4907]: I0313 15:26:17.383129 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Mar 13 15:26:17 crc kubenswrapper[4907]: I0313 15:26:17.828629 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Mar 13 15:26:17 crc kubenswrapper[4907]: W0313 15:26:17.839383 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podedb521bb_211e_48c0_b644_8dc44b235feb.slice/crio-0055e47f11b76f561cdcf0890df26a527b6eb47e8ee371daa37c8c6e62170061 WatchSource:0}: Error finding container 0055e47f11b76f561cdcf0890df26a527b6eb47e8ee371daa37c8c6e62170061: Status 404 returned error can't find the container with id 0055e47f11b76f561cdcf0890df26a527b6eb47e8ee371daa37c8c6e62170061 Mar 13 15:26:17 crc kubenswrapper[4907]: I0313 15:26:17.962376 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"edb521bb-211e-48c0-b644-8dc44b235feb","Type":"ContainerStarted","Data":"0055e47f11b76f561cdcf0890df26a527b6eb47e8ee371daa37c8c6e62170061"} Mar 13 15:26:17 crc kubenswrapper[4907]: I0313 15:26:17.963933 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"8c712e4f-4999-476b-970f-eae3189b4e59","Type":"ContainerStarted","Data":"72775b2f66df067aaded0627229fbd6c0b8d5f1a8301a47432b9b81d596998f8"} Mar 13 15:26:17 crc kubenswrapper[4907]: I0313 15:26:17.964817 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Mar 13 15:26:17 crc kubenswrapper[4907]: I0313 15:26:17.967781 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"585e3be7-1e17-443c-b430-0c03727167cb","Type":"ContainerStarted","Data":"f3b5304b835c93b9e62b5aa23ff48efa75463f10277227bd0d73800bbbe03088"} Mar 13 15:26:17 crc kubenswrapper[4907]: I0313 15:26:17.983662 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=2.983596097 podStartE2EDuration="2.983596097s" podCreationTimestamp="2026-03-13 15:26:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:26:17.980145642 +0000 UTC m=+4876.879933341" watchObservedRunningTime="2026-03-13 15:26:17.983596097 +0000 UTC m=+4876.883383776" Mar 13 15:26:18 crc kubenswrapper[4907]: I0313 15:26:18.041785 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 15:26:18 crc kubenswrapper[4907]: I0313 15:26:18.041851 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 15:26:18 crc kubenswrapper[4907]: I0313 15:26:18.979640 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"edb521bb-211e-48c0-b644-8dc44b235feb","Type":"ContainerStarted","Data":"0a58e80cfc5f151adf7f48c9420de170b24a2f153322cc550036bc388dca8316"} Mar 13 15:26:20 crc kubenswrapper[4907]: I0313 15:26:20.997150 4907 generic.go:334] "Generic (PLEG): container finished" podID="585e3be7-1e17-443c-b430-0c03727167cb" containerID="f3b5304b835c93b9e62b5aa23ff48efa75463f10277227bd0d73800bbbe03088" exitCode=0 Mar 13 15:26:20 crc kubenswrapper[4907]: I0313 15:26:20.997264 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"585e3be7-1e17-443c-b430-0c03727167cb","Type":"ContainerDied","Data":"f3b5304b835c93b9e62b5aa23ff48efa75463f10277227bd0d73800bbbe03088"} Mar 13 15:26:22 crc kubenswrapper[4907]: I0313 15:26:22.009272 4907 generic.go:334] "Generic (PLEG): container finished" podID="edb521bb-211e-48c0-b644-8dc44b235feb" containerID="0a58e80cfc5f151adf7f48c9420de170b24a2f153322cc550036bc388dca8316" exitCode=0 Mar 13 15:26:22 crc kubenswrapper[4907]: I0313 15:26:22.009382 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"edb521bb-211e-48c0-b644-8dc44b235feb","Type":"ContainerDied","Data":"0a58e80cfc5f151adf7f48c9420de170b24a2f153322cc550036bc388dca8316"} Mar 13 15:26:22 crc kubenswrapper[4907]: I0313 15:26:22.012684 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"585e3be7-1e17-443c-b430-0c03727167cb","Type":"ContainerStarted","Data":"8d2b4da29a65b464b3858d352aa2641c49967484e6c693a9d41c419714ea3020"} Mar 13 15:26:22 crc kubenswrapper[4907]: I0313 15:26:22.071464 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=8.071443857 podStartE2EDuration="8.071443857s" podCreationTimestamp="2026-03-13 15:26:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:26:22.061659299 +0000 UTC m=+4880.961446988" watchObservedRunningTime="2026-03-13 15:26:22.071443857 +0000 UTC m=+4880.971231546" Mar 13 15:26:23 crc kubenswrapper[4907]: I0313 15:26:23.020991 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"edb521bb-211e-48c0-b644-8dc44b235feb","Type":"ContainerStarted","Data":"775073a54525f805cb5e13eaea48657add5873f67ca298d457a515b2d40981c7"} Mar 13 15:26:23 crc kubenswrapper[4907]: I0313 15:26:23.049748 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=7.049726463 podStartE2EDuration="7.049726463s" podCreationTimestamp="2026-03-13 15:26:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:26:23.044343436 +0000 UTC m=+4881.944131145" watchObservedRunningTime="2026-03-13 15:26:23.049726463 +0000 UTC m=+4881.949514172" Mar 13 15:26:23 crc kubenswrapper[4907]: I0313 15:26:23.244167 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7fc7c884dc-98jxk" Mar 13 15:26:23 crc kubenswrapper[4907]: I0313 15:26:23.535494 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7c95686bd5-8kb7m" Mar 13 15:26:23 crc kubenswrapper[4907]: I0313 15:26:23.586971 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7fc7c884dc-98jxk"] Mar 13 15:26:24 crc kubenswrapper[4907]: I0313 15:26:24.026641 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7fc7c884dc-98jxk" podUID="2cdac990-bf65-48a3-8892-b375611f1e70" containerName="dnsmasq-dns" containerID="cri-o://4d2e0d92c2b302461c3a8b6d60213e7b81411578e2bc548ac86409412c045bf8" gracePeriod=10 Mar 13 15:26:24 crc kubenswrapper[4907]: I0313 15:26:24.426703 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fc7c884dc-98jxk" Mar 13 15:26:24 crc kubenswrapper[4907]: I0313 15:26:24.559158 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6c4r9\" (UniqueName: \"kubernetes.io/projected/2cdac990-bf65-48a3-8892-b375611f1e70-kube-api-access-6c4r9\") pod \"2cdac990-bf65-48a3-8892-b375611f1e70\" (UID: \"2cdac990-bf65-48a3-8892-b375611f1e70\") " Mar 13 15:26:24 crc kubenswrapper[4907]: I0313 15:26:24.559218 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2cdac990-bf65-48a3-8892-b375611f1e70-config\") pod \"2cdac990-bf65-48a3-8892-b375611f1e70\" (UID: \"2cdac990-bf65-48a3-8892-b375611f1e70\") " Mar 13 15:26:24 crc kubenswrapper[4907]: I0313 15:26:24.559266 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2cdac990-bf65-48a3-8892-b375611f1e70-dns-svc\") pod \"2cdac990-bf65-48a3-8892-b375611f1e70\" (UID: \"2cdac990-bf65-48a3-8892-b375611f1e70\") " Mar 13 15:26:24 crc kubenswrapper[4907]: I0313 15:26:24.564700 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2cdac990-bf65-48a3-8892-b375611f1e70-kube-api-access-6c4r9" (OuterVolumeSpecName: "kube-api-access-6c4r9") pod "2cdac990-bf65-48a3-8892-b375611f1e70" (UID: "2cdac990-bf65-48a3-8892-b375611f1e70"). InnerVolumeSpecName "kube-api-access-6c4r9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:26:24 crc kubenswrapper[4907]: I0313 15:26:24.599660 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2cdac990-bf65-48a3-8892-b375611f1e70-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "2cdac990-bf65-48a3-8892-b375611f1e70" (UID: "2cdac990-bf65-48a3-8892-b375611f1e70"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:26:24 crc kubenswrapper[4907]: I0313 15:26:24.599844 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2cdac990-bf65-48a3-8892-b375611f1e70-config" (OuterVolumeSpecName: "config") pod "2cdac990-bf65-48a3-8892-b375611f1e70" (UID: "2cdac990-bf65-48a3-8892-b375611f1e70"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:26:24 crc kubenswrapper[4907]: I0313 15:26:24.660659 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6c4r9\" (UniqueName: \"kubernetes.io/projected/2cdac990-bf65-48a3-8892-b375611f1e70-kube-api-access-6c4r9\") on node \"crc\" DevicePath \"\"" Mar 13 15:26:24 crc kubenswrapper[4907]: I0313 15:26:24.660869 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2cdac990-bf65-48a3-8892-b375611f1e70-config\") on node \"crc\" DevicePath \"\"" Mar 13 15:26:24 crc kubenswrapper[4907]: I0313 15:26:24.660990 4907 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2cdac990-bf65-48a3-8892-b375611f1e70-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 13 15:26:25 crc kubenswrapper[4907]: I0313 15:26:25.035175 4907 generic.go:334] "Generic (PLEG): container finished" podID="2cdac990-bf65-48a3-8892-b375611f1e70" containerID="4d2e0d92c2b302461c3a8b6d60213e7b81411578e2bc548ac86409412c045bf8" exitCode=0 Mar 13 15:26:25 crc kubenswrapper[4907]: I0313 15:26:25.035259 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fc7c884dc-98jxk" Mar 13 15:26:25 crc kubenswrapper[4907]: I0313 15:26:25.035245 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fc7c884dc-98jxk" event={"ID":"2cdac990-bf65-48a3-8892-b375611f1e70","Type":"ContainerDied","Data":"4d2e0d92c2b302461c3a8b6d60213e7b81411578e2bc548ac86409412c045bf8"} Mar 13 15:26:25 crc kubenswrapper[4907]: I0313 15:26:25.035460 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fc7c884dc-98jxk" event={"ID":"2cdac990-bf65-48a3-8892-b375611f1e70","Type":"ContainerDied","Data":"77a29a36e6edcd7453504c99bd8d6185ca0a664cb336729564e5ea2647a574ad"} Mar 13 15:26:25 crc kubenswrapper[4907]: I0313 15:26:25.035494 4907 scope.go:117] "RemoveContainer" containerID="4d2e0d92c2b302461c3a8b6d60213e7b81411578e2bc548ac86409412c045bf8" Mar 13 15:26:25 crc kubenswrapper[4907]: I0313 15:26:25.079136 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7fc7c884dc-98jxk"] Mar 13 15:26:25 crc kubenswrapper[4907]: I0313 15:26:25.083915 4907 scope.go:117] "RemoveContainer" containerID="380d0c10f65356a76638952335ee06865a37a23f94d9e43041a4522fb0545d73" Mar 13 15:26:25 crc kubenswrapper[4907]: I0313 15:26:25.085239 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7fc7c884dc-98jxk"] Mar 13 15:26:25 crc kubenswrapper[4907]: I0313 15:26:25.101061 4907 scope.go:117] "RemoveContainer" containerID="4d2e0d92c2b302461c3a8b6d60213e7b81411578e2bc548ac86409412c045bf8" Mar 13 15:26:25 crc kubenswrapper[4907]: E0313 15:26:25.101539 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4d2e0d92c2b302461c3a8b6d60213e7b81411578e2bc548ac86409412c045bf8\": container with ID starting with 4d2e0d92c2b302461c3a8b6d60213e7b81411578e2bc548ac86409412c045bf8 not found: ID does not exist" containerID="4d2e0d92c2b302461c3a8b6d60213e7b81411578e2bc548ac86409412c045bf8" Mar 13 15:26:25 crc kubenswrapper[4907]: I0313 15:26:25.101577 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4d2e0d92c2b302461c3a8b6d60213e7b81411578e2bc548ac86409412c045bf8"} err="failed to get container status \"4d2e0d92c2b302461c3a8b6d60213e7b81411578e2bc548ac86409412c045bf8\": rpc error: code = NotFound desc = could not find container \"4d2e0d92c2b302461c3a8b6d60213e7b81411578e2bc548ac86409412c045bf8\": container with ID starting with 4d2e0d92c2b302461c3a8b6d60213e7b81411578e2bc548ac86409412c045bf8 not found: ID does not exist" Mar 13 15:26:25 crc kubenswrapper[4907]: I0313 15:26:25.101603 4907 scope.go:117] "RemoveContainer" containerID="380d0c10f65356a76638952335ee06865a37a23f94d9e43041a4522fb0545d73" Mar 13 15:26:25 crc kubenswrapper[4907]: E0313 15:26:25.101895 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"380d0c10f65356a76638952335ee06865a37a23f94d9e43041a4522fb0545d73\": container with ID starting with 380d0c10f65356a76638952335ee06865a37a23f94d9e43041a4522fb0545d73 not found: ID does not exist" containerID="380d0c10f65356a76638952335ee06865a37a23f94d9e43041a4522fb0545d73" Mar 13 15:26:25 crc kubenswrapper[4907]: I0313 15:26:25.101922 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"380d0c10f65356a76638952335ee06865a37a23f94d9e43041a4522fb0545d73"} err="failed to get container status \"380d0c10f65356a76638952335ee06865a37a23f94d9e43041a4522fb0545d73\": rpc error: code = NotFound desc = could not find container \"380d0c10f65356a76638952335ee06865a37a23f94d9e43041a4522fb0545d73\": container with ID starting with 380d0c10f65356a76638952335ee06865a37a23f94d9e43041a4522fb0545d73 not found: ID does not exist" Mar 13 15:26:25 crc kubenswrapper[4907]: I0313 15:26:25.791940 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2cdac990-bf65-48a3-8892-b375611f1e70" path="/var/lib/kubelet/pods/2cdac990-bf65-48a3-8892-b375611f1e70/volumes" Mar 13 15:26:26 crc kubenswrapper[4907]: I0313 15:26:26.252360 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Mar 13 15:26:26 crc kubenswrapper[4907]: I0313 15:26:26.413218 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Mar 13 15:26:26 crc kubenswrapper[4907]: I0313 15:26:26.413273 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Mar 13 15:26:26 crc kubenswrapper[4907]: I0313 15:26:26.488754 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Mar 13 15:26:26 crc kubenswrapper[4907]: E0313 15:26:26.643016 4907 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.174:56862->38.102.83.174:33349: write tcp 38.102.83.174:56862->38.102.83.174:33349: write: connection reset by peer Mar 13 15:26:27 crc kubenswrapper[4907]: I0313 15:26:27.122622 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Mar 13 15:26:27 crc kubenswrapper[4907]: I0313 15:26:27.289929 4907 scope.go:117] "RemoveContainer" containerID="e5083f5e7cb4434ec8bff6d3c9df8bdd433195c0fc9dbe495454864580276a9b" Mar 13 15:26:27 crc kubenswrapper[4907]: I0313 15:26:27.384149 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Mar 13 15:26:27 crc kubenswrapper[4907]: I0313 15:26:27.384210 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Mar 13 15:26:29 crc kubenswrapper[4907]: I0313 15:26:29.675068 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Mar 13 15:26:29 crc kubenswrapper[4907]: I0313 15:26:29.764744 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Mar 13 15:26:34 crc kubenswrapper[4907]: I0313 15:26:34.448591 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/root-account-create-update-2gzz9"] Mar 13 15:26:34 crc kubenswrapper[4907]: E0313 15:26:34.449486 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2cdac990-bf65-48a3-8892-b375611f1e70" containerName="init" Mar 13 15:26:34 crc kubenswrapper[4907]: I0313 15:26:34.449504 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="2cdac990-bf65-48a3-8892-b375611f1e70" containerName="init" Mar 13 15:26:34 crc kubenswrapper[4907]: E0313 15:26:34.449524 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2cdac990-bf65-48a3-8892-b375611f1e70" containerName="dnsmasq-dns" Mar 13 15:26:34 crc kubenswrapper[4907]: I0313 15:26:34.449532 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="2cdac990-bf65-48a3-8892-b375611f1e70" containerName="dnsmasq-dns" Mar 13 15:26:34 crc kubenswrapper[4907]: I0313 15:26:34.449742 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="2cdac990-bf65-48a3-8892-b375611f1e70" containerName="dnsmasq-dns" Mar 13 15:26:34 crc kubenswrapper[4907]: I0313 15:26:34.450344 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-2gzz9" Mar 13 15:26:34 crc kubenswrapper[4907]: I0313 15:26:34.453312 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-mariadb-root-db-secret" Mar 13 15:26:34 crc kubenswrapper[4907]: I0313 15:26:34.455841 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-2gzz9"] Mar 13 15:26:34 crc kubenswrapper[4907]: I0313 15:26:34.519803 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b0ad87e8-d3b1-4f00-8f3d-b3b95e9be127-operator-scripts\") pod \"root-account-create-update-2gzz9\" (UID: \"b0ad87e8-d3b1-4f00-8f3d-b3b95e9be127\") " pod="openstack/root-account-create-update-2gzz9" Mar 13 15:26:34 crc kubenswrapper[4907]: I0313 15:26:34.519968 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t52tb\" (UniqueName: \"kubernetes.io/projected/b0ad87e8-d3b1-4f00-8f3d-b3b95e9be127-kube-api-access-t52tb\") pod \"root-account-create-update-2gzz9\" (UID: \"b0ad87e8-d3b1-4f00-8f3d-b3b95e9be127\") " pod="openstack/root-account-create-update-2gzz9" Mar 13 15:26:34 crc kubenswrapper[4907]: I0313 15:26:34.621680 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b0ad87e8-d3b1-4f00-8f3d-b3b95e9be127-operator-scripts\") pod \"root-account-create-update-2gzz9\" (UID: \"b0ad87e8-d3b1-4f00-8f3d-b3b95e9be127\") " pod="openstack/root-account-create-update-2gzz9" Mar 13 15:26:34 crc kubenswrapper[4907]: I0313 15:26:34.621851 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t52tb\" (UniqueName: \"kubernetes.io/projected/b0ad87e8-d3b1-4f00-8f3d-b3b95e9be127-kube-api-access-t52tb\") pod \"root-account-create-update-2gzz9\" (UID: \"b0ad87e8-d3b1-4f00-8f3d-b3b95e9be127\") " pod="openstack/root-account-create-update-2gzz9" Mar 13 15:26:34 crc kubenswrapper[4907]: I0313 15:26:34.622720 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b0ad87e8-d3b1-4f00-8f3d-b3b95e9be127-operator-scripts\") pod \"root-account-create-update-2gzz9\" (UID: \"b0ad87e8-d3b1-4f00-8f3d-b3b95e9be127\") " pod="openstack/root-account-create-update-2gzz9" Mar 13 15:26:34 crc kubenswrapper[4907]: I0313 15:26:34.643246 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t52tb\" (UniqueName: \"kubernetes.io/projected/b0ad87e8-d3b1-4f00-8f3d-b3b95e9be127-kube-api-access-t52tb\") pod \"root-account-create-update-2gzz9\" (UID: \"b0ad87e8-d3b1-4f00-8f3d-b3b95e9be127\") " pod="openstack/root-account-create-update-2gzz9" Mar 13 15:26:34 crc kubenswrapper[4907]: I0313 15:26:34.770497 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-2gzz9" Mar 13 15:26:35 crc kubenswrapper[4907]: I0313 15:26:35.184226 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-2gzz9"] Mar 13 15:26:35 crc kubenswrapper[4907]: W0313 15:26:35.195368 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb0ad87e8_d3b1_4f00_8f3d_b3b95e9be127.slice/crio-ea9cce4ceae5a244b75b04140ee1e964bc68e196b8c77a4abefc008e564e1a4b WatchSource:0}: Error finding container ea9cce4ceae5a244b75b04140ee1e964bc68e196b8c77a4abefc008e564e1a4b: Status 404 returned error can't find the container with id ea9cce4ceae5a244b75b04140ee1e964bc68e196b8c77a4abefc008e564e1a4b Mar 13 15:26:36 crc kubenswrapper[4907]: I0313 15:26:36.118932 4907 generic.go:334] "Generic (PLEG): container finished" podID="b0ad87e8-d3b1-4f00-8f3d-b3b95e9be127" containerID="8f8264e04faa67742d13e20b2a8a6a04d3b6912538ef54a5202a7e841e98d29e" exitCode=0 Mar 13 15:26:36 crc kubenswrapper[4907]: I0313 15:26:36.119067 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-2gzz9" event={"ID":"b0ad87e8-d3b1-4f00-8f3d-b3b95e9be127","Type":"ContainerDied","Data":"8f8264e04faa67742d13e20b2a8a6a04d3b6912538ef54a5202a7e841e98d29e"} Mar 13 15:26:36 crc kubenswrapper[4907]: I0313 15:26:36.119263 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-2gzz9" event={"ID":"b0ad87e8-d3b1-4f00-8f3d-b3b95e9be127","Type":"ContainerStarted","Data":"ea9cce4ceae5a244b75b04140ee1e964bc68e196b8c77a4abefc008e564e1a4b"} Mar 13 15:26:37 crc kubenswrapper[4907]: I0313 15:26:37.389310 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-2gzz9" Mar 13 15:26:37 crc kubenswrapper[4907]: I0313 15:26:37.560269 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b0ad87e8-d3b1-4f00-8f3d-b3b95e9be127-operator-scripts\") pod \"b0ad87e8-d3b1-4f00-8f3d-b3b95e9be127\" (UID: \"b0ad87e8-d3b1-4f00-8f3d-b3b95e9be127\") " Mar 13 15:26:37 crc kubenswrapper[4907]: I0313 15:26:37.560446 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t52tb\" (UniqueName: \"kubernetes.io/projected/b0ad87e8-d3b1-4f00-8f3d-b3b95e9be127-kube-api-access-t52tb\") pod \"b0ad87e8-d3b1-4f00-8f3d-b3b95e9be127\" (UID: \"b0ad87e8-d3b1-4f00-8f3d-b3b95e9be127\") " Mar 13 15:26:37 crc kubenswrapper[4907]: I0313 15:26:37.561108 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b0ad87e8-d3b1-4f00-8f3d-b3b95e9be127-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b0ad87e8-d3b1-4f00-8f3d-b3b95e9be127" (UID: "b0ad87e8-d3b1-4f00-8f3d-b3b95e9be127"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:26:37 crc kubenswrapper[4907]: I0313 15:26:37.571754 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b0ad87e8-d3b1-4f00-8f3d-b3b95e9be127-kube-api-access-t52tb" (OuterVolumeSpecName: "kube-api-access-t52tb") pod "b0ad87e8-d3b1-4f00-8f3d-b3b95e9be127" (UID: "b0ad87e8-d3b1-4f00-8f3d-b3b95e9be127"). InnerVolumeSpecName "kube-api-access-t52tb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:26:37 crc kubenswrapper[4907]: I0313 15:26:37.662499 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t52tb\" (UniqueName: \"kubernetes.io/projected/b0ad87e8-d3b1-4f00-8f3d-b3b95e9be127-kube-api-access-t52tb\") on node \"crc\" DevicePath \"\"" Mar 13 15:26:37 crc kubenswrapper[4907]: I0313 15:26:37.662532 4907 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b0ad87e8-d3b1-4f00-8f3d-b3b95e9be127-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 15:26:38 crc kubenswrapper[4907]: I0313 15:26:38.138513 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-2gzz9" event={"ID":"b0ad87e8-d3b1-4f00-8f3d-b3b95e9be127","Type":"ContainerDied","Data":"ea9cce4ceae5a244b75b04140ee1e964bc68e196b8c77a4abefc008e564e1a4b"} Mar 13 15:26:38 crc kubenswrapper[4907]: I0313 15:26:38.138739 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ea9cce4ceae5a244b75b04140ee1e964bc68e196b8c77a4abefc008e564e1a4b" Mar 13 15:26:38 crc kubenswrapper[4907]: I0313 15:26:38.138598 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-2gzz9" Mar 13 15:26:41 crc kubenswrapper[4907]: I0313 15:26:41.094610 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/root-account-create-update-2gzz9"] Mar 13 15:26:41 crc kubenswrapper[4907]: I0313 15:26:41.102337 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/root-account-create-update-2gzz9"] Mar 13 15:26:41 crc kubenswrapper[4907]: I0313 15:26:41.799569 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b0ad87e8-d3b1-4f00-8f3d-b3b95e9be127" path="/var/lib/kubelet/pods/b0ad87e8-d3b1-4f00-8f3d-b3b95e9be127/volumes" Mar 13 15:26:46 crc kubenswrapper[4907]: I0313 15:26:46.097060 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/root-account-create-update-l8vbt"] Mar 13 15:26:46 crc kubenswrapper[4907]: E0313 15:26:46.097605 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0ad87e8-d3b1-4f00-8f3d-b3b95e9be127" containerName="mariadb-account-create-update" Mar 13 15:26:46 crc kubenswrapper[4907]: I0313 15:26:46.097618 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0ad87e8-d3b1-4f00-8f3d-b3b95e9be127" containerName="mariadb-account-create-update" Mar 13 15:26:46 crc kubenswrapper[4907]: I0313 15:26:46.097803 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="b0ad87e8-d3b1-4f00-8f3d-b3b95e9be127" containerName="mariadb-account-create-update" Mar 13 15:26:46 crc kubenswrapper[4907]: I0313 15:26:46.098534 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-l8vbt" Mar 13 15:26:46 crc kubenswrapper[4907]: I0313 15:26:46.101418 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-mariadb-root-db-secret" Mar 13 15:26:46 crc kubenswrapper[4907]: I0313 15:26:46.148191 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-l8vbt"] Mar 13 15:26:46 crc kubenswrapper[4907]: I0313 15:26:46.196543 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-plc22\" (UniqueName: \"kubernetes.io/projected/1d5a861d-b4ca-4936-9fac-5aa23fcaa317-kube-api-access-plc22\") pod \"root-account-create-update-l8vbt\" (UID: \"1d5a861d-b4ca-4936-9fac-5aa23fcaa317\") " pod="openstack/root-account-create-update-l8vbt" Mar 13 15:26:46 crc kubenswrapper[4907]: I0313 15:26:46.196606 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1d5a861d-b4ca-4936-9fac-5aa23fcaa317-operator-scripts\") pod \"root-account-create-update-l8vbt\" (UID: \"1d5a861d-b4ca-4936-9fac-5aa23fcaa317\") " pod="openstack/root-account-create-update-l8vbt" Mar 13 15:26:46 crc kubenswrapper[4907]: I0313 15:26:46.298441 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-plc22\" (UniqueName: \"kubernetes.io/projected/1d5a861d-b4ca-4936-9fac-5aa23fcaa317-kube-api-access-plc22\") pod \"root-account-create-update-l8vbt\" (UID: \"1d5a861d-b4ca-4936-9fac-5aa23fcaa317\") " pod="openstack/root-account-create-update-l8vbt" Mar 13 15:26:46 crc kubenswrapper[4907]: I0313 15:26:46.298516 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1d5a861d-b4ca-4936-9fac-5aa23fcaa317-operator-scripts\") pod \"root-account-create-update-l8vbt\" (UID: \"1d5a861d-b4ca-4936-9fac-5aa23fcaa317\") " pod="openstack/root-account-create-update-l8vbt" Mar 13 15:26:46 crc kubenswrapper[4907]: I0313 15:26:46.299428 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1d5a861d-b4ca-4936-9fac-5aa23fcaa317-operator-scripts\") pod \"root-account-create-update-l8vbt\" (UID: \"1d5a861d-b4ca-4936-9fac-5aa23fcaa317\") " pod="openstack/root-account-create-update-l8vbt" Mar 13 15:26:46 crc kubenswrapper[4907]: I0313 15:26:46.320373 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-plc22\" (UniqueName: \"kubernetes.io/projected/1d5a861d-b4ca-4936-9fac-5aa23fcaa317-kube-api-access-plc22\") pod \"root-account-create-update-l8vbt\" (UID: \"1d5a861d-b4ca-4936-9fac-5aa23fcaa317\") " pod="openstack/root-account-create-update-l8vbt" Mar 13 15:26:46 crc kubenswrapper[4907]: I0313 15:26:46.431120 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-l8vbt" Mar 13 15:26:46 crc kubenswrapper[4907]: I0313 15:26:46.863367 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-l8vbt"] Mar 13 15:26:47 crc kubenswrapper[4907]: I0313 15:26:47.240580 4907 generic.go:334] "Generic (PLEG): container finished" podID="1d5a861d-b4ca-4936-9fac-5aa23fcaa317" containerID="bddac8c66ea0090f3f484cd37e970a8a28f550e7e44321363049a8343bf92762" exitCode=0 Mar 13 15:26:47 crc kubenswrapper[4907]: I0313 15:26:47.240715 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-l8vbt" event={"ID":"1d5a861d-b4ca-4936-9fac-5aa23fcaa317","Type":"ContainerDied","Data":"bddac8c66ea0090f3f484cd37e970a8a28f550e7e44321363049a8343bf92762"} Mar 13 15:26:47 crc kubenswrapper[4907]: I0313 15:26:47.241065 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-l8vbt" event={"ID":"1d5a861d-b4ca-4936-9fac-5aa23fcaa317","Type":"ContainerStarted","Data":"e1eea7ae7248a157798b49da39f331a2cd401dacad63dfa779a01db360a7bbda"} Mar 13 15:26:48 crc kubenswrapper[4907]: I0313 15:26:48.041851 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 15:26:48 crc kubenswrapper[4907]: I0313 15:26:48.041979 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 15:26:48 crc kubenswrapper[4907]: I0313 15:26:48.042059 4907 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" Mar 13 15:26:48 crc kubenswrapper[4907]: I0313 15:26:48.043224 4907 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b5bb3cf2979988d507615c0a856eebc9a02c4d2c3275ceb2cee45dcc6b625cfd"} pod="openshift-machine-config-operator/machine-config-daemon-hm56j" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 15:26:48 crc kubenswrapper[4907]: I0313 15:26:48.043360 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" containerID="cri-o://b5bb3cf2979988d507615c0a856eebc9a02c4d2c3275ceb2cee45dcc6b625cfd" gracePeriod=600 Mar 13 15:26:48 crc kubenswrapper[4907]: I0313 15:26:48.263348 4907 generic.go:334] "Generic (PLEG): container finished" podID="7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d" containerID="bf96e8e63801badfacccb754ff56342cdc5ba8417a4f20f022b8a5f76d1c9272" exitCode=0 Mar 13 15:26:48 crc kubenswrapper[4907]: I0313 15:26:48.263621 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d","Type":"ContainerDied","Data":"bf96e8e63801badfacccb754ff56342cdc5ba8417a4f20f022b8a5f76d1c9272"} Mar 13 15:26:48 crc kubenswrapper[4907]: I0313 15:26:48.268238 4907 generic.go:334] "Generic (PLEG): container finished" podID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerID="b5bb3cf2979988d507615c0a856eebc9a02c4d2c3275ceb2cee45dcc6b625cfd" exitCode=0 Mar 13 15:26:48 crc kubenswrapper[4907]: I0313 15:26:48.268298 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerDied","Data":"b5bb3cf2979988d507615c0a856eebc9a02c4d2c3275ceb2cee45dcc6b625cfd"} Mar 13 15:26:48 crc kubenswrapper[4907]: I0313 15:26:48.268346 4907 scope.go:117] "RemoveContainer" containerID="dd934d7c5fc7e95362f79c2301ff40920ce7e3112722cbc11a053a56c085e05c" Mar 13 15:26:48 crc kubenswrapper[4907]: I0313 15:26:48.497644 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-l8vbt" Mar 13 15:26:48 crc kubenswrapper[4907]: I0313 15:26:48.533306 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-plc22\" (UniqueName: \"kubernetes.io/projected/1d5a861d-b4ca-4936-9fac-5aa23fcaa317-kube-api-access-plc22\") pod \"1d5a861d-b4ca-4936-9fac-5aa23fcaa317\" (UID: \"1d5a861d-b4ca-4936-9fac-5aa23fcaa317\") " Mar 13 15:26:48 crc kubenswrapper[4907]: I0313 15:26:48.533466 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1d5a861d-b4ca-4936-9fac-5aa23fcaa317-operator-scripts\") pod \"1d5a861d-b4ca-4936-9fac-5aa23fcaa317\" (UID: \"1d5a861d-b4ca-4936-9fac-5aa23fcaa317\") " Mar 13 15:26:48 crc kubenswrapper[4907]: I0313 15:26:48.533849 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1d5a861d-b4ca-4936-9fac-5aa23fcaa317-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "1d5a861d-b4ca-4936-9fac-5aa23fcaa317" (UID: "1d5a861d-b4ca-4936-9fac-5aa23fcaa317"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:26:48 crc kubenswrapper[4907]: I0313 15:26:48.538021 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d5a861d-b4ca-4936-9fac-5aa23fcaa317-kube-api-access-plc22" (OuterVolumeSpecName: "kube-api-access-plc22") pod "1d5a861d-b4ca-4936-9fac-5aa23fcaa317" (UID: "1d5a861d-b4ca-4936-9fac-5aa23fcaa317"). InnerVolumeSpecName "kube-api-access-plc22". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:26:48 crc kubenswrapper[4907]: I0313 15:26:48.635267 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-plc22\" (UniqueName: \"kubernetes.io/projected/1d5a861d-b4ca-4936-9fac-5aa23fcaa317-kube-api-access-plc22\") on node \"crc\" DevicePath \"\"" Mar 13 15:26:48 crc kubenswrapper[4907]: I0313 15:26:48.635299 4907 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1d5a861d-b4ca-4936-9fac-5aa23fcaa317-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 15:26:49 crc kubenswrapper[4907]: I0313 15:26:49.280784 4907 generic.go:334] "Generic (PLEG): container finished" podID="75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd" containerID="0b4af08a62b26495d3442675a1ce03aa549cddc0c7a28e8f6cb98a7deb792d53" exitCode=0 Mar 13 15:26:49 crc kubenswrapper[4907]: I0313 15:26:49.280933 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd","Type":"ContainerDied","Data":"0b4af08a62b26495d3442675a1ce03aa549cddc0c7a28e8f6cb98a7deb792d53"} Mar 13 15:26:49 crc kubenswrapper[4907]: I0313 15:26:49.284376 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d","Type":"ContainerStarted","Data":"c2763a983f63aa1fe0a38c4e94c194af6a54549cd715d4e926bfe5396d362f36"} Mar 13 15:26:49 crc kubenswrapper[4907]: I0313 15:26:49.284583 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Mar 13 15:26:49 crc kubenswrapper[4907]: I0313 15:26:49.288523 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-l8vbt" Mar 13 15:26:49 crc kubenswrapper[4907]: I0313 15:26:49.288646 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-l8vbt" event={"ID":"1d5a861d-b4ca-4936-9fac-5aa23fcaa317","Type":"ContainerDied","Data":"e1eea7ae7248a157798b49da39f331a2cd401dacad63dfa779a01db360a7bbda"} Mar 13 15:26:49 crc kubenswrapper[4907]: I0313 15:26:49.288673 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e1eea7ae7248a157798b49da39f331a2cd401dacad63dfa779a01db360a7bbda" Mar 13 15:26:49 crc kubenswrapper[4907]: I0313 15:26:49.292181 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerStarted","Data":"592bf9f175f440e943369aab2abbb13ad2c26412cb0e0738c4a33a156b8a5596"} Mar 13 15:26:49 crc kubenswrapper[4907]: I0313 15:26:49.371111 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=36.3710826 podStartE2EDuration="36.3710826s" podCreationTimestamp="2026-03-13 15:26:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:26:49.365643401 +0000 UTC m=+4908.265431090" watchObservedRunningTime="2026-03-13 15:26:49.3710826 +0000 UTC m=+4908.270870299" Mar 13 15:26:50 crc kubenswrapper[4907]: I0313 15:26:50.299899 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd","Type":"ContainerStarted","Data":"83c101cd6c30ca361c7f00d6a637893118fddd69105f31e3b1c72e730b52960a"} Mar 13 15:26:50 crc kubenswrapper[4907]: I0313 15:26:50.300355 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:26:50 crc kubenswrapper[4907]: I0313 15:26:50.318923 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=37.318904363 podStartE2EDuration="37.318904363s" podCreationTimestamp="2026-03-13 15:26:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:26:50.317532926 +0000 UTC m=+4909.217320615" watchObservedRunningTime="2026-03-13 15:26:50.318904363 +0000 UTC m=+4909.218692052" Mar 13 15:27:04 crc kubenswrapper[4907]: I0313 15:27:04.472504 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Mar 13 15:27:04 crc kubenswrapper[4907]: I0313 15:27:04.723712 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:27:10 crc kubenswrapper[4907]: I0313 15:27:10.179088 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-684c864bc9-p4j95"] Mar 13 15:27:10 crc kubenswrapper[4907]: E0313 15:27:10.179957 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d5a861d-b4ca-4936-9fac-5aa23fcaa317" containerName="mariadb-account-create-update" Mar 13 15:27:10 crc kubenswrapper[4907]: I0313 15:27:10.179975 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d5a861d-b4ca-4936-9fac-5aa23fcaa317" containerName="mariadb-account-create-update" Mar 13 15:27:10 crc kubenswrapper[4907]: I0313 15:27:10.180196 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d5a861d-b4ca-4936-9fac-5aa23fcaa317" containerName="mariadb-account-create-update" Mar 13 15:27:10 crc kubenswrapper[4907]: I0313 15:27:10.181085 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-684c864bc9-p4j95" Mar 13 15:27:10 crc kubenswrapper[4907]: I0313 15:27:10.198010 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-684c864bc9-p4j95"] Mar 13 15:27:10 crc kubenswrapper[4907]: I0313 15:27:10.362832 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rhz6p\" (UniqueName: \"kubernetes.io/projected/6c367cd2-6ece-429d-a930-c1ec11175248-kube-api-access-rhz6p\") pod \"dnsmasq-dns-684c864bc9-p4j95\" (UID: \"6c367cd2-6ece-429d-a930-c1ec11175248\") " pod="openstack/dnsmasq-dns-684c864bc9-p4j95" Mar 13 15:27:10 crc kubenswrapper[4907]: I0313 15:27:10.362909 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6c367cd2-6ece-429d-a930-c1ec11175248-config\") pod \"dnsmasq-dns-684c864bc9-p4j95\" (UID: \"6c367cd2-6ece-429d-a930-c1ec11175248\") " pod="openstack/dnsmasq-dns-684c864bc9-p4j95" Mar 13 15:27:10 crc kubenswrapper[4907]: I0313 15:27:10.363254 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6c367cd2-6ece-429d-a930-c1ec11175248-dns-svc\") pod \"dnsmasq-dns-684c864bc9-p4j95\" (UID: \"6c367cd2-6ece-429d-a930-c1ec11175248\") " pod="openstack/dnsmasq-dns-684c864bc9-p4j95" Mar 13 15:27:10 crc kubenswrapper[4907]: I0313 15:27:10.464847 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6c367cd2-6ece-429d-a930-c1ec11175248-dns-svc\") pod \"dnsmasq-dns-684c864bc9-p4j95\" (UID: \"6c367cd2-6ece-429d-a930-c1ec11175248\") " pod="openstack/dnsmasq-dns-684c864bc9-p4j95" Mar 13 15:27:10 crc kubenswrapper[4907]: I0313 15:27:10.465021 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rhz6p\" (UniqueName: \"kubernetes.io/projected/6c367cd2-6ece-429d-a930-c1ec11175248-kube-api-access-rhz6p\") pod \"dnsmasq-dns-684c864bc9-p4j95\" (UID: \"6c367cd2-6ece-429d-a930-c1ec11175248\") " pod="openstack/dnsmasq-dns-684c864bc9-p4j95" Mar 13 15:27:10 crc kubenswrapper[4907]: I0313 15:27:10.465053 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6c367cd2-6ece-429d-a930-c1ec11175248-config\") pod \"dnsmasq-dns-684c864bc9-p4j95\" (UID: \"6c367cd2-6ece-429d-a930-c1ec11175248\") " pod="openstack/dnsmasq-dns-684c864bc9-p4j95" Mar 13 15:27:10 crc kubenswrapper[4907]: I0313 15:27:10.466424 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6c367cd2-6ece-429d-a930-c1ec11175248-config\") pod \"dnsmasq-dns-684c864bc9-p4j95\" (UID: \"6c367cd2-6ece-429d-a930-c1ec11175248\") " pod="openstack/dnsmasq-dns-684c864bc9-p4j95" Mar 13 15:27:10 crc kubenswrapper[4907]: I0313 15:27:10.466479 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6c367cd2-6ece-429d-a930-c1ec11175248-dns-svc\") pod \"dnsmasq-dns-684c864bc9-p4j95\" (UID: \"6c367cd2-6ece-429d-a930-c1ec11175248\") " pod="openstack/dnsmasq-dns-684c864bc9-p4j95" Mar 13 15:27:10 crc kubenswrapper[4907]: I0313 15:27:10.492429 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rhz6p\" (UniqueName: \"kubernetes.io/projected/6c367cd2-6ece-429d-a930-c1ec11175248-kube-api-access-rhz6p\") pod \"dnsmasq-dns-684c864bc9-p4j95\" (UID: \"6c367cd2-6ece-429d-a930-c1ec11175248\") " pod="openstack/dnsmasq-dns-684c864bc9-p4j95" Mar 13 15:27:10 crc kubenswrapper[4907]: I0313 15:27:10.500192 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-684c864bc9-p4j95" Mar 13 15:27:10 crc kubenswrapper[4907]: I0313 15:27:10.940809 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-684c864bc9-p4j95"] Mar 13 15:27:11 crc kubenswrapper[4907]: I0313 15:27:11.439747 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Mar 13 15:27:11 crc kubenswrapper[4907]: I0313 15:27:11.481050 4907 generic.go:334] "Generic (PLEG): container finished" podID="6c367cd2-6ece-429d-a930-c1ec11175248" containerID="dfbeef34d4d9c1c564438050b366f2c5cc11365a8117cb19955310b23bc22e05" exitCode=0 Mar 13 15:27:11 crc kubenswrapper[4907]: I0313 15:27:11.481352 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-684c864bc9-p4j95" event={"ID":"6c367cd2-6ece-429d-a930-c1ec11175248","Type":"ContainerDied","Data":"dfbeef34d4d9c1c564438050b366f2c5cc11365a8117cb19955310b23bc22e05"} Mar 13 15:27:11 crc kubenswrapper[4907]: I0313 15:27:11.482015 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-684c864bc9-p4j95" event={"ID":"6c367cd2-6ece-429d-a930-c1ec11175248","Type":"ContainerStarted","Data":"bc53c9c5cde7280652faad4d880d4ee4a2f7625ff891a79d0f242d66a55b312a"} Mar 13 15:27:11 crc kubenswrapper[4907]: I0313 15:27:11.492740 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 13 15:27:12 crc kubenswrapper[4907]: I0313 15:27:12.517854 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-684c864bc9-p4j95" event={"ID":"6c367cd2-6ece-429d-a930-c1ec11175248","Type":"ContainerStarted","Data":"1b6e2c3bc8a46bfa72edd14d9923048107755f57239082f4595ce1f2d9b0f3a2"} Mar 13 15:27:12 crc kubenswrapper[4907]: I0313 15:27:12.541955 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-684c864bc9-p4j95" podStartSLOduration=2.5419383939999998 podStartE2EDuration="2.541938394s" podCreationTimestamp="2026-03-13 15:27:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:27:12.536272549 +0000 UTC m=+4931.436060258" watchObservedRunningTime="2026-03-13 15:27:12.541938394 +0000 UTC m=+4931.441726083" Mar 13 15:27:13 crc kubenswrapper[4907]: I0313 15:27:13.172982 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d" containerName="rabbitmq" containerID="cri-o://c2763a983f63aa1fe0a38c4e94c194af6a54549cd715d4e926bfe5396d362f36" gracePeriod=604799 Mar 13 15:27:13 crc kubenswrapper[4907]: I0313 15:27:13.185204 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd" containerName="rabbitmq" containerID="cri-o://83c101cd6c30ca361c7f00d6a637893118fddd69105f31e3b1c72e730b52960a" gracePeriod=604799 Mar 13 15:27:13 crc kubenswrapper[4907]: I0313 15:27:13.525963 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-684c864bc9-p4j95" Mar 13 15:27:14 crc kubenswrapper[4907]: I0313 15:27:14.470905 4907 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.1.22:5672: connect: connection refused" Mar 13 15:27:14 crc kubenswrapper[4907]: I0313 15:27:14.722199 4907 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.1.23:5672: connect: connection refused" Mar 13 15:27:19 crc kubenswrapper[4907]: I0313 15:27:19.603868 4907 generic.go:334] "Generic (PLEG): container finished" podID="75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd" containerID="83c101cd6c30ca361c7f00d6a637893118fddd69105f31e3b1c72e730b52960a" exitCode=0 Mar 13 15:27:19 crc kubenswrapper[4907]: I0313 15:27:19.603922 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd","Type":"ContainerDied","Data":"83c101cd6c30ca361c7f00d6a637893118fddd69105f31e3b1c72e730b52960a"} Mar 13 15:27:19 crc kubenswrapper[4907]: I0313 15:27:19.604462 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd","Type":"ContainerDied","Data":"24618a7621fc23ef2173f3056fdb4f82fc6d9f9738f1b8f67a17813327203788"} Mar 13 15:27:19 crc kubenswrapper[4907]: I0313 15:27:19.604481 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="24618a7621fc23ef2173f3056fdb4f82fc6d9f9738f1b8f67a17813327203788" Mar 13 15:27:19 crc kubenswrapper[4907]: I0313 15:27:19.608276 4907 generic.go:334] "Generic (PLEG): container finished" podID="7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d" containerID="c2763a983f63aa1fe0a38c4e94c194af6a54549cd715d4e926bfe5396d362f36" exitCode=0 Mar 13 15:27:19 crc kubenswrapper[4907]: I0313 15:27:19.608322 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d","Type":"ContainerDied","Data":"c2763a983f63aa1fe0a38c4e94c194af6a54549cd715d4e926bfe5396d362f36"} Mar 13 15:27:19 crc kubenswrapper[4907]: I0313 15:27:19.613819 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:27:19 crc kubenswrapper[4907]: I0313 15:27:19.726531 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd-plugins-conf\") pod \"75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd\" (UID: \"75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd\") " Mar 13 15:27:19 crc kubenswrapper[4907]: I0313 15:27:19.726610 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vhr7j\" (UniqueName: \"kubernetes.io/projected/75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd-kube-api-access-vhr7j\") pod \"75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd\" (UID: \"75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd\") " Mar 13 15:27:19 crc kubenswrapper[4907]: I0313 15:27:19.726749 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7c39547e-1168-404d-865b-13cfd1526f6f\") pod \"75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd\" (UID: \"75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd\") " Mar 13 15:27:19 crc kubenswrapper[4907]: I0313 15:27:19.726791 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd-rabbitmq-plugins\") pod \"75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd\" (UID: \"75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd\") " Mar 13 15:27:19 crc kubenswrapper[4907]: I0313 15:27:19.726815 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd-rabbitmq-confd\") pod \"75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd\" (UID: \"75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd\") " Mar 13 15:27:19 crc kubenswrapper[4907]: I0313 15:27:19.726862 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd-pod-info\") pod \"75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd\" (UID: \"75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd\") " Mar 13 15:27:19 crc kubenswrapper[4907]: I0313 15:27:19.726914 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd-rabbitmq-erlang-cookie\") pod \"75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd\" (UID: \"75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd\") " Mar 13 15:27:19 crc kubenswrapper[4907]: I0313 15:27:19.726944 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd-erlang-cookie-secret\") pod \"75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd\" (UID: \"75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd\") " Mar 13 15:27:19 crc kubenswrapper[4907]: I0313 15:27:19.726991 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd-server-conf\") pod \"75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd\" (UID: \"75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd\") " Mar 13 15:27:19 crc kubenswrapper[4907]: I0313 15:27:19.727484 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd" (UID: "75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:27:19 crc kubenswrapper[4907]: I0313 15:27:19.728020 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd" (UID: "75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:27:19 crc kubenswrapper[4907]: I0313 15:27:19.728300 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd" (UID: "75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:27:19 crc kubenswrapper[4907]: I0313 15:27:19.734099 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd" (UID: "75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:27:19 crc kubenswrapper[4907]: I0313 15:27:19.735017 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd-pod-info" (OuterVolumeSpecName: "pod-info") pod "75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd" (UID: "75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Mar 13 15:27:19 crc kubenswrapper[4907]: I0313 15:27:19.735103 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd-kube-api-access-vhr7j" (OuterVolumeSpecName: "kube-api-access-vhr7j") pod "75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd" (UID: "75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd"). InnerVolumeSpecName "kube-api-access-vhr7j". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:27:19 crc kubenswrapper[4907]: I0313 15:27:19.737604 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7c39547e-1168-404d-865b-13cfd1526f6f" (OuterVolumeSpecName: "persistence") pod "75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd" (UID: "75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd"). InnerVolumeSpecName "pvc-7c39547e-1168-404d-865b-13cfd1526f6f". PluginName "kubernetes.io/csi", VolumeGidValue "" Mar 13 15:27:19 crc kubenswrapper[4907]: I0313 15:27:19.761374 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd-server-conf" (OuterVolumeSpecName: "server-conf") pod "75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd" (UID: "75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:27:19 crc kubenswrapper[4907]: I0313 15:27:19.773424 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Mar 13 15:27:19 crc kubenswrapper[4907]: I0313 15:27:19.803602 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd" (UID: "75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:27:19 crc kubenswrapper[4907]: I0313 15:27:19.830450 4907 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd-plugins-conf\") on node \"crc\" DevicePath \"\"" Mar 13 15:27:19 crc kubenswrapper[4907]: I0313 15:27:19.830483 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vhr7j\" (UniqueName: \"kubernetes.io/projected/75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd-kube-api-access-vhr7j\") on node \"crc\" DevicePath \"\"" Mar 13 15:27:19 crc kubenswrapper[4907]: I0313 15:27:19.830526 4907 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-7c39547e-1168-404d-865b-13cfd1526f6f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7c39547e-1168-404d-865b-13cfd1526f6f\") on node \"crc\" " Mar 13 15:27:19 crc kubenswrapper[4907]: I0313 15:27:19.830543 4907 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Mar 13 15:27:19 crc kubenswrapper[4907]: I0313 15:27:19.830555 4907 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Mar 13 15:27:19 crc kubenswrapper[4907]: I0313 15:27:19.830567 4907 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd-pod-info\") on node \"crc\" DevicePath \"\"" Mar 13 15:27:19 crc kubenswrapper[4907]: I0313 15:27:19.830577 4907 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Mar 13 15:27:19 crc kubenswrapper[4907]: I0313 15:27:19.830590 4907 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Mar 13 15:27:19 crc kubenswrapper[4907]: I0313 15:27:19.830601 4907 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd-server-conf\") on node \"crc\" DevicePath \"\"" Mar 13 15:27:19 crc kubenswrapper[4907]: I0313 15:27:19.848270 4907 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Mar 13 15:27:19 crc kubenswrapper[4907]: I0313 15:27:19.848432 4907 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-7c39547e-1168-404d-865b-13cfd1526f6f" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7c39547e-1168-404d-865b-13cfd1526f6f") on node "crc" Mar 13 15:27:19 crc kubenswrapper[4907]: I0313 15:27:19.931252 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d-server-conf\") pod \"7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d\" (UID: \"7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d\") " Mar 13 15:27:19 crc kubenswrapper[4907]: I0313 15:27:19.931356 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d-pod-info\") pod \"7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d\" (UID: \"7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d\") " Mar 13 15:27:19 crc kubenswrapper[4907]: I0313 15:27:19.931405 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vsj5v\" (UniqueName: \"kubernetes.io/projected/7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d-kube-api-access-vsj5v\") pod \"7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d\" (UID: \"7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d\") " Mar 13 15:27:19 crc kubenswrapper[4907]: I0313 15:27:19.931431 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d-rabbitmq-confd\") pod \"7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d\" (UID: \"7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d\") " Mar 13 15:27:19 crc kubenswrapper[4907]: I0313 15:27:19.931497 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d-erlang-cookie-secret\") pod \"7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d\" (UID: \"7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d\") " Mar 13 15:27:19 crc kubenswrapper[4907]: I0313 15:27:19.931529 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d-rabbitmq-erlang-cookie\") pod \"7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d\" (UID: \"7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d\") " Mar 13 15:27:19 crc kubenswrapper[4907]: I0313 15:27:19.931632 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-9187efe6-871d-4bbe-9376-bff5d8a151f2\") pod \"7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d\" (UID: \"7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d\") " Mar 13 15:27:19 crc kubenswrapper[4907]: I0313 15:27:19.931665 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d-rabbitmq-plugins\") pod \"7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d\" (UID: \"7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d\") " Mar 13 15:27:19 crc kubenswrapper[4907]: I0313 15:27:19.931695 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d-plugins-conf\") pod \"7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d\" (UID: \"7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d\") " Mar 13 15:27:19 crc kubenswrapper[4907]: I0313 15:27:19.932034 4907 reconciler_common.go:293] "Volume detached for volume \"pvc-7c39547e-1168-404d-865b-13cfd1526f6f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7c39547e-1168-404d-865b-13cfd1526f6f\") on node \"crc\" DevicePath \"\"" Mar 13 15:27:19 crc kubenswrapper[4907]: I0313 15:27:19.932417 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d" (UID: "7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:27:19 crc kubenswrapper[4907]: I0313 15:27:19.932444 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d" (UID: "7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:27:19 crc kubenswrapper[4907]: I0313 15:27:19.932459 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d" (UID: "7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:27:19 crc kubenswrapper[4907]: I0313 15:27:19.935376 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d-pod-info" (OuterVolumeSpecName: "pod-info") pod "7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d" (UID: "7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Mar 13 15:27:19 crc kubenswrapper[4907]: I0313 15:27:19.937568 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d" (UID: "7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:27:19 crc kubenswrapper[4907]: I0313 15:27:19.937639 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d-kube-api-access-vsj5v" (OuterVolumeSpecName: "kube-api-access-vsj5v") pod "7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d" (UID: "7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d"). InnerVolumeSpecName "kube-api-access-vsj5v". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:27:19 crc kubenswrapper[4907]: I0313 15:27:19.942435 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-9187efe6-871d-4bbe-9376-bff5d8a151f2" (OuterVolumeSpecName: "persistence") pod "7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d" (UID: "7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d"). InnerVolumeSpecName "pvc-9187efe6-871d-4bbe-9376-bff5d8a151f2". PluginName "kubernetes.io/csi", VolumeGidValue "" Mar 13 15:27:19 crc kubenswrapper[4907]: I0313 15:27:19.957708 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d-server-conf" (OuterVolumeSpecName: "server-conf") pod "7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d" (UID: "7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.014692 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d" (UID: "7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.032393 4907 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.032432 4907 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.032464 4907 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-9187efe6-871d-4bbe-9376-bff5d8a151f2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-9187efe6-871d-4bbe-9376-bff5d8a151f2\") on node \"crc\" " Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.032476 4907 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.032486 4907 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d-plugins-conf\") on node \"crc\" DevicePath \"\"" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.032495 4907 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d-server-conf\") on node \"crc\" DevicePath \"\"" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.032503 4907 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d-pod-info\") on node \"crc\" DevicePath \"\"" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.032515 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vsj5v\" (UniqueName: \"kubernetes.io/projected/7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d-kube-api-access-vsj5v\") on node \"crc\" DevicePath \"\"" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.032523 4907 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.047122 4907 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.047261 4907 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-9187efe6-871d-4bbe-9376-bff5d8a151f2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-9187efe6-871d-4bbe-9376-bff5d8a151f2") on node "crc" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.133327 4907 reconciler_common.go:293] "Volume detached for volume \"pvc-9187efe6-871d-4bbe-9376-bff5d8a151f2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-9187efe6-871d-4bbe-9376-bff5d8a151f2\") on node \"crc\" DevicePath \"\"" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.503061 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-684c864bc9-p4j95" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.574296 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7c95686bd5-8kb7m"] Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.575745 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7c95686bd5-8kb7m" podUID="d1b7f262-ac9a-4a7e-ae94-6ce13d4105a2" containerName="dnsmasq-dns" containerID="cri-o://7dc8560cb6dc85650a7b469d68aaea22e66e3e209dca62e699bcb1b0807b2fea" gracePeriod=10 Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.621299 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.621320 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.621326 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d","Type":"ContainerDied","Data":"7fa7486d25e0b1f49302f99040c36181735134e64166f989efe3f3ae26956d78"} Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.621415 4907 scope.go:117] "RemoveContainer" containerID="c2763a983f63aa1fe0a38c4e94c194af6a54549cd715d4e926bfe5396d362f36" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.672376 4907 scope.go:117] "RemoveContainer" containerID="bf96e8e63801badfacccb754ff56342cdc5ba8417a4f20f022b8a5f76d1c9272" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.682599 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.717770 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.725349 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.730063 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.734623 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 13 15:27:20 crc kubenswrapper[4907]: E0313 15:27:20.734907 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd" containerName="rabbitmq" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.734923 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd" containerName="rabbitmq" Mar 13 15:27:20 crc kubenswrapper[4907]: E0313 15:27:20.734936 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd" containerName="setup-container" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.734942 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd" containerName="setup-container" Mar 13 15:27:20 crc kubenswrapper[4907]: E0313 15:27:20.734961 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d" containerName="rabbitmq" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.734969 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d" containerName="rabbitmq" Mar 13 15:27:20 crc kubenswrapper[4907]: E0313 15:27:20.734978 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d" containerName="setup-container" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.734984 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d" containerName="setup-container" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.735102 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd" containerName="rabbitmq" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.735116 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d" containerName="rabbitmq" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.741071 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.741176 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.744867 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.745067 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.745204 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.745325 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-grjtp" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.752591 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.759084 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.765681 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.771230 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.771384 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.771490 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-vlpfs" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.784160 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.785172 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.792699 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.843621 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/fc486d5d-ec94-461c-acf2-dcca08b2a1b6-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"fc486d5d-ec94-461c-acf2-dcca08b2a1b6\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.843683 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/fc486d5d-ec94-461c-acf2-dcca08b2a1b6-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"fc486d5d-ec94-461c-acf2-dcca08b2a1b6\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.843701 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/fc486d5d-ec94-461c-acf2-dcca08b2a1b6-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"fc486d5d-ec94-461c-acf2-dcca08b2a1b6\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.843720 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/fc486d5d-ec94-461c-acf2-dcca08b2a1b6-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"fc486d5d-ec94-461c-acf2-dcca08b2a1b6\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.843738 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/fc486d5d-ec94-461c-acf2-dcca08b2a1b6-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"fc486d5d-ec94-461c-acf2-dcca08b2a1b6\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.843762 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/fc486d5d-ec94-461c-acf2-dcca08b2a1b6-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"fc486d5d-ec94-461c-acf2-dcca08b2a1b6\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.843821 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-7c39547e-1168-404d-865b-13cfd1526f6f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7c39547e-1168-404d-865b-13cfd1526f6f\") pod \"rabbitmq-cell1-server-0\" (UID: \"fc486d5d-ec94-461c-acf2-dcca08b2a1b6\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.843840 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p486x\" (UniqueName: \"kubernetes.io/projected/fc486d5d-ec94-461c-acf2-dcca08b2a1b6-kube-api-access-p486x\") pod \"rabbitmq-cell1-server-0\" (UID: \"fc486d5d-ec94-461c-acf2-dcca08b2a1b6\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.843858 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/fc486d5d-ec94-461c-acf2-dcca08b2a1b6-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"fc486d5d-ec94-461c-acf2-dcca08b2a1b6\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.945568 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/fc486d5d-ec94-461c-acf2-dcca08b2a1b6-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"fc486d5d-ec94-461c-acf2-dcca08b2a1b6\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.945645 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/2ae0c4fc-0b3a-47a3-a6db-3e0ae3d64dad-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"2ae0c4fc-0b3a-47a3-a6db-3e0ae3d64dad\") " pod="openstack/rabbitmq-server-0" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.945673 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/2ae0c4fc-0b3a-47a3-a6db-3e0ae3d64dad-server-conf\") pod \"rabbitmq-server-0\" (UID: \"2ae0c4fc-0b3a-47a3-a6db-3e0ae3d64dad\") " pod="openstack/rabbitmq-server-0" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.945699 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/2ae0c4fc-0b3a-47a3-a6db-3e0ae3d64dad-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"2ae0c4fc-0b3a-47a3-a6db-3e0ae3d64dad\") " pod="openstack/rabbitmq-server-0" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.945833 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mzkdl\" (UniqueName: \"kubernetes.io/projected/2ae0c4fc-0b3a-47a3-a6db-3e0ae3d64dad-kube-api-access-mzkdl\") pod \"rabbitmq-server-0\" (UID: \"2ae0c4fc-0b3a-47a3-a6db-3e0ae3d64dad\") " pod="openstack/rabbitmq-server-0" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.945898 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/2ae0c4fc-0b3a-47a3-a6db-3e0ae3d64dad-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"2ae0c4fc-0b3a-47a3-a6db-3e0ae3d64dad\") " pod="openstack/rabbitmq-server-0" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.945934 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-7c39547e-1168-404d-865b-13cfd1526f6f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7c39547e-1168-404d-865b-13cfd1526f6f\") pod \"rabbitmq-cell1-server-0\" (UID: \"fc486d5d-ec94-461c-acf2-dcca08b2a1b6\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.945958 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p486x\" (UniqueName: \"kubernetes.io/projected/fc486d5d-ec94-461c-acf2-dcca08b2a1b6-kube-api-access-p486x\") pod \"rabbitmq-cell1-server-0\" (UID: \"fc486d5d-ec94-461c-acf2-dcca08b2a1b6\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.945988 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/2ae0c4fc-0b3a-47a3-a6db-3e0ae3d64dad-pod-info\") pod \"rabbitmq-server-0\" (UID: \"2ae0c4fc-0b3a-47a3-a6db-3e0ae3d64dad\") " pod="openstack/rabbitmq-server-0" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.946008 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/fc486d5d-ec94-461c-acf2-dcca08b2a1b6-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"fc486d5d-ec94-461c-acf2-dcca08b2a1b6\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.946072 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/fc486d5d-ec94-461c-acf2-dcca08b2a1b6-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"fc486d5d-ec94-461c-acf2-dcca08b2a1b6\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.946130 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/fc486d5d-ec94-461c-acf2-dcca08b2a1b6-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"fc486d5d-ec94-461c-acf2-dcca08b2a1b6\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.946155 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/fc486d5d-ec94-461c-acf2-dcca08b2a1b6-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"fc486d5d-ec94-461c-acf2-dcca08b2a1b6\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.946181 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/fc486d5d-ec94-461c-acf2-dcca08b2a1b6-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"fc486d5d-ec94-461c-acf2-dcca08b2a1b6\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.946201 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/2ae0c4fc-0b3a-47a3-a6db-3e0ae3d64dad-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"2ae0c4fc-0b3a-47a3-a6db-3e0ae3d64dad\") " pod="openstack/rabbitmq-server-0" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.946236 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/fc486d5d-ec94-461c-acf2-dcca08b2a1b6-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"fc486d5d-ec94-461c-acf2-dcca08b2a1b6\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.946281 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/2ae0c4fc-0b3a-47a3-a6db-3e0ae3d64dad-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"2ae0c4fc-0b3a-47a3-a6db-3e0ae3d64dad\") " pod="openstack/rabbitmq-server-0" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.946301 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-9187efe6-871d-4bbe-9376-bff5d8a151f2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-9187efe6-871d-4bbe-9376-bff5d8a151f2\") pod \"rabbitmq-server-0\" (UID: \"2ae0c4fc-0b3a-47a3-a6db-3e0ae3d64dad\") " pod="openstack/rabbitmq-server-0" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.948268 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/fc486d5d-ec94-461c-acf2-dcca08b2a1b6-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"fc486d5d-ec94-461c-acf2-dcca08b2a1b6\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.948717 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/fc486d5d-ec94-461c-acf2-dcca08b2a1b6-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"fc486d5d-ec94-461c-acf2-dcca08b2a1b6\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.948942 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/fc486d5d-ec94-461c-acf2-dcca08b2a1b6-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"fc486d5d-ec94-461c-acf2-dcca08b2a1b6\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.949036 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/fc486d5d-ec94-461c-acf2-dcca08b2a1b6-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"fc486d5d-ec94-461c-acf2-dcca08b2a1b6\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.952337 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/fc486d5d-ec94-461c-acf2-dcca08b2a1b6-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"fc486d5d-ec94-461c-acf2-dcca08b2a1b6\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.954649 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/fc486d5d-ec94-461c-acf2-dcca08b2a1b6-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"fc486d5d-ec94-461c-acf2-dcca08b2a1b6\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.955743 4907 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.955784 4907 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-7c39547e-1168-404d-865b-13cfd1526f6f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7c39547e-1168-404d-865b-13cfd1526f6f\") pod \"rabbitmq-cell1-server-0\" (UID: \"fc486d5d-ec94-461c-acf2-dcca08b2a1b6\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/d84a704eae1ca42fe329ab2f9d170d2a410be00de88c5e0a709d7b62965c26cb/globalmount\"" pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.956661 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/fc486d5d-ec94-461c-acf2-dcca08b2a1b6-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"fc486d5d-ec94-461c-acf2-dcca08b2a1b6\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.965619 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p486x\" (UniqueName: \"kubernetes.io/projected/fc486d5d-ec94-461c-acf2-dcca08b2a1b6-kube-api-access-p486x\") pod \"rabbitmq-cell1-server-0\" (UID: \"fc486d5d-ec94-461c-acf2-dcca08b2a1b6\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.998306 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-7c39547e-1168-404d-865b-13cfd1526f6f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7c39547e-1168-404d-865b-13cfd1526f6f\") pod \"rabbitmq-cell1-server-0\" (UID: \"fc486d5d-ec94-461c-acf2-dcca08b2a1b6\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:27:21 crc kubenswrapper[4907]: I0313 15:27:21.026235 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c95686bd5-8kb7m" Mar 13 15:27:21 crc kubenswrapper[4907]: I0313 15:27:21.048008 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/2ae0c4fc-0b3a-47a3-a6db-3e0ae3d64dad-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"2ae0c4fc-0b3a-47a3-a6db-3e0ae3d64dad\") " pod="openstack/rabbitmq-server-0" Mar 13 15:27:21 crc kubenswrapper[4907]: I0313 15:27:21.048373 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/2ae0c4fc-0b3a-47a3-a6db-3e0ae3d64dad-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"2ae0c4fc-0b3a-47a3-a6db-3e0ae3d64dad\") " pod="openstack/rabbitmq-server-0" Mar 13 15:27:21 crc kubenswrapper[4907]: I0313 15:27:21.048398 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-9187efe6-871d-4bbe-9376-bff5d8a151f2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-9187efe6-871d-4bbe-9376-bff5d8a151f2\") pod \"rabbitmq-server-0\" (UID: \"2ae0c4fc-0b3a-47a3-a6db-3e0ae3d64dad\") " pod="openstack/rabbitmq-server-0" Mar 13 15:27:21 crc kubenswrapper[4907]: I0313 15:27:21.048438 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/2ae0c4fc-0b3a-47a3-a6db-3e0ae3d64dad-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"2ae0c4fc-0b3a-47a3-a6db-3e0ae3d64dad\") " pod="openstack/rabbitmq-server-0" Mar 13 15:27:21 crc kubenswrapper[4907]: I0313 15:27:21.048461 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/2ae0c4fc-0b3a-47a3-a6db-3e0ae3d64dad-server-conf\") pod \"rabbitmq-server-0\" (UID: \"2ae0c4fc-0b3a-47a3-a6db-3e0ae3d64dad\") " pod="openstack/rabbitmq-server-0" Mar 13 15:27:21 crc kubenswrapper[4907]: I0313 15:27:21.048487 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/2ae0c4fc-0b3a-47a3-a6db-3e0ae3d64dad-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"2ae0c4fc-0b3a-47a3-a6db-3e0ae3d64dad\") " pod="openstack/rabbitmq-server-0" Mar 13 15:27:21 crc kubenswrapper[4907]: I0313 15:27:21.048509 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mzkdl\" (UniqueName: \"kubernetes.io/projected/2ae0c4fc-0b3a-47a3-a6db-3e0ae3d64dad-kube-api-access-mzkdl\") pod \"rabbitmq-server-0\" (UID: \"2ae0c4fc-0b3a-47a3-a6db-3e0ae3d64dad\") " pod="openstack/rabbitmq-server-0" Mar 13 15:27:21 crc kubenswrapper[4907]: I0313 15:27:21.048527 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/2ae0c4fc-0b3a-47a3-a6db-3e0ae3d64dad-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"2ae0c4fc-0b3a-47a3-a6db-3e0ae3d64dad\") " pod="openstack/rabbitmq-server-0" Mar 13 15:27:21 crc kubenswrapper[4907]: I0313 15:27:21.048546 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/2ae0c4fc-0b3a-47a3-a6db-3e0ae3d64dad-pod-info\") pod \"rabbitmq-server-0\" (UID: \"2ae0c4fc-0b3a-47a3-a6db-3e0ae3d64dad\") " pod="openstack/rabbitmq-server-0" Mar 13 15:27:21 crc kubenswrapper[4907]: I0313 15:27:21.048995 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/2ae0c4fc-0b3a-47a3-a6db-3e0ae3d64dad-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"2ae0c4fc-0b3a-47a3-a6db-3e0ae3d64dad\") " pod="openstack/rabbitmq-server-0" Mar 13 15:27:21 crc kubenswrapper[4907]: I0313 15:27:21.049712 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/2ae0c4fc-0b3a-47a3-a6db-3e0ae3d64dad-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"2ae0c4fc-0b3a-47a3-a6db-3e0ae3d64dad\") " pod="openstack/rabbitmq-server-0" Mar 13 15:27:21 crc kubenswrapper[4907]: I0313 15:27:21.050182 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/2ae0c4fc-0b3a-47a3-a6db-3e0ae3d64dad-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"2ae0c4fc-0b3a-47a3-a6db-3e0ae3d64dad\") " pod="openstack/rabbitmq-server-0" Mar 13 15:27:21 crc kubenswrapper[4907]: I0313 15:27:21.051248 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/2ae0c4fc-0b3a-47a3-a6db-3e0ae3d64dad-server-conf\") pod \"rabbitmq-server-0\" (UID: \"2ae0c4fc-0b3a-47a3-a6db-3e0ae3d64dad\") " pod="openstack/rabbitmq-server-0" Mar 13 15:27:21 crc kubenswrapper[4907]: I0313 15:27:21.054099 4907 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 13 15:27:21 crc kubenswrapper[4907]: I0313 15:27:21.054139 4907 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-9187efe6-871d-4bbe-9376-bff5d8a151f2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-9187efe6-871d-4bbe-9376-bff5d8a151f2\") pod \"rabbitmq-server-0\" (UID: \"2ae0c4fc-0b3a-47a3-a6db-3e0ae3d64dad\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/c926015d5702d92351724596ba892321797e624e5f7a01b00404b7d1c8a17788/globalmount\"" pod="openstack/rabbitmq-server-0" Mar 13 15:27:21 crc kubenswrapper[4907]: I0313 15:27:21.056852 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/2ae0c4fc-0b3a-47a3-a6db-3e0ae3d64dad-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"2ae0c4fc-0b3a-47a3-a6db-3e0ae3d64dad\") " pod="openstack/rabbitmq-server-0" Mar 13 15:27:21 crc kubenswrapper[4907]: I0313 15:27:21.057748 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/2ae0c4fc-0b3a-47a3-a6db-3e0ae3d64dad-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"2ae0c4fc-0b3a-47a3-a6db-3e0ae3d64dad\") " pod="openstack/rabbitmq-server-0" Mar 13 15:27:21 crc kubenswrapper[4907]: I0313 15:27:21.064732 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/2ae0c4fc-0b3a-47a3-a6db-3e0ae3d64dad-pod-info\") pod \"rabbitmq-server-0\" (UID: \"2ae0c4fc-0b3a-47a3-a6db-3e0ae3d64dad\") " pod="openstack/rabbitmq-server-0" Mar 13 15:27:21 crc kubenswrapper[4907]: I0313 15:27:21.068211 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mzkdl\" (UniqueName: \"kubernetes.io/projected/2ae0c4fc-0b3a-47a3-a6db-3e0ae3d64dad-kube-api-access-mzkdl\") pod \"rabbitmq-server-0\" (UID: \"2ae0c4fc-0b3a-47a3-a6db-3e0ae3d64dad\") " pod="openstack/rabbitmq-server-0" Mar 13 15:27:21 crc kubenswrapper[4907]: I0313 15:27:21.075415 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:27:21 crc kubenswrapper[4907]: I0313 15:27:21.089316 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-9187efe6-871d-4bbe-9376-bff5d8a151f2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-9187efe6-871d-4bbe-9376-bff5d8a151f2\") pod \"rabbitmq-server-0\" (UID: \"2ae0c4fc-0b3a-47a3-a6db-3e0ae3d64dad\") " pod="openstack/rabbitmq-server-0" Mar 13 15:27:21 crc kubenswrapper[4907]: I0313 15:27:21.099456 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Mar 13 15:27:21 crc kubenswrapper[4907]: I0313 15:27:21.152482 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r9x6s\" (UniqueName: \"kubernetes.io/projected/d1b7f262-ac9a-4a7e-ae94-6ce13d4105a2-kube-api-access-r9x6s\") pod \"d1b7f262-ac9a-4a7e-ae94-6ce13d4105a2\" (UID: \"d1b7f262-ac9a-4a7e-ae94-6ce13d4105a2\") " Mar 13 15:27:21 crc kubenswrapper[4907]: I0313 15:27:21.152531 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d1b7f262-ac9a-4a7e-ae94-6ce13d4105a2-config\") pod \"d1b7f262-ac9a-4a7e-ae94-6ce13d4105a2\" (UID: \"d1b7f262-ac9a-4a7e-ae94-6ce13d4105a2\") " Mar 13 15:27:21 crc kubenswrapper[4907]: I0313 15:27:21.152628 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d1b7f262-ac9a-4a7e-ae94-6ce13d4105a2-dns-svc\") pod \"d1b7f262-ac9a-4a7e-ae94-6ce13d4105a2\" (UID: \"d1b7f262-ac9a-4a7e-ae94-6ce13d4105a2\") " Mar 13 15:27:21 crc kubenswrapper[4907]: I0313 15:27:21.158809 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d1b7f262-ac9a-4a7e-ae94-6ce13d4105a2-kube-api-access-r9x6s" (OuterVolumeSpecName: "kube-api-access-r9x6s") pod "d1b7f262-ac9a-4a7e-ae94-6ce13d4105a2" (UID: "d1b7f262-ac9a-4a7e-ae94-6ce13d4105a2"). InnerVolumeSpecName "kube-api-access-r9x6s". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:27:21 crc kubenswrapper[4907]: I0313 15:27:21.187622 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d1b7f262-ac9a-4a7e-ae94-6ce13d4105a2-config" (OuterVolumeSpecName: "config") pod "d1b7f262-ac9a-4a7e-ae94-6ce13d4105a2" (UID: "d1b7f262-ac9a-4a7e-ae94-6ce13d4105a2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:27:21 crc kubenswrapper[4907]: I0313 15:27:21.194957 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d1b7f262-ac9a-4a7e-ae94-6ce13d4105a2-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d1b7f262-ac9a-4a7e-ae94-6ce13d4105a2" (UID: "d1b7f262-ac9a-4a7e-ae94-6ce13d4105a2"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:27:21 crc kubenswrapper[4907]: I0313 15:27:21.254124 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r9x6s\" (UniqueName: \"kubernetes.io/projected/d1b7f262-ac9a-4a7e-ae94-6ce13d4105a2-kube-api-access-r9x6s\") on node \"crc\" DevicePath \"\"" Mar 13 15:27:21 crc kubenswrapper[4907]: I0313 15:27:21.254165 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d1b7f262-ac9a-4a7e-ae94-6ce13d4105a2-config\") on node \"crc\" DevicePath \"\"" Mar 13 15:27:21 crc kubenswrapper[4907]: I0313 15:27:21.254189 4907 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d1b7f262-ac9a-4a7e-ae94-6ce13d4105a2-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 13 15:27:21 crc kubenswrapper[4907]: I0313 15:27:21.581070 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 13 15:27:21 crc kubenswrapper[4907]: I0313 15:27:21.587863 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Mar 13 15:27:21 crc kubenswrapper[4907]: W0313 15:27:21.595853 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2ae0c4fc_0b3a_47a3_a6db_3e0ae3d64dad.slice/crio-b99ed7eeb540a88d6a8940684577ebb52b99ec7a8bb229f5d1502b3a7666409c WatchSource:0}: Error finding container b99ed7eeb540a88d6a8940684577ebb52b99ec7a8bb229f5d1502b3a7666409c: Status 404 returned error can't find the container with id b99ed7eeb540a88d6a8940684577ebb52b99ec7a8bb229f5d1502b3a7666409c Mar 13 15:27:21 crc kubenswrapper[4907]: I0313 15:27:21.635341 4907 generic.go:334] "Generic (PLEG): container finished" podID="d1b7f262-ac9a-4a7e-ae94-6ce13d4105a2" containerID="7dc8560cb6dc85650a7b469d68aaea22e66e3e209dca62e699bcb1b0807b2fea" exitCode=0 Mar 13 15:27:21 crc kubenswrapper[4907]: I0313 15:27:21.636140 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c95686bd5-8kb7m" Mar 13 15:27:21 crc kubenswrapper[4907]: I0313 15:27:21.636162 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c95686bd5-8kb7m" event={"ID":"d1b7f262-ac9a-4a7e-ae94-6ce13d4105a2","Type":"ContainerDied","Data":"7dc8560cb6dc85650a7b469d68aaea22e66e3e209dca62e699bcb1b0807b2fea"} Mar 13 15:27:21 crc kubenswrapper[4907]: I0313 15:27:21.636735 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c95686bd5-8kb7m" event={"ID":"d1b7f262-ac9a-4a7e-ae94-6ce13d4105a2","Type":"ContainerDied","Data":"18cfff04ebd6c2ab270b3fa1a352afaef9dc310f2e13e5df39289e1f0f968058"} Mar 13 15:27:21 crc kubenswrapper[4907]: I0313 15:27:21.636761 4907 scope.go:117] "RemoveContainer" containerID="7dc8560cb6dc85650a7b469d68aaea22e66e3e209dca62e699bcb1b0807b2fea" Mar 13 15:27:21 crc kubenswrapper[4907]: I0313 15:27:21.639444 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"fc486d5d-ec94-461c-acf2-dcca08b2a1b6","Type":"ContainerStarted","Data":"694eb1066ead945b234eab0fd1a5c138170fd26599f8651b6d0198397d809e5a"} Mar 13 15:27:21 crc kubenswrapper[4907]: I0313 15:27:21.641246 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"2ae0c4fc-0b3a-47a3-a6db-3e0ae3d64dad","Type":"ContainerStarted","Data":"b99ed7eeb540a88d6a8940684577ebb52b99ec7a8bb229f5d1502b3a7666409c"} Mar 13 15:27:21 crc kubenswrapper[4907]: I0313 15:27:21.673744 4907 scope.go:117] "RemoveContainer" containerID="8f7fad973a21b1188eda2310b2c3e66fb866528b33ce2f78d5f9e1986ac222d3" Mar 13 15:27:21 crc kubenswrapper[4907]: I0313 15:27:21.722398 4907 scope.go:117] "RemoveContainer" containerID="7dc8560cb6dc85650a7b469d68aaea22e66e3e209dca62e699bcb1b0807b2fea" Mar 13 15:27:21 crc kubenswrapper[4907]: E0313 15:27:21.723079 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7dc8560cb6dc85650a7b469d68aaea22e66e3e209dca62e699bcb1b0807b2fea\": container with ID starting with 7dc8560cb6dc85650a7b469d68aaea22e66e3e209dca62e699bcb1b0807b2fea not found: ID does not exist" containerID="7dc8560cb6dc85650a7b469d68aaea22e66e3e209dca62e699bcb1b0807b2fea" Mar 13 15:27:21 crc kubenswrapper[4907]: I0313 15:27:21.723219 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7dc8560cb6dc85650a7b469d68aaea22e66e3e209dca62e699bcb1b0807b2fea"} err="failed to get container status \"7dc8560cb6dc85650a7b469d68aaea22e66e3e209dca62e699bcb1b0807b2fea\": rpc error: code = NotFound desc = could not find container \"7dc8560cb6dc85650a7b469d68aaea22e66e3e209dca62e699bcb1b0807b2fea\": container with ID starting with 7dc8560cb6dc85650a7b469d68aaea22e66e3e209dca62e699bcb1b0807b2fea not found: ID does not exist" Mar 13 15:27:21 crc kubenswrapper[4907]: I0313 15:27:21.723351 4907 scope.go:117] "RemoveContainer" containerID="8f7fad973a21b1188eda2310b2c3e66fb866528b33ce2f78d5f9e1986ac222d3" Mar 13 15:27:21 crc kubenswrapper[4907]: E0313 15:27:21.727374 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8f7fad973a21b1188eda2310b2c3e66fb866528b33ce2f78d5f9e1986ac222d3\": container with ID starting with 8f7fad973a21b1188eda2310b2c3e66fb866528b33ce2f78d5f9e1986ac222d3 not found: ID does not exist" containerID="8f7fad973a21b1188eda2310b2c3e66fb866528b33ce2f78d5f9e1986ac222d3" Mar 13 15:27:21 crc kubenswrapper[4907]: I0313 15:27:21.727450 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f7fad973a21b1188eda2310b2c3e66fb866528b33ce2f78d5f9e1986ac222d3"} err="failed to get container status \"8f7fad973a21b1188eda2310b2c3e66fb866528b33ce2f78d5f9e1986ac222d3\": rpc error: code = NotFound desc = could not find container \"8f7fad973a21b1188eda2310b2c3e66fb866528b33ce2f78d5f9e1986ac222d3\": container with ID starting with 8f7fad973a21b1188eda2310b2c3e66fb866528b33ce2f78d5f9e1986ac222d3 not found: ID does not exist" Mar 13 15:27:21 crc kubenswrapper[4907]: I0313 15:27:21.756824 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7c95686bd5-8kb7m"] Mar 13 15:27:21 crc kubenswrapper[4907]: I0313 15:27:21.762869 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7c95686bd5-8kb7m"] Mar 13 15:27:21 crc kubenswrapper[4907]: I0313 15:27:21.793064 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd" path="/var/lib/kubelet/pods/75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd/volumes" Mar 13 15:27:21 crc kubenswrapper[4907]: I0313 15:27:21.794811 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d" path="/var/lib/kubelet/pods/7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d/volumes" Mar 13 15:27:21 crc kubenswrapper[4907]: I0313 15:27:21.796267 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d1b7f262-ac9a-4a7e-ae94-6ce13d4105a2" path="/var/lib/kubelet/pods/d1b7f262-ac9a-4a7e-ae94-6ce13d4105a2/volumes" Mar 13 15:27:23 crc kubenswrapper[4907]: I0313 15:27:23.662612 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"2ae0c4fc-0b3a-47a3-a6db-3e0ae3d64dad","Type":"ContainerStarted","Data":"4e6bfc55b73ea14dd81f54b09e66186a0fe8a4af94eb0f240b0b9d243ebb9b45"} Mar 13 15:27:23 crc kubenswrapper[4907]: I0313 15:27:23.666052 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"fc486d5d-ec94-461c-acf2-dcca08b2a1b6","Type":"ContainerStarted","Data":"9dc9569aced9e6a9d87eacfd4113edb23371b96cee87c58cb825f8e9ca84fd40"} Mar 13 15:27:37 crc kubenswrapper[4907]: I0313 15:27:37.346155 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-st569"] Mar 13 15:27:37 crc kubenswrapper[4907]: E0313 15:27:37.346935 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1b7f262-ac9a-4a7e-ae94-6ce13d4105a2" containerName="init" Mar 13 15:27:37 crc kubenswrapper[4907]: I0313 15:27:37.346946 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1b7f262-ac9a-4a7e-ae94-6ce13d4105a2" containerName="init" Mar 13 15:27:37 crc kubenswrapper[4907]: E0313 15:27:37.346967 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1b7f262-ac9a-4a7e-ae94-6ce13d4105a2" containerName="dnsmasq-dns" Mar 13 15:27:37 crc kubenswrapper[4907]: I0313 15:27:37.346973 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1b7f262-ac9a-4a7e-ae94-6ce13d4105a2" containerName="dnsmasq-dns" Mar 13 15:27:37 crc kubenswrapper[4907]: I0313 15:27:37.347114 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1b7f262-ac9a-4a7e-ae94-6ce13d4105a2" containerName="dnsmasq-dns" Mar 13 15:27:37 crc kubenswrapper[4907]: I0313 15:27:37.348106 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-st569" Mar 13 15:27:37 crc kubenswrapper[4907]: I0313 15:27:37.361198 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-st569"] Mar 13 15:27:37 crc kubenswrapper[4907]: I0313 15:27:37.399200 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9pqvf\" (UniqueName: \"kubernetes.io/projected/340af619-4557-47cd-89b6-86237b062783-kube-api-access-9pqvf\") pod \"certified-operators-st569\" (UID: \"340af619-4557-47cd-89b6-86237b062783\") " pod="openshift-marketplace/certified-operators-st569" Mar 13 15:27:37 crc kubenswrapper[4907]: I0313 15:27:37.399347 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/340af619-4557-47cd-89b6-86237b062783-catalog-content\") pod \"certified-operators-st569\" (UID: \"340af619-4557-47cd-89b6-86237b062783\") " pod="openshift-marketplace/certified-operators-st569" Mar 13 15:27:37 crc kubenswrapper[4907]: I0313 15:27:37.399368 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/340af619-4557-47cd-89b6-86237b062783-utilities\") pod \"certified-operators-st569\" (UID: \"340af619-4557-47cd-89b6-86237b062783\") " pod="openshift-marketplace/certified-operators-st569" Mar 13 15:27:37 crc kubenswrapper[4907]: I0313 15:27:37.500391 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/340af619-4557-47cd-89b6-86237b062783-catalog-content\") pod \"certified-operators-st569\" (UID: \"340af619-4557-47cd-89b6-86237b062783\") " pod="openshift-marketplace/certified-operators-st569" Mar 13 15:27:37 crc kubenswrapper[4907]: I0313 15:27:37.500674 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/340af619-4557-47cd-89b6-86237b062783-utilities\") pod \"certified-operators-st569\" (UID: \"340af619-4557-47cd-89b6-86237b062783\") " pod="openshift-marketplace/certified-operators-st569" Mar 13 15:27:37 crc kubenswrapper[4907]: I0313 15:27:37.500807 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9pqvf\" (UniqueName: \"kubernetes.io/projected/340af619-4557-47cd-89b6-86237b062783-kube-api-access-9pqvf\") pod \"certified-operators-st569\" (UID: \"340af619-4557-47cd-89b6-86237b062783\") " pod="openshift-marketplace/certified-operators-st569" Mar 13 15:27:37 crc kubenswrapper[4907]: I0313 15:27:37.500895 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/340af619-4557-47cd-89b6-86237b062783-catalog-content\") pod \"certified-operators-st569\" (UID: \"340af619-4557-47cd-89b6-86237b062783\") " pod="openshift-marketplace/certified-operators-st569" Mar 13 15:27:37 crc kubenswrapper[4907]: I0313 15:27:37.501083 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/340af619-4557-47cd-89b6-86237b062783-utilities\") pod \"certified-operators-st569\" (UID: \"340af619-4557-47cd-89b6-86237b062783\") " pod="openshift-marketplace/certified-operators-st569" Mar 13 15:27:37 crc kubenswrapper[4907]: I0313 15:27:37.523733 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9pqvf\" (UniqueName: \"kubernetes.io/projected/340af619-4557-47cd-89b6-86237b062783-kube-api-access-9pqvf\") pod \"certified-operators-st569\" (UID: \"340af619-4557-47cd-89b6-86237b062783\") " pod="openshift-marketplace/certified-operators-st569" Mar 13 15:27:37 crc kubenswrapper[4907]: I0313 15:27:37.666101 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-st569" Mar 13 15:27:38 crc kubenswrapper[4907]: I0313 15:27:38.122834 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-st569"] Mar 13 15:27:38 crc kubenswrapper[4907]: I0313 15:27:38.823306 4907 generic.go:334] "Generic (PLEG): container finished" podID="340af619-4557-47cd-89b6-86237b062783" containerID="1e3d0562c691019dbf46762447ac1cd38a0e5e62d058f6ab0fe85ff6f764b47a" exitCode=0 Mar 13 15:27:38 crc kubenswrapper[4907]: I0313 15:27:38.823397 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-st569" event={"ID":"340af619-4557-47cd-89b6-86237b062783","Type":"ContainerDied","Data":"1e3d0562c691019dbf46762447ac1cd38a0e5e62d058f6ab0fe85ff6f764b47a"} Mar 13 15:27:38 crc kubenswrapper[4907]: I0313 15:27:38.823633 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-st569" event={"ID":"340af619-4557-47cd-89b6-86237b062783","Type":"ContainerStarted","Data":"3e29c70762aed7c8ac6ec015bcd46c38075af57c46837c2bb98102a5838f0966"} Mar 13 15:27:39 crc kubenswrapper[4907]: I0313 15:27:39.834964 4907 generic.go:334] "Generic (PLEG): container finished" podID="340af619-4557-47cd-89b6-86237b062783" containerID="e7e8d7b2fd73572e0983d90f865c19570898239292fc89f15261b7908adb8b0b" exitCode=0 Mar 13 15:27:39 crc kubenswrapper[4907]: I0313 15:27:39.835064 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-st569" event={"ID":"340af619-4557-47cd-89b6-86237b062783","Type":"ContainerDied","Data":"e7e8d7b2fd73572e0983d90f865c19570898239292fc89f15261b7908adb8b0b"} Mar 13 15:27:40 crc kubenswrapper[4907]: I0313 15:27:40.848219 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-st569" event={"ID":"340af619-4557-47cd-89b6-86237b062783","Type":"ContainerStarted","Data":"77a6cca1c6bd5dff35a26d10feca15c2f38dbcf318403240f5c7990157c4ab46"} Mar 13 15:27:40 crc kubenswrapper[4907]: I0313 15:27:40.874534 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-st569" podStartSLOduration=2.46863519 podStartE2EDuration="3.874505066s" podCreationTimestamp="2026-03-13 15:27:37 +0000 UTC" firstStartedPulling="2026-03-13 15:27:38.825861787 +0000 UTC m=+4957.725649476" lastFinishedPulling="2026-03-13 15:27:40.231731663 +0000 UTC m=+4959.131519352" observedRunningTime="2026-03-13 15:27:40.86731209 +0000 UTC m=+4959.767099819" watchObservedRunningTime="2026-03-13 15:27:40.874505066 +0000 UTC m=+4959.774292795" Mar 13 15:27:47 crc kubenswrapper[4907]: I0313 15:27:47.666436 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-st569" Mar 13 15:27:47 crc kubenswrapper[4907]: I0313 15:27:47.667282 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-st569" Mar 13 15:27:47 crc kubenswrapper[4907]: I0313 15:27:47.704465 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-st569" Mar 13 15:27:47 crc kubenswrapper[4907]: I0313 15:27:47.943410 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-st569" Mar 13 15:27:47 crc kubenswrapper[4907]: I0313 15:27:47.996603 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-st569"] Mar 13 15:27:49 crc kubenswrapper[4907]: I0313 15:27:49.918195 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-st569" podUID="340af619-4557-47cd-89b6-86237b062783" containerName="registry-server" containerID="cri-o://77a6cca1c6bd5dff35a26d10feca15c2f38dbcf318403240f5c7990157c4ab46" gracePeriod=2 Mar 13 15:27:50 crc kubenswrapper[4907]: I0313 15:27:50.280235 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-st569" Mar 13 15:27:50 crc kubenswrapper[4907]: I0313 15:27:50.398137 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/340af619-4557-47cd-89b6-86237b062783-utilities\") pod \"340af619-4557-47cd-89b6-86237b062783\" (UID: \"340af619-4557-47cd-89b6-86237b062783\") " Mar 13 15:27:50 crc kubenswrapper[4907]: I0313 15:27:50.398247 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9pqvf\" (UniqueName: \"kubernetes.io/projected/340af619-4557-47cd-89b6-86237b062783-kube-api-access-9pqvf\") pod \"340af619-4557-47cd-89b6-86237b062783\" (UID: \"340af619-4557-47cd-89b6-86237b062783\") " Mar 13 15:27:50 crc kubenswrapper[4907]: I0313 15:27:50.398281 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/340af619-4557-47cd-89b6-86237b062783-catalog-content\") pod \"340af619-4557-47cd-89b6-86237b062783\" (UID: \"340af619-4557-47cd-89b6-86237b062783\") " Mar 13 15:27:50 crc kubenswrapper[4907]: I0313 15:27:50.400199 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/340af619-4557-47cd-89b6-86237b062783-utilities" (OuterVolumeSpecName: "utilities") pod "340af619-4557-47cd-89b6-86237b062783" (UID: "340af619-4557-47cd-89b6-86237b062783"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:27:50 crc kubenswrapper[4907]: I0313 15:27:50.403332 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/340af619-4557-47cd-89b6-86237b062783-kube-api-access-9pqvf" (OuterVolumeSpecName: "kube-api-access-9pqvf") pod "340af619-4557-47cd-89b6-86237b062783" (UID: "340af619-4557-47cd-89b6-86237b062783"). InnerVolumeSpecName "kube-api-access-9pqvf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:27:50 crc kubenswrapper[4907]: I0313 15:27:50.500998 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/340af619-4557-47cd-89b6-86237b062783-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 15:27:50 crc kubenswrapper[4907]: I0313 15:27:50.501295 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9pqvf\" (UniqueName: \"kubernetes.io/projected/340af619-4557-47cd-89b6-86237b062783-kube-api-access-9pqvf\") on node \"crc\" DevicePath \"\"" Mar 13 15:27:50 crc kubenswrapper[4907]: I0313 15:27:50.932658 4907 generic.go:334] "Generic (PLEG): container finished" podID="340af619-4557-47cd-89b6-86237b062783" containerID="77a6cca1c6bd5dff35a26d10feca15c2f38dbcf318403240f5c7990157c4ab46" exitCode=0 Mar 13 15:27:50 crc kubenswrapper[4907]: I0313 15:27:50.932740 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-st569" event={"ID":"340af619-4557-47cd-89b6-86237b062783","Type":"ContainerDied","Data":"77a6cca1c6bd5dff35a26d10feca15c2f38dbcf318403240f5c7990157c4ab46"} Mar 13 15:27:50 crc kubenswrapper[4907]: I0313 15:27:50.932799 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-st569" Mar 13 15:27:50 crc kubenswrapper[4907]: I0313 15:27:50.932830 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-st569" event={"ID":"340af619-4557-47cd-89b6-86237b062783","Type":"ContainerDied","Data":"3e29c70762aed7c8ac6ec015bcd46c38075af57c46837c2bb98102a5838f0966"} Mar 13 15:27:50 crc kubenswrapper[4907]: I0313 15:27:50.932952 4907 scope.go:117] "RemoveContainer" containerID="77a6cca1c6bd5dff35a26d10feca15c2f38dbcf318403240f5c7990157c4ab46" Mar 13 15:27:50 crc kubenswrapper[4907]: I0313 15:27:50.958685 4907 scope.go:117] "RemoveContainer" containerID="e7e8d7b2fd73572e0983d90f865c19570898239292fc89f15261b7908adb8b0b" Mar 13 15:27:51 crc kubenswrapper[4907]: I0313 15:27:51.001416 4907 scope.go:117] "RemoveContainer" containerID="1e3d0562c691019dbf46762447ac1cd38a0e5e62d058f6ab0fe85ff6f764b47a" Mar 13 15:27:51 crc kubenswrapper[4907]: I0313 15:27:51.025770 4907 scope.go:117] "RemoveContainer" containerID="77a6cca1c6bd5dff35a26d10feca15c2f38dbcf318403240f5c7990157c4ab46" Mar 13 15:27:51 crc kubenswrapper[4907]: E0313 15:27:51.026277 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"77a6cca1c6bd5dff35a26d10feca15c2f38dbcf318403240f5c7990157c4ab46\": container with ID starting with 77a6cca1c6bd5dff35a26d10feca15c2f38dbcf318403240f5c7990157c4ab46 not found: ID does not exist" containerID="77a6cca1c6bd5dff35a26d10feca15c2f38dbcf318403240f5c7990157c4ab46" Mar 13 15:27:51 crc kubenswrapper[4907]: I0313 15:27:51.026323 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"77a6cca1c6bd5dff35a26d10feca15c2f38dbcf318403240f5c7990157c4ab46"} err="failed to get container status \"77a6cca1c6bd5dff35a26d10feca15c2f38dbcf318403240f5c7990157c4ab46\": rpc error: code = NotFound desc = could not find container \"77a6cca1c6bd5dff35a26d10feca15c2f38dbcf318403240f5c7990157c4ab46\": container with ID starting with 77a6cca1c6bd5dff35a26d10feca15c2f38dbcf318403240f5c7990157c4ab46 not found: ID does not exist" Mar 13 15:27:51 crc kubenswrapper[4907]: I0313 15:27:51.026357 4907 scope.go:117] "RemoveContainer" containerID="e7e8d7b2fd73572e0983d90f865c19570898239292fc89f15261b7908adb8b0b" Mar 13 15:27:51 crc kubenswrapper[4907]: E0313 15:27:51.027017 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e7e8d7b2fd73572e0983d90f865c19570898239292fc89f15261b7908adb8b0b\": container with ID starting with e7e8d7b2fd73572e0983d90f865c19570898239292fc89f15261b7908adb8b0b not found: ID does not exist" containerID="e7e8d7b2fd73572e0983d90f865c19570898239292fc89f15261b7908adb8b0b" Mar 13 15:27:51 crc kubenswrapper[4907]: I0313 15:27:51.027073 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e7e8d7b2fd73572e0983d90f865c19570898239292fc89f15261b7908adb8b0b"} err="failed to get container status \"e7e8d7b2fd73572e0983d90f865c19570898239292fc89f15261b7908adb8b0b\": rpc error: code = NotFound desc = could not find container \"e7e8d7b2fd73572e0983d90f865c19570898239292fc89f15261b7908adb8b0b\": container with ID starting with e7e8d7b2fd73572e0983d90f865c19570898239292fc89f15261b7908adb8b0b not found: ID does not exist" Mar 13 15:27:51 crc kubenswrapper[4907]: I0313 15:27:51.027096 4907 scope.go:117] "RemoveContainer" containerID="1e3d0562c691019dbf46762447ac1cd38a0e5e62d058f6ab0fe85ff6f764b47a" Mar 13 15:27:51 crc kubenswrapper[4907]: E0313 15:27:51.027418 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1e3d0562c691019dbf46762447ac1cd38a0e5e62d058f6ab0fe85ff6f764b47a\": container with ID starting with 1e3d0562c691019dbf46762447ac1cd38a0e5e62d058f6ab0fe85ff6f764b47a not found: ID does not exist" containerID="1e3d0562c691019dbf46762447ac1cd38a0e5e62d058f6ab0fe85ff6f764b47a" Mar 13 15:27:51 crc kubenswrapper[4907]: I0313 15:27:51.027454 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1e3d0562c691019dbf46762447ac1cd38a0e5e62d058f6ab0fe85ff6f764b47a"} err="failed to get container status \"1e3d0562c691019dbf46762447ac1cd38a0e5e62d058f6ab0fe85ff6f764b47a\": rpc error: code = NotFound desc = could not find container \"1e3d0562c691019dbf46762447ac1cd38a0e5e62d058f6ab0fe85ff6f764b47a\": container with ID starting with 1e3d0562c691019dbf46762447ac1cd38a0e5e62d058f6ab0fe85ff6f764b47a not found: ID does not exist" Mar 13 15:27:51 crc kubenswrapper[4907]: I0313 15:27:51.304544 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/340af619-4557-47cd-89b6-86237b062783-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "340af619-4557-47cd-89b6-86237b062783" (UID: "340af619-4557-47cd-89b6-86237b062783"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:27:51 crc kubenswrapper[4907]: I0313 15:27:51.313283 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/340af619-4557-47cd-89b6-86237b062783-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 15:27:51 crc kubenswrapper[4907]: I0313 15:27:51.566559 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-st569"] Mar 13 15:27:51 crc kubenswrapper[4907]: I0313 15:27:51.575756 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-st569"] Mar 13 15:27:51 crc kubenswrapper[4907]: I0313 15:27:51.795859 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="340af619-4557-47cd-89b6-86237b062783" path="/var/lib/kubelet/pods/340af619-4557-47cd-89b6-86237b062783/volumes" Mar 13 15:27:54 crc kubenswrapper[4907]: I0313 15:27:54.964574 4907 generic.go:334] "Generic (PLEG): container finished" podID="fc486d5d-ec94-461c-acf2-dcca08b2a1b6" containerID="9dc9569aced9e6a9d87eacfd4113edb23371b96cee87c58cb825f8e9ca84fd40" exitCode=0 Mar 13 15:27:54 crc kubenswrapper[4907]: I0313 15:27:54.964688 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"fc486d5d-ec94-461c-acf2-dcca08b2a1b6","Type":"ContainerDied","Data":"9dc9569aced9e6a9d87eacfd4113edb23371b96cee87c58cb825f8e9ca84fd40"} Mar 13 15:27:55 crc kubenswrapper[4907]: I0313 15:27:55.976213 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"fc486d5d-ec94-461c-acf2-dcca08b2a1b6","Type":"ContainerStarted","Data":"49d332f53efdfc23759dba52b6372c5d1993e483aca345df376f98c207b6db3b"} Mar 13 15:27:55 crc kubenswrapper[4907]: I0313 15:27:55.977043 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:27:55 crc kubenswrapper[4907]: I0313 15:27:55.980815 4907 generic.go:334] "Generic (PLEG): container finished" podID="2ae0c4fc-0b3a-47a3-a6db-3e0ae3d64dad" containerID="4e6bfc55b73ea14dd81f54b09e66186a0fe8a4af94eb0f240b0b9d243ebb9b45" exitCode=0 Mar 13 15:27:55 crc kubenswrapper[4907]: I0313 15:27:55.980867 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"2ae0c4fc-0b3a-47a3-a6db-3e0ae3d64dad","Type":"ContainerDied","Data":"4e6bfc55b73ea14dd81f54b09e66186a0fe8a4af94eb0f240b0b9d243ebb9b45"} Mar 13 15:27:56 crc kubenswrapper[4907]: I0313 15:27:56.007434 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=36.007405306 podStartE2EDuration="36.007405306s" podCreationTimestamp="2026-03-13 15:27:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:27:56.005860813 +0000 UTC m=+4974.905648502" watchObservedRunningTime="2026-03-13 15:27:56.007405306 +0000 UTC m=+4974.907193015" Mar 13 15:27:56 crc kubenswrapper[4907]: I0313 15:27:56.989622 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"2ae0c4fc-0b3a-47a3-a6db-3e0ae3d64dad","Type":"ContainerStarted","Data":"99e651808a9607da115a9154bac22f68298b64908ab45c28005107610dea2392"} Mar 13 15:27:56 crc kubenswrapper[4907]: I0313 15:27:56.990196 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Mar 13 15:27:57 crc kubenswrapper[4907]: I0313 15:27:57.017231 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=37.017208514 podStartE2EDuration="37.017208514s" podCreationTimestamp="2026-03-13 15:27:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:27:57.013375969 +0000 UTC m=+4975.913163698" watchObservedRunningTime="2026-03-13 15:27:57.017208514 +0000 UTC m=+4975.916996213" Mar 13 15:28:00 crc kubenswrapper[4907]: I0313 15:28:00.143154 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556928-nh6lw"] Mar 13 15:28:00 crc kubenswrapper[4907]: E0313 15:28:00.143793 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="340af619-4557-47cd-89b6-86237b062783" containerName="extract-content" Mar 13 15:28:00 crc kubenswrapper[4907]: I0313 15:28:00.143807 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="340af619-4557-47cd-89b6-86237b062783" containerName="extract-content" Mar 13 15:28:00 crc kubenswrapper[4907]: E0313 15:28:00.143816 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="340af619-4557-47cd-89b6-86237b062783" containerName="extract-utilities" Mar 13 15:28:00 crc kubenswrapper[4907]: I0313 15:28:00.143823 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="340af619-4557-47cd-89b6-86237b062783" containerName="extract-utilities" Mar 13 15:28:00 crc kubenswrapper[4907]: E0313 15:28:00.143837 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="340af619-4557-47cd-89b6-86237b062783" containerName="registry-server" Mar 13 15:28:00 crc kubenswrapper[4907]: I0313 15:28:00.143843 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="340af619-4557-47cd-89b6-86237b062783" containerName="registry-server" Mar 13 15:28:00 crc kubenswrapper[4907]: I0313 15:28:00.144007 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="340af619-4557-47cd-89b6-86237b062783" containerName="registry-server" Mar 13 15:28:00 crc kubenswrapper[4907]: I0313 15:28:00.144514 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556928-nh6lw" Mar 13 15:28:00 crc kubenswrapper[4907]: I0313 15:28:00.146841 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 15:28:00 crc kubenswrapper[4907]: I0313 15:28:00.146849 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 15:28:00 crc kubenswrapper[4907]: I0313 15:28:00.153117 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 15:28:00 crc kubenswrapper[4907]: I0313 15:28:00.154759 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556928-nh6lw"] Mar 13 15:28:00 crc kubenswrapper[4907]: I0313 15:28:00.240835 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5l9fr\" (UniqueName: \"kubernetes.io/projected/566ddaa7-ea07-421b-a181-0aa768454ba6-kube-api-access-5l9fr\") pod \"auto-csr-approver-29556928-nh6lw\" (UID: \"566ddaa7-ea07-421b-a181-0aa768454ba6\") " pod="openshift-infra/auto-csr-approver-29556928-nh6lw" Mar 13 15:28:00 crc kubenswrapper[4907]: I0313 15:28:00.342228 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5l9fr\" (UniqueName: \"kubernetes.io/projected/566ddaa7-ea07-421b-a181-0aa768454ba6-kube-api-access-5l9fr\") pod \"auto-csr-approver-29556928-nh6lw\" (UID: \"566ddaa7-ea07-421b-a181-0aa768454ba6\") " pod="openshift-infra/auto-csr-approver-29556928-nh6lw" Mar 13 15:28:00 crc kubenswrapper[4907]: I0313 15:28:00.361821 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5l9fr\" (UniqueName: \"kubernetes.io/projected/566ddaa7-ea07-421b-a181-0aa768454ba6-kube-api-access-5l9fr\") pod \"auto-csr-approver-29556928-nh6lw\" (UID: \"566ddaa7-ea07-421b-a181-0aa768454ba6\") " pod="openshift-infra/auto-csr-approver-29556928-nh6lw" Mar 13 15:28:00 crc kubenswrapper[4907]: I0313 15:28:00.461179 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556928-nh6lw" Mar 13 15:28:00 crc kubenswrapper[4907]: I0313 15:28:00.885099 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556928-nh6lw"] Mar 13 15:28:00 crc kubenswrapper[4907]: W0313 15:28:00.892441 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod566ddaa7_ea07_421b_a181_0aa768454ba6.slice/crio-e2be2e36196a0df0aa7aaeb1f881a8eebe3037ab79d154d7fe5a1e31ed4297b7 WatchSource:0}: Error finding container e2be2e36196a0df0aa7aaeb1f881a8eebe3037ab79d154d7fe5a1e31ed4297b7: Status 404 returned error can't find the container with id e2be2e36196a0df0aa7aaeb1f881a8eebe3037ab79d154d7fe5a1e31ed4297b7 Mar 13 15:28:01 crc kubenswrapper[4907]: I0313 15:28:01.014716 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556928-nh6lw" event={"ID":"566ddaa7-ea07-421b-a181-0aa768454ba6","Type":"ContainerStarted","Data":"e2be2e36196a0df0aa7aaeb1f881a8eebe3037ab79d154d7fe5a1e31ed4297b7"} Mar 13 15:28:03 crc kubenswrapper[4907]: I0313 15:28:03.028840 4907 generic.go:334] "Generic (PLEG): container finished" podID="566ddaa7-ea07-421b-a181-0aa768454ba6" containerID="206bf50c83686d14437a447b0be4276ceccf75f1f0402647af114c3e1ce67140" exitCode=0 Mar 13 15:28:03 crc kubenswrapper[4907]: I0313 15:28:03.029128 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556928-nh6lw" event={"ID":"566ddaa7-ea07-421b-a181-0aa768454ba6","Type":"ContainerDied","Data":"206bf50c83686d14437a447b0be4276ceccf75f1f0402647af114c3e1ce67140"} Mar 13 15:28:04 crc kubenswrapper[4907]: I0313 15:28:04.291257 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556928-nh6lw" Mar 13 15:28:04 crc kubenswrapper[4907]: I0313 15:28:04.407480 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5l9fr\" (UniqueName: \"kubernetes.io/projected/566ddaa7-ea07-421b-a181-0aa768454ba6-kube-api-access-5l9fr\") pod \"566ddaa7-ea07-421b-a181-0aa768454ba6\" (UID: \"566ddaa7-ea07-421b-a181-0aa768454ba6\") " Mar 13 15:28:04 crc kubenswrapper[4907]: I0313 15:28:04.412714 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/566ddaa7-ea07-421b-a181-0aa768454ba6-kube-api-access-5l9fr" (OuterVolumeSpecName: "kube-api-access-5l9fr") pod "566ddaa7-ea07-421b-a181-0aa768454ba6" (UID: "566ddaa7-ea07-421b-a181-0aa768454ba6"). InnerVolumeSpecName "kube-api-access-5l9fr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:28:04 crc kubenswrapper[4907]: I0313 15:28:04.510077 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5l9fr\" (UniqueName: \"kubernetes.io/projected/566ddaa7-ea07-421b-a181-0aa768454ba6-kube-api-access-5l9fr\") on node \"crc\" DevicePath \"\"" Mar 13 15:28:05 crc kubenswrapper[4907]: I0313 15:28:05.045938 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556928-nh6lw" event={"ID":"566ddaa7-ea07-421b-a181-0aa768454ba6","Type":"ContainerDied","Data":"e2be2e36196a0df0aa7aaeb1f881a8eebe3037ab79d154d7fe5a1e31ed4297b7"} Mar 13 15:28:05 crc kubenswrapper[4907]: I0313 15:28:05.045978 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e2be2e36196a0df0aa7aaeb1f881a8eebe3037ab79d154d7fe5a1e31ed4297b7" Mar 13 15:28:05 crc kubenswrapper[4907]: I0313 15:28:05.046007 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556928-nh6lw" Mar 13 15:28:05 crc kubenswrapper[4907]: I0313 15:28:05.362002 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556922-b7qtk"] Mar 13 15:28:05 crc kubenswrapper[4907]: I0313 15:28:05.366756 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556922-b7qtk"] Mar 13 15:28:05 crc kubenswrapper[4907]: I0313 15:28:05.791941 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31cb9010-2aab-475d-aab3-9aa35a538581" path="/var/lib/kubelet/pods/31cb9010-2aab-475d-aab3-9aa35a538581/volumes" Mar 13 15:28:11 crc kubenswrapper[4907]: I0313 15:28:11.078132 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:28:11 crc kubenswrapper[4907]: I0313 15:28:11.102155 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Mar 13 15:28:22 crc kubenswrapper[4907]: I0313 15:28:22.942598 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client"] Mar 13 15:28:22 crc kubenswrapper[4907]: E0313 15:28:22.943554 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="566ddaa7-ea07-421b-a181-0aa768454ba6" containerName="oc" Mar 13 15:28:22 crc kubenswrapper[4907]: I0313 15:28:22.943573 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="566ddaa7-ea07-421b-a181-0aa768454ba6" containerName="oc" Mar 13 15:28:22 crc kubenswrapper[4907]: I0313 15:28:22.943732 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="566ddaa7-ea07-421b-a181-0aa768454ba6" containerName="oc" Mar 13 15:28:22 crc kubenswrapper[4907]: I0313 15:28:22.944409 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Mar 13 15:28:22 crc kubenswrapper[4907]: I0313 15:28:22.946781 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-mqs4s" Mar 13 15:28:22 crc kubenswrapper[4907]: I0313 15:28:22.952315 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Mar 13 15:28:23 crc kubenswrapper[4907]: I0313 15:28:23.087819 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sc7c7\" (UniqueName: \"kubernetes.io/projected/c39974e7-120a-410b-b78f-d3588b2cc22d-kube-api-access-sc7c7\") pod \"mariadb-client\" (UID: \"c39974e7-120a-410b-b78f-d3588b2cc22d\") " pod="openstack/mariadb-client" Mar 13 15:28:23 crc kubenswrapper[4907]: I0313 15:28:23.189588 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sc7c7\" (UniqueName: \"kubernetes.io/projected/c39974e7-120a-410b-b78f-d3588b2cc22d-kube-api-access-sc7c7\") pod \"mariadb-client\" (UID: \"c39974e7-120a-410b-b78f-d3588b2cc22d\") " pod="openstack/mariadb-client" Mar 13 15:28:23 crc kubenswrapper[4907]: I0313 15:28:23.207396 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sc7c7\" (UniqueName: \"kubernetes.io/projected/c39974e7-120a-410b-b78f-d3588b2cc22d-kube-api-access-sc7c7\") pod \"mariadb-client\" (UID: \"c39974e7-120a-410b-b78f-d3588b2cc22d\") " pod="openstack/mariadb-client" Mar 13 15:28:23 crc kubenswrapper[4907]: I0313 15:28:23.269528 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Mar 13 15:28:23 crc kubenswrapper[4907]: I0313 15:28:23.766961 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Mar 13 15:28:24 crc kubenswrapper[4907]: I0313 15:28:24.195494 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"c39974e7-120a-410b-b78f-d3588b2cc22d","Type":"ContainerStarted","Data":"82d4fd4ae31f9ad44ead28b840a4b0ae479a0cec913d3bdb2dc24a7f47cb13e9"} Mar 13 15:28:25 crc kubenswrapper[4907]: I0313 15:28:25.207828 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"c39974e7-120a-410b-b78f-d3588b2cc22d","Type":"ContainerStarted","Data":"076daa9876f3ca96d1b1fb0a5e45b1cf807c92541a65b07004b19037cc101e37"} Mar 13 15:28:25 crc kubenswrapper[4907]: I0313 15:28:25.244349 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/mariadb-client" podStartSLOduration=2.708066529 podStartE2EDuration="3.244327757s" podCreationTimestamp="2026-03-13 15:28:22 +0000 UTC" firstStartedPulling="2026-03-13 15:28:23.771782792 +0000 UTC m=+5002.671570481" lastFinishedPulling="2026-03-13 15:28:24.30804402 +0000 UTC m=+5003.207831709" observedRunningTime="2026-03-13 15:28:25.236061001 +0000 UTC m=+5004.135848760" watchObservedRunningTime="2026-03-13 15:28:25.244327757 +0000 UTC m=+5004.144115456" Mar 13 15:28:27 crc kubenswrapper[4907]: I0313 15:28:27.418291 4907 scope.go:117] "RemoveContainer" containerID="2cd2416cb09c17cb001811639f27834ed34890771029fd10f4c0a96794fa7519" Mar 13 15:28:27 crc kubenswrapper[4907]: I0313 15:28:27.480644 4907 scope.go:117] "RemoveContainer" containerID="d45e59416766d20c4f95b995286178dd988b4b6257768b1fef8c82dcc281cd6c" Mar 13 15:28:38 crc kubenswrapper[4907]: I0313 15:28:38.507483 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client"] Mar 13 15:28:38 crc kubenswrapper[4907]: I0313 15:28:38.508436 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/mariadb-client" podUID="c39974e7-120a-410b-b78f-d3588b2cc22d" containerName="mariadb-client" containerID="cri-o://076daa9876f3ca96d1b1fb0a5e45b1cf807c92541a65b07004b19037cc101e37" gracePeriod=30 Mar 13 15:28:39 crc kubenswrapper[4907]: I0313 15:28:39.030527 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Mar 13 15:28:39 crc kubenswrapper[4907]: I0313 15:28:39.145935 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sc7c7\" (UniqueName: \"kubernetes.io/projected/c39974e7-120a-410b-b78f-d3588b2cc22d-kube-api-access-sc7c7\") pod \"c39974e7-120a-410b-b78f-d3588b2cc22d\" (UID: \"c39974e7-120a-410b-b78f-d3588b2cc22d\") " Mar 13 15:28:39 crc kubenswrapper[4907]: I0313 15:28:39.153152 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c39974e7-120a-410b-b78f-d3588b2cc22d-kube-api-access-sc7c7" (OuterVolumeSpecName: "kube-api-access-sc7c7") pod "c39974e7-120a-410b-b78f-d3588b2cc22d" (UID: "c39974e7-120a-410b-b78f-d3588b2cc22d"). InnerVolumeSpecName "kube-api-access-sc7c7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:28:39 crc kubenswrapper[4907]: I0313 15:28:39.247753 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sc7c7\" (UniqueName: \"kubernetes.io/projected/c39974e7-120a-410b-b78f-d3588b2cc22d-kube-api-access-sc7c7\") on node \"crc\" DevicePath \"\"" Mar 13 15:28:39 crc kubenswrapper[4907]: I0313 15:28:39.320913 4907 generic.go:334] "Generic (PLEG): container finished" podID="c39974e7-120a-410b-b78f-d3588b2cc22d" containerID="076daa9876f3ca96d1b1fb0a5e45b1cf807c92541a65b07004b19037cc101e37" exitCode=143 Mar 13 15:28:39 crc kubenswrapper[4907]: I0313 15:28:39.320953 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Mar 13 15:28:39 crc kubenswrapper[4907]: I0313 15:28:39.320965 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"c39974e7-120a-410b-b78f-d3588b2cc22d","Type":"ContainerDied","Data":"076daa9876f3ca96d1b1fb0a5e45b1cf807c92541a65b07004b19037cc101e37"} Mar 13 15:28:39 crc kubenswrapper[4907]: I0313 15:28:39.320988 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"c39974e7-120a-410b-b78f-d3588b2cc22d","Type":"ContainerDied","Data":"82d4fd4ae31f9ad44ead28b840a4b0ae479a0cec913d3bdb2dc24a7f47cb13e9"} Mar 13 15:28:39 crc kubenswrapper[4907]: I0313 15:28:39.321004 4907 scope.go:117] "RemoveContainer" containerID="076daa9876f3ca96d1b1fb0a5e45b1cf807c92541a65b07004b19037cc101e37" Mar 13 15:28:39 crc kubenswrapper[4907]: I0313 15:28:39.338840 4907 scope.go:117] "RemoveContainer" containerID="076daa9876f3ca96d1b1fb0a5e45b1cf807c92541a65b07004b19037cc101e37" Mar 13 15:28:39 crc kubenswrapper[4907]: E0313 15:28:39.339212 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"076daa9876f3ca96d1b1fb0a5e45b1cf807c92541a65b07004b19037cc101e37\": container with ID starting with 076daa9876f3ca96d1b1fb0a5e45b1cf807c92541a65b07004b19037cc101e37 not found: ID does not exist" containerID="076daa9876f3ca96d1b1fb0a5e45b1cf807c92541a65b07004b19037cc101e37" Mar 13 15:28:39 crc kubenswrapper[4907]: I0313 15:28:39.339243 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"076daa9876f3ca96d1b1fb0a5e45b1cf807c92541a65b07004b19037cc101e37"} err="failed to get container status \"076daa9876f3ca96d1b1fb0a5e45b1cf807c92541a65b07004b19037cc101e37\": rpc error: code = NotFound desc = could not find container \"076daa9876f3ca96d1b1fb0a5e45b1cf807c92541a65b07004b19037cc101e37\": container with ID starting with 076daa9876f3ca96d1b1fb0a5e45b1cf807c92541a65b07004b19037cc101e37 not found: ID does not exist" Mar 13 15:28:39 crc kubenswrapper[4907]: I0313 15:28:39.353282 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client"] Mar 13 15:28:39 crc kubenswrapper[4907]: I0313 15:28:39.356180 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client"] Mar 13 15:28:39 crc kubenswrapper[4907]: I0313 15:28:39.797426 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c39974e7-120a-410b-b78f-d3588b2cc22d" path="/var/lib/kubelet/pods/c39974e7-120a-410b-b78f-d3588b2cc22d/volumes" Mar 13 15:28:40 crc kubenswrapper[4907]: I0313 15:28:40.762440 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-jzxtj"] Mar 13 15:28:40 crc kubenswrapper[4907]: E0313 15:28:40.762815 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c39974e7-120a-410b-b78f-d3588b2cc22d" containerName="mariadb-client" Mar 13 15:28:40 crc kubenswrapper[4907]: I0313 15:28:40.762831 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="c39974e7-120a-410b-b78f-d3588b2cc22d" containerName="mariadb-client" Mar 13 15:28:40 crc kubenswrapper[4907]: I0313 15:28:40.763031 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="c39974e7-120a-410b-b78f-d3588b2cc22d" containerName="mariadb-client" Mar 13 15:28:40 crc kubenswrapper[4907]: I0313 15:28:40.764286 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jzxtj" Mar 13 15:28:40 crc kubenswrapper[4907]: I0313 15:28:40.775052 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jzxtj"] Mar 13 15:28:40 crc kubenswrapper[4907]: I0313 15:28:40.868935 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g76wj\" (UniqueName: \"kubernetes.io/projected/471a271b-2d62-4bdb-956d-91e268d2e4b5-kube-api-access-g76wj\") pod \"redhat-operators-jzxtj\" (UID: \"471a271b-2d62-4bdb-956d-91e268d2e4b5\") " pod="openshift-marketplace/redhat-operators-jzxtj" Mar 13 15:28:40 crc kubenswrapper[4907]: I0313 15:28:40.868993 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/471a271b-2d62-4bdb-956d-91e268d2e4b5-catalog-content\") pod \"redhat-operators-jzxtj\" (UID: \"471a271b-2d62-4bdb-956d-91e268d2e4b5\") " pod="openshift-marketplace/redhat-operators-jzxtj" Mar 13 15:28:40 crc kubenswrapper[4907]: I0313 15:28:40.869066 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/471a271b-2d62-4bdb-956d-91e268d2e4b5-utilities\") pod \"redhat-operators-jzxtj\" (UID: \"471a271b-2d62-4bdb-956d-91e268d2e4b5\") " pod="openshift-marketplace/redhat-operators-jzxtj" Mar 13 15:28:40 crc kubenswrapper[4907]: I0313 15:28:40.970431 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/471a271b-2d62-4bdb-956d-91e268d2e4b5-utilities\") pod \"redhat-operators-jzxtj\" (UID: \"471a271b-2d62-4bdb-956d-91e268d2e4b5\") " pod="openshift-marketplace/redhat-operators-jzxtj" Mar 13 15:28:40 crc kubenswrapper[4907]: I0313 15:28:40.970568 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g76wj\" (UniqueName: \"kubernetes.io/projected/471a271b-2d62-4bdb-956d-91e268d2e4b5-kube-api-access-g76wj\") pod \"redhat-operators-jzxtj\" (UID: \"471a271b-2d62-4bdb-956d-91e268d2e4b5\") " pod="openshift-marketplace/redhat-operators-jzxtj" Mar 13 15:28:40 crc kubenswrapper[4907]: I0313 15:28:40.970615 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/471a271b-2d62-4bdb-956d-91e268d2e4b5-catalog-content\") pod \"redhat-operators-jzxtj\" (UID: \"471a271b-2d62-4bdb-956d-91e268d2e4b5\") " pod="openshift-marketplace/redhat-operators-jzxtj" Mar 13 15:28:40 crc kubenswrapper[4907]: I0313 15:28:40.970979 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/471a271b-2d62-4bdb-956d-91e268d2e4b5-utilities\") pod \"redhat-operators-jzxtj\" (UID: \"471a271b-2d62-4bdb-956d-91e268d2e4b5\") " pod="openshift-marketplace/redhat-operators-jzxtj" Mar 13 15:28:40 crc kubenswrapper[4907]: I0313 15:28:40.971052 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/471a271b-2d62-4bdb-956d-91e268d2e4b5-catalog-content\") pod \"redhat-operators-jzxtj\" (UID: \"471a271b-2d62-4bdb-956d-91e268d2e4b5\") " pod="openshift-marketplace/redhat-operators-jzxtj" Mar 13 15:28:41 crc kubenswrapper[4907]: I0313 15:28:41.003278 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g76wj\" (UniqueName: \"kubernetes.io/projected/471a271b-2d62-4bdb-956d-91e268d2e4b5-kube-api-access-g76wj\") pod \"redhat-operators-jzxtj\" (UID: \"471a271b-2d62-4bdb-956d-91e268d2e4b5\") " pod="openshift-marketplace/redhat-operators-jzxtj" Mar 13 15:28:41 crc kubenswrapper[4907]: I0313 15:28:41.082981 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jzxtj" Mar 13 15:28:41 crc kubenswrapper[4907]: I0313 15:28:41.518568 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jzxtj"] Mar 13 15:28:42 crc kubenswrapper[4907]: I0313 15:28:42.347133 4907 generic.go:334] "Generic (PLEG): container finished" podID="471a271b-2d62-4bdb-956d-91e268d2e4b5" containerID="717b6dbc97d7ec5e02f0d19fcad760520fac7eb81149b49eb75a2be915500239" exitCode=0 Mar 13 15:28:42 crc kubenswrapper[4907]: I0313 15:28:42.347185 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jzxtj" event={"ID":"471a271b-2d62-4bdb-956d-91e268d2e4b5","Type":"ContainerDied","Data":"717b6dbc97d7ec5e02f0d19fcad760520fac7eb81149b49eb75a2be915500239"} Mar 13 15:28:42 crc kubenswrapper[4907]: I0313 15:28:42.347212 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jzxtj" event={"ID":"471a271b-2d62-4bdb-956d-91e268d2e4b5","Type":"ContainerStarted","Data":"402cc4345879db4aec2b048632e5651fc97bb6231de9ed25658ecbdd0404e855"} Mar 13 15:28:44 crc kubenswrapper[4907]: I0313 15:28:44.365126 4907 generic.go:334] "Generic (PLEG): container finished" podID="471a271b-2d62-4bdb-956d-91e268d2e4b5" containerID="6aab4dd9c3bca6c8c51dfdb1e941984b5dcd418ed1ef52f4155b9cfe47b28ac7" exitCode=0 Mar 13 15:28:44 crc kubenswrapper[4907]: I0313 15:28:44.365174 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jzxtj" event={"ID":"471a271b-2d62-4bdb-956d-91e268d2e4b5","Type":"ContainerDied","Data":"6aab4dd9c3bca6c8c51dfdb1e941984b5dcd418ed1ef52f4155b9cfe47b28ac7"} Mar 13 15:28:45 crc kubenswrapper[4907]: I0313 15:28:45.375916 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jzxtj" event={"ID":"471a271b-2d62-4bdb-956d-91e268d2e4b5","Type":"ContainerStarted","Data":"23e6b0d481bf70d0a2451207b0ebb74ac5446201d8e102faf3f204db678a0312"} Mar 13 15:28:45 crc kubenswrapper[4907]: I0313 15:28:45.402486 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-jzxtj" podStartSLOduration=2.96754561 podStartE2EDuration="5.402465886s" podCreationTimestamp="2026-03-13 15:28:40 +0000 UTC" firstStartedPulling="2026-03-13 15:28:42.348805534 +0000 UTC m=+5021.248593223" lastFinishedPulling="2026-03-13 15:28:44.78372581 +0000 UTC m=+5023.683513499" observedRunningTime="2026-03-13 15:28:45.395032903 +0000 UTC m=+5024.294820592" watchObservedRunningTime="2026-03-13 15:28:45.402465886 +0000 UTC m=+5024.302253575" Mar 13 15:28:48 crc kubenswrapper[4907]: I0313 15:28:48.042245 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 15:28:48 crc kubenswrapper[4907]: I0313 15:28:48.042821 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 15:28:51 crc kubenswrapper[4907]: I0313 15:28:51.083955 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-jzxtj" Mar 13 15:28:51 crc kubenswrapper[4907]: I0313 15:28:51.085154 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-jzxtj" Mar 13 15:28:52 crc kubenswrapper[4907]: I0313 15:28:52.126839 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-jzxtj" podUID="471a271b-2d62-4bdb-956d-91e268d2e4b5" containerName="registry-server" probeResult="failure" output=< Mar 13 15:28:52 crc kubenswrapper[4907]: timeout: failed to connect service ":50051" within 1s Mar 13 15:28:52 crc kubenswrapper[4907]: > Mar 13 15:29:01 crc kubenswrapper[4907]: I0313 15:29:01.130849 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-jzxtj" Mar 13 15:29:01 crc kubenswrapper[4907]: I0313 15:29:01.275580 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-jzxtj" Mar 13 15:29:01 crc kubenswrapper[4907]: I0313 15:29:01.383328 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jzxtj"] Mar 13 15:29:02 crc kubenswrapper[4907]: I0313 15:29:02.492482 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-jzxtj" podUID="471a271b-2d62-4bdb-956d-91e268d2e4b5" containerName="registry-server" containerID="cri-o://23e6b0d481bf70d0a2451207b0ebb74ac5446201d8e102faf3f204db678a0312" gracePeriod=2 Mar 13 15:29:02 crc kubenswrapper[4907]: I0313 15:29:02.895638 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jzxtj" Mar 13 15:29:03 crc kubenswrapper[4907]: I0313 15:29:03.005631 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/471a271b-2d62-4bdb-956d-91e268d2e4b5-catalog-content\") pod \"471a271b-2d62-4bdb-956d-91e268d2e4b5\" (UID: \"471a271b-2d62-4bdb-956d-91e268d2e4b5\") " Mar 13 15:29:03 crc kubenswrapper[4907]: I0313 15:29:03.005717 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g76wj\" (UniqueName: \"kubernetes.io/projected/471a271b-2d62-4bdb-956d-91e268d2e4b5-kube-api-access-g76wj\") pod \"471a271b-2d62-4bdb-956d-91e268d2e4b5\" (UID: \"471a271b-2d62-4bdb-956d-91e268d2e4b5\") " Mar 13 15:29:03 crc kubenswrapper[4907]: I0313 15:29:03.005846 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/471a271b-2d62-4bdb-956d-91e268d2e4b5-utilities\") pod \"471a271b-2d62-4bdb-956d-91e268d2e4b5\" (UID: \"471a271b-2d62-4bdb-956d-91e268d2e4b5\") " Mar 13 15:29:03 crc kubenswrapper[4907]: I0313 15:29:03.006503 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/471a271b-2d62-4bdb-956d-91e268d2e4b5-utilities" (OuterVolumeSpecName: "utilities") pod "471a271b-2d62-4bdb-956d-91e268d2e4b5" (UID: "471a271b-2d62-4bdb-956d-91e268d2e4b5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:29:03 crc kubenswrapper[4907]: I0313 15:29:03.010741 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/471a271b-2d62-4bdb-956d-91e268d2e4b5-kube-api-access-g76wj" (OuterVolumeSpecName: "kube-api-access-g76wj") pod "471a271b-2d62-4bdb-956d-91e268d2e4b5" (UID: "471a271b-2d62-4bdb-956d-91e268d2e4b5"). InnerVolumeSpecName "kube-api-access-g76wj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:29:03 crc kubenswrapper[4907]: I0313 15:29:03.107640 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/471a271b-2d62-4bdb-956d-91e268d2e4b5-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 15:29:03 crc kubenswrapper[4907]: I0313 15:29:03.107704 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g76wj\" (UniqueName: \"kubernetes.io/projected/471a271b-2d62-4bdb-956d-91e268d2e4b5-kube-api-access-g76wj\") on node \"crc\" DevicePath \"\"" Mar 13 15:29:03 crc kubenswrapper[4907]: I0313 15:29:03.133723 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/471a271b-2d62-4bdb-956d-91e268d2e4b5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "471a271b-2d62-4bdb-956d-91e268d2e4b5" (UID: "471a271b-2d62-4bdb-956d-91e268d2e4b5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:29:03 crc kubenswrapper[4907]: I0313 15:29:03.209141 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/471a271b-2d62-4bdb-956d-91e268d2e4b5-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 15:29:03 crc kubenswrapper[4907]: I0313 15:29:03.503124 4907 generic.go:334] "Generic (PLEG): container finished" podID="471a271b-2d62-4bdb-956d-91e268d2e4b5" containerID="23e6b0d481bf70d0a2451207b0ebb74ac5446201d8e102faf3f204db678a0312" exitCode=0 Mar 13 15:29:03 crc kubenswrapper[4907]: I0313 15:29:03.503181 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jzxtj" event={"ID":"471a271b-2d62-4bdb-956d-91e268d2e4b5","Type":"ContainerDied","Data":"23e6b0d481bf70d0a2451207b0ebb74ac5446201d8e102faf3f204db678a0312"} Mar 13 15:29:03 crc kubenswrapper[4907]: I0313 15:29:03.503202 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jzxtj" Mar 13 15:29:03 crc kubenswrapper[4907]: I0313 15:29:03.503224 4907 scope.go:117] "RemoveContainer" containerID="23e6b0d481bf70d0a2451207b0ebb74ac5446201d8e102faf3f204db678a0312" Mar 13 15:29:03 crc kubenswrapper[4907]: I0313 15:29:03.503213 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jzxtj" event={"ID":"471a271b-2d62-4bdb-956d-91e268d2e4b5","Type":"ContainerDied","Data":"402cc4345879db4aec2b048632e5651fc97bb6231de9ed25658ecbdd0404e855"} Mar 13 15:29:03 crc kubenswrapper[4907]: I0313 15:29:03.520522 4907 scope.go:117] "RemoveContainer" containerID="6aab4dd9c3bca6c8c51dfdb1e941984b5dcd418ed1ef52f4155b9cfe47b28ac7" Mar 13 15:29:03 crc kubenswrapper[4907]: I0313 15:29:03.542562 4907 scope.go:117] "RemoveContainer" containerID="717b6dbc97d7ec5e02f0d19fcad760520fac7eb81149b49eb75a2be915500239" Mar 13 15:29:03 crc kubenswrapper[4907]: I0313 15:29:03.577560 4907 scope.go:117] "RemoveContainer" containerID="23e6b0d481bf70d0a2451207b0ebb74ac5446201d8e102faf3f204db678a0312" Mar 13 15:29:03 crc kubenswrapper[4907]: E0313 15:29:03.578261 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"23e6b0d481bf70d0a2451207b0ebb74ac5446201d8e102faf3f204db678a0312\": container with ID starting with 23e6b0d481bf70d0a2451207b0ebb74ac5446201d8e102faf3f204db678a0312 not found: ID does not exist" containerID="23e6b0d481bf70d0a2451207b0ebb74ac5446201d8e102faf3f204db678a0312" Mar 13 15:29:03 crc kubenswrapper[4907]: I0313 15:29:03.578320 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"23e6b0d481bf70d0a2451207b0ebb74ac5446201d8e102faf3f204db678a0312"} err="failed to get container status \"23e6b0d481bf70d0a2451207b0ebb74ac5446201d8e102faf3f204db678a0312\": rpc error: code = NotFound desc = could not find container \"23e6b0d481bf70d0a2451207b0ebb74ac5446201d8e102faf3f204db678a0312\": container with ID starting with 23e6b0d481bf70d0a2451207b0ebb74ac5446201d8e102faf3f204db678a0312 not found: ID does not exist" Mar 13 15:29:03 crc kubenswrapper[4907]: I0313 15:29:03.578358 4907 scope.go:117] "RemoveContainer" containerID="6aab4dd9c3bca6c8c51dfdb1e941984b5dcd418ed1ef52f4155b9cfe47b28ac7" Mar 13 15:29:03 crc kubenswrapper[4907]: E0313 15:29:03.578813 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6aab4dd9c3bca6c8c51dfdb1e941984b5dcd418ed1ef52f4155b9cfe47b28ac7\": container with ID starting with 6aab4dd9c3bca6c8c51dfdb1e941984b5dcd418ed1ef52f4155b9cfe47b28ac7 not found: ID does not exist" containerID="6aab4dd9c3bca6c8c51dfdb1e941984b5dcd418ed1ef52f4155b9cfe47b28ac7" Mar 13 15:29:03 crc kubenswrapper[4907]: I0313 15:29:03.578836 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6aab4dd9c3bca6c8c51dfdb1e941984b5dcd418ed1ef52f4155b9cfe47b28ac7"} err="failed to get container status \"6aab4dd9c3bca6c8c51dfdb1e941984b5dcd418ed1ef52f4155b9cfe47b28ac7\": rpc error: code = NotFound desc = could not find container \"6aab4dd9c3bca6c8c51dfdb1e941984b5dcd418ed1ef52f4155b9cfe47b28ac7\": container with ID starting with 6aab4dd9c3bca6c8c51dfdb1e941984b5dcd418ed1ef52f4155b9cfe47b28ac7 not found: ID does not exist" Mar 13 15:29:03 crc kubenswrapper[4907]: I0313 15:29:03.578852 4907 scope.go:117] "RemoveContainer" containerID="717b6dbc97d7ec5e02f0d19fcad760520fac7eb81149b49eb75a2be915500239" Mar 13 15:29:03 crc kubenswrapper[4907]: E0313 15:29:03.579501 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"717b6dbc97d7ec5e02f0d19fcad760520fac7eb81149b49eb75a2be915500239\": container with ID starting with 717b6dbc97d7ec5e02f0d19fcad760520fac7eb81149b49eb75a2be915500239 not found: ID does not exist" containerID="717b6dbc97d7ec5e02f0d19fcad760520fac7eb81149b49eb75a2be915500239" Mar 13 15:29:03 crc kubenswrapper[4907]: I0313 15:29:03.579540 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"717b6dbc97d7ec5e02f0d19fcad760520fac7eb81149b49eb75a2be915500239"} err="failed to get container status \"717b6dbc97d7ec5e02f0d19fcad760520fac7eb81149b49eb75a2be915500239\": rpc error: code = NotFound desc = could not find container \"717b6dbc97d7ec5e02f0d19fcad760520fac7eb81149b49eb75a2be915500239\": container with ID starting with 717b6dbc97d7ec5e02f0d19fcad760520fac7eb81149b49eb75a2be915500239 not found: ID does not exist" Mar 13 15:29:03 crc kubenswrapper[4907]: I0313 15:29:03.580162 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jzxtj"] Mar 13 15:29:03 crc kubenswrapper[4907]: I0313 15:29:03.586025 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-jzxtj"] Mar 13 15:29:03 crc kubenswrapper[4907]: I0313 15:29:03.802039 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="471a271b-2d62-4bdb-956d-91e268d2e4b5" path="/var/lib/kubelet/pods/471a271b-2d62-4bdb-956d-91e268d2e4b5/volumes" Mar 13 15:29:18 crc kubenswrapper[4907]: I0313 15:29:18.041368 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 15:29:18 crc kubenswrapper[4907]: I0313 15:29:18.042139 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 15:29:48 crc kubenswrapper[4907]: I0313 15:29:48.041575 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 15:29:48 crc kubenswrapper[4907]: I0313 15:29:48.042139 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 15:29:48 crc kubenswrapper[4907]: I0313 15:29:48.042181 4907 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" Mar 13 15:29:48 crc kubenswrapper[4907]: I0313 15:29:48.042642 4907 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"592bf9f175f440e943369aab2abbb13ad2c26412cb0e0738c4a33a156b8a5596"} pod="openshift-machine-config-operator/machine-config-daemon-hm56j" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 15:29:48 crc kubenswrapper[4907]: I0313 15:29:48.042684 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" containerID="cri-o://592bf9f175f440e943369aab2abbb13ad2c26412cb0e0738c4a33a156b8a5596" gracePeriod=600 Mar 13 15:29:48 crc kubenswrapper[4907]: E0313 15:29:48.165022 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:29:48 crc kubenswrapper[4907]: I0313 15:29:48.828922 4907 generic.go:334] "Generic (PLEG): container finished" podID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerID="592bf9f175f440e943369aab2abbb13ad2c26412cb0e0738c4a33a156b8a5596" exitCode=0 Mar 13 15:29:48 crc kubenswrapper[4907]: I0313 15:29:48.828970 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerDied","Data":"592bf9f175f440e943369aab2abbb13ad2c26412cb0e0738c4a33a156b8a5596"} Mar 13 15:29:48 crc kubenswrapper[4907]: I0313 15:29:48.829008 4907 scope.go:117] "RemoveContainer" containerID="b5bb3cf2979988d507615c0a856eebc9a02c4d2c3275ceb2cee45dcc6b625cfd" Mar 13 15:29:48 crc kubenswrapper[4907]: I0313 15:29:48.829659 4907 scope.go:117] "RemoveContainer" containerID="592bf9f175f440e943369aab2abbb13ad2c26412cb0e0738c4a33a156b8a5596" Mar 13 15:29:48 crc kubenswrapper[4907]: E0313 15:29:48.830036 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:30:00 crc kubenswrapper[4907]: I0313 15:30:00.145461 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556930-f852z"] Mar 13 15:30:00 crc kubenswrapper[4907]: E0313 15:30:00.146402 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="471a271b-2d62-4bdb-956d-91e268d2e4b5" containerName="extract-utilities" Mar 13 15:30:00 crc kubenswrapper[4907]: I0313 15:30:00.146420 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="471a271b-2d62-4bdb-956d-91e268d2e4b5" containerName="extract-utilities" Mar 13 15:30:00 crc kubenswrapper[4907]: E0313 15:30:00.146439 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="471a271b-2d62-4bdb-956d-91e268d2e4b5" containerName="registry-server" Mar 13 15:30:00 crc kubenswrapper[4907]: I0313 15:30:00.146447 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="471a271b-2d62-4bdb-956d-91e268d2e4b5" containerName="registry-server" Mar 13 15:30:00 crc kubenswrapper[4907]: E0313 15:30:00.146475 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="471a271b-2d62-4bdb-956d-91e268d2e4b5" containerName="extract-content" Mar 13 15:30:00 crc kubenswrapper[4907]: I0313 15:30:00.146483 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="471a271b-2d62-4bdb-956d-91e268d2e4b5" containerName="extract-content" Mar 13 15:30:00 crc kubenswrapper[4907]: I0313 15:30:00.146648 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="471a271b-2d62-4bdb-956d-91e268d2e4b5" containerName="registry-server" Mar 13 15:30:00 crc kubenswrapper[4907]: I0313 15:30:00.147330 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556930-f852z" Mar 13 15:30:00 crc kubenswrapper[4907]: I0313 15:30:00.149802 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 13 15:30:00 crc kubenswrapper[4907]: I0313 15:30:00.151367 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 13 15:30:00 crc kubenswrapper[4907]: I0313 15:30:00.156527 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556930-ndtqs"] Mar 13 15:30:00 crc kubenswrapper[4907]: I0313 15:30:00.162371 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556930-ndtqs" Mar 13 15:30:00 crc kubenswrapper[4907]: I0313 15:30:00.176236 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 15:30:00 crc kubenswrapper[4907]: I0313 15:30:00.177772 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 15:30:00 crc kubenswrapper[4907]: I0313 15:30:00.181739 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 15:30:00 crc kubenswrapper[4907]: I0313 15:30:00.193099 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556930-ndtqs"] Mar 13 15:30:00 crc kubenswrapper[4907]: I0313 15:30:00.197838 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556930-f852z"] Mar 13 15:30:00 crc kubenswrapper[4907]: I0313 15:30:00.269147 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2983956e-5d8d-4595-9fda-91759741010a-config-volume\") pod \"collect-profiles-29556930-f852z\" (UID: \"2983956e-5d8d-4595-9fda-91759741010a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556930-f852z" Mar 13 15:30:00 crc kubenswrapper[4907]: I0313 15:30:00.269203 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ctk6p\" (UniqueName: \"kubernetes.io/projected/df83f491-eb7d-4e1c-b814-28cc7ef9abee-kube-api-access-ctk6p\") pod \"auto-csr-approver-29556930-ndtqs\" (UID: \"df83f491-eb7d-4e1c-b814-28cc7ef9abee\") " pod="openshift-infra/auto-csr-approver-29556930-ndtqs" Mar 13 15:30:00 crc kubenswrapper[4907]: I0313 15:30:00.269235 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2983956e-5d8d-4595-9fda-91759741010a-secret-volume\") pod \"collect-profiles-29556930-f852z\" (UID: \"2983956e-5d8d-4595-9fda-91759741010a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556930-f852z" Mar 13 15:30:00 crc kubenswrapper[4907]: I0313 15:30:00.269537 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jrrsh\" (UniqueName: \"kubernetes.io/projected/2983956e-5d8d-4595-9fda-91759741010a-kube-api-access-jrrsh\") pod \"collect-profiles-29556930-f852z\" (UID: \"2983956e-5d8d-4595-9fda-91759741010a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556930-f852z" Mar 13 15:30:00 crc kubenswrapper[4907]: I0313 15:30:00.371043 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2983956e-5d8d-4595-9fda-91759741010a-config-volume\") pod \"collect-profiles-29556930-f852z\" (UID: \"2983956e-5d8d-4595-9fda-91759741010a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556930-f852z" Mar 13 15:30:00 crc kubenswrapper[4907]: I0313 15:30:00.371099 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ctk6p\" (UniqueName: \"kubernetes.io/projected/df83f491-eb7d-4e1c-b814-28cc7ef9abee-kube-api-access-ctk6p\") pod \"auto-csr-approver-29556930-ndtqs\" (UID: \"df83f491-eb7d-4e1c-b814-28cc7ef9abee\") " pod="openshift-infra/auto-csr-approver-29556930-ndtqs" Mar 13 15:30:00 crc kubenswrapper[4907]: I0313 15:30:00.371126 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2983956e-5d8d-4595-9fda-91759741010a-secret-volume\") pod \"collect-profiles-29556930-f852z\" (UID: \"2983956e-5d8d-4595-9fda-91759741010a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556930-f852z" Mar 13 15:30:00 crc kubenswrapper[4907]: I0313 15:30:00.371227 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jrrsh\" (UniqueName: \"kubernetes.io/projected/2983956e-5d8d-4595-9fda-91759741010a-kube-api-access-jrrsh\") pod \"collect-profiles-29556930-f852z\" (UID: \"2983956e-5d8d-4595-9fda-91759741010a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556930-f852z" Mar 13 15:30:00 crc kubenswrapper[4907]: I0313 15:30:00.372155 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2983956e-5d8d-4595-9fda-91759741010a-config-volume\") pod \"collect-profiles-29556930-f852z\" (UID: \"2983956e-5d8d-4595-9fda-91759741010a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556930-f852z" Mar 13 15:30:00 crc kubenswrapper[4907]: I0313 15:30:00.378200 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2983956e-5d8d-4595-9fda-91759741010a-secret-volume\") pod \"collect-profiles-29556930-f852z\" (UID: \"2983956e-5d8d-4595-9fda-91759741010a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556930-f852z" Mar 13 15:30:00 crc kubenswrapper[4907]: I0313 15:30:00.390217 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ctk6p\" (UniqueName: \"kubernetes.io/projected/df83f491-eb7d-4e1c-b814-28cc7ef9abee-kube-api-access-ctk6p\") pod \"auto-csr-approver-29556930-ndtqs\" (UID: \"df83f491-eb7d-4e1c-b814-28cc7ef9abee\") " pod="openshift-infra/auto-csr-approver-29556930-ndtqs" Mar 13 15:30:00 crc kubenswrapper[4907]: I0313 15:30:00.390304 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jrrsh\" (UniqueName: \"kubernetes.io/projected/2983956e-5d8d-4595-9fda-91759741010a-kube-api-access-jrrsh\") pod \"collect-profiles-29556930-f852z\" (UID: \"2983956e-5d8d-4595-9fda-91759741010a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556930-f852z" Mar 13 15:30:00 crc kubenswrapper[4907]: I0313 15:30:00.470631 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556930-f852z" Mar 13 15:30:00 crc kubenswrapper[4907]: I0313 15:30:00.490827 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556930-ndtqs" Mar 13 15:30:00 crc kubenswrapper[4907]: I0313 15:30:00.782726 4907 scope.go:117] "RemoveContainer" containerID="592bf9f175f440e943369aab2abbb13ad2c26412cb0e0738c4a33a156b8a5596" Mar 13 15:30:00 crc kubenswrapper[4907]: E0313 15:30:00.783154 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:30:00 crc kubenswrapper[4907]: I0313 15:30:00.905108 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556930-f852z"] Mar 13 15:30:00 crc kubenswrapper[4907]: I0313 15:30:00.929026 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556930-f852z" event={"ID":"2983956e-5d8d-4595-9fda-91759741010a","Type":"ContainerStarted","Data":"0617586958aed9d4fa53068330454c7a10480fc0c59b1ca5cc83cc8bc439a18b"} Mar 13 15:30:00 crc kubenswrapper[4907]: I0313 15:30:00.964766 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556930-ndtqs"] Mar 13 15:30:00 crc kubenswrapper[4907]: W0313 15:30:00.967138 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddf83f491_eb7d_4e1c_b814_28cc7ef9abee.slice/crio-3091fd9a1e1aec73474d08add55adf04d2c04dd62f815e3f3b5c66e65aca9591 WatchSource:0}: Error finding container 3091fd9a1e1aec73474d08add55adf04d2c04dd62f815e3f3b5c66e65aca9591: Status 404 returned error can't find the container with id 3091fd9a1e1aec73474d08add55adf04d2c04dd62f815e3f3b5c66e65aca9591 Mar 13 15:30:01 crc kubenswrapper[4907]: I0313 15:30:01.936454 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556930-ndtqs" event={"ID":"df83f491-eb7d-4e1c-b814-28cc7ef9abee","Type":"ContainerStarted","Data":"3091fd9a1e1aec73474d08add55adf04d2c04dd62f815e3f3b5c66e65aca9591"} Mar 13 15:30:01 crc kubenswrapper[4907]: I0313 15:30:01.937728 4907 generic.go:334] "Generic (PLEG): container finished" podID="2983956e-5d8d-4595-9fda-91759741010a" containerID="74d889f05e8d8fa3b0cfc7fed167cff28426e276bf0f0c9f759dd1496f3c1058" exitCode=0 Mar 13 15:30:01 crc kubenswrapper[4907]: I0313 15:30:01.937773 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556930-f852z" event={"ID":"2983956e-5d8d-4595-9fda-91759741010a","Type":"ContainerDied","Data":"74d889f05e8d8fa3b0cfc7fed167cff28426e276bf0f0c9f759dd1496f3c1058"} Mar 13 15:30:02 crc kubenswrapper[4907]: I0313 15:30:02.946778 4907 generic.go:334] "Generic (PLEG): container finished" podID="df83f491-eb7d-4e1c-b814-28cc7ef9abee" containerID="9c966e9f83d2d89ea300010e1e5e03c7e34fbdc9099449a4ffcb481299d87028" exitCode=0 Mar 13 15:30:02 crc kubenswrapper[4907]: I0313 15:30:02.947416 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556930-ndtqs" event={"ID":"df83f491-eb7d-4e1c-b814-28cc7ef9abee","Type":"ContainerDied","Data":"9c966e9f83d2d89ea300010e1e5e03c7e34fbdc9099449a4ffcb481299d87028"} Mar 13 15:30:03 crc kubenswrapper[4907]: I0313 15:30:03.219736 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556930-f852z" Mar 13 15:30:03 crc kubenswrapper[4907]: I0313 15:30:03.314114 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jrrsh\" (UniqueName: \"kubernetes.io/projected/2983956e-5d8d-4595-9fda-91759741010a-kube-api-access-jrrsh\") pod \"2983956e-5d8d-4595-9fda-91759741010a\" (UID: \"2983956e-5d8d-4595-9fda-91759741010a\") " Mar 13 15:30:03 crc kubenswrapper[4907]: I0313 15:30:03.314162 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2983956e-5d8d-4595-9fda-91759741010a-secret-volume\") pod \"2983956e-5d8d-4595-9fda-91759741010a\" (UID: \"2983956e-5d8d-4595-9fda-91759741010a\") " Mar 13 15:30:03 crc kubenswrapper[4907]: I0313 15:30:03.314205 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2983956e-5d8d-4595-9fda-91759741010a-config-volume\") pod \"2983956e-5d8d-4595-9fda-91759741010a\" (UID: \"2983956e-5d8d-4595-9fda-91759741010a\") " Mar 13 15:30:03 crc kubenswrapper[4907]: I0313 15:30:03.315190 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2983956e-5d8d-4595-9fda-91759741010a-config-volume" (OuterVolumeSpecName: "config-volume") pod "2983956e-5d8d-4595-9fda-91759741010a" (UID: "2983956e-5d8d-4595-9fda-91759741010a"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:30:03 crc kubenswrapper[4907]: I0313 15:30:03.320696 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2983956e-5d8d-4595-9fda-91759741010a-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "2983956e-5d8d-4595-9fda-91759741010a" (UID: "2983956e-5d8d-4595-9fda-91759741010a"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:30:03 crc kubenswrapper[4907]: I0313 15:30:03.321215 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2983956e-5d8d-4595-9fda-91759741010a-kube-api-access-jrrsh" (OuterVolumeSpecName: "kube-api-access-jrrsh") pod "2983956e-5d8d-4595-9fda-91759741010a" (UID: "2983956e-5d8d-4595-9fda-91759741010a"). InnerVolumeSpecName "kube-api-access-jrrsh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:30:03 crc kubenswrapper[4907]: I0313 15:30:03.416183 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jrrsh\" (UniqueName: \"kubernetes.io/projected/2983956e-5d8d-4595-9fda-91759741010a-kube-api-access-jrrsh\") on node \"crc\" DevicePath \"\"" Mar 13 15:30:03 crc kubenswrapper[4907]: I0313 15:30:03.416230 4907 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2983956e-5d8d-4595-9fda-91759741010a-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 13 15:30:03 crc kubenswrapper[4907]: I0313 15:30:03.416244 4907 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2983956e-5d8d-4595-9fda-91759741010a-config-volume\") on node \"crc\" DevicePath \"\"" Mar 13 15:30:03 crc kubenswrapper[4907]: I0313 15:30:03.955929 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556930-f852z" Mar 13 15:30:03 crc kubenswrapper[4907]: I0313 15:30:03.955930 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556930-f852z" event={"ID":"2983956e-5d8d-4595-9fda-91759741010a","Type":"ContainerDied","Data":"0617586958aed9d4fa53068330454c7a10480fc0c59b1ca5cc83cc8bc439a18b"} Mar 13 15:30:03 crc kubenswrapper[4907]: I0313 15:30:03.956353 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0617586958aed9d4fa53068330454c7a10480fc0c59b1ca5cc83cc8bc439a18b" Mar 13 15:30:04 crc kubenswrapper[4907]: I0313 15:30:04.233925 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556930-ndtqs" Mar 13 15:30:04 crc kubenswrapper[4907]: I0313 15:30:04.303910 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556885-9xsw6"] Mar 13 15:30:04 crc kubenswrapper[4907]: I0313 15:30:04.319139 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556885-9xsw6"] Mar 13 15:30:04 crc kubenswrapper[4907]: I0313 15:30:04.328832 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ctk6p\" (UniqueName: \"kubernetes.io/projected/df83f491-eb7d-4e1c-b814-28cc7ef9abee-kube-api-access-ctk6p\") pod \"df83f491-eb7d-4e1c-b814-28cc7ef9abee\" (UID: \"df83f491-eb7d-4e1c-b814-28cc7ef9abee\") " Mar 13 15:30:04 crc kubenswrapper[4907]: I0313 15:30:04.334088 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/df83f491-eb7d-4e1c-b814-28cc7ef9abee-kube-api-access-ctk6p" (OuterVolumeSpecName: "kube-api-access-ctk6p") pod "df83f491-eb7d-4e1c-b814-28cc7ef9abee" (UID: "df83f491-eb7d-4e1c-b814-28cc7ef9abee"). InnerVolumeSpecName "kube-api-access-ctk6p". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:30:04 crc kubenswrapper[4907]: I0313 15:30:04.430550 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ctk6p\" (UniqueName: \"kubernetes.io/projected/df83f491-eb7d-4e1c-b814-28cc7ef9abee-kube-api-access-ctk6p\") on node \"crc\" DevicePath \"\"" Mar 13 15:30:04 crc kubenswrapper[4907]: I0313 15:30:04.964735 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556930-ndtqs" event={"ID":"df83f491-eb7d-4e1c-b814-28cc7ef9abee","Type":"ContainerDied","Data":"3091fd9a1e1aec73474d08add55adf04d2c04dd62f815e3f3b5c66e65aca9591"} Mar 13 15:30:04 crc kubenswrapper[4907]: I0313 15:30:04.965041 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3091fd9a1e1aec73474d08add55adf04d2c04dd62f815e3f3b5c66e65aca9591" Mar 13 15:30:04 crc kubenswrapper[4907]: I0313 15:30:04.964776 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556930-ndtqs" Mar 13 15:30:05 crc kubenswrapper[4907]: I0313 15:30:05.295617 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556924-8z5rs"] Mar 13 15:30:05 crc kubenswrapper[4907]: I0313 15:30:05.303218 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556924-8z5rs"] Mar 13 15:30:05 crc kubenswrapper[4907]: I0313 15:30:05.793529 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="11a50278-3da1-4a63-ac7e-1fcb4ae77f52" path="/var/lib/kubelet/pods/11a50278-3da1-4a63-ac7e-1fcb4ae77f52/volumes" Mar 13 15:30:05 crc kubenswrapper[4907]: I0313 15:30:05.794645 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d79a5eae-26dc-4f69-aa63-9bd6c5788a03" path="/var/lib/kubelet/pods/d79a5eae-26dc-4f69-aa63-9bd6c5788a03/volumes" Mar 13 15:30:11 crc kubenswrapper[4907]: I0313 15:30:11.786140 4907 scope.go:117] "RemoveContainer" containerID="592bf9f175f440e943369aab2abbb13ad2c26412cb0e0738c4a33a156b8a5596" Mar 13 15:30:11 crc kubenswrapper[4907]: E0313 15:30:11.787658 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:30:24 crc kubenswrapper[4907]: I0313 15:30:24.783160 4907 scope.go:117] "RemoveContainer" containerID="592bf9f175f440e943369aab2abbb13ad2c26412cb0e0738c4a33a156b8a5596" Mar 13 15:30:24 crc kubenswrapper[4907]: E0313 15:30:24.783825 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:30:27 crc kubenswrapper[4907]: I0313 15:30:27.604319 4907 scope.go:117] "RemoveContainer" containerID="34346eaf9fc976e698b78600d581550e67e1b840574df19f7129d1f617293a7d" Mar 13 15:30:27 crc kubenswrapper[4907]: I0313 15:30:27.633308 4907 scope.go:117] "RemoveContainer" containerID="16a638d449eaf9553071c678dcf92f038cd2ddde8f659099ade22f5e78a0fcae" Mar 13 15:30:38 crc kubenswrapper[4907]: I0313 15:30:38.782687 4907 scope.go:117] "RemoveContainer" containerID="592bf9f175f440e943369aab2abbb13ad2c26412cb0e0738c4a33a156b8a5596" Mar 13 15:30:38 crc kubenswrapper[4907]: E0313 15:30:38.783366 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:30:50 crc kubenswrapper[4907]: I0313 15:30:50.782797 4907 scope.go:117] "RemoveContainer" containerID="592bf9f175f440e943369aab2abbb13ad2c26412cb0e0738c4a33a156b8a5596" Mar 13 15:30:50 crc kubenswrapper[4907]: E0313 15:30:50.783618 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:31:01 crc kubenswrapper[4907]: I0313 15:31:01.782373 4907 scope.go:117] "RemoveContainer" containerID="592bf9f175f440e943369aab2abbb13ad2c26412cb0e0738c4a33a156b8a5596" Mar 13 15:31:01 crc kubenswrapper[4907]: E0313 15:31:01.783033 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:31:13 crc kubenswrapper[4907]: I0313 15:31:13.782674 4907 scope.go:117] "RemoveContainer" containerID="592bf9f175f440e943369aab2abbb13ad2c26412cb0e0738c4a33a156b8a5596" Mar 13 15:31:13 crc kubenswrapper[4907]: E0313 15:31:13.783426 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:31:27 crc kubenswrapper[4907]: I0313 15:31:27.782572 4907 scope.go:117] "RemoveContainer" containerID="592bf9f175f440e943369aab2abbb13ad2c26412cb0e0738c4a33a156b8a5596" Mar 13 15:31:27 crc kubenswrapper[4907]: E0313 15:31:27.785538 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:31:42 crc kubenswrapper[4907]: I0313 15:31:42.782927 4907 scope.go:117] "RemoveContainer" containerID="592bf9f175f440e943369aab2abbb13ad2c26412cb0e0738c4a33a156b8a5596" Mar 13 15:31:42 crc kubenswrapper[4907]: E0313 15:31:42.783725 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:31:54 crc kubenswrapper[4907]: I0313 15:31:54.783389 4907 scope.go:117] "RemoveContainer" containerID="592bf9f175f440e943369aab2abbb13ad2c26412cb0e0738c4a33a156b8a5596" Mar 13 15:31:54 crc kubenswrapper[4907]: E0313 15:31:54.784226 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:32:00 crc kubenswrapper[4907]: I0313 15:32:00.136504 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556932-fxhnm"] Mar 13 15:32:00 crc kubenswrapper[4907]: E0313 15:32:00.138788 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df83f491-eb7d-4e1c-b814-28cc7ef9abee" containerName="oc" Mar 13 15:32:00 crc kubenswrapper[4907]: I0313 15:32:00.138970 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="df83f491-eb7d-4e1c-b814-28cc7ef9abee" containerName="oc" Mar 13 15:32:00 crc kubenswrapper[4907]: E0313 15:32:00.139100 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2983956e-5d8d-4595-9fda-91759741010a" containerName="collect-profiles" Mar 13 15:32:00 crc kubenswrapper[4907]: I0313 15:32:00.139183 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="2983956e-5d8d-4595-9fda-91759741010a" containerName="collect-profiles" Mar 13 15:32:00 crc kubenswrapper[4907]: I0313 15:32:00.139433 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="df83f491-eb7d-4e1c-b814-28cc7ef9abee" containerName="oc" Mar 13 15:32:00 crc kubenswrapper[4907]: I0313 15:32:00.139534 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="2983956e-5d8d-4595-9fda-91759741010a" containerName="collect-profiles" Mar 13 15:32:00 crc kubenswrapper[4907]: I0313 15:32:00.140302 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556932-fxhnm" Mar 13 15:32:00 crc kubenswrapper[4907]: I0313 15:32:00.142917 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 15:32:00 crc kubenswrapper[4907]: I0313 15:32:00.143102 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 15:32:00 crc kubenswrapper[4907]: I0313 15:32:00.143566 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 15:32:00 crc kubenswrapper[4907]: I0313 15:32:00.145282 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556932-fxhnm"] Mar 13 15:32:00 crc kubenswrapper[4907]: I0313 15:32:00.324159 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n2cs8\" (UniqueName: \"kubernetes.io/projected/e5efd747-3795-4922-9f81-3ef02dcb2f91-kube-api-access-n2cs8\") pod \"auto-csr-approver-29556932-fxhnm\" (UID: \"e5efd747-3795-4922-9f81-3ef02dcb2f91\") " pod="openshift-infra/auto-csr-approver-29556932-fxhnm" Mar 13 15:32:00 crc kubenswrapper[4907]: I0313 15:32:00.426076 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n2cs8\" (UniqueName: \"kubernetes.io/projected/e5efd747-3795-4922-9f81-3ef02dcb2f91-kube-api-access-n2cs8\") pod \"auto-csr-approver-29556932-fxhnm\" (UID: \"e5efd747-3795-4922-9f81-3ef02dcb2f91\") " pod="openshift-infra/auto-csr-approver-29556932-fxhnm" Mar 13 15:32:00 crc kubenswrapper[4907]: I0313 15:32:00.445999 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n2cs8\" (UniqueName: \"kubernetes.io/projected/e5efd747-3795-4922-9f81-3ef02dcb2f91-kube-api-access-n2cs8\") pod \"auto-csr-approver-29556932-fxhnm\" (UID: \"e5efd747-3795-4922-9f81-3ef02dcb2f91\") " pod="openshift-infra/auto-csr-approver-29556932-fxhnm" Mar 13 15:32:00 crc kubenswrapper[4907]: I0313 15:32:00.474024 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556932-fxhnm" Mar 13 15:32:00 crc kubenswrapper[4907]: I0313 15:32:00.897674 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556932-fxhnm"] Mar 13 15:32:00 crc kubenswrapper[4907]: I0313 15:32:00.905683 4907 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 13 15:32:01 crc kubenswrapper[4907]: I0313 15:32:01.090957 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556932-fxhnm" event={"ID":"e5efd747-3795-4922-9f81-3ef02dcb2f91","Type":"ContainerStarted","Data":"8ff7e377b3f47cd070a55919264b03011577973ad91350e5e107b8bfd0e083f2"} Mar 13 15:32:02 crc kubenswrapper[4907]: I0313 15:32:02.101733 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556932-fxhnm" event={"ID":"e5efd747-3795-4922-9f81-3ef02dcb2f91","Type":"ContainerStarted","Data":"44715012ac71de6a61b777f99c8ba8523d3ece17a5f46795a3605b96d66aee6e"} Mar 13 15:32:02 crc kubenswrapper[4907]: I0313 15:32:02.120167 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29556932-fxhnm" podStartSLOduration=1.206299333 podStartE2EDuration="2.120142975s" podCreationTimestamp="2026-03-13 15:32:00 +0000 UTC" firstStartedPulling="2026-03-13 15:32:00.905443528 +0000 UTC m=+5219.805231217" lastFinishedPulling="2026-03-13 15:32:01.81928714 +0000 UTC m=+5220.719074859" observedRunningTime="2026-03-13 15:32:02.11190537 +0000 UTC m=+5221.011693089" watchObservedRunningTime="2026-03-13 15:32:02.120142975 +0000 UTC m=+5221.019930664" Mar 13 15:32:03 crc kubenswrapper[4907]: I0313 15:32:03.111078 4907 generic.go:334] "Generic (PLEG): container finished" podID="e5efd747-3795-4922-9f81-3ef02dcb2f91" containerID="44715012ac71de6a61b777f99c8ba8523d3ece17a5f46795a3605b96d66aee6e" exitCode=0 Mar 13 15:32:03 crc kubenswrapper[4907]: I0313 15:32:03.111236 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556932-fxhnm" event={"ID":"e5efd747-3795-4922-9f81-3ef02dcb2f91","Type":"ContainerDied","Data":"44715012ac71de6a61b777f99c8ba8523d3ece17a5f46795a3605b96d66aee6e"} Mar 13 15:32:04 crc kubenswrapper[4907]: I0313 15:32:04.405749 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556932-fxhnm" Mar 13 15:32:04 crc kubenswrapper[4907]: I0313 15:32:04.491700 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n2cs8\" (UniqueName: \"kubernetes.io/projected/e5efd747-3795-4922-9f81-3ef02dcb2f91-kube-api-access-n2cs8\") pod \"e5efd747-3795-4922-9f81-3ef02dcb2f91\" (UID: \"e5efd747-3795-4922-9f81-3ef02dcb2f91\") " Mar 13 15:32:04 crc kubenswrapper[4907]: I0313 15:32:04.496971 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e5efd747-3795-4922-9f81-3ef02dcb2f91-kube-api-access-n2cs8" (OuterVolumeSpecName: "kube-api-access-n2cs8") pod "e5efd747-3795-4922-9f81-3ef02dcb2f91" (UID: "e5efd747-3795-4922-9f81-3ef02dcb2f91"). InnerVolumeSpecName "kube-api-access-n2cs8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:32:04 crc kubenswrapper[4907]: I0313 15:32:04.593427 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n2cs8\" (UniqueName: \"kubernetes.io/projected/e5efd747-3795-4922-9f81-3ef02dcb2f91-kube-api-access-n2cs8\") on node \"crc\" DevicePath \"\"" Mar 13 15:32:04 crc kubenswrapper[4907]: I0313 15:32:04.886954 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556926-r9tzc"] Mar 13 15:32:04 crc kubenswrapper[4907]: I0313 15:32:04.896433 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556926-r9tzc"] Mar 13 15:32:05 crc kubenswrapper[4907]: I0313 15:32:05.131074 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556932-fxhnm" event={"ID":"e5efd747-3795-4922-9f81-3ef02dcb2f91","Type":"ContainerDied","Data":"8ff7e377b3f47cd070a55919264b03011577973ad91350e5e107b8bfd0e083f2"} Mar 13 15:32:05 crc kubenswrapper[4907]: I0313 15:32:05.131148 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556932-fxhnm" Mar 13 15:32:05 crc kubenswrapper[4907]: I0313 15:32:05.131157 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8ff7e377b3f47cd070a55919264b03011577973ad91350e5e107b8bfd0e083f2" Mar 13 15:32:05 crc kubenswrapper[4907]: I0313 15:32:05.791565 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="81a286dc-1af0-43f3-ad06-0e7d42141a0a" path="/var/lib/kubelet/pods/81a286dc-1af0-43f3-ad06-0e7d42141a0a/volumes" Mar 13 15:32:06 crc kubenswrapper[4907]: I0313 15:32:06.783279 4907 scope.go:117] "RemoveContainer" containerID="592bf9f175f440e943369aab2abbb13ad2c26412cb0e0738c4a33a156b8a5596" Mar 13 15:32:06 crc kubenswrapper[4907]: E0313 15:32:06.783570 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:32:21 crc kubenswrapper[4907]: I0313 15:32:21.787439 4907 scope.go:117] "RemoveContainer" containerID="592bf9f175f440e943369aab2abbb13ad2c26412cb0e0738c4a33a156b8a5596" Mar 13 15:32:21 crc kubenswrapper[4907]: E0313 15:32:21.789211 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:32:27 crc kubenswrapper[4907]: I0313 15:32:27.744104 4907 scope.go:117] "RemoveContainer" containerID="0b4af08a62b26495d3442675a1ce03aa549cddc0c7a28e8f6cb98a7deb792d53" Mar 13 15:32:27 crc kubenswrapper[4907]: I0313 15:32:27.765128 4907 scope.go:117] "RemoveContainer" containerID="fdb72e54eff786808e8d908105afbd91c2b9b22c84cb13b6b15ad7f081283322" Mar 13 15:32:31 crc kubenswrapper[4907]: I0313 15:32:31.301579 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-5nvvj"] Mar 13 15:32:31 crc kubenswrapper[4907]: E0313 15:32:31.302306 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5efd747-3795-4922-9f81-3ef02dcb2f91" containerName="oc" Mar 13 15:32:31 crc kubenswrapper[4907]: I0313 15:32:31.302320 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5efd747-3795-4922-9f81-3ef02dcb2f91" containerName="oc" Mar 13 15:32:31 crc kubenswrapper[4907]: I0313 15:32:31.302472 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="e5efd747-3795-4922-9f81-3ef02dcb2f91" containerName="oc" Mar 13 15:32:31 crc kubenswrapper[4907]: I0313 15:32:31.303634 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5nvvj" Mar 13 15:32:31 crc kubenswrapper[4907]: I0313 15:32:31.316550 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5nvvj"] Mar 13 15:32:31 crc kubenswrapper[4907]: I0313 15:32:31.441455 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/94531b92-db3d-4e91-acbc-ce321d2363a6-utilities\") pod \"redhat-marketplace-5nvvj\" (UID: \"94531b92-db3d-4e91-acbc-ce321d2363a6\") " pod="openshift-marketplace/redhat-marketplace-5nvvj" Mar 13 15:32:31 crc kubenswrapper[4907]: I0313 15:32:31.441599 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vjw7p\" (UniqueName: \"kubernetes.io/projected/94531b92-db3d-4e91-acbc-ce321d2363a6-kube-api-access-vjw7p\") pod \"redhat-marketplace-5nvvj\" (UID: \"94531b92-db3d-4e91-acbc-ce321d2363a6\") " pod="openshift-marketplace/redhat-marketplace-5nvvj" Mar 13 15:32:31 crc kubenswrapper[4907]: I0313 15:32:31.441644 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/94531b92-db3d-4e91-acbc-ce321d2363a6-catalog-content\") pod \"redhat-marketplace-5nvvj\" (UID: \"94531b92-db3d-4e91-acbc-ce321d2363a6\") " pod="openshift-marketplace/redhat-marketplace-5nvvj" Mar 13 15:32:31 crc kubenswrapper[4907]: I0313 15:32:31.543625 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vjw7p\" (UniqueName: \"kubernetes.io/projected/94531b92-db3d-4e91-acbc-ce321d2363a6-kube-api-access-vjw7p\") pod \"redhat-marketplace-5nvvj\" (UID: \"94531b92-db3d-4e91-acbc-ce321d2363a6\") " pod="openshift-marketplace/redhat-marketplace-5nvvj" Mar 13 15:32:31 crc kubenswrapper[4907]: I0313 15:32:31.543690 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/94531b92-db3d-4e91-acbc-ce321d2363a6-catalog-content\") pod \"redhat-marketplace-5nvvj\" (UID: \"94531b92-db3d-4e91-acbc-ce321d2363a6\") " pod="openshift-marketplace/redhat-marketplace-5nvvj" Mar 13 15:32:31 crc kubenswrapper[4907]: I0313 15:32:31.543747 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/94531b92-db3d-4e91-acbc-ce321d2363a6-utilities\") pod \"redhat-marketplace-5nvvj\" (UID: \"94531b92-db3d-4e91-acbc-ce321d2363a6\") " pod="openshift-marketplace/redhat-marketplace-5nvvj" Mar 13 15:32:31 crc kubenswrapper[4907]: I0313 15:32:31.544406 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/94531b92-db3d-4e91-acbc-ce321d2363a6-catalog-content\") pod \"redhat-marketplace-5nvvj\" (UID: \"94531b92-db3d-4e91-acbc-ce321d2363a6\") " pod="openshift-marketplace/redhat-marketplace-5nvvj" Mar 13 15:32:31 crc kubenswrapper[4907]: I0313 15:32:31.544461 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/94531b92-db3d-4e91-acbc-ce321d2363a6-utilities\") pod \"redhat-marketplace-5nvvj\" (UID: \"94531b92-db3d-4e91-acbc-ce321d2363a6\") " pod="openshift-marketplace/redhat-marketplace-5nvvj" Mar 13 15:32:31 crc kubenswrapper[4907]: I0313 15:32:31.565693 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vjw7p\" (UniqueName: \"kubernetes.io/projected/94531b92-db3d-4e91-acbc-ce321d2363a6-kube-api-access-vjw7p\") pod \"redhat-marketplace-5nvvj\" (UID: \"94531b92-db3d-4e91-acbc-ce321d2363a6\") " pod="openshift-marketplace/redhat-marketplace-5nvvj" Mar 13 15:32:31 crc kubenswrapper[4907]: I0313 15:32:31.621444 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5nvvj" Mar 13 15:32:31 crc kubenswrapper[4907]: I0313 15:32:31.882285 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5nvvj"] Mar 13 15:32:32 crc kubenswrapper[4907]: I0313 15:32:32.318153 4907 generic.go:334] "Generic (PLEG): container finished" podID="94531b92-db3d-4e91-acbc-ce321d2363a6" containerID="1287840fe6e969d4fb6a3ec192005abc0045588fa52933a0bfda715eb3a10a3b" exitCode=0 Mar 13 15:32:32 crc kubenswrapper[4907]: I0313 15:32:32.318237 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5nvvj" event={"ID":"94531b92-db3d-4e91-acbc-ce321d2363a6","Type":"ContainerDied","Data":"1287840fe6e969d4fb6a3ec192005abc0045588fa52933a0bfda715eb3a10a3b"} Mar 13 15:32:32 crc kubenswrapper[4907]: I0313 15:32:32.318506 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5nvvj" event={"ID":"94531b92-db3d-4e91-acbc-ce321d2363a6","Type":"ContainerStarted","Data":"b9169ee22d40d10babc6f49b1bfe9a71b6b37f17774a82002ce20e9b46ed79d4"} Mar 13 15:32:33 crc kubenswrapper[4907]: I0313 15:32:33.328297 4907 generic.go:334] "Generic (PLEG): container finished" podID="94531b92-db3d-4e91-acbc-ce321d2363a6" containerID="1f597e88c310a274ade7c6495c51734c8934b5e89f3a0a2ab10dee76116abb50" exitCode=0 Mar 13 15:32:33 crc kubenswrapper[4907]: I0313 15:32:33.328396 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5nvvj" event={"ID":"94531b92-db3d-4e91-acbc-ce321d2363a6","Type":"ContainerDied","Data":"1f597e88c310a274ade7c6495c51734c8934b5e89f3a0a2ab10dee76116abb50"} Mar 13 15:32:34 crc kubenswrapper[4907]: I0313 15:32:34.341582 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5nvvj" event={"ID":"94531b92-db3d-4e91-acbc-ce321d2363a6","Type":"ContainerStarted","Data":"28a2eb5a8c2e8b36c1aa29be80cf9445e6ac6f4b49f4d62964f9c7fe98f4f182"} Mar 13 15:32:34 crc kubenswrapper[4907]: I0313 15:32:34.360699 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-5nvvj" podStartSLOduration=1.937383814 podStartE2EDuration="3.360677472s" podCreationTimestamp="2026-03-13 15:32:31 +0000 UTC" firstStartedPulling="2026-03-13 15:32:32.320031367 +0000 UTC m=+5251.219819046" lastFinishedPulling="2026-03-13 15:32:33.743325015 +0000 UTC m=+5252.643112704" observedRunningTime="2026-03-13 15:32:34.357287449 +0000 UTC m=+5253.257075148" watchObservedRunningTime="2026-03-13 15:32:34.360677472 +0000 UTC m=+5253.260465171" Mar 13 15:32:35 crc kubenswrapper[4907]: I0313 15:32:35.782476 4907 scope.go:117] "RemoveContainer" containerID="592bf9f175f440e943369aab2abbb13ad2c26412cb0e0738c4a33a156b8a5596" Mar 13 15:32:35 crc kubenswrapper[4907]: E0313 15:32:35.783106 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:32:41 crc kubenswrapper[4907]: I0313 15:32:41.621750 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-5nvvj" Mar 13 15:32:41 crc kubenswrapper[4907]: I0313 15:32:41.622343 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-5nvvj" Mar 13 15:32:41 crc kubenswrapper[4907]: I0313 15:32:41.668490 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-5nvvj" Mar 13 15:32:42 crc kubenswrapper[4907]: I0313 15:32:42.441739 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-5nvvj" Mar 13 15:32:42 crc kubenswrapper[4907]: I0313 15:32:42.493135 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5nvvj"] Mar 13 15:32:44 crc kubenswrapper[4907]: I0313 15:32:44.415796 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-5nvvj" podUID="94531b92-db3d-4e91-acbc-ce321d2363a6" containerName="registry-server" containerID="cri-o://28a2eb5a8c2e8b36c1aa29be80cf9445e6ac6f4b49f4d62964f9c7fe98f4f182" gracePeriod=2 Mar 13 15:32:44 crc kubenswrapper[4907]: I0313 15:32:44.850518 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5nvvj" Mar 13 15:32:44 crc kubenswrapper[4907]: I0313 15:32:44.939780 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/94531b92-db3d-4e91-acbc-ce321d2363a6-catalog-content\") pod \"94531b92-db3d-4e91-acbc-ce321d2363a6\" (UID: \"94531b92-db3d-4e91-acbc-ce321d2363a6\") " Mar 13 15:32:44 crc kubenswrapper[4907]: I0313 15:32:44.939841 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/94531b92-db3d-4e91-acbc-ce321d2363a6-utilities\") pod \"94531b92-db3d-4e91-acbc-ce321d2363a6\" (UID: \"94531b92-db3d-4e91-acbc-ce321d2363a6\") " Mar 13 15:32:44 crc kubenswrapper[4907]: I0313 15:32:44.939974 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vjw7p\" (UniqueName: \"kubernetes.io/projected/94531b92-db3d-4e91-acbc-ce321d2363a6-kube-api-access-vjw7p\") pod \"94531b92-db3d-4e91-acbc-ce321d2363a6\" (UID: \"94531b92-db3d-4e91-acbc-ce321d2363a6\") " Mar 13 15:32:44 crc kubenswrapper[4907]: I0313 15:32:44.940783 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/94531b92-db3d-4e91-acbc-ce321d2363a6-utilities" (OuterVolumeSpecName: "utilities") pod "94531b92-db3d-4e91-acbc-ce321d2363a6" (UID: "94531b92-db3d-4e91-acbc-ce321d2363a6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:32:44 crc kubenswrapper[4907]: I0313 15:32:44.946164 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/94531b92-db3d-4e91-acbc-ce321d2363a6-kube-api-access-vjw7p" (OuterVolumeSpecName: "kube-api-access-vjw7p") pod "94531b92-db3d-4e91-acbc-ce321d2363a6" (UID: "94531b92-db3d-4e91-acbc-ce321d2363a6"). InnerVolumeSpecName "kube-api-access-vjw7p". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:32:44 crc kubenswrapper[4907]: I0313 15:32:44.978644 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/94531b92-db3d-4e91-acbc-ce321d2363a6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "94531b92-db3d-4e91-acbc-ce321d2363a6" (UID: "94531b92-db3d-4e91-acbc-ce321d2363a6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:32:45 crc kubenswrapper[4907]: I0313 15:32:45.041630 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/94531b92-db3d-4e91-acbc-ce321d2363a6-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 15:32:45 crc kubenswrapper[4907]: I0313 15:32:45.041669 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/94531b92-db3d-4e91-acbc-ce321d2363a6-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 15:32:45 crc kubenswrapper[4907]: I0313 15:32:45.041680 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vjw7p\" (UniqueName: \"kubernetes.io/projected/94531b92-db3d-4e91-acbc-ce321d2363a6-kube-api-access-vjw7p\") on node \"crc\" DevicePath \"\"" Mar 13 15:32:45 crc kubenswrapper[4907]: I0313 15:32:45.423707 4907 generic.go:334] "Generic (PLEG): container finished" podID="94531b92-db3d-4e91-acbc-ce321d2363a6" containerID="28a2eb5a8c2e8b36c1aa29be80cf9445e6ac6f4b49f4d62964f9c7fe98f4f182" exitCode=0 Mar 13 15:32:45 crc kubenswrapper[4907]: I0313 15:32:45.423766 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5nvvj" event={"ID":"94531b92-db3d-4e91-acbc-ce321d2363a6","Type":"ContainerDied","Data":"28a2eb5a8c2e8b36c1aa29be80cf9445e6ac6f4b49f4d62964f9c7fe98f4f182"} Mar 13 15:32:45 crc kubenswrapper[4907]: I0313 15:32:45.423800 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5nvvj" event={"ID":"94531b92-db3d-4e91-acbc-ce321d2363a6","Type":"ContainerDied","Data":"b9169ee22d40d10babc6f49b1bfe9a71b6b37f17774a82002ce20e9b46ed79d4"} Mar 13 15:32:45 crc kubenswrapper[4907]: I0313 15:32:45.423822 4907 scope.go:117] "RemoveContainer" containerID="28a2eb5a8c2e8b36c1aa29be80cf9445e6ac6f4b49f4d62964f9c7fe98f4f182" Mar 13 15:32:45 crc kubenswrapper[4907]: I0313 15:32:45.423977 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5nvvj" Mar 13 15:32:45 crc kubenswrapper[4907]: I0313 15:32:45.444188 4907 scope.go:117] "RemoveContainer" containerID="1f597e88c310a274ade7c6495c51734c8934b5e89f3a0a2ab10dee76116abb50" Mar 13 15:32:45 crc kubenswrapper[4907]: I0313 15:32:45.462841 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5nvvj"] Mar 13 15:32:45 crc kubenswrapper[4907]: I0313 15:32:45.464731 4907 scope.go:117] "RemoveContainer" containerID="1287840fe6e969d4fb6a3ec192005abc0045588fa52933a0bfda715eb3a10a3b" Mar 13 15:32:45 crc kubenswrapper[4907]: I0313 15:32:45.473619 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-5nvvj"] Mar 13 15:32:45 crc kubenswrapper[4907]: I0313 15:32:45.519425 4907 scope.go:117] "RemoveContainer" containerID="28a2eb5a8c2e8b36c1aa29be80cf9445e6ac6f4b49f4d62964f9c7fe98f4f182" Mar 13 15:32:45 crc kubenswrapper[4907]: E0313 15:32:45.520232 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"28a2eb5a8c2e8b36c1aa29be80cf9445e6ac6f4b49f4d62964f9c7fe98f4f182\": container with ID starting with 28a2eb5a8c2e8b36c1aa29be80cf9445e6ac6f4b49f4d62964f9c7fe98f4f182 not found: ID does not exist" containerID="28a2eb5a8c2e8b36c1aa29be80cf9445e6ac6f4b49f4d62964f9c7fe98f4f182" Mar 13 15:32:45 crc kubenswrapper[4907]: I0313 15:32:45.520270 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"28a2eb5a8c2e8b36c1aa29be80cf9445e6ac6f4b49f4d62964f9c7fe98f4f182"} err="failed to get container status \"28a2eb5a8c2e8b36c1aa29be80cf9445e6ac6f4b49f4d62964f9c7fe98f4f182\": rpc error: code = NotFound desc = could not find container \"28a2eb5a8c2e8b36c1aa29be80cf9445e6ac6f4b49f4d62964f9c7fe98f4f182\": container with ID starting with 28a2eb5a8c2e8b36c1aa29be80cf9445e6ac6f4b49f4d62964f9c7fe98f4f182 not found: ID does not exist" Mar 13 15:32:45 crc kubenswrapper[4907]: I0313 15:32:45.520295 4907 scope.go:117] "RemoveContainer" containerID="1f597e88c310a274ade7c6495c51734c8934b5e89f3a0a2ab10dee76116abb50" Mar 13 15:32:45 crc kubenswrapper[4907]: E0313 15:32:45.520630 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1f597e88c310a274ade7c6495c51734c8934b5e89f3a0a2ab10dee76116abb50\": container with ID starting with 1f597e88c310a274ade7c6495c51734c8934b5e89f3a0a2ab10dee76116abb50 not found: ID does not exist" containerID="1f597e88c310a274ade7c6495c51734c8934b5e89f3a0a2ab10dee76116abb50" Mar 13 15:32:45 crc kubenswrapper[4907]: I0313 15:32:45.520660 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1f597e88c310a274ade7c6495c51734c8934b5e89f3a0a2ab10dee76116abb50"} err="failed to get container status \"1f597e88c310a274ade7c6495c51734c8934b5e89f3a0a2ab10dee76116abb50\": rpc error: code = NotFound desc = could not find container \"1f597e88c310a274ade7c6495c51734c8934b5e89f3a0a2ab10dee76116abb50\": container with ID starting with 1f597e88c310a274ade7c6495c51734c8934b5e89f3a0a2ab10dee76116abb50 not found: ID does not exist" Mar 13 15:32:45 crc kubenswrapper[4907]: I0313 15:32:45.520703 4907 scope.go:117] "RemoveContainer" containerID="1287840fe6e969d4fb6a3ec192005abc0045588fa52933a0bfda715eb3a10a3b" Mar 13 15:32:45 crc kubenswrapper[4907]: E0313 15:32:45.521106 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1287840fe6e969d4fb6a3ec192005abc0045588fa52933a0bfda715eb3a10a3b\": container with ID starting with 1287840fe6e969d4fb6a3ec192005abc0045588fa52933a0bfda715eb3a10a3b not found: ID does not exist" containerID="1287840fe6e969d4fb6a3ec192005abc0045588fa52933a0bfda715eb3a10a3b" Mar 13 15:32:45 crc kubenswrapper[4907]: I0313 15:32:45.521159 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1287840fe6e969d4fb6a3ec192005abc0045588fa52933a0bfda715eb3a10a3b"} err="failed to get container status \"1287840fe6e969d4fb6a3ec192005abc0045588fa52933a0bfda715eb3a10a3b\": rpc error: code = NotFound desc = could not find container \"1287840fe6e969d4fb6a3ec192005abc0045588fa52933a0bfda715eb3a10a3b\": container with ID starting with 1287840fe6e969d4fb6a3ec192005abc0045588fa52933a0bfda715eb3a10a3b not found: ID does not exist" Mar 13 15:32:45 crc kubenswrapper[4907]: I0313 15:32:45.792743 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="94531b92-db3d-4e91-acbc-ce321d2363a6" path="/var/lib/kubelet/pods/94531b92-db3d-4e91-acbc-ce321d2363a6/volumes" Mar 13 15:32:49 crc kubenswrapper[4907]: I0313 15:32:49.782756 4907 scope.go:117] "RemoveContainer" containerID="592bf9f175f440e943369aab2abbb13ad2c26412cb0e0738c4a33a156b8a5596" Mar 13 15:32:49 crc kubenswrapper[4907]: E0313 15:32:49.783569 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:32:53 crc kubenswrapper[4907]: I0313 15:32:53.352086 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-copy-data"] Mar 13 15:32:53 crc kubenswrapper[4907]: E0313 15:32:53.352704 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94531b92-db3d-4e91-acbc-ce321d2363a6" containerName="registry-server" Mar 13 15:32:53 crc kubenswrapper[4907]: I0313 15:32:53.352720 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="94531b92-db3d-4e91-acbc-ce321d2363a6" containerName="registry-server" Mar 13 15:32:53 crc kubenswrapper[4907]: E0313 15:32:53.352759 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94531b92-db3d-4e91-acbc-ce321d2363a6" containerName="extract-content" Mar 13 15:32:53 crc kubenswrapper[4907]: I0313 15:32:53.352770 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="94531b92-db3d-4e91-acbc-ce321d2363a6" containerName="extract-content" Mar 13 15:32:53 crc kubenswrapper[4907]: E0313 15:32:53.352794 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94531b92-db3d-4e91-acbc-ce321d2363a6" containerName="extract-utilities" Mar 13 15:32:53 crc kubenswrapper[4907]: I0313 15:32:53.352804 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="94531b92-db3d-4e91-acbc-ce321d2363a6" containerName="extract-utilities" Mar 13 15:32:53 crc kubenswrapper[4907]: I0313 15:32:53.353068 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="94531b92-db3d-4e91-acbc-ce321d2363a6" containerName="registry-server" Mar 13 15:32:53 crc kubenswrapper[4907]: I0313 15:32:53.353735 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-copy-data" Mar 13 15:32:53 crc kubenswrapper[4907]: I0313 15:32:53.361661 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-mqs4s" Mar 13 15:32:53 crc kubenswrapper[4907]: I0313 15:32:53.374181 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-copy-data"] Mar 13 15:32:53 crc kubenswrapper[4907]: I0313 15:32:53.471337 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-3b4b3b59-8b2d-44c3-bc88-a79545d23d7c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3b4b3b59-8b2d-44c3-bc88-a79545d23d7c\") pod \"mariadb-copy-data\" (UID: \"6b679563-ed42-418a-80ef-0b24e225d5ef\") " pod="openstack/mariadb-copy-data" Mar 13 15:32:53 crc kubenswrapper[4907]: I0313 15:32:53.471393 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rzv64\" (UniqueName: \"kubernetes.io/projected/6b679563-ed42-418a-80ef-0b24e225d5ef-kube-api-access-rzv64\") pod \"mariadb-copy-data\" (UID: \"6b679563-ed42-418a-80ef-0b24e225d5ef\") " pod="openstack/mariadb-copy-data" Mar 13 15:32:53 crc kubenswrapper[4907]: I0313 15:32:53.572767 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-3b4b3b59-8b2d-44c3-bc88-a79545d23d7c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3b4b3b59-8b2d-44c3-bc88-a79545d23d7c\") pod \"mariadb-copy-data\" (UID: \"6b679563-ed42-418a-80ef-0b24e225d5ef\") " pod="openstack/mariadb-copy-data" Mar 13 15:32:53 crc kubenswrapper[4907]: I0313 15:32:53.572808 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rzv64\" (UniqueName: \"kubernetes.io/projected/6b679563-ed42-418a-80ef-0b24e225d5ef-kube-api-access-rzv64\") pod \"mariadb-copy-data\" (UID: \"6b679563-ed42-418a-80ef-0b24e225d5ef\") " pod="openstack/mariadb-copy-data" Mar 13 15:32:53 crc kubenswrapper[4907]: I0313 15:32:53.576253 4907 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 13 15:32:53 crc kubenswrapper[4907]: I0313 15:32:53.576319 4907 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-3b4b3b59-8b2d-44c3-bc88-a79545d23d7c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3b4b3b59-8b2d-44c3-bc88-a79545d23d7c\") pod \"mariadb-copy-data\" (UID: \"6b679563-ed42-418a-80ef-0b24e225d5ef\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/f298a5b7b791478733605b03fd2767b8f1abec427d368f68c943bb89527cfeeb/globalmount\"" pod="openstack/mariadb-copy-data" Mar 13 15:32:53 crc kubenswrapper[4907]: I0313 15:32:53.595783 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rzv64\" (UniqueName: \"kubernetes.io/projected/6b679563-ed42-418a-80ef-0b24e225d5ef-kube-api-access-rzv64\") pod \"mariadb-copy-data\" (UID: \"6b679563-ed42-418a-80ef-0b24e225d5ef\") " pod="openstack/mariadb-copy-data" Mar 13 15:32:53 crc kubenswrapper[4907]: I0313 15:32:53.606479 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-3b4b3b59-8b2d-44c3-bc88-a79545d23d7c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3b4b3b59-8b2d-44c3-bc88-a79545d23d7c\") pod \"mariadb-copy-data\" (UID: \"6b679563-ed42-418a-80ef-0b24e225d5ef\") " pod="openstack/mariadb-copy-data" Mar 13 15:32:53 crc kubenswrapper[4907]: I0313 15:32:53.682296 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-copy-data" Mar 13 15:32:54 crc kubenswrapper[4907]: I0313 15:32:54.231001 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-copy-data"] Mar 13 15:32:54 crc kubenswrapper[4907]: I0313 15:32:54.485008 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-copy-data" event={"ID":"6b679563-ed42-418a-80ef-0b24e225d5ef","Type":"ContainerStarted","Data":"be1062c79f4023088d8eb4fb41e0f370f57e80c1813e6265704fdadc605e804a"} Mar 13 15:32:54 crc kubenswrapper[4907]: I0313 15:32:54.485057 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-copy-data" event={"ID":"6b679563-ed42-418a-80ef-0b24e225d5ef","Type":"ContainerStarted","Data":"763a1aa9a599cf6e70f4f78d5d2da41195b4bd8f467a89e3d789f85b36605af1"} Mar 13 15:32:54 crc kubenswrapper[4907]: I0313 15:32:54.504629 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/mariadb-copy-data" podStartSLOduration=2.50460878 podStartE2EDuration="2.50460878s" podCreationTimestamp="2026-03-13 15:32:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:32:54.500233161 +0000 UTC m=+5273.400020870" watchObservedRunningTime="2026-03-13 15:32:54.50460878 +0000 UTC m=+5273.404396469" Mar 13 15:32:57 crc kubenswrapper[4907]: I0313 15:32:57.359899 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client"] Mar 13 15:32:57 crc kubenswrapper[4907]: I0313 15:32:57.361670 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Mar 13 15:32:57 crc kubenswrapper[4907]: I0313 15:32:57.367874 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Mar 13 15:32:57 crc kubenswrapper[4907]: I0313 15:32:57.530914 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6tkbl\" (UniqueName: \"kubernetes.io/projected/4ba8510f-f827-4fc5-a32b-f64153e33f5e-kube-api-access-6tkbl\") pod \"mariadb-client\" (UID: \"4ba8510f-f827-4fc5-a32b-f64153e33f5e\") " pod="openstack/mariadb-client" Mar 13 15:32:57 crc kubenswrapper[4907]: I0313 15:32:57.632467 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6tkbl\" (UniqueName: \"kubernetes.io/projected/4ba8510f-f827-4fc5-a32b-f64153e33f5e-kube-api-access-6tkbl\") pod \"mariadb-client\" (UID: \"4ba8510f-f827-4fc5-a32b-f64153e33f5e\") " pod="openstack/mariadb-client" Mar 13 15:32:57 crc kubenswrapper[4907]: I0313 15:32:57.660719 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6tkbl\" (UniqueName: \"kubernetes.io/projected/4ba8510f-f827-4fc5-a32b-f64153e33f5e-kube-api-access-6tkbl\") pod \"mariadb-client\" (UID: \"4ba8510f-f827-4fc5-a32b-f64153e33f5e\") " pod="openstack/mariadb-client" Mar 13 15:32:57 crc kubenswrapper[4907]: I0313 15:32:57.714214 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Mar 13 15:32:58 crc kubenswrapper[4907]: I0313 15:32:58.123807 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Mar 13 15:32:58 crc kubenswrapper[4907]: W0313 15:32:58.126659 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4ba8510f_f827_4fc5_a32b_f64153e33f5e.slice/crio-b642f87690b3f00088c572d86d8921572a5e11af8fab4078310397ab7d0aaeab WatchSource:0}: Error finding container b642f87690b3f00088c572d86d8921572a5e11af8fab4078310397ab7d0aaeab: Status 404 returned error can't find the container with id b642f87690b3f00088c572d86d8921572a5e11af8fab4078310397ab7d0aaeab Mar 13 15:32:58 crc kubenswrapper[4907]: I0313 15:32:58.525931 4907 generic.go:334] "Generic (PLEG): container finished" podID="4ba8510f-f827-4fc5-a32b-f64153e33f5e" containerID="3c35c6992ec9b87a4cfcf6f80130463db4ea23212e37b3ceef2bd0131f1b10b1" exitCode=0 Mar 13 15:32:58 crc kubenswrapper[4907]: I0313 15:32:58.526018 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"4ba8510f-f827-4fc5-a32b-f64153e33f5e","Type":"ContainerDied","Data":"3c35c6992ec9b87a4cfcf6f80130463db4ea23212e37b3ceef2bd0131f1b10b1"} Mar 13 15:32:58 crc kubenswrapper[4907]: I0313 15:32:58.526255 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"4ba8510f-f827-4fc5-a32b-f64153e33f5e","Type":"ContainerStarted","Data":"b642f87690b3f00088c572d86d8921572a5e11af8fab4078310397ab7d0aaeab"} Mar 13 15:32:59 crc kubenswrapper[4907]: I0313 15:32:59.872041 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Mar 13 15:32:59 crc kubenswrapper[4907]: I0313 15:32:59.896843 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client_4ba8510f-f827-4fc5-a32b-f64153e33f5e/mariadb-client/0.log" Mar 13 15:32:59 crc kubenswrapper[4907]: I0313 15:32:59.930390 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client"] Mar 13 15:32:59 crc kubenswrapper[4907]: I0313 15:32:59.938041 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client"] Mar 13 15:32:59 crc kubenswrapper[4907]: I0313 15:32:59.963637 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6tkbl\" (UniqueName: \"kubernetes.io/projected/4ba8510f-f827-4fc5-a32b-f64153e33f5e-kube-api-access-6tkbl\") pod \"4ba8510f-f827-4fc5-a32b-f64153e33f5e\" (UID: \"4ba8510f-f827-4fc5-a32b-f64153e33f5e\") " Mar 13 15:32:59 crc kubenswrapper[4907]: I0313 15:32:59.971806 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ba8510f-f827-4fc5-a32b-f64153e33f5e-kube-api-access-6tkbl" (OuterVolumeSpecName: "kube-api-access-6tkbl") pod "4ba8510f-f827-4fc5-a32b-f64153e33f5e" (UID: "4ba8510f-f827-4fc5-a32b-f64153e33f5e"). InnerVolumeSpecName "kube-api-access-6tkbl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:33:00 crc kubenswrapper[4907]: I0313 15:33:00.063801 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client"] Mar 13 15:33:00 crc kubenswrapper[4907]: E0313 15:33:00.064164 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ba8510f-f827-4fc5-a32b-f64153e33f5e" containerName="mariadb-client" Mar 13 15:33:00 crc kubenswrapper[4907]: I0313 15:33:00.064185 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ba8510f-f827-4fc5-a32b-f64153e33f5e" containerName="mariadb-client" Mar 13 15:33:00 crc kubenswrapper[4907]: I0313 15:33:00.064405 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ba8510f-f827-4fc5-a32b-f64153e33f5e" containerName="mariadb-client" Mar 13 15:33:00 crc kubenswrapper[4907]: I0313 15:33:00.064937 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Mar 13 15:33:00 crc kubenswrapper[4907]: I0313 15:33:00.066296 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6tkbl\" (UniqueName: \"kubernetes.io/projected/4ba8510f-f827-4fc5-a32b-f64153e33f5e-kube-api-access-6tkbl\") on node \"crc\" DevicePath \"\"" Mar 13 15:33:00 crc kubenswrapper[4907]: I0313 15:33:00.079401 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Mar 13 15:33:00 crc kubenswrapper[4907]: I0313 15:33:00.168048 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8f58c\" (UniqueName: \"kubernetes.io/projected/6616b212-937e-4100-9007-c529da51469a-kube-api-access-8f58c\") pod \"mariadb-client\" (UID: \"6616b212-937e-4100-9007-c529da51469a\") " pod="openstack/mariadb-client" Mar 13 15:33:00 crc kubenswrapper[4907]: I0313 15:33:00.269583 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8f58c\" (UniqueName: \"kubernetes.io/projected/6616b212-937e-4100-9007-c529da51469a-kube-api-access-8f58c\") pod \"mariadb-client\" (UID: \"6616b212-937e-4100-9007-c529da51469a\") " pod="openstack/mariadb-client" Mar 13 15:33:00 crc kubenswrapper[4907]: I0313 15:33:00.285445 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8f58c\" (UniqueName: \"kubernetes.io/projected/6616b212-937e-4100-9007-c529da51469a-kube-api-access-8f58c\") pod \"mariadb-client\" (UID: \"6616b212-937e-4100-9007-c529da51469a\") " pod="openstack/mariadb-client" Mar 13 15:33:00 crc kubenswrapper[4907]: I0313 15:33:00.388035 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Mar 13 15:33:00 crc kubenswrapper[4907]: I0313 15:33:00.541796 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b642f87690b3f00088c572d86d8921572a5e11af8fab4078310397ab7d0aaeab" Mar 13 15:33:00 crc kubenswrapper[4907]: I0313 15:33:00.541863 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Mar 13 15:33:00 crc kubenswrapper[4907]: I0313 15:33:00.575237 4907 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/mariadb-client" oldPodUID="4ba8510f-f827-4fc5-a32b-f64153e33f5e" podUID="6616b212-937e-4100-9007-c529da51469a" Mar 13 15:33:00 crc kubenswrapper[4907]: I0313 15:33:00.784995 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Mar 13 15:33:00 crc kubenswrapper[4907]: W0313 15:33:00.787208 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6616b212_937e_4100_9007_c529da51469a.slice/crio-bbea2db874a2a8327fe7e7e46748c830df370fb5bb85720bc7d40582c81c4bb1 WatchSource:0}: Error finding container bbea2db874a2a8327fe7e7e46748c830df370fb5bb85720bc7d40582c81c4bb1: Status 404 returned error can't find the container with id bbea2db874a2a8327fe7e7e46748c830df370fb5bb85720bc7d40582c81c4bb1 Mar 13 15:33:01 crc kubenswrapper[4907]: I0313 15:33:01.549580 4907 generic.go:334] "Generic (PLEG): container finished" podID="6616b212-937e-4100-9007-c529da51469a" containerID="2c2f32fc7b86119b10010b256cedf2bb75451fdf3804d5c1b2c61e62707d4b83" exitCode=0 Mar 13 15:33:01 crc kubenswrapper[4907]: I0313 15:33:01.549622 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"6616b212-937e-4100-9007-c529da51469a","Type":"ContainerDied","Data":"2c2f32fc7b86119b10010b256cedf2bb75451fdf3804d5c1b2c61e62707d4b83"} Mar 13 15:33:01 crc kubenswrapper[4907]: I0313 15:33:01.549655 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"6616b212-937e-4100-9007-c529da51469a","Type":"ContainerStarted","Data":"bbea2db874a2a8327fe7e7e46748c830df370fb5bb85720bc7d40582c81c4bb1"} Mar 13 15:33:01 crc kubenswrapper[4907]: I0313 15:33:01.793149 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4ba8510f-f827-4fc5-a32b-f64153e33f5e" path="/var/lib/kubelet/pods/4ba8510f-f827-4fc5-a32b-f64153e33f5e/volumes" Mar 13 15:33:02 crc kubenswrapper[4907]: I0313 15:33:02.784092 4907 scope.go:117] "RemoveContainer" containerID="592bf9f175f440e943369aab2abbb13ad2c26412cb0e0738c4a33a156b8a5596" Mar 13 15:33:02 crc kubenswrapper[4907]: E0313 15:33:02.785207 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:33:02 crc kubenswrapper[4907]: I0313 15:33:02.864427 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Mar 13 15:33:02 crc kubenswrapper[4907]: I0313 15:33:02.881518 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client_6616b212-937e-4100-9007-c529da51469a/mariadb-client/0.log" Mar 13 15:33:02 crc kubenswrapper[4907]: I0313 15:33:02.906075 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client"] Mar 13 15:33:02 crc kubenswrapper[4907]: I0313 15:33:02.909529 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8f58c\" (UniqueName: \"kubernetes.io/projected/6616b212-937e-4100-9007-c529da51469a-kube-api-access-8f58c\") pod \"6616b212-937e-4100-9007-c529da51469a\" (UID: \"6616b212-937e-4100-9007-c529da51469a\") " Mar 13 15:33:02 crc kubenswrapper[4907]: I0313 15:33:02.914318 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client"] Mar 13 15:33:02 crc kubenswrapper[4907]: I0313 15:33:02.916220 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6616b212-937e-4100-9007-c529da51469a-kube-api-access-8f58c" (OuterVolumeSpecName: "kube-api-access-8f58c") pod "6616b212-937e-4100-9007-c529da51469a" (UID: "6616b212-937e-4100-9007-c529da51469a"). InnerVolumeSpecName "kube-api-access-8f58c". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:33:03 crc kubenswrapper[4907]: I0313 15:33:03.010607 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8f58c\" (UniqueName: \"kubernetes.io/projected/6616b212-937e-4100-9007-c529da51469a-kube-api-access-8f58c\") on node \"crc\" DevicePath \"\"" Mar 13 15:33:03 crc kubenswrapper[4907]: I0313 15:33:03.571233 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bbea2db874a2a8327fe7e7e46748c830df370fb5bb85720bc7d40582c81c4bb1" Mar 13 15:33:03 crc kubenswrapper[4907]: I0313 15:33:03.571539 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Mar 13 15:33:03 crc kubenswrapper[4907]: I0313 15:33:03.792984 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6616b212-937e-4100-9007-c529da51469a" path="/var/lib/kubelet/pods/6616b212-937e-4100-9007-c529da51469a/volumes" Mar 13 15:33:13 crc kubenswrapper[4907]: I0313 15:33:13.782966 4907 scope.go:117] "RemoveContainer" containerID="592bf9f175f440e943369aab2abbb13ad2c26412cb0e0738c4a33a156b8a5596" Mar 13 15:33:13 crc kubenswrapper[4907]: E0313 15:33:13.783720 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:33:26 crc kubenswrapper[4907]: I0313 15:33:26.783291 4907 scope.go:117] "RemoveContainer" containerID="592bf9f175f440e943369aab2abbb13ad2c26412cb0e0738c4a33a156b8a5596" Mar 13 15:33:26 crc kubenswrapper[4907]: E0313 15:33:26.786875 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:33:27 crc kubenswrapper[4907]: I0313 15:33:27.835940 4907 scope.go:117] "RemoveContainer" containerID="83c101cd6c30ca361c7f00d6a637893118fddd69105f31e3b1c72e730b52960a" Mar 13 15:33:27 crc kubenswrapper[4907]: I0313 15:33:27.856287 4907 scope.go:117] "RemoveContainer" containerID="8f8264e04faa67742d13e20b2a8a6a04d3b6912538ef54a5202a7e841e98d29e" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.133323 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Mar 13 15:33:34 crc kubenswrapper[4907]: E0313 15:33:34.134634 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6616b212-937e-4100-9007-c529da51469a" containerName="mariadb-client" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.134653 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="6616b212-937e-4100-9007-c529da51469a" containerName="mariadb-client" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.134955 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="6616b212-937e-4100-9007-c529da51469a" containerName="mariadb-client" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.135767 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.138259 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.138289 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-s57jd" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.138464 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.154131 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-2"] Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.155894 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-2" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.161718 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-1"] Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.167581 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-1" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.200534 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.210598 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-2"] Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.225290 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-1"] Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.270469 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bffaea76-e210-44d1-8caa-ebf82edb8ef0-config\") pod \"ovsdbserver-sb-0\" (UID: \"bffaea76-e210-44d1-8caa-ebf82edb8ef0\") " pod="openstack/ovsdbserver-sb-0" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.270561 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-3bdb22ed-1952-4cca-8e2a-e4c4e2597e8c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3bdb22ed-1952-4cca-8e2a-e4c4e2597e8c\") pod \"ovsdbserver-sb-1\" (UID: \"e5ec94c1-f60b-4156-b3ba-278db827c9b8\") " pod="openstack/ovsdbserver-sb-1" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.270593 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cdlpw\" (UniqueName: \"kubernetes.io/projected/e5ec94c1-f60b-4156-b3ba-278db827c9b8-kube-api-access-cdlpw\") pod \"ovsdbserver-sb-1\" (UID: \"e5ec94c1-f60b-4156-b3ba-278db827c9b8\") " pod="openstack/ovsdbserver-sb-1" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.270619 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xw2xr\" (UniqueName: \"kubernetes.io/projected/b188658f-1e5c-49d4-be19-0a888caddb43-kube-api-access-xw2xr\") pod \"ovsdbserver-sb-2\" (UID: \"b188658f-1e5c-49d4-be19-0a888caddb43\") " pod="openstack/ovsdbserver-sb-2" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.270638 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b188658f-1e5c-49d4-be19-0a888caddb43-combined-ca-bundle\") pod \"ovsdbserver-sb-2\" (UID: \"b188658f-1e5c-49d4-be19-0a888caddb43\") " pod="openstack/ovsdbserver-sb-2" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.270655 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bffaea76-e210-44d1-8caa-ebf82edb8ef0-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"bffaea76-e210-44d1-8caa-ebf82edb8ef0\") " pod="openstack/ovsdbserver-sb-0" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.270673 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b188658f-1e5c-49d4-be19-0a888caddb43-config\") pod \"ovsdbserver-sb-2\" (UID: \"b188658f-1e5c-49d4-be19-0a888caddb43\") " pod="openstack/ovsdbserver-sb-2" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.270697 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/e5ec94c1-f60b-4156-b3ba-278db827c9b8-ovsdb-rundir\") pod \"ovsdbserver-sb-1\" (UID: \"e5ec94c1-f60b-4156-b3ba-278db827c9b8\") " pod="openstack/ovsdbserver-sb-1" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.270721 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5ec94c1-f60b-4156-b3ba-278db827c9b8-combined-ca-bundle\") pod \"ovsdbserver-sb-1\" (UID: \"e5ec94c1-f60b-4156-b3ba-278db827c9b8\") " pod="openstack/ovsdbserver-sb-1" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.270834 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bffaea76-e210-44d1-8caa-ebf82edb8ef0-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"bffaea76-e210-44d1-8caa-ebf82edb8ef0\") " pod="openstack/ovsdbserver-sb-0" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.270995 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e5ec94c1-f60b-4156-b3ba-278db827c9b8-config\") pod \"ovsdbserver-sb-1\" (UID: \"e5ec94c1-f60b-4156-b3ba-278db827c9b8\") " pod="openstack/ovsdbserver-sb-1" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.271182 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e5ec94c1-f60b-4156-b3ba-278db827c9b8-scripts\") pod \"ovsdbserver-sb-1\" (UID: \"e5ec94c1-f60b-4156-b3ba-278db827c9b8\") " pod="openstack/ovsdbserver-sb-1" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.271270 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/b188658f-1e5c-49d4-be19-0a888caddb43-ovsdb-rundir\") pod \"ovsdbserver-sb-2\" (UID: \"b188658f-1e5c-49d4-be19-0a888caddb43\") " pod="openstack/ovsdbserver-sb-2" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.271349 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-8e01f892-be99-49f8-b8aa-bf80a68ae42b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-8e01f892-be99-49f8-b8aa-bf80a68ae42b\") pod \"ovsdbserver-sb-2\" (UID: \"b188658f-1e5c-49d4-be19-0a888caddb43\") " pod="openstack/ovsdbserver-sb-2" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.271408 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/bffaea76-e210-44d1-8caa-ebf82edb8ef0-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"bffaea76-e210-44d1-8caa-ebf82edb8ef0\") " pod="openstack/ovsdbserver-sb-0" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.271471 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-shc8z\" (UniqueName: \"kubernetes.io/projected/bffaea76-e210-44d1-8caa-ebf82edb8ef0-kube-api-access-shc8z\") pod \"ovsdbserver-sb-0\" (UID: \"bffaea76-e210-44d1-8caa-ebf82edb8ef0\") " pod="openstack/ovsdbserver-sb-0" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.271516 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-4247ec19-7024-4129-80de-23a6ba6ce376\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-4247ec19-7024-4129-80de-23a6ba6ce376\") pod \"ovsdbserver-sb-0\" (UID: \"bffaea76-e210-44d1-8caa-ebf82edb8ef0\") " pod="openstack/ovsdbserver-sb-0" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.271635 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b188658f-1e5c-49d4-be19-0a888caddb43-scripts\") pod \"ovsdbserver-sb-2\" (UID: \"b188658f-1e5c-49d4-be19-0a888caddb43\") " pod="openstack/ovsdbserver-sb-2" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.343111 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.345050 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.347709 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-tpm7x" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.347718 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.352660 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.355319 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.372424 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-1"] Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.383120 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-1" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.390177 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-2"] Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.390446 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e5ec94c1-f60b-4156-b3ba-278db827c9b8-scripts\") pod \"ovsdbserver-sb-1\" (UID: \"e5ec94c1-f60b-4156-b3ba-278db827c9b8\") " pod="openstack/ovsdbserver-sb-1" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.390572 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/b188658f-1e5c-49d4-be19-0a888caddb43-ovsdb-rundir\") pod \"ovsdbserver-sb-2\" (UID: \"b188658f-1e5c-49d4-be19-0a888caddb43\") " pod="openstack/ovsdbserver-sb-2" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.390653 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-8e01f892-be99-49f8-b8aa-bf80a68ae42b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-8e01f892-be99-49f8-b8aa-bf80a68ae42b\") pod \"ovsdbserver-sb-2\" (UID: \"b188658f-1e5c-49d4-be19-0a888caddb43\") " pod="openstack/ovsdbserver-sb-2" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.390703 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/bffaea76-e210-44d1-8caa-ebf82edb8ef0-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"bffaea76-e210-44d1-8caa-ebf82edb8ef0\") " pod="openstack/ovsdbserver-sb-0" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.390794 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-shc8z\" (UniqueName: \"kubernetes.io/projected/bffaea76-e210-44d1-8caa-ebf82edb8ef0-kube-api-access-shc8z\") pod \"ovsdbserver-sb-0\" (UID: \"bffaea76-e210-44d1-8caa-ebf82edb8ef0\") " pod="openstack/ovsdbserver-sb-0" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.390828 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-4247ec19-7024-4129-80de-23a6ba6ce376\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-4247ec19-7024-4129-80de-23a6ba6ce376\") pod \"ovsdbserver-sb-0\" (UID: \"bffaea76-e210-44d1-8caa-ebf82edb8ef0\") " pod="openstack/ovsdbserver-sb-0" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.390977 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b188658f-1e5c-49d4-be19-0a888caddb43-scripts\") pod \"ovsdbserver-sb-2\" (UID: \"b188658f-1e5c-49d4-be19-0a888caddb43\") " pod="openstack/ovsdbserver-sb-2" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.391032 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bffaea76-e210-44d1-8caa-ebf82edb8ef0-config\") pod \"ovsdbserver-sb-0\" (UID: \"bffaea76-e210-44d1-8caa-ebf82edb8ef0\") " pod="openstack/ovsdbserver-sb-0" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.391087 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-3bdb22ed-1952-4cca-8e2a-e4c4e2597e8c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3bdb22ed-1952-4cca-8e2a-e4c4e2597e8c\") pod \"ovsdbserver-sb-1\" (UID: \"e5ec94c1-f60b-4156-b3ba-278db827c9b8\") " pod="openstack/ovsdbserver-sb-1" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.391136 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cdlpw\" (UniqueName: \"kubernetes.io/projected/e5ec94c1-f60b-4156-b3ba-278db827c9b8-kube-api-access-cdlpw\") pod \"ovsdbserver-sb-1\" (UID: \"e5ec94c1-f60b-4156-b3ba-278db827c9b8\") " pod="openstack/ovsdbserver-sb-1" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.391210 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xw2xr\" (UniqueName: \"kubernetes.io/projected/b188658f-1e5c-49d4-be19-0a888caddb43-kube-api-access-xw2xr\") pod \"ovsdbserver-sb-2\" (UID: \"b188658f-1e5c-49d4-be19-0a888caddb43\") " pod="openstack/ovsdbserver-sb-2" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.391220 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/b188658f-1e5c-49d4-be19-0a888caddb43-ovsdb-rundir\") pod \"ovsdbserver-sb-2\" (UID: \"b188658f-1e5c-49d4-be19-0a888caddb43\") " pod="openstack/ovsdbserver-sb-2" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.391252 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b188658f-1e5c-49d4-be19-0a888caddb43-combined-ca-bundle\") pod \"ovsdbserver-sb-2\" (UID: \"b188658f-1e5c-49d4-be19-0a888caddb43\") " pod="openstack/ovsdbserver-sb-2" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.391279 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bffaea76-e210-44d1-8caa-ebf82edb8ef0-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"bffaea76-e210-44d1-8caa-ebf82edb8ef0\") " pod="openstack/ovsdbserver-sb-0" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.391307 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b188658f-1e5c-49d4-be19-0a888caddb43-config\") pod \"ovsdbserver-sb-2\" (UID: \"b188658f-1e5c-49d4-be19-0a888caddb43\") " pod="openstack/ovsdbserver-sb-2" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.391335 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/e5ec94c1-f60b-4156-b3ba-278db827c9b8-ovsdb-rundir\") pod \"ovsdbserver-sb-1\" (UID: \"e5ec94c1-f60b-4156-b3ba-278db827c9b8\") " pod="openstack/ovsdbserver-sb-1" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.391362 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5ec94c1-f60b-4156-b3ba-278db827c9b8-combined-ca-bundle\") pod \"ovsdbserver-sb-1\" (UID: \"e5ec94c1-f60b-4156-b3ba-278db827c9b8\") " pod="openstack/ovsdbserver-sb-1" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.391386 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bffaea76-e210-44d1-8caa-ebf82edb8ef0-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"bffaea76-e210-44d1-8caa-ebf82edb8ef0\") " pod="openstack/ovsdbserver-sb-0" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.391389 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/bffaea76-e210-44d1-8caa-ebf82edb8ef0-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"bffaea76-e210-44d1-8caa-ebf82edb8ef0\") " pod="openstack/ovsdbserver-sb-0" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.391420 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e5ec94c1-f60b-4156-b3ba-278db827c9b8-config\") pod \"ovsdbserver-sb-1\" (UID: \"e5ec94c1-f60b-4156-b3ba-278db827c9b8\") " pod="openstack/ovsdbserver-sb-1" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.392004 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bffaea76-e210-44d1-8caa-ebf82edb8ef0-config\") pod \"ovsdbserver-sb-0\" (UID: \"bffaea76-e210-44d1-8caa-ebf82edb8ef0\") " pod="openstack/ovsdbserver-sb-0" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.396523 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b188658f-1e5c-49d4-be19-0a888caddb43-config\") pod \"ovsdbserver-sb-2\" (UID: \"b188658f-1e5c-49d4-be19-0a888caddb43\") " pod="openstack/ovsdbserver-sb-2" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.406762 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5ec94c1-f60b-4156-b3ba-278db827c9b8-combined-ca-bundle\") pod \"ovsdbserver-sb-1\" (UID: \"e5ec94c1-f60b-4156-b3ba-278db827c9b8\") " pod="openstack/ovsdbserver-sb-1" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.409503 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-2" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.411562 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b188658f-1e5c-49d4-be19-0a888caddb43-scripts\") pod \"ovsdbserver-sb-2\" (UID: \"b188658f-1e5c-49d4-be19-0a888caddb43\") " pod="openstack/ovsdbserver-sb-2" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.411607 4907 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.411636 4907 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-3bdb22ed-1952-4cca-8e2a-e4c4e2597e8c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3bdb22ed-1952-4cca-8e2a-e4c4e2597e8c\") pod \"ovsdbserver-sb-1\" (UID: \"e5ec94c1-f60b-4156-b3ba-278db827c9b8\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/8198106df7b96a617bc2acc54b8057faf23b369987a114c32f994974a6cdd7ae/globalmount\"" pod="openstack/ovsdbserver-sb-1" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.412453 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e5ec94c1-f60b-4156-b3ba-278db827c9b8-config\") pod \"ovsdbserver-sb-1\" (UID: \"e5ec94c1-f60b-4156-b3ba-278db827c9b8\") " pod="openstack/ovsdbserver-sb-1" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.413641 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/e5ec94c1-f60b-4156-b3ba-278db827c9b8-ovsdb-rundir\") pod \"ovsdbserver-sb-1\" (UID: \"e5ec94c1-f60b-4156-b3ba-278db827c9b8\") " pod="openstack/ovsdbserver-sb-1" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.413656 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bffaea76-e210-44d1-8caa-ebf82edb8ef0-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"bffaea76-e210-44d1-8caa-ebf82edb8ef0\") " pod="openstack/ovsdbserver-sb-0" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.413895 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e5ec94c1-f60b-4156-b3ba-278db827c9b8-scripts\") pod \"ovsdbserver-sb-1\" (UID: \"e5ec94c1-f60b-4156-b3ba-278db827c9b8\") " pod="openstack/ovsdbserver-sb-1" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.415158 4907 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.415359 4907 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-4247ec19-7024-4129-80de-23a6ba6ce376\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-4247ec19-7024-4129-80de-23a6ba6ce376\") pod \"ovsdbserver-sb-0\" (UID: \"bffaea76-e210-44d1-8caa-ebf82edb8ef0\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/22668b1ed8c08d70322619ed0b00585c03f06e051ce689e13cc883c7d890ed36/globalmount\"" pod="openstack/ovsdbserver-sb-0" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.417553 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bffaea76-e210-44d1-8caa-ebf82edb8ef0-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"bffaea76-e210-44d1-8caa-ebf82edb8ef0\") " pod="openstack/ovsdbserver-sb-0" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.419570 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xw2xr\" (UniqueName: \"kubernetes.io/projected/b188658f-1e5c-49d4-be19-0a888caddb43-kube-api-access-xw2xr\") pod \"ovsdbserver-sb-2\" (UID: \"b188658f-1e5c-49d4-be19-0a888caddb43\") " pod="openstack/ovsdbserver-sb-2" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.424540 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b188658f-1e5c-49d4-be19-0a888caddb43-combined-ca-bundle\") pod \"ovsdbserver-sb-2\" (UID: \"b188658f-1e5c-49d4-be19-0a888caddb43\") " pod="openstack/ovsdbserver-sb-2" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.432314 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-1"] Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.434995 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-shc8z\" (UniqueName: \"kubernetes.io/projected/bffaea76-e210-44d1-8caa-ebf82edb8ef0-kube-api-access-shc8z\") pod \"ovsdbserver-sb-0\" (UID: \"bffaea76-e210-44d1-8caa-ebf82edb8ef0\") " pod="openstack/ovsdbserver-sb-0" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.437792 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cdlpw\" (UniqueName: \"kubernetes.io/projected/e5ec94c1-f60b-4156-b3ba-278db827c9b8-kube-api-access-cdlpw\") pod \"ovsdbserver-sb-1\" (UID: \"e5ec94c1-f60b-4156-b3ba-278db827c9b8\") " pod="openstack/ovsdbserver-sb-1" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.440920 4907 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.440953 4907 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-8e01f892-be99-49f8-b8aa-bf80a68ae42b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-8e01f892-be99-49f8-b8aa-bf80a68ae42b\") pod \"ovsdbserver-sb-2\" (UID: \"b188658f-1e5c-49d4-be19-0a888caddb43\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/00f7424f6f9b73ded1a9336a9e06d79a0db8f52c63e13ea4f24769687d2a94b2/globalmount\"" pod="openstack/ovsdbserver-sb-2" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.449950 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-2"] Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.460938 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-3bdb22ed-1952-4cca-8e2a-e4c4e2597e8c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3bdb22ed-1952-4cca-8e2a-e4c4e2597e8c\") pod \"ovsdbserver-sb-1\" (UID: \"e5ec94c1-f60b-4156-b3ba-278db827c9b8\") " pod="openstack/ovsdbserver-sb-1" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.465272 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-4247ec19-7024-4129-80de-23a6ba6ce376\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-4247ec19-7024-4129-80de-23a6ba6ce376\") pod \"ovsdbserver-sb-0\" (UID: \"bffaea76-e210-44d1-8caa-ebf82edb8ef0\") " pod="openstack/ovsdbserver-sb-0" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.468146 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-8e01f892-be99-49f8-b8aa-bf80a68ae42b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-8e01f892-be99-49f8-b8aa-bf80a68ae42b\") pod \"ovsdbserver-sb-2\" (UID: \"b188658f-1e5c-49d4-be19-0a888caddb43\") " pod="openstack/ovsdbserver-sb-2" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.478522 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.494638 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a167eb31-6b82-44f9-bf66-465b64a91226-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"a167eb31-6b82-44f9-bf66-465b64a91226\") " pod="openstack/ovsdbserver-nb-0" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.494675 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-be1fde5e-0257-4035-803c-6c276dcd538c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-be1fde5e-0257-4035-803c-6c276dcd538c\") pod \"ovsdbserver-nb-2\" (UID: \"3ca7bf8f-4514-493e-9385-75cfcd0807f9\") " pod="openstack/ovsdbserver-nb-2" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.494854 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-f5e5ec9c-2696-40da-aa58-a0cdbb969cd9\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f5e5ec9c-2696-40da-aa58-a0cdbb969cd9\") pod \"ovsdbserver-nb-0\" (UID: \"a167eb31-6b82-44f9-bf66-465b64a91226\") " pod="openstack/ovsdbserver-nb-0" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.494921 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e270de1-912f-48f9-bfc7-e415fa5c7b5d-combined-ca-bundle\") pod \"ovsdbserver-nb-1\" (UID: \"7e270de1-912f-48f9-bfc7-e415fa5c7b5d\") " pod="openstack/ovsdbserver-nb-1" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.494995 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qfvnc\" (UniqueName: \"kubernetes.io/projected/a167eb31-6b82-44f9-bf66-465b64a91226-kube-api-access-qfvnc\") pod \"ovsdbserver-nb-0\" (UID: \"a167eb31-6b82-44f9-bf66-465b64a91226\") " pod="openstack/ovsdbserver-nb-0" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.495034 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-45mq2\" (UniqueName: \"kubernetes.io/projected/3ca7bf8f-4514-493e-9385-75cfcd0807f9-kube-api-access-45mq2\") pod \"ovsdbserver-nb-2\" (UID: \"3ca7bf8f-4514-493e-9385-75cfcd0807f9\") " pod="openstack/ovsdbserver-nb-2" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.495078 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a167eb31-6b82-44f9-bf66-465b64a91226-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"a167eb31-6b82-44f9-bf66-465b64a91226\") " pod="openstack/ovsdbserver-nb-0" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.495116 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a167eb31-6b82-44f9-bf66-465b64a91226-config\") pod \"ovsdbserver-nb-0\" (UID: \"a167eb31-6b82-44f9-bf66-465b64a91226\") " pod="openstack/ovsdbserver-nb-0" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.495153 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e270de1-912f-48f9-bfc7-e415fa5c7b5d-config\") pod \"ovsdbserver-nb-1\" (UID: \"7e270de1-912f-48f9-bfc7-e415fa5c7b5d\") " pod="openstack/ovsdbserver-nb-1" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.495209 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-4b7e3d76-850f-42aa-9232-939baf06bab0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-4b7e3d76-850f-42aa-9232-939baf06bab0\") pod \"ovsdbserver-nb-1\" (UID: \"7e270de1-912f-48f9-bfc7-e415fa5c7b5d\") " pod="openstack/ovsdbserver-nb-1" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.495286 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ca7bf8f-4514-493e-9385-75cfcd0807f9-combined-ca-bundle\") pod \"ovsdbserver-nb-2\" (UID: \"3ca7bf8f-4514-493e-9385-75cfcd0807f9\") " pod="openstack/ovsdbserver-nb-2" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.495317 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/a167eb31-6b82-44f9-bf66-465b64a91226-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"a167eb31-6b82-44f9-bf66-465b64a91226\") " pod="openstack/ovsdbserver-nb-0" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.495417 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3ca7bf8f-4514-493e-9385-75cfcd0807f9-scripts\") pod \"ovsdbserver-nb-2\" (UID: \"3ca7bf8f-4514-493e-9385-75cfcd0807f9\") " pod="openstack/ovsdbserver-nb-2" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.495931 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/3ca7bf8f-4514-493e-9385-75cfcd0807f9-ovsdb-rundir\") pod \"ovsdbserver-nb-2\" (UID: \"3ca7bf8f-4514-493e-9385-75cfcd0807f9\") " pod="openstack/ovsdbserver-nb-2" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.496014 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7e270de1-912f-48f9-bfc7-e415fa5c7b5d-scripts\") pod \"ovsdbserver-nb-1\" (UID: \"7e270de1-912f-48f9-bfc7-e415fa5c7b5d\") " pod="openstack/ovsdbserver-nb-1" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.496147 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-2" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.497102 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3ca7bf8f-4514-493e-9385-75cfcd0807f9-config\") pod \"ovsdbserver-nb-2\" (UID: \"3ca7bf8f-4514-493e-9385-75cfcd0807f9\") " pod="openstack/ovsdbserver-nb-2" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.497855 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/7e270de1-912f-48f9-bfc7-e415fa5c7b5d-ovsdb-rundir\") pod \"ovsdbserver-nb-1\" (UID: \"7e270de1-912f-48f9-bfc7-e415fa5c7b5d\") " pod="openstack/ovsdbserver-nb-1" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.497921 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lbbbq\" (UniqueName: \"kubernetes.io/projected/7e270de1-912f-48f9-bfc7-e415fa5c7b5d-kube-api-access-lbbbq\") pod \"ovsdbserver-nb-1\" (UID: \"7e270de1-912f-48f9-bfc7-e415fa5c7b5d\") " pod="openstack/ovsdbserver-nb-1" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.506544 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-1" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.599151 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a167eb31-6b82-44f9-bf66-465b64a91226-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"a167eb31-6b82-44f9-bf66-465b64a91226\") " pod="openstack/ovsdbserver-nb-0" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.600400 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-be1fde5e-0257-4035-803c-6c276dcd538c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-be1fde5e-0257-4035-803c-6c276dcd538c\") pod \"ovsdbserver-nb-2\" (UID: \"3ca7bf8f-4514-493e-9385-75cfcd0807f9\") " pod="openstack/ovsdbserver-nb-2" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.600451 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-f5e5ec9c-2696-40da-aa58-a0cdbb969cd9\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f5e5ec9c-2696-40da-aa58-a0cdbb969cd9\") pod \"ovsdbserver-nb-0\" (UID: \"a167eb31-6b82-44f9-bf66-465b64a91226\") " pod="openstack/ovsdbserver-nb-0" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.600470 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e270de1-912f-48f9-bfc7-e415fa5c7b5d-combined-ca-bundle\") pod \"ovsdbserver-nb-1\" (UID: \"7e270de1-912f-48f9-bfc7-e415fa5c7b5d\") " pod="openstack/ovsdbserver-nb-1" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.600493 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qfvnc\" (UniqueName: \"kubernetes.io/projected/a167eb31-6b82-44f9-bf66-465b64a91226-kube-api-access-qfvnc\") pod \"ovsdbserver-nb-0\" (UID: \"a167eb31-6b82-44f9-bf66-465b64a91226\") " pod="openstack/ovsdbserver-nb-0" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.600509 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-45mq2\" (UniqueName: \"kubernetes.io/projected/3ca7bf8f-4514-493e-9385-75cfcd0807f9-kube-api-access-45mq2\") pod \"ovsdbserver-nb-2\" (UID: \"3ca7bf8f-4514-493e-9385-75cfcd0807f9\") " pod="openstack/ovsdbserver-nb-2" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.600547 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a167eb31-6b82-44f9-bf66-465b64a91226-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"a167eb31-6b82-44f9-bf66-465b64a91226\") " pod="openstack/ovsdbserver-nb-0" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.600569 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a167eb31-6b82-44f9-bf66-465b64a91226-config\") pod \"ovsdbserver-nb-0\" (UID: \"a167eb31-6b82-44f9-bf66-465b64a91226\") " pod="openstack/ovsdbserver-nb-0" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.600592 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e270de1-912f-48f9-bfc7-e415fa5c7b5d-config\") pod \"ovsdbserver-nb-1\" (UID: \"7e270de1-912f-48f9-bfc7-e415fa5c7b5d\") " pod="openstack/ovsdbserver-nb-1" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.600617 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-4b7e3d76-850f-42aa-9232-939baf06bab0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-4b7e3d76-850f-42aa-9232-939baf06bab0\") pod \"ovsdbserver-nb-1\" (UID: \"7e270de1-912f-48f9-bfc7-e415fa5c7b5d\") " pod="openstack/ovsdbserver-nb-1" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.600643 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ca7bf8f-4514-493e-9385-75cfcd0807f9-combined-ca-bundle\") pod \"ovsdbserver-nb-2\" (UID: \"3ca7bf8f-4514-493e-9385-75cfcd0807f9\") " pod="openstack/ovsdbserver-nb-2" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.600665 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/a167eb31-6b82-44f9-bf66-465b64a91226-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"a167eb31-6b82-44f9-bf66-465b64a91226\") " pod="openstack/ovsdbserver-nb-0" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.600703 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3ca7bf8f-4514-493e-9385-75cfcd0807f9-scripts\") pod \"ovsdbserver-nb-2\" (UID: \"3ca7bf8f-4514-493e-9385-75cfcd0807f9\") " pod="openstack/ovsdbserver-nb-2" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.600730 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/3ca7bf8f-4514-493e-9385-75cfcd0807f9-ovsdb-rundir\") pod \"ovsdbserver-nb-2\" (UID: \"3ca7bf8f-4514-493e-9385-75cfcd0807f9\") " pod="openstack/ovsdbserver-nb-2" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.600764 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7e270de1-912f-48f9-bfc7-e415fa5c7b5d-scripts\") pod \"ovsdbserver-nb-1\" (UID: \"7e270de1-912f-48f9-bfc7-e415fa5c7b5d\") " pod="openstack/ovsdbserver-nb-1" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.600849 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3ca7bf8f-4514-493e-9385-75cfcd0807f9-config\") pod \"ovsdbserver-nb-2\" (UID: \"3ca7bf8f-4514-493e-9385-75cfcd0807f9\") " pod="openstack/ovsdbserver-nb-2" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.600911 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/7e270de1-912f-48f9-bfc7-e415fa5c7b5d-ovsdb-rundir\") pod \"ovsdbserver-nb-1\" (UID: \"7e270de1-912f-48f9-bfc7-e415fa5c7b5d\") " pod="openstack/ovsdbserver-nb-1" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.600946 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lbbbq\" (UniqueName: \"kubernetes.io/projected/7e270de1-912f-48f9-bfc7-e415fa5c7b5d-kube-api-access-lbbbq\") pod \"ovsdbserver-nb-1\" (UID: \"7e270de1-912f-48f9-bfc7-e415fa5c7b5d\") " pod="openstack/ovsdbserver-nb-1" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.604558 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/7e270de1-912f-48f9-bfc7-e415fa5c7b5d-ovsdb-rundir\") pod \"ovsdbserver-nb-1\" (UID: \"7e270de1-912f-48f9-bfc7-e415fa5c7b5d\") " pod="openstack/ovsdbserver-nb-1" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.604857 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7e270de1-912f-48f9-bfc7-e415fa5c7b5d-scripts\") pod \"ovsdbserver-nb-1\" (UID: \"7e270de1-912f-48f9-bfc7-e415fa5c7b5d\") " pod="openstack/ovsdbserver-nb-1" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.605197 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3ca7bf8f-4514-493e-9385-75cfcd0807f9-config\") pod \"ovsdbserver-nb-2\" (UID: \"3ca7bf8f-4514-493e-9385-75cfcd0807f9\") " pod="openstack/ovsdbserver-nb-2" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.605223 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/3ca7bf8f-4514-493e-9385-75cfcd0807f9-ovsdb-rundir\") pod \"ovsdbserver-nb-2\" (UID: \"3ca7bf8f-4514-493e-9385-75cfcd0807f9\") " pod="openstack/ovsdbserver-nb-2" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.607453 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3ca7bf8f-4514-493e-9385-75cfcd0807f9-scripts\") pod \"ovsdbserver-nb-2\" (UID: \"3ca7bf8f-4514-493e-9385-75cfcd0807f9\") " pod="openstack/ovsdbserver-nb-2" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.608078 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a167eb31-6b82-44f9-bf66-465b64a91226-config\") pod \"ovsdbserver-nb-0\" (UID: \"a167eb31-6b82-44f9-bf66-465b64a91226\") " pod="openstack/ovsdbserver-nb-0" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.609675 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ca7bf8f-4514-493e-9385-75cfcd0807f9-combined-ca-bundle\") pod \"ovsdbserver-nb-2\" (UID: \"3ca7bf8f-4514-493e-9385-75cfcd0807f9\") " pod="openstack/ovsdbserver-nb-2" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.609681 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e270de1-912f-48f9-bfc7-e415fa5c7b5d-combined-ca-bundle\") pod \"ovsdbserver-nb-1\" (UID: \"7e270de1-912f-48f9-bfc7-e415fa5c7b5d\") " pod="openstack/ovsdbserver-nb-1" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.610190 4907 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.610219 4907 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-be1fde5e-0257-4035-803c-6c276dcd538c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-be1fde5e-0257-4035-803c-6c276dcd538c\") pod \"ovsdbserver-nb-2\" (UID: \"3ca7bf8f-4514-493e-9385-75cfcd0807f9\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/554de8a28393bb5481936a1456b52e6abac7129d07afc1025c887f49696ce35e/globalmount\"" pod="openstack/ovsdbserver-nb-2" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.611012 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e270de1-912f-48f9-bfc7-e415fa5c7b5d-config\") pod \"ovsdbserver-nb-1\" (UID: \"7e270de1-912f-48f9-bfc7-e415fa5c7b5d\") " pod="openstack/ovsdbserver-nb-1" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.611231 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/a167eb31-6b82-44f9-bf66-465b64a91226-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"a167eb31-6b82-44f9-bf66-465b64a91226\") " pod="openstack/ovsdbserver-nb-0" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.611551 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a167eb31-6b82-44f9-bf66-465b64a91226-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"a167eb31-6b82-44f9-bf66-465b64a91226\") " pod="openstack/ovsdbserver-nb-0" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.614034 4907 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.614076 4907 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-4b7e3d76-850f-42aa-9232-939baf06bab0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-4b7e3d76-850f-42aa-9232-939baf06bab0\") pod \"ovsdbserver-nb-1\" (UID: \"7e270de1-912f-48f9-bfc7-e415fa5c7b5d\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/99ba545b23bc3775acd4c535d721325e522f269a911518ef755286a88dbac8f2/globalmount\"" pod="openstack/ovsdbserver-nb-1" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.614192 4907 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.614242 4907 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-f5e5ec9c-2696-40da-aa58-a0cdbb969cd9\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f5e5ec9c-2696-40da-aa58-a0cdbb969cd9\") pod \"ovsdbserver-nb-0\" (UID: \"a167eb31-6b82-44f9-bf66-465b64a91226\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/7bde764fd3ca106780c183d470e3dca511bd76d320905b722bf14f86ea815a69/globalmount\"" pod="openstack/ovsdbserver-nb-0" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.624588 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a167eb31-6b82-44f9-bf66-465b64a91226-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"a167eb31-6b82-44f9-bf66-465b64a91226\") " pod="openstack/ovsdbserver-nb-0" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.624963 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qfvnc\" (UniqueName: \"kubernetes.io/projected/a167eb31-6b82-44f9-bf66-465b64a91226-kube-api-access-qfvnc\") pod \"ovsdbserver-nb-0\" (UID: \"a167eb31-6b82-44f9-bf66-465b64a91226\") " pod="openstack/ovsdbserver-nb-0" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.626502 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lbbbq\" (UniqueName: \"kubernetes.io/projected/7e270de1-912f-48f9-bfc7-e415fa5c7b5d-kube-api-access-lbbbq\") pod \"ovsdbserver-nb-1\" (UID: \"7e270de1-912f-48f9-bfc7-e415fa5c7b5d\") " pod="openstack/ovsdbserver-nb-1" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.629604 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-45mq2\" (UniqueName: \"kubernetes.io/projected/3ca7bf8f-4514-493e-9385-75cfcd0807f9-kube-api-access-45mq2\") pod \"ovsdbserver-nb-2\" (UID: \"3ca7bf8f-4514-493e-9385-75cfcd0807f9\") " pod="openstack/ovsdbserver-nb-2" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.661215 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-f5e5ec9c-2696-40da-aa58-a0cdbb969cd9\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f5e5ec9c-2696-40da-aa58-a0cdbb969cd9\") pod \"ovsdbserver-nb-0\" (UID: \"a167eb31-6b82-44f9-bf66-465b64a91226\") " pod="openstack/ovsdbserver-nb-0" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.668829 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-4b7e3d76-850f-42aa-9232-939baf06bab0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-4b7e3d76-850f-42aa-9232-939baf06bab0\") pod \"ovsdbserver-nb-1\" (UID: \"7e270de1-912f-48f9-bfc7-e415fa5c7b5d\") " pod="openstack/ovsdbserver-nb-1" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.674257 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-be1fde5e-0257-4035-803c-6c276dcd538c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-be1fde5e-0257-4035-803c-6c276dcd538c\") pod \"ovsdbserver-nb-2\" (UID: \"3ca7bf8f-4514-493e-9385-75cfcd0807f9\") " pod="openstack/ovsdbserver-nb-2" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.858055 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-1" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.872320 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-2" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.963517 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Mar 13 15:33:35 crc kubenswrapper[4907]: I0313 15:33:35.017494 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Mar 13 15:33:35 crc kubenswrapper[4907]: W0313 15:33:35.124954 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb188658f_1e5c_49d4_be19_0a888caddb43.slice/crio-674402c7ad6b0314499d0c1c017cc7ebeb988958ee9224c14b3b9c8ef682dfba WatchSource:0}: Error finding container 674402c7ad6b0314499d0c1c017cc7ebeb988958ee9224c14b3b9c8ef682dfba: Status 404 returned error can't find the container with id 674402c7ad6b0314499d0c1c017cc7ebeb988958ee9224c14b3b9c8ef682dfba Mar 13 15:33:35 crc kubenswrapper[4907]: I0313 15:33:35.130507 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-2"] Mar 13 15:33:35 crc kubenswrapper[4907]: I0313 15:33:35.206445 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-1"] Mar 13 15:33:35 crc kubenswrapper[4907]: I0313 15:33:35.418136 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-1"] Mar 13 15:33:35 crc kubenswrapper[4907]: I0313 15:33:35.529205 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-2"] Mar 13 15:33:35 crc kubenswrapper[4907]: W0313 15:33:35.535802 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3ca7bf8f_4514_493e_9385_75cfcd0807f9.slice/crio-3b780eb2bb9a20e04ad7d5ac0beab1e2e4343780f0a2d6530af68108f6b4beef WatchSource:0}: Error finding container 3b780eb2bb9a20e04ad7d5ac0beab1e2e4343780f0a2d6530af68108f6b4beef: Status 404 returned error can't find the container with id 3b780eb2bb9a20e04ad7d5ac0beab1e2e4343780f0a2d6530af68108f6b4beef Mar 13 15:33:35 crc kubenswrapper[4907]: W0313 15:33:35.610076 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda167eb31_6b82_44f9_bf66_465b64a91226.slice/crio-79ebf4c6ba715cb4043d5908d97a193e951453a828adb94d17d744b98922a037 WatchSource:0}: Error finding container 79ebf4c6ba715cb4043d5908d97a193e951453a828adb94d17d744b98922a037: Status 404 returned error can't find the container with id 79ebf4c6ba715cb4043d5908d97a193e951453a828adb94d17d744b98922a037 Mar 13 15:33:35 crc kubenswrapper[4907]: I0313 15:33:35.612099 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Mar 13 15:33:35 crc kubenswrapper[4907]: I0313 15:33:35.816992 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-1" event={"ID":"e5ec94c1-f60b-4156-b3ba-278db827c9b8","Type":"ContainerStarted","Data":"8f22dd4ec76846c8d2a37b6af861b79cc8fcee078f77e42f7ce5d93872a72cf3"} Mar 13 15:33:35 crc kubenswrapper[4907]: I0313 15:33:35.817035 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-1" event={"ID":"e5ec94c1-f60b-4156-b3ba-278db827c9b8","Type":"ContainerStarted","Data":"2c3fa3c8d0e82f5efce1785d752c20bdf28c4cd1e70d951f529de9c7869a9e3b"} Mar 13 15:33:35 crc kubenswrapper[4907]: I0313 15:33:35.817045 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-1" event={"ID":"e5ec94c1-f60b-4156-b3ba-278db827c9b8","Type":"ContainerStarted","Data":"0bff93d9d737cb5a76a3af81265b6b53584369519d92c7569bfd52f6414fe9d3"} Mar 13 15:33:35 crc kubenswrapper[4907]: I0313 15:33:35.820950 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"a167eb31-6b82-44f9-bf66-465b64a91226","Type":"ContainerStarted","Data":"4ba727733c83b392d0aa81d74209df8981f0e659132779ee0ceb6a05ef7d5103"} Mar 13 15:33:35 crc kubenswrapper[4907]: I0313 15:33:35.821002 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"a167eb31-6b82-44f9-bf66-465b64a91226","Type":"ContainerStarted","Data":"79ebf4c6ba715cb4043d5908d97a193e951453a828adb94d17d744b98922a037"} Mar 13 15:33:35 crc kubenswrapper[4907]: I0313 15:33:35.828177 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"bffaea76-e210-44d1-8caa-ebf82edb8ef0","Type":"ContainerStarted","Data":"90dadbe657fd114fe2e914afa903244ef949d9111d57f162f668daadb25f1921"} Mar 13 15:33:35 crc kubenswrapper[4907]: I0313 15:33:35.828233 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"bffaea76-e210-44d1-8caa-ebf82edb8ef0","Type":"ContainerStarted","Data":"d4d1ad2a84a27593b08932e7a490f0cb17eb191199bc9374857ed81b7661c44e"} Mar 13 15:33:35 crc kubenswrapper[4907]: I0313 15:33:35.828245 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"bffaea76-e210-44d1-8caa-ebf82edb8ef0","Type":"ContainerStarted","Data":"59294b2cb6419dec7fe1f1dbada2c9df28cf8f11feb23127830df996d8ed5535"} Mar 13 15:33:35 crc kubenswrapper[4907]: I0313 15:33:35.842955 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-2" event={"ID":"b188658f-1e5c-49d4-be19-0a888caddb43","Type":"ContainerStarted","Data":"0670fcbcd8be942fc9083ff1c66a6253e07eaec25465a3bf14ed1e0c14798409"} Mar 13 15:33:35 crc kubenswrapper[4907]: I0313 15:33:35.843001 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-2" event={"ID":"b188658f-1e5c-49d4-be19-0a888caddb43","Type":"ContainerStarted","Data":"aef67d6a1ec2fcb17e8dcb7f26ca4c158448ccf469b72e5cca1a5e99b7cd2d21"} Mar 13 15:33:35 crc kubenswrapper[4907]: I0313 15:33:35.843013 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-2" event={"ID":"b188658f-1e5c-49d4-be19-0a888caddb43","Type":"ContainerStarted","Data":"674402c7ad6b0314499d0c1c017cc7ebeb988958ee9224c14b3b9c8ef682dfba"} Mar 13 15:33:35 crc kubenswrapper[4907]: I0313 15:33:35.844956 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-1" podStartSLOduration=2.844935366 podStartE2EDuration="2.844935366s" podCreationTimestamp="2026-03-13 15:33:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:33:35.837669516 +0000 UTC m=+5314.737457225" watchObservedRunningTime="2026-03-13 15:33:35.844935366 +0000 UTC m=+5314.744723065" Mar 13 15:33:35 crc kubenswrapper[4907]: I0313 15:33:35.845907 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-1" event={"ID":"7e270de1-912f-48f9-bfc7-e415fa5c7b5d","Type":"ContainerStarted","Data":"d7819948441d190c863d4f57f53dd980b3bbb30335e7487d4366124beb0e6c1d"} Mar 13 15:33:35 crc kubenswrapper[4907]: I0313 15:33:35.845947 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-1" event={"ID":"7e270de1-912f-48f9-bfc7-e415fa5c7b5d","Type":"ContainerStarted","Data":"147d938f69e2e32e5d6fee7c068767956ae44d04924674daefa96fdf56a2a532"} Mar 13 15:33:35 crc kubenswrapper[4907]: I0313 15:33:35.851057 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-2" event={"ID":"3ca7bf8f-4514-493e-9385-75cfcd0807f9","Type":"ContainerStarted","Data":"8404c331259c77750125e3e64294e712b997d439ae3bb0cde654a622897811bd"} Mar 13 15:33:35 crc kubenswrapper[4907]: I0313 15:33:35.851095 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-2" event={"ID":"3ca7bf8f-4514-493e-9385-75cfcd0807f9","Type":"ContainerStarted","Data":"3b780eb2bb9a20e04ad7d5ac0beab1e2e4343780f0a2d6530af68108f6b4beef"} Mar 13 15:33:35 crc kubenswrapper[4907]: I0313 15:33:35.861496 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=2.861480548 podStartE2EDuration="2.861480548s" podCreationTimestamp="2026-03-13 15:33:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:33:35.855303799 +0000 UTC m=+5314.755091488" watchObservedRunningTime="2026-03-13 15:33:35.861480548 +0000 UTC m=+5314.761268237" Mar 13 15:33:35 crc kubenswrapper[4907]: I0313 15:33:35.884217 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-2" podStartSLOduration=2.88419758 podStartE2EDuration="2.88419758s" podCreationTimestamp="2026-03-13 15:33:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:33:35.873859387 +0000 UTC m=+5314.773647066" watchObservedRunningTime="2026-03-13 15:33:35.88419758 +0000 UTC m=+5314.783985269" Mar 13 15:33:36 crc kubenswrapper[4907]: I0313 15:33:36.858582 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-1" event={"ID":"7e270de1-912f-48f9-bfc7-e415fa5c7b5d","Type":"ContainerStarted","Data":"e04df1932d1e43003ccafc8b156c0365035744d253a0b242b932fc2d50b2a961"} Mar 13 15:33:36 crc kubenswrapper[4907]: I0313 15:33:36.861556 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-2" event={"ID":"3ca7bf8f-4514-493e-9385-75cfcd0807f9","Type":"ContainerStarted","Data":"953ee43d80e6b216eca1936519f349ce31b34a7eaa34a251834139aa0548d049"} Mar 13 15:33:36 crc kubenswrapper[4907]: I0313 15:33:36.863997 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"a167eb31-6b82-44f9-bf66-465b64a91226","Type":"ContainerStarted","Data":"1fda18c9bb6492a6795eff525886132c05c148df97ad8119fd84e436dbecec7e"} Mar 13 15:33:36 crc kubenswrapper[4907]: I0313 15:33:36.877237 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-1" podStartSLOduration=3.87720873 podStartE2EDuration="3.87720873s" podCreationTimestamp="2026-03-13 15:33:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:33:36.877112027 +0000 UTC m=+5315.776899716" watchObservedRunningTime="2026-03-13 15:33:36.87720873 +0000 UTC m=+5315.776996419" Mar 13 15:33:36 crc kubenswrapper[4907]: I0313 15:33:36.899251 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=3.899232342 podStartE2EDuration="3.899232342s" podCreationTimestamp="2026-03-13 15:33:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:33:36.893567717 +0000 UTC m=+5315.793355426" watchObservedRunningTime="2026-03-13 15:33:36.899232342 +0000 UTC m=+5315.799020031" Mar 13 15:33:36 crc kubenswrapper[4907]: I0313 15:33:36.918501 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-2" podStartSLOduration=3.918443188 podStartE2EDuration="3.918443188s" podCreationTimestamp="2026-03-13 15:33:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:33:36.908859876 +0000 UTC m=+5315.808647585" watchObservedRunningTime="2026-03-13 15:33:36.918443188 +0000 UTC m=+5315.818230877" Mar 13 15:33:37 crc kubenswrapper[4907]: I0313 15:33:37.479045 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Mar 13 15:33:37 crc kubenswrapper[4907]: I0313 15:33:37.496774 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-2" Mar 13 15:33:37 crc kubenswrapper[4907]: I0313 15:33:37.506971 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-1" Mar 13 15:33:37 crc kubenswrapper[4907]: I0313 15:33:37.859084 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-1" Mar 13 15:33:37 crc kubenswrapper[4907]: I0313 15:33:37.873626 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-2" Mar 13 15:33:37 crc kubenswrapper[4907]: I0313 15:33:37.963951 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Mar 13 15:33:39 crc kubenswrapper[4907]: I0313 15:33:39.478805 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Mar 13 15:33:39 crc kubenswrapper[4907]: I0313 15:33:39.497390 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-2" Mar 13 15:33:39 crc kubenswrapper[4907]: I0313 15:33:39.507128 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-1" Mar 13 15:33:39 crc kubenswrapper[4907]: I0313 15:33:39.859101 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-1" Mar 13 15:33:39 crc kubenswrapper[4907]: I0313 15:33:39.873832 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-2" Mar 13 15:33:39 crc kubenswrapper[4907]: I0313 15:33:39.963988 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Mar 13 15:33:40 crc kubenswrapper[4907]: I0313 15:33:40.535759 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Mar 13 15:33:40 crc kubenswrapper[4907]: I0313 15:33:40.542773 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-1" Mar 13 15:33:40 crc kubenswrapper[4907]: I0313 15:33:40.555656 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-2" Mar 13 15:33:40 crc kubenswrapper[4907]: I0313 15:33:40.584867 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Mar 13 15:33:40 crc kubenswrapper[4907]: I0313 15:33:40.598108 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-1" Mar 13 15:33:40 crc kubenswrapper[4907]: I0313 15:33:40.598794 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-2" Mar 13 15:33:40 crc kubenswrapper[4907]: I0313 15:33:40.783477 4907 scope.go:117] "RemoveContainer" containerID="592bf9f175f440e943369aab2abbb13ad2c26412cb0e0738c4a33a156b8a5596" Mar 13 15:33:40 crc kubenswrapper[4907]: E0313 15:33:40.785302 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:33:40 crc kubenswrapper[4907]: I0313 15:33:40.843485 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-794d85d9cf-xl4f8"] Mar 13 15:33:40 crc kubenswrapper[4907]: I0313 15:33:40.845412 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-794d85d9cf-xl4f8" Mar 13 15:33:40 crc kubenswrapper[4907]: I0313 15:33:40.847506 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Mar 13 15:33:40 crc kubenswrapper[4907]: I0313 15:33:40.852616 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-794d85d9cf-xl4f8"] Mar 13 15:33:40 crc kubenswrapper[4907]: I0313 15:33:40.897595 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-1" Mar 13 15:33:40 crc kubenswrapper[4907]: I0313 15:33:40.914375 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-2" Mar 13 15:33:40 crc kubenswrapper[4907]: I0313 15:33:40.952899 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-1" Mar 13 15:33:40 crc kubenswrapper[4907]: I0313 15:33:40.960723 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-2" Mar 13 15:33:40 crc kubenswrapper[4907]: I0313 15:33:40.966725 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/72846279-1333-4762-b5c9-cd7b25492843-dns-svc\") pod \"dnsmasq-dns-794d85d9cf-xl4f8\" (UID: \"72846279-1333-4762-b5c9-cd7b25492843\") " pod="openstack/dnsmasq-dns-794d85d9cf-xl4f8" Mar 13 15:33:40 crc kubenswrapper[4907]: I0313 15:33:40.966775 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/72846279-1333-4762-b5c9-cd7b25492843-config\") pod \"dnsmasq-dns-794d85d9cf-xl4f8\" (UID: \"72846279-1333-4762-b5c9-cd7b25492843\") " pod="openstack/dnsmasq-dns-794d85d9cf-xl4f8" Mar 13 15:33:40 crc kubenswrapper[4907]: I0313 15:33:40.966933 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/72846279-1333-4762-b5c9-cd7b25492843-ovsdbserver-sb\") pod \"dnsmasq-dns-794d85d9cf-xl4f8\" (UID: \"72846279-1333-4762-b5c9-cd7b25492843\") " pod="openstack/dnsmasq-dns-794d85d9cf-xl4f8" Mar 13 15:33:40 crc kubenswrapper[4907]: I0313 15:33:40.966969 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6jdqp\" (UniqueName: \"kubernetes.io/projected/72846279-1333-4762-b5c9-cd7b25492843-kube-api-access-6jdqp\") pod \"dnsmasq-dns-794d85d9cf-xl4f8\" (UID: \"72846279-1333-4762-b5c9-cd7b25492843\") " pod="openstack/dnsmasq-dns-794d85d9cf-xl4f8" Mar 13 15:33:41 crc kubenswrapper[4907]: I0313 15:33:41.006006 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Mar 13 15:33:41 crc kubenswrapper[4907]: I0313 15:33:41.047854 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Mar 13 15:33:41 crc kubenswrapper[4907]: I0313 15:33:41.067930 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/72846279-1333-4762-b5c9-cd7b25492843-dns-svc\") pod \"dnsmasq-dns-794d85d9cf-xl4f8\" (UID: \"72846279-1333-4762-b5c9-cd7b25492843\") " pod="openstack/dnsmasq-dns-794d85d9cf-xl4f8" Mar 13 15:33:41 crc kubenswrapper[4907]: I0313 15:33:41.068010 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/72846279-1333-4762-b5c9-cd7b25492843-config\") pod \"dnsmasq-dns-794d85d9cf-xl4f8\" (UID: \"72846279-1333-4762-b5c9-cd7b25492843\") " pod="openstack/dnsmasq-dns-794d85d9cf-xl4f8" Mar 13 15:33:41 crc kubenswrapper[4907]: I0313 15:33:41.068863 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/72846279-1333-4762-b5c9-cd7b25492843-dns-svc\") pod \"dnsmasq-dns-794d85d9cf-xl4f8\" (UID: \"72846279-1333-4762-b5c9-cd7b25492843\") " pod="openstack/dnsmasq-dns-794d85d9cf-xl4f8" Mar 13 15:33:41 crc kubenswrapper[4907]: I0313 15:33:41.068863 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/72846279-1333-4762-b5c9-cd7b25492843-config\") pod \"dnsmasq-dns-794d85d9cf-xl4f8\" (UID: \"72846279-1333-4762-b5c9-cd7b25492843\") " pod="openstack/dnsmasq-dns-794d85d9cf-xl4f8" Mar 13 15:33:41 crc kubenswrapper[4907]: I0313 15:33:41.069047 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/72846279-1333-4762-b5c9-cd7b25492843-ovsdbserver-sb\") pod \"dnsmasq-dns-794d85d9cf-xl4f8\" (UID: \"72846279-1333-4762-b5c9-cd7b25492843\") " pod="openstack/dnsmasq-dns-794d85d9cf-xl4f8" Mar 13 15:33:41 crc kubenswrapper[4907]: I0313 15:33:41.069090 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6jdqp\" (UniqueName: \"kubernetes.io/projected/72846279-1333-4762-b5c9-cd7b25492843-kube-api-access-6jdqp\") pod \"dnsmasq-dns-794d85d9cf-xl4f8\" (UID: \"72846279-1333-4762-b5c9-cd7b25492843\") " pod="openstack/dnsmasq-dns-794d85d9cf-xl4f8" Mar 13 15:33:41 crc kubenswrapper[4907]: I0313 15:33:41.069745 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/72846279-1333-4762-b5c9-cd7b25492843-ovsdbserver-sb\") pod \"dnsmasq-dns-794d85d9cf-xl4f8\" (UID: \"72846279-1333-4762-b5c9-cd7b25492843\") " pod="openstack/dnsmasq-dns-794d85d9cf-xl4f8" Mar 13 15:33:41 crc kubenswrapper[4907]: I0313 15:33:41.086140 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6jdqp\" (UniqueName: \"kubernetes.io/projected/72846279-1333-4762-b5c9-cd7b25492843-kube-api-access-6jdqp\") pod \"dnsmasq-dns-794d85d9cf-xl4f8\" (UID: \"72846279-1333-4762-b5c9-cd7b25492843\") " pod="openstack/dnsmasq-dns-794d85d9cf-xl4f8" Mar 13 15:33:41 crc kubenswrapper[4907]: I0313 15:33:41.168655 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-794d85d9cf-xl4f8" Mar 13 15:33:41 crc kubenswrapper[4907]: I0313 15:33:41.369465 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-794d85d9cf-xl4f8"] Mar 13 15:33:41 crc kubenswrapper[4907]: I0313 15:33:41.408798 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-65b7b7b995-wl96j"] Mar 13 15:33:41 crc kubenswrapper[4907]: I0313 15:33:41.419310 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-65b7b7b995-wl96j" Mar 13 15:33:41 crc kubenswrapper[4907]: I0313 15:33:41.424764 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Mar 13 15:33:41 crc kubenswrapper[4907]: I0313 15:33:41.448326 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-65b7b7b995-wl96j"] Mar 13 15:33:41 crc kubenswrapper[4907]: I0313 15:33:41.476522 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/49e01f82-fb82-4351-a637-3b885e663e0e-dns-svc\") pod \"dnsmasq-dns-65b7b7b995-wl96j\" (UID: \"49e01f82-fb82-4351-a637-3b885e663e0e\") " pod="openstack/dnsmasq-dns-65b7b7b995-wl96j" Mar 13 15:33:41 crc kubenswrapper[4907]: I0313 15:33:41.476579 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/49e01f82-fb82-4351-a637-3b885e663e0e-ovsdbserver-sb\") pod \"dnsmasq-dns-65b7b7b995-wl96j\" (UID: \"49e01f82-fb82-4351-a637-3b885e663e0e\") " pod="openstack/dnsmasq-dns-65b7b7b995-wl96j" Mar 13 15:33:41 crc kubenswrapper[4907]: I0313 15:33:41.476726 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/49e01f82-fb82-4351-a637-3b885e663e0e-ovsdbserver-nb\") pod \"dnsmasq-dns-65b7b7b995-wl96j\" (UID: \"49e01f82-fb82-4351-a637-3b885e663e0e\") " pod="openstack/dnsmasq-dns-65b7b7b995-wl96j" Mar 13 15:33:41 crc kubenswrapper[4907]: I0313 15:33:41.476825 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lk79c\" (UniqueName: \"kubernetes.io/projected/49e01f82-fb82-4351-a637-3b885e663e0e-kube-api-access-lk79c\") pod \"dnsmasq-dns-65b7b7b995-wl96j\" (UID: \"49e01f82-fb82-4351-a637-3b885e663e0e\") " pod="openstack/dnsmasq-dns-65b7b7b995-wl96j" Mar 13 15:33:41 crc kubenswrapper[4907]: I0313 15:33:41.476872 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/49e01f82-fb82-4351-a637-3b885e663e0e-config\") pod \"dnsmasq-dns-65b7b7b995-wl96j\" (UID: \"49e01f82-fb82-4351-a637-3b885e663e0e\") " pod="openstack/dnsmasq-dns-65b7b7b995-wl96j" Mar 13 15:33:41 crc kubenswrapper[4907]: I0313 15:33:41.578049 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/49e01f82-fb82-4351-a637-3b885e663e0e-ovsdbserver-nb\") pod \"dnsmasq-dns-65b7b7b995-wl96j\" (UID: \"49e01f82-fb82-4351-a637-3b885e663e0e\") " pod="openstack/dnsmasq-dns-65b7b7b995-wl96j" Mar 13 15:33:41 crc kubenswrapper[4907]: I0313 15:33:41.578125 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lk79c\" (UniqueName: \"kubernetes.io/projected/49e01f82-fb82-4351-a637-3b885e663e0e-kube-api-access-lk79c\") pod \"dnsmasq-dns-65b7b7b995-wl96j\" (UID: \"49e01f82-fb82-4351-a637-3b885e663e0e\") " pod="openstack/dnsmasq-dns-65b7b7b995-wl96j" Mar 13 15:33:41 crc kubenswrapper[4907]: I0313 15:33:41.578158 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/49e01f82-fb82-4351-a637-3b885e663e0e-config\") pod \"dnsmasq-dns-65b7b7b995-wl96j\" (UID: \"49e01f82-fb82-4351-a637-3b885e663e0e\") " pod="openstack/dnsmasq-dns-65b7b7b995-wl96j" Mar 13 15:33:41 crc kubenswrapper[4907]: I0313 15:33:41.578220 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/49e01f82-fb82-4351-a637-3b885e663e0e-dns-svc\") pod \"dnsmasq-dns-65b7b7b995-wl96j\" (UID: \"49e01f82-fb82-4351-a637-3b885e663e0e\") " pod="openstack/dnsmasq-dns-65b7b7b995-wl96j" Mar 13 15:33:41 crc kubenswrapper[4907]: I0313 15:33:41.578270 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/49e01f82-fb82-4351-a637-3b885e663e0e-ovsdbserver-sb\") pod \"dnsmasq-dns-65b7b7b995-wl96j\" (UID: \"49e01f82-fb82-4351-a637-3b885e663e0e\") " pod="openstack/dnsmasq-dns-65b7b7b995-wl96j" Mar 13 15:33:41 crc kubenswrapper[4907]: I0313 15:33:41.578998 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/49e01f82-fb82-4351-a637-3b885e663e0e-ovsdbserver-nb\") pod \"dnsmasq-dns-65b7b7b995-wl96j\" (UID: \"49e01f82-fb82-4351-a637-3b885e663e0e\") " pod="openstack/dnsmasq-dns-65b7b7b995-wl96j" Mar 13 15:33:41 crc kubenswrapper[4907]: I0313 15:33:41.579176 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/49e01f82-fb82-4351-a637-3b885e663e0e-ovsdbserver-sb\") pod \"dnsmasq-dns-65b7b7b995-wl96j\" (UID: \"49e01f82-fb82-4351-a637-3b885e663e0e\") " pod="openstack/dnsmasq-dns-65b7b7b995-wl96j" Mar 13 15:33:41 crc kubenswrapper[4907]: I0313 15:33:41.579344 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/49e01f82-fb82-4351-a637-3b885e663e0e-config\") pod \"dnsmasq-dns-65b7b7b995-wl96j\" (UID: \"49e01f82-fb82-4351-a637-3b885e663e0e\") " pod="openstack/dnsmasq-dns-65b7b7b995-wl96j" Mar 13 15:33:41 crc kubenswrapper[4907]: I0313 15:33:41.579627 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/49e01f82-fb82-4351-a637-3b885e663e0e-dns-svc\") pod \"dnsmasq-dns-65b7b7b995-wl96j\" (UID: \"49e01f82-fb82-4351-a637-3b885e663e0e\") " pod="openstack/dnsmasq-dns-65b7b7b995-wl96j" Mar 13 15:33:41 crc kubenswrapper[4907]: I0313 15:33:41.596987 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lk79c\" (UniqueName: \"kubernetes.io/projected/49e01f82-fb82-4351-a637-3b885e663e0e-kube-api-access-lk79c\") pod \"dnsmasq-dns-65b7b7b995-wl96j\" (UID: \"49e01f82-fb82-4351-a637-3b885e663e0e\") " pod="openstack/dnsmasq-dns-65b7b7b995-wl96j" Mar 13 15:33:41 crc kubenswrapper[4907]: I0313 15:33:41.665181 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-794d85d9cf-xl4f8"] Mar 13 15:33:41 crc kubenswrapper[4907]: I0313 15:33:41.746528 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-65b7b7b995-wl96j" Mar 13 15:33:41 crc kubenswrapper[4907]: I0313 15:33:41.924240 4907 generic.go:334] "Generic (PLEG): container finished" podID="72846279-1333-4762-b5c9-cd7b25492843" containerID="b312e10aee1a09c0f414d5ab55a14272c1683d9e1cf46fbb198ff6477ba0ca4f" exitCode=0 Mar 13 15:33:41 crc kubenswrapper[4907]: I0313 15:33:41.924392 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-794d85d9cf-xl4f8" event={"ID":"72846279-1333-4762-b5c9-cd7b25492843","Type":"ContainerDied","Data":"b312e10aee1a09c0f414d5ab55a14272c1683d9e1cf46fbb198ff6477ba0ca4f"} Mar 13 15:33:41 crc kubenswrapper[4907]: I0313 15:33:41.924657 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-794d85d9cf-xl4f8" event={"ID":"72846279-1333-4762-b5c9-cd7b25492843","Type":"ContainerStarted","Data":"847c3ddce6826b442d1b5ec210f83bfadfb60c76eab28f422fd1cc1f341d21d1"} Mar 13 15:33:42 crc kubenswrapper[4907]: I0313 15:33:42.210763 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-794d85d9cf-xl4f8" Mar 13 15:33:42 crc kubenswrapper[4907]: I0313 15:33:42.212384 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-65b7b7b995-wl96j"] Mar 13 15:33:42 crc kubenswrapper[4907]: W0313 15:33:42.217069 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod49e01f82_fb82_4351_a637_3b885e663e0e.slice/crio-2cc3bc2c8a0a2c514a94d228c418e25f81f45826dc803a6c0203288a2ddf2101 WatchSource:0}: Error finding container 2cc3bc2c8a0a2c514a94d228c418e25f81f45826dc803a6c0203288a2ddf2101: Status 404 returned error can't find the container with id 2cc3bc2c8a0a2c514a94d228c418e25f81f45826dc803a6c0203288a2ddf2101 Mar 13 15:33:42 crc kubenswrapper[4907]: I0313 15:33:42.388890 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6jdqp\" (UniqueName: \"kubernetes.io/projected/72846279-1333-4762-b5c9-cd7b25492843-kube-api-access-6jdqp\") pod \"72846279-1333-4762-b5c9-cd7b25492843\" (UID: \"72846279-1333-4762-b5c9-cd7b25492843\") " Mar 13 15:33:42 crc kubenswrapper[4907]: I0313 15:33:42.388954 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/72846279-1333-4762-b5c9-cd7b25492843-ovsdbserver-sb\") pod \"72846279-1333-4762-b5c9-cd7b25492843\" (UID: \"72846279-1333-4762-b5c9-cd7b25492843\") " Mar 13 15:33:42 crc kubenswrapper[4907]: I0313 15:33:42.389331 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/72846279-1333-4762-b5c9-cd7b25492843-config\") pod \"72846279-1333-4762-b5c9-cd7b25492843\" (UID: \"72846279-1333-4762-b5c9-cd7b25492843\") " Mar 13 15:33:42 crc kubenswrapper[4907]: I0313 15:33:42.389386 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/72846279-1333-4762-b5c9-cd7b25492843-dns-svc\") pod \"72846279-1333-4762-b5c9-cd7b25492843\" (UID: \"72846279-1333-4762-b5c9-cd7b25492843\") " Mar 13 15:33:42 crc kubenswrapper[4907]: I0313 15:33:42.396374 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/72846279-1333-4762-b5c9-cd7b25492843-kube-api-access-6jdqp" (OuterVolumeSpecName: "kube-api-access-6jdqp") pod "72846279-1333-4762-b5c9-cd7b25492843" (UID: "72846279-1333-4762-b5c9-cd7b25492843"). InnerVolumeSpecName "kube-api-access-6jdqp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:33:42 crc kubenswrapper[4907]: I0313 15:33:42.412176 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/72846279-1333-4762-b5c9-cd7b25492843-config" (OuterVolumeSpecName: "config") pod "72846279-1333-4762-b5c9-cd7b25492843" (UID: "72846279-1333-4762-b5c9-cd7b25492843"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:33:42 crc kubenswrapper[4907]: I0313 15:33:42.420654 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/72846279-1333-4762-b5c9-cd7b25492843-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "72846279-1333-4762-b5c9-cd7b25492843" (UID: "72846279-1333-4762-b5c9-cd7b25492843"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:33:42 crc kubenswrapper[4907]: I0313 15:33:42.420767 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/72846279-1333-4762-b5c9-cd7b25492843-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "72846279-1333-4762-b5c9-cd7b25492843" (UID: "72846279-1333-4762-b5c9-cd7b25492843"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:33:42 crc kubenswrapper[4907]: I0313 15:33:42.492070 4907 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/72846279-1333-4762-b5c9-cd7b25492843-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 13 15:33:42 crc kubenswrapper[4907]: I0313 15:33:42.492407 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6jdqp\" (UniqueName: \"kubernetes.io/projected/72846279-1333-4762-b5c9-cd7b25492843-kube-api-access-6jdqp\") on node \"crc\" DevicePath \"\"" Mar 13 15:33:42 crc kubenswrapper[4907]: I0313 15:33:42.492478 4907 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/72846279-1333-4762-b5c9-cd7b25492843-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 13 15:33:42 crc kubenswrapper[4907]: I0313 15:33:42.492539 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/72846279-1333-4762-b5c9-cd7b25492843-config\") on node \"crc\" DevicePath \"\"" Mar 13 15:33:42 crc kubenswrapper[4907]: I0313 15:33:42.933568 4907 generic.go:334] "Generic (PLEG): container finished" podID="49e01f82-fb82-4351-a637-3b885e663e0e" containerID="7829d0510750d5cae8df3422b54b042bd33e40f132a8e3e16f6eec0a8952963b" exitCode=0 Mar 13 15:33:42 crc kubenswrapper[4907]: I0313 15:33:42.933638 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-65b7b7b995-wl96j" event={"ID":"49e01f82-fb82-4351-a637-3b885e663e0e","Type":"ContainerDied","Data":"7829d0510750d5cae8df3422b54b042bd33e40f132a8e3e16f6eec0a8952963b"} Mar 13 15:33:42 crc kubenswrapper[4907]: I0313 15:33:42.933666 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-65b7b7b995-wl96j" event={"ID":"49e01f82-fb82-4351-a637-3b885e663e0e","Type":"ContainerStarted","Data":"2cc3bc2c8a0a2c514a94d228c418e25f81f45826dc803a6c0203288a2ddf2101"} Mar 13 15:33:42 crc kubenswrapper[4907]: I0313 15:33:42.935139 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-794d85d9cf-xl4f8" event={"ID":"72846279-1333-4762-b5c9-cd7b25492843","Type":"ContainerDied","Data":"847c3ddce6826b442d1b5ec210f83bfadfb60c76eab28f422fd1cc1f341d21d1"} Mar 13 15:33:42 crc kubenswrapper[4907]: I0313 15:33:42.935178 4907 scope.go:117] "RemoveContainer" containerID="b312e10aee1a09c0f414d5ab55a14272c1683d9e1cf46fbb198ff6477ba0ca4f" Mar 13 15:33:42 crc kubenswrapper[4907]: I0313 15:33:42.935186 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-794d85d9cf-xl4f8" Mar 13 15:33:43 crc kubenswrapper[4907]: I0313 15:33:43.006986 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-794d85d9cf-xl4f8"] Mar 13 15:33:43 crc kubenswrapper[4907]: I0313 15:33:43.022343 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-794d85d9cf-xl4f8"] Mar 13 15:33:43 crc kubenswrapper[4907]: I0313 15:33:43.795158 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="72846279-1333-4762-b5c9-cd7b25492843" path="/var/lib/kubelet/pods/72846279-1333-4762-b5c9-cd7b25492843/volumes" Mar 13 15:33:43 crc kubenswrapper[4907]: I0313 15:33:43.916598 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-copy-data"] Mar 13 15:33:43 crc kubenswrapper[4907]: E0313 15:33:43.916936 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72846279-1333-4762-b5c9-cd7b25492843" containerName="init" Mar 13 15:33:43 crc kubenswrapper[4907]: I0313 15:33:43.916952 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="72846279-1333-4762-b5c9-cd7b25492843" containerName="init" Mar 13 15:33:43 crc kubenswrapper[4907]: I0313 15:33:43.917292 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="72846279-1333-4762-b5c9-cd7b25492843" containerName="init" Mar 13 15:33:43 crc kubenswrapper[4907]: I0313 15:33:43.918543 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-copy-data" Mar 13 15:33:43 crc kubenswrapper[4907]: I0313 15:33:43.921165 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovn-data-cert" Mar 13 15:33:43 crc kubenswrapper[4907]: I0313 15:33:43.926942 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-copy-data"] Mar 13 15:33:43 crc kubenswrapper[4907]: I0313 15:33:43.952182 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-65b7b7b995-wl96j" event={"ID":"49e01f82-fb82-4351-a637-3b885e663e0e","Type":"ContainerStarted","Data":"a3109c92b6a4f47975f53b04558486733a8cd9a2cb768f638be94b37e0295965"} Mar 13 15:33:43 crc kubenswrapper[4907]: I0313 15:33:43.952321 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-65b7b7b995-wl96j" Mar 13 15:33:43 crc kubenswrapper[4907]: I0313 15:33:43.975163 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-65b7b7b995-wl96j" podStartSLOduration=2.975146467 podStartE2EDuration="2.975146467s" podCreationTimestamp="2026-03-13 15:33:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:33:43.969126812 +0000 UTC m=+5322.868914551" watchObservedRunningTime="2026-03-13 15:33:43.975146467 +0000 UTC m=+5322.874934156" Mar 13 15:33:44 crc kubenswrapper[4907]: I0313 15:33:44.014539 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-1a945566-190e-43e6-ae94-58c483d5c2e4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1a945566-190e-43e6-ae94-58c483d5c2e4\") pod \"ovn-copy-data\" (UID: \"e0c0f09a-8646-4326-86e8-5bd9123d78ca\") " pod="openstack/ovn-copy-data" Mar 13 15:33:44 crc kubenswrapper[4907]: I0313 15:33:44.014589 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ctpt2\" (UniqueName: \"kubernetes.io/projected/e0c0f09a-8646-4326-86e8-5bd9123d78ca-kube-api-access-ctpt2\") pod \"ovn-copy-data\" (UID: \"e0c0f09a-8646-4326-86e8-5bd9123d78ca\") " pod="openstack/ovn-copy-data" Mar 13 15:33:44 crc kubenswrapper[4907]: I0313 15:33:44.014659 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/e0c0f09a-8646-4326-86e8-5bd9123d78ca-ovn-data-cert\") pod \"ovn-copy-data\" (UID: \"e0c0f09a-8646-4326-86e8-5bd9123d78ca\") " pod="openstack/ovn-copy-data" Mar 13 15:33:44 crc kubenswrapper[4907]: I0313 15:33:44.117091 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/e0c0f09a-8646-4326-86e8-5bd9123d78ca-ovn-data-cert\") pod \"ovn-copy-data\" (UID: \"e0c0f09a-8646-4326-86e8-5bd9123d78ca\") " pod="openstack/ovn-copy-data" Mar 13 15:33:44 crc kubenswrapper[4907]: I0313 15:33:44.117234 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-1a945566-190e-43e6-ae94-58c483d5c2e4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1a945566-190e-43e6-ae94-58c483d5c2e4\") pod \"ovn-copy-data\" (UID: \"e0c0f09a-8646-4326-86e8-5bd9123d78ca\") " pod="openstack/ovn-copy-data" Mar 13 15:33:44 crc kubenswrapper[4907]: I0313 15:33:44.117268 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ctpt2\" (UniqueName: \"kubernetes.io/projected/e0c0f09a-8646-4326-86e8-5bd9123d78ca-kube-api-access-ctpt2\") pod \"ovn-copy-data\" (UID: \"e0c0f09a-8646-4326-86e8-5bd9123d78ca\") " pod="openstack/ovn-copy-data" Mar 13 15:33:44 crc kubenswrapper[4907]: I0313 15:33:44.121535 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/e0c0f09a-8646-4326-86e8-5bd9123d78ca-ovn-data-cert\") pod \"ovn-copy-data\" (UID: \"e0c0f09a-8646-4326-86e8-5bd9123d78ca\") " pod="openstack/ovn-copy-data" Mar 13 15:33:44 crc kubenswrapper[4907]: I0313 15:33:44.122308 4907 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 13 15:33:44 crc kubenswrapper[4907]: I0313 15:33:44.122344 4907 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-1a945566-190e-43e6-ae94-58c483d5c2e4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1a945566-190e-43e6-ae94-58c483d5c2e4\") pod \"ovn-copy-data\" (UID: \"e0c0f09a-8646-4326-86e8-5bd9123d78ca\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/5d7cb02ae805cd0ef340a8d0e965e05c32a025a0afcab88074ddc006f8992ebd/globalmount\"" pod="openstack/ovn-copy-data" Mar 13 15:33:44 crc kubenswrapper[4907]: I0313 15:33:44.140377 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ctpt2\" (UniqueName: \"kubernetes.io/projected/e0c0f09a-8646-4326-86e8-5bd9123d78ca-kube-api-access-ctpt2\") pod \"ovn-copy-data\" (UID: \"e0c0f09a-8646-4326-86e8-5bd9123d78ca\") " pod="openstack/ovn-copy-data" Mar 13 15:33:44 crc kubenswrapper[4907]: I0313 15:33:44.149971 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-1a945566-190e-43e6-ae94-58c483d5c2e4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1a945566-190e-43e6-ae94-58c483d5c2e4\") pod \"ovn-copy-data\" (UID: \"e0c0f09a-8646-4326-86e8-5bd9123d78ca\") " pod="openstack/ovn-copy-data" Mar 13 15:33:44 crc kubenswrapper[4907]: I0313 15:33:44.238162 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-copy-data" Mar 13 15:33:44 crc kubenswrapper[4907]: I0313 15:33:44.709764 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-copy-data"] Mar 13 15:33:44 crc kubenswrapper[4907]: W0313 15:33:44.717765 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode0c0f09a_8646_4326_86e8_5bd9123d78ca.slice/crio-afb18a83c45725fa475076eb88bc1f59d47ff74d8c364938ed9892b1ba42b1be WatchSource:0}: Error finding container afb18a83c45725fa475076eb88bc1f59d47ff74d8c364938ed9892b1ba42b1be: Status 404 returned error can't find the container with id afb18a83c45725fa475076eb88bc1f59d47ff74d8c364938ed9892b1ba42b1be Mar 13 15:33:44 crc kubenswrapper[4907]: I0313 15:33:44.961388 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-copy-data" event={"ID":"e0c0f09a-8646-4326-86e8-5bd9123d78ca","Type":"ContainerStarted","Data":"afb18a83c45725fa475076eb88bc1f59d47ff74d8c364938ed9892b1ba42b1be"} Mar 13 15:33:45 crc kubenswrapper[4907]: I0313 15:33:45.973577 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-copy-data" event={"ID":"e0c0f09a-8646-4326-86e8-5bd9123d78ca","Type":"ContainerStarted","Data":"d115558c971717b0e8dd16d1b2ccb9b5004c901031eda65c32db4372508a5b7c"} Mar 13 15:33:45 crc kubenswrapper[4907]: I0313 15:33:45.994600 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-copy-data" podStartSLOduration=3.45972068 podStartE2EDuration="3.99457321s" podCreationTimestamp="2026-03-13 15:33:42 +0000 UTC" firstStartedPulling="2026-03-13 15:33:44.720876628 +0000 UTC m=+5323.620664317" lastFinishedPulling="2026-03-13 15:33:45.255729158 +0000 UTC m=+5324.155516847" observedRunningTime="2026-03-13 15:33:45.993075579 +0000 UTC m=+5324.892863288" watchObservedRunningTime="2026-03-13 15:33:45.99457321 +0000 UTC m=+5324.894360929" Mar 13 15:33:51 crc kubenswrapper[4907]: I0313 15:33:51.058444 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Mar 13 15:33:51 crc kubenswrapper[4907]: I0313 15:33:51.073689 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Mar 13 15:33:51 crc kubenswrapper[4907]: I0313 15:33:51.079756 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Mar 13 15:33:51 crc kubenswrapper[4907]: I0313 15:33:51.082647 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-rgmnm" Mar 13 15:33:51 crc kubenswrapper[4907]: I0313 15:33:51.082708 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Mar 13 15:33:51 crc kubenswrapper[4907]: I0313 15:33:51.082871 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Mar 13 15:33:51 crc kubenswrapper[4907]: I0313 15:33:51.134525 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b95d939-59f1-4ba9-b39a-eb255ab7aa77-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"1b95d939-59f1-4ba9-b39a-eb255ab7aa77\") " pod="openstack/ovn-northd-0" Mar 13 15:33:51 crc kubenswrapper[4907]: I0313 15:33:51.134582 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cf6kz\" (UniqueName: \"kubernetes.io/projected/1b95d939-59f1-4ba9-b39a-eb255ab7aa77-kube-api-access-cf6kz\") pod \"ovn-northd-0\" (UID: \"1b95d939-59f1-4ba9-b39a-eb255ab7aa77\") " pod="openstack/ovn-northd-0" Mar 13 15:33:51 crc kubenswrapper[4907]: I0313 15:33:51.134634 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/1b95d939-59f1-4ba9-b39a-eb255ab7aa77-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"1b95d939-59f1-4ba9-b39a-eb255ab7aa77\") " pod="openstack/ovn-northd-0" Mar 13 15:33:51 crc kubenswrapper[4907]: I0313 15:33:51.134653 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1b95d939-59f1-4ba9-b39a-eb255ab7aa77-config\") pod \"ovn-northd-0\" (UID: \"1b95d939-59f1-4ba9-b39a-eb255ab7aa77\") " pod="openstack/ovn-northd-0" Mar 13 15:33:51 crc kubenswrapper[4907]: I0313 15:33:51.134772 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1b95d939-59f1-4ba9-b39a-eb255ab7aa77-scripts\") pod \"ovn-northd-0\" (UID: \"1b95d939-59f1-4ba9-b39a-eb255ab7aa77\") " pod="openstack/ovn-northd-0" Mar 13 15:33:51 crc kubenswrapper[4907]: I0313 15:33:51.236289 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cf6kz\" (UniqueName: \"kubernetes.io/projected/1b95d939-59f1-4ba9-b39a-eb255ab7aa77-kube-api-access-cf6kz\") pod \"ovn-northd-0\" (UID: \"1b95d939-59f1-4ba9-b39a-eb255ab7aa77\") " pod="openstack/ovn-northd-0" Mar 13 15:33:51 crc kubenswrapper[4907]: I0313 15:33:51.236382 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/1b95d939-59f1-4ba9-b39a-eb255ab7aa77-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"1b95d939-59f1-4ba9-b39a-eb255ab7aa77\") " pod="openstack/ovn-northd-0" Mar 13 15:33:51 crc kubenswrapper[4907]: I0313 15:33:51.236410 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1b95d939-59f1-4ba9-b39a-eb255ab7aa77-config\") pod \"ovn-northd-0\" (UID: \"1b95d939-59f1-4ba9-b39a-eb255ab7aa77\") " pod="openstack/ovn-northd-0" Mar 13 15:33:51 crc kubenswrapper[4907]: I0313 15:33:51.236472 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1b95d939-59f1-4ba9-b39a-eb255ab7aa77-scripts\") pod \"ovn-northd-0\" (UID: \"1b95d939-59f1-4ba9-b39a-eb255ab7aa77\") " pod="openstack/ovn-northd-0" Mar 13 15:33:51 crc kubenswrapper[4907]: I0313 15:33:51.236564 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b95d939-59f1-4ba9-b39a-eb255ab7aa77-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"1b95d939-59f1-4ba9-b39a-eb255ab7aa77\") " pod="openstack/ovn-northd-0" Mar 13 15:33:51 crc kubenswrapper[4907]: I0313 15:33:51.236990 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/1b95d939-59f1-4ba9-b39a-eb255ab7aa77-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"1b95d939-59f1-4ba9-b39a-eb255ab7aa77\") " pod="openstack/ovn-northd-0" Mar 13 15:33:51 crc kubenswrapper[4907]: I0313 15:33:51.237358 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1b95d939-59f1-4ba9-b39a-eb255ab7aa77-scripts\") pod \"ovn-northd-0\" (UID: \"1b95d939-59f1-4ba9-b39a-eb255ab7aa77\") " pod="openstack/ovn-northd-0" Mar 13 15:33:51 crc kubenswrapper[4907]: I0313 15:33:51.237485 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1b95d939-59f1-4ba9-b39a-eb255ab7aa77-config\") pod \"ovn-northd-0\" (UID: \"1b95d939-59f1-4ba9-b39a-eb255ab7aa77\") " pod="openstack/ovn-northd-0" Mar 13 15:33:51 crc kubenswrapper[4907]: I0313 15:33:51.242589 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b95d939-59f1-4ba9-b39a-eb255ab7aa77-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"1b95d939-59f1-4ba9-b39a-eb255ab7aa77\") " pod="openstack/ovn-northd-0" Mar 13 15:33:51 crc kubenswrapper[4907]: I0313 15:33:51.253025 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cf6kz\" (UniqueName: \"kubernetes.io/projected/1b95d939-59f1-4ba9-b39a-eb255ab7aa77-kube-api-access-cf6kz\") pod \"ovn-northd-0\" (UID: \"1b95d939-59f1-4ba9-b39a-eb255ab7aa77\") " pod="openstack/ovn-northd-0" Mar 13 15:33:51 crc kubenswrapper[4907]: I0313 15:33:51.405937 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Mar 13 15:33:51 crc kubenswrapper[4907]: I0313 15:33:51.748126 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-65b7b7b995-wl96j" Mar 13 15:33:51 crc kubenswrapper[4907]: I0313 15:33:51.849792 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-684c864bc9-p4j95"] Mar 13 15:33:51 crc kubenswrapper[4907]: I0313 15:33:51.850024 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-684c864bc9-p4j95" podUID="6c367cd2-6ece-429d-a930-c1ec11175248" containerName="dnsmasq-dns" containerID="cri-o://1b6e2c3bc8a46bfa72edd14d9923048107755f57239082f4595ce1f2d9b0f3a2" gracePeriod=10 Mar 13 15:33:51 crc kubenswrapper[4907]: I0313 15:33:51.890633 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Mar 13 15:33:51 crc kubenswrapper[4907]: W0313 15:33:51.944474 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1b95d939_59f1_4ba9_b39a_eb255ab7aa77.slice/crio-f1360a96df510c54ad081cf7e101eb797e728625968b75461bc0e672bb69bda4 WatchSource:0}: Error finding container f1360a96df510c54ad081cf7e101eb797e728625968b75461bc0e672bb69bda4: Status 404 returned error can't find the container with id f1360a96df510c54ad081cf7e101eb797e728625968b75461bc0e672bb69bda4 Mar 13 15:33:52 crc kubenswrapper[4907]: I0313 15:33:52.024711 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"1b95d939-59f1-4ba9-b39a-eb255ab7aa77","Type":"ContainerStarted","Data":"f1360a96df510c54ad081cf7e101eb797e728625968b75461bc0e672bb69bda4"} Mar 13 15:33:52 crc kubenswrapper[4907]: I0313 15:33:52.026413 4907 generic.go:334] "Generic (PLEG): container finished" podID="6c367cd2-6ece-429d-a930-c1ec11175248" containerID="1b6e2c3bc8a46bfa72edd14d9923048107755f57239082f4595ce1f2d9b0f3a2" exitCode=0 Mar 13 15:33:52 crc kubenswrapper[4907]: I0313 15:33:52.026449 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-684c864bc9-p4j95" event={"ID":"6c367cd2-6ece-429d-a930-c1ec11175248","Type":"ContainerDied","Data":"1b6e2c3bc8a46bfa72edd14d9923048107755f57239082f4595ce1f2d9b0f3a2"} Mar 13 15:33:52 crc kubenswrapper[4907]: I0313 15:33:52.252291 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-684c864bc9-p4j95" Mar 13 15:33:52 crc kubenswrapper[4907]: I0313 15:33:52.368616 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6c367cd2-6ece-429d-a930-c1ec11175248-config\") pod \"6c367cd2-6ece-429d-a930-c1ec11175248\" (UID: \"6c367cd2-6ece-429d-a930-c1ec11175248\") " Mar 13 15:33:52 crc kubenswrapper[4907]: I0313 15:33:52.368703 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rhz6p\" (UniqueName: \"kubernetes.io/projected/6c367cd2-6ece-429d-a930-c1ec11175248-kube-api-access-rhz6p\") pod \"6c367cd2-6ece-429d-a930-c1ec11175248\" (UID: \"6c367cd2-6ece-429d-a930-c1ec11175248\") " Mar 13 15:33:52 crc kubenswrapper[4907]: I0313 15:33:52.368803 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6c367cd2-6ece-429d-a930-c1ec11175248-dns-svc\") pod \"6c367cd2-6ece-429d-a930-c1ec11175248\" (UID: \"6c367cd2-6ece-429d-a930-c1ec11175248\") " Mar 13 15:33:52 crc kubenswrapper[4907]: I0313 15:33:52.373478 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6c367cd2-6ece-429d-a930-c1ec11175248-kube-api-access-rhz6p" (OuterVolumeSpecName: "kube-api-access-rhz6p") pod "6c367cd2-6ece-429d-a930-c1ec11175248" (UID: "6c367cd2-6ece-429d-a930-c1ec11175248"). InnerVolumeSpecName "kube-api-access-rhz6p". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:33:52 crc kubenswrapper[4907]: I0313 15:33:52.401621 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6c367cd2-6ece-429d-a930-c1ec11175248-config" (OuterVolumeSpecName: "config") pod "6c367cd2-6ece-429d-a930-c1ec11175248" (UID: "6c367cd2-6ece-429d-a930-c1ec11175248"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:33:52 crc kubenswrapper[4907]: I0313 15:33:52.406746 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6c367cd2-6ece-429d-a930-c1ec11175248-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "6c367cd2-6ece-429d-a930-c1ec11175248" (UID: "6c367cd2-6ece-429d-a930-c1ec11175248"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:33:52 crc kubenswrapper[4907]: I0313 15:33:52.470277 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rhz6p\" (UniqueName: \"kubernetes.io/projected/6c367cd2-6ece-429d-a930-c1ec11175248-kube-api-access-rhz6p\") on node \"crc\" DevicePath \"\"" Mar 13 15:33:52 crc kubenswrapper[4907]: I0313 15:33:52.470306 4907 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6c367cd2-6ece-429d-a930-c1ec11175248-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 13 15:33:52 crc kubenswrapper[4907]: I0313 15:33:52.470315 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6c367cd2-6ece-429d-a930-c1ec11175248-config\") on node \"crc\" DevicePath \"\"" Mar 13 15:33:53 crc kubenswrapper[4907]: I0313 15:33:53.042419 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-684c864bc9-p4j95" event={"ID":"6c367cd2-6ece-429d-a930-c1ec11175248","Type":"ContainerDied","Data":"bc53c9c5cde7280652faad4d880d4ee4a2f7625ff891a79d0f242d66a55b312a"} Mar 13 15:33:53 crc kubenswrapper[4907]: I0313 15:33:53.042474 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-684c864bc9-p4j95" Mar 13 15:33:53 crc kubenswrapper[4907]: I0313 15:33:53.042482 4907 scope.go:117] "RemoveContainer" containerID="1b6e2c3bc8a46bfa72edd14d9923048107755f57239082f4595ce1f2d9b0f3a2" Mar 13 15:33:53 crc kubenswrapper[4907]: I0313 15:33:53.045529 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"1b95d939-59f1-4ba9-b39a-eb255ab7aa77","Type":"ContainerStarted","Data":"4e605736636de74c497de794d5fba08d765bee51520f0a7d4c6c7d3da9d0341f"} Mar 13 15:33:53 crc kubenswrapper[4907]: I0313 15:33:53.045562 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"1b95d939-59f1-4ba9-b39a-eb255ab7aa77","Type":"ContainerStarted","Data":"1b1a680cf63cb8b00e032e250e346fc2a5d3bfc5566f746214754e555cb13ce5"} Mar 13 15:33:53 crc kubenswrapper[4907]: I0313 15:33:53.046353 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Mar 13 15:33:53 crc kubenswrapper[4907]: I0313 15:33:53.060408 4907 scope.go:117] "RemoveContainer" containerID="dfbeef34d4d9c1c564438050b366f2c5cc11365a8117cb19955310b23bc22e05" Mar 13 15:33:53 crc kubenswrapper[4907]: I0313 15:33:53.070725 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=2.07070575 podStartE2EDuration="2.07070575s" podCreationTimestamp="2026-03-13 15:33:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:33:53.063276047 +0000 UTC m=+5331.963063776" watchObservedRunningTime="2026-03-13 15:33:53.07070575 +0000 UTC m=+5331.970493439" Mar 13 15:33:53 crc kubenswrapper[4907]: I0313 15:33:53.107293 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-684c864bc9-p4j95"] Mar 13 15:33:53 crc kubenswrapper[4907]: I0313 15:33:53.119532 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-684c864bc9-p4j95"] Mar 13 15:33:53 crc kubenswrapper[4907]: I0313 15:33:53.790959 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6c367cd2-6ece-429d-a930-c1ec11175248" path="/var/lib/kubelet/pods/6c367cd2-6ece-429d-a930-c1ec11175248/volumes" Mar 13 15:33:55 crc kubenswrapper[4907]: I0313 15:33:55.782956 4907 scope.go:117] "RemoveContainer" containerID="592bf9f175f440e943369aab2abbb13ad2c26412cb0e0738c4a33a156b8a5596" Mar 13 15:33:55 crc kubenswrapper[4907]: E0313 15:33:55.783422 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:33:56 crc kubenswrapper[4907]: I0313 15:33:56.652691 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-gxjv5"] Mar 13 15:33:56 crc kubenswrapper[4907]: E0313 15:33:56.653405 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c367cd2-6ece-429d-a930-c1ec11175248" containerName="init" Mar 13 15:33:56 crc kubenswrapper[4907]: I0313 15:33:56.653434 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c367cd2-6ece-429d-a930-c1ec11175248" containerName="init" Mar 13 15:33:56 crc kubenswrapper[4907]: E0313 15:33:56.653463 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c367cd2-6ece-429d-a930-c1ec11175248" containerName="dnsmasq-dns" Mar 13 15:33:56 crc kubenswrapper[4907]: I0313 15:33:56.653472 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c367cd2-6ece-429d-a930-c1ec11175248" containerName="dnsmasq-dns" Mar 13 15:33:56 crc kubenswrapper[4907]: I0313 15:33:56.653690 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="6c367cd2-6ece-429d-a930-c1ec11175248" containerName="dnsmasq-dns" Mar 13 15:33:56 crc kubenswrapper[4907]: I0313 15:33:56.654306 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-gxjv5" Mar 13 15:33:56 crc kubenswrapper[4907]: I0313 15:33:56.663960 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-gxjv5"] Mar 13 15:33:56 crc kubenswrapper[4907]: I0313 15:33:56.739683 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9c51cd5f-cfaf-4c1a-9aa2-e7ade1b79058-operator-scripts\") pod \"keystone-db-create-gxjv5\" (UID: \"9c51cd5f-cfaf-4c1a-9aa2-e7ade1b79058\") " pod="openstack/keystone-db-create-gxjv5" Mar 13 15:33:56 crc kubenswrapper[4907]: I0313 15:33:56.739863 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q45f8\" (UniqueName: \"kubernetes.io/projected/9c51cd5f-cfaf-4c1a-9aa2-e7ade1b79058-kube-api-access-q45f8\") pod \"keystone-db-create-gxjv5\" (UID: \"9c51cd5f-cfaf-4c1a-9aa2-e7ade1b79058\") " pod="openstack/keystone-db-create-gxjv5" Mar 13 15:33:56 crc kubenswrapper[4907]: I0313 15:33:56.763130 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-48f2-account-create-update-wvks8"] Mar 13 15:33:56 crc kubenswrapper[4907]: I0313 15:33:56.764425 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-48f2-account-create-update-wvks8" Mar 13 15:33:56 crc kubenswrapper[4907]: I0313 15:33:56.771295 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-48f2-account-create-update-wvks8"] Mar 13 15:33:56 crc kubenswrapper[4907]: I0313 15:33:56.773484 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Mar 13 15:33:56 crc kubenswrapper[4907]: I0313 15:33:56.841092 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b03a1328-cfc5-4ff6-88b9-6523f4378708-operator-scripts\") pod \"keystone-48f2-account-create-update-wvks8\" (UID: \"b03a1328-cfc5-4ff6-88b9-6523f4378708\") " pod="openstack/keystone-48f2-account-create-update-wvks8" Mar 13 15:33:56 crc kubenswrapper[4907]: I0313 15:33:56.841177 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9c51cd5f-cfaf-4c1a-9aa2-e7ade1b79058-operator-scripts\") pod \"keystone-db-create-gxjv5\" (UID: \"9c51cd5f-cfaf-4c1a-9aa2-e7ade1b79058\") " pod="openstack/keystone-db-create-gxjv5" Mar 13 15:33:56 crc kubenswrapper[4907]: I0313 15:33:56.841257 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q45f8\" (UniqueName: \"kubernetes.io/projected/9c51cd5f-cfaf-4c1a-9aa2-e7ade1b79058-kube-api-access-q45f8\") pod \"keystone-db-create-gxjv5\" (UID: \"9c51cd5f-cfaf-4c1a-9aa2-e7ade1b79058\") " pod="openstack/keystone-db-create-gxjv5" Mar 13 15:33:56 crc kubenswrapper[4907]: I0313 15:33:56.841295 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-979dz\" (UniqueName: \"kubernetes.io/projected/b03a1328-cfc5-4ff6-88b9-6523f4378708-kube-api-access-979dz\") pod \"keystone-48f2-account-create-update-wvks8\" (UID: \"b03a1328-cfc5-4ff6-88b9-6523f4378708\") " pod="openstack/keystone-48f2-account-create-update-wvks8" Mar 13 15:33:56 crc kubenswrapper[4907]: I0313 15:33:56.841995 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9c51cd5f-cfaf-4c1a-9aa2-e7ade1b79058-operator-scripts\") pod \"keystone-db-create-gxjv5\" (UID: \"9c51cd5f-cfaf-4c1a-9aa2-e7ade1b79058\") " pod="openstack/keystone-db-create-gxjv5" Mar 13 15:33:56 crc kubenswrapper[4907]: I0313 15:33:56.868356 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q45f8\" (UniqueName: \"kubernetes.io/projected/9c51cd5f-cfaf-4c1a-9aa2-e7ade1b79058-kube-api-access-q45f8\") pod \"keystone-db-create-gxjv5\" (UID: \"9c51cd5f-cfaf-4c1a-9aa2-e7ade1b79058\") " pod="openstack/keystone-db-create-gxjv5" Mar 13 15:33:56 crc kubenswrapper[4907]: I0313 15:33:56.942716 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-979dz\" (UniqueName: \"kubernetes.io/projected/b03a1328-cfc5-4ff6-88b9-6523f4378708-kube-api-access-979dz\") pod \"keystone-48f2-account-create-update-wvks8\" (UID: \"b03a1328-cfc5-4ff6-88b9-6523f4378708\") " pod="openstack/keystone-48f2-account-create-update-wvks8" Mar 13 15:33:56 crc kubenswrapper[4907]: I0313 15:33:56.942756 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b03a1328-cfc5-4ff6-88b9-6523f4378708-operator-scripts\") pod \"keystone-48f2-account-create-update-wvks8\" (UID: \"b03a1328-cfc5-4ff6-88b9-6523f4378708\") " pod="openstack/keystone-48f2-account-create-update-wvks8" Mar 13 15:33:56 crc kubenswrapper[4907]: I0313 15:33:56.943411 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b03a1328-cfc5-4ff6-88b9-6523f4378708-operator-scripts\") pod \"keystone-48f2-account-create-update-wvks8\" (UID: \"b03a1328-cfc5-4ff6-88b9-6523f4378708\") " pod="openstack/keystone-48f2-account-create-update-wvks8" Mar 13 15:33:56 crc kubenswrapper[4907]: I0313 15:33:56.960392 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-979dz\" (UniqueName: \"kubernetes.io/projected/b03a1328-cfc5-4ff6-88b9-6523f4378708-kube-api-access-979dz\") pod \"keystone-48f2-account-create-update-wvks8\" (UID: \"b03a1328-cfc5-4ff6-88b9-6523f4378708\") " pod="openstack/keystone-48f2-account-create-update-wvks8" Mar 13 15:33:56 crc kubenswrapper[4907]: I0313 15:33:56.969174 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-gxjv5" Mar 13 15:33:57 crc kubenswrapper[4907]: I0313 15:33:57.081956 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-48f2-account-create-update-wvks8" Mar 13 15:33:58 crc kubenswrapper[4907]: I0313 15:33:58.022447 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-gxjv5"] Mar 13 15:33:58 crc kubenswrapper[4907]: I0313 15:33:58.089819 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-gxjv5" event={"ID":"9c51cd5f-cfaf-4c1a-9aa2-e7ade1b79058","Type":"ContainerStarted","Data":"ad6e40e290e5e7fd73118a9466156d34be11e432c2810a7a56144057ee355e41"} Mar 13 15:33:58 crc kubenswrapper[4907]: I0313 15:33:58.114409 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-48f2-account-create-update-wvks8"] Mar 13 15:33:58 crc kubenswrapper[4907]: W0313 15:33:58.128475 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb03a1328_cfc5_4ff6_88b9_6523f4378708.slice/crio-82735ca0c85dbbd92ff63172f6177cf41836602cab104a6008e63c9d13624f5b WatchSource:0}: Error finding container 82735ca0c85dbbd92ff63172f6177cf41836602cab104a6008e63c9d13624f5b: Status 404 returned error can't find the container with id 82735ca0c85dbbd92ff63172f6177cf41836602cab104a6008e63c9d13624f5b Mar 13 15:33:59 crc kubenswrapper[4907]: I0313 15:33:59.100341 4907 generic.go:334] "Generic (PLEG): container finished" podID="b03a1328-cfc5-4ff6-88b9-6523f4378708" containerID="d95988efdbf9f67c502e797aeeb6493599a1bcc1156251e60584683248ab36b0" exitCode=0 Mar 13 15:33:59 crc kubenswrapper[4907]: I0313 15:33:59.100441 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-48f2-account-create-update-wvks8" event={"ID":"b03a1328-cfc5-4ff6-88b9-6523f4378708","Type":"ContainerDied","Data":"d95988efdbf9f67c502e797aeeb6493599a1bcc1156251e60584683248ab36b0"} Mar 13 15:33:59 crc kubenswrapper[4907]: I0313 15:33:59.100721 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-48f2-account-create-update-wvks8" event={"ID":"b03a1328-cfc5-4ff6-88b9-6523f4378708","Type":"ContainerStarted","Data":"82735ca0c85dbbd92ff63172f6177cf41836602cab104a6008e63c9d13624f5b"} Mar 13 15:33:59 crc kubenswrapper[4907]: I0313 15:33:59.104241 4907 generic.go:334] "Generic (PLEG): container finished" podID="9c51cd5f-cfaf-4c1a-9aa2-e7ade1b79058" containerID="6f94031107f6bbf3f09d35d19fba9b0dfe08f6a055661f1e04241d246d42472d" exitCode=0 Mar 13 15:33:59 crc kubenswrapper[4907]: I0313 15:33:59.104288 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-gxjv5" event={"ID":"9c51cd5f-cfaf-4c1a-9aa2-e7ade1b79058","Type":"ContainerDied","Data":"6f94031107f6bbf3f09d35d19fba9b0dfe08f6a055661f1e04241d246d42472d"} Mar 13 15:34:00 crc kubenswrapper[4907]: I0313 15:34:00.139819 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556934-wf7zr"] Mar 13 15:34:00 crc kubenswrapper[4907]: I0313 15:34:00.141326 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556934-wf7zr" Mar 13 15:34:00 crc kubenswrapper[4907]: I0313 15:34:00.143677 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 15:34:00 crc kubenswrapper[4907]: I0313 15:34:00.143870 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 15:34:00 crc kubenswrapper[4907]: I0313 15:34:00.144067 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 15:34:00 crc kubenswrapper[4907]: I0313 15:34:00.176033 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556934-wf7zr"] Mar 13 15:34:00 crc kubenswrapper[4907]: I0313 15:34:00.202315 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kwpz2\" (UniqueName: \"kubernetes.io/projected/041cab28-d0e1-4ec9-a0cb-92176be369aa-kube-api-access-kwpz2\") pod \"auto-csr-approver-29556934-wf7zr\" (UID: \"041cab28-d0e1-4ec9-a0cb-92176be369aa\") " pod="openshift-infra/auto-csr-approver-29556934-wf7zr" Mar 13 15:34:00 crc kubenswrapper[4907]: I0313 15:34:00.303937 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kwpz2\" (UniqueName: \"kubernetes.io/projected/041cab28-d0e1-4ec9-a0cb-92176be369aa-kube-api-access-kwpz2\") pod \"auto-csr-approver-29556934-wf7zr\" (UID: \"041cab28-d0e1-4ec9-a0cb-92176be369aa\") " pod="openshift-infra/auto-csr-approver-29556934-wf7zr" Mar 13 15:34:00 crc kubenswrapper[4907]: I0313 15:34:00.333570 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kwpz2\" (UniqueName: \"kubernetes.io/projected/041cab28-d0e1-4ec9-a0cb-92176be369aa-kube-api-access-kwpz2\") pod \"auto-csr-approver-29556934-wf7zr\" (UID: \"041cab28-d0e1-4ec9-a0cb-92176be369aa\") " pod="openshift-infra/auto-csr-approver-29556934-wf7zr" Mar 13 15:34:00 crc kubenswrapper[4907]: I0313 15:34:00.469066 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556934-wf7zr" Mar 13 15:34:00 crc kubenswrapper[4907]: I0313 15:34:00.569098 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-48f2-account-create-update-wvks8" Mar 13 15:34:00 crc kubenswrapper[4907]: I0313 15:34:00.573533 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-gxjv5" Mar 13 15:34:00 crc kubenswrapper[4907]: I0313 15:34:00.710101 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b03a1328-cfc5-4ff6-88b9-6523f4378708-operator-scripts\") pod \"b03a1328-cfc5-4ff6-88b9-6523f4378708\" (UID: \"b03a1328-cfc5-4ff6-88b9-6523f4378708\") " Mar 13 15:34:00 crc kubenswrapper[4907]: I0313 15:34:00.710158 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-979dz\" (UniqueName: \"kubernetes.io/projected/b03a1328-cfc5-4ff6-88b9-6523f4378708-kube-api-access-979dz\") pod \"b03a1328-cfc5-4ff6-88b9-6523f4378708\" (UID: \"b03a1328-cfc5-4ff6-88b9-6523f4378708\") " Mar 13 15:34:00 crc kubenswrapper[4907]: I0313 15:34:00.710290 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9c51cd5f-cfaf-4c1a-9aa2-e7ade1b79058-operator-scripts\") pod \"9c51cd5f-cfaf-4c1a-9aa2-e7ade1b79058\" (UID: \"9c51cd5f-cfaf-4c1a-9aa2-e7ade1b79058\") " Mar 13 15:34:00 crc kubenswrapper[4907]: I0313 15:34:00.710351 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q45f8\" (UniqueName: \"kubernetes.io/projected/9c51cd5f-cfaf-4c1a-9aa2-e7ade1b79058-kube-api-access-q45f8\") pod \"9c51cd5f-cfaf-4c1a-9aa2-e7ade1b79058\" (UID: \"9c51cd5f-cfaf-4c1a-9aa2-e7ade1b79058\") " Mar 13 15:34:00 crc kubenswrapper[4907]: I0313 15:34:00.710957 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9c51cd5f-cfaf-4c1a-9aa2-e7ade1b79058-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "9c51cd5f-cfaf-4c1a-9aa2-e7ade1b79058" (UID: "9c51cd5f-cfaf-4c1a-9aa2-e7ade1b79058"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:34:00 crc kubenswrapper[4907]: I0313 15:34:00.711487 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b03a1328-cfc5-4ff6-88b9-6523f4378708-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b03a1328-cfc5-4ff6-88b9-6523f4378708" (UID: "b03a1328-cfc5-4ff6-88b9-6523f4378708"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:34:00 crc kubenswrapper[4907]: I0313 15:34:00.714545 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b03a1328-cfc5-4ff6-88b9-6523f4378708-kube-api-access-979dz" (OuterVolumeSpecName: "kube-api-access-979dz") pod "b03a1328-cfc5-4ff6-88b9-6523f4378708" (UID: "b03a1328-cfc5-4ff6-88b9-6523f4378708"). InnerVolumeSpecName "kube-api-access-979dz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:34:00 crc kubenswrapper[4907]: I0313 15:34:00.714622 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c51cd5f-cfaf-4c1a-9aa2-e7ade1b79058-kube-api-access-q45f8" (OuterVolumeSpecName: "kube-api-access-q45f8") pod "9c51cd5f-cfaf-4c1a-9aa2-e7ade1b79058" (UID: "9c51cd5f-cfaf-4c1a-9aa2-e7ade1b79058"). InnerVolumeSpecName "kube-api-access-q45f8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:34:00 crc kubenswrapper[4907]: I0313 15:34:00.812259 4907 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b03a1328-cfc5-4ff6-88b9-6523f4378708-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 15:34:00 crc kubenswrapper[4907]: I0313 15:34:00.812300 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-979dz\" (UniqueName: \"kubernetes.io/projected/b03a1328-cfc5-4ff6-88b9-6523f4378708-kube-api-access-979dz\") on node \"crc\" DevicePath \"\"" Mar 13 15:34:00 crc kubenswrapper[4907]: I0313 15:34:00.812313 4907 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9c51cd5f-cfaf-4c1a-9aa2-e7ade1b79058-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 15:34:00 crc kubenswrapper[4907]: I0313 15:34:00.812322 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q45f8\" (UniqueName: \"kubernetes.io/projected/9c51cd5f-cfaf-4c1a-9aa2-e7ade1b79058-kube-api-access-q45f8\") on node \"crc\" DevicePath \"\"" Mar 13 15:34:00 crc kubenswrapper[4907]: I0313 15:34:00.897778 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556934-wf7zr"] Mar 13 15:34:00 crc kubenswrapper[4907]: W0313 15:34:00.899640 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod041cab28_d0e1_4ec9_a0cb_92176be369aa.slice/crio-1deb6d11c2b99bf690d558f4ecc15f659c14ba85d596c123916d00f2b8567d4c WatchSource:0}: Error finding container 1deb6d11c2b99bf690d558f4ecc15f659c14ba85d596c123916d00f2b8567d4c: Status 404 returned error can't find the container with id 1deb6d11c2b99bf690d558f4ecc15f659c14ba85d596c123916d00f2b8567d4c Mar 13 15:34:01 crc kubenswrapper[4907]: I0313 15:34:01.121764 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-48f2-account-create-update-wvks8" event={"ID":"b03a1328-cfc5-4ff6-88b9-6523f4378708","Type":"ContainerDied","Data":"82735ca0c85dbbd92ff63172f6177cf41836602cab104a6008e63c9d13624f5b"} Mar 13 15:34:01 crc kubenswrapper[4907]: I0313 15:34:01.121804 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-48f2-account-create-update-wvks8" Mar 13 15:34:01 crc kubenswrapper[4907]: I0313 15:34:01.121815 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="82735ca0c85dbbd92ff63172f6177cf41836602cab104a6008e63c9d13624f5b" Mar 13 15:34:01 crc kubenswrapper[4907]: I0313 15:34:01.126933 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-gxjv5" Mar 13 15:34:01 crc kubenswrapper[4907]: I0313 15:34:01.126980 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-gxjv5" event={"ID":"9c51cd5f-cfaf-4c1a-9aa2-e7ade1b79058","Type":"ContainerDied","Data":"ad6e40e290e5e7fd73118a9466156d34be11e432c2810a7a56144057ee355e41"} Mar 13 15:34:01 crc kubenswrapper[4907]: I0313 15:34:01.127122 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ad6e40e290e5e7fd73118a9466156d34be11e432c2810a7a56144057ee355e41" Mar 13 15:34:01 crc kubenswrapper[4907]: I0313 15:34:01.128204 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556934-wf7zr" event={"ID":"041cab28-d0e1-4ec9-a0cb-92176be369aa","Type":"ContainerStarted","Data":"1deb6d11c2b99bf690d558f4ecc15f659c14ba85d596c123916d00f2b8567d4c"} Mar 13 15:34:02 crc kubenswrapper[4907]: I0313 15:34:02.389376 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-4gtrw"] Mar 13 15:34:02 crc kubenswrapper[4907]: E0313 15:34:02.390366 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b03a1328-cfc5-4ff6-88b9-6523f4378708" containerName="mariadb-account-create-update" Mar 13 15:34:02 crc kubenswrapper[4907]: I0313 15:34:02.390386 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="b03a1328-cfc5-4ff6-88b9-6523f4378708" containerName="mariadb-account-create-update" Mar 13 15:34:02 crc kubenswrapper[4907]: E0313 15:34:02.390429 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c51cd5f-cfaf-4c1a-9aa2-e7ade1b79058" containerName="mariadb-database-create" Mar 13 15:34:02 crc kubenswrapper[4907]: I0313 15:34:02.390438 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c51cd5f-cfaf-4c1a-9aa2-e7ade1b79058" containerName="mariadb-database-create" Mar 13 15:34:02 crc kubenswrapper[4907]: I0313 15:34:02.390828 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="b03a1328-cfc5-4ff6-88b9-6523f4378708" containerName="mariadb-account-create-update" Mar 13 15:34:02 crc kubenswrapper[4907]: I0313 15:34:02.390862 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c51cd5f-cfaf-4c1a-9aa2-e7ade1b79058" containerName="mariadb-database-create" Mar 13 15:34:02 crc kubenswrapper[4907]: I0313 15:34:02.392307 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-4gtrw" Mar 13 15:34:02 crc kubenswrapper[4907]: I0313 15:34:02.394580 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Mar 13 15:34:02 crc kubenswrapper[4907]: I0313 15:34:02.394810 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-gjbw7" Mar 13 15:34:02 crc kubenswrapper[4907]: I0313 15:34:02.395059 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Mar 13 15:34:02 crc kubenswrapper[4907]: I0313 15:34:02.395399 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Mar 13 15:34:02 crc kubenswrapper[4907]: I0313 15:34:02.400529 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-4gtrw"] Mar 13 15:34:02 crc kubenswrapper[4907]: I0313 15:34:02.442845 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f23a73ee-57ff-4d58-8812-b72f624b7739-config-data\") pod \"keystone-db-sync-4gtrw\" (UID: \"f23a73ee-57ff-4d58-8812-b72f624b7739\") " pod="openstack/keystone-db-sync-4gtrw" Mar 13 15:34:02 crc kubenswrapper[4907]: I0313 15:34:02.442943 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wjdfl\" (UniqueName: \"kubernetes.io/projected/f23a73ee-57ff-4d58-8812-b72f624b7739-kube-api-access-wjdfl\") pod \"keystone-db-sync-4gtrw\" (UID: \"f23a73ee-57ff-4d58-8812-b72f624b7739\") " pod="openstack/keystone-db-sync-4gtrw" Mar 13 15:34:02 crc kubenswrapper[4907]: I0313 15:34:02.443081 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f23a73ee-57ff-4d58-8812-b72f624b7739-combined-ca-bundle\") pod \"keystone-db-sync-4gtrw\" (UID: \"f23a73ee-57ff-4d58-8812-b72f624b7739\") " pod="openstack/keystone-db-sync-4gtrw" Mar 13 15:34:02 crc kubenswrapper[4907]: I0313 15:34:02.544670 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f23a73ee-57ff-4d58-8812-b72f624b7739-combined-ca-bundle\") pod \"keystone-db-sync-4gtrw\" (UID: \"f23a73ee-57ff-4d58-8812-b72f624b7739\") " pod="openstack/keystone-db-sync-4gtrw" Mar 13 15:34:02 crc kubenswrapper[4907]: I0313 15:34:02.544729 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f23a73ee-57ff-4d58-8812-b72f624b7739-config-data\") pod \"keystone-db-sync-4gtrw\" (UID: \"f23a73ee-57ff-4d58-8812-b72f624b7739\") " pod="openstack/keystone-db-sync-4gtrw" Mar 13 15:34:02 crc kubenswrapper[4907]: I0313 15:34:02.544762 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wjdfl\" (UniqueName: \"kubernetes.io/projected/f23a73ee-57ff-4d58-8812-b72f624b7739-kube-api-access-wjdfl\") pod \"keystone-db-sync-4gtrw\" (UID: \"f23a73ee-57ff-4d58-8812-b72f624b7739\") " pod="openstack/keystone-db-sync-4gtrw" Mar 13 15:34:02 crc kubenswrapper[4907]: I0313 15:34:02.558649 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f23a73ee-57ff-4d58-8812-b72f624b7739-combined-ca-bundle\") pod \"keystone-db-sync-4gtrw\" (UID: \"f23a73ee-57ff-4d58-8812-b72f624b7739\") " pod="openstack/keystone-db-sync-4gtrw" Mar 13 15:34:02 crc kubenswrapper[4907]: I0313 15:34:02.565843 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f23a73ee-57ff-4d58-8812-b72f624b7739-config-data\") pod \"keystone-db-sync-4gtrw\" (UID: \"f23a73ee-57ff-4d58-8812-b72f624b7739\") " pod="openstack/keystone-db-sync-4gtrw" Mar 13 15:34:02 crc kubenswrapper[4907]: I0313 15:34:02.568177 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wjdfl\" (UniqueName: \"kubernetes.io/projected/f23a73ee-57ff-4d58-8812-b72f624b7739-kube-api-access-wjdfl\") pod \"keystone-db-sync-4gtrw\" (UID: \"f23a73ee-57ff-4d58-8812-b72f624b7739\") " pod="openstack/keystone-db-sync-4gtrw" Mar 13 15:34:02 crc kubenswrapper[4907]: I0313 15:34:02.718379 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-4gtrw" Mar 13 15:34:03 crc kubenswrapper[4907]: I0313 15:34:03.143867 4907 generic.go:334] "Generic (PLEG): container finished" podID="041cab28-d0e1-4ec9-a0cb-92176be369aa" containerID="1eda03ad73ea12900c383cf8f9715d8d1c4690e9fdc026cd78850ed84ad83b37" exitCode=0 Mar 13 15:34:03 crc kubenswrapper[4907]: I0313 15:34:03.143975 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556934-wf7zr" event={"ID":"041cab28-d0e1-4ec9-a0cb-92176be369aa","Type":"ContainerDied","Data":"1eda03ad73ea12900c383cf8f9715d8d1c4690e9fdc026cd78850ed84ad83b37"} Mar 13 15:34:03 crc kubenswrapper[4907]: I0313 15:34:03.170065 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-4gtrw"] Mar 13 15:34:04 crc kubenswrapper[4907]: I0313 15:34:04.163457 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-4gtrw" event={"ID":"f23a73ee-57ff-4d58-8812-b72f624b7739","Type":"ContainerStarted","Data":"a3a50fd2f676041cc1b8a25989baf35e7feed38466f29e552c042a8c33760d83"} Mar 13 15:34:04 crc kubenswrapper[4907]: I0313 15:34:04.163763 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-4gtrw" event={"ID":"f23a73ee-57ff-4d58-8812-b72f624b7739","Type":"ContainerStarted","Data":"3c42a8c243c0093bfc9f1f8cf6bfd418cc51eaed61e0ec45a9fbb1802050c2c9"} Mar 13 15:34:04 crc kubenswrapper[4907]: I0313 15:34:04.526136 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556934-wf7zr" Mar 13 15:34:04 crc kubenswrapper[4907]: I0313 15:34:04.542094 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-4gtrw" podStartSLOduration=2.542073573 podStartE2EDuration="2.542073573s" podCreationTimestamp="2026-03-13 15:34:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:34:04.18279993 +0000 UTC m=+5343.082587619" watchObservedRunningTime="2026-03-13 15:34:04.542073573 +0000 UTC m=+5343.441861262" Mar 13 15:34:04 crc kubenswrapper[4907]: I0313 15:34:04.575061 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kwpz2\" (UniqueName: \"kubernetes.io/projected/041cab28-d0e1-4ec9-a0cb-92176be369aa-kube-api-access-kwpz2\") pod \"041cab28-d0e1-4ec9-a0cb-92176be369aa\" (UID: \"041cab28-d0e1-4ec9-a0cb-92176be369aa\") " Mar 13 15:34:04 crc kubenswrapper[4907]: I0313 15:34:04.583520 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/041cab28-d0e1-4ec9-a0cb-92176be369aa-kube-api-access-kwpz2" (OuterVolumeSpecName: "kube-api-access-kwpz2") pod "041cab28-d0e1-4ec9-a0cb-92176be369aa" (UID: "041cab28-d0e1-4ec9-a0cb-92176be369aa"). InnerVolumeSpecName "kube-api-access-kwpz2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:34:04 crc kubenswrapper[4907]: I0313 15:34:04.677337 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kwpz2\" (UniqueName: \"kubernetes.io/projected/041cab28-d0e1-4ec9-a0cb-92176be369aa-kube-api-access-kwpz2\") on node \"crc\" DevicePath \"\"" Mar 13 15:34:05 crc kubenswrapper[4907]: I0313 15:34:05.172275 4907 generic.go:334] "Generic (PLEG): container finished" podID="f23a73ee-57ff-4d58-8812-b72f624b7739" containerID="a3a50fd2f676041cc1b8a25989baf35e7feed38466f29e552c042a8c33760d83" exitCode=0 Mar 13 15:34:05 crc kubenswrapper[4907]: I0313 15:34:05.172329 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-4gtrw" event={"ID":"f23a73ee-57ff-4d58-8812-b72f624b7739","Type":"ContainerDied","Data":"a3a50fd2f676041cc1b8a25989baf35e7feed38466f29e552c042a8c33760d83"} Mar 13 15:34:05 crc kubenswrapper[4907]: I0313 15:34:05.174458 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556934-wf7zr" event={"ID":"041cab28-d0e1-4ec9-a0cb-92176be369aa","Type":"ContainerDied","Data":"1deb6d11c2b99bf690d558f4ecc15f659c14ba85d596c123916d00f2b8567d4c"} Mar 13 15:34:05 crc kubenswrapper[4907]: I0313 15:34:05.174488 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1deb6d11c2b99bf690d558f4ecc15f659c14ba85d596c123916d00f2b8567d4c" Mar 13 15:34:05 crc kubenswrapper[4907]: I0313 15:34:05.174540 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556934-wf7zr" Mar 13 15:34:05 crc kubenswrapper[4907]: I0313 15:34:05.591523 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556928-nh6lw"] Mar 13 15:34:05 crc kubenswrapper[4907]: I0313 15:34:05.598946 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556928-nh6lw"] Mar 13 15:34:05 crc kubenswrapper[4907]: I0313 15:34:05.790793 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="566ddaa7-ea07-421b-a181-0aa768454ba6" path="/var/lib/kubelet/pods/566ddaa7-ea07-421b-a181-0aa768454ba6/volumes" Mar 13 15:34:06 crc kubenswrapper[4907]: I0313 15:34:06.588896 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-4gtrw" Mar 13 15:34:06 crc kubenswrapper[4907]: I0313 15:34:06.712795 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f23a73ee-57ff-4d58-8812-b72f624b7739-config-data\") pod \"f23a73ee-57ff-4d58-8812-b72f624b7739\" (UID: \"f23a73ee-57ff-4d58-8812-b72f624b7739\") " Mar 13 15:34:06 crc kubenswrapper[4907]: I0313 15:34:06.712854 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f23a73ee-57ff-4d58-8812-b72f624b7739-combined-ca-bundle\") pod \"f23a73ee-57ff-4d58-8812-b72f624b7739\" (UID: \"f23a73ee-57ff-4d58-8812-b72f624b7739\") " Mar 13 15:34:06 crc kubenswrapper[4907]: I0313 15:34:06.713014 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wjdfl\" (UniqueName: \"kubernetes.io/projected/f23a73ee-57ff-4d58-8812-b72f624b7739-kube-api-access-wjdfl\") pod \"f23a73ee-57ff-4d58-8812-b72f624b7739\" (UID: \"f23a73ee-57ff-4d58-8812-b72f624b7739\") " Mar 13 15:34:06 crc kubenswrapper[4907]: I0313 15:34:06.718156 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f23a73ee-57ff-4d58-8812-b72f624b7739-kube-api-access-wjdfl" (OuterVolumeSpecName: "kube-api-access-wjdfl") pod "f23a73ee-57ff-4d58-8812-b72f624b7739" (UID: "f23a73ee-57ff-4d58-8812-b72f624b7739"). InnerVolumeSpecName "kube-api-access-wjdfl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:34:06 crc kubenswrapper[4907]: I0313 15:34:06.736649 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f23a73ee-57ff-4d58-8812-b72f624b7739-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f23a73ee-57ff-4d58-8812-b72f624b7739" (UID: "f23a73ee-57ff-4d58-8812-b72f624b7739"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:34:06 crc kubenswrapper[4907]: I0313 15:34:06.761083 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f23a73ee-57ff-4d58-8812-b72f624b7739-config-data" (OuterVolumeSpecName: "config-data") pod "f23a73ee-57ff-4d58-8812-b72f624b7739" (UID: "f23a73ee-57ff-4d58-8812-b72f624b7739"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:34:06 crc kubenswrapper[4907]: I0313 15:34:06.782291 4907 scope.go:117] "RemoveContainer" containerID="592bf9f175f440e943369aab2abbb13ad2c26412cb0e0738c4a33a156b8a5596" Mar 13 15:34:06 crc kubenswrapper[4907]: E0313 15:34:06.782907 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:34:06 crc kubenswrapper[4907]: I0313 15:34:06.815164 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wjdfl\" (UniqueName: \"kubernetes.io/projected/f23a73ee-57ff-4d58-8812-b72f624b7739-kube-api-access-wjdfl\") on node \"crc\" DevicePath \"\"" Mar 13 15:34:06 crc kubenswrapper[4907]: I0313 15:34:06.815205 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f23a73ee-57ff-4d58-8812-b72f624b7739-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 15:34:06 crc kubenswrapper[4907]: I0313 15:34:06.815220 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f23a73ee-57ff-4d58-8812-b72f624b7739-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 15:34:07 crc kubenswrapper[4907]: I0313 15:34:07.191553 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-4gtrw" event={"ID":"f23a73ee-57ff-4d58-8812-b72f624b7739","Type":"ContainerDied","Data":"3c42a8c243c0093bfc9f1f8cf6bfd418cc51eaed61e0ec45a9fbb1802050c2c9"} Mar 13 15:34:07 crc kubenswrapper[4907]: I0313 15:34:07.191862 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3c42a8c243c0093bfc9f1f8cf6bfd418cc51eaed61e0ec45a9fbb1802050c2c9" Mar 13 15:34:07 crc kubenswrapper[4907]: I0313 15:34:07.191618 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-4gtrw" Mar 13 15:34:07 crc kubenswrapper[4907]: I0313 15:34:07.427056 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6c5b6fdc47-txgc2"] Mar 13 15:34:07 crc kubenswrapper[4907]: E0313 15:34:07.427683 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="041cab28-d0e1-4ec9-a0cb-92176be369aa" containerName="oc" Mar 13 15:34:07 crc kubenswrapper[4907]: I0313 15:34:07.427701 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="041cab28-d0e1-4ec9-a0cb-92176be369aa" containerName="oc" Mar 13 15:34:07 crc kubenswrapper[4907]: E0313 15:34:07.427743 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f23a73ee-57ff-4d58-8812-b72f624b7739" containerName="keystone-db-sync" Mar 13 15:34:07 crc kubenswrapper[4907]: I0313 15:34:07.427753 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="f23a73ee-57ff-4d58-8812-b72f624b7739" containerName="keystone-db-sync" Mar 13 15:34:07 crc kubenswrapper[4907]: I0313 15:34:07.427960 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="f23a73ee-57ff-4d58-8812-b72f624b7739" containerName="keystone-db-sync" Mar 13 15:34:07 crc kubenswrapper[4907]: I0313 15:34:07.428000 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="041cab28-d0e1-4ec9-a0cb-92176be369aa" containerName="oc" Mar 13 15:34:07 crc kubenswrapper[4907]: I0313 15:34:07.429166 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c5b6fdc47-txgc2" Mar 13 15:34:07 crc kubenswrapper[4907]: I0313 15:34:07.442031 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6c5b6fdc47-txgc2"] Mar 13 15:34:07 crc kubenswrapper[4907]: I0313 15:34:07.502940 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-mqwdv"] Mar 13 15:34:07 crc kubenswrapper[4907]: I0313 15:34:07.504297 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-mqwdv" Mar 13 15:34:07 crc kubenswrapper[4907]: I0313 15:34:07.509801 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-mqwdv"] Mar 13 15:34:07 crc kubenswrapper[4907]: I0313 15:34:07.510639 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Mar 13 15:34:07 crc kubenswrapper[4907]: I0313 15:34:07.510826 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Mar 13 15:34:07 crc kubenswrapper[4907]: I0313 15:34:07.511029 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Mar 13 15:34:07 crc kubenswrapper[4907]: I0313 15:34:07.511187 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-gjbw7" Mar 13 15:34:07 crc kubenswrapper[4907]: I0313 15:34:07.511382 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Mar 13 15:34:07 crc kubenswrapper[4907]: I0313 15:34:07.531899 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c26e3cb6-4f00-49d9-8bc2-9ec347574580-config\") pod \"dnsmasq-dns-6c5b6fdc47-txgc2\" (UID: \"c26e3cb6-4f00-49d9-8bc2-9ec347574580\") " pod="openstack/dnsmasq-dns-6c5b6fdc47-txgc2" Mar 13 15:34:07 crc kubenswrapper[4907]: I0313 15:34:07.531976 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c26e3cb6-4f00-49d9-8bc2-9ec347574580-dns-svc\") pod \"dnsmasq-dns-6c5b6fdc47-txgc2\" (UID: \"c26e3cb6-4f00-49d9-8bc2-9ec347574580\") " pod="openstack/dnsmasq-dns-6c5b6fdc47-txgc2" Mar 13 15:34:07 crc kubenswrapper[4907]: I0313 15:34:07.532107 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c26e3cb6-4f00-49d9-8bc2-9ec347574580-ovsdbserver-sb\") pod \"dnsmasq-dns-6c5b6fdc47-txgc2\" (UID: \"c26e3cb6-4f00-49d9-8bc2-9ec347574580\") " pod="openstack/dnsmasq-dns-6c5b6fdc47-txgc2" Mar 13 15:34:07 crc kubenswrapper[4907]: I0313 15:34:07.532140 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c26e3cb6-4f00-49d9-8bc2-9ec347574580-ovsdbserver-nb\") pod \"dnsmasq-dns-6c5b6fdc47-txgc2\" (UID: \"c26e3cb6-4f00-49d9-8bc2-9ec347574580\") " pod="openstack/dnsmasq-dns-6c5b6fdc47-txgc2" Mar 13 15:34:07 crc kubenswrapper[4907]: I0313 15:34:07.532166 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hmslv\" (UniqueName: \"kubernetes.io/projected/c26e3cb6-4f00-49d9-8bc2-9ec347574580-kube-api-access-hmslv\") pod \"dnsmasq-dns-6c5b6fdc47-txgc2\" (UID: \"c26e3cb6-4f00-49d9-8bc2-9ec347574580\") " pod="openstack/dnsmasq-dns-6c5b6fdc47-txgc2" Mar 13 15:34:07 crc kubenswrapper[4907]: I0313 15:34:07.633677 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1838c955-2910-4589-a64f-7fe9f865d991-credential-keys\") pod \"keystone-bootstrap-mqwdv\" (UID: \"1838c955-2910-4589-a64f-7fe9f865d991\") " pod="openstack/keystone-bootstrap-mqwdv" Mar 13 15:34:07 crc kubenswrapper[4907]: I0313 15:34:07.633731 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c26e3cb6-4f00-49d9-8bc2-9ec347574580-ovsdbserver-nb\") pod \"dnsmasq-dns-6c5b6fdc47-txgc2\" (UID: \"c26e3cb6-4f00-49d9-8bc2-9ec347574580\") " pod="openstack/dnsmasq-dns-6c5b6fdc47-txgc2" Mar 13 15:34:07 crc kubenswrapper[4907]: I0313 15:34:07.633757 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hmslv\" (UniqueName: \"kubernetes.io/projected/c26e3cb6-4f00-49d9-8bc2-9ec347574580-kube-api-access-hmslv\") pod \"dnsmasq-dns-6c5b6fdc47-txgc2\" (UID: \"c26e3cb6-4f00-49d9-8bc2-9ec347574580\") " pod="openstack/dnsmasq-dns-6c5b6fdc47-txgc2" Mar 13 15:34:07 crc kubenswrapper[4907]: I0313 15:34:07.633783 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kf2mt\" (UniqueName: \"kubernetes.io/projected/1838c955-2910-4589-a64f-7fe9f865d991-kube-api-access-kf2mt\") pod \"keystone-bootstrap-mqwdv\" (UID: \"1838c955-2910-4589-a64f-7fe9f865d991\") " pod="openstack/keystone-bootstrap-mqwdv" Mar 13 15:34:07 crc kubenswrapper[4907]: I0313 15:34:07.633806 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c26e3cb6-4f00-49d9-8bc2-9ec347574580-config\") pod \"dnsmasq-dns-6c5b6fdc47-txgc2\" (UID: \"c26e3cb6-4f00-49d9-8bc2-9ec347574580\") " pod="openstack/dnsmasq-dns-6c5b6fdc47-txgc2" Mar 13 15:34:07 crc kubenswrapper[4907]: I0313 15:34:07.633824 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1838c955-2910-4589-a64f-7fe9f865d991-fernet-keys\") pod \"keystone-bootstrap-mqwdv\" (UID: \"1838c955-2910-4589-a64f-7fe9f865d991\") " pod="openstack/keystone-bootstrap-mqwdv" Mar 13 15:34:07 crc kubenswrapper[4907]: I0313 15:34:07.633862 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c26e3cb6-4f00-49d9-8bc2-9ec347574580-dns-svc\") pod \"dnsmasq-dns-6c5b6fdc47-txgc2\" (UID: \"c26e3cb6-4f00-49d9-8bc2-9ec347574580\") " pod="openstack/dnsmasq-dns-6c5b6fdc47-txgc2" Mar 13 15:34:07 crc kubenswrapper[4907]: I0313 15:34:07.633904 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1838c955-2910-4589-a64f-7fe9f865d991-combined-ca-bundle\") pod \"keystone-bootstrap-mqwdv\" (UID: \"1838c955-2910-4589-a64f-7fe9f865d991\") " pod="openstack/keystone-bootstrap-mqwdv" Mar 13 15:34:07 crc kubenswrapper[4907]: I0313 15:34:07.634020 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1838c955-2910-4589-a64f-7fe9f865d991-config-data\") pod \"keystone-bootstrap-mqwdv\" (UID: \"1838c955-2910-4589-a64f-7fe9f865d991\") " pod="openstack/keystone-bootstrap-mqwdv" Mar 13 15:34:07 crc kubenswrapper[4907]: I0313 15:34:07.634132 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1838c955-2910-4589-a64f-7fe9f865d991-scripts\") pod \"keystone-bootstrap-mqwdv\" (UID: \"1838c955-2910-4589-a64f-7fe9f865d991\") " pod="openstack/keystone-bootstrap-mqwdv" Mar 13 15:34:07 crc kubenswrapper[4907]: I0313 15:34:07.634817 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c26e3cb6-4f00-49d9-8bc2-9ec347574580-ovsdbserver-nb\") pod \"dnsmasq-dns-6c5b6fdc47-txgc2\" (UID: \"c26e3cb6-4f00-49d9-8bc2-9ec347574580\") " pod="openstack/dnsmasq-dns-6c5b6fdc47-txgc2" Mar 13 15:34:07 crc kubenswrapper[4907]: I0313 15:34:07.637038 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c26e3cb6-4f00-49d9-8bc2-9ec347574580-ovsdbserver-sb\") pod \"dnsmasq-dns-6c5b6fdc47-txgc2\" (UID: \"c26e3cb6-4f00-49d9-8bc2-9ec347574580\") " pod="openstack/dnsmasq-dns-6c5b6fdc47-txgc2" Mar 13 15:34:07 crc kubenswrapper[4907]: I0313 15:34:07.638367 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c26e3cb6-4f00-49d9-8bc2-9ec347574580-config\") pod \"dnsmasq-dns-6c5b6fdc47-txgc2\" (UID: \"c26e3cb6-4f00-49d9-8bc2-9ec347574580\") " pod="openstack/dnsmasq-dns-6c5b6fdc47-txgc2" Mar 13 15:34:07 crc kubenswrapper[4907]: I0313 15:34:07.638397 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c26e3cb6-4f00-49d9-8bc2-9ec347574580-ovsdbserver-sb\") pod \"dnsmasq-dns-6c5b6fdc47-txgc2\" (UID: \"c26e3cb6-4f00-49d9-8bc2-9ec347574580\") " pod="openstack/dnsmasq-dns-6c5b6fdc47-txgc2" Mar 13 15:34:07 crc kubenswrapper[4907]: I0313 15:34:07.638420 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c26e3cb6-4f00-49d9-8bc2-9ec347574580-dns-svc\") pod \"dnsmasq-dns-6c5b6fdc47-txgc2\" (UID: \"c26e3cb6-4f00-49d9-8bc2-9ec347574580\") " pod="openstack/dnsmasq-dns-6c5b6fdc47-txgc2" Mar 13 15:34:07 crc kubenswrapper[4907]: I0313 15:34:07.659755 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hmslv\" (UniqueName: \"kubernetes.io/projected/c26e3cb6-4f00-49d9-8bc2-9ec347574580-kube-api-access-hmslv\") pod \"dnsmasq-dns-6c5b6fdc47-txgc2\" (UID: \"c26e3cb6-4f00-49d9-8bc2-9ec347574580\") " pod="openstack/dnsmasq-dns-6c5b6fdc47-txgc2" Mar 13 15:34:07 crc kubenswrapper[4907]: I0313 15:34:07.738921 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1838c955-2910-4589-a64f-7fe9f865d991-config-data\") pod \"keystone-bootstrap-mqwdv\" (UID: \"1838c955-2910-4589-a64f-7fe9f865d991\") " pod="openstack/keystone-bootstrap-mqwdv" Mar 13 15:34:07 crc kubenswrapper[4907]: I0313 15:34:07.739304 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1838c955-2910-4589-a64f-7fe9f865d991-scripts\") pod \"keystone-bootstrap-mqwdv\" (UID: \"1838c955-2910-4589-a64f-7fe9f865d991\") " pod="openstack/keystone-bootstrap-mqwdv" Mar 13 15:34:07 crc kubenswrapper[4907]: I0313 15:34:07.739340 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1838c955-2910-4589-a64f-7fe9f865d991-credential-keys\") pod \"keystone-bootstrap-mqwdv\" (UID: \"1838c955-2910-4589-a64f-7fe9f865d991\") " pod="openstack/keystone-bootstrap-mqwdv" Mar 13 15:34:07 crc kubenswrapper[4907]: I0313 15:34:07.739372 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kf2mt\" (UniqueName: \"kubernetes.io/projected/1838c955-2910-4589-a64f-7fe9f865d991-kube-api-access-kf2mt\") pod \"keystone-bootstrap-mqwdv\" (UID: \"1838c955-2910-4589-a64f-7fe9f865d991\") " pod="openstack/keystone-bootstrap-mqwdv" Mar 13 15:34:07 crc kubenswrapper[4907]: I0313 15:34:07.739393 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1838c955-2910-4589-a64f-7fe9f865d991-fernet-keys\") pod \"keystone-bootstrap-mqwdv\" (UID: \"1838c955-2910-4589-a64f-7fe9f865d991\") " pod="openstack/keystone-bootstrap-mqwdv" Mar 13 15:34:07 crc kubenswrapper[4907]: I0313 15:34:07.739429 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1838c955-2910-4589-a64f-7fe9f865d991-combined-ca-bundle\") pod \"keystone-bootstrap-mqwdv\" (UID: \"1838c955-2910-4589-a64f-7fe9f865d991\") " pod="openstack/keystone-bootstrap-mqwdv" Mar 13 15:34:07 crc kubenswrapper[4907]: I0313 15:34:07.742839 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1838c955-2910-4589-a64f-7fe9f865d991-combined-ca-bundle\") pod \"keystone-bootstrap-mqwdv\" (UID: \"1838c955-2910-4589-a64f-7fe9f865d991\") " pod="openstack/keystone-bootstrap-mqwdv" Mar 13 15:34:07 crc kubenswrapper[4907]: I0313 15:34:07.743067 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1838c955-2910-4589-a64f-7fe9f865d991-fernet-keys\") pod \"keystone-bootstrap-mqwdv\" (UID: \"1838c955-2910-4589-a64f-7fe9f865d991\") " pod="openstack/keystone-bootstrap-mqwdv" Mar 13 15:34:07 crc kubenswrapper[4907]: I0313 15:34:07.743173 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1838c955-2910-4589-a64f-7fe9f865d991-credential-keys\") pod \"keystone-bootstrap-mqwdv\" (UID: \"1838c955-2910-4589-a64f-7fe9f865d991\") " pod="openstack/keystone-bootstrap-mqwdv" Mar 13 15:34:07 crc kubenswrapper[4907]: I0313 15:34:07.745003 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1838c955-2910-4589-a64f-7fe9f865d991-scripts\") pod \"keystone-bootstrap-mqwdv\" (UID: \"1838c955-2910-4589-a64f-7fe9f865d991\") " pod="openstack/keystone-bootstrap-mqwdv" Mar 13 15:34:07 crc kubenswrapper[4907]: I0313 15:34:07.745015 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1838c955-2910-4589-a64f-7fe9f865d991-config-data\") pod \"keystone-bootstrap-mqwdv\" (UID: \"1838c955-2910-4589-a64f-7fe9f865d991\") " pod="openstack/keystone-bootstrap-mqwdv" Mar 13 15:34:07 crc kubenswrapper[4907]: I0313 15:34:07.755324 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c5b6fdc47-txgc2" Mar 13 15:34:07 crc kubenswrapper[4907]: I0313 15:34:07.759721 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kf2mt\" (UniqueName: \"kubernetes.io/projected/1838c955-2910-4589-a64f-7fe9f865d991-kube-api-access-kf2mt\") pod \"keystone-bootstrap-mqwdv\" (UID: \"1838c955-2910-4589-a64f-7fe9f865d991\") " pod="openstack/keystone-bootstrap-mqwdv" Mar 13 15:34:07 crc kubenswrapper[4907]: I0313 15:34:07.834902 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-mqwdv" Mar 13 15:34:08 crc kubenswrapper[4907]: I0313 15:34:08.217913 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6c5b6fdc47-txgc2"] Mar 13 15:34:08 crc kubenswrapper[4907]: W0313 15:34:08.221636 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc26e3cb6_4f00_49d9_8bc2_9ec347574580.slice/crio-5651c9bf703897955b69287b03043d49d9012737130119f792c71677cc9d3fe5 WatchSource:0}: Error finding container 5651c9bf703897955b69287b03043d49d9012737130119f792c71677cc9d3fe5: Status 404 returned error can't find the container with id 5651c9bf703897955b69287b03043d49d9012737130119f792c71677cc9d3fe5 Mar 13 15:34:08 crc kubenswrapper[4907]: W0313 15:34:08.347160 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1838c955_2910_4589_a64f_7fe9f865d991.slice/crio-caca426201cac3321e9d42ef962695a8f59b09bc5a76701b84aff5acd2e5bb8f WatchSource:0}: Error finding container caca426201cac3321e9d42ef962695a8f59b09bc5a76701b84aff5acd2e5bb8f: Status 404 returned error can't find the container with id caca426201cac3321e9d42ef962695a8f59b09bc5a76701b84aff5acd2e5bb8f Mar 13 15:34:08 crc kubenswrapper[4907]: I0313 15:34:08.351537 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-mqwdv"] Mar 13 15:34:09 crc kubenswrapper[4907]: I0313 15:34:09.210275 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-mqwdv" event={"ID":"1838c955-2910-4589-a64f-7fe9f865d991","Type":"ContainerStarted","Data":"47fa5a52c38336d054a9aed448aae48b3dd0ee12065b60aaad1abb03635d9d32"} Mar 13 15:34:09 crc kubenswrapper[4907]: I0313 15:34:09.210967 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-mqwdv" event={"ID":"1838c955-2910-4589-a64f-7fe9f865d991","Type":"ContainerStarted","Data":"caca426201cac3321e9d42ef962695a8f59b09bc5a76701b84aff5acd2e5bb8f"} Mar 13 15:34:09 crc kubenswrapper[4907]: I0313 15:34:09.212198 4907 generic.go:334] "Generic (PLEG): container finished" podID="c26e3cb6-4f00-49d9-8bc2-9ec347574580" containerID="de00b39199d31a11cec915e1367446617b6ca45fa3186e7f080356f4a1332db1" exitCode=0 Mar 13 15:34:09 crc kubenswrapper[4907]: I0313 15:34:09.212236 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6c5b6fdc47-txgc2" event={"ID":"c26e3cb6-4f00-49d9-8bc2-9ec347574580","Type":"ContainerDied","Data":"de00b39199d31a11cec915e1367446617b6ca45fa3186e7f080356f4a1332db1"} Mar 13 15:34:09 crc kubenswrapper[4907]: I0313 15:34:09.212260 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6c5b6fdc47-txgc2" event={"ID":"c26e3cb6-4f00-49d9-8bc2-9ec347574580","Type":"ContainerStarted","Data":"5651c9bf703897955b69287b03043d49d9012737130119f792c71677cc9d3fe5"} Mar 13 15:34:09 crc kubenswrapper[4907]: I0313 15:34:09.236345 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-mqwdv" podStartSLOduration=2.236323909 podStartE2EDuration="2.236323909s" podCreationTimestamp="2026-03-13 15:34:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:34:09.235769184 +0000 UTC m=+5348.135556873" watchObservedRunningTime="2026-03-13 15:34:09.236323909 +0000 UTC m=+5348.136111618" Mar 13 15:34:10 crc kubenswrapper[4907]: I0313 15:34:10.220030 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6c5b6fdc47-txgc2" event={"ID":"c26e3cb6-4f00-49d9-8bc2-9ec347574580","Type":"ContainerStarted","Data":"f3459463f090073ba747ae28c0f5e112ee217c5b14b17182f95119ff18ca2ea2"} Mar 13 15:34:10 crc kubenswrapper[4907]: I0313 15:34:10.220654 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6c5b6fdc47-txgc2" Mar 13 15:34:10 crc kubenswrapper[4907]: I0313 15:34:10.248466 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6c5b6fdc47-txgc2" podStartSLOduration=3.248444712 podStartE2EDuration="3.248444712s" podCreationTimestamp="2026-03-13 15:34:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:34:10.24546877 +0000 UTC m=+5349.145256469" watchObservedRunningTime="2026-03-13 15:34:10.248444712 +0000 UTC m=+5349.148232411" Mar 13 15:34:11 crc kubenswrapper[4907]: I0313 15:34:11.475643 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Mar 13 15:34:12 crc kubenswrapper[4907]: I0313 15:34:12.237138 4907 generic.go:334] "Generic (PLEG): container finished" podID="1838c955-2910-4589-a64f-7fe9f865d991" containerID="47fa5a52c38336d054a9aed448aae48b3dd0ee12065b60aaad1abb03635d9d32" exitCode=0 Mar 13 15:34:12 crc kubenswrapper[4907]: I0313 15:34:12.237176 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-mqwdv" event={"ID":"1838c955-2910-4589-a64f-7fe9f865d991","Type":"ContainerDied","Data":"47fa5a52c38336d054a9aed448aae48b3dd0ee12065b60aaad1abb03635d9d32"} Mar 13 15:34:13 crc kubenswrapper[4907]: I0313 15:34:13.624291 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-mqwdv" Mar 13 15:34:13 crc kubenswrapper[4907]: I0313 15:34:13.785432 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1838c955-2910-4589-a64f-7fe9f865d991-combined-ca-bundle\") pod \"1838c955-2910-4589-a64f-7fe9f865d991\" (UID: \"1838c955-2910-4589-a64f-7fe9f865d991\") " Mar 13 15:34:13 crc kubenswrapper[4907]: I0313 15:34:13.785489 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1838c955-2910-4589-a64f-7fe9f865d991-fernet-keys\") pod \"1838c955-2910-4589-a64f-7fe9f865d991\" (UID: \"1838c955-2910-4589-a64f-7fe9f865d991\") " Mar 13 15:34:13 crc kubenswrapper[4907]: I0313 15:34:13.785573 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1838c955-2910-4589-a64f-7fe9f865d991-credential-keys\") pod \"1838c955-2910-4589-a64f-7fe9f865d991\" (UID: \"1838c955-2910-4589-a64f-7fe9f865d991\") " Mar 13 15:34:13 crc kubenswrapper[4907]: I0313 15:34:13.785653 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kf2mt\" (UniqueName: \"kubernetes.io/projected/1838c955-2910-4589-a64f-7fe9f865d991-kube-api-access-kf2mt\") pod \"1838c955-2910-4589-a64f-7fe9f865d991\" (UID: \"1838c955-2910-4589-a64f-7fe9f865d991\") " Mar 13 15:34:13 crc kubenswrapper[4907]: I0313 15:34:13.785686 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1838c955-2910-4589-a64f-7fe9f865d991-scripts\") pod \"1838c955-2910-4589-a64f-7fe9f865d991\" (UID: \"1838c955-2910-4589-a64f-7fe9f865d991\") " Mar 13 15:34:13 crc kubenswrapper[4907]: I0313 15:34:13.785977 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1838c955-2910-4589-a64f-7fe9f865d991-config-data\") pod \"1838c955-2910-4589-a64f-7fe9f865d991\" (UID: \"1838c955-2910-4589-a64f-7fe9f865d991\") " Mar 13 15:34:13 crc kubenswrapper[4907]: I0313 15:34:13.791422 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1838c955-2910-4589-a64f-7fe9f865d991-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "1838c955-2910-4589-a64f-7fe9f865d991" (UID: "1838c955-2910-4589-a64f-7fe9f865d991"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:34:13 crc kubenswrapper[4907]: I0313 15:34:13.791446 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1838c955-2910-4589-a64f-7fe9f865d991-kube-api-access-kf2mt" (OuterVolumeSpecName: "kube-api-access-kf2mt") pod "1838c955-2910-4589-a64f-7fe9f865d991" (UID: "1838c955-2910-4589-a64f-7fe9f865d991"). InnerVolumeSpecName "kube-api-access-kf2mt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:34:13 crc kubenswrapper[4907]: I0313 15:34:13.793712 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1838c955-2910-4589-a64f-7fe9f865d991-scripts" (OuterVolumeSpecName: "scripts") pod "1838c955-2910-4589-a64f-7fe9f865d991" (UID: "1838c955-2910-4589-a64f-7fe9f865d991"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:34:13 crc kubenswrapper[4907]: I0313 15:34:13.794470 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1838c955-2910-4589-a64f-7fe9f865d991-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "1838c955-2910-4589-a64f-7fe9f865d991" (UID: "1838c955-2910-4589-a64f-7fe9f865d991"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:34:13 crc kubenswrapper[4907]: I0313 15:34:13.811077 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1838c955-2910-4589-a64f-7fe9f865d991-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1838c955-2910-4589-a64f-7fe9f865d991" (UID: "1838c955-2910-4589-a64f-7fe9f865d991"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:34:13 crc kubenswrapper[4907]: I0313 15:34:13.814492 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1838c955-2910-4589-a64f-7fe9f865d991-config-data" (OuterVolumeSpecName: "config-data") pod "1838c955-2910-4589-a64f-7fe9f865d991" (UID: "1838c955-2910-4589-a64f-7fe9f865d991"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:34:13 crc kubenswrapper[4907]: I0313 15:34:13.888656 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1838c955-2910-4589-a64f-7fe9f865d991-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 15:34:13 crc kubenswrapper[4907]: I0313 15:34:13.888700 4907 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1838c955-2910-4589-a64f-7fe9f865d991-fernet-keys\") on node \"crc\" DevicePath \"\"" Mar 13 15:34:13 crc kubenswrapper[4907]: I0313 15:34:13.888718 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1838c955-2910-4589-a64f-7fe9f865d991-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 15:34:13 crc kubenswrapper[4907]: I0313 15:34:13.888734 4907 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1838c955-2910-4589-a64f-7fe9f865d991-credential-keys\") on node \"crc\" DevicePath \"\"" Mar 13 15:34:13 crc kubenswrapper[4907]: I0313 15:34:13.888747 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kf2mt\" (UniqueName: \"kubernetes.io/projected/1838c955-2910-4589-a64f-7fe9f865d991-kube-api-access-kf2mt\") on node \"crc\" DevicePath \"\"" Mar 13 15:34:13 crc kubenswrapper[4907]: I0313 15:34:13.888762 4907 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1838c955-2910-4589-a64f-7fe9f865d991-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 15:34:14 crc kubenswrapper[4907]: I0313 15:34:14.256174 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-mqwdv" event={"ID":"1838c955-2910-4589-a64f-7fe9f865d991","Type":"ContainerDied","Data":"caca426201cac3321e9d42ef962695a8f59b09bc5a76701b84aff5acd2e5bb8f"} Mar 13 15:34:14 crc kubenswrapper[4907]: I0313 15:34:14.256225 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="caca426201cac3321e9d42ef962695a8f59b09bc5a76701b84aff5acd2e5bb8f" Mar 13 15:34:14 crc kubenswrapper[4907]: I0313 15:34:14.256272 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-mqwdv" Mar 13 15:34:14 crc kubenswrapper[4907]: I0313 15:34:14.341090 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-mqwdv"] Mar 13 15:34:14 crc kubenswrapper[4907]: I0313 15:34:14.350565 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-mqwdv"] Mar 13 15:34:14 crc kubenswrapper[4907]: I0313 15:34:14.416989 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-mf86w"] Mar 13 15:34:14 crc kubenswrapper[4907]: E0313 15:34:14.417360 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1838c955-2910-4589-a64f-7fe9f865d991" containerName="keystone-bootstrap" Mar 13 15:34:14 crc kubenswrapper[4907]: I0313 15:34:14.417384 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="1838c955-2910-4589-a64f-7fe9f865d991" containerName="keystone-bootstrap" Mar 13 15:34:14 crc kubenswrapper[4907]: I0313 15:34:14.417589 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="1838c955-2910-4589-a64f-7fe9f865d991" containerName="keystone-bootstrap" Mar 13 15:34:14 crc kubenswrapper[4907]: I0313 15:34:14.418256 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-mf86w" Mar 13 15:34:14 crc kubenswrapper[4907]: I0313 15:34:14.422332 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-gjbw7" Mar 13 15:34:14 crc kubenswrapper[4907]: I0313 15:34:14.422554 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Mar 13 15:34:14 crc kubenswrapper[4907]: I0313 15:34:14.423915 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Mar 13 15:34:14 crc kubenswrapper[4907]: I0313 15:34:14.423961 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Mar 13 15:34:14 crc kubenswrapper[4907]: I0313 15:34:14.423970 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Mar 13 15:34:14 crc kubenswrapper[4907]: I0313 15:34:14.428217 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-mf86w"] Mar 13 15:34:14 crc kubenswrapper[4907]: I0313 15:34:14.600031 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0056ce52-b92e-4f2d-b3b0-a7dff01207ff-combined-ca-bundle\") pod \"keystone-bootstrap-mf86w\" (UID: \"0056ce52-b92e-4f2d-b3b0-a7dff01207ff\") " pod="openstack/keystone-bootstrap-mf86w" Mar 13 15:34:14 crc kubenswrapper[4907]: I0313 15:34:14.600095 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0056ce52-b92e-4f2d-b3b0-a7dff01207ff-config-data\") pod \"keystone-bootstrap-mf86w\" (UID: \"0056ce52-b92e-4f2d-b3b0-a7dff01207ff\") " pod="openstack/keystone-bootstrap-mf86w" Mar 13 15:34:14 crc kubenswrapper[4907]: I0313 15:34:14.600121 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/0056ce52-b92e-4f2d-b3b0-a7dff01207ff-fernet-keys\") pod \"keystone-bootstrap-mf86w\" (UID: \"0056ce52-b92e-4f2d-b3b0-a7dff01207ff\") " pod="openstack/keystone-bootstrap-mf86w" Mar 13 15:34:14 crc kubenswrapper[4907]: I0313 15:34:14.600172 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7qbqf\" (UniqueName: \"kubernetes.io/projected/0056ce52-b92e-4f2d-b3b0-a7dff01207ff-kube-api-access-7qbqf\") pod \"keystone-bootstrap-mf86w\" (UID: \"0056ce52-b92e-4f2d-b3b0-a7dff01207ff\") " pod="openstack/keystone-bootstrap-mf86w" Mar 13 15:34:14 crc kubenswrapper[4907]: I0313 15:34:14.600226 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0056ce52-b92e-4f2d-b3b0-a7dff01207ff-scripts\") pod \"keystone-bootstrap-mf86w\" (UID: \"0056ce52-b92e-4f2d-b3b0-a7dff01207ff\") " pod="openstack/keystone-bootstrap-mf86w" Mar 13 15:34:14 crc kubenswrapper[4907]: I0313 15:34:14.600261 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/0056ce52-b92e-4f2d-b3b0-a7dff01207ff-credential-keys\") pod \"keystone-bootstrap-mf86w\" (UID: \"0056ce52-b92e-4f2d-b3b0-a7dff01207ff\") " pod="openstack/keystone-bootstrap-mf86w" Mar 13 15:34:14 crc kubenswrapper[4907]: I0313 15:34:14.700947 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/0056ce52-b92e-4f2d-b3b0-a7dff01207ff-credential-keys\") pod \"keystone-bootstrap-mf86w\" (UID: \"0056ce52-b92e-4f2d-b3b0-a7dff01207ff\") " pod="openstack/keystone-bootstrap-mf86w" Mar 13 15:34:14 crc kubenswrapper[4907]: I0313 15:34:14.701025 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0056ce52-b92e-4f2d-b3b0-a7dff01207ff-combined-ca-bundle\") pod \"keystone-bootstrap-mf86w\" (UID: \"0056ce52-b92e-4f2d-b3b0-a7dff01207ff\") " pod="openstack/keystone-bootstrap-mf86w" Mar 13 15:34:14 crc kubenswrapper[4907]: I0313 15:34:14.701068 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0056ce52-b92e-4f2d-b3b0-a7dff01207ff-config-data\") pod \"keystone-bootstrap-mf86w\" (UID: \"0056ce52-b92e-4f2d-b3b0-a7dff01207ff\") " pod="openstack/keystone-bootstrap-mf86w" Mar 13 15:34:14 crc kubenswrapper[4907]: I0313 15:34:14.701091 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/0056ce52-b92e-4f2d-b3b0-a7dff01207ff-fernet-keys\") pod \"keystone-bootstrap-mf86w\" (UID: \"0056ce52-b92e-4f2d-b3b0-a7dff01207ff\") " pod="openstack/keystone-bootstrap-mf86w" Mar 13 15:34:14 crc kubenswrapper[4907]: I0313 15:34:14.701140 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7qbqf\" (UniqueName: \"kubernetes.io/projected/0056ce52-b92e-4f2d-b3b0-a7dff01207ff-kube-api-access-7qbqf\") pod \"keystone-bootstrap-mf86w\" (UID: \"0056ce52-b92e-4f2d-b3b0-a7dff01207ff\") " pod="openstack/keystone-bootstrap-mf86w" Mar 13 15:34:14 crc kubenswrapper[4907]: I0313 15:34:14.701189 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0056ce52-b92e-4f2d-b3b0-a7dff01207ff-scripts\") pod \"keystone-bootstrap-mf86w\" (UID: \"0056ce52-b92e-4f2d-b3b0-a7dff01207ff\") " pod="openstack/keystone-bootstrap-mf86w" Mar 13 15:34:14 crc kubenswrapper[4907]: I0313 15:34:14.705097 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0056ce52-b92e-4f2d-b3b0-a7dff01207ff-scripts\") pod \"keystone-bootstrap-mf86w\" (UID: \"0056ce52-b92e-4f2d-b3b0-a7dff01207ff\") " pod="openstack/keystone-bootstrap-mf86w" Mar 13 15:34:14 crc kubenswrapper[4907]: I0313 15:34:14.705318 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/0056ce52-b92e-4f2d-b3b0-a7dff01207ff-fernet-keys\") pod \"keystone-bootstrap-mf86w\" (UID: \"0056ce52-b92e-4f2d-b3b0-a7dff01207ff\") " pod="openstack/keystone-bootstrap-mf86w" Mar 13 15:34:14 crc kubenswrapper[4907]: I0313 15:34:14.706210 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0056ce52-b92e-4f2d-b3b0-a7dff01207ff-config-data\") pod \"keystone-bootstrap-mf86w\" (UID: \"0056ce52-b92e-4f2d-b3b0-a7dff01207ff\") " pod="openstack/keystone-bootstrap-mf86w" Mar 13 15:34:14 crc kubenswrapper[4907]: I0313 15:34:14.707069 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0056ce52-b92e-4f2d-b3b0-a7dff01207ff-combined-ca-bundle\") pod \"keystone-bootstrap-mf86w\" (UID: \"0056ce52-b92e-4f2d-b3b0-a7dff01207ff\") " pod="openstack/keystone-bootstrap-mf86w" Mar 13 15:34:14 crc kubenswrapper[4907]: I0313 15:34:14.708245 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/0056ce52-b92e-4f2d-b3b0-a7dff01207ff-credential-keys\") pod \"keystone-bootstrap-mf86w\" (UID: \"0056ce52-b92e-4f2d-b3b0-a7dff01207ff\") " pod="openstack/keystone-bootstrap-mf86w" Mar 13 15:34:14 crc kubenswrapper[4907]: I0313 15:34:14.717669 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7qbqf\" (UniqueName: \"kubernetes.io/projected/0056ce52-b92e-4f2d-b3b0-a7dff01207ff-kube-api-access-7qbqf\") pod \"keystone-bootstrap-mf86w\" (UID: \"0056ce52-b92e-4f2d-b3b0-a7dff01207ff\") " pod="openstack/keystone-bootstrap-mf86w" Mar 13 15:34:14 crc kubenswrapper[4907]: I0313 15:34:14.734707 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-mf86w" Mar 13 15:34:15 crc kubenswrapper[4907]: I0313 15:34:15.177544 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-mf86w"] Mar 13 15:34:15 crc kubenswrapper[4907]: I0313 15:34:15.267373 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-mf86w" event={"ID":"0056ce52-b92e-4f2d-b3b0-a7dff01207ff","Type":"ContainerStarted","Data":"d609585b4e370b2fba0b235b1daafc7fcb8099494f72d1a80e4d09fbc6068dc3"} Mar 13 15:34:15 crc kubenswrapper[4907]: I0313 15:34:15.792419 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1838c955-2910-4589-a64f-7fe9f865d991" path="/var/lib/kubelet/pods/1838c955-2910-4589-a64f-7fe9f865d991/volumes" Mar 13 15:34:16 crc kubenswrapper[4907]: I0313 15:34:16.278237 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-mf86w" event={"ID":"0056ce52-b92e-4f2d-b3b0-a7dff01207ff","Type":"ContainerStarted","Data":"e0b7f12342f112aa71fb3ae1454f11d3b6411a1575b99feb9bc8f40c4e53c905"} Mar 13 15:34:16 crc kubenswrapper[4907]: I0313 15:34:16.301625 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-mf86w" podStartSLOduration=2.301608373 podStartE2EDuration="2.301608373s" podCreationTimestamp="2026-03-13 15:34:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:34:16.29710399 +0000 UTC m=+5355.196891679" watchObservedRunningTime="2026-03-13 15:34:16.301608373 +0000 UTC m=+5355.201396062" Mar 13 15:34:17 crc kubenswrapper[4907]: I0313 15:34:17.757091 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6c5b6fdc47-txgc2" Mar 13 15:34:17 crc kubenswrapper[4907]: I0313 15:34:17.782904 4907 scope.go:117] "RemoveContainer" containerID="592bf9f175f440e943369aab2abbb13ad2c26412cb0e0738c4a33a156b8a5596" Mar 13 15:34:17 crc kubenswrapper[4907]: E0313 15:34:17.783102 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:34:17 crc kubenswrapper[4907]: I0313 15:34:17.838852 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-65b7b7b995-wl96j"] Mar 13 15:34:17 crc kubenswrapper[4907]: I0313 15:34:17.839168 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-65b7b7b995-wl96j" podUID="49e01f82-fb82-4351-a637-3b885e663e0e" containerName="dnsmasq-dns" containerID="cri-o://a3109c92b6a4f47975f53b04558486733a8cd9a2cb768f638be94b37e0295965" gracePeriod=10 Mar 13 15:34:18 crc kubenswrapper[4907]: I0313 15:34:18.300129 4907 generic.go:334] "Generic (PLEG): container finished" podID="49e01f82-fb82-4351-a637-3b885e663e0e" containerID="a3109c92b6a4f47975f53b04558486733a8cd9a2cb768f638be94b37e0295965" exitCode=0 Mar 13 15:34:18 crc kubenswrapper[4907]: I0313 15:34:18.300230 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-65b7b7b995-wl96j" event={"ID":"49e01f82-fb82-4351-a637-3b885e663e0e","Type":"ContainerDied","Data":"a3109c92b6a4f47975f53b04558486733a8cd9a2cb768f638be94b37e0295965"} Mar 13 15:34:18 crc kubenswrapper[4907]: I0313 15:34:18.300271 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-65b7b7b995-wl96j" event={"ID":"49e01f82-fb82-4351-a637-3b885e663e0e","Type":"ContainerDied","Data":"2cc3bc2c8a0a2c514a94d228c418e25f81f45826dc803a6c0203288a2ddf2101"} Mar 13 15:34:18 crc kubenswrapper[4907]: I0313 15:34:18.300290 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2cc3bc2c8a0a2c514a94d228c418e25f81f45826dc803a6c0203288a2ddf2101" Mar 13 15:34:18 crc kubenswrapper[4907]: I0313 15:34:18.302166 4907 generic.go:334] "Generic (PLEG): container finished" podID="0056ce52-b92e-4f2d-b3b0-a7dff01207ff" containerID="e0b7f12342f112aa71fb3ae1454f11d3b6411a1575b99feb9bc8f40c4e53c905" exitCode=0 Mar 13 15:34:18 crc kubenswrapper[4907]: I0313 15:34:18.302199 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-mf86w" event={"ID":"0056ce52-b92e-4f2d-b3b0-a7dff01207ff","Type":"ContainerDied","Data":"e0b7f12342f112aa71fb3ae1454f11d3b6411a1575b99feb9bc8f40c4e53c905"} Mar 13 15:34:18 crc kubenswrapper[4907]: I0313 15:34:18.329485 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-65b7b7b995-wl96j" Mar 13 15:34:18 crc kubenswrapper[4907]: I0313 15:34:18.463016 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lk79c\" (UniqueName: \"kubernetes.io/projected/49e01f82-fb82-4351-a637-3b885e663e0e-kube-api-access-lk79c\") pod \"49e01f82-fb82-4351-a637-3b885e663e0e\" (UID: \"49e01f82-fb82-4351-a637-3b885e663e0e\") " Mar 13 15:34:18 crc kubenswrapper[4907]: I0313 15:34:18.463104 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/49e01f82-fb82-4351-a637-3b885e663e0e-config\") pod \"49e01f82-fb82-4351-a637-3b885e663e0e\" (UID: \"49e01f82-fb82-4351-a637-3b885e663e0e\") " Mar 13 15:34:18 crc kubenswrapper[4907]: I0313 15:34:18.463142 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/49e01f82-fb82-4351-a637-3b885e663e0e-ovsdbserver-sb\") pod \"49e01f82-fb82-4351-a637-3b885e663e0e\" (UID: \"49e01f82-fb82-4351-a637-3b885e663e0e\") " Mar 13 15:34:18 crc kubenswrapper[4907]: I0313 15:34:18.463201 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/49e01f82-fb82-4351-a637-3b885e663e0e-dns-svc\") pod \"49e01f82-fb82-4351-a637-3b885e663e0e\" (UID: \"49e01f82-fb82-4351-a637-3b885e663e0e\") " Mar 13 15:34:18 crc kubenswrapper[4907]: I0313 15:34:18.463252 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/49e01f82-fb82-4351-a637-3b885e663e0e-ovsdbserver-nb\") pod \"49e01f82-fb82-4351-a637-3b885e663e0e\" (UID: \"49e01f82-fb82-4351-a637-3b885e663e0e\") " Mar 13 15:34:18 crc kubenswrapper[4907]: I0313 15:34:18.478154 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49e01f82-fb82-4351-a637-3b885e663e0e-kube-api-access-lk79c" (OuterVolumeSpecName: "kube-api-access-lk79c") pod "49e01f82-fb82-4351-a637-3b885e663e0e" (UID: "49e01f82-fb82-4351-a637-3b885e663e0e"). InnerVolumeSpecName "kube-api-access-lk79c". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:34:18 crc kubenswrapper[4907]: I0313 15:34:18.504423 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49e01f82-fb82-4351-a637-3b885e663e0e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "49e01f82-fb82-4351-a637-3b885e663e0e" (UID: "49e01f82-fb82-4351-a637-3b885e663e0e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:34:18 crc kubenswrapper[4907]: I0313 15:34:18.504474 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49e01f82-fb82-4351-a637-3b885e663e0e-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "49e01f82-fb82-4351-a637-3b885e663e0e" (UID: "49e01f82-fb82-4351-a637-3b885e663e0e"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:34:18 crc kubenswrapper[4907]: I0313 15:34:18.516261 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49e01f82-fb82-4351-a637-3b885e663e0e-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "49e01f82-fb82-4351-a637-3b885e663e0e" (UID: "49e01f82-fb82-4351-a637-3b885e663e0e"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:34:18 crc kubenswrapper[4907]: I0313 15:34:18.525079 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49e01f82-fb82-4351-a637-3b885e663e0e-config" (OuterVolumeSpecName: "config") pod "49e01f82-fb82-4351-a637-3b885e663e0e" (UID: "49e01f82-fb82-4351-a637-3b885e663e0e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:34:18 crc kubenswrapper[4907]: I0313 15:34:18.565469 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/49e01f82-fb82-4351-a637-3b885e663e0e-config\") on node \"crc\" DevicePath \"\"" Mar 13 15:34:18 crc kubenswrapper[4907]: I0313 15:34:18.565506 4907 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/49e01f82-fb82-4351-a637-3b885e663e0e-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 13 15:34:18 crc kubenswrapper[4907]: I0313 15:34:18.565518 4907 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/49e01f82-fb82-4351-a637-3b885e663e0e-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 13 15:34:18 crc kubenswrapper[4907]: I0313 15:34:18.565525 4907 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/49e01f82-fb82-4351-a637-3b885e663e0e-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 13 15:34:18 crc kubenswrapper[4907]: I0313 15:34:18.565535 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lk79c\" (UniqueName: \"kubernetes.io/projected/49e01f82-fb82-4351-a637-3b885e663e0e-kube-api-access-lk79c\") on node \"crc\" DevicePath \"\"" Mar 13 15:34:19 crc kubenswrapper[4907]: I0313 15:34:19.309925 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-65b7b7b995-wl96j" Mar 13 15:34:19 crc kubenswrapper[4907]: I0313 15:34:19.349606 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-65b7b7b995-wl96j"] Mar 13 15:34:19 crc kubenswrapper[4907]: I0313 15:34:19.357395 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-65b7b7b995-wl96j"] Mar 13 15:34:19 crc kubenswrapper[4907]: I0313 15:34:19.627607 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-mf86w" Mar 13 15:34:19 crc kubenswrapper[4907]: I0313 15:34:19.787984 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/0056ce52-b92e-4f2d-b3b0-a7dff01207ff-credential-keys\") pod \"0056ce52-b92e-4f2d-b3b0-a7dff01207ff\" (UID: \"0056ce52-b92e-4f2d-b3b0-a7dff01207ff\") " Mar 13 15:34:19 crc kubenswrapper[4907]: I0313 15:34:19.788024 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7qbqf\" (UniqueName: \"kubernetes.io/projected/0056ce52-b92e-4f2d-b3b0-a7dff01207ff-kube-api-access-7qbqf\") pod \"0056ce52-b92e-4f2d-b3b0-a7dff01207ff\" (UID: \"0056ce52-b92e-4f2d-b3b0-a7dff01207ff\") " Mar 13 15:34:19 crc kubenswrapper[4907]: I0313 15:34:19.788091 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0056ce52-b92e-4f2d-b3b0-a7dff01207ff-combined-ca-bundle\") pod \"0056ce52-b92e-4f2d-b3b0-a7dff01207ff\" (UID: \"0056ce52-b92e-4f2d-b3b0-a7dff01207ff\") " Mar 13 15:34:19 crc kubenswrapper[4907]: I0313 15:34:19.788124 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/0056ce52-b92e-4f2d-b3b0-a7dff01207ff-fernet-keys\") pod \"0056ce52-b92e-4f2d-b3b0-a7dff01207ff\" (UID: \"0056ce52-b92e-4f2d-b3b0-a7dff01207ff\") " Mar 13 15:34:19 crc kubenswrapper[4907]: I0313 15:34:19.788177 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0056ce52-b92e-4f2d-b3b0-a7dff01207ff-scripts\") pod \"0056ce52-b92e-4f2d-b3b0-a7dff01207ff\" (UID: \"0056ce52-b92e-4f2d-b3b0-a7dff01207ff\") " Mar 13 15:34:19 crc kubenswrapper[4907]: I0313 15:34:19.788207 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0056ce52-b92e-4f2d-b3b0-a7dff01207ff-config-data\") pod \"0056ce52-b92e-4f2d-b3b0-a7dff01207ff\" (UID: \"0056ce52-b92e-4f2d-b3b0-a7dff01207ff\") " Mar 13 15:34:19 crc kubenswrapper[4907]: I0313 15:34:19.793478 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0056ce52-b92e-4f2d-b3b0-a7dff01207ff-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "0056ce52-b92e-4f2d-b3b0-a7dff01207ff" (UID: "0056ce52-b92e-4f2d-b3b0-a7dff01207ff"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:34:19 crc kubenswrapper[4907]: I0313 15:34:19.793970 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0056ce52-b92e-4f2d-b3b0-a7dff01207ff-scripts" (OuterVolumeSpecName: "scripts") pod "0056ce52-b92e-4f2d-b3b0-a7dff01207ff" (UID: "0056ce52-b92e-4f2d-b3b0-a7dff01207ff"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:34:19 crc kubenswrapper[4907]: I0313 15:34:19.794342 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0056ce52-b92e-4f2d-b3b0-a7dff01207ff-kube-api-access-7qbqf" (OuterVolumeSpecName: "kube-api-access-7qbqf") pod "0056ce52-b92e-4f2d-b3b0-a7dff01207ff" (UID: "0056ce52-b92e-4f2d-b3b0-a7dff01207ff"). InnerVolumeSpecName "kube-api-access-7qbqf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:34:19 crc kubenswrapper[4907]: I0313 15:34:19.794398 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0056ce52-b92e-4f2d-b3b0-a7dff01207ff-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "0056ce52-b92e-4f2d-b3b0-a7dff01207ff" (UID: "0056ce52-b92e-4f2d-b3b0-a7dff01207ff"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:34:19 crc kubenswrapper[4907]: I0313 15:34:19.799348 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49e01f82-fb82-4351-a637-3b885e663e0e" path="/var/lib/kubelet/pods/49e01f82-fb82-4351-a637-3b885e663e0e/volumes" Mar 13 15:34:19 crc kubenswrapper[4907]: I0313 15:34:19.810042 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0056ce52-b92e-4f2d-b3b0-a7dff01207ff-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0056ce52-b92e-4f2d-b3b0-a7dff01207ff" (UID: "0056ce52-b92e-4f2d-b3b0-a7dff01207ff"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:34:19 crc kubenswrapper[4907]: I0313 15:34:19.811908 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0056ce52-b92e-4f2d-b3b0-a7dff01207ff-config-data" (OuterVolumeSpecName: "config-data") pod "0056ce52-b92e-4f2d-b3b0-a7dff01207ff" (UID: "0056ce52-b92e-4f2d-b3b0-a7dff01207ff"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:34:19 crc kubenswrapper[4907]: I0313 15:34:19.891233 4907 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0056ce52-b92e-4f2d-b3b0-a7dff01207ff-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 15:34:19 crc kubenswrapper[4907]: I0313 15:34:19.891681 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0056ce52-b92e-4f2d-b3b0-a7dff01207ff-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 15:34:19 crc kubenswrapper[4907]: I0313 15:34:19.891710 4907 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/0056ce52-b92e-4f2d-b3b0-a7dff01207ff-credential-keys\") on node \"crc\" DevicePath \"\"" Mar 13 15:34:19 crc kubenswrapper[4907]: I0313 15:34:19.891736 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7qbqf\" (UniqueName: \"kubernetes.io/projected/0056ce52-b92e-4f2d-b3b0-a7dff01207ff-kube-api-access-7qbqf\") on node \"crc\" DevicePath \"\"" Mar 13 15:34:19 crc kubenswrapper[4907]: I0313 15:34:19.891761 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0056ce52-b92e-4f2d-b3b0-a7dff01207ff-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 15:34:19 crc kubenswrapper[4907]: I0313 15:34:19.891783 4907 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/0056ce52-b92e-4f2d-b3b0-a7dff01207ff-fernet-keys\") on node \"crc\" DevicePath \"\"" Mar 13 15:34:20 crc kubenswrapper[4907]: I0313 15:34:20.321770 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-mf86w" event={"ID":"0056ce52-b92e-4f2d-b3b0-a7dff01207ff","Type":"ContainerDied","Data":"d609585b4e370b2fba0b235b1daafc7fcb8099494f72d1a80e4d09fbc6068dc3"} Mar 13 15:34:20 crc kubenswrapper[4907]: I0313 15:34:20.321820 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d609585b4e370b2fba0b235b1daafc7fcb8099494f72d1a80e4d09fbc6068dc3" Mar 13 15:34:20 crc kubenswrapper[4907]: I0313 15:34:20.322082 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-mf86w" Mar 13 15:34:20 crc kubenswrapper[4907]: I0313 15:34:20.416850 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-58b5f55f68-nlrbn"] Mar 13 15:34:20 crc kubenswrapper[4907]: E0313 15:34:20.417958 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49e01f82-fb82-4351-a637-3b885e663e0e" containerName="init" Mar 13 15:34:20 crc kubenswrapper[4907]: I0313 15:34:20.417985 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="49e01f82-fb82-4351-a637-3b885e663e0e" containerName="init" Mar 13 15:34:20 crc kubenswrapper[4907]: E0313 15:34:20.418026 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49e01f82-fb82-4351-a637-3b885e663e0e" containerName="dnsmasq-dns" Mar 13 15:34:20 crc kubenswrapper[4907]: I0313 15:34:20.418034 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="49e01f82-fb82-4351-a637-3b885e663e0e" containerName="dnsmasq-dns" Mar 13 15:34:20 crc kubenswrapper[4907]: E0313 15:34:20.418052 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0056ce52-b92e-4f2d-b3b0-a7dff01207ff" containerName="keystone-bootstrap" Mar 13 15:34:20 crc kubenswrapper[4907]: I0313 15:34:20.418064 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="0056ce52-b92e-4f2d-b3b0-a7dff01207ff" containerName="keystone-bootstrap" Mar 13 15:34:20 crc kubenswrapper[4907]: I0313 15:34:20.418266 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="0056ce52-b92e-4f2d-b3b0-a7dff01207ff" containerName="keystone-bootstrap" Mar 13 15:34:20 crc kubenswrapper[4907]: I0313 15:34:20.418284 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="49e01f82-fb82-4351-a637-3b885e663e0e" containerName="dnsmasq-dns" Mar 13 15:34:20 crc kubenswrapper[4907]: I0313 15:34:20.418926 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-58b5f55f68-nlrbn" Mar 13 15:34:20 crc kubenswrapper[4907]: I0313 15:34:20.420989 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-gjbw7" Mar 13 15:34:20 crc kubenswrapper[4907]: I0313 15:34:20.421002 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Mar 13 15:34:20 crc kubenswrapper[4907]: I0313 15:34:20.421299 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Mar 13 15:34:20 crc kubenswrapper[4907]: I0313 15:34:20.430378 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-58b5f55f68-nlrbn"] Mar 13 15:34:20 crc kubenswrapper[4907]: I0313 15:34:20.432249 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Mar 13 15:34:20 crc kubenswrapper[4907]: I0313 15:34:20.601914 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8b987\" (UniqueName: \"kubernetes.io/projected/121ed53e-99ad-41a2-b8e6-56e4f56520c6-kube-api-access-8b987\") pod \"keystone-58b5f55f68-nlrbn\" (UID: \"121ed53e-99ad-41a2-b8e6-56e4f56520c6\") " pod="openstack/keystone-58b5f55f68-nlrbn" Mar 13 15:34:20 crc kubenswrapper[4907]: I0313 15:34:20.602176 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/121ed53e-99ad-41a2-b8e6-56e4f56520c6-fernet-keys\") pod \"keystone-58b5f55f68-nlrbn\" (UID: \"121ed53e-99ad-41a2-b8e6-56e4f56520c6\") " pod="openstack/keystone-58b5f55f68-nlrbn" Mar 13 15:34:20 crc kubenswrapper[4907]: I0313 15:34:20.602238 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/121ed53e-99ad-41a2-b8e6-56e4f56520c6-credential-keys\") pod \"keystone-58b5f55f68-nlrbn\" (UID: \"121ed53e-99ad-41a2-b8e6-56e4f56520c6\") " pod="openstack/keystone-58b5f55f68-nlrbn" Mar 13 15:34:20 crc kubenswrapper[4907]: I0313 15:34:20.602279 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/121ed53e-99ad-41a2-b8e6-56e4f56520c6-combined-ca-bundle\") pod \"keystone-58b5f55f68-nlrbn\" (UID: \"121ed53e-99ad-41a2-b8e6-56e4f56520c6\") " pod="openstack/keystone-58b5f55f68-nlrbn" Mar 13 15:34:20 crc kubenswrapper[4907]: I0313 15:34:20.602331 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/121ed53e-99ad-41a2-b8e6-56e4f56520c6-scripts\") pod \"keystone-58b5f55f68-nlrbn\" (UID: \"121ed53e-99ad-41a2-b8e6-56e4f56520c6\") " pod="openstack/keystone-58b5f55f68-nlrbn" Mar 13 15:34:20 crc kubenswrapper[4907]: I0313 15:34:20.602397 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/121ed53e-99ad-41a2-b8e6-56e4f56520c6-config-data\") pod \"keystone-58b5f55f68-nlrbn\" (UID: \"121ed53e-99ad-41a2-b8e6-56e4f56520c6\") " pod="openstack/keystone-58b5f55f68-nlrbn" Mar 13 15:34:20 crc kubenswrapper[4907]: I0313 15:34:20.703504 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/121ed53e-99ad-41a2-b8e6-56e4f56520c6-config-data\") pod \"keystone-58b5f55f68-nlrbn\" (UID: \"121ed53e-99ad-41a2-b8e6-56e4f56520c6\") " pod="openstack/keystone-58b5f55f68-nlrbn" Mar 13 15:34:20 crc kubenswrapper[4907]: I0313 15:34:20.703592 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8b987\" (UniqueName: \"kubernetes.io/projected/121ed53e-99ad-41a2-b8e6-56e4f56520c6-kube-api-access-8b987\") pod \"keystone-58b5f55f68-nlrbn\" (UID: \"121ed53e-99ad-41a2-b8e6-56e4f56520c6\") " pod="openstack/keystone-58b5f55f68-nlrbn" Mar 13 15:34:20 crc kubenswrapper[4907]: I0313 15:34:20.703700 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/121ed53e-99ad-41a2-b8e6-56e4f56520c6-fernet-keys\") pod \"keystone-58b5f55f68-nlrbn\" (UID: \"121ed53e-99ad-41a2-b8e6-56e4f56520c6\") " pod="openstack/keystone-58b5f55f68-nlrbn" Mar 13 15:34:20 crc kubenswrapper[4907]: I0313 15:34:20.703732 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/121ed53e-99ad-41a2-b8e6-56e4f56520c6-credential-keys\") pod \"keystone-58b5f55f68-nlrbn\" (UID: \"121ed53e-99ad-41a2-b8e6-56e4f56520c6\") " pod="openstack/keystone-58b5f55f68-nlrbn" Mar 13 15:34:20 crc kubenswrapper[4907]: I0313 15:34:20.703757 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/121ed53e-99ad-41a2-b8e6-56e4f56520c6-combined-ca-bundle\") pod \"keystone-58b5f55f68-nlrbn\" (UID: \"121ed53e-99ad-41a2-b8e6-56e4f56520c6\") " pod="openstack/keystone-58b5f55f68-nlrbn" Mar 13 15:34:20 crc kubenswrapper[4907]: I0313 15:34:20.704302 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/121ed53e-99ad-41a2-b8e6-56e4f56520c6-scripts\") pod \"keystone-58b5f55f68-nlrbn\" (UID: \"121ed53e-99ad-41a2-b8e6-56e4f56520c6\") " pod="openstack/keystone-58b5f55f68-nlrbn" Mar 13 15:34:20 crc kubenswrapper[4907]: I0313 15:34:20.707777 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/121ed53e-99ad-41a2-b8e6-56e4f56520c6-config-data\") pod \"keystone-58b5f55f68-nlrbn\" (UID: \"121ed53e-99ad-41a2-b8e6-56e4f56520c6\") " pod="openstack/keystone-58b5f55f68-nlrbn" Mar 13 15:34:20 crc kubenswrapper[4907]: I0313 15:34:20.714397 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/121ed53e-99ad-41a2-b8e6-56e4f56520c6-scripts\") pod \"keystone-58b5f55f68-nlrbn\" (UID: \"121ed53e-99ad-41a2-b8e6-56e4f56520c6\") " pod="openstack/keystone-58b5f55f68-nlrbn" Mar 13 15:34:20 crc kubenswrapper[4907]: I0313 15:34:20.714396 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/121ed53e-99ad-41a2-b8e6-56e4f56520c6-credential-keys\") pod \"keystone-58b5f55f68-nlrbn\" (UID: \"121ed53e-99ad-41a2-b8e6-56e4f56520c6\") " pod="openstack/keystone-58b5f55f68-nlrbn" Mar 13 15:34:20 crc kubenswrapper[4907]: I0313 15:34:20.714701 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/121ed53e-99ad-41a2-b8e6-56e4f56520c6-combined-ca-bundle\") pod \"keystone-58b5f55f68-nlrbn\" (UID: \"121ed53e-99ad-41a2-b8e6-56e4f56520c6\") " pod="openstack/keystone-58b5f55f68-nlrbn" Mar 13 15:34:20 crc kubenswrapper[4907]: I0313 15:34:20.715411 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/121ed53e-99ad-41a2-b8e6-56e4f56520c6-fernet-keys\") pod \"keystone-58b5f55f68-nlrbn\" (UID: \"121ed53e-99ad-41a2-b8e6-56e4f56520c6\") " pod="openstack/keystone-58b5f55f68-nlrbn" Mar 13 15:34:20 crc kubenswrapper[4907]: I0313 15:34:20.729418 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8b987\" (UniqueName: \"kubernetes.io/projected/121ed53e-99ad-41a2-b8e6-56e4f56520c6-kube-api-access-8b987\") pod \"keystone-58b5f55f68-nlrbn\" (UID: \"121ed53e-99ad-41a2-b8e6-56e4f56520c6\") " pod="openstack/keystone-58b5f55f68-nlrbn" Mar 13 15:34:20 crc kubenswrapper[4907]: I0313 15:34:20.738356 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-58b5f55f68-nlrbn" Mar 13 15:34:21 crc kubenswrapper[4907]: I0313 15:34:21.171018 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-58b5f55f68-nlrbn"] Mar 13 15:34:21 crc kubenswrapper[4907]: I0313 15:34:21.339107 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-58b5f55f68-nlrbn" event={"ID":"121ed53e-99ad-41a2-b8e6-56e4f56520c6","Type":"ContainerStarted","Data":"47fb56f43f328a13f175803cd3cef1e5f392538a9ed6b02f2b0e2451c024895b"} Mar 13 15:34:22 crc kubenswrapper[4907]: I0313 15:34:22.350687 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-58b5f55f68-nlrbn" event={"ID":"121ed53e-99ad-41a2-b8e6-56e4f56520c6","Type":"ContainerStarted","Data":"62d31481988e6763e874bc01945c073a90df2def5706e0b04fd7660a70a1cf83"} Mar 13 15:34:22 crc kubenswrapper[4907]: I0313 15:34:22.350840 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-58b5f55f68-nlrbn" Mar 13 15:34:22 crc kubenswrapper[4907]: I0313 15:34:22.380764 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-58b5f55f68-nlrbn" podStartSLOduration=2.380739015 podStartE2EDuration="2.380739015s" podCreationTimestamp="2026-03-13 15:34:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:34:22.370347011 +0000 UTC m=+5361.270134700" watchObservedRunningTime="2026-03-13 15:34:22.380739015 +0000 UTC m=+5361.280526704" Mar 13 15:34:27 crc kubenswrapper[4907]: I0313 15:34:27.934717 4907 scope.go:117] "RemoveContainer" containerID="206bf50c83686d14437a447b0be4276ceccf75f1f0402647af114c3e1ce67140" Mar 13 15:34:29 crc kubenswrapper[4907]: I0313 15:34:29.781920 4907 scope.go:117] "RemoveContainer" containerID="592bf9f175f440e943369aab2abbb13ad2c26412cb0e0738c4a33a156b8a5596" Mar 13 15:34:29 crc kubenswrapper[4907]: E0313 15:34:29.782478 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:34:44 crc kubenswrapper[4907]: I0313 15:34:44.782589 4907 scope.go:117] "RemoveContainer" containerID="592bf9f175f440e943369aab2abbb13ad2c26412cb0e0738c4a33a156b8a5596" Mar 13 15:34:44 crc kubenswrapper[4907]: E0313 15:34:44.783338 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:34:52 crc kubenswrapper[4907]: I0313 15:34:52.168903 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-58b5f55f68-nlrbn" Mar 13 15:34:56 crc kubenswrapper[4907]: I0313 15:34:56.656203 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Mar 13 15:34:56 crc kubenswrapper[4907]: I0313 15:34:56.657561 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Mar 13 15:34:56 crc kubenswrapper[4907]: I0313 15:34:56.660032 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Mar 13 15:34:56 crc kubenswrapper[4907]: I0313 15:34:56.660045 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Mar 13 15:34:56 crc kubenswrapper[4907]: I0313 15:34:56.668009 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-ff76w" Mar 13 15:34:56 crc kubenswrapper[4907]: I0313 15:34:56.680744 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Mar 13 15:34:56 crc kubenswrapper[4907]: I0313 15:34:56.782707 4907 scope.go:117] "RemoveContainer" containerID="592bf9f175f440e943369aab2abbb13ad2c26412cb0e0738c4a33a156b8a5596" Mar 13 15:34:56 crc kubenswrapper[4907]: I0313 15:34:56.859086 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/74a5c619-2dde-4f75-9106-cad6a39265ce-openstack-config\") pod \"openstackclient\" (UID: \"74a5c619-2dde-4f75-9106-cad6a39265ce\") " pod="openstack/openstackclient" Mar 13 15:34:56 crc kubenswrapper[4907]: I0313 15:34:56.859321 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-85lwz\" (UniqueName: \"kubernetes.io/projected/74a5c619-2dde-4f75-9106-cad6a39265ce-kube-api-access-85lwz\") pod \"openstackclient\" (UID: \"74a5c619-2dde-4f75-9106-cad6a39265ce\") " pod="openstack/openstackclient" Mar 13 15:34:56 crc kubenswrapper[4907]: I0313 15:34:56.859438 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/74a5c619-2dde-4f75-9106-cad6a39265ce-openstack-config-secret\") pod \"openstackclient\" (UID: \"74a5c619-2dde-4f75-9106-cad6a39265ce\") " pod="openstack/openstackclient" Mar 13 15:34:56 crc kubenswrapper[4907]: I0313 15:34:56.963395 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/74a5c619-2dde-4f75-9106-cad6a39265ce-openstack-config-secret\") pod \"openstackclient\" (UID: \"74a5c619-2dde-4f75-9106-cad6a39265ce\") " pod="openstack/openstackclient" Mar 13 15:34:56 crc kubenswrapper[4907]: I0313 15:34:56.964100 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/74a5c619-2dde-4f75-9106-cad6a39265ce-openstack-config\") pod \"openstackclient\" (UID: \"74a5c619-2dde-4f75-9106-cad6a39265ce\") " pod="openstack/openstackclient" Mar 13 15:34:56 crc kubenswrapper[4907]: I0313 15:34:56.964166 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-85lwz\" (UniqueName: \"kubernetes.io/projected/74a5c619-2dde-4f75-9106-cad6a39265ce-kube-api-access-85lwz\") pod \"openstackclient\" (UID: \"74a5c619-2dde-4f75-9106-cad6a39265ce\") " pod="openstack/openstackclient" Mar 13 15:34:56 crc kubenswrapper[4907]: I0313 15:34:56.964934 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/74a5c619-2dde-4f75-9106-cad6a39265ce-openstack-config\") pod \"openstackclient\" (UID: \"74a5c619-2dde-4f75-9106-cad6a39265ce\") " pod="openstack/openstackclient" Mar 13 15:34:56 crc kubenswrapper[4907]: I0313 15:34:56.969507 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/74a5c619-2dde-4f75-9106-cad6a39265ce-openstack-config-secret\") pod \"openstackclient\" (UID: \"74a5c619-2dde-4f75-9106-cad6a39265ce\") " pod="openstack/openstackclient" Mar 13 15:34:56 crc kubenswrapper[4907]: I0313 15:34:56.986545 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-85lwz\" (UniqueName: \"kubernetes.io/projected/74a5c619-2dde-4f75-9106-cad6a39265ce-kube-api-access-85lwz\") pod \"openstackclient\" (UID: \"74a5c619-2dde-4f75-9106-cad6a39265ce\") " pod="openstack/openstackclient" Mar 13 15:34:57 crc kubenswrapper[4907]: I0313 15:34:57.277437 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Mar 13 15:34:57 crc kubenswrapper[4907]: I0313 15:34:57.649953 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerStarted","Data":"90a39cf81ab6b6b036397a9c751a635e5f4cd834dccdc53b2ff02dd0144698a4"} Mar 13 15:34:57 crc kubenswrapper[4907]: I0313 15:34:57.719842 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Mar 13 15:34:57 crc kubenswrapper[4907]: W0313 15:34:57.723027 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod74a5c619_2dde_4f75_9106_cad6a39265ce.slice/crio-677d1ff690530322a0aa44c9900ffa8ce8057cfe59c98e9b289689a68f5fd8ae WatchSource:0}: Error finding container 677d1ff690530322a0aa44c9900ffa8ce8057cfe59c98e9b289689a68f5fd8ae: Status 404 returned error can't find the container with id 677d1ff690530322a0aa44c9900ffa8ce8057cfe59c98e9b289689a68f5fd8ae Mar 13 15:34:58 crc kubenswrapper[4907]: I0313 15:34:58.658822 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"74a5c619-2dde-4f75-9106-cad6a39265ce","Type":"ContainerStarted","Data":"e99b6a0dae893428515991f160f7db931ea2980ec0fe9b55c8e67e61a9c011ff"} Mar 13 15:34:58 crc kubenswrapper[4907]: I0313 15:34:58.659129 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"74a5c619-2dde-4f75-9106-cad6a39265ce","Type":"ContainerStarted","Data":"677d1ff690530322a0aa44c9900ffa8ce8057cfe59c98e9b289689a68f5fd8ae"} Mar 13 15:34:58 crc kubenswrapper[4907]: I0313 15:34:58.679050 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.679003232 podStartE2EDuration="2.679003232s" podCreationTimestamp="2026-03-13 15:34:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:34:58.67455922 +0000 UTC m=+5397.574346949" watchObservedRunningTime="2026-03-13 15:34:58.679003232 +0000 UTC m=+5397.578790921" Mar 13 15:35:50 crc kubenswrapper[4907]: I0313 15:35:50.665758 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-vkhz4"] Mar 13 15:35:50 crc kubenswrapper[4907]: I0313 15:35:50.668231 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vkhz4" Mar 13 15:35:50 crc kubenswrapper[4907]: I0313 15:35:50.685780 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vkhz4"] Mar 13 15:35:50 crc kubenswrapper[4907]: I0313 15:35:50.799810 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7213ba2d-c84c-43f8-99a6-c08f87c5c0aa-utilities\") pod \"community-operators-vkhz4\" (UID: \"7213ba2d-c84c-43f8-99a6-c08f87c5c0aa\") " pod="openshift-marketplace/community-operators-vkhz4" Mar 13 15:35:50 crc kubenswrapper[4907]: I0313 15:35:50.799863 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7213ba2d-c84c-43f8-99a6-c08f87c5c0aa-catalog-content\") pod \"community-operators-vkhz4\" (UID: \"7213ba2d-c84c-43f8-99a6-c08f87c5c0aa\") " pod="openshift-marketplace/community-operators-vkhz4" Mar 13 15:35:50 crc kubenswrapper[4907]: I0313 15:35:50.799920 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w7d4m\" (UniqueName: \"kubernetes.io/projected/7213ba2d-c84c-43f8-99a6-c08f87c5c0aa-kube-api-access-w7d4m\") pod \"community-operators-vkhz4\" (UID: \"7213ba2d-c84c-43f8-99a6-c08f87c5c0aa\") " pod="openshift-marketplace/community-operators-vkhz4" Mar 13 15:35:50 crc kubenswrapper[4907]: I0313 15:35:50.901645 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w7d4m\" (UniqueName: \"kubernetes.io/projected/7213ba2d-c84c-43f8-99a6-c08f87c5c0aa-kube-api-access-w7d4m\") pod \"community-operators-vkhz4\" (UID: \"7213ba2d-c84c-43f8-99a6-c08f87c5c0aa\") " pod="openshift-marketplace/community-operators-vkhz4" Mar 13 15:35:50 crc kubenswrapper[4907]: I0313 15:35:50.902095 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7213ba2d-c84c-43f8-99a6-c08f87c5c0aa-utilities\") pod \"community-operators-vkhz4\" (UID: \"7213ba2d-c84c-43f8-99a6-c08f87c5c0aa\") " pod="openshift-marketplace/community-operators-vkhz4" Mar 13 15:35:50 crc kubenswrapper[4907]: I0313 15:35:50.902417 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7213ba2d-c84c-43f8-99a6-c08f87c5c0aa-catalog-content\") pod \"community-operators-vkhz4\" (UID: \"7213ba2d-c84c-43f8-99a6-c08f87c5c0aa\") " pod="openshift-marketplace/community-operators-vkhz4" Mar 13 15:35:50 crc kubenswrapper[4907]: I0313 15:35:50.902847 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7213ba2d-c84c-43f8-99a6-c08f87c5c0aa-catalog-content\") pod \"community-operators-vkhz4\" (UID: \"7213ba2d-c84c-43f8-99a6-c08f87c5c0aa\") " pod="openshift-marketplace/community-operators-vkhz4" Mar 13 15:35:50 crc kubenswrapper[4907]: I0313 15:35:50.902845 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7213ba2d-c84c-43f8-99a6-c08f87c5c0aa-utilities\") pod \"community-operators-vkhz4\" (UID: \"7213ba2d-c84c-43f8-99a6-c08f87c5c0aa\") " pod="openshift-marketplace/community-operators-vkhz4" Mar 13 15:35:50 crc kubenswrapper[4907]: I0313 15:35:50.920221 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w7d4m\" (UniqueName: \"kubernetes.io/projected/7213ba2d-c84c-43f8-99a6-c08f87c5c0aa-kube-api-access-w7d4m\") pod \"community-operators-vkhz4\" (UID: \"7213ba2d-c84c-43f8-99a6-c08f87c5c0aa\") " pod="openshift-marketplace/community-operators-vkhz4" Mar 13 15:35:50 crc kubenswrapper[4907]: I0313 15:35:50.987602 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vkhz4" Mar 13 15:35:51 crc kubenswrapper[4907]: I0313 15:35:51.513795 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vkhz4"] Mar 13 15:35:52 crc kubenswrapper[4907]: I0313 15:35:52.100334 4907 generic.go:334] "Generic (PLEG): container finished" podID="7213ba2d-c84c-43f8-99a6-c08f87c5c0aa" containerID="378658932bb4f93263f9c7c834e5eb6a55511ab6e84b9fd5f0fe1ea37d58141a" exitCode=0 Mar 13 15:35:52 crc kubenswrapper[4907]: I0313 15:35:52.100388 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vkhz4" event={"ID":"7213ba2d-c84c-43f8-99a6-c08f87c5c0aa","Type":"ContainerDied","Data":"378658932bb4f93263f9c7c834e5eb6a55511ab6e84b9fd5f0fe1ea37d58141a"} Mar 13 15:35:52 crc kubenswrapper[4907]: I0313 15:35:52.100617 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vkhz4" event={"ID":"7213ba2d-c84c-43f8-99a6-c08f87c5c0aa","Type":"ContainerStarted","Data":"d2d3846d126727c38882c055157fb114b51e62e91302d46c000f07a01a3197ba"} Mar 13 15:35:53 crc kubenswrapper[4907]: I0313 15:35:53.114553 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vkhz4" event={"ID":"7213ba2d-c84c-43f8-99a6-c08f87c5c0aa","Type":"ContainerStarted","Data":"57388821445c4b768505f257c2ec6fe32c44ef455b2d02ef9106d2a78dd49da5"} Mar 13 15:35:54 crc kubenswrapper[4907]: I0313 15:35:54.124819 4907 generic.go:334] "Generic (PLEG): container finished" podID="7213ba2d-c84c-43f8-99a6-c08f87c5c0aa" containerID="57388821445c4b768505f257c2ec6fe32c44ef455b2d02ef9106d2a78dd49da5" exitCode=0 Mar 13 15:35:54 crc kubenswrapper[4907]: I0313 15:35:54.124962 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vkhz4" event={"ID":"7213ba2d-c84c-43f8-99a6-c08f87c5c0aa","Type":"ContainerDied","Data":"57388821445c4b768505f257c2ec6fe32c44ef455b2d02ef9106d2a78dd49da5"} Mar 13 15:35:55 crc kubenswrapper[4907]: I0313 15:35:55.133932 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vkhz4" event={"ID":"7213ba2d-c84c-43f8-99a6-c08f87c5c0aa","Type":"ContainerStarted","Data":"b0d053e1bf868544e4714cfc718cdfdb940c35699b616905f2247bf747bbb8a8"} Mar 13 15:35:55 crc kubenswrapper[4907]: I0313 15:35:55.156010 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-vkhz4" podStartSLOduration=2.527085146 podStartE2EDuration="5.155987781s" podCreationTimestamp="2026-03-13 15:35:50 +0000 UTC" firstStartedPulling="2026-03-13 15:35:52.102850594 +0000 UTC m=+5451.002638283" lastFinishedPulling="2026-03-13 15:35:54.731753229 +0000 UTC m=+5453.631540918" observedRunningTime="2026-03-13 15:35:55.149224606 +0000 UTC m=+5454.049012295" watchObservedRunningTime="2026-03-13 15:35:55.155987781 +0000 UTC m=+5454.055775460" Mar 13 15:36:00 crc kubenswrapper[4907]: I0313 15:36:00.147552 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556936-bwsf9"] Mar 13 15:36:00 crc kubenswrapper[4907]: I0313 15:36:00.150546 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556936-bwsf9" Mar 13 15:36:00 crc kubenswrapper[4907]: I0313 15:36:00.153683 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 15:36:00 crc kubenswrapper[4907]: I0313 15:36:00.155099 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 15:36:00 crc kubenswrapper[4907]: I0313 15:36:00.158583 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 15:36:00 crc kubenswrapper[4907]: I0313 15:36:00.172947 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556936-bwsf9"] Mar 13 15:36:00 crc kubenswrapper[4907]: I0313 15:36:00.252251 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ngwfh\" (UniqueName: \"kubernetes.io/projected/38ab6eb6-9c07-41f4-87e4-d74cb4e5aff7-kube-api-access-ngwfh\") pod \"auto-csr-approver-29556936-bwsf9\" (UID: \"38ab6eb6-9c07-41f4-87e4-d74cb4e5aff7\") " pod="openshift-infra/auto-csr-approver-29556936-bwsf9" Mar 13 15:36:00 crc kubenswrapper[4907]: I0313 15:36:00.354763 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ngwfh\" (UniqueName: \"kubernetes.io/projected/38ab6eb6-9c07-41f4-87e4-d74cb4e5aff7-kube-api-access-ngwfh\") pod \"auto-csr-approver-29556936-bwsf9\" (UID: \"38ab6eb6-9c07-41f4-87e4-d74cb4e5aff7\") " pod="openshift-infra/auto-csr-approver-29556936-bwsf9" Mar 13 15:36:00 crc kubenswrapper[4907]: I0313 15:36:00.398501 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ngwfh\" (UniqueName: \"kubernetes.io/projected/38ab6eb6-9c07-41f4-87e4-d74cb4e5aff7-kube-api-access-ngwfh\") pod \"auto-csr-approver-29556936-bwsf9\" (UID: \"38ab6eb6-9c07-41f4-87e4-d74cb4e5aff7\") " pod="openshift-infra/auto-csr-approver-29556936-bwsf9" Mar 13 15:36:00 crc kubenswrapper[4907]: I0313 15:36:00.477572 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556936-bwsf9" Mar 13 15:36:00 crc kubenswrapper[4907]: I0313 15:36:00.903057 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556936-bwsf9"] Mar 13 15:36:00 crc kubenswrapper[4907]: I0313 15:36:00.988530 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-vkhz4" Mar 13 15:36:00 crc kubenswrapper[4907]: I0313 15:36:00.988570 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-vkhz4" Mar 13 15:36:01 crc kubenswrapper[4907]: I0313 15:36:01.028008 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-vkhz4" Mar 13 15:36:01 crc kubenswrapper[4907]: I0313 15:36:01.191048 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556936-bwsf9" event={"ID":"38ab6eb6-9c07-41f4-87e4-d74cb4e5aff7","Type":"ContainerStarted","Data":"8ffd45b87adb147d65f6690cb310171e5230c18b1399582fbe6807ed0ee79621"} Mar 13 15:36:01 crc kubenswrapper[4907]: I0313 15:36:01.228700 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-vkhz4" Mar 13 15:36:02 crc kubenswrapper[4907]: I0313 15:36:02.164948 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vkhz4"] Mar 13 15:36:03 crc kubenswrapper[4907]: I0313 15:36:03.216672 4907 generic.go:334] "Generic (PLEG): container finished" podID="38ab6eb6-9c07-41f4-87e4-d74cb4e5aff7" containerID="a02ae219ab9b5a1fd678d76ac3b1c7e41a88bec1b6ae3e7a16334b35c97e0864" exitCode=0 Mar 13 15:36:03 crc kubenswrapper[4907]: I0313 15:36:03.216789 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556936-bwsf9" event={"ID":"38ab6eb6-9c07-41f4-87e4-d74cb4e5aff7","Type":"ContainerDied","Data":"a02ae219ab9b5a1fd678d76ac3b1c7e41a88bec1b6ae3e7a16334b35c97e0864"} Mar 13 15:36:03 crc kubenswrapper[4907]: I0313 15:36:03.217617 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-vkhz4" podUID="7213ba2d-c84c-43f8-99a6-c08f87c5c0aa" containerName="registry-server" containerID="cri-o://b0d053e1bf868544e4714cfc718cdfdb940c35699b616905f2247bf747bbb8a8" gracePeriod=2 Mar 13 15:36:03 crc kubenswrapper[4907]: I0313 15:36:03.649855 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vkhz4" Mar 13 15:36:03 crc kubenswrapper[4907]: I0313 15:36:03.709992 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7213ba2d-c84c-43f8-99a6-c08f87c5c0aa-utilities\") pod \"7213ba2d-c84c-43f8-99a6-c08f87c5c0aa\" (UID: \"7213ba2d-c84c-43f8-99a6-c08f87c5c0aa\") " Mar 13 15:36:03 crc kubenswrapper[4907]: I0313 15:36:03.710138 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7d4m\" (UniqueName: \"kubernetes.io/projected/7213ba2d-c84c-43f8-99a6-c08f87c5c0aa-kube-api-access-w7d4m\") pod \"7213ba2d-c84c-43f8-99a6-c08f87c5c0aa\" (UID: \"7213ba2d-c84c-43f8-99a6-c08f87c5c0aa\") " Mar 13 15:36:03 crc kubenswrapper[4907]: I0313 15:36:03.710313 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7213ba2d-c84c-43f8-99a6-c08f87c5c0aa-catalog-content\") pod \"7213ba2d-c84c-43f8-99a6-c08f87c5c0aa\" (UID: \"7213ba2d-c84c-43f8-99a6-c08f87c5c0aa\") " Mar 13 15:36:03 crc kubenswrapper[4907]: I0313 15:36:03.711166 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7213ba2d-c84c-43f8-99a6-c08f87c5c0aa-utilities" (OuterVolumeSpecName: "utilities") pod "7213ba2d-c84c-43f8-99a6-c08f87c5c0aa" (UID: "7213ba2d-c84c-43f8-99a6-c08f87c5c0aa"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:36:03 crc kubenswrapper[4907]: I0313 15:36:03.717341 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7213ba2d-c84c-43f8-99a6-c08f87c5c0aa-kube-api-access-w7d4m" (OuterVolumeSpecName: "kube-api-access-w7d4m") pod "7213ba2d-c84c-43f8-99a6-c08f87c5c0aa" (UID: "7213ba2d-c84c-43f8-99a6-c08f87c5c0aa"). InnerVolumeSpecName "kube-api-access-w7d4m". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:36:03 crc kubenswrapper[4907]: I0313 15:36:03.769740 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7213ba2d-c84c-43f8-99a6-c08f87c5c0aa-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7213ba2d-c84c-43f8-99a6-c08f87c5c0aa" (UID: "7213ba2d-c84c-43f8-99a6-c08f87c5c0aa"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:36:03 crc kubenswrapper[4907]: I0313 15:36:03.812799 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7d4m\" (UniqueName: \"kubernetes.io/projected/7213ba2d-c84c-43f8-99a6-c08f87c5c0aa-kube-api-access-w7d4m\") on node \"crc\" DevicePath \"\"" Mar 13 15:36:03 crc kubenswrapper[4907]: I0313 15:36:03.812825 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7213ba2d-c84c-43f8-99a6-c08f87c5c0aa-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 15:36:03 crc kubenswrapper[4907]: I0313 15:36:03.812834 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7213ba2d-c84c-43f8-99a6-c08f87c5c0aa-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 15:36:04 crc kubenswrapper[4907]: I0313 15:36:04.225957 4907 generic.go:334] "Generic (PLEG): container finished" podID="7213ba2d-c84c-43f8-99a6-c08f87c5c0aa" containerID="b0d053e1bf868544e4714cfc718cdfdb940c35699b616905f2247bf747bbb8a8" exitCode=0 Mar 13 15:36:04 crc kubenswrapper[4907]: I0313 15:36:04.226019 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vkhz4" Mar 13 15:36:04 crc kubenswrapper[4907]: I0313 15:36:04.226027 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vkhz4" event={"ID":"7213ba2d-c84c-43f8-99a6-c08f87c5c0aa","Type":"ContainerDied","Data":"b0d053e1bf868544e4714cfc718cdfdb940c35699b616905f2247bf747bbb8a8"} Mar 13 15:36:04 crc kubenswrapper[4907]: I0313 15:36:04.226105 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vkhz4" event={"ID":"7213ba2d-c84c-43f8-99a6-c08f87c5c0aa","Type":"ContainerDied","Data":"d2d3846d126727c38882c055157fb114b51e62e91302d46c000f07a01a3197ba"} Mar 13 15:36:04 crc kubenswrapper[4907]: I0313 15:36:04.226135 4907 scope.go:117] "RemoveContainer" containerID="b0d053e1bf868544e4714cfc718cdfdb940c35699b616905f2247bf747bbb8a8" Mar 13 15:36:04 crc kubenswrapper[4907]: I0313 15:36:04.252988 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vkhz4"] Mar 13 15:36:04 crc kubenswrapper[4907]: I0313 15:36:04.258126 4907 scope.go:117] "RemoveContainer" containerID="57388821445c4b768505f257c2ec6fe32c44ef455b2d02ef9106d2a78dd49da5" Mar 13 15:36:04 crc kubenswrapper[4907]: I0313 15:36:04.261456 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-vkhz4"] Mar 13 15:36:04 crc kubenswrapper[4907]: I0313 15:36:04.306937 4907 scope.go:117] "RemoveContainer" containerID="378658932bb4f93263f9c7c834e5eb6a55511ab6e84b9fd5f0fe1ea37d58141a" Mar 13 15:36:04 crc kubenswrapper[4907]: I0313 15:36:04.323472 4907 scope.go:117] "RemoveContainer" containerID="b0d053e1bf868544e4714cfc718cdfdb940c35699b616905f2247bf747bbb8a8" Mar 13 15:36:04 crc kubenswrapper[4907]: E0313 15:36:04.324509 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b0d053e1bf868544e4714cfc718cdfdb940c35699b616905f2247bf747bbb8a8\": container with ID starting with b0d053e1bf868544e4714cfc718cdfdb940c35699b616905f2247bf747bbb8a8 not found: ID does not exist" containerID="b0d053e1bf868544e4714cfc718cdfdb940c35699b616905f2247bf747bbb8a8" Mar 13 15:36:04 crc kubenswrapper[4907]: I0313 15:36:04.324557 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b0d053e1bf868544e4714cfc718cdfdb940c35699b616905f2247bf747bbb8a8"} err="failed to get container status \"b0d053e1bf868544e4714cfc718cdfdb940c35699b616905f2247bf747bbb8a8\": rpc error: code = NotFound desc = could not find container \"b0d053e1bf868544e4714cfc718cdfdb940c35699b616905f2247bf747bbb8a8\": container with ID starting with b0d053e1bf868544e4714cfc718cdfdb940c35699b616905f2247bf747bbb8a8 not found: ID does not exist" Mar 13 15:36:04 crc kubenswrapper[4907]: I0313 15:36:04.324590 4907 scope.go:117] "RemoveContainer" containerID="57388821445c4b768505f257c2ec6fe32c44ef455b2d02ef9106d2a78dd49da5" Mar 13 15:36:04 crc kubenswrapper[4907]: E0313 15:36:04.324997 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"57388821445c4b768505f257c2ec6fe32c44ef455b2d02ef9106d2a78dd49da5\": container with ID starting with 57388821445c4b768505f257c2ec6fe32c44ef455b2d02ef9106d2a78dd49da5 not found: ID does not exist" containerID="57388821445c4b768505f257c2ec6fe32c44ef455b2d02ef9106d2a78dd49da5" Mar 13 15:36:04 crc kubenswrapper[4907]: I0313 15:36:04.325097 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"57388821445c4b768505f257c2ec6fe32c44ef455b2d02ef9106d2a78dd49da5"} err="failed to get container status \"57388821445c4b768505f257c2ec6fe32c44ef455b2d02ef9106d2a78dd49da5\": rpc error: code = NotFound desc = could not find container \"57388821445c4b768505f257c2ec6fe32c44ef455b2d02ef9106d2a78dd49da5\": container with ID starting with 57388821445c4b768505f257c2ec6fe32c44ef455b2d02ef9106d2a78dd49da5 not found: ID does not exist" Mar 13 15:36:04 crc kubenswrapper[4907]: I0313 15:36:04.325142 4907 scope.go:117] "RemoveContainer" containerID="378658932bb4f93263f9c7c834e5eb6a55511ab6e84b9fd5f0fe1ea37d58141a" Mar 13 15:36:04 crc kubenswrapper[4907]: E0313 15:36:04.325837 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"378658932bb4f93263f9c7c834e5eb6a55511ab6e84b9fd5f0fe1ea37d58141a\": container with ID starting with 378658932bb4f93263f9c7c834e5eb6a55511ab6e84b9fd5f0fe1ea37d58141a not found: ID does not exist" containerID="378658932bb4f93263f9c7c834e5eb6a55511ab6e84b9fd5f0fe1ea37d58141a" Mar 13 15:36:04 crc kubenswrapper[4907]: I0313 15:36:04.325869 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"378658932bb4f93263f9c7c834e5eb6a55511ab6e84b9fd5f0fe1ea37d58141a"} err="failed to get container status \"378658932bb4f93263f9c7c834e5eb6a55511ab6e84b9fd5f0fe1ea37d58141a\": rpc error: code = NotFound desc = could not find container \"378658932bb4f93263f9c7c834e5eb6a55511ab6e84b9fd5f0fe1ea37d58141a\": container with ID starting with 378658932bb4f93263f9c7c834e5eb6a55511ab6e84b9fd5f0fe1ea37d58141a not found: ID does not exist" Mar 13 15:36:04 crc kubenswrapper[4907]: I0313 15:36:04.572595 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556936-bwsf9" Mar 13 15:36:04 crc kubenswrapper[4907]: I0313 15:36:04.628788 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngwfh\" (UniqueName: \"kubernetes.io/projected/38ab6eb6-9c07-41f4-87e4-d74cb4e5aff7-kube-api-access-ngwfh\") pod \"38ab6eb6-9c07-41f4-87e4-d74cb4e5aff7\" (UID: \"38ab6eb6-9c07-41f4-87e4-d74cb4e5aff7\") " Mar 13 15:36:04 crc kubenswrapper[4907]: I0313 15:36:04.635110 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/38ab6eb6-9c07-41f4-87e4-d74cb4e5aff7-kube-api-access-ngwfh" (OuterVolumeSpecName: "kube-api-access-ngwfh") pod "38ab6eb6-9c07-41f4-87e4-d74cb4e5aff7" (UID: "38ab6eb6-9c07-41f4-87e4-d74cb4e5aff7"). InnerVolumeSpecName "kube-api-access-ngwfh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:36:04 crc kubenswrapper[4907]: I0313 15:36:04.730755 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngwfh\" (UniqueName: \"kubernetes.io/projected/38ab6eb6-9c07-41f4-87e4-d74cb4e5aff7-kube-api-access-ngwfh\") on node \"crc\" DevicePath \"\"" Mar 13 15:36:05 crc kubenswrapper[4907]: I0313 15:36:05.235174 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556936-bwsf9" event={"ID":"38ab6eb6-9c07-41f4-87e4-d74cb4e5aff7","Type":"ContainerDied","Data":"8ffd45b87adb147d65f6690cb310171e5230c18b1399582fbe6807ed0ee79621"} Mar 13 15:36:05 crc kubenswrapper[4907]: I0313 15:36:05.235215 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8ffd45b87adb147d65f6690cb310171e5230c18b1399582fbe6807ed0ee79621" Mar 13 15:36:05 crc kubenswrapper[4907]: I0313 15:36:05.235231 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556936-bwsf9" Mar 13 15:36:05 crc kubenswrapper[4907]: I0313 15:36:05.662568 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556930-ndtqs"] Mar 13 15:36:05 crc kubenswrapper[4907]: I0313 15:36:05.670214 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556930-ndtqs"] Mar 13 15:36:05 crc kubenswrapper[4907]: I0313 15:36:05.791797 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7213ba2d-c84c-43f8-99a6-c08f87c5c0aa" path="/var/lib/kubelet/pods/7213ba2d-c84c-43f8-99a6-c08f87c5c0aa/volumes" Mar 13 15:36:05 crc kubenswrapper[4907]: I0313 15:36:05.792589 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="df83f491-eb7d-4e1c-b814-28cc7ef9abee" path="/var/lib/kubelet/pods/df83f491-eb7d-4e1c-b814-28cc7ef9abee/volumes" Mar 13 15:36:28 crc kubenswrapper[4907]: I0313 15:36:28.062300 4907 scope.go:117] "RemoveContainer" containerID="9c966e9f83d2d89ea300010e1e5e03c7e34fbdc9099449a4ffcb481299d87028" Mar 13 15:36:40 crc kubenswrapper[4907]: I0313 15:36:40.463689 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-4jnct"] Mar 13 15:36:40 crc kubenswrapper[4907]: E0313 15:36:40.464735 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7213ba2d-c84c-43f8-99a6-c08f87c5c0aa" containerName="extract-content" Mar 13 15:36:40 crc kubenswrapper[4907]: I0313 15:36:40.464752 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="7213ba2d-c84c-43f8-99a6-c08f87c5c0aa" containerName="extract-content" Mar 13 15:36:40 crc kubenswrapper[4907]: E0313 15:36:40.464793 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7213ba2d-c84c-43f8-99a6-c08f87c5c0aa" containerName="extract-utilities" Mar 13 15:36:40 crc kubenswrapper[4907]: I0313 15:36:40.464802 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="7213ba2d-c84c-43f8-99a6-c08f87c5c0aa" containerName="extract-utilities" Mar 13 15:36:40 crc kubenswrapper[4907]: E0313 15:36:40.464833 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38ab6eb6-9c07-41f4-87e4-d74cb4e5aff7" containerName="oc" Mar 13 15:36:40 crc kubenswrapper[4907]: I0313 15:36:40.464842 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="38ab6eb6-9c07-41f4-87e4-d74cb4e5aff7" containerName="oc" Mar 13 15:36:40 crc kubenswrapper[4907]: E0313 15:36:40.464858 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7213ba2d-c84c-43f8-99a6-c08f87c5c0aa" containerName="registry-server" Mar 13 15:36:40 crc kubenswrapper[4907]: I0313 15:36:40.464866 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="7213ba2d-c84c-43f8-99a6-c08f87c5c0aa" containerName="registry-server" Mar 13 15:36:40 crc kubenswrapper[4907]: I0313 15:36:40.465110 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="7213ba2d-c84c-43f8-99a6-c08f87c5c0aa" containerName="registry-server" Mar 13 15:36:40 crc kubenswrapper[4907]: I0313 15:36:40.465136 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="38ab6eb6-9c07-41f4-87e4-d74cb4e5aff7" containerName="oc" Mar 13 15:36:40 crc kubenswrapper[4907]: I0313 15:36:40.465938 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-4jnct" Mar 13 15:36:40 crc kubenswrapper[4907]: I0313 15:36:40.471417 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-b654-account-create-update-4jjkj"] Mar 13 15:36:40 crc kubenswrapper[4907]: I0313 15:36:40.473176 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-b654-account-create-update-4jjkj" Mar 13 15:36:40 crc kubenswrapper[4907]: I0313 15:36:40.475892 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Mar 13 15:36:40 crc kubenswrapper[4907]: I0313 15:36:40.499508 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-4jnct"] Mar 13 15:36:40 crc kubenswrapper[4907]: I0313 15:36:40.526713 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-b654-account-create-update-4jjkj"] Mar 13 15:36:40 crc kubenswrapper[4907]: I0313 15:36:40.670083 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1e73852b-fdd0-4cd2-a268-96f6ed551011-operator-scripts\") pod \"barbican-b654-account-create-update-4jjkj\" (UID: \"1e73852b-fdd0-4cd2-a268-96f6ed551011\") " pod="openstack/barbican-b654-account-create-update-4jjkj" Mar 13 15:36:40 crc kubenswrapper[4907]: I0313 15:36:40.670168 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rwpg2\" (UniqueName: \"kubernetes.io/projected/1e73852b-fdd0-4cd2-a268-96f6ed551011-kube-api-access-rwpg2\") pod \"barbican-b654-account-create-update-4jjkj\" (UID: \"1e73852b-fdd0-4cd2-a268-96f6ed551011\") " pod="openstack/barbican-b654-account-create-update-4jjkj" Mar 13 15:36:40 crc kubenswrapper[4907]: I0313 15:36:40.670219 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/523d3b7c-7098-44bc-92a6-d48f3ae49b65-operator-scripts\") pod \"barbican-db-create-4jnct\" (UID: \"523d3b7c-7098-44bc-92a6-d48f3ae49b65\") " pod="openstack/barbican-db-create-4jnct" Mar 13 15:36:40 crc kubenswrapper[4907]: I0313 15:36:40.670313 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zkhl8\" (UniqueName: \"kubernetes.io/projected/523d3b7c-7098-44bc-92a6-d48f3ae49b65-kube-api-access-zkhl8\") pod \"barbican-db-create-4jnct\" (UID: \"523d3b7c-7098-44bc-92a6-d48f3ae49b65\") " pod="openstack/barbican-db-create-4jnct" Mar 13 15:36:40 crc kubenswrapper[4907]: I0313 15:36:40.771664 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zkhl8\" (UniqueName: \"kubernetes.io/projected/523d3b7c-7098-44bc-92a6-d48f3ae49b65-kube-api-access-zkhl8\") pod \"barbican-db-create-4jnct\" (UID: \"523d3b7c-7098-44bc-92a6-d48f3ae49b65\") " pod="openstack/barbican-db-create-4jnct" Mar 13 15:36:40 crc kubenswrapper[4907]: I0313 15:36:40.771984 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1e73852b-fdd0-4cd2-a268-96f6ed551011-operator-scripts\") pod \"barbican-b654-account-create-update-4jjkj\" (UID: \"1e73852b-fdd0-4cd2-a268-96f6ed551011\") " pod="openstack/barbican-b654-account-create-update-4jjkj" Mar 13 15:36:40 crc kubenswrapper[4907]: I0313 15:36:40.772107 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rwpg2\" (UniqueName: \"kubernetes.io/projected/1e73852b-fdd0-4cd2-a268-96f6ed551011-kube-api-access-rwpg2\") pod \"barbican-b654-account-create-update-4jjkj\" (UID: \"1e73852b-fdd0-4cd2-a268-96f6ed551011\") " pod="openstack/barbican-b654-account-create-update-4jjkj" Mar 13 15:36:40 crc kubenswrapper[4907]: I0313 15:36:40.772204 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/523d3b7c-7098-44bc-92a6-d48f3ae49b65-operator-scripts\") pod \"barbican-db-create-4jnct\" (UID: \"523d3b7c-7098-44bc-92a6-d48f3ae49b65\") " pod="openstack/barbican-db-create-4jnct" Mar 13 15:36:40 crc kubenswrapper[4907]: I0313 15:36:40.772783 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1e73852b-fdd0-4cd2-a268-96f6ed551011-operator-scripts\") pod \"barbican-b654-account-create-update-4jjkj\" (UID: \"1e73852b-fdd0-4cd2-a268-96f6ed551011\") " pod="openstack/barbican-b654-account-create-update-4jjkj" Mar 13 15:36:40 crc kubenswrapper[4907]: I0313 15:36:40.772871 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/523d3b7c-7098-44bc-92a6-d48f3ae49b65-operator-scripts\") pod \"barbican-db-create-4jnct\" (UID: \"523d3b7c-7098-44bc-92a6-d48f3ae49b65\") " pod="openstack/barbican-db-create-4jnct" Mar 13 15:36:40 crc kubenswrapper[4907]: I0313 15:36:40.789208 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rwpg2\" (UniqueName: \"kubernetes.io/projected/1e73852b-fdd0-4cd2-a268-96f6ed551011-kube-api-access-rwpg2\") pod \"barbican-b654-account-create-update-4jjkj\" (UID: \"1e73852b-fdd0-4cd2-a268-96f6ed551011\") " pod="openstack/barbican-b654-account-create-update-4jjkj" Mar 13 15:36:40 crc kubenswrapper[4907]: I0313 15:36:40.795545 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zkhl8\" (UniqueName: \"kubernetes.io/projected/523d3b7c-7098-44bc-92a6-d48f3ae49b65-kube-api-access-zkhl8\") pod \"barbican-db-create-4jnct\" (UID: \"523d3b7c-7098-44bc-92a6-d48f3ae49b65\") " pod="openstack/barbican-db-create-4jnct" Mar 13 15:36:40 crc kubenswrapper[4907]: I0313 15:36:40.795974 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-4jnct" Mar 13 15:36:40 crc kubenswrapper[4907]: I0313 15:36:40.812145 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-b654-account-create-update-4jjkj" Mar 13 15:36:41 crc kubenswrapper[4907]: I0313 15:36:41.242901 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-b654-account-create-update-4jjkj"] Mar 13 15:36:41 crc kubenswrapper[4907]: I0313 15:36:41.323574 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-4jnct"] Mar 13 15:36:41 crc kubenswrapper[4907]: W0313 15:36:41.334486 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod523d3b7c_7098_44bc_92a6_d48f3ae49b65.slice/crio-56a67057374a475d911f887833cd0ea7420af1041fbdd3d80eddab806d07786a WatchSource:0}: Error finding container 56a67057374a475d911f887833cd0ea7420af1041fbdd3d80eddab806d07786a: Status 404 returned error can't find the container with id 56a67057374a475d911f887833cd0ea7420af1041fbdd3d80eddab806d07786a Mar 13 15:36:41 crc kubenswrapper[4907]: I0313 15:36:41.545623 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-b654-account-create-update-4jjkj" event={"ID":"1e73852b-fdd0-4cd2-a268-96f6ed551011","Type":"ContainerStarted","Data":"0e271cdd3f12ce6ad2a6fc124e7bdced8c6ac6c38d7c47a03deae9761bf3f57a"} Mar 13 15:36:41 crc kubenswrapper[4907]: I0313 15:36:41.545665 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-b654-account-create-update-4jjkj" event={"ID":"1e73852b-fdd0-4cd2-a268-96f6ed551011","Type":"ContainerStarted","Data":"fd2873409a27ed99fa509e75a3b1acb251ee1ed6a6086a5abc56230de487ac8f"} Mar 13 15:36:41 crc kubenswrapper[4907]: I0313 15:36:41.548548 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-4jnct" event={"ID":"523d3b7c-7098-44bc-92a6-d48f3ae49b65","Type":"ContainerStarted","Data":"ca617bcf57734b16a051ae67c35004e1d491c4d4b7be39c47467fe81b31f47b6"} Mar 13 15:36:41 crc kubenswrapper[4907]: I0313 15:36:41.548599 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-4jnct" event={"ID":"523d3b7c-7098-44bc-92a6-d48f3ae49b65","Type":"ContainerStarted","Data":"56a67057374a475d911f887833cd0ea7420af1041fbdd3d80eddab806d07786a"} Mar 13 15:36:41 crc kubenswrapper[4907]: I0313 15:36:41.560013 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-b654-account-create-update-4jjkj" podStartSLOduration=1.559994263 podStartE2EDuration="1.559994263s" podCreationTimestamp="2026-03-13 15:36:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:36:41.558753269 +0000 UTC m=+5500.458540978" watchObservedRunningTime="2026-03-13 15:36:41.559994263 +0000 UTC m=+5500.459781952" Mar 13 15:36:41 crc kubenswrapper[4907]: I0313 15:36:41.572601 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-create-4jnct" podStartSLOduration=1.572582347 podStartE2EDuration="1.572582347s" podCreationTimestamp="2026-03-13 15:36:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:36:41.572348831 +0000 UTC m=+5500.472136520" watchObservedRunningTime="2026-03-13 15:36:41.572582347 +0000 UTC m=+5500.472370036" Mar 13 15:36:42 crc kubenswrapper[4907]: I0313 15:36:42.559348 4907 generic.go:334] "Generic (PLEG): container finished" podID="1e73852b-fdd0-4cd2-a268-96f6ed551011" containerID="0e271cdd3f12ce6ad2a6fc124e7bdced8c6ac6c38d7c47a03deae9761bf3f57a" exitCode=0 Mar 13 15:36:42 crc kubenswrapper[4907]: I0313 15:36:42.559443 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-b654-account-create-update-4jjkj" event={"ID":"1e73852b-fdd0-4cd2-a268-96f6ed551011","Type":"ContainerDied","Data":"0e271cdd3f12ce6ad2a6fc124e7bdced8c6ac6c38d7c47a03deae9761bf3f57a"} Mar 13 15:36:42 crc kubenswrapper[4907]: I0313 15:36:42.561636 4907 generic.go:334] "Generic (PLEG): container finished" podID="523d3b7c-7098-44bc-92a6-d48f3ae49b65" containerID="ca617bcf57734b16a051ae67c35004e1d491c4d4b7be39c47467fe81b31f47b6" exitCode=0 Mar 13 15:36:42 crc kubenswrapper[4907]: I0313 15:36:42.561697 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-4jnct" event={"ID":"523d3b7c-7098-44bc-92a6-d48f3ae49b65","Type":"ContainerDied","Data":"ca617bcf57734b16a051ae67c35004e1d491c4d4b7be39c47467fe81b31f47b6"} Mar 13 15:36:43 crc kubenswrapper[4907]: I0313 15:36:43.920119 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-b654-account-create-update-4jjkj" Mar 13 15:36:43 crc kubenswrapper[4907]: I0313 15:36:43.927391 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rwpg2\" (UniqueName: \"kubernetes.io/projected/1e73852b-fdd0-4cd2-a268-96f6ed551011-kube-api-access-rwpg2\") pod \"1e73852b-fdd0-4cd2-a268-96f6ed551011\" (UID: \"1e73852b-fdd0-4cd2-a268-96f6ed551011\") " Mar 13 15:36:43 crc kubenswrapper[4907]: I0313 15:36:43.927498 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1e73852b-fdd0-4cd2-a268-96f6ed551011-operator-scripts\") pod \"1e73852b-fdd0-4cd2-a268-96f6ed551011\" (UID: \"1e73852b-fdd0-4cd2-a268-96f6ed551011\") " Mar 13 15:36:43 crc kubenswrapper[4907]: I0313 15:36:43.929539 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1e73852b-fdd0-4cd2-a268-96f6ed551011-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "1e73852b-fdd0-4cd2-a268-96f6ed551011" (UID: "1e73852b-fdd0-4cd2-a268-96f6ed551011"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:36:43 crc kubenswrapper[4907]: I0313 15:36:43.933027 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1e73852b-fdd0-4cd2-a268-96f6ed551011-kube-api-access-rwpg2" (OuterVolumeSpecName: "kube-api-access-rwpg2") pod "1e73852b-fdd0-4cd2-a268-96f6ed551011" (UID: "1e73852b-fdd0-4cd2-a268-96f6ed551011"). InnerVolumeSpecName "kube-api-access-rwpg2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:36:43 crc kubenswrapper[4907]: I0313 15:36:43.935312 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-4jnct" Mar 13 15:36:44 crc kubenswrapper[4907]: I0313 15:36:44.029293 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rwpg2\" (UniqueName: \"kubernetes.io/projected/1e73852b-fdd0-4cd2-a268-96f6ed551011-kube-api-access-rwpg2\") on node \"crc\" DevicePath \"\"" Mar 13 15:36:44 crc kubenswrapper[4907]: I0313 15:36:44.029333 4907 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1e73852b-fdd0-4cd2-a268-96f6ed551011-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 15:36:44 crc kubenswrapper[4907]: I0313 15:36:44.130194 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkhl8\" (UniqueName: \"kubernetes.io/projected/523d3b7c-7098-44bc-92a6-d48f3ae49b65-kube-api-access-zkhl8\") pod \"523d3b7c-7098-44bc-92a6-d48f3ae49b65\" (UID: \"523d3b7c-7098-44bc-92a6-d48f3ae49b65\") " Mar 13 15:36:44 crc kubenswrapper[4907]: I0313 15:36:44.130285 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/523d3b7c-7098-44bc-92a6-d48f3ae49b65-operator-scripts\") pod \"523d3b7c-7098-44bc-92a6-d48f3ae49b65\" (UID: \"523d3b7c-7098-44bc-92a6-d48f3ae49b65\") " Mar 13 15:36:44 crc kubenswrapper[4907]: I0313 15:36:44.130986 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/523d3b7c-7098-44bc-92a6-d48f3ae49b65-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "523d3b7c-7098-44bc-92a6-d48f3ae49b65" (UID: "523d3b7c-7098-44bc-92a6-d48f3ae49b65"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:36:44 crc kubenswrapper[4907]: I0313 15:36:44.135420 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/523d3b7c-7098-44bc-92a6-d48f3ae49b65-kube-api-access-zkhl8" (OuterVolumeSpecName: "kube-api-access-zkhl8") pod "523d3b7c-7098-44bc-92a6-d48f3ae49b65" (UID: "523d3b7c-7098-44bc-92a6-d48f3ae49b65"). InnerVolumeSpecName "kube-api-access-zkhl8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:36:44 crc kubenswrapper[4907]: I0313 15:36:44.232654 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkhl8\" (UniqueName: \"kubernetes.io/projected/523d3b7c-7098-44bc-92a6-d48f3ae49b65-kube-api-access-zkhl8\") on node \"crc\" DevicePath \"\"" Mar 13 15:36:44 crc kubenswrapper[4907]: I0313 15:36:44.232719 4907 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/523d3b7c-7098-44bc-92a6-d48f3ae49b65-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 15:36:44 crc kubenswrapper[4907]: I0313 15:36:44.576947 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-b654-account-create-update-4jjkj" Mar 13 15:36:44 crc kubenswrapper[4907]: I0313 15:36:44.576943 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-b654-account-create-update-4jjkj" event={"ID":"1e73852b-fdd0-4cd2-a268-96f6ed551011","Type":"ContainerDied","Data":"fd2873409a27ed99fa509e75a3b1acb251ee1ed6a6086a5abc56230de487ac8f"} Mar 13 15:36:44 crc kubenswrapper[4907]: I0313 15:36:44.577084 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fd2873409a27ed99fa509e75a3b1acb251ee1ed6a6086a5abc56230de487ac8f" Mar 13 15:36:44 crc kubenswrapper[4907]: I0313 15:36:44.578944 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-4jnct" event={"ID":"523d3b7c-7098-44bc-92a6-d48f3ae49b65","Type":"ContainerDied","Data":"56a67057374a475d911f887833cd0ea7420af1041fbdd3d80eddab806d07786a"} Mar 13 15:36:44 crc kubenswrapper[4907]: I0313 15:36:44.578969 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-4jnct" Mar 13 15:36:44 crc kubenswrapper[4907]: I0313 15:36:44.578990 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="56a67057374a475d911f887833cd0ea7420af1041fbdd3d80eddab806d07786a" Mar 13 15:36:45 crc kubenswrapper[4907]: I0313 15:36:45.677715 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-xrjlt"] Mar 13 15:36:45 crc kubenswrapper[4907]: E0313 15:36:45.678306 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="523d3b7c-7098-44bc-92a6-d48f3ae49b65" containerName="mariadb-database-create" Mar 13 15:36:45 crc kubenswrapper[4907]: I0313 15:36:45.678321 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="523d3b7c-7098-44bc-92a6-d48f3ae49b65" containerName="mariadb-database-create" Mar 13 15:36:45 crc kubenswrapper[4907]: E0313 15:36:45.678350 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e73852b-fdd0-4cd2-a268-96f6ed551011" containerName="mariadb-account-create-update" Mar 13 15:36:45 crc kubenswrapper[4907]: I0313 15:36:45.678358 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e73852b-fdd0-4cd2-a268-96f6ed551011" containerName="mariadb-account-create-update" Mar 13 15:36:45 crc kubenswrapper[4907]: I0313 15:36:45.678575 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e73852b-fdd0-4cd2-a268-96f6ed551011" containerName="mariadb-account-create-update" Mar 13 15:36:45 crc kubenswrapper[4907]: I0313 15:36:45.678597 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="523d3b7c-7098-44bc-92a6-d48f3ae49b65" containerName="mariadb-database-create" Mar 13 15:36:45 crc kubenswrapper[4907]: I0313 15:36:45.679294 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-xrjlt" Mar 13 15:36:45 crc kubenswrapper[4907]: I0313 15:36:45.681384 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-9gtqn" Mar 13 15:36:45 crc kubenswrapper[4907]: I0313 15:36:45.686658 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Mar 13 15:36:45 crc kubenswrapper[4907]: I0313 15:36:45.690277 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-xrjlt"] Mar 13 15:36:45 crc kubenswrapper[4907]: I0313 15:36:45.757943 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/df0d68f5-c073-4dff-8f70-f2897aaef9dc-db-sync-config-data\") pod \"barbican-db-sync-xrjlt\" (UID: \"df0d68f5-c073-4dff-8f70-f2897aaef9dc\") " pod="openstack/barbican-db-sync-xrjlt" Mar 13 15:36:45 crc kubenswrapper[4907]: I0313 15:36:45.758004 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hdl8q\" (UniqueName: \"kubernetes.io/projected/df0d68f5-c073-4dff-8f70-f2897aaef9dc-kube-api-access-hdl8q\") pod \"barbican-db-sync-xrjlt\" (UID: \"df0d68f5-c073-4dff-8f70-f2897aaef9dc\") " pod="openstack/barbican-db-sync-xrjlt" Mar 13 15:36:45 crc kubenswrapper[4907]: I0313 15:36:45.758117 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df0d68f5-c073-4dff-8f70-f2897aaef9dc-combined-ca-bundle\") pod \"barbican-db-sync-xrjlt\" (UID: \"df0d68f5-c073-4dff-8f70-f2897aaef9dc\") " pod="openstack/barbican-db-sync-xrjlt" Mar 13 15:36:45 crc kubenswrapper[4907]: I0313 15:36:45.859087 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hdl8q\" (UniqueName: \"kubernetes.io/projected/df0d68f5-c073-4dff-8f70-f2897aaef9dc-kube-api-access-hdl8q\") pod \"barbican-db-sync-xrjlt\" (UID: \"df0d68f5-c073-4dff-8f70-f2897aaef9dc\") " pod="openstack/barbican-db-sync-xrjlt" Mar 13 15:36:45 crc kubenswrapper[4907]: I0313 15:36:45.859204 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df0d68f5-c073-4dff-8f70-f2897aaef9dc-combined-ca-bundle\") pod \"barbican-db-sync-xrjlt\" (UID: \"df0d68f5-c073-4dff-8f70-f2897aaef9dc\") " pod="openstack/barbican-db-sync-xrjlt" Mar 13 15:36:45 crc kubenswrapper[4907]: I0313 15:36:45.859265 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/df0d68f5-c073-4dff-8f70-f2897aaef9dc-db-sync-config-data\") pod \"barbican-db-sync-xrjlt\" (UID: \"df0d68f5-c073-4dff-8f70-f2897aaef9dc\") " pod="openstack/barbican-db-sync-xrjlt" Mar 13 15:36:45 crc kubenswrapper[4907]: I0313 15:36:45.864780 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df0d68f5-c073-4dff-8f70-f2897aaef9dc-combined-ca-bundle\") pod \"barbican-db-sync-xrjlt\" (UID: \"df0d68f5-c073-4dff-8f70-f2897aaef9dc\") " pod="openstack/barbican-db-sync-xrjlt" Mar 13 15:36:45 crc kubenswrapper[4907]: I0313 15:36:45.864799 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/df0d68f5-c073-4dff-8f70-f2897aaef9dc-db-sync-config-data\") pod \"barbican-db-sync-xrjlt\" (UID: \"df0d68f5-c073-4dff-8f70-f2897aaef9dc\") " pod="openstack/barbican-db-sync-xrjlt" Mar 13 15:36:45 crc kubenswrapper[4907]: I0313 15:36:45.893562 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hdl8q\" (UniqueName: \"kubernetes.io/projected/df0d68f5-c073-4dff-8f70-f2897aaef9dc-kube-api-access-hdl8q\") pod \"barbican-db-sync-xrjlt\" (UID: \"df0d68f5-c073-4dff-8f70-f2897aaef9dc\") " pod="openstack/barbican-db-sync-xrjlt" Mar 13 15:36:45 crc kubenswrapper[4907]: I0313 15:36:45.995627 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-xrjlt" Mar 13 15:36:46 crc kubenswrapper[4907]: I0313 15:36:46.614636 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-xrjlt"] Mar 13 15:36:47 crc kubenswrapper[4907]: I0313 15:36:47.602053 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-xrjlt" event={"ID":"df0d68f5-c073-4dff-8f70-f2897aaef9dc","Type":"ContainerStarted","Data":"29bfa2eaa41b750e3b4b9e54874f96fd9d36ce47ad55875b7c1d9bd6b63f1eef"} Mar 13 15:36:47 crc kubenswrapper[4907]: I0313 15:36:47.603473 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-xrjlt" event={"ID":"df0d68f5-c073-4dff-8f70-f2897aaef9dc","Type":"ContainerStarted","Data":"d8afe419c9f005bbe4627b889eb2b0b424eb7134bbdb9b7658ade7f6e310e864"} Mar 13 15:36:47 crc kubenswrapper[4907]: I0313 15:36:47.624560 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-xrjlt" podStartSLOduration=2.624541947 podStartE2EDuration="2.624541947s" podCreationTimestamp="2026-03-13 15:36:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:36:47.618540452 +0000 UTC m=+5506.518328161" watchObservedRunningTime="2026-03-13 15:36:47.624541947 +0000 UTC m=+5506.524329656" Mar 13 15:36:48 crc kubenswrapper[4907]: I0313 15:36:48.611507 4907 generic.go:334] "Generic (PLEG): container finished" podID="df0d68f5-c073-4dff-8f70-f2897aaef9dc" containerID="29bfa2eaa41b750e3b4b9e54874f96fd9d36ce47ad55875b7c1d9bd6b63f1eef" exitCode=0 Mar 13 15:36:48 crc kubenswrapper[4907]: I0313 15:36:48.611548 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-xrjlt" event={"ID":"df0d68f5-c073-4dff-8f70-f2897aaef9dc","Type":"ContainerDied","Data":"29bfa2eaa41b750e3b4b9e54874f96fd9d36ce47ad55875b7c1d9bd6b63f1eef"} Mar 13 15:36:49 crc kubenswrapper[4907]: I0313 15:36:49.083192 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/root-account-create-update-l8vbt"] Mar 13 15:36:49 crc kubenswrapper[4907]: I0313 15:36:49.089614 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/root-account-create-update-l8vbt"] Mar 13 15:36:49 crc kubenswrapper[4907]: I0313 15:36:49.791228 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d5a861d-b4ca-4936-9fac-5aa23fcaa317" path="/var/lib/kubelet/pods/1d5a861d-b4ca-4936-9fac-5aa23fcaa317/volumes" Mar 13 15:36:49 crc kubenswrapper[4907]: I0313 15:36:49.915116 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-xrjlt" Mar 13 15:36:49 crc kubenswrapper[4907]: I0313 15:36:49.946862 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/df0d68f5-c073-4dff-8f70-f2897aaef9dc-db-sync-config-data\") pod \"df0d68f5-c073-4dff-8f70-f2897aaef9dc\" (UID: \"df0d68f5-c073-4dff-8f70-f2897aaef9dc\") " Mar 13 15:36:49 crc kubenswrapper[4907]: I0313 15:36:49.947133 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df0d68f5-c073-4dff-8f70-f2897aaef9dc-combined-ca-bundle\") pod \"df0d68f5-c073-4dff-8f70-f2897aaef9dc\" (UID: \"df0d68f5-c073-4dff-8f70-f2897aaef9dc\") " Mar 13 15:36:49 crc kubenswrapper[4907]: I0313 15:36:49.947179 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hdl8q\" (UniqueName: \"kubernetes.io/projected/df0d68f5-c073-4dff-8f70-f2897aaef9dc-kube-api-access-hdl8q\") pod \"df0d68f5-c073-4dff-8f70-f2897aaef9dc\" (UID: \"df0d68f5-c073-4dff-8f70-f2897aaef9dc\") " Mar 13 15:36:49 crc kubenswrapper[4907]: I0313 15:36:49.952595 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df0d68f5-c073-4dff-8f70-f2897aaef9dc-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "df0d68f5-c073-4dff-8f70-f2897aaef9dc" (UID: "df0d68f5-c073-4dff-8f70-f2897aaef9dc"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:36:49 crc kubenswrapper[4907]: I0313 15:36:49.953465 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/df0d68f5-c073-4dff-8f70-f2897aaef9dc-kube-api-access-hdl8q" (OuterVolumeSpecName: "kube-api-access-hdl8q") pod "df0d68f5-c073-4dff-8f70-f2897aaef9dc" (UID: "df0d68f5-c073-4dff-8f70-f2897aaef9dc"). InnerVolumeSpecName "kube-api-access-hdl8q". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:36:49 crc kubenswrapper[4907]: I0313 15:36:49.973346 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df0d68f5-c073-4dff-8f70-f2897aaef9dc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "df0d68f5-c073-4dff-8f70-f2897aaef9dc" (UID: "df0d68f5-c073-4dff-8f70-f2897aaef9dc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:36:50 crc kubenswrapper[4907]: I0313 15:36:50.049234 4907 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/df0d68f5-c073-4dff-8f70-f2897aaef9dc-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 15:36:50 crc kubenswrapper[4907]: I0313 15:36:50.049270 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df0d68f5-c073-4dff-8f70-f2897aaef9dc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 15:36:50 crc kubenswrapper[4907]: I0313 15:36:50.049280 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hdl8q\" (UniqueName: \"kubernetes.io/projected/df0d68f5-c073-4dff-8f70-f2897aaef9dc-kube-api-access-hdl8q\") on node \"crc\" DevicePath \"\"" Mar 13 15:36:50 crc kubenswrapper[4907]: I0313 15:36:50.632140 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-xrjlt" event={"ID":"df0d68f5-c073-4dff-8f70-f2897aaef9dc","Type":"ContainerDied","Data":"d8afe419c9f005bbe4627b889eb2b0b424eb7134bbdb9b7658ade7f6e310e864"} Mar 13 15:36:50 crc kubenswrapper[4907]: I0313 15:36:50.632488 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d8afe419c9f005bbe4627b889eb2b0b424eb7134bbdb9b7658ade7f6e310e864" Mar 13 15:36:50 crc kubenswrapper[4907]: I0313 15:36:50.632187 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-xrjlt" Mar 13 15:36:50 crc kubenswrapper[4907]: I0313 15:36:50.854352 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-68d9445c58-bjlh2"] Mar 13 15:36:50 crc kubenswrapper[4907]: E0313 15:36:50.856039 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df0d68f5-c073-4dff-8f70-f2897aaef9dc" containerName="barbican-db-sync" Mar 13 15:36:50 crc kubenswrapper[4907]: I0313 15:36:50.857157 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="df0d68f5-c073-4dff-8f70-f2897aaef9dc" containerName="barbican-db-sync" Mar 13 15:36:50 crc kubenswrapper[4907]: I0313 15:36:50.857574 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="df0d68f5-c073-4dff-8f70-f2897aaef9dc" containerName="barbican-db-sync" Mar 13 15:36:50 crc kubenswrapper[4907]: I0313 15:36:50.858705 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-68d9445c58-bjlh2" Mar 13 15:36:50 crc kubenswrapper[4907]: I0313 15:36:50.861409 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-868b698d7f-wbtl4"] Mar 13 15:36:50 crc kubenswrapper[4907]: I0313 15:36:50.862438 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-9gtqn" Mar 13 15:36:50 crc kubenswrapper[4907]: I0313 15:36:50.862870 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Mar 13 15:36:50 crc kubenswrapper[4907]: I0313 15:36:50.862942 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-868b698d7f-wbtl4" Mar 13 15:36:50 crc kubenswrapper[4907]: I0313 15:36:50.863032 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Mar 13 15:36:50 crc kubenswrapper[4907]: I0313 15:36:50.865080 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Mar 13 15:36:50 crc kubenswrapper[4907]: I0313 15:36:50.880153 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-868b698d7f-wbtl4"] Mar 13 15:36:50 crc kubenswrapper[4907]: I0313 15:36:50.895116 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-68d9445c58-bjlh2"] Mar 13 15:36:50 crc kubenswrapper[4907]: I0313 15:36:50.963947 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6zts7\" (UniqueName: \"kubernetes.io/projected/3599175f-0bbd-4645-ab11-07e26d7fb98d-kube-api-access-6zts7\") pod \"barbican-worker-868b698d7f-wbtl4\" (UID: \"3599175f-0bbd-4645-ab11-07e26d7fb98d\") " pod="openstack/barbican-worker-868b698d7f-wbtl4" Mar 13 15:36:50 crc kubenswrapper[4907]: I0313 15:36:50.964088 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/87b4d8b9-57da-4c61-9dd9-d746fc62db84-logs\") pod \"barbican-keystone-listener-68d9445c58-bjlh2\" (UID: \"87b4d8b9-57da-4c61-9dd9-d746fc62db84\") " pod="openstack/barbican-keystone-listener-68d9445c58-bjlh2" Mar 13 15:36:50 crc kubenswrapper[4907]: I0313 15:36:50.964111 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87b4d8b9-57da-4c61-9dd9-d746fc62db84-config-data\") pod \"barbican-keystone-listener-68d9445c58-bjlh2\" (UID: \"87b4d8b9-57da-4c61-9dd9-d746fc62db84\") " pod="openstack/barbican-keystone-listener-68d9445c58-bjlh2" Mar 13 15:36:50 crc kubenswrapper[4907]: I0313 15:36:50.964142 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3599175f-0bbd-4645-ab11-07e26d7fb98d-logs\") pod \"barbican-worker-868b698d7f-wbtl4\" (UID: \"3599175f-0bbd-4645-ab11-07e26d7fb98d\") " pod="openstack/barbican-worker-868b698d7f-wbtl4" Mar 13 15:36:50 crc kubenswrapper[4907]: I0313 15:36:50.964161 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3599175f-0bbd-4645-ab11-07e26d7fb98d-combined-ca-bundle\") pod \"barbican-worker-868b698d7f-wbtl4\" (UID: \"3599175f-0bbd-4645-ab11-07e26d7fb98d\") " pod="openstack/barbican-worker-868b698d7f-wbtl4" Mar 13 15:36:50 crc kubenswrapper[4907]: I0313 15:36:50.964177 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9rqph\" (UniqueName: \"kubernetes.io/projected/87b4d8b9-57da-4c61-9dd9-d746fc62db84-kube-api-access-9rqph\") pod \"barbican-keystone-listener-68d9445c58-bjlh2\" (UID: \"87b4d8b9-57da-4c61-9dd9-d746fc62db84\") " pod="openstack/barbican-keystone-listener-68d9445c58-bjlh2" Mar 13 15:36:50 crc kubenswrapper[4907]: I0313 15:36:50.964215 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3599175f-0bbd-4645-ab11-07e26d7fb98d-config-data-custom\") pod \"barbican-worker-868b698d7f-wbtl4\" (UID: \"3599175f-0bbd-4645-ab11-07e26d7fb98d\") " pod="openstack/barbican-worker-868b698d7f-wbtl4" Mar 13 15:36:50 crc kubenswrapper[4907]: I0313 15:36:50.964230 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3599175f-0bbd-4645-ab11-07e26d7fb98d-config-data\") pod \"barbican-worker-868b698d7f-wbtl4\" (UID: \"3599175f-0bbd-4645-ab11-07e26d7fb98d\") " pod="openstack/barbican-worker-868b698d7f-wbtl4" Mar 13 15:36:50 crc kubenswrapper[4907]: I0313 15:36:50.964244 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/87b4d8b9-57da-4c61-9dd9-d746fc62db84-config-data-custom\") pod \"barbican-keystone-listener-68d9445c58-bjlh2\" (UID: \"87b4d8b9-57da-4c61-9dd9-d746fc62db84\") " pod="openstack/barbican-keystone-listener-68d9445c58-bjlh2" Mar 13 15:36:50 crc kubenswrapper[4907]: I0313 15:36:50.964267 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87b4d8b9-57da-4c61-9dd9-d746fc62db84-combined-ca-bundle\") pod \"barbican-keystone-listener-68d9445c58-bjlh2\" (UID: \"87b4d8b9-57da-4c61-9dd9-d746fc62db84\") " pod="openstack/barbican-keystone-listener-68d9445c58-bjlh2" Mar 13 15:36:50 crc kubenswrapper[4907]: I0313 15:36:50.980977 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-bd6bb59b7-mfqq8"] Mar 13 15:36:50 crc kubenswrapper[4907]: I0313 15:36:50.982585 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bd6bb59b7-mfqq8" Mar 13 15:36:50 crc kubenswrapper[4907]: I0313 15:36:50.989071 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-bd6bb59b7-mfqq8"] Mar 13 15:36:51 crc kubenswrapper[4907]: I0313 15:36:51.060801 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-5564857dbd-vn7xm"] Mar 13 15:36:51 crc kubenswrapper[4907]: I0313 15:36:51.062109 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5564857dbd-vn7xm" Mar 13 15:36:51 crc kubenswrapper[4907]: I0313 15:36:51.068563 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3599175f-0bbd-4645-ab11-07e26d7fb98d-config-data-custom\") pod \"barbican-worker-868b698d7f-wbtl4\" (UID: \"3599175f-0bbd-4645-ab11-07e26d7fb98d\") " pod="openstack/barbican-worker-868b698d7f-wbtl4" Mar 13 15:36:51 crc kubenswrapper[4907]: I0313 15:36:51.068611 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3599175f-0bbd-4645-ab11-07e26d7fb98d-config-data\") pod \"barbican-worker-868b698d7f-wbtl4\" (UID: \"3599175f-0bbd-4645-ab11-07e26d7fb98d\") " pod="openstack/barbican-worker-868b698d7f-wbtl4" Mar 13 15:36:51 crc kubenswrapper[4907]: I0313 15:36:51.068639 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/87b4d8b9-57da-4c61-9dd9-d746fc62db84-config-data-custom\") pod \"barbican-keystone-listener-68d9445c58-bjlh2\" (UID: \"87b4d8b9-57da-4c61-9dd9-d746fc62db84\") " pod="openstack/barbican-keystone-listener-68d9445c58-bjlh2" Mar 13 15:36:51 crc kubenswrapper[4907]: I0313 15:36:51.068701 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/96134178-5784-41c4-bfb2-da280292c4ee-ovsdbserver-sb\") pod \"dnsmasq-dns-bd6bb59b7-mfqq8\" (UID: \"96134178-5784-41c4-bfb2-da280292c4ee\") " pod="openstack/dnsmasq-dns-bd6bb59b7-mfqq8" Mar 13 15:36:51 crc kubenswrapper[4907]: I0313 15:36:51.068729 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87b4d8b9-57da-4c61-9dd9-d746fc62db84-combined-ca-bundle\") pod \"barbican-keystone-listener-68d9445c58-bjlh2\" (UID: \"87b4d8b9-57da-4c61-9dd9-d746fc62db84\") " pod="openstack/barbican-keystone-listener-68d9445c58-bjlh2" Mar 13 15:36:51 crc kubenswrapper[4907]: I0313 15:36:51.068773 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/96134178-5784-41c4-bfb2-da280292c4ee-dns-svc\") pod \"dnsmasq-dns-bd6bb59b7-mfqq8\" (UID: \"96134178-5784-41c4-bfb2-da280292c4ee\") " pod="openstack/dnsmasq-dns-bd6bb59b7-mfqq8" Mar 13 15:36:51 crc kubenswrapper[4907]: I0313 15:36:51.068823 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6zts7\" (UniqueName: \"kubernetes.io/projected/3599175f-0bbd-4645-ab11-07e26d7fb98d-kube-api-access-6zts7\") pod \"barbican-worker-868b698d7f-wbtl4\" (UID: \"3599175f-0bbd-4645-ab11-07e26d7fb98d\") " pod="openstack/barbican-worker-868b698d7f-wbtl4" Mar 13 15:36:51 crc kubenswrapper[4907]: I0313 15:36:51.068920 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/96134178-5784-41c4-bfb2-da280292c4ee-config\") pod \"dnsmasq-dns-bd6bb59b7-mfqq8\" (UID: \"96134178-5784-41c4-bfb2-da280292c4ee\") " pod="openstack/dnsmasq-dns-bd6bb59b7-mfqq8" Mar 13 15:36:51 crc kubenswrapper[4907]: I0313 15:36:51.068951 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/87b4d8b9-57da-4c61-9dd9-d746fc62db84-logs\") pod \"barbican-keystone-listener-68d9445c58-bjlh2\" (UID: \"87b4d8b9-57da-4c61-9dd9-d746fc62db84\") " pod="openstack/barbican-keystone-listener-68d9445c58-bjlh2" Mar 13 15:36:51 crc kubenswrapper[4907]: I0313 15:36:51.068976 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87b4d8b9-57da-4c61-9dd9-d746fc62db84-config-data\") pod \"barbican-keystone-listener-68d9445c58-bjlh2\" (UID: \"87b4d8b9-57da-4c61-9dd9-d746fc62db84\") " pod="openstack/barbican-keystone-listener-68d9445c58-bjlh2" Mar 13 15:36:51 crc kubenswrapper[4907]: I0313 15:36:51.069006 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6vmfr\" (UniqueName: \"kubernetes.io/projected/96134178-5784-41c4-bfb2-da280292c4ee-kube-api-access-6vmfr\") pod \"dnsmasq-dns-bd6bb59b7-mfqq8\" (UID: \"96134178-5784-41c4-bfb2-da280292c4ee\") " pod="openstack/dnsmasq-dns-bd6bb59b7-mfqq8" Mar 13 15:36:51 crc kubenswrapper[4907]: I0313 15:36:51.069038 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3599175f-0bbd-4645-ab11-07e26d7fb98d-logs\") pod \"barbican-worker-868b698d7f-wbtl4\" (UID: \"3599175f-0bbd-4645-ab11-07e26d7fb98d\") " pod="openstack/barbican-worker-868b698d7f-wbtl4" Mar 13 15:36:51 crc kubenswrapper[4907]: I0313 15:36:51.069059 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3599175f-0bbd-4645-ab11-07e26d7fb98d-combined-ca-bundle\") pod \"barbican-worker-868b698d7f-wbtl4\" (UID: \"3599175f-0bbd-4645-ab11-07e26d7fb98d\") " pod="openstack/barbican-worker-868b698d7f-wbtl4" Mar 13 15:36:51 crc kubenswrapper[4907]: I0313 15:36:51.069082 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9rqph\" (UniqueName: \"kubernetes.io/projected/87b4d8b9-57da-4c61-9dd9-d746fc62db84-kube-api-access-9rqph\") pod \"barbican-keystone-listener-68d9445c58-bjlh2\" (UID: \"87b4d8b9-57da-4c61-9dd9-d746fc62db84\") " pod="openstack/barbican-keystone-listener-68d9445c58-bjlh2" Mar 13 15:36:51 crc kubenswrapper[4907]: I0313 15:36:51.069106 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/96134178-5784-41c4-bfb2-da280292c4ee-ovsdbserver-nb\") pod \"dnsmasq-dns-bd6bb59b7-mfqq8\" (UID: \"96134178-5784-41c4-bfb2-da280292c4ee\") " pod="openstack/dnsmasq-dns-bd6bb59b7-mfqq8" Mar 13 15:36:51 crc kubenswrapper[4907]: I0313 15:36:51.070731 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/87b4d8b9-57da-4c61-9dd9-d746fc62db84-logs\") pod \"barbican-keystone-listener-68d9445c58-bjlh2\" (UID: \"87b4d8b9-57da-4c61-9dd9-d746fc62db84\") " pod="openstack/barbican-keystone-listener-68d9445c58-bjlh2" Mar 13 15:36:51 crc kubenswrapper[4907]: I0313 15:36:51.077029 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3599175f-0bbd-4645-ab11-07e26d7fb98d-logs\") pod \"barbican-worker-868b698d7f-wbtl4\" (UID: \"3599175f-0bbd-4645-ab11-07e26d7fb98d\") " pod="openstack/barbican-worker-868b698d7f-wbtl4" Mar 13 15:36:51 crc kubenswrapper[4907]: I0313 15:36:51.077410 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Mar 13 15:36:51 crc kubenswrapper[4907]: I0313 15:36:51.084345 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-5564857dbd-vn7xm"] Mar 13 15:36:51 crc kubenswrapper[4907]: I0313 15:36:51.088194 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/87b4d8b9-57da-4c61-9dd9-d746fc62db84-config-data-custom\") pod \"barbican-keystone-listener-68d9445c58-bjlh2\" (UID: \"87b4d8b9-57da-4c61-9dd9-d746fc62db84\") " pod="openstack/barbican-keystone-listener-68d9445c58-bjlh2" Mar 13 15:36:51 crc kubenswrapper[4907]: I0313 15:36:51.088584 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87b4d8b9-57da-4c61-9dd9-d746fc62db84-config-data\") pod \"barbican-keystone-listener-68d9445c58-bjlh2\" (UID: \"87b4d8b9-57da-4c61-9dd9-d746fc62db84\") " pod="openstack/barbican-keystone-listener-68d9445c58-bjlh2" Mar 13 15:36:51 crc kubenswrapper[4907]: I0313 15:36:51.088679 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87b4d8b9-57da-4c61-9dd9-d746fc62db84-combined-ca-bundle\") pod \"barbican-keystone-listener-68d9445c58-bjlh2\" (UID: \"87b4d8b9-57da-4c61-9dd9-d746fc62db84\") " pod="openstack/barbican-keystone-listener-68d9445c58-bjlh2" Mar 13 15:36:51 crc kubenswrapper[4907]: I0313 15:36:51.091091 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3599175f-0bbd-4645-ab11-07e26d7fb98d-config-data-custom\") pod \"barbican-worker-868b698d7f-wbtl4\" (UID: \"3599175f-0bbd-4645-ab11-07e26d7fb98d\") " pod="openstack/barbican-worker-868b698d7f-wbtl4" Mar 13 15:36:51 crc kubenswrapper[4907]: I0313 15:36:51.093578 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3599175f-0bbd-4645-ab11-07e26d7fb98d-combined-ca-bundle\") pod \"barbican-worker-868b698d7f-wbtl4\" (UID: \"3599175f-0bbd-4645-ab11-07e26d7fb98d\") " pod="openstack/barbican-worker-868b698d7f-wbtl4" Mar 13 15:36:51 crc kubenswrapper[4907]: I0313 15:36:51.094345 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3599175f-0bbd-4645-ab11-07e26d7fb98d-config-data\") pod \"barbican-worker-868b698d7f-wbtl4\" (UID: \"3599175f-0bbd-4645-ab11-07e26d7fb98d\") " pod="openstack/barbican-worker-868b698d7f-wbtl4" Mar 13 15:36:51 crc kubenswrapper[4907]: I0313 15:36:51.096563 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9rqph\" (UniqueName: \"kubernetes.io/projected/87b4d8b9-57da-4c61-9dd9-d746fc62db84-kube-api-access-9rqph\") pod \"barbican-keystone-listener-68d9445c58-bjlh2\" (UID: \"87b4d8b9-57da-4c61-9dd9-d746fc62db84\") " pod="openstack/barbican-keystone-listener-68d9445c58-bjlh2" Mar 13 15:36:51 crc kubenswrapper[4907]: I0313 15:36:51.098391 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6zts7\" (UniqueName: \"kubernetes.io/projected/3599175f-0bbd-4645-ab11-07e26d7fb98d-kube-api-access-6zts7\") pod \"barbican-worker-868b698d7f-wbtl4\" (UID: \"3599175f-0bbd-4645-ab11-07e26d7fb98d\") " pod="openstack/barbican-worker-868b698d7f-wbtl4" Mar 13 15:36:51 crc kubenswrapper[4907]: I0313 15:36:51.170678 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d83fca2b-ed84-4516-8510-6e0ff4b2cfef-config-data-custom\") pod \"barbican-api-5564857dbd-vn7xm\" (UID: \"d83fca2b-ed84-4516-8510-6e0ff4b2cfef\") " pod="openstack/barbican-api-5564857dbd-vn7xm" Mar 13 15:36:51 crc kubenswrapper[4907]: I0313 15:36:51.170720 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d83fca2b-ed84-4516-8510-6e0ff4b2cfef-config-data\") pod \"barbican-api-5564857dbd-vn7xm\" (UID: \"d83fca2b-ed84-4516-8510-6e0ff4b2cfef\") " pod="openstack/barbican-api-5564857dbd-vn7xm" Mar 13 15:36:51 crc kubenswrapper[4907]: I0313 15:36:51.170782 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/96134178-5784-41c4-bfb2-da280292c4ee-config\") pod \"dnsmasq-dns-bd6bb59b7-mfqq8\" (UID: \"96134178-5784-41c4-bfb2-da280292c4ee\") " pod="openstack/dnsmasq-dns-bd6bb59b7-mfqq8" Mar 13 15:36:51 crc kubenswrapper[4907]: I0313 15:36:51.170815 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6vmfr\" (UniqueName: \"kubernetes.io/projected/96134178-5784-41c4-bfb2-da280292c4ee-kube-api-access-6vmfr\") pod \"dnsmasq-dns-bd6bb59b7-mfqq8\" (UID: \"96134178-5784-41c4-bfb2-da280292c4ee\") " pod="openstack/dnsmasq-dns-bd6bb59b7-mfqq8" Mar 13 15:36:51 crc kubenswrapper[4907]: I0313 15:36:51.170840 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/96134178-5784-41c4-bfb2-da280292c4ee-ovsdbserver-nb\") pod \"dnsmasq-dns-bd6bb59b7-mfqq8\" (UID: \"96134178-5784-41c4-bfb2-da280292c4ee\") " pod="openstack/dnsmasq-dns-bd6bb59b7-mfqq8" Mar 13 15:36:51 crc kubenswrapper[4907]: I0313 15:36:51.170872 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/96134178-5784-41c4-bfb2-da280292c4ee-ovsdbserver-sb\") pod \"dnsmasq-dns-bd6bb59b7-mfqq8\" (UID: \"96134178-5784-41c4-bfb2-da280292c4ee\") " pod="openstack/dnsmasq-dns-bd6bb59b7-mfqq8" Mar 13 15:36:51 crc kubenswrapper[4907]: I0313 15:36:51.170905 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d83fca2b-ed84-4516-8510-6e0ff4b2cfef-logs\") pod \"barbican-api-5564857dbd-vn7xm\" (UID: \"d83fca2b-ed84-4516-8510-6e0ff4b2cfef\") " pod="openstack/barbican-api-5564857dbd-vn7xm" Mar 13 15:36:51 crc kubenswrapper[4907]: I0313 15:36:51.170927 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jzlbb\" (UniqueName: \"kubernetes.io/projected/d83fca2b-ed84-4516-8510-6e0ff4b2cfef-kube-api-access-jzlbb\") pod \"barbican-api-5564857dbd-vn7xm\" (UID: \"d83fca2b-ed84-4516-8510-6e0ff4b2cfef\") " pod="openstack/barbican-api-5564857dbd-vn7xm" Mar 13 15:36:51 crc kubenswrapper[4907]: I0313 15:36:51.170953 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/96134178-5784-41c4-bfb2-da280292c4ee-dns-svc\") pod \"dnsmasq-dns-bd6bb59b7-mfqq8\" (UID: \"96134178-5784-41c4-bfb2-da280292c4ee\") " pod="openstack/dnsmasq-dns-bd6bb59b7-mfqq8" Mar 13 15:36:51 crc kubenswrapper[4907]: I0313 15:36:51.170972 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d83fca2b-ed84-4516-8510-6e0ff4b2cfef-combined-ca-bundle\") pod \"barbican-api-5564857dbd-vn7xm\" (UID: \"d83fca2b-ed84-4516-8510-6e0ff4b2cfef\") " pod="openstack/barbican-api-5564857dbd-vn7xm" Mar 13 15:36:51 crc kubenswrapper[4907]: I0313 15:36:51.171782 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/96134178-5784-41c4-bfb2-da280292c4ee-config\") pod \"dnsmasq-dns-bd6bb59b7-mfqq8\" (UID: \"96134178-5784-41c4-bfb2-da280292c4ee\") " pod="openstack/dnsmasq-dns-bd6bb59b7-mfqq8" Mar 13 15:36:51 crc kubenswrapper[4907]: I0313 15:36:51.172602 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/96134178-5784-41c4-bfb2-da280292c4ee-ovsdbserver-nb\") pod \"dnsmasq-dns-bd6bb59b7-mfqq8\" (UID: \"96134178-5784-41c4-bfb2-da280292c4ee\") " pod="openstack/dnsmasq-dns-bd6bb59b7-mfqq8" Mar 13 15:36:51 crc kubenswrapper[4907]: I0313 15:36:51.172782 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/96134178-5784-41c4-bfb2-da280292c4ee-ovsdbserver-sb\") pod \"dnsmasq-dns-bd6bb59b7-mfqq8\" (UID: \"96134178-5784-41c4-bfb2-da280292c4ee\") " pod="openstack/dnsmasq-dns-bd6bb59b7-mfqq8" Mar 13 15:36:51 crc kubenswrapper[4907]: I0313 15:36:51.172862 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/96134178-5784-41c4-bfb2-da280292c4ee-dns-svc\") pod \"dnsmasq-dns-bd6bb59b7-mfqq8\" (UID: \"96134178-5784-41c4-bfb2-da280292c4ee\") " pod="openstack/dnsmasq-dns-bd6bb59b7-mfqq8" Mar 13 15:36:51 crc kubenswrapper[4907]: I0313 15:36:51.176592 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-68d9445c58-bjlh2" Mar 13 15:36:51 crc kubenswrapper[4907]: I0313 15:36:51.185728 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-868b698d7f-wbtl4" Mar 13 15:36:51 crc kubenswrapper[4907]: I0313 15:36:51.190356 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6vmfr\" (UniqueName: \"kubernetes.io/projected/96134178-5784-41c4-bfb2-da280292c4ee-kube-api-access-6vmfr\") pod \"dnsmasq-dns-bd6bb59b7-mfqq8\" (UID: \"96134178-5784-41c4-bfb2-da280292c4ee\") " pod="openstack/dnsmasq-dns-bd6bb59b7-mfqq8" Mar 13 15:36:51 crc kubenswrapper[4907]: I0313 15:36:51.272379 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d83fca2b-ed84-4516-8510-6e0ff4b2cfef-logs\") pod \"barbican-api-5564857dbd-vn7xm\" (UID: \"d83fca2b-ed84-4516-8510-6e0ff4b2cfef\") " pod="openstack/barbican-api-5564857dbd-vn7xm" Mar 13 15:36:51 crc kubenswrapper[4907]: I0313 15:36:51.272463 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jzlbb\" (UniqueName: \"kubernetes.io/projected/d83fca2b-ed84-4516-8510-6e0ff4b2cfef-kube-api-access-jzlbb\") pod \"barbican-api-5564857dbd-vn7xm\" (UID: \"d83fca2b-ed84-4516-8510-6e0ff4b2cfef\") " pod="openstack/barbican-api-5564857dbd-vn7xm" Mar 13 15:36:51 crc kubenswrapper[4907]: I0313 15:36:51.272526 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d83fca2b-ed84-4516-8510-6e0ff4b2cfef-combined-ca-bundle\") pod \"barbican-api-5564857dbd-vn7xm\" (UID: \"d83fca2b-ed84-4516-8510-6e0ff4b2cfef\") " pod="openstack/barbican-api-5564857dbd-vn7xm" Mar 13 15:36:51 crc kubenswrapper[4907]: I0313 15:36:51.272582 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d83fca2b-ed84-4516-8510-6e0ff4b2cfef-config-data-custom\") pod \"barbican-api-5564857dbd-vn7xm\" (UID: \"d83fca2b-ed84-4516-8510-6e0ff4b2cfef\") " pod="openstack/barbican-api-5564857dbd-vn7xm" Mar 13 15:36:51 crc kubenswrapper[4907]: I0313 15:36:51.272627 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d83fca2b-ed84-4516-8510-6e0ff4b2cfef-config-data\") pod \"barbican-api-5564857dbd-vn7xm\" (UID: \"d83fca2b-ed84-4516-8510-6e0ff4b2cfef\") " pod="openstack/barbican-api-5564857dbd-vn7xm" Mar 13 15:36:51 crc kubenswrapper[4907]: I0313 15:36:51.273192 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d83fca2b-ed84-4516-8510-6e0ff4b2cfef-logs\") pod \"barbican-api-5564857dbd-vn7xm\" (UID: \"d83fca2b-ed84-4516-8510-6e0ff4b2cfef\") " pod="openstack/barbican-api-5564857dbd-vn7xm" Mar 13 15:36:51 crc kubenswrapper[4907]: I0313 15:36:51.279749 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d83fca2b-ed84-4516-8510-6e0ff4b2cfef-config-data\") pod \"barbican-api-5564857dbd-vn7xm\" (UID: \"d83fca2b-ed84-4516-8510-6e0ff4b2cfef\") " pod="openstack/barbican-api-5564857dbd-vn7xm" Mar 13 15:36:51 crc kubenswrapper[4907]: I0313 15:36:51.280305 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d83fca2b-ed84-4516-8510-6e0ff4b2cfef-config-data-custom\") pod \"barbican-api-5564857dbd-vn7xm\" (UID: \"d83fca2b-ed84-4516-8510-6e0ff4b2cfef\") " pod="openstack/barbican-api-5564857dbd-vn7xm" Mar 13 15:36:51 crc kubenswrapper[4907]: I0313 15:36:51.280306 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d83fca2b-ed84-4516-8510-6e0ff4b2cfef-combined-ca-bundle\") pod \"barbican-api-5564857dbd-vn7xm\" (UID: \"d83fca2b-ed84-4516-8510-6e0ff4b2cfef\") " pod="openstack/barbican-api-5564857dbd-vn7xm" Mar 13 15:36:51 crc kubenswrapper[4907]: I0313 15:36:51.296725 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bd6bb59b7-mfqq8" Mar 13 15:36:51 crc kubenswrapper[4907]: I0313 15:36:51.298247 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jzlbb\" (UniqueName: \"kubernetes.io/projected/d83fca2b-ed84-4516-8510-6e0ff4b2cfef-kube-api-access-jzlbb\") pod \"barbican-api-5564857dbd-vn7xm\" (UID: \"d83fca2b-ed84-4516-8510-6e0ff4b2cfef\") " pod="openstack/barbican-api-5564857dbd-vn7xm" Mar 13 15:36:51 crc kubenswrapper[4907]: I0313 15:36:51.460571 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5564857dbd-vn7xm" Mar 13 15:36:51 crc kubenswrapper[4907]: I0313 15:36:51.696349 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-68d9445c58-bjlh2"] Mar 13 15:36:51 crc kubenswrapper[4907]: I0313 15:36:51.706788 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-868b698d7f-wbtl4"] Mar 13 15:36:51 crc kubenswrapper[4907]: W0313 15:36:51.724901 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3599175f_0bbd_4645_ab11_07e26d7fb98d.slice/crio-70f8f900b1131883a154adee5909b47faa634913da47d0ca1dcba98372e75717 WatchSource:0}: Error finding container 70f8f900b1131883a154adee5909b47faa634913da47d0ca1dcba98372e75717: Status 404 returned error can't find the container with id 70f8f900b1131883a154adee5909b47faa634913da47d0ca1dcba98372e75717 Mar 13 15:36:51 crc kubenswrapper[4907]: I0313 15:36:51.862612 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-bd6bb59b7-mfqq8"] Mar 13 15:36:51 crc kubenswrapper[4907]: W0313 15:36:51.867480 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod96134178_5784_41c4_bfb2_da280292c4ee.slice/crio-feefca8ed61869667fc95d91e0bb79d91fef75b784e7cbb8c5d167534ad176c0 WatchSource:0}: Error finding container feefca8ed61869667fc95d91e0bb79d91fef75b784e7cbb8c5d167534ad176c0: Status 404 returned error can't find the container with id feefca8ed61869667fc95d91e0bb79d91fef75b784e7cbb8c5d167534ad176c0 Mar 13 15:36:51 crc kubenswrapper[4907]: I0313 15:36:51.971366 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-5564857dbd-vn7xm"] Mar 13 15:36:52 crc kubenswrapper[4907]: I0313 15:36:52.650279 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5564857dbd-vn7xm" event={"ID":"d83fca2b-ed84-4516-8510-6e0ff4b2cfef","Type":"ContainerStarted","Data":"eb9beef76927f7ed790dbd2612952eb33c40a8f5966b2bc7ca005b02f2176cd0"} Mar 13 15:36:52 crc kubenswrapper[4907]: I0313 15:36:52.650769 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5564857dbd-vn7xm" Mar 13 15:36:52 crc kubenswrapper[4907]: I0313 15:36:52.650845 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5564857dbd-vn7xm" event={"ID":"d83fca2b-ed84-4516-8510-6e0ff4b2cfef","Type":"ContainerStarted","Data":"a05c32904e57c5a2b807e53693b620924bb5ac41933bd588aad9c8e26572947b"} Mar 13 15:36:52 crc kubenswrapper[4907]: I0313 15:36:52.650937 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5564857dbd-vn7xm" event={"ID":"d83fca2b-ed84-4516-8510-6e0ff4b2cfef","Type":"ContainerStarted","Data":"9e14830182c886273f0d1485e37dea5076045dd81323eaacf2d8eec0acdc38e9"} Mar 13 15:36:52 crc kubenswrapper[4907]: I0313 15:36:52.652449 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-68d9445c58-bjlh2" event={"ID":"87b4d8b9-57da-4c61-9dd9-d746fc62db84","Type":"ContainerStarted","Data":"a5db3d582632269b451134f4ed459f116cf65a1a60b5091612a4c60f98350fb9"} Mar 13 15:36:52 crc kubenswrapper[4907]: I0313 15:36:52.652491 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-68d9445c58-bjlh2" event={"ID":"87b4d8b9-57da-4c61-9dd9-d746fc62db84","Type":"ContainerStarted","Data":"185cd70b5b58c41af371e1d5271c3806e31be590b16086db7b6711d2e98cc6ce"} Mar 13 15:36:52 crc kubenswrapper[4907]: I0313 15:36:52.652502 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-68d9445c58-bjlh2" event={"ID":"87b4d8b9-57da-4c61-9dd9-d746fc62db84","Type":"ContainerStarted","Data":"e0c2059048ef60d27cd987200d115afaea617344cf5c9390a36015f744db9317"} Mar 13 15:36:52 crc kubenswrapper[4907]: I0313 15:36:52.654704 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-868b698d7f-wbtl4" event={"ID":"3599175f-0bbd-4645-ab11-07e26d7fb98d","Type":"ContainerStarted","Data":"0264c77b446d8dbb95b35312638da246b5431cf6b293a9faa6a0099662264a46"} Mar 13 15:36:52 crc kubenswrapper[4907]: I0313 15:36:52.654742 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-868b698d7f-wbtl4" event={"ID":"3599175f-0bbd-4645-ab11-07e26d7fb98d","Type":"ContainerStarted","Data":"7c0f7e2f1cdaaf04ee701eb036ae9ee0a5c60418c67eb89e7d28d3e4b3d2e5ae"} Mar 13 15:36:52 crc kubenswrapper[4907]: I0313 15:36:52.654759 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-868b698d7f-wbtl4" event={"ID":"3599175f-0bbd-4645-ab11-07e26d7fb98d","Type":"ContainerStarted","Data":"70f8f900b1131883a154adee5909b47faa634913da47d0ca1dcba98372e75717"} Mar 13 15:36:52 crc kubenswrapper[4907]: I0313 15:36:52.656339 4907 generic.go:334] "Generic (PLEG): container finished" podID="96134178-5784-41c4-bfb2-da280292c4ee" containerID="344113e606bf545fcc172c4d2e36f56d58e31f6999f405a725dfb2c2de7e6e6c" exitCode=0 Mar 13 15:36:52 crc kubenswrapper[4907]: I0313 15:36:52.656379 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bd6bb59b7-mfqq8" event={"ID":"96134178-5784-41c4-bfb2-da280292c4ee","Type":"ContainerDied","Data":"344113e606bf545fcc172c4d2e36f56d58e31f6999f405a725dfb2c2de7e6e6c"} Mar 13 15:36:52 crc kubenswrapper[4907]: I0313 15:36:52.656399 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bd6bb59b7-mfqq8" event={"ID":"96134178-5784-41c4-bfb2-da280292c4ee","Type":"ContainerStarted","Data":"feefca8ed61869667fc95d91e0bb79d91fef75b784e7cbb8c5d167534ad176c0"} Mar 13 15:36:52 crc kubenswrapper[4907]: I0313 15:36:52.682867 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-5564857dbd-vn7xm" podStartSLOduration=1.682836918 podStartE2EDuration="1.682836918s" podCreationTimestamp="2026-03-13 15:36:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:36:52.67739928 +0000 UTC m=+5511.577186969" watchObservedRunningTime="2026-03-13 15:36:52.682836918 +0000 UTC m=+5511.582624607" Mar 13 15:36:52 crc kubenswrapper[4907]: I0313 15:36:52.727332 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-868b698d7f-wbtl4" podStartSLOduration=2.727311493 podStartE2EDuration="2.727311493s" podCreationTimestamp="2026-03-13 15:36:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:36:52.723436467 +0000 UTC m=+5511.623224166" watchObservedRunningTime="2026-03-13 15:36:52.727311493 +0000 UTC m=+5511.627099182" Mar 13 15:36:52 crc kubenswrapper[4907]: I0313 15:36:52.759353 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-68d9445c58-bjlh2" podStartSLOduration=2.759335868 podStartE2EDuration="2.759335868s" podCreationTimestamp="2026-03-13 15:36:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:36:52.758960898 +0000 UTC m=+5511.658748587" watchObservedRunningTime="2026-03-13 15:36:52.759335868 +0000 UTC m=+5511.659123557" Mar 13 15:36:53 crc kubenswrapper[4907]: I0313 15:36:53.666474 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bd6bb59b7-mfqq8" event={"ID":"96134178-5784-41c4-bfb2-da280292c4ee","Type":"ContainerStarted","Data":"c953880ea09d4c9e0288bb5fb9aab81c29c152c5094c1b1a2b88b549e9c408e2"} Mar 13 15:36:53 crc kubenswrapper[4907]: I0313 15:36:53.667860 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5564857dbd-vn7xm" Mar 13 15:36:53 crc kubenswrapper[4907]: I0313 15:36:53.667993 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-bd6bb59b7-mfqq8" Mar 13 15:36:53 crc kubenswrapper[4907]: I0313 15:36:53.686540 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-bd6bb59b7-mfqq8" podStartSLOduration=3.686390327 podStartE2EDuration="3.686390327s" podCreationTimestamp="2026-03-13 15:36:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:36:53.681791681 +0000 UTC m=+5512.581579380" watchObservedRunningTime="2026-03-13 15:36:53.686390327 +0000 UTC m=+5512.586178026" Mar 13 15:37:01 crc kubenswrapper[4907]: I0313 15:37:01.299131 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-bd6bb59b7-mfqq8" Mar 13 15:37:01 crc kubenswrapper[4907]: I0313 15:37:01.361632 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6c5b6fdc47-txgc2"] Mar 13 15:37:01 crc kubenswrapper[4907]: I0313 15:37:01.361929 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6c5b6fdc47-txgc2" podUID="c26e3cb6-4f00-49d9-8bc2-9ec347574580" containerName="dnsmasq-dns" containerID="cri-o://f3459463f090073ba747ae28c0f5e112ee217c5b14b17182f95119ff18ca2ea2" gracePeriod=10 Mar 13 15:37:01 crc kubenswrapper[4907]: I0313 15:37:01.740086 4907 generic.go:334] "Generic (PLEG): container finished" podID="c26e3cb6-4f00-49d9-8bc2-9ec347574580" containerID="f3459463f090073ba747ae28c0f5e112ee217c5b14b17182f95119ff18ca2ea2" exitCode=0 Mar 13 15:37:01 crc kubenswrapper[4907]: I0313 15:37:01.740175 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6c5b6fdc47-txgc2" event={"ID":"c26e3cb6-4f00-49d9-8bc2-9ec347574580","Type":"ContainerDied","Data":"f3459463f090073ba747ae28c0f5e112ee217c5b14b17182f95119ff18ca2ea2"} Mar 13 15:37:01 crc kubenswrapper[4907]: I0313 15:37:01.940111 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c5b6fdc47-txgc2" Mar 13 15:37:02 crc kubenswrapper[4907]: I0313 15:37:02.062472 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c26e3cb6-4f00-49d9-8bc2-9ec347574580-dns-svc\") pod \"c26e3cb6-4f00-49d9-8bc2-9ec347574580\" (UID: \"c26e3cb6-4f00-49d9-8bc2-9ec347574580\") " Mar 13 15:37:02 crc kubenswrapper[4907]: I0313 15:37:02.063136 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c26e3cb6-4f00-49d9-8bc2-9ec347574580-config\") pod \"c26e3cb6-4f00-49d9-8bc2-9ec347574580\" (UID: \"c26e3cb6-4f00-49d9-8bc2-9ec347574580\") " Mar 13 15:37:02 crc kubenswrapper[4907]: I0313 15:37:02.063205 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c26e3cb6-4f00-49d9-8bc2-9ec347574580-ovsdbserver-sb\") pod \"c26e3cb6-4f00-49d9-8bc2-9ec347574580\" (UID: \"c26e3cb6-4f00-49d9-8bc2-9ec347574580\") " Mar 13 15:37:02 crc kubenswrapper[4907]: I0313 15:37:02.063376 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c26e3cb6-4f00-49d9-8bc2-9ec347574580-ovsdbserver-nb\") pod \"c26e3cb6-4f00-49d9-8bc2-9ec347574580\" (UID: \"c26e3cb6-4f00-49d9-8bc2-9ec347574580\") " Mar 13 15:37:02 crc kubenswrapper[4907]: I0313 15:37:02.063409 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hmslv\" (UniqueName: \"kubernetes.io/projected/c26e3cb6-4f00-49d9-8bc2-9ec347574580-kube-api-access-hmslv\") pod \"c26e3cb6-4f00-49d9-8bc2-9ec347574580\" (UID: \"c26e3cb6-4f00-49d9-8bc2-9ec347574580\") " Mar 13 15:37:02 crc kubenswrapper[4907]: I0313 15:37:02.073100 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c26e3cb6-4f00-49d9-8bc2-9ec347574580-kube-api-access-hmslv" (OuterVolumeSpecName: "kube-api-access-hmslv") pod "c26e3cb6-4f00-49d9-8bc2-9ec347574580" (UID: "c26e3cb6-4f00-49d9-8bc2-9ec347574580"). InnerVolumeSpecName "kube-api-access-hmslv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:37:02 crc kubenswrapper[4907]: I0313 15:37:02.108019 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c26e3cb6-4f00-49d9-8bc2-9ec347574580-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "c26e3cb6-4f00-49d9-8bc2-9ec347574580" (UID: "c26e3cb6-4f00-49d9-8bc2-9ec347574580"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:37:02 crc kubenswrapper[4907]: I0313 15:37:02.108932 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c26e3cb6-4f00-49d9-8bc2-9ec347574580-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "c26e3cb6-4f00-49d9-8bc2-9ec347574580" (UID: "c26e3cb6-4f00-49d9-8bc2-9ec347574580"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:37:02 crc kubenswrapper[4907]: I0313 15:37:02.122304 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c26e3cb6-4f00-49d9-8bc2-9ec347574580-config" (OuterVolumeSpecName: "config") pod "c26e3cb6-4f00-49d9-8bc2-9ec347574580" (UID: "c26e3cb6-4f00-49d9-8bc2-9ec347574580"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:37:02 crc kubenswrapper[4907]: I0313 15:37:02.132539 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c26e3cb6-4f00-49d9-8bc2-9ec347574580-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c26e3cb6-4f00-49d9-8bc2-9ec347574580" (UID: "c26e3cb6-4f00-49d9-8bc2-9ec347574580"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:37:02 crc kubenswrapper[4907]: I0313 15:37:02.166015 4907 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c26e3cb6-4f00-49d9-8bc2-9ec347574580-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 13 15:37:02 crc kubenswrapper[4907]: I0313 15:37:02.166066 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hmslv\" (UniqueName: \"kubernetes.io/projected/c26e3cb6-4f00-49d9-8bc2-9ec347574580-kube-api-access-hmslv\") on node \"crc\" DevicePath \"\"" Mar 13 15:37:02 crc kubenswrapper[4907]: I0313 15:37:02.166081 4907 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c26e3cb6-4f00-49d9-8bc2-9ec347574580-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 13 15:37:02 crc kubenswrapper[4907]: I0313 15:37:02.166098 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c26e3cb6-4f00-49d9-8bc2-9ec347574580-config\") on node \"crc\" DevicePath \"\"" Mar 13 15:37:02 crc kubenswrapper[4907]: I0313 15:37:02.166109 4907 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c26e3cb6-4f00-49d9-8bc2-9ec347574580-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 13 15:37:02 crc kubenswrapper[4907]: I0313 15:37:02.750365 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6c5b6fdc47-txgc2" event={"ID":"c26e3cb6-4f00-49d9-8bc2-9ec347574580","Type":"ContainerDied","Data":"5651c9bf703897955b69287b03043d49d9012737130119f792c71677cc9d3fe5"} Mar 13 15:37:02 crc kubenswrapper[4907]: I0313 15:37:02.750429 4907 scope.go:117] "RemoveContainer" containerID="f3459463f090073ba747ae28c0f5e112ee217c5b14b17182f95119ff18ca2ea2" Mar 13 15:37:02 crc kubenswrapper[4907]: I0313 15:37:02.750586 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c5b6fdc47-txgc2" Mar 13 15:37:02 crc kubenswrapper[4907]: I0313 15:37:02.773690 4907 scope.go:117] "RemoveContainer" containerID="de00b39199d31a11cec915e1367446617b6ca45fa3186e7f080356f4a1332db1" Mar 13 15:37:02 crc kubenswrapper[4907]: I0313 15:37:02.789683 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6c5b6fdc47-txgc2"] Mar 13 15:37:02 crc kubenswrapper[4907]: I0313 15:37:02.807682 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6c5b6fdc47-txgc2"] Mar 13 15:37:02 crc kubenswrapper[4907]: I0313 15:37:02.979867 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-5564857dbd-vn7xm" Mar 13 15:37:02 crc kubenswrapper[4907]: I0313 15:37:02.998795 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-5564857dbd-vn7xm" Mar 13 15:37:03 crc kubenswrapper[4907]: I0313 15:37:03.816294 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c26e3cb6-4f00-49d9-8bc2-9ec347574580" path="/var/lib/kubelet/pods/c26e3cb6-4f00-49d9-8bc2-9ec347574580/volumes" Mar 13 15:37:14 crc kubenswrapper[4907]: I0313 15:37:14.633811 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-jh7h4"] Mar 13 15:37:14 crc kubenswrapper[4907]: E0313 15:37:14.634587 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c26e3cb6-4f00-49d9-8bc2-9ec347574580" containerName="init" Mar 13 15:37:14 crc kubenswrapper[4907]: I0313 15:37:14.634599 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="c26e3cb6-4f00-49d9-8bc2-9ec347574580" containerName="init" Mar 13 15:37:14 crc kubenswrapper[4907]: E0313 15:37:14.634610 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c26e3cb6-4f00-49d9-8bc2-9ec347574580" containerName="dnsmasq-dns" Mar 13 15:37:14 crc kubenswrapper[4907]: I0313 15:37:14.634616 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="c26e3cb6-4f00-49d9-8bc2-9ec347574580" containerName="dnsmasq-dns" Mar 13 15:37:14 crc kubenswrapper[4907]: I0313 15:37:14.634767 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="c26e3cb6-4f00-49d9-8bc2-9ec347574580" containerName="dnsmasq-dns" Mar 13 15:37:14 crc kubenswrapper[4907]: I0313 15:37:14.635334 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-jh7h4" Mar 13 15:37:14 crc kubenswrapper[4907]: I0313 15:37:14.647752 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-jh7h4"] Mar 13 15:37:14 crc kubenswrapper[4907]: I0313 15:37:14.704600 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3cbbccb7-b0c0-49be-a969-ffa18b07bd13-operator-scripts\") pod \"neutron-db-create-jh7h4\" (UID: \"3cbbccb7-b0c0-49be-a969-ffa18b07bd13\") " pod="openstack/neutron-db-create-jh7h4" Mar 13 15:37:14 crc kubenswrapper[4907]: I0313 15:37:14.704704 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7hv4b\" (UniqueName: \"kubernetes.io/projected/3cbbccb7-b0c0-49be-a969-ffa18b07bd13-kube-api-access-7hv4b\") pod \"neutron-db-create-jh7h4\" (UID: \"3cbbccb7-b0c0-49be-a969-ffa18b07bd13\") " pod="openstack/neutron-db-create-jh7h4" Mar 13 15:37:14 crc kubenswrapper[4907]: I0313 15:37:14.738952 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-629b-account-create-update-rptv2"] Mar 13 15:37:14 crc kubenswrapper[4907]: I0313 15:37:14.740134 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-629b-account-create-update-rptv2" Mar 13 15:37:14 crc kubenswrapper[4907]: I0313 15:37:14.742306 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Mar 13 15:37:14 crc kubenswrapper[4907]: I0313 15:37:14.752334 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-629b-account-create-update-rptv2"] Mar 13 15:37:14 crc kubenswrapper[4907]: I0313 15:37:14.806146 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3cbbccb7-b0c0-49be-a969-ffa18b07bd13-operator-scripts\") pod \"neutron-db-create-jh7h4\" (UID: \"3cbbccb7-b0c0-49be-a969-ffa18b07bd13\") " pod="openstack/neutron-db-create-jh7h4" Mar 13 15:37:14 crc kubenswrapper[4907]: I0313 15:37:14.806305 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7hv4b\" (UniqueName: \"kubernetes.io/projected/3cbbccb7-b0c0-49be-a969-ffa18b07bd13-kube-api-access-7hv4b\") pod \"neutron-db-create-jh7h4\" (UID: \"3cbbccb7-b0c0-49be-a969-ffa18b07bd13\") " pod="openstack/neutron-db-create-jh7h4" Mar 13 15:37:14 crc kubenswrapper[4907]: I0313 15:37:14.806352 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/608c9419-22d5-4915-b5c6-8be70eebdb23-operator-scripts\") pod \"neutron-629b-account-create-update-rptv2\" (UID: \"608c9419-22d5-4915-b5c6-8be70eebdb23\") " pod="openstack/neutron-629b-account-create-update-rptv2" Mar 13 15:37:14 crc kubenswrapper[4907]: I0313 15:37:14.806379 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qh4pd\" (UniqueName: \"kubernetes.io/projected/608c9419-22d5-4915-b5c6-8be70eebdb23-kube-api-access-qh4pd\") pod \"neutron-629b-account-create-update-rptv2\" (UID: \"608c9419-22d5-4915-b5c6-8be70eebdb23\") " pod="openstack/neutron-629b-account-create-update-rptv2" Mar 13 15:37:14 crc kubenswrapper[4907]: I0313 15:37:14.806970 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3cbbccb7-b0c0-49be-a969-ffa18b07bd13-operator-scripts\") pod \"neutron-db-create-jh7h4\" (UID: \"3cbbccb7-b0c0-49be-a969-ffa18b07bd13\") " pod="openstack/neutron-db-create-jh7h4" Mar 13 15:37:14 crc kubenswrapper[4907]: I0313 15:37:14.830526 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7hv4b\" (UniqueName: \"kubernetes.io/projected/3cbbccb7-b0c0-49be-a969-ffa18b07bd13-kube-api-access-7hv4b\") pod \"neutron-db-create-jh7h4\" (UID: \"3cbbccb7-b0c0-49be-a969-ffa18b07bd13\") " pod="openstack/neutron-db-create-jh7h4" Mar 13 15:37:14 crc kubenswrapper[4907]: I0313 15:37:14.907845 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/608c9419-22d5-4915-b5c6-8be70eebdb23-operator-scripts\") pod \"neutron-629b-account-create-update-rptv2\" (UID: \"608c9419-22d5-4915-b5c6-8be70eebdb23\") " pod="openstack/neutron-629b-account-create-update-rptv2" Mar 13 15:37:14 crc kubenswrapper[4907]: I0313 15:37:14.908201 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qh4pd\" (UniqueName: \"kubernetes.io/projected/608c9419-22d5-4915-b5c6-8be70eebdb23-kube-api-access-qh4pd\") pod \"neutron-629b-account-create-update-rptv2\" (UID: \"608c9419-22d5-4915-b5c6-8be70eebdb23\") " pod="openstack/neutron-629b-account-create-update-rptv2" Mar 13 15:37:14 crc kubenswrapper[4907]: I0313 15:37:14.908794 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/608c9419-22d5-4915-b5c6-8be70eebdb23-operator-scripts\") pod \"neutron-629b-account-create-update-rptv2\" (UID: \"608c9419-22d5-4915-b5c6-8be70eebdb23\") " pod="openstack/neutron-629b-account-create-update-rptv2" Mar 13 15:37:14 crc kubenswrapper[4907]: I0313 15:37:14.940933 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qh4pd\" (UniqueName: \"kubernetes.io/projected/608c9419-22d5-4915-b5c6-8be70eebdb23-kube-api-access-qh4pd\") pod \"neutron-629b-account-create-update-rptv2\" (UID: \"608c9419-22d5-4915-b5c6-8be70eebdb23\") " pod="openstack/neutron-629b-account-create-update-rptv2" Mar 13 15:37:14 crc kubenswrapper[4907]: I0313 15:37:14.962276 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-jh7h4" Mar 13 15:37:15 crc kubenswrapper[4907]: I0313 15:37:15.055069 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-629b-account-create-update-rptv2" Mar 13 15:37:15 crc kubenswrapper[4907]: I0313 15:37:15.461781 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-jh7h4"] Mar 13 15:37:15 crc kubenswrapper[4907]: I0313 15:37:15.563641 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-629b-account-create-update-rptv2"] Mar 13 15:37:15 crc kubenswrapper[4907]: W0313 15:37:15.569019 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod608c9419_22d5_4915_b5c6_8be70eebdb23.slice/crio-72837ec872dcb1aeb63d3b6252b15b3fff4ae6b7c18ef8ec42f445d9fd0a0e5a WatchSource:0}: Error finding container 72837ec872dcb1aeb63d3b6252b15b3fff4ae6b7c18ef8ec42f445d9fd0a0e5a: Status 404 returned error can't find the container with id 72837ec872dcb1aeb63d3b6252b15b3fff4ae6b7c18ef8ec42f445d9fd0a0e5a Mar 13 15:37:15 crc kubenswrapper[4907]: I0313 15:37:15.872264 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-629b-account-create-update-rptv2" event={"ID":"608c9419-22d5-4915-b5c6-8be70eebdb23","Type":"ContainerStarted","Data":"2d8f130696e7f9d5999a1090959bb47d9418a9bc337a9f062febc9b4224a5eb9"} Mar 13 15:37:15 crc kubenswrapper[4907]: I0313 15:37:15.872650 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-629b-account-create-update-rptv2" event={"ID":"608c9419-22d5-4915-b5c6-8be70eebdb23","Type":"ContainerStarted","Data":"72837ec872dcb1aeb63d3b6252b15b3fff4ae6b7c18ef8ec42f445d9fd0a0e5a"} Mar 13 15:37:15 crc kubenswrapper[4907]: I0313 15:37:15.876028 4907 generic.go:334] "Generic (PLEG): container finished" podID="3cbbccb7-b0c0-49be-a969-ffa18b07bd13" containerID="6b2aa6e8ccbdaa4de43a0268ebdb8841de04a3498e495de6ed723c183300253b" exitCode=0 Mar 13 15:37:15 crc kubenswrapper[4907]: I0313 15:37:15.876070 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-jh7h4" event={"ID":"3cbbccb7-b0c0-49be-a969-ffa18b07bd13","Type":"ContainerDied","Data":"6b2aa6e8ccbdaa4de43a0268ebdb8841de04a3498e495de6ed723c183300253b"} Mar 13 15:37:15 crc kubenswrapper[4907]: I0313 15:37:15.876087 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-jh7h4" event={"ID":"3cbbccb7-b0c0-49be-a969-ffa18b07bd13","Type":"ContainerStarted","Data":"874bae6b236dfc82c48bb081ac8b4f317afe5f9203d0173666c9a64b2661229b"} Mar 13 15:37:15 crc kubenswrapper[4907]: I0313 15:37:15.890133 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-629b-account-create-update-rptv2" podStartSLOduration=1.890108781 podStartE2EDuration="1.890108781s" podCreationTimestamp="2026-03-13 15:37:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:37:15.888356863 +0000 UTC m=+5534.788144562" watchObservedRunningTime="2026-03-13 15:37:15.890108781 +0000 UTC m=+5534.789896470" Mar 13 15:37:16 crc kubenswrapper[4907]: I0313 15:37:16.887961 4907 generic.go:334] "Generic (PLEG): container finished" podID="608c9419-22d5-4915-b5c6-8be70eebdb23" containerID="2d8f130696e7f9d5999a1090959bb47d9418a9bc337a9f062febc9b4224a5eb9" exitCode=0 Mar 13 15:37:16 crc kubenswrapper[4907]: I0313 15:37:16.888161 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-629b-account-create-update-rptv2" event={"ID":"608c9419-22d5-4915-b5c6-8be70eebdb23","Type":"ContainerDied","Data":"2d8f130696e7f9d5999a1090959bb47d9418a9bc337a9f062febc9b4224a5eb9"} Mar 13 15:37:17 crc kubenswrapper[4907]: I0313 15:37:17.283187 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-jh7h4" Mar 13 15:37:17 crc kubenswrapper[4907]: I0313 15:37:17.347851 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3cbbccb7-b0c0-49be-a969-ffa18b07bd13-operator-scripts\") pod \"3cbbccb7-b0c0-49be-a969-ffa18b07bd13\" (UID: \"3cbbccb7-b0c0-49be-a969-ffa18b07bd13\") " Mar 13 15:37:17 crc kubenswrapper[4907]: I0313 15:37:17.348101 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7hv4b\" (UniqueName: \"kubernetes.io/projected/3cbbccb7-b0c0-49be-a969-ffa18b07bd13-kube-api-access-7hv4b\") pod \"3cbbccb7-b0c0-49be-a969-ffa18b07bd13\" (UID: \"3cbbccb7-b0c0-49be-a969-ffa18b07bd13\") " Mar 13 15:37:17 crc kubenswrapper[4907]: I0313 15:37:17.348749 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cbbccb7-b0c0-49be-a969-ffa18b07bd13-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "3cbbccb7-b0c0-49be-a969-ffa18b07bd13" (UID: "3cbbccb7-b0c0-49be-a969-ffa18b07bd13"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:37:17 crc kubenswrapper[4907]: I0313 15:37:17.353385 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cbbccb7-b0c0-49be-a969-ffa18b07bd13-kube-api-access-7hv4b" (OuterVolumeSpecName: "kube-api-access-7hv4b") pod "3cbbccb7-b0c0-49be-a969-ffa18b07bd13" (UID: "3cbbccb7-b0c0-49be-a969-ffa18b07bd13"). InnerVolumeSpecName "kube-api-access-7hv4b". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:37:17 crc kubenswrapper[4907]: I0313 15:37:17.449751 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7hv4b\" (UniqueName: \"kubernetes.io/projected/3cbbccb7-b0c0-49be-a969-ffa18b07bd13-kube-api-access-7hv4b\") on node \"crc\" DevicePath \"\"" Mar 13 15:37:17 crc kubenswrapper[4907]: I0313 15:37:17.449792 4907 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3cbbccb7-b0c0-49be-a969-ffa18b07bd13-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 15:37:17 crc kubenswrapper[4907]: I0313 15:37:17.897387 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-jh7h4" Mar 13 15:37:17 crc kubenswrapper[4907]: I0313 15:37:17.898049 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-jh7h4" event={"ID":"3cbbccb7-b0c0-49be-a969-ffa18b07bd13","Type":"ContainerDied","Data":"874bae6b236dfc82c48bb081ac8b4f317afe5f9203d0173666c9a64b2661229b"} Mar 13 15:37:17 crc kubenswrapper[4907]: I0313 15:37:17.898124 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="874bae6b236dfc82c48bb081ac8b4f317afe5f9203d0173666c9a64b2661229b" Mar 13 15:37:18 crc kubenswrapper[4907]: I0313 15:37:18.043993 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 15:37:18 crc kubenswrapper[4907]: I0313 15:37:18.044298 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 15:37:18 crc kubenswrapper[4907]: I0313 15:37:18.200691 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-629b-account-create-update-rptv2" Mar 13 15:37:18 crc kubenswrapper[4907]: I0313 15:37:18.265605 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qh4pd\" (UniqueName: \"kubernetes.io/projected/608c9419-22d5-4915-b5c6-8be70eebdb23-kube-api-access-qh4pd\") pod \"608c9419-22d5-4915-b5c6-8be70eebdb23\" (UID: \"608c9419-22d5-4915-b5c6-8be70eebdb23\") " Mar 13 15:37:18 crc kubenswrapper[4907]: I0313 15:37:18.265780 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/608c9419-22d5-4915-b5c6-8be70eebdb23-operator-scripts\") pod \"608c9419-22d5-4915-b5c6-8be70eebdb23\" (UID: \"608c9419-22d5-4915-b5c6-8be70eebdb23\") " Mar 13 15:37:18 crc kubenswrapper[4907]: I0313 15:37:18.266183 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/608c9419-22d5-4915-b5c6-8be70eebdb23-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "608c9419-22d5-4915-b5c6-8be70eebdb23" (UID: "608c9419-22d5-4915-b5c6-8be70eebdb23"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:37:18 crc kubenswrapper[4907]: I0313 15:37:18.270572 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/608c9419-22d5-4915-b5c6-8be70eebdb23-kube-api-access-qh4pd" (OuterVolumeSpecName: "kube-api-access-qh4pd") pod "608c9419-22d5-4915-b5c6-8be70eebdb23" (UID: "608c9419-22d5-4915-b5c6-8be70eebdb23"). InnerVolumeSpecName "kube-api-access-qh4pd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:37:18 crc kubenswrapper[4907]: I0313 15:37:18.367228 4907 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/608c9419-22d5-4915-b5c6-8be70eebdb23-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 15:37:18 crc kubenswrapper[4907]: I0313 15:37:18.367262 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qh4pd\" (UniqueName: \"kubernetes.io/projected/608c9419-22d5-4915-b5c6-8be70eebdb23-kube-api-access-qh4pd\") on node \"crc\" DevicePath \"\"" Mar 13 15:37:18 crc kubenswrapper[4907]: I0313 15:37:18.921389 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-629b-account-create-update-rptv2" event={"ID":"608c9419-22d5-4915-b5c6-8be70eebdb23","Type":"ContainerDied","Data":"72837ec872dcb1aeb63d3b6252b15b3fff4ae6b7c18ef8ec42f445d9fd0a0e5a"} Mar 13 15:37:18 crc kubenswrapper[4907]: I0313 15:37:18.921798 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="72837ec872dcb1aeb63d3b6252b15b3fff4ae6b7c18ef8ec42f445d9fd0a0e5a" Mar 13 15:37:18 crc kubenswrapper[4907]: I0313 15:37:18.922148 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-629b-account-create-update-rptv2" Mar 13 15:37:19 crc kubenswrapper[4907]: I0313 15:37:19.955841 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-jtsxr"] Mar 13 15:37:19 crc kubenswrapper[4907]: E0313 15:37:19.956251 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="608c9419-22d5-4915-b5c6-8be70eebdb23" containerName="mariadb-account-create-update" Mar 13 15:37:19 crc kubenswrapper[4907]: I0313 15:37:19.956267 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="608c9419-22d5-4915-b5c6-8be70eebdb23" containerName="mariadb-account-create-update" Mar 13 15:37:19 crc kubenswrapper[4907]: E0313 15:37:19.956277 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3cbbccb7-b0c0-49be-a969-ffa18b07bd13" containerName="mariadb-database-create" Mar 13 15:37:19 crc kubenswrapper[4907]: I0313 15:37:19.956284 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="3cbbccb7-b0c0-49be-a969-ffa18b07bd13" containerName="mariadb-database-create" Mar 13 15:37:19 crc kubenswrapper[4907]: I0313 15:37:19.956493 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="608c9419-22d5-4915-b5c6-8be70eebdb23" containerName="mariadb-account-create-update" Mar 13 15:37:19 crc kubenswrapper[4907]: I0313 15:37:19.956517 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="3cbbccb7-b0c0-49be-a969-ffa18b07bd13" containerName="mariadb-database-create" Mar 13 15:37:19 crc kubenswrapper[4907]: I0313 15:37:19.957158 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-jtsxr" Mar 13 15:37:19 crc kubenswrapper[4907]: I0313 15:37:19.966716 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Mar 13 15:37:19 crc kubenswrapper[4907]: I0313 15:37:19.966940 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-rdwdw" Mar 13 15:37:19 crc kubenswrapper[4907]: I0313 15:37:19.967178 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Mar 13 15:37:19 crc kubenswrapper[4907]: I0313 15:37:19.968252 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-jtsxr"] Mar 13 15:37:20 crc kubenswrapper[4907]: I0313 15:37:20.005748 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4vzb5\" (UniqueName: \"kubernetes.io/projected/1af3ec26-0a7c-44c4-8073-fc93093c0cf7-kube-api-access-4vzb5\") pod \"neutron-db-sync-jtsxr\" (UID: \"1af3ec26-0a7c-44c4-8073-fc93093c0cf7\") " pod="openstack/neutron-db-sync-jtsxr" Mar 13 15:37:20 crc kubenswrapper[4907]: I0313 15:37:20.006353 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/1af3ec26-0a7c-44c4-8073-fc93093c0cf7-config\") pod \"neutron-db-sync-jtsxr\" (UID: \"1af3ec26-0a7c-44c4-8073-fc93093c0cf7\") " pod="openstack/neutron-db-sync-jtsxr" Mar 13 15:37:20 crc kubenswrapper[4907]: I0313 15:37:20.006552 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1af3ec26-0a7c-44c4-8073-fc93093c0cf7-combined-ca-bundle\") pod \"neutron-db-sync-jtsxr\" (UID: \"1af3ec26-0a7c-44c4-8073-fc93093c0cf7\") " pod="openstack/neutron-db-sync-jtsxr" Mar 13 15:37:20 crc kubenswrapper[4907]: I0313 15:37:20.108535 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4vzb5\" (UniqueName: \"kubernetes.io/projected/1af3ec26-0a7c-44c4-8073-fc93093c0cf7-kube-api-access-4vzb5\") pod \"neutron-db-sync-jtsxr\" (UID: \"1af3ec26-0a7c-44c4-8073-fc93093c0cf7\") " pod="openstack/neutron-db-sync-jtsxr" Mar 13 15:37:20 crc kubenswrapper[4907]: I0313 15:37:20.108599 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/1af3ec26-0a7c-44c4-8073-fc93093c0cf7-config\") pod \"neutron-db-sync-jtsxr\" (UID: \"1af3ec26-0a7c-44c4-8073-fc93093c0cf7\") " pod="openstack/neutron-db-sync-jtsxr" Mar 13 15:37:20 crc kubenswrapper[4907]: I0313 15:37:20.108704 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1af3ec26-0a7c-44c4-8073-fc93093c0cf7-combined-ca-bundle\") pod \"neutron-db-sync-jtsxr\" (UID: \"1af3ec26-0a7c-44c4-8073-fc93093c0cf7\") " pod="openstack/neutron-db-sync-jtsxr" Mar 13 15:37:20 crc kubenswrapper[4907]: I0313 15:37:20.112343 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1af3ec26-0a7c-44c4-8073-fc93093c0cf7-combined-ca-bundle\") pod \"neutron-db-sync-jtsxr\" (UID: \"1af3ec26-0a7c-44c4-8073-fc93093c0cf7\") " pod="openstack/neutron-db-sync-jtsxr" Mar 13 15:37:20 crc kubenswrapper[4907]: I0313 15:37:20.112444 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/1af3ec26-0a7c-44c4-8073-fc93093c0cf7-config\") pod \"neutron-db-sync-jtsxr\" (UID: \"1af3ec26-0a7c-44c4-8073-fc93093c0cf7\") " pod="openstack/neutron-db-sync-jtsxr" Mar 13 15:37:20 crc kubenswrapper[4907]: I0313 15:37:20.123569 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4vzb5\" (UniqueName: \"kubernetes.io/projected/1af3ec26-0a7c-44c4-8073-fc93093c0cf7-kube-api-access-4vzb5\") pod \"neutron-db-sync-jtsxr\" (UID: \"1af3ec26-0a7c-44c4-8073-fc93093c0cf7\") " pod="openstack/neutron-db-sync-jtsxr" Mar 13 15:37:20 crc kubenswrapper[4907]: I0313 15:37:20.278909 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-jtsxr" Mar 13 15:37:20 crc kubenswrapper[4907]: I0313 15:37:20.751032 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-jtsxr"] Mar 13 15:37:20 crc kubenswrapper[4907]: I0313 15:37:20.938875 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-jtsxr" event={"ID":"1af3ec26-0a7c-44c4-8073-fc93093c0cf7","Type":"ContainerStarted","Data":"970fccf6e2bfb3136160282e85ca44140a4f0863ea34492c0b1a3eb07ca112e9"} Mar 13 15:37:21 crc kubenswrapper[4907]: I0313 15:37:21.948197 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-jtsxr" event={"ID":"1af3ec26-0a7c-44c4-8073-fc93093c0cf7","Type":"ContainerStarted","Data":"eb5f297398bf730822a16b0adfb1835959f9d7d48e8a8a2e6d33b8a74b86db31"} Mar 13 15:37:21 crc kubenswrapper[4907]: I0313 15:37:21.962924 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-jtsxr" podStartSLOduration=2.96286247 podStartE2EDuration="2.96286247s" podCreationTimestamp="2026-03-13 15:37:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:37:21.961194463 +0000 UTC m=+5540.860982152" watchObservedRunningTime="2026-03-13 15:37:21.96286247 +0000 UTC m=+5540.862650149" Mar 13 15:37:24 crc kubenswrapper[4907]: I0313 15:37:24.991357 4907 generic.go:334] "Generic (PLEG): container finished" podID="1af3ec26-0a7c-44c4-8073-fc93093c0cf7" containerID="eb5f297398bf730822a16b0adfb1835959f9d7d48e8a8a2e6d33b8a74b86db31" exitCode=0 Mar 13 15:37:24 crc kubenswrapper[4907]: I0313 15:37:24.991437 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-jtsxr" event={"ID":"1af3ec26-0a7c-44c4-8073-fc93093c0cf7","Type":"ContainerDied","Data":"eb5f297398bf730822a16b0adfb1835959f9d7d48e8a8a2e6d33b8a74b86db31"} Mar 13 15:37:26 crc kubenswrapper[4907]: I0313 15:37:26.292442 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-jtsxr" Mar 13 15:37:26 crc kubenswrapper[4907]: I0313 15:37:26.340704 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/1af3ec26-0a7c-44c4-8073-fc93093c0cf7-config\") pod \"1af3ec26-0a7c-44c4-8073-fc93093c0cf7\" (UID: \"1af3ec26-0a7c-44c4-8073-fc93093c0cf7\") " Mar 13 15:37:26 crc kubenswrapper[4907]: I0313 15:37:26.340816 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4vzb5\" (UniqueName: \"kubernetes.io/projected/1af3ec26-0a7c-44c4-8073-fc93093c0cf7-kube-api-access-4vzb5\") pod \"1af3ec26-0a7c-44c4-8073-fc93093c0cf7\" (UID: \"1af3ec26-0a7c-44c4-8073-fc93093c0cf7\") " Mar 13 15:37:26 crc kubenswrapper[4907]: I0313 15:37:26.340860 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1af3ec26-0a7c-44c4-8073-fc93093c0cf7-combined-ca-bundle\") pod \"1af3ec26-0a7c-44c4-8073-fc93093c0cf7\" (UID: \"1af3ec26-0a7c-44c4-8073-fc93093c0cf7\") " Mar 13 15:37:26 crc kubenswrapper[4907]: I0313 15:37:26.346541 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1af3ec26-0a7c-44c4-8073-fc93093c0cf7-kube-api-access-4vzb5" (OuterVolumeSpecName: "kube-api-access-4vzb5") pod "1af3ec26-0a7c-44c4-8073-fc93093c0cf7" (UID: "1af3ec26-0a7c-44c4-8073-fc93093c0cf7"). InnerVolumeSpecName "kube-api-access-4vzb5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:37:26 crc kubenswrapper[4907]: I0313 15:37:26.364275 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1af3ec26-0a7c-44c4-8073-fc93093c0cf7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1af3ec26-0a7c-44c4-8073-fc93093c0cf7" (UID: "1af3ec26-0a7c-44c4-8073-fc93093c0cf7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:37:26 crc kubenswrapper[4907]: I0313 15:37:26.366995 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1af3ec26-0a7c-44c4-8073-fc93093c0cf7-config" (OuterVolumeSpecName: "config") pod "1af3ec26-0a7c-44c4-8073-fc93093c0cf7" (UID: "1af3ec26-0a7c-44c4-8073-fc93093c0cf7"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:37:26 crc kubenswrapper[4907]: I0313 15:37:26.443013 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4vzb5\" (UniqueName: \"kubernetes.io/projected/1af3ec26-0a7c-44c4-8073-fc93093c0cf7-kube-api-access-4vzb5\") on node \"crc\" DevicePath \"\"" Mar 13 15:37:26 crc kubenswrapper[4907]: I0313 15:37:26.443063 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1af3ec26-0a7c-44c4-8073-fc93093c0cf7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 15:37:26 crc kubenswrapper[4907]: I0313 15:37:26.443075 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/1af3ec26-0a7c-44c4-8073-fc93093c0cf7-config\") on node \"crc\" DevicePath \"\"" Mar 13 15:37:27 crc kubenswrapper[4907]: I0313 15:37:27.011334 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-jtsxr" event={"ID":"1af3ec26-0a7c-44c4-8073-fc93093c0cf7","Type":"ContainerDied","Data":"970fccf6e2bfb3136160282e85ca44140a4f0863ea34492c0b1a3eb07ca112e9"} Mar 13 15:37:27 crc kubenswrapper[4907]: I0313 15:37:27.011372 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="970fccf6e2bfb3136160282e85ca44140a4f0863ea34492c0b1a3eb07ca112e9" Mar 13 15:37:27 crc kubenswrapper[4907]: I0313 15:37:27.011427 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-jtsxr" Mar 13 15:37:27 crc kubenswrapper[4907]: I0313 15:37:27.239212 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-db9bd6759-f9ndc"] Mar 13 15:37:27 crc kubenswrapper[4907]: E0313 15:37:27.240157 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1af3ec26-0a7c-44c4-8073-fc93093c0cf7" containerName="neutron-db-sync" Mar 13 15:37:27 crc kubenswrapper[4907]: I0313 15:37:27.240181 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="1af3ec26-0a7c-44c4-8073-fc93093c0cf7" containerName="neutron-db-sync" Mar 13 15:37:27 crc kubenswrapper[4907]: I0313 15:37:27.240425 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="1af3ec26-0a7c-44c4-8073-fc93093c0cf7" containerName="neutron-db-sync" Mar 13 15:37:27 crc kubenswrapper[4907]: I0313 15:37:27.241860 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-db9bd6759-f9ndc" Mar 13 15:37:27 crc kubenswrapper[4907]: I0313 15:37:27.257080 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-db9bd6759-f9ndc"] Mar 13 15:37:27 crc kubenswrapper[4907]: I0313 15:37:27.332871 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-564986d4bf-zhh4c"] Mar 13 15:37:27 crc kubenswrapper[4907]: I0313 15:37:27.335251 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-564986d4bf-zhh4c" Mar 13 15:37:27 crc kubenswrapper[4907]: I0313 15:37:27.338056 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Mar 13 15:37:27 crc kubenswrapper[4907]: I0313 15:37:27.338326 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Mar 13 15:37:27 crc kubenswrapper[4907]: I0313 15:37:27.338552 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-rdwdw" Mar 13 15:37:27 crc kubenswrapper[4907]: I0313 15:37:27.356780 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-564986d4bf-zhh4c"] Mar 13 15:37:27 crc kubenswrapper[4907]: I0313 15:37:27.371998 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b43d065a-571d-416f-96d3-f1713f3d3bdb-ovsdbserver-sb\") pod \"dnsmasq-dns-db9bd6759-f9ndc\" (UID: \"b43d065a-571d-416f-96d3-f1713f3d3bdb\") " pod="openstack/dnsmasq-dns-db9bd6759-f9ndc" Mar 13 15:37:27 crc kubenswrapper[4907]: I0313 15:37:27.372068 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b43d065a-571d-416f-96d3-f1713f3d3bdb-config\") pod \"dnsmasq-dns-db9bd6759-f9ndc\" (UID: \"b43d065a-571d-416f-96d3-f1713f3d3bdb\") " pod="openstack/dnsmasq-dns-db9bd6759-f9ndc" Mar 13 15:37:27 crc kubenswrapper[4907]: I0313 15:37:27.372193 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b43d065a-571d-416f-96d3-f1713f3d3bdb-dns-svc\") pod \"dnsmasq-dns-db9bd6759-f9ndc\" (UID: \"b43d065a-571d-416f-96d3-f1713f3d3bdb\") " pod="openstack/dnsmasq-dns-db9bd6759-f9ndc" Mar 13 15:37:27 crc kubenswrapper[4907]: I0313 15:37:27.372258 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b43d065a-571d-416f-96d3-f1713f3d3bdb-ovsdbserver-nb\") pod \"dnsmasq-dns-db9bd6759-f9ndc\" (UID: \"b43d065a-571d-416f-96d3-f1713f3d3bdb\") " pod="openstack/dnsmasq-dns-db9bd6759-f9ndc" Mar 13 15:37:27 crc kubenswrapper[4907]: I0313 15:37:27.372291 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-99knh\" (UniqueName: \"kubernetes.io/projected/b43d065a-571d-416f-96d3-f1713f3d3bdb-kube-api-access-99knh\") pod \"dnsmasq-dns-db9bd6759-f9ndc\" (UID: \"b43d065a-571d-416f-96d3-f1713f3d3bdb\") " pod="openstack/dnsmasq-dns-db9bd6759-f9ndc" Mar 13 15:37:27 crc kubenswrapper[4907]: I0313 15:37:27.473483 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b43d065a-571d-416f-96d3-f1713f3d3bdb-ovsdbserver-nb\") pod \"dnsmasq-dns-db9bd6759-f9ndc\" (UID: \"b43d065a-571d-416f-96d3-f1713f3d3bdb\") " pod="openstack/dnsmasq-dns-db9bd6759-f9ndc" Mar 13 15:37:27 crc kubenswrapper[4907]: I0313 15:37:27.473582 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-99knh\" (UniqueName: \"kubernetes.io/projected/b43d065a-571d-416f-96d3-f1713f3d3bdb-kube-api-access-99knh\") pod \"dnsmasq-dns-db9bd6759-f9ndc\" (UID: \"b43d065a-571d-416f-96d3-f1713f3d3bdb\") " pod="openstack/dnsmasq-dns-db9bd6759-f9ndc" Mar 13 15:37:27 crc kubenswrapper[4907]: I0313 15:37:27.474065 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/3ce663fe-b525-44ea-a6d4-33f6b2366f46-httpd-config\") pod \"neutron-564986d4bf-zhh4c\" (UID: \"3ce663fe-b525-44ea-a6d4-33f6b2366f46\") " pod="openstack/neutron-564986d4bf-zhh4c" Mar 13 15:37:27 crc kubenswrapper[4907]: I0313 15:37:27.474179 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b43d065a-571d-416f-96d3-f1713f3d3bdb-ovsdbserver-sb\") pod \"dnsmasq-dns-db9bd6759-f9ndc\" (UID: \"b43d065a-571d-416f-96d3-f1713f3d3bdb\") " pod="openstack/dnsmasq-dns-db9bd6759-f9ndc" Mar 13 15:37:27 crc kubenswrapper[4907]: I0313 15:37:27.474223 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6chw4\" (UniqueName: \"kubernetes.io/projected/3ce663fe-b525-44ea-a6d4-33f6b2366f46-kube-api-access-6chw4\") pod \"neutron-564986d4bf-zhh4c\" (UID: \"3ce663fe-b525-44ea-a6d4-33f6b2366f46\") " pod="openstack/neutron-564986d4bf-zhh4c" Mar 13 15:37:27 crc kubenswrapper[4907]: I0313 15:37:27.474291 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b43d065a-571d-416f-96d3-f1713f3d3bdb-config\") pod \"dnsmasq-dns-db9bd6759-f9ndc\" (UID: \"b43d065a-571d-416f-96d3-f1713f3d3bdb\") " pod="openstack/dnsmasq-dns-db9bd6759-f9ndc" Mar 13 15:37:27 crc kubenswrapper[4907]: I0313 15:37:27.474309 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b43d065a-571d-416f-96d3-f1713f3d3bdb-ovsdbserver-nb\") pod \"dnsmasq-dns-db9bd6759-f9ndc\" (UID: \"b43d065a-571d-416f-96d3-f1713f3d3bdb\") " pod="openstack/dnsmasq-dns-db9bd6759-f9ndc" Mar 13 15:37:27 crc kubenswrapper[4907]: I0313 15:37:27.474361 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ce663fe-b525-44ea-a6d4-33f6b2366f46-combined-ca-bundle\") pod \"neutron-564986d4bf-zhh4c\" (UID: \"3ce663fe-b525-44ea-a6d4-33f6b2366f46\") " pod="openstack/neutron-564986d4bf-zhh4c" Mar 13 15:37:27 crc kubenswrapper[4907]: I0313 15:37:27.474467 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/3ce663fe-b525-44ea-a6d4-33f6b2366f46-config\") pod \"neutron-564986d4bf-zhh4c\" (UID: \"3ce663fe-b525-44ea-a6d4-33f6b2366f46\") " pod="openstack/neutron-564986d4bf-zhh4c" Mar 13 15:37:27 crc kubenswrapper[4907]: I0313 15:37:27.474504 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b43d065a-571d-416f-96d3-f1713f3d3bdb-dns-svc\") pod \"dnsmasq-dns-db9bd6759-f9ndc\" (UID: \"b43d065a-571d-416f-96d3-f1713f3d3bdb\") " pod="openstack/dnsmasq-dns-db9bd6759-f9ndc" Mar 13 15:37:27 crc kubenswrapper[4907]: I0313 15:37:27.474843 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b43d065a-571d-416f-96d3-f1713f3d3bdb-config\") pod \"dnsmasq-dns-db9bd6759-f9ndc\" (UID: \"b43d065a-571d-416f-96d3-f1713f3d3bdb\") " pod="openstack/dnsmasq-dns-db9bd6759-f9ndc" Mar 13 15:37:27 crc kubenswrapper[4907]: I0313 15:37:27.475215 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b43d065a-571d-416f-96d3-f1713f3d3bdb-ovsdbserver-sb\") pod \"dnsmasq-dns-db9bd6759-f9ndc\" (UID: \"b43d065a-571d-416f-96d3-f1713f3d3bdb\") " pod="openstack/dnsmasq-dns-db9bd6759-f9ndc" Mar 13 15:37:27 crc kubenswrapper[4907]: I0313 15:37:27.475387 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b43d065a-571d-416f-96d3-f1713f3d3bdb-dns-svc\") pod \"dnsmasq-dns-db9bd6759-f9ndc\" (UID: \"b43d065a-571d-416f-96d3-f1713f3d3bdb\") " pod="openstack/dnsmasq-dns-db9bd6759-f9ndc" Mar 13 15:37:27 crc kubenswrapper[4907]: I0313 15:37:27.499926 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-99knh\" (UniqueName: \"kubernetes.io/projected/b43d065a-571d-416f-96d3-f1713f3d3bdb-kube-api-access-99knh\") pod \"dnsmasq-dns-db9bd6759-f9ndc\" (UID: \"b43d065a-571d-416f-96d3-f1713f3d3bdb\") " pod="openstack/dnsmasq-dns-db9bd6759-f9ndc" Mar 13 15:37:27 crc kubenswrapper[4907]: I0313 15:37:27.565909 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-db9bd6759-f9ndc" Mar 13 15:37:27 crc kubenswrapper[4907]: I0313 15:37:27.576162 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6chw4\" (UniqueName: \"kubernetes.io/projected/3ce663fe-b525-44ea-a6d4-33f6b2366f46-kube-api-access-6chw4\") pod \"neutron-564986d4bf-zhh4c\" (UID: \"3ce663fe-b525-44ea-a6d4-33f6b2366f46\") " pod="openstack/neutron-564986d4bf-zhh4c" Mar 13 15:37:27 crc kubenswrapper[4907]: I0313 15:37:27.576226 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ce663fe-b525-44ea-a6d4-33f6b2366f46-combined-ca-bundle\") pod \"neutron-564986d4bf-zhh4c\" (UID: \"3ce663fe-b525-44ea-a6d4-33f6b2366f46\") " pod="openstack/neutron-564986d4bf-zhh4c" Mar 13 15:37:27 crc kubenswrapper[4907]: I0313 15:37:27.576282 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/3ce663fe-b525-44ea-a6d4-33f6b2366f46-config\") pod \"neutron-564986d4bf-zhh4c\" (UID: \"3ce663fe-b525-44ea-a6d4-33f6b2366f46\") " pod="openstack/neutron-564986d4bf-zhh4c" Mar 13 15:37:27 crc kubenswrapper[4907]: I0313 15:37:27.576375 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/3ce663fe-b525-44ea-a6d4-33f6b2366f46-httpd-config\") pod \"neutron-564986d4bf-zhh4c\" (UID: \"3ce663fe-b525-44ea-a6d4-33f6b2366f46\") " pod="openstack/neutron-564986d4bf-zhh4c" Mar 13 15:37:27 crc kubenswrapper[4907]: I0313 15:37:27.582559 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/3ce663fe-b525-44ea-a6d4-33f6b2366f46-httpd-config\") pod \"neutron-564986d4bf-zhh4c\" (UID: \"3ce663fe-b525-44ea-a6d4-33f6b2366f46\") " pod="openstack/neutron-564986d4bf-zhh4c" Mar 13 15:37:27 crc kubenswrapper[4907]: I0313 15:37:27.586549 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ce663fe-b525-44ea-a6d4-33f6b2366f46-combined-ca-bundle\") pod \"neutron-564986d4bf-zhh4c\" (UID: \"3ce663fe-b525-44ea-a6d4-33f6b2366f46\") " pod="openstack/neutron-564986d4bf-zhh4c" Mar 13 15:37:27 crc kubenswrapper[4907]: I0313 15:37:27.596666 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/3ce663fe-b525-44ea-a6d4-33f6b2366f46-config\") pod \"neutron-564986d4bf-zhh4c\" (UID: \"3ce663fe-b525-44ea-a6d4-33f6b2366f46\") " pod="openstack/neutron-564986d4bf-zhh4c" Mar 13 15:37:27 crc kubenswrapper[4907]: I0313 15:37:27.607071 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6chw4\" (UniqueName: \"kubernetes.io/projected/3ce663fe-b525-44ea-a6d4-33f6b2366f46-kube-api-access-6chw4\") pod \"neutron-564986d4bf-zhh4c\" (UID: \"3ce663fe-b525-44ea-a6d4-33f6b2366f46\") " pod="openstack/neutron-564986d4bf-zhh4c" Mar 13 15:37:27 crc kubenswrapper[4907]: I0313 15:37:27.661529 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-564986d4bf-zhh4c" Mar 13 15:37:28 crc kubenswrapper[4907]: I0313 15:37:28.146108 4907 scope.go:117] "RemoveContainer" containerID="bddac8c66ea0090f3f484cd37e970a8a28f550e7e44321363049a8343bf92762" Mar 13 15:37:28 crc kubenswrapper[4907]: I0313 15:37:28.153076 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-db9bd6759-f9ndc"] Mar 13 15:37:28 crc kubenswrapper[4907]: I0313 15:37:28.314125 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-564986d4bf-zhh4c"] Mar 13 15:37:28 crc kubenswrapper[4907]: W0313 15:37:28.320760 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3ce663fe_b525_44ea_a6d4_33f6b2366f46.slice/crio-c0afd6d8d5bedb0a51cafa4e726c26baac20fd293ee0e8cf9f73f3fe64e24c0a WatchSource:0}: Error finding container c0afd6d8d5bedb0a51cafa4e726c26baac20fd293ee0e8cf9f73f3fe64e24c0a: Status 404 returned error can't find the container with id c0afd6d8d5bedb0a51cafa4e726c26baac20fd293ee0e8cf9f73f3fe64e24c0a Mar 13 15:37:29 crc kubenswrapper[4907]: I0313 15:37:29.026986 4907 generic.go:334] "Generic (PLEG): container finished" podID="b43d065a-571d-416f-96d3-f1713f3d3bdb" containerID="f003308b3b3b337edbb1d567791fd2916dac4c39389f2b18c7276a22c4799e5d" exitCode=0 Mar 13 15:37:29 crc kubenswrapper[4907]: I0313 15:37:29.027210 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-db9bd6759-f9ndc" event={"ID":"b43d065a-571d-416f-96d3-f1713f3d3bdb","Type":"ContainerDied","Data":"f003308b3b3b337edbb1d567791fd2916dac4c39389f2b18c7276a22c4799e5d"} Mar 13 15:37:29 crc kubenswrapper[4907]: I0313 15:37:29.027385 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-db9bd6759-f9ndc" event={"ID":"b43d065a-571d-416f-96d3-f1713f3d3bdb","Type":"ContainerStarted","Data":"89449f9bc1a34a373899c9c8f376372efe1d7f3851750d5a63cfd6c8a1ef4335"} Mar 13 15:37:29 crc kubenswrapper[4907]: I0313 15:37:29.030032 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-564986d4bf-zhh4c" event={"ID":"3ce663fe-b525-44ea-a6d4-33f6b2366f46","Type":"ContainerStarted","Data":"4852d549b802ea4097d262c3246a3ef7bebacfab6f360b646e235ad408bb2f24"} Mar 13 15:37:29 crc kubenswrapper[4907]: I0313 15:37:29.030075 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-564986d4bf-zhh4c" event={"ID":"3ce663fe-b525-44ea-a6d4-33f6b2366f46","Type":"ContainerStarted","Data":"6dc6b0fe7218c94a5f2886858f2d8f14a2cb91737d9ced5752f7e83123baa22d"} Mar 13 15:37:29 crc kubenswrapper[4907]: I0313 15:37:29.030085 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-564986d4bf-zhh4c" event={"ID":"3ce663fe-b525-44ea-a6d4-33f6b2366f46","Type":"ContainerStarted","Data":"c0afd6d8d5bedb0a51cafa4e726c26baac20fd293ee0e8cf9f73f3fe64e24c0a"} Mar 13 15:37:29 crc kubenswrapper[4907]: I0313 15:37:29.030566 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-564986d4bf-zhh4c" Mar 13 15:37:29 crc kubenswrapper[4907]: I0313 15:37:29.093739 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-564986d4bf-zhh4c" podStartSLOduration=2.093715497 podStartE2EDuration="2.093715497s" podCreationTimestamp="2026-03-13 15:37:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:37:29.073162135 +0000 UTC m=+5547.972949824" watchObservedRunningTime="2026-03-13 15:37:29.093715497 +0000 UTC m=+5547.993503186" Mar 13 15:37:30 crc kubenswrapper[4907]: I0313 15:37:30.039319 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-db9bd6759-f9ndc" event={"ID":"b43d065a-571d-416f-96d3-f1713f3d3bdb","Type":"ContainerStarted","Data":"181ece190f085ef12291d2abeb4a964937fb06d189ff825afc0a6770f3e932ba"} Mar 13 15:37:30 crc kubenswrapper[4907]: I0313 15:37:30.059768 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-db9bd6759-f9ndc" podStartSLOduration=3.05975089 podStartE2EDuration="3.05975089s" podCreationTimestamp="2026-03-13 15:37:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:37:30.054390504 +0000 UTC m=+5548.954178203" watchObservedRunningTime="2026-03-13 15:37:30.05975089 +0000 UTC m=+5548.959538579" Mar 13 15:37:31 crc kubenswrapper[4907]: I0313 15:37:31.048192 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-db9bd6759-f9ndc" Mar 13 15:37:37 crc kubenswrapper[4907]: I0313 15:37:37.567462 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-db9bd6759-f9ndc" Mar 13 15:37:37 crc kubenswrapper[4907]: I0313 15:37:37.621817 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bd6bb59b7-mfqq8"] Mar 13 15:37:37 crc kubenswrapper[4907]: I0313 15:37:37.622072 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-bd6bb59b7-mfqq8" podUID="96134178-5784-41c4-bfb2-da280292c4ee" containerName="dnsmasq-dns" containerID="cri-o://c953880ea09d4c9e0288bb5fb9aab81c29c152c5094c1b1a2b88b549e9c408e2" gracePeriod=10 Mar 13 15:37:38 crc kubenswrapper[4907]: I0313 15:37:38.097122 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bd6bb59b7-mfqq8" Mar 13 15:37:38 crc kubenswrapper[4907]: I0313 15:37:38.103817 4907 generic.go:334] "Generic (PLEG): container finished" podID="96134178-5784-41c4-bfb2-da280292c4ee" containerID="c953880ea09d4c9e0288bb5fb9aab81c29c152c5094c1b1a2b88b549e9c408e2" exitCode=0 Mar 13 15:37:38 crc kubenswrapper[4907]: I0313 15:37:38.103863 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bd6bb59b7-mfqq8" event={"ID":"96134178-5784-41c4-bfb2-da280292c4ee","Type":"ContainerDied","Data":"c953880ea09d4c9e0288bb5fb9aab81c29c152c5094c1b1a2b88b549e9c408e2"} Mar 13 15:37:38 crc kubenswrapper[4907]: I0313 15:37:38.103919 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bd6bb59b7-mfqq8" event={"ID":"96134178-5784-41c4-bfb2-da280292c4ee","Type":"ContainerDied","Data":"feefca8ed61869667fc95d91e0bb79d91fef75b784e7cbb8c5d167534ad176c0"} Mar 13 15:37:38 crc kubenswrapper[4907]: I0313 15:37:38.103941 4907 scope.go:117] "RemoveContainer" containerID="c953880ea09d4c9e0288bb5fb9aab81c29c152c5094c1b1a2b88b549e9c408e2" Mar 13 15:37:38 crc kubenswrapper[4907]: I0313 15:37:38.104023 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bd6bb59b7-mfqq8" Mar 13 15:37:38 crc kubenswrapper[4907]: I0313 15:37:38.144098 4907 scope.go:117] "RemoveContainer" containerID="344113e606bf545fcc172c4d2e36f56d58e31f6999f405a725dfb2c2de7e6e6c" Mar 13 15:37:38 crc kubenswrapper[4907]: I0313 15:37:38.163344 4907 scope.go:117] "RemoveContainer" containerID="c953880ea09d4c9e0288bb5fb9aab81c29c152c5094c1b1a2b88b549e9c408e2" Mar 13 15:37:38 crc kubenswrapper[4907]: E0313 15:37:38.163721 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c953880ea09d4c9e0288bb5fb9aab81c29c152c5094c1b1a2b88b549e9c408e2\": container with ID starting with c953880ea09d4c9e0288bb5fb9aab81c29c152c5094c1b1a2b88b549e9c408e2 not found: ID does not exist" containerID="c953880ea09d4c9e0288bb5fb9aab81c29c152c5094c1b1a2b88b549e9c408e2" Mar 13 15:37:38 crc kubenswrapper[4907]: I0313 15:37:38.163756 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c953880ea09d4c9e0288bb5fb9aab81c29c152c5094c1b1a2b88b549e9c408e2"} err="failed to get container status \"c953880ea09d4c9e0288bb5fb9aab81c29c152c5094c1b1a2b88b549e9c408e2\": rpc error: code = NotFound desc = could not find container \"c953880ea09d4c9e0288bb5fb9aab81c29c152c5094c1b1a2b88b549e9c408e2\": container with ID starting with c953880ea09d4c9e0288bb5fb9aab81c29c152c5094c1b1a2b88b549e9c408e2 not found: ID does not exist" Mar 13 15:37:38 crc kubenswrapper[4907]: I0313 15:37:38.163779 4907 scope.go:117] "RemoveContainer" containerID="344113e606bf545fcc172c4d2e36f56d58e31f6999f405a725dfb2c2de7e6e6c" Mar 13 15:37:38 crc kubenswrapper[4907]: E0313 15:37:38.164077 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"344113e606bf545fcc172c4d2e36f56d58e31f6999f405a725dfb2c2de7e6e6c\": container with ID starting with 344113e606bf545fcc172c4d2e36f56d58e31f6999f405a725dfb2c2de7e6e6c not found: ID does not exist" containerID="344113e606bf545fcc172c4d2e36f56d58e31f6999f405a725dfb2c2de7e6e6c" Mar 13 15:37:38 crc kubenswrapper[4907]: I0313 15:37:38.164102 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"344113e606bf545fcc172c4d2e36f56d58e31f6999f405a725dfb2c2de7e6e6c"} err="failed to get container status \"344113e606bf545fcc172c4d2e36f56d58e31f6999f405a725dfb2c2de7e6e6c\": rpc error: code = NotFound desc = could not find container \"344113e606bf545fcc172c4d2e36f56d58e31f6999f405a725dfb2c2de7e6e6c\": container with ID starting with 344113e606bf545fcc172c4d2e36f56d58e31f6999f405a725dfb2c2de7e6e6c not found: ID does not exist" Mar 13 15:37:38 crc kubenswrapper[4907]: I0313 15:37:38.167954 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/96134178-5784-41c4-bfb2-da280292c4ee-config\") pod \"96134178-5784-41c4-bfb2-da280292c4ee\" (UID: \"96134178-5784-41c4-bfb2-da280292c4ee\") " Mar 13 15:37:38 crc kubenswrapper[4907]: I0313 15:37:38.168068 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/96134178-5784-41c4-bfb2-da280292c4ee-ovsdbserver-nb\") pod \"96134178-5784-41c4-bfb2-da280292c4ee\" (UID: \"96134178-5784-41c4-bfb2-da280292c4ee\") " Mar 13 15:37:38 crc kubenswrapper[4907]: I0313 15:37:38.168133 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/96134178-5784-41c4-bfb2-da280292c4ee-dns-svc\") pod \"96134178-5784-41c4-bfb2-da280292c4ee\" (UID: \"96134178-5784-41c4-bfb2-da280292c4ee\") " Mar 13 15:37:38 crc kubenswrapper[4907]: I0313 15:37:38.168156 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6vmfr\" (UniqueName: \"kubernetes.io/projected/96134178-5784-41c4-bfb2-da280292c4ee-kube-api-access-6vmfr\") pod \"96134178-5784-41c4-bfb2-da280292c4ee\" (UID: \"96134178-5784-41c4-bfb2-da280292c4ee\") " Mar 13 15:37:38 crc kubenswrapper[4907]: I0313 15:37:38.168200 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/96134178-5784-41c4-bfb2-da280292c4ee-ovsdbserver-sb\") pod \"96134178-5784-41c4-bfb2-da280292c4ee\" (UID: \"96134178-5784-41c4-bfb2-da280292c4ee\") " Mar 13 15:37:38 crc kubenswrapper[4907]: I0313 15:37:38.175035 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96134178-5784-41c4-bfb2-da280292c4ee-kube-api-access-6vmfr" (OuterVolumeSpecName: "kube-api-access-6vmfr") pod "96134178-5784-41c4-bfb2-da280292c4ee" (UID: "96134178-5784-41c4-bfb2-da280292c4ee"). InnerVolumeSpecName "kube-api-access-6vmfr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:37:38 crc kubenswrapper[4907]: I0313 15:37:38.214138 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/96134178-5784-41c4-bfb2-da280292c4ee-config" (OuterVolumeSpecName: "config") pod "96134178-5784-41c4-bfb2-da280292c4ee" (UID: "96134178-5784-41c4-bfb2-da280292c4ee"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:37:38 crc kubenswrapper[4907]: I0313 15:37:38.219087 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/96134178-5784-41c4-bfb2-da280292c4ee-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "96134178-5784-41c4-bfb2-da280292c4ee" (UID: "96134178-5784-41c4-bfb2-da280292c4ee"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:37:38 crc kubenswrapper[4907]: I0313 15:37:38.219215 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/96134178-5784-41c4-bfb2-da280292c4ee-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "96134178-5784-41c4-bfb2-da280292c4ee" (UID: "96134178-5784-41c4-bfb2-da280292c4ee"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:37:38 crc kubenswrapper[4907]: I0313 15:37:38.227057 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/96134178-5784-41c4-bfb2-da280292c4ee-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "96134178-5784-41c4-bfb2-da280292c4ee" (UID: "96134178-5784-41c4-bfb2-da280292c4ee"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:37:38 crc kubenswrapper[4907]: I0313 15:37:38.270852 4907 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/96134178-5784-41c4-bfb2-da280292c4ee-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 13 15:37:38 crc kubenswrapper[4907]: I0313 15:37:38.270943 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6vmfr\" (UniqueName: \"kubernetes.io/projected/96134178-5784-41c4-bfb2-da280292c4ee-kube-api-access-6vmfr\") on node \"crc\" DevicePath \"\"" Mar 13 15:37:38 crc kubenswrapper[4907]: I0313 15:37:38.270962 4907 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/96134178-5784-41c4-bfb2-da280292c4ee-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 13 15:37:38 crc kubenswrapper[4907]: I0313 15:37:38.270978 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/96134178-5784-41c4-bfb2-da280292c4ee-config\") on node \"crc\" DevicePath \"\"" Mar 13 15:37:38 crc kubenswrapper[4907]: I0313 15:37:38.270989 4907 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/96134178-5784-41c4-bfb2-da280292c4ee-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 13 15:37:38 crc kubenswrapper[4907]: I0313 15:37:38.450834 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bd6bb59b7-mfqq8"] Mar 13 15:37:38 crc kubenswrapper[4907]: I0313 15:37:38.459484 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-bd6bb59b7-mfqq8"] Mar 13 15:37:39 crc kubenswrapper[4907]: I0313 15:37:39.793666 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96134178-5784-41c4-bfb2-da280292c4ee" path="/var/lib/kubelet/pods/96134178-5784-41c4-bfb2-da280292c4ee/volumes" Mar 13 15:37:48 crc kubenswrapper[4907]: I0313 15:37:48.041737 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 15:37:48 crc kubenswrapper[4907]: I0313 15:37:48.042196 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 15:37:53 crc kubenswrapper[4907]: I0313 15:37:53.537137 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-5k9nb"] Mar 13 15:37:53 crc kubenswrapper[4907]: E0313 15:37:53.538445 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96134178-5784-41c4-bfb2-da280292c4ee" containerName="dnsmasq-dns" Mar 13 15:37:53 crc kubenswrapper[4907]: I0313 15:37:53.538470 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="96134178-5784-41c4-bfb2-da280292c4ee" containerName="dnsmasq-dns" Mar 13 15:37:53 crc kubenswrapper[4907]: E0313 15:37:53.538515 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96134178-5784-41c4-bfb2-da280292c4ee" containerName="init" Mar 13 15:37:53 crc kubenswrapper[4907]: I0313 15:37:53.538530 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="96134178-5784-41c4-bfb2-da280292c4ee" containerName="init" Mar 13 15:37:53 crc kubenswrapper[4907]: I0313 15:37:53.538801 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="96134178-5784-41c4-bfb2-da280292c4ee" containerName="dnsmasq-dns" Mar 13 15:37:53 crc kubenswrapper[4907]: I0313 15:37:53.545116 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5k9nb" Mar 13 15:37:53 crc kubenswrapper[4907]: I0313 15:37:53.548853 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5k9nb"] Mar 13 15:37:53 crc kubenswrapper[4907]: I0313 15:37:53.670957 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/46d919ab-084b-46cf-a55a-80060d0f89e5-utilities\") pod \"certified-operators-5k9nb\" (UID: \"46d919ab-084b-46cf-a55a-80060d0f89e5\") " pod="openshift-marketplace/certified-operators-5k9nb" Mar 13 15:37:53 crc kubenswrapper[4907]: I0313 15:37:53.671086 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/46d919ab-084b-46cf-a55a-80060d0f89e5-catalog-content\") pod \"certified-operators-5k9nb\" (UID: \"46d919ab-084b-46cf-a55a-80060d0f89e5\") " pod="openshift-marketplace/certified-operators-5k9nb" Mar 13 15:37:53 crc kubenswrapper[4907]: I0313 15:37:53.671160 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pm6kc\" (UniqueName: \"kubernetes.io/projected/46d919ab-084b-46cf-a55a-80060d0f89e5-kube-api-access-pm6kc\") pod \"certified-operators-5k9nb\" (UID: \"46d919ab-084b-46cf-a55a-80060d0f89e5\") " pod="openshift-marketplace/certified-operators-5k9nb" Mar 13 15:37:53 crc kubenswrapper[4907]: I0313 15:37:53.772041 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/46d919ab-084b-46cf-a55a-80060d0f89e5-utilities\") pod \"certified-operators-5k9nb\" (UID: \"46d919ab-084b-46cf-a55a-80060d0f89e5\") " pod="openshift-marketplace/certified-operators-5k9nb" Mar 13 15:37:53 crc kubenswrapper[4907]: I0313 15:37:53.772119 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/46d919ab-084b-46cf-a55a-80060d0f89e5-catalog-content\") pod \"certified-operators-5k9nb\" (UID: \"46d919ab-084b-46cf-a55a-80060d0f89e5\") " pod="openshift-marketplace/certified-operators-5k9nb" Mar 13 15:37:53 crc kubenswrapper[4907]: I0313 15:37:53.772178 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pm6kc\" (UniqueName: \"kubernetes.io/projected/46d919ab-084b-46cf-a55a-80060d0f89e5-kube-api-access-pm6kc\") pod \"certified-operators-5k9nb\" (UID: \"46d919ab-084b-46cf-a55a-80060d0f89e5\") " pod="openshift-marketplace/certified-operators-5k9nb" Mar 13 15:37:53 crc kubenswrapper[4907]: I0313 15:37:53.772560 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/46d919ab-084b-46cf-a55a-80060d0f89e5-utilities\") pod \"certified-operators-5k9nb\" (UID: \"46d919ab-084b-46cf-a55a-80060d0f89e5\") " pod="openshift-marketplace/certified-operators-5k9nb" Mar 13 15:37:53 crc kubenswrapper[4907]: I0313 15:37:53.772714 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/46d919ab-084b-46cf-a55a-80060d0f89e5-catalog-content\") pod \"certified-operators-5k9nb\" (UID: \"46d919ab-084b-46cf-a55a-80060d0f89e5\") " pod="openshift-marketplace/certified-operators-5k9nb" Mar 13 15:37:53 crc kubenswrapper[4907]: I0313 15:37:53.797162 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pm6kc\" (UniqueName: \"kubernetes.io/projected/46d919ab-084b-46cf-a55a-80060d0f89e5-kube-api-access-pm6kc\") pod \"certified-operators-5k9nb\" (UID: \"46d919ab-084b-46cf-a55a-80060d0f89e5\") " pod="openshift-marketplace/certified-operators-5k9nb" Mar 13 15:37:53 crc kubenswrapper[4907]: I0313 15:37:53.882203 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5k9nb" Mar 13 15:37:54 crc kubenswrapper[4907]: I0313 15:37:54.407392 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5k9nb"] Mar 13 15:37:55 crc kubenswrapper[4907]: I0313 15:37:55.249661 4907 generic.go:334] "Generic (PLEG): container finished" podID="46d919ab-084b-46cf-a55a-80060d0f89e5" containerID="07e6a739eb2cd151b3d0c8f4d016361ec920ef5cee0ff3c0cf3353c888330ada" exitCode=0 Mar 13 15:37:55 crc kubenswrapper[4907]: I0313 15:37:55.249762 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5k9nb" event={"ID":"46d919ab-084b-46cf-a55a-80060d0f89e5","Type":"ContainerDied","Data":"07e6a739eb2cd151b3d0c8f4d016361ec920ef5cee0ff3c0cf3353c888330ada"} Mar 13 15:37:55 crc kubenswrapper[4907]: I0313 15:37:55.249991 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5k9nb" event={"ID":"46d919ab-084b-46cf-a55a-80060d0f89e5","Type":"ContainerStarted","Data":"6bf1cbbbca2dc292f4b7fb3a637b22871e833710be9542b6cacb01ec4e9d5066"} Mar 13 15:37:55 crc kubenswrapper[4907]: I0313 15:37:55.251406 4907 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 13 15:37:56 crc kubenswrapper[4907]: I0313 15:37:56.258672 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5k9nb" event={"ID":"46d919ab-084b-46cf-a55a-80060d0f89e5","Type":"ContainerStarted","Data":"b930f886e923a7aeedd0f47b3c8af96617296760b49fcc8a4cd563a5557bb45c"} Mar 13 15:37:57 crc kubenswrapper[4907]: I0313 15:37:57.269982 4907 generic.go:334] "Generic (PLEG): container finished" podID="46d919ab-084b-46cf-a55a-80060d0f89e5" containerID="b930f886e923a7aeedd0f47b3c8af96617296760b49fcc8a4cd563a5557bb45c" exitCode=0 Mar 13 15:37:57 crc kubenswrapper[4907]: I0313 15:37:57.270040 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5k9nb" event={"ID":"46d919ab-084b-46cf-a55a-80060d0f89e5","Type":"ContainerDied","Data":"b930f886e923a7aeedd0f47b3c8af96617296760b49fcc8a4cd563a5557bb45c"} Mar 13 15:37:57 crc kubenswrapper[4907]: I0313 15:37:57.672838 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-564986d4bf-zhh4c" Mar 13 15:37:58 crc kubenswrapper[4907]: I0313 15:37:58.282849 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5k9nb" event={"ID":"46d919ab-084b-46cf-a55a-80060d0f89e5","Type":"ContainerStarted","Data":"1c7e727a5c90422b9d55f69cb636beca42802c6044e4986c0e34da16172b738a"} Mar 13 15:37:58 crc kubenswrapper[4907]: I0313 15:37:58.306030 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-5k9nb" podStartSLOduration=2.492291323 podStartE2EDuration="5.306007688s" podCreationTimestamp="2026-03-13 15:37:53 +0000 UTC" firstStartedPulling="2026-03-13 15:37:55.251197866 +0000 UTC m=+5574.150985555" lastFinishedPulling="2026-03-13 15:37:58.064914241 +0000 UTC m=+5576.964701920" observedRunningTime="2026-03-13 15:37:58.296632803 +0000 UTC m=+5577.196420502" watchObservedRunningTime="2026-03-13 15:37:58.306007688 +0000 UTC m=+5577.205795377" Mar 13 15:38:00 crc kubenswrapper[4907]: I0313 15:38:00.139636 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556938-ptksb"] Mar 13 15:38:00 crc kubenswrapper[4907]: I0313 15:38:00.141069 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556938-ptksb" Mar 13 15:38:00 crc kubenswrapper[4907]: I0313 15:38:00.144468 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 15:38:00 crc kubenswrapper[4907]: I0313 15:38:00.144474 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 15:38:00 crc kubenswrapper[4907]: I0313 15:38:00.144667 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 15:38:00 crc kubenswrapper[4907]: I0313 15:38:00.148876 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556938-ptksb"] Mar 13 15:38:00 crc kubenswrapper[4907]: I0313 15:38:00.282443 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-97bz7\" (UniqueName: \"kubernetes.io/projected/6a5048c0-3add-450a-94c0-c068ceca72a4-kube-api-access-97bz7\") pod \"auto-csr-approver-29556938-ptksb\" (UID: \"6a5048c0-3add-450a-94c0-c068ceca72a4\") " pod="openshift-infra/auto-csr-approver-29556938-ptksb" Mar 13 15:38:00 crc kubenswrapper[4907]: I0313 15:38:00.384708 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-97bz7\" (UniqueName: \"kubernetes.io/projected/6a5048c0-3add-450a-94c0-c068ceca72a4-kube-api-access-97bz7\") pod \"auto-csr-approver-29556938-ptksb\" (UID: \"6a5048c0-3add-450a-94c0-c068ceca72a4\") " pod="openshift-infra/auto-csr-approver-29556938-ptksb" Mar 13 15:38:00 crc kubenswrapper[4907]: I0313 15:38:00.419948 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-97bz7\" (UniqueName: \"kubernetes.io/projected/6a5048c0-3add-450a-94c0-c068ceca72a4-kube-api-access-97bz7\") pod \"auto-csr-approver-29556938-ptksb\" (UID: \"6a5048c0-3add-450a-94c0-c068ceca72a4\") " pod="openshift-infra/auto-csr-approver-29556938-ptksb" Mar 13 15:38:00 crc kubenswrapper[4907]: I0313 15:38:00.460823 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556938-ptksb" Mar 13 15:38:00 crc kubenswrapper[4907]: I0313 15:38:00.905840 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556938-ptksb"] Mar 13 15:38:01 crc kubenswrapper[4907]: I0313 15:38:01.305441 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556938-ptksb" event={"ID":"6a5048c0-3add-450a-94c0-c068ceca72a4","Type":"ContainerStarted","Data":"e969e95aa6638b032cb7230dee7feb590118185e18e1ed4bf4c49a061143dfc1"} Mar 13 15:38:02 crc kubenswrapper[4907]: I0313 15:38:02.316852 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556938-ptksb" event={"ID":"6a5048c0-3add-450a-94c0-c068ceca72a4","Type":"ContainerStarted","Data":"19b5bd889c57ba8c1711b31659d515b279612f88bb0e0f645828fd40237e4402"} Mar 13 15:38:02 crc kubenswrapper[4907]: I0313 15:38:02.337489 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29556938-ptksb" podStartSLOduration=1.290600144 podStartE2EDuration="2.337460525s" podCreationTimestamp="2026-03-13 15:38:00 +0000 UTC" firstStartedPulling="2026-03-13 15:38:00.903940539 +0000 UTC m=+5579.803728218" lastFinishedPulling="2026-03-13 15:38:01.95080089 +0000 UTC m=+5580.850588599" observedRunningTime="2026-03-13 15:38:02.330208966 +0000 UTC m=+5581.229996655" watchObservedRunningTime="2026-03-13 15:38:02.337460525 +0000 UTC m=+5581.237248214" Mar 13 15:38:03 crc kubenswrapper[4907]: I0313 15:38:03.325350 4907 generic.go:334] "Generic (PLEG): container finished" podID="6a5048c0-3add-450a-94c0-c068ceca72a4" containerID="19b5bd889c57ba8c1711b31659d515b279612f88bb0e0f645828fd40237e4402" exitCode=0 Mar 13 15:38:03 crc kubenswrapper[4907]: I0313 15:38:03.325577 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556938-ptksb" event={"ID":"6a5048c0-3add-450a-94c0-c068ceca72a4","Type":"ContainerDied","Data":"19b5bd889c57ba8c1711b31659d515b279612f88bb0e0f645828fd40237e4402"} Mar 13 15:38:03 crc kubenswrapper[4907]: I0313 15:38:03.883369 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-5k9nb" Mar 13 15:38:03 crc kubenswrapper[4907]: I0313 15:38:03.883424 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-5k9nb" Mar 13 15:38:03 crc kubenswrapper[4907]: I0313 15:38:03.944291 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-5k9nb" Mar 13 15:38:04 crc kubenswrapper[4907]: I0313 15:38:04.392276 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-5k9nb" Mar 13 15:38:04 crc kubenswrapper[4907]: I0313 15:38:04.448346 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5k9nb"] Mar 13 15:38:04 crc kubenswrapper[4907]: I0313 15:38:04.662387 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556938-ptksb" Mar 13 15:38:04 crc kubenswrapper[4907]: I0313 15:38:04.771967 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-97bz7\" (UniqueName: \"kubernetes.io/projected/6a5048c0-3add-450a-94c0-c068ceca72a4-kube-api-access-97bz7\") pod \"6a5048c0-3add-450a-94c0-c068ceca72a4\" (UID: \"6a5048c0-3add-450a-94c0-c068ceca72a4\") " Mar 13 15:38:04 crc kubenswrapper[4907]: I0313 15:38:04.788875 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6a5048c0-3add-450a-94c0-c068ceca72a4-kube-api-access-97bz7" (OuterVolumeSpecName: "kube-api-access-97bz7") pod "6a5048c0-3add-450a-94c0-c068ceca72a4" (UID: "6a5048c0-3add-450a-94c0-c068ceca72a4"). InnerVolumeSpecName "kube-api-access-97bz7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:38:04 crc kubenswrapper[4907]: I0313 15:38:04.838680 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-kxfg2"] Mar 13 15:38:04 crc kubenswrapper[4907]: E0313 15:38:04.839142 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a5048c0-3add-450a-94c0-c068ceca72a4" containerName="oc" Mar 13 15:38:04 crc kubenswrapper[4907]: I0313 15:38:04.839166 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a5048c0-3add-450a-94c0-c068ceca72a4" containerName="oc" Mar 13 15:38:04 crc kubenswrapper[4907]: I0313 15:38:04.839407 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a5048c0-3add-450a-94c0-c068ceca72a4" containerName="oc" Mar 13 15:38:04 crc kubenswrapper[4907]: I0313 15:38:04.840095 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-kxfg2" Mar 13 15:38:04 crc kubenswrapper[4907]: I0313 15:38:04.863919 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-kxfg2"] Mar 13 15:38:04 crc kubenswrapper[4907]: I0313 15:38:04.875039 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-97bz7\" (UniqueName: \"kubernetes.io/projected/6a5048c0-3add-450a-94c0-c068ceca72a4-kube-api-access-97bz7\") on node \"crc\" DevicePath \"\"" Mar 13 15:38:04 crc kubenswrapper[4907]: I0313 15:38:04.909390 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556932-fxhnm"] Mar 13 15:38:04 crc kubenswrapper[4907]: I0313 15:38:04.916221 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556932-fxhnm"] Mar 13 15:38:04 crc kubenswrapper[4907]: I0313 15:38:04.939299 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-fd12-account-create-update-kxbcs"] Mar 13 15:38:04 crc kubenswrapper[4907]: I0313 15:38:04.940647 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-fd12-account-create-update-kxbcs" Mar 13 15:38:04 crc kubenswrapper[4907]: I0313 15:38:04.942726 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Mar 13 15:38:04 crc kubenswrapper[4907]: I0313 15:38:04.947416 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-fd12-account-create-update-kxbcs"] Mar 13 15:38:04 crc kubenswrapper[4907]: I0313 15:38:04.976843 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/440d4795-da82-461c-b460-bdd5c271de4d-operator-scripts\") pod \"glance-db-create-kxfg2\" (UID: \"440d4795-da82-461c-b460-bdd5c271de4d\") " pod="openstack/glance-db-create-kxfg2" Mar 13 15:38:04 crc kubenswrapper[4907]: I0313 15:38:04.976953 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tzph2\" (UniqueName: \"kubernetes.io/projected/440d4795-da82-461c-b460-bdd5c271de4d-kube-api-access-tzph2\") pod \"glance-db-create-kxfg2\" (UID: \"440d4795-da82-461c-b460-bdd5c271de4d\") " pod="openstack/glance-db-create-kxfg2" Mar 13 15:38:05 crc kubenswrapper[4907]: I0313 15:38:05.078041 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-99dgh\" (UniqueName: \"kubernetes.io/projected/0d6a6021-7cc4-42f9-848b-5bbfc3650ee7-kube-api-access-99dgh\") pod \"glance-fd12-account-create-update-kxbcs\" (UID: \"0d6a6021-7cc4-42f9-848b-5bbfc3650ee7\") " pod="openstack/glance-fd12-account-create-update-kxbcs" Mar 13 15:38:05 crc kubenswrapper[4907]: I0313 15:38:05.078264 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0d6a6021-7cc4-42f9-848b-5bbfc3650ee7-operator-scripts\") pod \"glance-fd12-account-create-update-kxbcs\" (UID: \"0d6a6021-7cc4-42f9-848b-5bbfc3650ee7\") " pod="openstack/glance-fd12-account-create-update-kxbcs" Mar 13 15:38:05 crc kubenswrapper[4907]: I0313 15:38:05.078390 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/440d4795-da82-461c-b460-bdd5c271de4d-operator-scripts\") pod \"glance-db-create-kxfg2\" (UID: \"440d4795-da82-461c-b460-bdd5c271de4d\") " pod="openstack/glance-db-create-kxfg2" Mar 13 15:38:05 crc kubenswrapper[4907]: I0313 15:38:05.078429 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tzph2\" (UniqueName: \"kubernetes.io/projected/440d4795-da82-461c-b460-bdd5c271de4d-kube-api-access-tzph2\") pod \"glance-db-create-kxfg2\" (UID: \"440d4795-da82-461c-b460-bdd5c271de4d\") " pod="openstack/glance-db-create-kxfg2" Mar 13 15:38:05 crc kubenswrapper[4907]: I0313 15:38:05.079095 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/440d4795-da82-461c-b460-bdd5c271de4d-operator-scripts\") pod \"glance-db-create-kxfg2\" (UID: \"440d4795-da82-461c-b460-bdd5c271de4d\") " pod="openstack/glance-db-create-kxfg2" Mar 13 15:38:05 crc kubenswrapper[4907]: I0313 15:38:05.107406 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tzph2\" (UniqueName: \"kubernetes.io/projected/440d4795-da82-461c-b460-bdd5c271de4d-kube-api-access-tzph2\") pod \"glance-db-create-kxfg2\" (UID: \"440d4795-da82-461c-b460-bdd5c271de4d\") " pod="openstack/glance-db-create-kxfg2" Mar 13 15:38:05 crc kubenswrapper[4907]: I0313 15:38:05.158982 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-kxfg2" Mar 13 15:38:05 crc kubenswrapper[4907]: I0313 15:38:05.179822 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-99dgh\" (UniqueName: \"kubernetes.io/projected/0d6a6021-7cc4-42f9-848b-5bbfc3650ee7-kube-api-access-99dgh\") pod \"glance-fd12-account-create-update-kxbcs\" (UID: \"0d6a6021-7cc4-42f9-848b-5bbfc3650ee7\") " pod="openstack/glance-fd12-account-create-update-kxbcs" Mar 13 15:38:05 crc kubenswrapper[4907]: I0313 15:38:05.179928 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0d6a6021-7cc4-42f9-848b-5bbfc3650ee7-operator-scripts\") pod \"glance-fd12-account-create-update-kxbcs\" (UID: \"0d6a6021-7cc4-42f9-848b-5bbfc3650ee7\") " pod="openstack/glance-fd12-account-create-update-kxbcs" Mar 13 15:38:05 crc kubenswrapper[4907]: I0313 15:38:05.180841 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0d6a6021-7cc4-42f9-848b-5bbfc3650ee7-operator-scripts\") pod \"glance-fd12-account-create-update-kxbcs\" (UID: \"0d6a6021-7cc4-42f9-848b-5bbfc3650ee7\") " pod="openstack/glance-fd12-account-create-update-kxbcs" Mar 13 15:38:05 crc kubenswrapper[4907]: I0313 15:38:05.198116 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-99dgh\" (UniqueName: \"kubernetes.io/projected/0d6a6021-7cc4-42f9-848b-5bbfc3650ee7-kube-api-access-99dgh\") pod \"glance-fd12-account-create-update-kxbcs\" (UID: \"0d6a6021-7cc4-42f9-848b-5bbfc3650ee7\") " pod="openstack/glance-fd12-account-create-update-kxbcs" Mar 13 15:38:05 crc kubenswrapper[4907]: I0313 15:38:05.263599 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-fd12-account-create-update-kxbcs" Mar 13 15:38:05 crc kubenswrapper[4907]: I0313 15:38:05.342668 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556938-ptksb" event={"ID":"6a5048c0-3add-450a-94c0-c068ceca72a4","Type":"ContainerDied","Data":"e969e95aa6638b032cb7230dee7feb590118185e18e1ed4bf4c49a061143dfc1"} Mar 13 15:38:05 crc kubenswrapper[4907]: I0313 15:38:05.342717 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e969e95aa6638b032cb7230dee7feb590118185e18e1ed4bf4c49a061143dfc1" Mar 13 15:38:05 crc kubenswrapper[4907]: I0313 15:38:05.342691 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556938-ptksb" Mar 13 15:38:05 crc kubenswrapper[4907]: I0313 15:38:05.606015 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-kxfg2"] Mar 13 15:38:05 crc kubenswrapper[4907]: I0313 15:38:05.705808 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-fd12-account-create-update-kxbcs"] Mar 13 15:38:05 crc kubenswrapper[4907]: W0313 15:38:05.712328 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0d6a6021_7cc4_42f9_848b_5bbfc3650ee7.slice/crio-83481b88571f9dd6758c4725b16318418738a3b1dc14c5db9de531f76d808495 WatchSource:0}: Error finding container 83481b88571f9dd6758c4725b16318418738a3b1dc14c5db9de531f76d808495: Status 404 returned error can't find the container with id 83481b88571f9dd6758c4725b16318418738a3b1dc14c5db9de531f76d808495 Mar 13 15:38:05 crc kubenswrapper[4907]: I0313 15:38:05.792669 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e5efd747-3795-4922-9f81-3ef02dcb2f91" path="/var/lib/kubelet/pods/e5efd747-3795-4922-9f81-3ef02dcb2f91/volumes" Mar 13 15:38:06 crc kubenswrapper[4907]: I0313 15:38:06.353040 4907 generic.go:334] "Generic (PLEG): container finished" podID="440d4795-da82-461c-b460-bdd5c271de4d" containerID="7414c52bdeec9a9b8ecea00c6ccd0ca4cd0ecf4d6c333567e9430b7214c0182e" exitCode=0 Mar 13 15:38:06 crc kubenswrapper[4907]: I0313 15:38:06.353193 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-kxfg2" event={"ID":"440d4795-da82-461c-b460-bdd5c271de4d","Type":"ContainerDied","Data":"7414c52bdeec9a9b8ecea00c6ccd0ca4cd0ecf4d6c333567e9430b7214c0182e"} Mar 13 15:38:06 crc kubenswrapper[4907]: I0313 15:38:06.353514 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-kxfg2" event={"ID":"440d4795-da82-461c-b460-bdd5c271de4d","Type":"ContainerStarted","Data":"7a689c1f01fe0ef145be6f24241adc9dc1ca896caf778fafb7281e1050d46874"} Mar 13 15:38:06 crc kubenswrapper[4907]: I0313 15:38:06.356091 4907 generic.go:334] "Generic (PLEG): container finished" podID="0d6a6021-7cc4-42f9-848b-5bbfc3650ee7" containerID="38713faad1d72bf574ce7d6ae2bd740312997c935666c1fb81c567bd339dc0d0" exitCode=0 Mar 13 15:38:06 crc kubenswrapper[4907]: I0313 15:38:06.356158 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-fd12-account-create-update-kxbcs" event={"ID":"0d6a6021-7cc4-42f9-848b-5bbfc3650ee7","Type":"ContainerDied","Data":"38713faad1d72bf574ce7d6ae2bd740312997c935666c1fb81c567bd339dc0d0"} Mar 13 15:38:06 crc kubenswrapper[4907]: I0313 15:38:06.356252 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-fd12-account-create-update-kxbcs" event={"ID":"0d6a6021-7cc4-42f9-848b-5bbfc3650ee7","Type":"ContainerStarted","Data":"83481b88571f9dd6758c4725b16318418738a3b1dc14c5db9de531f76d808495"} Mar 13 15:38:06 crc kubenswrapper[4907]: I0313 15:38:06.356463 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-5k9nb" podUID="46d919ab-084b-46cf-a55a-80060d0f89e5" containerName="registry-server" containerID="cri-o://1c7e727a5c90422b9d55f69cb636beca42802c6044e4986c0e34da16172b738a" gracePeriod=2 Mar 13 15:38:07 crc kubenswrapper[4907]: I0313 15:38:07.312982 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5k9nb" Mar 13 15:38:07 crc kubenswrapper[4907]: I0313 15:38:07.367649 4907 generic.go:334] "Generic (PLEG): container finished" podID="46d919ab-084b-46cf-a55a-80060d0f89e5" containerID="1c7e727a5c90422b9d55f69cb636beca42802c6044e4986c0e34da16172b738a" exitCode=0 Mar 13 15:38:07 crc kubenswrapper[4907]: I0313 15:38:07.367695 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5k9nb" event={"ID":"46d919ab-084b-46cf-a55a-80060d0f89e5","Type":"ContainerDied","Data":"1c7e727a5c90422b9d55f69cb636beca42802c6044e4986c0e34da16172b738a"} Mar 13 15:38:07 crc kubenswrapper[4907]: I0313 15:38:07.367739 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5k9nb" event={"ID":"46d919ab-084b-46cf-a55a-80060d0f89e5","Type":"ContainerDied","Data":"6bf1cbbbca2dc292f4b7fb3a637b22871e833710be9542b6cacb01ec4e9d5066"} Mar 13 15:38:07 crc kubenswrapper[4907]: I0313 15:38:07.367738 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5k9nb" Mar 13 15:38:07 crc kubenswrapper[4907]: I0313 15:38:07.367755 4907 scope.go:117] "RemoveContainer" containerID="1c7e727a5c90422b9d55f69cb636beca42802c6044e4986c0e34da16172b738a" Mar 13 15:38:07 crc kubenswrapper[4907]: I0313 15:38:07.389499 4907 scope.go:117] "RemoveContainer" containerID="b930f886e923a7aeedd0f47b3c8af96617296760b49fcc8a4cd563a5557bb45c" Mar 13 15:38:07 crc kubenswrapper[4907]: I0313 15:38:07.408513 4907 scope.go:117] "RemoveContainer" containerID="07e6a739eb2cd151b3d0c8f4d016361ec920ef5cee0ff3c0cf3353c888330ada" Mar 13 15:38:07 crc kubenswrapper[4907]: I0313 15:38:07.421322 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/46d919ab-084b-46cf-a55a-80060d0f89e5-utilities\") pod \"46d919ab-084b-46cf-a55a-80060d0f89e5\" (UID: \"46d919ab-084b-46cf-a55a-80060d0f89e5\") " Mar 13 15:38:07 crc kubenswrapper[4907]: I0313 15:38:07.421438 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pm6kc\" (UniqueName: \"kubernetes.io/projected/46d919ab-084b-46cf-a55a-80060d0f89e5-kube-api-access-pm6kc\") pod \"46d919ab-084b-46cf-a55a-80060d0f89e5\" (UID: \"46d919ab-084b-46cf-a55a-80060d0f89e5\") " Mar 13 15:38:07 crc kubenswrapper[4907]: I0313 15:38:07.421549 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/46d919ab-084b-46cf-a55a-80060d0f89e5-catalog-content\") pod \"46d919ab-084b-46cf-a55a-80060d0f89e5\" (UID: \"46d919ab-084b-46cf-a55a-80060d0f89e5\") " Mar 13 15:38:07 crc kubenswrapper[4907]: I0313 15:38:07.422696 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/46d919ab-084b-46cf-a55a-80060d0f89e5-utilities" (OuterVolumeSpecName: "utilities") pod "46d919ab-084b-46cf-a55a-80060d0f89e5" (UID: "46d919ab-084b-46cf-a55a-80060d0f89e5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:38:07 crc kubenswrapper[4907]: I0313 15:38:07.428755 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/46d919ab-084b-46cf-a55a-80060d0f89e5-kube-api-access-pm6kc" (OuterVolumeSpecName: "kube-api-access-pm6kc") pod "46d919ab-084b-46cf-a55a-80060d0f89e5" (UID: "46d919ab-084b-46cf-a55a-80060d0f89e5"). InnerVolumeSpecName "kube-api-access-pm6kc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:38:07 crc kubenswrapper[4907]: I0313 15:38:07.490861 4907 scope.go:117] "RemoveContainer" containerID="1c7e727a5c90422b9d55f69cb636beca42802c6044e4986c0e34da16172b738a" Mar 13 15:38:07 crc kubenswrapper[4907]: E0313 15:38:07.492978 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1c7e727a5c90422b9d55f69cb636beca42802c6044e4986c0e34da16172b738a\": container with ID starting with 1c7e727a5c90422b9d55f69cb636beca42802c6044e4986c0e34da16172b738a not found: ID does not exist" containerID="1c7e727a5c90422b9d55f69cb636beca42802c6044e4986c0e34da16172b738a" Mar 13 15:38:07 crc kubenswrapper[4907]: I0313 15:38:07.493057 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c7e727a5c90422b9d55f69cb636beca42802c6044e4986c0e34da16172b738a"} err="failed to get container status \"1c7e727a5c90422b9d55f69cb636beca42802c6044e4986c0e34da16172b738a\": rpc error: code = NotFound desc = could not find container \"1c7e727a5c90422b9d55f69cb636beca42802c6044e4986c0e34da16172b738a\": container with ID starting with 1c7e727a5c90422b9d55f69cb636beca42802c6044e4986c0e34da16172b738a not found: ID does not exist" Mar 13 15:38:07 crc kubenswrapper[4907]: I0313 15:38:07.493090 4907 scope.go:117] "RemoveContainer" containerID="b930f886e923a7aeedd0f47b3c8af96617296760b49fcc8a4cd563a5557bb45c" Mar 13 15:38:07 crc kubenswrapper[4907]: E0313 15:38:07.493548 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b930f886e923a7aeedd0f47b3c8af96617296760b49fcc8a4cd563a5557bb45c\": container with ID starting with b930f886e923a7aeedd0f47b3c8af96617296760b49fcc8a4cd563a5557bb45c not found: ID does not exist" containerID="b930f886e923a7aeedd0f47b3c8af96617296760b49fcc8a4cd563a5557bb45c" Mar 13 15:38:07 crc kubenswrapper[4907]: I0313 15:38:07.493578 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b930f886e923a7aeedd0f47b3c8af96617296760b49fcc8a4cd563a5557bb45c"} err="failed to get container status \"b930f886e923a7aeedd0f47b3c8af96617296760b49fcc8a4cd563a5557bb45c\": rpc error: code = NotFound desc = could not find container \"b930f886e923a7aeedd0f47b3c8af96617296760b49fcc8a4cd563a5557bb45c\": container with ID starting with b930f886e923a7aeedd0f47b3c8af96617296760b49fcc8a4cd563a5557bb45c not found: ID does not exist" Mar 13 15:38:07 crc kubenswrapper[4907]: I0313 15:38:07.493597 4907 scope.go:117] "RemoveContainer" containerID="07e6a739eb2cd151b3d0c8f4d016361ec920ef5cee0ff3c0cf3353c888330ada" Mar 13 15:38:07 crc kubenswrapper[4907]: E0313 15:38:07.493920 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"07e6a739eb2cd151b3d0c8f4d016361ec920ef5cee0ff3c0cf3353c888330ada\": container with ID starting with 07e6a739eb2cd151b3d0c8f4d016361ec920ef5cee0ff3c0cf3353c888330ada not found: ID does not exist" containerID="07e6a739eb2cd151b3d0c8f4d016361ec920ef5cee0ff3c0cf3353c888330ada" Mar 13 15:38:07 crc kubenswrapper[4907]: I0313 15:38:07.493939 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"07e6a739eb2cd151b3d0c8f4d016361ec920ef5cee0ff3c0cf3353c888330ada"} err="failed to get container status \"07e6a739eb2cd151b3d0c8f4d016361ec920ef5cee0ff3c0cf3353c888330ada\": rpc error: code = NotFound desc = could not find container \"07e6a739eb2cd151b3d0c8f4d016361ec920ef5cee0ff3c0cf3353c888330ada\": container with ID starting with 07e6a739eb2cd151b3d0c8f4d016361ec920ef5cee0ff3c0cf3353c888330ada not found: ID does not exist" Mar 13 15:38:07 crc kubenswrapper[4907]: I0313 15:38:07.497831 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/46d919ab-084b-46cf-a55a-80060d0f89e5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "46d919ab-084b-46cf-a55a-80060d0f89e5" (UID: "46d919ab-084b-46cf-a55a-80060d0f89e5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:38:07 crc kubenswrapper[4907]: I0313 15:38:07.524661 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pm6kc\" (UniqueName: \"kubernetes.io/projected/46d919ab-084b-46cf-a55a-80060d0f89e5-kube-api-access-pm6kc\") on node \"crc\" DevicePath \"\"" Mar 13 15:38:07 crc kubenswrapper[4907]: I0313 15:38:07.524705 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/46d919ab-084b-46cf-a55a-80060d0f89e5-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 15:38:07 crc kubenswrapper[4907]: I0313 15:38:07.524718 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/46d919ab-084b-46cf-a55a-80060d0f89e5-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 15:38:07 crc kubenswrapper[4907]: I0313 15:38:07.765923 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-kxfg2" Mar 13 15:38:07 crc kubenswrapper[4907]: I0313 15:38:07.781264 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-fd12-account-create-update-kxbcs" Mar 13 15:38:07 crc kubenswrapper[4907]: I0313 15:38:07.810832 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5k9nb"] Mar 13 15:38:07 crc kubenswrapper[4907]: I0313 15:38:07.810901 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-5k9nb"] Mar 13 15:38:07 crc kubenswrapper[4907]: I0313 15:38:07.829847 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0d6a6021-7cc4-42f9-848b-5bbfc3650ee7-operator-scripts\") pod \"0d6a6021-7cc4-42f9-848b-5bbfc3650ee7\" (UID: \"0d6a6021-7cc4-42f9-848b-5bbfc3650ee7\") " Mar 13 15:38:07 crc kubenswrapper[4907]: I0313 15:38:07.830678 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0d6a6021-7cc4-42f9-848b-5bbfc3650ee7-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "0d6a6021-7cc4-42f9-848b-5bbfc3650ee7" (UID: "0d6a6021-7cc4-42f9-848b-5bbfc3650ee7"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:38:07 crc kubenswrapper[4907]: I0313 15:38:07.933127 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tzph2\" (UniqueName: \"kubernetes.io/projected/440d4795-da82-461c-b460-bdd5c271de4d-kube-api-access-tzph2\") pod \"440d4795-da82-461c-b460-bdd5c271de4d\" (UID: \"440d4795-da82-461c-b460-bdd5c271de4d\") " Mar 13 15:38:07 crc kubenswrapper[4907]: I0313 15:38:07.933261 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-99dgh\" (UniqueName: \"kubernetes.io/projected/0d6a6021-7cc4-42f9-848b-5bbfc3650ee7-kube-api-access-99dgh\") pod \"0d6a6021-7cc4-42f9-848b-5bbfc3650ee7\" (UID: \"0d6a6021-7cc4-42f9-848b-5bbfc3650ee7\") " Mar 13 15:38:07 crc kubenswrapper[4907]: I0313 15:38:07.933367 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/440d4795-da82-461c-b460-bdd5c271de4d-operator-scripts\") pod \"440d4795-da82-461c-b460-bdd5c271de4d\" (UID: \"440d4795-da82-461c-b460-bdd5c271de4d\") " Mar 13 15:38:07 crc kubenswrapper[4907]: I0313 15:38:07.934034 4907 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0d6a6021-7cc4-42f9-848b-5bbfc3650ee7-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 15:38:07 crc kubenswrapper[4907]: I0313 15:38:07.934034 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/440d4795-da82-461c-b460-bdd5c271de4d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "440d4795-da82-461c-b460-bdd5c271de4d" (UID: "440d4795-da82-461c-b460-bdd5c271de4d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:38:07 crc kubenswrapper[4907]: I0313 15:38:07.938976 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/440d4795-da82-461c-b460-bdd5c271de4d-kube-api-access-tzph2" (OuterVolumeSpecName: "kube-api-access-tzph2") pod "440d4795-da82-461c-b460-bdd5c271de4d" (UID: "440d4795-da82-461c-b460-bdd5c271de4d"). InnerVolumeSpecName "kube-api-access-tzph2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:38:07 crc kubenswrapper[4907]: I0313 15:38:07.940016 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0d6a6021-7cc4-42f9-848b-5bbfc3650ee7-kube-api-access-99dgh" (OuterVolumeSpecName: "kube-api-access-99dgh") pod "0d6a6021-7cc4-42f9-848b-5bbfc3650ee7" (UID: "0d6a6021-7cc4-42f9-848b-5bbfc3650ee7"). InnerVolumeSpecName "kube-api-access-99dgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:38:08 crc kubenswrapper[4907]: I0313 15:38:08.034516 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tzph2\" (UniqueName: \"kubernetes.io/projected/440d4795-da82-461c-b460-bdd5c271de4d-kube-api-access-tzph2\") on node \"crc\" DevicePath \"\"" Mar 13 15:38:08 crc kubenswrapper[4907]: I0313 15:38:08.034555 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-99dgh\" (UniqueName: \"kubernetes.io/projected/0d6a6021-7cc4-42f9-848b-5bbfc3650ee7-kube-api-access-99dgh\") on node \"crc\" DevicePath \"\"" Mar 13 15:38:08 crc kubenswrapper[4907]: I0313 15:38:08.034565 4907 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/440d4795-da82-461c-b460-bdd5c271de4d-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 15:38:08 crc kubenswrapper[4907]: I0313 15:38:08.379695 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-kxfg2" Mar 13 15:38:08 crc kubenswrapper[4907]: I0313 15:38:08.379731 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-kxfg2" event={"ID":"440d4795-da82-461c-b460-bdd5c271de4d","Type":"ContainerDied","Data":"7a689c1f01fe0ef145be6f24241adc9dc1ca896caf778fafb7281e1050d46874"} Mar 13 15:38:08 crc kubenswrapper[4907]: I0313 15:38:08.379783 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7a689c1f01fe0ef145be6f24241adc9dc1ca896caf778fafb7281e1050d46874" Mar 13 15:38:08 crc kubenswrapper[4907]: I0313 15:38:08.383066 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-fd12-account-create-update-kxbcs" event={"ID":"0d6a6021-7cc4-42f9-848b-5bbfc3650ee7","Type":"ContainerDied","Data":"83481b88571f9dd6758c4725b16318418738a3b1dc14c5db9de531f76d808495"} Mar 13 15:38:08 crc kubenswrapper[4907]: I0313 15:38:08.383124 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="83481b88571f9dd6758c4725b16318418738a3b1dc14c5db9de531f76d808495" Mar 13 15:38:08 crc kubenswrapper[4907]: I0313 15:38:08.383079 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-fd12-account-create-update-kxbcs" Mar 13 15:38:09 crc kubenswrapper[4907]: I0313 15:38:09.796586 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="46d919ab-084b-46cf-a55a-80060d0f89e5" path="/var/lib/kubelet/pods/46d919ab-084b-46cf-a55a-80060d0f89e5/volumes" Mar 13 15:38:10 crc kubenswrapper[4907]: I0313 15:38:10.062604 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-tlb57"] Mar 13 15:38:10 crc kubenswrapper[4907]: E0313 15:38:10.062969 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46d919ab-084b-46cf-a55a-80060d0f89e5" containerName="extract-utilities" Mar 13 15:38:10 crc kubenswrapper[4907]: I0313 15:38:10.062984 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="46d919ab-084b-46cf-a55a-80060d0f89e5" containerName="extract-utilities" Mar 13 15:38:10 crc kubenswrapper[4907]: E0313 15:38:10.062998 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="440d4795-da82-461c-b460-bdd5c271de4d" containerName="mariadb-database-create" Mar 13 15:38:10 crc kubenswrapper[4907]: I0313 15:38:10.063008 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="440d4795-da82-461c-b460-bdd5c271de4d" containerName="mariadb-database-create" Mar 13 15:38:10 crc kubenswrapper[4907]: E0313 15:38:10.063023 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d6a6021-7cc4-42f9-848b-5bbfc3650ee7" containerName="mariadb-account-create-update" Mar 13 15:38:10 crc kubenswrapper[4907]: I0313 15:38:10.063032 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d6a6021-7cc4-42f9-848b-5bbfc3650ee7" containerName="mariadb-account-create-update" Mar 13 15:38:10 crc kubenswrapper[4907]: E0313 15:38:10.063049 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46d919ab-084b-46cf-a55a-80060d0f89e5" containerName="extract-content" Mar 13 15:38:10 crc kubenswrapper[4907]: I0313 15:38:10.063056 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="46d919ab-084b-46cf-a55a-80060d0f89e5" containerName="extract-content" Mar 13 15:38:10 crc kubenswrapper[4907]: E0313 15:38:10.063067 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46d919ab-084b-46cf-a55a-80060d0f89e5" containerName="registry-server" Mar 13 15:38:10 crc kubenswrapper[4907]: I0313 15:38:10.063074 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="46d919ab-084b-46cf-a55a-80060d0f89e5" containerName="registry-server" Mar 13 15:38:10 crc kubenswrapper[4907]: I0313 15:38:10.063232 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d6a6021-7cc4-42f9-848b-5bbfc3650ee7" containerName="mariadb-account-create-update" Mar 13 15:38:10 crc kubenswrapper[4907]: I0313 15:38:10.063245 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="440d4795-da82-461c-b460-bdd5c271de4d" containerName="mariadb-database-create" Mar 13 15:38:10 crc kubenswrapper[4907]: I0313 15:38:10.063252 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="46d919ab-084b-46cf-a55a-80060d0f89e5" containerName="registry-server" Mar 13 15:38:10 crc kubenswrapper[4907]: I0313 15:38:10.063793 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-tlb57" Mar 13 15:38:10 crc kubenswrapper[4907]: I0313 15:38:10.065648 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-5c589" Mar 13 15:38:10 crc kubenswrapper[4907]: I0313 15:38:10.065726 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Mar 13 15:38:10 crc kubenswrapper[4907]: I0313 15:38:10.069637 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9t6d2\" (UniqueName: \"kubernetes.io/projected/13c6b679-96d2-407c-8b94-cd908d61945b-kube-api-access-9t6d2\") pod \"glance-db-sync-tlb57\" (UID: \"13c6b679-96d2-407c-8b94-cd908d61945b\") " pod="openstack/glance-db-sync-tlb57" Mar 13 15:38:10 crc kubenswrapper[4907]: I0313 15:38:10.069696 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/13c6b679-96d2-407c-8b94-cd908d61945b-db-sync-config-data\") pod \"glance-db-sync-tlb57\" (UID: \"13c6b679-96d2-407c-8b94-cd908d61945b\") " pod="openstack/glance-db-sync-tlb57" Mar 13 15:38:10 crc kubenswrapper[4907]: I0313 15:38:10.069867 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/13c6b679-96d2-407c-8b94-cd908d61945b-config-data\") pod \"glance-db-sync-tlb57\" (UID: \"13c6b679-96d2-407c-8b94-cd908d61945b\") " pod="openstack/glance-db-sync-tlb57" Mar 13 15:38:10 crc kubenswrapper[4907]: I0313 15:38:10.069947 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13c6b679-96d2-407c-8b94-cd908d61945b-combined-ca-bundle\") pod \"glance-db-sync-tlb57\" (UID: \"13c6b679-96d2-407c-8b94-cd908d61945b\") " pod="openstack/glance-db-sync-tlb57" Mar 13 15:38:10 crc kubenswrapper[4907]: I0313 15:38:10.087525 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-tlb57"] Mar 13 15:38:10 crc kubenswrapper[4907]: I0313 15:38:10.171631 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/13c6b679-96d2-407c-8b94-cd908d61945b-config-data\") pod \"glance-db-sync-tlb57\" (UID: \"13c6b679-96d2-407c-8b94-cd908d61945b\") " pod="openstack/glance-db-sync-tlb57" Mar 13 15:38:10 crc kubenswrapper[4907]: I0313 15:38:10.171720 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13c6b679-96d2-407c-8b94-cd908d61945b-combined-ca-bundle\") pod \"glance-db-sync-tlb57\" (UID: \"13c6b679-96d2-407c-8b94-cd908d61945b\") " pod="openstack/glance-db-sync-tlb57" Mar 13 15:38:10 crc kubenswrapper[4907]: I0313 15:38:10.171801 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9t6d2\" (UniqueName: \"kubernetes.io/projected/13c6b679-96d2-407c-8b94-cd908d61945b-kube-api-access-9t6d2\") pod \"glance-db-sync-tlb57\" (UID: \"13c6b679-96d2-407c-8b94-cd908d61945b\") " pod="openstack/glance-db-sync-tlb57" Mar 13 15:38:10 crc kubenswrapper[4907]: I0313 15:38:10.172243 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/13c6b679-96d2-407c-8b94-cd908d61945b-db-sync-config-data\") pod \"glance-db-sync-tlb57\" (UID: \"13c6b679-96d2-407c-8b94-cd908d61945b\") " pod="openstack/glance-db-sync-tlb57" Mar 13 15:38:10 crc kubenswrapper[4907]: I0313 15:38:10.176680 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/13c6b679-96d2-407c-8b94-cd908d61945b-config-data\") pod \"glance-db-sync-tlb57\" (UID: \"13c6b679-96d2-407c-8b94-cd908d61945b\") " pod="openstack/glance-db-sync-tlb57" Mar 13 15:38:10 crc kubenswrapper[4907]: I0313 15:38:10.177123 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/13c6b679-96d2-407c-8b94-cd908d61945b-db-sync-config-data\") pod \"glance-db-sync-tlb57\" (UID: \"13c6b679-96d2-407c-8b94-cd908d61945b\") " pod="openstack/glance-db-sync-tlb57" Mar 13 15:38:10 crc kubenswrapper[4907]: I0313 15:38:10.178287 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13c6b679-96d2-407c-8b94-cd908d61945b-combined-ca-bundle\") pod \"glance-db-sync-tlb57\" (UID: \"13c6b679-96d2-407c-8b94-cd908d61945b\") " pod="openstack/glance-db-sync-tlb57" Mar 13 15:38:10 crc kubenswrapper[4907]: I0313 15:38:10.191036 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9t6d2\" (UniqueName: \"kubernetes.io/projected/13c6b679-96d2-407c-8b94-cd908d61945b-kube-api-access-9t6d2\") pod \"glance-db-sync-tlb57\" (UID: \"13c6b679-96d2-407c-8b94-cd908d61945b\") " pod="openstack/glance-db-sync-tlb57" Mar 13 15:38:10 crc kubenswrapper[4907]: I0313 15:38:10.386579 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-tlb57" Mar 13 15:38:11 crc kubenswrapper[4907]: I0313 15:38:11.017481 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-tlb57"] Mar 13 15:38:11 crc kubenswrapper[4907]: I0313 15:38:11.431840 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-tlb57" event={"ID":"13c6b679-96d2-407c-8b94-cd908d61945b","Type":"ContainerStarted","Data":"3232e9eb498f3e8fa1c6c1dce83de18cc94e2bb946c98b463971e7acd7900377"} Mar 13 15:38:12 crc kubenswrapper[4907]: I0313 15:38:12.439953 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-tlb57" event={"ID":"13c6b679-96d2-407c-8b94-cd908d61945b","Type":"ContainerStarted","Data":"75dd2c3ec3d1e72efcbe86eb956d5d57de584abd904ef3f55fe25aad75bb5498"} Mar 13 15:38:12 crc kubenswrapper[4907]: I0313 15:38:12.461989 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-tlb57" podStartSLOduration=2.461972963 podStartE2EDuration="2.461972963s" podCreationTimestamp="2026-03-13 15:38:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:38:12.453130921 +0000 UTC m=+5591.352918610" watchObservedRunningTime="2026-03-13 15:38:12.461972963 +0000 UTC m=+5591.361760642" Mar 13 15:38:15 crc kubenswrapper[4907]: I0313 15:38:15.474788 4907 generic.go:334] "Generic (PLEG): container finished" podID="13c6b679-96d2-407c-8b94-cd908d61945b" containerID="75dd2c3ec3d1e72efcbe86eb956d5d57de584abd904ef3f55fe25aad75bb5498" exitCode=0 Mar 13 15:38:15 crc kubenswrapper[4907]: I0313 15:38:15.474899 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-tlb57" event={"ID":"13c6b679-96d2-407c-8b94-cd908d61945b","Type":"ContainerDied","Data":"75dd2c3ec3d1e72efcbe86eb956d5d57de584abd904ef3f55fe25aad75bb5498"} Mar 13 15:38:16 crc kubenswrapper[4907]: I0313 15:38:16.853325 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-tlb57" Mar 13 15:38:16 crc kubenswrapper[4907]: I0313 15:38:16.890124 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13c6b679-96d2-407c-8b94-cd908d61945b-combined-ca-bundle\") pod \"13c6b679-96d2-407c-8b94-cd908d61945b\" (UID: \"13c6b679-96d2-407c-8b94-cd908d61945b\") " Mar 13 15:38:16 crc kubenswrapper[4907]: I0313 15:38:16.890225 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/13c6b679-96d2-407c-8b94-cd908d61945b-db-sync-config-data\") pod \"13c6b679-96d2-407c-8b94-cd908d61945b\" (UID: \"13c6b679-96d2-407c-8b94-cd908d61945b\") " Mar 13 15:38:16 crc kubenswrapper[4907]: I0313 15:38:16.890297 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/13c6b679-96d2-407c-8b94-cd908d61945b-config-data\") pod \"13c6b679-96d2-407c-8b94-cd908d61945b\" (UID: \"13c6b679-96d2-407c-8b94-cd908d61945b\") " Mar 13 15:38:16 crc kubenswrapper[4907]: I0313 15:38:16.890362 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9t6d2\" (UniqueName: \"kubernetes.io/projected/13c6b679-96d2-407c-8b94-cd908d61945b-kube-api-access-9t6d2\") pod \"13c6b679-96d2-407c-8b94-cd908d61945b\" (UID: \"13c6b679-96d2-407c-8b94-cd908d61945b\") " Mar 13 15:38:16 crc kubenswrapper[4907]: I0313 15:38:16.899222 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/13c6b679-96d2-407c-8b94-cd908d61945b-kube-api-access-9t6d2" (OuterVolumeSpecName: "kube-api-access-9t6d2") pod "13c6b679-96d2-407c-8b94-cd908d61945b" (UID: "13c6b679-96d2-407c-8b94-cd908d61945b"). InnerVolumeSpecName "kube-api-access-9t6d2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:38:16 crc kubenswrapper[4907]: I0313 15:38:16.901638 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/13c6b679-96d2-407c-8b94-cd908d61945b-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "13c6b679-96d2-407c-8b94-cd908d61945b" (UID: "13c6b679-96d2-407c-8b94-cd908d61945b"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:38:16 crc kubenswrapper[4907]: I0313 15:38:16.927806 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/13c6b679-96d2-407c-8b94-cd908d61945b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "13c6b679-96d2-407c-8b94-cd908d61945b" (UID: "13c6b679-96d2-407c-8b94-cd908d61945b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:38:16 crc kubenswrapper[4907]: I0313 15:38:16.940973 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/13c6b679-96d2-407c-8b94-cd908d61945b-config-data" (OuterVolumeSpecName: "config-data") pod "13c6b679-96d2-407c-8b94-cd908d61945b" (UID: "13c6b679-96d2-407c-8b94-cd908d61945b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:38:16 crc kubenswrapper[4907]: I0313 15:38:16.991583 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9t6d2\" (UniqueName: \"kubernetes.io/projected/13c6b679-96d2-407c-8b94-cd908d61945b-kube-api-access-9t6d2\") on node \"crc\" DevicePath \"\"" Mar 13 15:38:16 crc kubenswrapper[4907]: I0313 15:38:16.991626 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13c6b679-96d2-407c-8b94-cd908d61945b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 15:38:16 crc kubenswrapper[4907]: I0313 15:38:16.991640 4907 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/13c6b679-96d2-407c-8b94-cd908d61945b-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 15:38:16 crc kubenswrapper[4907]: I0313 15:38:16.991652 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/13c6b679-96d2-407c-8b94-cd908d61945b-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 15:38:17 crc kubenswrapper[4907]: I0313 15:38:17.489354 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-tlb57" event={"ID":"13c6b679-96d2-407c-8b94-cd908d61945b","Type":"ContainerDied","Data":"3232e9eb498f3e8fa1c6c1dce83de18cc94e2bb946c98b463971e7acd7900377"} Mar 13 15:38:17 crc kubenswrapper[4907]: I0313 15:38:17.489399 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3232e9eb498f3e8fa1c6c1dce83de18cc94e2bb946c98b463971e7acd7900377" Mar 13 15:38:17 crc kubenswrapper[4907]: I0313 15:38:17.489433 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-tlb57" Mar 13 15:38:17 crc kubenswrapper[4907]: I0313 15:38:17.753089 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Mar 13 15:38:17 crc kubenswrapper[4907]: E0313 15:38:17.753465 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13c6b679-96d2-407c-8b94-cd908d61945b" containerName="glance-db-sync" Mar 13 15:38:17 crc kubenswrapper[4907]: I0313 15:38:17.753483 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="13c6b679-96d2-407c-8b94-cd908d61945b" containerName="glance-db-sync" Mar 13 15:38:17 crc kubenswrapper[4907]: I0313 15:38:17.753636 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="13c6b679-96d2-407c-8b94-cd908d61945b" containerName="glance-db-sync" Mar 13 15:38:17 crc kubenswrapper[4907]: I0313 15:38:17.756779 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 13 15:38:17 crc kubenswrapper[4907]: I0313 15:38:17.761300 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Mar 13 15:38:17 crc kubenswrapper[4907]: I0313 15:38:17.761393 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-5c589" Mar 13 15:38:17 crc kubenswrapper[4907]: I0313 15:38:17.761473 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Mar 13 15:38:17 crc kubenswrapper[4907]: I0313 15:38:17.761761 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Mar 13 15:38:17 crc kubenswrapper[4907]: I0313 15:38:17.774754 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 13 15:38:17 crc kubenswrapper[4907]: I0313 15:38:17.803970 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-whgs7\" (UniqueName: \"kubernetes.io/projected/d38a5143-6b9b-4f3e-a41a-63668c7516ee-kube-api-access-whgs7\") pod \"glance-default-external-api-0\" (UID: \"d38a5143-6b9b-4f3e-a41a-63668c7516ee\") " pod="openstack/glance-default-external-api-0" Mar 13 15:38:17 crc kubenswrapper[4907]: I0313 15:38:17.804045 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d38a5143-6b9b-4f3e-a41a-63668c7516ee-logs\") pod \"glance-default-external-api-0\" (UID: \"d38a5143-6b9b-4f3e-a41a-63668c7516ee\") " pod="openstack/glance-default-external-api-0" Mar 13 15:38:17 crc kubenswrapper[4907]: I0313 15:38:17.804070 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d38a5143-6b9b-4f3e-a41a-63668c7516ee-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"d38a5143-6b9b-4f3e-a41a-63668c7516ee\") " pod="openstack/glance-default-external-api-0" Mar 13 15:38:17 crc kubenswrapper[4907]: I0313 15:38:17.804104 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/d38a5143-6b9b-4f3e-a41a-63668c7516ee-ceph\") pod \"glance-default-external-api-0\" (UID: \"d38a5143-6b9b-4f3e-a41a-63668c7516ee\") " pod="openstack/glance-default-external-api-0" Mar 13 15:38:17 crc kubenswrapper[4907]: I0313 15:38:17.804121 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d38a5143-6b9b-4f3e-a41a-63668c7516ee-scripts\") pod \"glance-default-external-api-0\" (UID: \"d38a5143-6b9b-4f3e-a41a-63668c7516ee\") " pod="openstack/glance-default-external-api-0" Mar 13 15:38:17 crc kubenswrapper[4907]: I0313 15:38:17.804148 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d38a5143-6b9b-4f3e-a41a-63668c7516ee-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"d38a5143-6b9b-4f3e-a41a-63668c7516ee\") " pod="openstack/glance-default-external-api-0" Mar 13 15:38:17 crc kubenswrapper[4907]: I0313 15:38:17.804168 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d38a5143-6b9b-4f3e-a41a-63668c7516ee-config-data\") pod \"glance-default-external-api-0\" (UID: \"d38a5143-6b9b-4f3e-a41a-63668c7516ee\") " pod="openstack/glance-default-external-api-0" Mar 13 15:38:17 crc kubenswrapper[4907]: I0313 15:38:17.883766 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-746ff5d579-xk972"] Mar 13 15:38:17 crc kubenswrapper[4907]: I0313 15:38:17.887458 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-746ff5d579-xk972" Mar 13 15:38:17 crc kubenswrapper[4907]: I0313 15:38:17.905816 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d38a5143-6b9b-4f3e-a41a-63668c7516ee-config-data\") pod \"glance-default-external-api-0\" (UID: \"d38a5143-6b9b-4f3e-a41a-63668c7516ee\") " pod="openstack/glance-default-external-api-0" Mar 13 15:38:17 crc kubenswrapper[4907]: I0313 15:38:17.905919 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-whgs7\" (UniqueName: \"kubernetes.io/projected/d38a5143-6b9b-4f3e-a41a-63668c7516ee-kube-api-access-whgs7\") pod \"glance-default-external-api-0\" (UID: \"d38a5143-6b9b-4f3e-a41a-63668c7516ee\") " pod="openstack/glance-default-external-api-0" Mar 13 15:38:17 crc kubenswrapper[4907]: I0313 15:38:17.906801 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d38a5143-6b9b-4f3e-a41a-63668c7516ee-logs\") pod \"glance-default-external-api-0\" (UID: \"d38a5143-6b9b-4f3e-a41a-63668c7516ee\") " pod="openstack/glance-default-external-api-0" Mar 13 15:38:17 crc kubenswrapper[4907]: I0313 15:38:17.906875 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d38a5143-6b9b-4f3e-a41a-63668c7516ee-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"d38a5143-6b9b-4f3e-a41a-63668c7516ee\") " pod="openstack/glance-default-external-api-0" Mar 13 15:38:17 crc kubenswrapper[4907]: I0313 15:38:17.907138 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/d38a5143-6b9b-4f3e-a41a-63668c7516ee-ceph\") pod \"glance-default-external-api-0\" (UID: \"d38a5143-6b9b-4f3e-a41a-63668c7516ee\") " pod="openstack/glance-default-external-api-0" Mar 13 15:38:17 crc kubenswrapper[4907]: I0313 15:38:17.907177 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d38a5143-6b9b-4f3e-a41a-63668c7516ee-scripts\") pod \"glance-default-external-api-0\" (UID: \"d38a5143-6b9b-4f3e-a41a-63668c7516ee\") " pod="openstack/glance-default-external-api-0" Mar 13 15:38:17 crc kubenswrapper[4907]: I0313 15:38:17.907380 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d38a5143-6b9b-4f3e-a41a-63668c7516ee-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"d38a5143-6b9b-4f3e-a41a-63668c7516ee\") " pod="openstack/glance-default-external-api-0" Mar 13 15:38:17 crc kubenswrapper[4907]: I0313 15:38:17.907497 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d38a5143-6b9b-4f3e-a41a-63668c7516ee-logs\") pod \"glance-default-external-api-0\" (UID: \"d38a5143-6b9b-4f3e-a41a-63668c7516ee\") " pod="openstack/glance-default-external-api-0" Mar 13 15:38:17 crc kubenswrapper[4907]: I0313 15:38:17.907704 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d38a5143-6b9b-4f3e-a41a-63668c7516ee-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"d38a5143-6b9b-4f3e-a41a-63668c7516ee\") " pod="openstack/glance-default-external-api-0" Mar 13 15:38:17 crc kubenswrapper[4907]: I0313 15:38:17.914074 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/d38a5143-6b9b-4f3e-a41a-63668c7516ee-ceph\") pod \"glance-default-external-api-0\" (UID: \"d38a5143-6b9b-4f3e-a41a-63668c7516ee\") " pod="openstack/glance-default-external-api-0" Mar 13 15:38:17 crc kubenswrapper[4907]: I0313 15:38:17.914767 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d38a5143-6b9b-4f3e-a41a-63668c7516ee-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"d38a5143-6b9b-4f3e-a41a-63668c7516ee\") " pod="openstack/glance-default-external-api-0" Mar 13 15:38:17 crc kubenswrapper[4907]: I0313 15:38:17.917773 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d38a5143-6b9b-4f3e-a41a-63668c7516ee-config-data\") pod \"glance-default-external-api-0\" (UID: \"d38a5143-6b9b-4f3e-a41a-63668c7516ee\") " pod="openstack/glance-default-external-api-0" Mar 13 15:38:17 crc kubenswrapper[4907]: I0313 15:38:17.924766 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-746ff5d579-xk972"] Mar 13 15:38:17 crc kubenswrapper[4907]: I0313 15:38:17.934701 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d38a5143-6b9b-4f3e-a41a-63668c7516ee-scripts\") pod \"glance-default-external-api-0\" (UID: \"d38a5143-6b9b-4f3e-a41a-63668c7516ee\") " pod="openstack/glance-default-external-api-0" Mar 13 15:38:17 crc kubenswrapper[4907]: I0313 15:38:17.940669 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-whgs7\" (UniqueName: \"kubernetes.io/projected/d38a5143-6b9b-4f3e-a41a-63668c7516ee-kube-api-access-whgs7\") pod \"glance-default-external-api-0\" (UID: \"d38a5143-6b9b-4f3e-a41a-63668c7516ee\") " pod="openstack/glance-default-external-api-0" Mar 13 15:38:17 crc kubenswrapper[4907]: I0313 15:38:17.973200 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 13 15:38:17 crc kubenswrapper[4907]: I0313 15:38:17.978217 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 13 15:38:17 crc kubenswrapper[4907]: I0313 15:38:17.986904 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Mar 13 15:38:17 crc kubenswrapper[4907]: I0313 15:38:17.988289 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 13 15:38:18 crc kubenswrapper[4907]: I0313 15:38:18.009429 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4cf4ac52-db94-4e0c-a990-c2c6c64f3167-config\") pod \"dnsmasq-dns-746ff5d579-xk972\" (UID: \"4cf4ac52-db94-4e0c-a990-c2c6c64f3167\") " pod="openstack/dnsmasq-dns-746ff5d579-xk972" Mar 13 15:38:18 crc kubenswrapper[4907]: I0313 15:38:18.009530 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4cf4ac52-db94-4e0c-a990-c2c6c64f3167-dns-svc\") pod \"dnsmasq-dns-746ff5d579-xk972\" (UID: \"4cf4ac52-db94-4e0c-a990-c2c6c64f3167\") " pod="openstack/dnsmasq-dns-746ff5d579-xk972" Mar 13 15:38:18 crc kubenswrapper[4907]: I0313 15:38:18.009556 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4cf4ac52-db94-4e0c-a990-c2c6c64f3167-ovsdbserver-nb\") pod \"dnsmasq-dns-746ff5d579-xk972\" (UID: \"4cf4ac52-db94-4e0c-a990-c2c6c64f3167\") " pod="openstack/dnsmasq-dns-746ff5d579-xk972" Mar 13 15:38:18 crc kubenswrapper[4907]: I0313 15:38:18.009655 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4h66s\" (UniqueName: \"kubernetes.io/projected/4cf4ac52-db94-4e0c-a990-c2c6c64f3167-kube-api-access-4h66s\") pod \"dnsmasq-dns-746ff5d579-xk972\" (UID: \"4cf4ac52-db94-4e0c-a990-c2c6c64f3167\") " pod="openstack/dnsmasq-dns-746ff5d579-xk972" Mar 13 15:38:18 crc kubenswrapper[4907]: I0313 15:38:18.009724 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4cf4ac52-db94-4e0c-a990-c2c6c64f3167-ovsdbserver-sb\") pod \"dnsmasq-dns-746ff5d579-xk972\" (UID: \"4cf4ac52-db94-4e0c-a990-c2c6c64f3167\") " pod="openstack/dnsmasq-dns-746ff5d579-xk972" Mar 13 15:38:18 crc kubenswrapper[4907]: I0313 15:38:18.042531 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 15:38:18 crc kubenswrapper[4907]: I0313 15:38:18.042606 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 15:38:18 crc kubenswrapper[4907]: I0313 15:38:18.042660 4907 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" Mar 13 15:38:18 crc kubenswrapper[4907]: I0313 15:38:18.043399 4907 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"90a39cf81ab6b6b036397a9c751a635e5f4cd834dccdc53b2ff02dd0144698a4"} pod="openshift-machine-config-operator/machine-config-daemon-hm56j" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 15:38:18 crc kubenswrapper[4907]: I0313 15:38:18.043470 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" containerID="cri-o://90a39cf81ab6b6b036397a9c751a635e5f4cd834dccdc53b2ff02dd0144698a4" gracePeriod=600 Mar 13 15:38:18 crc kubenswrapper[4907]: I0313 15:38:18.080861 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 13 15:38:18 crc kubenswrapper[4907]: I0313 15:38:18.112037 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4h66s\" (UniqueName: \"kubernetes.io/projected/4cf4ac52-db94-4e0c-a990-c2c6c64f3167-kube-api-access-4h66s\") pod \"dnsmasq-dns-746ff5d579-xk972\" (UID: \"4cf4ac52-db94-4e0c-a990-c2c6c64f3167\") " pod="openstack/dnsmasq-dns-746ff5d579-xk972" Mar 13 15:38:18 crc kubenswrapper[4907]: I0313 15:38:18.112121 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4cf4ac52-db94-4e0c-a990-c2c6c64f3167-ovsdbserver-sb\") pod \"dnsmasq-dns-746ff5d579-xk972\" (UID: \"4cf4ac52-db94-4e0c-a990-c2c6c64f3167\") " pod="openstack/dnsmasq-dns-746ff5d579-xk972" Mar 13 15:38:18 crc kubenswrapper[4907]: I0313 15:38:18.112159 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3431e634-aed2-44d7-a5a2-251956514c4b-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"3431e634-aed2-44d7-a5a2-251956514c4b\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:38:18 crc kubenswrapper[4907]: I0313 15:38:18.112205 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3431e634-aed2-44d7-a5a2-251956514c4b-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"3431e634-aed2-44d7-a5a2-251956514c4b\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:38:18 crc kubenswrapper[4907]: I0313 15:38:18.112254 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vkjx6\" (UniqueName: \"kubernetes.io/projected/3431e634-aed2-44d7-a5a2-251956514c4b-kube-api-access-vkjx6\") pod \"glance-default-internal-api-0\" (UID: \"3431e634-aed2-44d7-a5a2-251956514c4b\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:38:18 crc kubenswrapper[4907]: I0313 15:38:18.112285 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3431e634-aed2-44d7-a5a2-251956514c4b-scripts\") pod \"glance-default-internal-api-0\" (UID: \"3431e634-aed2-44d7-a5a2-251956514c4b\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:38:18 crc kubenswrapper[4907]: I0313 15:38:18.112377 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4cf4ac52-db94-4e0c-a990-c2c6c64f3167-config\") pod \"dnsmasq-dns-746ff5d579-xk972\" (UID: \"4cf4ac52-db94-4e0c-a990-c2c6c64f3167\") " pod="openstack/dnsmasq-dns-746ff5d579-xk972" Mar 13 15:38:18 crc kubenswrapper[4907]: I0313 15:38:18.112400 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/3431e634-aed2-44d7-a5a2-251956514c4b-ceph\") pod \"glance-default-internal-api-0\" (UID: \"3431e634-aed2-44d7-a5a2-251956514c4b\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:38:18 crc kubenswrapper[4907]: I0313 15:38:18.112424 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3431e634-aed2-44d7-a5a2-251956514c4b-config-data\") pod \"glance-default-internal-api-0\" (UID: \"3431e634-aed2-44d7-a5a2-251956514c4b\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:38:18 crc kubenswrapper[4907]: I0313 15:38:18.112459 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3431e634-aed2-44d7-a5a2-251956514c4b-logs\") pod \"glance-default-internal-api-0\" (UID: \"3431e634-aed2-44d7-a5a2-251956514c4b\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:38:18 crc kubenswrapper[4907]: I0313 15:38:18.112488 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4cf4ac52-db94-4e0c-a990-c2c6c64f3167-dns-svc\") pod \"dnsmasq-dns-746ff5d579-xk972\" (UID: \"4cf4ac52-db94-4e0c-a990-c2c6c64f3167\") " pod="openstack/dnsmasq-dns-746ff5d579-xk972" Mar 13 15:38:18 crc kubenswrapper[4907]: I0313 15:38:18.112515 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4cf4ac52-db94-4e0c-a990-c2c6c64f3167-ovsdbserver-nb\") pod \"dnsmasq-dns-746ff5d579-xk972\" (UID: \"4cf4ac52-db94-4e0c-a990-c2c6c64f3167\") " pod="openstack/dnsmasq-dns-746ff5d579-xk972" Mar 13 15:38:18 crc kubenswrapper[4907]: I0313 15:38:18.113173 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4cf4ac52-db94-4e0c-a990-c2c6c64f3167-ovsdbserver-sb\") pod \"dnsmasq-dns-746ff5d579-xk972\" (UID: \"4cf4ac52-db94-4e0c-a990-c2c6c64f3167\") " pod="openstack/dnsmasq-dns-746ff5d579-xk972" Mar 13 15:38:18 crc kubenswrapper[4907]: I0313 15:38:18.113191 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4cf4ac52-db94-4e0c-a990-c2c6c64f3167-config\") pod \"dnsmasq-dns-746ff5d579-xk972\" (UID: \"4cf4ac52-db94-4e0c-a990-c2c6c64f3167\") " pod="openstack/dnsmasq-dns-746ff5d579-xk972" Mar 13 15:38:18 crc kubenswrapper[4907]: I0313 15:38:18.117630 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4cf4ac52-db94-4e0c-a990-c2c6c64f3167-ovsdbserver-nb\") pod \"dnsmasq-dns-746ff5d579-xk972\" (UID: \"4cf4ac52-db94-4e0c-a990-c2c6c64f3167\") " pod="openstack/dnsmasq-dns-746ff5d579-xk972" Mar 13 15:38:18 crc kubenswrapper[4907]: I0313 15:38:18.117793 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4cf4ac52-db94-4e0c-a990-c2c6c64f3167-dns-svc\") pod \"dnsmasq-dns-746ff5d579-xk972\" (UID: \"4cf4ac52-db94-4e0c-a990-c2c6c64f3167\") " pod="openstack/dnsmasq-dns-746ff5d579-xk972" Mar 13 15:38:18 crc kubenswrapper[4907]: I0313 15:38:18.131169 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4h66s\" (UniqueName: \"kubernetes.io/projected/4cf4ac52-db94-4e0c-a990-c2c6c64f3167-kube-api-access-4h66s\") pod \"dnsmasq-dns-746ff5d579-xk972\" (UID: \"4cf4ac52-db94-4e0c-a990-c2c6c64f3167\") " pod="openstack/dnsmasq-dns-746ff5d579-xk972" Mar 13 15:38:18 crc kubenswrapper[4907]: I0313 15:38:18.203654 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-746ff5d579-xk972" Mar 13 15:38:18 crc kubenswrapper[4907]: I0313 15:38:18.233623 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3431e634-aed2-44d7-a5a2-251956514c4b-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"3431e634-aed2-44d7-a5a2-251956514c4b\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:38:18 crc kubenswrapper[4907]: I0313 15:38:18.233700 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3431e634-aed2-44d7-a5a2-251956514c4b-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"3431e634-aed2-44d7-a5a2-251956514c4b\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:38:18 crc kubenswrapper[4907]: I0313 15:38:18.233776 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vkjx6\" (UniqueName: \"kubernetes.io/projected/3431e634-aed2-44d7-a5a2-251956514c4b-kube-api-access-vkjx6\") pod \"glance-default-internal-api-0\" (UID: \"3431e634-aed2-44d7-a5a2-251956514c4b\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:38:18 crc kubenswrapper[4907]: I0313 15:38:18.233811 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3431e634-aed2-44d7-a5a2-251956514c4b-scripts\") pod \"glance-default-internal-api-0\" (UID: \"3431e634-aed2-44d7-a5a2-251956514c4b\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:38:18 crc kubenswrapper[4907]: I0313 15:38:18.233868 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/3431e634-aed2-44d7-a5a2-251956514c4b-ceph\") pod \"glance-default-internal-api-0\" (UID: \"3431e634-aed2-44d7-a5a2-251956514c4b\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:38:18 crc kubenswrapper[4907]: I0313 15:38:18.233900 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3431e634-aed2-44d7-a5a2-251956514c4b-config-data\") pod \"glance-default-internal-api-0\" (UID: \"3431e634-aed2-44d7-a5a2-251956514c4b\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:38:18 crc kubenswrapper[4907]: I0313 15:38:18.233938 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3431e634-aed2-44d7-a5a2-251956514c4b-logs\") pod \"glance-default-internal-api-0\" (UID: \"3431e634-aed2-44d7-a5a2-251956514c4b\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:38:18 crc kubenswrapper[4907]: I0313 15:38:18.234467 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3431e634-aed2-44d7-a5a2-251956514c4b-logs\") pod \"glance-default-internal-api-0\" (UID: \"3431e634-aed2-44d7-a5a2-251956514c4b\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:38:18 crc kubenswrapper[4907]: I0313 15:38:18.234930 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3431e634-aed2-44d7-a5a2-251956514c4b-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"3431e634-aed2-44d7-a5a2-251956514c4b\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:38:18 crc kubenswrapper[4907]: I0313 15:38:18.240392 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3431e634-aed2-44d7-a5a2-251956514c4b-scripts\") pod \"glance-default-internal-api-0\" (UID: \"3431e634-aed2-44d7-a5a2-251956514c4b\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:38:18 crc kubenswrapper[4907]: I0313 15:38:18.242983 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/3431e634-aed2-44d7-a5a2-251956514c4b-ceph\") pod \"glance-default-internal-api-0\" (UID: \"3431e634-aed2-44d7-a5a2-251956514c4b\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:38:18 crc kubenswrapper[4907]: I0313 15:38:18.243081 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3431e634-aed2-44d7-a5a2-251956514c4b-config-data\") pod \"glance-default-internal-api-0\" (UID: \"3431e634-aed2-44d7-a5a2-251956514c4b\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:38:18 crc kubenswrapper[4907]: I0313 15:38:18.243104 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3431e634-aed2-44d7-a5a2-251956514c4b-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"3431e634-aed2-44d7-a5a2-251956514c4b\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:38:18 crc kubenswrapper[4907]: I0313 15:38:18.255680 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vkjx6\" (UniqueName: \"kubernetes.io/projected/3431e634-aed2-44d7-a5a2-251956514c4b-kube-api-access-vkjx6\") pod \"glance-default-internal-api-0\" (UID: \"3431e634-aed2-44d7-a5a2-251956514c4b\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:38:18 crc kubenswrapper[4907]: I0313 15:38:18.309842 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 13 15:38:18 crc kubenswrapper[4907]: I0313 15:38:18.451605 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 13 15:38:18 crc kubenswrapper[4907]: I0313 15:38:18.510240 4907 generic.go:334] "Generic (PLEG): container finished" podID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerID="90a39cf81ab6b6b036397a9c751a635e5f4cd834dccdc53b2ff02dd0144698a4" exitCode=0 Mar 13 15:38:18 crc kubenswrapper[4907]: I0313 15:38:18.510337 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerDied","Data":"90a39cf81ab6b6b036397a9c751a635e5f4cd834dccdc53b2ff02dd0144698a4"} Mar 13 15:38:18 crc kubenswrapper[4907]: I0313 15:38:18.510372 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerStarted","Data":"68ffe92dd03078633ec89d7ef29da9bbdc4c5f04c956a1e8487e3f2e01c5d1c5"} Mar 13 15:38:18 crc kubenswrapper[4907]: I0313 15:38:18.510393 4907 scope.go:117] "RemoveContainer" containerID="592bf9f175f440e943369aab2abbb13ad2c26412cb0e0738c4a33a156b8a5596" Mar 13 15:38:18 crc kubenswrapper[4907]: I0313 15:38:18.511752 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d38a5143-6b9b-4f3e-a41a-63668c7516ee","Type":"ContainerStarted","Data":"4fce01782f7e90bf4e850aaa9469f5e98dfe763d1a2f2b92bdd79b3911f5c894"} Mar 13 15:38:18 crc kubenswrapper[4907]: I0313 15:38:18.751363 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-746ff5d579-xk972"] Mar 13 15:38:18 crc kubenswrapper[4907]: I0313 15:38:18.819997 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 13 15:38:18 crc kubenswrapper[4907]: I0313 15:38:18.975486 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 13 15:38:19 crc kubenswrapper[4907]: I0313 15:38:19.534008 4907 generic.go:334] "Generic (PLEG): container finished" podID="4cf4ac52-db94-4e0c-a990-c2c6c64f3167" containerID="2d77589298c0e5cd48847476ea24e4e7c3f37cb59a1ce9b8eb1bd699e22f7f97" exitCode=0 Mar 13 15:38:19 crc kubenswrapper[4907]: I0313 15:38:19.534326 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-746ff5d579-xk972" event={"ID":"4cf4ac52-db94-4e0c-a990-c2c6c64f3167","Type":"ContainerDied","Data":"2d77589298c0e5cd48847476ea24e4e7c3f37cb59a1ce9b8eb1bd699e22f7f97"} Mar 13 15:38:19 crc kubenswrapper[4907]: I0313 15:38:19.534413 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-746ff5d579-xk972" event={"ID":"4cf4ac52-db94-4e0c-a990-c2c6c64f3167","Type":"ContainerStarted","Data":"ab14195333d30dbd417719105a07ed83adbe0a377eeede3562757ec9ea3f4db9"} Mar 13 15:38:19 crc kubenswrapper[4907]: I0313 15:38:19.540608 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"3431e634-aed2-44d7-a5a2-251956514c4b","Type":"ContainerStarted","Data":"d459823cf8e274f3ff30a8db80d9e6b1c4d9208bd24b1d7094389f4a3c7c74dd"} Mar 13 15:38:19 crc kubenswrapper[4907]: I0313 15:38:19.553949 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d38a5143-6b9b-4f3e-a41a-63668c7516ee","Type":"ContainerStarted","Data":"43c65f745fbc9ed6cd5669f25b8b192f01f2ec6518de198f3e6e582ee13ac564"} Mar 13 15:38:20 crc kubenswrapper[4907]: I0313 15:38:20.565067 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-746ff5d579-xk972" event={"ID":"4cf4ac52-db94-4e0c-a990-c2c6c64f3167","Type":"ContainerStarted","Data":"be908740b0d62c157a0b40660ddc6349bfde1dfa92e3ea3447a7f72e7e36459e"} Mar 13 15:38:20 crc kubenswrapper[4907]: I0313 15:38:20.566448 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-746ff5d579-xk972" Mar 13 15:38:20 crc kubenswrapper[4907]: I0313 15:38:20.567760 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"3431e634-aed2-44d7-a5a2-251956514c4b","Type":"ContainerStarted","Data":"052c7c2589da49f678cd5ba63c3a814c27749ed0800ee5b95fc97c6b92f33830"} Mar 13 15:38:20 crc kubenswrapper[4907]: I0313 15:38:20.567793 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"3431e634-aed2-44d7-a5a2-251956514c4b","Type":"ContainerStarted","Data":"ffa291e0576f78b774635c00fdaf08366bb2adfaf8a71181fa700bc295bbbdb6"} Mar 13 15:38:20 crc kubenswrapper[4907]: I0313 15:38:20.570587 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d38a5143-6b9b-4f3e-a41a-63668c7516ee","Type":"ContainerStarted","Data":"11786c2adfad844744530cf8121b7e49795de27cf3bb1b2d63adb0c8946c6ed3"} Mar 13 15:38:20 crc kubenswrapper[4907]: I0313 15:38:20.570729 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="d38a5143-6b9b-4f3e-a41a-63668c7516ee" containerName="glance-log" containerID="cri-o://43c65f745fbc9ed6cd5669f25b8b192f01f2ec6518de198f3e6e582ee13ac564" gracePeriod=30 Mar 13 15:38:20 crc kubenswrapper[4907]: I0313 15:38:20.571095 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="d38a5143-6b9b-4f3e-a41a-63668c7516ee" containerName="glance-httpd" containerID="cri-o://11786c2adfad844744530cf8121b7e49795de27cf3bb1b2d63adb0c8946c6ed3" gracePeriod=30 Mar 13 15:38:20 crc kubenswrapper[4907]: I0313 15:38:20.594302 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-746ff5d579-xk972" podStartSLOduration=3.5942807549999998 podStartE2EDuration="3.594280755s" podCreationTimestamp="2026-03-13 15:38:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:38:20.584397295 +0000 UTC m=+5599.484184984" watchObservedRunningTime="2026-03-13 15:38:20.594280755 +0000 UTC m=+5599.494068454" Mar 13 15:38:20 crc kubenswrapper[4907]: I0313 15:38:20.609380 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.609361427 podStartE2EDuration="3.609361427s" podCreationTimestamp="2026-03-13 15:38:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:38:20.605995294 +0000 UTC m=+5599.505782983" watchObservedRunningTime="2026-03-13 15:38:20.609361427 +0000 UTC m=+5599.509149116" Mar 13 15:38:20 crc kubenswrapper[4907]: I0313 15:38:20.624990 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.6249688129999997 podStartE2EDuration="3.624968813s" podCreationTimestamp="2026-03-13 15:38:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:38:20.624347296 +0000 UTC m=+5599.524134985" watchObservedRunningTime="2026-03-13 15:38:20.624968813 +0000 UTC m=+5599.524756502" Mar 13 15:38:20 crc kubenswrapper[4907]: I0313 15:38:20.787330 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.152211 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.285356 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-whgs7\" (UniqueName: \"kubernetes.io/projected/d38a5143-6b9b-4f3e-a41a-63668c7516ee-kube-api-access-whgs7\") pod \"d38a5143-6b9b-4f3e-a41a-63668c7516ee\" (UID: \"d38a5143-6b9b-4f3e-a41a-63668c7516ee\") " Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.285427 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d38a5143-6b9b-4f3e-a41a-63668c7516ee-config-data\") pod \"d38a5143-6b9b-4f3e-a41a-63668c7516ee\" (UID: \"d38a5143-6b9b-4f3e-a41a-63668c7516ee\") " Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.285471 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d38a5143-6b9b-4f3e-a41a-63668c7516ee-httpd-run\") pod \"d38a5143-6b9b-4f3e-a41a-63668c7516ee\" (UID: \"d38a5143-6b9b-4f3e-a41a-63668c7516ee\") " Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.285516 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/d38a5143-6b9b-4f3e-a41a-63668c7516ee-ceph\") pod \"d38a5143-6b9b-4f3e-a41a-63668c7516ee\" (UID: \"d38a5143-6b9b-4f3e-a41a-63668c7516ee\") " Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.285715 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d38a5143-6b9b-4f3e-a41a-63668c7516ee-logs\") pod \"d38a5143-6b9b-4f3e-a41a-63668c7516ee\" (UID: \"d38a5143-6b9b-4f3e-a41a-63668c7516ee\") " Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.285750 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d38a5143-6b9b-4f3e-a41a-63668c7516ee-scripts\") pod \"d38a5143-6b9b-4f3e-a41a-63668c7516ee\" (UID: \"d38a5143-6b9b-4f3e-a41a-63668c7516ee\") " Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.285801 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d38a5143-6b9b-4f3e-a41a-63668c7516ee-combined-ca-bundle\") pod \"d38a5143-6b9b-4f3e-a41a-63668c7516ee\" (UID: \"d38a5143-6b9b-4f3e-a41a-63668c7516ee\") " Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.286031 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d38a5143-6b9b-4f3e-a41a-63668c7516ee-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "d38a5143-6b9b-4f3e-a41a-63668c7516ee" (UID: "d38a5143-6b9b-4f3e-a41a-63668c7516ee"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.286089 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d38a5143-6b9b-4f3e-a41a-63668c7516ee-logs" (OuterVolumeSpecName: "logs") pod "d38a5143-6b9b-4f3e-a41a-63668c7516ee" (UID: "d38a5143-6b9b-4f3e-a41a-63668c7516ee"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.286632 4907 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d38a5143-6b9b-4f3e-a41a-63668c7516ee-httpd-run\") on node \"crc\" DevicePath \"\"" Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.286662 4907 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d38a5143-6b9b-4f3e-a41a-63668c7516ee-logs\") on node \"crc\" DevicePath \"\"" Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.291936 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d38a5143-6b9b-4f3e-a41a-63668c7516ee-kube-api-access-whgs7" (OuterVolumeSpecName: "kube-api-access-whgs7") pod "d38a5143-6b9b-4f3e-a41a-63668c7516ee" (UID: "d38a5143-6b9b-4f3e-a41a-63668c7516ee"). InnerVolumeSpecName "kube-api-access-whgs7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.302042 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d38a5143-6b9b-4f3e-a41a-63668c7516ee-scripts" (OuterVolumeSpecName: "scripts") pod "d38a5143-6b9b-4f3e-a41a-63668c7516ee" (UID: "d38a5143-6b9b-4f3e-a41a-63668c7516ee"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.305149 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d38a5143-6b9b-4f3e-a41a-63668c7516ee-ceph" (OuterVolumeSpecName: "ceph") pod "d38a5143-6b9b-4f3e-a41a-63668c7516ee" (UID: "d38a5143-6b9b-4f3e-a41a-63668c7516ee"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.315575 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d38a5143-6b9b-4f3e-a41a-63668c7516ee-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d38a5143-6b9b-4f3e-a41a-63668c7516ee" (UID: "d38a5143-6b9b-4f3e-a41a-63668c7516ee"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.331821 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d38a5143-6b9b-4f3e-a41a-63668c7516ee-config-data" (OuterVolumeSpecName: "config-data") pod "d38a5143-6b9b-4f3e-a41a-63668c7516ee" (UID: "d38a5143-6b9b-4f3e-a41a-63668c7516ee"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.389025 4907 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/d38a5143-6b9b-4f3e-a41a-63668c7516ee-ceph\") on node \"crc\" DevicePath \"\"" Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.389067 4907 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d38a5143-6b9b-4f3e-a41a-63668c7516ee-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.389078 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d38a5143-6b9b-4f3e-a41a-63668c7516ee-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.389100 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-whgs7\" (UniqueName: \"kubernetes.io/projected/d38a5143-6b9b-4f3e-a41a-63668c7516ee-kube-api-access-whgs7\") on node \"crc\" DevicePath \"\"" Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.389119 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d38a5143-6b9b-4f3e-a41a-63668c7516ee-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.578249 4907 generic.go:334] "Generic (PLEG): container finished" podID="d38a5143-6b9b-4f3e-a41a-63668c7516ee" containerID="11786c2adfad844744530cf8121b7e49795de27cf3bb1b2d63adb0c8946c6ed3" exitCode=0 Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.578290 4907 generic.go:334] "Generic (PLEG): container finished" podID="d38a5143-6b9b-4f3e-a41a-63668c7516ee" containerID="43c65f745fbc9ed6cd5669f25b8b192f01f2ec6518de198f3e6e582ee13ac564" exitCode=143 Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.578301 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.578396 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d38a5143-6b9b-4f3e-a41a-63668c7516ee","Type":"ContainerDied","Data":"11786c2adfad844744530cf8121b7e49795de27cf3bb1b2d63adb0c8946c6ed3"} Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.578428 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d38a5143-6b9b-4f3e-a41a-63668c7516ee","Type":"ContainerDied","Data":"43c65f745fbc9ed6cd5669f25b8b192f01f2ec6518de198f3e6e582ee13ac564"} Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.578457 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d38a5143-6b9b-4f3e-a41a-63668c7516ee","Type":"ContainerDied","Data":"4fce01782f7e90bf4e850aaa9469f5e98dfe763d1a2f2b92bdd79b3911f5c894"} Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.578474 4907 scope.go:117] "RemoveContainer" containerID="11786c2adfad844744530cf8121b7e49795de27cf3bb1b2d63adb0c8946c6ed3" Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.610778 4907 scope.go:117] "RemoveContainer" containerID="43c65f745fbc9ed6cd5669f25b8b192f01f2ec6518de198f3e6e582ee13ac564" Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.618007 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.625130 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.631718 4907 scope.go:117] "RemoveContainer" containerID="11786c2adfad844744530cf8121b7e49795de27cf3bb1b2d63adb0c8946c6ed3" Mar 13 15:38:21 crc kubenswrapper[4907]: E0313 15:38:21.632509 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"11786c2adfad844744530cf8121b7e49795de27cf3bb1b2d63adb0c8946c6ed3\": container with ID starting with 11786c2adfad844744530cf8121b7e49795de27cf3bb1b2d63adb0c8946c6ed3 not found: ID does not exist" containerID="11786c2adfad844744530cf8121b7e49795de27cf3bb1b2d63adb0c8946c6ed3" Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.632550 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"11786c2adfad844744530cf8121b7e49795de27cf3bb1b2d63adb0c8946c6ed3"} err="failed to get container status \"11786c2adfad844744530cf8121b7e49795de27cf3bb1b2d63adb0c8946c6ed3\": rpc error: code = NotFound desc = could not find container \"11786c2adfad844744530cf8121b7e49795de27cf3bb1b2d63adb0c8946c6ed3\": container with ID starting with 11786c2adfad844744530cf8121b7e49795de27cf3bb1b2d63adb0c8946c6ed3 not found: ID does not exist" Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.632579 4907 scope.go:117] "RemoveContainer" containerID="43c65f745fbc9ed6cd5669f25b8b192f01f2ec6518de198f3e6e582ee13ac564" Mar 13 15:38:21 crc kubenswrapper[4907]: E0313 15:38:21.632870 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"43c65f745fbc9ed6cd5669f25b8b192f01f2ec6518de198f3e6e582ee13ac564\": container with ID starting with 43c65f745fbc9ed6cd5669f25b8b192f01f2ec6518de198f3e6e582ee13ac564 not found: ID does not exist" containerID="43c65f745fbc9ed6cd5669f25b8b192f01f2ec6518de198f3e6e582ee13ac564" Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.632936 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"43c65f745fbc9ed6cd5669f25b8b192f01f2ec6518de198f3e6e582ee13ac564"} err="failed to get container status \"43c65f745fbc9ed6cd5669f25b8b192f01f2ec6518de198f3e6e582ee13ac564\": rpc error: code = NotFound desc = could not find container \"43c65f745fbc9ed6cd5669f25b8b192f01f2ec6518de198f3e6e582ee13ac564\": container with ID starting with 43c65f745fbc9ed6cd5669f25b8b192f01f2ec6518de198f3e6e582ee13ac564 not found: ID does not exist" Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.632951 4907 scope.go:117] "RemoveContainer" containerID="11786c2adfad844744530cf8121b7e49795de27cf3bb1b2d63adb0c8946c6ed3" Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.633377 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"11786c2adfad844744530cf8121b7e49795de27cf3bb1b2d63adb0c8946c6ed3"} err="failed to get container status \"11786c2adfad844744530cf8121b7e49795de27cf3bb1b2d63adb0c8946c6ed3\": rpc error: code = NotFound desc = could not find container \"11786c2adfad844744530cf8121b7e49795de27cf3bb1b2d63adb0c8946c6ed3\": container with ID starting with 11786c2adfad844744530cf8121b7e49795de27cf3bb1b2d63adb0c8946c6ed3 not found: ID does not exist" Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.633397 4907 scope.go:117] "RemoveContainer" containerID="43c65f745fbc9ed6cd5669f25b8b192f01f2ec6518de198f3e6e582ee13ac564" Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.633646 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"43c65f745fbc9ed6cd5669f25b8b192f01f2ec6518de198f3e6e582ee13ac564"} err="failed to get container status \"43c65f745fbc9ed6cd5669f25b8b192f01f2ec6518de198f3e6e582ee13ac564\": rpc error: code = NotFound desc = could not find container \"43c65f745fbc9ed6cd5669f25b8b192f01f2ec6518de198f3e6e582ee13ac564\": container with ID starting with 43c65f745fbc9ed6cd5669f25b8b192f01f2ec6518de198f3e6e582ee13ac564 not found: ID does not exist" Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.640441 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Mar 13 15:38:21 crc kubenswrapper[4907]: E0313 15:38:21.640785 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d38a5143-6b9b-4f3e-a41a-63668c7516ee" containerName="glance-log" Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.640800 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="d38a5143-6b9b-4f3e-a41a-63668c7516ee" containerName="glance-log" Mar 13 15:38:21 crc kubenswrapper[4907]: E0313 15:38:21.640819 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d38a5143-6b9b-4f3e-a41a-63668c7516ee" containerName="glance-httpd" Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.640828 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="d38a5143-6b9b-4f3e-a41a-63668c7516ee" containerName="glance-httpd" Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.641016 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="d38a5143-6b9b-4f3e-a41a-63668c7516ee" containerName="glance-log" Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.641032 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="d38a5143-6b9b-4f3e-a41a-63668c7516ee" containerName="glance-httpd" Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.641875 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.643420 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.676856 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.798513 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4srpn\" (UniqueName: \"kubernetes.io/projected/86794caf-bf6d-4394-bc1c-3c516bdbada8-kube-api-access-4srpn\") pod \"glance-default-external-api-0\" (UID: \"86794caf-bf6d-4394-bc1c-3c516bdbada8\") " pod="openstack/glance-default-external-api-0" Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.798581 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/86794caf-bf6d-4394-bc1c-3c516bdbada8-scripts\") pod \"glance-default-external-api-0\" (UID: \"86794caf-bf6d-4394-bc1c-3c516bdbada8\") " pod="openstack/glance-default-external-api-0" Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.798621 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/86794caf-bf6d-4394-bc1c-3c516bdbada8-logs\") pod \"glance-default-external-api-0\" (UID: \"86794caf-bf6d-4394-bc1c-3c516bdbada8\") " pod="openstack/glance-default-external-api-0" Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.798645 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86794caf-bf6d-4394-bc1c-3c516bdbada8-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"86794caf-bf6d-4394-bc1c-3c516bdbada8\") " pod="openstack/glance-default-external-api-0" Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.798710 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/86794caf-bf6d-4394-bc1c-3c516bdbada8-config-data\") pod \"glance-default-external-api-0\" (UID: \"86794caf-bf6d-4394-bc1c-3c516bdbada8\") " pod="openstack/glance-default-external-api-0" Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.798742 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/86794caf-bf6d-4394-bc1c-3c516bdbada8-ceph\") pod \"glance-default-external-api-0\" (UID: \"86794caf-bf6d-4394-bc1c-3c516bdbada8\") " pod="openstack/glance-default-external-api-0" Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.798774 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/86794caf-bf6d-4394-bc1c-3c516bdbada8-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"86794caf-bf6d-4394-bc1c-3c516bdbada8\") " pod="openstack/glance-default-external-api-0" Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.806760 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d38a5143-6b9b-4f3e-a41a-63668c7516ee" path="/var/lib/kubelet/pods/d38a5143-6b9b-4f3e-a41a-63668c7516ee/volumes" Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.900220 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/86794caf-bf6d-4394-bc1c-3c516bdbada8-scripts\") pod \"glance-default-external-api-0\" (UID: \"86794caf-bf6d-4394-bc1c-3c516bdbada8\") " pod="openstack/glance-default-external-api-0" Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.900299 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/86794caf-bf6d-4394-bc1c-3c516bdbada8-logs\") pod \"glance-default-external-api-0\" (UID: \"86794caf-bf6d-4394-bc1c-3c516bdbada8\") " pod="openstack/glance-default-external-api-0" Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.900335 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86794caf-bf6d-4394-bc1c-3c516bdbada8-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"86794caf-bf6d-4394-bc1c-3c516bdbada8\") " pod="openstack/glance-default-external-api-0" Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.900477 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/86794caf-bf6d-4394-bc1c-3c516bdbada8-config-data\") pod \"glance-default-external-api-0\" (UID: \"86794caf-bf6d-4394-bc1c-3c516bdbada8\") " pod="openstack/glance-default-external-api-0" Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.900516 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/86794caf-bf6d-4394-bc1c-3c516bdbada8-ceph\") pod \"glance-default-external-api-0\" (UID: \"86794caf-bf6d-4394-bc1c-3c516bdbada8\") " pod="openstack/glance-default-external-api-0" Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.900544 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/86794caf-bf6d-4394-bc1c-3c516bdbada8-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"86794caf-bf6d-4394-bc1c-3c516bdbada8\") " pod="openstack/glance-default-external-api-0" Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.900613 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4srpn\" (UniqueName: \"kubernetes.io/projected/86794caf-bf6d-4394-bc1c-3c516bdbada8-kube-api-access-4srpn\") pod \"glance-default-external-api-0\" (UID: \"86794caf-bf6d-4394-bc1c-3c516bdbada8\") " pod="openstack/glance-default-external-api-0" Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.900848 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/86794caf-bf6d-4394-bc1c-3c516bdbada8-logs\") pod \"glance-default-external-api-0\" (UID: \"86794caf-bf6d-4394-bc1c-3c516bdbada8\") " pod="openstack/glance-default-external-api-0" Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.902310 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/86794caf-bf6d-4394-bc1c-3c516bdbada8-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"86794caf-bf6d-4394-bc1c-3c516bdbada8\") " pod="openstack/glance-default-external-api-0" Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.920180 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/86794caf-bf6d-4394-bc1c-3c516bdbada8-ceph\") pod \"glance-default-external-api-0\" (UID: \"86794caf-bf6d-4394-bc1c-3c516bdbada8\") " pod="openstack/glance-default-external-api-0" Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.921450 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86794caf-bf6d-4394-bc1c-3c516bdbada8-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"86794caf-bf6d-4394-bc1c-3c516bdbada8\") " pod="openstack/glance-default-external-api-0" Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.922180 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/86794caf-bf6d-4394-bc1c-3c516bdbada8-config-data\") pod \"glance-default-external-api-0\" (UID: \"86794caf-bf6d-4394-bc1c-3c516bdbada8\") " pod="openstack/glance-default-external-api-0" Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.923420 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/86794caf-bf6d-4394-bc1c-3c516bdbada8-scripts\") pod \"glance-default-external-api-0\" (UID: \"86794caf-bf6d-4394-bc1c-3c516bdbada8\") " pod="openstack/glance-default-external-api-0" Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.924733 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4srpn\" (UniqueName: \"kubernetes.io/projected/86794caf-bf6d-4394-bc1c-3c516bdbada8-kube-api-access-4srpn\") pod \"glance-default-external-api-0\" (UID: \"86794caf-bf6d-4394-bc1c-3c516bdbada8\") " pod="openstack/glance-default-external-api-0" Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.966245 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 13 15:38:22 crc kubenswrapper[4907]: I0313 15:38:22.445748 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 13 15:38:22 crc kubenswrapper[4907]: I0313 15:38:22.589282 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"86794caf-bf6d-4394-bc1c-3c516bdbada8","Type":"ContainerStarted","Data":"e6d5e941f6575995d025c2bf565c8da48222d14bbb2fbb1710f7cd79e1394fae"} Mar 13 15:38:22 crc kubenswrapper[4907]: I0313 15:38:22.591294 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="3431e634-aed2-44d7-a5a2-251956514c4b" containerName="glance-log" containerID="cri-o://ffa291e0576f78b774635c00fdaf08366bb2adfaf8a71181fa700bc295bbbdb6" gracePeriod=30 Mar 13 15:38:22 crc kubenswrapper[4907]: I0313 15:38:22.591406 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="3431e634-aed2-44d7-a5a2-251956514c4b" containerName="glance-httpd" containerID="cri-o://052c7c2589da49f678cd5ba63c3a814c27749ed0800ee5b95fc97c6b92f33830" gracePeriod=30 Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.293349 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.438611 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3431e634-aed2-44d7-a5a2-251956514c4b-config-data\") pod \"3431e634-aed2-44d7-a5a2-251956514c4b\" (UID: \"3431e634-aed2-44d7-a5a2-251956514c4b\") " Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.438674 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3431e634-aed2-44d7-a5a2-251956514c4b-combined-ca-bundle\") pod \"3431e634-aed2-44d7-a5a2-251956514c4b\" (UID: \"3431e634-aed2-44d7-a5a2-251956514c4b\") " Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.438724 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3431e634-aed2-44d7-a5a2-251956514c4b-httpd-run\") pod \"3431e634-aed2-44d7-a5a2-251956514c4b\" (UID: \"3431e634-aed2-44d7-a5a2-251956514c4b\") " Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.438755 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vkjx6\" (UniqueName: \"kubernetes.io/projected/3431e634-aed2-44d7-a5a2-251956514c4b-kube-api-access-vkjx6\") pod \"3431e634-aed2-44d7-a5a2-251956514c4b\" (UID: \"3431e634-aed2-44d7-a5a2-251956514c4b\") " Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.438867 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3431e634-aed2-44d7-a5a2-251956514c4b-scripts\") pod \"3431e634-aed2-44d7-a5a2-251956514c4b\" (UID: \"3431e634-aed2-44d7-a5a2-251956514c4b\") " Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.438951 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3431e634-aed2-44d7-a5a2-251956514c4b-logs\") pod \"3431e634-aed2-44d7-a5a2-251956514c4b\" (UID: \"3431e634-aed2-44d7-a5a2-251956514c4b\") " Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.439022 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/3431e634-aed2-44d7-a5a2-251956514c4b-ceph\") pod \"3431e634-aed2-44d7-a5a2-251956514c4b\" (UID: \"3431e634-aed2-44d7-a5a2-251956514c4b\") " Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.441248 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3431e634-aed2-44d7-a5a2-251956514c4b-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "3431e634-aed2-44d7-a5a2-251956514c4b" (UID: "3431e634-aed2-44d7-a5a2-251956514c4b"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.457188 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3431e634-aed2-44d7-a5a2-251956514c4b-ceph" (OuterVolumeSpecName: "ceph") pod "3431e634-aed2-44d7-a5a2-251956514c4b" (UID: "3431e634-aed2-44d7-a5a2-251956514c4b"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.458156 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3431e634-aed2-44d7-a5a2-251956514c4b-scripts" (OuterVolumeSpecName: "scripts") pod "3431e634-aed2-44d7-a5a2-251956514c4b" (UID: "3431e634-aed2-44d7-a5a2-251956514c4b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.459075 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3431e634-aed2-44d7-a5a2-251956514c4b-kube-api-access-vkjx6" (OuterVolumeSpecName: "kube-api-access-vkjx6") pod "3431e634-aed2-44d7-a5a2-251956514c4b" (UID: "3431e634-aed2-44d7-a5a2-251956514c4b"). InnerVolumeSpecName "kube-api-access-vkjx6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.461122 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3431e634-aed2-44d7-a5a2-251956514c4b-logs" (OuterVolumeSpecName: "logs") pod "3431e634-aed2-44d7-a5a2-251956514c4b" (UID: "3431e634-aed2-44d7-a5a2-251956514c4b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.501164 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3431e634-aed2-44d7-a5a2-251956514c4b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3431e634-aed2-44d7-a5a2-251956514c4b" (UID: "3431e634-aed2-44d7-a5a2-251956514c4b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.541064 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3431e634-aed2-44d7-a5a2-251956514c4b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.541294 4907 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3431e634-aed2-44d7-a5a2-251956514c4b-httpd-run\") on node \"crc\" DevicePath \"\"" Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.541389 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vkjx6\" (UniqueName: \"kubernetes.io/projected/3431e634-aed2-44d7-a5a2-251956514c4b-kube-api-access-vkjx6\") on node \"crc\" DevicePath \"\"" Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.541461 4907 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3431e634-aed2-44d7-a5a2-251956514c4b-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.541552 4907 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3431e634-aed2-44d7-a5a2-251956514c4b-logs\") on node \"crc\" DevicePath \"\"" Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.541621 4907 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/3431e634-aed2-44d7-a5a2-251956514c4b-ceph\") on node \"crc\" DevicePath \"\"" Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.575046 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3431e634-aed2-44d7-a5a2-251956514c4b-config-data" (OuterVolumeSpecName: "config-data") pod "3431e634-aed2-44d7-a5a2-251956514c4b" (UID: "3431e634-aed2-44d7-a5a2-251956514c4b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.601240 4907 generic.go:334] "Generic (PLEG): container finished" podID="3431e634-aed2-44d7-a5a2-251956514c4b" containerID="052c7c2589da49f678cd5ba63c3a814c27749ed0800ee5b95fc97c6b92f33830" exitCode=0 Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.601272 4907 generic.go:334] "Generic (PLEG): container finished" podID="3431e634-aed2-44d7-a5a2-251956514c4b" containerID="ffa291e0576f78b774635c00fdaf08366bb2adfaf8a71181fa700bc295bbbdb6" exitCode=143 Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.601309 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"3431e634-aed2-44d7-a5a2-251956514c4b","Type":"ContainerDied","Data":"052c7c2589da49f678cd5ba63c3a814c27749ed0800ee5b95fc97c6b92f33830"} Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.601334 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"3431e634-aed2-44d7-a5a2-251956514c4b","Type":"ContainerDied","Data":"ffa291e0576f78b774635c00fdaf08366bb2adfaf8a71181fa700bc295bbbdb6"} Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.601345 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"3431e634-aed2-44d7-a5a2-251956514c4b","Type":"ContainerDied","Data":"d459823cf8e274f3ff30a8db80d9e6b1c4d9208bd24b1d7094389f4a3c7c74dd"} Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.601365 4907 scope.go:117] "RemoveContainer" containerID="052c7c2589da49f678cd5ba63c3a814c27749ed0800ee5b95fc97c6b92f33830" Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.601513 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.605377 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"86794caf-bf6d-4394-bc1c-3c516bdbada8","Type":"ContainerStarted","Data":"56ff9c5433a9cbee0386dea3d88b4e88f870fa5dcf67d5221e12245f174e8f85"} Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.637254 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.643177 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3431e634-aed2-44d7-a5a2-251956514c4b-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.644185 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.654709 4907 scope.go:117] "RemoveContainer" containerID="ffa291e0576f78b774635c00fdaf08366bb2adfaf8a71181fa700bc295bbbdb6" Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.658382 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 13 15:38:23 crc kubenswrapper[4907]: E0313 15:38:23.659137 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3431e634-aed2-44d7-a5a2-251956514c4b" containerName="glance-httpd" Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.659160 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="3431e634-aed2-44d7-a5a2-251956514c4b" containerName="glance-httpd" Mar 13 15:38:23 crc kubenswrapper[4907]: E0313 15:38:23.659174 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3431e634-aed2-44d7-a5a2-251956514c4b" containerName="glance-log" Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.659181 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="3431e634-aed2-44d7-a5a2-251956514c4b" containerName="glance-log" Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.659325 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="3431e634-aed2-44d7-a5a2-251956514c4b" containerName="glance-httpd" Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.659342 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="3431e634-aed2-44d7-a5a2-251956514c4b" containerName="glance-log" Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.660368 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.666848 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.675844 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.705059 4907 scope.go:117] "RemoveContainer" containerID="052c7c2589da49f678cd5ba63c3a814c27749ed0800ee5b95fc97c6b92f33830" Mar 13 15:38:23 crc kubenswrapper[4907]: E0313 15:38:23.706410 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"052c7c2589da49f678cd5ba63c3a814c27749ed0800ee5b95fc97c6b92f33830\": container with ID starting with 052c7c2589da49f678cd5ba63c3a814c27749ed0800ee5b95fc97c6b92f33830 not found: ID does not exist" containerID="052c7c2589da49f678cd5ba63c3a814c27749ed0800ee5b95fc97c6b92f33830" Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.706448 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"052c7c2589da49f678cd5ba63c3a814c27749ed0800ee5b95fc97c6b92f33830"} err="failed to get container status \"052c7c2589da49f678cd5ba63c3a814c27749ed0800ee5b95fc97c6b92f33830\": rpc error: code = NotFound desc = could not find container \"052c7c2589da49f678cd5ba63c3a814c27749ed0800ee5b95fc97c6b92f33830\": container with ID starting with 052c7c2589da49f678cd5ba63c3a814c27749ed0800ee5b95fc97c6b92f33830 not found: ID does not exist" Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.706473 4907 scope.go:117] "RemoveContainer" containerID="ffa291e0576f78b774635c00fdaf08366bb2adfaf8a71181fa700bc295bbbdb6" Mar 13 15:38:23 crc kubenswrapper[4907]: E0313 15:38:23.706866 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ffa291e0576f78b774635c00fdaf08366bb2adfaf8a71181fa700bc295bbbdb6\": container with ID starting with ffa291e0576f78b774635c00fdaf08366bb2adfaf8a71181fa700bc295bbbdb6 not found: ID does not exist" containerID="ffa291e0576f78b774635c00fdaf08366bb2adfaf8a71181fa700bc295bbbdb6" Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.706902 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ffa291e0576f78b774635c00fdaf08366bb2adfaf8a71181fa700bc295bbbdb6"} err="failed to get container status \"ffa291e0576f78b774635c00fdaf08366bb2adfaf8a71181fa700bc295bbbdb6\": rpc error: code = NotFound desc = could not find container \"ffa291e0576f78b774635c00fdaf08366bb2adfaf8a71181fa700bc295bbbdb6\": container with ID starting with ffa291e0576f78b774635c00fdaf08366bb2adfaf8a71181fa700bc295bbbdb6 not found: ID does not exist" Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.706917 4907 scope.go:117] "RemoveContainer" containerID="052c7c2589da49f678cd5ba63c3a814c27749ed0800ee5b95fc97c6b92f33830" Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.710225 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"052c7c2589da49f678cd5ba63c3a814c27749ed0800ee5b95fc97c6b92f33830"} err="failed to get container status \"052c7c2589da49f678cd5ba63c3a814c27749ed0800ee5b95fc97c6b92f33830\": rpc error: code = NotFound desc = could not find container \"052c7c2589da49f678cd5ba63c3a814c27749ed0800ee5b95fc97c6b92f33830\": container with ID starting with 052c7c2589da49f678cd5ba63c3a814c27749ed0800ee5b95fc97c6b92f33830 not found: ID does not exist" Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.710264 4907 scope.go:117] "RemoveContainer" containerID="ffa291e0576f78b774635c00fdaf08366bb2adfaf8a71181fa700bc295bbbdb6" Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.711172 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ffa291e0576f78b774635c00fdaf08366bb2adfaf8a71181fa700bc295bbbdb6"} err="failed to get container status \"ffa291e0576f78b774635c00fdaf08366bb2adfaf8a71181fa700bc295bbbdb6\": rpc error: code = NotFound desc = could not find container \"ffa291e0576f78b774635c00fdaf08366bb2adfaf8a71181fa700bc295bbbdb6\": container with ID starting with ffa291e0576f78b774635c00fdaf08366bb2adfaf8a71181fa700bc295bbbdb6 not found: ID does not exist" Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.798206 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3431e634-aed2-44d7-a5a2-251956514c4b" path="/var/lib/kubelet/pods/3431e634-aed2-44d7-a5a2-251956514c4b/volumes" Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.846585 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/096d3713-ea4f-43d3-bfb9-9170e8958ed0-config-data\") pod \"glance-default-internal-api-0\" (UID: \"096d3713-ea4f-43d3-bfb9-9170e8958ed0\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.846911 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/096d3713-ea4f-43d3-bfb9-9170e8958ed0-logs\") pod \"glance-default-internal-api-0\" (UID: \"096d3713-ea4f-43d3-bfb9-9170e8958ed0\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.846933 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/096d3713-ea4f-43d3-bfb9-9170e8958ed0-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"096d3713-ea4f-43d3-bfb9-9170e8958ed0\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.846973 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h7zl5\" (UniqueName: \"kubernetes.io/projected/096d3713-ea4f-43d3-bfb9-9170e8958ed0-kube-api-access-h7zl5\") pod \"glance-default-internal-api-0\" (UID: \"096d3713-ea4f-43d3-bfb9-9170e8958ed0\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.847171 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/096d3713-ea4f-43d3-bfb9-9170e8958ed0-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"096d3713-ea4f-43d3-bfb9-9170e8958ed0\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.847482 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/096d3713-ea4f-43d3-bfb9-9170e8958ed0-scripts\") pod \"glance-default-internal-api-0\" (UID: \"096d3713-ea4f-43d3-bfb9-9170e8958ed0\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.847552 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/096d3713-ea4f-43d3-bfb9-9170e8958ed0-ceph\") pod \"glance-default-internal-api-0\" (UID: \"096d3713-ea4f-43d3-bfb9-9170e8958ed0\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.948984 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/096d3713-ea4f-43d3-bfb9-9170e8958ed0-scripts\") pod \"glance-default-internal-api-0\" (UID: \"096d3713-ea4f-43d3-bfb9-9170e8958ed0\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.949046 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/096d3713-ea4f-43d3-bfb9-9170e8958ed0-ceph\") pod \"glance-default-internal-api-0\" (UID: \"096d3713-ea4f-43d3-bfb9-9170e8958ed0\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.949126 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/096d3713-ea4f-43d3-bfb9-9170e8958ed0-config-data\") pod \"glance-default-internal-api-0\" (UID: \"096d3713-ea4f-43d3-bfb9-9170e8958ed0\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.949158 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/096d3713-ea4f-43d3-bfb9-9170e8958ed0-logs\") pod \"glance-default-internal-api-0\" (UID: \"096d3713-ea4f-43d3-bfb9-9170e8958ed0\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.949185 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/096d3713-ea4f-43d3-bfb9-9170e8958ed0-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"096d3713-ea4f-43d3-bfb9-9170e8958ed0\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.949222 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h7zl5\" (UniqueName: \"kubernetes.io/projected/096d3713-ea4f-43d3-bfb9-9170e8958ed0-kube-api-access-h7zl5\") pod \"glance-default-internal-api-0\" (UID: \"096d3713-ea4f-43d3-bfb9-9170e8958ed0\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.949263 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/096d3713-ea4f-43d3-bfb9-9170e8958ed0-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"096d3713-ea4f-43d3-bfb9-9170e8958ed0\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.950007 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/096d3713-ea4f-43d3-bfb9-9170e8958ed0-logs\") pod \"glance-default-internal-api-0\" (UID: \"096d3713-ea4f-43d3-bfb9-9170e8958ed0\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.950010 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/096d3713-ea4f-43d3-bfb9-9170e8958ed0-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"096d3713-ea4f-43d3-bfb9-9170e8958ed0\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.954421 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/096d3713-ea4f-43d3-bfb9-9170e8958ed0-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"096d3713-ea4f-43d3-bfb9-9170e8958ed0\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.954633 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/096d3713-ea4f-43d3-bfb9-9170e8958ed0-scripts\") pod \"glance-default-internal-api-0\" (UID: \"096d3713-ea4f-43d3-bfb9-9170e8958ed0\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.955064 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/096d3713-ea4f-43d3-bfb9-9170e8958ed0-ceph\") pod \"glance-default-internal-api-0\" (UID: \"096d3713-ea4f-43d3-bfb9-9170e8958ed0\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.955648 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/096d3713-ea4f-43d3-bfb9-9170e8958ed0-config-data\") pod \"glance-default-internal-api-0\" (UID: \"096d3713-ea4f-43d3-bfb9-9170e8958ed0\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.967761 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h7zl5\" (UniqueName: \"kubernetes.io/projected/096d3713-ea4f-43d3-bfb9-9170e8958ed0-kube-api-access-h7zl5\") pod \"glance-default-internal-api-0\" (UID: \"096d3713-ea4f-43d3-bfb9-9170e8958ed0\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.986045 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 13 15:38:24 crc kubenswrapper[4907]: I0313 15:38:24.511495 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 13 15:38:24 crc kubenswrapper[4907]: W0313 15:38:24.515920 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod096d3713_ea4f_43d3_bfb9_9170e8958ed0.slice/crio-13bb6e3535193b961d70131bf3aebbe0d658a8a82ddb3d84795b101e8ae9b66a WatchSource:0}: Error finding container 13bb6e3535193b961d70131bf3aebbe0d658a8a82ddb3d84795b101e8ae9b66a: Status 404 returned error can't find the container with id 13bb6e3535193b961d70131bf3aebbe0d658a8a82ddb3d84795b101e8ae9b66a Mar 13 15:38:24 crc kubenswrapper[4907]: I0313 15:38:24.617394 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"86794caf-bf6d-4394-bc1c-3c516bdbada8","Type":"ContainerStarted","Data":"d5e6285fe8e2b22da125d5022a7924d7086ba68515ab90ca41e35d0607040a03"} Mar 13 15:38:24 crc kubenswrapper[4907]: I0313 15:38:24.619746 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"096d3713-ea4f-43d3-bfb9-9170e8958ed0","Type":"ContainerStarted","Data":"13bb6e3535193b961d70131bf3aebbe0d658a8a82ddb3d84795b101e8ae9b66a"} Mar 13 15:38:25 crc kubenswrapper[4907]: I0313 15:38:25.628978 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"096d3713-ea4f-43d3-bfb9-9170e8958ed0","Type":"ContainerStarted","Data":"27209680abb804b6c48ccdaa33b44ab4f54951d1df1401058af333f852827d0f"} Mar 13 15:38:25 crc kubenswrapper[4907]: I0313 15:38:25.629509 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"096d3713-ea4f-43d3-bfb9-9170e8958ed0","Type":"ContainerStarted","Data":"02ef5ab885e1e984508608add8cbc5f616e0cf1ed8d42cf7ded2960a6fa681f8"} Mar 13 15:38:25 crc kubenswrapper[4907]: I0313 15:38:25.653168 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=2.653147532 podStartE2EDuration="2.653147532s" podCreationTimestamp="2026-03-13 15:38:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:38:25.645347748 +0000 UTC m=+5604.545135457" watchObservedRunningTime="2026-03-13 15:38:25.653147532 +0000 UTC m=+5604.552935221" Mar 13 15:38:25 crc kubenswrapper[4907]: I0313 15:38:25.654029 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=4.654019426 podStartE2EDuration="4.654019426s" podCreationTimestamp="2026-03-13 15:38:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:38:24.648193744 +0000 UTC m=+5603.547981433" watchObservedRunningTime="2026-03-13 15:38:25.654019426 +0000 UTC m=+5604.553807125" Mar 13 15:38:28 crc kubenswrapper[4907]: I0313 15:38:28.205018 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-746ff5d579-xk972" Mar 13 15:38:28 crc kubenswrapper[4907]: I0313 15:38:28.224962 4907 scope.go:117] "RemoveContainer" containerID="44715012ac71de6a61b777f99c8ba8523d3ece17a5f46795a3605b96d66aee6e" Mar 13 15:38:28 crc kubenswrapper[4907]: I0313 15:38:28.275829 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-db9bd6759-f9ndc"] Mar 13 15:38:28 crc kubenswrapper[4907]: I0313 15:38:28.276075 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-db9bd6759-f9ndc" podUID="b43d065a-571d-416f-96d3-f1713f3d3bdb" containerName="dnsmasq-dns" containerID="cri-o://181ece190f085ef12291d2abeb4a964937fb06d189ff825afc0a6770f3e932ba" gracePeriod=10 Mar 13 15:38:28 crc kubenswrapper[4907]: I0313 15:38:28.656991 4907 generic.go:334] "Generic (PLEG): container finished" podID="b43d065a-571d-416f-96d3-f1713f3d3bdb" containerID="181ece190f085ef12291d2abeb4a964937fb06d189ff825afc0a6770f3e932ba" exitCode=0 Mar 13 15:38:28 crc kubenswrapper[4907]: I0313 15:38:28.657085 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-db9bd6759-f9ndc" event={"ID":"b43d065a-571d-416f-96d3-f1713f3d3bdb","Type":"ContainerDied","Data":"181ece190f085ef12291d2abeb4a964937fb06d189ff825afc0a6770f3e932ba"} Mar 13 15:38:28 crc kubenswrapper[4907]: I0313 15:38:28.760139 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-db9bd6759-f9ndc" Mar 13 15:38:28 crc kubenswrapper[4907]: I0313 15:38:28.935514 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b43d065a-571d-416f-96d3-f1713f3d3bdb-ovsdbserver-nb\") pod \"b43d065a-571d-416f-96d3-f1713f3d3bdb\" (UID: \"b43d065a-571d-416f-96d3-f1713f3d3bdb\") " Mar 13 15:38:28 crc kubenswrapper[4907]: I0313 15:38:28.935575 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b43d065a-571d-416f-96d3-f1713f3d3bdb-ovsdbserver-sb\") pod \"b43d065a-571d-416f-96d3-f1713f3d3bdb\" (UID: \"b43d065a-571d-416f-96d3-f1713f3d3bdb\") " Mar 13 15:38:28 crc kubenswrapper[4907]: I0313 15:38:28.935600 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-99knh\" (UniqueName: \"kubernetes.io/projected/b43d065a-571d-416f-96d3-f1713f3d3bdb-kube-api-access-99knh\") pod \"b43d065a-571d-416f-96d3-f1713f3d3bdb\" (UID: \"b43d065a-571d-416f-96d3-f1713f3d3bdb\") " Mar 13 15:38:28 crc kubenswrapper[4907]: I0313 15:38:28.935767 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b43d065a-571d-416f-96d3-f1713f3d3bdb-dns-svc\") pod \"b43d065a-571d-416f-96d3-f1713f3d3bdb\" (UID: \"b43d065a-571d-416f-96d3-f1713f3d3bdb\") " Mar 13 15:38:28 crc kubenswrapper[4907]: I0313 15:38:28.935820 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b43d065a-571d-416f-96d3-f1713f3d3bdb-config\") pod \"b43d065a-571d-416f-96d3-f1713f3d3bdb\" (UID: \"b43d065a-571d-416f-96d3-f1713f3d3bdb\") " Mar 13 15:38:28 crc kubenswrapper[4907]: I0313 15:38:28.944280 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b43d065a-571d-416f-96d3-f1713f3d3bdb-kube-api-access-99knh" (OuterVolumeSpecName: "kube-api-access-99knh") pod "b43d065a-571d-416f-96d3-f1713f3d3bdb" (UID: "b43d065a-571d-416f-96d3-f1713f3d3bdb"). InnerVolumeSpecName "kube-api-access-99knh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:38:28 crc kubenswrapper[4907]: I0313 15:38:28.979250 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b43d065a-571d-416f-96d3-f1713f3d3bdb-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "b43d065a-571d-416f-96d3-f1713f3d3bdb" (UID: "b43d065a-571d-416f-96d3-f1713f3d3bdb"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:38:28 crc kubenswrapper[4907]: I0313 15:38:28.980739 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b43d065a-571d-416f-96d3-f1713f3d3bdb-config" (OuterVolumeSpecName: "config") pod "b43d065a-571d-416f-96d3-f1713f3d3bdb" (UID: "b43d065a-571d-416f-96d3-f1713f3d3bdb"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:38:28 crc kubenswrapper[4907]: I0313 15:38:28.983668 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b43d065a-571d-416f-96d3-f1713f3d3bdb-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b43d065a-571d-416f-96d3-f1713f3d3bdb" (UID: "b43d065a-571d-416f-96d3-f1713f3d3bdb"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:38:28 crc kubenswrapper[4907]: I0313 15:38:28.989362 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b43d065a-571d-416f-96d3-f1713f3d3bdb-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "b43d065a-571d-416f-96d3-f1713f3d3bdb" (UID: "b43d065a-571d-416f-96d3-f1713f3d3bdb"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:38:29 crc kubenswrapper[4907]: I0313 15:38:29.037471 4907 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b43d065a-571d-416f-96d3-f1713f3d3bdb-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 13 15:38:29 crc kubenswrapper[4907]: I0313 15:38:29.037504 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b43d065a-571d-416f-96d3-f1713f3d3bdb-config\") on node \"crc\" DevicePath \"\"" Mar 13 15:38:29 crc kubenswrapper[4907]: I0313 15:38:29.037514 4907 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b43d065a-571d-416f-96d3-f1713f3d3bdb-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 13 15:38:29 crc kubenswrapper[4907]: I0313 15:38:29.037525 4907 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b43d065a-571d-416f-96d3-f1713f3d3bdb-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 13 15:38:29 crc kubenswrapper[4907]: I0313 15:38:29.037536 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-99knh\" (UniqueName: \"kubernetes.io/projected/b43d065a-571d-416f-96d3-f1713f3d3bdb-kube-api-access-99knh\") on node \"crc\" DevicePath \"\"" Mar 13 15:38:29 crc kubenswrapper[4907]: I0313 15:38:29.669154 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-db9bd6759-f9ndc" event={"ID":"b43d065a-571d-416f-96d3-f1713f3d3bdb","Type":"ContainerDied","Data":"89449f9bc1a34a373899c9c8f376372efe1d7f3851750d5a63cfd6c8a1ef4335"} Mar 13 15:38:29 crc kubenswrapper[4907]: I0313 15:38:29.669223 4907 scope.go:117] "RemoveContainer" containerID="181ece190f085ef12291d2abeb4a964937fb06d189ff825afc0a6770f3e932ba" Mar 13 15:38:29 crc kubenswrapper[4907]: I0313 15:38:29.669232 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-db9bd6759-f9ndc" Mar 13 15:38:29 crc kubenswrapper[4907]: I0313 15:38:29.718766 4907 scope.go:117] "RemoveContainer" containerID="f003308b3b3b337edbb1d567791fd2916dac4c39389f2b18c7276a22c4799e5d" Mar 13 15:38:29 crc kubenswrapper[4907]: I0313 15:38:29.727585 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-db9bd6759-f9ndc"] Mar 13 15:38:29 crc kubenswrapper[4907]: I0313 15:38:29.753142 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-db9bd6759-f9ndc"] Mar 13 15:38:29 crc kubenswrapper[4907]: I0313 15:38:29.799945 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b43d065a-571d-416f-96d3-f1713f3d3bdb" path="/var/lib/kubelet/pods/b43d065a-571d-416f-96d3-f1713f3d3bdb/volumes" Mar 13 15:38:31 crc kubenswrapper[4907]: I0313 15:38:31.967608 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Mar 13 15:38:31 crc kubenswrapper[4907]: I0313 15:38:31.967935 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Mar 13 15:38:32 crc kubenswrapper[4907]: I0313 15:38:32.007147 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Mar 13 15:38:32 crc kubenswrapper[4907]: I0313 15:38:32.016582 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Mar 13 15:38:32 crc kubenswrapper[4907]: I0313 15:38:32.707307 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Mar 13 15:38:32 crc kubenswrapper[4907]: I0313 15:38:32.707357 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Mar 13 15:38:33 crc kubenswrapper[4907]: I0313 15:38:33.987071 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Mar 13 15:38:33 crc kubenswrapper[4907]: I0313 15:38:33.987398 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Mar 13 15:38:34 crc kubenswrapper[4907]: I0313 15:38:34.015229 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Mar 13 15:38:34 crc kubenswrapper[4907]: I0313 15:38:34.045976 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Mar 13 15:38:34 crc kubenswrapper[4907]: I0313 15:38:34.739382 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Mar 13 15:38:34 crc kubenswrapper[4907]: I0313 15:38:34.739737 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Mar 13 15:38:34 crc kubenswrapper[4907]: I0313 15:38:34.818126 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Mar 13 15:38:34 crc kubenswrapper[4907]: I0313 15:38:34.818243 4907 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 13 15:38:34 crc kubenswrapper[4907]: I0313 15:38:34.856566 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Mar 13 15:38:36 crc kubenswrapper[4907]: I0313 15:38:36.725794 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Mar 13 15:38:36 crc kubenswrapper[4907]: I0313 15:38:36.730137 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Mar 13 15:38:46 crc kubenswrapper[4907]: I0313 15:38:46.373243 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-5svds"] Mar 13 15:38:46 crc kubenswrapper[4907]: E0313 15:38:46.374193 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b43d065a-571d-416f-96d3-f1713f3d3bdb" containerName="init" Mar 13 15:38:46 crc kubenswrapper[4907]: I0313 15:38:46.374208 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="b43d065a-571d-416f-96d3-f1713f3d3bdb" containerName="init" Mar 13 15:38:46 crc kubenswrapper[4907]: E0313 15:38:46.374229 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b43d065a-571d-416f-96d3-f1713f3d3bdb" containerName="dnsmasq-dns" Mar 13 15:38:46 crc kubenswrapper[4907]: I0313 15:38:46.374235 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="b43d065a-571d-416f-96d3-f1713f3d3bdb" containerName="dnsmasq-dns" Mar 13 15:38:46 crc kubenswrapper[4907]: I0313 15:38:46.374420 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="b43d065a-571d-416f-96d3-f1713f3d3bdb" containerName="dnsmasq-dns" Mar 13 15:38:46 crc kubenswrapper[4907]: I0313 15:38:46.375013 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-5svds" Mar 13 15:38:46 crc kubenswrapper[4907]: I0313 15:38:46.384185 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-5svds"] Mar 13 15:38:46 crc kubenswrapper[4907]: I0313 15:38:46.434145 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9x5m8\" (UniqueName: \"kubernetes.io/projected/978c19e3-b11d-44ef-86d5-6b0674e488d1-kube-api-access-9x5m8\") pod \"placement-db-create-5svds\" (UID: \"978c19e3-b11d-44ef-86d5-6b0674e488d1\") " pod="openstack/placement-db-create-5svds" Mar 13 15:38:46 crc kubenswrapper[4907]: I0313 15:38:46.434204 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/978c19e3-b11d-44ef-86d5-6b0674e488d1-operator-scripts\") pod \"placement-db-create-5svds\" (UID: \"978c19e3-b11d-44ef-86d5-6b0674e488d1\") " pod="openstack/placement-db-create-5svds" Mar 13 15:38:46 crc kubenswrapper[4907]: I0313 15:38:46.477707 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-c036-account-create-update-26jl8"] Mar 13 15:38:46 crc kubenswrapper[4907]: I0313 15:38:46.479209 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-c036-account-create-update-26jl8" Mar 13 15:38:46 crc kubenswrapper[4907]: I0313 15:38:46.481479 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Mar 13 15:38:46 crc kubenswrapper[4907]: I0313 15:38:46.489774 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-c036-account-create-update-26jl8"] Mar 13 15:38:46 crc kubenswrapper[4907]: I0313 15:38:46.535760 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0e98ef6f-8765-4e4d-ab91-7dcdcb3f788d-operator-scripts\") pod \"placement-c036-account-create-update-26jl8\" (UID: \"0e98ef6f-8765-4e4d-ab91-7dcdcb3f788d\") " pod="openstack/placement-c036-account-create-update-26jl8" Mar 13 15:38:46 crc kubenswrapper[4907]: I0313 15:38:46.535807 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pq9rc\" (UniqueName: \"kubernetes.io/projected/0e98ef6f-8765-4e4d-ab91-7dcdcb3f788d-kube-api-access-pq9rc\") pod \"placement-c036-account-create-update-26jl8\" (UID: \"0e98ef6f-8765-4e4d-ab91-7dcdcb3f788d\") " pod="openstack/placement-c036-account-create-update-26jl8" Mar 13 15:38:46 crc kubenswrapper[4907]: I0313 15:38:46.535867 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9x5m8\" (UniqueName: \"kubernetes.io/projected/978c19e3-b11d-44ef-86d5-6b0674e488d1-kube-api-access-9x5m8\") pod \"placement-db-create-5svds\" (UID: \"978c19e3-b11d-44ef-86d5-6b0674e488d1\") " pod="openstack/placement-db-create-5svds" Mar 13 15:38:46 crc kubenswrapper[4907]: I0313 15:38:46.535995 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/978c19e3-b11d-44ef-86d5-6b0674e488d1-operator-scripts\") pod \"placement-db-create-5svds\" (UID: \"978c19e3-b11d-44ef-86d5-6b0674e488d1\") " pod="openstack/placement-db-create-5svds" Mar 13 15:38:46 crc kubenswrapper[4907]: I0313 15:38:46.537033 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/978c19e3-b11d-44ef-86d5-6b0674e488d1-operator-scripts\") pod \"placement-db-create-5svds\" (UID: \"978c19e3-b11d-44ef-86d5-6b0674e488d1\") " pod="openstack/placement-db-create-5svds" Mar 13 15:38:46 crc kubenswrapper[4907]: I0313 15:38:46.554643 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9x5m8\" (UniqueName: \"kubernetes.io/projected/978c19e3-b11d-44ef-86d5-6b0674e488d1-kube-api-access-9x5m8\") pod \"placement-db-create-5svds\" (UID: \"978c19e3-b11d-44ef-86d5-6b0674e488d1\") " pod="openstack/placement-db-create-5svds" Mar 13 15:38:46 crc kubenswrapper[4907]: I0313 15:38:46.643811 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0e98ef6f-8765-4e4d-ab91-7dcdcb3f788d-operator-scripts\") pod \"placement-c036-account-create-update-26jl8\" (UID: \"0e98ef6f-8765-4e4d-ab91-7dcdcb3f788d\") " pod="openstack/placement-c036-account-create-update-26jl8" Mar 13 15:38:46 crc kubenswrapper[4907]: I0313 15:38:46.643899 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pq9rc\" (UniqueName: \"kubernetes.io/projected/0e98ef6f-8765-4e4d-ab91-7dcdcb3f788d-kube-api-access-pq9rc\") pod \"placement-c036-account-create-update-26jl8\" (UID: \"0e98ef6f-8765-4e4d-ab91-7dcdcb3f788d\") " pod="openstack/placement-c036-account-create-update-26jl8" Mar 13 15:38:46 crc kubenswrapper[4907]: I0313 15:38:46.645193 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0e98ef6f-8765-4e4d-ab91-7dcdcb3f788d-operator-scripts\") pod \"placement-c036-account-create-update-26jl8\" (UID: \"0e98ef6f-8765-4e4d-ab91-7dcdcb3f788d\") " pod="openstack/placement-c036-account-create-update-26jl8" Mar 13 15:38:46 crc kubenswrapper[4907]: I0313 15:38:46.665534 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pq9rc\" (UniqueName: \"kubernetes.io/projected/0e98ef6f-8765-4e4d-ab91-7dcdcb3f788d-kube-api-access-pq9rc\") pod \"placement-c036-account-create-update-26jl8\" (UID: \"0e98ef6f-8765-4e4d-ab91-7dcdcb3f788d\") " pod="openstack/placement-c036-account-create-update-26jl8" Mar 13 15:38:46 crc kubenswrapper[4907]: I0313 15:38:46.692214 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-5svds" Mar 13 15:38:46 crc kubenswrapper[4907]: I0313 15:38:46.807146 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-c036-account-create-update-26jl8" Mar 13 15:38:47 crc kubenswrapper[4907]: I0313 15:38:47.156172 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-5svds"] Mar 13 15:38:47 crc kubenswrapper[4907]: I0313 15:38:47.260694 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-c036-account-create-update-26jl8"] Mar 13 15:38:47 crc kubenswrapper[4907]: W0313 15:38:47.267319 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0e98ef6f_8765_4e4d_ab91_7dcdcb3f788d.slice/crio-4c1252f178308a87af8cf89fb1bf3b3a1cf8b33192c8980c6d5095afcec1f78c WatchSource:0}: Error finding container 4c1252f178308a87af8cf89fb1bf3b3a1cf8b33192c8980c6d5095afcec1f78c: Status 404 returned error can't find the container with id 4c1252f178308a87af8cf89fb1bf3b3a1cf8b33192c8980c6d5095afcec1f78c Mar 13 15:38:47 crc kubenswrapper[4907]: I0313 15:38:47.880942 4907 generic.go:334] "Generic (PLEG): container finished" podID="0e98ef6f-8765-4e4d-ab91-7dcdcb3f788d" containerID="dcfa157fb4134a5c5ee1865cf6f3b6a035a94e4eeb4f7a317b410b36548c37aa" exitCode=0 Mar 13 15:38:47 crc kubenswrapper[4907]: I0313 15:38:47.881056 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-c036-account-create-update-26jl8" event={"ID":"0e98ef6f-8765-4e4d-ab91-7dcdcb3f788d","Type":"ContainerDied","Data":"dcfa157fb4134a5c5ee1865cf6f3b6a035a94e4eeb4f7a317b410b36548c37aa"} Mar 13 15:38:47 crc kubenswrapper[4907]: I0313 15:38:47.881102 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-c036-account-create-update-26jl8" event={"ID":"0e98ef6f-8765-4e4d-ab91-7dcdcb3f788d","Type":"ContainerStarted","Data":"4c1252f178308a87af8cf89fb1bf3b3a1cf8b33192c8980c6d5095afcec1f78c"} Mar 13 15:38:47 crc kubenswrapper[4907]: I0313 15:38:47.882773 4907 generic.go:334] "Generic (PLEG): container finished" podID="978c19e3-b11d-44ef-86d5-6b0674e488d1" containerID="48b0748e550ebe6b33142e46ea370e9b43cfd8df0c8482529bda2ee099868df0" exitCode=0 Mar 13 15:38:47 crc kubenswrapper[4907]: I0313 15:38:47.882804 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-5svds" event={"ID":"978c19e3-b11d-44ef-86d5-6b0674e488d1","Type":"ContainerDied","Data":"48b0748e550ebe6b33142e46ea370e9b43cfd8df0c8482529bda2ee099868df0"} Mar 13 15:38:47 crc kubenswrapper[4907]: I0313 15:38:47.882818 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-5svds" event={"ID":"978c19e3-b11d-44ef-86d5-6b0674e488d1","Type":"ContainerStarted","Data":"1868172bc20ad95b4823c0ddc65847d8216cf5daabbe1eacc5e6555c36492028"} Mar 13 15:38:49 crc kubenswrapper[4907]: I0313 15:38:49.308800 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-5svds" Mar 13 15:38:49 crc kubenswrapper[4907]: I0313 15:38:49.316011 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-c036-account-create-update-26jl8" Mar 13 15:38:49 crc kubenswrapper[4907]: I0313 15:38:49.393253 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0e98ef6f-8765-4e4d-ab91-7dcdcb3f788d-operator-scripts\") pod \"0e98ef6f-8765-4e4d-ab91-7dcdcb3f788d\" (UID: \"0e98ef6f-8765-4e4d-ab91-7dcdcb3f788d\") " Mar 13 15:38:49 crc kubenswrapper[4907]: I0313 15:38:49.393347 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9x5m8\" (UniqueName: \"kubernetes.io/projected/978c19e3-b11d-44ef-86d5-6b0674e488d1-kube-api-access-9x5m8\") pod \"978c19e3-b11d-44ef-86d5-6b0674e488d1\" (UID: \"978c19e3-b11d-44ef-86d5-6b0674e488d1\") " Mar 13 15:38:49 crc kubenswrapper[4907]: I0313 15:38:49.393378 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pq9rc\" (UniqueName: \"kubernetes.io/projected/0e98ef6f-8765-4e4d-ab91-7dcdcb3f788d-kube-api-access-pq9rc\") pod \"0e98ef6f-8765-4e4d-ab91-7dcdcb3f788d\" (UID: \"0e98ef6f-8765-4e4d-ab91-7dcdcb3f788d\") " Mar 13 15:38:49 crc kubenswrapper[4907]: I0313 15:38:49.393552 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/978c19e3-b11d-44ef-86d5-6b0674e488d1-operator-scripts\") pod \"978c19e3-b11d-44ef-86d5-6b0674e488d1\" (UID: \"978c19e3-b11d-44ef-86d5-6b0674e488d1\") " Mar 13 15:38:49 crc kubenswrapper[4907]: I0313 15:38:49.394107 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0e98ef6f-8765-4e4d-ab91-7dcdcb3f788d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "0e98ef6f-8765-4e4d-ab91-7dcdcb3f788d" (UID: "0e98ef6f-8765-4e4d-ab91-7dcdcb3f788d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:38:49 crc kubenswrapper[4907]: I0313 15:38:49.394303 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/978c19e3-b11d-44ef-86d5-6b0674e488d1-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "978c19e3-b11d-44ef-86d5-6b0674e488d1" (UID: "978c19e3-b11d-44ef-86d5-6b0674e488d1"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:38:49 crc kubenswrapper[4907]: I0313 15:38:49.401121 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/978c19e3-b11d-44ef-86d5-6b0674e488d1-kube-api-access-9x5m8" (OuterVolumeSpecName: "kube-api-access-9x5m8") pod "978c19e3-b11d-44ef-86d5-6b0674e488d1" (UID: "978c19e3-b11d-44ef-86d5-6b0674e488d1"). InnerVolumeSpecName "kube-api-access-9x5m8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:38:49 crc kubenswrapper[4907]: I0313 15:38:49.401178 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0e98ef6f-8765-4e4d-ab91-7dcdcb3f788d-kube-api-access-pq9rc" (OuterVolumeSpecName: "kube-api-access-pq9rc") pod "0e98ef6f-8765-4e4d-ab91-7dcdcb3f788d" (UID: "0e98ef6f-8765-4e4d-ab91-7dcdcb3f788d"). InnerVolumeSpecName "kube-api-access-pq9rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:38:49 crc kubenswrapper[4907]: I0313 15:38:49.495376 4907 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/978c19e3-b11d-44ef-86d5-6b0674e488d1-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 15:38:49 crc kubenswrapper[4907]: I0313 15:38:49.495667 4907 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0e98ef6f-8765-4e4d-ab91-7dcdcb3f788d-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 15:38:49 crc kubenswrapper[4907]: I0313 15:38:49.495681 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9x5m8\" (UniqueName: \"kubernetes.io/projected/978c19e3-b11d-44ef-86d5-6b0674e488d1-kube-api-access-9x5m8\") on node \"crc\" DevicePath \"\"" Mar 13 15:38:49 crc kubenswrapper[4907]: I0313 15:38:49.495693 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pq9rc\" (UniqueName: \"kubernetes.io/projected/0e98ef6f-8765-4e4d-ab91-7dcdcb3f788d-kube-api-access-pq9rc\") on node \"crc\" DevicePath \"\"" Mar 13 15:38:49 crc kubenswrapper[4907]: I0313 15:38:49.900717 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-c036-account-create-update-26jl8" event={"ID":"0e98ef6f-8765-4e4d-ab91-7dcdcb3f788d","Type":"ContainerDied","Data":"4c1252f178308a87af8cf89fb1bf3b3a1cf8b33192c8980c6d5095afcec1f78c"} Mar 13 15:38:49 crc kubenswrapper[4907]: I0313 15:38:49.900755 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4c1252f178308a87af8cf89fb1bf3b3a1cf8b33192c8980c6d5095afcec1f78c" Mar 13 15:38:49 crc kubenswrapper[4907]: I0313 15:38:49.900765 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-c036-account-create-update-26jl8" Mar 13 15:38:49 crc kubenswrapper[4907]: I0313 15:38:49.902305 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-5svds" event={"ID":"978c19e3-b11d-44ef-86d5-6b0674e488d1","Type":"ContainerDied","Data":"1868172bc20ad95b4823c0ddc65847d8216cf5daabbe1eacc5e6555c36492028"} Mar 13 15:38:49 crc kubenswrapper[4907]: I0313 15:38:49.902325 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1868172bc20ad95b4823c0ddc65847d8216cf5daabbe1eacc5e6555c36492028" Mar 13 15:38:49 crc kubenswrapper[4907]: I0313 15:38:49.902336 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-5svds" Mar 13 15:38:51 crc kubenswrapper[4907]: I0313 15:38:51.845809 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5dd7c58f95-579qq"] Mar 13 15:38:51 crc kubenswrapper[4907]: E0313 15:38:51.846555 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e98ef6f-8765-4e4d-ab91-7dcdcb3f788d" containerName="mariadb-account-create-update" Mar 13 15:38:51 crc kubenswrapper[4907]: I0313 15:38:51.846574 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e98ef6f-8765-4e4d-ab91-7dcdcb3f788d" containerName="mariadb-account-create-update" Mar 13 15:38:51 crc kubenswrapper[4907]: E0313 15:38:51.846609 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="978c19e3-b11d-44ef-86d5-6b0674e488d1" containerName="mariadb-database-create" Mar 13 15:38:51 crc kubenswrapper[4907]: I0313 15:38:51.846618 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="978c19e3-b11d-44ef-86d5-6b0674e488d1" containerName="mariadb-database-create" Mar 13 15:38:51 crc kubenswrapper[4907]: I0313 15:38:51.846828 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e98ef6f-8765-4e4d-ab91-7dcdcb3f788d" containerName="mariadb-account-create-update" Mar 13 15:38:51 crc kubenswrapper[4907]: I0313 15:38:51.846848 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="978c19e3-b11d-44ef-86d5-6b0674e488d1" containerName="mariadb-database-create" Mar 13 15:38:51 crc kubenswrapper[4907]: I0313 15:38:51.849933 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5dd7c58f95-579qq" Mar 13 15:38:51 crc kubenswrapper[4907]: I0313 15:38:51.877908 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5dd7c58f95-579qq"] Mar 13 15:38:51 crc kubenswrapper[4907]: I0313 15:38:51.910967 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-z7wl4"] Mar 13 15:38:51 crc kubenswrapper[4907]: I0313 15:38:51.912819 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-z7wl4" Mar 13 15:38:51 crc kubenswrapper[4907]: I0313 15:38:51.917210 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-6c87x" Mar 13 15:38:51 crc kubenswrapper[4907]: I0313 15:38:51.917260 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Mar 13 15:38:51 crc kubenswrapper[4907]: I0313 15:38:51.917416 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Mar 13 15:38:51 crc kubenswrapper[4907]: I0313 15:38:51.937987 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5153eb33-98f7-4452-b4f5-751a6d49c54b-dns-svc\") pod \"dnsmasq-dns-5dd7c58f95-579qq\" (UID: \"5153eb33-98f7-4452-b4f5-751a6d49c54b\") " pod="openstack/dnsmasq-dns-5dd7c58f95-579qq" Mar 13 15:38:51 crc kubenswrapper[4907]: I0313 15:38:51.938042 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5153eb33-98f7-4452-b4f5-751a6d49c54b-ovsdbserver-sb\") pod \"dnsmasq-dns-5dd7c58f95-579qq\" (UID: \"5153eb33-98f7-4452-b4f5-751a6d49c54b\") " pod="openstack/dnsmasq-dns-5dd7c58f95-579qq" Mar 13 15:38:51 crc kubenswrapper[4907]: I0313 15:38:51.938081 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5153eb33-98f7-4452-b4f5-751a6d49c54b-config\") pod \"dnsmasq-dns-5dd7c58f95-579qq\" (UID: \"5153eb33-98f7-4452-b4f5-751a6d49c54b\") " pod="openstack/dnsmasq-dns-5dd7c58f95-579qq" Mar 13 15:38:51 crc kubenswrapper[4907]: I0313 15:38:51.938109 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wh4kf\" (UniqueName: \"kubernetes.io/projected/a43a8ffb-4f15-4524-b750-517442c1f561-kube-api-access-wh4kf\") pod \"placement-db-sync-z7wl4\" (UID: \"a43a8ffb-4f15-4524-b750-517442c1f561\") " pod="openstack/placement-db-sync-z7wl4" Mar 13 15:38:51 crc kubenswrapper[4907]: I0313 15:38:51.938131 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gskhr\" (UniqueName: \"kubernetes.io/projected/5153eb33-98f7-4452-b4f5-751a6d49c54b-kube-api-access-gskhr\") pod \"dnsmasq-dns-5dd7c58f95-579qq\" (UID: \"5153eb33-98f7-4452-b4f5-751a6d49c54b\") " pod="openstack/dnsmasq-dns-5dd7c58f95-579qq" Mar 13 15:38:51 crc kubenswrapper[4907]: I0313 15:38:51.938187 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5153eb33-98f7-4452-b4f5-751a6d49c54b-ovsdbserver-nb\") pod \"dnsmasq-dns-5dd7c58f95-579qq\" (UID: \"5153eb33-98f7-4452-b4f5-751a6d49c54b\") " pod="openstack/dnsmasq-dns-5dd7c58f95-579qq" Mar 13 15:38:51 crc kubenswrapper[4907]: I0313 15:38:51.938225 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a43a8ffb-4f15-4524-b750-517442c1f561-combined-ca-bundle\") pod \"placement-db-sync-z7wl4\" (UID: \"a43a8ffb-4f15-4524-b750-517442c1f561\") " pod="openstack/placement-db-sync-z7wl4" Mar 13 15:38:51 crc kubenswrapper[4907]: I0313 15:38:51.938250 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a43a8ffb-4f15-4524-b750-517442c1f561-scripts\") pod \"placement-db-sync-z7wl4\" (UID: \"a43a8ffb-4f15-4524-b750-517442c1f561\") " pod="openstack/placement-db-sync-z7wl4" Mar 13 15:38:51 crc kubenswrapper[4907]: I0313 15:38:51.938272 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a43a8ffb-4f15-4524-b750-517442c1f561-config-data\") pod \"placement-db-sync-z7wl4\" (UID: \"a43a8ffb-4f15-4524-b750-517442c1f561\") " pod="openstack/placement-db-sync-z7wl4" Mar 13 15:38:51 crc kubenswrapper[4907]: I0313 15:38:51.938295 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a43a8ffb-4f15-4524-b750-517442c1f561-logs\") pod \"placement-db-sync-z7wl4\" (UID: \"a43a8ffb-4f15-4524-b750-517442c1f561\") " pod="openstack/placement-db-sync-z7wl4" Mar 13 15:38:51 crc kubenswrapper[4907]: I0313 15:38:51.940536 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-z7wl4"] Mar 13 15:38:52 crc kubenswrapper[4907]: I0313 15:38:52.049827 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5153eb33-98f7-4452-b4f5-751a6d49c54b-ovsdbserver-nb\") pod \"dnsmasq-dns-5dd7c58f95-579qq\" (UID: \"5153eb33-98f7-4452-b4f5-751a6d49c54b\") " pod="openstack/dnsmasq-dns-5dd7c58f95-579qq" Mar 13 15:38:52 crc kubenswrapper[4907]: I0313 15:38:52.049909 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a43a8ffb-4f15-4524-b750-517442c1f561-combined-ca-bundle\") pod \"placement-db-sync-z7wl4\" (UID: \"a43a8ffb-4f15-4524-b750-517442c1f561\") " pod="openstack/placement-db-sync-z7wl4" Mar 13 15:38:52 crc kubenswrapper[4907]: I0313 15:38:52.049937 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a43a8ffb-4f15-4524-b750-517442c1f561-scripts\") pod \"placement-db-sync-z7wl4\" (UID: \"a43a8ffb-4f15-4524-b750-517442c1f561\") " pod="openstack/placement-db-sync-z7wl4" Mar 13 15:38:52 crc kubenswrapper[4907]: I0313 15:38:52.049964 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a43a8ffb-4f15-4524-b750-517442c1f561-config-data\") pod \"placement-db-sync-z7wl4\" (UID: \"a43a8ffb-4f15-4524-b750-517442c1f561\") " pod="openstack/placement-db-sync-z7wl4" Mar 13 15:38:52 crc kubenswrapper[4907]: I0313 15:38:52.049986 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a43a8ffb-4f15-4524-b750-517442c1f561-logs\") pod \"placement-db-sync-z7wl4\" (UID: \"a43a8ffb-4f15-4524-b750-517442c1f561\") " pod="openstack/placement-db-sync-z7wl4" Mar 13 15:38:52 crc kubenswrapper[4907]: I0313 15:38:52.050020 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5153eb33-98f7-4452-b4f5-751a6d49c54b-dns-svc\") pod \"dnsmasq-dns-5dd7c58f95-579qq\" (UID: \"5153eb33-98f7-4452-b4f5-751a6d49c54b\") " pod="openstack/dnsmasq-dns-5dd7c58f95-579qq" Mar 13 15:38:52 crc kubenswrapper[4907]: I0313 15:38:52.050034 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5153eb33-98f7-4452-b4f5-751a6d49c54b-ovsdbserver-sb\") pod \"dnsmasq-dns-5dd7c58f95-579qq\" (UID: \"5153eb33-98f7-4452-b4f5-751a6d49c54b\") " pod="openstack/dnsmasq-dns-5dd7c58f95-579qq" Mar 13 15:38:52 crc kubenswrapper[4907]: I0313 15:38:52.050052 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5153eb33-98f7-4452-b4f5-751a6d49c54b-config\") pod \"dnsmasq-dns-5dd7c58f95-579qq\" (UID: \"5153eb33-98f7-4452-b4f5-751a6d49c54b\") " pod="openstack/dnsmasq-dns-5dd7c58f95-579qq" Mar 13 15:38:52 crc kubenswrapper[4907]: I0313 15:38:52.050073 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wh4kf\" (UniqueName: \"kubernetes.io/projected/a43a8ffb-4f15-4524-b750-517442c1f561-kube-api-access-wh4kf\") pod \"placement-db-sync-z7wl4\" (UID: \"a43a8ffb-4f15-4524-b750-517442c1f561\") " pod="openstack/placement-db-sync-z7wl4" Mar 13 15:38:52 crc kubenswrapper[4907]: I0313 15:38:52.050092 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gskhr\" (UniqueName: \"kubernetes.io/projected/5153eb33-98f7-4452-b4f5-751a6d49c54b-kube-api-access-gskhr\") pod \"dnsmasq-dns-5dd7c58f95-579qq\" (UID: \"5153eb33-98f7-4452-b4f5-751a6d49c54b\") " pod="openstack/dnsmasq-dns-5dd7c58f95-579qq" Mar 13 15:38:52 crc kubenswrapper[4907]: I0313 15:38:52.051172 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5153eb33-98f7-4452-b4f5-751a6d49c54b-ovsdbserver-nb\") pod \"dnsmasq-dns-5dd7c58f95-579qq\" (UID: \"5153eb33-98f7-4452-b4f5-751a6d49c54b\") " pod="openstack/dnsmasq-dns-5dd7c58f95-579qq" Mar 13 15:38:52 crc kubenswrapper[4907]: I0313 15:38:52.051600 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a43a8ffb-4f15-4524-b750-517442c1f561-logs\") pod \"placement-db-sync-z7wl4\" (UID: \"a43a8ffb-4f15-4524-b750-517442c1f561\") " pod="openstack/placement-db-sync-z7wl4" Mar 13 15:38:52 crc kubenswrapper[4907]: I0313 15:38:52.051800 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5153eb33-98f7-4452-b4f5-751a6d49c54b-dns-svc\") pod \"dnsmasq-dns-5dd7c58f95-579qq\" (UID: \"5153eb33-98f7-4452-b4f5-751a6d49c54b\") " pod="openstack/dnsmasq-dns-5dd7c58f95-579qq" Mar 13 15:38:52 crc kubenswrapper[4907]: I0313 15:38:52.052355 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5153eb33-98f7-4452-b4f5-751a6d49c54b-ovsdbserver-sb\") pod \"dnsmasq-dns-5dd7c58f95-579qq\" (UID: \"5153eb33-98f7-4452-b4f5-751a6d49c54b\") " pod="openstack/dnsmasq-dns-5dd7c58f95-579qq" Mar 13 15:38:52 crc kubenswrapper[4907]: I0313 15:38:52.052905 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5153eb33-98f7-4452-b4f5-751a6d49c54b-config\") pod \"dnsmasq-dns-5dd7c58f95-579qq\" (UID: \"5153eb33-98f7-4452-b4f5-751a6d49c54b\") " pod="openstack/dnsmasq-dns-5dd7c58f95-579qq" Mar 13 15:38:52 crc kubenswrapper[4907]: I0313 15:38:52.057577 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a43a8ffb-4f15-4524-b750-517442c1f561-scripts\") pod \"placement-db-sync-z7wl4\" (UID: \"a43a8ffb-4f15-4524-b750-517442c1f561\") " pod="openstack/placement-db-sync-z7wl4" Mar 13 15:38:52 crc kubenswrapper[4907]: I0313 15:38:52.058025 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a43a8ffb-4f15-4524-b750-517442c1f561-combined-ca-bundle\") pod \"placement-db-sync-z7wl4\" (UID: \"a43a8ffb-4f15-4524-b750-517442c1f561\") " pod="openstack/placement-db-sync-z7wl4" Mar 13 15:38:52 crc kubenswrapper[4907]: I0313 15:38:52.068964 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a43a8ffb-4f15-4524-b750-517442c1f561-config-data\") pod \"placement-db-sync-z7wl4\" (UID: \"a43a8ffb-4f15-4524-b750-517442c1f561\") " pod="openstack/placement-db-sync-z7wl4" Mar 13 15:38:52 crc kubenswrapper[4907]: I0313 15:38:52.069545 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gskhr\" (UniqueName: \"kubernetes.io/projected/5153eb33-98f7-4452-b4f5-751a6d49c54b-kube-api-access-gskhr\") pod \"dnsmasq-dns-5dd7c58f95-579qq\" (UID: \"5153eb33-98f7-4452-b4f5-751a6d49c54b\") " pod="openstack/dnsmasq-dns-5dd7c58f95-579qq" Mar 13 15:38:52 crc kubenswrapper[4907]: I0313 15:38:52.070318 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wh4kf\" (UniqueName: \"kubernetes.io/projected/a43a8ffb-4f15-4524-b750-517442c1f561-kube-api-access-wh4kf\") pod \"placement-db-sync-z7wl4\" (UID: \"a43a8ffb-4f15-4524-b750-517442c1f561\") " pod="openstack/placement-db-sync-z7wl4" Mar 13 15:38:52 crc kubenswrapper[4907]: I0313 15:38:52.182362 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5dd7c58f95-579qq" Mar 13 15:38:52 crc kubenswrapper[4907]: I0313 15:38:52.254406 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-z7wl4" Mar 13 15:38:52 crc kubenswrapper[4907]: I0313 15:38:52.709624 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5dd7c58f95-579qq"] Mar 13 15:38:52 crc kubenswrapper[4907]: W0313 15:38:52.711289 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5153eb33_98f7_4452_b4f5_751a6d49c54b.slice/crio-1f607dd9ca9be04a0a9e3a628ae846054d952f5ec49fd5e3e8f6f1422eb9fe8b WatchSource:0}: Error finding container 1f607dd9ca9be04a0a9e3a628ae846054d952f5ec49fd5e3e8f6f1422eb9fe8b: Status 404 returned error can't find the container with id 1f607dd9ca9be04a0a9e3a628ae846054d952f5ec49fd5e3e8f6f1422eb9fe8b Mar 13 15:38:52 crc kubenswrapper[4907]: I0313 15:38:52.807105 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-z7wl4"] Mar 13 15:38:52 crc kubenswrapper[4907]: I0313 15:38:52.944667 4907 generic.go:334] "Generic (PLEG): container finished" podID="5153eb33-98f7-4452-b4f5-751a6d49c54b" containerID="dc68df440400c3300857779e0180b105a0536c85adecc574069e0c384c4a2b78" exitCode=0 Mar 13 15:38:52 crc kubenswrapper[4907]: I0313 15:38:52.945001 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5dd7c58f95-579qq" event={"ID":"5153eb33-98f7-4452-b4f5-751a6d49c54b","Type":"ContainerDied","Data":"dc68df440400c3300857779e0180b105a0536c85adecc574069e0c384c4a2b78"} Mar 13 15:38:52 crc kubenswrapper[4907]: I0313 15:38:52.945061 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5dd7c58f95-579qq" event={"ID":"5153eb33-98f7-4452-b4f5-751a6d49c54b","Type":"ContainerStarted","Data":"1f607dd9ca9be04a0a9e3a628ae846054d952f5ec49fd5e3e8f6f1422eb9fe8b"} Mar 13 15:38:52 crc kubenswrapper[4907]: I0313 15:38:52.949431 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-z7wl4" event={"ID":"a43a8ffb-4f15-4524-b750-517442c1f561","Type":"ContainerStarted","Data":"2f496bc28dfefe3747a49ec89a56b7e3d709569e49ca85c48f559a57416fa409"} Mar 13 15:38:53 crc kubenswrapper[4907]: I0313 15:38:53.960571 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-z7wl4" event={"ID":"a43a8ffb-4f15-4524-b750-517442c1f561","Type":"ContainerStarted","Data":"10110c9c9bf7e0c30409af7a457a19b276708bc01363735a9e6dff6935d07eb0"} Mar 13 15:38:53 crc kubenswrapper[4907]: I0313 15:38:53.964156 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5dd7c58f95-579qq" event={"ID":"5153eb33-98f7-4452-b4f5-751a6d49c54b","Type":"ContainerStarted","Data":"4d47e9b4f5705f84e359bebffdaf6b40c2a5a7d780d8daa8048b0a559dabed59"} Mar 13 15:38:53 crc kubenswrapper[4907]: I0313 15:38:53.964343 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5dd7c58f95-579qq" Mar 13 15:38:53 crc kubenswrapper[4907]: I0313 15:38:53.982362 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-z7wl4" podStartSLOduration=2.982342353 podStartE2EDuration="2.982342353s" podCreationTimestamp="2026-03-13 15:38:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:38:53.973134831 +0000 UTC m=+5632.872922620" watchObservedRunningTime="2026-03-13 15:38:53.982342353 +0000 UTC m=+5632.882130052" Mar 13 15:38:53 crc kubenswrapper[4907]: I0313 15:38:53.994907 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5dd7c58f95-579qq" podStartSLOduration=2.994873926 podStartE2EDuration="2.994873926s" podCreationTimestamp="2026-03-13 15:38:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:38:53.992604083 +0000 UTC m=+5632.892391802" watchObservedRunningTime="2026-03-13 15:38:53.994873926 +0000 UTC m=+5632.894661625" Mar 13 15:38:54 crc kubenswrapper[4907]: I0313 15:38:54.977530 4907 generic.go:334] "Generic (PLEG): container finished" podID="a43a8ffb-4f15-4524-b750-517442c1f561" containerID="10110c9c9bf7e0c30409af7a457a19b276708bc01363735a9e6dff6935d07eb0" exitCode=0 Mar 13 15:38:54 crc kubenswrapper[4907]: I0313 15:38:54.977614 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-z7wl4" event={"ID":"a43a8ffb-4f15-4524-b750-517442c1f561","Type":"ContainerDied","Data":"10110c9c9bf7e0c30409af7a457a19b276708bc01363735a9e6dff6935d07eb0"} Mar 13 15:38:56 crc kubenswrapper[4907]: I0313 15:38:56.325459 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-z7wl4" Mar 13 15:38:56 crc kubenswrapper[4907]: I0313 15:38:56.440383 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a43a8ffb-4f15-4524-b750-517442c1f561-combined-ca-bundle\") pod \"a43a8ffb-4f15-4524-b750-517442c1f561\" (UID: \"a43a8ffb-4f15-4524-b750-517442c1f561\") " Mar 13 15:38:56 crc kubenswrapper[4907]: I0313 15:38:56.440548 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a43a8ffb-4f15-4524-b750-517442c1f561-logs\") pod \"a43a8ffb-4f15-4524-b750-517442c1f561\" (UID: \"a43a8ffb-4f15-4524-b750-517442c1f561\") " Mar 13 15:38:56 crc kubenswrapper[4907]: I0313 15:38:56.440619 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a43a8ffb-4f15-4524-b750-517442c1f561-scripts\") pod \"a43a8ffb-4f15-4524-b750-517442c1f561\" (UID: \"a43a8ffb-4f15-4524-b750-517442c1f561\") " Mar 13 15:38:56 crc kubenswrapper[4907]: I0313 15:38:56.440720 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a43a8ffb-4f15-4524-b750-517442c1f561-config-data\") pod \"a43a8ffb-4f15-4524-b750-517442c1f561\" (UID: \"a43a8ffb-4f15-4524-b750-517442c1f561\") " Mar 13 15:38:56 crc kubenswrapper[4907]: I0313 15:38:56.440798 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wh4kf\" (UniqueName: \"kubernetes.io/projected/a43a8ffb-4f15-4524-b750-517442c1f561-kube-api-access-wh4kf\") pod \"a43a8ffb-4f15-4524-b750-517442c1f561\" (UID: \"a43a8ffb-4f15-4524-b750-517442c1f561\") " Mar 13 15:38:56 crc kubenswrapper[4907]: I0313 15:38:56.440919 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a43a8ffb-4f15-4524-b750-517442c1f561-logs" (OuterVolumeSpecName: "logs") pod "a43a8ffb-4f15-4524-b750-517442c1f561" (UID: "a43a8ffb-4f15-4524-b750-517442c1f561"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:38:56 crc kubenswrapper[4907]: I0313 15:38:56.441620 4907 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a43a8ffb-4f15-4524-b750-517442c1f561-logs\") on node \"crc\" DevicePath \"\"" Mar 13 15:38:56 crc kubenswrapper[4907]: I0313 15:38:56.460084 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a43a8ffb-4f15-4524-b750-517442c1f561-kube-api-access-wh4kf" (OuterVolumeSpecName: "kube-api-access-wh4kf") pod "a43a8ffb-4f15-4524-b750-517442c1f561" (UID: "a43a8ffb-4f15-4524-b750-517442c1f561"). InnerVolumeSpecName "kube-api-access-wh4kf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:38:56 crc kubenswrapper[4907]: I0313 15:38:56.460213 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a43a8ffb-4f15-4524-b750-517442c1f561-scripts" (OuterVolumeSpecName: "scripts") pod "a43a8ffb-4f15-4524-b750-517442c1f561" (UID: "a43a8ffb-4f15-4524-b750-517442c1f561"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:38:56 crc kubenswrapper[4907]: I0313 15:38:56.465140 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a43a8ffb-4f15-4524-b750-517442c1f561-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a43a8ffb-4f15-4524-b750-517442c1f561" (UID: "a43a8ffb-4f15-4524-b750-517442c1f561"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:38:56 crc kubenswrapper[4907]: I0313 15:38:56.475213 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a43a8ffb-4f15-4524-b750-517442c1f561-config-data" (OuterVolumeSpecName: "config-data") pod "a43a8ffb-4f15-4524-b750-517442c1f561" (UID: "a43a8ffb-4f15-4524-b750-517442c1f561"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:38:56 crc kubenswrapper[4907]: I0313 15:38:56.543414 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a43a8ffb-4f15-4524-b750-517442c1f561-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 15:38:56 crc kubenswrapper[4907]: I0313 15:38:56.543458 4907 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a43a8ffb-4f15-4524-b750-517442c1f561-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 15:38:56 crc kubenswrapper[4907]: I0313 15:38:56.543470 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a43a8ffb-4f15-4524-b750-517442c1f561-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 15:38:56 crc kubenswrapper[4907]: I0313 15:38:56.543481 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wh4kf\" (UniqueName: \"kubernetes.io/projected/a43a8ffb-4f15-4524-b750-517442c1f561-kube-api-access-wh4kf\") on node \"crc\" DevicePath \"\"" Mar 13 15:38:56 crc kubenswrapper[4907]: I0313 15:38:56.994659 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-z7wl4" event={"ID":"a43a8ffb-4f15-4524-b750-517442c1f561","Type":"ContainerDied","Data":"2f496bc28dfefe3747a49ec89a56b7e3d709569e49ca85c48f559a57416fa409"} Mar 13 15:38:56 crc kubenswrapper[4907]: I0313 15:38:56.994697 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2f496bc28dfefe3747a49ec89a56b7e3d709569e49ca85c48f559a57416fa409" Mar 13 15:38:56 crc kubenswrapper[4907]: I0313 15:38:56.994753 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-z7wl4" Mar 13 15:38:57 crc kubenswrapper[4907]: I0313 15:38:57.065268 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-6575d7f87d-t4csd"] Mar 13 15:38:57 crc kubenswrapper[4907]: E0313 15:38:57.065651 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a43a8ffb-4f15-4524-b750-517442c1f561" containerName="placement-db-sync" Mar 13 15:38:57 crc kubenswrapper[4907]: I0313 15:38:57.065669 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="a43a8ffb-4f15-4524-b750-517442c1f561" containerName="placement-db-sync" Mar 13 15:38:57 crc kubenswrapper[4907]: I0313 15:38:57.065834 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="a43a8ffb-4f15-4524-b750-517442c1f561" containerName="placement-db-sync" Mar 13 15:38:57 crc kubenswrapper[4907]: I0313 15:38:57.066754 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-6575d7f87d-t4csd" Mar 13 15:38:57 crc kubenswrapper[4907]: I0313 15:38:57.070850 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Mar 13 15:38:57 crc kubenswrapper[4907]: I0313 15:38:57.071074 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Mar 13 15:38:57 crc kubenswrapper[4907]: I0313 15:38:57.071949 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-6c87x" Mar 13 15:38:57 crc kubenswrapper[4907]: I0313 15:38:57.087211 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-6575d7f87d-t4csd"] Mar 13 15:38:57 crc kubenswrapper[4907]: I0313 15:38:57.153618 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t2dbv\" (UniqueName: \"kubernetes.io/projected/12e05309-4563-4b78-a598-e4e8ced9cc8d-kube-api-access-t2dbv\") pod \"placement-6575d7f87d-t4csd\" (UID: \"12e05309-4563-4b78-a598-e4e8ced9cc8d\") " pod="openstack/placement-6575d7f87d-t4csd" Mar 13 15:38:57 crc kubenswrapper[4907]: I0313 15:38:57.153756 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/12e05309-4563-4b78-a598-e4e8ced9cc8d-scripts\") pod \"placement-6575d7f87d-t4csd\" (UID: \"12e05309-4563-4b78-a598-e4e8ced9cc8d\") " pod="openstack/placement-6575d7f87d-t4csd" Mar 13 15:38:57 crc kubenswrapper[4907]: I0313 15:38:57.153814 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12e05309-4563-4b78-a598-e4e8ced9cc8d-combined-ca-bundle\") pod \"placement-6575d7f87d-t4csd\" (UID: \"12e05309-4563-4b78-a598-e4e8ced9cc8d\") " pod="openstack/placement-6575d7f87d-t4csd" Mar 13 15:38:57 crc kubenswrapper[4907]: I0313 15:38:57.153849 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12e05309-4563-4b78-a598-e4e8ced9cc8d-config-data\") pod \"placement-6575d7f87d-t4csd\" (UID: \"12e05309-4563-4b78-a598-e4e8ced9cc8d\") " pod="openstack/placement-6575d7f87d-t4csd" Mar 13 15:38:57 crc kubenswrapper[4907]: I0313 15:38:57.153870 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/12e05309-4563-4b78-a598-e4e8ced9cc8d-logs\") pod \"placement-6575d7f87d-t4csd\" (UID: \"12e05309-4563-4b78-a598-e4e8ced9cc8d\") " pod="openstack/placement-6575d7f87d-t4csd" Mar 13 15:38:57 crc kubenswrapper[4907]: I0313 15:38:57.255091 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12e05309-4563-4b78-a598-e4e8ced9cc8d-combined-ca-bundle\") pod \"placement-6575d7f87d-t4csd\" (UID: \"12e05309-4563-4b78-a598-e4e8ced9cc8d\") " pod="openstack/placement-6575d7f87d-t4csd" Mar 13 15:38:57 crc kubenswrapper[4907]: I0313 15:38:57.255166 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12e05309-4563-4b78-a598-e4e8ced9cc8d-config-data\") pod \"placement-6575d7f87d-t4csd\" (UID: \"12e05309-4563-4b78-a598-e4e8ced9cc8d\") " pod="openstack/placement-6575d7f87d-t4csd" Mar 13 15:38:57 crc kubenswrapper[4907]: I0313 15:38:57.255192 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/12e05309-4563-4b78-a598-e4e8ced9cc8d-logs\") pod \"placement-6575d7f87d-t4csd\" (UID: \"12e05309-4563-4b78-a598-e4e8ced9cc8d\") " pod="openstack/placement-6575d7f87d-t4csd" Mar 13 15:38:57 crc kubenswrapper[4907]: I0313 15:38:57.255278 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t2dbv\" (UniqueName: \"kubernetes.io/projected/12e05309-4563-4b78-a598-e4e8ced9cc8d-kube-api-access-t2dbv\") pod \"placement-6575d7f87d-t4csd\" (UID: \"12e05309-4563-4b78-a598-e4e8ced9cc8d\") " pod="openstack/placement-6575d7f87d-t4csd" Mar 13 15:38:57 crc kubenswrapper[4907]: I0313 15:38:57.255326 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/12e05309-4563-4b78-a598-e4e8ced9cc8d-scripts\") pod \"placement-6575d7f87d-t4csd\" (UID: \"12e05309-4563-4b78-a598-e4e8ced9cc8d\") " pod="openstack/placement-6575d7f87d-t4csd" Mar 13 15:38:57 crc kubenswrapper[4907]: I0313 15:38:57.255714 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/12e05309-4563-4b78-a598-e4e8ced9cc8d-logs\") pod \"placement-6575d7f87d-t4csd\" (UID: \"12e05309-4563-4b78-a598-e4e8ced9cc8d\") " pod="openstack/placement-6575d7f87d-t4csd" Mar 13 15:38:57 crc kubenswrapper[4907]: I0313 15:38:57.259491 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12e05309-4563-4b78-a598-e4e8ced9cc8d-combined-ca-bundle\") pod \"placement-6575d7f87d-t4csd\" (UID: \"12e05309-4563-4b78-a598-e4e8ced9cc8d\") " pod="openstack/placement-6575d7f87d-t4csd" Mar 13 15:38:57 crc kubenswrapper[4907]: I0313 15:38:57.260822 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/12e05309-4563-4b78-a598-e4e8ced9cc8d-scripts\") pod \"placement-6575d7f87d-t4csd\" (UID: \"12e05309-4563-4b78-a598-e4e8ced9cc8d\") " pod="openstack/placement-6575d7f87d-t4csd" Mar 13 15:38:57 crc kubenswrapper[4907]: I0313 15:38:57.269955 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12e05309-4563-4b78-a598-e4e8ced9cc8d-config-data\") pod \"placement-6575d7f87d-t4csd\" (UID: \"12e05309-4563-4b78-a598-e4e8ced9cc8d\") " pod="openstack/placement-6575d7f87d-t4csd" Mar 13 15:38:57 crc kubenswrapper[4907]: I0313 15:38:57.270683 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t2dbv\" (UniqueName: \"kubernetes.io/projected/12e05309-4563-4b78-a598-e4e8ced9cc8d-kube-api-access-t2dbv\") pod \"placement-6575d7f87d-t4csd\" (UID: \"12e05309-4563-4b78-a598-e4e8ced9cc8d\") " pod="openstack/placement-6575d7f87d-t4csd" Mar 13 15:38:57 crc kubenswrapper[4907]: I0313 15:38:57.391707 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-6575d7f87d-t4csd" Mar 13 15:38:57 crc kubenswrapper[4907]: I0313 15:38:57.849214 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-6575d7f87d-t4csd"] Mar 13 15:38:57 crc kubenswrapper[4907]: W0313 15:38:57.855238 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod12e05309_4563_4b78_a598_e4e8ced9cc8d.slice/crio-3ccd101742905092f11d499e1eb4ca102c9c0edc448d3af36ebab529b04074f0 WatchSource:0}: Error finding container 3ccd101742905092f11d499e1eb4ca102c9c0edc448d3af36ebab529b04074f0: Status 404 returned error can't find the container with id 3ccd101742905092f11d499e1eb4ca102c9c0edc448d3af36ebab529b04074f0 Mar 13 15:38:58 crc kubenswrapper[4907]: I0313 15:38:58.007760 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6575d7f87d-t4csd" event={"ID":"12e05309-4563-4b78-a598-e4e8ced9cc8d","Type":"ContainerStarted","Data":"3ccd101742905092f11d499e1eb4ca102c9c0edc448d3af36ebab529b04074f0"} Mar 13 15:38:59 crc kubenswrapper[4907]: I0313 15:38:59.018050 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6575d7f87d-t4csd" event={"ID":"12e05309-4563-4b78-a598-e4e8ced9cc8d","Type":"ContainerStarted","Data":"c86e60fc008ccf1955bb3b2212ab51aaed4913a4393b406fcbdfd6f106d7909f"} Mar 13 15:38:59 crc kubenswrapper[4907]: I0313 15:38:59.019966 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6575d7f87d-t4csd" event={"ID":"12e05309-4563-4b78-a598-e4e8ced9cc8d","Type":"ContainerStarted","Data":"84c89fa87368ab0d7170f4175671afe4126cfa61e8fd73f4ce834716a16cef38"} Mar 13 15:38:59 crc kubenswrapper[4907]: I0313 15:38:59.019996 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-6575d7f87d-t4csd" Mar 13 15:38:59 crc kubenswrapper[4907]: I0313 15:38:59.037642 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-6575d7f87d-t4csd" podStartSLOduration=2.037616873 podStartE2EDuration="2.037616873s" podCreationTimestamp="2026-03-13 15:38:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:38:59.033059197 +0000 UTC m=+5637.932846926" watchObservedRunningTime="2026-03-13 15:38:59.037616873 +0000 UTC m=+5637.937404592" Mar 13 15:39:00 crc kubenswrapper[4907]: I0313 15:39:00.027711 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-6575d7f87d-t4csd" Mar 13 15:39:02 crc kubenswrapper[4907]: I0313 15:39:02.184095 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5dd7c58f95-579qq" Mar 13 15:39:02 crc kubenswrapper[4907]: I0313 15:39:02.243287 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-746ff5d579-xk972"] Mar 13 15:39:02 crc kubenswrapper[4907]: I0313 15:39:02.252590 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-746ff5d579-xk972" podUID="4cf4ac52-db94-4e0c-a990-c2c6c64f3167" containerName="dnsmasq-dns" containerID="cri-o://be908740b0d62c157a0b40660ddc6349bfde1dfa92e3ea3447a7f72e7e36459e" gracePeriod=10 Mar 13 15:39:02 crc kubenswrapper[4907]: I0313 15:39:02.691630 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-746ff5d579-xk972" Mar 13 15:39:02 crc kubenswrapper[4907]: I0313 15:39:02.767547 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4cf4ac52-db94-4e0c-a990-c2c6c64f3167-ovsdbserver-nb\") pod \"4cf4ac52-db94-4e0c-a990-c2c6c64f3167\" (UID: \"4cf4ac52-db94-4e0c-a990-c2c6c64f3167\") " Mar 13 15:39:02 crc kubenswrapper[4907]: I0313 15:39:02.767699 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4cf4ac52-db94-4e0c-a990-c2c6c64f3167-config\") pod \"4cf4ac52-db94-4e0c-a990-c2c6c64f3167\" (UID: \"4cf4ac52-db94-4e0c-a990-c2c6c64f3167\") " Mar 13 15:39:02 crc kubenswrapper[4907]: I0313 15:39:02.767791 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4cf4ac52-db94-4e0c-a990-c2c6c64f3167-ovsdbserver-sb\") pod \"4cf4ac52-db94-4e0c-a990-c2c6c64f3167\" (UID: \"4cf4ac52-db94-4e0c-a990-c2c6c64f3167\") " Mar 13 15:39:02 crc kubenswrapper[4907]: I0313 15:39:02.767896 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4cf4ac52-db94-4e0c-a990-c2c6c64f3167-dns-svc\") pod \"4cf4ac52-db94-4e0c-a990-c2c6c64f3167\" (UID: \"4cf4ac52-db94-4e0c-a990-c2c6c64f3167\") " Mar 13 15:39:02 crc kubenswrapper[4907]: I0313 15:39:02.767956 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4h66s\" (UniqueName: \"kubernetes.io/projected/4cf4ac52-db94-4e0c-a990-c2c6c64f3167-kube-api-access-4h66s\") pod \"4cf4ac52-db94-4e0c-a990-c2c6c64f3167\" (UID: \"4cf4ac52-db94-4e0c-a990-c2c6c64f3167\") " Mar 13 15:39:02 crc kubenswrapper[4907]: I0313 15:39:02.774316 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4cf4ac52-db94-4e0c-a990-c2c6c64f3167-kube-api-access-4h66s" (OuterVolumeSpecName: "kube-api-access-4h66s") pod "4cf4ac52-db94-4e0c-a990-c2c6c64f3167" (UID: "4cf4ac52-db94-4e0c-a990-c2c6c64f3167"). InnerVolumeSpecName "kube-api-access-4h66s". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:39:02 crc kubenswrapper[4907]: I0313 15:39:02.811426 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4cf4ac52-db94-4e0c-a990-c2c6c64f3167-config" (OuterVolumeSpecName: "config") pod "4cf4ac52-db94-4e0c-a990-c2c6c64f3167" (UID: "4cf4ac52-db94-4e0c-a990-c2c6c64f3167"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:39:02 crc kubenswrapper[4907]: I0313 15:39:02.812132 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4cf4ac52-db94-4e0c-a990-c2c6c64f3167-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "4cf4ac52-db94-4e0c-a990-c2c6c64f3167" (UID: "4cf4ac52-db94-4e0c-a990-c2c6c64f3167"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:39:02 crc kubenswrapper[4907]: I0313 15:39:02.815457 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4cf4ac52-db94-4e0c-a990-c2c6c64f3167-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "4cf4ac52-db94-4e0c-a990-c2c6c64f3167" (UID: "4cf4ac52-db94-4e0c-a990-c2c6c64f3167"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:39:02 crc kubenswrapper[4907]: I0313 15:39:02.818491 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4cf4ac52-db94-4e0c-a990-c2c6c64f3167-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "4cf4ac52-db94-4e0c-a990-c2c6c64f3167" (UID: "4cf4ac52-db94-4e0c-a990-c2c6c64f3167"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:39:02 crc kubenswrapper[4907]: I0313 15:39:02.869301 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4h66s\" (UniqueName: \"kubernetes.io/projected/4cf4ac52-db94-4e0c-a990-c2c6c64f3167-kube-api-access-4h66s\") on node \"crc\" DevicePath \"\"" Mar 13 15:39:02 crc kubenswrapper[4907]: I0313 15:39:02.869335 4907 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4cf4ac52-db94-4e0c-a990-c2c6c64f3167-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 13 15:39:02 crc kubenswrapper[4907]: I0313 15:39:02.869348 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4cf4ac52-db94-4e0c-a990-c2c6c64f3167-config\") on node \"crc\" DevicePath \"\"" Mar 13 15:39:02 crc kubenswrapper[4907]: I0313 15:39:02.869357 4907 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4cf4ac52-db94-4e0c-a990-c2c6c64f3167-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 13 15:39:02 crc kubenswrapper[4907]: I0313 15:39:02.869366 4907 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4cf4ac52-db94-4e0c-a990-c2c6c64f3167-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 13 15:39:03 crc kubenswrapper[4907]: I0313 15:39:03.066719 4907 generic.go:334] "Generic (PLEG): container finished" podID="4cf4ac52-db94-4e0c-a990-c2c6c64f3167" containerID="be908740b0d62c157a0b40660ddc6349bfde1dfa92e3ea3447a7f72e7e36459e" exitCode=0 Mar 13 15:39:03 crc kubenswrapper[4907]: I0313 15:39:03.066764 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-746ff5d579-xk972" Mar 13 15:39:03 crc kubenswrapper[4907]: I0313 15:39:03.066779 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-746ff5d579-xk972" event={"ID":"4cf4ac52-db94-4e0c-a990-c2c6c64f3167","Type":"ContainerDied","Data":"be908740b0d62c157a0b40660ddc6349bfde1dfa92e3ea3447a7f72e7e36459e"} Mar 13 15:39:03 crc kubenswrapper[4907]: I0313 15:39:03.066817 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-746ff5d579-xk972" event={"ID":"4cf4ac52-db94-4e0c-a990-c2c6c64f3167","Type":"ContainerDied","Data":"ab14195333d30dbd417719105a07ed83adbe0a377eeede3562757ec9ea3f4db9"} Mar 13 15:39:03 crc kubenswrapper[4907]: I0313 15:39:03.066846 4907 scope.go:117] "RemoveContainer" containerID="be908740b0d62c157a0b40660ddc6349bfde1dfa92e3ea3447a7f72e7e36459e" Mar 13 15:39:03 crc kubenswrapper[4907]: I0313 15:39:03.086196 4907 scope.go:117] "RemoveContainer" containerID="2d77589298c0e5cd48847476ea24e4e7c3f37cb59a1ce9b8eb1bd699e22f7f97" Mar 13 15:39:03 crc kubenswrapper[4907]: I0313 15:39:03.101047 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-746ff5d579-xk972"] Mar 13 15:39:03 crc kubenswrapper[4907]: I0313 15:39:03.107590 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-746ff5d579-xk972"] Mar 13 15:39:03 crc kubenswrapper[4907]: I0313 15:39:03.111397 4907 scope.go:117] "RemoveContainer" containerID="be908740b0d62c157a0b40660ddc6349bfde1dfa92e3ea3447a7f72e7e36459e" Mar 13 15:39:03 crc kubenswrapper[4907]: E0313 15:39:03.111863 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"be908740b0d62c157a0b40660ddc6349bfde1dfa92e3ea3447a7f72e7e36459e\": container with ID starting with be908740b0d62c157a0b40660ddc6349bfde1dfa92e3ea3447a7f72e7e36459e not found: ID does not exist" containerID="be908740b0d62c157a0b40660ddc6349bfde1dfa92e3ea3447a7f72e7e36459e" Mar 13 15:39:03 crc kubenswrapper[4907]: I0313 15:39:03.111927 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"be908740b0d62c157a0b40660ddc6349bfde1dfa92e3ea3447a7f72e7e36459e"} err="failed to get container status \"be908740b0d62c157a0b40660ddc6349bfde1dfa92e3ea3447a7f72e7e36459e\": rpc error: code = NotFound desc = could not find container \"be908740b0d62c157a0b40660ddc6349bfde1dfa92e3ea3447a7f72e7e36459e\": container with ID starting with be908740b0d62c157a0b40660ddc6349bfde1dfa92e3ea3447a7f72e7e36459e not found: ID does not exist" Mar 13 15:39:03 crc kubenswrapper[4907]: I0313 15:39:03.111958 4907 scope.go:117] "RemoveContainer" containerID="2d77589298c0e5cd48847476ea24e4e7c3f37cb59a1ce9b8eb1bd699e22f7f97" Mar 13 15:39:03 crc kubenswrapper[4907]: E0313 15:39:03.112305 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2d77589298c0e5cd48847476ea24e4e7c3f37cb59a1ce9b8eb1bd699e22f7f97\": container with ID starting with 2d77589298c0e5cd48847476ea24e4e7c3f37cb59a1ce9b8eb1bd699e22f7f97 not found: ID does not exist" containerID="2d77589298c0e5cd48847476ea24e4e7c3f37cb59a1ce9b8eb1bd699e22f7f97" Mar 13 15:39:03 crc kubenswrapper[4907]: I0313 15:39:03.112352 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2d77589298c0e5cd48847476ea24e4e7c3f37cb59a1ce9b8eb1bd699e22f7f97"} err="failed to get container status \"2d77589298c0e5cd48847476ea24e4e7c3f37cb59a1ce9b8eb1bd699e22f7f97\": rpc error: code = NotFound desc = could not find container \"2d77589298c0e5cd48847476ea24e4e7c3f37cb59a1ce9b8eb1bd699e22f7f97\": container with ID starting with 2d77589298c0e5cd48847476ea24e4e7c3f37cb59a1ce9b8eb1bd699e22f7f97 not found: ID does not exist" Mar 13 15:39:03 crc kubenswrapper[4907]: I0313 15:39:03.796367 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4cf4ac52-db94-4e0c-a990-c2c6c64f3167" path="/var/lib/kubelet/pods/4cf4ac52-db94-4e0c-a990-c2c6c64f3167/volumes" Mar 13 15:39:28 crc kubenswrapper[4907]: I0313 15:39:28.387644 4907 scope.go:117] "RemoveContainer" containerID="3c35c6992ec9b87a4cfcf6f80130463db4ea23212e37b3ceef2bd0131f1b10b1" Mar 13 15:39:28 crc kubenswrapper[4907]: I0313 15:39:28.392247 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-6575d7f87d-t4csd" Mar 13 15:39:28 crc kubenswrapper[4907]: I0313 15:39:28.405083 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-6575d7f87d-t4csd" Mar 13 15:39:28 crc kubenswrapper[4907]: I0313 15:39:28.406056 4907 scope.go:117] "RemoveContainer" containerID="2c2f32fc7b86119b10010b256cedf2bb75451fdf3804d5c1b2c61e62707d4b83" Mar 13 15:39:51 crc kubenswrapper[4907]: I0313 15:39:51.727654 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-5r4tb"] Mar 13 15:39:51 crc kubenswrapper[4907]: E0313 15:39:51.728570 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4cf4ac52-db94-4e0c-a990-c2c6c64f3167" containerName="dnsmasq-dns" Mar 13 15:39:51 crc kubenswrapper[4907]: I0313 15:39:51.728585 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="4cf4ac52-db94-4e0c-a990-c2c6c64f3167" containerName="dnsmasq-dns" Mar 13 15:39:51 crc kubenswrapper[4907]: E0313 15:39:51.728612 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4cf4ac52-db94-4e0c-a990-c2c6c64f3167" containerName="init" Mar 13 15:39:51 crc kubenswrapper[4907]: I0313 15:39:51.728638 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="4cf4ac52-db94-4e0c-a990-c2c6c64f3167" containerName="init" Mar 13 15:39:51 crc kubenswrapper[4907]: I0313 15:39:51.728836 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="4cf4ac52-db94-4e0c-a990-c2c6c64f3167" containerName="dnsmasq-dns" Mar 13 15:39:51 crc kubenswrapper[4907]: I0313 15:39:51.729543 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-5r4tb" Mar 13 15:39:51 crc kubenswrapper[4907]: I0313 15:39:51.738980 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-5r4tb"] Mar 13 15:39:51 crc kubenswrapper[4907]: I0313 15:39:51.827434 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-brbzj"] Mar 13 15:39:51 crc kubenswrapper[4907]: I0313 15:39:51.828773 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-brbzj" Mar 13 15:39:51 crc kubenswrapper[4907]: I0313 15:39:51.836960 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-brbzj"] Mar 13 15:39:51 crc kubenswrapper[4907]: I0313 15:39:51.855358 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b4jdc\" (UniqueName: \"kubernetes.io/projected/e5aab76f-4bcc-47be-989d-7d9d6ecf371b-kube-api-access-b4jdc\") pod \"nova-api-db-create-5r4tb\" (UID: \"e5aab76f-4bcc-47be-989d-7d9d6ecf371b\") " pod="openstack/nova-api-db-create-5r4tb" Mar 13 15:39:51 crc kubenswrapper[4907]: I0313 15:39:51.855416 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e5aab76f-4bcc-47be-989d-7d9d6ecf371b-operator-scripts\") pod \"nova-api-db-create-5r4tb\" (UID: \"e5aab76f-4bcc-47be-989d-7d9d6ecf371b\") " pod="openstack/nova-api-db-create-5r4tb" Mar 13 15:39:51 crc kubenswrapper[4907]: I0313 15:39:51.951654 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-7djpm"] Mar 13 15:39:51 crc kubenswrapper[4907]: I0313 15:39:51.953076 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-7djpm" Mar 13 15:39:51 crc kubenswrapper[4907]: I0313 15:39:51.956752 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b4jdc\" (UniqueName: \"kubernetes.io/projected/e5aab76f-4bcc-47be-989d-7d9d6ecf371b-kube-api-access-b4jdc\") pod \"nova-api-db-create-5r4tb\" (UID: \"e5aab76f-4bcc-47be-989d-7d9d6ecf371b\") " pod="openstack/nova-api-db-create-5r4tb" Mar 13 15:39:51 crc kubenswrapper[4907]: I0313 15:39:51.956803 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e5aab76f-4bcc-47be-989d-7d9d6ecf371b-operator-scripts\") pod \"nova-api-db-create-5r4tb\" (UID: \"e5aab76f-4bcc-47be-989d-7d9d6ecf371b\") " pod="openstack/nova-api-db-create-5r4tb" Mar 13 15:39:51 crc kubenswrapper[4907]: I0313 15:39:51.956835 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vrtdp\" (UniqueName: \"kubernetes.io/projected/18fa9304-50a2-45cf-b983-9d459f0dcb5f-kube-api-access-vrtdp\") pod \"nova-cell0-db-create-brbzj\" (UID: \"18fa9304-50a2-45cf-b983-9d459f0dcb5f\") " pod="openstack/nova-cell0-db-create-brbzj" Mar 13 15:39:51 crc kubenswrapper[4907]: I0313 15:39:51.956935 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/18fa9304-50a2-45cf-b983-9d459f0dcb5f-operator-scripts\") pod \"nova-cell0-db-create-brbzj\" (UID: \"18fa9304-50a2-45cf-b983-9d459f0dcb5f\") " pod="openstack/nova-cell0-db-create-brbzj" Mar 13 15:39:51 crc kubenswrapper[4907]: I0313 15:39:51.957530 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e5aab76f-4bcc-47be-989d-7d9d6ecf371b-operator-scripts\") pod \"nova-api-db-create-5r4tb\" (UID: \"e5aab76f-4bcc-47be-989d-7d9d6ecf371b\") " pod="openstack/nova-api-db-create-5r4tb" Mar 13 15:39:51 crc kubenswrapper[4907]: I0313 15:39:51.959482 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-62d0-account-create-update-vf8sr"] Mar 13 15:39:51 crc kubenswrapper[4907]: I0313 15:39:51.960541 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-62d0-account-create-update-vf8sr" Mar 13 15:39:51 crc kubenswrapper[4907]: I0313 15:39:51.962175 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Mar 13 15:39:51 crc kubenswrapper[4907]: I0313 15:39:51.970505 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-62d0-account-create-update-vf8sr"] Mar 13 15:39:51 crc kubenswrapper[4907]: I0313 15:39:51.983793 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-7djpm"] Mar 13 15:39:52 crc kubenswrapper[4907]: I0313 15:39:52.012514 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b4jdc\" (UniqueName: \"kubernetes.io/projected/e5aab76f-4bcc-47be-989d-7d9d6ecf371b-kube-api-access-b4jdc\") pod \"nova-api-db-create-5r4tb\" (UID: \"e5aab76f-4bcc-47be-989d-7d9d6ecf371b\") " pod="openstack/nova-api-db-create-5r4tb" Mar 13 15:39:52 crc kubenswrapper[4907]: I0313 15:39:52.051240 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-5r4tb" Mar 13 15:39:52 crc kubenswrapper[4907]: I0313 15:39:52.065079 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/18fa9304-50a2-45cf-b983-9d459f0dcb5f-operator-scripts\") pod \"nova-cell0-db-create-brbzj\" (UID: \"18fa9304-50a2-45cf-b983-9d459f0dcb5f\") " pod="openstack/nova-cell0-db-create-brbzj" Mar 13 15:39:52 crc kubenswrapper[4907]: I0313 15:39:52.065142 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lxptq\" (UniqueName: \"kubernetes.io/projected/c2c52fb5-c4ba-4c24-95af-ce75cd609126-kube-api-access-lxptq\") pod \"nova-cell1-db-create-7djpm\" (UID: \"c2c52fb5-c4ba-4c24-95af-ce75cd609126\") " pod="openstack/nova-cell1-db-create-7djpm" Mar 13 15:39:52 crc kubenswrapper[4907]: I0313 15:39:52.065174 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/94ba887c-f88a-48e7-b4bc-2f171d21fa47-operator-scripts\") pod \"nova-api-62d0-account-create-update-vf8sr\" (UID: \"94ba887c-f88a-48e7-b4bc-2f171d21fa47\") " pod="openstack/nova-api-62d0-account-create-update-vf8sr" Mar 13 15:39:52 crc kubenswrapper[4907]: I0313 15:39:52.065233 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c2c52fb5-c4ba-4c24-95af-ce75cd609126-operator-scripts\") pod \"nova-cell1-db-create-7djpm\" (UID: \"c2c52fb5-c4ba-4c24-95af-ce75cd609126\") " pod="openstack/nova-cell1-db-create-7djpm" Mar 13 15:39:52 crc kubenswrapper[4907]: I0313 15:39:52.065260 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vrtdp\" (UniqueName: \"kubernetes.io/projected/18fa9304-50a2-45cf-b983-9d459f0dcb5f-kube-api-access-vrtdp\") pod \"nova-cell0-db-create-brbzj\" (UID: \"18fa9304-50a2-45cf-b983-9d459f0dcb5f\") " pod="openstack/nova-cell0-db-create-brbzj" Mar 13 15:39:52 crc kubenswrapper[4907]: I0313 15:39:52.065308 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-64c2b\" (UniqueName: \"kubernetes.io/projected/94ba887c-f88a-48e7-b4bc-2f171d21fa47-kube-api-access-64c2b\") pod \"nova-api-62d0-account-create-update-vf8sr\" (UID: \"94ba887c-f88a-48e7-b4bc-2f171d21fa47\") " pod="openstack/nova-api-62d0-account-create-update-vf8sr" Mar 13 15:39:52 crc kubenswrapper[4907]: I0313 15:39:52.066343 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/18fa9304-50a2-45cf-b983-9d459f0dcb5f-operator-scripts\") pod \"nova-cell0-db-create-brbzj\" (UID: \"18fa9304-50a2-45cf-b983-9d459f0dcb5f\") " pod="openstack/nova-cell0-db-create-brbzj" Mar 13 15:39:52 crc kubenswrapper[4907]: I0313 15:39:52.110801 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vrtdp\" (UniqueName: \"kubernetes.io/projected/18fa9304-50a2-45cf-b983-9d459f0dcb5f-kube-api-access-vrtdp\") pod \"nova-cell0-db-create-brbzj\" (UID: \"18fa9304-50a2-45cf-b983-9d459f0dcb5f\") " pod="openstack/nova-cell0-db-create-brbzj" Mar 13 15:39:52 crc kubenswrapper[4907]: I0313 15:39:52.144621 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-brbzj" Mar 13 15:39:52 crc kubenswrapper[4907]: I0313 15:39:52.157743 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-356d-account-create-update-kb8lr"] Mar 13 15:39:52 crc kubenswrapper[4907]: I0313 15:39:52.159852 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-356d-account-create-update-kb8lr" Mar 13 15:39:52 crc kubenswrapper[4907]: I0313 15:39:52.165204 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Mar 13 15:39:52 crc kubenswrapper[4907]: I0313 15:39:52.166614 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-64c2b\" (UniqueName: \"kubernetes.io/projected/94ba887c-f88a-48e7-b4bc-2f171d21fa47-kube-api-access-64c2b\") pod \"nova-api-62d0-account-create-update-vf8sr\" (UID: \"94ba887c-f88a-48e7-b4bc-2f171d21fa47\") " pod="openstack/nova-api-62d0-account-create-update-vf8sr" Mar 13 15:39:52 crc kubenswrapper[4907]: I0313 15:39:52.166720 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lxptq\" (UniqueName: \"kubernetes.io/projected/c2c52fb5-c4ba-4c24-95af-ce75cd609126-kube-api-access-lxptq\") pod \"nova-cell1-db-create-7djpm\" (UID: \"c2c52fb5-c4ba-4c24-95af-ce75cd609126\") " pod="openstack/nova-cell1-db-create-7djpm" Mar 13 15:39:52 crc kubenswrapper[4907]: I0313 15:39:52.166745 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/94ba887c-f88a-48e7-b4bc-2f171d21fa47-operator-scripts\") pod \"nova-api-62d0-account-create-update-vf8sr\" (UID: \"94ba887c-f88a-48e7-b4bc-2f171d21fa47\") " pod="openstack/nova-api-62d0-account-create-update-vf8sr" Mar 13 15:39:52 crc kubenswrapper[4907]: I0313 15:39:52.166793 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c2c52fb5-c4ba-4c24-95af-ce75cd609126-operator-scripts\") pod \"nova-cell1-db-create-7djpm\" (UID: \"c2c52fb5-c4ba-4c24-95af-ce75cd609126\") " pod="openstack/nova-cell1-db-create-7djpm" Mar 13 15:39:52 crc kubenswrapper[4907]: I0313 15:39:52.167481 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c2c52fb5-c4ba-4c24-95af-ce75cd609126-operator-scripts\") pod \"nova-cell1-db-create-7djpm\" (UID: \"c2c52fb5-c4ba-4c24-95af-ce75cd609126\") " pod="openstack/nova-cell1-db-create-7djpm" Mar 13 15:39:52 crc kubenswrapper[4907]: I0313 15:39:52.168315 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/94ba887c-f88a-48e7-b4bc-2f171d21fa47-operator-scripts\") pod \"nova-api-62d0-account-create-update-vf8sr\" (UID: \"94ba887c-f88a-48e7-b4bc-2f171d21fa47\") " pod="openstack/nova-api-62d0-account-create-update-vf8sr" Mar 13 15:39:52 crc kubenswrapper[4907]: I0313 15:39:52.169684 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-356d-account-create-update-kb8lr"] Mar 13 15:39:52 crc kubenswrapper[4907]: I0313 15:39:52.192701 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lxptq\" (UniqueName: \"kubernetes.io/projected/c2c52fb5-c4ba-4c24-95af-ce75cd609126-kube-api-access-lxptq\") pod \"nova-cell1-db-create-7djpm\" (UID: \"c2c52fb5-c4ba-4c24-95af-ce75cd609126\") " pod="openstack/nova-cell1-db-create-7djpm" Mar 13 15:39:52 crc kubenswrapper[4907]: I0313 15:39:52.193190 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-64c2b\" (UniqueName: \"kubernetes.io/projected/94ba887c-f88a-48e7-b4bc-2f171d21fa47-kube-api-access-64c2b\") pod \"nova-api-62d0-account-create-update-vf8sr\" (UID: \"94ba887c-f88a-48e7-b4bc-2f171d21fa47\") " pod="openstack/nova-api-62d0-account-create-update-vf8sr" Mar 13 15:39:52 crc kubenswrapper[4907]: I0313 15:39:52.268523 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b66093c1-e494-4975-b8d1-217844060204-operator-scripts\") pod \"nova-cell0-356d-account-create-update-kb8lr\" (UID: \"b66093c1-e494-4975-b8d1-217844060204\") " pod="openstack/nova-cell0-356d-account-create-update-kb8lr" Mar 13 15:39:52 crc kubenswrapper[4907]: I0313 15:39:52.268583 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wcjg9\" (UniqueName: \"kubernetes.io/projected/b66093c1-e494-4975-b8d1-217844060204-kube-api-access-wcjg9\") pod \"nova-cell0-356d-account-create-update-kb8lr\" (UID: \"b66093c1-e494-4975-b8d1-217844060204\") " pod="openstack/nova-cell0-356d-account-create-update-kb8lr" Mar 13 15:39:52 crc kubenswrapper[4907]: I0313 15:39:52.271360 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-7djpm" Mar 13 15:39:52 crc kubenswrapper[4907]: I0313 15:39:52.284448 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-62d0-account-create-update-vf8sr" Mar 13 15:39:52 crc kubenswrapper[4907]: I0313 15:39:52.345469 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-5a99-account-create-update-nkg4x"] Mar 13 15:39:52 crc kubenswrapper[4907]: I0313 15:39:52.348783 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-5a99-account-create-update-nkg4x" Mar 13 15:39:52 crc kubenswrapper[4907]: I0313 15:39:52.351010 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Mar 13 15:39:52 crc kubenswrapper[4907]: I0313 15:39:52.354759 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-5a99-account-create-update-nkg4x"] Mar 13 15:39:52 crc kubenswrapper[4907]: I0313 15:39:52.370231 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wcjg9\" (UniqueName: \"kubernetes.io/projected/b66093c1-e494-4975-b8d1-217844060204-kube-api-access-wcjg9\") pod \"nova-cell0-356d-account-create-update-kb8lr\" (UID: \"b66093c1-e494-4975-b8d1-217844060204\") " pod="openstack/nova-cell0-356d-account-create-update-kb8lr" Mar 13 15:39:52 crc kubenswrapper[4907]: I0313 15:39:52.370489 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b66093c1-e494-4975-b8d1-217844060204-operator-scripts\") pod \"nova-cell0-356d-account-create-update-kb8lr\" (UID: \"b66093c1-e494-4975-b8d1-217844060204\") " pod="openstack/nova-cell0-356d-account-create-update-kb8lr" Mar 13 15:39:52 crc kubenswrapper[4907]: I0313 15:39:52.371500 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b66093c1-e494-4975-b8d1-217844060204-operator-scripts\") pod \"nova-cell0-356d-account-create-update-kb8lr\" (UID: \"b66093c1-e494-4975-b8d1-217844060204\") " pod="openstack/nova-cell0-356d-account-create-update-kb8lr" Mar 13 15:39:52 crc kubenswrapper[4907]: I0313 15:39:52.386557 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wcjg9\" (UniqueName: \"kubernetes.io/projected/b66093c1-e494-4975-b8d1-217844060204-kube-api-access-wcjg9\") pod \"nova-cell0-356d-account-create-update-kb8lr\" (UID: \"b66093c1-e494-4975-b8d1-217844060204\") " pod="openstack/nova-cell0-356d-account-create-update-kb8lr" Mar 13 15:39:52 crc kubenswrapper[4907]: I0313 15:39:52.473378 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/40763271-36d9-4bd2-8ae8-82140648fcf4-operator-scripts\") pod \"nova-cell1-5a99-account-create-update-nkg4x\" (UID: \"40763271-36d9-4bd2-8ae8-82140648fcf4\") " pod="openstack/nova-cell1-5a99-account-create-update-nkg4x" Mar 13 15:39:52 crc kubenswrapper[4907]: I0313 15:39:52.473491 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pkzvl\" (UniqueName: \"kubernetes.io/projected/40763271-36d9-4bd2-8ae8-82140648fcf4-kube-api-access-pkzvl\") pod \"nova-cell1-5a99-account-create-update-nkg4x\" (UID: \"40763271-36d9-4bd2-8ae8-82140648fcf4\") " pod="openstack/nova-cell1-5a99-account-create-update-nkg4x" Mar 13 15:39:52 crc kubenswrapper[4907]: I0313 15:39:52.574321 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-356d-account-create-update-kb8lr" Mar 13 15:39:52 crc kubenswrapper[4907]: I0313 15:39:52.575781 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/40763271-36d9-4bd2-8ae8-82140648fcf4-operator-scripts\") pod \"nova-cell1-5a99-account-create-update-nkg4x\" (UID: \"40763271-36d9-4bd2-8ae8-82140648fcf4\") " pod="openstack/nova-cell1-5a99-account-create-update-nkg4x" Mar 13 15:39:52 crc kubenswrapper[4907]: I0313 15:39:52.575972 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pkzvl\" (UniqueName: \"kubernetes.io/projected/40763271-36d9-4bd2-8ae8-82140648fcf4-kube-api-access-pkzvl\") pod \"nova-cell1-5a99-account-create-update-nkg4x\" (UID: \"40763271-36d9-4bd2-8ae8-82140648fcf4\") " pod="openstack/nova-cell1-5a99-account-create-update-nkg4x" Mar 13 15:39:52 crc kubenswrapper[4907]: I0313 15:39:52.577260 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/40763271-36d9-4bd2-8ae8-82140648fcf4-operator-scripts\") pod \"nova-cell1-5a99-account-create-update-nkg4x\" (UID: \"40763271-36d9-4bd2-8ae8-82140648fcf4\") " pod="openstack/nova-cell1-5a99-account-create-update-nkg4x" Mar 13 15:39:52 crc kubenswrapper[4907]: I0313 15:39:52.592897 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pkzvl\" (UniqueName: \"kubernetes.io/projected/40763271-36d9-4bd2-8ae8-82140648fcf4-kube-api-access-pkzvl\") pod \"nova-cell1-5a99-account-create-update-nkg4x\" (UID: \"40763271-36d9-4bd2-8ae8-82140648fcf4\") " pod="openstack/nova-cell1-5a99-account-create-update-nkg4x" Mar 13 15:39:52 crc kubenswrapper[4907]: I0313 15:39:52.629768 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-5r4tb"] Mar 13 15:39:52 crc kubenswrapper[4907]: I0313 15:39:52.685085 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-5a99-account-create-update-nkg4x" Mar 13 15:39:52 crc kubenswrapper[4907]: W0313 15:39:52.734997 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod18fa9304_50a2_45cf_b983_9d459f0dcb5f.slice/crio-e2b380c63ddd685de31e5a747984baea0335628ccbc6b3f1af44debb13f13653 WatchSource:0}: Error finding container e2b380c63ddd685de31e5a747984baea0335628ccbc6b3f1af44debb13f13653: Status 404 returned error can't find the container with id e2b380c63ddd685de31e5a747984baea0335628ccbc6b3f1af44debb13f13653 Mar 13 15:39:52 crc kubenswrapper[4907]: I0313 15:39:52.737407 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-brbzj"] Mar 13 15:39:52 crc kubenswrapper[4907]: I0313 15:39:52.847360 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-62d0-account-create-update-vf8sr"] Mar 13 15:39:52 crc kubenswrapper[4907]: W0313 15:39:52.859755 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc2c52fb5_c4ba_4c24_95af_ce75cd609126.slice/crio-0e23da64c584ff8030a07246c197bc4df16cf45b1b7b65a2a1e9a6373f844c90 WatchSource:0}: Error finding container 0e23da64c584ff8030a07246c197bc4df16cf45b1b7b65a2a1e9a6373f844c90: Status 404 returned error can't find the container with id 0e23da64c584ff8030a07246c197bc4df16cf45b1b7b65a2a1e9a6373f844c90 Mar 13 15:39:52 crc kubenswrapper[4907]: I0313 15:39:52.869758 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-7djpm"] Mar 13 15:39:53 crc kubenswrapper[4907]: I0313 15:39:53.089899 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-356d-account-create-update-kb8lr"] Mar 13 15:39:53 crc kubenswrapper[4907]: W0313 15:39:53.131412 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb66093c1_e494_4975_b8d1_217844060204.slice/crio-0bf11b5c16f026ec4133b29e8d2e6b391b0da007fcd4acde47500c6e6fdc42e7 WatchSource:0}: Error finding container 0bf11b5c16f026ec4133b29e8d2e6b391b0da007fcd4acde47500c6e6fdc42e7: Status 404 returned error can't find the container with id 0bf11b5c16f026ec4133b29e8d2e6b391b0da007fcd4acde47500c6e6fdc42e7 Mar 13 15:39:53 crc kubenswrapper[4907]: I0313 15:39:53.203682 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-5a99-account-create-update-nkg4x"] Mar 13 15:39:53 crc kubenswrapper[4907]: W0313 15:39:53.225181 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod40763271_36d9_4bd2_8ae8_82140648fcf4.slice/crio-543db5418a330957dba2de6b68ab5a198dc372131f880d599abcd7de199ca1e8 WatchSource:0}: Error finding container 543db5418a330957dba2de6b68ab5a198dc372131f880d599abcd7de199ca1e8: Status 404 returned error can't find the container with id 543db5418a330957dba2de6b68ab5a198dc372131f880d599abcd7de199ca1e8 Mar 13 15:39:53 crc kubenswrapper[4907]: I0313 15:39:53.508675 4907 generic.go:334] "Generic (PLEG): container finished" podID="18fa9304-50a2-45cf-b983-9d459f0dcb5f" containerID="f18dc1660972c456737f08b933ea252844a2bf45813b5a1dae54e428ebc166a5" exitCode=0 Mar 13 15:39:53 crc kubenswrapper[4907]: I0313 15:39:53.508742 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-brbzj" event={"ID":"18fa9304-50a2-45cf-b983-9d459f0dcb5f","Type":"ContainerDied","Data":"f18dc1660972c456737f08b933ea252844a2bf45813b5a1dae54e428ebc166a5"} Mar 13 15:39:53 crc kubenswrapper[4907]: I0313 15:39:53.508769 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-brbzj" event={"ID":"18fa9304-50a2-45cf-b983-9d459f0dcb5f","Type":"ContainerStarted","Data":"e2b380c63ddd685de31e5a747984baea0335628ccbc6b3f1af44debb13f13653"} Mar 13 15:39:53 crc kubenswrapper[4907]: I0313 15:39:53.510938 4907 generic.go:334] "Generic (PLEG): container finished" podID="94ba887c-f88a-48e7-b4bc-2f171d21fa47" containerID="c18542c3b44a6f6eb28a26dfdcfcff6b4ceb92d4f5c1a2a581f9880ab136aa90" exitCode=0 Mar 13 15:39:53 crc kubenswrapper[4907]: I0313 15:39:53.510988 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-62d0-account-create-update-vf8sr" event={"ID":"94ba887c-f88a-48e7-b4bc-2f171d21fa47","Type":"ContainerDied","Data":"c18542c3b44a6f6eb28a26dfdcfcff6b4ceb92d4f5c1a2a581f9880ab136aa90"} Mar 13 15:39:53 crc kubenswrapper[4907]: I0313 15:39:53.511004 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-62d0-account-create-update-vf8sr" event={"ID":"94ba887c-f88a-48e7-b4bc-2f171d21fa47","Type":"ContainerStarted","Data":"4a40897775cd1b3c4b3543b2070f41a576ca4e865bdb987c81c613efd9443cc6"} Mar 13 15:39:53 crc kubenswrapper[4907]: I0313 15:39:53.512773 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-356d-account-create-update-kb8lr" event={"ID":"b66093c1-e494-4975-b8d1-217844060204","Type":"ContainerStarted","Data":"0bf11b5c16f026ec4133b29e8d2e6b391b0da007fcd4acde47500c6e6fdc42e7"} Mar 13 15:39:53 crc kubenswrapper[4907]: I0313 15:39:53.514550 4907 generic.go:334] "Generic (PLEG): container finished" podID="e5aab76f-4bcc-47be-989d-7d9d6ecf371b" containerID="d3e4c590d97414f576974cec65543e931ac60a3bcf12387dbe9df0a1a39e2c28" exitCode=0 Mar 13 15:39:53 crc kubenswrapper[4907]: I0313 15:39:53.514602 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-5r4tb" event={"ID":"e5aab76f-4bcc-47be-989d-7d9d6ecf371b","Type":"ContainerDied","Data":"d3e4c590d97414f576974cec65543e931ac60a3bcf12387dbe9df0a1a39e2c28"} Mar 13 15:39:53 crc kubenswrapper[4907]: I0313 15:39:53.514621 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-5r4tb" event={"ID":"e5aab76f-4bcc-47be-989d-7d9d6ecf371b","Type":"ContainerStarted","Data":"dd3f82f5adfcd093fbfe4c95a182789207aea1df8bcb504fdcb2fe97dbfbdb96"} Mar 13 15:39:53 crc kubenswrapper[4907]: I0313 15:39:53.517154 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-5a99-account-create-update-nkg4x" event={"ID":"40763271-36d9-4bd2-8ae8-82140648fcf4","Type":"ContainerStarted","Data":"543db5418a330957dba2de6b68ab5a198dc372131f880d599abcd7de199ca1e8"} Mar 13 15:39:53 crc kubenswrapper[4907]: I0313 15:39:53.519103 4907 generic.go:334] "Generic (PLEG): container finished" podID="c2c52fb5-c4ba-4c24-95af-ce75cd609126" containerID="9d15efc6444fb5cd991e1e807468f40b653afac37426c92c9f3e8218e37197cf" exitCode=0 Mar 13 15:39:53 crc kubenswrapper[4907]: I0313 15:39:53.519149 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-7djpm" event={"ID":"c2c52fb5-c4ba-4c24-95af-ce75cd609126","Type":"ContainerDied","Data":"9d15efc6444fb5cd991e1e807468f40b653afac37426c92c9f3e8218e37197cf"} Mar 13 15:39:53 crc kubenswrapper[4907]: I0313 15:39:53.519170 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-7djpm" event={"ID":"c2c52fb5-c4ba-4c24-95af-ce75cd609126","Type":"ContainerStarted","Data":"0e23da64c584ff8030a07246c197bc4df16cf45b1b7b65a2a1e9a6373f844c90"} Mar 13 15:39:53 crc kubenswrapper[4907]: E0313 15:39:53.912491 4907 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb66093c1_e494_4975_b8d1_217844060204.slice/crio-conmon-c3aa022d7c9308a16e2c36ea9708c7b13f760c092c5e791147b871724f8b2c96.scope\": RecentStats: unable to find data in memory cache]" Mar 13 15:39:54 crc kubenswrapper[4907]: I0313 15:39:54.531553 4907 generic.go:334] "Generic (PLEG): container finished" podID="40763271-36d9-4bd2-8ae8-82140648fcf4" containerID="0948541af6cc99af21bd6ee88e49185bc604c42dbf31f8661316b429337a2c7f" exitCode=0 Mar 13 15:39:54 crc kubenswrapper[4907]: I0313 15:39:54.531649 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-5a99-account-create-update-nkg4x" event={"ID":"40763271-36d9-4bd2-8ae8-82140648fcf4","Type":"ContainerDied","Data":"0948541af6cc99af21bd6ee88e49185bc604c42dbf31f8661316b429337a2c7f"} Mar 13 15:39:54 crc kubenswrapper[4907]: I0313 15:39:54.535573 4907 generic.go:334] "Generic (PLEG): container finished" podID="b66093c1-e494-4975-b8d1-217844060204" containerID="c3aa022d7c9308a16e2c36ea9708c7b13f760c092c5e791147b871724f8b2c96" exitCode=0 Mar 13 15:39:54 crc kubenswrapper[4907]: I0313 15:39:54.535643 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-356d-account-create-update-kb8lr" event={"ID":"b66093c1-e494-4975-b8d1-217844060204","Type":"ContainerDied","Data":"c3aa022d7c9308a16e2c36ea9708c7b13f760c092c5e791147b871724f8b2c96"} Mar 13 15:39:54 crc kubenswrapper[4907]: I0313 15:39:54.874190 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-5r4tb" Mar 13 15:39:54 crc kubenswrapper[4907]: I0313 15:39:54.947123 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b4jdc\" (UniqueName: \"kubernetes.io/projected/e5aab76f-4bcc-47be-989d-7d9d6ecf371b-kube-api-access-b4jdc\") pod \"e5aab76f-4bcc-47be-989d-7d9d6ecf371b\" (UID: \"e5aab76f-4bcc-47be-989d-7d9d6ecf371b\") " Mar 13 15:39:54 crc kubenswrapper[4907]: I0313 15:39:54.947218 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e5aab76f-4bcc-47be-989d-7d9d6ecf371b-operator-scripts\") pod \"e5aab76f-4bcc-47be-989d-7d9d6ecf371b\" (UID: \"e5aab76f-4bcc-47be-989d-7d9d6ecf371b\") " Mar 13 15:39:54 crc kubenswrapper[4907]: I0313 15:39:54.947987 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e5aab76f-4bcc-47be-989d-7d9d6ecf371b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e5aab76f-4bcc-47be-989d-7d9d6ecf371b" (UID: "e5aab76f-4bcc-47be-989d-7d9d6ecf371b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:39:54 crc kubenswrapper[4907]: I0313 15:39:54.949488 4907 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e5aab76f-4bcc-47be-989d-7d9d6ecf371b-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 15:39:54 crc kubenswrapper[4907]: I0313 15:39:54.953155 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e5aab76f-4bcc-47be-989d-7d9d6ecf371b-kube-api-access-b4jdc" (OuterVolumeSpecName: "kube-api-access-b4jdc") pod "e5aab76f-4bcc-47be-989d-7d9d6ecf371b" (UID: "e5aab76f-4bcc-47be-989d-7d9d6ecf371b"). InnerVolumeSpecName "kube-api-access-b4jdc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:39:55 crc kubenswrapper[4907]: I0313 15:39:55.041663 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-62d0-account-create-update-vf8sr" Mar 13 15:39:55 crc kubenswrapper[4907]: I0313 15:39:55.046438 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-7djpm" Mar 13 15:39:55 crc kubenswrapper[4907]: I0313 15:39:55.051275 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-brbzj" Mar 13 15:39:55 crc kubenswrapper[4907]: I0313 15:39:55.051802 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b4jdc\" (UniqueName: \"kubernetes.io/projected/e5aab76f-4bcc-47be-989d-7d9d6ecf371b-kube-api-access-b4jdc\") on node \"crc\" DevicePath \"\"" Mar 13 15:39:55 crc kubenswrapper[4907]: I0313 15:39:55.152778 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c2c52fb5-c4ba-4c24-95af-ce75cd609126-operator-scripts\") pod \"c2c52fb5-c4ba-4c24-95af-ce75cd609126\" (UID: \"c2c52fb5-c4ba-4c24-95af-ce75cd609126\") " Mar 13 15:39:55 crc kubenswrapper[4907]: I0313 15:39:55.152928 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-64c2b\" (UniqueName: \"kubernetes.io/projected/94ba887c-f88a-48e7-b4bc-2f171d21fa47-kube-api-access-64c2b\") pod \"94ba887c-f88a-48e7-b4bc-2f171d21fa47\" (UID: \"94ba887c-f88a-48e7-b4bc-2f171d21fa47\") " Mar 13 15:39:55 crc kubenswrapper[4907]: I0313 15:39:55.152955 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lxptq\" (UniqueName: \"kubernetes.io/projected/c2c52fb5-c4ba-4c24-95af-ce75cd609126-kube-api-access-lxptq\") pod \"c2c52fb5-c4ba-4c24-95af-ce75cd609126\" (UID: \"c2c52fb5-c4ba-4c24-95af-ce75cd609126\") " Mar 13 15:39:55 crc kubenswrapper[4907]: I0313 15:39:55.153002 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/18fa9304-50a2-45cf-b983-9d459f0dcb5f-operator-scripts\") pod \"18fa9304-50a2-45cf-b983-9d459f0dcb5f\" (UID: \"18fa9304-50a2-45cf-b983-9d459f0dcb5f\") " Mar 13 15:39:55 crc kubenswrapper[4907]: I0313 15:39:55.153063 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/94ba887c-f88a-48e7-b4bc-2f171d21fa47-operator-scripts\") pod \"94ba887c-f88a-48e7-b4bc-2f171d21fa47\" (UID: \"94ba887c-f88a-48e7-b4bc-2f171d21fa47\") " Mar 13 15:39:55 crc kubenswrapper[4907]: I0313 15:39:55.153111 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vrtdp\" (UniqueName: \"kubernetes.io/projected/18fa9304-50a2-45cf-b983-9d459f0dcb5f-kube-api-access-vrtdp\") pod \"18fa9304-50a2-45cf-b983-9d459f0dcb5f\" (UID: \"18fa9304-50a2-45cf-b983-9d459f0dcb5f\") " Mar 13 15:39:55 crc kubenswrapper[4907]: I0313 15:39:55.153385 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c2c52fb5-c4ba-4c24-95af-ce75cd609126-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c2c52fb5-c4ba-4c24-95af-ce75cd609126" (UID: "c2c52fb5-c4ba-4c24-95af-ce75cd609126"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:39:55 crc kubenswrapper[4907]: I0313 15:39:55.153455 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/18fa9304-50a2-45cf-b983-9d459f0dcb5f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "18fa9304-50a2-45cf-b983-9d459f0dcb5f" (UID: "18fa9304-50a2-45cf-b983-9d459f0dcb5f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:39:55 crc kubenswrapper[4907]: I0313 15:39:55.153478 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/94ba887c-f88a-48e7-b4bc-2f171d21fa47-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "94ba887c-f88a-48e7-b4bc-2f171d21fa47" (UID: "94ba887c-f88a-48e7-b4bc-2f171d21fa47"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:39:55 crc kubenswrapper[4907]: I0313 15:39:55.153895 4907 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c2c52fb5-c4ba-4c24-95af-ce75cd609126-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 15:39:55 crc kubenswrapper[4907]: I0313 15:39:55.153914 4907 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/18fa9304-50a2-45cf-b983-9d459f0dcb5f-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 15:39:55 crc kubenswrapper[4907]: I0313 15:39:55.153923 4907 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/94ba887c-f88a-48e7-b4bc-2f171d21fa47-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 15:39:55 crc kubenswrapper[4907]: I0313 15:39:55.156066 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/18fa9304-50a2-45cf-b983-9d459f0dcb5f-kube-api-access-vrtdp" (OuterVolumeSpecName: "kube-api-access-vrtdp") pod "18fa9304-50a2-45cf-b983-9d459f0dcb5f" (UID: "18fa9304-50a2-45cf-b983-9d459f0dcb5f"). InnerVolumeSpecName "kube-api-access-vrtdp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:39:55 crc kubenswrapper[4907]: I0313 15:39:55.156114 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/94ba887c-f88a-48e7-b4bc-2f171d21fa47-kube-api-access-64c2b" (OuterVolumeSpecName: "kube-api-access-64c2b") pod "94ba887c-f88a-48e7-b4bc-2f171d21fa47" (UID: "94ba887c-f88a-48e7-b4bc-2f171d21fa47"). InnerVolumeSpecName "kube-api-access-64c2b". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:39:55 crc kubenswrapper[4907]: I0313 15:39:55.157783 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c2c52fb5-c4ba-4c24-95af-ce75cd609126-kube-api-access-lxptq" (OuterVolumeSpecName: "kube-api-access-lxptq") pod "c2c52fb5-c4ba-4c24-95af-ce75cd609126" (UID: "c2c52fb5-c4ba-4c24-95af-ce75cd609126"). InnerVolumeSpecName "kube-api-access-lxptq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:39:55 crc kubenswrapper[4907]: I0313 15:39:55.255844 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-64c2b\" (UniqueName: \"kubernetes.io/projected/94ba887c-f88a-48e7-b4bc-2f171d21fa47-kube-api-access-64c2b\") on node \"crc\" DevicePath \"\"" Mar 13 15:39:55 crc kubenswrapper[4907]: I0313 15:39:55.255942 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lxptq\" (UniqueName: \"kubernetes.io/projected/c2c52fb5-c4ba-4c24-95af-ce75cd609126-kube-api-access-lxptq\") on node \"crc\" DevicePath \"\"" Mar 13 15:39:55 crc kubenswrapper[4907]: I0313 15:39:55.255953 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vrtdp\" (UniqueName: \"kubernetes.io/projected/18fa9304-50a2-45cf-b983-9d459f0dcb5f-kube-api-access-vrtdp\") on node \"crc\" DevicePath \"\"" Mar 13 15:39:55 crc kubenswrapper[4907]: I0313 15:39:55.549233 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-5r4tb" event={"ID":"e5aab76f-4bcc-47be-989d-7d9d6ecf371b","Type":"ContainerDied","Data":"dd3f82f5adfcd093fbfe4c95a182789207aea1df8bcb504fdcb2fe97dbfbdb96"} Mar 13 15:39:55 crc kubenswrapper[4907]: I0313 15:39:55.549287 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dd3f82f5adfcd093fbfe4c95a182789207aea1df8bcb504fdcb2fe97dbfbdb96" Mar 13 15:39:55 crc kubenswrapper[4907]: I0313 15:39:55.549351 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-5r4tb" Mar 13 15:39:55 crc kubenswrapper[4907]: I0313 15:39:55.555478 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-7djpm" event={"ID":"c2c52fb5-c4ba-4c24-95af-ce75cd609126","Type":"ContainerDied","Data":"0e23da64c584ff8030a07246c197bc4df16cf45b1b7b65a2a1e9a6373f844c90"} Mar 13 15:39:55 crc kubenswrapper[4907]: I0313 15:39:55.555540 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0e23da64c584ff8030a07246c197bc4df16cf45b1b7b65a2a1e9a6373f844c90" Mar 13 15:39:55 crc kubenswrapper[4907]: I0313 15:39:55.555626 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-7djpm" Mar 13 15:39:55 crc kubenswrapper[4907]: I0313 15:39:55.561122 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-brbzj" event={"ID":"18fa9304-50a2-45cf-b983-9d459f0dcb5f","Type":"ContainerDied","Data":"e2b380c63ddd685de31e5a747984baea0335628ccbc6b3f1af44debb13f13653"} Mar 13 15:39:55 crc kubenswrapper[4907]: I0313 15:39:55.561168 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e2b380c63ddd685de31e5a747984baea0335628ccbc6b3f1af44debb13f13653" Mar 13 15:39:55 crc kubenswrapper[4907]: I0313 15:39:55.561224 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-brbzj" Mar 13 15:39:55 crc kubenswrapper[4907]: I0313 15:39:55.564310 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-62d0-account-create-update-vf8sr" Mar 13 15:39:55 crc kubenswrapper[4907]: I0313 15:39:55.564968 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-62d0-account-create-update-vf8sr" event={"ID":"94ba887c-f88a-48e7-b4bc-2f171d21fa47","Type":"ContainerDied","Data":"4a40897775cd1b3c4b3543b2070f41a576ca4e865bdb987c81c613efd9443cc6"} Mar 13 15:39:55 crc kubenswrapper[4907]: I0313 15:39:55.565024 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4a40897775cd1b3c4b3543b2070f41a576ca4e865bdb987c81c613efd9443cc6" Mar 13 15:39:55 crc kubenswrapper[4907]: I0313 15:39:55.925042 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-5a99-account-create-update-nkg4x" Mar 13 15:39:55 crc kubenswrapper[4907]: I0313 15:39:55.930684 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-356d-account-create-update-kb8lr" Mar 13 15:39:56 crc kubenswrapper[4907]: I0313 15:39:56.069865 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/40763271-36d9-4bd2-8ae8-82140648fcf4-operator-scripts\") pod \"40763271-36d9-4bd2-8ae8-82140648fcf4\" (UID: \"40763271-36d9-4bd2-8ae8-82140648fcf4\") " Mar 13 15:39:56 crc kubenswrapper[4907]: I0313 15:39:56.070202 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b66093c1-e494-4975-b8d1-217844060204-operator-scripts\") pod \"b66093c1-e494-4975-b8d1-217844060204\" (UID: \"b66093c1-e494-4975-b8d1-217844060204\") " Mar 13 15:39:56 crc kubenswrapper[4907]: I0313 15:39:56.070252 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pkzvl\" (UniqueName: \"kubernetes.io/projected/40763271-36d9-4bd2-8ae8-82140648fcf4-kube-api-access-pkzvl\") pod \"40763271-36d9-4bd2-8ae8-82140648fcf4\" (UID: \"40763271-36d9-4bd2-8ae8-82140648fcf4\") " Mar 13 15:39:56 crc kubenswrapper[4907]: I0313 15:39:56.070548 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wcjg9\" (UniqueName: \"kubernetes.io/projected/b66093c1-e494-4975-b8d1-217844060204-kube-api-access-wcjg9\") pod \"b66093c1-e494-4975-b8d1-217844060204\" (UID: \"b66093c1-e494-4975-b8d1-217844060204\") " Mar 13 15:39:56 crc kubenswrapper[4907]: I0313 15:39:56.073222 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b66093c1-e494-4975-b8d1-217844060204-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b66093c1-e494-4975-b8d1-217844060204" (UID: "b66093c1-e494-4975-b8d1-217844060204"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:39:56 crc kubenswrapper[4907]: I0313 15:39:56.073543 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/40763271-36d9-4bd2-8ae8-82140648fcf4-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "40763271-36d9-4bd2-8ae8-82140648fcf4" (UID: "40763271-36d9-4bd2-8ae8-82140648fcf4"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:39:56 crc kubenswrapper[4907]: I0313 15:39:56.081004 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/40763271-36d9-4bd2-8ae8-82140648fcf4-kube-api-access-pkzvl" (OuterVolumeSpecName: "kube-api-access-pkzvl") pod "40763271-36d9-4bd2-8ae8-82140648fcf4" (UID: "40763271-36d9-4bd2-8ae8-82140648fcf4"). InnerVolumeSpecName "kube-api-access-pkzvl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:39:56 crc kubenswrapper[4907]: I0313 15:39:56.087300 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b66093c1-e494-4975-b8d1-217844060204-kube-api-access-wcjg9" (OuterVolumeSpecName: "kube-api-access-wcjg9") pod "b66093c1-e494-4975-b8d1-217844060204" (UID: "b66093c1-e494-4975-b8d1-217844060204"). InnerVolumeSpecName "kube-api-access-wcjg9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:39:56 crc kubenswrapper[4907]: I0313 15:39:56.173590 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wcjg9\" (UniqueName: \"kubernetes.io/projected/b66093c1-e494-4975-b8d1-217844060204-kube-api-access-wcjg9\") on node \"crc\" DevicePath \"\"" Mar 13 15:39:56 crc kubenswrapper[4907]: I0313 15:39:56.173640 4907 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/40763271-36d9-4bd2-8ae8-82140648fcf4-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 15:39:56 crc kubenswrapper[4907]: I0313 15:39:56.173653 4907 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b66093c1-e494-4975-b8d1-217844060204-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 15:39:56 crc kubenswrapper[4907]: I0313 15:39:56.173665 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pkzvl\" (UniqueName: \"kubernetes.io/projected/40763271-36d9-4bd2-8ae8-82140648fcf4-kube-api-access-pkzvl\") on node \"crc\" DevicePath \"\"" Mar 13 15:39:56 crc kubenswrapper[4907]: I0313 15:39:56.575802 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-356d-account-create-update-kb8lr" Mar 13 15:39:56 crc kubenswrapper[4907]: I0313 15:39:56.575825 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-356d-account-create-update-kb8lr" event={"ID":"b66093c1-e494-4975-b8d1-217844060204","Type":"ContainerDied","Data":"0bf11b5c16f026ec4133b29e8d2e6b391b0da007fcd4acde47500c6e6fdc42e7"} Mar 13 15:39:56 crc kubenswrapper[4907]: I0313 15:39:56.575937 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0bf11b5c16f026ec4133b29e8d2e6b391b0da007fcd4acde47500c6e6fdc42e7" Mar 13 15:39:56 crc kubenswrapper[4907]: I0313 15:39:56.577682 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-5a99-account-create-update-nkg4x" event={"ID":"40763271-36d9-4bd2-8ae8-82140648fcf4","Type":"ContainerDied","Data":"543db5418a330957dba2de6b68ab5a198dc372131f880d599abcd7de199ca1e8"} Mar 13 15:39:56 crc kubenswrapper[4907]: I0313 15:39:56.577731 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="543db5418a330957dba2de6b68ab5a198dc372131f880d599abcd7de199ca1e8" Mar 13 15:39:56 crc kubenswrapper[4907]: I0313 15:39:56.577790 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-5a99-account-create-update-nkg4x" Mar 13 15:39:57 crc kubenswrapper[4907]: I0313 15:39:57.330285 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-mwt47"] Mar 13 15:39:57 crc kubenswrapper[4907]: E0313 15:39:57.330637 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5aab76f-4bcc-47be-989d-7d9d6ecf371b" containerName="mariadb-database-create" Mar 13 15:39:57 crc kubenswrapper[4907]: I0313 15:39:57.330650 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5aab76f-4bcc-47be-989d-7d9d6ecf371b" containerName="mariadb-database-create" Mar 13 15:39:57 crc kubenswrapper[4907]: E0313 15:39:57.330675 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b66093c1-e494-4975-b8d1-217844060204" containerName="mariadb-account-create-update" Mar 13 15:39:57 crc kubenswrapper[4907]: I0313 15:39:57.330681 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="b66093c1-e494-4975-b8d1-217844060204" containerName="mariadb-account-create-update" Mar 13 15:39:57 crc kubenswrapper[4907]: E0313 15:39:57.330689 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2c52fb5-c4ba-4c24-95af-ce75cd609126" containerName="mariadb-database-create" Mar 13 15:39:57 crc kubenswrapper[4907]: I0313 15:39:57.330695 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2c52fb5-c4ba-4c24-95af-ce75cd609126" containerName="mariadb-database-create" Mar 13 15:39:57 crc kubenswrapper[4907]: E0313 15:39:57.330703 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40763271-36d9-4bd2-8ae8-82140648fcf4" containerName="mariadb-account-create-update" Mar 13 15:39:57 crc kubenswrapper[4907]: I0313 15:39:57.330709 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="40763271-36d9-4bd2-8ae8-82140648fcf4" containerName="mariadb-account-create-update" Mar 13 15:39:57 crc kubenswrapper[4907]: E0313 15:39:57.330721 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18fa9304-50a2-45cf-b983-9d459f0dcb5f" containerName="mariadb-database-create" Mar 13 15:39:57 crc kubenswrapper[4907]: I0313 15:39:57.330727 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="18fa9304-50a2-45cf-b983-9d459f0dcb5f" containerName="mariadb-database-create" Mar 13 15:39:57 crc kubenswrapper[4907]: E0313 15:39:57.330746 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94ba887c-f88a-48e7-b4bc-2f171d21fa47" containerName="mariadb-account-create-update" Mar 13 15:39:57 crc kubenswrapper[4907]: I0313 15:39:57.330752 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="94ba887c-f88a-48e7-b4bc-2f171d21fa47" containerName="mariadb-account-create-update" Mar 13 15:39:57 crc kubenswrapper[4907]: I0313 15:39:57.330910 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="e5aab76f-4bcc-47be-989d-7d9d6ecf371b" containerName="mariadb-database-create" Mar 13 15:39:57 crc kubenswrapper[4907]: I0313 15:39:57.330923 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="94ba887c-f88a-48e7-b4bc-2f171d21fa47" containerName="mariadb-account-create-update" Mar 13 15:39:57 crc kubenswrapper[4907]: I0313 15:39:57.330935 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="40763271-36d9-4bd2-8ae8-82140648fcf4" containerName="mariadb-account-create-update" Mar 13 15:39:57 crc kubenswrapper[4907]: I0313 15:39:57.330949 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="b66093c1-e494-4975-b8d1-217844060204" containerName="mariadb-account-create-update" Mar 13 15:39:57 crc kubenswrapper[4907]: I0313 15:39:57.330960 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="c2c52fb5-c4ba-4c24-95af-ce75cd609126" containerName="mariadb-database-create" Mar 13 15:39:57 crc kubenswrapper[4907]: I0313 15:39:57.330970 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="18fa9304-50a2-45cf-b983-9d459f0dcb5f" containerName="mariadb-database-create" Mar 13 15:39:57 crc kubenswrapper[4907]: I0313 15:39:57.331826 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-mwt47" Mar 13 15:39:57 crc kubenswrapper[4907]: I0313 15:39:57.334751 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Mar 13 15:39:57 crc kubenswrapper[4907]: I0313 15:39:57.334800 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-fmnpb" Mar 13 15:39:57 crc kubenswrapper[4907]: I0313 15:39:57.335023 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Mar 13 15:39:57 crc kubenswrapper[4907]: I0313 15:39:57.343414 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-mwt47"] Mar 13 15:39:57 crc kubenswrapper[4907]: I0313 15:39:57.394143 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rxd2t\" (UniqueName: \"kubernetes.io/projected/0336c3c0-cc18-4bb3-afe3-7e2a0232c889-kube-api-access-rxd2t\") pod \"nova-cell0-conductor-db-sync-mwt47\" (UID: \"0336c3c0-cc18-4bb3-afe3-7e2a0232c889\") " pod="openstack/nova-cell0-conductor-db-sync-mwt47" Mar 13 15:39:57 crc kubenswrapper[4907]: I0313 15:39:57.394202 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0336c3c0-cc18-4bb3-afe3-7e2a0232c889-config-data\") pod \"nova-cell0-conductor-db-sync-mwt47\" (UID: \"0336c3c0-cc18-4bb3-afe3-7e2a0232c889\") " pod="openstack/nova-cell0-conductor-db-sync-mwt47" Mar 13 15:39:57 crc kubenswrapper[4907]: I0313 15:39:57.394319 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0336c3c0-cc18-4bb3-afe3-7e2a0232c889-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-mwt47\" (UID: \"0336c3c0-cc18-4bb3-afe3-7e2a0232c889\") " pod="openstack/nova-cell0-conductor-db-sync-mwt47" Mar 13 15:39:57 crc kubenswrapper[4907]: I0313 15:39:57.394340 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0336c3c0-cc18-4bb3-afe3-7e2a0232c889-scripts\") pod \"nova-cell0-conductor-db-sync-mwt47\" (UID: \"0336c3c0-cc18-4bb3-afe3-7e2a0232c889\") " pod="openstack/nova-cell0-conductor-db-sync-mwt47" Mar 13 15:39:57 crc kubenswrapper[4907]: I0313 15:39:57.495292 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0336c3c0-cc18-4bb3-afe3-7e2a0232c889-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-mwt47\" (UID: \"0336c3c0-cc18-4bb3-afe3-7e2a0232c889\") " pod="openstack/nova-cell0-conductor-db-sync-mwt47" Mar 13 15:39:57 crc kubenswrapper[4907]: I0313 15:39:57.495349 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0336c3c0-cc18-4bb3-afe3-7e2a0232c889-scripts\") pod \"nova-cell0-conductor-db-sync-mwt47\" (UID: \"0336c3c0-cc18-4bb3-afe3-7e2a0232c889\") " pod="openstack/nova-cell0-conductor-db-sync-mwt47" Mar 13 15:39:57 crc kubenswrapper[4907]: I0313 15:39:57.495400 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rxd2t\" (UniqueName: \"kubernetes.io/projected/0336c3c0-cc18-4bb3-afe3-7e2a0232c889-kube-api-access-rxd2t\") pod \"nova-cell0-conductor-db-sync-mwt47\" (UID: \"0336c3c0-cc18-4bb3-afe3-7e2a0232c889\") " pod="openstack/nova-cell0-conductor-db-sync-mwt47" Mar 13 15:39:57 crc kubenswrapper[4907]: I0313 15:39:57.495434 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0336c3c0-cc18-4bb3-afe3-7e2a0232c889-config-data\") pod \"nova-cell0-conductor-db-sync-mwt47\" (UID: \"0336c3c0-cc18-4bb3-afe3-7e2a0232c889\") " pod="openstack/nova-cell0-conductor-db-sync-mwt47" Mar 13 15:39:57 crc kubenswrapper[4907]: I0313 15:39:57.499194 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0336c3c0-cc18-4bb3-afe3-7e2a0232c889-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-mwt47\" (UID: \"0336c3c0-cc18-4bb3-afe3-7e2a0232c889\") " pod="openstack/nova-cell0-conductor-db-sync-mwt47" Mar 13 15:39:57 crc kubenswrapper[4907]: I0313 15:39:57.499846 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0336c3c0-cc18-4bb3-afe3-7e2a0232c889-scripts\") pod \"nova-cell0-conductor-db-sync-mwt47\" (UID: \"0336c3c0-cc18-4bb3-afe3-7e2a0232c889\") " pod="openstack/nova-cell0-conductor-db-sync-mwt47" Mar 13 15:39:57 crc kubenswrapper[4907]: I0313 15:39:57.501031 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0336c3c0-cc18-4bb3-afe3-7e2a0232c889-config-data\") pod \"nova-cell0-conductor-db-sync-mwt47\" (UID: \"0336c3c0-cc18-4bb3-afe3-7e2a0232c889\") " pod="openstack/nova-cell0-conductor-db-sync-mwt47" Mar 13 15:39:57 crc kubenswrapper[4907]: I0313 15:39:57.517152 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rxd2t\" (UniqueName: \"kubernetes.io/projected/0336c3c0-cc18-4bb3-afe3-7e2a0232c889-kube-api-access-rxd2t\") pod \"nova-cell0-conductor-db-sync-mwt47\" (UID: \"0336c3c0-cc18-4bb3-afe3-7e2a0232c889\") " pod="openstack/nova-cell0-conductor-db-sync-mwt47" Mar 13 15:39:57 crc kubenswrapper[4907]: I0313 15:39:57.648950 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-mwt47" Mar 13 15:39:58 crc kubenswrapper[4907]: I0313 15:39:58.086611 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-mwt47"] Mar 13 15:39:58 crc kubenswrapper[4907]: W0313 15:39:58.092382 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0336c3c0_cc18_4bb3_afe3_7e2a0232c889.slice/crio-fd3776051992ad0e194814f0af106e536e760a3d9887d7bff525ead01fdacb92 WatchSource:0}: Error finding container fd3776051992ad0e194814f0af106e536e760a3d9887d7bff525ead01fdacb92: Status 404 returned error can't find the container with id fd3776051992ad0e194814f0af106e536e760a3d9887d7bff525ead01fdacb92 Mar 13 15:39:58 crc kubenswrapper[4907]: I0313 15:39:58.596045 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-mwt47" event={"ID":"0336c3c0-cc18-4bb3-afe3-7e2a0232c889","Type":"ContainerStarted","Data":"e0422eb37278e23b8d6a6f631b14d7a520f0f17ab9fd9d3ee373391ac6276186"} Mar 13 15:39:58 crc kubenswrapper[4907]: I0313 15:39:58.596343 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-mwt47" event={"ID":"0336c3c0-cc18-4bb3-afe3-7e2a0232c889","Type":"ContainerStarted","Data":"fd3776051992ad0e194814f0af106e536e760a3d9887d7bff525ead01fdacb92"} Mar 13 15:39:58 crc kubenswrapper[4907]: I0313 15:39:58.624031 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-mwt47" podStartSLOduration=1.6240110250000002 podStartE2EDuration="1.624011025s" podCreationTimestamp="2026-03-13 15:39:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:39:58.612334395 +0000 UTC m=+5697.512122084" watchObservedRunningTime="2026-03-13 15:39:58.624011025 +0000 UTC m=+5697.523798714" Mar 13 15:40:00 crc kubenswrapper[4907]: I0313 15:40:00.137469 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556940-5rmt4"] Mar 13 15:40:00 crc kubenswrapper[4907]: I0313 15:40:00.139224 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556940-5rmt4" Mar 13 15:40:00 crc kubenswrapper[4907]: I0313 15:40:00.145583 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556940-5rmt4"] Mar 13 15:40:00 crc kubenswrapper[4907]: I0313 15:40:00.145635 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 15:40:00 crc kubenswrapper[4907]: I0313 15:40:00.146041 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 15:40:00 crc kubenswrapper[4907]: I0313 15:40:00.147306 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 15:40:00 crc kubenswrapper[4907]: I0313 15:40:00.252269 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v7mcj\" (UniqueName: \"kubernetes.io/projected/72d20fdf-a093-459f-9328-0549ef7b48c4-kube-api-access-v7mcj\") pod \"auto-csr-approver-29556940-5rmt4\" (UID: \"72d20fdf-a093-459f-9328-0549ef7b48c4\") " pod="openshift-infra/auto-csr-approver-29556940-5rmt4" Mar 13 15:40:00 crc kubenswrapper[4907]: I0313 15:40:00.354023 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v7mcj\" (UniqueName: \"kubernetes.io/projected/72d20fdf-a093-459f-9328-0549ef7b48c4-kube-api-access-v7mcj\") pod \"auto-csr-approver-29556940-5rmt4\" (UID: \"72d20fdf-a093-459f-9328-0549ef7b48c4\") " pod="openshift-infra/auto-csr-approver-29556940-5rmt4" Mar 13 15:40:00 crc kubenswrapper[4907]: I0313 15:40:00.371090 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v7mcj\" (UniqueName: \"kubernetes.io/projected/72d20fdf-a093-459f-9328-0549ef7b48c4-kube-api-access-v7mcj\") pod \"auto-csr-approver-29556940-5rmt4\" (UID: \"72d20fdf-a093-459f-9328-0549ef7b48c4\") " pod="openshift-infra/auto-csr-approver-29556940-5rmt4" Mar 13 15:40:00 crc kubenswrapper[4907]: I0313 15:40:00.465977 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556940-5rmt4" Mar 13 15:40:00 crc kubenswrapper[4907]: I0313 15:40:00.889875 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556940-5rmt4"] Mar 13 15:40:00 crc kubenswrapper[4907]: W0313 15:40:00.895296 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod72d20fdf_a093_459f_9328_0549ef7b48c4.slice/crio-d2776116aa3026d03d3e50e4d891cc0f1286bb990c15bcd8292e00f25ab105e4 WatchSource:0}: Error finding container d2776116aa3026d03d3e50e4d891cc0f1286bb990c15bcd8292e00f25ab105e4: Status 404 returned error can't find the container with id d2776116aa3026d03d3e50e4d891cc0f1286bb990c15bcd8292e00f25ab105e4 Mar 13 15:40:01 crc kubenswrapper[4907]: I0313 15:40:01.629184 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556940-5rmt4" event={"ID":"72d20fdf-a093-459f-9328-0549ef7b48c4","Type":"ContainerStarted","Data":"d2776116aa3026d03d3e50e4d891cc0f1286bb990c15bcd8292e00f25ab105e4"} Mar 13 15:40:02 crc kubenswrapper[4907]: I0313 15:40:02.640145 4907 generic.go:334] "Generic (PLEG): container finished" podID="72d20fdf-a093-459f-9328-0549ef7b48c4" containerID="3f861dd2c3ff8fa5314ce9c35d45445320ad82d6381b4138837ad16c289319f5" exitCode=0 Mar 13 15:40:02 crc kubenswrapper[4907]: I0313 15:40:02.640205 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556940-5rmt4" event={"ID":"72d20fdf-a093-459f-9328-0549ef7b48c4","Type":"ContainerDied","Data":"3f861dd2c3ff8fa5314ce9c35d45445320ad82d6381b4138837ad16c289319f5"} Mar 13 15:40:03 crc kubenswrapper[4907]: I0313 15:40:03.651224 4907 generic.go:334] "Generic (PLEG): container finished" podID="0336c3c0-cc18-4bb3-afe3-7e2a0232c889" containerID="e0422eb37278e23b8d6a6f631b14d7a520f0f17ab9fd9d3ee373391ac6276186" exitCode=0 Mar 13 15:40:03 crc kubenswrapper[4907]: I0313 15:40:03.651302 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-mwt47" event={"ID":"0336c3c0-cc18-4bb3-afe3-7e2a0232c889","Type":"ContainerDied","Data":"e0422eb37278e23b8d6a6f631b14d7a520f0f17ab9fd9d3ee373391ac6276186"} Mar 13 15:40:03 crc kubenswrapper[4907]: I0313 15:40:03.949648 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556940-5rmt4" Mar 13 15:40:04 crc kubenswrapper[4907]: I0313 15:40:04.014991 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v7mcj\" (UniqueName: \"kubernetes.io/projected/72d20fdf-a093-459f-9328-0549ef7b48c4-kube-api-access-v7mcj\") pod \"72d20fdf-a093-459f-9328-0549ef7b48c4\" (UID: \"72d20fdf-a093-459f-9328-0549ef7b48c4\") " Mar 13 15:40:04 crc kubenswrapper[4907]: I0313 15:40:04.020922 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/72d20fdf-a093-459f-9328-0549ef7b48c4-kube-api-access-v7mcj" (OuterVolumeSpecName: "kube-api-access-v7mcj") pod "72d20fdf-a093-459f-9328-0549ef7b48c4" (UID: "72d20fdf-a093-459f-9328-0549ef7b48c4"). InnerVolumeSpecName "kube-api-access-v7mcj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:40:04 crc kubenswrapper[4907]: I0313 15:40:04.116565 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v7mcj\" (UniqueName: \"kubernetes.io/projected/72d20fdf-a093-459f-9328-0549ef7b48c4-kube-api-access-v7mcj\") on node \"crc\" DevicePath \"\"" Mar 13 15:40:04 crc kubenswrapper[4907]: I0313 15:40:04.660437 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556940-5rmt4" event={"ID":"72d20fdf-a093-459f-9328-0549ef7b48c4","Type":"ContainerDied","Data":"d2776116aa3026d03d3e50e4d891cc0f1286bb990c15bcd8292e00f25ab105e4"} Mar 13 15:40:04 crc kubenswrapper[4907]: I0313 15:40:04.660480 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d2776116aa3026d03d3e50e4d891cc0f1286bb990c15bcd8292e00f25ab105e4" Mar 13 15:40:04 crc kubenswrapper[4907]: I0313 15:40:04.660489 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556940-5rmt4" Mar 13 15:40:05 crc kubenswrapper[4907]: I0313 15:40:05.016668 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-mwt47" Mar 13 15:40:05 crc kubenswrapper[4907]: I0313 15:40:05.020466 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556934-wf7zr"] Mar 13 15:40:05 crc kubenswrapper[4907]: I0313 15:40:05.027713 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556934-wf7zr"] Mar 13 15:40:05 crc kubenswrapper[4907]: I0313 15:40:05.136411 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rxd2t\" (UniqueName: \"kubernetes.io/projected/0336c3c0-cc18-4bb3-afe3-7e2a0232c889-kube-api-access-rxd2t\") pod \"0336c3c0-cc18-4bb3-afe3-7e2a0232c889\" (UID: \"0336c3c0-cc18-4bb3-afe3-7e2a0232c889\") " Mar 13 15:40:05 crc kubenswrapper[4907]: I0313 15:40:05.136724 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0336c3c0-cc18-4bb3-afe3-7e2a0232c889-combined-ca-bundle\") pod \"0336c3c0-cc18-4bb3-afe3-7e2a0232c889\" (UID: \"0336c3c0-cc18-4bb3-afe3-7e2a0232c889\") " Mar 13 15:40:05 crc kubenswrapper[4907]: I0313 15:40:05.136830 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0336c3c0-cc18-4bb3-afe3-7e2a0232c889-scripts\") pod \"0336c3c0-cc18-4bb3-afe3-7e2a0232c889\" (UID: \"0336c3c0-cc18-4bb3-afe3-7e2a0232c889\") " Mar 13 15:40:05 crc kubenswrapper[4907]: I0313 15:40:05.137041 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0336c3c0-cc18-4bb3-afe3-7e2a0232c889-config-data\") pod \"0336c3c0-cc18-4bb3-afe3-7e2a0232c889\" (UID: \"0336c3c0-cc18-4bb3-afe3-7e2a0232c889\") " Mar 13 15:40:05 crc kubenswrapper[4907]: I0313 15:40:05.141736 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0336c3c0-cc18-4bb3-afe3-7e2a0232c889-scripts" (OuterVolumeSpecName: "scripts") pod "0336c3c0-cc18-4bb3-afe3-7e2a0232c889" (UID: "0336c3c0-cc18-4bb3-afe3-7e2a0232c889"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:40:05 crc kubenswrapper[4907]: I0313 15:40:05.142039 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0336c3c0-cc18-4bb3-afe3-7e2a0232c889-kube-api-access-rxd2t" (OuterVolumeSpecName: "kube-api-access-rxd2t") pod "0336c3c0-cc18-4bb3-afe3-7e2a0232c889" (UID: "0336c3c0-cc18-4bb3-afe3-7e2a0232c889"). InnerVolumeSpecName "kube-api-access-rxd2t". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:40:05 crc kubenswrapper[4907]: I0313 15:40:05.163387 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0336c3c0-cc18-4bb3-afe3-7e2a0232c889-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0336c3c0-cc18-4bb3-afe3-7e2a0232c889" (UID: "0336c3c0-cc18-4bb3-afe3-7e2a0232c889"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:40:05 crc kubenswrapper[4907]: I0313 15:40:05.165593 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0336c3c0-cc18-4bb3-afe3-7e2a0232c889-config-data" (OuterVolumeSpecName: "config-data") pod "0336c3c0-cc18-4bb3-afe3-7e2a0232c889" (UID: "0336c3c0-cc18-4bb3-afe3-7e2a0232c889"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:40:05 crc kubenswrapper[4907]: I0313 15:40:05.238921 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0336c3c0-cc18-4bb3-afe3-7e2a0232c889-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 15:40:05 crc kubenswrapper[4907]: I0313 15:40:05.238968 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rxd2t\" (UniqueName: \"kubernetes.io/projected/0336c3c0-cc18-4bb3-afe3-7e2a0232c889-kube-api-access-rxd2t\") on node \"crc\" DevicePath \"\"" Mar 13 15:40:05 crc kubenswrapper[4907]: I0313 15:40:05.238982 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0336c3c0-cc18-4bb3-afe3-7e2a0232c889-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 15:40:05 crc kubenswrapper[4907]: I0313 15:40:05.238993 4907 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0336c3c0-cc18-4bb3-afe3-7e2a0232c889-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 15:40:05 crc kubenswrapper[4907]: I0313 15:40:05.669775 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-mwt47" event={"ID":"0336c3c0-cc18-4bb3-afe3-7e2a0232c889","Type":"ContainerDied","Data":"fd3776051992ad0e194814f0af106e536e760a3d9887d7bff525ead01fdacb92"} Mar 13 15:40:05 crc kubenswrapper[4907]: I0313 15:40:05.669808 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fd3776051992ad0e194814f0af106e536e760a3d9887d7bff525ead01fdacb92" Mar 13 15:40:05 crc kubenswrapper[4907]: I0313 15:40:05.669896 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-mwt47" Mar 13 15:40:05 crc kubenswrapper[4907]: I0313 15:40:05.741176 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Mar 13 15:40:05 crc kubenswrapper[4907]: E0313 15:40:05.742173 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0336c3c0-cc18-4bb3-afe3-7e2a0232c889" containerName="nova-cell0-conductor-db-sync" Mar 13 15:40:05 crc kubenswrapper[4907]: I0313 15:40:05.742206 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="0336c3c0-cc18-4bb3-afe3-7e2a0232c889" containerName="nova-cell0-conductor-db-sync" Mar 13 15:40:05 crc kubenswrapper[4907]: E0313 15:40:05.742269 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72d20fdf-a093-459f-9328-0549ef7b48c4" containerName="oc" Mar 13 15:40:05 crc kubenswrapper[4907]: I0313 15:40:05.742281 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="72d20fdf-a093-459f-9328-0549ef7b48c4" containerName="oc" Mar 13 15:40:05 crc kubenswrapper[4907]: I0313 15:40:05.742521 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="72d20fdf-a093-459f-9328-0549ef7b48c4" containerName="oc" Mar 13 15:40:05 crc kubenswrapper[4907]: I0313 15:40:05.742563 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="0336c3c0-cc18-4bb3-afe3-7e2a0232c889" containerName="nova-cell0-conductor-db-sync" Mar 13 15:40:05 crc kubenswrapper[4907]: I0313 15:40:05.743481 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Mar 13 15:40:05 crc kubenswrapper[4907]: I0313 15:40:05.745613 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-fmnpb" Mar 13 15:40:05 crc kubenswrapper[4907]: I0313 15:40:05.762115 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Mar 13 15:40:05 crc kubenswrapper[4907]: I0313 15:40:05.765348 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Mar 13 15:40:05 crc kubenswrapper[4907]: I0313 15:40:05.805568 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="041cab28-d0e1-4ec9-a0cb-92176be369aa" path="/var/lib/kubelet/pods/041cab28-d0e1-4ec9-a0cb-92176be369aa/volumes" Mar 13 15:40:05 crc kubenswrapper[4907]: I0313 15:40:05.847313 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8047522-cc05-44a3-b57d-2362e6457d49-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"a8047522-cc05-44a3-b57d-2362e6457d49\") " pod="openstack/nova-cell0-conductor-0" Mar 13 15:40:05 crc kubenswrapper[4907]: I0313 15:40:05.847400 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-77rtk\" (UniqueName: \"kubernetes.io/projected/a8047522-cc05-44a3-b57d-2362e6457d49-kube-api-access-77rtk\") pod \"nova-cell0-conductor-0\" (UID: \"a8047522-cc05-44a3-b57d-2362e6457d49\") " pod="openstack/nova-cell0-conductor-0" Mar 13 15:40:05 crc kubenswrapper[4907]: I0313 15:40:05.847717 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8047522-cc05-44a3-b57d-2362e6457d49-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"a8047522-cc05-44a3-b57d-2362e6457d49\") " pod="openstack/nova-cell0-conductor-0" Mar 13 15:40:05 crc kubenswrapper[4907]: I0313 15:40:05.949460 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8047522-cc05-44a3-b57d-2362e6457d49-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"a8047522-cc05-44a3-b57d-2362e6457d49\") " pod="openstack/nova-cell0-conductor-0" Mar 13 15:40:05 crc kubenswrapper[4907]: I0313 15:40:05.949535 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8047522-cc05-44a3-b57d-2362e6457d49-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"a8047522-cc05-44a3-b57d-2362e6457d49\") " pod="openstack/nova-cell0-conductor-0" Mar 13 15:40:05 crc kubenswrapper[4907]: I0313 15:40:05.949564 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-77rtk\" (UniqueName: \"kubernetes.io/projected/a8047522-cc05-44a3-b57d-2362e6457d49-kube-api-access-77rtk\") pod \"nova-cell0-conductor-0\" (UID: \"a8047522-cc05-44a3-b57d-2362e6457d49\") " pod="openstack/nova-cell0-conductor-0" Mar 13 15:40:05 crc kubenswrapper[4907]: I0313 15:40:05.953373 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8047522-cc05-44a3-b57d-2362e6457d49-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"a8047522-cc05-44a3-b57d-2362e6457d49\") " pod="openstack/nova-cell0-conductor-0" Mar 13 15:40:05 crc kubenswrapper[4907]: I0313 15:40:05.953962 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8047522-cc05-44a3-b57d-2362e6457d49-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"a8047522-cc05-44a3-b57d-2362e6457d49\") " pod="openstack/nova-cell0-conductor-0" Mar 13 15:40:05 crc kubenswrapper[4907]: I0313 15:40:05.967506 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-77rtk\" (UniqueName: \"kubernetes.io/projected/a8047522-cc05-44a3-b57d-2362e6457d49-kube-api-access-77rtk\") pod \"nova-cell0-conductor-0\" (UID: \"a8047522-cc05-44a3-b57d-2362e6457d49\") " pod="openstack/nova-cell0-conductor-0" Mar 13 15:40:06 crc kubenswrapper[4907]: I0313 15:40:06.069834 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Mar 13 15:40:06 crc kubenswrapper[4907]: I0313 15:40:06.490604 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Mar 13 15:40:06 crc kubenswrapper[4907]: I0313 15:40:06.678927 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"a8047522-cc05-44a3-b57d-2362e6457d49","Type":"ContainerStarted","Data":"4efeb4b84a961f5a72c12fc7db6d3a11b565d9a8fc5e00e1bd6bd862f0227c34"} Mar 13 15:40:06 crc kubenswrapper[4907]: I0313 15:40:06.678977 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"a8047522-cc05-44a3-b57d-2362e6457d49","Type":"ContainerStarted","Data":"78d4b76c3a1fd63fabbe50a18b7db60eba6e1e2f95e60b44b880d052c99e1044"} Mar 13 15:40:06 crc kubenswrapper[4907]: I0313 15:40:06.679073 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Mar 13 15:40:06 crc kubenswrapper[4907]: I0313 15:40:06.703740 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=1.703722275 podStartE2EDuration="1.703722275s" podCreationTimestamp="2026-03-13 15:40:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:40:06.694908715 +0000 UTC m=+5705.594696404" watchObservedRunningTime="2026-03-13 15:40:06.703722275 +0000 UTC m=+5705.603509954" Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.106677 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.535369 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-lvbcr"] Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.536784 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-lvbcr" Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.539080 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.539174 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.581645 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-lvbcr"] Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.655385 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/89840ecf-ddb9-4a66-855b-0a1ae524bd6e-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-lvbcr\" (UID: \"89840ecf-ddb9-4a66-855b-0a1ae524bd6e\") " pod="openstack/nova-cell0-cell-mapping-lvbcr" Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.655439 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lvkgk\" (UniqueName: \"kubernetes.io/projected/89840ecf-ddb9-4a66-855b-0a1ae524bd6e-kube-api-access-lvkgk\") pod \"nova-cell0-cell-mapping-lvbcr\" (UID: \"89840ecf-ddb9-4a66-855b-0a1ae524bd6e\") " pod="openstack/nova-cell0-cell-mapping-lvbcr" Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.655498 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/89840ecf-ddb9-4a66-855b-0a1ae524bd6e-config-data\") pod \"nova-cell0-cell-mapping-lvbcr\" (UID: \"89840ecf-ddb9-4a66-855b-0a1ae524bd6e\") " pod="openstack/nova-cell0-cell-mapping-lvbcr" Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.655533 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/89840ecf-ddb9-4a66-855b-0a1ae524bd6e-scripts\") pod \"nova-cell0-cell-mapping-lvbcr\" (UID: \"89840ecf-ddb9-4a66-855b-0a1ae524bd6e\") " pod="openstack/nova-cell0-cell-mapping-lvbcr" Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.704187 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.705601 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.711618 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.736752 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.745583 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.747079 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.756667 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.757280 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/89840ecf-ddb9-4a66-855b-0a1ae524bd6e-config-data\") pod \"nova-cell0-cell-mapping-lvbcr\" (UID: \"89840ecf-ddb9-4a66-855b-0a1ae524bd6e\") " pod="openstack/nova-cell0-cell-mapping-lvbcr" Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.757322 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/89840ecf-ddb9-4a66-855b-0a1ae524bd6e-scripts\") pod \"nova-cell0-cell-mapping-lvbcr\" (UID: \"89840ecf-ddb9-4a66-855b-0a1ae524bd6e\") " pod="openstack/nova-cell0-cell-mapping-lvbcr" Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.757385 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6dmkv\" (UniqueName: \"kubernetes.io/projected/188c81b4-cc35-444d-a5f1-f1633bae67a0-kube-api-access-6dmkv\") pod \"nova-api-0\" (UID: \"188c81b4-cc35-444d-a5f1-f1633bae67a0\") " pod="openstack/nova-api-0" Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.757424 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/188c81b4-cc35-444d-a5f1-f1633bae67a0-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"188c81b4-cc35-444d-a5f1-f1633bae67a0\") " pod="openstack/nova-api-0" Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.757445 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/188c81b4-cc35-444d-a5f1-f1633bae67a0-config-data\") pod \"nova-api-0\" (UID: \"188c81b4-cc35-444d-a5f1-f1633bae67a0\") " pod="openstack/nova-api-0" Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.757469 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/188c81b4-cc35-444d-a5f1-f1633bae67a0-logs\") pod \"nova-api-0\" (UID: \"188c81b4-cc35-444d-a5f1-f1633bae67a0\") " pod="openstack/nova-api-0" Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.757498 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/89840ecf-ddb9-4a66-855b-0a1ae524bd6e-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-lvbcr\" (UID: \"89840ecf-ddb9-4a66-855b-0a1ae524bd6e\") " pod="openstack/nova-cell0-cell-mapping-lvbcr" Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.757517 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lvkgk\" (UniqueName: \"kubernetes.io/projected/89840ecf-ddb9-4a66-855b-0a1ae524bd6e-kube-api-access-lvkgk\") pod \"nova-cell0-cell-mapping-lvbcr\" (UID: \"89840ecf-ddb9-4a66-855b-0a1ae524bd6e\") " pod="openstack/nova-cell0-cell-mapping-lvbcr" Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.763968 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/89840ecf-ddb9-4a66-855b-0a1ae524bd6e-config-data\") pod \"nova-cell0-cell-mapping-lvbcr\" (UID: \"89840ecf-ddb9-4a66-855b-0a1ae524bd6e\") " pod="openstack/nova-cell0-cell-mapping-lvbcr" Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.764162 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/89840ecf-ddb9-4a66-855b-0a1ae524bd6e-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-lvbcr\" (UID: \"89840ecf-ddb9-4a66-855b-0a1ae524bd6e\") " pod="openstack/nova-cell0-cell-mapping-lvbcr" Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.764580 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/89840ecf-ddb9-4a66-855b-0a1ae524bd6e-scripts\") pod \"nova-cell0-cell-mapping-lvbcr\" (UID: \"89840ecf-ddb9-4a66-855b-0a1ae524bd6e\") " pod="openstack/nova-cell0-cell-mapping-lvbcr" Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.789399 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lvkgk\" (UniqueName: \"kubernetes.io/projected/89840ecf-ddb9-4a66-855b-0a1ae524bd6e-kube-api-access-lvkgk\") pod \"nova-cell0-cell-mapping-lvbcr\" (UID: \"89840ecf-ddb9-4a66-855b-0a1ae524bd6e\") " pod="openstack/nova-cell0-cell-mapping-lvbcr" Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.811561 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.817333 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.819678 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.822467 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.859151 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc125541-0216-45f4-8276-520470f141fb-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"bc125541-0216-45f4-8276-520470f141fb\") " pod="openstack/nova-metadata-0" Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.859199 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/188c81b4-cc35-444d-a5f1-f1633bae67a0-logs\") pod \"nova-api-0\" (UID: \"188c81b4-cc35-444d-a5f1-f1633bae67a0\") " pod="openstack/nova-api-0" Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.859267 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5192153-c160-40bc-9375-a730e800e76e-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"e5192153-c160-40bc-9375-a730e800e76e\") " pod="openstack/nova-scheduler-0" Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.859325 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5192153-c160-40bc-9375-a730e800e76e-config-data\") pod \"nova-scheduler-0\" (UID: \"e5192153-c160-40bc-9375-a730e800e76e\") " pod="openstack/nova-scheduler-0" Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.859358 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6dmkv\" (UniqueName: \"kubernetes.io/projected/188c81b4-cc35-444d-a5f1-f1633bae67a0-kube-api-access-6dmkv\") pod \"nova-api-0\" (UID: \"188c81b4-cc35-444d-a5f1-f1633bae67a0\") " pod="openstack/nova-api-0" Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.859379 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc125541-0216-45f4-8276-520470f141fb-config-data\") pod \"nova-metadata-0\" (UID: \"bc125541-0216-45f4-8276-520470f141fb\") " pod="openstack/nova-metadata-0" Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.859396 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bc125541-0216-45f4-8276-520470f141fb-logs\") pod \"nova-metadata-0\" (UID: \"bc125541-0216-45f4-8276-520470f141fb\") " pod="openstack/nova-metadata-0" Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.859424 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hwqn2\" (UniqueName: \"kubernetes.io/projected/bc125541-0216-45f4-8276-520470f141fb-kube-api-access-hwqn2\") pod \"nova-metadata-0\" (UID: \"bc125541-0216-45f4-8276-520470f141fb\") " pod="openstack/nova-metadata-0" Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.859449 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/188c81b4-cc35-444d-a5f1-f1633bae67a0-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"188c81b4-cc35-444d-a5f1-f1633bae67a0\") " pod="openstack/nova-api-0" Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.859467 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5kk9j\" (UniqueName: \"kubernetes.io/projected/e5192153-c160-40bc-9375-a730e800e76e-kube-api-access-5kk9j\") pod \"nova-scheduler-0\" (UID: \"e5192153-c160-40bc-9375-a730e800e76e\") " pod="openstack/nova-scheduler-0" Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.859485 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/188c81b4-cc35-444d-a5f1-f1633bae67a0-config-data\") pod \"nova-api-0\" (UID: \"188c81b4-cc35-444d-a5f1-f1633bae67a0\") " pod="openstack/nova-api-0" Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.861003 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/188c81b4-cc35-444d-a5f1-f1633bae67a0-logs\") pod \"nova-api-0\" (UID: \"188c81b4-cc35-444d-a5f1-f1633bae67a0\") " pod="openstack/nova-api-0" Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.862678 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/188c81b4-cc35-444d-a5f1-f1633bae67a0-config-data\") pod \"nova-api-0\" (UID: \"188c81b4-cc35-444d-a5f1-f1633bae67a0\") " pod="openstack/nova-api-0" Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.862821 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-lvbcr" Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.873923 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.882063 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/188c81b4-cc35-444d-a5f1-f1633bae67a0-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"188c81b4-cc35-444d-a5f1-f1633bae67a0\") " pod="openstack/nova-api-0" Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.897950 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6dmkv\" (UniqueName: \"kubernetes.io/projected/188c81b4-cc35-444d-a5f1-f1633bae67a0-kube-api-access-6dmkv\") pod \"nova-api-0\" (UID: \"188c81b4-cc35-444d-a5f1-f1633bae67a0\") " pod="openstack/nova-api-0" Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.950312 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-68dbfc785f-kjm62"] Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.952020 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-68dbfc785f-kjm62" Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.960721 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc125541-0216-45f4-8276-520470f141fb-config-data\") pod \"nova-metadata-0\" (UID: \"bc125541-0216-45f4-8276-520470f141fb\") " pod="openstack/nova-metadata-0" Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.960778 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bc125541-0216-45f4-8276-520470f141fb-logs\") pod \"nova-metadata-0\" (UID: \"bc125541-0216-45f4-8276-520470f141fb\") " pod="openstack/nova-metadata-0" Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.960809 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hwqn2\" (UniqueName: \"kubernetes.io/projected/bc125541-0216-45f4-8276-520470f141fb-kube-api-access-hwqn2\") pod \"nova-metadata-0\" (UID: \"bc125541-0216-45f4-8276-520470f141fb\") " pod="openstack/nova-metadata-0" Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.960835 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5kk9j\" (UniqueName: \"kubernetes.io/projected/e5192153-c160-40bc-9375-a730e800e76e-kube-api-access-5kk9j\") pod \"nova-scheduler-0\" (UID: \"e5192153-c160-40bc-9375-a730e800e76e\") " pod="openstack/nova-scheduler-0" Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.960863 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc125541-0216-45f4-8276-520470f141fb-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"bc125541-0216-45f4-8276-520470f141fb\") " pod="openstack/nova-metadata-0" Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.960968 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5192153-c160-40bc-9375-a730e800e76e-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"e5192153-c160-40bc-9375-a730e800e76e\") " pod="openstack/nova-scheduler-0" Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.961052 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5192153-c160-40bc-9375-a730e800e76e-config-data\") pod \"nova-scheduler-0\" (UID: \"e5192153-c160-40bc-9375-a730e800e76e\") " pod="openstack/nova-scheduler-0" Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.964564 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bc125541-0216-45f4-8276-520470f141fb-logs\") pod \"nova-metadata-0\" (UID: \"bc125541-0216-45f4-8276-520470f141fb\") " pod="openstack/nova-metadata-0" Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.965189 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5192153-c160-40bc-9375-a730e800e76e-config-data\") pod \"nova-scheduler-0\" (UID: \"e5192153-c160-40bc-9375-a730e800e76e\") " pod="openstack/nova-scheduler-0" Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.970959 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc125541-0216-45f4-8276-520470f141fb-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"bc125541-0216-45f4-8276-520470f141fb\") " pod="openstack/nova-metadata-0" Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.971762 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5192153-c160-40bc-9375-a730e800e76e-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"e5192153-c160-40bc-9375-a730e800e76e\") " pod="openstack/nova-scheduler-0" Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.971801 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.973028 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.977169 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.981026 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc125541-0216-45f4-8276-520470f141fb-config-data\") pod \"nova-metadata-0\" (UID: \"bc125541-0216-45f4-8276-520470f141fb\") " pod="openstack/nova-metadata-0" Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.983785 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hwqn2\" (UniqueName: \"kubernetes.io/projected/bc125541-0216-45f4-8276-520470f141fb-kube-api-access-hwqn2\") pod \"nova-metadata-0\" (UID: \"bc125541-0216-45f4-8276-520470f141fb\") " pod="openstack/nova-metadata-0" Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.985008 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-68dbfc785f-kjm62"] Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.987126 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5kk9j\" (UniqueName: \"kubernetes.io/projected/e5192153-c160-40bc-9375-a730e800e76e-kube-api-access-5kk9j\") pod \"nova-scheduler-0\" (UID: \"e5192153-c160-40bc-9375-a730e800e76e\") " pod="openstack/nova-scheduler-0" Mar 13 15:40:12 crc kubenswrapper[4907]: I0313 15:40:12.001175 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 13 15:40:12 crc kubenswrapper[4907]: I0313 15:40:12.006721 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 13 15:40:12 crc kubenswrapper[4907]: I0313 15:40:12.025475 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 13 15:40:12 crc kubenswrapper[4907]: I0313 15:40:12.062858 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6244358c-3ba3-41fa-bf94-0b2bfb6ef956-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"6244358c-3ba3-41fa-bf94-0b2bfb6ef956\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 15:40:12 crc kubenswrapper[4907]: I0313 15:40:12.063193 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/690a67d3-7115-4897-ba26-edd5eeff0521-dns-svc\") pod \"dnsmasq-dns-68dbfc785f-kjm62\" (UID: \"690a67d3-7115-4897-ba26-edd5eeff0521\") " pod="openstack/dnsmasq-dns-68dbfc785f-kjm62" Mar 13 15:40:12 crc kubenswrapper[4907]: I0313 15:40:12.063219 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xl2nr\" (UniqueName: \"kubernetes.io/projected/6244358c-3ba3-41fa-bf94-0b2bfb6ef956-kube-api-access-xl2nr\") pod \"nova-cell1-novncproxy-0\" (UID: \"6244358c-3ba3-41fa-bf94-0b2bfb6ef956\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 15:40:12 crc kubenswrapper[4907]: I0313 15:40:12.063248 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6244358c-3ba3-41fa-bf94-0b2bfb6ef956-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"6244358c-3ba3-41fa-bf94-0b2bfb6ef956\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 15:40:12 crc kubenswrapper[4907]: I0313 15:40:12.063269 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/690a67d3-7115-4897-ba26-edd5eeff0521-config\") pod \"dnsmasq-dns-68dbfc785f-kjm62\" (UID: \"690a67d3-7115-4897-ba26-edd5eeff0521\") " pod="openstack/dnsmasq-dns-68dbfc785f-kjm62" Mar 13 15:40:12 crc kubenswrapper[4907]: I0313 15:40:12.063300 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/690a67d3-7115-4897-ba26-edd5eeff0521-ovsdbserver-sb\") pod \"dnsmasq-dns-68dbfc785f-kjm62\" (UID: \"690a67d3-7115-4897-ba26-edd5eeff0521\") " pod="openstack/dnsmasq-dns-68dbfc785f-kjm62" Mar 13 15:40:12 crc kubenswrapper[4907]: I0313 15:40:12.063320 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x55kq\" (UniqueName: \"kubernetes.io/projected/690a67d3-7115-4897-ba26-edd5eeff0521-kube-api-access-x55kq\") pod \"dnsmasq-dns-68dbfc785f-kjm62\" (UID: \"690a67d3-7115-4897-ba26-edd5eeff0521\") " pod="openstack/dnsmasq-dns-68dbfc785f-kjm62" Mar 13 15:40:12 crc kubenswrapper[4907]: I0313 15:40:12.063381 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/690a67d3-7115-4897-ba26-edd5eeff0521-ovsdbserver-nb\") pod \"dnsmasq-dns-68dbfc785f-kjm62\" (UID: \"690a67d3-7115-4897-ba26-edd5eeff0521\") " pod="openstack/dnsmasq-dns-68dbfc785f-kjm62" Mar 13 15:40:12 crc kubenswrapper[4907]: I0313 15:40:12.161381 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 13 15:40:12 crc kubenswrapper[4907]: I0313 15:40:12.165194 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/690a67d3-7115-4897-ba26-edd5eeff0521-ovsdbserver-sb\") pod \"dnsmasq-dns-68dbfc785f-kjm62\" (UID: \"690a67d3-7115-4897-ba26-edd5eeff0521\") " pod="openstack/dnsmasq-dns-68dbfc785f-kjm62" Mar 13 15:40:12 crc kubenswrapper[4907]: I0313 15:40:12.165252 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x55kq\" (UniqueName: \"kubernetes.io/projected/690a67d3-7115-4897-ba26-edd5eeff0521-kube-api-access-x55kq\") pod \"dnsmasq-dns-68dbfc785f-kjm62\" (UID: \"690a67d3-7115-4897-ba26-edd5eeff0521\") " pod="openstack/dnsmasq-dns-68dbfc785f-kjm62" Mar 13 15:40:12 crc kubenswrapper[4907]: I0313 15:40:12.165334 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/690a67d3-7115-4897-ba26-edd5eeff0521-ovsdbserver-nb\") pod \"dnsmasq-dns-68dbfc785f-kjm62\" (UID: \"690a67d3-7115-4897-ba26-edd5eeff0521\") " pod="openstack/dnsmasq-dns-68dbfc785f-kjm62" Mar 13 15:40:12 crc kubenswrapper[4907]: I0313 15:40:12.165401 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6244358c-3ba3-41fa-bf94-0b2bfb6ef956-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"6244358c-3ba3-41fa-bf94-0b2bfb6ef956\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 15:40:12 crc kubenswrapper[4907]: I0313 15:40:12.165470 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/690a67d3-7115-4897-ba26-edd5eeff0521-dns-svc\") pod \"dnsmasq-dns-68dbfc785f-kjm62\" (UID: \"690a67d3-7115-4897-ba26-edd5eeff0521\") " pod="openstack/dnsmasq-dns-68dbfc785f-kjm62" Mar 13 15:40:12 crc kubenswrapper[4907]: I0313 15:40:12.165502 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xl2nr\" (UniqueName: \"kubernetes.io/projected/6244358c-3ba3-41fa-bf94-0b2bfb6ef956-kube-api-access-xl2nr\") pod \"nova-cell1-novncproxy-0\" (UID: \"6244358c-3ba3-41fa-bf94-0b2bfb6ef956\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 15:40:12 crc kubenswrapper[4907]: I0313 15:40:12.165528 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6244358c-3ba3-41fa-bf94-0b2bfb6ef956-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"6244358c-3ba3-41fa-bf94-0b2bfb6ef956\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 15:40:12 crc kubenswrapper[4907]: I0313 15:40:12.165560 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/690a67d3-7115-4897-ba26-edd5eeff0521-config\") pod \"dnsmasq-dns-68dbfc785f-kjm62\" (UID: \"690a67d3-7115-4897-ba26-edd5eeff0521\") " pod="openstack/dnsmasq-dns-68dbfc785f-kjm62" Mar 13 15:40:12 crc kubenswrapper[4907]: I0313 15:40:12.167021 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/690a67d3-7115-4897-ba26-edd5eeff0521-config\") pod \"dnsmasq-dns-68dbfc785f-kjm62\" (UID: \"690a67d3-7115-4897-ba26-edd5eeff0521\") " pod="openstack/dnsmasq-dns-68dbfc785f-kjm62" Mar 13 15:40:12 crc kubenswrapper[4907]: I0313 15:40:12.167652 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/690a67d3-7115-4897-ba26-edd5eeff0521-ovsdbserver-sb\") pod \"dnsmasq-dns-68dbfc785f-kjm62\" (UID: \"690a67d3-7115-4897-ba26-edd5eeff0521\") " pod="openstack/dnsmasq-dns-68dbfc785f-kjm62" Mar 13 15:40:12 crc kubenswrapper[4907]: I0313 15:40:12.169053 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/690a67d3-7115-4897-ba26-edd5eeff0521-ovsdbserver-nb\") pod \"dnsmasq-dns-68dbfc785f-kjm62\" (UID: \"690a67d3-7115-4897-ba26-edd5eeff0521\") " pod="openstack/dnsmasq-dns-68dbfc785f-kjm62" Mar 13 15:40:12 crc kubenswrapper[4907]: I0313 15:40:12.169310 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/690a67d3-7115-4897-ba26-edd5eeff0521-dns-svc\") pod \"dnsmasq-dns-68dbfc785f-kjm62\" (UID: \"690a67d3-7115-4897-ba26-edd5eeff0521\") " pod="openstack/dnsmasq-dns-68dbfc785f-kjm62" Mar 13 15:40:12 crc kubenswrapper[4907]: I0313 15:40:12.175927 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6244358c-3ba3-41fa-bf94-0b2bfb6ef956-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"6244358c-3ba3-41fa-bf94-0b2bfb6ef956\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 15:40:12 crc kubenswrapper[4907]: I0313 15:40:12.185925 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6244358c-3ba3-41fa-bf94-0b2bfb6ef956-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"6244358c-3ba3-41fa-bf94-0b2bfb6ef956\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 15:40:12 crc kubenswrapper[4907]: I0313 15:40:12.186144 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xl2nr\" (UniqueName: \"kubernetes.io/projected/6244358c-3ba3-41fa-bf94-0b2bfb6ef956-kube-api-access-xl2nr\") pod \"nova-cell1-novncproxy-0\" (UID: \"6244358c-3ba3-41fa-bf94-0b2bfb6ef956\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 15:40:12 crc kubenswrapper[4907]: I0313 15:40:12.186196 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x55kq\" (UniqueName: \"kubernetes.io/projected/690a67d3-7115-4897-ba26-edd5eeff0521-kube-api-access-x55kq\") pod \"dnsmasq-dns-68dbfc785f-kjm62\" (UID: \"690a67d3-7115-4897-ba26-edd5eeff0521\") " pod="openstack/dnsmasq-dns-68dbfc785f-kjm62" Mar 13 15:40:12 crc kubenswrapper[4907]: I0313 15:40:12.319039 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-68dbfc785f-kjm62" Mar 13 15:40:12 crc kubenswrapper[4907]: I0313 15:40:12.329255 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Mar 13 15:40:12 crc kubenswrapper[4907]: I0313 15:40:12.491828 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-lvbcr"] Mar 13 15:40:12 crc kubenswrapper[4907]: I0313 15:40:12.599532 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-c2mhb"] Mar 13 15:40:12 crc kubenswrapper[4907]: I0313 15:40:12.600562 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-c2mhb" Mar 13 15:40:12 crc kubenswrapper[4907]: I0313 15:40:12.602652 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Mar 13 15:40:12 crc kubenswrapper[4907]: I0313 15:40:12.602842 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Mar 13 15:40:12 crc kubenswrapper[4907]: I0313 15:40:12.630249 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-c2mhb"] Mar 13 15:40:12 crc kubenswrapper[4907]: W0313 15:40:12.636377 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod188c81b4_cc35_444d_a5f1_f1633bae67a0.slice/crio-d9e105ef07cb22db95da3ee8ed6caefc5f9cc867634b8baf1d1dfb328f44cec9 WatchSource:0}: Error finding container d9e105ef07cb22db95da3ee8ed6caefc5f9cc867634b8baf1d1dfb328f44cec9: Status 404 returned error can't find the container with id d9e105ef07cb22db95da3ee8ed6caefc5f9cc867634b8baf1d1dfb328f44cec9 Mar 13 15:40:12 crc kubenswrapper[4907]: W0313 15:40:12.640043 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode5192153_c160_40bc_9375_a730e800e76e.slice/crio-a6f0a44a2f3b94386ebd8b9c603db045564e6d31764a5593228f18f153118242 WatchSource:0}: Error finding container a6f0a44a2f3b94386ebd8b9c603db045564e6d31764a5593228f18f153118242: Status 404 returned error can't find the container with id a6f0a44a2f3b94386ebd8b9c603db045564e6d31764a5593228f18f153118242 Mar 13 15:40:12 crc kubenswrapper[4907]: I0313 15:40:12.645483 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 13 15:40:12 crc kubenswrapper[4907]: I0313 15:40:12.662661 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 15:40:12 crc kubenswrapper[4907]: I0313 15:40:12.677907 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e935140e-65a6-48ff-9dda-fdf57edb9a4d-config-data\") pod \"nova-cell1-conductor-db-sync-c2mhb\" (UID: \"e935140e-65a6-48ff-9dda-fdf57edb9a4d\") " pod="openstack/nova-cell1-conductor-db-sync-c2mhb" Mar 13 15:40:12 crc kubenswrapper[4907]: I0313 15:40:12.677968 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-798cm\" (UniqueName: \"kubernetes.io/projected/e935140e-65a6-48ff-9dda-fdf57edb9a4d-kube-api-access-798cm\") pod \"nova-cell1-conductor-db-sync-c2mhb\" (UID: \"e935140e-65a6-48ff-9dda-fdf57edb9a4d\") " pod="openstack/nova-cell1-conductor-db-sync-c2mhb" Mar 13 15:40:12 crc kubenswrapper[4907]: I0313 15:40:12.678291 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e935140e-65a6-48ff-9dda-fdf57edb9a4d-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-c2mhb\" (UID: \"e935140e-65a6-48ff-9dda-fdf57edb9a4d\") " pod="openstack/nova-cell1-conductor-db-sync-c2mhb" Mar 13 15:40:12 crc kubenswrapper[4907]: I0313 15:40:12.678413 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e935140e-65a6-48ff-9dda-fdf57edb9a4d-scripts\") pod \"nova-cell1-conductor-db-sync-c2mhb\" (UID: \"e935140e-65a6-48ff-9dda-fdf57edb9a4d\") " pod="openstack/nova-cell1-conductor-db-sync-c2mhb" Mar 13 15:40:12 crc kubenswrapper[4907]: I0313 15:40:12.725470 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Mar 13 15:40:12 crc kubenswrapper[4907]: I0313 15:40:12.733357 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"188c81b4-cc35-444d-a5f1-f1633bae67a0","Type":"ContainerStarted","Data":"d9e105ef07cb22db95da3ee8ed6caefc5f9cc867634b8baf1d1dfb328f44cec9"} Mar 13 15:40:12 crc kubenswrapper[4907]: I0313 15:40:12.737254 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"e5192153-c160-40bc-9375-a730e800e76e","Type":"ContainerStarted","Data":"a6f0a44a2f3b94386ebd8b9c603db045564e6d31764a5593228f18f153118242"} Mar 13 15:40:12 crc kubenswrapper[4907]: I0313 15:40:12.738854 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-lvbcr" event={"ID":"89840ecf-ddb9-4a66-855b-0a1ae524bd6e","Type":"ContainerStarted","Data":"2e59f7eb348a0c98bd44ebc16e7f2b9007773ae7074558512509321e58dadcfc"} Mar 13 15:40:12 crc kubenswrapper[4907]: I0313 15:40:12.738920 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-lvbcr" event={"ID":"89840ecf-ddb9-4a66-855b-0a1ae524bd6e","Type":"ContainerStarted","Data":"59f93412cb46a9a06d3f2d6b59dd6e308e6731badd1d9e181954832e856363ec"} Mar 13 15:40:12 crc kubenswrapper[4907]: W0313 15:40:12.744181 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbc125541_0216_45f4_8276_520470f141fb.slice/crio-88a55c77e8165ad6fc6f2a546ffecd9fd6b597852354859b4d4ea8640528506f WatchSource:0}: Error finding container 88a55c77e8165ad6fc6f2a546ffecd9fd6b597852354859b4d4ea8640528506f: Status 404 returned error can't find the container with id 88a55c77e8165ad6fc6f2a546ffecd9fd6b597852354859b4d4ea8640528506f Mar 13 15:40:12 crc kubenswrapper[4907]: I0313 15:40:12.762220 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-lvbcr" podStartSLOduration=1.762199324 podStartE2EDuration="1.762199324s" podCreationTimestamp="2026-03-13 15:40:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:40:12.755658836 +0000 UTC m=+5711.655446525" watchObservedRunningTime="2026-03-13 15:40:12.762199324 +0000 UTC m=+5711.661987013" Mar 13 15:40:12 crc kubenswrapper[4907]: I0313 15:40:12.780636 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e935140e-65a6-48ff-9dda-fdf57edb9a4d-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-c2mhb\" (UID: \"e935140e-65a6-48ff-9dda-fdf57edb9a4d\") " pod="openstack/nova-cell1-conductor-db-sync-c2mhb" Mar 13 15:40:12 crc kubenswrapper[4907]: I0313 15:40:12.780717 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e935140e-65a6-48ff-9dda-fdf57edb9a4d-scripts\") pod \"nova-cell1-conductor-db-sync-c2mhb\" (UID: \"e935140e-65a6-48ff-9dda-fdf57edb9a4d\") " pod="openstack/nova-cell1-conductor-db-sync-c2mhb" Mar 13 15:40:12 crc kubenswrapper[4907]: I0313 15:40:12.780794 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e935140e-65a6-48ff-9dda-fdf57edb9a4d-config-data\") pod \"nova-cell1-conductor-db-sync-c2mhb\" (UID: \"e935140e-65a6-48ff-9dda-fdf57edb9a4d\") " pod="openstack/nova-cell1-conductor-db-sync-c2mhb" Mar 13 15:40:12 crc kubenswrapper[4907]: I0313 15:40:12.780838 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-798cm\" (UniqueName: \"kubernetes.io/projected/e935140e-65a6-48ff-9dda-fdf57edb9a4d-kube-api-access-798cm\") pod \"nova-cell1-conductor-db-sync-c2mhb\" (UID: \"e935140e-65a6-48ff-9dda-fdf57edb9a4d\") " pod="openstack/nova-cell1-conductor-db-sync-c2mhb" Mar 13 15:40:12 crc kubenswrapper[4907]: I0313 15:40:12.787210 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e935140e-65a6-48ff-9dda-fdf57edb9a4d-scripts\") pod \"nova-cell1-conductor-db-sync-c2mhb\" (UID: \"e935140e-65a6-48ff-9dda-fdf57edb9a4d\") " pod="openstack/nova-cell1-conductor-db-sync-c2mhb" Mar 13 15:40:12 crc kubenswrapper[4907]: I0313 15:40:12.787818 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e935140e-65a6-48ff-9dda-fdf57edb9a4d-config-data\") pod \"nova-cell1-conductor-db-sync-c2mhb\" (UID: \"e935140e-65a6-48ff-9dda-fdf57edb9a4d\") " pod="openstack/nova-cell1-conductor-db-sync-c2mhb" Mar 13 15:40:12 crc kubenswrapper[4907]: I0313 15:40:12.788216 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e935140e-65a6-48ff-9dda-fdf57edb9a4d-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-c2mhb\" (UID: \"e935140e-65a6-48ff-9dda-fdf57edb9a4d\") " pod="openstack/nova-cell1-conductor-db-sync-c2mhb" Mar 13 15:40:12 crc kubenswrapper[4907]: I0313 15:40:12.797229 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-798cm\" (UniqueName: \"kubernetes.io/projected/e935140e-65a6-48ff-9dda-fdf57edb9a4d-kube-api-access-798cm\") pod \"nova-cell1-conductor-db-sync-c2mhb\" (UID: \"e935140e-65a6-48ff-9dda-fdf57edb9a4d\") " pod="openstack/nova-cell1-conductor-db-sync-c2mhb" Mar 13 15:40:12 crc kubenswrapper[4907]: I0313 15:40:12.869974 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 13 15:40:12 crc kubenswrapper[4907]: I0313 15:40:12.881588 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-68dbfc785f-kjm62"] Mar 13 15:40:12 crc kubenswrapper[4907]: I0313 15:40:12.938434 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-c2mhb" Mar 13 15:40:13 crc kubenswrapper[4907]: W0313 15:40:13.454507 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode935140e_65a6_48ff_9dda_fdf57edb9a4d.slice/crio-3b30e6b6c78b39d41057f4a23478e828e3d6cb01d9da82b539e5864e0f6f7f4b WatchSource:0}: Error finding container 3b30e6b6c78b39d41057f4a23478e828e3d6cb01d9da82b539e5864e0f6f7f4b: Status 404 returned error can't find the container with id 3b30e6b6c78b39d41057f4a23478e828e3d6cb01d9da82b539e5864e0f6f7f4b Mar 13 15:40:13 crc kubenswrapper[4907]: I0313 15:40:13.455134 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-c2mhb"] Mar 13 15:40:13 crc kubenswrapper[4907]: I0313 15:40:13.751284 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"188c81b4-cc35-444d-a5f1-f1633bae67a0","Type":"ContainerStarted","Data":"74fbc4900d4783f6ccc5e6fc3b37fb82375ab950f69f896b8c992183e571e4e4"} Mar 13 15:40:13 crc kubenswrapper[4907]: I0313 15:40:13.751332 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"188c81b4-cc35-444d-a5f1-f1633bae67a0","Type":"ContainerStarted","Data":"08fb2b974765a711e6b2fa7513d50a1469142e3763a394f769251a01b141f79b"} Mar 13 15:40:13 crc kubenswrapper[4907]: I0313 15:40:13.753007 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"bc125541-0216-45f4-8276-520470f141fb","Type":"ContainerStarted","Data":"1ca5f3e9f3e3e7e3febfd017c86076ba77be5580f4a8b79cbb7bc8952ab76373"} Mar 13 15:40:13 crc kubenswrapper[4907]: I0313 15:40:13.753057 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"bc125541-0216-45f4-8276-520470f141fb","Type":"ContainerStarted","Data":"47b10daf3ce966a6fd75bc5149bce0256a8ca53a79611f00cde77948ce92e7ba"} Mar 13 15:40:13 crc kubenswrapper[4907]: I0313 15:40:13.753074 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"bc125541-0216-45f4-8276-520470f141fb","Type":"ContainerStarted","Data":"88a55c77e8165ad6fc6f2a546ffecd9fd6b597852354859b4d4ea8640528506f"} Mar 13 15:40:13 crc kubenswrapper[4907]: I0313 15:40:13.754169 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"e5192153-c160-40bc-9375-a730e800e76e","Type":"ContainerStarted","Data":"9ca8fe863388471bc1a4b09ca8e4b518970b726d2835186fd66e2f90d2cda6ca"} Mar 13 15:40:13 crc kubenswrapper[4907]: I0313 15:40:13.755352 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-c2mhb" event={"ID":"e935140e-65a6-48ff-9dda-fdf57edb9a4d","Type":"ContainerStarted","Data":"04b92be87927c1c66e4f4d3a575acbb3afb714960c495c7fe45aa93f70f86d7e"} Mar 13 15:40:13 crc kubenswrapper[4907]: I0313 15:40:13.755376 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-c2mhb" event={"ID":"e935140e-65a6-48ff-9dda-fdf57edb9a4d","Type":"ContainerStarted","Data":"3b30e6b6c78b39d41057f4a23478e828e3d6cb01d9da82b539e5864e0f6f7f4b"} Mar 13 15:40:13 crc kubenswrapper[4907]: I0313 15:40:13.757770 4907 generic.go:334] "Generic (PLEG): container finished" podID="690a67d3-7115-4897-ba26-edd5eeff0521" containerID="5f757938e0484b76f29e92050d686900e13810e4b83c689080ec66591a9b77ed" exitCode=0 Mar 13 15:40:13 crc kubenswrapper[4907]: I0313 15:40:13.757862 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-68dbfc785f-kjm62" event={"ID":"690a67d3-7115-4897-ba26-edd5eeff0521","Type":"ContainerDied","Data":"5f757938e0484b76f29e92050d686900e13810e4b83c689080ec66591a9b77ed"} Mar 13 15:40:13 crc kubenswrapper[4907]: I0313 15:40:13.757896 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-68dbfc785f-kjm62" event={"ID":"690a67d3-7115-4897-ba26-edd5eeff0521","Type":"ContainerStarted","Data":"4598382ec750564a06ecd379f171c64e71e4ac2c6f182c30bec9ff89463c8af3"} Mar 13 15:40:13 crc kubenswrapper[4907]: I0313 15:40:13.765715 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"6244358c-3ba3-41fa-bf94-0b2bfb6ef956","Type":"ContainerStarted","Data":"2ff24611962f784f04b97d2b6de0e1621289a9ce82da636a0133b47ec0d27610"} Mar 13 15:40:13 crc kubenswrapper[4907]: I0313 15:40:13.765755 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"6244358c-3ba3-41fa-bf94-0b2bfb6ef956","Type":"ContainerStarted","Data":"97ff1fdcfd5fd91a658ff67b990106cf5c21eac69b149055f325885b224d558e"} Mar 13 15:40:13 crc kubenswrapper[4907]: I0313 15:40:13.800013 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.799997609 podStartE2EDuration="2.799997609s" podCreationTimestamp="2026-03-13 15:40:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:40:13.799349581 +0000 UTC m=+5712.699137270" watchObservedRunningTime="2026-03-13 15:40:13.799997609 +0000 UTC m=+5712.699785298" Mar 13 15:40:13 crc kubenswrapper[4907]: I0313 15:40:13.857134 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.857117359 podStartE2EDuration="2.857117359s" podCreationTimestamp="2026-03-13 15:40:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:40:13.845022589 +0000 UTC m=+5712.744810278" watchObservedRunningTime="2026-03-13 15:40:13.857117359 +0000 UTC m=+5712.756905048" Mar 13 15:40:13 crc kubenswrapper[4907]: I0313 15:40:13.874807 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.874790152 podStartE2EDuration="2.874790152s" podCreationTimestamp="2026-03-13 15:40:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:40:13.871799721 +0000 UTC m=+5712.771587410" watchObservedRunningTime="2026-03-13 15:40:13.874790152 +0000 UTC m=+5712.774577841" Mar 13 15:40:13 crc kubenswrapper[4907]: I0313 15:40:13.925480 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-c2mhb" podStartSLOduration=1.925455747 podStartE2EDuration="1.925455747s" podCreationTimestamp="2026-03-13 15:40:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:40:13.921385865 +0000 UTC m=+5712.821173564" watchObservedRunningTime="2026-03-13 15:40:13.925455747 +0000 UTC m=+5712.825243436" Mar 13 15:40:13 crc kubenswrapper[4907]: I0313 15:40:13.949488 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.949468842 podStartE2EDuration="2.949468842s" podCreationTimestamp="2026-03-13 15:40:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:40:13.947327604 +0000 UTC m=+5712.847115313" watchObservedRunningTime="2026-03-13 15:40:13.949468842 +0000 UTC m=+5712.849256531" Mar 13 15:40:14 crc kubenswrapper[4907]: I0313 15:40:14.776229 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-68dbfc785f-kjm62" event={"ID":"690a67d3-7115-4897-ba26-edd5eeff0521","Type":"ContainerStarted","Data":"b0ec28b6beaa68b816d49abaf6a1073e21a0e38ace3bf382a509d710008100ce"} Mar 13 15:40:14 crc kubenswrapper[4907]: I0313 15:40:14.778053 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-68dbfc785f-kjm62" Mar 13 15:40:14 crc kubenswrapper[4907]: I0313 15:40:14.802656 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-68dbfc785f-kjm62" podStartSLOduration=3.8026372630000003 podStartE2EDuration="3.802637263s" podCreationTimestamp="2026-03-13 15:40:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:40:14.793543564 +0000 UTC m=+5713.693331253" watchObservedRunningTime="2026-03-13 15:40:14.802637263 +0000 UTC m=+5713.702424952" Mar 13 15:40:16 crc kubenswrapper[4907]: I0313 15:40:16.792537 4907 generic.go:334] "Generic (PLEG): container finished" podID="e935140e-65a6-48ff-9dda-fdf57edb9a4d" containerID="04b92be87927c1c66e4f4d3a575acbb3afb714960c495c7fe45aa93f70f86d7e" exitCode=0 Mar 13 15:40:16 crc kubenswrapper[4907]: I0313 15:40:16.792687 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-c2mhb" event={"ID":"e935140e-65a6-48ff-9dda-fdf57edb9a4d","Type":"ContainerDied","Data":"04b92be87927c1c66e4f4d3a575acbb3afb714960c495c7fe45aa93f70f86d7e"} Mar 13 15:40:17 crc kubenswrapper[4907]: I0313 15:40:17.007753 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Mar 13 15:40:17 crc kubenswrapper[4907]: I0313 15:40:17.329610 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Mar 13 15:40:17 crc kubenswrapper[4907]: I0313 15:40:17.802006 4907 generic.go:334] "Generic (PLEG): container finished" podID="89840ecf-ddb9-4a66-855b-0a1ae524bd6e" containerID="2e59f7eb348a0c98bd44ebc16e7f2b9007773ae7074558512509321e58dadcfc" exitCode=0 Mar 13 15:40:17 crc kubenswrapper[4907]: I0313 15:40:17.802099 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-lvbcr" event={"ID":"89840ecf-ddb9-4a66-855b-0a1ae524bd6e","Type":"ContainerDied","Data":"2e59f7eb348a0c98bd44ebc16e7f2b9007773ae7074558512509321e58dadcfc"} Mar 13 15:40:18 crc kubenswrapper[4907]: I0313 15:40:18.041714 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 15:40:18 crc kubenswrapper[4907]: I0313 15:40:18.042085 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 15:40:18 crc kubenswrapper[4907]: I0313 15:40:18.249133 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-c2mhb" Mar 13 15:40:18 crc kubenswrapper[4907]: I0313 15:40:18.294577 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e935140e-65a6-48ff-9dda-fdf57edb9a4d-scripts\") pod \"e935140e-65a6-48ff-9dda-fdf57edb9a4d\" (UID: \"e935140e-65a6-48ff-9dda-fdf57edb9a4d\") " Mar 13 15:40:18 crc kubenswrapper[4907]: I0313 15:40:18.294651 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e935140e-65a6-48ff-9dda-fdf57edb9a4d-config-data\") pod \"e935140e-65a6-48ff-9dda-fdf57edb9a4d\" (UID: \"e935140e-65a6-48ff-9dda-fdf57edb9a4d\") " Mar 13 15:40:18 crc kubenswrapper[4907]: I0313 15:40:18.294692 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e935140e-65a6-48ff-9dda-fdf57edb9a4d-combined-ca-bundle\") pod \"e935140e-65a6-48ff-9dda-fdf57edb9a4d\" (UID: \"e935140e-65a6-48ff-9dda-fdf57edb9a4d\") " Mar 13 15:40:18 crc kubenswrapper[4907]: I0313 15:40:18.294717 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-798cm\" (UniqueName: \"kubernetes.io/projected/e935140e-65a6-48ff-9dda-fdf57edb9a4d-kube-api-access-798cm\") pod \"e935140e-65a6-48ff-9dda-fdf57edb9a4d\" (UID: \"e935140e-65a6-48ff-9dda-fdf57edb9a4d\") " Mar 13 15:40:18 crc kubenswrapper[4907]: I0313 15:40:18.304016 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e935140e-65a6-48ff-9dda-fdf57edb9a4d-scripts" (OuterVolumeSpecName: "scripts") pod "e935140e-65a6-48ff-9dda-fdf57edb9a4d" (UID: "e935140e-65a6-48ff-9dda-fdf57edb9a4d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:40:18 crc kubenswrapper[4907]: I0313 15:40:18.304038 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e935140e-65a6-48ff-9dda-fdf57edb9a4d-kube-api-access-798cm" (OuterVolumeSpecName: "kube-api-access-798cm") pod "e935140e-65a6-48ff-9dda-fdf57edb9a4d" (UID: "e935140e-65a6-48ff-9dda-fdf57edb9a4d"). InnerVolumeSpecName "kube-api-access-798cm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:40:18 crc kubenswrapper[4907]: I0313 15:40:18.321505 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e935140e-65a6-48ff-9dda-fdf57edb9a4d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e935140e-65a6-48ff-9dda-fdf57edb9a4d" (UID: "e935140e-65a6-48ff-9dda-fdf57edb9a4d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:40:18 crc kubenswrapper[4907]: I0313 15:40:18.339830 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e935140e-65a6-48ff-9dda-fdf57edb9a4d-config-data" (OuterVolumeSpecName: "config-data") pod "e935140e-65a6-48ff-9dda-fdf57edb9a4d" (UID: "e935140e-65a6-48ff-9dda-fdf57edb9a4d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:40:18 crc kubenswrapper[4907]: I0313 15:40:18.397978 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e935140e-65a6-48ff-9dda-fdf57edb9a4d-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 15:40:18 crc kubenswrapper[4907]: I0313 15:40:18.398012 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e935140e-65a6-48ff-9dda-fdf57edb9a4d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 15:40:18 crc kubenswrapper[4907]: I0313 15:40:18.398027 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-798cm\" (UniqueName: \"kubernetes.io/projected/e935140e-65a6-48ff-9dda-fdf57edb9a4d-kube-api-access-798cm\") on node \"crc\" DevicePath \"\"" Mar 13 15:40:18 crc kubenswrapper[4907]: I0313 15:40:18.398041 4907 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e935140e-65a6-48ff-9dda-fdf57edb9a4d-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 15:40:18 crc kubenswrapper[4907]: I0313 15:40:18.813689 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-c2mhb" event={"ID":"e935140e-65a6-48ff-9dda-fdf57edb9a4d","Type":"ContainerDied","Data":"3b30e6b6c78b39d41057f4a23478e828e3d6cb01d9da82b539e5864e0f6f7f4b"} Mar 13 15:40:18 crc kubenswrapper[4907]: I0313 15:40:18.813745 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3b30e6b6c78b39d41057f4a23478e828e3d6cb01d9da82b539e5864e0f6f7f4b" Mar 13 15:40:18 crc kubenswrapper[4907]: I0313 15:40:18.813700 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-c2mhb" Mar 13 15:40:18 crc kubenswrapper[4907]: I0313 15:40:18.922254 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Mar 13 15:40:18 crc kubenswrapper[4907]: E0313 15:40:18.923336 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e935140e-65a6-48ff-9dda-fdf57edb9a4d" containerName="nova-cell1-conductor-db-sync" Mar 13 15:40:18 crc kubenswrapper[4907]: I0313 15:40:18.923356 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="e935140e-65a6-48ff-9dda-fdf57edb9a4d" containerName="nova-cell1-conductor-db-sync" Mar 13 15:40:18 crc kubenswrapper[4907]: I0313 15:40:18.923611 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="e935140e-65a6-48ff-9dda-fdf57edb9a4d" containerName="nova-cell1-conductor-db-sync" Mar 13 15:40:18 crc kubenswrapper[4907]: I0313 15:40:18.925444 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Mar 13 15:40:18 crc kubenswrapper[4907]: I0313 15:40:18.928042 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Mar 13 15:40:18 crc kubenswrapper[4907]: I0313 15:40:18.935102 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Mar 13 15:40:19 crc kubenswrapper[4907]: I0313 15:40:19.009339 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d8d0d2c8-d0f8-4dc1-b65e-e694571a6072-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"d8d0d2c8-d0f8-4dc1-b65e-e694571a6072\") " pod="openstack/nova-cell1-conductor-0" Mar 13 15:40:19 crc kubenswrapper[4907]: I0313 15:40:19.009392 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-47msf\" (UniqueName: \"kubernetes.io/projected/d8d0d2c8-d0f8-4dc1-b65e-e694571a6072-kube-api-access-47msf\") pod \"nova-cell1-conductor-0\" (UID: \"d8d0d2c8-d0f8-4dc1-b65e-e694571a6072\") " pod="openstack/nova-cell1-conductor-0" Mar 13 15:40:19 crc kubenswrapper[4907]: I0313 15:40:19.009525 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8d0d2c8-d0f8-4dc1-b65e-e694571a6072-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"d8d0d2c8-d0f8-4dc1-b65e-e694571a6072\") " pod="openstack/nova-cell1-conductor-0" Mar 13 15:40:19 crc kubenswrapper[4907]: I0313 15:40:19.111032 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d8d0d2c8-d0f8-4dc1-b65e-e694571a6072-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"d8d0d2c8-d0f8-4dc1-b65e-e694571a6072\") " pod="openstack/nova-cell1-conductor-0" Mar 13 15:40:19 crc kubenswrapper[4907]: I0313 15:40:19.111275 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-47msf\" (UniqueName: \"kubernetes.io/projected/d8d0d2c8-d0f8-4dc1-b65e-e694571a6072-kube-api-access-47msf\") pod \"nova-cell1-conductor-0\" (UID: \"d8d0d2c8-d0f8-4dc1-b65e-e694571a6072\") " pod="openstack/nova-cell1-conductor-0" Mar 13 15:40:19 crc kubenswrapper[4907]: I0313 15:40:19.111418 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8d0d2c8-d0f8-4dc1-b65e-e694571a6072-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"d8d0d2c8-d0f8-4dc1-b65e-e694571a6072\") " pod="openstack/nova-cell1-conductor-0" Mar 13 15:40:19 crc kubenswrapper[4907]: I0313 15:40:19.116269 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8d0d2c8-d0f8-4dc1-b65e-e694571a6072-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"d8d0d2c8-d0f8-4dc1-b65e-e694571a6072\") " pod="openstack/nova-cell1-conductor-0" Mar 13 15:40:19 crc kubenswrapper[4907]: I0313 15:40:19.116275 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d8d0d2c8-d0f8-4dc1-b65e-e694571a6072-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"d8d0d2c8-d0f8-4dc1-b65e-e694571a6072\") " pod="openstack/nova-cell1-conductor-0" Mar 13 15:40:19 crc kubenswrapper[4907]: I0313 15:40:19.127293 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-47msf\" (UniqueName: \"kubernetes.io/projected/d8d0d2c8-d0f8-4dc1-b65e-e694571a6072-kube-api-access-47msf\") pod \"nova-cell1-conductor-0\" (UID: \"d8d0d2c8-d0f8-4dc1-b65e-e694571a6072\") " pod="openstack/nova-cell1-conductor-0" Mar 13 15:40:19 crc kubenswrapper[4907]: I0313 15:40:19.185698 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-lvbcr" Mar 13 15:40:19 crc kubenswrapper[4907]: I0313 15:40:19.213361 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/89840ecf-ddb9-4a66-855b-0a1ae524bd6e-scripts\") pod \"89840ecf-ddb9-4a66-855b-0a1ae524bd6e\" (UID: \"89840ecf-ddb9-4a66-855b-0a1ae524bd6e\") " Mar 13 15:40:19 crc kubenswrapper[4907]: I0313 15:40:19.213845 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lvkgk\" (UniqueName: \"kubernetes.io/projected/89840ecf-ddb9-4a66-855b-0a1ae524bd6e-kube-api-access-lvkgk\") pod \"89840ecf-ddb9-4a66-855b-0a1ae524bd6e\" (UID: \"89840ecf-ddb9-4a66-855b-0a1ae524bd6e\") " Mar 13 15:40:19 crc kubenswrapper[4907]: I0313 15:40:19.213969 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/89840ecf-ddb9-4a66-855b-0a1ae524bd6e-combined-ca-bundle\") pod \"89840ecf-ddb9-4a66-855b-0a1ae524bd6e\" (UID: \"89840ecf-ddb9-4a66-855b-0a1ae524bd6e\") " Mar 13 15:40:19 crc kubenswrapper[4907]: I0313 15:40:19.214173 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/89840ecf-ddb9-4a66-855b-0a1ae524bd6e-config-data\") pod \"89840ecf-ddb9-4a66-855b-0a1ae524bd6e\" (UID: \"89840ecf-ddb9-4a66-855b-0a1ae524bd6e\") " Mar 13 15:40:19 crc kubenswrapper[4907]: I0313 15:40:19.217560 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/89840ecf-ddb9-4a66-855b-0a1ae524bd6e-scripts" (OuterVolumeSpecName: "scripts") pod "89840ecf-ddb9-4a66-855b-0a1ae524bd6e" (UID: "89840ecf-ddb9-4a66-855b-0a1ae524bd6e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:40:19 crc kubenswrapper[4907]: I0313 15:40:19.221050 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/89840ecf-ddb9-4a66-855b-0a1ae524bd6e-kube-api-access-lvkgk" (OuterVolumeSpecName: "kube-api-access-lvkgk") pod "89840ecf-ddb9-4a66-855b-0a1ae524bd6e" (UID: "89840ecf-ddb9-4a66-855b-0a1ae524bd6e"). InnerVolumeSpecName "kube-api-access-lvkgk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:40:19 crc kubenswrapper[4907]: I0313 15:40:19.241369 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/89840ecf-ddb9-4a66-855b-0a1ae524bd6e-config-data" (OuterVolumeSpecName: "config-data") pod "89840ecf-ddb9-4a66-855b-0a1ae524bd6e" (UID: "89840ecf-ddb9-4a66-855b-0a1ae524bd6e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:40:19 crc kubenswrapper[4907]: I0313 15:40:19.241438 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/89840ecf-ddb9-4a66-855b-0a1ae524bd6e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "89840ecf-ddb9-4a66-855b-0a1ae524bd6e" (UID: "89840ecf-ddb9-4a66-855b-0a1ae524bd6e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:40:19 crc kubenswrapper[4907]: I0313 15:40:19.248840 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Mar 13 15:40:19 crc kubenswrapper[4907]: I0313 15:40:19.316610 4907 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/89840ecf-ddb9-4a66-855b-0a1ae524bd6e-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 15:40:19 crc kubenswrapper[4907]: I0313 15:40:19.316644 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lvkgk\" (UniqueName: \"kubernetes.io/projected/89840ecf-ddb9-4a66-855b-0a1ae524bd6e-kube-api-access-lvkgk\") on node \"crc\" DevicePath \"\"" Mar 13 15:40:19 crc kubenswrapper[4907]: I0313 15:40:19.316657 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/89840ecf-ddb9-4a66-855b-0a1ae524bd6e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 15:40:19 crc kubenswrapper[4907]: I0313 15:40:19.316665 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/89840ecf-ddb9-4a66-855b-0a1ae524bd6e-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 15:40:19 crc kubenswrapper[4907]: I0313 15:40:19.684786 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Mar 13 15:40:19 crc kubenswrapper[4907]: I0313 15:40:19.823893 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"d8d0d2c8-d0f8-4dc1-b65e-e694571a6072","Type":"ContainerStarted","Data":"cbddd5aedb9a61c15acca2a6a299e779028d30303050e66d3356b0d2ec6abc1d"} Mar 13 15:40:19 crc kubenswrapper[4907]: I0313 15:40:19.828763 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-lvbcr" event={"ID":"89840ecf-ddb9-4a66-855b-0a1ae524bd6e","Type":"ContainerDied","Data":"59f93412cb46a9a06d3f2d6b59dd6e308e6731badd1d9e181954832e856363ec"} Mar 13 15:40:19 crc kubenswrapper[4907]: I0313 15:40:19.828804 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="59f93412cb46a9a06d3f2d6b59dd6e308e6731badd1d9e181954832e856363ec" Mar 13 15:40:19 crc kubenswrapper[4907]: I0313 15:40:19.828866 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-lvbcr" Mar 13 15:40:19 crc kubenswrapper[4907]: I0313 15:40:19.999235 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Mar 13 15:40:19 crc kubenswrapper[4907]: I0313 15:40:19.999784 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="188c81b4-cc35-444d-a5f1-f1633bae67a0" containerName="nova-api-log" containerID="cri-o://08fb2b974765a711e6b2fa7513d50a1469142e3763a394f769251a01b141f79b" gracePeriod=30 Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.000003 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="188c81b4-cc35-444d-a5f1-f1633bae67a0" containerName="nova-api-api" containerID="cri-o://74fbc4900d4783f6ccc5e6fc3b37fb82375ab950f69f896b8c992183e571e4e4" gracePeriod=30 Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.008577 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.008776 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="e5192153-c160-40bc-9375-a730e800e76e" containerName="nova-scheduler-scheduler" containerID="cri-o://9ca8fe863388471bc1a4b09ca8e4b518970b726d2835186fd66e2f90d2cda6ca" gracePeriod=30 Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.065512 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.065733 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="bc125541-0216-45f4-8276-520470f141fb" containerName="nova-metadata-log" containerID="cri-o://47b10daf3ce966a6fd75bc5149bce0256a8ca53a79611f00cde77948ce92e7ba" gracePeriod=30 Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.065800 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="bc125541-0216-45f4-8276-520470f141fb" containerName="nova-metadata-metadata" containerID="cri-o://1ca5f3e9f3e3e7e3febfd017c86076ba77be5580f4a8b79cbb7bc8952ab76373" gracePeriod=30 Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.529937 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.621524 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.646131 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/188c81b4-cc35-444d-a5f1-f1633bae67a0-config-data\") pod \"188c81b4-cc35-444d-a5f1-f1633bae67a0\" (UID: \"188c81b4-cc35-444d-a5f1-f1633bae67a0\") " Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.646213 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/188c81b4-cc35-444d-a5f1-f1633bae67a0-combined-ca-bundle\") pod \"188c81b4-cc35-444d-a5f1-f1633bae67a0\" (UID: \"188c81b4-cc35-444d-a5f1-f1633bae67a0\") " Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.646290 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6dmkv\" (UniqueName: \"kubernetes.io/projected/188c81b4-cc35-444d-a5f1-f1633bae67a0-kube-api-access-6dmkv\") pod \"188c81b4-cc35-444d-a5f1-f1633bae67a0\" (UID: \"188c81b4-cc35-444d-a5f1-f1633bae67a0\") " Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.646308 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/188c81b4-cc35-444d-a5f1-f1633bae67a0-logs\") pod \"188c81b4-cc35-444d-a5f1-f1633bae67a0\" (UID: \"188c81b4-cc35-444d-a5f1-f1633bae67a0\") " Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.647159 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/188c81b4-cc35-444d-a5f1-f1633bae67a0-logs" (OuterVolumeSpecName: "logs") pod "188c81b4-cc35-444d-a5f1-f1633bae67a0" (UID: "188c81b4-cc35-444d-a5f1-f1633bae67a0"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.651369 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/188c81b4-cc35-444d-a5f1-f1633bae67a0-kube-api-access-6dmkv" (OuterVolumeSpecName: "kube-api-access-6dmkv") pod "188c81b4-cc35-444d-a5f1-f1633bae67a0" (UID: "188c81b4-cc35-444d-a5f1-f1633bae67a0"). InnerVolumeSpecName "kube-api-access-6dmkv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.679104 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/188c81b4-cc35-444d-a5f1-f1633bae67a0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "188c81b4-cc35-444d-a5f1-f1633bae67a0" (UID: "188c81b4-cc35-444d-a5f1-f1633bae67a0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.689245 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/188c81b4-cc35-444d-a5f1-f1633bae67a0-config-data" (OuterVolumeSpecName: "config-data") pod "188c81b4-cc35-444d-a5f1-f1633bae67a0" (UID: "188c81b4-cc35-444d-a5f1-f1633bae67a0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.747912 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hwqn2\" (UniqueName: \"kubernetes.io/projected/bc125541-0216-45f4-8276-520470f141fb-kube-api-access-hwqn2\") pod \"bc125541-0216-45f4-8276-520470f141fb\" (UID: \"bc125541-0216-45f4-8276-520470f141fb\") " Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.748323 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc125541-0216-45f4-8276-520470f141fb-combined-ca-bundle\") pod \"bc125541-0216-45f4-8276-520470f141fb\" (UID: \"bc125541-0216-45f4-8276-520470f141fb\") " Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.748531 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc125541-0216-45f4-8276-520470f141fb-config-data\") pod \"bc125541-0216-45f4-8276-520470f141fb\" (UID: \"bc125541-0216-45f4-8276-520470f141fb\") " Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.748775 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bc125541-0216-45f4-8276-520470f141fb-logs\") pod \"bc125541-0216-45f4-8276-520470f141fb\" (UID: \"bc125541-0216-45f4-8276-520470f141fb\") " Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.749299 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc125541-0216-45f4-8276-520470f141fb-logs" (OuterVolumeSpecName: "logs") pod "bc125541-0216-45f4-8276-520470f141fb" (UID: "bc125541-0216-45f4-8276-520470f141fb"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.749779 4907 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bc125541-0216-45f4-8276-520470f141fb-logs\") on node \"crc\" DevicePath \"\"" Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.749936 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/188c81b4-cc35-444d-a5f1-f1633bae67a0-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.750078 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/188c81b4-cc35-444d-a5f1-f1633bae67a0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.750170 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6dmkv\" (UniqueName: \"kubernetes.io/projected/188c81b4-cc35-444d-a5f1-f1633bae67a0-kube-api-access-6dmkv\") on node \"crc\" DevicePath \"\"" Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.750251 4907 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/188c81b4-cc35-444d-a5f1-f1633bae67a0-logs\") on node \"crc\" DevicePath \"\"" Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.753020 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc125541-0216-45f4-8276-520470f141fb-kube-api-access-hwqn2" (OuterVolumeSpecName: "kube-api-access-hwqn2") pod "bc125541-0216-45f4-8276-520470f141fb" (UID: "bc125541-0216-45f4-8276-520470f141fb"). InnerVolumeSpecName "kube-api-access-hwqn2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.771809 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc125541-0216-45f4-8276-520470f141fb-config-data" (OuterVolumeSpecName: "config-data") pod "bc125541-0216-45f4-8276-520470f141fb" (UID: "bc125541-0216-45f4-8276-520470f141fb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.776163 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc125541-0216-45f4-8276-520470f141fb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bc125541-0216-45f4-8276-520470f141fb" (UID: "bc125541-0216-45f4-8276-520470f141fb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.840705 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"d8d0d2c8-d0f8-4dc1-b65e-e694571a6072","Type":"ContainerStarted","Data":"716a10e355d921d8594fa1927d76d58c79889f5d1c1505d7df88ab1927fd9db4"} Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.841566 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.843873 4907 generic.go:334] "Generic (PLEG): container finished" podID="188c81b4-cc35-444d-a5f1-f1633bae67a0" containerID="74fbc4900d4783f6ccc5e6fc3b37fb82375ab950f69f896b8c992183e571e4e4" exitCode=0 Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.843933 4907 generic.go:334] "Generic (PLEG): container finished" podID="188c81b4-cc35-444d-a5f1-f1633bae67a0" containerID="08fb2b974765a711e6b2fa7513d50a1469142e3763a394f769251a01b141f79b" exitCode=143 Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.843943 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"188c81b4-cc35-444d-a5f1-f1633bae67a0","Type":"ContainerDied","Data":"74fbc4900d4783f6ccc5e6fc3b37fb82375ab950f69f896b8c992183e571e4e4"} Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.844062 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"188c81b4-cc35-444d-a5f1-f1633bae67a0","Type":"ContainerDied","Data":"08fb2b974765a711e6b2fa7513d50a1469142e3763a394f769251a01b141f79b"} Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.844086 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"188c81b4-cc35-444d-a5f1-f1633bae67a0","Type":"ContainerDied","Data":"d9e105ef07cb22db95da3ee8ed6caefc5f9cc867634b8baf1d1dfb328f44cec9"} Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.844145 4907 scope.go:117] "RemoveContainer" containerID="74fbc4900d4783f6ccc5e6fc3b37fb82375ab950f69f896b8c992183e571e4e4" Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.846350 4907 generic.go:334] "Generic (PLEG): container finished" podID="bc125541-0216-45f4-8276-520470f141fb" containerID="1ca5f3e9f3e3e7e3febfd017c86076ba77be5580f4a8b79cbb7bc8952ab76373" exitCode=0 Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.846372 4907 generic.go:334] "Generic (PLEG): container finished" podID="bc125541-0216-45f4-8276-520470f141fb" containerID="47b10daf3ce966a6fd75bc5149bce0256a8ca53a79611f00cde77948ce92e7ba" exitCode=143 Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.846393 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"bc125541-0216-45f4-8276-520470f141fb","Type":"ContainerDied","Data":"1ca5f3e9f3e3e7e3febfd017c86076ba77be5580f4a8b79cbb7bc8952ab76373"} Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.846415 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"bc125541-0216-45f4-8276-520470f141fb","Type":"ContainerDied","Data":"47b10daf3ce966a6fd75bc5149bce0256a8ca53a79611f00cde77948ce92e7ba"} Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.846425 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"bc125541-0216-45f4-8276-520470f141fb","Type":"ContainerDied","Data":"88a55c77e8165ad6fc6f2a546ffecd9fd6b597852354859b4d4ea8640528506f"} Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.846468 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.851067 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.852562 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hwqn2\" (UniqueName: \"kubernetes.io/projected/bc125541-0216-45f4-8276-520470f141fb-kube-api-access-hwqn2\") on node \"crc\" DevicePath \"\"" Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.852587 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc125541-0216-45f4-8276-520470f141fb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.852596 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc125541-0216-45f4-8276-520470f141fb-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.861510 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.861491661 podStartE2EDuration="2.861491661s" podCreationTimestamp="2026-03-13 15:40:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:40:20.858172451 +0000 UTC m=+5719.757960140" watchObservedRunningTime="2026-03-13 15:40:20.861491661 +0000 UTC m=+5719.761279350" Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.881718 4907 scope.go:117] "RemoveContainer" containerID="08fb2b974765a711e6b2fa7513d50a1469142e3763a394f769251a01b141f79b" Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.899482 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.914504 4907 scope.go:117] "RemoveContainer" containerID="74fbc4900d4783f6ccc5e6fc3b37fb82375ab950f69f896b8c992183e571e4e4" Mar 13 15:40:20 crc kubenswrapper[4907]: E0313 15:40:20.915494 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"74fbc4900d4783f6ccc5e6fc3b37fb82375ab950f69f896b8c992183e571e4e4\": container with ID starting with 74fbc4900d4783f6ccc5e6fc3b37fb82375ab950f69f896b8c992183e571e4e4 not found: ID does not exist" containerID="74fbc4900d4783f6ccc5e6fc3b37fb82375ab950f69f896b8c992183e571e4e4" Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.915568 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"74fbc4900d4783f6ccc5e6fc3b37fb82375ab950f69f896b8c992183e571e4e4"} err="failed to get container status \"74fbc4900d4783f6ccc5e6fc3b37fb82375ab950f69f896b8c992183e571e4e4\": rpc error: code = NotFound desc = could not find container \"74fbc4900d4783f6ccc5e6fc3b37fb82375ab950f69f896b8c992183e571e4e4\": container with ID starting with 74fbc4900d4783f6ccc5e6fc3b37fb82375ab950f69f896b8c992183e571e4e4 not found: ID does not exist" Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.915627 4907 scope.go:117] "RemoveContainer" containerID="08fb2b974765a711e6b2fa7513d50a1469142e3763a394f769251a01b141f79b" Mar 13 15:40:20 crc kubenswrapper[4907]: E0313 15:40:20.916270 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"08fb2b974765a711e6b2fa7513d50a1469142e3763a394f769251a01b141f79b\": container with ID starting with 08fb2b974765a711e6b2fa7513d50a1469142e3763a394f769251a01b141f79b not found: ID does not exist" containerID="08fb2b974765a711e6b2fa7513d50a1469142e3763a394f769251a01b141f79b" Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.916335 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"08fb2b974765a711e6b2fa7513d50a1469142e3763a394f769251a01b141f79b"} err="failed to get container status \"08fb2b974765a711e6b2fa7513d50a1469142e3763a394f769251a01b141f79b\": rpc error: code = NotFound desc = could not find container \"08fb2b974765a711e6b2fa7513d50a1469142e3763a394f769251a01b141f79b\": container with ID starting with 08fb2b974765a711e6b2fa7513d50a1469142e3763a394f769251a01b141f79b not found: ID does not exist" Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.916353 4907 scope.go:117] "RemoveContainer" containerID="74fbc4900d4783f6ccc5e6fc3b37fb82375ab950f69f896b8c992183e571e4e4" Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.917253 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"74fbc4900d4783f6ccc5e6fc3b37fb82375ab950f69f896b8c992183e571e4e4"} err="failed to get container status \"74fbc4900d4783f6ccc5e6fc3b37fb82375ab950f69f896b8c992183e571e4e4\": rpc error: code = NotFound desc = could not find container \"74fbc4900d4783f6ccc5e6fc3b37fb82375ab950f69f896b8c992183e571e4e4\": container with ID starting with 74fbc4900d4783f6ccc5e6fc3b37fb82375ab950f69f896b8c992183e571e4e4 not found: ID does not exist" Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.917280 4907 scope.go:117] "RemoveContainer" containerID="08fb2b974765a711e6b2fa7513d50a1469142e3763a394f769251a01b141f79b" Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.917492 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"08fb2b974765a711e6b2fa7513d50a1469142e3763a394f769251a01b141f79b"} err="failed to get container status \"08fb2b974765a711e6b2fa7513d50a1469142e3763a394f769251a01b141f79b\": rpc error: code = NotFound desc = could not find container \"08fb2b974765a711e6b2fa7513d50a1469142e3763a394f769251a01b141f79b\": container with ID starting with 08fb2b974765a711e6b2fa7513d50a1469142e3763a394f769251a01b141f79b not found: ID does not exist" Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.917513 4907 scope.go:117] "RemoveContainer" containerID="1ca5f3e9f3e3e7e3febfd017c86076ba77be5580f4a8b79cbb7bc8952ab76373" Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.920034 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.932787 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Mar 13 15:40:20 crc kubenswrapper[4907]: E0313 15:40:20.933311 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="188c81b4-cc35-444d-a5f1-f1633bae67a0" containerName="nova-api-api" Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.933335 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="188c81b4-cc35-444d-a5f1-f1633bae67a0" containerName="nova-api-api" Mar 13 15:40:20 crc kubenswrapper[4907]: E0313 15:40:20.933351 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="188c81b4-cc35-444d-a5f1-f1633bae67a0" containerName="nova-api-log" Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.933359 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="188c81b4-cc35-444d-a5f1-f1633bae67a0" containerName="nova-api-log" Mar 13 15:40:20 crc kubenswrapper[4907]: E0313 15:40:20.933372 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89840ecf-ddb9-4a66-855b-0a1ae524bd6e" containerName="nova-manage" Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.933379 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="89840ecf-ddb9-4a66-855b-0a1ae524bd6e" containerName="nova-manage" Mar 13 15:40:20 crc kubenswrapper[4907]: E0313 15:40:20.933396 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc125541-0216-45f4-8276-520470f141fb" containerName="nova-metadata-metadata" Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.933404 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc125541-0216-45f4-8276-520470f141fb" containerName="nova-metadata-metadata" Mar 13 15:40:20 crc kubenswrapper[4907]: E0313 15:40:20.933433 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc125541-0216-45f4-8276-520470f141fb" containerName="nova-metadata-log" Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.933441 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc125541-0216-45f4-8276-520470f141fb" containerName="nova-metadata-log" Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.933647 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="188c81b4-cc35-444d-a5f1-f1633bae67a0" containerName="nova-api-log" Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.933664 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="188c81b4-cc35-444d-a5f1-f1633bae67a0" containerName="nova-api-api" Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.933678 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc125541-0216-45f4-8276-520470f141fb" containerName="nova-metadata-metadata" Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.933696 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="89840ecf-ddb9-4a66-855b-0a1ae524bd6e" containerName="nova-manage" Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.933720 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc125541-0216-45f4-8276-520470f141fb" containerName="nova-metadata-log" Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.935026 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.942610 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.944248 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.949541 4907 scope.go:117] "RemoveContainer" containerID="47b10daf3ce966a6fd75bc5149bce0256a8ca53a79611f00cde77948ce92e7ba" Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.960604 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.980242 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.984777 4907 scope.go:117] "RemoveContainer" containerID="1ca5f3e9f3e3e7e3febfd017c86076ba77be5580f4a8b79cbb7bc8952ab76373" Mar 13 15:40:20 crc kubenswrapper[4907]: E0313 15:40:20.985276 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1ca5f3e9f3e3e7e3febfd017c86076ba77be5580f4a8b79cbb7bc8952ab76373\": container with ID starting with 1ca5f3e9f3e3e7e3febfd017c86076ba77be5580f4a8b79cbb7bc8952ab76373 not found: ID does not exist" containerID="1ca5f3e9f3e3e7e3febfd017c86076ba77be5580f4a8b79cbb7bc8952ab76373" Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.985310 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ca5f3e9f3e3e7e3febfd017c86076ba77be5580f4a8b79cbb7bc8952ab76373"} err="failed to get container status \"1ca5f3e9f3e3e7e3febfd017c86076ba77be5580f4a8b79cbb7bc8952ab76373\": rpc error: code = NotFound desc = could not find container \"1ca5f3e9f3e3e7e3febfd017c86076ba77be5580f4a8b79cbb7bc8952ab76373\": container with ID starting with 1ca5f3e9f3e3e7e3febfd017c86076ba77be5580f4a8b79cbb7bc8952ab76373 not found: ID does not exist" Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.985333 4907 scope.go:117] "RemoveContainer" containerID="47b10daf3ce966a6fd75bc5149bce0256a8ca53a79611f00cde77948ce92e7ba" Mar 13 15:40:20 crc kubenswrapper[4907]: E0313 15:40:20.985725 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"47b10daf3ce966a6fd75bc5149bce0256a8ca53a79611f00cde77948ce92e7ba\": container with ID starting with 47b10daf3ce966a6fd75bc5149bce0256a8ca53a79611f00cde77948ce92e7ba not found: ID does not exist" containerID="47b10daf3ce966a6fd75bc5149bce0256a8ca53a79611f00cde77948ce92e7ba" Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.985745 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"47b10daf3ce966a6fd75bc5149bce0256a8ca53a79611f00cde77948ce92e7ba"} err="failed to get container status \"47b10daf3ce966a6fd75bc5149bce0256a8ca53a79611f00cde77948ce92e7ba\": rpc error: code = NotFound desc = could not find container \"47b10daf3ce966a6fd75bc5149bce0256a8ca53a79611f00cde77948ce92e7ba\": container with ID starting with 47b10daf3ce966a6fd75bc5149bce0256a8ca53a79611f00cde77948ce92e7ba not found: ID does not exist" Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.985759 4907 scope.go:117] "RemoveContainer" containerID="1ca5f3e9f3e3e7e3febfd017c86076ba77be5580f4a8b79cbb7bc8952ab76373" Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.986109 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ca5f3e9f3e3e7e3febfd017c86076ba77be5580f4a8b79cbb7bc8952ab76373"} err="failed to get container status \"1ca5f3e9f3e3e7e3febfd017c86076ba77be5580f4a8b79cbb7bc8952ab76373\": rpc error: code = NotFound desc = could not find container \"1ca5f3e9f3e3e7e3febfd017c86076ba77be5580f4a8b79cbb7bc8952ab76373\": container with ID starting with 1ca5f3e9f3e3e7e3febfd017c86076ba77be5580f4a8b79cbb7bc8952ab76373 not found: ID does not exist" Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.986138 4907 scope.go:117] "RemoveContainer" containerID="47b10daf3ce966a6fd75bc5149bce0256a8ca53a79611f00cde77948ce92e7ba" Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.986451 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"47b10daf3ce966a6fd75bc5149bce0256a8ca53a79611f00cde77948ce92e7ba"} err="failed to get container status \"47b10daf3ce966a6fd75bc5149bce0256a8ca53a79611f00cde77948ce92e7ba\": rpc error: code = NotFound desc = could not find container \"47b10daf3ce966a6fd75bc5149bce0256a8ca53a79611f00cde77948ce92e7ba\": container with ID starting with 47b10daf3ce966a6fd75bc5149bce0256a8ca53a79611f00cde77948ce92e7ba not found: ID does not exist" Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.994131 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.995738 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.997907 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.997976 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 13 15:40:21 crc kubenswrapper[4907]: I0313 15:40:21.055254 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x6msv\" (UniqueName: \"kubernetes.io/projected/ad3c8be4-d237-4c1e-bd98-963acae4ac32-kube-api-access-x6msv\") pod \"nova-api-0\" (UID: \"ad3c8be4-d237-4c1e-bd98-963acae4ac32\") " pod="openstack/nova-api-0" Mar 13 15:40:21 crc kubenswrapper[4907]: I0313 15:40:21.055303 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0\") " pod="openstack/nova-metadata-0" Mar 13 15:40:21 crc kubenswrapper[4907]: I0313 15:40:21.055330 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ad3c8be4-d237-4c1e-bd98-963acae4ac32-logs\") pod \"nova-api-0\" (UID: \"ad3c8be4-d237-4c1e-bd98-963acae4ac32\") " pod="openstack/nova-api-0" Mar 13 15:40:21 crc kubenswrapper[4907]: I0313 15:40:21.055428 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad3c8be4-d237-4c1e-bd98-963acae4ac32-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"ad3c8be4-d237-4c1e-bd98-963acae4ac32\") " pod="openstack/nova-api-0" Mar 13 15:40:21 crc kubenswrapper[4907]: I0313 15:40:21.055562 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0-logs\") pod \"nova-metadata-0\" (UID: \"8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0\") " pod="openstack/nova-metadata-0" Mar 13 15:40:21 crc kubenswrapper[4907]: I0313 15:40:21.055617 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad3c8be4-d237-4c1e-bd98-963acae4ac32-config-data\") pod \"nova-api-0\" (UID: \"ad3c8be4-d237-4c1e-bd98-963acae4ac32\") " pod="openstack/nova-api-0" Mar 13 15:40:21 crc kubenswrapper[4907]: I0313 15:40:21.055715 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nm2jk\" (UniqueName: \"kubernetes.io/projected/8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0-kube-api-access-nm2jk\") pod \"nova-metadata-0\" (UID: \"8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0\") " pod="openstack/nova-metadata-0" Mar 13 15:40:21 crc kubenswrapper[4907]: I0313 15:40:21.055840 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0-config-data\") pod \"nova-metadata-0\" (UID: \"8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0\") " pod="openstack/nova-metadata-0" Mar 13 15:40:21 crc kubenswrapper[4907]: I0313 15:40:21.158079 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nm2jk\" (UniqueName: \"kubernetes.io/projected/8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0-kube-api-access-nm2jk\") pod \"nova-metadata-0\" (UID: \"8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0\") " pod="openstack/nova-metadata-0" Mar 13 15:40:21 crc kubenswrapper[4907]: I0313 15:40:21.158140 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0-config-data\") pod \"nova-metadata-0\" (UID: \"8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0\") " pod="openstack/nova-metadata-0" Mar 13 15:40:21 crc kubenswrapper[4907]: I0313 15:40:21.158203 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x6msv\" (UniqueName: \"kubernetes.io/projected/ad3c8be4-d237-4c1e-bd98-963acae4ac32-kube-api-access-x6msv\") pod \"nova-api-0\" (UID: \"ad3c8be4-d237-4c1e-bd98-963acae4ac32\") " pod="openstack/nova-api-0" Mar 13 15:40:21 crc kubenswrapper[4907]: I0313 15:40:21.158221 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0\") " pod="openstack/nova-metadata-0" Mar 13 15:40:21 crc kubenswrapper[4907]: I0313 15:40:21.158247 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ad3c8be4-d237-4c1e-bd98-963acae4ac32-logs\") pod \"nova-api-0\" (UID: \"ad3c8be4-d237-4c1e-bd98-963acae4ac32\") " pod="openstack/nova-api-0" Mar 13 15:40:21 crc kubenswrapper[4907]: I0313 15:40:21.158288 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad3c8be4-d237-4c1e-bd98-963acae4ac32-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"ad3c8be4-d237-4c1e-bd98-963acae4ac32\") " pod="openstack/nova-api-0" Mar 13 15:40:21 crc kubenswrapper[4907]: I0313 15:40:21.158583 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0-logs\") pod \"nova-metadata-0\" (UID: \"8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0\") " pod="openstack/nova-metadata-0" Mar 13 15:40:21 crc kubenswrapper[4907]: I0313 15:40:21.158603 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad3c8be4-d237-4c1e-bd98-963acae4ac32-config-data\") pod \"nova-api-0\" (UID: \"ad3c8be4-d237-4c1e-bd98-963acae4ac32\") " pod="openstack/nova-api-0" Mar 13 15:40:21 crc kubenswrapper[4907]: I0313 15:40:21.158791 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ad3c8be4-d237-4c1e-bd98-963acae4ac32-logs\") pod \"nova-api-0\" (UID: \"ad3c8be4-d237-4c1e-bd98-963acae4ac32\") " pod="openstack/nova-api-0" Mar 13 15:40:21 crc kubenswrapper[4907]: I0313 15:40:21.159204 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0-logs\") pod \"nova-metadata-0\" (UID: \"8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0\") " pod="openstack/nova-metadata-0" Mar 13 15:40:21 crc kubenswrapper[4907]: I0313 15:40:21.162743 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0-config-data\") pod \"nova-metadata-0\" (UID: \"8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0\") " pod="openstack/nova-metadata-0" Mar 13 15:40:21 crc kubenswrapper[4907]: I0313 15:40:21.163349 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad3c8be4-d237-4c1e-bd98-963acae4ac32-config-data\") pod \"nova-api-0\" (UID: \"ad3c8be4-d237-4c1e-bd98-963acae4ac32\") " pod="openstack/nova-api-0" Mar 13 15:40:21 crc kubenswrapper[4907]: I0313 15:40:21.164237 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0\") " pod="openstack/nova-metadata-0" Mar 13 15:40:21 crc kubenswrapper[4907]: I0313 15:40:21.174660 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad3c8be4-d237-4c1e-bd98-963acae4ac32-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"ad3c8be4-d237-4c1e-bd98-963acae4ac32\") " pod="openstack/nova-api-0" Mar 13 15:40:21 crc kubenswrapper[4907]: I0313 15:40:21.175097 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nm2jk\" (UniqueName: \"kubernetes.io/projected/8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0-kube-api-access-nm2jk\") pod \"nova-metadata-0\" (UID: \"8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0\") " pod="openstack/nova-metadata-0" Mar 13 15:40:21 crc kubenswrapper[4907]: I0313 15:40:21.178926 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x6msv\" (UniqueName: \"kubernetes.io/projected/ad3c8be4-d237-4c1e-bd98-963acae4ac32-kube-api-access-x6msv\") pod \"nova-api-0\" (UID: \"ad3c8be4-d237-4c1e-bd98-963acae4ac32\") " pod="openstack/nova-api-0" Mar 13 15:40:21 crc kubenswrapper[4907]: I0313 15:40:21.271107 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 13 15:40:21 crc kubenswrapper[4907]: I0313 15:40:21.319956 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 13 15:40:21 crc kubenswrapper[4907]: I0313 15:40:21.718862 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Mar 13 15:40:21 crc kubenswrapper[4907]: I0313 15:40:21.797239 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="188c81b4-cc35-444d-a5f1-f1633bae67a0" path="/var/lib/kubelet/pods/188c81b4-cc35-444d-a5f1-f1633bae67a0/volumes" Mar 13 15:40:21 crc kubenswrapper[4907]: I0313 15:40:21.798804 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc125541-0216-45f4-8276-520470f141fb" path="/var/lib/kubelet/pods/bc125541-0216-45f4-8276-520470f141fb/volumes" Mar 13 15:40:21 crc kubenswrapper[4907]: I0313 15:40:21.799946 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 13 15:40:21 crc kubenswrapper[4907]: I0313 15:40:21.854586 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0","Type":"ContainerStarted","Data":"0d9dea632a506015513087187b39d9560b7d9208b5328332f68e9033c985b609"} Mar 13 15:40:21 crc kubenswrapper[4907]: I0313 15:40:21.858316 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ad3c8be4-d237-4c1e-bd98-963acae4ac32","Type":"ContainerStarted","Data":"7d15b64a01f946210967bf64f6d3e77361db086303536eb844ce8563c1a6da6d"} Mar 13 15:40:22 crc kubenswrapper[4907]: I0313 15:40:22.320086 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-68dbfc785f-kjm62" Mar 13 15:40:22 crc kubenswrapper[4907]: I0313 15:40:22.330483 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Mar 13 15:40:22 crc kubenswrapper[4907]: I0313 15:40:22.346176 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Mar 13 15:40:22 crc kubenswrapper[4907]: I0313 15:40:22.399778 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5dd7c58f95-579qq"] Mar 13 15:40:22 crc kubenswrapper[4907]: I0313 15:40:22.400954 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5dd7c58f95-579qq" podUID="5153eb33-98f7-4452-b4f5-751a6d49c54b" containerName="dnsmasq-dns" containerID="cri-o://4d47e9b4f5705f84e359bebffdaf6b40c2a5a7d780d8daa8048b0a559dabed59" gracePeriod=10 Mar 13 15:40:22 crc kubenswrapper[4907]: I0313 15:40:22.874111 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5dd7c58f95-579qq" Mar 13 15:40:22 crc kubenswrapper[4907]: I0313 15:40:22.881222 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ad3c8be4-d237-4c1e-bd98-963acae4ac32","Type":"ContainerStarted","Data":"e833237c58d59cd7a382ff7644c38c3979cbe4fa6a8a9e044f9040c86f3823df"} Mar 13 15:40:22 crc kubenswrapper[4907]: I0313 15:40:22.881259 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ad3c8be4-d237-4c1e-bd98-963acae4ac32","Type":"ContainerStarted","Data":"a9b52fcfd1633f1cfd4fe4e130d7daf6a6a0f261d982b7490f4c49360b713147"} Mar 13 15:40:22 crc kubenswrapper[4907]: I0313 15:40:22.884620 4907 generic.go:334] "Generic (PLEG): container finished" podID="5153eb33-98f7-4452-b4f5-751a6d49c54b" containerID="4d47e9b4f5705f84e359bebffdaf6b40c2a5a7d780d8daa8048b0a559dabed59" exitCode=0 Mar 13 15:40:22 crc kubenswrapper[4907]: I0313 15:40:22.884679 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5dd7c58f95-579qq" event={"ID":"5153eb33-98f7-4452-b4f5-751a6d49c54b","Type":"ContainerDied","Data":"4d47e9b4f5705f84e359bebffdaf6b40c2a5a7d780d8daa8048b0a559dabed59"} Mar 13 15:40:22 crc kubenswrapper[4907]: I0313 15:40:22.884849 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5dd7c58f95-579qq" Mar 13 15:40:22 crc kubenswrapper[4907]: I0313 15:40:22.885083 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5dd7c58f95-579qq" event={"ID":"5153eb33-98f7-4452-b4f5-751a6d49c54b","Type":"ContainerDied","Data":"1f607dd9ca9be04a0a9e3a628ae846054d952f5ec49fd5e3e8f6f1422eb9fe8b"} Mar 13 15:40:22 crc kubenswrapper[4907]: I0313 15:40:22.885106 4907 scope.go:117] "RemoveContainer" containerID="4d47e9b4f5705f84e359bebffdaf6b40c2a5a7d780d8daa8048b0a559dabed59" Mar 13 15:40:22 crc kubenswrapper[4907]: I0313 15:40:22.889059 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0","Type":"ContainerStarted","Data":"9d4f1d5de2313cc0cacaa0de90599bf7cc29209144c55b94e7e5f63f34cd569d"} Mar 13 15:40:22 crc kubenswrapper[4907]: I0313 15:40:22.889083 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0","Type":"ContainerStarted","Data":"12ad1f3ba8429b32fe4517953f072fa40cf2ef694f54ab5cea4b98ef8339414c"} Mar 13 15:40:22 crc kubenswrapper[4907]: I0313 15:40:22.901790 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Mar 13 15:40:22 crc kubenswrapper[4907]: I0313 15:40:22.938010 4907 scope.go:117] "RemoveContainer" containerID="dc68df440400c3300857779e0180b105a0536c85adecc574069e0c384c4a2b78" Mar 13 15:40:22 crc kubenswrapper[4907]: I0313 15:40:22.948447 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.94842752 podStartE2EDuration="2.94842752s" podCreationTimestamp="2026-03-13 15:40:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:40:22.924119085 +0000 UTC m=+5721.823906794" watchObservedRunningTime="2026-03-13 15:40:22.94842752 +0000 UTC m=+5721.848215209" Mar 13 15:40:22 crc kubenswrapper[4907]: I0313 15:40:22.950239 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.950231039 podStartE2EDuration="2.950231039s" podCreationTimestamp="2026-03-13 15:40:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:40:22.947371561 +0000 UTC m=+5721.847159250" watchObservedRunningTime="2026-03-13 15:40:22.950231039 +0000 UTC m=+5721.850018728" Mar 13 15:40:22 crc kubenswrapper[4907]: I0313 15:40:22.968983 4907 scope.go:117] "RemoveContainer" containerID="4d47e9b4f5705f84e359bebffdaf6b40c2a5a7d780d8daa8048b0a559dabed59" Mar 13 15:40:22 crc kubenswrapper[4907]: E0313 15:40:22.975512 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4d47e9b4f5705f84e359bebffdaf6b40c2a5a7d780d8daa8048b0a559dabed59\": container with ID starting with 4d47e9b4f5705f84e359bebffdaf6b40c2a5a7d780d8daa8048b0a559dabed59 not found: ID does not exist" containerID="4d47e9b4f5705f84e359bebffdaf6b40c2a5a7d780d8daa8048b0a559dabed59" Mar 13 15:40:22 crc kubenswrapper[4907]: I0313 15:40:22.975565 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4d47e9b4f5705f84e359bebffdaf6b40c2a5a7d780d8daa8048b0a559dabed59"} err="failed to get container status \"4d47e9b4f5705f84e359bebffdaf6b40c2a5a7d780d8daa8048b0a559dabed59\": rpc error: code = NotFound desc = could not find container \"4d47e9b4f5705f84e359bebffdaf6b40c2a5a7d780d8daa8048b0a559dabed59\": container with ID starting with 4d47e9b4f5705f84e359bebffdaf6b40c2a5a7d780d8daa8048b0a559dabed59 not found: ID does not exist" Mar 13 15:40:22 crc kubenswrapper[4907]: I0313 15:40:22.975597 4907 scope.go:117] "RemoveContainer" containerID="dc68df440400c3300857779e0180b105a0536c85adecc574069e0c384c4a2b78" Mar 13 15:40:22 crc kubenswrapper[4907]: E0313 15:40:22.975992 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dc68df440400c3300857779e0180b105a0536c85adecc574069e0c384c4a2b78\": container with ID starting with dc68df440400c3300857779e0180b105a0536c85adecc574069e0c384c4a2b78 not found: ID does not exist" containerID="dc68df440400c3300857779e0180b105a0536c85adecc574069e0c384c4a2b78" Mar 13 15:40:22 crc kubenswrapper[4907]: I0313 15:40:22.976019 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc68df440400c3300857779e0180b105a0536c85adecc574069e0c384c4a2b78"} err="failed to get container status \"dc68df440400c3300857779e0180b105a0536c85adecc574069e0c384c4a2b78\": rpc error: code = NotFound desc = could not find container \"dc68df440400c3300857779e0180b105a0536c85adecc574069e0c384c4a2b78\": container with ID starting with dc68df440400c3300857779e0180b105a0536c85adecc574069e0c384c4a2b78 not found: ID does not exist" Mar 13 15:40:22 crc kubenswrapper[4907]: I0313 15:40:22.989755 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5153eb33-98f7-4452-b4f5-751a6d49c54b-ovsdbserver-nb\") pod \"5153eb33-98f7-4452-b4f5-751a6d49c54b\" (UID: \"5153eb33-98f7-4452-b4f5-751a6d49c54b\") " Mar 13 15:40:22 crc kubenswrapper[4907]: I0313 15:40:22.989832 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5153eb33-98f7-4452-b4f5-751a6d49c54b-config\") pod \"5153eb33-98f7-4452-b4f5-751a6d49c54b\" (UID: \"5153eb33-98f7-4452-b4f5-751a6d49c54b\") " Mar 13 15:40:22 crc kubenswrapper[4907]: I0313 15:40:22.989968 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5153eb33-98f7-4452-b4f5-751a6d49c54b-dns-svc\") pod \"5153eb33-98f7-4452-b4f5-751a6d49c54b\" (UID: \"5153eb33-98f7-4452-b4f5-751a6d49c54b\") " Mar 13 15:40:22 crc kubenswrapper[4907]: I0313 15:40:22.990006 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5153eb33-98f7-4452-b4f5-751a6d49c54b-ovsdbserver-sb\") pod \"5153eb33-98f7-4452-b4f5-751a6d49c54b\" (UID: \"5153eb33-98f7-4452-b4f5-751a6d49c54b\") " Mar 13 15:40:22 crc kubenswrapper[4907]: I0313 15:40:22.990032 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gskhr\" (UniqueName: \"kubernetes.io/projected/5153eb33-98f7-4452-b4f5-751a6d49c54b-kube-api-access-gskhr\") pod \"5153eb33-98f7-4452-b4f5-751a6d49c54b\" (UID: \"5153eb33-98f7-4452-b4f5-751a6d49c54b\") " Mar 13 15:40:23 crc kubenswrapper[4907]: I0313 15:40:23.009008 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5153eb33-98f7-4452-b4f5-751a6d49c54b-kube-api-access-gskhr" (OuterVolumeSpecName: "kube-api-access-gskhr") pod "5153eb33-98f7-4452-b4f5-751a6d49c54b" (UID: "5153eb33-98f7-4452-b4f5-751a6d49c54b"). InnerVolumeSpecName "kube-api-access-gskhr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:40:23 crc kubenswrapper[4907]: I0313 15:40:23.052157 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5153eb33-98f7-4452-b4f5-751a6d49c54b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "5153eb33-98f7-4452-b4f5-751a6d49c54b" (UID: "5153eb33-98f7-4452-b4f5-751a6d49c54b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:40:23 crc kubenswrapper[4907]: I0313 15:40:23.053963 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5153eb33-98f7-4452-b4f5-751a6d49c54b-config" (OuterVolumeSpecName: "config") pod "5153eb33-98f7-4452-b4f5-751a6d49c54b" (UID: "5153eb33-98f7-4452-b4f5-751a6d49c54b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:40:23 crc kubenswrapper[4907]: I0313 15:40:23.061630 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5153eb33-98f7-4452-b4f5-751a6d49c54b-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "5153eb33-98f7-4452-b4f5-751a6d49c54b" (UID: "5153eb33-98f7-4452-b4f5-751a6d49c54b"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:40:23 crc kubenswrapper[4907]: I0313 15:40:23.064667 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5153eb33-98f7-4452-b4f5-751a6d49c54b-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "5153eb33-98f7-4452-b4f5-751a6d49c54b" (UID: "5153eb33-98f7-4452-b4f5-751a6d49c54b"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:40:23 crc kubenswrapper[4907]: I0313 15:40:23.091683 4907 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5153eb33-98f7-4452-b4f5-751a6d49c54b-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 13 15:40:23 crc kubenswrapper[4907]: I0313 15:40:23.091710 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gskhr\" (UniqueName: \"kubernetes.io/projected/5153eb33-98f7-4452-b4f5-751a6d49c54b-kube-api-access-gskhr\") on node \"crc\" DevicePath \"\"" Mar 13 15:40:23 crc kubenswrapper[4907]: I0313 15:40:23.091722 4907 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5153eb33-98f7-4452-b4f5-751a6d49c54b-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 13 15:40:23 crc kubenswrapper[4907]: I0313 15:40:23.091731 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5153eb33-98f7-4452-b4f5-751a6d49c54b-config\") on node \"crc\" DevicePath \"\"" Mar 13 15:40:23 crc kubenswrapper[4907]: I0313 15:40:23.091741 4907 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5153eb33-98f7-4452-b4f5-751a6d49c54b-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 13 15:40:23 crc kubenswrapper[4907]: I0313 15:40:23.227405 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5dd7c58f95-579qq"] Mar 13 15:40:23 crc kubenswrapper[4907]: I0313 15:40:23.235281 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5dd7c58f95-579qq"] Mar 13 15:40:23 crc kubenswrapper[4907]: I0313 15:40:23.792390 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5153eb33-98f7-4452-b4f5-751a6d49c54b" path="/var/lib/kubelet/pods/5153eb33-98f7-4452-b4f5-751a6d49c54b/volumes" Mar 13 15:40:24 crc kubenswrapper[4907]: I0313 15:40:24.277578 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Mar 13 15:40:24 crc kubenswrapper[4907]: I0313 15:40:24.698839 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 13 15:40:24 crc kubenswrapper[4907]: I0313 15:40:24.767414 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-d9b97"] Mar 13 15:40:24 crc kubenswrapper[4907]: E0313 15:40:24.767859 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5153eb33-98f7-4452-b4f5-751a6d49c54b" containerName="init" Mar 13 15:40:24 crc kubenswrapper[4907]: I0313 15:40:24.767958 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="5153eb33-98f7-4452-b4f5-751a6d49c54b" containerName="init" Mar 13 15:40:24 crc kubenswrapper[4907]: E0313 15:40:24.767988 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5153eb33-98f7-4452-b4f5-751a6d49c54b" containerName="dnsmasq-dns" Mar 13 15:40:24 crc kubenswrapper[4907]: I0313 15:40:24.767999 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="5153eb33-98f7-4452-b4f5-751a6d49c54b" containerName="dnsmasq-dns" Mar 13 15:40:24 crc kubenswrapper[4907]: E0313 15:40:24.768018 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5192153-c160-40bc-9375-a730e800e76e" containerName="nova-scheduler-scheduler" Mar 13 15:40:24 crc kubenswrapper[4907]: I0313 15:40:24.768027 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5192153-c160-40bc-9375-a730e800e76e" containerName="nova-scheduler-scheduler" Mar 13 15:40:24 crc kubenswrapper[4907]: I0313 15:40:24.768280 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="5153eb33-98f7-4452-b4f5-751a6d49c54b" containerName="dnsmasq-dns" Mar 13 15:40:24 crc kubenswrapper[4907]: I0313 15:40:24.768309 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="e5192153-c160-40bc-9375-a730e800e76e" containerName="nova-scheduler-scheduler" Mar 13 15:40:24 crc kubenswrapper[4907]: I0313 15:40:24.769042 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-d9b97" Mar 13 15:40:24 crc kubenswrapper[4907]: I0313 15:40:24.777332 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Mar 13 15:40:24 crc kubenswrapper[4907]: I0313 15:40:24.777560 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Mar 13 15:40:24 crc kubenswrapper[4907]: I0313 15:40:24.780715 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-d9b97"] Mar 13 15:40:24 crc kubenswrapper[4907]: I0313 15:40:24.824777 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5kk9j\" (UniqueName: \"kubernetes.io/projected/e5192153-c160-40bc-9375-a730e800e76e-kube-api-access-5kk9j\") pod \"e5192153-c160-40bc-9375-a730e800e76e\" (UID: \"e5192153-c160-40bc-9375-a730e800e76e\") " Mar 13 15:40:24 crc kubenswrapper[4907]: I0313 15:40:24.824861 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5192153-c160-40bc-9375-a730e800e76e-combined-ca-bundle\") pod \"e5192153-c160-40bc-9375-a730e800e76e\" (UID: \"e5192153-c160-40bc-9375-a730e800e76e\") " Mar 13 15:40:24 crc kubenswrapper[4907]: I0313 15:40:24.824977 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5192153-c160-40bc-9375-a730e800e76e-config-data\") pod \"e5192153-c160-40bc-9375-a730e800e76e\" (UID: \"e5192153-c160-40bc-9375-a730e800e76e\") " Mar 13 15:40:24 crc kubenswrapper[4907]: I0313 15:40:24.838786 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e5192153-c160-40bc-9375-a730e800e76e-kube-api-access-5kk9j" (OuterVolumeSpecName: "kube-api-access-5kk9j") pod "e5192153-c160-40bc-9375-a730e800e76e" (UID: "e5192153-c160-40bc-9375-a730e800e76e"). InnerVolumeSpecName "kube-api-access-5kk9j". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:40:24 crc kubenswrapper[4907]: I0313 15:40:24.851754 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e5192153-c160-40bc-9375-a730e800e76e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e5192153-c160-40bc-9375-a730e800e76e" (UID: "e5192153-c160-40bc-9375-a730e800e76e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:40:24 crc kubenswrapper[4907]: I0313 15:40:24.852137 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e5192153-c160-40bc-9375-a730e800e76e-config-data" (OuterVolumeSpecName: "config-data") pod "e5192153-c160-40bc-9375-a730e800e76e" (UID: "e5192153-c160-40bc-9375-a730e800e76e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:40:24 crc kubenswrapper[4907]: I0313 15:40:24.910596 4907 generic.go:334] "Generic (PLEG): container finished" podID="e5192153-c160-40bc-9375-a730e800e76e" containerID="9ca8fe863388471bc1a4b09ca8e4b518970b726d2835186fd66e2f90d2cda6ca" exitCode=0 Mar 13 15:40:24 crc kubenswrapper[4907]: I0313 15:40:24.910643 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"e5192153-c160-40bc-9375-a730e800e76e","Type":"ContainerDied","Data":"9ca8fe863388471bc1a4b09ca8e4b518970b726d2835186fd66e2f90d2cda6ca"} Mar 13 15:40:24 crc kubenswrapper[4907]: I0313 15:40:24.910673 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"e5192153-c160-40bc-9375-a730e800e76e","Type":"ContainerDied","Data":"a6f0a44a2f3b94386ebd8b9c603db045564e6d31764a5593228f18f153118242"} Mar 13 15:40:24 crc kubenswrapper[4907]: I0313 15:40:24.910690 4907 scope.go:117] "RemoveContainer" containerID="9ca8fe863388471bc1a4b09ca8e4b518970b726d2835186fd66e2f90d2cda6ca" Mar 13 15:40:24 crc kubenswrapper[4907]: I0313 15:40:24.910817 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 13 15:40:24 crc kubenswrapper[4907]: I0313 15:40:24.929019 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6e87a99-1744-4002-a08b-dabf8a97e48b-config-data\") pod \"nova-cell1-cell-mapping-d9b97\" (UID: \"f6e87a99-1744-4002-a08b-dabf8a97e48b\") " pod="openstack/nova-cell1-cell-mapping-d9b97" Mar 13 15:40:24 crc kubenswrapper[4907]: I0313 15:40:24.929253 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vl2kv\" (UniqueName: \"kubernetes.io/projected/f6e87a99-1744-4002-a08b-dabf8a97e48b-kube-api-access-vl2kv\") pod \"nova-cell1-cell-mapping-d9b97\" (UID: \"f6e87a99-1744-4002-a08b-dabf8a97e48b\") " pod="openstack/nova-cell1-cell-mapping-d9b97" Mar 13 15:40:24 crc kubenswrapper[4907]: I0313 15:40:24.930048 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6e87a99-1744-4002-a08b-dabf8a97e48b-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-d9b97\" (UID: \"f6e87a99-1744-4002-a08b-dabf8a97e48b\") " pod="openstack/nova-cell1-cell-mapping-d9b97" Mar 13 15:40:24 crc kubenswrapper[4907]: I0313 15:40:24.932706 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f6e87a99-1744-4002-a08b-dabf8a97e48b-scripts\") pod \"nova-cell1-cell-mapping-d9b97\" (UID: \"f6e87a99-1744-4002-a08b-dabf8a97e48b\") " pod="openstack/nova-cell1-cell-mapping-d9b97" Mar 13 15:40:24 crc kubenswrapper[4907]: I0313 15:40:24.932895 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5192153-c160-40bc-9375-a730e800e76e-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 15:40:24 crc kubenswrapper[4907]: I0313 15:40:24.932919 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5kk9j\" (UniqueName: \"kubernetes.io/projected/e5192153-c160-40bc-9375-a730e800e76e-kube-api-access-5kk9j\") on node \"crc\" DevicePath \"\"" Mar 13 15:40:24 crc kubenswrapper[4907]: I0313 15:40:24.932933 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5192153-c160-40bc-9375-a730e800e76e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 15:40:24 crc kubenswrapper[4907]: I0313 15:40:24.948559 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 15:40:24 crc kubenswrapper[4907]: I0313 15:40:24.954525 4907 scope.go:117] "RemoveContainer" containerID="9ca8fe863388471bc1a4b09ca8e4b518970b726d2835186fd66e2f90d2cda6ca" Mar 13 15:40:24 crc kubenswrapper[4907]: E0313 15:40:24.955146 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9ca8fe863388471bc1a4b09ca8e4b518970b726d2835186fd66e2f90d2cda6ca\": container with ID starting with 9ca8fe863388471bc1a4b09ca8e4b518970b726d2835186fd66e2f90d2cda6ca not found: ID does not exist" containerID="9ca8fe863388471bc1a4b09ca8e4b518970b726d2835186fd66e2f90d2cda6ca" Mar 13 15:40:24 crc kubenswrapper[4907]: I0313 15:40:24.955192 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9ca8fe863388471bc1a4b09ca8e4b518970b726d2835186fd66e2f90d2cda6ca"} err="failed to get container status \"9ca8fe863388471bc1a4b09ca8e4b518970b726d2835186fd66e2f90d2cda6ca\": rpc error: code = NotFound desc = could not find container \"9ca8fe863388471bc1a4b09ca8e4b518970b726d2835186fd66e2f90d2cda6ca\": container with ID starting with 9ca8fe863388471bc1a4b09ca8e4b518970b726d2835186fd66e2f90d2cda6ca not found: ID does not exist" Mar 13 15:40:24 crc kubenswrapper[4907]: I0313 15:40:24.967608 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 15:40:24 crc kubenswrapper[4907]: I0313 15:40:24.977243 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 15:40:24 crc kubenswrapper[4907]: I0313 15:40:24.978596 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 13 15:40:24 crc kubenswrapper[4907]: I0313 15:40:24.981150 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Mar 13 15:40:25 crc kubenswrapper[4907]: I0313 15:40:25.002629 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 15:40:25 crc kubenswrapper[4907]: I0313 15:40:25.033323 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vl2kv\" (UniqueName: \"kubernetes.io/projected/f6e87a99-1744-4002-a08b-dabf8a97e48b-kube-api-access-vl2kv\") pod \"nova-cell1-cell-mapping-d9b97\" (UID: \"f6e87a99-1744-4002-a08b-dabf8a97e48b\") " pod="openstack/nova-cell1-cell-mapping-d9b97" Mar 13 15:40:25 crc kubenswrapper[4907]: I0313 15:40:25.033367 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b2e0902-2ec8-4712-8215-71643998f9ae-config-data\") pod \"nova-scheduler-0\" (UID: \"6b2e0902-2ec8-4712-8215-71643998f9ae\") " pod="openstack/nova-scheduler-0" Mar 13 15:40:25 crc kubenswrapper[4907]: I0313 15:40:25.033418 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6e87a99-1744-4002-a08b-dabf8a97e48b-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-d9b97\" (UID: \"f6e87a99-1744-4002-a08b-dabf8a97e48b\") " pod="openstack/nova-cell1-cell-mapping-d9b97" Mar 13 15:40:25 crc kubenswrapper[4907]: I0313 15:40:25.033444 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f6e87a99-1744-4002-a08b-dabf8a97e48b-scripts\") pod \"nova-cell1-cell-mapping-d9b97\" (UID: \"f6e87a99-1744-4002-a08b-dabf8a97e48b\") " pod="openstack/nova-cell1-cell-mapping-d9b97" Mar 13 15:40:25 crc kubenswrapper[4907]: I0313 15:40:25.033560 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b2e0902-2ec8-4712-8215-71643998f9ae-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"6b2e0902-2ec8-4712-8215-71643998f9ae\") " pod="openstack/nova-scheduler-0" Mar 13 15:40:25 crc kubenswrapper[4907]: I0313 15:40:25.033666 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6e87a99-1744-4002-a08b-dabf8a97e48b-config-data\") pod \"nova-cell1-cell-mapping-d9b97\" (UID: \"f6e87a99-1744-4002-a08b-dabf8a97e48b\") " pod="openstack/nova-cell1-cell-mapping-d9b97" Mar 13 15:40:25 crc kubenswrapper[4907]: I0313 15:40:25.033726 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lpzzl\" (UniqueName: \"kubernetes.io/projected/6b2e0902-2ec8-4712-8215-71643998f9ae-kube-api-access-lpzzl\") pod \"nova-scheduler-0\" (UID: \"6b2e0902-2ec8-4712-8215-71643998f9ae\") " pod="openstack/nova-scheduler-0" Mar 13 15:40:25 crc kubenswrapper[4907]: I0313 15:40:25.037262 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6e87a99-1744-4002-a08b-dabf8a97e48b-config-data\") pod \"nova-cell1-cell-mapping-d9b97\" (UID: \"f6e87a99-1744-4002-a08b-dabf8a97e48b\") " pod="openstack/nova-cell1-cell-mapping-d9b97" Mar 13 15:40:25 crc kubenswrapper[4907]: I0313 15:40:25.037292 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6e87a99-1744-4002-a08b-dabf8a97e48b-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-d9b97\" (UID: \"f6e87a99-1744-4002-a08b-dabf8a97e48b\") " pod="openstack/nova-cell1-cell-mapping-d9b97" Mar 13 15:40:25 crc kubenswrapper[4907]: I0313 15:40:25.037454 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f6e87a99-1744-4002-a08b-dabf8a97e48b-scripts\") pod \"nova-cell1-cell-mapping-d9b97\" (UID: \"f6e87a99-1744-4002-a08b-dabf8a97e48b\") " pod="openstack/nova-cell1-cell-mapping-d9b97" Mar 13 15:40:25 crc kubenswrapper[4907]: I0313 15:40:25.047836 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vl2kv\" (UniqueName: \"kubernetes.io/projected/f6e87a99-1744-4002-a08b-dabf8a97e48b-kube-api-access-vl2kv\") pod \"nova-cell1-cell-mapping-d9b97\" (UID: \"f6e87a99-1744-4002-a08b-dabf8a97e48b\") " pod="openstack/nova-cell1-cell-mapping-d9b97" Mar 13 15:40:25 crc kubenswrapper[4907]: I0313 15:40:25.095903 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-d9b97" Mar 13 15:40:25 crc kubenswrapper[4907]: I0313 15:40:25.137012 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b2e0902-2ec8-4712-8215-71643998f9ae-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"6b2e0902-2ec8-4712-8215-71643998f9ae\") " pod="openstack/nova-scheduler-0" Mar 13 15:40:25 crc kubenswrapper[4907]: I0313 15:40:25.137154 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lpzzl\" (UniqueName: \"kubernetes.io/projected/6b2e0902-2ec8-4712-8215-71643998f9ae-kube-api-access-lpzzl\") pod \"nova-scheduler-0\" (UID: \"6b2e0902-2ec8-4712-8215-71643998f9ae\") " pod="openstack/nova-scheduler-0" Mar 13 15:40:25 crc kubenswrapper[4907]: I0313 15:40:25.137379 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b2e0902-2ec8-4712-8215-71643998f9ae-config-data\") pod \"nova-scheduler-0\" (UID: \"6b2e0902-2ec8-4712-8215-71643998f9ae\") " pod="openstack/nova-scheduler-0" Mar 13 15:40:25 crc kubenswrapper[4907]: I0313 15:40:25.158495 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b2e0902-2ec8-4712-8215-71643998f9ae-config-data\") pod \"nova-scheduler-0\" (UID: \"6b2e0902-2ec8-4712-8215-71643998f9ae\") " pod="openstack/nova-scheduler-0" Mar 13 15:40:25 crc kubenswrapper[4907]: I0313 15:40:25.158619 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lpzzl\" (UniqueName: \"kubernetes.io/projected/6b2e0902-2ec8-4712-8215-71643998f9ae-kube-api-access-lpzzl\") pod \"nova-scheduler-0\" (UID: \"6b2e0902-2ec8-4712-8215-71643998f9ae\") " pod="openstack/nova-scheduler-0" Mar 13 15:40:25 crc kubenswrapper[4907]: I0313 15:40:25.159346 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b2e0902-2ec8-4712-8215-71643998f9ae-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"6b2e0902-2ec8-4712-8215-71643998f9ae\") " pod="openstack/nova-scheduler-0" Mar 13 15:40:25 crc kubenswrapper[4907]: I0313 15:40:25.310761 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 13 15:40:25 crc kubenswrapper[4907]: I0313 15:40:25.566287 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-d9b97"] Mar 13 15:40:25 crc kubenswrapper[4907]: I0313 15:40:25.823449 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e5192153-c160-40bc-9375-a730e800e76e" path="/var/lib/kubelet/pods/e5192153-c160-40bc-9375-a730e800e76e/volumes" Mar 13 15:40:25 crc kubenswrapper[4907]: I0313 15:40:25.860998 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 15:40:25 crc kubenswrapper[4907]: W0313 15:40:25.866961 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6b2e0902_2ec8_4712_8215_71643998f9ae.slice/crio-f8adfb940d43d79ae7a52f858d549a2c2afbabbdf09ea298e87ae615f8153c93 WatchSource:0}: Error finding container f8adfb940d43d79ae7a52f858d549a2c2afbabbdf09ea298e87ae615f8153c93: Status 404 returned error can't find the container with id f8adfb940d43d79ae7a52f858d549a2c2afbabbdf09ea298e87ae615f8153c93 Mar 13 15:40:25 crc kubenswrapper[4907]: I0313 15:40:25.923387 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-d9b97" event={"ID":"f6e87a99-1744-4002-a08b-dabf8a97e48b","Type":"ContainerStarted","Data":"d3ac5726910e7451a7d9a4ee3678b7f325b341f8f577f2d9ded3dcb9a624377a"} Mar 13 15:40:25 crc kubenswrapper[4907]: I0313 15:40:25.925325 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"6b2e0902-2ec8-4712-8215-71643998f9ae","Type":"ContainerStarted","Data":"f8adfb940d43d79ae7a52f858d549a2c2afbabbdf09ea298e87ae615f8153c93"} Mar 13 15:40:26 crc kubenswrapper[4907]: I0313 15:40:26.945337 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-d9b97" event={"ID":"f6e87a99-1744-4002-a08b-dabf8a97e48b","Type":"ContainerStarted","Data":"0bdd4c179fd6402b48e39e0498ccf620cd42daa3f95e10bf8dc41cdb2673e660"} Mar 13 15:40:26 crc kubenswrapper[4907]: I0313 15:40:26.948498 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"6b2e0902-2ec8-4712-8215-71643998f9ae","Type":"ContainerStarted","Data":"1f04df211f07bf45e057e613b89c2ab160d4ab249aeb981eda1a7e92b8868b32"} Mar 13 15:40:26 crc kubenswrapper[4907]: I0313 15:40:26.971592 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-d9b97" podStartSLOduration=2.971573059 podStartE2EDuration="2.971573059s" podCreationTimestamp="2026-03-13 15:40:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:40:26.966520331 +0000 UTC m=+5725.866308040" watchObservedRunningTime="2026-03-13 15:40:26.971573059 +0000 UTC m=+5725.871360738" Mar 13 15:40:26 crc kubenswrapper[4907]: I0313 15:40:26.988331 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.988307336 podStartE2EDuration="2.988307336s" podCreationTimestamp="2026-03-13 15:40:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:40:26.983132435 +0000 UTC m=+5725.882920134" watchObservedRunningTime="2026-03-13 15:40:26.988307336 +0000 UTC m=+5725.888095025" Mar 13 15:40:28 crc kubenswrapper[4907]: I0313 15:40:28.660444 4907 scope.go:117] "RemoveContainer" containerID="a3109c92b6a4f47975f53b04558486733a8cd9a2cb768f638be94b37e0295965" Mar 13 15:40:28 crc kubenswrapper[4907]: I0313 15:40:28.684245 4907 scope.go:117] "RemoveContainer" containerID="47fa5a52c38336d054a9aed448aae48b3dd0ee12065b60aaad1abb03635d9d32" Mar 13 15:40:28 crc kubenswrapper[4907]: I0313 15:40:28.750167 4907 scope.go:117] "RemoveContainer" containerID="7829d0510750d5cae8df3422b54b042bd33e40f132a8e3e16f6eec0a8952963b" Mar 13 15:40:28 crc kubenswrapper[4907]: I0313 15:40:28.781547 4907 scope.go:117] "RemoveContainer" containerID="1eda03ad73ea12900c383cf8f9715d8d1c4690e9fdc026cd78850ed84ad83b37" Mar 13 15:40:30 crc kubenswrapper[4907]: I0313 15:40:30.311855 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Mar 13 15:40:30 crc kubenswrapper[4907]: I0313 15:40:30.995711 4907 generic.go:334] "Generic (PLEG): container finished" podID="f6e87a99-1744-4002-a08b-dabf8a97e48b" containerID="0bdd4c179fd6402b48e39e0498ccf620cd42daa3f95e10bf8dc41cdb2673e660" exitCode=0 Mar 13 15:40:30 crc kubenswrapper[4907]: I0313 15:40:30.995804 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-d9b97" event={"ID":"f6e87a99-1744-4002-a08b-dabf8a97e48b","Type":"ContainerDied","Data":"0bdd4c179fd6402b48e39e0498ccf620cd42daa3f95e10bf8dc41cdb2673e660"} Mar 13 15:40:31 crc kubenswrapper[4907]: I0313 15:40:31.272317 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Mar 13 15:40:31 crc kubenswrapper[4907]: I0313 15:40:31.273598 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Mar 13 15:40:31 crc kubenswrapper[4907]: I0313 15:40:31.320539 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Mar 13 15:40:31 crc kubenswrapper[4907]: I0313 15:40:31.320593 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Mar 13 15:40:32 crc kubenswrapper[4907]: I0313 15:40:32.355085 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.1.108:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 15:40:32 crc kubenswrapper[4907]: I0313 15:40:32.355119 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.1.108:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 15:40:32 crc kubenswrapper[4907]: I0313 15:40:32.360704 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-d9b97" Mar 13 15:40:32 crc kubenswrapper[4907]: I0313 15:40:32.438165 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="ad3c8be4-d237-4c1e-bd98-963acae4ac32" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.1.109:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 15:40:32 crc kubenswrapper[4907]: I0313 15:40:32.438766 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="ad3c8be4-d237-4c1e-bd98-963acae4ac32" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.1.109:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 15:40:32 crc kubenswrapper[4907]: I0313 15:40:32.480246 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6e87a99-1744-4002-a08b-dabf8a97e48b-combined-ca-bundle\") pod \"f6e87a99-1744-4002-a08b-dabf8a97e48b\" (UID: \"f6e87a99-1744-4002-a08b-dabf8a97e48b\") " Mar 13 15:40:32 crc kubenswrapper[4907]: I0313 15:40:32.480305 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vl2kv\" (UniqueName: \"kubernetes.io/projected/f6e87a99-1744-4002-a08b-dabf8a97e48b-kube-api-access-vl2kv\") pod \"f6e87a99-1744-4002-a08b-dabf8a97e48b\" (UID: \"f6e87a99-1744-4002-a08b-dabf8a97e48b\") " Mar 13 15:40:32 crc kubenswrapper[4907]: I0313 15:40:32.480373 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f6e87a99-1744-4002-a08b-dabf8a97e48b-scripts\") pod \"f6e87a99-1744-4002-a08b-dabf8a97e48b\" (UID: \"f6e87a99-1744-4002-a08b-dabf8a97e48b\") " Mar 13 15:40:32 crc kubenswrapper[4907]: I0313 15:40:32.480397 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6e87a99-1744-4002-a08b-dabf8a97e48b-config-data\") pod \"f6e87a99-1744-4002-a08b-dabf8a97e48b\" (UID: \"f6e87a99-1744-4002-a08b-dabf8a97e48b\") " Mar 13 15:40:32 crc kubenswrapper[4907]: I0313 15:40:32.486998 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6e87a99-1744-4002-a08b-dabf8a97e48b-scripts" (OuterVolumeSpecName: "scripts") pod "f6e87a99-1744-4002-a08b-dabf8a97e48b" (UID: "f6e87a99-1744-4002-a08b-dabf8a97e48b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:40:32 crc kubenswrapper[4907]: I0313 15:40:32.487131 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f6e87a99-1744-4002-a08b-dabf8a97e48b-kube-api-access-vl2kv" (OuterVolumeSpecName: "kube-api-access-vl2kv") pod "f6e87a99-1744-4002-a08b-dabf8a97e48b" (UID: "f6e87a99-1744-4002-a08b-dabf8a97e48b"). InnerVolumeSpecName "kube-api-access-vl2kv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:40:32 crc kubenswrapper[4907]: I0313 15:40:32.517152 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6e87a99-1744-4002-a08b-dabf8a97e48b-config-data" (OuterVolumeSpecName: "config-data") pod "f6e87a99-1744-4002-a08b-dabf8a97e48b" (UID: "f6e87a99-1744-4002-a08b-dabf8a97e48b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:40:32 crc kubenswrapper[4907]: I0313 15:40:32.525747 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6e87a99-1744-4002-a08b-dabf8a97e48b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f6e87a99-1744-4002-a08b-dabf8a97e48b" (UID: "f6e87a99-1744-4002-a08b-dabf8a97e48b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:40:32 crc kubenswrapper[4907]: I0313 15:40:32.581836 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6e87a99-1744-4002-a08b-dabf8a97e48b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 15:40:32 crc kubenswrapper[4907]: I0313 15:40:32.581870 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vl2kv\" (UniqueName: \"kubernetes.io/projected/f6e87a99-1744-4002-a08b-dabf8a97e48b-kube-api-access-vl2kv\") on node \"crc\" DevicePath \"\"" Mar 13 15:40:32 crc kubenswrapper[4907]: I0313 15:40:32.581898 4907 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f6e87a99-1744-4002-a08b-dabf8a97e48b-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 15:40:32 crc kubenswrapper[4907]: I0313 15:40:32.581907 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6e87a99-1744-4002-a08b-dabf8a97e48b-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 15:40:33 crc kubenswrapper[4907]: I0313 15:40:33.014727 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-d9b97" event={"ID":"f6e87a99-1744-4002-a08b-dabf8a97e48b","Type":"ContainerDied","Data":"d3ac5726910e7451a7d9a4ee3678b7f325b341f8f577f2d9ded3dcb9a624377a"} Mar 13 15:40:33 crc kubenswrapper[4907]: I0313 15:40:33.014769 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d3ac5726910e7451a7d9a4ee3678b7f325b341f8f577f2d9ded3dcb9a624377a" Mar 13 15:40:33 crc kubenswrapper[4907]: I0313 15:40:33.014775 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-d9b97" Mar 13 15:40:33 crc kubenswrapper[4907]: I0313 15:40:33.209929 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Mar 13 15:40:33 crc kubenswrapper[4907]: I0313 15:40:33.210434 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="ad3c8be4-d237-4c1e-bd98-963acae4ac32" containerName="nova-api-log" containerID="cri-o://a9b52fcfd1633f1cfd4fe4e130d7daf6a6a0f261d982b7490f4c49360b713147" gracePeriod=30 Mar 13 15:40:33 crc kubenswrapper[4907]: I0313 15:40:33.210500 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="ad3c8be4-d237-4c1e-bd98-963acae4ac32" containerName="nova-api-api" containerID="cri-o://e833237c58d59cd7a382ff7644c38c3979cbe4fa6a8a9e044f9040c86f3823df" gracePeriod=30 Mar 13 15:40:33 crc kubenswrapper[4907]: I0313 15:40:33.256164 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 15:40:33 crc kubenswrapper[4907]: I0313 15:40:33.256392 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="6b2e0902-2ec8-4712-8215-71643998f9ae" containerName="nova-scheduler-scheduler" containerID="cri-o://1f04df211f07bf45e057e613b89c2ab160d4ab249aeb981eda1a7e92b8868b32" gracePeriod=30 Mar 13 15:40:33 crc kubenswrapper[4907]: I0313 15:40:33.272259 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Mar 13 15:40:33 crc kubenswrapper[4907]: I0313 15:40:33.272550 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0" containerName="nova-metadata-log" containerID="cri-o://12ad1f3ba8429b32fe4517953f072fa40cf2ef694f54ab5cea4b98ef8339414c" gracePeriod=30 Mar 13 15:40:33 crc kubenswrapper[4907]: I0313 15:40:33.272739 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0" containerName="nova-metadata-metadata" containerID="cri-o://9d4f1d5de2313cc0cacaa0de90599bf7cc29209144c55b94e7e5f63f34cd569d" gracePeriod=30 Mar 13 15:40:34 crc kubenswrapper[4907]: I0313 15:40:34.025451 4907 generic.go:334] "Generic (PLEG): container finished" podID="ad3c8be4-d237-4c1e-bd98-963acae4ac32" containerID="a9b52fcfd1633f1cfd4fe4e130d7daf6a6a0f261d982b7490f4c49360b713147" exitCode=143 Mar 13 15:40:34 crc kubenswrapper[4907]: I0313 15:40:34.025531 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ad3c8be4-d237-4c1e-bd98-963acae4ac32","Type":"ContainerDied","Data":"a9b52fcfd1633f1cfd4fe4e130d7daf6a6a0f261d982b7490f4c49360b713147"} Mar 13 15:40:34 crc kubenswrapper[4907]: I0313 15:40:34.027452 4907 generic.go:334] "Generic (PLEG): container finished" podID="8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0" containerID="12ad1f3ba8429b32fe4517953f072fa40cf2ef694f54ab5cea4b98ef8339414c" exitCode=143 Mar 13 15:40:34 crc kubenswrapper[4907]: I0313 15:40:34.027490 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0","Type":"ContainerDied","Data":"12ad1f3ba8429b32fe4517953f072fa40cf2ef694f54ab5cea4b98ef8339414c"} Mar 13 15:40:37 crc kubenswrapper[4907]: I0313 15:40:37.636732 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 13 15:40:37 crc kubenswrapper[4907]: I0313 15:40:37.782854 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b2e0902-2ec8-4712-8215-71643998f9ae-combined-ca-bundle\") pod \"6b2e0902-2ec8-4712-8215-71643998f9ae\" (UID: \"6b2e0902-2ec8-4712-8215-71643998f9ae\") " Mar 13 15:40:37 crc kubenswrapper[4907]: I0313 15:40:37.782967 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lpzzl\" (UniqueName: \"kubernetes.io/projected/6b2e0902-2ec8-4712-8215-71643998f9ae-kube-api-access-lpzzl\") pod \"6b2e0902-2ec8-4712-8215-71643998f9ae\" (UID: \"6b2e0902-2ec8-4712-8215-71643998f9ae\") " Mar 13 15:40:37 crc kubenswrapper[4907]: I0313 15:40:37.783161 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b2e0902-2ec8-4712-8215-71643998f9ae-config-data\") pod \"6b2e0902-2ec8-4712-8215-71643998f9ae\" (UID: \"6b2e0902-2ec8-4712-8215-71643998f9ae\") " Mar 13 15:40:37 crc kubenswrapper[4907]: I0313 15:40:37.788473 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6b2e0902-2ec8-4712-8215-71643998f9ae-kube-api-access-lpzzl" (OuterVolumeSpecName: "kube-api-access-lpzzl") pod "6b2e0902-2ec8-4712-8215-71643998f9ae" (UID: "6b2e0902-2ec8-4712-8215-71643998f9ae"). InnerVolumeSpecName "kube-api-access-lpzzl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:40:37 crc kubenswrapper[4907]: I0313 15:40:37.809130 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b2e0902-2ec8-4712-8215-71643998f9ae-config-data" (OuterVolumeSpecName: "config-data") pod "6b2e0902-2ec8-4712-8215-71643998f9ae" (UID: "6b2e0902-2ec8-4712-8215-71643998f9ae"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:40:37 crc kubenswrapper[4907]: I0313 15:40:37.810926 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b2e0902-2ec8-4712-8215-71643998f9ae-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6b2e0902-2ec8-4712-8215-71643998f9ae" (UID: "6b2e0902-2ec8-4712-8215-71643998f9ae"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:40:37 crc kubenswrapper[4907]: I0313 15:40:37.885676 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b2e0902-2ec8-4712-8215-71643998f9ae-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 15:40:37 crc kubenswrapper[4907]: I0313 15:40:37.885699 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b2e0902-2ec8-4712-8215-71643998f9ae-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 15:40:37 crc kubenswrapper[4907]: I0313 15:40:37.885709 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lpzzl\" (UniqueName: \"kubernetes.io/projected/6b2e0902-2ec8-4712-8215-71643998f9ae-kube-api-access-lpzzl\") on node \"crc\" DevicePath \"\"" Mar 13 15:40:38 crc kubenswrapper[4907]: I0313 15:40:38.063931 4907 generic.go:334] "Generic (PLEG): container finished" podID="6b2e0902-2ec8-4712-8215-71643998f9ae" containerID="1f04df211f07bf45e057e613b89c2ab160d4ab249aeb981eda1a7e92b8868b32" exitCode=0 Mar 13 15:40:38 crc kubenswrapper[4907]: I0313 15:40:38.064014 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 13 15:40:38 crc kubenswrapper[4907]: I0313 15:40:38.064014 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"6b2e0902-2ec8-4712-8215-71643998f9ae","Type":"ContainerDied","Data":"1f04df211f07bf45e057e613b89c2ab160d4ab249aeb981eda1a7e92b8868b32"} Mar 13 15:40:38 crc kubenswrapper[4907]: I0313 15:40:38.064165 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"6b2e0902-2ec8-4712-8215-71643998f9ae","Type":"ContainerDied","Data":"f8adfb940d43d79ae7a52f858d549a2c2afbabbdf09ea298e87ae615f8153c93"} Mar 13 15:40:38 crc kubenswrapper[4907]: I0313 15:40:38.064200 4907 scope.go:117] "RemoveContainer" containerID="1f04df211f07bf45e057e613b89c2ab160d4ab249aeb981eda1a7e92b8868b32" Mar 13 15:40:38 crc kubenswrapper[4907]: I0313 15:40:38.067571 4907 generic.go:334] "Generic (PLEG): container finished" podID="ad3c8be4-d237-4c1e-bd98-963acae4ac32" containerID="e833237c58d59cd7a382ff7644c38c3979cbe4fa6a8a9e044f9040c86f3823df" exitCode=0 Mar 13 15:40:38 crc kubenswrapper[4907]: I0313 15:40:38.067626 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ad3c8be4-d237-4c1e-bd98-963acae4ac32","Type":"ContainerDied","Data":"e833237c58d59cd7a382ff7644c38c3979cbe4fa6a8a9e044f9040c86f3823df"} Mar 13 15:40:38 crc kubenswrapper[4907]: I0313 15:40:38.071467 4907 generic.go:334] "Generic (PLEG): container finished" podID="8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0" containerID="9d4f1d5de2313cc0cacaa0de90599bf7cc29209144c55b94e7e5f63f34cd569d" exitCode=0 Mar 13 15:40:38 crc kubenswrapper[4907]: I0313 15:40:38.071626 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0","Type":"ContainerDied","Data":"9d4f1d5de2313cc0cacaa0de90599bf7cc29209144c55b94e7e5f63f34cd569d"} Mar 13 15:40:38 crc kubenswrapper[4907]: I0313 15:40:38.088231 4907 scope.go:117] "RemoveContainer" containerID="1f04df211f07bf45e057e613b89c2ab160d4ab249aeb981eda1a7e92b8868b32" Mar 13 15:40:38 crc kubenswrapper[4907]: E0313 15:40:38.093839 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1f04df211f07bf45e057e613b89c2ab160d4ab249aeb981eda1a7e92b8868b32\": container with ID starting with 1f04df211f07bf45e057e613b89c2ab160d4ab249aeb981eda1a7e92b8868b32 not found: ID does not exist" containerID="1f04df211f07bf45e057e613b89c2ab160d4ab249aeb981eda1a7e92b8868b32" Mar 13 15:40:38 crc kubenswrapper[4907]: I0313 15:40:38.093947 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1f04df211f07bf45e057e613b89c2ab160d4ab249aeb981eda1a7e92b8868b32"} err="failed to get container status \"1f04df211f07bf45e057e613b89c2ab160d4ab249aeb981eda1a7e92b8868b32\": rpc error: code = NotFound desc = could not find container \"1f04df211f07bf45e057e613b89c2ab160d4ab249aeb981eda1a7e92b8868b32\": container with ID starting with 1f04df211f07bf45e057e613b89c2ab160d4ab249aeb981eda1a7e92b8868b32 not found: ID does not exist" Mar 13 15:40:38 crc kubenswrapper[4907]: I0313 15:40:38.105191 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 15:40:38 crc kubenswrapper[4907]: I0313 15:40:38.126660 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 15:40:38 crc kubenswrapper[4907]: I0313 15:40:38.128453 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 13 15:40:38 crc kubenswrapper[4907]: I0313 15:40:38.138938 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 15:40:38 crc kubenswrapper[4907]: E0313 15:40:38.139345 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0" containerName="nova-metadata-log" Mar 13 15:40:38 crc kubenswrapper[4907]: I0313 15:40:38.139359 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0" containerName="nova-metadata-log" Mar 13 15:40:38 crc kubenswrapper[4907]: E0313 15:40:38.139371 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0" containerName="nova-metadata-metadata" Mar 13 15:40:38 crc kubenswrapper[4907]: I0313 15:40:38.139378 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0" containerName="nova-metadata-metadata" Mar 13 15:40:38 crc kubenswrapper[4907]: E0313 15:40:38.139390 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b2e0902-2ec8-4712-8215-71643998f9ae" containerName="nova-scheduler-scheduler" Mar 13 15:40:38 crc kubenswrapper[4907]: I0313 15:40:38.139396 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b2e0902-2ec8-4712-8215-71643998f9ae" containerName="nova-scheduler-scheduler" Mar 13 15:40:38 crc kubenswrapper[4907]: E0313 15:40:38.139422 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6e87a99-1744-4002-a08b-dabf8a97e48b" containerName="nova-manage" Mar 13 15:40:38 crc kubenswrapper[4907]: I0313 15:40:38.139428 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6e87a99-1744-4002-a08b-dabf8a97e48b" containerName="nova-manage" Mar 13 15:40:38 crc kubenswrapper[4907]: I0313 15:40:38.139581 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0" containerName="nova-metadata-metadata" Mar 13 15:40:38 crc kubenswrapper[4907]: I0313 15:40:38.139591 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6e87a99-1744-4002-a08b-dabf8a97e48b" containerName="nova-manage" Mar 13 15:40:38 crc kubenswrapper[4907]: I0313 15:40:38.139607 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b2e0902-2ec8-4712-8215-71643998f9ae" containerName="nova-scheduler-scheduler" Mar 13 15:40:38 crc kubenswrapper[4907]: I0313 15:40:38.139622 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0" containerName="nova-metadata-log" Mar 13 15:40:38 crc kubenswrapper[4907]: I0313 15:40:38.140263 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 13 15:40:38 crc kubenswrapper[4907]: I0313 15:40:38.142330 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Mar 13 15:40:38 crc kubenswrapper[4907]: I0313 15:40:38.148555 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 15:40:38 crc kubenswrapper[4907]: I0313 15:40:38.214273 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0-logs\") pod \"8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0\" (UID: \"8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0\") " Mar 13 15:40:38 crc kubenswrapper[4907]: I0313 15:40:38.214338 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0-config-data\") pod \"8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0\" (UID: \"8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0\") " Mar 13 15:40:38 crc kubenswrapper[4907]: I0313 15:40:38.214495 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nm2jk\" (UniqueName: \"kubernetes.io/projected/8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0-kube-api-access-nm2jk\") pod \"8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0\" (UID: \"8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0\") " Mar 13 15:40:38 crc kubenswrapper[4907]: I0313 15:40:38.214542 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0-combined-ca-bundle\") pod \"8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0\" (UID: \"8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0\") " Mar 13 15:40:38 crc kubenswrapper[4907]: I0313 15:40:38.214733 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/61a2ddd9-1fcc-427e-aaa2-f092ebb7bd30-config-data\") pod \"nova-scheduler-0\" (UID: \"61a2ddd9-1fcc-427e-aaa2-f092ebb7bd30\") " pod="openstack/nova-scheduler-0" Mar 13 15:40:38 crc kubenswrapper[4907]: I0313 15:40:38.214871 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p7hsl\" (UniqueName: \"kubernetes.io/projected/61a2ddd9-1fcc-427e-aaa2-f092ebb7bd30-kube-api-access-p7hsl\") pod \"nova-scheduler-0\" (UID: \"61a2ddd9-1fcc-427e-aaa2-f092ebb7bd30\") " pod="openstack/nova-scheduler-0" Mar 13 15:40:38 crc kubenswrapper[4907]: I0313 15:40:38.214914 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61a2ddd9-1fcc-427e-aaa2-f092ebb7bd30-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"61a2ddd9-1fcc-427e-aaa2-f092ebb7bd30\") " pod="openstack/nova-scheduler-0" Mar 13 15:40:38 crc kubenswrapper[4907]: I0313 15:40:38.215608 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0-logs" (OuterVolumeSpecName: "logs") pod "8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0" (UID: "8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:40:38 crc kubenswrapper[4907]: I0313 15:40:38.221917 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0-kube-api-access-nm2jk" (OuterVolumeSpecName: "kube-api-access-nm2jk") pod "8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0" (UID: "8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0"). InnerVolumeSpecName "kube-api-access-nm2jk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:40:38 crc kubenswrapper[4907]: I0313 15:40:38.242748 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0" (UID: "8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:40:38 crc kubenswrapper[4907]: I0313 15:40:38.253138 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 13 15:40:38 crc kubenswrapper[4907]: I0313 15:40:38.253844 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0-config-data" (OuterVolumeSpecName: "config-data") pod "8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0" (UID: "8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:40:38 crc kubenswrapper[4907]: I0313 15:40:38.316055 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p7hsl\" (UniqueName: \"kubernetes.io/projected/61a2ddd9-1fcc-427e-aaa2-f092ebb7bd30-kube-api-access-p7hsl\") pod \"nova-scheduler-0\" (UID: \"61a2ddd9-1fcc-427e-aaa2-f092ebb7bd30\") " pod="openstack/nova-scheduler-0" Mar 13 15:40:38 crc kubenswrapper[4907]: I0313 15:40:38.316100 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61a2ddd9-1fcc-427e-aaa2-f092ebb7bd30-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"61a2ddd9-1fcc-427e-aaa2-f092ebb7bd30\") " pod="openstack/nova-scheduler-0" Mar 13 15:40:38 crc kubenswrapper[4907]: I0313 15:40:38.316153 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/61a2ddd9-1fcc-427e-aaa2-f092ebb7bd30-config-data\") pod \"nova-scheduler-0\" (UID: \"61a2ddd9-1fcc-427e-aaa2-f092ebb7bd30\") " pod="openstack/nova-scheduler-0" Mar 13 15:40:38 crc kubenswrapper[4907]: I0313 15:40:38.316234 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nm2jk\" (UniqueName: \"kubernetes.io/projected/8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0-kube-api-access-nm2jk\") on node \"crc\" DevicePath \"\"" Mar 13 15:40:38 crc kubenswrapper[4907]: I0313 15:40:38.316245 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 15:40:38 crc kubenswrapper[4907]: I0313 15:40:38.316255 4907 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0-logs\") on node \"crc\" DevicePath \"\"" Mar 13 15:40:38 crc kubenswrapper[4907]: I0313 15:40:38.316263 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 15:40:38 crc kubenswrapper[4907]: I0313 15:40:38.319684 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/61a2ddd9-1fcc-427e-aaa2-f092ebb7bd30-config-data\") pod \"nova-scheduler-0\" (UID: \"61a2ddd9-1fcc-427e-aaa2-f092ebb7bd30\") " pod="openstack/nova-scheduler-0" Mar 13 15:40:38 crc kubenswrapper[4907]: I0313 15:40:38.323807 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61a2ddd9-1fcc-427e-aaa2-f092ebb7bd30-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"61a2ddd9-1fcc-427e-aaa2-f092ebb7bd30\") " pod="openstack/nova-scheduler-0" Mar 13 15:40:38 crc kubenswrapper[4907]: I0313 15:40:38.331336 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p7hsl\" (UniqueName: \"kubernetes.io/projected/61a2ddd9-1fcc-427e-aaa2-f092ebb7bd30-kube-api-access-p7hsl\") pod \"nova-scheduler-0\" (UID: \"61a2ddd9-1fcc-427e-aaa2-f092ebb7bd30\") " pod="openstack/nova-scheduler-0" Mar 13 15:40:38 crc kubenswrapper[4907]: I0313 15:40:38.417260 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad3c8be4-d237-4c1e-bd98-963acae4ac32-combined-ca-bundle\") pod \"ad3c8be4-d237-4c1e-bd98-963acae4ac32\" (UID: \"ad3c8be4-d237-4c1e-bd98-963acae4ac32\") " Mar 13 15:40:38 crc kubenswrapper[4907]: I0313 15:40:38.417352 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ad3c8be4-d237-4c1e-bd98-963acae4ac32-logs\") pod \"ad3c8be4-d237-4c1e-bd98-963acae4ac32\" (UID: \"ad3c8be4-d237-4c1e-bd98-963acae4ac32\") " Mar 13 15:40:38 crc kubenswrapper[4907]: I0313 15:40:38.417381 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x6msv\" (UniqueName: \"kubernetes.io/projected/ad3c8be4-d237-4c1e-bd98-963acae4ac32-kube-api-access-x6msv\") pod \"ad3c8be4-d237-4c1e-bd98-963acae4ac32\" (UID: \"ad3c8be4-d237-4c1e-bd98-963acae4ac32\") " Mar 13 15:40:38 crc kubenswrapper[4907]: I0313 15:40:38.417482 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad3c8be4-d237-4c1e-bd98-963acae4ac32-config-data\") pod \"ad3c8be4-d237-4c1e-bd98-963acae4ac32\" (UID: \"ad3c8be4-d237-4c1e-bd98-963acae4ac32\") " Mar 13 15:40:38 crc kubenswrapper[4907]: I0313 15:40:38.418011 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ad3c8be4-d237-4c1e-bd98-963acae4ac32-logs" (OuterVolumeSpecName: "logs") pod "ad3c8be4-d237-4c1e-bd98-963acae4ac32" (UID: "ad3c8be4-d237-4c1e-bd98-963acae4ac32"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:40:38 crc kubenswrapper[4907]: I0313 15:40:38.421011 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ad3c8be4-d237-4c1e-bd98-963acae4ac32-kube-api-access-x6msv" (OuterVolumeSpecName: "kube-api-access-x6msv") pod "ad3c8be4-d237-4c1e-bd98-963acae4ac32" (UID: "ad3c8be4-d237-4c1e-bd98-963acae4ac32"). InnerVolumeSpecName "kube-api-access-x6msv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:40:38 crc kubenswrapper[4907]: I0313 15:40:38.439492 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ad3c8be4-d237-4c1e-bd98-963acae4ac32-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ad3c8be4-d237-4c1e-bd98-963acae4ac32" (UID: "ad3c8be4-d237-4c1e-bd98-963acae4ac32"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:40:38 crc kubenswrapper[4907]: I0313 15:40:38.439555 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ad3c8be4-d237-4c1e-bd98-963acae4ac32-config-data" (OuterVolumeSpecName: "config-data") pod "ad3c8be4-d237-4c1e-bd98-963acae4ac32" (UID: "ad3c8be4-d237-4c1e-bd98-963acae4ac32"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:40:38 crc kubenswrapper[4907]: I0313 15:40:38.519495 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad3c8be4-d237-4c1e-bd98-963acae4ac32-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 15:40:38 crc kubenswrapper[4907]: I0313 15:40:38.519541 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad3c8be4-d237-4c1e-bd98-963acae4ac32-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 15:40:38 crc kubenswrapper[4907]: I0313 15:40:38.519558 4907 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ad3c8be4-d237-4c1e-bd98-963acae4ac32-logs\") on node \"crc\" DevicePath \"\"" Mar 13 15:40:38 crc kubenswrapper[4907]: I0313 15:40:38.519570 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x6msv\" (UniqueName: \"kubernetes.io/projected/ad3c8be4-d237-4c1e-bd98-963acae4ac32-kube-api-access-x6msv\") on node \"crc\" DevicePath \"\"" Mar 13 15:40:38 crc kubenswrapper[4907]: I0313 15:40:38.520570 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 13 15:40:38 crc kubenswrapper[4907]: I0313 15:40:38.957327 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 15:40:38 crc kubenswrapper[4907]: W0313 15:40:38.960555 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod61a2ddd9_1fcc_427e_aaa2_f092ebb7bd30.slice/crio-605697270f0438aa6a3971d720e6b66130d7e9f9c333f5d331abb06ba749d725 WatchSource:0}: Error finding container 605697270f0438aa6a3971d720e6b66130d7e9f9c333f5d331abb06ba749d725: Status 404 returned error can't find the container with id 605697270f0438aa6a3971d720e6b66130d7e9f9c333f5d331abb06ba749d725 Mar 13 15:40:39 crc kubenswrapper[4907]: I0313 15:40:39.082051 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"61a2ddd9-1fcc-427e-aaa2-f092ebb7bd30","Type":"ContainerStarted","Data":"605697270f0438aa6a3971d720e6b66130d7e9f9c333f5d331abb06ba749d725"} Mar 13 15:40:39 crc kubenswrapper[4907]: I0313 15:40:39.084441 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ad3c8be4-d237-4c1e-bd98-963acae4ac32","Type":"ContainerDied","Data":"7d15b64a01f946210967bf64f6d3e77361db086303536eb844ce8563c1a6da6d"} Mar 13 15:40:39 crc kubenswrapper[4907]: I0313 15:40:39.084484 4907 scope.go:117] "RemoveContainer" containerID="e833237c58d59cd7a382ff7644c38c3979cbe4fa6a8a9e044f9040c86f3823df" Mar 13 15:40:39 crc kubenswrapper[4907]: I0313 15:40:39.084522 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 13 15:40:39 crc kubenswrapper[4907]: I0313 15:40:39.090987 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0","Type":"ContainerDied","Data":"0d9dea632a506015513087187b39d9560b7d9208b5328332f68e9033c985b609"} Mar 13 15:40:39 crc kubenswrapper[4907]: I0313 15:40:39.091130 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 13 15:40:39 crc kubenswrapper[4907]: I0313 15:40:39.123092 4907 scope.go:117] "RemoveContainer" containerID="a9b52fcfd1633f1cfd4fe4e130d7daf6a6a0f261d982b7490f4c49360b713147" Mar 13 15:40:39 crc kubenswrapper[4907]: I0313 15:40:39.144976 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Mar 13 15:40:39 crc kubenswrapper[4907]: I0313 15:40:39.162163 4907 scope.go:117] "RemoveContainer" containerID="9d4f1d5de2313cc0cacaa0de90599bf7cc29209144c55b94e7e5f63f34cd569d" Mar 13 15:40:39 crc kubenswrapper[4907]: I0313 15:40:39.173214 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Mar 13 15:40:39 crc kubenswrapper[4907]: I0313 15:40:39.185359 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Mar 13 15:40:39 crc kubenswrapper[4907]: I0313 15:40:39.201533 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Mar 13 15:40:39 crc kubenswrapper[4907]: E0313 15:40:39.201973 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad3c8be4-d237-4c1e-bd98-963acae4ac32" containerName="nova-api-log" Mar 13 15:40:39 crc kubenswrapper[4907]: I0313 15:40:39.201992 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad3c8be4-d237-4c1e-bd98-963acae4ac32" containerName="nova-api-log" Mar 13 15:40:39 crc kubenswrapper[4907]: E0313 15:40:39.202031 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad3c8be4-d237-4c1e-bd98-963acae4ac32" containerName="nova-api-api" Mar 13 15:40:39 crc kubenswrapper[4907]: I0313 15:40:39.202038 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad3c8be4-d237-4c1e-bd98-963acae4ac32" containerName="nova-api-api" Mar 13 15:40:39 crc kubenswrapper[4907]: I0313 15:40:39.202210 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="ad3c8be4-d237-4c1e-bd98-963acae4ac32" containerName="nova-api-log" Mar 13 15:40:39 crc kubenswrapper[4907]: I0313 15:40:39.202232 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="ad3c8be4-d237-4c1e-bd98-963acae4ac32" containerName="nova-api-api" Mar 13 15:40:39 crc kubenswrapper[4907]: I0313 15:40:39.203227 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 13 15:40:39 crc kubenswrapper[4907]: I0313 15:40:39.203963 4907 scope.go:117] "RemoveContainer" containerID="12ad1f3ba8429b32fe4517953f072fa40cf2ef694f54ab5cea4b98ef8339414c" Mar 13 15:40:39 crc kubenswrapper[4907]: I0313 15:40:39.212074 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Mar 13 15:40:39 crc kubenswrapper[4907]: I0313 15:40:39.215105 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Mar 13 15:40:39 crc kubenswrapper[4907]: I0313 15:40:39.228388 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 13 15:40:39 crc kubenswrapper[4907]: I0313 15:40:39.232512 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xct78\" (UniqueName: \"kubernetes.io/projected/81ef3cd7-e06f-49f8-bc01-41b32fbb5546-kube-api-access-xct78\") pod \"nova-api-0\" (UID: \"81ef3cd7-e06f-49f8-bc01-41b32fbb5546\") " pod="openstack/nova-api-0" Mar 13 15:40:39 crc kubenswrapper[4907]: I0313 15:40:39.232663 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/81ef3cd7-e06f-49f8-bc01-41b32fbb5546-logs\") pod \"nova-api-0\" (UID: \"81ef3cd7-e06f-49f8-bc01-41b32fbb5546\") " pod="openstack/nova-api-0" Mar 13 15:40:39 crc kubenswrapper[4907]: I0313 15:40:39.232766 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81ef3cd7-e06f-49f8-bc01-41b32fbb5546-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"81ef3cd7-e06f-49f8-bc01-41b32fbb5546\") " pod="openstack/nova-api-0" Mar 13 15:40:39 crc kubenswrapper[4907]: I0313 15:40:39.232923 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81ef3cd7-e06f-49f8-bc01-41b32fbb5546-config-data\") pod \"nova-api-0\" (UID: \"81ef3cd7-e06f-49f8-bc01-41b32fbb5546\") " pod="openstack/nova-api-0" Mar 13 15:40:39 crc kubenswrapper[4907]: I0313 15:40:39.236705 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Mar 13 15:40:39 crc kubenswrapper[4907]: I0313 15:40:39.238615 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 13 15:40:39 crc kubenswrapper[4907]: I0313 15:40:39.242718 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Mar 13 15:40:39 crc kubenswrapper[4907]: I0313 15:40:39.246483 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Mar 13 15:40:39 crc kubenswrapper[4907]: I0313 15:40:39.335053 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81ef3cd7-e06f-49f8-bc01-41b32fbb5546-config-data\") pod \"nova-api-0\" (UID: \"81ef3cd7-e06f-49f8-bc01-41b32fbb5546\") " pod="openstack/nova-api-0" Mar 13 15:40:39 crc kubenswrapper[4907]: I0313 15:40:39.335111 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37198210-23ed-4025-8cd9-e284c6573318-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"37198210-23ed-4025-8cd9-e284c6573318\") " pod="openstack/nova-metadata-0" Mar 13 15:40:39 crc kubenswrapper[4907]: I0313 15:40:39.335156 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bdf6j\" (UniqueName: \"kubernetes.io/projected/37198210-23ed-4025-8cd9-e284c6573318-kube-api-access-bdf6j\") pod \"nova-metadata-0\" (UID: \"37198210-23ed-4025-8cd9-e284c6573318\") " pod="openstack/nova-metadata-0" Mar 13 15:40:39 crc kubenswrapper[4907]: I0313 15:40:39.335315 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xct78\" (UniqueName: \"kubernetes.io/projected/81ef3cd7-e06f-49f8-bc01-41b32fbb5546-kube-api-access-xct78\") pod \"nova-api-0\" (UID: \"81ef3cd7-e06f-49f8-bc01-41b32fbb5546\") " pod="openstack/nova-api-0" Mar 13 15:40:39 crc kubenswrapper[4907]: I0313 15:40:39.335392 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/37198210-23ed-4025-8cd9-e284c6573318-logs\") pod \"nova-metadata-0\" (UID: \"37198210-23ed-4025-8cd9-e284c6573318\") " pod="openstack/nova-metadata-0" Mar 13 15:40:39 crc kubenswrapper[4907]: I0313 15:40:39.335473 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/81ef3cd7-e06f-49f8-bc01-41b32fbb5546-logs\") pod \"nova-api-0\" (UID: \"81ef3cd7-e06f-49f8-bc01-41b32fbb5546\") " pod="openstack/nova-api-0" Mar 13 15:40:39 crc kubenswrapper[4907]: I0313 15:40:39.335585 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81ef3cd7-e06f-49f8-bc01-41b32fbb5546-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"81ef3cd7-e06f-49f8-bc01-41b32fbb5546\") " pod="openstack/nova-api-0" Mar 13 15:40:39 crc kubenswrapper[4907]: I0313 15:40:39.335902 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/81ef3cd7-e06f-49f8-bc01-41b32fbb5546-logs\") pod \"nova-api-0\" (UID: \"81ef3cd7-e06f-49f8-bc01-41b32fbb5546\") " pod="openstack/nova-api-0" Mar 13 15:40:39 crc kubenswrapper[4907]: I0313 15:40:39.338427 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37198210-23ed-4025-8cd9-e284c6573318-config-data\") pod \"nova-metadata-0\" (UID: \"37198210-23ed-4025-8cd9-e284c6573318\") " pod="openstack/nova-metadata-0" Mar 13 15:40:39 crc kubenswrapper[4907]: I0313 15:40:39.339758 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81ef3cd7-e06f-49f8-bc01-41b32fbb5546-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"81ef3cd7-e06f-49f8-bc01-41b32fbb5546\") " pod="openstack/nova-api-0" Mar 13 15:40:39 crc kubenswrapper[4907]: I0313 15:40:39.340452 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81ef3cd7-e06f-49f8-bc01-41b32fbb5546-config-data\") pod \"nova-api-0\" (UID: \"81ef3cd7-e06f-49f8-bc01-41b32fbb5546\") " pod="openstack/nova-api-0" Mar 13 15:40:39 crc kubenswrapper[4907]: I0313 15:40:39.351321 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xct78\" (UniqueName: \"kubernetes.io/projected/81ef3cd7-e06f-49f8-bc01-41b32fbb5546-kube-api-access-xct78\") pod \"nova-api-0\" (UID: \"81ef3cd7-e06f-49f8-bc01-41b32fbb5546\") " pod="openstack/nova-api-0" Mar 13 15:40:39 crc kubenswrapper[4907]: I0313 15:40:39.439700 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37198210-23ed-4025-8cd9-e284c6573318-config-data\") pod \"nova-metadata-0\" (UID: \"37198210-23ed-4025-8cd9-e284c6573318\") " pod="openstack/nova-metadata-0" Mar 13 15:40:39 crc kubenswrapper[4907]: I0313 15:40:39.439780 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37198210-23ed-4025-8cd9-e284c6573318-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"37198210-23ed-4025-8cd9-e284c6573318\") " pod="openstack/nova-metadata-0" Mar 13 15:40:39 crc kubenswrapper[4907]: I0313 15:40:39.439823 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bdf6j\" (UniqueName: \"kubernetes.io/projected/37198210-23ed-4025-8cd9-e284c6573318-kube-api-access-bdf6j\") pod \"nova-metadata-0\" (UID: \"37198210-23ed-4025-8cd9-e284c6573318\") " pod="openstack/nova-metadata-0" Mar 13 15:40:39 crc kubenswrapper[4907]: I0313 15:40:39.439949 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/37198210-23ed-4025-8cd9-e284c6573318-logs\") pod \"nova-metadata-0\" (UID: \"37198210-23ed-4025-8cd9-e284c6573318\") " pod="openstack/nova-metadata-0" Mar 13 15:40:39 crc kubenswrapper[4907]: I0313 15:40:39.440427 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/37198210-23ed-4025-8cd9-e284c6573318-logs\") pod \"nova-metadata-0\" (UID: \"37198210-23ed-4025-8cd9-e284c6573318\") " pod="openstack/nova-metadata-0" Mar 13 15:40:39 crc kubenswrapper[4907]: I0313 15:40:39.443474 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37198210-23ed-4025-8cd9-e284c6573318-config-data\") pod \"nova-metadata-0\" (UID: \"37198210-23ed-4025-8cd9-e284c6573318\") " pod="openstack/nova-metadata-0" Mar 13 15:40:39 crc kubenswrapper[4907]: I0313 15:40:39.452251 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37198210-23ed-4025-8cd9-e284c6573318-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"37198210-23ed-4025-8cd9-e284c6573318\") " pod="openstack/nova-metadata-0" Mar 13 15:40:39 crc kubenswrapper[4907]: I0313 15:40:39.455337 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bdf6j\" (UniqueName: \"kubernetes.io/projected/37198210-23ed-4025-8cd9-e284c6573318-kube-api-access-bdf6j\") pod \"nova-metadata-0\" (UID: \"37198210-23ed-4025-8cd9-e284c6573318\") " pod="openstack/nova-metadata-0" Mar 13 15:40:39 crc kubenswrapper[4907]: I0313 15:40:39.525150 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 13 15:40:39 crc kubenswrapper[4907]: I0313 15:40:39.558441 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 13 15:40:39 crc kubenswrapper[4907]: I0313 15:40:39.814615 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6b2e0902-2ec8-4712-8215-71643998f9ae" path="/var/lib/kubelet/pods/6b2e0902-2ec8-4712-8215-71643998f9ae/volumes" Mar 13 15:40:39 crc kubenswrapper[4907]: I0313 15:40:39.818522 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0" path="/var/lib/kubelet/pods/8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0/volumes" Mar 13 15:40:39 crc kubenswrapper[4907]: I0313 15:40:39.821033 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ad3c8be4-d237-4c1e-bd98-963acae4ac32" path="/var/lib/kubelet/pods/ad3c8be4-d237-4c1e-bd98-963acae4ac32/volumes" Mar 13 15:40:40 crc kubenswrapper[4907]: I0313 15:40:40.008110 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 13 15:40:40 crc kubenswrapper[4907]: W0313 15:40:40.009630 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod81ef3cd7_e06f_49f8_bc01_41b32fbb5546.slice/crio-c85aa3e7cd1bf81f73dff61921375db95cb273392a8693e2f1fe5ad2b1d643c9 WatchSource:0}: Error finding container c85aa3e7cd1bf81f73dff61921375db95cb273392a8693e2f1fe5ad2b1d643c9: Status 404 returned error can't find the container with id c85aa3e7cd1bf81f73dff61921375db95cb273392a8693e2f1fe5ad2b1d643c9 Mar 13 15:40:40 crc kubenswrapper[4907]: I0313 15:40:40.016148 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Mar 13 15:40:40 crc kubenswrapper[4907]: W0313 15:40:40.023958 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod37198210_23ed_4025_8cd9_e284c6573318.slice/crio-54a994af9272c8e084ae758a74263408bf4549d1118439e6fd32c7e4d5aaa9e5 WatchSource:0}: Error finding container 54a994af9272c8e084ae758a74263408bf4549d1118439e6fd32c7e4d5aaa9e5: Status 404 returned error can't find the container with id 54a994af9272c8e084ae758a74263408bf4549d1118439e6fd32c7e4d5aaa9e5 Mar 13 15:40:40 crc kubenswrapper[4907]: I0313 15:40:40.116429 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"37198210-23ed-4025-8cd9-e284c6573318","Type":"ContainerStarted","Data":"54a994af9272c8e084ae758a74263408bf4549d1118439e6fd32c7e4d5aaa9e5"} Mar 13 15:40:40 crc kubenswrapper[4907]: I0313 15:40:40.119482 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"61a2ddd9-1fcc-427e-aaa2-f092ebb7bd30","Type":"ContainerStarted","Data":"ce84852f1d320a52f408594e4b89f2cbf753717a67ce179872f364593d90c469"} Mar 13 15:40:40 crc kubenswrapper[4907]: I0313 15:40:40.125018 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"81ef3cd7-e06f-49f8-bc01-41b32fbb5546","Type":"ContainerStarted","Data":"c85aa3e7cd1bf81f73dff61921375db95cb273392a8693e2f1fe5ad2b1d643c9"} Mar 13 15:40:40 crc kubenswrapper[4907]: I0313 15:40:40.146788 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.146746347 podStartE2EDuration="2.146746347s" podCreationTimestamp="2026-03-13 15:40:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:40:40.135067898 +0000 UTC m=+5739.034855577" watchObservedRunningTime="2026-03-13 15:40:40.146746347 +0000 UTC m=+5739.046534036" Mar 13 15:40:41 crc kubenswrapper[4907]: I0313 15:40:41.134174 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"37198210-23ed-4025-8cd9-e284c6573318","Type":"ContainerStarted","Data":"ae907961f503a65c4dfb28c60282669570b0349aea1f6830f2e1f6e44d773cdb"} Mar 13 15:40:41 crc kubenswrapper[4907]: I0313 15:40:41.134502 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"37198210-23ed-4025-8cd9-e284c6573318","Type":"ContainerStarted","Data":"6dcdbcfacbeb1b7e6dcbc3fb09e947521ce095a95f3551396c6c38c4e6a8b77f"} Mar 13 15:40:41 crc kubenswrapper[4907]: I0313 15:40:41.136432 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"81ef3cd7-e06f-49f8-bc01-41b32fbb5546","Type":"ContainerStarted","Data":"da08a8ab4ad1ce90fd28c0cf97955538dcb49b21c61a30f6afb435225b153254"} Mar 13 15:40:41 crc kubenswrapper[4907]: I0313 15:40:41.136476 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"81ef3cd7-e06f-49f8-bc01-41b32fbb5546","Type":"ContainerStarted","Data":"697da8faed8339f44e9c8f069a46c98c95d40fcca3ba8965d84a7f51ddc0a358"} Mar 13 15:40:41 crc kubenswrapper[4907]: I0313 15:40:41.155936 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.155911819 podStartE2EDuration="2.155911819s" podCreationTimestamp="2026-03-13 15:40:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:40:41.149894134 +0000 UTC m=+5740.049681823" watchObservedRunningTime="2026-03-13 15:40:41.155911819 +0000 UTC m=+5740.055699508" Mar 13 15:40:41 crc kubenswrapper[4907]: I0313 15:40:41.196659 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.196630962 podStartE2EDuration="2.196630962s" podCreationTimestamp="2026-03-13 15:40:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:40:41.176211704 +0000 UTC m=+5740.075999393" watchObservedRunningTime="2026-03-13 15:40:41.196630962 +0000 UTC m=+5740.096418651" Mar 13 15:40:43 crc kubenswrapper[4907]: I0313 15:40:43.521583 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Mar 13 15:40:48 crc kubenswrapper[4907]: I0313 15:40:48.041374 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 15:40:48 crc kubenswrapper[4907]: I0313 15:40:48.043349 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 15:40:48 crc kubenswrapper[4907]: I0313 15:40:48.521301 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Mar 13 15:40:48 crc kubenswrapper[4907]: I0313 15:40:48.551369 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Mar 13 15:40:49 crc kubenswrapper[4907]: I0313 15:40:49.235972 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Mar 13 15:40:49 crc kubenswrapper[4907]: I0313 15:40:49.526079 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Mar 13 15:40:49 crc kubenswrapper[4907]: I0313 15:40:49.526426 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Mar 13 15:40:49 crc kubenswrapper[4907]: I0313 15:40:49.559421 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Mar 13 15:40:49 crc kubenswrapper[4907]: I0313 15:40:49.559617 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Mar 13 15:40:50 crc kubenswrapper[4907]: I0313 15:40:50.609166 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="81ef3cd7-e06f-49f8-bc01-41b32fbb5546" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.1.113:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 15:40:50 crc kubenswrapper[4907]: I0313 15:40:50.692133 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="37198210-23ed-4025-8cd9-e284c6573318" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.1.114:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 15:40:50 crc kubenswrapper[4907]: I0313 15:40:50.692315 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="37198210-23ed-4025-8cd9-e284c6573318" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.1.114:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 15:40:50 crc kubenswrapper[4907]: I0313 15:40:50.692133 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="81ef3cd7-e06f-49f8-bc01-41b32fbb5546" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.1.113:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 15:40:57 crc kubenswrapper[4907]: I0313 15:40:57.526203 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Mar 13 15:40:57 crc kubenswrapper[4907]: I0313 15:40:57.526612 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Mar 13 15:40:57 crc kubenswrapper[4907]: I0313 15:40:57.559842 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Mar 13 15:40:57 crc kubenswrapper[4907]: I0313 15:40:57.560015 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Mar 13 15:40:59 crc kubenswrapper[4907]: I0313 15:40:59.529560 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Mar 13 15:40:59 crc kubenswrapper[4907]: I0313 15:40:59.532141 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Mar 13 15:40:59 crc kubenswrapper[4907]: I0313 15:40:59.534607 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Mar 13 15:40:59 crc kubenswrapper[4907]: I0313 15:40:59.562091 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Mar 13 15:40:59 crc kubenswrapper[4907]: I0313 15:40:59.563480 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Mar 13 15:40:59 crc kubenswrapper[4907]: I0313 15:40:59.569124 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Mar 13 15:41:00 crc kubenswrapper[4907]: I0313 15:41:00.294931 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Mar 13 15:41:00 crc kubenswrapper[4907]: I0313 15:41:00.295845 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Mar 13 15:41:00 crc kubenswrapper[4907]: I0313 15:41:00.496466 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-f9699695f-pwxvd"] Mar 13 15:41:00 crc kubenswrapper[4907]: I0313 15:41:00.499662 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f9699695f-pwxvd" Mar 13 15:41:00 crc kubenswrapper[4907]: I0313 15:41:00.526545 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-f9699695f-pwxvd"] Mar 13 15:41:00 crc kubenswrapper[4907]: I0313 15:41:00.625638 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a122794c-3b28-4344-9817-f05a9e663110-ovsdbserver-nb\") pod \"dnsmasq-dns-f9699695f-pwxvd\" (UID: \"a122794c-3b28-4344-9817-f05a9e663110\") " pod="openstack/dnsmasq-dns-f9699695f-pwxvd" Mar 13 15:41:00 crc kubenswrapper[4907]: I0313 15:41:00.625696 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a122794c-3b28-4344-9817-f05a9e663110-ovsdbserver-sb\") pod \"dnsmasq-dns-f9699695f-pwxvd\" (UID: \"a122794c-3b28-4344-9817-f05a9e663110\") " pod="openstack/dnsmasq-dns-f9699695f-pwxvd" Mar 13 15:41:00 crc kubenswrapper[4907]: I0313 15:41:00.625775 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a122794c-3b28-4344-9817-f05a9e663110-dns-svc\") pod \"dnsmasq-dns-f9699695f-pwxvd\" (UID: \"a122794c-3b28-4344-9817-f05a9e663110\") " pod="openstack/dnsmasq-dns-f9699695f-pwxvd" Mar 13 15:41:00 crc kubenswrapper[4907]: I0313 15:41:00.625809 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a122794c-3b28-4344-9817-f05a9e663110-config\") pod \"dnsmasq-dns-f9699695f-pwxvd\" (UID: \"a122794c-3b28-4344-9817-f05a9e663110\") " pod="openstack/dnsmasq-dns-f9699695f-pwxvd" Mar 13 15:41:00 crc kubenswrapper[4907]: I0313 15:41:00.625929 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d2fkn\" (UniqueName: \"kubernetes.io/projected/a122794c-3b28-4344-9817-f05a9e663110-kube-api-access-d2fkn\") pod \"dnsmasq-dns-f9699695f-pwxvd\" (UID: \"a122794c-3b28-4344-9817-f05a9e663110\") " pod="openstack/dnsmasq-dns-f9699695f-pwxvd" Mar 13 15:41:00 crc kubenswrapper[4907]: I0313 15:41:00.727192 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a122794c-3b28-4344-9817-f05a9e663110-ovsdbserver-nb\") pod \"dnsmasq-dns-f9699695f-pwxvd\" (UID: \"a122794c-3b28-4344-9817-f05a9e663110\") " pod="openstack/dnsmasq-dns-f9699695f-pwxvd" Mar 13 15:41:00 crc kubenswrapper[4907]: I0313 15:41:00.727252 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a122794c-3b28-4344-9817-f05a9e663110-ovsdbserver-sb\") pod \"dnsmasq-dns-f9699695f-pwxvd\" (UID: \"a122794c-3b28-4344-9817-f05a9e663110\") " pod="openstack/dnsmasq-dns-f9699695f-pwxvd" Mar 13 15:41:00 crc kubenswrapper[4907]: I0313 15:41:00.727287 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a122794c-3b28-4344-9817-f05a9e663110-dns-svc\") pod \"dnsmasq-dns-f9699695f-pwxvd\" (UID: \"a122794c-3b28-4344-9817-f05a9e663110\") " pod="openstack/dnsmasq-dns-f9699695f-pwxvd" Mar 13 15:41:00 crc kubenswrapper[4907]: I0313 15:41:00.727311 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a122794c-3b28-4344-9817-f05a9e663110-config\") pod \"dnsmasq-dns-f9699695f-pwxvd\" (UID: \"a122794c-3b28-4344-9817-f05a9e663110\") " pod="openstack/dnsmasq-dns-f9699695f-pwxvd" Mar 13 15:41:00 crc kubenswrapper[4907]: I0313 15:41:00.727349 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d2fkn\" (UniqueName: \"kubernetes.io/projected/a122794c-3b28-4344-9817-f05a9e663110-kube-api-access-d2fkn\") pod \"dnsmasq-dns-f9699695f-pwxvd\" (UID: \"a122794c-3b28-4344-9817-f05a9e663110\") " pod="openstack/dnsmasq-dns-f9699695f-pwxvd" Mar 13 15:41:00 crc kubenswrapper[4907]: I0313 15:41:00.728379 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a122794c-3b28-4344-9817-f05a9e663110-dns-svc\") pod \"dnsmasq-dns-f9699695f-pwxvd\" (UID: \"a122794c-3b28-4344-9817-f05a9e663110\") " pod="openstack/dnsmasq-dns-f9699695f-pwxvd" Mar 13 15:41:00 crc kubenswrapper[4907]: I0313 15:41:00.728407 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a122794c-3b28-4344-9817-f05a9e663110-ovsdbserver-sb\") pod \"dnsmasq-dns-f9699695f-pwxvd\" (UID: \"a122794c-3b28-4344-9817-f05a9e663110\") " pod="openstack/dnsmasq-dns-f9699695f-pwxvd" Mar 13 15:41:00 crc kubenswrapper[4907]: I0313 15:41:00.728447 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a122794c-3b28-4344-9817-f05a9e663110-ovsdbserver-nb\") pod \"dnsmasq-dns-f9699695f-pwxvd\" (UID: \"a122794c-3b28-4344-9817-f05a9e663110\") " pod="openstack/dnsmasq-dns-f9699695f-pwxvd" Mar 13 15:41:00 crc kubenswrapper[4907]: I0313 15:41:00.728480 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a122794c-3b28-4344-9817-f05a9e663110-config\") pod \"dnsmasq-dns-f9699695f-pwxvd\" (UID: \"a122794c-3b28-4344-9817-f05a9e663110\") " pod="openstack/dnsmasq-dns-f9699695f-pwxvd" Mar 13 15:41:00 crc kubenswrapper[4907]: I0313 15:41:00.749102 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d2fkn\" (UniqueName: \"kubernetes.io/projected/a122794c-3b28-4344-9817-f05a9e663110-kube-api-access-d2fkn\") pod \"dnsmasq-dns-f9699695f-pwxvd\" (UID: \"a122794c-3b28-4344-9817-f05a9e663110\") " pod="openstack/dnsmasq-dns-f9699695f-pwxvd" Mar 13 15:41:00 crc kubenswrapper[4907]: I0313 15:41:00.831050 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f9699695f-pwxvd" Mar 13 15:41:01 crc kubenswrapper[4907]: I0313 15:41:01.330451 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-f9699695f-pwxvd"] Mar 13 15:41:01 crc kubenswrapper[4907]: W0313 15:41:01.334988 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda122794c_3b28_4344_9817_f05a9e663110.slice/crio-8f46474c0bd8af139725a1ba39a29f7abd9cd2e12d0840cb8535991ab1cd685e WatchSource:0}: Error finding container 8f46474c0bd8af139725a1ba39a29f7abd9cd2e12d0840cb8535991ab1cd685e: Status 404 returned error can't find the container with id 8f46474c0bd8af139725a1ba39a29f7abd9cd2e12d0840cb8535991ab1cd685e Mar 13 15:41:02 crc kubenswrapper[4907]: I0313 15:41:02.312035 4907 generic.go:334] "Generic (PLEG): container finished" podID="a122794c-3b28-4344-9817-f05a9e663110" containerID="ec3db5cd0d09990bedd2c273757078f6b84cc7be565b5f045a0e95e3abe11b83" exitCode=0 Mar 13 15:41:02 crc kubenswrapper[4907]: I0313 15:41:02.312153 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f9699695f-pwxvd" event={"ID":"a122794c-3b28-4344-9817-f05a9e663110","Type":"ContainerDied","Data":"ec3db5cd0d09990bedd2c273757078f6b84cc7be565b5f045a0e95e3abe11b83"} Mar 13 15:41:02 crc kubenswrapper[4907]: I0313 15:41:02.312601 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f9699695f-pwxvd" event={"ID":"a122794c-3b28-4344-9817-f05a9e663110","Type":"ContainerStarted","Data":"8f46474c0bd8af139725a1ba39a29f7abd9cd2e12d0840cb8535991ab1cd685e"} Mar 13 15:41:03 crc kubenswrapper[4907]: I0313 15:41:03.322564 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f9699695f-pwxvd" event={"ID":"a122794c-3b28-4344-9817-f05a9e663110","Type":"ContainerStarted","Data":"73bbc4d60c70852afe1d527317bf4861e321326bd2af88b210073cad5a3608bd"} Mar 13 15:41:03 crc kubenswrapper[4907]: I0313 15:41:03.322922 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-f9699695f-pwxvd" Mar 13 15:41:03 crc kubenswrapper[4907]: I0313 15:41:03.340069 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-f9699695f-pwxvd" podStartSLOduration=3.340045119 podStartE2EDuration="3.340045119s" podCreationTimestamp="2026-03-13 15:41:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:41:03.337305744 +0000 UTC m=+5762.237093433" watchObservedRunningTime="2026-03-13 15:41:03.340045119 +0000 UTC m=+5762.239832848" Mar 13 15:41:10 crc kubenswrapper[4907]: I0313 15:41:10.833039 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-f9699695f-pwxvd" Mar 13 15:41:10 crc kubenswrapper[4907]: I0313 15:41:10.893298 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-68dbfc785f-kjm62"] Mar 13 15:41:10 crc kubenswrapper[4907]: I0313 15:41:10.893529 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-68dbfc785f-kjm62" podUID="690a67d3-7115-4897-ba26-edd5eeff0521" containerName="dnsmasq-dns" containerID="cri-o://b0ec28b6beaa68b816d49abaf6a1073e21a0e38ace3bf382a509d710008100ce" gracePeriod=10 Mar 13 15:41:11 crc kubenswrapper[4907]: I0313 15:41:11.419807 4907 generic.go:334] "Generic (PLEG): container finished" podID="690a67d3-7115-4897-ba26-edd5eeff0521" containerID="b0ec28b6beaa68b816d49abaf6a1073e21a0e38ace3bf382a509d710008100ce" exitCode=0 Mar 13 15:41:11 crc kubenswrapper[4907]: I0313 15:41:11.420112 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-68dbfc785f-kjm62" event={"ID":"690a67d3-7115-4897-ba26-edd5eeff0521","Type":"ContainerDied","Data":"b0ec28b6beaa68b816d49abaf6a1073e21a0e38ace3bf382a509d710008100ce"} Mar 13 15:41:11 crc kubenswrapper[4907]: I0313 15:41:11.549990 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-68dbfc785f-kjm62" Mar 13 15:41:11 crc kubenswrapper[4907]: I0313 15:41:11.634813 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/690a67d3-7115-4897-ba26-edd5eeff0521-dns-svc\") pod \"690a67d3-7115-4897-ba26-edd5eeff0521\" (UID: \"690a67d3-7115-4897-ba26-edd5eeff0521\") " Mar 13 15:41:11 crc kubenswrapper[4907]: I0313 15:41:11.634932 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x55kq\" (UniqueName: \"kubernetes.io/projected/690a67d3-7115-4897-ba26-edd5eeff0521-kube-api-access-x55kq\") pod \"690a67d3-7115-4897-ba26-edd5eeff0521\" (UID: \"690a67d3-7115-4897-ba26-edd5eeff0521\") " Mar 13 15:41:11 crc kubenswrapper[4907]: I0313 15:41:11.634964 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/690a67d3-7115-4897-ba26-edd5eeff0521-ovsdbserver-nb\") pod \"690a67d3-7115-4897-ba26-edd5eeff0521\" (UID: \"690a67d3-7115-4897-ba26-edd5eeff0521\") " Mar 13 15:41:11 crc kubenswrapper[4907]: I0313 15:41:11.635036 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/690a67d3-7115-4897-ba26-edd5eeff0521-ovsdbserver-sb\") pod \"690a67d3-7115-4897-ba26-edd5eeff0521\" (UID: \"690a67d3-7115-4897-ba26-edd5eeff0521\") " Mar 13 15:41:11 crc kubenswrapper[4907]: I0313 15:41:11.635082 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/690a67d3-7115-4897-ba26-edd5eeff0521-config\") pod \"690a67d3-7115-4897-ba26-edd5eeff0521\" (UID: \"690a67d3-7115-4897-ba26-edd5eeff0521\") " Mar 13 15:41:11 crc kubenswrapper[4907]: I0313 15:41:11.642636 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/690a67d3-7115-4897-ba26-edd5eeff0521-kube-api-access-x55kq" (OuterVolumeSpecName: "kube-api-access-x55kq") pod "690a67d3-7115-4897-ba26-edd5eeff0521" (UID: "690a67d3-7115-4897-ba26-edd5eeff0521"). InnerVolumeSpecName "kube-api-access-x55kq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:41:11 crc kubenswrapper[4907]: I0313 15:41:11.690801 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/690a67d3-7115-4897-ba26-edd5eeff0521-config" (OuterVolumeSpecName: "config") pod "690a67d3-7115-4897-ba26-edd5eeff0521" (UID: "690a67d3-7115-4897-ba26-edd5eeff0521"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:41:11 crc kubenswrapper[4907]: I0313 15:41:11.702813 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/690a67d3-7115-4897-ba26-edd5eeff0521-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "690a67d3-7115-4897-ba26-edd5eeff0521" (UID: "690a67d3-7115-4897-ba26-edd5eeff0521"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:41:11 crc kubenswrapper[4907]: I0313 15:41:11.708770 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/690a67d3-7115-4897-ba26-edd5eeff0521-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "690a67d3-7115-4897-ba26-edd5eeff0521" (UID: "690a67d3-7115-4897-ba26-edd5eeff0521"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:41:11 crc kubenswrapper[4907]: I0313 15:41:11.727571 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/690a67d3-7115-4897-ba26-edd5eeff0521-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "690a67d3-7115-4897-ba26-edd5eeff0521" (UID: "690a67d3-7115-4897-ba26-edd5eeff0521"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:41:11 crc kubenswrapper[4907]: I0313 15:41:11.742588 4907 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/690a67d3-7115-4897-ba26-edd5eeff0521-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 13 15:41:11 crc kubenswrapper[4907]: I0313 15:41:11.742628 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/690a67d3-7115-4897-ba26-edd5eeff0521-config\") on node \"crc\" DevicePath \"\"" Mar 13 15:41:11 crc kubenswrapper[4907]: I0313 15:41:11.742664 4907 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/690a67d3-7115-4897-ba26-edd5eeff0521-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 13 15:41:11 crc kubenswrapper[4907]: I0313 15:41:11.742677 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x55kq\" (UniqueName: \"kubernetes.io/projected/690a67d3-7115-4897-ba26-edd5eeff0521-kube-api-access-x55kq\") on node \"crc\" DevicePath \"\"" Mar 13 15:41:11 crc kubenswrapper[4907]: I0313 15:41:11.742691 4907 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/690a67d3-7115-4897-ba26-edd5eeff0521-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 13 15:41:12 crc kubenswrapper[4907]: I0313 15:41:12.429621 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-68dbfc785f-kjm62" event={"ID":"690a67d3-7115-4897-ba26-edd5eeff0521","Type":"ContainerDied","Data":"4598382ec750564a06ecd379f171c64e71e4ac2c6f182c30bec9ff89463c8af3"} Mar 13 15:41:12 crc kubenswrapper[4907]: I0313 15:41:12.429937 4907 scope.go:117] "RemoveContainer" containerID="b0ec28b6beaa68b816d49abaf6a1073e21a0e38ace3bf382a509d710008100ce" Mar 13 15:41:12 crc kubenswrapper[4907]: I0313 15:41:12.429673 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-68dbfc785f-kjm62" Mar 13 15:41:12 crc kubenswrapper[4907]: I0313 15:41:12.456807 4907 scope.go:117] "RemoveContainer" containerID="5f757938e0484b76f29e92050d686900e13810e4b83c689080ec66591a9b77ed" Mar 13 15:41:12 crc kubenswrapper[4907]: I0313 15:41:12.485261 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-68dbfc785f-kjm62"] Mar 13 15:41:12 crc kubenswrapper[4907]: I0313 15:41:12.494785 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-68dbfc785f-kjm62"] Mar 13 15:41:12 crc kubenswrapper[4907]: I0313 15:41:12.833001 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-pvpwm"] Mar 13 15:41:12 crc kubenswrapper[4907]: E0313 15:41:12.833361 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="690a67d3-7115-4897-ba26-edd5eeff0521" containerName="init" Mar 13 15:41:12 crc kubenswrapper[4907]: I0313 15:41:12.833378 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="690a67d3-7115-4897-ba26-edd5eeff0521" containerName="init" Mar 13 15:41:12 crc kubenswrapper[4907]: E0313 15:41:12.833411 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="690a67d3-7115-4897-ba26-edd5eeff0521" containerName="dnsmasq-dns" Mar 13 15:41:12 crc kubenswrapper[4907]: I0313 15:41:12.833417 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="690a67d3-7115-4897-ba26-edd5eeff0521" containerName="dnsmasq-dns" Mar 13 15:41:12 crc kubenswrapper[4907]: I0313 15:41:12.833587 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="690a67d3-7115-4897-ba26-edd5eeff0521" containerName="dnsmasq-dns" Mar 13 15:41:12 crc kubenswrapper[4907]: I0313 15:41:12.834160 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-pvpwm" Mar 13 15:41:12 crc kubenswrapper[4907]: I0313 15:41:12.850281 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-pvpwm"] Mar 13 15:41:12 crc kubenswrapper[4907]: I0313 15:41:12.861157 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/499bc96f-8fa9-4900-bd92-2a507215791f-operator-scripts\") pod \"cinder-db-create-pvpwm\" (UID: \"499bc96f-8fa9-4900-bd92-2a507215791f\") " pod="openstack/cinder-db-create-pvpwm" Mar 13 15:41:12 crc kubenswrapper[4907]: I0313 15:41:12.861420 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-52hkg\" (UniqueName: \"kubernetes.io/projected/499bc96f-8fa9-4900-bd92-2a507215791f-kube-api-access-52hkg\") pod \"cinder-db-create-pvpwm\" (UID: \"499bc96f-8fa9-4900-bd92-2a507215791f\") " pod="openstack/cinder-db-create-pvpwm" Mar 13 15:41:12 crc kubenswrapper[4907]: I0313 15:41:12.940294 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-9290-account-create-update-qws42"] Mar 13 15:41:12 crc kubenswrapper[4907]: I0313 15:41:12.941743 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-9290-account-create-update-qws42" Mar 13 15:41:12 crc kubenswrapper[4907]: I0313 15:41:12.943770 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Mar 13 15:41:12 crc kubenswrapper[4907]: I0313 15:41:12.959925 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-9290-account-create-update-qws42"] Mar 13 15:41:12 crc kubenswrapper[4907]: I0313 15:41:12.963429 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-52hkg\" (UniqueName: \"kubernetes.io/projected/499bc96f-8fa9-4900-bd92-2a507215791f-kube-api-access-52hkg\") pod \"cinder-db-create-pvpwm\" (UID: \"499bc96f-8fa9-4900-bd92-2a507215791f\") " pod="openstack/cinder-db-create-pvpwm" Mar 13 15:41:12 crc kubenswrapper[4907]: I0313 15:41:12.963478 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/499bc96f-8fa9-4900-bd92-2a507215791f-operator-scripts\") pod \"cinder-db-create-pvpwm\" (UID: \"499bc96f-8fa9-4900-bd92-2a507215791f\") " pod="openstack/cinder-db-create-pvpwm" Mar 13 15:41:12 crc kubenswrapper[4907]: I0313 15:41:12.964261 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/499bc96f-8fa9-4900-bd92-2a507215791f-operator-scripts\") pod \"cinder-db-create-pvpwm\" (UID: \"499bc96f-8fa9-4900-bd92-2a507215791f\") " pod="openstack/cinder-db-create-pvpwm" Mar 13 15:41:12 crc kubenswrapper[4907]: I0313 15:41:12.981253 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-52hkg\" (UniqueName: \"kubernetes.io/projected/499bc96f-8fa9-4900-bd92-2a507215791f-kube-api-access-52hkg\") pod \"cinder-db-create-pvpwm\" (UID: \"499bc96f-8fa9-4900-bd92-2a507215791f\") " pod="openstack/cinder-db-create-pvpwm" Mar 13 15:41:13 crc kubenswrapper[4907]: I0313 15:41:13.065340 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rvbzz\" (UniqueName: \"kubernetes.io/projected/0ab8c7fa-9565-4b6c-b8f8-033cfdf444db-kube-api-access-rvbzz\") pod \"cinder-9290-account-create-update-qws42\" (UID: \"0ab8c7fa-9565-4b6c-b8f8-033cfdf444db\") " pod="openstack/cinder-9290-account-create-update-qws42" Mar 13 15:41:13 crc kubenswrapper[4907]: I0313 15:41:13.065387 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0ab8c7fa-9565-4b6c-b8f8-033cfdf444db-operator-scripts\") pod \"cinder-9290-account-create-update-qws42\" (UID: \"0ab8c7fa-9565-4b6c-b8f8-033cfdf444db\") " pod="openstack/cinder-9290-account-create-update-qws42" Mar 13 15:41:13 crc kubenswrapper[4907]: I0313 15:41:13.150178 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-pvpwm" Mar 13 15:41:13 crc kubenswrapper[4907]: I0313 15:41:13.167252 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0ab8c7fa-9565-4b6c-b8f8-033cfdf444db-operator-scripts\") pod \"cinder-9290-account-create-update-qws42\" (UID: \"0ab8c7fa-9565-4b6c-b8f8-033cfdf444db\") " pod="openstack/cinder-9290-account-create-update-qws42" Mar 13 15:41:13 crc kubenswrapper[4907]: I0313 15:41:13.167423 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rvbzz\" (UniqueName: \"kubernetes.io/projected/0ab8c7fa-9565-4b6c-b8f8-033cfdf444db-kube-api-access-rvbzz\") pod \"cinder-9290-account-create-update-qws42\" (UID: \"0ab8c7fa-9565-4b6c-b8f8-033cfdf444db\") " pod="openstack/cinder-9290-account-create-update-qws42" Mar 13 15:41:13 crc kubenswrapper[4907]: I0313 15:41:13.168214 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0ab8c7fa-9565-4b6c-b8f8-033cfdf444db-operator-scripts\") pod \"cinder-9290-account-create-update-qws42\" (UID: \"0ab8c7fa-9565-4b6c-b8f8-033cfdf444db\") " pod="openstack/cinder-9290-account-create-update-qws42" Mar 13 15:41:13 crc kubenswrapper[4907]: I0313 15:41:13.184260 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rvbzz\" (UniqueName: \"kubernetes.io/projected/0ab8c7fa-9565-4b6c-b8f8-033cfdf444db-kube-api-access-rvbzz\") pod \"cinder-9290-account-create-update-qws42\" (UID: \"0ab8c7fa-9565-4b6c-b8f8-033cfdf444db\") " pod="openstack/cinder-9290-account-create-update-qws42" Mar 13 15:41:13 crc kubenswrapper[4907]: I0313 15:41:13.263376 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-9290-account-create-update-qws42" Mar 13 15:41:13 crc kubenswrapper[4907]: I0313 15:41:13.600496 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-pvpwm"] Mar 13 15:41:13 crc kubenswrapper[4907]: I0313 15:41:13.709761 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-9290-account-create-update-qws42"] Mar 13 15:41:13 crc kubenswrapper[4907]: W0313 15:41:13.719411 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0ab8c7fa_9565_4b6c_b8f8_033cfdf444db.slice/crio-80540ba04d58c62659d9d51465d46822d741fb9ff2667a93002e21e7b5205881 WatchSource:0}: Error finding container 80540ba04d58c62659d9d51465d46822d741fb9ff2667a93002e21e7b5205881: Status 404 returned error can't find the container with id 80540ba04d58c62659d9d51465d46822d741fb9ff2667a93002e21e7b5205881 Mar 13 15:41:13 crc kubenswrapper[4907]: I0313 15:41:13.800676 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="690a67d3-7115-4897-ba26-edd5eeff0521" path="/var/lib/kubelet/pods/690a67d3-7115-4897-ba26-edd5eeff0521/volumes" Mar 13 15:41:14 crc kubenswrapper[4907]: I0313 15:41:14.471946 4907 generic.go:334] "Generic (PLEG): container finished" podID="499bc96f-8fa9-4900-bd92-2a507215791f" containerID="5140f9e51d2f4e22927da88ea13b8b76a7a069356e41a43262ca41c166a3779c" exitCode=0 Mar 13 15:41:14 crc kubenswrapper[4907]: I0313 15:41:14.473165 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-pvpwm" event={"ID":"499bc96f-8fa9-4900-bd92-2a507215791f","Type":"ContainerDied","Data":"5140f9e51d2f4e22927da88ea13b8b76a7a069356e41a43262ca41c166a3779c"} Mar 13 15:41:14 crc kubenswrapper[4907]: I0313 15:41:14.473276 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-pvpwm" event={"ID":"499bc96f-8fa9-4900-bd92-2a507215791f","Type":"ContainerStarted","Data":"4f157b815232d30b0784ee5079591971c6c531c2992a38589627a781beae23d9"} Mar 13 15:41:14 crc kubenswrapper[4907]: I0313 15:41:14.476256 4907 generic.go:334] "Generic (PLEG): container finished" podID="0ab8c7fa-9565-4b6c-b8f8-033cfdf444db" containerID="0559fde0f7d2daa5d36de2042211ef456904ceea0804232b86a43826c351c91c" exitCode=0 Mar 13 15:41:14 crc kubenswrapper[4907]: I0313 15:41:14.476265 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-9290-account-create-update-qws42" event={"ID":"0ab8c7fa-9565-4b6c-b8f8-033cfdf444db","Type":"ContainerDied","Data":"0559fde0f7d2daa5d36de2042211ef456904ceea0804232b86a43826c351c91c"} Mar 13 15:41:14 crc kubenswrapper[4907]: I0313 15:41:14.476338 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-9290-account-create-update-qws42" event={"ID":"0ab8c7fa-9565-4b6c-b8f8-033cfdf444db","Type":"ContainerStarted","Data":"80540ba04d58c62659d9d51465d46822d741fb9ff2667a93002e21e7b5205881"} Mar 13 15:41:15 crc kubenswrapper[4907]: I0313 15:41:15.956105 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-9290-account-create-update-qws42" Mar 13 15:41:15 crc kubenswrapper[4907]: I0313 15:41:15.966020 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-pvpwm" Mar 13 15:41:16 crc kubenswrapper[4907]: I0313 15:41:16.019584 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-52hkg\" (UniqueName: \"kubernetes.io/projected/499bc96f-8fa9-4900-bd92-2a507215791f-kube-api-access-52hkg\") pod \"499bc96f-8fa9-4900-bd92-2a507215791f\" (UID: \"499bc96f-8fa9-4900-bd92-2a507215791f\") " Mar 13 15:41:16 crc kubenswrapper[4907]: I0313 15:41:16.019644 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0ab8c7fa-9565-4b6c-b8f8-033cfdf444db-operator-scripts\") pod \"0ab8c7fa-9565-4b6c-b8f8-033cfdf444db\" (UID: \"0ab8c7fa-9565-4b6c-b8f8-033cfdf444db\") " Mar 13 15:41:16 crc kubenswrapper[4907]: I0313 15:41:16.019673 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rvbzz\" (UniqueName: \"kubernetes.io/projected/0ab8c7fa-9565-4b6c-b8f8-033cfdf444db-kube-api-access-rvbzz\") pod \"0ab8c7fa-9565-4b6c-b8f8-033cfdf444db\" (UID: \"0ab8c7fa-9565-4b6c-b8f8-033cfdf444db\") " Mar 13 15:41:16 crc kubenswrapper[4907]: I0313 15:41:16.019778 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/499bc96f-8fa9-4900-bd92-2a507215791f-operator-scripts\") pod \"499bc96f-8fa9-4900-bd92-2a507215791f\" (UID: \"499bc96f-8fa9-4900-bd92-2a507215791f\") " Mar 13 15:41:16 crc kubenswrapper[4907]: I0313 15:41:16.021034 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/499bc96f-8fa9-4900-bd92-2a507215791f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "499bc96f-8fa9-4900-bd92-2a507215791f" (UID: "499bc96f-8fa9-4900-bd92-2a507215791f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:41:16 crc kubenswrapper[4907]: I0313 15:41:16.021130 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0ab8c7fa-9565-4b6c-b8f8-033cfdf444db-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "0ab8c7fa-9565-4b6c-b8f8-033cfdf444db" (UID: "0ab8c7fa-9565-4b6c-b8f8-033cfdf444db"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:41:16 crc kubenswrapper[4907]: I0313 15:41:16.027303 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0ab8c7fa-9565-4b6c-b8f8-033cfdf444db-kube-api-access-rvbzz" (OuterVolumeSpecName: "kube-api-access-rvbzz") pod "0ab8c7fa-9565-4b6c-b8f8-033cfdf444db" (UID: "0ab8c7fa-9565-4b6c-b8f8-033cfdf444db"). InnerVolumeSpecName "kube-api-access-rvbzz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:41:16 crc kubenswrapper[4907]: I0313 15:41:16.027659 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/499bc96f-8fa9-4900-bd92-2a507215791f-kube-api-access-52hkg" (OuterVolumeSpecName: "kube-api-access-52hkg") pod "499bc96f-8fa9-4900-bd92-2a507215791f" (UID: "499bc96f-8fa9-4900-bd92-2a507215791f"). InnerVolumeSpecName "kube-api-access-52hkg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:41:16 crc kubenswrapper[4907]: I0313 15:41:16.121659 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-52hkg\" (UniqueName: \"kubernetes.io/projected/499bc96f-8fa9-4900-bd92-2a507215791f-kube-api-access-52hkg\") on node \"crc\" DevicePath \"\"" Mar 13 15:41:16 crc kubenswrapper[4907]: I0313 15:41:16.121698 4907 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0ab8c7fa-9565-4b6c-b8f8-033cfdf444db-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 15:41:16 crc kubenswrapper[4907]: I0313 15:41:16.121707 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rvbzz\" (UniqueName: \"kubernetes.io/projected/0ab8c7fa-9565-4b6c-b8f8-033cfdf444db-kube-api-access-rvbzz\") on node \"crc\" DevicePath \"\"" Mar 13 15:41:16 crc kubenswrapper[4907]: I0313 15:41:16.121715 4907 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/499bc96f-8fa9-4900-bd92-2a507215791f-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 15:41:16 crc kubenswrapper[4907]: I0313 15:41:16.496137 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-pvpwm" event={"ID":"499bc96f-8fa9-4900-bd92-2a507215791f","Type":"ContainerDied","Data":"4f157b815232d30b0784ee5079591971c6c531c2992a38589627a781beae23d9"} Mar 13 15:41:16 crc kubenswrapper[4907]: I0313 15:41:16.496478 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4f157b815232d30b0784ee5079591971c6c531c2992a38589627a781beae23d9" Mar 13 15:41:16 crc kubenswrapper[4907]: I0313 15:41:16.496191 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-pvpwm" Mar 13 15:41:16 crc kubenswrapper[4907]: I0313 15:41:16.497718 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-9290-account-create-update-qws42" event={"ID":"0ab8c7fa-9565-4b6c-b8f8-033cfdf444db","Type":"ContainerDied","Data":"80540ba04d58c62659d9d51465d46822d741fb9ff2667a93002e21e7b5205881"} Mar 13 15:41:16 crc kubenswrapper[4907]: I0313 15:41:16.497746 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="80540ba04d58c62659d9d51465d46822d741fb9ff2667a93002e21e7b5205881" Mar 13 15:41:16 crc kubenswrapper[4907]: I0313 15:41:16.497797 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-9290-account-create-update-qws42" Mar 13 15:41:18 crc kubenswrapper[4907]: I0313 15:41:18.041336 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 15:41:18 crc kubenswrapper[4907]: I0313 15:41:18.041402 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 15:41:18 crc kubenswrapper[4907]: I0313 15:41:18.041457 4907 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" Mar 13 15:41:18 crc kubenswrapper[4907]: I0313 15:41:18.042265 4907 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"68ffe92dd03078633ec89d7ef29da9bbdc4c5f04c956a1e8487e3f2e01c5d1c5"} pod="openshift-machine-config-operator/machine-config-daemon-hm56j" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 15:41:18 crc kubenswrapper[4907]: I0313 15:41:18.042320 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" containerID="cri-o://68ffe92dd03078633ec89d7ef29da9bbdc4c5f04c956a1e8487e3f2e01c5d1c5" gracePeriod=600 Mar 13 15:41:18 crc kubenswrapper[4907]: E0313 15:41:18.190757 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:41:18 crc kubenswrapper[4907]: I0313 15:41:18.197980 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-hv2mf"] Mar 13 15:41:18 crc kubenswrapper[4907]: E0313 15:41:18.198777 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="499bc96f-8fa9-4900-bd92-2a507215791f" containerName="mariadb-database-create" Mar 13 15:41:18 crc kubenswrapper[4907]: I0313 15:41:18.198807 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="499bc96f-8fa9-4900-bd92-2a507215791f" containerName="mariadb-database-create" Mar 13 15:41:18 crc kubenswrapper[4907]: E0313 15:41:18.198851 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ab8c7fa-9565-4b6c-b8f8-033cfdf444db" containerName="mariadb-account-create-update" Mar 13 15:41:18 crc kubenswrapper[4907]: I0313 15:41:18.198861 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ab8c7fa-9565-4b6c-b8f8-033cfdf444db" containerName="mariadb-account-create-update" Mar 13 15:41:18 crc kubenswrapper[4907]: I0313 15:41:18.201102 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="499bc96f-8fa9-4900-bd92-2a507215791f" containerName="mariadb-database-create" Mar 13 15:41:18 crc kubenswrapper[4907]: I0313 15:41:18.201173 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="0ab8c7fa-9565-4b6c-b8f8-033cfdf444db" containerName="mariadb-account-create-update" Mar 13 15:41:18 crc kubenswrapper[4907]: I0313 15:41:18.204790 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-hv2mf" Mar 13 15:41:18 crc kubenswrapper[4907]: I0313 15:41:18.209148 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-v5jht" Mar 13 15:41:18 crc kubenswrapper[4907]: I0313 15:41:18.211000 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Mar 13 15:41:18 crc kubenswrapper[4907]: I0313 15:41:18.213914 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Mar 13 15:41:18 crc kubenswrapper[4907]: I0313 15:41:18.219607 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-hv2mf"] Mar 13 15:41:18 crc kubenswrapper[4907]: I0313 15:41:18.260555 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ef105bad-6763-4234-b52b-6d2820d48b02-db-sync-config-data\") pod \"cinder-db-sync-hv2mf\" (UID: \"ef105bad-6763-4234-b52b-6d2820d48b02\") " pod="openstack/cinder-db-sync-hv2mf" Mar 13 15:41:18 crc kubenswrapper[4907]: I0313 15:41:18.260595 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef105bad-6763-4234-b52b-6d2820d48b02-combined-ca-bundle\") pod \"cinder-db-sync-hv2mf\" (UID: \"ef105bad-6763-4234-b52b-6d2820d48b02\") " pod="openstack/cinder-db-sync-hv2mf" Mar 13 15:41:18 crc kubenswrapper[4907]: I0313 15:41:18.260657 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ef105bad-6763-4234-b52b-6d2820d48b02-scripts\") pod \"cinder-db-sync-hv2mf\" (UID: \"ef105bad-6763-4234-b52b-6d2820d48b02\") " pod="openstack/cinder-db-sync-hv2mf" Mar 13 15:41:18 crc kubenswrapper[4907]: I0313 15:41:18.260787 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef105bad-6763-4234-b52b-6d2820d48b02-config-data\") pod \"cinder-db-sync-hv2mf\" (UID: \"ef105bad-6763-4234-b52b-6d2820d48b02\") " pod="openstack/cinder-db-sync-hv2mf" Mar 13 15:41:18 crc kubenswrapper[4907]: I0313 15:41:18.260991 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-skdq4\" (UniqueName: \"kubernetes.io/projected/ef105bad-6763-4234-b52b-6d2820d48b02-kube-api-access-skdq4\") pod \"cinder-db-sync-hv2mf\" (UID: \"ef105bad-6763-4234-b52b-6d2820d48b02\") " pod="openstack/cinder-db-sync-hv2mf" Mar 13 15:41:18 crc kubenswrapper[4907]: I0313 15:41:18.261199 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ef105bad-6763-4234-b52b-6d2820d48b02-etc-machine-id\") pod \"cinder-db-sync-hv2mf\" (UID: \"ef105bad-6763-4234-b52b-6d2820d48b02\") " pod="openstack/cinder-db-sync-hv2mf" Mar 13 15:41:18 crc kubenswrapper[4907]: I0313 15:41:18.363472 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-skdq4\" (UniqueName: \"kubernetes.io/projected/ef105bad-6763-4234-b52b-6d2820d48b02-kube-api-access-skdq4\") pod \"cinder-db-sync-hv2mf\" (UID: \"ef105bad-6763-4234-b52b-6d2820d48b02\") " pod="openstack/cinder-db-sync-hv2mf" Mar 13 15:41:18 crc kubenswrapper[4907]: I0313 15:41:18.363549 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ef105bad-6763-4234-b52b-6d2820d48b02-etc-machine-id\") pod \"cinder-db-sync-hv2mf\" (UID: \"ef105bad-6763-4234-b52b-6d2820d48b02\") " pod="openstack/cinder-db-sync-hv2mf" Mar 13 15:41:18 crc kubenswrapper[4907]: I0313 15:41:18.363616 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ef105bad-6763-4234-b52b-6d2820d48b02-db-sync-config-data\") pod \"cinder-db-sync-hv2mf\" (UID: \"ef105bad-6763-4234-b52b-6d2820d48b02\") " pod="openstack/cinder-db-sync-hv2mf" Mar 13 15:41:18 crc kubenswrapper[4907]: I0313 15:41:18.363633 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef105bad-6763-4234-b52b-6d2820d48b02-combined-ca-bundle\") pod \"cinder-db-sync-hv2mf\" (UID: \"ef105bad-6763-4234-b52b-6d2820d48b02\") " pod="openstack/cinder-db-sync-hv2mf" Mar 13 15:41:18 crc kubenswrapper[4907]: I0313 15:41:18.363651 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ef105bad-6763-4234-b52b-6d2820d48b02-scripts\") pod \"cinder-db-sync-hv2mf\" (UID: \"ef105bad-6763-4234-b52b-6d2820d48b02\") " pod="openstack/cinder-db-sync-hv2mf" Mar 13 15:41:18 crc kubenswrapper[4907]: I0313 15:41:18.363688 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef105bad-6763-4234-b52b-6d2820d48b02-config-data\") pod \"cinder-db-sync-hv2mf\" (UID: \"ef105bad-6763-4234-b52b-6d2820d48b02\") " pod="openstack/cinder-db-sync-hv2mf" Mar 13 15:41:18 crc kubenswrapper[4907]: I0313 15:41:18.363710 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ef105bad-6763-4234-b52b-6d2820d48b02-etc-machine-id\") pod \"cinder-db-sync-hv2mf\" (UID: \"ef105bad-6763-4234-b52b-6d2820d48b02\") " pod="openstack/cinder-db-sync-hv2mf" Mar 13 15:41:18 crc kubenswrapper[4907]: I0313 15:41:18.368501 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ef105bad-6763-4234-b52b-6d2820d48b02-db-sync-config-data\") pod \"cinder-db-sync-hv2mf\" (UID: \"ef105bad-6763-4234-b52b-6d2820d48b02\") " pod="openstack/cinder-db-sync-hv2mf" Mar 13 15:41:18 crc kubenswrapper[4907]: I0313 15:41:18.368737 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef105bad-6763-4234-b52b-6d2820d48b02-combined-ca-bundle\") pod \"cinder-db-sync-hv2mf\" (UID: \"ef105bad-6763-4234-b52b-6d2820d48b02\") " pod="openstack/cinder-db-sync-hv2mf" Mar 13 15:41:18 crc kubenswrapper[4907]: I0313 15:41:18.369214 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef105bad-6763-4234-b52b-6d2820d48b02-config-data\") pod \"cinder-db-sync-hv2mf\" (UID: \"ef105bad-6763-4234-b52b-6d2820d48b02\") " pod="openstack/cinder-db-sync-hv2mf" Mar 13 15:41:18 crc kubenswrapper[4907]: I0313 15:41:18.375429 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ef105bad-6763-4234-b52b-6d2820d48b02-scripts\") pod \"cinder-db-sync-hv2mf\" (UID: \"ef105bad-6763-4234-b52b-6d2820d48b02\") " pod="openstack/cinder-db-sync-hv2mf" Mar 13 15:41:18 crc kubenswrapper[4907]: I0313 15:41:18.380506 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-skdq4\" (UniqueName: \"kubernetes.io/projected/ef105bad-6763-4234-b52b-6d2820d48b02-kube-api-access-skdq4\") pod \"cinder-db-sync-hv2mf\" (UID: \"ef105bad-6763-4234-b52b-6d2820d48b02\") " pod="openstack/cinder-db-sync-hv2mf" Mar 13 15:41:18 crc kubenswrapper[4907]: I0313 15:41:18.515021 4907 generic.go:334] "Generic (PLEG): container finished" podID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerID="68ffe92dd03078633ec89d7ef29da9bbdc4c5f04c956a1e8487e3f2e01c5d1c5" exitCode=0 Mar 13 15:41:18 crc kubenswrapper[4907]: I0313 15:41:18.515089 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerDied","Data":"68ffe92dd03078633ec89d7ef29da9bbdc4c5f04c956a1e8487e3f2e01c5d1c5"} Mar 13 15:41:18 crc kubenswrapper[4907]: I0313 15:41:18.515140 4907 scope.go:117] "RemoveContainer" containerID="90a39cf81ab6b6b036397a9c751a635e5f4cd834dccdc53b2ff02dd0144698a4" Mar 13 15:41:18 crc kubenswrapper[4907]: I0313 15:41:18.516178 4907 scope.go:117] "RemoveContainer" containerID="68ffe92dd03078633ec89d7ef29da9bbdc4c5f04c956a1e8487e3f2e01c5d1c5" Mar 13 15:41:18 crc kubenswrapper[4907]: E0313 15:41:18.516490 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:41:18 crc kubenswrapper[4907]: I0313 15:41:18.530870 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-hv2mf" Mar 13 15:41:19 crc kubenswrapper[4907]: I0313 15:41:19.038356 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-hv2mf"] Mar 13 15:41:19 crc kubenswrapper[4907]: I0313 15:41:19.523846 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-hv2mf" event={"ID":"ef105bad-6763-4234-b52b-6d2820d48b02","Type":"ContainerStarted","Data":"d1248fb6962eeda0ac2a1465b0501825fb69861c03d06ab1f3e3c6b244ab9af9"} Mar 13 15:41:20 crc kubenswrapper[4907]: I0313 15:41:20.533954 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-hv2mf" event={"ID":"ef105bad-6763-4234-b52b-6d2820d48b02","Type":"ContainerStarted","Data":"cfb2b66bd790bad49fda00e0dee115de187df8f601026658c19c9100dc1ee877"} Mar 13 15:41:20 crc kubenswrapper[4907]: I0313 15:41:20.564741 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-hv2mf" podStartSLOduration=2.564716546 podStartE2EDuration="2.564716546s" podCreationTimestamp="2026-03-13 15:41:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:41:20.553754687 +0000 UTC m=+5779.453542376" watchObservedRunningTime="2026-03-13 15:41:20.564716546 +0000 UTC m=+5779.464504265" Mar 13 15:41:22 crc kubenswrapper[4907]: I0313 15:41:22.552053 4907 generic.go:334] "Generic (PLEG): container finished" podID="ef105bad-6763-4234-b52b-6d2820d48b02" containerID="cfb2b66bd790bad49fda00e0dee115de187df8f601026658c19c9100dc1ee877" exitCode=0 Mar 13 15:41:22 crc kubenswrapper[4907]: I0313 15:41:22.552374 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-hv2mf" event={"ID":"ef105bad-6763-4234-b52b-6d2820d48b02","Type":"ContainerDied","Data":"cfb2b66bd790bad49fda00e0dee115de187df8f601026658c19c9100dc1ee877"} Mar 13 15:41:23 crc kubenswrapper[4907]: I0313 15:41:23.927413 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-hv2mf" Mar 13 15:41:23 crc kubenswrapper[4907]: I0313 15:41:23.974310 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ef105bad-6763-4234-b52b-6d2820d48b02-scripts\") pod \"ef105bad-6763-4234-b52b-6d2820d48b02\" (UID: \"ef105bad-6763-4234-b52b-6d2820d48b02\") " Mar 13 15:41:23 crc kubenswrapper[4907]: I0313 15:41:23.974463 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef105bad-6763-4234-b52b-6d2820d48b02-config-data\") pod \"ef105bad-6763-4234-b52b-6d2820d48b02\" (UID: \"ef105bad-6763-4234-b52b-6d2820d48b02\") " Mar 13 15:41:23 crc kubenswrapper[4907]: I0313 15:41:23.974530 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-skdq4\" (UniqueName: \"kubernetes.io/projected/ef105bad-6763-4234-b52b-6d2820d48b02-kube-api-access-skdq4\") pod \"ef105bad-6763-4234-b52b-6d2820d48b02\" (UID: \"ef105bad-6763-4234-b52b-6d2820d48b02\") " Mar 13 15:41:23 crc kubenswrapper[4907]: I0313 15:41:23.974581 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ef105bad-6763-4234-b52b-6d2820d48b02-db-sync-config-data\") pod \"ef105bad-6763-4234-b52b-6d2820d48b02\" (UID: \"ef105bad-6763-4234-b52b-6d2820d48b02\") " Mar 13 15:41:23 crc kubenswrapper[4907]: I0313 15:41:23.974640 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef105bad-6763-4234-b52b-6d2820d48b02-combined-ca-bundle\") pod \"ef105bad-6763-4234-b52b-6d2820d48b02\" (UID: \"ef105bad-6763-4234-b52b-6d2820d48b02\") " Mar 13 15:41:23 crc kubenswrapper[4907]: I0313 15:41:23.974667 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ef105bad-6763-4234-b52b-6d2820d48b02-etc-machine-id\") pod \"ef105bad-6763-4234-b52b-6d2820d48b02\" (UID: \"ef105bad-6763-4234-b52b-6d2820d48b02\") " Mar 13 15:41:23 crc kubenswrapper[4907]: I0313 15:41:23.975026 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ef105bad-6763-4234-b52b-6d2820d48b02-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "ef105bad-6763-4234-b52b-6d2820d48b02" (UID: "ef105bad-6763-4234-b52b-6d2820d48b02"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 15:41:23 crc kubenswrapper[4907]: I0313 15:41:23.980290 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef105bad-6763-4234-b52b-6d2820d48b02-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "ef105bad-6763-4234-b52b-6d2820d48b02" (UID: "ef105bad-6763-4234-b52b-6d2820d48b02"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:41:23 crc kubenswrapper[4907]: I0313 15:41:23.981794 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef105bad-6763-4234-b52b-6d2820d48b02-kube-api-access-skdq4" (OuterVolumeSpecName: "kube-api-access-skdq4") pod "ef105bad-6763-4234-b52b-6d2820d48b02" (UID: "ef105bad-6763-4234-b52b-6d2820d48b02"). InnerVolumeSpecName "kube-api-access-skdq4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:41:23 crc kubenswrapper[4907]: I0313 15:41:23.993436 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef105bad-6763-4234-b52b-6d2820d48b02-scripts" (OuterVolumeSpecName: "scripts") pod "ef105bad-6763-4234-b52b-6d2820d48b02" (UID: "ef105bad-6763-4234-b52b-6d2820d48b02"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:41:24 crc kubenswrapper[4907]: I0313 15:41:24.006234 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef105bad-6763-4234-b52b-6d2820d48b02-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ef105bad-6763-4234-b52b-6d2820d48b02" (UID: "ef105bad-6763-4234-b52b-6d2820d48b02"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:41:24 crc kubenswrapper[4907]: I0313 15:41:24.037229 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef105bad-6763-4234-b52b-6d2820d48b02-config-data" (OuterVolumeSpecName: "config-data") pod "ef105bad-6763-4234-b52b-6d2820d48b02" (UID: "ef105bad-6763-4234-b52b-6d2820d48b02"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:41:24 crc kubenswrapper[4907]: I0313 15:41:24.075854 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef105bad-6763-4234-b52b-6d2820d48b02-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 15:41:24 crc kubenswrapper[4907]: I0313 15:41:24.075909 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-skdq4\" (UniqueName: \"kubernetes.io/projected/ef105bad-6763-4234-b52b-6d2820d48b02-kube-api-access-skdq4\") on node \"crc\" DevicePath \"\"" Mar 13 15:41:24 crc kubenswrapper[4907]: I0313 15:41:24.075924 4907 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ef105bad-6763-4234-b52b-6d2820d48b02-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 15:41:24 crc kubenswrapper[4907]: I0313 15:41:24.075936 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef105bad-6763-4234-b52b-6d2820d48b02-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 15:41:24 crc kubenswrapper[4907]: I0313 15:41:24.075948 4907 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ef105bad-6763-4234-b52b-6d2820d48b02-etc-machine-id\") on node \"crc\" DevicePath \"\"" Mar 13 15:41:24 crc kubenswrapper[4907]: I0313 15:41:24.075958 4907 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ef105bad-6763-4234-b52b-6d2820d48b02-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 15:41:24 crc kubenswrapper[4907]: I0313 15:41:24.572404 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-hv2mf" event={"ID":"ef105bad-6763-4234-b52b-6d2820d48b02","Type":"ContainerDied","Data":"d1248fb6962eeda0ac2a1465b0501825fb69861c03d06ab1f3e3c6b244ab9af9"} Mar 13 15:41:24 crc kubenswrapper[4907]: I0313 15:41:24.572444 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d1248fb6962eeda0ac2a1465b0501825fb69861c03d06ab1f3e3c6b244ab9af9" Mar 13 15:41:24 crc kubenswrapper[4907]: I0313 15:41:24.572453 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-hv2mf" Mar 13 15:41:24 crc kubenswrapper[4907]: I0313 15:41:24.923230 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6d59c8ff75-hm7rf"] Mar 13 15:41:24 crc kubenswrapper[4907]: E0313 15:41:24.924101 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef105bad-6763-4234-b52b-6d2820d48b02" containerName="cinder-db-sync" Mar 13 15:41:24 crc kubenswrapper[4907]: I0313 15:41:24.924125 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef105bad-6763-4234-b52b-6d2820d48b02" containerName="cinder-db-sync" Mar 13 15:41:24 crc kubenswrapper[4907]: I0313 15:41:24.924413 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef105bad-6763-4234-b52b-6d2820d48b02" containerName="cinder-db-sync" Mar 13 15:41:24 crc kubenswrapper[4907]: I0313 15:41:24.925668 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d59c8ff75-hm7rf" Mar 13 15:41:24 crc kubenswrapper[4907]: I0313 15:41:24.933743 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6d59c8ff75-hm7rf"] Mar 13 15:41:25 crc kubenswrapper[4907]: I0313 15:41:25.106254 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bb294da7-acf1-4984-a401-bbfc15ee97b7-dns-svc\") pod \"dnsmasq-dns-6d59c8ff75-hm7rf\" (UID: \"bb294da7-acf1-4984-a401-bbfc15ee97b7\") " pod="openstack/dnsmasq-dns-6d59c8ff75-hm7rf" Mar 13 15:41:25 crc kubenswrapper[4907]: I0313 15:41:25.106777 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bb294da7-acf1-4984-a401-bbfc15ee97b7-ovsdbserver-sb\") pod \"dnsmasq-dns-6d59c8ff75-hm7rf\" (UID: \"bb294da7-acf1-4984-a401-bbfc15ee97b7\") " pod="openstack/dnsmasq-dns-6d59c8ff75-hm7rf" Mar 13 15:41:25 crc kubenswrapper[4907]: I0313 15:41:25.106979 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bb294da7-acf1-4984-a401-bbfc15ee97b7-config\") pod \"dnsmasq-dns-6d59c8ff75-hm7rf\" (UID: \"bb294da7-acf1-4984-a401-bbfc15ee97b7\") " pod="openstack/dnsmasq-dns-6d59c8ff75-hm7rf" Mar 13 15:41:25 crc kubenswrapper[4907]: I0313 15:41:25.107100 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jrkzt\" (UniqueName: \"kubernetes.io/projected/bb294da7-acf1-4984-a401-bbfc15ee97b7-kube-api-access-jrkzt\") pod \"dnsmasq-dns-6d59c8ff75-hm7rf\" (UID: \"bb294da7-acf1-4984-a401-bbfc15ee97b7\") " pod="openstack/dnsmasq-dns-6d59c8ff75-hm7rf" Mar 13 15:41:25 crc kubenswrapper[4907]: I0313 15:41:25.107208 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bb294da7-acf1-4984-a401-bbfc15ee97b7-ovsdbserver-nb\") pod \"dnsmasq-dns-6d59c8ff75-hm7rf\" (UID: \"bb294da7-acf1-4984-a401-bbfc15ee97b7\") " pod="openstack/dnsmasq-dns-6d59c8ff75-hm7rf" Mar 13 15:41:25 crc kubenswrapper[4907]: I0313 15:41:25.183582 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Mar 13 15:41:25 crc kubenswrapper[4907]: I0313 15:41:25.185242 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Mar 13 15:41:25 crc kubenswrapper[4907]: I0313 15:41:25.189712 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Mar 13 15:41:25 crc kubenswrapper[4907]: I0313 15:41:25.190155 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Mar 13 15:41:25 crc kubenswrapper[4907]: I0313 15:41:25.190320 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Mar 13 15:41:25 crc kubenswrapper[4907]: I0313 15:41:25.191033 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-v5jht" Mar 13 15:41:25 crc kubenswrapper[4907]: I0313 15:41:25.209126 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bb294da7-acf1-4984-a401-bbfc15ee97b7-ovsdbserver-sb\") pod \"dnsmasq-dns-6d59c8ff75-hm7rf\" (UID: \"bb294da7-acf1-4984-a401-bbfc15ee97b7\") " pod="openstack/dnsmasq-dns-6d59c8ff75-hm7rf" Mar 13 15:41:25 crc kubenswrapper[4907]: I0313 15:41:25.210675 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bb294da7-acf1-4984-a401-bbfc15ee97b7-ovsdbserver-sb\") pod \"dnsmasq-dns-6d59c8ff75-hm7rf\" (UID: \"bb294da7-acf1-4984-a401-bbfc15ee97b7\") " pod="openstack/dnsmasq-dns-6d59c8ff75-hm7rf" Mar 13 15:41:25 crc kubenswrapper[4907]: I0313 15:41:25.216409 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bb294da7-acf1-4984-a401-bbfc15ee97b7-config\") pod \"dnsmasq-dns-6d59c8ff75-hm7rf\" (UID: \"bb294da7-acf1-4984-a401-bbfc15ee97b7\") " pod="openstack/dnsmasq-dns-6d59c8ff75-hm7rf" Mar 13 15:41:25 crc kubenswrapper[4907]: I0313 15:41:25.216543 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jrkzt\" (UniqueName: \"kubernetes.io/projected/bb294da7-acf1-4984-a401-bbfc15ee97b7-kube-api-access-jrkzt\") pod \"dnsmasq-dns-6d59c8ff75-hm7rf\" (UID: \"bb294da7-acf1-4984-a401-bbfc15ee97b7\") " pod="openstack/dnsmasq-dns-6d59c8ff75-hm7rf" Mar 13 15:41:25 crc kubenswrapper[4907]: I0313 15:41:25.216739 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bb294da7-acf1-4984-a401-bbfc15ee97b7-ovsdbserver-nb\") pod \"dnsmasq-dns-6d59c8ff75-hm7rf\" (UID: \"bb294da7-acf1-4984-a401-bbfc15ee97b7\") " pod="openstack/dnsmasq-dns-6d59c8ff75-hm7rf" Mar 13 15:41:25 crc kubenswrapper[4907]: I0313 15:41:25.217295 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bb294da7-acf1-4984-a401-bbfc15ee97b7-config\") pod \"dnsmasq-dns-6d59c8ff75-hm7rf\" (UID: \"bb294da7-acf1-4984-a401-bbfc15ee97b7\") " pod="openstack/dnsmasq-dns-6d59c8ff75-hm7rf" Mar 13 15:41:25 crc kubenswrapper[4907]: I0313 15:41:25.217511 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bb294da7-acf1-4984-a401-bbfc15ee97b7-ovsdbserver-nb\") pod \"dnsmasq-dns-6d59c8ff75-hm7rf\" (UID: \"bb294da7-acf1-4984-a401-bbfc15ee97b7\") " pod="openstack/dnsmasq-dns-6d59c8ff75-hm7rf" Mar 13 15:41:25 crc kubenswrapper[4907]: I0313 15:41:25.217954 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bb294da7-acf1-4984-a401-bbfc15ee97b7-dns-svc\") pod \"dnsmasq-dns-6d59c8ff75-hm7rf\" (UID: \"bb294da7-acf1-4984-a401-bbfc15ee97b7\") " pod="openstack/dnsmasq-dns-6d59c8ff75-hm7rf" Mar 13 15:41:25 crc kubenswrapper[4907]: I0313 15:41:25.218568 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bb294da7-acf1-4984-a401-bbfc15ee97b7-dns-svc\") pod \"dnsmasq-dns-6d59c8ff75-hm7rf\" (UID: \"bb294da7-acf1-4984-a401-bbfc15ee97b7\") " pod="openstack/dnsmasq-dns-6d59c8ff75-hm7rf" Mar 13 15:41:25 crc kubenswrapper[4907]: I0313 15:41:25.221408 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Mar 13 15:41:25 crc kubenswrapper[4907]: I0313 15:41:25.249373 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jrkzt\" (UniqueName: \"kubernetes.io/projected/bb294da7-acf1-4984-a401-bbfc15ee97b7-kube-api-access-jrkzt\") pod \"dnsmasq-dns-6d59c8ff75-hm7rf\" (UID: \"bb294da7-acf1-4984-a401-bbfc15ee97b7\") " pod="openstack/dnsmasq-dns-6d59c8ff75-hm7rf" Mar 13 15:41:25 crc kubenswrapper[4907]: I0313 15:41:25.261499 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d59c8ff75-hm7rf" Mar 13 15:41:25 crc kubenswrapper[4907]: I0313 15:41:25.320390 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12616424-013a-47c8-9ed3-b407791537fe-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"12616424-013a-47c8-9ed3-b407791537fe\") " pod="openstack/cinder-api-0" Mar 13 15:41:25 crc kubenswrapper[4907]: I0313 15:41:25.320802 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/12616424-013a-47c8-9ed3-b407791537fe-etc-machine-id\") pod \"cinder-api-0\" (UID: \"12616424-013a-47c8-9ed3-b407791537fe\") " pod="openstack/cinder-api-0" Mar 13 15:41:25 crc kubenswrapper[4907]: I0313 15:41:25.320978 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/12616424-013a-47c8-9ed3-b407791537fe-logs\") pod \"cinder-api-0\" (UID: \"12616424-013a-47c8-9ed3-b407791537fe\") " pod="openstack/cinder-api-0" Mar 13 15:41:25 crc kubenswrapper[4907]: I0313 15:41:25.321199 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12616424-013a-47c8-9ed3-b407791537fe-config-data\") pod \"cinder-api-0\" (UID: \"12616424-013a-47c8-9ed3-b407791537fe\") " pod="openstack/cinder-api-0" Mar 13 15:41:25 crc kubenswrapper[4907]: I0313 15:41:25.321362 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/12616424-013a-47c8-9ed3-b407791537fe-config-data-custom\") pod \"cinder-api-0\" (UID: \"12616424-013a-47c8-9ed3-b407791537fe\") " pod="openstack/cinder-api-0" Mar 13 15:41:25 crc kubenswrapper[4907]: I0313 15:41:25.321536 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dh7tk\" (UniqueName: \"kubernetes.io/projected/12616424-013a-47c8-9ed3-b407791537fe-kube-api-access-dh7tk\") pod \"cinder-api-0\" (UID: \"12616424-013a-47c8-9ed3-b407791537fe\") " pod="openstack/cinder-api-0" Mar 13 15:41:25 crc kubenswrapper[4907]: I0313 15:41:25.321699 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/12616424-013a-47c8-9ed3-b407791537fe-scripts\") pod \"cinder-api-0\" (UID: \"12616424-013a-47c8-9ed3-b407791537fe\") " pod="openstack/cinder-api-0" Mar 13 15:41:25 crc kubenswrapper[4907]: I0313 15:41:25.423207 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/12616424-013a-47c8-9ed3-b407791537fe-config-data-custom\") pod \"cinder-api-0\" (UID: \"12616424-013a-47c8-9ed3-b407791537fe\") " pod="openstack/cinder-api-0" Mar 13 15:41:25 crc kubenswrapper[4907]: I0313 15:41:25.423267 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dh7tk\" (UniqueName: \"kubernetes.io/projected/12616424-013a-47c8-9ed3-b407791537fe-kube-api-access-dh7tk\") pod \"cinder-api-0\" (UID: \"12616424-013a-47c8-9ed3-b407791537fe\") " pod="openstack/cinder-api-0" Mar 13 15:41:25 crc kubenswrapper[4907]: I0313 15:41:25.423292 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/12616424-013a-47c8-9ed3-b407791537fe-scripts\") pod \"cinder-api-0\" (UID: \"12616424-013a-47c8-9ed3-b407791537fe\") " pod="openstack/cinder-api-0" Mar 13 15:41:25 crc kubenswrapper[4907]: I0313 15:41:25.423341 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12616424-013a-47c8-9ed3-b407791537fe-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"12616424-013a-47c8-9ed3-b407791537fe\") " pod="openstack/cinder-api-0" Mar 13 15:41:25 crc kubenswrapper[4907]: I0313 15:41:25.423378 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/12616424-013a-47c8-9ed3-b407791537fe-etc-machine-id\") pod \"cinder-api-0\" (UID: \"12616424-013a-47c8-9ed3-b407791537fe\") " pod="openstack/cinder-api-0" Mar 13 15:41:25 crc kubenswrapper[4907]: I0313 15:41:25.423401 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/12616424-013a-47c8-9ed3-b407791537fe-logs\") pod \"cinder-api-0\" (UID: \"12616424-013a-47c8-9ed3-b407791537fe\") " pod="openstack/cinder-api-0" Mar 13 15:41:25 crc kubenswrapper[4907]: I0313 15:41:25.423454 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12616424-013a-47c8-9ed3-b407791537fe-config-data\") pod \"cinder-api-0\" (UID: \"12616424-013a-47c8-9ed3-b407791537fe\") " pod="openstack/cinder-api-0" Mar 13 15:41:25 crc kubenswrapper[4907]: I0313 15:41:25.423799 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/12616424-013a-47c8-9ed3-b407791537fe-etc-machine-id\") pod \"cinder-api-0\" (UID: \"12616424-013a-47c8-9ed3-b407791537fe\") " pod="openstack/cinder-api-0" Mar 13 15:41:25 crc kubenswrapper[4907]: I0313 15:41:25.424456 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/12616424-013a-47c8-9ed3-b407791537fe-logs\") pod \"cinder-api-0\" (UID: \"12616424-013a-47c8-9ed3-b407791537fe\") " pod="openstack/cinder-api-0" Mar 13 15:41:25 crc kubenswrapper[4907]: I0313 15:41:25.428628 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/12616424-013a-47c8-9ed3-b407791537fe-scripts\") pod \"cinder-api-0\" (UID: \"12616424-013a-47c8-9ed3-b407791537fe\") " pod="openstack/cinder-api-0" Mar 13 15:41:25 crc kubenswrapper[4907]: I0313 15:41:25.428946 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12616424-013a-47c8-9ed3-b407791537fe-config-data\") pod \"cinder-api-0\" (UID: \"12616424-013a-47c8-9ed3-b407791537fe\") " pod="openstack/cinder-api-0" Mar 13 15:41:25 crc kubenswrapper[4907]: I0313 15:41:25.440868 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/12616424-013a-47c8-9ed3-b407791537fe-config-data-custom\") pod \"cinder-api-0\" (UID: \"12616424-013a-47c8-9ed3-b407791537fe\") " pod="openstack/cinder-api-0" Mar 13 15:41:25 crc kubenswrapper[4907]: I0313 15:41:25.442584 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dh7tk\" (UniqueName: \"kubernetes.io/projected/12616424-013a-47c8-9ed3-b407791537fe-kube-api-access-dh7tk\") pod \"cinder-api-0\" (UID: \"12616424-013a-47c8-9ed3-b407791537fe\") " pod="openstack/cinder-api-0" Mar 13 15:41:25 crc kubenswrapper[4907]: I0313 15:41:25.442612 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12616424-013a-47c8-9ed3-b407791537fe-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"12616424-013a-47c8-9ed3-b407791537fe\") " pod="openstack/cinder-api-0" Mar 13 15:41:25 crc kubenswrapper[4907]: I0313 15:41:25.503262 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Mar 13 15:41:25 crc kubenswrapper[4907]: I0313 15:41:25.841546 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6d59c8ff75-hm7rf"] Mar 13 15:41:26 crc kubenswrapper[4907]: I0313 15:41:26.023449 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Mar 13 15:41:26 crc kubenswrapper[4907]: W0313 15:41:26.024326 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod12616424_013a_47c8_9ed3_b407791537fe.slice/crio-32e9d54b2cd43cd1f900bf7ababaaace7fcabaccb1a6e6993088fcc8c885c896 WatchSource:0}: Error finding container 32e9d54b2cd43cd1f900bf7ababaaace7fcabaccb1a6e6993088fcc8c885c896: Status 404 returned error can't find the container with id 32e9d54b2cd43cd1f900bf7ababaaace7fcabaccb1a6e6993088fcc8c885c896 Mar 13 15:41:26 crc kubenswrapper[4907]: I0313 15:41:26.599040 4907 generic.go:334] "Generic (PLEG): container finished" podID="bb294da7-acf1-4984-a401-bbfc15ee97b7" containerID="3d3465a0228af3b611e5ca1d2d283d5739801f17bcfa78429f88dbe78da94add" exitCode=0 Mar 13 15:41:26 crc kubenswrapper[4907]: I0313 15:41:26.599350 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d59c8ff75-hm7rf" event={"ID":"bb294da7-acf1-4984-a401-bbfc15ee97b7","Type":"ContainerDied","Data":"3d3465a0228af3b611e5ca1d2d283d5739801f17bcfa78429f88dbe78da94add"} Mar 13 15:41:26 crc kubenswrapper[4907]: I0313 15:41:26.599382 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d59c8ff75-hm7rf" event={"ID":"bb294da7-acf1-4984-a401-bbfc15ee97b7","Type":"ContainerStarted","Data":"062befe692770a942fb75e8e6fd540c55d77667654c81f8c1cedf1f5d48f5908"} Mar 13 15:41:26 crc kubenswrapper[4907]: I0313 15:41:26.602326 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"12616424-013a-47c8-9ed3-b407791537fe","Type":"ContainerStarted","Data":"32e9d54b2cd43cd1f900bf7ababaaace7fcabaccb1a6e6993088fcc8c885c896"} Mar 13 15:41:27 crc kubenswrapper[4907]: I0313 15:41:27.610863 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"12616424-013a-47c8-9ed3-b407791537fe","Type":"ContainerStarted","Data":"f234443780f86c06ddc613fb1859ac940ed1d9617b56539195ffdc071bbc882b"} Mar 13 15:41:27 crc kubenswrapper[4907]: I0313 15:41:27.611393 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"12616424-013a-47c8-9ed3-b407791537fe","Type":"ContainerStarted","Data":"b5b4ab9a472db32a9c0ae91b2509ae28f1254ad915d9737b1279509e0c8d4717"} Mar 13 15:41:27 crc kubenswrapper[4907]: I0313 15:41:27.611414 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Mar 13 15:41:27 crc kubenswrapper[4907]: I0313 15:41:27.613189 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d59c8ff75-hm7rf" event={"ID":"bb294da7-acf1-4984-a401-bbfc15ee97b7","Type":"ContainerStarted","Data":"04739d4af66a61e29f63286c048d55b12a764eb2ad7e3456b1fff0c9a19ff45a"} Mar 13 15:41:27 crc kubenswrapper[4907]: I0313 15:41:27.613334 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6d59c8ff75-hm7rf" Mar 13 15:41:27 crc kubenswrapper[4907]: I0313 15:41:27.630393 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=2.630372634 podStartE2EDuration="2.630372634s" podCreationTimestamp="2026-03-13 15:41:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:41:27.629083838 +0000 UTC m=+5786.528871527" watchObservedRunningTime="2026-03-13 15:41:27.630372634 +0000 UTC m=+5786.530160323" Mar 13 15:41:27 crc kubenswrapper[4907]: I0313 15:41:27.652098 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6d59c8ff75-hm7rf" podStartSLOduration=3.6520765170000002 podStartE2EDuration="3.652076517s" podCreationTimestamp="2026-03-13 15:41:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:41:27.642833534 +0000 UTC m=+5786.542621223" watchObservedRunningTime="2026-03-13 15:41:27.652076517 +0000 UTC m=+5786.551864206" Mar 13 15:41:30 crc kubenswrapper[4907]: I0313 15:41:30.783865 4907 scope.go:117] "RemoveContainer" containerID="68ffe92dd03078633ec89d7ef29da9bbdc4c5f04c956a1e8487e3f2e01c5d1c5" Mar 13 15:41:30 crc kubenswrapper[4907]: E0313 15:41:30.784528 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:41:35 crc kubenswrapper[4907]: I0313 15:41:35.263096 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6d59c8ff75-hm7rf" Mar 13 15:41:35 crc kubenswrapper[4907]: I0313 15:41:35.336477 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-f9699695f-pwxvd"] Mar 13 15:41:35 crc kubenswrapper[4907]: I0313 15:41:35.336742 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-f9699695f-pwxvd" podUID="a122794c-3b28-4344-9817-f05a9e663110" containerName="dnsmasq-dns" containerID="cri-o://73bbc4d60c70852afe1d527317bf4861e321326bd2af88b210073cad5a3608bd" gracePeriod=10 Mar 13 15:41:35 crc kubenswrapper[4907]: I0313 15:41:35.685728 4907 generic.go:334] "Generic (PLEG): container finished" podID="a122794c-3b28-4344-9817-f05a9e663110" containerID="73bbc4d60c70852afe1d527317bf4861e321326bd2af88b210073cad5a3608bd" exitCode=0 Mar 13 15:41:35 crc kubenswrapper[4907]: I0313 15:41:35.686033 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f9699695f-pwxvd" event={"ID":"a122794c-3b28-4344-9817-f05a9e663110","Type":"ContainerDied","Data":"73bbc4d60c70852afe1d527317bf4861e321326bd2af88b210073cad5a3608bd"} Mar 13 15:41:35 crc kubenswrapper[4907]: I0313 15:41:35.868487 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f9699695f-pwxvd" Mar 13 15:41:35 crc kubenswrapper[4907]: I0313 15:41:35.948374 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a122794c-3b28-4344-9817-f05a9e663110-ovsdbserver-nb\") pod \"a122794c-3b28-4344-9817-f05a9e663110\" (UID: \"a122794c-3b28-4344-9817-f05a9e663110\") " Mar 13 15:41:35 crc kubenswrapper[4907]: I0313 15:41:35.948429 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a122794c-3b28-4344-9817-f05a9e663110-dns-svc\") pod \"a122794c-3b28-4344-9817-f05a9e663110\" (UID: \"a122794c-3b28-4344-9817-f05a9e663110\") " Mar 13 15:41:35 crc kubenswrapper[4907]: I0313 15:41:35.948451 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a122794c-3b28-4344-9817-f05a9e663110-config\") pod \"a122794c-3b28-4344-9817-f05a9e663110\" (UID: \"a122794c-3b28-4344-9817-f05a9e663110\") " Mar 13 15:41:35 crc kubenswrapper[4907]: I0313 15:41:35.949150 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a122794c-3b28-4344-9817-f05a9e663110-ovsdbserver-sb\") pod \"a122794c-3b28-4344-9817-f05a9e663110\" (UID: \"a122794c-3b28-4344-9817-f05a9e663110\") " Mar 13 15:41:35 crc kubenswrapper[4907]: I0313 15:41:35.949206 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d2fkn\" (UniqueName: \"kubernetes.io/projected/a122794c-3b28-4344-9817-f05a9e663110-kube-api-access-d2fkn\") pod \"a122794c-3b28-4344-9817-f05a9e663110\" (UID: \"a122794c-3b28-4344-9817-f05a9e663110\") " Mar 13 15:41:35 crc kubenswrapper[4907]: I0313 15:41:35.958173 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a122794c-3b28-4344-9817-f05a9e663110-kube-api-access-d2fkn" (OuterVolumeSpecName: "kube-api-access-d2fkn") pod "a122794c-3b28-4344-9817-f05a9e663110" (UID: "a122794c-3b28-4344-9817-f05a9e663110"). InnerVolumeSpecName "kube-api-access-d2fkn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:41:36 crc kubenswrapper[4907]: I0313 15:41:36.051082 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d2fkn\" (UniqueName: \"kubernetes.io/projected/a122794c-3b28-4344-9817-f05a9e663110-kube-api-access-d2fkn\") on node \"crc\" DevicePath \"\"" Mar 13 15:41:36 crc kubenswrapper[4907]: I0313 15:41:36.052078 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a122794c-3b28-4344-9817-f05a9e663110-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "a122794c-3b28-4344-9817-f05a9e663110" (UID: "a122794c-3b28-4344-9817-f05a9e663110"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:41:36 crc kubenswrapper[4907]: I0313 15:41:36.138261 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a122794c-3b28-4344-9817-f05a9e663110-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "a122794c-3b28-4344-9817-f05a9e663110" (UID: "a122794c-3b28-4344-9817-f05a9e663110"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:41:36 crc kubenswrapper[4907]: I0313 15:41:36.153978 4907 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a122794c-3b28-4344-9817-f05a9e663110-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 13 15:41:36 crc kubenswrapper[4907]: I0313 15:41:36.154011 4907 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a122794c-3b28-4344-9817-f05a9e663110-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 13 15:41:36 crc kubenswrapper[4907]: I0313 15:41:36.168648 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a122794c-3b28-4344-9817-f05a9e663110-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a122794c-3b28-4344-9817-f05a9e663110" (UID: "a122794c-3b28-4344-9817-f05a9e663110"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:41:36 crc kubenswrapper[4907]: I0313 15:41:36.173314 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a122794c-3b28-4344-9817-f05a9e663110-config" (OuterVolumeSpecName: "config") pod "a122794c-3b28-4344-9817-f05a9e663110" (UID: "a122794c-3b28-4344-9817-f05a9e663110"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:41:36 crc kubenswrapper[4907]: I0313 15:41:36.255724 4907 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a122794c-3b28-4344-9817-f05a9e663110-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 13 15:41:36 crc kubenswrapper[4907]: I0313 15:41:36.255753 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a122794c-3b28-4344-9817-f05a9e663110-config\") on node \"crc\" DevicePath \"\"" Mar 13 15:41:36 crc kubenswrapper[4907]: I0313 15:41:36.697032 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f9699695f-pwxvd" event={"ID":"a122794c-3b28-4344-9817-f05a9e663110","Type":"ContainerDied","Data":"8f46474c0bd8af139725a1ba39a29f7abd9cd2e12d0840cb8535991ab1cd685e"} Mar 13 15:41:36 crc kubenswrapper[4907]: I0313 15:41:36.697083 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f9699695f-pwxvd" Mar 13 15:41:36 crc kubenswrapper[4907]: I0313 15:41:36.697098 4907 scope.go:117] "RemoveContainer" containerID="73bbc4d60c70852afe1d527317bf4861e321326bd2af88b210073cad5a3608bd" Mar 13 15:41:36 crc kubenswrapper[4907]: I0313 15:41:36.722944 4907 scope.go:117] "RemoveContainer" containerID="ec3db5cd0d09990bedd2c273757078f6b84cc7be565b5f045a0e95e3abe11b83" Mar 13 15:41:36 crc kubenswrapper[4907]: I0313 15:41:36.732089 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-f9699695f-pwxvd"] Mar 13 15:41:36 crc kubenswrapper[4907]: I0313 15:41:36.749195 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-f9699695f-pwxvd"] Mar 13 15:41:37 crc kubenswrapper[4907]: I0313 15:41:37.104037 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 13 15:41:37 crc kubenswrapper[4907]: I0313 15:41:37.104904 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="6244358c-3ba3-41fa-bf94-0b2bfb6ef956" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://2ff24611962f784f04b97d2b6de0e1621289a9ce82da636a0133b47ec0d27610" gracePeriod=30 Mar 13 15:41:37 crc kubenswrapper[4907]: I0313 15:41:37.113698 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 15:41:37 crc kubenswrapper[4907]: I0313 15:41:37.114034 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="61a2ddd9-1fcc-427e-aaa2-f092ebb7bd30" containerName="nova-scheduler-scheduler" containerID="cri-o://ce84852f1d320a52f408594e4b89f2cbf753717a67ce179872f364593d90c469" gracePeriod=30 Mar 13 15:41:37 crc kubenswrapper[4907]: I0313 15:41:37.129946 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Mar 13 15:41:37 crc kubenswrapper[4907]: I0313 15:41:37.130255 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="81ef3cd7-e06f-49f8-bc01-41b32fbb5546" containerName="nova-api-log" containerID="cri-o://697da8faed8339f44e9c8f069a46c98c95d40fcca3ba8965d84a7f51ddc0a358" gracePeriod=30 Mar 13 15:41:37 crc kubenswrapper[4907]: I0313 15:41:37.130423 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="81ef3cd7-e06f-49f8-bc01-41b32fbb5546" containerName="nova-api-api" containerID="cri-o://da08a8ab4ad1ce90fd28c0cf97955538dcb49b21c61a30f6afb435225b153254" gracePeriod=30 Mar 13 15:41:37 crc kubenswrapper[4907]: I0313 15:41:37.142147 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Mar 13 15:41:37 crc kubenswrapper[4907]: I0313 15:41:37.142759 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="37198210-23ed-4025-8cd9-e284c6573318" containerName="nova-metadata-log" containerID="cri-o://6dcdbcfacbeb1b7e6dcbc3fb09e947521ce095a95f3551396c6c38c4e6a8b77f" gracePeriod=30 Mar 13 15:41:37 crc kubenswrapper[4907]: I0313 15:41:37.143025 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="37198210-23ed-4025-8cd9-e284c6573318" containerName="nova-metadata-metadata" containerID="cri-o://ae907961f503a65c4dfb28c60282669570b0349aea1f6830f2e1f6e44d773cdb" gracePeriod=30 Mar 13 15:41:37 crc kubenswrapper[4907]: I0313 15:41:37.157067 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Mar 13 15:41:37 crc kubenswrapper[4907]: I0313 15:41:37.157296 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell0-conductor-0" podUID="a8047522-cc05-44a3-b57d-2362e6457d49" containerName="nova-cell0-conductor-conductor" containerID="cri-o://4efeb4b84a961f5a72c12fc7db6d3a11b565d9a8fc5e00e1bd6bd862f0227c34" gracePeriod=30 Mar 13 15:41:37 crc kubenswrapper[4907]: I0313 15:41:37.190735 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Mar 13 15:41:37 crc kubenswrapper[4907]: I0313 15:41:37.190959 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-conductor-0" podUID="d8d0d2c8-d0f8-4dc1-b65e-e694571a6072" containerName="nova-cell1-conductor-conductor" containerID="cri-o://716a10e355d921d8594fa1927d76d58c79889f5d1c1505d7df88ab1927fd9db4" gracePeriod=30 Mar 13 15:41:37 crc kubenswrapper[4907]: I0313 15:41:37.330609 4907 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-cell1-novncproxy-0" podUID="6244358c-3ba3-41fa-bf94-0b2bfb6ef956" containerName="nova-cell1-novncproxy-novncproxy" probeResult="failure" output="Get \"http://10.217.1.105:6080/vnc_lite.html\": dial tcp 10.217.1.105:6080: connect: connection refused" Mar 13 15:41:37 crc kubenswrapper[4907]: I0313 15:41:37.714981 4907 generic.go:334] "Generic (PLEG): container finished" podID="37198210-23ed-4025-8cd9-e284c6573318" containerID="6dcdbcfacbeb1b7e6dcbc3fb09e947521ce095a95f3551396c6c38c4e6a8b77f" exitCode=143 Mar 13 15:41:37 crc kubenswrapper[4907]: I0313 15:41:37.715056 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"37198210-23ed-4025-8cd9-e284c6573318","Type":"ContainerDied","Data":"6dcdbcfacbeb1b7e6dcbc3fb09e947521ce095a95f3551396c6c38c4e6a8b77f"} Mar 13 15:41:37 crc kubenswrapper[4907]: I0313 15:41:37.719500 4907 generic.go:334] "Generic (PLEG): container finished" podID="81ef3cd7-e06f-49f8-bc01-41b32fbb5546" containerID="697da8faed8339f44e9c8f069a46c98c95d40fcca3ba8965d84a7f51ddc0a358" exitCode=143 Mar 13 15:41:37 crc kubenswrapper[4907]: I0313 15:41:37.719586 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"81ef3cd7-e06f-49f8-bc01-41b32fbb5546","Type":"ContainerDied","Data":"697da8faed8339f44e9c8f069a46c98c95d40fcca3ba8965d84a7f51ddc0a358"} Mar 13 15:41:37 crc kubenswrapper[4907]: I0313 15:41:37.721045 4907 generic.go:334] "Generic (PLEG): container finished" podID="6244358c-3ba3-41fa-bf94-0b2bfb6ef956" containerID="2ff24611962f784f04b97d2b6de0e1621289a9ce82da636a0133b47ec0d27610" exitCode=0 Mar 13 15:41:37 crc kubenswrapper[4907]: I0313 15:41:37.721074 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"6244358c-3ba3-41fa-bf94-0b2bfb6ef956","Type":"ContainerDied","Data":"2ff24611962f784f04b97d2b6de0e1621289a9ce82da636a0133b47ec0d27610"} Mar 13 15:41:37 crc kubenswrapper[4907]: I0313 15:41:37.796379 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a122794c-3b28-4344-9817-f05a9e663110" path="/var/lib/kubelet/pods/a122794c-3b28-4344-9817-f05a9e663110/volumes" Mar 13 15:41:37 crc kubenswrapper[4907]: I0313 15:41:37.937358 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Mar 13 15:41:38 crc kubenswrapper[4907]: I0313 15:41:38.071967 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Mar 13 15:41:38 crc kubenswrapper[4907]: I0313 15:41:38.193441 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6244358c-3ba3-41fa-bf94-0b2bfb6ef956-combined-ca-bundle\") pod \"6244358c-3ba3-41fa-bf94-0b2bfb6ef956\" (UID: \"6244358c-3ba3-41fa-bf94-0b2bfb6ef956\") " Mar 13 15:41:38 crc kubenswrapper[4907]: I0313 15:41:38.193592 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6244358c-3ba3-41fa-bf94-0b2bfb6ef956-config-data\") pod \"6244358c-3ba3-41fa-bf94-0b2bfb6ef956\" (UID: \"6244358c-3ba3-41fa-bf94-0b2bfb6ef956\") " Mar 13 15:41:38 crc kubenswrapper[4907]: I0313 15:41:38.193752 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xl2nr\" (UniqueName: \"kubernetes.io/projected/6244358c-3ba3-41fa-bf94-0b2bfb6ef956-kube-api-access-xl2nr\") pod \"6244358c-3ba3-41fa-bf94-0b2bfb6ef956\" (UID: \"6244358c-3ba3-41fa-bf94-0b2bfb6ef956\") " Mar 13 15:41:38 crc kubenswrapper[4907]: I0313 15:41:38.230550 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6244358c-3ba3-41fa-bf94-0b2bfb6ef956-config-data" (OuterVolumeSpecName: "config-data") pod "6244358c-3ba3-41fa-bf94-0b2bfb6ef956" (UID: "6244358c-3ba3-41fa-bf94-0b2bfb6ef956"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:41:38 crc kubenswrapper[4907]: I0313 15:41:38.230694 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6244358c-3ba3-41fa-bf94-0b2bfb6ef956-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6244358c-3ba3-41fa-bf94-0b2bfb6ef956" (UID: "6244358c-3ba3-41fa-bf94-0b2bfb6ef956"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:41:38 crc kubenswrapper[4907]: I0313 15:41:38.231046 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6244358c-3ba3-41fa-bf94-0b2bfb6ef956-kube-api-access-xl2nr" (OuterVolumeSpecName: "kube-api-access-xl2nr") pod "6244358c-3ba3-41fa-bf94-0b2bfb6ef956" (UID: "6244358c-3ba3-41fa-bf94-0b2bfb6ef956"). InnerVolumeSpecName "kube-api-access-xl2nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:41:38 crc kubenswrapper[4907]: I0313 15:41:38.305549 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6244358c-3ba3-41fa-bf94-0b2bfb6ef956-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 15:41:38 crc kubenswrapper[4907]: I0313 15:41:38.305581 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6244358c-3ba3-41fa-bf94-0b2bfb6ef956-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 15:41:38 crc kubenswrapper[4907]: I0313 15:41:38.305605 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xl2nr\" (UniqueName: \"kubernetes.io/projected/6244358c-3ba3-41fa-bf94-0b2bfb6ef956-kube-api-access-xl2nr\") on node \"crc\" DevicePath \"\"" Mar 13 15:41:38 crc kubenswrapper[4907]: I0313 15:41:38.450097 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Mar 13 15:41:38 crc kubenswrapper[4907]: E0313 15:41:38.534195 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="ce84852f1d320a52f408594e4b89f2cbf753717a67ce179872f364593d90c469" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Mar 13 15:41:38 crc kubenswrapper[4907]: E0313 15:41:38.536632 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="ce84852f1d320a52f408594e4b89f2cbf753717a67ce179872f364593d90c469" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Mar 13 15:41:38 crc kubenswrapper[4907]: E0313 15:41:38.538078 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="ce84852f1d320a52f408594e4b89f2cbf753717a67ce179872f364593d90c469" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Mar 13 15:41:38 crc kubenswrapper[4907]: E0313 15:41:38.538120 4907 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="61a2ddd9-1fcc-427e-aaa2-f092ebb7bd30" containerName="nova-scheduler-scheduler" Mar 13 15:41:38 crc kubenswrapper[4907]: I0313 15:41:38.633531 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d8d0d2c8-d0f8-4dc1-b65e-e694571a6072-config-data\") pod \"d8d0d2c8-d0f8-4dc1-b65e-e694571a6072\" (UID: \"d8d0d2c8-d0f8-4dc1-b65e-e694571a6072\") " Mar 13 15:41:38 crc kubenswrapper[4907]: I0313 15:41:38.633694 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8d0d2c8-d0f8-4dc1-b65e-e694571a6072-combined-ca-bundle\") pod \"d8d0d2c8-d0f8-4dc1-b65e-e694571a6072\" (UID: \"d8d0d2c8-d0f8-4dc1-b65e-e694571a6072\") " Mar 13 15:41:38 crc kubenswrapper[4907]: I0313 15:41:38.633743 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-47msf\" (UniqueName: \"kubernetes.io/projected/d8d0d2c8-d0f8-4dc1-b65e-e694571a6072-kube-api-access-47msf\") pod \"d8d0d2c8-d0f8-4dc1-b65e-e694571a6072\" (UID: \"d8d0d2c8-d0f8-4dc1-b65e-e694571a6072\") " Mar 13 15:41:38 crc kubenswrapper[4907]: I0313 15:41:38.640715 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d8d0d2c8-d0f8-4dc1-b65e-e694571a6072-kube-api-access-47msf" (OuterVolumeSpecName: "kube-api-access-47msf") pod "d8d0d2c8-d0f8-4dc1-b65e-e694571a6072" (UID: "d8d0d2c8-d0f8-4dc1-b65e-e694571a6072"). InnerVolumeSpecName "kube-api-access-47msf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:41:38 crc kubenswrapper[4907]: I0313 15:41:38.670640 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8d0d2c8-d0f8-4dc1-b65e-e694571a6072-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d8d0d2c8-d0f8-4dc1-b65e-e694571a6072" (UID: "d8d0d2c8-d0f8-4dc1-b65e-e694571a6072"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:41:38 crc kubenswrapper[4907]: I0313 15:41:38.677067 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8d0d2c8-d0f8-4dc1-b65e-e694571a6072-config-data" (OuterVolumeSpecName: "config-data") pod "d8d0d2c8-d0f8-4dc1-b65e-e694571a6072" (UID: "d8d0d2c8-d0f8-4dc1-b65e-e694571a6072"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:41:38 crc kubenswrapper[4907]: I0313 15:41:38.733479 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Mar 13 15:41:38 crc kubenswrapper[4907]: I0313 15:41:38.743698 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"6244358c-3ba3-41fa-bf94-0b2bfb6ef956","Type":"ContainerDied","Data":"97ff1fdcfd5fd91a658ff67b990106cf5c21eac69b149055f325885b224d558e"} Mar 13 15:41:38 crc kubenswrapper[4907]: I0313 15:41:38.748838 4907 scope.go:117] "RemoveContainer" containerID="2ff24611962f784f04b97d2b6de0e1621289a9ce82da636a0133b47ec0d27610" Mar 13 15:41:38 crc kubenswrapper[4907]: I0313 15:41:38.748932 4907 generic.go:334] "Generic (PLEG): container finished" podID="d8d0d2c8-d0f8-4dc1-b65e-e694571a6072" containerID="716a10e355d921d8594fa1927d76d58c79889f5d1c1505d7df88ab1927fd9db4" exitCode=0 Mar 13 15:41:38 crc kubenswrapper[4907]: I0313 15:41:38.748509 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d8d0d2c8-d0f8-4dc1-b65e-e694571a6072-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 15:41:38 crc kubenswrapper[4907]: I0313 15:41:38.748974 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"d8d0d2c8-d0f8-4dc1-b65e-e694571a6072","Type":"ContainerDied","Data":"716a10e355d921d8594fa1927d76d58c79889f5d1c1505d7df88ab1927fd9db4"} Mar 13 15:41:38 crc kubenswrapper[4907]: I0313 15:41:38.748994 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8d0d2c8-d0f8-4dc1-b65e-e694571a6072-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 15:41:38 crc kubenswrapper[4907]: I0313 15:41:38.749006 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"d8d0d2c8-d0f8-4dc1-b65e-e694571a6072","Type":"ContainerDied","Data":"cbddd5aedb9a61c15acca2a6a299e779028d30303050e66d3356b0d2ec6abc1d"} Mar 13 15:41:38 crc kubenswrapper[4907]: I0313 15:41:38.749010 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-47msf\" (UniqueName: \"kubernetes.io/projected/d8d0d2c8-d0f8-4dc1-b65e-e694571a6072-kube-api-access-47msf\") on node \"crc\" DevicePath \"\"" Mar 13 15:41:38 crc kubenswrapper[4907]: I0313 15:41:38.749087 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Mar 13 15:41:38 crc kubenswrapper[4907]: I0313 15:41:38.812942 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 13 15:41:38 crc kubenswrapper[4907]: I0313 15:41:38.825197 4907 scope.go:117] "RemoveContainer" containerID="716a10e355d921d8594fa1927d76d58c79889f5d1c1505d7df88ab1927fd9db4" Mar 13 15:41:38 crc kubenswrapper[4907]: I0313 15:41:38.840084 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 13 15:41:38 crc kubenswrapper[4907]: I0313 15:41:38.857152 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Mar 13 15:41:38 crc kubenswrapper[4907]: I0313 15:41:38.868469 4907 scope.go:117] "RemoveContainer" containerID="716a10e355d921d8594fa1927d76d58c79889f5d1c1505d7df88ab1927fd9db4" Mar 13 15:41:38 crc kubenswrapper[4907]: E0313 15:41:38.869913 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"716a10e355d921d8594fa1927d76d58c79889f5d1c1505d7df88ab1927fd9db4\": container with ID starting with 716a10e355d921d8594fa1927d76d58c79889f5d1c1505d7df88ab1927fd9db4 not found: ID does not exist" containerID="716a10e355d921d8594fa1927d76d58c79889f5d1c1505d7df88ab1927fd9db4" Mar 13 15:41:38 crc kubenswrapper[4907]: I0313 15:41:38.869977 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"716a10e355d921d8594fa1927d76d58c79889f5d1c1505d7df88ab1927fd9db4"} err="failed to get container status \"716a10e355d921d8594fa1927d76d58c79889f5d1c1505d7df88ab1927fd9db4\": rpc error: code = NotFound desc = could not find container \"716a10e355d921d8594fa1927d76d58c79889f5d1c1505d7df88ab1927fd9db4\": container with ID starting with 716a10e355d921d8594fa1927d76d58c79889f5d1c1505d7df88ab1927fd9db4 not found: ID does not exist" Mar 13 15:41:38 crc kubenswrapper[4907]: I0313 15:41:38.899940 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-0"] Mar 13 15:41:38 crc kubenswrapper[4907]: I0313 15:41:38.917627 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 13 15:41:38 crc kubenswrapper[4907]: E0313 15:41:38.918139 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6244358c-3ba3-41fa-bf94-0b2bfb6ef956" containerName="nova-cell1-novncproxy-novncproxy" Mar 13 15:41:38 crc kubenswrapper[4907]: I0313 15:41:38.918166 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="6244358c-3ba3-41fa-bf94-0b2bfb6ef956" containerName="nova-cell1-novncproxy-novncproxy" Mar 13 15:41:38 crc kubenswrapper[4907]: E0313 15:41:38.918196 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a122794c-3b28-4344-9817-f05a9e663110" containerName="init" Mar 13 15:41:38 crc kubenswrapper[4907]: I0313 15:41:38.918204 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="a122794c-3b28-4344-9817-f05a9e663110" containerName="init" Mar 13 15:41:38 crc kubenswrapper[4907]: E0313 15:41:38.918220 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a122794c-3b28-4344-9817-f05a9e663110" containerName="dnsmasq-dns" Mar 13 15:41:38 crc kubenswrapper[4907]: I0313 15:41:38.918228 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="a122794c-3b28-4344-9817-f05a9e663110" containerName="dnsmasq-dns" Mar 13 15:41:38 crc kubenswrapper[4907]: E0313 15:41:38.918263 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8d0d2c8-d0f8-4dc1-b65e-e694571a6072" containerName="nova-cell1-conductor-conductor" Mar 13 15:41:38 crc kubenswrapper[4907]: I0313 15:41:38.918271 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8d0d2c8-d0f8-4dc1-b65e-e694571a6072" containerName="nova-cell1-conductor-conductor" Mar 13 15:41:38 crc kubenswrapper[4907]: I0313 15:41:38.918465 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="d8d0d2c8-d0f8-4dc1-b65e-e694571a6072" containerName="nova-cell1-conductor-conductor" Mar 13 15:41:38 crc kubenswrapper[4907]: I0313 15:41:38.918495 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="a122794c-3b28-4344-9817-f05a9e663110" containerName="dnsmasq-dns" Mar 13 15:41:38 crc kubenswrapper[4907]: I0313 15:41:38.918509 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="6244358c-3ba3-41fa-bf94-0b2bfb6ef956" containerName="nova-cell1-novncproxy-novncproxy" Mar 13 15:41:38 crc kubenswrapper[4907]: I0313 15:41:38.919267 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Mar 13 15:41:38 crc kubenswrapper[4907]: I0313 15:41:38.934734 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Mar 13 15:41:38 crc kubenswrapper[4907]: I0313 15:41:38.940071 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Mar 13 15:41:38 crc kubenswrapper[4907]: I0313 15:41:38.941352 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Mar 13 15:41:38 crc kubenswrapper[4907]: I0313 15:41:38.945379 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Mar 13 15:41:38 crc kubenswrapper[4907]: I0313 15:41:38.951490 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 13 15:41:38 crc kubenswrapper[4907]: I0313 15:41:38.952697 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/317b6fd4-e642-4858-bc3d-2cce128ea0e7-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"317b6fd4-e642-4858-bc3d-2cce128ea0e7\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 15:41:38 crc kubenswrapper[4907]: I0313 15:41:38.952759 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/317b6fd4-e642-4858-bc3d-2cce128ea0e7-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"317b6fd4-e642-4858-bc3d-2cce128ea0e7\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 15:41:38 crc kubenswrapper[4907]: I0313 15:41:38.952784 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hh9bc\" (UniqueName: \"kubernetes.io/projected/eded65eb-7a04-4f90-ad82-be65db9014eb-kube-api-access-hh9bc\") pod \"nova-cell1-conductor-0\" (UID: \"eded65eb-7a04-4f90-ad82-be65db9014eb\") " pod="openstack/nova-cell1-conductor-0" Mar 13 15:41:38 crc kubenswrapper[4907]: I0313 15:41:38.952842 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eded65eb-7a04-4f90-ad82-be65db9014eb-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"eded65eb-7a04-4f90-ad82-be65db9014eb\") " pod="openstack/nova-cell1-conductor-0" Mar 13 15:41:38 crc kubenswrapper[4907]: I0313 15:41:38.952891 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h69pj\" (UniqueName: \"kubernetes.io/projected/317b6fd4-e642-4858-bc3d-2cce128ea0e7-kube-api-access-h69pj\") pod \"nova-cell1-novncproxy-0\" (UID: \"317b6fd4-e642-4858-bc3d-2cce128ea0e7\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 15:41:38 crc kubenswrapper[4907]: I0313 15:41:38.952915 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eded65eb-7a04-4f90-ad82-be65db9014eb-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"eded65eb-7a04-4f90-ad82-be65db9014eb\") " pod="openstack/nova-cell1-conductor-0" Mar 13 15:41:38 crc kubenswrapper[4907]: I0313 15:41:38.969701 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Mar 13 15:41:39 crc kubenswrapper[4907]: I0313 15:41:39.054673 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/317b6fd4-e642-4858-bc3d-2cce128ea0e7-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"317b6fd4-e642-4858-bc3d-2cce128ea0e7\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 15:41:39 crc kubenswrapper[4907]: I0313 15:41:39.054747 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/317b6fd4-e642-4858-bc3d-2cce128ea0e7-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"317b6fd4-e642-4858-bc3d-2cce128ea0e7\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 15:41:39 crc kubenswrapper[4907]: I0313 15:41:39.054779 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hh9bc\" (UniqueName: \"kubernetes.io/projected/eded65eb-7a04-4f90-ad82-be65db9014eb-kube-api-access-hh9bc\") pod \"nova-cell1-conductor-0\" (UID: \"eded65eb-7a04-4f90-ad82-be65db9014eb\") " pod="openstack/nova-cell1-conductor-0" Mar 13 15:41:39 crc kubenswrapper[4907]: I0313 15:41:39.054849 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eded65eb-7a04-4f90-ad82-be65db9014eb-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"eded65eb-7a04-4f90-ad82-be65db9014eb\") " pod="openstack/nova-cell1-conductor-0" Mar 13 15:41:39 crc kubenswrapper[4907]: I0313 15:41:39.054902 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h69pj\" (UniqueName: \"kubernetes.io/projected/317b6fd4-e642-4858-bc3d-2cce128ea0e7-kube-api-access-h69pj\") pod \"nova-cell1-novncproxy-0\" (UID: \"317b6fd4-e642-4858-bc3d-2cce128ea0e7\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 15:41:39 crc kubenswrapper[4907]: I0313 15:41:39.054935 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eded65eb-7a04-4f90-ad82-be65db9014eb-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"eded65eb-7a04-4f90-ad82-be65db9014eb\") " pod="openstack/nova-cell1-conductor-0" Mar 13 15:41:39 crc kubenswrapper[4907]: I0313 15:41:39.061690 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/317b6fd4-e642-4858-bc3d-2cce128ea0e7-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"317b6fd4-e642-4858-bc3d-2cce128ea0e7\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 15:41:39 crc kubenswrapper[4907]: I0313 15:41:39.061711 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eded65eb-7a04-4f90-ad82-be65db9014eb-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"eded65eb-7a04-4f90-ad82-be65db9014eb\") " pod="openstack/nova-cell1-conductor-0" Mar 13 15:41:39 crc kubenswrapper[4907]: I0313 15:41:39.068316 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/317b6fd4-e642-4858-bc3d-2cce128ea0e7-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"317b6fd4-e642-4858-bc3d-2cce128ea0e7\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 15:41:39 crc kubenswrapper[4907]: I0313 15:41:39.071153 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hh9bc\" (UniqueName: \"kubernetes.io/projected/eded65eb-7a04-4f90-ad82-be65db9014eb-kube-api-access-hh9bc\") pod \"nova-cell1-conductor-0\" (UID: \"eded65eb-7a04-4f90-ad82-be65db9014eb\") " pod="openstack/nova-cell1-conductor-0" Mar 13 15:41:39 crc kubenswrapper[4907]: I0313 15:41:39.071452 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h69pj\" (UniqueName: \"kubernetes.io/projected/317b6fd4-e642-4858-bc3d-2cce128ea0e7-kube-api-access-h69pj\") pod \"nova-cell1-novncproxy-0\" (UID: \"317b6fd4-e642-4858-bc3d-2cce128ea0e7\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 15:41:39 crc kubenswrapper[4907]: I0313 15:41:39.079843 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eded65eb-7a04-4f90-ad82-be65db9014eb-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"eded65eb-7a04-4f90-ad82-be65db9014eb\") " pod="openstack/nova-cell1-conductor-0" Mar 13 15:41:39 crc kubenswrapper[4907]: I0313 15:41:39.237081 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Mar 13 15:41:39 crc kubenswrapper[4907]: I0313 15:41:39.261567 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Mar 13 15:41:40 crc kubenswrapper[4907]: I0313 15:41:39.769022 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 13 15:41:40 crc kubenswrapper[4907]: I0313 15:41:39.808311 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6244358c-3ba3-41fa-bf94-0b2bfb6ef956" path="/var/lib/kubelet/pods/6244358c-3ba3-41fa-bf94-0b2bfb6ef956/volumes" Mar 13 15:41:40 crc kubenswrapper[4907]: I0313 15:41:39.812999 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d8d0d2c8-d0f8-4dc1-b65e-e694571a6072" path="/var/lib/kubelet/pods/d8d0d2c8-d0f8-4dc1-b65e-e694571a6072/volumes" Mar 13 15:41:40 crc kubenswrapper[4907]: I0313 15:41:40.522034 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Mar 13 15:41:40 crc kubenswrapper[4907]: I0313 15:41:40.795388 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 13 15:41:40 crc kubenswrapper[4907]: I0313 15:41:40.804220 4907 generic.go:334] "Generic (PLEG): container finished" podID="81ef3cd7-e06f-49f8-bc01-41b32fbb5546" containerID="da08a8ab4ad1ce90fd28c0cf97955538dcb49b21c61a30f6afb435225b153254" exitCode=0 Mar 13 15:41:40 crc kubenswrapper[4907]: I0313 15:41:40.804286 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"81ef3cd7-e06f-49f8-bc01-41b32fbb5546","Type":"ContainerDied","Data":"da08a8ab4ad1ce90fd28c0cf97955538dcb49b21c61a30f6afb435225b153254"} Mar 13 15:41:40 crc kubenswrapper[4907]: I0313 15:41:40.804312 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"81ef3cd7-e06f-49f8-bc01-41b32fbb5546","Type":"ContainerDied","Data":"c85aa3e7cd1bf81f73dff61921375db95cb273392a8693e2f1fe5ad2b1d643c9"} Mar 13 15:41:40 crc kubenswrapper[4907]: I0313 15:41:40.804328 4907 scope.go:117] "RemoveContainer" containerID="da08a8ab4ad1ce90fd28c0cf97955538dcb49b21c61a30f6afb435225b153254" Mar 13 15:41:40 crc kubenswrapper[4907]: I0313 15:41:40.806918 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"317b6fd4-e642-4858-bc3d-2cce128ea0e7","Type":"ContainerStarted","Data":"70f28ec9e0de121bbf148130d45c1213b4aeefcb940ccc1c2c8bb04c732a2477"} Mar 13 15:41:40 crc kubenswrapper[4907]: I0313 15:41:40.806952 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"317b6fd4-e642-4858-bc3d-2cce128ea0e7","Type":"ContainerStarted","Data":"a5faa6bbb08bed4fc6f210842e67428795ae39508c3318118605a1b438f13212"} Mar 13 15:41:40 crc kubenswrapper[4907]: I0313 15:41:40.818495 4907 generic.go:334] "Generic (PLEG): container finished" podID="37198210-23ed-4025-8cd9-e284c6573318" containerID="ae907961f503a65c4dfb28c60282669570b0349aea1f6830f2e1f6e44d773cdb" exitCode=0 Mar 13 15:41:40 crc kubenswrapper[4907]: I0313 15:41:40.818558 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"37198210-23ed-4025-8cd9-e284c6573318","Type":"ContainerDied","Data":"ae907961f503a65c4dfb28c60282669570b0349aea1f6830f2e1f6e44d773cdb"} Mar 13 15:41:40 crc kubenswrapper[4907]: I0313 15:41:40.819495 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"eded65eb-7a04-4f90-ad82-be65db9014eb","Type":"ContainerStarted","Data":"4c8aa8b2ae3cb083bba62c66c737dee676b8b626c0c8f53960f20765d91bb7b2"} Mar 13 15:41:40 crc kubenswrapper[4907]: I0313 15:41:40.832213 4907 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-f9699695f-pwxvd" podUID="a122794c-3b28-4344-9817-f05a9e663110" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.1.115:5353: i/o timeout" Mar 13 15:41:40 crc kubenswrapper[4907]: I0313 15:41:40.845011 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.84499128 podStartE2EDuration="2.84499128s" podCreationTimestamp="2026-03-13 15:41:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:41:40.83763362 +0000 UTC m=+5799.737421309" watchObservedRunningTime="2026-03-13 15:41:40.84499128 +0000 UTC m=+5799.744778969" Mar 13 15:41:40 crc kubenswrapper[4907]: I0313 15:41:40.919748 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81ef3cd7-e06f-49f8-bc01-41b32fbb5546-combined-ca-bundle\") pod \"81ef3cd7-e06f-49f8-bc01-41b32fbb5546\" (UID: \"81ef3cd7-e06f-49f8-bc01-41b32fbb5546\") " Mar 13 15:41:40 crc kubenswrapper[4907]: I0313 15:41:40.919924 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xct78\" (UniqueName: \"kubernetes.io/projected/81ef3cd7-e06f-49f8-bc01-41b32fbb5546-kube-api-access-xct78\") pod \"81ef3cd7-e06f-49f8-bc01-41b32fbb5546\" (UID: \"81ef3cd7-e06f-49f8-bc01-41b32fbb5546\") " Mar 13 15:41:40 crc kubenswrapper[4907]: I0313 15:41:40.919978 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81ef3cd7-e06f-49f8-bc01-41b32fbb5546-config-data\") pod \"81ef3cd7-e06f-49f8-bc01-41b32fbb5546\" (UID: \"81ef3cd7-e06f-49f8-bc01-41b32fbb5546\") " Mar 13 15:41:40 crc kubenswrapper[4907]: I0313 15:41:40.920036 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/81ef3cd7-e06f-49f8-bc01-41b32fbb5546-logs\") pod \"81ef3cd7-e06f-49f8-bc01-41b32fbb5546\" (UID: \"81ef3cd7-e06f-49f8-bc01-41b32fbb5546\") " Mar 13 15:41:40 crc kubenswrapper[4907]: I0313 15:41:40.922294 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/81ef3cd7-e06f-49f8-bc01-41b32fbb5546-logs" (OuterVolumeSpecName: "logs") pod "81ef3cd7-e06f-49f8-bc01-41b32fbb5546" (UID: "81ef3cd7-e06f-49f8-bc01-41b32fbb5546"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:41:40 crc kubenswrapper[4907]: I0313 15:41:40.928060 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/81ef3cd7-e06f-49f8-bc01-41b32fbb5546-kube-api-access-xct78" (OuterVolumeSpecName: "kube-api-access-xct78") pod "81ef3cd7-e06f-49f8-bc01-41b32fbb5546" (UID: "81ef3cd7-e06f-49f8-bc01-41b32fbb5546"). InnerVolumeSpecName "kube-api-access-xct78". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:41:40 crc kubenswrapper[4907]: I0313 15:41:40.940411 4907 scope.go:117] "RemoveContainer" containerID="697da8faed8339f44e9c8f069a46c98c95d40fcca3ba8965d84a7f51ddc0a358" Mar 13 15:41:40 crc kubenswrapper[4907]: I0313 15:41:40.946219 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 13 15:41:40 crc kubenswrapper[4907]: I0313 15:41:40.950345 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/81ef3cd7-e06f-49f8-bc01-41b32fbb5546-config-data" (OuterVolumeSpecName: "config-data") pod "81ef3cd7-e06f-49f8-bc01-41b32fbb5546" (UID: "81ef3cd7-e06f-49f8-bc01-41b32fbb5546"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:41:40 crc kubenswrapper[4907]: I0313 15:41:40.965033 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/81ef3cd7-e06f-49f8-bc01-41b32fbb5546-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "81ef3cd7-e06f-49f8-bc01-41b32fbb5546" (UID: "81ef3cd7-e06f-49f8-bc01-41b32fbb5546"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:41:41 crc kubenswrapper[4907]: I0313 15:41:41.018236 4907 scope.go:117] "RemoveContainer" containerID="da08a8ab4ad1ce90fd28c0cf97955538dcb49b21c61a30f6afb435225b153254" Mar 13 15:41:41 crc kubenswrapper[4907]: E0313 15:41:41.019074 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"da08a8ab4ad1ce90fd28c0cf97955538dcb49b21c61a30f6afb435225b153254\": container with ID starting with da08a8ab4ad1ce90fd28c0cf97955538dcb49b21c61a30f6afb435225b153254 not found: ID does not exist" containerID="da08a8ab4ad1ce90fd28c0cf97955538dcb49b21c61a30f6afb435225b153254" Mar 13 15:41:41 crc kubenswrapper[4907]: I0313 15:41:41.019120 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da08a8ab4ad1ce90fd28c0cf97955538dcb49b21c61a30f6afb435225b153254"} err="failed to get container status \"da08a8ab4ad1ce90fd28c0cf97955538dcb49b21c61a30f6afb435225b153254\": rpc error: code = NotFound desc = could not find container \"da08a8ab4ad1ce90fd28c0cf97955538dcb49b21c61a30f6afb435225b153254\": container with ID starting with da08a8ab4ad1ce90fd28c0cf97955538dcb49b21c61a30f6afb435225b153254 not found: ID does not exist" Mar 13 15:41:41 crc kubenswrapper[4907]: I0313 15:41:41.019149 4907 scope.go:117] "RemoveContainer" containerID="697da8faed8339f44e9c8f069a46c98c95d40fcca3ba8965d84a7f51ddc0a358" Mar 13 15:41:41 crc kubenswrapper[4907]: E0313 15:41:41.020544 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"697da8faed8339f44e9c8f069a46c98c95d40fcca3ba8965d84a7f51ddc0a358\": container with ID starting with 697da8faed8339f44e9c8f069a46c98c95d40fcca3ba8965d84a7f51ddc0a358 not found: ID does not exist" containerID="697da8faed8339f44e9c8f069a46c98c95d40fcca3ba8965d84a7f51ddc0a358" Mar 13 15:41:41 crc kubenswrapper[4907]: I0313 15:41:41.020578 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"697da8faed8339f44e9c8f069a46c98c95d40fcca3ba8965d84a7f51ddc0a358"} err="failed to get container status \"697da8faed8339f44e9c8f069a46c98c95d40fcca3ba8965d84a7f51ddc0a358\": rpc error: code = NotFound desc = could not find container \"697da8faed8339f44e9c8f069a46c98c95d40fcca3ba8965d84a7f51ddc0a358\": container with ID starting with 697da8faed8339f44e9c8f069a46c98c95d40fcca3ba8965d84a7f51ddc0a358 not found: ID does not exist" Mar 13 15:41:41 crc kubenswrapper[4907]: I0313 15:41:41.022384 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37198210-23ed-4025-8cd9-e284c6573318-combined-ca-bundle\") pod \"37198210-23ed-4025-8cd9-e284c6573318\" (UID: \"37198210-23ed-4025-8cd9-e284c6573318\") " Mar 13 15:41:41 crc kubenswrapper[4907]: I0313 15:41:41.022544 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bdf6j\" (UniqueName: \"kubernetes.io/projected/37198210-23ed-4025-8cd9-e284c6573318-kube-api-access-bdf6j\") pod \"37198210-23ed-4025-8cd9-e284c6573318\" (UID: \"37198210-23ed-4025-8cd9-e284c6573318\") " Mar 13 15:41:41 crc kubenswrapper[4907]: I0313 15:41:41.022587 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37198210-23ed-4025-8cd9-e284c6573318-config-data\") pod \"37198210-23ed-4025-8cd9-e284c6573318\" (UID: \"37198210-23ed-4025-8cd9-e284c6573318\") " Mar 13 15:41:41 crc kubenswrapper[4907]: I0313 15:41:41.022703 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/37198210-23ed-4025-8cd9-e284c6573318-logs\") pod \"37198210-23ed-4025-8cd9-e284c6573318\" (UID: \"37198210-23ed-4025-8cd9-e284c6573318\") " Mar 13 15:41:41 crc kubenswrapper[4907]: I0313 15:41:41.023221 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81ef3cd7-e06f-49f8-bc01-41b32fbb5546-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 15:41:41 crc kubenswrapper[4907]: I0313 15:41:41.023247 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xct78\" (UniqueName: \"kubernetes.io/projected/81ef3cd7-e06f-49f8-bc01-41b32fbb5546-kube-api-access-xct78\") on node \"crc\" DevicePath \"\"" Mar 13 15:41:41 crc kubenswrapper[4907]: I0313 15:41:41.023260 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81ef3cd7-e06f-49f8-bc01-41b32fbb5546-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 15:41:41 crc kubenswrapper[4907]: I0313 15:41:41.023271 4907 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/81ef3cd7-e06f-49f8-bc01-41b32fbb5546-logs\") on node \"crc\" DevicePath \"\"" Mar 13 15:41:41 crc kubenswrapper[4907]: I0313 15:41:41.025106 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/37198210-23ed-4025-8cd9-e284c6573318-logs" (OuterVolumeSpecName: "logs") pod "37198210-23ed-4025-8cd9-e284c6573318" (UID: "37198210-23ed-4025-8cd9-e284c6573318"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:41:41 crc kubenswrapper[4907]: I0313 15:41:41.031160 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/37198210-23ed-4025-8cd9-e284c6573318-kube-api-access-bdf6j" (OuterVolumeSpecName: "kube-api-access-bdf6j") pod "37198210-23ed-4025-8cd9-e284c6573318" (UID: "37198210-23ed-4025-8cd9-e284c6573318"). InnerVolumeSpecName "kube-api-access-bdf6j". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:41:41 crc kubenswrapper[4907]: I0313 15:41:41.042414 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37198210-23ed-4025-8cd9-e284c6573318-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "37198210-23ed-4025-8cd9-e284c6573318" (UID: "37198210-23ed-4025-8cd9-e284c6573318"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:41:41 crc kubenswrapper[4907]: I0313 15:41:41.054094 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37198210-23ed-4025-8cd9-e284c6573318-config-data" (OuterVolumeSpecName: "config-data") pod "37198210-23ed-4025-8cd9-e284c6573318" (UID: "37198210-23ed-4025-8cd9-e284c6573318"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:41:41 crc kubenswrapper[4907]: E0313 15:41:41.072025 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="4efeb4b84a961f5a72c12fc7db6d3a11b565d9a8fc5e00e1bd6bd862f0227c34" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Mar 13 15:41:41 crc kubenswrapper[4907]: E0313 15:41:41.073444 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="4efeb4b84a961f5a72c12fc7db6d3a11b565d9a8fc5e00e1bd6bd862f0227c34" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Mar 13 15:41:41 crc kubenswrapper[4907]: E0313 15:41:41.079137 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="4efeb4b84a961f5a72c12fc7db6d3a11b565d9a8fc5e00e1bd6bd862f0227c34" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Mar 13 15:41:41 crc kubenswrapper[4907]: E0313 15:41:41.079200 4907 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell0-conductor-0" podUID="a8047522-cc05-44a3-b57d-2362e6457d49" containerName="nova-cell0-conductor-conductor" Mar 13 15:41:41 crc kubenswrapper[4907]: I0313 15:41:41.125154 4907 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/37198210-23ed-4025-8cd9-e284c6573318-logs\") on node \"crc\" DevicePath \"\"" Mar 13 15:41:41 crc kubenswrapper[4907]: I0313 15:41:41.125196 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37198210-23ed-4025-8cd9-e284c6573318-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 15:41:41 crc kubenswrapper[4907]: I0313 15:41:41.125211 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bdf6j\" (UniqueName: \"kubernetes.io/projected/37198210-23ed-4025-8cd9-e284c6573318-kube-api-access-bdf6j\") on node \"crc\" DevicePath \"\"" Mar 13 15:41:41 crc kubenswrapper[4907]: I0313 15:41:41.125223 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37198210-23ed-4025-8cd9-e284c6573318-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 15:41:41 crc kubenswrapper[4907]: I0313 15:41:41.833280 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 13 15:41:41 crc kubenswrapper[4907]: I0313 15:41:41.834679 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"37198210-23ed-4025-8cd9-e284c6573318","Type":"ContainerDied","Data":"54a994af9272c8e084ae758a74263408bf4549d1118439e6fd32c7e4d5aaa9e5"} Mar 13 15:41:41 crc kubenswrapper[4907]: I0313 15:41:41.834723 4907 scope.go:117] "RemoveContainer" containerID="ae907961f503a65c4dfb28c60282669570b0349aea1f6830f2e1f6e44d773cdb" Mar 13 15:41:41 crc kubenswrapper[4907]: I0313 15:41:41.834844 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 13 15:41:41 crc kubenswrapper[4907]: I0313 15:41:41.843332 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"eded65eb-7a04-4f90-ad82-be65db9014eb","Type":"ContainerStarted","Data":"e6d65819607c1b5f8189419bb95b4c9b19ca06cb630c59aa03a5474fcce9aa3d"} Mar 13 15:41:41 crc kubenswrapper[4907]: I0313 15:41:41.843537 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Mar 13 15:41:41 crc kubenswrapper[4907]: I0313 15:41:41.869614 4907 scope.go:117] "RemoveContainer" containerID="6dcdbcfacbeb1b7e6dcbc3fb09e947521ce095a95f3551396c6c38c4e6a8b77f" Mar 13 15:41:41 crc kubenswrapper[4907]: I0313 15:41:41.875775 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=3.875751232 podStartE2EDuration="3.875751232s" podCreationTimestamp="2026-03-13 15:41:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:41:41.863245761 +0000 UTC m=+5800.763033440" watchObservedRunningTime="2026-03-13 15:41:41.875751232 +0000 UTC m=+5800.775538921" Mar 13 15:41:41 crc kubenswrapper[4907]: I0313 15:41:41.891750 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Mar 13 15:41:41 crc kubenswrapper[4907]: I0313 15:41:41.918953 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Mar 13 15:41:41 crc kubenswrapper[4907]: I0313 15:41:41.929172 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Mar 13 15:41:41 crc kubenswrapper[4907]: I0313 15:41:41.953952 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Mar 13 15:41:41 crc kubenswrapper[4907]: E0313 15:41:41.954332 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37198210-23ed-4025-8cd9-e284c6573318" containerName="nova-metadata-metadata" Mar 13 15:41:41 crc kubenswrapper[4907]: I0313 15:41:41.954348 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="37198210-23ed-4025-8cd9-e284c6573318" containerName="nova-metadata-metadata" Mar 13 15:41:41 crc kubenswrapper[4907]: E0313 15:41:41.954369 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81ef3cd7-e06f-49f8-bc01-41b32fbb5546" containerName="nova-api-api" Mar 13 15:41:41 crc kubenswrapper[4907]: I0313 15:41:41.954376 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="81ef3cd7-e06f-49f8-bc01-41b32fbb5546" containerName="nova-api-api" Mar 13 15:41:41 crc kubenswrapper[4907]: E0313 15:41:41.954397 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81ef3cd7-e06f-49f8-bc01-41b32fbb5546" containerName="nova-api-log" Mar 13 15:41:41 crc kubenswrapper[4907]: I0313 15:41:41.954404 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="81ef3cd7-e06f-49f8-bc01-41b32fbb5546" containerName="nova-api-log" Mar 13 15:41:41 crc kubenswrapper[4907]: E0313 15:41:41.954412 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37198210-23ed-4025-8cd9-e284c6573318" containerName="nova-metadata-log" Mar 13 15:41:41 crc kubenswrapper[4907]: I0313 15:41:41.954419 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="37198210-23ed-4025-8cd9-e284c6573318" containerName="nova-metadata-log" Mar 13 15:41:41 crc kubenswrapper[4907]: I0313 15:41:41.954579 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="81ef3cd7-e06f-49f8-bc01-41b32fbb5546" containerName="nova-api-log" Mar 13 15:41:41 crc kubenswrapper[4907]: I0313 15:41:41.954593 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="37198210-23ed-4025-8cd9-e284c6573318" containerName="nova-metadata-metadata" Mar 13 15:41:41 crc kubenswrapper[4907]: I0313 15:41:41.954604 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="37198210-23ed-4025-8cd9-e284c6573318" containerName="nova-metadata-log" Mar 13 15:41:41 crc kubenswrapper[4907]: I0313 15:41:41.954614 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="81ef3cd7-e06f-49f8-bc01-41b32fbb5546" containerName="nova-api-api" Mar 13 15:41:41 crc kubenswrapper[4907]: I0313 15:41:41.955480 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 13 15:41:41 crc kubenswrapper[4907]: I0313 15:41:41.961268 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Mar 13 15:41:41 crc kubenswrapper[4907]: I0313 15:41:41.967990 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Mar 13 15:41:41 crc kubenswrapper[4907]: I0313 15:41:41.979102 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 13 15:41:42 crc kubenswrapper[4907]: I0313 15:41:42.006300 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Mar 13 15:41:42 crc kubenswrapper[4907]: I0313 15:41:42.008472 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 13 15:41:42 crc kubenswrapper[4907]: I0313 15:41:42.011297 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Mar 13 15:41:42 crc kubenswrapper[4907]: I0313 15:41:42.018941 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Mar 13 15:41:42 crc kubenswrapper[4907]: I0313 15:41:42.041213 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b98ee93f-f32c-4199-8045-62ab95761a6a-logs\") pod \"nova-api-0\" (UID: \"b98ee93f-f32c-4199-8045-62ab95761a6a\") " pod="openstack/nova-api-0" Mar 13 15:41:42 crc kubenswrapper[4907]: I0313 15:41:42.041371 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b98ee93f-f32c-4199-8045-62ab95761a6a-config-data\") pod \"nova-api-0\" (UID: \"b98ee93f-f32c-4199-8045-62ab95761a6a\") " pod="openstack/nova-api-0" Mar 13 15:41:42 crc kubenswrapper[4907]: I0313 15:41:42.041404 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t2t6r\" (UniqueName: \"kubernetes.io/projected/b98ee93f-f32c-4199-8045-62ab95761a6a-kube-api-access-t2t6r\") pod \"nova-api-0\" (UID: \"b98ee93f-f32c-4199-8045-62ab95761a6a\") " pod="openstack/nova-api-0" Mar 13 15:41:42 crc kubenswrapper[4907]: I0313 15:41:42.041455 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07f475b7-1542-4689-b3c1-f7f4470d4a67-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"07f475b7-1542-4689-b3c1-f7f4470d4a67\") " pod="openstack/nova-metadata-0" Mar 13 15:41:42 crc kubenswrapper[4907]: I0313 15:41:42.041491 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b98ee93f-f32c-4199-8045-62ab95761a6a-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"b98ee93f-f32c-4199-8045-62ab95761a6a\") " pod="openstack/nova-api-0" Mar 13 15:41:42 crc kubenswrapper[4907]: I0313 15:41:42.041516 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/07f475b7-1542-4689-b3c1-f7f4470d4a67-logs\") pod \"nova-metadata-0\" (UID: \"07f475b7-1542-4689-b3c1-f7f4470d4a67\") " pod="openstack/nova-metadata-0" Mar 13 15:41:42 crc kubenswrapper[4907]: I0313 15:41:42.041560 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/07f475b7-1542-4689-b3c1-f7f4470d4a67-config-data\") pod \"nova-metadata-0\" (UID: \"07f475b7-1542-4689-b3c1-f7f4470d4a67\") " pod="openstack/nova-metadata-0" Mar 13 15:41:42 crc kubenswrapper[4907]: I0313 15:41:42.041590 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7fwbz\" (UniqueName: \"kubernetes.io/projected/07f475b7-1542-4689-b3c1-f7f4470d4a67-kube-api-access-7fwbz\") pod \"nova-metadata-0\" (UID: \"07f475b7-1542-4689-b3c1-f7f4470d4a67\") " pod="openstack/nova-metadata-0" Mar 13 15:41:42 crc kubenswrapper[4907]: I0313 15:41:42.143659 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/07f475b7-1542-4689-b3c1-f7f4470d4a67-config-data\") pod \"nova-metadata-0\" (UID: \"07f475b7-1542-4689-b3c1-f7f4470d4a67\") " pod="openstack/nova-metadata-0" Mar 13 15:41:42 crc kubenswrapper[4907]: I0313 15:41:42.143721 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7fwbz\" (UniqueName: \"kubernetes.io/projected/07f475b7-1542-4689-b3c1-f7f4470d4a67-kube-api-access-7fwbz\") pod \"nova-metadata-0\" (UID: \"07f475b7-1542-4689-b3c1-f7f4470d4a67\") " pod="openstack/nova-metadata-0" Mar 13 15:41:42 crc kubenswrapper[4907]: I0313 15:41:42.143776 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b98ee93f-f32c-4199-8045-62ab95761a6a-logs\") pod \"nova-api-0\" (UID: \"b98ee93f-f32c-4199-8045-62ab95761a6a\") " pod="openstack/nova-api-0" Mar 13 15:41:42 crc kubenswrapper[4907]: I0313 15:41:42.143870 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b98ee93f-f32c-4199-8045-62ab95761a6a-config-data\") pod \"nova-api-0\" (UID: \"b98ee93f-f32c-4199-8045-62ab95761a6a\") " pod="openstack/nova-api-0" Mar 13 15:41:42 crc kubenswrapper[4907]: I0313 15:41:42.143913 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t2t6r\" (UniqueName: \"kubernetes.io/projected/b98ee93f-f32c-4199-8045-62ab95761a6a-kube-api-access-t2t6r\") pod \"nova-api-0\" (UID: \"b98ee93f-f32c-4199-8045-62ab95761a6a\") " pod="openstack/nova-api-0" Mar 13 15:41:42 crc kubenswrapper[4907]: I0313 15:41:42.143960 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07f475b7-1542-4689-b3c1-f7f4470d4a67-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"07f475b7-1542-4689-b3c1-f7f4470d4a67\") " pod="openstack/nova-metadata-0" Mar 13 15:41:42 crc kubenswrapper[4907]: I0313 15:41:42.143996 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b98ee93f-f32c-4199-8045-62ab95761a6a-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"b98ee93f-f32c-4199-8045-62ab95761a6a\") " pod="openstack/nova-api-0" Mar 13 15:41:42 crc kubenswrapper[4907]: I0313 15:41:42.144019 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/07f475b7-1542-4689-b3c1-f7f4470d4a67-logs\") pod \"nova-metadata-0\" (UID: \"07f475b7-1542-4689-b3c1-f7f4470d4a67\") " pod="openstack/nova-metadata-0" Mar 13 15:41:42 crc kubenswrapper[4907]: I0313 15:41:42.144443 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/07f475b7-1542-4689-b3c1-f7f4470d4a67-logs\") pod \"nova-metadata-0\" (UID: \"07f475b7-1542-4689-b3c1-f7f4470d4a67\") " pod="openstack/nova-metadata-0" Mar 13 15:41:42 crc kubenswrapper[4907]: I0313 15:41:42.144841 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b98ee93f-f32c-4199-8045-62ab95761a6a-logs\") pod \"nova-api-0\" (UID: \"b98ee93f-f32c-4199-8045-62ab95761a6a\") " pod="openstack/nova-api-0" Mar 13 15:41:42 crc kubenswrapper[4907]: I0313 15:41:42.148620 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07f475b7-1542-4689-b3c1-f7f4470d4a67-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"07f475b7-1542-4689-b3c1-f7f4470d4a67\") " pod="openstack/nova-metadata-0" Mar 13 15:41:42 crc kubenswrapper[4907]: I0313 15:41:42.149658 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/07f475b7-1542-4689-b3c1-f7f4470d4a67-config-data\") pod \"nova-metadata-0\" (UID: \"07f475b7-1542-4689-b3c1-f7f4470d4a67\") " pod="openstack/nova-metadata-0" Mar 13 15:41:42 crc kubenswrapper[4907]: I0313 15:41:42.169672 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t2t6r\" (UniqueName: \"kubernetes.io/projected/b98ee93f-f32c-4199-8045-62ab95761a6a-kube-api-access-t2t6r\") pod \"nova-api-0\" (UID: \"b98ee93f-f32c-4199-8045-62ab95761a6a\") " pod="openstack/nova-api-0" Mar 13 15:41:42 crc kubenswrapper[4907]: I0313 15:41:42.173626 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b98ee93f-f32c-4199-8045-62ab95761a6a-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"b98ee93f-f32c-4199-8045-62ab95761a6a\") " pod="openstack/nova-api-0" Mar 13 15:41:42 crc kubenswrapper[4907]: I0313 15:41:42.173766 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b98ee93f-f32c-4199-8045-62ab95761a6a-config-data\") pod \"nova-api-0\" (UID: \"b98ee93f-f32c-4199-8045-62ab95761a6a\") " pod="openstack/nova-api-0" Mar 13 15:41:42 crc kubenswrapper[4907]: I0313 15:41:42.207520 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7fwbz\" (UniqueName: \"kubernetes.io/projected/07f475b7-1542-4689-b3c1-f7f4470d4a67-kube-api-access-7fwbz\") pod \"nova-metadata-0\" (UID: \"07f475b7-1542-4689-b3c1-f7f4470d4a67\") " pod="openstack/nova-metadata-0" Mar 13 15:41:42 crc kubenswrapper[4907]: I0313 15:41:42.274259 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 13 15:41:42 crc kubenswrapper[4907]: I0313 15:41:42.327898 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 13 15:41:42 crc kubenswrapper[4907]: I0313 15:41:42.824949 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Mar 13 15:41:42 crc kubenswrapper[4907]: I0313 15:41:42.855512 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"07f475b7-1542-4689-b3c1-f7f4470d4a67","Type":"ContainerStarted","Data":"55a13d8a20cd160563d8a2c5f472cc7693d88cf83045ad53675342ed80dab573"} Mar 13 15:41:42 crc kubenswrapper[4907]: I0313 15:41:42.965354 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 13 15:41:43 crc kubenswrapper[4907]: E0313 15:41:43.523131 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="ce84852f1d320a52f408594e4b89f2cbf753717a67ce179872f364593d90c469" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Mar 13 15:41:43 crc kubenswrapper[4907]: E0313 15:41:43.524534 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="ce84852f1d320a52f408594e4b89f2cbf753717a67ce179872f364593d90c469" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Mar 13 15:41:43 crc kubenswrapper[4907]: E0313 15:41:43.529640 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="ce84852f1d320a52f408594e4b89f2cbf753717a67ce179872f364593d90c469" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Mar 13 15:41:43 crc kubenswrapper[4907]: E0313 15:41:43.529705 4907 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="61a2ddd9-1fcc-427e-aaa2-f092ebb7bd30" containerName="nova-scheduler-scheduler" Mar 13 15:41:43 crc kubenswrapper[4907]: I0313 15:41:43.794610 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="37198210-23ed-4025-8cd9-e284c6573318" path="/var/lib/kubelet/pods/37198210-23ed-4025-8cd9-e284c6573318/volumes" Mar 13 15:41:43 crc kubenswrapper[4907]: I0313 15:41:43.795305 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="81ef3cd7-e06f-49f8-bc01-41b32fbb5546" path="/var/lib/kubelet/pods/81ef3cd7-e06f-49f8-bc01-41b32fbb5546/volumes" Mar 13 15:41:43 crc kubenswrapper[4907]: I0313 15:41:43.865039 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b98ee93f-f32c-4199-8045-62ab95761a6a","Type":"ContainerStarted","Data":"19d030a6fb6fc190b4d45542a5b426271778e2fd40eba0953eec528508e6f63a"} Mar 13 15:41:43 crc kubenswrapper[4907]: I0313 15:41:43.865109 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b98ee93f-f32c-4199-8045-62ab95761a6a","Type":"ContainerStarted","Data":"fa43cc779c11f32664399bd6d5fb18c420f2b54971c3216eb69f8689384ae861"} Mar 13 15:41:43 crc kubenswrapper[4907]: I0313 15:41:43.865507 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b98ee93f-f32c-4199-8045-62ab95761a6a","Type":"ContainerStarted","Data":"6c6eae74def50447208be0f6d41599241acef5069efd81dc6d9fc3a28d2dd298"} Mar 13 15:41:43 crc kubenswrapper[4907]: I0313 15:41:43.867523 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"07f475b7-1542-4689-b3c1-f7f4470d4a67","Type":"ContainerStarted","Data":"047fde4a00c176dcd1e1415fa32aa7d47b1a371c3c13b528d5f7c550e42353f4"} Mar 13 15:41:43 crc kubenswrapper[4907]: I0313 15:41:43.867554 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"07f475b7-1542-4689-b3c1-f7f4470d4a67","Type":"ContainerStarted","Data":"c29e40d545e0c96f28dbfbc899c2ea7a2bf52224e0171c5a63ab9504fa49a30d"} Mar 13 15:41:43 crc kubenswrapper[4907]: I0313 15:41:43.885007 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.884986958 podStartE2EDuration="2.884986958s" podCreationTimestamp="2026-03-13 15:41:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:41:43.883443906 +0000 UTC m=+5802.783231595" watchObservedRunningTime="2026-03-13 15:41:43.884986958 +0000 UTC m=+5802.784774647" Mar 13 15:41:43 crc kubenswrapper[4907]: I0313 15:41:43.922080 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.922056321 podStartE2EDuration="2.922056321s" podCreationTimestamp="2026-03-13 15:41:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:41:43.908513371 +0000 UTC m=+5802.808301070" watchObservedRunningTime="2026-03-13 15:41:43.922056321 +0000 UTC m=+5802.821844010" Mar 13 15:41:44 crc kubenswrapper[4907]: I0313 15:41:44.237606 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Mar 13 15:41:44 crc kubenswrapper[4907]: I0313 15:41:44.810544 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Mar 13 15:41:44 crc kubenswrapper[4907]: I0313 15:41:44.884516 4907 generic.go:334] "Generic (PLEG): container finished" podID="a8047522-cc05-44a3-b57d-2362e6457d49" containerID="4efeb4b84a961f5a72c12fc7db6d3a11b565d9a8fc5e00e1bd6bd862f0227c34" exitCode=0 Mar 13 15:41:44 crc kubenswrapper[4907]: I0313 15:41:44.884561 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"a8047522-cc05-44a3-b57d-2362e6457d49","Type":"ContainerDied","Data":"4efeb4b84a961f5a72c12fc7db6d3a11b565d9a8fc5e00e1bd6bd862f0227c34"} Mar 13 15:41:44 crc kubenswrapper[4907]: I0313 15:41:44.884587 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Mar 13 15:41:44 crc kubenswrapper[4907]: I0313 15:41:44.884602 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"a8047522-cc05-44a3-b57d-2362e6457d49","Type":"ContainerDied","Data":"78d4b76c3a1fd63fabbe50a18b7db60eba6e1e2f95e60b44b880d052c99e1044"} Mar 13 15:41:44 crc kubenswrapper[4907]: I0313 15:41:44.884624 4907 scope.go:117] "RemoveContainer" containerID="4efeb4b84a961f5a72c12fc7db6d3a11b565d9a8fc5e00e1bd6bd862f0227c34" Mar 13 15:41:44 crc kubenswrapper[4907]: I0313 15:41:44.913691 4907 scope.go:117] "RemoveContainer" containerID="4efeb4b84a961f5a72c12fc7db6d3a11b565d9a8fc5e00e1bd6bd862f0227c34" Mar 13 15:41:44 crc kubenswrapper[4907]: E0313 15:41:44.914233 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4efeb4b84a961f5a72c12fc7db6d3a11b565d9a8fc5e00e1bd6bd862f0227c34\": container with ID starting with 4efeb4b84a961f5a72c12fc7db6d3a11b565d9a8fc5e00e1bd6bd862f0227c34 not found: ID does not exist" containerID="4efeb4b84a961f5a72c12fc7db6d3a11b565d9a8fc5e00e1bd6bd862f0227c34" Mar 13 15:41:44 crc kubenswrapper[4907]: I0313 15:41:44.914263 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4efeb4b84a961f5a72c12fc7db6d3a11b565d9a8fc5e00e1bd6bd862f0227c34"} err="failed to get container status \"4efeb4b84a961f5a72c12fc7db6d3a11b565d9a8fc5e00e1bd6bd862f0227c34\": rpc error: code = NotFound desc = could not find container \"4efeb4b84a961f5a72c12fc7db6d3a11b565d9a8fc5e00e1bd6bd862f0227c34\": container with ID starting with 4efeb4b84a961f5a72c12fc7db6d3a11b565d9a8fc5e00e1bd6bd862f0227c34 not found: ID does not exist" Mar 13 15:41:44 crc kubenswrapper[4907]: I0313 15:41:44.997929 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8047522-cc05-44a3-b57d-2362e6457d49-config-data\") pod \"a8047522-cc05-44a3-b57d-2362e6457d49\" (UID: \"a8047522-cc05-44a3-b57d-2362e6457d49\") " Mar 13 15:41:44 crc kubenswrapper[4907]: I0313 15:41:44.999949 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8047522-cc05-44a3-b57d-2362e6457d49-combined-ca-bundle\") pod \"a8047522-cc05-44a3-b57d-2362e6457d49\" (UID: \"a8047522-cc05-44a3-b57d-2362e6457d49\") " Mar 13 15:41:45 crc kubenswrapper[4907]: I0313 15:41:45.000051 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-77rtk\" (UniqueName: \"kubernetes.io/projected/a8047522-cc05-44a3-b57d-2362e6457d49-kube-api-access-77rtk\") pod \"a8047522-cc05-44a3-b57d-2362e6457d49\" (UID: \"a8047522-cc05-44a3-b57d-2362e6457d49\") " Mar 13 15:41:45 crc kubenswrapper[4907]: I0313 15:41:45.004310 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a8047522-cc05-44a3-b57d-2362e6457d49-kube-api-access-77rtk" (OuterVolumeSpecName: "kube-api-access-77rtk") pod "a8047522-cc05-44a3-b57d-2362e6457d49" (UID: "a8047522-cc05-44a3-b57d-2362e6457d49"). InnerVolumeSpecName "kube-api-access-77rtk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:41:45 crc kubenswrapper[4907]: I0313 15:41:45.028553 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8047522-cc05-44a3-b57d-2362e6457d49-config-data" (OuterVolumeSpecName: "config-data") pod "a8047522-cc05-44a3-b57d-2362e6457d49" (UID: "a8047522-cc05-44a3-b57d-2362e6457d49"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:41:45 crc kubenswrapper[4907]: I0313 15:41:45.029091 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8047522-cc05-44a3-b57d-2362e6457d49-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a8047522-cc05-44a3-b57d-2362e6457d49" (UID: "a8047522-cc05-44a3-b57d-2362e6457d49"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:41:45 crc kubenswrapper[4907]: I0313 15:41:45.102517 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8047522-cc05-44a3-b57d-2362e6457d49-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 15:41:45 crc kubenswrapper[4907]: I0313 15:41:45.102838 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8047522-cc05-44a3-b57d-2362e6457d49-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 15:41:45 crc kubenswrapper[4907]: I0313 15:41:45.102849 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-77rtk\" (UniqueName: \"kubernetes.io/projected/a8047522-cc05-44a3-b57d-2362e6457d49-kube-api-access-77rtk\") on node \"crc\" DevicePath \"\"" Mar 13 15:41:45 crc kubenswrapper[4907]: I0313 15:41:45.221949 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Mar 13 15:41:45 crc kubenswrapper[4907]: I0313 15:41:45.235838 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-0"] Mar 13 15:41:45 crc kubenswrapper[4907]: I0313 15:41:45.245709 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Mar 13 15:41:45 crc kubenswrapper[4907]: E0313 15:41:45.246129 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8047522-cc05-44a3-b57d-2362e6457d49" containerName="nova-cell0-conductor-conductor" Mar 13 15:41:45 crc kubenswrapper[4907]: I0313 15:41:45.246147 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8047522-cc05-44a3-b57d-2362e6457d49" containerName="nova-cell0-conductor-conductor" Mar 13 15:41:45 crc kubenswrapper[4907]: I0313 15:41:45.246311 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="a8047522-cc05-44a3-b57d-2362e6457d49" containerName="nova-cell0-conductor-conductor" Mar 13 15:41:45 crc kubenswrapper[4907]: I0313 15:41:45.246969 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Mar 13 15:41:45 crc kubenswrapper[4907]: I0313 15:41:45.253219 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Mar 13 15:41:45 crc kubenswrapper[4907]: I0313 15:41:45.280691 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Mar 13 15:41:45 crc kubenswrapper[4907]: I0313 15:41:45.408261 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/647720e3-d21e-4f41-aeb8-d4e8f092cbb5-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"647720e3-d21e-4f41-aeb8-d4e8f092cbb5\") " pod="openstack/nova-cell0-conductor-0" Mar 13 15:41:45 crc kubenswrapper[4907]: I0313 15:41:45.409208 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s7nfv\" (UniqueName: \"kubernetes.io/projected/647720e3-d21e-4f41-aeb8-d4e8f092cbb5-kube-api-access-s7nfv\") pod \"nova-cell0-conductor-0\" (UID: \"647720e3-d21e-4f41-aeb8-d4e8f092cbb5\") " pod="openstack/nova-cell0-conductor-0" Mar 13 15:41:45 crc kubenswrapper[4907]: I0313 15:41:45.409264 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/647720e3-d21e-4f41-aeb8-d4e8f092cbb5-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"647720e3-d21e-4f41-aeb8-d4e8f092cbb5\") " pod="openstack/nova-cell0-conductor-0" Mar 13 15:41:45 crc kubenswrapper[4907]: I0313 15:41:45.510928 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s7nfv\" (UniqueName: \"kubernetes.io/projected/647720e3-d21e-4f41-aeb8-d4e8f092cbb5-kube-api-access-s7nfv\") pod \"nova-cell0-conductor-0\" (UID: \"647720e3-d21e-4f41-aeb8-d4e8f092cbb5\") " pod="openstack/nova-cell0-conductor-0" Mar 13 15:41:45 crc kubenswrapper[4907]: I0313 15:41:45.510981 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/647720e3-d21e-4f41-aeb8-d4e8f092cbb5-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"647720e3-d21e-4f41-aeb8-d4e8f092cbb5\") " pod="openstack/nova-cell0-conductor-0" Mar 13 15:41:45 crc kubenswrapper[4907]: I0313 15:41:45.511030 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/647720e3-d21e-4f41-aeb8-d4e8f092cbb5-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"647720e3-d21e-4f41-aeb8-d4e8f092cbb5\") " pod="openstack/nova-cell0-conductor-0" Mar 13 15:41:45 crc kubenswrapper[4907]: I0313 15:41:45.514951 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/647720e3-d21e-4f41-aeb8-d4e8f092cbb5-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"647720e3-d21e-4f41-aeb8-d4e8f092cbb5\") " pod="openstack/nova-cell0-conductor-0" Mar 13 15:41:45 crc kubenswrapper[4907]: I0313 15:41:45.519530 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/647720e3-d21e-4f41-aeb8-d4e8f092cbb5-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"647720e3-d21e-4f41-aeb8-d4e8f092cbb5\") " pod="openstack/nova-cell0-conductor-0" Mar 13 15:41:45 crc kubenswrapper[4907]: I0313 15:41:45.528252 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s7nfv\" (UniqueName: \"kubernetes.io/projected/647720e3-d21e-4f41-aeb8-d4e8f092cbb5-kube-api-access-s7nfv\") pod \"nova-cell0-conductor-0\" (UID: \"647720e3-d21e-4f41-aeb8-d4e8f092cbb5\") " pod="openstack/nova-cell0-conductor-0" Mar 13 15:41:45 crc kubenswrapper[4907]: I0313 15:41:45.609628 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Mar 13 15:41:45 crc kubenswrapper[4907]: I0313 15:41:45.782995 4907 scope.go:117] "RemoveContainer" containerID="68ffe92dd03078633ec89d7ef29da9bbdc4c5f04c956a1e8487e3f2e01c5d1c5" Mar 13 15:41:45 crc kubenswrapper[4907]: E0313 15:41:45.783541 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:41:45 crc kubenswrapper[4907]: I0313 15:41:45.812990 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a8047522-cc05-44a3-b57d-2362e6457d49" path="/var/lib/kubelet/pods/a8047522-cc05-44a3-b57d-2362e6457d49/volumes" Mar 13 15:41:46 crc kubenswrapper[4907]: I0313 15:41:46.169020 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Mar 13 15:41:46 crc kubenswrapper[4907]: I0313 15:41:46.907660 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"647720e3-d21e-4f41-aeb8-d4e8f092cbb5","Type":"ContainerStarted","Data":"c20788b6b1aecba8b32acff82d1977637a4b4a170211b700ca1e5ce58e45cffe"} Mar 13 15:41:46 crc kubenswrapper[4907]: I0313 15:41:46.908003 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"647720e3-d21e-4f41-aeb8-d4e8f092cbb5","Type":"ContainerStarted","Data":"b4d79e83e6b122d073a9e884515b0623866f06609025b076fcf8513ad8311abc"} Mar 13 15:41:46 crc kubenswrapper[4907]: I0313 15:41:46.908039 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Mar 13 15:41:46 crc kubenswrapper[4907]: I0313 15:41:46.933192 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=1.9331698400000001 podStartE2EDuration="1.93316984s" podCreationTimestamp="2026-03-13 15:41:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:41:46.927247718 +0000 UTC m=+5805.827035407" watchObservedRunningTime="2026-03-13 15:41:46.93316984 +0000 UTC m=+5805.832957539" Mar 13 15:41:47 crc kubenswrapper[4907]: I0313 15:41:47.556479 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 13 15:41:47 crc kubenswrapper[4907]: I0313 15:41:47.751019 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61a2ddd9-1fcc-427e-aaa2-f092ebb7bd30-combined-ca-bundle\") pod \"61a2ddd9-1fcc-427e-aaa2-f092ebb7bd30\" (UID: \"61a2ddd9-1fcc-427e-aaa2-f092ebb7bd30\") " Mar 13 15:41:47 crc kubenswrapper[4907]: I0313 15:41:47.751189 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p7hsl\" (UniqueName: \"kubernetes.io/projected/61a2ddd9-1fcc-427e-aaa2-f092ebb7bd30-kube-api-access-p7hsl\") pod \"61a2ddd9-1fcc-427e-aaa2-f092ebb7bd30\" (UID: \"61a2ddd9-1fcc-427e-aaa2-f092ebb7bd30\") " Mar 13 15:41:47 crc kubenswrapper[4907]: I0313 15:41:47.751310 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/61a2ddd9-1fcc-427e-aaa2-f092ebb7bd30-config-data\") pod \"61a2ddd9-1fcc-427e-aaa2-f092ebb7bd30\" (UID: \"61a2ddd9-1fcc-427e-aaa2-f092ebb7bd30\") " Mar 13 15:41:47 crc kubenswrapper[4907]: I0313 15:41:47.764130 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/61a2ddd9-1fcc-427e-aaa2-f092ebb7bd30-kube-api-access-p7hsl" (OuterVolumeSpecName: "kube-api-access-p7hsl") pod "61a2ddd9-1fcc-427e-aaa2-f092ebb7bd30" (UID: "61a2ddd9-1fcc-427e-aaa2-f092ebb7bd30"). InnerVolumeSpecName "kube-api-access-p7hsl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:41:47 crc kubenswrapper[4907]: I0313 15:41:47.776373 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/61a2ddd9-1fcc-427e-aaa2-f092ebb7bd30-config-data" (OuterVolumeSpecName: "config-data") pod "61a2ddd9-1fcc-427e-aaa2-f092ebb7bd30" (UID: "61a2ddd9-1fcc-427e-aaa2-f092ebb7bd30"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:41:47 crc kubenswrapper[4907]: I0313 15:41:47.778825 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/61a2ddd9-1fcc-427e-aaa2-f092ebb7bd30-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "61a2ddd9-1fcc-427e-aaa2-f092ebb7bd30" (UID: "61a2ddd9-1fcc-427e-aaa2-f092ebb7bd30"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:41:47 crc kubenswrapper[4907]: I0313 15:41:47.853027 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/61a2ddd9-1fcc-427e-aaa2-f092ebb7bd30-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 15:41:47 crc kubenswrapper[4907]: I0313 15:41:47.853065 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61a2ddd9-1fcc-427e-aaa2-f092ebb7bd30-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 15:41:47 crc kubenswrapper[4907]: I0313 15:41:47.853079 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p7hsl\" (UniqueName: \"kubernetes.io/projected/61a2ddd9-1fcc-427e-aaa2-f092ebb7bd30-kube-api-access-p7hsl\") on node \"crc\" DevicePath \"\"" Mar 13 15:41:47 crc kubenswrapper[4907]: I0313 15:41:47.915986 4907 generic.go:334] "Generic (PLEG): container finished" podID="61a2ddd9-1fcc-427e-aaa2-f092ebb7bd30" containerID="ce84852f1d320a52f408594e4b89f2cbf753717a67ce179872f364593d90c469" exitCode=0 Mar 13 15:41:47 crc kubenswrapper[4907]: I0313 15:41:47.916049 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 13 15:41:47 crc kubenswrapper[4907]: I0313 15:41:47.916049 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"61a2ddd9-1fcc-427e-aaa2-f092ebb7bd30","Type":"ContainerDied","Data":"ce84852f1d320a52f408594e4b89f2cbf753717a67ce179872f364593d90c469"} Mar 13 15:41:47 crc kubenswrapper[4907]: I0313 15:41:47.916113 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"61a2ddd9-1fcc-427e-aaa2-f092ebb7bd30","Type":"ContainerDied","Data":"605697270f0438aa6a3971d720e6b66130d7e9f9c333f5d331abb06ba749d725"} Mar 13 15:41:47 crc kubenswrapper[4907]: I0313 15:41:47.916142 4907 scope.go:117] "RemoveContainer" containerID="ce84852f1d320a52f408594e4b89f2cbf753717a67ce179872f364593d90c469" Mar 13 15:41:47 crc kubenswrapper[4907]: I0313 15:41:47.946178 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 15:41:47 crc kubenswrapper[4907]: I0313 15:41:47.947275 4907 scope.go:117] "RemoveContainer" containerID="ce84852f1d320a52f408594e4b89f2cbf753717a67ce179872f364593d90c469" Mar 13 15:41:47 crc kubenswrapper[4907]: E0313 15:41:47.947701 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ce84852f1d320a52f408594e4b89f2cbf753717a67ce179872f364593d90c469\": container with ID starting with ce84852f1d320a52f408594e4b89f2cbf753717a67ce179872f364593d90c469 not found: ID does not exist" containerID="ce84852f1d320a52f408594e4b89f2cbf753717a67ce179872f364593d90c469" Mar 13 15:41:47 crc kubenswrapper[4907]: I0313 15:41:47.947730 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ce84852f1d320a52f408594e4b89f2cbf753717a67ce179872f364593d90c469"} err="failed to get container status \"ce84852f1d320a52f408594e4b89f2cbf753717a67ce179872f364593d90c469\": rpc error: code = NotFound desc = could not find container \"ce84852f1d320a52f408594e4b89f2cbf753717a67ce179872f364593d90c469\": container with ID starting with ce84852f1d320a52f408594e4b89f2cbf753717a67ce179872f364593d90c469 not found: ID does not exist" Mar 13 15:41:47 crc kubenswrapper[4907]: I0313 15:41:47.962070 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 15:41:47 crc kubenswrapper[4907]: I0313 15:41:47.970028 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 15:41:47 crc kubenswrapper[4907]: E0313 15:41:47.970612 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61a2ddd9-1fcc-427e-aaa2-f092ebb7bd30" containerName="nova-scheduler-scheduler" Mar 13 15:41:47 crc kubenswrapper[4907]: I0313 15:41:47.970636 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="61a2ddd9-1fcc-427e-aaa2-f092ebb7bd30" containerName="nova-scheduler-scheduler" Mar 13 15:41:47 crc kubenswrapper[4907]: I0313 15:41:47.970919 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="61a2ddd9-1fcc-427e-aaa2-f092ebb7bd30" containerName="nova-scheduler-scheduler" Mar 13 15:41:47 crc kubenswrapper[4907]: I0313 15:41:47.971862 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 13 15:41:47 crc kubenswrapper[4907]: I0313 15:41:47.974182 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Mar 13 15:41:47 crc kubenswrapper[4907]: I0313 15:41:47.978700 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 15:41:48 crc kubenswrapper[4907]: I0313 15:41:48.162491 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/48adeb28-f850-4e7c-a546-2c0b32d1b16c-config-data\") pod \"nova-scheduler-0\" (UID: \"48adeb28-f850-4e7c-a546-2c0b32d1b16c\") " pod="openstack/nova-scheduler-0" Mar 13 15:41:48 crc kubenswrapper[4907]: I0313 15:41:48.162667 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/48adeb28-f850-4e7c-a546-2c0b32d1b16c-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"48adeb28-f850-4e7c-a546-2c0b32d1b16c\") " pod="openstack/nova-scheduler-0" Mar 13 15:41:48 crc kubenswrapper[4907]: I0313 15:41:48.162969 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bm6cq\" (UniqueName: \"kubernetes.io/projected/48adeb28-f850-4e7c-a546-2c0b32d1b16c-kube-api-access-bm6cq\") pod \"nova-scheduler-0\" (UID: \"48adeb28-f850-4e7c-a546-2c0b32d1b16c\") " pod="openstack/nova-scheduler-0" Mar 13 15:41:48 crc kubenswrapper[4907]: I0313 15:41:48.264460 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/48adeb28-f850-4e7c-a546-2c0b32d1b16c-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"48adeb28-f850-4e7c-a546-2c0b32d1b16c\") " pod="openstack/nova-scheduler-0" Mar 13 15:41:48 crc kubenswrapper[4907]: I0313 15:41:48.264809 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bm6cq\" (UniqueName: \"kubernetes.io/projected/48adeb28-f850-4e7c-a546-2c0b32d1b16c-kube-api-access-bm6cq\") pod \"nova-scheduler-0\" (UID: \"48adeb28-f850-4e7c-a546-2c0b32d1b16c\") " pod="openstack/nova-scheduler-0" Mar 13 15:41:48 crc kubenswrapper[4907]: I0313 15:41:48.264893 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/48adeb28-f850-4e7c-a546-2c0b32d1b16c-config-data\") pod \"nova-scheduler-0\" (UID: \"48adeb28-f850-4e7c-a546-2c0b32d1b16c\") " pod="openstack/nova-scheduler-0" Mar 13 15:41:48 crc kubenswrapper[4907]: I0313 15:41:48.268583 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/48adeb28-f850-4e7c-a546-2c0b32d1b16c-config-data\") pod \"nova-scheduler-0\" (UID: \"48adeb28-f850-4e7c-a546-2c0b32d1b16c\") " pod="openstack/nova-scheduler-0" Mar 13 15:41:48 crc kubenswrapper[4907]: I0313 15:41:48.275596 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/48adeb28-f850-4e7c-a546-2c0b32d1b16c-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"48adeb28-f850-4e7c-a546-2c0b32d1b16c\") " pod="openstack/nova-scheduler-0" Mar 13 15:41:48 crc kubenswrapper[4907]: I0313 15:41:48.297417 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bm6cq\" (UniqueName: \"kubernetes.io/projected/48adeb28-f850-4e7c-a546-2c0b32d1b16c-kube-api-access-bm6cq\") pod \"nova-scheduler-0\" (UID: \"48adeb28-f850-4e7c-a546-2c0b32d1b16c\") " pod="openstack/nova-scheduler-0" Mar 13 15:41:48 crc kubenswrapper[4907]: I0313 15:41:48.302424 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 13 15:41:48 crc kubenswrapper[4907]: I0313 15:41:48.795769 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 15:41:48 crc kubenswrapper[4907]: I0313 15:41:48.925220 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"48adeb28-f850-4e7c-a546-2c0b32d1b16c","Type":"ContainerStarted","Data":"bba3814d9581f3c03937b4a23ce7f341cbe05c058698e236d16b4dfa33146288"} Mar 13 15:41:49 crc kubenswrapper[4907]: I0313 15:41:49.237855 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Mar 13 15:41:49 crc kubenswrapper[4907]: I0313 15:41:49.248989 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Mar 13 15:41:49 crc kubenswrapper[4907]: I0313 15:41:49.298268 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Mar 13 15:41:49 crc kubenswrapper[4907]: I0313 15:41:49.792461 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="61a2ddd9-1fcc-427e-aaa2-f092ebb7bd30" path="/var/lib/kubelet/pods/61a2ddd9-1fcc-427e-aaa2-f092ebb7bd30/volumes" Mar 13 15:41:49 crc kubenswrapper[4907]: I0313 15:41:49.939028 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"48adeb28-f850-4e7c-a546-2c0b32d1b16c","Type":"ContainerStarted","Data":"5de9510392b1c0241095d60bc966293a382f722cf2d981565b650d57c091ce88"} Mar 13 15:41:49 crc kubenswrapper[4907]: I0313 15:41:49.950023 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Mar 13 15:41:49 crc kubenswrapper[4907]: I0313 15:41:49.965702 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.965687213 podStartE2EDuration="2.965687213s" podCreationTimestamp="2026-03-13 15:41:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:41:49.962342092 +0000 UTC m=+5808.862129771" watchObservedRunningTime="2026-03-13 15:41:49.965687213 +0000 UTC m=+5808.865474902" Mar 13 15:41:52 crc kubenswrapper[4907]: I0313 15:41:52.275170 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Mar 13 15:41:52 crc kubenswrapper[4907]: I0313 15:41:52.275519 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Mar 13 15:41:52 crc kubenswrapper[4907]: I0313 15:41:52.328942 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Mar 13 15:41:52 crc kubenswrapper[4907]: I0313 15:41:52.330504 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Mar 13 15:41:53 crc kubenswrapper[4907]: I0313 15:41:53.302941 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Mar 13 15:41:53 crc kubenswrapper[4907]: I0313 15:41:53.358183 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="b98ee93f-f32c-4199-8045-62ab95761a6a" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.1.123:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 15:41:53 crc kubenswrapper[4907]: I0313 15:41:53.440082 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="07f475b7-1542-4689-b3c1-f7f4470d4a67" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.1.124:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 15:41:53 crc kubenswrapper[4907]: I0313 15:41:53.440196 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="07f475b7-1542-4689-b3c1-f7f4470d4a67" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.1.124:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 15:41:53 crc kubenswrapper[4907]: I0313 15:41:53.440083 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="b98ee93f-f32c-4199-8045-62ab95761a6a" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.1.123:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 15:41:55 crc kubenswrapper[4907]: I0313 15:41:55.651769 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Mar 13 15:41:55 crc kubenswrapper[4907]: I0313 15:41:55.713046 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Mar 13 15:41:55 crc kubenswrapper[4907]: I0313 15:41:55.715349 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Mar 13 15:41:55 crc kubenswrapper[4907]: I0313 15:41:55.731270 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Mar 13 15:41:55 crc kubenswrapper[4907]: I0313 15:41:55.765830 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Mar 13 15:41:55 crc kubenswrapper[4907]: I0313 15:41:55.905206 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/bfe30c39-e9da-49f4-a9da-ebbce147b863-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"bfe30c39-e9da-49f4-a9da-ebbce147b863\") " pod="openstack/cinder-scheduler-0" Mar 13 15:41:55 crc kubenswrapper[4907]: I0313 15:41:55.905275 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d5t84\" (UniqueName: \"kubernetes.io/projected/bfe30c39-e9da-49f4-a9da-ebbce147b863-kube-api-access-d5t84\") pod \"cinder-scheduler-0\" (UID: \"bfe30c39-e9da-49f4-a9da-ebbce147b863\") " pod="openstack/cinder-scheduler-0" Mar 13 15:41:55 crc kubenswrapper[4907]: I0313 15:41:55.905313 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bfe30c39-e9da-49f4-a9da-ebbce147b863-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"bfe30c39-e9da-49f4-a9da-ebbce147b863\") " pod="openstack/cinder-scheduler-0" Mar 13 15:41:55 crc kubenswrapper[4907]: I0313 15:41:55.906006 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfe30c39-e9da-49f4-a9da-ebbce147b863-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"bfe30c39-e9da-49f4-a9da-ebbce147b863\") " pod="openstack/cinder-scheduler-0" Mar 13 15:41:55 crc kubenswrapper[4907]: I0313 15:41:55.906234 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bfe30c39-e9da-49f4-a9da-ebbce147b863-config-data\") pod \"cinder-scheduler-0\" (UID: \"bfe30c39-e9da-49f4-a9da-ebbce147b863\") " pod="openstack/cinder-scheduler-0" Mar 13 15:41:55 crc kubenswrapper[4907]: I0313 15:41:55.906380 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bfe30c39-e9da-49f4-a9da-ebbce147b863-scripts\") pod \"cinder-scheduler-0\" (UID: \"bfe30c39-e9da-49f4-a9da-ebbce147b863\") " pod="openstack/cinder-scheduler-0" Mar 13 15:41:56 crc kubenswrapper[4907]: I0313 15:41:56.008049 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bfe30c39-e9da-49f4-a9da-ebbce147b863-scripts\") pod \"cinder-scheduler-0\" (UID: \"bfe30c39-e9da-49f4-a9da-ebbce147b863\") " pod="openstack/cinder-scheduler-0" Mar 13 15:41:56 crc kubenswrapper[4907]: I0313 15:41:56.008176 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/bfe30c39-e9da-49f4-a9da-ebbce147b863-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"bfe30c39-e9da-49f4-a9da-ebbce147b863\") " pod="openstack/cinder-scheduler-0" Mar 13 15:41:56 crc kubenswrapper[4907]: I0313 15:41:56.008209 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d5t84\" (UniqueName: \"kubernetes.io/projected/bfe30c39-e9da-49f4-a9da-ebbce147b863-kube-api-access-d5t84\") pod \"cinder-scheduler-0\" (UID: \"bfe30c39-e9da-49f4-a9da-ebbce147b863\") " pod="openstack/cinder-scheduler-0" Mar 13 15:41:56 crc kubenswrapper[4907]: I0313 15:41:56.008255 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bfe30c39-e9da-49f4-a9da-ebbce147b863-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"bfe30c39-e9da-49f4-a9da-ebbce147b863\") " pod="openstack/cinder-scheduler-0" Mar 13 15:41:56 crc kubenswrapper[4907]: I0313 15:41:56.008369 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfe30c39-e9da-49f4-a9da-ebbce147b863-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"bfe30c39-e9da-49f4-a9da-ebbce147b863\") " pod="openstack/cinder-scheduler-0" Mar 13 15:41:56 crc kubenswrapper[4907]: I0313 15:41:56.008373 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/bfe30c39-e9da-49f4-a9da-ebbce147b863-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"bfe30c39-e9da-49f4-a9da-ebbce147b863\") " pod="openstack/cinder-scheduler-0" Mar 13 15:41:56 crc kubenswrapper[4907]: I0313 15:41:56.008423 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bfe30c39-e9da-49f4-a9da-ebbce147b863-config-data\") pod \"cinder-scheduler-0\" (UID: \"bfe30c39-e9da-49f4-a9da-ebbce147b863\") " pod="openstack/cinder-scheduler-0" Mar 13 15:41:56 crc kubenswrapper[4907]: I0313 15:41:56.013596 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bfe30c39-e9da-49f4-a9da-ebbce147b863-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"bfe30c39-e9da-49f4-a9da-ebbce147b863\") " pod="openstack/cinder-scheduler-0" Mar 13 15:41:56 crc kubenswrapper[4907]: I0313 15:41:56.015385 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfe30c39-e9da-49f4-a9da-ebbce147b863-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"bfe30c39-e9da-49f4-a9da-ebbce147b863\") " pod="openstack/cinder-scheduler-0" Mar 13 15:41:56 crc kubenswrapper[4907]: I0313 15:41:56.017370 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bfe30c39-e9da-49f4-a9da-ebbce147b863-config-data\") pod \"cinder-scheduler-0\" (UID: \"bfe30c39-e9da-49f4-a9da-ebbce147b863\") " pod="openstack/cinder-scheduler-0" Mar 13 15:41:56 crc kubenswrapper[4907]: I0313 15:41:56.027463 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bfe30c39-e9da-49f4-a9da-ebbce147b863-scripts\") pod \"cinder-scheduler-0\" (UID: \"bfe30c39-e9da-49f4-a9da-ebbce147b863\") " pod="openstack/cinder-scheduler-0" Mar 13 15:41:56 crc kubenswrapper[4907]: I0313 15:41:56.028534 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d5t84\" (UniqueName: \"kubernetes.io/projected/bfe30c39-e9da-49f4-a9da-ebbce147b863-kube-api-access-d5t84\") pod \"cinder-scheduler-0\" (UID: \"bfe30c39-e9da-49f4-a9da-ebbce147b863\") " pod="openstack/cinder-scheduler-0" Mar 13 15:41:56 crc kubenswrapper[4907]: I0313 15:41:56.063470 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Mar 13 15:41:56 crc kubenswrapper[4907]: I0313 15:41:56.667212 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Mar 13 15:41:56 crc kubenswrapper[4907]: W0313 15:41:56.672179 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbfe30c39_e9da_49f4_a9da_ebbce147b863.slice/crio-36d4f2be9f852fdbc463601ca568186a38921963adf0237624ff766e5451ebed WatchSource:0}: Error finding container 36d4f2be9f852fdbc463601ca568186a38921963adf0237624ff766e5451ebed: Status 404 returned error can't find the container with id 36d4f2be9f852fdbc463601ca568186a38921963adf0237624ff766e5451ebed Mar 13 15:41:56 crc kubenswrapper[4907]: I0313 15:41:56.782935 4907 scope.go:117] "RemoveContainer" containerID="68ffe92dd03078633ec89d7ef29da9bbdc4c5f04c956a1e8487e3f2e01c5d1c5" Mar 13 15:41:56 crc kubenswrapper[4907]: E0313 15:41:56.783553 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:41:56 crc kubenswrapper[4907]: I0313 15:41:56.988084 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Mar 13 15:41:56 crc kubenswrapper[4907]: I0313 15:41:56.988332 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="12616424-013a-47c8-9ed3-b407791537fe" containerName="cinder-api-log" containerID="cri-o://f234443780f86c06ddc613fb1859ac940ed1d9617b56539195ffdc071bbc882b" gracePeriod=30 Mar 13 15:41:56 crc kubenswrapper[4907]: I0313 15:41:56.988439 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="12616424-013a-47c8-9ed3-b407791537fe" containerName="cinder-api" containerID="cri-o://b5b4ab9a472db32a9c0ae91b2509ae28f1254ad915d9737b1279509e0c8d4717" gracePeriod=30 Mar 13 15:41:57 crc kubenswrapper[4907]: I0313 15:41:57.035710 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"bfe30c39-e9da-49f4-a9da-ebbce147b863","Type":"ContainerStarted","Data":"36d4f2be9f852fdbc463601ca568186a38921963adf0237624ff766e5451ebed"} Mar 13 15:41:57 crc kubenswrapper[4907]: I0313 15:41:57.477845 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-volume-volume1-0"] Mar 13 15:41:57 crc kubenswrapper[4907]: I0313 15:41:57.479759 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-volume-volume1-0" Mar 13 15:41:57 crc kubenswrapper[4907]: I0313 15:41:57.481519 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-volume-volume1-config-data" Mar 13 15:41:57 crc kubenswrapper[4907]: I0313 15:41:57.527759 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-volume-volume1-0"] Mar 13 15:41:57 crc kubenswrapper[4907]: I0313 15:41:57.553038 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/697762c8-b121-410a-a5db-21916ca1791f-sys\") pod \"cinder-volume-volume1-0\" (UID: \"697762c8-b121-410a-a5db-21916ca1791f\") " pod="openstack/cinder-volume-volume1-0" Mar 13 15:41:57 crc kubenswrapper[4907]: I0313 15:41:57.553091 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/697762c8-b121-410a-a5db-21916ca1791f-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"697762c8-b121-410a-a5db-21916ca1791f\") " pod="openstack/cinder-volume-volume1-0" Mar 13 15:41:57 crc kubenswrapper[4907]: I0313 15:41:57.553115 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/697762c8-b121-410a-a5db-21916ca1791f-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"697762c8-b121-410a-a5db-21916ca1791f\") " pod="openstack/cinder-volume-volume1-0" Mar 13 15:41:57 crc kubenswrapper[4907]: I0313 15:41:57.553256 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/697762c8-b121-410a-a5db-21916ca1791f-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"697762c8-b121-410a-a5db-21916ca1791f\") " pod="openstack/cinder-volume-volume1-0" Mar 13 15:41:57 crc kubenswrapper[4907]: I0313 15:41:57.553329 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/697762c8-b121-410a-a5db-21916ca1791f-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"697762c8-b121-410a-a5db-21916ca1791f\") " pod="openstack/cinder-volume-volume1-0" Mar 13 15:41:57 crc kubenswrapper[4907]: I0313 15:41:57.553368 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/697762c8-b121-410a-a5db-21916ca1791f-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"697762c8-b121-410a-a5db-21916ca1791f\") " pod="openstack/cinder-volume-volume1-0" Mar 13 15:41:57 crc kubenswrapper[4907]: I0313 15:41:57.553408 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/697762c8-b121-410a-a5db-21916ca1791f-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"697762c8-b121-410a-a5db-21916ca1791f\") " pod="openstack/cinder-volume-volume1-0" Mar 13 15:41:57 crc kubenswrapper[4907]: I0313 15:41:57.553507 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/697762c8-b121-410a-a5db-21916ca1791f-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"697762c8-b121-410a-a5db-21916ca1791f\") " pod="openstack/cinder-volume-volume1-0" Mar 13 15:41:57 crc kubenswrapper[4907]: I0313 15:41:57.553533 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/697762c8-b121-410a-a5db-21916ca1791f-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"697762c8-b121-410a-a5db-21916ca1791f\") " pod="openstack/cinder-volume-volume1-0" Mar 13 15:41:57 crc kubenswrapper[4907]: I0313 15:41:57.553593 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z4d56\" (UniqueName: \"kubernetes.io/projected/697762c8-b121-410a-a5db-21916ca1791f-kube-api-access-z4d56\") pod \"cinder-volume-volume1-0\" (UID: \"697762c8-b121-410a-a5db-21916ca1791f\") " pod="openstack/cinder-volume-volume1-0" Mar 13 15:41:57 crc kubenswrapper[4907]: I0313 15:41:57.553954 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/697762c8-b121-410a-a5db-21916ca1791f-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"697762c8-b121-410a-a5db-21916ca1791f\") " pod="openstack/cinder-volume-volume1-0" Mar 13 15:41:57 crc kubenswrapper[4907]: I0313 15:41:57.554012 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/697762c8-b121-410a-a5db-21916ca1791f-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"697762c8-b121-410a-a5db-21916ca1791f\") " pod="openstack/cinder-volume-volume1-0" Mar 13 15:41:57 crc kubenswrapper[4907]: I0313 15:41:57.554115 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/697762c8-b121-410a-a5db-21916ca1791f-run\") pod \"cinder-volume-volume1-0\" (UID: \"697762c8-b121-410a-a5db-21916ca1791f\") " pod="openstack/cinder-volume-volume1-0" Mar 13 15:41:57 crc kubenswrapper[4907]: I0313 15:41:57.554182 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/697762c8-b121-410a-a5db-21916ca1791f-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"697762c8-b121-410a-a5db-21916ca1791f\") " pod="openstack/cinder-volume-volume1-0" Mar 13 15:41:57 crc kubenswrapper[4907]: I0313 15:41:57.554240 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/697762c8-b121-410a-a5db-21916ca1791f-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"697762c8-b121-410a-a5db-21916ca1791f\") " pod="openstack/cinder-volume-volume1-0" Mar 13 15:41:57 crc kubenswrapper[4907]: I0313 15:41:57.554267 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/697762c8-b121-410a-a5db-21916ca1791f-dev\") pod \"cinder-volume-volume1-0\" (UID: \"697762c8-b121-410a-a5db-21916ca1791f\") " pod="openstack/cinder-volume-volume1-0" Mar 13 15:41:57 crc kubenswrapper[4907]: I0313 15:41:57.655512 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/697762c8-b121-410a-a5db-21916ca1791f-run\") pod \"cinder-volume-volume1-0\" (UID: \"697762c8-b121-410a-a5db-21916ca1791f\") " pod="openstack/cinder-volume-volume1-0" Mar 13 15:41:57 crc kubenswrapper[4907]: I0313 15:41:57.655574 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/697762c8-b121-410a-a5db-21916ca1791f-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"697762c8-b121-410a-a5db-21916ca1791f\") " pod="openstack/cinder-volume-volume1-0" Mar 13 15:41:57 crc kubenswrapper[4907]: I0313 15:41:57.655606 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/697762c8-b121-410a-a5db-21916ca1791f-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"697762c8-b121-410a-a5db-21916ca1791f\") " pod="openstack/cinder-volume-volume1-0" Mar 13 15:41:57 crc kubenswrapper[4907]: I0313 15:41:57.655629 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/697762c8-b121-410a-a5db-21916ca1791f-dev\") pod \"cinder-volume-volume1-0\" (UID: \"697762c8-b121-410a-a5db-21916ca1791f\") " pod="openstack/cinder-volume-volume1-0" Mar 13 15:41:57 crc kubenswrapper[4907]: I0313 15:41:57.655657 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/697762c8-b121-410a-a5db-21916ca1791f-sys\") pod \"cinder-volume-volume1-0\" (UID: \"697762c8-b121-410a-a5db-21916ca1791f\") " pod="openstack/cinder-volume-volume1-0" Mar 13 15:41:57 crc kubenswrapper[4907]: I0313 15:41:57.655687 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/697762c8-b121-410a-a5db-21916ca1791f-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"697762c8-b121-410a-a5db-21916ca1791f\") " pod="openstack/cinder-volume-volume1-0" Mar 13 15:41:57 crc kubenswrapper[4907]: I0313 15:41:57.655713 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/697762c8-b121-410a-a5db-21916ca1791f-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"697762c8-b121-410a-a5db-21916ca1791f\") " pod="openstack/cinder-volume-volume1-0" Mar 13 15:41:57 crc kubenswrapper[4907]: I0313 15:41:57.655709 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/697762c8-b121-410a-a5db-21916ca1791f-run\") pod \"cinder-volume-volume1-0\" (UID: \"697762c8-b121-410a-a5db-21916ca1791f\") " pod="openstack/cinder-volume-volume1-0" Mar 13 15:41:57 crc kubenswrapper[4907]: I0313 15:41:57.655746 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/697762c8-b121-410a-a5db-21916ca1791f-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"697762c8-b121-410a-a5db-21916ca1791f\") " pod="openstack/cinder-volume-volume1-0" Mar 13 15:41:57 crc kubenswrapper[4907]: I0313 15:41:57.655872 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/697762c8-b121-410a-a5db-21916ca1791f-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"697762c8-b121-410a-a5db-21916ca1791f\") " pod="openstack/cinder-volume-volume1-0" Mar 13 15:41:57 crc kubenswrapper[4907]: I0313 15:41:57.655959 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/697762c8-b121-410a-a5db-21916ca1791f-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"697762c8-b121-410a-a5db-21916ca1791f\") " pod="openstack/cinder-volume-volume1-0" Mar 13 15:41:57 crc kubenswrapper[4907]: I0313 15:41:57.656019 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/697762c8-b121-410a-a5db-21916ca1791f-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"697762c8-b121-410a-a5db-21916ca1791f\") " pod="openstack/cinder-volume-volume1-0" Mar 13 15:41:57 crc kubenswrapper[4907]: I0313 15:41:57.656215 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/697762c8-b121-410a-a5db-21916ca1791f-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"697762c8-b121-410a-a5db-21916ca1791f\") " pod="openstack/cinder-volume-volume1-0" Mar 13 15:41:57 crc kubenswrapper[4907]: I0313 15:41:57.656247 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/697762c8-b121-410a-a5db-21916ca1791f-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"697762c8-b121-410a-a5db-21916ca1791f\") " pod="openstack/cinder-volume-volume1-0" Mar 13 15:41:57 crc kubenswrapper[4907]: I0313 15:41:57.656299 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z4d56\" (UniqueName: \"kubernetes.io/projected/697762c8-b121-410a-a5db-21916ca1791f-kube-api-access-z4d56\") pod \"cinder-volume-volume1-0\" (UID: \"697762c8-b121-410a-a5db-21916ca1791f\") " pod="openstack/cinder-volume-volume1-0" Mar 13 15:41:57 crc kubenswrapper[4907]: I0313 15:41:57.656484 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/697762c8-b121-410a-a5db-21916ca1791f-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"697762c8-b121-410a-a5db-21916ca1791f\") " pod="openstack/cinder-volume-volume1-0" Mar 13 15:41:57 crc kubenswrapper[4907]: I0313 15:41:57.656541 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/697762c8-b121-410a-a5db-21916ca1791f-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"697762c8-b121-410a-a5db-21916ca1791f\") " pod="openstack/cinder-volume-volume1-0" Mar 13 15:41:57 crc kubenswrapper[4907]: I0313 15:41:57.656634 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/697762c8-b121-410a-a5db-21916ca1791f-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"697762c8-b121-410a-a5db-21916ca1791f\") " pod="openstack/cinder-volume-volume1-0" Mar 13 15:41:57 crc kubenswrapper[4907]: I0313 15:41:57.656778 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/697762c8-b121-410a-a5db-21916ca1791f-dev\") pod \"cinder-volume-volume1-0\" (UID: \"697762c8-b121-410a-a5db-21916ca1791f\") " pod="openstack/cinder-volume-volume1-0" Mar 13 15:41:57 crc kubenswrapper[4907]: I0313 15:41:57.657267 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/697762c8-b121-410a-a5db-21916ca1791f-sys\") pod \"cinder-volume-volume1-0\" (UID: \"697762c8-b121-410a-a5db-21916ca1791f\") " pod="openstack/cinder-volume-volume1-0" Mar 13 15:41:57 crc kubenswrapper[4907]: I0313 15:41:57.657392 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/697762c8-b121-410a-a5db-21916ca1791f-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"697762c8-b121-410a-a5db-21916ca1791f\") " pod="openstack/cinder-volume-volume1-0" Mar 13 15:41:57 crc kubenswrapper[4907]: I0313 15:41:57.657394 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/697762c8-b121-410a-a5db-21916ca1791f-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"697762c8-b121-410a-a5db-21916ca1791f\") " pod="openstack/cinder-volume-volume1-0" Mar 13 15:41:57 crc kubenswrapper[4907]: I0313 15:41:57.657552 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/697762c8-b121-410a-a5db-21916ca1791f-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"697762c8-b121-410a-a5db-21916ca1791f\") " pod="openstack/cinder-volume-volume1-0" Mar 13 15:41:57 crc kubenswrapper[4907]: I0313 15:41:57.657597 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/697762c8-b121-410a-a5db-21916ca1791f-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"697762c8-b121-410a-a5db-21916ca1791f\") " pod="openstack/cinder-volume-volume1-0" Mar 13 15:41:57 crc kubenswrapper[4907]: I0313 15:41:57.657683 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/697762c8-b121-410a-a5db-21916ca1791f-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"697762c8-b121-410a-a5db-21916ca1791f\") " pod="openstack/cinder-volume-volume1-0" Mar 13 15:41:57 crc kubenswrapper[4907]: I0313 15:41:57.658126 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/697762c8-b121-410a-a5db-21916ca1791f-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"697762c8-b121-410a-a5db-21916ca1791f\") " pod="openstack/cinder-volume-volume1-0" Mar 13 15:41:57 crc kubenswrapper[4907]: I0313 15:41:57.660444 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/697762c8-b121-410a-a5db-21916ca1791f-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"697762c8-b121-410a-a5db-21916ca1791f\") " pod="openstack/cinder-volume-volume1-0" Mar 13 15:41:57 crc kubenswrapper[4907]: I0313 15:41:57.660639 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/697762c8-b121-410a-a5db-21916ca1791f-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"697762c8-b121-410a-a5db-21916ca1791f\") " pod="openstack/cinder-volume-volume1-0" Mar 13 15:41:57 crc kubenswrapper[4907]: I0313 15:41:57.660678 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/697762c8-b121-410a-a5db-21916ca1791f-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"697762c8-b121-410a-a5db-21916ca1791f\") " pod="openstack/cinder-volume-volume1-0" Mar 13 15:41:57 crc kubenswrapper[4907]: I0313 15:41:57.660691 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/697762c8-b121-410a-a5db-21916ca1791f-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"697762c8-b121-410a-a5db-21916ca1791f\") " pod="openstack/cinder-volume-volume1-0" Mar 13 15:41:57 crc kubenswrapper[4907]: I0313 15:41:57.664378 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/697762c8-b121-410a-a5db-21916ca1791f-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"697762c8-b121-410a-a5db-21916ca1791f\") " pod="openstack/cinder-volume-volume1-0" Mar 13 15:41:57 crc kubenswrapper[4907]: I0313 15:41:57.674070 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z4d56\" (UniqueName: \"kubernetes.io/projected/697762c8-b121-410a-a5db-21916ca1791f-kube-api-access-z4d56\") pod \"cinder-volume-volume1-0\" (UID: \"697762c8-b121-410a-a5db-21916ca1791f\") " pod="openstack/cinder-volume-volume1-0" Mar 13 15:41:57 crc kubenswrapper[4907]: I0313 15:41:57.833987 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-volume-volume1-0" Mar 13 15:41:58 crc kubenswrapper[4907]: I0313 15:41:58.074091 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"bfe30c39-e9da-49f4-a9da-ebbce147b863","Type":"ContainerStarted","Data":"fbe67e4b2a6e6d29dd61f35ad336b329273792eab74bfcf4dd2ade78056599e6"} Mar 13 15:41:58 crc kubenswrapper[4907]: I0313 15:41:58.074698 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"bfe30c39-e9da-49f4-a9da-ebbce147b863","Type":"ContainerStarted","Data":"24af8c70da41dc6cb88e8d25d6b679b4b7c4b0986765337bf4ab3605a412ae27"} Mar 13 15:41:58 crc kubenswrapper[4907]: I0313 15:41:58.080579 4907 generic.go:334] "Generic (PLEG): container finished" podID="12616424-013a-47c8-9ed3-b407791537fe" containerID="f234443780f86c06ddc613fb1859ac940ed1d9617b56539195ffdc071bbc882b" exitCode=143 Mar 13 15:41:58 crc kubenswrapper[4907]: I0313 15:41:58.080633 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"12616424-013a-47c8-9ed3-b407791537fe","Type":"ContainerDied","Data":"f234443780f86c06ddc613fb1859ac940ed1d9617b56539195ffdc071bbc882b"} Mar 13 15:41:58 crc kubenswrapper[4907]: I0313 15:41:58.107183 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.107165302 podStartE2EDuration="3.107165302s" podCreationTimestamp="2026-03-13 15:41:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:41:58.100084059 +0000 UTC m=+5816.999871748" watchObservedRunningTime="2026-03-13 15:41:58.107165302 +0000 UTC m=+5817.006952991" Mar 13 15:41:58 crc kubenswrapper[4907]: I0313 15:41:58.302730 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Mar 13 15:41:58 crc kubenswrapper[4907]: I0313 15:41:58.327901 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-backup-0"] Mar 13 15:41:58 crc kubenswrapper[4907]: I0313 15:41:58.329700 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-backup-0" Mar 13 15:41:58 crc kubenswrapper[4907]: I0313 15:41:58.335999 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-backup-config-data" Mar 13 15:41:58 crc kubenswrapper[4907]: I0313 15:41:58.353554 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Mar 13 15:41:58 crc kubenswrapper[4907]: I0313 15:41:58.365835 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-backup-0"] Mar 13 15:41:58 crc kubenswrapper[4907]: I0313 15:41:58.373700 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/d92d5136-2169-44cc-9900-c3335a4b69eb-sys\") pod \"cinder-backup-0\" (UID: \"d92d5136-2169-44cc-9900-c3335a4b69eb\") " pod="openstack/cinder-backup-0" Mar 13 15:41:58 crc kubenswrapper[4907]: I0313 15:41:58.373774 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d92d5136-2169-44cc-9900-c3335a4b69eb-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"d92d5136-2169-44cc-9900-c3335a4b69eb\") " pod="openstack/cinder-backup-0" Mar 13 15:41:58 crc kubenswrapper[4907]: I0313 15:41:58.373843 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/d92d5136-2169-44cc-9900-c3335a4b69eb-run\") pod \"cinder-backup-0\" (UID: \"d92d5136-2169-44cc-9900-c3335a4b69eb\") " pod="openstack/cinder-backup-0" Mar 13 15:41:58 crc kubenswrapper[4907]: I0313 15:41:58.374052 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/d92d5136-2169-44cc-9900-c3335a4b69eb-dev\") pod \"cinder-backup-0\" (UID: \"d92d5136-2169-44cc-9900-c3335a4b69eb\") " pod="openstack/cinder-backup-0" Mar 13 15:41:58 crc kubenswrapper[4907]: I0313 15:41:58.374162 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/d92d5136-2169-44cc-9900-c3335a4b69eb-ceph\") pod \"cinder-backup-0\" (UID: \"d92d5136-2169-44cc-9900-c3335a4b69eb\") " pod="openstack/cinder-backup-0" Mar 13 15:41:58 crc kubenswrapper[4907]: I0313 15:41:58.374253 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kls8z\" (UniqueName: \"kubernetes.io/projected/d92d5136-2169-44cc-9900-c3335a4b69eb-kube-api-access-kls8z\") pod \"cinder-backup-0\" (UID: \"d92d5136-2169-44cc-9900-c3335a4b69eb\") " pod="openstack/cinder-backup-0" Mar 13 15:41:58 crc kubenswrapper[4907]: I0313 15:41:58.374332 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d92d5136-2169-44cc-9900-c3335a4b69eb-config-data\") pod \"cinder-backup-0\" (UID: \"d92d5136-2169-44cc-9900-c3335a4b69eb\") " pod="openstack/cinder-backup-0" Mar 13 15:41:58 crc kubenswrapper[4907]: I0313 15:41:58.374403 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d92d5136-2169-44cc-9900-c3335a4b69eb-scripts\") pod \"cinder-backup-0\" (UID: \"d92d5136-2169-44cc-9900-c3335a4b69eb\") " pod="openstack/cinder-backup-0" Mar 13 15:41:58 crc kubenswrapper[4907]: I0313 15:41:58.374485 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/d92d5136-2169-44cc-9900-c3335a4b69eb-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"d92d5136-2169-44cc-9900-c3335a4b69eb\") " pod="openstack/cinder-backup-0" Mar 13 15:41:58 crc kubenswrapper[4907]: I0313 15:41:58.374669 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d92d5136-2169-44cc-9900-c3335a4b69eb-config-data-custom\") pod \"cinder-backup-0\" (UID: \"d92d5136-2169-44cc-9900-c3335a4b69eb\") " pod="openstack/cinder-backup-0" Mar 13 15:41:58 crc kubenswrapper[4907]: I0313 15:41:58.374746 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/d92d5136-2169-44cc-9900-c3335a4b69eb-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"d92d5136-2169-44cc-9900-c3335a4b69eb\") " pod="openstack/cinder-backup-0" Mar 13 15:41:58 crc kubenswrapper[4907]: I0313 15:41:58.374819 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d92d5136-2169-44cc-9900-c3335a4b69eb-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"d92d5136-2169-44cc-9900-c3335a4b69eb\") " pod="openstack/cinder-backup-0" Mar 13 15:41:58 crc kubenswrapper[4907]: I0313 15:41:58.375132 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/d92d5136-2169-44cc-9900-c3335a4b69eb-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"d92d5136-2169-44cc-9900-c3335a4b69eb\") " pod="openstack/cinder-backup-0" Mar 13 15:41:58 crc kubenswrapper[4907]: I0313 15:41:58.375273 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/d92d5136-2169-44cc-9900-c3335a4b69eb-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"d92d5136-2169-44cc-9900-c3335a4b69eb\") " pod="openstack/cinder-backup-0" Mar 13 15:41:58 crc kubenswrapper[4907]: I0313 15:41:58.375349 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/d92d5136-2169-44cc-9900-c3335a4b69eb-lib-modules\") pod \"cinder-backup-0\" (UID: \"d92d5136-2169-44cc-9900-c3335a4b69eb\") " pod="openstack/cinder-backup-0" Mar 13 15:41:58 crc kubenswrapper[4907]: I0313 15:41:58.375412 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/d92d5136-2169-44cc-9900-c3335a4b69eb-etc-nvme\") pod \"cinder-backup-0\" (UID: \"d92d5136-2169-44cc-9900-c3335a4b69eb\") " pod="openstack/cinder-backup-0" Mar 13 15:41:58 crc kubenswrapper[4907]: I0313 15:41:58.478517 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kls8z\" (UniqueName: \"kubernetes.io/projected/d92d5136-2169-44cc-9900-c3335a4b69eb-kube-api-access-kls8z\") pod \"cinder-backup-0\" (UID: \"d92d5136-2169-44cc-9900-c3335a4b69eb\") " pod="openstack/cinder-backup-0" Mar 13 15:41:58 crc kubenswrapper[4907]: I0313 15:41:58.478560 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d92d5136-2169-44cc-9900-c3335a4b69eb-config-data\") pod \"cinder-backup-0\" (UID: \"d92d5136-2169-44cc-9900-c3335a4b69eb\") " pod="openstack/cinder-backup-0" Mar 13 15:41:58 crc kubenswrapper[4907]: I0313 15:41:58.478580 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d92d5136-2169-44cc-9900-c3335a4b69eb-scripts\") pod \"cinder-backup-0\" (UID: \"d92d5136-2169-44cc-9900-c3335a4b69eb\") " pod="openstack/cinder-backup-0" Mar 13 15:41:58 crc kubenswrapper[4907]: I0313 15:41:58.478601 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/d92d5136-2169-44cc-9900-c3335a4b69eb-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"d92d5136-2169-44cc-9900-c3335a4b69eb\") " pod="openstack/cinder-backup-0" Mar 13 15:41:58 crc kubenswrapper[4907]: I0313 15:41:58.478617 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d92d5136-2169-44cc-9900-c3335a4b69eb-config-data-custom\") pod \"cinder-backup-0\" (UID: \"d92d5136-2169-44cc-9900-c3335a4b69eb\") " pod="openstack/cinder-backup-0" Mar 13 15:41:58 crc kubenswrapper[4907]: I0313 15:41:58.478635 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/d92d5136-2169-44cc-9900-c3335a4b69eb-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"d92d5136-2169-44cc-9900-c3335a4b69eb\") " pod="openstack/cinder-backup-0" Mar 13 15:41:58 crc kubenswrapper[4907]: I0313 15:41:58.478654 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d92d5136-2169-44cc-9900-c3335a4b69eb-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"d92d5136-2169-44cc-9900-c3335a4b69eb\") " pod="openstack/cinder-backup-0" Mar 13 15:41:58 crc kubenswrapper[4907]: I0313 15:41:58.478676 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/d92d5136-2169-44cc-9900-c3335a4b69eb-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"d92d5136-2169-44cc-9900-c3335a4b69eb\") " pod="openstack/cinder-backup-0" Mar 13 15:41:58 crc kubenswrapper[4907]: I0313 15:41:58.478716 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/d92d5136-2169-44cc-9900-c3335a4b69eb-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"d92d5136-2169-44cc-9900-c3335a4b69eb\") " pod="openstack/cinder-backup-0" Mar 13 15:41:58 crc kubenswrapper[4907]: I0313 15:41:58.478733 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/d92d5136-2169-44cc-9900-c3335a4b69eb-lib-modules\") pod \"cinder-backup-0\" (UID: \"d92d5136-2169-44cc-9900-c3335a4b69eb\") " pod="openstack/cinder-backup-0" Mar 13 15:41:58 crc kubenswrapper[4907]: I0313 15:41:58.478749 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/d92d5136-2169-44cc-9900-c3335a4b69eb-etc-nvme\") pod \"cinder-backup-0\" (UID: \"d92d5136-2169-44cc-9900-c3335a4b69eb\") " pod="openstack/cinder-backup-0" Mar 13 15:41:58 crc kubenswrapper[4907]: I0313 15:41:58.478810 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/d92d5136-2169-44cc-9900-c3335a4b69eb-sys\") pod \"cinder-backup-0\" (UID: \"d92d5136-2169-44cc-9900-c3335a4b69eb\") " pod="openstack/cinder-backup-0" Mar 13 15:41:58 crc kubenswrapper[4907]: I0313 15:41:58.478825 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d92d5136-2169-44cc-9900-c3335a4b69eb-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"d92d5136-2169-44cc-9900-c3335a4b69eb\") " pod="openstack/cinder-backup-0" Mar 13 15:41:58 crc kubenswrapper[4907]: I0313 15:41:58.478846 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/d92d5136-2169-44cc-9900-c3335a4b69eb-run\") pod \"cinder-backup-0\" (UID: \"d92d5136-2169-44cc-9900-c3335a4b69eb\") " pod="openstack/cinder-backup-0" Mar 13 15:41:58 crc kubenswrapper[4907]: I0313 15:41:58.478866 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/d92d5136-2169-44cc-9900-c3335a4b69eb-dev\") pod \"cinder-backup-0\" (UID: \"d92d5136-2169-44cc-9900-c3335a4b69eb\") " pod="openstack/cinder-backup-0" Mar 13 15:41:58 crc kubenswrapper[4907]: I0313 15:41:58.478894 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/d92d5136-2169-44cc-9900-c3335a4b69eb-ceph\") pod \"cinder-backup-0\" (UID: \"d92d5136-2169-44cc-9900-c3335a4b69eb\") " pod="openstack/cinder-backup-0" Mar 13 15:41:58 crc kubenswrapper[4907]: I0313 15:41:58.479341 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/d92d5136-2169-44cc-9900-c3335a4b69eb-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"d92d5136-2169-44cc-9900-c3335a4b69eb\") " pod="openstack/cinder-backup-0" Mar 13 15:41:58 crc kubenswrapper[4907]: I0313 15:41:58.482794 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-volume-volume1-0"] Mar 13 15:41:58 crc kubenswrapper[4907]: I0313 15:41:58.484777 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/d92d5136-2169-44cc-9900-c3335a4b69eb-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"d92d5136-2169-44cc-9900-c3335a4b69eb\") " pod="openstack/cinder-backup-0" Mar 13 15:41:58 crc kubenswrapper[4907]: I0313 15:41:58.484848 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/d92d5136-2169-44cc-9900-c3335a4b69eb-sys\") pod \"cinder-backup-0\" (UID: \"d92d5136-2169-44cc-9900-c3335a4b69eb\") " pod="openstack/cinder-backup-0" Mar 13 15:41:58 crc kubenswrapper[4907]: I0313 15:41:58.484912 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/d92d5136-2169-44cc-9900-c3335a4b69eb-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"d92d5136-2169-44cc-9900-c3335a4b69eb\") " pod="openstack/cinder-backup-0" Mar 13 15:41:58 crc kubenswrapper[4907]: I0313 15:41:58.484944 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/d92d5136-2169-44cc-9900-c3335a4b69eb-lib-modules\") pod \"cinder-backup-0\" (UID: \"d92d5136-2169-44cc-9900-c3335a4b69eb\") " pod="openstack/cinder-backup-0" Mar 13 15:41:58 crc kubenswrapper[4907]: I0313 15:41:58.484979 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/d92d5136-2169-44cc-9900-c3335a4b69eb-etc-nvme\") pod \"cinder-backup-0\" (UID: \"d92d5136-2169-44cc-9900-c3335a4b69eb\") " pod="openstack/cinder-backup-0" Mar 13 15:41:58 crc kubenswrapper[4907]: I0313 15:41:58.485177 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/d92d5136-2169-44cc-9900-c3335a4b69eb-run\") pod \"cinder-backup-0\" (UID: \"d92d5136-2169-44cc-9900-c3335a4b69eb\") " pod="openstack/cinder-backup-0" Mar 13 15:41:58 crc kubenswrapper[4907]: I0313 15:41:58.485244 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d92d5136-2169-44cc-9900-c3335a4b69eb-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"d92d5136-2169-44cc-9900-c3335a4b69eb\") " pod="openstack/cinder-backup-0" Mar 13 15:41:58 crc kubenswrapper[4907]: I0313 15:41:58.486667 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/d92d5136-2169-44cc-9900-c3335a4b69eb-ceph\") pod \"cinder-backup-0\" (UID: \"d92d5136-2169-44cc-9900-c3335a4b69eb\") " pod="openstack/cinder-backup-0" Mar 13 15:41:58 crc kubenswrapper[4907]: I0313 15:41:58.486735 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/d92d5136-2169-44cc-9900-c3335a4b69eb-dev\") pod \"cinder-backup-0\" (UID: \"d92d5136-2169-44cc-9900-c3335a4b69eb\") " pod="openstack/cinder-backup-0" Mar 13 15:41:58 crc kubenswrapper[4907]: I0313 15:41:58.486788 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/d92d5136-2169-44cc-9900-c3335a4b69eb-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"d92d5136-2169-44cc-9900-c3335a4b69eb\") " pod="openstack/cinder-backup-0" Mar 13 15:41:58 crc kubenswrapper[4907]: I0313 15:41:58.489680 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d92d5136-2169-44cc-9900-c3335a4b69eb-config-data\") pod \"cinder-backup-0\" (UID: \"d92d5136-2169-44cc-9900-c3335a4b69eb\") " pod="openstack/cinder-backup-0" Mar 13 15:41:58 crc kubenswrapper[4907]: I0313 15:41:58.499659 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d92d5136-2169-44cc-9900-c3335a4b69eb-scripts\") pod \"cinder-backup-0\" (UID: \"d92d5136-2169-44cc-9900-c3335a4b69eb\") " pod="openstack/cinder-backup-0" Mar 13 15:41:58 crc kubenswrapper[4907]: I0313 15:41:58.503443 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d92d5136-2169-44cc-9900-c3335a4b69eb-config-data-custom\") pod \"cinder-backup-0\" (UID: \"d92d5136-2169-44cc-9900-c3335a4b69eb\") " pod="openstack/cinder-backup-0" Mar 13 15:41:58 crc kubenswrapper[4907]: I0313 15:41:58.503566 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kls8z\" (UniqueName: \"kubernetes.io/projected/d92d5136-2169-44cc-9900-c3335a4b69eb-kube-api-access-kls8z\") pod \"cinder-backup-0\" (UID: \"d92d5136-2169-44cc-9900-c3335a4b69eb\") " pod="openstack/cinder-backup-0" Mar 13 15:41:58 crc kubenswrapper[4907]: I0313 15:41:58.506390 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d92d5136-2169-44cc-9900-c3335a4b69eb-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"d92d5136-2169-44cc-9900-c3335a4b69eb\") " pod="openstack/cinder-backup-0" Mar 13 15:41:58 crc kubenswrapper[4907]: I0313 15:41:58.655958 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-backup-0" Mar 13 15:41:59 crc kubenswrapper[4907]: I0313 15:41:59.090794 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"697762c8-b121-410a-a5db-21916ca1791f","Type":"ContainerStarted","Data":"dd76292c1d3583e1de82c7463dec626e3e766b926d7d02c412e867389a2f86d2"} Mar 13 15:41:59 crc kubenswrapper[4907]: I0313 15:41:59.117662 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Mar 13 15:41:59 crc kubenswrapper[4907]: I0313 15:41:59.255762 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-backup-0"] Mar 13 15:42:00 crc kubenswrapper[4907]: I0313 15:42:00.099178 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"697762c8-b121-410a-a5db-21916ca1791f","Type":"ContainerStarted","Data":"67e59dc847d580867e158938a2fb3b51424233c6ff48024374cf07e953a69756"} Mar 13 15:42:00 crc kubenswrapper[4907]: I0313 15:42:00.099809 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"697762c8-b121-410a-a5db-21916ca1791f","Type":"ContainerStarted","Data":"bd5431dfa5b4119b17ac9f6b0b76ab0c54b5f6b19e74d192b265bb7cb3e2db8a"} Mar 13 15:42:00 crc kubenswrapper[4907]: I0313 15:42:00.103659 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"d92d5136-2169-44cc-9900-c3335a4b69eb","Type":"ContainerStarted","Data":"14a65b0f6bf8a07e477af765e4bdab5239237933ab32aa4a7cef322ba35a3dc1"} Mar 13 15:42:00 crc kubenswrapper[4907]: I0313 15:42:00.126347 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-volume-volume1-0" podStartSLOduration=2.13408908 podStartE2EDuration="3.126328539s" podCreationTimestamp="2026-03-13 15:41:57 +0000 UTC" firstStartedPulling="2026-03-13 15:41:58.485773647 +0000 UTC m=+5817.385561336" lastFinishedPulling="2026-03-13 15:41:59.478013106 +0000 UTC m=+5818.377800795" observedRunningTime="2026-03-13 15:42:00.124643643 +0000 UTC m=+5819.024431352" watchObservedRunningTime="2026-03-13 15:42:00.126328539 +0000 UTC m=+5819.026116228" Mar 13 15:42:00 crc kubenswrapper[4907]: I0313 15:42:00.150139 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556942-mdddr"] Mar 13 15:42:00 crc kubenswrapper[4907]: I0313 15:42:00.151725 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556942-mdddr" Mar 13 15:42:00 crc kubenswrapper[4907]: I0313 15:42:00.156993 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556942-mdddr"] Mar 13 15:42:00 crc kubenswrapper[4907]: I0313 15:42:00.157477 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 15:42:00 crc kubenswrapper[4907]: I0313 15:42:00.157477 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 15:42:00 crc kubenswrapper[4907]: I0313 15:42:00.157844 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 15:42:00 crc kubenswrapper[4907]: I0313 15:42:00.274421 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Mar 13 15:42:00 crc kubenswrapper[4907]: I0313 15:42:00.274482 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Mar 13 15:42:00 crc kubenswrapper[4907]: I0313 15:42:00.328463 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Mar 13 15:42:00 crc kubenswrapper[4907]: I0313 15:42:00.328811 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Mar 13 15:42:00 crc kubenswrapper[4907]: I0313 15:42:00.329852 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s8lgh\" (UniqueName: \"kubernetes.io/projected/ed0ed6c7-e5d4-4770-a9aa-308c0853573e-kube-api-access-s8lgh\") pod \"auto-csr-approver-29556942-mdddr\" (UID: \"ed0ed6c7-e5d4-4770-a9aa-308c0853573e\") " pod="openshift-infra/auto-csr-approver-29556942-mdddr" Mar 13 15:42:00 crc kubenswrapper[4907]: I0313 15:42:00.432301 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s8lgh\" (UniqueName: \"kubernetes.io/projected/ed0ed6c7-e5d4-4770-a9aa-308c0853573e-kube-api-access-s8lgh\") pod \"auto-csr-approver-29556942-mdddr\" (UID: \"ed0ed6c7-e5d4-4770-a9aa-308c0853573e\") " pod="openshift-infra/auto-csr-approver-29556942-mdddr" Mar 13 15:42:00 crc kubenswrapper[4907]: I0313 15:42:00.455461 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s8lgh\" (UniqueName: \"kubernetes.io/projected/ed0ed6c7-e5d4-4770-a9aa-308c0853573e-kube-api-access-s8lgh\") pod \"auto-csr-approver-29556942-mdddr\" (UID: \"ed0ed6c7-e5d4-4770-a9aa-308c0853573e\") " pod="openshift-infra/auto-csr-approver-29556942-mdddr" Mar 13 15:42:00 crc kubenswrapper[4907]: I0313 15:42:00.482996 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556942-mdddr" Mar 13 15:42:00 crc kubenswrapper[4907]: I0313 15:42:00.660768 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Mar 13 15:42:00 crc kubenswrapper[4907]: I0313 15:42:00.844628 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dh7tk\" (UniqueName: \"kubernetes.io/projected/12616424-013a-47c8-9ed3-b407791537fe-kube-api-access-dh7tk\") pod \"12616424-013a-47c8-9ed3-b407791537fe\" (UID: \"12616424-013a-47c8-9ed3-b407791537fe\") " Mar 13 15:42:00 crc kubenswrapper[4907]: I0313 15:42:00.844710 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/12616424-013a-47c8-9ed3-b407791537fe-etc-machine-id\") pod \"12616424-013a-47c8-9ed3-b407791537fe\" (UID: \"12616424-013a-47c8-9ed3-b407791537fe\") " Mar 13 15:42:00 crc kubenswrapper[4907]: I0313 15:42:00.844747 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/12616424-013a-47c8-9ed3-b407791537fe-config-data-custom\") pod \"12616424-013a-47c8-9ed3-b407791537fe\" (UID: \"12616424-013a-47c8-9ed3-b407791537fe\") " Mar 13 15:42:00 crc kubenswrapper[4907]: I0313 15:42:00.844825 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12616424-013a-47c8-9ed3-b407791537fe-combined-ca-bundle\") pod \"12616424-013a-47c8-9ed3-b407791537fe\" (UID: \"12616424-013a-47c8-9ed3-b407791537fe\") " Mar 13 15:42:00 crc kubenswrapper[4907]: I0313 15:42:00.844975 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/12616424-013a-47c8-9ed3-b407791537fe-scripts\") pod \"12616424-013a-47c8-9ed3-b407791537fe\" (UID: \"12616424-013a-47c8-9ed3-b407791537fe\") " Mar 13 15:42:00 crc kubenswrapper[4907]: I0313 15:42:00.845087 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12616424-013a-47c8-9ed3-b407791537fe-config-data\") pod \"12616424-013a-47c8-9ed3-b407791537fe\" (UID: \"12616424-013a-47c8-9ed3-b407791537fe\") " Mar 13 15:42:00 crc kubenswrapper[4907]: I0313 15:42:00.845119 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/12616424-013a-47c8-9ed3-b407791537fe-logs\") pod \"12616424-013a-47c8-9ed3-b407791537fe\" (UID: \"12616424-013a-47c8-9ed3-b407791537fe\") " Mar 13 15:42:00 crc kubenswrapper[4907]: I0313 15:42:00.845802 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/12616424-013a-47c8-9ed3-b407791537fe-logs" (OuterVolumeSpecName: "logs") pod "12616424-013a-47c8-9ed3-b407791537fe" (UID: "12616424-013a-47c8-9ed3-b407791537fe"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:42:00 crc kubenswrapper[4907]: I0313 15:42:00.848661 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/12616424-013a-47c8-9ed3-b407791537fe-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "12616424-013a-47c8-9ed3-b407791537fe" (UID: "12616424-013a-47c8-9ed3-b407791537fe"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 15:42:00 crc kubenswrapper[4907]: I0313 15:42:00.851970 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/12616424-013a-47c8-9ed3-b407791537fe-kube-api-access-dh7tk" (OuterVolumeSpecName: "kube-api-access-dh7tk") pod "12616424-013a-47c8-9ed3-b407791537fe" (UID: "12616424-013a-47c8-9ed3-b407791537fe"). InnerVolumeSpecName "kube-api-access-dh7tk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:42:00 crc kubenswrapper[4907]: I0313 15:42:00.852092 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/12616424-013a-47c8-9ed3-b407791537fe-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "12616424-013a-47c8-9ed3-b407791537fe" (UID: "12616424-013a-47c8-9ed3-b407791537fe"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:42:00 crc kubenswrapper[4907]: I0313 15:42:00.857224 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/12616424-013a-47c8-9ed3-b407791537fe-scripts" (OuterVolumeSpecName: "scripts") pod "12616424-013a-47c8-9ed3-b407791537fe" (UID: "12616424-013a-47c8-9ed3-b407791537fe"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:42:00 crc kubenswrapper[4907]: I0313 15:42:00.904423 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/12616424-013a-47c8-9ed3-b407791537fe-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "12616424-013a-47c8-9ed3-b407791537fe" (UID: "12616424-013a-47c8-9ed3-b407791537fe"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:42:00 crc kubenswrapper[4907]: I0313 15:42:00.923499 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/12616424-013a-47c8-9ed3-b407791537fe-config-data" (OuterVolumeSpecName: "config-data") pod "12616424-013a-47c8-9ed3-b407791537fe" (UID: "12616424-013a-47c8-9ed3-b407791537fe"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:42:00 crc kubenswrapper[4907]: I0313 15:42:00.947678 4907 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/12616424-013a-47c8-9ed3-b407791537fe-etc-machine-id\") on node \"crc\" DevicePath \"\"" Mar 13 15:42:00 crc kubenswrapper[4907]: I0313 15:42:00.947712 4907 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/12616424-013a-47c8-9ed3-b407791537fe-config-data-custom\") on node \"crc\" DevicePath \"\"" Mar 13 15:42:00 crc kubenswrapper[4907]: I0313 15:42:00.947723 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12616424-013a-47c8-9ed3-b407791537fe-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 15:42:00 crc kubenswrapper[4907]: I0313 15:42:00.947731 4907 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/12616424-013a-47c8-9ed3-b407791537fe-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 15:42:00 crc kubenswrapper[4907]: I0313 15:42:00.947740 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12616424-013a-47c8-9ed3-b407791537fe-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 15:42:00 crc kubenswrapper[4907]: I0313 15:42:00.947747 4907 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/12616424-013a-47c8-9ed3-b407791537fe-logs\") on node \"crc\" DevicePath \"\"" Mar 13 15:42:00 crc kubenswrapper[4907]: I0313 15:42:00.947758 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dh7tk\" (UniqueName: \"kubernetes.io/projected/12616424-013a-47c8-9ed3-b407791537fe-kube-api-access-dh7tk\") on node \"crc\" DevicePath \"\"" Mar 13 15:42:01 crc kubenswrapper[4907]: I0313 15:42:01.059761 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556942-mdddr"] Mar 13 15:42:01 crc kubenswrapper[4907]: I0313 15:42:01.065038 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Mar 13 15:42:01 crc kubenswrapper[4907]: I0313 15:42:01.116619 4907 generic.go:334] "Generic (PLEG): container finished" podID="12616424-013a-47c8-9ed3-b407791537fe" containerID="b5b4ab9a472db32a9c0ae91b2509ae28f1254ad915d9737b1279509e0c8d4717" exitCode=0 Mar 13 15:42:01 crc kubenswrapper[4907]: I0313 15:42:01.116680 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"12616424-013a-47c8-9ed3-b407791537fe","Type":"ContainerDied","Data":"b5b4ab9a472db32a9c0ae91b2509ae28f1254ad915d9737b1279509e0c8d4717"} Mar 13 15:42:01 crc kubenswrapper[4907]: I0313 15:42:01.116707 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"12616424-013a-47c8-9ed3-b407791537fe","Type":"ContainerDied","Data":"32e9d54b2cd43cd1f900bf7ababaaace7fcabaccb1a6e6993088fcc8c885c896"} Mar 13 15:42:01 crc kubenswrapper[4907]: I0313 15:42:01.116723 4907 scope.go:117] "RemoveContainer" containerID="b5b4ab9a472db32a9c0ae91b2509ae28f1254ad915d9737b1279509e0c8d4717" Mar 13 15:42:01 crc kubenswrapper[4907]: I0313 15:42:01.116837 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Mar 13 15:42:01 crc kubenswrapper[4907]: I0313 15:42:01.135718 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556942-mdddr" event={"ID":"ed0ed6c7-e5d4-4770-a9aa-308c0853573e","Type":"ContainerStarted","Data":"763973579a2d2ec12ba365a5be5b38596e6d77a76e685b5bbe27b06cc46b881e"} Mar 13 15:42:01 crc kubenswrapper[4907]: I0313 15:42:01.140423 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"d92d5136-2169-44cc-9900-c3335a4b69eb","Type":"ContainerStarted","Data":"29454c8b38fc418834a59d22c32579d841bc25e689c84de54872ee146876dd55"} Mar 13 15:42:01 crc kubenswrapper[4907]: I0313 15:42:01.189045 4907 scope.go:117] "RemoveContainer" containerID="f234443780f86c06ddc613fb1859ac940ed1d9617b56539195ffdc071bbc882b" Mar 13 15:42:01 crc kubenswrapper[4907]: I0313 15:42:01.197167 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Mar 13 15:42:01 crc kubenswrapper[4907]: I0313 15:42:01.207128 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Mar 13 15:42:01 crc kubenswrapper[4907]: I0313 15:42:01.225782 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Mar 13 15:42:01 crc kubenswrapper[4907]: E0313 15:42:01.226733 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12616424-013a-47c8-9ed3-b407791537fe" containerName="cinder-api" Mar 13 15:42:01 crc kubenswrapper[4907]: I0313 15:42:01.226772 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="12616424-013a-47c8-9ed3-b407791537fe" containerName="cinder-api" Mar 13 15:42:01 crc kubenswrapper[4907]: E0313 15:42:01.226792 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12616424-013a-47c8-9ed3-b407791537fe" containerName="cinder-api-log" Mar 13 15:42:01 crc kubenswrapper[4907]: I0313 15:42:01.226800 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="12616424-013a-47c8-9ed3-b407791537fe" containerName="cinder-api-log" Mar 13 15:42:01 crc kubenswrapper[4907]: I0313 15:42:01.227064 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="12616424-013a-47c8-9ed3-b407791537fe" containerName="cinder-api-log" Mar 13 15:42:01 crc kubenswrapper[4907]: I0313 15:42:01.227094 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="12616424-013a-47c8-9ed3-b407791537fe" containerName="cinder-api" Mar 13 15:42:01 crc kubenswrapper[4907]: I0313 15:42:01.228399 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Mar 13 15:42:01 crc kubenswrapper[4907]: I0313 15:42:01.234942 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Mar 13 15:42:01 crc kubenswrapper[4907]: I0313 15:42:01.242510 4907 scope.go:117] "RemoveContainer" containerID="b5b4ab9a472db32a9c0ae91b2509ae28f1254ad915d9737b1279509e0c8d4717" Mar 13 15:42:01 crc kubenswrapper[4907]: E0313 15:42:01.247450 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b5b4ab9a472db32a9c0ae91b2509ae28f1254ad915d9737b1279509e0c8d4717\": container with ID starting with b5b4ab9a472db32a9c0ae91b2509ae28f1254ad915d9737b1279509e0c8d4717 not found: ID does not exist" containerID="b5b4ab9a472db32a9c0ae91b2509ae28f1254ad915d9737b1279509e0c8d4717" Mar 13 15:42:01 crc kubenswrapper[4907]: I0313 15:42:01.247487 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b5b4ab9a472db32a9c0ae91b2509ae28f1254ad915d9737b1279509e0c8d4717"} err="failed to get container status \"b5b4ab9a472db32a9c0ae91b2509ae28f1254ad915d9737b1279509e0c8d4717\": rpc error: code = NotFound desc = could not find container \"b5b4ab9a472db32a9c0ae91b2509ae28f1254ad915d9737b1279509e0c8d4717\": container with ID starting with b5b4ab9a472db32a9c0ae91b2509ae28f1254ad915d9737b1279509e0c8d4717 not found: ID does not exist" Mar 13 15:42:01 crc kubenswrapper[4907]: I0313 15:42:01.247509 4907 scope.go:117] "RemoveContainer" containerID="f234443780f86c06ddc613fb1859ac940ed1d9617b56539195ffdc071bbc882b" Mar 13 15:42:01 crc kubenswrapper[4907]: E0313 15:42:01.249901 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f234443780f86c06ddc613fb1859ac940ed1d9617b56539195ffdc071bbc882b\": container with ID starting with f234443780f86c06ddc613fb1859ac940ed1d9617b56539195ffdc071bbc882b not found: ID does not exist" containerID="f234443780f86c06ddc613fb1859ac940ed1d9617b56539195ffdc071bbc882b" Mar 13 15:42:01 crc kubenswrapper[4907]: I0313 15:42:01.249929 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f234443780f86c06ddc613fb1859ac940ed1d9617b56539195ffdc071bbc882b"} err="failed to get container status \"f234443780f86c06ddc613fb1859ac940ed1d9617b56539195ffdc071bbc882b\": rpc error: code = NotFound desc = could not find container \"f234443780f86c06ddc613fb1859ac940ed1d9617b56539195ffdc071bbc882b\": container with ID starting with f234443780f86c06ddc613fb1859ac940ed1d9617b56539195ffdc071bbc882b not found: ID does not exist" Mar 13 15:42:01 crc kubenswrapper[4907]: I0313 15:42:01.254204 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Mar 13 15:42:01 crc kubenswrapper[4907]: I0313 15:42:01.264711 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c82a596-df2b-4e69-abe9-5f4662a3c6da-config-data\") pod \"cinder-api-0\" (UID: \"8c82a596-df2b-4e69-abe9-5f4662a3c6da\") " pod="openstack/cinder-api-0" Mar 13 15:42:01 crc kubenswrapper[4907]: I0313 15:42:01.265394 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8c82a596-df2b-4e69-abe9-5f4662a3c6da-logs\") pod \"cinder-api-0\" (UID: \"8c82a596-df2b-4e69-abe9-5f4662a3c6da\") " pod="openstack/cinder-api-0" Mar 13 15:42:01 crc kubenswrapper[4907]: I0313 15:42:01.265656 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8c82a596-df2b-4e69-abe9-5f4662a3c6da-scripts\") pod \"cinder-api-0\" (UID: \"8c82a596-df2b-4e69-abe9-5f4662a3c6da\") " pod="openstack/cinder-api-0" Mar 13 15:42:01 crc kubenswrapper[4907]: I0313 15:42:01.265730 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c82a596-df2b-4e69-abe9-5f4662a3c6da-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"8c82a596-df2b-4e69-abe9-5f4662a3c6da\") " pod="openstack/cinder-api-0" Mar 13 15:42:01 crc kubenswrapper[4907]: I0313 15:42:01.265760 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8c82a596-df2b-4e69-abe9-5f4662a3c6da-etc-machine-id\") pod \"cinder-api-0\" (UID: \"8c82a596-df2b-4e69-abe9-5f4662a3c6da\") " pod="openstack/cinder-api-0" Mar 13 15:42:01 crc kubenswrapper[4907]: I0313 15:42:01.266238 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8c82a596-df2b-4e69-abe9-5f4662a3c6da-config-data-custom\") pod \"cinder-api-0\" (UID: \"8c82a596-df2b-4e69-abe9-5f4662a3c6da\") " pod="openstack/cinder-api-0" Mar 13 15:42:01 crc kubenswrapper[4907]: I0313 15:42:01.266303 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mf67l\" (UniqueName: \"kubernetes.io/projected/8c82a596-df2b-4e69-abe9-5f4662a3c6da-kube-api-access-mf67l\") pod \"cinder-api-0\" (UID: \"8c82a596-df2b-4e69-abe9-5f4662a3c6da\") " pod="openstack/cinder-api-0" Mar 13 15:42:01 crc kubenswrapper[4907]: I0313 15:42:01.371215 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8c82a596-df2b-4e69-abe9-5f4662a3c6da-config-data-custom\") pod \"cinder-api-0\" (UID: \"8c82a596-df2b-4e69-abe9-5f4662a3c6da\") " pod="openstack/cinder-api-0" Mar 13 15:42:01 crc kubenswrapper[4907]: I0313 15:42:01.371287 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mf67l\" (UniqueName: \"kubernetes.io/projected/8c82a596-df2b-4e69-abe9-5f4662a3c6da-kube-api-access-mf67l\") pod \"cinder-api-0\" (UID: \"8c82a596-df2b-4e69-abe9-5f4662a3c6da\") " pod="openstack/cinder-api-0" Mar 13 15:42:01 crc kubenswrapper[4907]: I0313 15:42:01.372521 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c82a596-df2b-4e69-abe9-5f4662a3c6da-config-data\") pod \"cinder-api-0\" (UID: \"8c82a596-df2b-4e69-abe9-5f4662a3c6da\") " pod="openstack/cinder-api-0" Mar 13 15:42:01 crc kubenswrapper[4907]: I0313 15:42:01.372698 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8c82a596-df2b-4e69-abe9-5f4662a3c6da-logs\") pod \"cinder-api-0\" (UID: \"8c82a596-df2b-4e69-abe9-5f4662a3c6da\") " pod="openstack/cinder-api-0" Mar 13 15:42:01 crc kubenswrapper[4907]: I0313 15:42:01.374458 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8c82a596-df2b-4e69-abe9-5f4662a3c6da-scripts\") pod \"cinder-api-0\" (UID: \"8c82a596-df2b-4e69-abe9-5f4662a3c6da\") " pod="openstack/cinder-api-0" Mar 13 15:42:01 crc kubenswrapper[4907]: I0313 15:42:01.374502 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c82a596-df2b-4e69-abe9-5f4662a3c6da-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"8c82a596-df2b-4e69-abe9-5f4662a3c6da\") " pod="openstack/cinder-api-0" Mar 13 15:42:01 crc kubenswrapper[4907]: I0313 15:42:01.374645 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8c82a596-df2b-4e69-abe9-5f4662a3c6da-etc-machine-id\") pod \"cinder-api-0\" (UID: \"8c82a596-df2b-4e69-abe9-5f4662a3c6da\") " pod="openstack/cinder-api-0" Mar 13 15:42:01 crc kubenswrapper[4907]: I0313 15:42:01.376991 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8c82a596-df2b-4e69-abe9-5f4662a3c6da-etc-machine-id\") pod \"cinder-api-0\" (UID: \"8c82a596-df2b-4e69-abe9-5f4662a3c6da\") " pod="openstack/cinder-api-0" Mar 13 15:42:01 crc kubenswrapper[4907]: I0313 15:42:01.378193 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8c82a596-df2b-4e69-abe9-5f4662a3c6da-logs\") pod \"cinder-api-0\" (UID: \"8c82a596-df2b-4e69-abe9-5f4662a3c6da\") " pod="openstack/cinder-api-0" Mar 13 15:42:01 crc kubenswrapper[4907]: I0313 15:42:01.382861 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8c82a596-df2b-4e69-abe9-5f4662a3c6da-scripts\") pod \"cinder-api-0\" (UID: \"8c82a596-df2b-4e69-abe9-5f4662a3c6da\") " pod="openstack/cinder-api-0" Mar 13 15:42:01 crc kubenswrapper[4907]: I0313 15:42:01.396625 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mf67l\" (UniqueName: \"kubernetes.io/projected/8c82a596-df2b-4e69-abe9-5f4662a3c6da-kube-api-access-mf67l\") pod \"cinder-api-0\" (UID: \"8c82a596-df2b-4e69-abe9-5f4662a3c6da\") " pod="openstack/cinder-api-0" Mar 13 15:42:01 crc kubenswrapper[4907]: I0313 15:42:01.396950 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8c82a596-df2b-4e69-abe9-5f4662a3c6da-config-data-custom\") pod \"cinder-api-0\" (UID: \"8c82a596-df2b-4e69-abe9-5f4662a3c6da\") " pod="openstack/cinder-api-0" Mar 13 15:42:01 crc kubenswrapper[4907]: I0313 15:42:01.398019 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c82a596-df2b-4e69-abe9-5f4662a3c6da-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"8c82a596-df2b-4e69-abe9-5f4662a3c6da\") " pod="openstack/cinder-api-0" Mar 13 15:42:01 crc kubenswrapper[4907]: I0313 15:42:01.402112 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c82a596-df2b-4e69-abe9-5f4662a3c6da-config-data\") pod \"cinder-api-0\" (UID: \"8c82a596-df2b-4e69-abe9-5f4662a3c6da\") " pod="openstack/cinder-api-0" Mar 13 15:42:01 crc kubenswrapper[4907]: I0313 15:42:01.616415 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Mar 13 15:42:01 crc kubenswrapper[4907]: I0313 15:42:01.811482 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="12616424-013a-47c8-9ed3-b407791537fe" path="/var/lib/kubelet/pods/12616424-013a-47c8-9ed3-b407791537fe/volumes" Mar 13 15:42:02 crc kubenswrapper[4907]: W0313 15:42:02.083691 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8c82a596_df2b_4e69_abe9_5f4662a3c6da.slice/crio-ea21a86a7210843a61ba8d606f21ee3219af37b3433a82cdba8aae6e4a77d1b7 WatchSource:0}: Error finding container ea21a86a7210843a61ba8d606f21ee3219af37b3433a82cdba8aae6e4a77d1b7: Status 404 returned error can't find the container with id ea21a86a7210843a61ba8d606f21ee3219af37b3433a82cdba8aae6e4a77d1b7 Mar 13 15:42:02 crc kubenswrapper[4907]: I0313 15:42:02.090637 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Mar 13 15:42:02 crc kubenswrapper[4907]: I0313 15:42:02.152816 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"8c82a596-df2b-4e69-abe9-5f4662a3c6da","Type":"ContainerStarted","Data":"ea21a86a7210843a61ba8d606f21ee3219af37b3433a82cdba8aae6e4a77d1b7"} Mar 13 15:42:02 crc kubenswrapper[4907]: I0313 15:42:02.159660 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"d92d5136-2169-44cc-9900-c3335a4b69eb","Type":"ContainerStarted","Data":"427089c1bbd58a37d7507fb363c5b14d82a397d44285e0e4981cb8c2241be726"} Mar 13 15:42:02 crc kubenswrapper[4907]: I0313 15:42:02.204383 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-backup-0" podStartSLOduration=2.701627639 podStartE2EDuration="4.204341225s" podCreationTimestamp="2026-03-13 15:41:58 +0000 UTC" firstStartedPulling="2026-03-13 15:41:59.262118248 +0000 UTC m=+5818.161905937" lastFinishedPulling="2026-03-13 15:42:00.764831834 +0000 UTC m=+5819.664619523" observedRunningTime="2026-03-13 15:42:02.196312396 +0000 UTC m=+5821.096100085" watchObservedRunningTime="2026-03-13 15:42:02.204341225 +0000 UTC m=+5821.104128914" Mar 13 15:42:02 crc kubenswrapper[4907]: I0313 15:42:02.281979 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Mar 13 15:42:02 crc kubenswrapper[4907]: I0313 15:42:02.285821 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Mar 13 15:42:02 crc kubenswrapper[4907]: I0313 15:42:02.285970 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Mar 13 15:42:02 crc kubenswrapper[4907]: I0313 15:42:02.332055 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Mar 13 15:42:02 crc kubenswrapper[4907]: I0313 15:42:02.334365 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Mar 13 15:42:02 crc kubenswrapper[4907]: I0313 15:42:02.334842 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Mar 13 15:42:02 crc kubenswrapper[4907]: I0313 15:42:02.835411 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-volume-volume1-0" Mar 13 15:42:03 crc kubenswrapper[4907]: I0313 15:42:03.171242 4907 generic.go:334] "Generic (PLEG): container finished" podID="ed0ed6c7-e5d4-4770-a9aa-308c0853573e" containerID="7c596675c8272c2dfe2507df7b83ab578498cef4ae7b5f27995638598d3ea046" exitCode=0 Mar 13 15:42:03 crc kubenswrapper[4907]: I0313 15:42:03.171334 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556942-mdddr" event={"ID":"ed0ed6c7-e5d4-4770-a9aa-308c0853573e","Type":"ContainerDied","Data":"7c596675c8272c2dfe2507df7b83ab578498cef4ae7b5f27995638598d3ea046"} Mar 13 15:42:03 crc kubenswrapper[4907]: I0313 15:42:03.187693 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"8c82a596-df2b-4e69-abe9-5f4662a3c6da","Type":"ContainerStarted","Data":"268cc9bc4e1fb3e66dd707585b30d6d2cb836b5a40afb999a88784cf686be14c"} Mar 13 15:42:03 crc kubenswrapper[4907]: I0313 15:42:03.191114 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Mar 13 15:42:03 crc kubenswrapper[4907]: I0313 15:42:03.192636 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Mar 13 15:42:03 crc kubenswrapper[4907]: I0313 15:42:03.657143 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-backup-0" Mar 13 15:42:04 crc kubenswrapper[4907]: I0313 15:42:04.196219 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"8c82a596-df2b-4e69-abe9-5f4662a3c6da","Type":"ContainerStarted","Data":"a7a8a2bc0d7150294189b41f6abfff6c5259b80a09c5f094de918c944f25e0b2"} Mar 13 15:42:04 crc kubenswrapper[4907]: I0313 15:42:04.232392 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.232373354 podStartE2EDuration="3.232373354s" podCreationTimestamp="2026-03-13 15:42:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:42:04.21721359 +0000 UTC m=+5823.117001279" watchObservedRunningTime="2026-03-13 15:42:04.232373354 +0000 UTC m=+5823.132161043" Mar 13 15:42:04 crc kubenswrapper[4907]: I0313 15:42:04.555723 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556942-mdddr" Mar 13 15:42:04 crc kubenswrapper[4907]: I0313 15:42:04.645726 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s8lgh\" (UniqueName: \"kubernetes.io/projected/ed0ed6c7-e5d4-4770-a9aa-308c0853573e-kube-api-access-s8lgh\") pod \"ed0ed6c7-e5d4-4770-a9aa-308c0853573e\" (UID: \"ed0ed6c7-e5d4-4770-a9aa-308c0853573e\") " Mar 13 15:42:04 crc kubenswrapper[4907]: I0313 15:42:04.651451 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ed0ed6c7-e5d4-4770-a9aa-308c0853573e-kube-api-access-s8lgh" (OuterVolumeSpecName: "kube-api-access-s8lgh") pod "ed0ed6c7-e5d4-4770-a9aa-308c0853573e" (UID: "ed0ed6c7-e5d4-4770-a9aa-308c0853573e"). InnerVolumeSpecName "kube-api-access-s8lgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:42:04 crc kubenswrapper[4907]: I0313 15:42:04.748344 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s8lgh\" (UniqueName: \"kubernetes.io/projected/ed0ed6c7-e5d4-4770-a9aa-308c0853573e-kube-api-access-s8lgh\") on node \"crc\" DevicePath \"\"" Mar 13 15:42:05 crc kubenswrapper[4907]: I0313 15:42:05.208319 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556942-mdddr" Mar 13 15:42:05 crc kubenswrapper[4907]: I0313 15:42:05.208340 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556942-mdddr" event={"ID":"ed0ed6c7-e5d4-4770-a9aa-308c0853573e","Type":"ContainerDied","Data":"763973579a2d2ec12ba365a5be5b38596e6d77a76e685b5bbe27b06cc46b881e"} Mar 13 15:42:05 crc kubenswrapper[4907]: I0313 15:42:05.208370 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="763973579a2d2ec12ba365a5be5b38596e6d77a76e685b5bbe27b06cc46b881e" Mar 13 15:42:05 crc kubenswrapper[4907]: I0313 15:42:05.208540 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Mar 13 15:42:05 crc kubenswrapper[4907]: I0313 15:42:05.504967 4907 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cinder-api-0" podUID="12616424-013a-47c8-9ed3-b407791537fe" containerName="cinder-api" probeResult="failure" output="Get \"http://10.217.1.120:8776/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 15:42:05 crc kubenswrapper[4907]: I0313 15:42:05.648178 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556936-bwsf9"] Mar 13 15:42:05 crc kubenswrapper[4907]: I0313 15:42:05.656246 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556936-bwsf9"] Mar 13 15:42:05 crc kubenswrapper[4907]: I0313 15:42:05.796088 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="38ab6eb6-9c07-41f4-87e4-d74cb4e5aff7" path="/var/lib/kubelet/pods/38ab6eb6-9c07-41f4-87e4-d74cb4e5aff7/volumes" Mar 13 15:42:06 crc kubenswrapper[4907]: I0313 15:42:06.262083 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Mar 13 15:42:06 crc kubenswrapper[4907]: I0313 15:42:06.302479 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Mar 13 15:42:07 crc kubenswrapper[4907]: I0313 15:42:07.228705 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="bfe30c39-e9da-49f4-a9da-ebbce147b863" containerName="cinder-scheduler" containerID="cri-o://24af8c70da41dc6cb88e8d25d6b679b4b7c4b0986765337bf4ab3605a412ae27" gracePeriod=30 Mar 13 15:42:07 crc kubenswrapper[4907]: I0313 15:42:07.228873 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="bfe30c39-e9da-49f4-a9da-ebbce147b863" containerName="probe" containerID="cri-o://fbe67e4b2a6e6d29dd61f35ad336b329273792eab74bfcf4dd2ade78056599e6" gracePeriod=30 Mar 13 15:42:08 crc kubenswrapper[4907]: I0313 15:42:08.053095 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-volume-volume1-0" Mar 13 15:42:08 crc kubenswrapper[4907]: I0313 15:42:08.239606 4907 generic.go:334] "Generic (PLEG): container finished" podID="bfe30c39-e9da-49f4-a9da-ebbce147b863" containerID="fbe67e4b2a6e6d29dd61f35ad336b329273792eab74bfcf4dd2ade78056599e6" exitCode=0 Mar 13 15:42:08 crc kubenswrapper[4907]: I0313 15:42:08.239704 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"bfe30c39-e9da-49f4-a9da-ebbce147b863","Type":"ContainerDied","Data":"fbe67e4b2a6e6d29dd61f35ad336b329273792eab74bfcf4dd2ade78056599e6"} Mar 13 15:42:08 crc kubenswrapper[4907]: I0313 15:42:08.872665 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-backup-0" Mar 13 15:42:08 crc kubenswrapper[4907]: I0313 15:42:08.888413 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Mar 13 15:42:09 crc kubenswrapper[4907]: I0313 15:42:09.023448 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d5t84\" (UniqueName: \"kubernetes.io/projected/bfe30c39-e9da-49f4-a9da-ebbce147b863-kube-api-access-d5t84\") pod \"bfe30c39-e9da-49f4-a9da-ebbce147b863\" (UID: \"bfe30c39-e9da-49f4-a9da-ebbce147b863\") " Mar 13 15:42:09 crc kubenswrapper[4907]: I0313 15:42:09.023858 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/bfe30c39-e9da-49f4-a9da-ebbce147b863-etc-machine-id\") pod \"bfe30c39-e9da-49f4-a9da-ebbce147b863\" (UID: \"bfe30c39-e9da-49f4-a9da-ebbce147b863\") " Mar 13 15:42:09 crc kubenswrapper[4907]: I0313 15:42:09.023938 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bfe30c39-e9da-49f4-a9da-ebbce147b863-config-data-custom\") pod \"bfe30c39-e9da-49f4-a9da-ebbce147b863\" (UID: \"bfe30c39-e9da-49f4-a9da-ebbce147b863\") " Mar 13 15:42:09 crc kubenswrapper[4907]: I0313 15:42:09.023961 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bfe30c39-e9da-49f4-a9da-ebbce147b863-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "bfe30c39-e9da-49f4-a9da-ebbce147b863" (UID: "bfe30c39-e9da-49f4-a9da-ebbce147b863"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 15:42:09 crc kubenswrapper[4907]: I0313 15:42:09.023989 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bfe30c39-e9da-49f4-a9da-ebbce147b863-scripts\") pod \"bfe30c39-e9da-49f4-a9da-ebbce147b863\" (UID: \"bfe30c39-e9da-49f4-a9da-ebbce147b863\") " Mar 13 15:42:09 crc kubenswrapper[4907]: I0313 15:42:09.024018 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfe30c39-e9da-49f4-a9da-ebbce147b863-combined-ca-bundle\") pod \"bfe30c39-e9da-49f4-a9da-ebbce147b863\" (UID: \"bfe30c39-e9da-49f4-a9da-ebbce147b863\") " Mar 13 15:42:09 crc kubenswrapper[4907]: I0313 15:42:09.024077 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bfe30c39-e9da-49f4-a9da-ebbce147b863-config-data\") pod \"bfe30c39-e9da-49f4-a9da-ebbce147b863\" (UID: \"bfe30c39-e9da-49f4-a9da-ebbce147b863\") " Mar 13 15:42:09 crc kubenswrapper[4907]: I0313 15:42:09.024920 4907 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/bfe30c39-e9da-49f4-a9da-ebbce147b863-etc-machine-id\") on node \"crc\" DevicePath \"\"" Mar 13 15:42:09 crc kubenswrapper[4907]: I0313 15:42:09.029152 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bfe30c39-e9da-49f4-a9da-ebbce147b863-scripts" (OuterVolumeSpecName: "scripts") pod "bfe30c39-e9da-49f4-a9da-ebbce147b863" (UID: "bfe30c39-e9da-49f4-a9da-ebbce147b863"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:42:09 crc kubenswrapper[4907]: I0313 15:42:09.033157 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bfe30c39-e9da-49f4-a9da-ebbce147b863-kube-api-access-d5t84" (OuterVolumeSpecName: "kube-api-access-d5t84") pod "bfe30c39-e9da-49f4-a9da-ebbce147b863" (UID: "bfe30c39-e9da-49f4-a9da-ebbce147b863"). InnerVolumeSpecName "kube-api-access-d5t84". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:42:09 crc kubenswrapper[4907]: I0313 15:42:09.042135 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bfe30c39-e9da-49f4-a9da-ebbce147b863-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "bfe30c39-e9da-49f4-a9da-ebbce147b863" (UID: "bfe30c39-e9da-49f4-a9da-ebbce147b863"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:42:09 crc kubenswrapper[4907]: I0313 15:42:09.087675 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bfe30c39-e9da-49f4-a9da-ebbce147b863-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bfe30c39-e9da-49f4-a9da-ebbce147b863" (UID: "bfe30c39-e9da-49f4-a9da-ebbce147b863"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:42:09 crc kubenswrapper[4907]: I0313 15:42:09.126135 4907 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bfe30c39-e9da-49f4-a9da-ebbce147b863-config-data-custom\") on node \"crc\" DevicePath \"\"" Mar 13 15:42:09 crc kubenswrapper[4907]: I0313 15:42:09.126167 4907 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bfe30c39-e9da-49f4-a9da-ebbce147b863-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 15:42:09 crc kubenswrapper[4907]: I0313 15:42:09.126179 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfe30c39-e9da-49f4-a9da-ebbce147b863-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 15:42:09 crc kubenswrapper[4907]: I0313 15:42:09.126190 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d5t84\" (UniqueName: \"kubernetes.io/projected/bfe30c39-e9da-49f4-a9da-ebbce147b863-kube-api-access-d5t84\") on node \"crc\" DevicePath \"\"" Mar 13 15:42:09 crc kubenswrapper[4907]: I0313 15:42:09.141184 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bfe30c39-e9da-49f4-a9da-ebbce147b863-config-data" (OuterVolumeSpecName: "config-data") pod "bfe30c39-e9da-49f4-a9da-ebbce147b863" (UID: "bfe30c39-e9da-49f4-a9da-ebbce147b863"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:42:09 crc kubenswrapper[4907]: I0313 15:42:09.228796 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bfe30c39-e9da-49f4-a9da-ebbce147b863-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 15:42:09 crc kubenswrapper[4907]: I0313 15:42:09.250939 4907 generic.go:334] "Generic (PLEG): container finished" podID="bfe30c39-e9da-49f4-a9da-ebbce147b863" containerID="24af8c70da41dc6cb88e8d25d6b679b4b7c4b0986765337bf4ab3605a412ae27" exitCode=0 Mar 13 15:42:09 crc kubenswrapper[4907]: I0313 15:42:09.250992 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"bfe30c39-e9da-49f4-a9da-ebbce147b863","Type":"ContainerDied","Data":"24af8c70da41dc6cb88e8d25d6b679b4b7c4b0986765337bf4ab3605a412ae27"} Mar 13 15:42:09 crc kubenswrapper[4907]: I0313 15:42:09.251029 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"bfe30c39-e9da-49f4-a9da-ebbce147b863","Type":"ContainerDied","Data":"36d4f2be9f852fdbc463601ca568186a38921963adf0237624ff766e5451ebed"} Mar 13 15:42:09 crc kubenswrapper[4907]: I0313 15:42:09.251049 4907 scope.go:117] "RemoveContainer" containerID="fbe67e4b2a6e6d29dd61f35ad336b329273792eab74bfcf4dd2ade78056599e6" Mar 13 15:42:09 crc kubenswrapper[4907]: I0313 15:42:09.251329 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Mar 13 15:42:09 crc kubenswrapper[4907]: I0313 15:42:09.291721 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Mar 13 15:42:09 crc kubenswrapper[4907]: I0313 15:42:09.296158 4907 scope.go:117] "RemoveContainer" containerID="24af8c70da41dc6cb88e8d25d6b679b4b7c4b0986765337bf4ab3605a412ae27" Mar 13 15:42:09 crc kubenswrapper[4907]: I0313 15:42:09.297642 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Mar 13 15:42:09 crc kubenswrapper[4907]: I0313 15:42:09.318141 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Mar 13 15:42:09 crc kubenswrapper[4907]: E0313 15:42:09.318601 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bfe30c39-e9da-49f4-a9da-ebbce147b863" containerName="probe" Mar 13 15:42:09 crc kubenswrapper[4907]: I0313 15:42:09.318621 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="bfe30c39-e9da-49f4-a9da-ebbce147b863" containerName="probe" Mar 13 15:42:09 crc kubenswrapper[4907]: E0313 15:42:09.318650 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bfe30c39-e9da-49f4-a9da-ebbce147b863" containerName="cinder-scheduler" Mar 13 15:42:09 crc kubenswrapper[4907]: I0313 15:42:09.318657 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="bfe30c39-e9da-49f4-a9da-ebbce147b863" containerName="cinder-scheduler" Mar 13 15:42:09 crc kubenswrapper[4907]: E0313 15:42:09.318671 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed0ed6c7-e5d4-4770-a9aa-308c0853573e" containerName="oc" Mar 13 15:42:09 crc kubenswrapper[4907]: I0313 15:42:09.318678 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed0ed6c7-e5d4-4770-a9aa-308c0853573e" containerName="oc" Mar 13 15:42:09 crc kubenswrapper[4907]: I0313 15:42:09.318903 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="bfe30c39-e9da-49f4-a9da-ebbce147b863" containerName="probe" Mar 13 15:42:09 crc kubenswrapper[4907]: I0313 15:42:09.318931 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed0ed6c7-e5d4-4770-a9aa-308c0853573e" containerName="oc" Mar 13 15:42:09 crc kubenswrapper[4907]: I0313 15:42:09.318942 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="bfe30c39-e9da-49f4-a9da-ebbce147b863" containerName="cinder-scheduler" Mar 13 15:42:09 crc kubenswrapper[4907]: I0313 15:42:09.320203 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Mar 13 15:42:09 crc kubenswrapper[4907]: I0313 15:42:09.325125 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Mar 13 15:42:09 crc kubenswrapper[4907]: I0313 15:42:09.329779 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Mar 13 15:42:09 crc kubenswrapper[4907]: I0313 15:42:09.330545 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3a999cb1-5df0-43ae-946f-eae3f2c53848-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"3a999cb1-5df0-43ae-946f-eae3f2c53848\") " pod="openstack/cinder-scheduler-0" Mar 13 15:42:09 crc kubenswrapper[4907]: I0313 15:42:09.330604 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a999cb1-5df0-43ae-946f-eae3f2c53848-config-data\") pod \"cinder-scheduler-0\" (UID: \"3a999cb1-5df0-43ae-946f-eae3f2c53848\") " pod="openstack/cinder-scheduler-0" Mar 13 15:42:09 crc kubenswrapper[4907]: I0313 15:42:09.330662 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wwxzz\" (UniqueName: \"kubernetes.io/projected/3a999cb1-5df0-43ae-946f-eae3f2c53848-kube-api-access-wwxzz\") pod \"cinder-scheduler-0\" (UID: \"3a999cb1-5df0-43ae-946f-eae3f2c53848\") " pod="openstack/cinder-scheduler-0" Mar 13 15:42:09 crc kubenswrapper[4907]: I0313 15:42:09.330686 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3a999cb1-5df0-43ae-946f-eae3f2c53848-scripts\") pod \"cinder-scheduler-0\" (UID: \"3a999cb1-5df0-43ae-946f-eae3f2c53848\") " pod="openstack/cinder-scheduler-0" Mar 13 15:42:09 crc kubenswrapper[4907]: I0313 15:42:09.330704 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a999cb1-5df0-43ae-946f-eae3f2c53848-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"3a999cb1-5df0-43ae-946f-eae3f2c53848\") " pod="openstack/cinder-scheduler-0" Mar 13 15:42:09 crc kubenswrapper[4907]: I0313 15:42:09.330764 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3a999cb1-5df0-43ae-946f-eae3f2c53848-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"3a999cb1-5df0-43ae-946f-eae3f2c53848\") " pod="openstack/cinder-scheduler-0" Mar 13 15:42:09 crc kubenswrapper[4907]: I0313 15:42:09.358026 4907 scope.go:117] "RemoveContainer" containerID="fbe67e4b2a6e6d29dd61f35ad336b329273792eab74bfcf4dd2ade78056599e6" Mar 13 15:42:09 crc kubenswrapper[4907]: E0313 15:42:09.359755 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fbe67e4b2a6e6d29dd61f35ad336b329273792eab74bfcf4dd2ade78056599e6\": container with ID starting with fbe67e4b2a6e6d29dd61f35ad336b329273792eab74bfcf4dd2ade78056599e6 not found: ID does not exist" containerID="fbe67e4b2a6e6d29dd61f35ad336b329273792eab74bfcf4dd2ade78056599e6" Mar 13 15:42:09 crc kubenswrapper[4907]: I0313 15:42:09.359808 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fbe67e4b2a6e6d29dd61f35ad336b329273792eab74bfcf4dd2ade78056599e6"} err="failed to get container status \"fbe67e4b2a6e6d29dd61f35ad336b329273792eab74bfcf4dd2ade78056599e6\": rpc error: code = NotFound desc = could not find container \"fbe67e4b2a6e6d29dd61f35ad336b329273792eab74bfcf4dd2ade78056599e6\": container with ID starting with fbe67e4b2a6e6d29dd61f35ad336b329273792eab74bfcf4dd2ade78056599e6 not found: ID does not exist" Mar 13 15:42:09 crc kubenswrapper[4907]: I0313 15:42:09.359838 4907 scope.go:117] "RemoveContainer" containerID="24af8c70da41dc6cb88e8d25d6b679b4b7c4b0986765337bf4ab3605a412ae27" Mar 13 15:42:09 crc kubenswrapper[4907]: E0313 15:42:09.361292 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"24af8c70da41dc6cb88e8d25d6b679b4b7c4b0986765337bf4ab3605a412ae27\": container with ID starting with 24af8c70da41dc6cb88e8d25d6b679b4b7c4b0986765337bf4ab3605a412ae27 not found: ID does not exist" containerID="24af8c70da41dc6cb88e8d25d6b679b4b7c4b0986765337bf4ab3605a412ae27" Mar 13 15:42:09 crc kubenswrapper[4907]: I0313 15:42:09.361332 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"24af8c70da41dc6cb88e8d25d6b679b4b7c4b0986765337bf4ab3605a412ae27"} err="failed to get container status \"24af8c70da41dc6cb88e8d25d6b679b4b7c4b0986765337bf4ab3605a412ae27\": rpc error: code = NotFound desc = could not find container \"24af8c70da41dc6cb88e8d25d6b679b4b7c4b0986765337bf4ab3605a412ae27\": container with ID starting with 24af8c70da41dc6cb88e8d25d6b679b4b7c4b0986765337bf4ab3605a412ae27 not found: ID does not exist" Mar 13 15:42:09 crc kubenswrapper[4907]: I0313 15:42:09.432914 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a999cb1-5df0-43ae-946f-eae3f2c53848-config-data\") pod \"cinder-scheduler-0\" (UID: \"3a999cb1-5df0-43ae-946f-eae3f2c53848\") " pod="openstack/cinder-scheduler-0" Mar 13 15:42:09 crc kubenswrapper[4907]: I0313 15:42:09.433042 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wwxzz\" (UniqueName: \"kubernetes.io/projected/3a999cb1-5df0-43ae-946f-eae3f2c53848-kube-api-access-wwxzz\") pod \"cinder-scheduler-0\" (UID: \"3a999cb1-5df0-43ae-946f-eae3f2c53848\") " pod="openstack/cinder-scheduler-0" Mar 13 15:42:09 crc kubenswrapper[4907]: I0313 15:42:09.433085 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3a999cb1-5df0-43ae-946f-eae3f2c53848-scripts\") pod \"cinder-scheduler-0\" (UID: \"3a999cb1-5df0-43ae-946f-eae3f2c53848\") " pod="openstack/cinder-scheduler-0" Mar 13 15:42:09 crc kubenswrapper[4907]: I0313 15:42:09.433108 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a999cb1-5df0-43ae-946f-eae3f2c53848-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"3a999cb1-5df0-43ae-946f-eae3f2c53848\") " pod="openstack/cinder-scheduler-0" Mar 13 15:42:09 crc kubenswrapper[4907]: I0313 15:42:09.433195 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3a999cb1-5df0-43ae-946f-eae3f2c53848-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"3a999cb1-5df0-43ae-946f-eae3f2c53848\") " pod="openstack/cinder-scheduler-0" Mar 13 15:42:09 crc kubenswrapper[4907]: I0313 15:42:09.433277 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3a999cb1-5df0-43ae-946f-eae3f2c53848-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"3a999cb1-5df0-43ae-946f-eae3f2c53848\") " pod="openstack/cinder-scheduler-0" Mar 13 15:42:09 crc kubenswrapper[4907]: I0313 15:42:09.433494 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3a999cb1-5df0-43ae-946f-eae3f2c53848-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"3a999cb1-5df0-43ae-946f-eae3f2c53848\") " pod="openstack/cinder-scheduler-0" Mar 13 15:42:09 crc kubenswrapper[4907]: I0313 15:42:09.437260 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a999cb1-5df0-43ae-946f-eae3f2c53848-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"3a999cb1-5df0-43ae-946f-eae3f2c53848\") " pod="openstack/cinder-scheduler-0" Mar 13 15:42:09 crc kubenswrapper[4907]: I0313 15:42:09.437526 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3a999cb1-5df0-43ae-946f-eae3f2c53848-scripts\") pod \"cinder-scheduler-0\" (UID: \"3a999cb1-5df0-43ae-946f-eae3f2c53848\") " pod="openstack/cinder-scheduler-0" Mar 13 15:42:09 crc kubenswrapper[4907]: I0313 15:42:09.438001 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a999cb1-5df0-43ae-946f-eae3f2c53848-config-data\") pod \"cinder-scheduler-0\" (UID: \"3a999cb1-5df0-43ae-946f-eae3f2c53848\") " pod="openstack/cinder-scheduler-0" Mar 13 15:42:09 crc kubenswrapper[4907]: I0313 15:42:09.442349 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3a999cb1-5df0-43ae-946f-eae3f2c53848-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"3a999cb1-5df0-43ae-946f-eae3f2c53848\") " pod="openstack/cinder-scheduler-0" Mar 13 15:42:09 crc kubenswrapper[4907]: I0313 15:42:09.450541 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wwxzz\" (UniqueName: \"kubernetes.io/projected/3a999cb1-5df0-43ae-946f-eae3f2c53848-kube-api-access-wwxzz\") pod \"cinder-scheduler-0\" (UID: \"3a999cb1-5df0-43ae-946f-eae3f2c53848\") " pod="openstack/cinder-scheduler-0" Mar 13 15:42:09 crc kubenswrapper[4907]: I0313 15:42:09.637069 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Mar 13 15:42:09 crc kubenswrapper[4907]: I0313 15:42:09.796589 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bfe30c39-e9da-49f4-a9da-ebbce147b863" path="/var/lib/kubelet/pods/bfe30c39-e9da-49f4-a9da-ebbce147b863/volumes" Mar 13 15:42:10 crc kubenswrapper[4907]: I0313 15:42:10.054128 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Mar 13 15:42:10 crc kubenswrapper[4907]: W0313 15:42:10.062349 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3a999cb1_5df0_43ae_946f_eae3f2c53848.slice/crio-61ecef7c54a689fe4149ac7c8e9ba61eb21ae1c4bc42a3f8e894b735b5cb869e WatchSource:0}: Error finding container 61ecef7c54a689fe4149ac7c8e9ba61eb21ae1c4bc42a3f8e894b735b5cb869e: Status 404 returned error can't find the container with id 61ecef7c54a689fe4149ac7c8e9ba61eb21ae1c4bc42a3f8e894b735b5cb869e Mar 13 15:42:10 crc kubenswrapper[4907]: I0313 15:42:10.260653 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"3a999cb1-5df0-43ae-946f-eae3f2c53848","Type":"ContainerStarted","Data":"61ecef7c54a689fe4149ac7c8e9ba61eb21ae1c4bc42a3f8e894b735b5cb869e"} Mar 13 15:42:11 crc kubenswrapper[4907]: I0313 15:42:11.274597 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"3a999cb1-5df0-43ae-946f-eae3f2c53848","Type":"ContainerStarted","Data":"3504104e47cbc2a010fa4b4eaa305d2b337e3a5bec3d3fca454e6618c2ac801b"} Mar 13 15:42:11 crc kubenswrapper[4907]: I0313 15:42:11.275099 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"3a999cb1-5df0-43ae-946f-eae3f2c53848","Type":"ContainerStarted","Data":"bf04a7270e06252486e2cbf6625bd906eddc33b84f8b5fb6a27a97ccc0c75eaf"} Mar 13 15:42:11 crc kubenswrapper[4907]: I0313 15:42:11.298170 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=2.298154383 podStartE2EDuration="2.298154383s" podCreationTimestamp="2026-03-13 15:42:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:42:11.297558508 +0000 UTC m=+5830.197346217" watchObservedRunningTime="2026-03-13 15:42:11.298154383 +0000 UTC m=+5830.197942072" Mar 13 15:42:11 crc kubenswrapper[4907]: I0313 15:42:11.786503 4907 scope.go:117] "RemoveContainer" containerID="68ffe92dd03078633ec89d7ef29da9bbdc4c5f04c956a1e8487e3f2e01c5d1c5" Mar 13 15:42:11 crc kubenswrapper[4907]: E0313 15:42:11.787113 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:42:13 crc kubenswrapper[4907]: I0313 15:42:13.468849 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Mar 13 15:42:14 crc kubenswrapper[4907]: I0313 15:42:14.637696 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Mar 13 15:42:19 crc kubenswrapper[4907]: I0313 15:42:19.822267 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Mar 13 15:42:22 crc kubenswrapper[4907]: I0313 15:42:22.782550 4907 scope.go:117] "RemoveContainer" containerID="68ffe92dd03078633ec89d7ef29da9bbdc4c5f04c956a1e8487e3f2e01c5d1c5" Mar 13 15:42:22 crc kubenswrapper[4907]: E0313 15:42:22.783091 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:42:29 crc kubenswrapper[4907]: I0313 15:42:29.064280 4907 scope.go:117] "RemoveContainer" containerID="a02ae219ab9b5a1fd678d76ac3b1c7e41a88bec1b6ae3e7a16334b35c97e0864" Mar 13 15:42:34 crc kubenswrapper[4907]: I0313 15:42:34.782254 4907 scope.go:117] "RemoveContainer" containerID="68ffe92dd03078633ec89d7ef29da9bbdc4c5f04c956a1e8487e3f2e01c5d1c5" Mar 13 15:42:34 crc kubenswrapper[4907]: E0313 15:42:34.782844 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:42:49 crc kubenswrapper[4907]: I0313 15:42:49.782998 4907 scope.go:117] "RemoveContainer" containerID="68ffe92dd03078633ec89d7ef29da9bbdc4c5f04c956a1e8487e3f2e01c5d1c5" Mar 13 15:42:49 crc kubenswrapper[4907]: E0313 15:42:49.783842 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:42:54 crc kubenswrapper[4907]: I0313 15:42:54.485222 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-4cjst"] Mar 13 15:42:54 crc kubenswrapper[4907]: I0313 15:42:54.490028 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4cjst" Mar 13 15:42:54 crc kubenswrapper[4907]: I0313 15:42:54.501742 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4cjst"] Mar 13 15:42:54 crc kubenswrapper[4907]: I0313 15:42:54.638512 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/24f03607-91fb-432b-afba-5d7c7634e98e-catalog-content\") pod \"redhat-marketplace-4cjst\" (UID: \"24f03607-91fb-432b-afba-5d7c7634e98e\") " pod="openshift-marketplace/redhat-marketplace-4cjst" Mar 13 15:42:54 crc kubenswrapper[4907]: I0313 15:42:54.639030 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lr6wt\" (UniqueName: \"kubernetes.io/projected/24f03607-91fb-432b-afba-5d7c7634e98e-kube-api-access-lr6wt\") pod \"redhat-marketplace-4cjst\" (UID: \"24f03607-91fb-432b-afba-5d7c7634e98e\") " pod="openshift-marketplace/redhat-marketplace-4cjst" Mar 13 15:42:54 crc kubenswrapper[4907]: I0313 15:42:54.639077 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/24f03607-91fb-432b-afba-5d7c7634e98e-utilities\") pod \"redhat-marketplace-4cjst\" (UID: \"24f03607-91fb-432b-afba-5d7c7634e98e\") " pod="openshift-marketplace/redhat-marketplace-4cjst" Mar 13 15:42:54 crc kubenswrapper[4907]: I0313 15:42:54.764764 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/24f03607-91fb-432b-afba-5d7c7634e98e-catalog-content\") pod \"redhat-marketplace-4cjst\" (UID: \"24f03607-91fb-432b-afba-5d7c7634e98e\") " pod="openshift-marketplace/redhat-marketplace-4cjst" Mar 13 15:42:54 crc kubenswrapper[4907]: I0313 15:42:54.764986 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lr6wt\" (UniqueName: \"kubernetes.io/projected/24f03607-91fb-432b-afba-5d7c7634e98e-kube-api-access-lr6wt\") pod \"redhat-marketplace-4cjst\" (UID: \"24f03607-91fb-432b-afba-5d7c7634e98e\") " pod="openshift-marketplace/redhat-marketplace-4cjst" Mar 13 15:42:54 crc kubenswrapper[4907]: I0313 15:42:54.765025 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/24f03607-91fb-432b-afba-5d7c7634e98e-utilities\") pod \"redhat-marketplace-4cjst\" (UID: \"24f03607-91fb-432b-afba-5d7c7634e98e\") " pod="openshift-marketplace/redhat-marketplace-4cjst" Mar 13 15:42:54 crc kubenswrapper[4907]: I0313 15:42:54.765388 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/24f03607-91fb-432b-afba-5d7c7634e98e-catalog-content\") pod \"redhat-marketplace-4cjst\" (UID: \"24f03607-91fb-432b-afba-5d7c7634e98e\") " pod="openshift-marketplace/redhat-marketplace-4cjst" Mar 13 15:42:54 crc kubenswrapper[4907]: I0313 15:42:54.765486 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/24f03607-91fb-432b-afba-5d7c7634e98e-utilities\") pod \"redhat-marketplace-4cjst\" (UID: \"24f03607-91fb-432b-afba-5d7c7634e98e\") " pod="openshift-marketplace/redhat-marketplace-4cjst" Mar 13 15:42:54 crc kubenswrapper[4907]: I0313 15:42:54.792986 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lr6wt\" (UniqueName: \"kubernetes.io/projected/24f03607-91fb-432b-afba-5d7c7634e98e-kube-api-access-lr6wt\") pod \"redhat-marketplace-4cjst\" (UID: \"24f03607-91fb-432b-afba-5d7c7634e98e\") " pod="openshift-marketplace/redhat-marketplace-4cjst" Mar 13 15:42:54 crc kubenswrapper[4907]: I0313 15:42:54.826297 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4cjst" Mar 13 15:42:55 crc kubenswrapper[4907]: I0313 15:42:55.361207 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4cjst"] Mar 13 15:42:55 crc kubenswrapper[4907]: W0313 15:42:55.364726 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod24f03607_91fb_432b_afba_5d7c7634e98e.slice/crio-7965e8c5457fc504594c88a465076bcea4651faf50c777df09b4aa7e11a81d7c WatchSource:0}: Error finding container 7965e8c5457fc504594c88a465076bcea4651faf50c777df09b4aa7e11a81d7c: Status 404 returned error can't find the container with id 7965e8c5457fc504594c88a465076bcea4651faf50c777df09b4aa7e11a81d7c Mar 13 15:42:55 crc kubenswrapper[4907]: I0313 15:42:55.691289 4907 generic.go:334] "Generic (PLEG): container finished" podID="24f03607-91fb-432b-afba-5d7c7634e98e" containerID="5ec383916b83bb60b7b67c55360eb9d838cf4da064b8916d7f76d953c53403c8" exitCode=0 Mar 13 15:42:55 crc kubenswrapper[4907]: I0313 15:42:55.691397 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4cjst" event={"ID":"24f03607-91fb-432b-afba-5d7c7634e98e","Type":"ContainerDied","Data":"5ec383916b83bb60b7b67c55360eb9d838cf4da064b8916d7f76d953c53403c8"} Mar 13 15:42:55 crc kubenswrapper[4907]: I0313 15:42:55.691668 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4cjst" event={"ID":"24f03607-91fb-432b-afba-5d7c7634e98e","Type":"ContainerStarted","Data":"7965e8c5457fc504594c88a465076bcea4651faf50c777df09b4aa7e11a81d7c"} Mar 13 15:42:55 crc kubenswrapper[4907]: I0313 15:42:55.693238 4907 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 13 15:42:57 crc kubenswrapper[4907]: I0313 15:42:57.714363 4907 generic.go:334] "Generic (PLEG): container finished" podID="24f03607-91fb-432b-afba-5d7c7634e98e" containerID="f1e605d4688b3741b8cfb0f2118afdc7b3de9f3ec5a9a3b8e5ba6a60f2026d01" exitCode=0 Mar 13 15:42:57 crc kubenswrapper[4907]: I0313 15:42:57.714540 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4cjst" event={"ID":"24f03607-91fb-432b-afba-5d7c7634e98e","Type":"ContainerDied","Data":"f1e605d4688b3741b8cfb0f2118afdc7b3de9f3ec5a9a3b8e5ba6a60f2026d01"} Mar 13 15:42:58 crc kubenswrapper[4907]: I0313 15:42:58.724064 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4cjst" event={"ID":"24f03607-91fb-432b-afba-5d7c7634e98e","Type":"ContainerStarted","Data":"ec5a19c81a6873a70bb62001169af40470c3baefad28d2afe99af6e6991ef588"} Mar 13 15:42:58 crc kubenswrapper[4907]: I0313 15:42:58.745686 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-4cjst" podStartSLOduration=2.312450315 podStartE2EDuration="4.745661094s" podCreationTimestamp="2026-03-13 15:42:54 +0000 UTC" firstStartedPulling="2026-03-13 15:42:55.693013771 +0000 UTC m=+5874.592801450" lastFinishedPulling="2026-03-13 15:42:58.12622454 +0000 UTC m=+5877.026012229" observedRunningTime="2026-03-13 15:42:58.738740415 +0000 UTC m=+5877.638528114" watchObservedRunningTime="2026-03-13 15:42:58.745661094 +0000 UTC m=+5877.645448793" Mar 13 15:43:02 crc kubenswrapper[4907]: I0313 15:43:02.782552 4907 scope.go:117] "RemoveContainer" containerID="68ffe92dd03078633ec89d7ef29da9bbdc4c5f04c956a1e8487e3f2e01c5d1c5" Mar 13 15:43:02 crc kubenswrapper[4907]: E0313 15:43:02.783599 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:43:04 crc kubenswrapper[4907]: I0313 15:43:04.827467 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-4cjst" Mar 13 15:43:04 crc kubenswrapper[4907]: I0313 15:43:04.827787 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-4cjst" Mar 13 15:43:04 crc kubenswrapper[4907]: I0313 15:43:04.870845 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-4cjst" Mar 13 15:43:05 crc kubenswrapper[4907]: I0313 15:43:05.835532 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-4cjst" Mar 13 15:43:05 crc kubenswrapper[4907]: I0313 15:43:05.893349 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4cjst"] Mar 13 15:43:07 crc kubenswrapper[4907]: I0313 15:43:07.811690 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-4cjst" podUID="24f03607-91fb-432b-afba-5d7c7634e98e" containerName="registry-server" containerID="cri-o://ec5a19c81a6873a70bb62001169af40470c3baefad28d2afe99af6e6991ef588" gracePeriod=2 Mar 13 15:43:08 crc kubenswrapper[4907]: I0313 15:43:08.296093 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4cjst" Mar 13 15:43:08 crc kubenswrapper[4907]: I0313 15:43:08.471236 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/24f03607-91fb-432b-afba-5d7c7634e98e-utilities\") pod \"24f03607-91fb-432b-afba-5d7c7634e98e\" (UID: \"24f03607-91fb-432b-afba-5d7c7634e98e\") " Mar 13 15:43:08 crc kubenswrapper[4907]: I0313 15:43:08.471289 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/24f03607-91fb-432b-afba-5d7c7634e98e-catalog-content\") pod \"24f03607-91fb-432b-afba-5d7c7634e98e\" (UID: \"24f03607-91fb-432b-afba-5d7c7634e98e\") " Mar 13 15:43:08 crc kubenswrapper[4907]: I0313 15:43:08.471422 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lr6wt\" (UniqueName: \"kubernetes.io/projected/24f03607-91fb-432b-afba-5d7c7634e98e-kube-api-access-lr6wt\") pod \"24f03607-91fb-432b-afba-5d7c7634e98e\" (UID: \"24f03607-91fb-432b-afba-5d7c7634e98e\") " Mar 13 15:43:08 crc kubenswrapper[4907]: I0313 15:43:08.472469 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/24f03607-91fb-432b-afba-5d7c7634e98e-utilities" (OuterVolumeSpecName: "utilities") pod "24f03607-91fb-432b-afba-5d7c7634e98e" (UID: "24f03607-91fb-432b-afba-5d7c7634e98e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:43:08 crc kubenswrapper[4907]: I0313 15:43:08.477607 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/24f03607-91fb-432b-afba-5d7c7634e98e-kube-api-access-lr6wt" (OuterVolumeSpecName: "kube-api-access-lr6wt") pod "24f03607-91fb-432b-afba-5d7c7634e98e" (UID: "24f03607-91fb-432b-afba-5d7c7634e98e"). InnerVolumeSpecName "kube-api-access-lr6wt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:43:08 crc kubenswrapper[4907]: I0313 15:43:08.495894 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/24f03607-91fb-432b-afba-5d7c7634e98e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "24f03607-91fb-432b-afba-5d7c7634e98e" (UID: "24f03607-91fb-432b-afba-5d7c7634e98e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:43:08 crc kubenswrapper[4907]: I0313 15:43:08.573510 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/24f03607-91fb-432b-afba-5d7c7634e98e-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 15:43:08 crc kubenswrapper[4907]: I0313 15:43:08.573553 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/24f03607-91fb-432b-afba-5d7c7634e98e-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 15:43:08 crc kubenswrapper[4907]: I0313 15:43:08.573566 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lr6wt\" (UniqueName: \"kubernetes.io/projected/24f03607-91fb-432b-afba-5d7c7634e98e-kube-api-access-lr6wt\") on node \"crc\" DevicePath \"\"" Mar 13 15:43:08 crc kubenswrapper[4907]: I0313 15:43:08.827134 4907 generic.go:334] "Generic (PLEG): container finished" podID="24f03607-91fb-432b-afba-5d7c7634e98e" containerID="ec5a19c81a6873a70bb62001169af40470c3baefad28d2afe99af6e6991ef588" exitCode=0 Mar 13 15:43:08 crc kubenswrapper[4907]: I0313 15:43:08.827190 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4cjst" Mar 13 15:43:08 crc kubenswrapper[4907]: I0313 15:43:08.827200 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4cjst" event={"ID":"24f03607-91fb-432b-afba-5d7c7634e98e","Type":"ContainerDied","Data":"ec5a19c81a6873a70bb62001169af40470c3baefad28d2afe99af6e6991ef588"} Mar 13 15:43:08 crc kubenswrapper[4907]: I0313 15:43:08.827242 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4cjst" event={"ID":"24f03607-91fb-432b-afba-5d7c7634e98e","Type":"ContainerDied","Data":"7965e8c5457fc504594c88a465076bcea4651faf50c777df09b4aa7e11a81d7c"} Mar 13 15:43:08 crc kubenswrapper[4907]: I0313 15:43:08.827270 4907 scope.go:117] "RemoveContainer" containerID="ec5a19c81a6873a70bb62001169af40470c3baefad28d2afe99af6e6991ef588" Mar 13 15:43:08 crc kubenswrapper[4907]: I0313 15:43:08.849013 4907 scope.go:117] "RemoveContainer" containerID="f1e605d4688b3741b8cfb0f2118afdc7b3de9f3ec5a9a3b8e5ba6a60f2026d01" Mar 13 15:43:08 crc kubenswrapper[4907]: I0313 15:43:08.869805 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4cjst"] Mar 13 15:43:08 crc kubenswrapper[4907]: I0313 15:43:08.882961 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-4cjst"] Mar 13 15:43:08 crc kubenswrapper[4907]: I0313 15:43:08.897439 4907 scope.go:117] "RemoveContainer" containerID="5ec383916b83bb60b7b67c55360eb9d838cf4da064b8916d7f76d953c53403c8" Mar 13 15:43:08 crc kubenswrapper[4907]: I0313 15:43:08.921895 4907 scope.go:117] "RemoveContainer" containerID="ec5a19c81a6873a70bb62001169af40470c3baefad28d2afe99af6e6991ef588" Mar 13 15:43:08 crc kubenswrapper[4907]: E0313 15:43:08.922394 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ec5a19c81a6873a70bb62001169af40470c3baefad28d2afe99af6e6991ef588\": container with ID starting with ec5a19c81a6873a70bb62001169af40470c3baefad28d2afe99af6e6991ef588 not found: ID does not exist" containerID="ec5a19c81a6873a70bb62001169af40470c3baefad28d2afe99af6e6991ef588" Mar 13 15:43:08 crc kubenswrapper[4907]: I0313 15:43:08.922426 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec5a19c81a6873a70bb62001169af40470c3baefad28d2afe99af6e6991ef588"} err="failed to get container status \"ec5a19c81a6873a70bb62001169af40470c3baefad28d2afe99af6e6991ef588\": rpc error: code = NotFound desc = could not find container \"ec5a19c81a6873a70bb62001169af40470c3baefad28d2afe99af6e6991ef588\": container with ID starting with ec5a19c81a6873a70bb62001169af40470c3baefad28d2afe99af6e6991ef588 not found: ID does not exist" Mar 13 15:43:08 crc kubenswrapper[4907]: I0313 15:43:08.922448 4907 scope.go:117] "RemoveContainer" containerID="f1e605d4688b3741b8cfb0f2118afdc7b3de9f3ec5a9a3b8e5ba6a60f2026d01" Mar 13 15:43:08 crc kubenswrapper[4907]: E0313 15:43:08.922834 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f1e605d4688b3741b8cfb0f2118afdc7b3de9f3ec5a9a3b8e5ba6a60f2026d01\": container with ID starting with f1e605d4688b3741b8cfb0f2118afdc7b3de9f3ec5a9a3b8e5ba6a60f2026d01 not found: ID does not exist" containerID="f1e605d4688b3741b8cfb0f2118afdc7b3de9f3ec5a9a3b8e5ba6a60f2026d01" Mar 13 15:43:08 crc kubenswrapper[4907]: I0313 15:43:08.922892 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f1e605d4688b3741b8cfb0f2118afdc7b3de9f3ec5a9a3b8e5ba6a60f2026d01"} err="failed to get container status \"f1e605d4688b3741b8cfb0f2118afdc7b3de9f3ec5a9a3b8e5ba6a60f2026d01\": rpc error: code = NotFound desc = could not find container \"f1e605d4688b3741b8cfb0f2118afdc7b3de9f3ec5a9a3b8e5ba6a60f2026d01\": container with ID starting with f1e605d4688b3741b8cfb0f2118afdc7b3de9f3ec5a9a3b8e5ba6a60f2026d01 not found: ID does not exist" Mar 13 15:43:08 crc kubenswrapper[4907]: I0313 15:43:08.922924 4907 scope.go:117] "RemoveContainer" containerID="5ec383916b83bb60b7b67c55360eb9d838cf4da064b8916d7f76d953c53403c8" Mar 13 15:43:08 crc kubenswrapper[4907]: E0313 15:43:08.923223 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5ec383916b83bb60b7b67c55360eb9d838cf4da064b8916d7f76d953c53403c8\": container with ID starting with 5ec383916b83bb60b7b67c55360eb9d838cf4da064b8916d7f76d953c53403c8 not found: ID does not exist" containerID="5ec383916b83bb60b7b67c55360eb9d838cf4da064b8916d7f76d953c53403c8" Mar 13 15:43:08 crc kubenswrapper[4907]: I0313 15:43:08.923248 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5ec383916b83bb60b7b67c55360eb9d838cf4da064b8916d7f76d953c53403c8"} err="failed to get container status \"5ec383916b83bb60b7b67c55360eb9d838cf4da064b8916d7f76d953c53403c8\": rpc error: code = NotFound desc = could not find container \"5ec383916b83bb60b7b67c55360eb9d838cf4da064b8916d7f76d953c53403c8\": container with ID starting with 5ec383916b83bb60b7b67c55360eb9d838cf4da064b8916d7f76d953c53403c8 not found: ID does not exist" Mar 13 15:43:09 crc kubenswrapper[4907]: I0313 15:43:09.795445 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="24f03607-91fb-432b-afba-5d7c7634e98e" path="/var/lib/kubelet/pods/24f03607-91fb-432b-afba-5d7c7634e98e/volumes" Mar 13 15:43:15 crc kubenswrapper[4907]: I0313 15:43:15.782903 4907 scope.go:117] "RemoveContainer" containerID="68ffe92dd03078633ec89d7ef29da9bbdc4c5f04c956a1e8487e3f2e01c5d1c5" Mar 13 15:43:15 crc kubenswrapper[4907]: E0313 15:43:15.783679 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:43:27 crc kubenswrapper[4907]: I0313 15:43:27.782450 4907 scope.go:117] "RemoveContainer" containerID="68ffe92dd03078633ec89d7ef29da9bbdc4c5f04c956a1e8487e3f2e01c5d1c5" Mar 13 15:43:27 crc kubenswrapper[4907]: E0313 15:43:27.783304 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:43:38 crc kubenswrapper[4907]: I0313 15:43:38.783079 4907 scope.go:117] "RemoveContainer" containerID="68ffe92dd03078633ec89d7ef29da9bbdc4c5f04c956a1e8487e3f2e01c5d1c5" Mar 13 15:43:38 crc kubenswrapper[4907]: E0313 15:43:38.783988 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:43:49 crc kubenswrapper[4907]: I0313 15:43:49.787991 4907 scope.go:117] "RemoveContainer" containerID="68ffe92dd03078633ec89d7ef29da9bbdc4c5f04c956a1e8487e3f2e01c5d1c5" Mar 13 15:43:49 crc kubenswrapper[4907]: E0313 15:43:49.788696 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:43:56 crc kubenswrapper[4907]: I0313 15:43:56.753203 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-zb9k2"] Mar 13 15:43:56 crc kubenswrapper[4907]: E0313 15:43:56.754108 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24f03607-91fb-432b-afba-5d7c7634e98e" containerName="extract-content" Mar 13 15:43:56 crc kubenswrapper[4907]: I0313 15:43:56.754122 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="24f03607-91fb-432b-afba-5d7c7634e98e" containerName="extract-content" Mar 13 15:43:56 crc kubenswrapper[4907]: E0313 15:43:56.754140 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24f03607-91fb-432b-afba-5d7c7634e98e" containerName="extract-utilities" Mar 13 15:43:56 crc kubenswrapper[4907]: I0313 15:43:56.754147 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="24f03607-91fb-432b-afba-5d7c7634e98e" containerName="extract-utilities" Mar 13 15:43:56 crc kubenswrapper[4907]: E0313 15:43:56.754168 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24f03607-91fb-432b-afba-5d7c7634e98e" containerName="registry-server" Mar 13 15:43:56 crc kubenswrapper[4907]: I0313 15:43:56.754175 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="24f03607-91fb-432b-afba-5d7c7634e98e" containerName="registry-server" Mar 13 15:43:56 crc kubenswrapper[4907]: I0313 15:43:56.754337 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="24f03607-91fb-432b-afba-5d7c7634e98e" containerName="registry-server" Mar 13 15:43:56 crc kubenswrapper[4907]: I0313 15:43:56.754924 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-zb9k2" Mar 13 15:43:56 crc kubenswrapper[4907]: I0313 15:43:56.761890 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-w5dfv" Mar 13 15:43:56 crc kubenswrapper[4907]: I0313 15:43:56.762336 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Mar 13 15:43:56 crc kubenswrapper[4907]: I0313 15:43:56.772805 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-v5b57"] Mar 13 15:43:56 crc kubenswrapper[4907]: I0313 15:43:56.775529 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-v5b57" Mar 13 15:43:56 crc kubenswrapper[4907]: I0313 15:43:56.789421 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-zb9k2"] Mar 13 15:43:56 crc kubenswrapper[4907]: I0313 15:43:56.805177 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-v5b57"] Mar 13 15:43:56 crc kubenswrapper[4907]: I0313 15:43:56.884491 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/1cf43c88-f066-4979-8590-b8e6668aab09-var-run\") pod \"ovn-controller-zb9k2\" (UID: \"1cf43c88-f066-4979-8590-b8e6668aab09\") " pod="openstack/ovn-controller-zb9k2" Mar 13 15:43:56 crc kubenswrapper[4907]: I0313 15:43:56.884537 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/34b897ab-ce41-4cd5-bcae-cea3c0dea9e9-etc-ovs\") pod \"ovn-controller-ovs-v5b57\" (UID: \"34b897ab-ce41-4cd5-bcae-cea3c0dea9e9\") " pod="openstack/ovn-controller-ovs-v5b57" Mar 13 15:43:56 crc kubenswrapper[4907]: I0313 15:43:56.884566 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/34b897ab-ce41-4cd5-bcae-cea3c0dea9e9-var-log\") pod \"ovn-controller-ovs-v5b57\" (UID: \"34b897ab-ce41-4cd5-bcae-cea3c0dea9e9\") " pod="openstack/ovn-controller-ovs-v5b57" Mar 13 15:43:56 crc kubenswrapper[4907]: I0313 15:43:56.884659 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/34b897ab-ce41-4cd5-bcae-cea3c0dea9e9-var-lib\") pod \"ovn-controller-ovs-v5b57\" (UID: \"34b897ab-ce41-4cd5-bcae-cea3c0dea9e9\") " pod="openstack/ovn-controller-ovs-v5b57" Mar 13 15:43:56 crc kubenswrapper[4907]: I0313 15:43:56.884698 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gzxds\" (UniqueName: \"kubernetes.io/projected/1cf43c88-f066-4979-8590-b8e6668aab09-kube-api-access-gzxds\") pod \"ovn-controller-zb9k2\" (UID: \"1cf43c88-f066-4979-8590-b8e6668aab09\") " pod="openstack/ovn-controller-zb9k2" Mar 13 15:43:56 crc kubenswrapper[4907]: I0313 15:43:56.884724 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1cf43c88-f066-4979-8590-b8e6668aab09-scripts\") pod \"ovn-controller-zb9k2\" (UID: \"1cf43c88-f066-4979-8590-b8e6668aab09\") " pod="openstack/ovn-controller-zb9k2" Mar 13 15:43:56 crc kubenswrapper[4907]: I0313 15:43:56.884750 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/34b897ab-ce41-4cd5-bcae-cea3c0dea9e9-scripts\") pod \"ovn-controller-ovs-v5b57\" (UID: \"34b897ab-ce41-4cd5-bcae-cea3c0dea9e9\") " pod="openstack/ovn-controller-ovs-v5b57" Mar 13 15:43:56 crc kubenswrapper[4907]: I0313 15:43:56.884784 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/1cf43c88-f066-4979-8590-b8e6668aab09-var-log-ovn\") pod \"ovn-controller-zb9k2\" (UID: \"1cf43c88-f066-4979-8590-b8e6668aab09\") " pod="openstack/ovn-controller-zb9k2" Mar 13 15:43:56 crc kubenswrapper[4907]: I0313 15:43:56.884854 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/1cf43c88-f066-4979-8590-b8e6668aab09-var-run-ovn\") pod \"ovn-controller-zb9k2\" (UID: \"1cf43c88-f066-4979-8590-b8e6668aab09\") " pod="openstack/ovn-controller-zb9k2" Mar 13 15:43:56 crc kubenswrapper[4907]: I0313 15:43:56.884914 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/34b897ab-ce41-4cd5-bcae-cea3c0dea9e9-var-run\") pod \"ovn-controller-ovs-v5b57\" (UID: \"34b897ab-ce41-4cd5-bcae-cea3c0dea9e9\") " pod="openstack/ovn-controller-ovs-v5b57" Mar 13 15:43:56 crc kubenswrapper[4907]: I0313 15:43:56.884952 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-92p4x\" (UniqueName: \"kubernetes.io/projected/34b897ab-ce41-4cd5-bcae-cea3c0dea9e9-kube-api-access-92p4x\") pod \"ovn-controller-ovs-v5b57\" (UID: \"34b897ab-ce41-4cd5-bcae-cea3c0dea9e9\") " pod="openstack/ovn-controller-ovs-v5b57" Mar 13 15:43:56 crc kubenswrapper[4907]: I0313 15:43:56.986379 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/1cf43c88-f066-4979-8590-b8e6668aab09-var-run\") pod \"ovn-controller-zb9k2\" (UID: \"1cf43c88-f066-4979-8590-b8e6668aab09\") " pod="openstack/ovn-controller-zb9k2" Mar 13 15:43:56 crc kubenswrapper[4907]: I0313 15:43:56.986431 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/34b897ab-ce41-4cd5-bcae-cea3c0dea9e9-etc-ovs\") pod \"ovn-controller-ovs-v5b57\" (UID: \"34b897ab-ce41-4cd5-bcae-cea3c0dea9e9\") " pod="openstack/ovn-controller-ovs-v5b57" Mar 13 15:43:56 crc kubenswrapper[4907]: I0313 15:43:56.986460 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/34b897ab-ce41-4cd5-bcae-cea3c0dea9e9-var-log\") pod \"ovn-controller-ovs-v5b57\" (UID: \"34b897ab-ce41-4cd5-bcae-cea3c0dea9e9\") " pod="openstack/ovn-controller-ovs-v5b57" Mar 13 15:43:56 crc kubenswrapper[4907]: I0313 15:43:56.986518 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/34b897ab-ce41-4cd5-bcae-cea3c0dea9e9-var-lib\") pod \"ovn-controller-ovs-v5b57\" (UID: \"34b897ab-ce41-4cd5-bcae-cea3c0dea9e9\") " pod="openstack/ovn-controller-ovs-v5b57" Mar 13 15:43:56 crc kubenswrapper[4907]: I0313 15:43:56.986558 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gzxds\" (UniqueName: \"kubernetes.io/projected/1cf43c88-f066-4979-8590-b8e6668aab09-kube-api-access-gzxds\") pod \"ovn-controller-zb9k2\" (UID: \"1cf43c88-f066-4979-8590-b8e6668aab09\") " pod="openstack/ovn-controller-zb9k2" Mar 13 15:43:56 crc kubenswrapper[4907]: I0313 15:43:56.986585 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1cf43c88-f066-4979-8590-b8e6668aab09-scripts\") pod \"ovn-controller-zb9k2\" (UID: \"1cf43c88-f066-4979-8590-b8e6668aab09\") " pod="openstack/ovn-controller-zb9k2" Mar 13 15:43:56 crc kubenswrapper[4907]: I0313 15:43:56.986612 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/34b897ab-ce41-4cd5-bcae-cea3c0dea9e9-scripts\") pod \"ovn-controller-ovs-v5b57\" (UID: \"34b897ab-ce41-4cd5-bcae-cea3c0dea9e9\") " pod="openstack/ovn-controller-ovs-v5b57" Mar 13 15:43:56 crc kubenswrapper[4907]: I0313 15:43:56.986645 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/1cf43c88-f066-4979-8590-b8e6668aab09-var-log-ovn\") pod \"ovn-controller-zb9k2\" (UID: \"1cf43c88-f066-4979-8590-b8e6668aab09\") " pod="openstack/ovn-controller-zb9k2" Mar 13 15:43:56 crc kubenswrapper[4907]: I0313 15:43:56.986680 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/1cf43c88-f066-4979-8590-b8e6668aab09-var-run-ovn\") pod \"ovn-controller-zb9k2\" (UID: \"1cf43c88-f066-4979-8590-b8e6668aab09\") " pod="openstack/ovn-controller-zb9k2" Mar 13 15:43:56 crc kubenswrapper[4907]: I0313 15:43:56.986707 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/34b897ab-ce41-4cd5-bcae-cea3c0dea9e9-var-run\") pod \"ovn-controller-ovs-v5b57\" (UID: \"34b897ab-ce41-4cd5-bcae-cea3c0dea9e9\") " pod="openstack/ovn-controller-ovs-v5b57" Mar 13 15:43:56 crc kubenswrapper[4907]: I0313 15:43:56.986738 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-92p4x\" (UniqueName: \"kubernetes.io/projected/34b897ab-ce41-4cd5-bcae-cea3c0dea9e9-kube-api-access-92p4x\") pod \"ovn-controller-ovs-v5b57\" (UID: \"34b897ab-ce41-4cd5-bcae-cea3c0dea9e9\") " pod="openstack/ovn-controller-ovs-v5b57" Mar 13 15:43:56 crc kubenswrapper[4907]: I0313 15:43:56.987381 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/34b897ab-ce41-4cd5-bcae-cea3c0dea9e9-etc-ovs\") pod \"ovn-controller-ovs-v5b57\" (UID: \"34b897ab-ce41-4cd5-bcae-cea3c0dea9e9\") " pod="openstack/ovn-controller-ovs-v5b57" Mar 13 15:43:56 crc kubenswrapper[4907]: I0313 15:43:56.987406 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/1cf43c88-f066-4979-8590-b8e6668aab09-var-run\") pod \"ovn-controller-zb9k2\" (UID: \"1cf43c88-f066-4979-8590-b8e6668aab09\") " pod="openstack/ovn-controller-zb9k2" Mar 13 15:43:56 crc kubenswrapper[4907]: I0313 15:43:56.987394 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/34b897ab-ce41-4cd5-bcae-cea3c0dea9e9-var-log\") pod \"ovn-controller-ovs-v5b57\" (UID: \"34b897ab-ce41-4cd5-bcae-cea3c0dea9e9\") " pod="openstack/ovn-controller-ovs-v5b57" Mar 13 15:43:56 crc kubenswrapper[4907]: I0313 15:43:56.987491 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/1cf43c88-f066-4979-8590-b8e6668aab09-var-log-ovn\") pod \"ovn-controller-zb9k2\" (UID: \"1cf43c88-f066-4979-8590-b8e6668aab09\") " pod="openstack/ovn-controller-zb9k2" Mar 13 15:43:56 crc kubenswrapper[4907]: I0313 15:43:56.987508 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/1cf43c88-f066-4979-8590-b8e6668aab09-var-run-ovn\") pod \"ovn-controller-zb9k2\" (UID: \"1cf43c88-f066-4979-8590-b8e6668aab09\") " pod="openstack/ovn-controller-zb9k2" Mar 13 15:43:56 crc kubenswrapper[4907]: I0313 15:43:56.987574 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/34b897ab-ce41-4cd5-bcae-cea3c0dea9e9-var-run\") pod \"ovn-controller-ovs-v5b57\" (UID: \"34b897ab-ce41-4cd5-bcae-cea3c0dea9e9\") " pod="openstack/ovn-controller-ovs-v5b57" Mar 13 15:43:56 crc kubenswrapper[4907]: I0313 15:43:56.987680 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/34b897ab-ce41-4cd5-bcae-cea3c0dea9e9-var-lib\") pod \"ovn-controller-ovs-v5b57\" (UID: \"34b897ab-ce41-4cd5-bcae-cea3c0dea9e9\") " pod="openstack/ovn-controller-ovs-v5b57" Mar 13 15:43:56 crc kubenswrapper[4907]: I0313 15:43:56.989500 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/34b897ab-ce41-4cd5-bcae-cea3c0dea9e9-scripts\") pod \"ovn-controller-ovs-v5b57\" (UID: \"34b897ab-ce41-4cd5-bcae-cea3c0dea9e9\") " pod="openstack/ovn-controller-ovs-v5b57" Mar 13 15:43:56 crc kubenswrapper[4907]: I0313 15:43:56.989797 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1cf43c88-f066-4979-8590-b8e6668aab09-scripts\") pod \"ovn-controller-zb9k2\" (UID: \"1cf43c88-f066-4979-8590-b8e6668aab09\") " pod="openstack/ovn-controller-zb9k2" Mar 13 15:43:57 crc kubenswrapper[4907]: I0313 15:43:57.006850 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gzxds\" (UniqueName: \"kubernetes.io/projected/1cf43c88-f066-4979-8590-b8e6668aab09-kube-api-access-gzxds\") pod \"ovn-controller-zb9k2\" (UID: \"1cf43c88-f066-4979-8590-b8e6668aab09\") " pod="openstack/ovn-controller-zb9k2" Mar 13 15:43:57 crc kubenswrapper[4907]: I0313 15:43:57.015458 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-92p4x\" (UniqueName: \"kubernetes.io/projected/34b897ab-ce41-4cd5-bcae-cea3c0dea9e9-kube-api-access-92p4x\") pod \"ovn-controller-ovs-v5b57\" (UID: \"34b897ab-ce41-4cd5-bcae-cea3c0dea9e9\") " pod="openstack/ovn-controller-ovs-v5b57" Mar 13 15:43:57 crc kubenswrapper[4907]: I0313 15:43:57.075107 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-zb9k2" Mar 13 15:43:57 crc kubenswrapper[4907]: I0313 15:43:57.091061 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-v5b57" Mar 13 15:43:57 crc kubenswrapper[4907]: I0313 15:43:57.554767 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-zb9k2"] Mar 13 15:43:58 crc kubenswrapper[4907]: I0313 15:43:58.062153 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-v5b57"] Mar 13 15:43:58 crc kubenswrapper[4907]: W0313 15:43:58.066324 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod34b897ab_ce41_4cd5_bcae_cea3c0dea9e9.slice/crio-16325fb744d63b685dea9ff2553f2408a5735d291ee63bfc31b4a34d96d3664c WatchSource:0}: Error finding container 16325fb744d63b685dea9ff2553f2408a5735d291ee63bfc31b4a34d96d3664c: Status 404 returned error can't find the container with id 16325fb744d63b685dea9ff2553f2408a5735d291ee63bfc31b4a34d96d3664c Mar 13 15:43:58 crc kubenswrapper[4907]: I0313 15:43:58.335630 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-v5b57" event={"ID":"34b897ab-ce41-4cd5-bcae-cea3c0dea9e9","Type":"ContainerStarted","Data":"16325fb744d63b685dea9ff2553f2408a5735d291ee63bfc31b4a34d96d3664c"} Mar 13 15:43:58 crc kubenswrapper[4907]: I0313 15:43:58.337193 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-zb9k2" event={"ID":"1cf43c88-f066-4979-8590-b8e6668aab09","Type":"ContainerStarted","Data":"666a17c67dac296199ae01f33215fddd1a37e44ab70056e8156c835ff9a3812b"} Mar 13 15:43:58 crc kubenswrapper[4907]: I0313 15:43:58.337236 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-zb9k2" event={"ID":"1cf43c88-f066-4979-8590-b8e6668aab09","Type":"ContainerStarted","Data":"3b7c23409163ae5d10332502c897514fd1d764456aac8e68f45261995b6934b5"} Mar 13 15:43:58 crc kubenswrapper[4907]: I0313 15:43:58.337533 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-zb9k2" Mar 13 15:43:58 crc kubenswrapper[4907]: I0313 15:43:58.360319 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-zb9k2" podStartSLOduration=2.360303212 podStartE2EDuration="2.360303212s" podCreationTimestamp="2026-03-13 15:43:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:43:58.355567592 +0000 UTC m=+5937.255355321" watchObservedRunningTime="2026-03-13 15:43:58.360303212 +0000 UTC m=+5937.260090901" Mar 13 15:43:59 crc kubenswrapper[4907]: I0313 15:43:59.369967 4907 generic.go:334] "Generic (PLEG): container finished" podID="34b897ab-ce41-4cd5-bcae-cea3c0dea9e9" containerID="fecc41f131bc004354ba494350d5e8f1e25a264511e3c8936fa8c28deae60120" exitCode=0 Mar 13 15:43:59 crc kubenswrapper[4907]: I0313 15:43:59.370977 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-v5b57" event={"ID":"34b897ab-ce41-4cd5-bcae-cea3c0dea9e9","Type":"ContainerDied","Data":"fecc41f131bc004354ba494350d5e8f1e25a264511e3c8936fa8c28deae60120"} Mar 13 15:43:59 crc kubenswrapper[4907]: I0313 15:43:59.374038 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-xstlm"] Mar 13 15:43:59 crc kubenswrapper[4907]: I0313 15:43:59.397331 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-xstlm" Mar 13 15:43:59 crc kubenswrapper[4907]: I0313 15:43:59.409318 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Mar 13 15:43:59 crc kubenswrapper[4907]: I0313 15:43:59.422988 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-xstlm"] Mar 13 15:43:59 crc kubenswrapper[4907]: I0313 15:43:59.444371 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pvq2t\" (UniqueName: \"kubernetes.io/projected/f9398439-2570-42ec-b6fc-f9770b988c73-kube-api-access-pvq2t\") pod \"ovn-controller-metrics-xstlm\" (UID: \"f9398439-2570-42ec-b6fc-f9770b988c73\") " pod="openstack/ovn-controller-metrics-xstlm" Mar 13 15:43:59 crc kubenswrapper[4907]: I0313 15:43:59.444592 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f9398439-2570-42ec-b6fc-f9770b988c73-config\") pod \"ovn-controller-metrics-xstlm\" (UID: \"f9398439-2570-42ec-b6fc-f9770b988c73\") " pod="openstack/ovn-controller-metrics-xstlm" Mar 13 15:43:59 crc kubenswrapper[4907]: I0313 15:43:59.444661 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/f9398439-2570-42ec-b6fc-f9770b988c73-ovs-rundir\") pod \"ovn-controller-metrics-xstlm\" (UID: \"f9398439-2570-42ec-b6fc-f9770b988c73\") " pod="openstack/ovn-controller-metrics-xstlm" Mar 13 15:43:59 crc kubenswrapper[4907]: I0313 15:43:59.444711 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/f9398439-2570-42ec-b6fc-f9770b988c73-ovn-rundir\") pod \"ovn-controller-metrics-xstlm\" (UID: \"f9398439-2570-42ec-b6fc-f9770b988c73\") " pod="openstack/ovn-controller-metrics-xstlm" Mar 13 15:43:59 crc kubenswrapper[4907]: I0313 15:43:59.546006 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pvq2t\" (UniqueName: \"kubernetes.io/projected/f9398439-2570-42ec-b6fc-f9770b988c73-kube-api-access-pvq2t\") pod \"ovn-controller-metrics-xstlm\" (UID: \"f9398439-2570-42ec-b6fc-f9770b988c73\") " pod="openstack/ovn-controller-metrics-xstlm" Mar 13 15:43:59 crc kubenswrapper[4907]: I0313 15:43:59.546094 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f9398439-2570-42ec-b6fc-f9770b988c73-config\") pod \"ovn-controller-metrics-xstlm\" (UID: \"f9398439-2570-42ec-b6fc-f9770b988c73\") " pod="openstack/ovn-controller-metrics-xstlm" Mar 13 15:43:59 crc kubenswrapper[4907]: I0313 15:43:59.546129 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/f9398439-2570-42ec-b6fc-f9770b988c73-ovs-rundir\") pod \"ovn-controller-metrics-xstlm\" (UID: \"f9398439-2570-42ec-b6fc-f9770b988c73\") " pod="openstack/ovn-controller-metrics-xstlm" Mar 13 15:43:59 crc kubenswrapper[4907]: I0313 15:43:59.546158 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/f9398439-2570-42ec-b6fc-f9770b988c73-ovn-rundir\") pod \"ovn-controller-metrics-xstlm\" (UID: \"f9398439-2570-42ec-b6fc-f9770b988c73\") " pod="openstack/ovn-controller-metrics-xstlm" Mar 13 15:43:59 crc kubenswrapper[4907]: I0313 15:43:59.546440 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/f9398439-2570-42ec-b6fc-f9770b988c73-ovn-rundir\") pod \"ovn-controller-metrics-xstlm\" (UID: \"f9398439-2570-42ec-b6fc-f9770b988c73\") " pod="openstack/ovn-controller-metrics-xstlm" Mar 13 15:43:59 crc kubenswrapper[4907]: I0313 15:43:59.546810 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/f9398439-2570-42ec-b6fc-f9770b988c73-ovs-rundir\") pod \"ovn-controller-metrics-xstlm\" (UID: \"f9398439-2570-42ec-b6fc-f9770b988c73\") " pod="openstack/ovn-controller-metrics-xstlm" Mar 13 15:43:59 crc kubenswrapper[4907]: I0313 15:43:59.547517 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f9398439-2570-42ec-b6fc-f9770b988c73-config\") pod \"ovn-controller-metrics-xstlm\" (UID: \"f9398439-2570-42ec-b6fc-f9770b988c73\") " pod="openstack/ovn-controller-metrics-xstlm" Mar 13 15:43:59 crc kubenswrapper[4907]: I0313 15:43:59.575226 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pvq2t\" (UniqueName: \"kubernetes.io/projected/f9398439-2570-42ec-b6fc-f9770b988c73-kube-api-access-pvq2t\") pod \"ovn-controller-metrics-xstlm\" (UID: \"f9398439-2570-42ec-b6fc-f9770b988c73\") " pod="openstack/ovn-controller-metrics-xstlm" Mar 13 15:43:59 crc kubenswrapper[4907]: I0313 15:43:59.744605 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-xstlm" Mar 13 15:44:00 crc kubenswrapper[4907]: I0313 15:44:00.129651 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556944-gcp6x"] Mar 13 15:44:00 crc kubenswrapper[4907]: I0313 15:44:00.130969 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556944-gcp6x" Mar 13 15:44:00 crc kubenswrapper[4907]: I0313 15:44:00.133447 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 15:44:00 crc kubenswrapper[4907]: I0313 15:44:00.133688 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 15:44:00 crc kubenswrapper[4907]: I0313 15:44:00.133912 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 15:44:00 crc kubenswrapper[4907]: I0313 15:44:00.147652 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556944-gcp6x"] Mar 13 15:44:00 crc kubenswrapper[4907]: I0313 15:44:00.207986 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-xstlm"] Mar 13 15:44:00 crc kubenswrapper[4907]: W0313 15:44:00.214410 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf9398439_2570_42ec_b6fc_f9770b988c73.slice/crio-8afc6916ec9a9c7dd33fc43ed1b15de0c05e625f1672700c2abd0b3864a725b3 WatchSource:0}: Error finding container 8afc6916ec9a9c7dd33fc43ed1b15de0c05e625f1672700c2abd0b3864a725b3: Status 404 returned error can't find the container with id 8afc6916ec9a9c7dd33fc43ed1b15de0c05e625f1672700c2abd0b3864a725b3 Mar 13 15:44:00 crc kubenswrapper[4907]: I0313 15:44:00.257276 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-76z5k\" (UniqueName: \"kubernetes.io/projected/04ca577b-ab8d-44a6-a598-1297bd5ab664-kube-api-access-76z5k\") pod \"auto-csr-approver-29556944-gcp6x\" (UID: \"04ca577b-ab8d-44a6-a598-1297bd5ab664\") " pod="openshift-infra/auto-csr-approver-29556944-gcp6x" Mar 13 15:44:00 crc kubenswrapper[4907]: I0313 15:44:00.359027 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-76z5k\" (UniqueName: \"kubernetes.io/projected/04ca577b-ab8d-44a6-a598-1297bd5ab664-kube-api-access-76z5k\") pod \"auto-csr-approver-29556944-gcp6x\" (UID: \"04ca577b-ab8d-44a6-a598-1297bd5ab664\") " pod="openshift-infra/auto-csr-approver-29556944-gcp6x" Mar 13 15:44:00 crc kubenswrapper[4907]: I0313 15:44:00.378591 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-76z5k\" (UniqueName: \"kubernetes.io/projected/04ca577b-ab8d-44a6-a598-1297bd5ab664-kube-api-access-76z5k\") pod \"auto-csr-approver-29556944-gcp6x\" (UID: \"04ca577b-ab8d-44a6-a598-1297bd5ab664\") " pod="openshift-infra/auto-csr-approver-29556944-gcp6x" Mar 13 15:44:00 crc kubenswrapper[4907]: I0313 15:44:00.383047 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-v5b57" event={"ID":"34b897ab-ce41-4cd5-bcae-cea3c0dea9e9","Type":"ContainerStarted","Data":"c66b2e74b53f59150cce15506d2f46e588dab64d0fa13fd691a426405cb7b6a1"} Mar 13 15:44:00 crc kubenswrapper[4907]: I0313 15:44:00.383094 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-v5b57" event={"ID":"34b897ab-ce41-4cd5-bcae-cea3c0dea9e9","Type":"ContainerStarted","Data":"d1fcb80a9f83604cd9cc526d40f3aeec6026f7c8772347b1064ff90d72289853"} Mar 13 15:44:00 crc kubenswrapper[4907]: I0313 15:44:00.383428 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-v5b57" Mar 13 15:44:00 crc kubenswrapper[4907]: I0313 15:44:00.383462 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-v5b57" Mar 13 15:44:00 crc kubenswrapper[4907]: I0313 15:44:00.384759 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-xstlm" event={"ID":"f9398439-2570-42ec-b6fc-f9770b988c73","Type":"ContainerStarted","Data":"8afc6916ec9a9c7dd33fc43ed1b15de0c05e625f1672700c2abd0b3864a725b3"} Mar 13 15:44:00 crc kubenswrapper[4907]: I0313 15:44:00.469473 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556944-gcp6x" Mar 13 15:44:00 crc kubenswrapper[4907]: I0313 15:44:00.923733 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-v5b57" podStartSLOduration=4.923709548 podStartE2EDuration="4.923709548s" podCreationTimestamp="2026-03-13 15:43:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:44:00.407113624 +0000 UTC m=+5939.306901353" watchObservedRunningTime="2026-03-13 15:44:00.923709548 +0000 UTC m=+5939.823497247" Mar 13 15:44:00 crc kubenswrapper[4907]: I0313 15:44:00.935827 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556944-gcp6x"] Mar 13 15:44:00 crc kubenswrapper[4907]: W0313 15:44:00.948212 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod04ca577b_ab8d_44a6_a598_1297bd5ab664.slice/crio-bd9a2d3454311f65870318872731b5b1e0ae357a02e57d212f00b5e59db8956f WatchSource:0}: Error finding container bd9a2d3454311f65870318872731b5b1e0ae357a02e57d212f00b5e59db8956f: Status 404 returned error can't find the container with id bd9a2d3454311f65870318872731b5b1e0ae357a02e57d212f00b5e59db8956f Mar 13 15:44:01 crc kubenswrapper[4907]: I0313 15:44:01.050923 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-48f2-account-create-update-wvks8"] Mar 13 15:44:01 crc kubenswrapper[4907]: I0313 15:44:01.059898 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-gxjv5"] Mar 13 15:44:01 crc kubenswrapper[4907]: I0313 15:44:01.070482 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-48f2-account-create-update-wvks8"] Mar 13 15:44:01 crc kubenswrapper[4907]: I0313 15:44:01.079063 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-gxjv5"] Mar 13 15:44:01 crc kubenswrapper[4907]: I0313 15:44:01.396071 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-xstlm" event={"ID":"f9398439-2570-42ec-b6fc-f9770b988c73","Type":"ContainerStarted","Data":"b2340f62039b5ff0165b80754d58582bd36311f660eac8476fec48aa23ac78bf"} Mar 13 15:44:01 crc kubenswrapper[4907]: I0313 15:44:01.397918 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556944-gcp6x" event={"ID":"04ca577b-ab8d-44a6-a598-1297bd5ab664","Type":"ContainerStarted","Data":"bd9a2d3454311f65870318872731b5b1e0ae357a02e57d212f00b5e59db8956f"} Mar 13 15:44:01 crc kubenswrapper[4907]: I0313 15:44:01.419908 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-xstlm" podStartSLOduration=2.419872174 podStartE2EDuration="2.419872174s" podCreationTimestamp="2026-03-13 15:43:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:44:01.410057406 +0000 UTC m=+5940.309845095" watchObservedRunningTime="2026-03-13 15:44:01.419872174 +0000 UTC m=+5940.319659883" Mar 13 15:44:01 crc kubenswrapper[4907]: I0313 15:44:01.795361 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9c51cd5f-cfaf-4c1a-9aa2-e7ade1b79058" path="/var/lib/kubelet/pods/9c51cd5f-cfaf-4c1a-9aa2-e7ade1b79058/volumes" Mar 13 15:44:01 crc kubenswrapper[4907]: I0313 15:44:01.796808 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b03a1328-cfc5-4ff6-88b9-6523f4378708" path="/var/lib/kubelet/pods/b03a1328-cfc5-4ff6-88b9-6523f4378708/volumes" Mar 13 15:44:02 crc kubenswrapper[4907]: I0313 15:44:02.406294 4907 generic.go:334] "Generic (PLEG): container finished" podID="04ca577b-ab8d-44a6-a598-1297bd5ab664" containerID="7157d1dc97b867b0493ef78bf3ed0dea2af629c4ae1bace22c4fa35387dfbe4f" exitCode=0 Mar 13 15:44:02 crc kubenswrapper[4907]: I0313 15:44:02.406381 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556944-gcp6x" event={"ID":"04ca577b-ab8d-44a6-a598-1297bd5ab664","Type":"ContainerDied","Data":"7157d1dc97b867b0493ef78bf3ed0dea2af629c4ae1bace22c4fa35387dfbe4f"} Mar 13 15:44:03 crc kubenswrapper[4907]: I0313 15:44:03.702082 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556944-gcp6x" Mar 13 15:44:03 crc kubenswrapper[4907]: I0313 15:44:03.831151 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-76z5k\" (UniqueName: \"kubernetes.io/projected/04ca577b-ab8d-44a6-a598-1297bd5ab664-kube-api-access-76z5k\") pod \"04ca577b-ab8d-44a6-a598-1297bd5ab664\" (UID: \"04ca577b-ab8d-44a6-a598-1297bd5ab664\") " Mar 13 15:44:03 crc kubenswrapper[4907]: I0313 15:44:03.836109 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/04ca577b-ab8d-44a6-a598-1297bd5ab664-kube-api-access-76z5k" (OuterVolumeSpecName: "kube-api-access-76z5k") pod "04ca577b-ab8d-44a6-a598-1297bd5ab664" (UID: "04ca577b-ab8d-44a6-a598-1297bd5ab664"). InnerVolumeSpecName "kube-api-access-76z5k". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:44:03 crc kubenswrapper[4907]: I0313 15:44:03.933388 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-76z5k\" (UniqueName: \"kubernetes.io/projected/04ca577b-ab8d-44a6-a598-1297bd5ab664-kube-api-access-76z5k\") on node \"crc\" DevicePath \"\"" Mar 13 15:44:04 crc kubenswrapper[4907]: I0313 15:44:04.434608 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556944-gcp6x" event={"ID":"04ca577b-ab8d-44a6-a598-1297bd5ab664","Type":"ContainerDied","Data":"bd9a2d3454311f65870318872731b5b1e0ae357a02e57d212f00b5e59db8956f"} Mar 13 15:44:04 crc kubenswrapper[4907]: I0313 15:44:04.434658 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bd9a2d3454311f65870318872731b5b1e0ae357a02e57d212f00b5e59db8956f" Mar 13 15:44:04 crc kubenswrapper[4907]: I0313 15:44:04.434667 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556944-gcp6x" Mar 13 15:44:04 crc kubenswrapper[4907]: I0313 15:44:04.758693 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556938-ptksb"] Mar 13 15:44:04 crc kubenswrapper[4907]: I0313 15:44:04.766045 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556938-ptksb"] Mar 13 15:44:04 crc kubenswrapper[4907]: I0313 15:44:04.783090 4907 scope.go:117] "RemoveContainer" containerID="68ffe92dd03078633ec89d7ef29da9bbdc4c5f04c956a1e8487e3f2e01c5d1c5" Mar 13 15:44:04 crc kubenswrapper[4907]: E0313 15:44:04.783484 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:44:05 crc kubenswrapper[4907]: I0313 15:44:05.795000 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6a5048c0-3add-450a-94c0-c068ceca72a4" path="/var/lib/kubelet/pods/6a5048c0-3add-450a-94c0-c068ceca72a4/volumes" Mar 13 15:44:07 crc kubenswrapper[4907]: I0313 15:44:07.029982 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-4gtrw"] Mar 13 15:44:07 crc kubenswrapper[4907]: I0313 15:44:07.042097 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-4gtrw"] Mar 13 15:44:07 crc kubenswrapper[4907]: I0313 15:44:07.796872 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f23a73ee-57ff-4d58-8812-b72f624b7739" path="/var/lib/kubelet/pods/f23a73ee-57ff-4d58-8812-b72f624b7739/volumes" Mar 13 15:44:19 crc kubenswrapper[4907]: I0313 15:44:19.786511 4907 scope.go:117] "RemoveContainer" containerID="68ffe92dd03078633ec89d7ef29da9bbdc4c5f04c956a1e8487e3f2e01c5d1c5" Mar 13 15:44:19 crc kubenswrapper[4907]: E0313 15:44:19.787126 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:44:20 crc kubenswrapper[4907]: I0313 15:44:20.060103 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-mf86w"] Mar 13 15:44:20 crc kubenswrapper[4907]: I0313 15:44:20.067440 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-mf86w"] Mar 13 15:44:21 crc kubenswrapper[4907]: I0313 15:44:21.194858 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-db-create-ghczh"] Mar 13 15:44:21 crc kubenswrapper[4907]: E0313 15:44:21.195583 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04ca577b-ab8d-44a6-a598-1297bd5ab664" containerName="oc" Mar 13 15:44:21 crc kubenswrapper[4907]: I0313 15:44:21.195599 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="04ca577b-ab8d-44a6-a598-1297bd5ab664" containerName="oc" Mar 13 15:44:21 crc kubenswrapper[4907]: I0313 15:44:21.195851 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="04ca577b-ab8d-44a6-a598-1297bd5ab664" containerName="oc" Mar 13 15:44:21 crc kubenswrapper[4907]: I0313 15:44:21.196595 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-create-ghczh" Mar 13 15:44:21 crc kubenswrapper[4907]: I0313 15:44:21.207231 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-db-create-ghczh"] Mar 13 15:44:21 crc kubenswrapper[4907]: I0313 15:44:21.279169 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l5kw2\" (UniqueName: \"kubernetes.io/projected/0aa6e501-c175-44be-af00-02b8aac76bab-kube-api-access-l5kw2\") pod \"octavia-db-create-ghczh\" (UID: \"0aa6e501-c175-44be-af00-02b8aac76bab\") " pod="openstack/octavia-db-create-ghczh" Mar 13 15:44:21 crc kubenswrapper[4907]: I0313 15:44:21.279292 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0aa6e501-c175-44be-af00-02b8aac76bab-operator-scripts\") pod \"octavia-db-create-ghczh\" (UID: \"0aa6e501-c175-44be-af00-02b8aac76bab\") " pod="openstack/octavia-db-create-ghczh" Mar 13 15:44:21 crc kubenswrapper[4907]: I0313 15:44:21.381539 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l5kw2\" (UniqueName: \"kubernetes.io/projected/0aa6e501-c175-44be-af00-02b8aac76bab-kube-api-access-l5kw2\") pod \"octavia-db-create-ghczh\" (UID: \"0aa6e501-c175-44be-af00-02b8aac76bab\") " pod="openstack/octavia-db-create-ghczh" Mar 13 15:44:21 crc kubenswrapper[4907]: I0313 15:44:21.382017 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0aa6e501-c175-44be-af00-02b8aac76bab-operator-scripts\") pod \"octavia-db-create-ghczh\" (UID: \"0aa6e501-c175-44be-af00-02b8aac76bab\") " pod="openstack/octavia-db-create-ghczh" Mar 13 15:44:21 crc kubenswrapper[4907]: I0313 15:44:21.382732 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0aa6e501-c175-44be-af00-02b8aac76bab-operator-scripts\") pod \"octavia-db-create-ghczh\" (UID: \"0aa6e501-c175-44be-af00-02b8aac76bab\") " pod="openstack/octavia-db-create-ghczh" Mar 13 15:44:21 crc kubenswrapper[4907]: I0313 15:44:21.399504 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l5kw2\" (UniqueName: \"kubernetes.io/projected/0aa6e501-c175-44be-af00-02b8aac76bab-kube-api-access-l5kw2\") pod \"octavia-db-create-ghczh\" (UID: \"0aa6e501-c175-44be-af00-02b8aac76bab\") " pod="openstack/octavia-db-create-ghczh" Mar 13 15:44:21 crc kubenswrapper[4907]: I0313 15:44:21.566216 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-create-ghczh" Mar 13 15:44:21 crc kubenswrapper[4907]: I0313 15:44:21.795198 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0056ce52-b92e-4f2d-b3b0-a7dff01207ff" path="/var/lib/kubelet/pods/0056ce52-b92e-4f2d-b3b0-a7dff01207ff/volumes" Mar 13 15:44:22 crc kubenswrapper[4907]: I0313 15:44:22.048379 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-db-create-ghczh"] Mar 13 15:44:22 crc kubenswrapper[4907]: W0313 15:44:22.050494 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0aa6e501_c175_44be_af00_02b8aac76bab.slice/crio-df9a2353d33e2427fe0c752dfa6b84b5852af4f10c0bc4157236881d9895340f WatchSource:0}: Error finding container df9a2353d33e2427fe0c752dfa6b84b5852af4f10c0bc4157236881d9895340f: Status 404 returned error can't find the container with id df9a2353d33e2427fe0c752dfa6b84b5852af4f10c0bc4157236881d9895340f Mar 13 15:44:22 crc kubenswrapper[4907]: I0313 15:44:22.535098 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-6234-account-create-update-wjdx9"] Mar 13 15:44:22 crc kubenswrapper[4907]: I0313 15:44:22.536674 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-6234-account-create-update-wjdx9" Mar 13 15:44:22 crc kubenswrapper[4907]: I0313 15:44:22.538271 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-db-secret" Mar 13 15:44:22 crc kubenswrapper[4907]: I0313 15:44:22.559272 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-6234-account-create-update-wjdx9"] Mar 13 15:44:22 crc kubenswrapper[4907]: I0313 15:44:22.601594 4907 generic.go:334] "Generic (PLEG): container finished" podID="0aa6e501-c175-44be-af00-02b8aac76bab" containerID="ef160b1245008062ab4f675d742f68b73acd696e5db3bbb49ff009cadb324f56" exitCode=0 Mar 13 15:44:22 crc kubenswrapper[4907]: I0313 15:44:22.601638 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-create-ghczh" event={"ID":"0aa6e501-c175-44be-af00-02b8aac76bab","Type":"ContainerDied","Data":"ef160b1245008062ab4f675d742f68b73acd696e5db3bbb49ff009cadb324f56"} Mar 13 15:44:22 crc kubenswrapper[4907]: I0313 15:44:22.601664 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-create-ghczh" event={"ID":"0aa6e501-c175-44be-af00-02b8aac76bab","Type":"ContainerStarted","Data":"df9a2353d33e2427fe0c752dfa6b84b5852af4f10c0bc4157236881d9895340f"} Mar 13 15:44:22 crc kubenswrapper[4907]: I0313 15:44:22.604681 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c420aeff-7a0b-4efa-b9c0-a93791eb0378-operator-scripts\") pod \"octavia-6234-account-create-update-wjdx9\" (UID: \"c420aeff-7a0b-4efa-b9c0-a93791eb0378\") " pod="openstack/octavia-6234-account-create-update-wjdx9" Mar 13 15:44:22 crc kubenswrapper[4907]: I0313 15:44:22.604894 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jgf6b\" (UniqueName: \"kubernetes.io/projected/c420aeff-7a0b-4efa-b9c0-a93791eb0378-kube-api-access-jgf6b\") pod \"octavia-6234-account-create-update-wjdx9\" (UID: \"c420aeff-7a0b-4efa-b9c0-a93791eb0378\") " pod="openstack/octavia-6234-account-create-update-wjdx9" Mar 13 15:44:22 crc kubenswrapper[4907]: I0313 15:44:22.707201 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c420aeff-7a0b-4efa-b9c0-a93791eb0378-operator-scripts\") pod \"octavia-6234-account-create-update-wjdx9\" (UID: \"c420aeff-7a0b-4efa-b9c0-a93791eb0378\") " pod="openstack/octavia-6234-account-create-update-wjdx9" Mar 13 15:44:22 crc kubenswrapper[4907]: I0313 15:44:22.707282 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jgf6b\" (UniqueName: \"kubernetes.io/projected/c420aeff-7a0b-4efa-b9c0-a93791eb0378-kube-api-access-jgf6b\") pod \"octavia-6234-account-create-update-wjdx9\" (UID: \"c420aeff-7a0b-4efa-b9c0-a93791eb0378\") " pod="openstack/octavia-6234-account-create-update-wjdx9" Mar 13 15:44:22 crc kubenswrapper[4907]: I0313 15:44:22.708108 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c420aeff-7a0b-4efa-b9c0-a93791eb0378-operator-scripts\") pod \"octavia-6234-account-create-update-wjdx9\" (UID: \"c420aeff-7a0b-4efa-b9c0-a93791eb0378\") " pod="openstack/octavia-6234-account-create-update-wjdx9" Mar 13 15:44:22 crc kubenswrapper[4907]: I0313 15:44:22.741089 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jgf6b\" (UniqueName: \"kubernetes.io/projected/c420aeff-7a0b-4efa-b9c0-a93791eb0378-kube-api-access-jgf6b\") pod \"octavia-6234-account-create-update-wjdx9\" (UID: \"c420aeff-7a0b-4efa-b9c0-a93791eb0378\") " pod="openstack/octavia-6234-account-create-update-wjdx9" Mar 13 15:44:22 crc kubenswrapper[4907]: I0313 15:44:22.857255 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-6234-account-create-update-wjdx9" Mar 13 15:44:23 crc kubenswrapper[4907]: I0313 15:44:23.324936 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-6234-account-create-update-wjdx9"] Mar 13 15:44:23 crc kubenswrapper[4907]: I0313 15:44:23.614567 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-6234-account-create-update-wjdx9" event={"ID":"c420aeff-7a0b-4efa-b9c0-a93791eb0378","Type":"ContainerStarted","Data":"1d71acbd6f06389aec2ad567011e53ff359b678fb6ad5c86cc822259fbdd55b9"} Mar 13 15:44:23 crc kubenswrapper[4907]: I0313 15:44:23.614935 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-6234-account-create-update-wjdx9" event={"ID":"c420aeff-7a0b-4efa-b9c0-a93791eb0378","Type":"ContainerStarted","Data":"e27a1243847151a896b96d511cbc60921c81660ade39e55cfd361a3a6ded5d7a"} Mar 13 15:44:23 crc kubenswrapper[4907]: I0313 15:44:23.632208 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-6234-account-create-update-wjdx9" podStartSLOduration=1.632187283 podStartE2EDuration="1.632187283s" podCreationTimestamp="2026-03-13 15:44:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:44:23.629463598 +0000 UTC m=+5962.529251287" watchObservedRunningTime="2026-03-13 15:44:23.632187283 +0000 UTC m=+5962.531974982" Mar 13 15:44:23 crc kubenswrapper[4907]: I0313 15:44:23.952524 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-create-ghczh" Mar 13 15:44:24 crc kubenswrapper[4907]: I0313 15:44:24.047476 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l5kw2\" (UniqueName: \"kubernetes.io/projected/0aa6e501-c175-44be-af00-02b8aac76bab-kube-api-access-l5kw2\") pod \"0aa6e501-c175-44be-af00-02b8aac76bab\" (UID: \"0aa6e501-c175-44be-af00-02b8aac76bab\") " Mar 13 15:44:24 crc kubenswrapper[4907]: I0313 15:44:24.047515 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0aa6e501-c175-44be-af00-02b8aac76bab-operator-scripts\") pod \"0aa6e501-c175-44be-af00-02b8aac76bab\" (UID: \"0aa6e501-c175-44be-af00-02b8aac76bab\") " Mar 13 15:44:24 crc kubenswrapper[4907]: I0313 15:44:24.048233 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0aa6e501-c175-44be-af00-02b8aac76bab-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "0aa6e501-c175-44be-af00-02b8aac76bab" (UID: "0aa6e501-c175-44be-af00-02b8aac76bab"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:44:24 crc kubenswrapper[4907]: I0313 15:44:24.053256 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0aa6e501-c175-44be-af00-02b8aac76bab-kube-api-access-l5kw2" (OuterVolumeSpecName: "kube-api-access-l5kw2") pod "0aa6e501-c175-44be-af00-02b8aac76bab" (UID: "0aa6e501-c175-44be-af00-02b8aac76bab"). InnerVolumeSpecName "kube-api-access-l5kw2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:44:24 crc kubenswrapper[4907]: I0313 15:44:24.149336 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l5kw2\" (UniqueName: \"kubernetes.io/projected/0aa6e501-c175-44be-af00-02b8aac76bab-kube-api-access-l5kw2\") on node \"crc\" DevicePath \"\"" Mar 13 15:44:24 crc kubenswrapper[4907]: I0313 15:44:24.149374 4907 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0aa6e501-c175-44be-af00-02b8aac76bab-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 15:44:24 crc kubenswrapper[4907]: I0313 15:44:24.622767 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-create-ghczh" event={"ID":"0aa6e501-c175-44be-af00-02b8aac76bab","Type":"ContainerDied","Data":"df9a2353d33e2427fe0c752dfa6b84b5852af4f10c0bc4157236881d9895340f"} Mar 13 15:44:24 crc kubenswrapper[4907]: I0313 15:44:24.623147 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="df9a2353d33e2427fe0c752dfa6b84b5852af4f10c0bc4157236881d9895340f" Mar 13 15:44:24 crc kubenswrapper[4907]: I0313 15:44:24.622816 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-create-ghczh" Mar 13 15:44:24 crc kubenswrapper[4907]: I0313 15:44:24.625783 4907 generic.go:334] "Generic (PLEG): container finished" podID="c420aeff-7a0b-4efa-b9c0-a93791eb0378" containerID="1d71acbd6f06389aec2ad567011e53ff359b678fb6ad5c86cc822259fbdd55b9" exitCode=0 Mar 13 15:44:24 crc kubenswrapper[4907]: I0313 15:44:24.625844 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-6234-account-create-update-wjdx9" event={"ID":"c420aeff-7a0b-4efa-b9c0-a93791eb0378","Type":"ContainerDied","Data":"1d71acbd6f06389aec2ad567011e53ff359b678fb6ad5c86cc822259fbdd55b9"} Mar 13 15:44:26 crc kubenswrapper[4907]: I0313 15:44:26.023847 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-6234-account-create-update-wjdx9" Mar 13 15:44:26 crc kubenswrapper[4907]: I0313 15:44:26.085804 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c420aeff-7a0b-4efa-b9c0-a93791eb0378-operator-scripts\") pod \"c420aeff-7a0b-4efa-b9c0-a93791eb0378\" (UID: \"c420aeff-7a0b-4efa-b9c0-a93791eb0378\") " Mar 13 15:44:26 crc kubenswrapper[4907]: I0313 15:44:26.085848 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jgf6b\" (UniqueName: \"kubernetes.io/projected/c420aeff-7a0b-4efa-b9c0-a93791eb0378-kube-api-access-jgf6b\") pod \"c420aeff-7a0b-4efa-b9c0-a93791eb0378\" (UID: \"c420aeff-7a0b-4efa-b9c0-a93791eb0378\") " Mar 13 15:44:26 crc kubenswrapper[4907]: I0313 15:44:26.086219 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c420aeff-7a0b-4efa-b9c0-a93791eb0378-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c420aeff-7a0b-4efa-b9c0-a93791eb0378" (UID: "c420aeff-7a0b-4efa-b9c0-a93791eb0378"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:44:26 crc kubenswrapper[4907]: I0313 15:44:26.086396 4907 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c420aeff-7a0b-4efa-b9c0-a93791eb0378-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 15:44:26 crc kubenswrapper[4907]: I0313 15:44:26.090630 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c420aeff-7a0b-4efa-b9c0-a93791eb0378-kube-api-access-jgf6b" (OuterVolumeSpecName: "kube-api-access-jgf6b") pod "c420aeff-7a0b-4efa-b9c0-a93791eb0378" (UID: "c420aeff-7a0b-4efa-b9c0-a93791eb0378"). InnerVolumeSpecName "kube-api-access-jgf6b". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:44:26 crc kubenswrapper[4907]: I0313 15:44:26.188302 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jgf6b\" (UniqueName: \"kubernetes.io/projected/c420aeff-7a0b-4efa-b9c0-a93791eb0378-kube-api-access-jgf6b\") on node \"crc\" DevicePath \"\"" Mar 13 15:44:26 crc kubenswrapper[4907]: I0313 15:44:26.644853 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-6234-account-create-update-wjdx9" event={"ID":"c420aeff-7a0b-4efa-b9c0-a93791eb0378","Type":"ContainerDied","Data":"e27a1243847151a896b96d511cbc60921c81660ade39e55cfd361a3a6ded5d7a"} Mar 13 15:44:26 crc kubenswrapper[4907]: I0313 15:44:26.645170 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e27a1243847151a896b96d511cbc60921c81660ade39e55cfd361a3a6ded5d7a" Mar 13 15:44:26 crc kubenswrapper[4907]: I0313 15:44:26.644914 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-6234-account-create-update-wjdx9" Mar 13 15:44:28 crc kubenswrapper[4907]: I0313 15:44:28.225921 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-persistence-db-create-p9hf2"] Mar 13 15:44:28 crc kubenswrapper[4907]: E0313 15:44:28.226311 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0aa6e501-c175-44be-af00-02b8aac76bab" containerName="mariadb-database-create" Mar 13 15:44:28 crc kubenswrapper[4907]: I0313 15:44:28.226322 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="0aa6e501-c175-44be-af00-02b8aac76bab" containerName="mariadb-database-create" Mar 13 15:44:28 crc kubenswrapper[4907]: E0313 15:44:28.226331 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c420aeff-7a0b-4efa-b9c0-a93791eb0378" containerName="mariadb-account-create-update" Mar 13 15:44:28 crc kubenswrapper[4907]: I0313 15:44:28.226337 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="c420aeff-7a0b-4efa-b9c0-a93791eb0378" containerName="mariadb-account-create-update" Mar 13 15:44:28 crc kubenswrapper[4907]: I0313 15:44:28.226557 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="0aa6e501-c175-44be-af00-02b8aac76bab" containerName="mariadb-database-create" Mar 13 15:44:28 crc kubenswrapper[4907]: I0313 15:44:28.226565 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="c420aeff-7a0b-4efa-b9c0-a93791eb0378" containerName="mariadb-account-create-update" Mar 13 15:44:28 crc kubenswrapper[4907]: I0313 15:44:28.227217 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-persistence-db-create-p9hf2" Mar 13 15:44:28 crc kubenswrapper[4907]: I0313 15:44:28.241792 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-persistence-db-create-p9hf2"] Mar 13 15:44:28 crc kubenswrapper[4907]: I0313 15:44:28.328859 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jsfdx\" (UniqueName: \"kubernetes.io/projected/d218d11d-2afc-4f21-add2-af15cd8c50ad-kube-api-access-jsfdx\") pod \"octavia-persistence-db-create-p9hf2\" (UID: \"d218d11d-2afc-4f21-add2-af15cd8c50ad\") " pod="openstack/octavia-persistence-db-create-p9hf2" Mar 13 15:44:28 crc kubenswrapper[4907]: I0313 15:44:28.328958 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d218d11d-2afc-4f21-add2-af15cd8c50ad-operator-scripts\") pod \"octavia-persistence-db-create-p9hf2\" (UID: \"d218d11d-2afc-4f21-add2-af15cd8c50ad\") " pod="openstack/octavia-persistence-db-create-p9hf2" Mar 13 15:44:28 crc kubenswrapper[4907]: I0313 15:44:28.430958 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jsfdx\" (UniqueName: \"kubernetes.io/projected/d218d11d-2afc-4f21-add2-af15cd8c50ad-kube-api-access-jsfdx\") pod \"octavia-persistence-db-create-p9hf2\" (UID: \"d218d11d-2afc-4f21-add2-af15cd8c50ad\") " pod="openstack/octavia-persistence-db-create-p9hf2" Mar 13 15:44:28 crc kubenswrapper[4907]: I0313 15:44:28.431039 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d218d11d-2afc-4f21-add2-af15cd8c50ad-operator-scripts\") pod \"octavia-persistence-db-create-p9hf2\" (UID: \"d218d11d-2afc-4f21-add2-af15cd8c50ad\") " pod="openstack/octavia-persistence-db-create-p9hf2" Mar 13 15:44:28 crc kubenswrapper[4907]: I0313 15:44:28.431829 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d218d11d-2afc-4f21-add2-af15cd8c50ad-operator-scripts\") pod \"octavia-persistence-db-create-p9hf2\" (UID: \"d218d11d-2afc-4f21-add2-af15cd8c50ad\") " pod="openstack/octavia-persistence-db-create-p9hf2" Mar 13 15:44:28 crc kubenswrapper[4907]: I0313 15:44:28.448786 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jsfdx\" (UniqueName: \"kubernetes.io/projected/d218d11d-2afc-4f21-add2-af15cd8c50ad-kube-api-access-jsfdx\") pod \"octavia-persistence-db-create-p9hf2\" (UID: \"d218d11d-2afc-4f21-add2-af15cd8c50ad\") " pod="openstack/octavia-persistence-db-create-p9hf2" Mar 13 15:44:28 crc kubenswrapper[4907]: I0313 15:44:28.546631 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-persistence-db-create-p9hf2" Mar 13 15:44:28 crc kubenswrapper[4907]: I0313 15:44:28.738018 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-1da8-account-create-update-82kpd"] Mar 13 15:44:28 crc kubenswrapper[4907]: I0313 15:44:28.739610 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-1da8-account-create-update-82kpd" Mar 13 15:44:28 crc kubenswrapper[4907]: I0313 15:44:28.745801 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-1da8-account-create-update-82kpd"] Mar 13 15:44:28 crc kubenswrapper[4907]: I0313 15:44:28.747424 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-persistence-db-secret" Mar 13 15:44:28 crc kubenswrapper[4907]: I0313 15:44:28.838481 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3b46488e-1fef-479f-94b6-1958e5174219-operator-scripts\") pod \"octavia-1da8-account-create-update-82kpd\" (UID: \"3b46488e-1fef-479f-94b6-1958e5174219\") " pod="openstack/octavia-1da8-account-create-update-82kpd" Mar 13 15:44:28 crc kubenswrapper[4907]: I0313 15:44:28.838789 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jz5hg\" (UniqueName: \"kubernetes.io/projected/3b46488e-1fef-479f-94b6-1958e5174219-kube-api-access-jz5hg\") pod \"octavia-1da8-account-create-update-82kpd\" (UID: \"3b46488e-1fef-479f-94b6-1958e5174219\") " pod="openstack/octavia-1da8-account-create-update-82kpd" Mar 13 15:44:28 crc kubenswrapper[4907]: I0313 15:44:28.940298 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jz5hg\" (UniqueName: \"kubernetes.io/projected/3b46488e-1fef-479f-94b6-1958e5174219-kube-api-access-jz5hg\") pod \"octavia-1da8-account-create-update-82kpd\" (UID: \"3b46488e-1fef-479f-94b6-1958e5174219\") " pod="openstack/octavia-1da8-account-create-update-82kpd" Mar 13 15:44:28 crc kubenswrapper[4907]: I0313 15:44:28.940436 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3b46488e-1fef-479f-94b6-1958e5174219-operator-scripts\") pod \"octavia-1da8-account-create-update-82kpd\" (UID: \"3b46488e-1fef-479f-94b6-1958e5174219\") " pod="openstack/octavia-1da8-account-create-update-82kpd" Mar 13 15:44:28 crc kubenswrapper[4907]: I0313 15:44:28.941230 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3b46488e-1fef-479f-94b6-1958e5174219-operator-scripts\") pod \"octavia-1da8-account-create-update-82kpd\" (UID: \"3b46488e-1fef-479f-94b6-1958e5174219\") " pod="openstack/octavia-1da8-account-create-update-82kpd" Mar 13 15:44:28 crc kubenswrapper[4907]: I0313 15:44:28.957588 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jz5hg\" (UniqueName: \"kubernetes.io/projected/3b46488e-1fef-479f-94b6-1958e5174219-kube-api-access-jz5hg\") pod \"octavia-1da8-account-create-update-82kpd\" (UID: \"3b46488e-1fef-479f-94b6-1958e5174219\") " pod="openstack/octavia-1da8-account-create-update-82kpd" Mar 13 15:44:29 crc kubenswrapper[4907]: I0313 15:44:29.008796 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-persistence-db-create-p9hf2"] Mar 13 15:44:29 crc kubenswrapper[4907]: I0313 15:44:29.061056 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-1da8-account-create-update-82kpd" Mar 13 15:44:29 crc kubenswrapper[4907]: I0313 15:44:29.305140 4907 scope.go:117] "RemoveContainer" containerID="e0b7f12342f112aa71fb3ae1454f11d3b6411a1575b99feb9bc8f40c4e53c905" Mar 13 15:44:29 crc kubenswrapper[4907]: I0313 15:44:29.354982 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-1da8-account-create-update-82kpd"] Mar 13 15:44:29 crc kubenswrapper[4907]: I0313 15:44:29.363800 4907 scope.go:117] "RemoveContainer" containerID="6f94031107f6bbf3f09d35d19fba9b0dfe08f6a055661f1e04241d246d42472d" Mar 13 15:44:29 crc kubenswrapper[4907]: W0313 15:44:29.369393 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3b46488e_1fef_479f_94b6_1958e5174219.slice/crio-eeb8ce48309afbf39df362ceed325c415fd044fc1e43aef2b30ca4f7acab66dd WatchSource:0}: Error finding container eeb8ce48309afbf39df362ceed325c415fd044fc1e43aef2b30ca4f7acab66dd: Status 404 returned error can't find the container with id eeb8ce48309afbf39df362ceed325c415fd044fc1e43aef2b30ca4f7acab66dd Mar 13 15:44:29 crc kubenswrapper[4907]: I0313 15:44:29.388157 4907 scope.go:117] "RemoveContainer" containerID="d95988efdbf9f67c502e797aeeb6493599a1bcc1156251e60584683248ab36b0" Mar 13 15:44:29 crc kubenswrapper[4907]: I0313 15:44:29.441523 4907 scope.go:117] "RemoveContainer" containerID="19b5bd889c57ba8c1711b31659d515b279612f88bb0e0f645828fd40237e4402" Mar 13 15:44:29 crc kubenswrapper[4907]: I0313 15:44:29.542109 4907 scope.go:117] "RemoveContainer" containerID="a3a50fd2f676041cc1b8a25989baf35e7feed38466f29e552c042a8c33760d83" Mar 13 15:44:29 crc kubenswrapper[4907]: I0313 15:44:29.677478 4907 generic.go:334] "Generic (PLEG): container finished" podID="d218d11d-2afc-4f21-add2-af15cd8c50ad" containerID="f56058903e847c31c47f45cadeb76ef300892a82a5c55a4fd28b47f0d9e1eeec" exitCode=0 Mar 13 15:44:29 crc kubenswrapper[4907]: I0313 15:44:29.677547 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-persistence-db-create-p9hf2" event={"ID":"d218d11d-2afc-4f21-add2-af15cd8c50ad","Type":"ContainerDied","Data":"f56058903e847c31c47f45cadeb76ef300892a82a5c55a4fd28b47f0d9e1eeec"} Mar 13 15:44:29 crc kubenswrapper[4907]: I0313 15:44:29.677572 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-persistence-db-create-p9hf2" event={"ID":"d218d11d-2afc-4f21-add2-af15cd8c50ad","Type":"ContainerStarted","Data":"22a8b1c8ecbde85926fedcdd711c6c938300b78064fcaf19ed68711cc58f703a"} Mar 13 15:44:29 crc kubenswrapper[4907]: I0313 15:44:29.680135 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-1da8-account-create-update-82kpd" event={"ID":"3b46488e-1fef-479f-94b6-1958e5174219","Type":"ContainerStarted","Data":"e63bf49f2a3afc218c4b126f220439d3e3bcaa22e7a77cab733b2ce5385658bb"} Mar 13 15:44:29 crc kubenswrapper[4907]: I0313 15:44:29.680161 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-1da8-account-create-update-82kpd" event={"ID":"3b46488e-1fef-479f-94b6-1958e5174219","Type":"ContainerStarted","Data":"eeb8ce48309afbf39df362ceed325c415fd044fc1e43aef2b30ca4f7acab66dd"} Mar 13 15:44:29 crc kubenswrapper[4907]: I0313 15:44:29.727166 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-1da8-account-create-update-82kpd" podStartSLOduration=1.7271425969999998 podStartE2EDuration="1.727142597s" podCreationTimestamp="2026-03-13 15:44:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:44:29.716540887 +0000 UTC m=+5968.616328576" watchObservedRunningTime="2026-03-13 15:44:29.727142597 +0000 UTC m=+5968.626930286" Mar 13 15:44:30 crc kubenswrapper[4907]: I0313 15:44:30.690753 4907 generic.go:334] "Generic (PLEG): container finished" podID="3b46488e-1fef-479f-94b6-1958e5174219" containerID="e63bf49f2a3afc218c4b126f220439d3e3bcaa22e7a77cab733b2ce5385658bb" exitCode=0 Mar 13 15:44:30 crc kubenswrapper[4907]: I0313 15:44:30.690828 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-1da8-account-create-update-82kpd" event={"ID":"3b46488e-1fef-479f-94b6-1958e5174219","Type":"ContainerDied","Data":"e63bf49f2a3afc218c4b126f220439d3e3bcaa22e7a77cab733b2ce5385658bb"} Mar 13 15:44:31 crc kubenswrapper[4907]: I0313 15:44:31.019511 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-persistence-db-create-p9hf2" Mar 13 15:44:31 crc kubenswrapper[4907]: I0313 15:44:31.182679 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jsfdx\" (UniqueName: \"kubernetes.io/projected/d218d11d-2afc-4f21-add2-af15cd8c50ad-kube-api-access-jsfdx\") pod \"d218d11d-2afc-4f21-add2-af15cd8c50ad\" (UID: \"d218d11d-2afc-4f21-add2-af15cd8c50ad\") " Mar 13 15:44:31 crc kubenswrapper[4907]: I0313 15:44:31.182855 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d218d11d-2afc-4f21-add2-af15cd8c50ad-operator-scripts\") pod \"d218d11d-2afc-4f21-add2-af15cd8c50ad\" (UID: \"d218d11d-2afc-4f21-add2-af15cd8c50ad\") " Mar 13 15:44:31 crc kubenswrapper[4907]: I0313 15:44:31.183296 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d218d11d-2afc-4f21-add2-af15cd8c50ad-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d218d11d-2afc-4f21-add2-af15cd8c50ad" (UID: "d218d11d-2afc-4f21-add2-af15cd8c50ad"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:44:31 crc kubenswrapper[4907]: I0313 15:44:31.183554 4907 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d218d11d-2afc-4f21-add2-af15cd8c50ad-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 15:44:31 crc kubenswrapper[4907]: I0313 15:44:31.189360 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d218d11d-2afc-4f21-add2-af15cd8c50ad-kube-api-access-jsfdx" (OuterVolumeSpecName: "kube-api-access-jsfdx") pod "d218d11d-2afc-4f21-add2-af15cd8c50ad" (UID: "d218d11d-2afc-4f21-add2-af15cd8c50ad"). InnerVolumeSpecName "kube-api-access-jsfdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:44:31 crc kubenswrapper[4907]: I0313 15:44:31.285340 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jsfdx\" (UniqueName: \"kubernetes.io/projected/d218d11d-2afc-4f21-add2-af15cd8c50ad-kube-api-access-jsfdx\") on node \"crc\" DevicePath \"\"" Mar 13 15:44:31 crc kubenswrapper[4907]: I0313 15:44:31.701942 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-persistence-db-create-p9hf2" Mar 13 15:44:31 crc kubenswrapper[4907]: I0313 15:44:31.701937 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-persistence-db-create-p9hf2" event={"ID":"d218d11d-2afc-4f21-add2-af15cd8c50ad","Type":"ContainerDied","Data":"22a8b1c8ecbde85926fedcdd711c6c938300b78064fcaf19ed68711cc58f703a"} Mar 13 15:44:31 crc kubenswrapper[4907]: I0313 15:44:31.702425 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="22a8b1c8ecbde85926fedcdd711c6c938300b78064fcaf19ed68711cc58f703a" Mar 13 15:44:31 crc kubenswrapper[4907]: I0313 15:44:31.794386 4907 scope.go:117] "RemoveContainer" containerID="68ffe92dd03078633ec89d7ef29da9bbdc4c5f04c956a1e8487e3f2e01c5d1c5" Mar 13 15:44:31 crc kubenswrapper[4907]: E0313 15:44:31.794775 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:44:32 crc kubenswrapper[4907]: I0313 15:44:32.088008 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-1da8-account-create-update-82kpd" Mar 13 15:44:32 crc kubenswrapper[4907]: I0313 15:44:32.115022 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3b46488e-1fef-479f-94b6-1958e5174219-operator-scripts\") pod \"3b46488e-1fef-479f-94b6-1958e5174219\" (UID: \"3b46488e-1fef-479f-94b6-1958e5174219\") " Mar 13 15:44:32 crc kubenswrapper[4907]: I0313 15:44:32.115243 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jz5hg\" (UniqueName: \"kubernetes.io/projected/3b46488e-1fef-479f-94b6-1958e5174219-kube-api-access-jz5hg\") pod \"3b46488e-1fef-479f-94b6-1958e5174219\" (UID: \"3b46488e-1fef-479f-94b6-1958e5174219\") " Mar 13 15:44:32 crc kubenswrapper[4907]: I0313 15:44:32.115777 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3b46488e-1fef-479f-94b6-1958e5174219-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "3b46488e-1fef-479f-94b6-1958e5174219" (UID: "3b46488e-1fef-479f-94b6-1958e5174219"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:44:32 crc kubenswrapper[4907]: I0313 15:44:32.116189 4907 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3b46488e-1fef-479f-94b6-1958e5174219-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 15:44:32 crc kubenswrapper[4907]: I0313 15:44:32.120598 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3b46488e-1fef-479f-94b6-1958e5174219-kube-api-access-jz5hg" (OuterVolumeSpecName: "kube-api-access-jz5hg") pod "3b46488e-1fef-479f-94b6-1958e5174219" (UID: "3b46488e-1fef-479f-94b6-1958e5174219"). InnerVolumeSpecName "kube-api-access-jz5hg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:44:32 crc kubenswrapper[4907]: I0313 15:44:32.124385 4907 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-zb9k2" podUID="1cf43c88-f066-4979-8590-b8e6668aab09" containerName="ovn-controller" probeResult="failure" output=< Mar 13 15:44:32 crc kubenswrapper[4907]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Mar 13 15:44:32 crc kubenswrapper[4907]: > Mar 13 15:44:32 crc kubenswrapper[4907]: I0313 15:44:32.137546 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-v5b57" Mar 13 15:44:32 crc kubenswrapper[4907]: I0313 15:44:32.161868 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-v5b57" Mar 13 15:44:32 crc kubenswrapper[4907]: I0313 15:44:32.219213 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jz5hg\" (UniqueName: \"kubernetes.io/projected/3b46488e-1fef-479f-94b6-1958e5174219-kube-api-access-jz5hg\") on node \"crc\" DevicePath \"\"" Mar 13 15:44:32 crc kubenswrapper[4907]: I0313 15:44:32.254525 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-zb9k2-config-4n289"] Mar 13 15:44:32 crc kubenswrapper[4907]: E0313 15:44:32.255076 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b46488e-1fef-479f-94b6-1958e5174219" containerName="mariadb-account-create-update" Mar 13 15:44:32 crc kubenswrapper[4907]: I0313 15:44:32.255095 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b46488e-1fef-479f-94b6-1958e5174219" containerName="mariadb-account-create-update" Mar 13 15:44:32 crc kubenswrapper[4907]: E0313 15:44:32.255124 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d218d11d-2afc-4f21-add2-af15cd8c50ad" containerName="mariadb-database-create" Mar 13 15:44:32 crc kubenswrapper[4907]: I0313 15:44:32.255132 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="d218d11d-2afc-4f21-add2-af15cd8c50ad" containerName="mariadb-database-create" Mar 13 15:44:32 crc kubenswrapper[4907]: I0313 15:44:32.255417 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="d218d11d-2afc-4f21-add2-af15cd8c50ad" containerName="mariadb-database-create" Mar 13 15:44:32 crc kubenswrapper[4907]: I0313 15:44:32.255437 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b46488e-1fef-479f-94b6-1958e5174219" containerName="mariadb-account-create-update" Mar 13 15:44:32 crc kubenswrapper[4907]: I0313 15:44:32.256407 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-zb9k2-config-4n289" Mar 13 15:44:32 crc kubenswrapper[4907]: I0313 15:44:32.258538 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Mar 13 15:44:32 crc kubenswrapper[4907]: I0313 15:44:32.271530 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-zb9k2-config-4n289"] Mar 13 15:44:32 crc kubenswrapper[4907]: I0313 15:44:32.320343 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/ecd299e1-64f9-4ad1-9ff8-af388e993db9-var-log-ovn\") pod \"ovn-controller-zb9k2-config-4n289\" (UID: \"ecd299e1-64f9-4ad1-9ff8-af388e993db9\") " pod="openstack/ovn-controller-zb9k2-config-4n289" Mar 13 15:44:32 crc kubenswrapper[4907]: I0313 15:44:32.320403 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/ecd299e1-64f9-4ad1-9ff8-af388e993db9-var-run\") pod \"ovn-controller-zb9k2-config-4n289\" (UID: \"ecd299e1-64f9-4ad1-9ff8-af388e993db9\") " pod="openstack/ovn-controller-zb9k2-config-4n289" Mar 13 15:44:32 crc kubenswrapper[4907]: I0313 15:44:32.320430 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ecd299e1-64f9-4ad1-9ff8-af388e993db9-scripts\") pod \"ovn-controller-zb9k2-config-4n289\" (UID: \"ecd299e1-64f9-4ad1-9ff8-af388e993db9\") " pod="openstack/ovn-controller-zb9k2-config-4n289" Mar 13 15:44:32 crc kubenswrapper[4907]: I0313 15:44:32.320496 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/ecd299e1-64f9-4ad1-9ff8-af388e993db9-var-run-ovn\") pod \"ovn-controller-zb9k2-config-4n289\" (UID: \"ecd299e1-64f9-4ad1-9ff8-af388e993db9\") " pod="openstack/ovn-controller-zb9k2-config-4n289" Mar 13 15:44:32 crc kubenswrapper[4907]: I0313 15:44:32.320530 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sf9zj\" (UniqueName: \"kubernetes.io/projected/ecd299e1-64f9-4ad1-9ff8-af388e993db9-kube-api-access-sf9zj\") pod \"ovn-controller-zb9k2-config-4n289\" (UID: \"ecd299e1-64f9-4ad1-9ff8-af388e993db9\") " pod="openstack/ovn-controller-zb9k2-config-4n289" Mar 13 15:44:32 crc kubenswrapper[4907]: I0313 15:44:32.320617 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/ecd299e1-64f9-4ad1-9ff8-af388e993db9-additional-scripts\") pod \"ovn-controller-zb9k2-config-4n289\" (UID: \"ecd299e1-64f9-4ad1-9ff8-af388e993db9\") " pod="openstack/ovn-controller-zb9k2-config-4n289" Mar 13 15:44:32 crc kubenswrapper[4907]: I0313 15:44:32.422201 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sf9zj\" (UniqueName: \"kubernetes.io/projected/ecd299e1-64f9-4ad1-9ff8-af388e993db9-kube-api-access-sf9zj\") pod \"ovn-controller-zb9k2-config-4n289\" (UID: \"ecd299e1-64f9-4ad1-9ff8-af388e993db9\") " pod="openstack/ovn-controller-zb9k2-config-4n289" Mar 13 15:44:32 crc kubenswrapper[4907]: I0313 15:44:32.422333 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/ecd299e1-64f9-4ad1-9ff8-af388e993db9-additional-scripts\") pod \"ovn-controller-zb9k2-config-4n289\" (UID: \"ecd299e1-64f9-4ad1-9ff8-af388e993db9\") " pod="openstack/ovn-controller-zb9k2-config-4n289" Mar 13 15:44:32 crc kubenswrapper[4907]: I0313 15:44:32.422409 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/ecd299e1-64f9-4ad1-9ff8-af388e993db9-var-log-ovn\") pod \"ovn-controller-zb9k2-config-4n289\" (UID: \"ecd299e1-64f9-4ad1-9ff8-af388e993db9\") " pod="openstack/ovn-controller-zb9k2-config-4n289" Mar 13 15:44:32 crc kubenswrapper[4907]: I0313 15:44:32.422442 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/ecd299e1-64f9-4ad1-9ff8-af388e993db9-var-run\") pod \"ovn-controller-zb9k2-config-4n289\" (UID: \"ecd299e1-64f9-4ad1-9ff8-af388e993db9\") " pod="openstack/ovn-controller-zb9k2-config-4n289" Mar 13 15:44:32 crc kubenswrapper[4907]: I0313 15:44:32.422460 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ecd299e1-64f9-4ad1-9ff8-af388e993db9-scripts\") pod \"ovn-controller-zb9k2-config-4n289\" (UID: \"ecd299e1-64f9-4ad1-9ff8-af388e993db9\") " pod="openstack/ovn-controller-zb9k2-config-4n289" Mar 13 15:44:32 crc kubenswrapper[4907]: I0313 15:44:32.422483 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/ecd299e1-64f9-4ad1-9ff8-af388e993db9-var-run-ovn\") pod \"ovn-controller-zb9k2-config-4n289\" (UID: \"ecd299e1-64f9-4ad1-9ff8-af388e993db9\") " pod="openstack/ovn-controller-zb9k2-config-4n289" Mar 13 15:44:32 crc kubenswrapper[4907]: I0313 15:44:32.422775 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/ecd299e1-64f9-4ad1-9ff8-af388e993db9-var-run-ovn\") pod \"ovn-controller-zb9k2-config-4n289\" (UID: \"ecd299e1-64f9-4ad1-9ff8-af388e993db9\") " pod="openstack/ovn-controller-zb9k2-config-4n289" Mar 13 15:44:32 crc kubenswrapper[4907]: I0313 15:44:32.423798 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/ecd299e1-64f9-4ad1-9ff8-af388e993db9-additional-scripts\") pod \"ovn-controller-zb9k2-config-4n289\" (UID: \"ecd299e1-64f9-4ad1-9ff8-af388e993db9\") " pod="openstack/ovn-controller-zb9k2-config-4n289" Mar 13 15:44:32 crc kubenswrapper[4907]: I0313 15:44:32.423858 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/ecd299e1-64f9-4ad1-9ff8-af388e993db9-var-log-ovn\") pod \"ovn-controller-zb9k2-config-4n289\" (UID: \"ecd299e1-64f9-4ad1-9ff8-af388e993db9\") " pod="openstack/ovn-controller-zb9k2-config-4n289" Mar 13 15:44:32 crc kubenswrapper[4907]: I0313 15:44:32.423911 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/ecd299e1-64f9-4ad1-9ff8-af388e993db9-var-run\") pod \"ovn-controller-zb9k2-config-4n289\" (UID: \"ecd299e1-64f9-4ad1-9ff8-af388e993db9\") " pod="openstack/ovn-controller-zb9k2-config-4n289" Mar 13 15:44:32 crc kubenswrapper[4907]: I0313 15:44:32.425560 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ecd299e1-64f9-4ad1-9ff8-af388e993db9-scripts\") pod \"ovn-controller-zb9k2-config-4n289\" (UID: \"ecd299e1-64f9-4ad1-9ff8-af388e993db9\") " pod="openstack/ovn-controller-zb9k2-config-4n289" Mar 13 15:44:32 crc kubenswrapper[4907]: I0313 15:44:32.455695 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sf9zj\" (UniqueName: \"kubernetes.io/projected/ecd299e1-64f9-4ad1-9ff8-af388e993db9-kube-api-access-sf9zj\") pod \"ovn-controller-zb9k2-config-4n289\" (UID: \"ecd299e1-64f9-4ad1-9ff8-af388e993db9\") " pod="openstack/ovn-controller-zb9k2-config-4n289" Mar 13 15:44:32 crc kubenswrapper[4907]: I0313 15:44:32.583134 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-zb9k2-config-4n289" Mar 13 15:44:32 crc kubenswrapper[4907]: I0313 15:44:32.740826 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-1da8-account-create-update-82kpd" event={"ID":"3b46488e-1fef-479f-94b6-1958e5174219","Type":"ContainerDied","Data":"eeb8ce48309afbf39df362ceed325c415fd044fc1e43aef2b30ca4f7acab66dd"} Mar 13 15:44:32 crc kubenswrapper[4907]: I0313 15:44:32.741190 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eeb8ce48309afbf39df362ceed325c415fd044fc1e43aef2b30ca4f7acab66dd" Mar 13 15:44:32 crc kubenswrapper[4907]: I0313 15:44:32.740864 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-1da8-account-create-update-82kpd" Mar 13 15:44:33 crc kubenswrapper[4907]: I0313 15:44:33.101517 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-zb9k2-config-4n289"] Mar 13 15:44:33 crc kubenswrapper[4907]: W0313 15:44:33.102898 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podecd299e1_64f9_4ad1_9ff8_af388e993db9.slice/crio-c4a92902f8ae073c889d6b0292d8ab146828a907c2ec89705db070f340ffd437 WatchSource:0}: Error finding container c4a92902f8ae073c889d6b0292d8ab146828a907c2ec89705db070f340ffd437: Status 404 returned error can't find the container with id c4a92902f8ae073c889d6b0292d8ab146828a907c2ec89705db070f340ffd437 Mar 13 15:44:33 crc kubenswrapper[4907]: I0313 15:44:33.751697 4907 generic.go:334] "Generic (PLEG): container finished" podID="ecd299e1-64f9-4ad1-9ff8-af388e993db9" containerID="219cbe1415b39a75ce8171bb61ada0538ae87eb0c8b470a088b03fdc9289f0b1" exitCode=0 Mar 13 15:44:33 crc kubenswrapper[4907]: I0313 15:44:33.751763 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-zb9k2-config-4n289" event={"ID":"ecd299e1-64f9-4ad1-9ff8-af388e993db9","Type":"ContainerDied","Data":"219cbe1415b39a75ce8171bb61ada0538ae87eb0c8b470a088b03fdc9289f0b1"} Mar 13 15:44:33 crc kubenswrapper[4907]: I0313 15:44:33.752042 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-zb9k2-config-4n289" event={"ID":"ecd299e1-64f9-4ad1-9ff8-af388e993db9","Type":"ContainerStarted","Data":"c4a92902f8ae073c889d6b0292d8ab146828a907c2ec89705db070f340ffd437"} Mar 13 15:44:34 crc kubenswrapper[4907]: I0313 15:44:34.165068 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-api-6db876fcbd-4wtkj"] Mar 13 15:44:34 crc kubenswrapper[4907]: I0313 15:44:34.168738 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-api-6db876fcbd-4wtkj" Mar 13 15:44:34 crc kubenswrapper[4907]: I0313 15:44:34.171873 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-api-config-data" Mar 13 15:44:34 crc kubenswrapper[4907]: I0313 15:44:34.172080 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-octavia-dockercfg-m6wbw" Mar 13 15:44:34 crc kubenswrapper[4907]: I0313 15:44:34.178310 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-api-scripts" Mar 13 15:44:34 crc kubenswrapper[4907]: I0313 15:44:34.180171 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-api-6db876fcbd-4wtkj"] Mar 13 15:44:34 crc kubenswrapper[4907]: I0313 15:44:34.258847 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/1c868ec4-5e49-4be1-a1eb-c4748753aecb-config-data-merged\") pod \"octavia-api-6db876fcbd-4wtkj\" (UID: \"1c868ec4-5e49-4be1-a1eb-c4748753aecb\") " pod="openstack/octavia-api-6db876fcbd-4wtkj" Mar 13 15:44:34 crc kubenswrapper[4907]: I0313 15:44:34.259009 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1c868ec4-5e49-4be1-a1eb-c4748753aecb-scripts\") pod \"octavia-api-6db876fcbd-4wtkj\" (UID: \"1c868ec4-5e49-4be1-a1eb-c4748753aecb\") " pod="openstack/octavia-api-6db876fcbd-4wtkj" Mar 13 15:44:34 crc kubenswrapper[4907]: I0313 15:44:34.259045 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1c868ec4-5e49-4be1-a1eb-c4748753aecb-config-data\") pod \"octavia-api-6db876fcbd-4wtkj\" (UID: \"1c868ec4-5e49-4be1-a1eb-c4748753aecb\") " pod="openstack/octavia-api-6db876fcbd-4wtkj" Mar 13 15:44:34 crc kubenswrapper[4907]: I0313 15:44:34.259303 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"octavia-run\" (UniqueName: \"kubernetes.io/empty-dir/1c868ec4-5e49-4be1-a1eb-c4748753aecb-octavia-run\") pod \"octavia-api-6db876fcbd-4wtkj\" (UID: \"1c868ec4-5e49-4be1-a1eb-c4748753aecb\") " pod="openstack/octavia-api-6db876fcbd-4wtkj" Mar 13 15:44:34 crc kubenswrapper[4907]: I0313 15:44:34.259399 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c868ec4-5e49-4be1-a1eb-c4748753aecb-combined-ca-bundle\") pod \"octavia-api-6db876fcbd-4wtkj\" (UID: \"1c868ec4-5e49-4be1-a1eb-c4748753aecb\") " pod="openstack/octavia-api-6db876fcbd-4wtkj" Mar 13 15:44:34 crc kubenswrapper[4907]: I0313 15:44:34.360933 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1c868ec4-5e49-4be1-a1eb-c4748753aecb-scripts\") pod \"octavia-api-6db876fcbd-4wtkj\" (UID: \"1c868ec4-5e49-4be1-a1eb-c4748753aecb\") " pod="openstack/octavia-api-6db876fcbd-4wtkj" Mar 13 15:44:34 crc kubenswrapper[4907]: I0313 15:44:34.360983 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1c868ec4-5e49-4be1-a1eb-c4748753aecb-config-data\") pod \"octavia-api-6db876fcbd-4wtkj\" (UID: \"1c868ec4-5e49-4be1-a1eb-c4748753aecb\") " pod="openstack/octavia-api-6db876fcbd-4wtkj" Mar 13 15:44:34 crc kubenswrapper[4907]: I0313 15:44:34.361088 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"octavia-run\" (UniqueName: \"kubernetes.io/empty-dir/1c868ec4-5e49-4be1-a1eb-c4748753aecb-octavia-run\") pod \"octavia-api-6db876fcbd-4wtkj\" (UID: \"1c868ec4-5e49-4be1-a1eb-c4748753aecb\") " pod="openstack/octavia-api-6db876fcbd-4wtkj" Mar 13 15:44:34 crc kubenswrapper[4907]: I0313 15:44:34.361126 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c868ec4-5e49-4be1-a1eb-c4748753aecb-combined-ca-bundle\") pod \"octavia-api-6db876fcbd-4wtkj\" (UID: \"1c868ec4-5e49-4be1-a1eb-c4748753aecb\") " pod="openstack/octavia-api-6db876fcbd-4wtkj" Mar 13 15:44:34 crc kubenswrapper[4907]: I0313 15:44:34.361240 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/1c868ec4-5e49-4be1-a1eb-c4748753aecb-config-data-merged\") pod \"octavia-api-6db876fcbd-4wtkj\" (UID: \"1c868ec4-5e49-4be1-a1eb-c4748753aecb\") " pod="openstack/octavia-api-6db876fcbd-4wtkj" Mar 13 15:44:34 crc kubenswrapper[4907]: I0313 15:44:34.361779 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"octavia-run\" (UniqueName: \"kubernetes.io/empty-dir/1c868ec4-5e49-4be1-a1eb-c4748753aecb-octavia-run\") pod \"octavia-api-6db876fcbd-4wtkj\" (UID: \"1c868ec4-5e49-4be1-a1eb-c4748753aecb\") " pod="openstack/octavia-api-6db876fcbd-4wtkj" Mar 13 15:44:34 crc kubenswrapper[4907]: I0313 15:44:34.361780 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/1c868ec4-5e49-4be1-a1eb-c4748753aecb-config-data-merged\") pod \"octavia-api-6db876fcbd-4wtkj\" (UID: \"1c868ec4-5e49-4be1-a1eb-c4748753aecb\") " pod="openstack/octavia-api-6db876fcbd-4wtkj" Mar 13 15:44:34 crc kubenswrapper[4907]: I0313 15:44:34.367325 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c868ec4-5e49-4be1-a1eb-c4748753aecb-combined-ca-bundle\") pod \"octavia-api-6db876fcbd-4wtkj\" (UID: \"1c868ec4-5e49-4be1-a1eb-c4748753aecb\") " pod="openstack/octavia-api-6db876fcbd-4wtkj" Mar 13 15:44:34 crc kubenswrapper[4907]: I0313 15:44:34.367614 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1c868ec4-5e49-4be1-a1eb-c4748753aecb-config-data\") pod \"octavia-api-6db876fcbd-4wtkj\" (UID: \"1c868ec4-5e49-4be1-a1eb-c4748753aecb\") " pod="openstack/octavia-api-6db876fcbd-4wtkj" Mar 13 15:44:34 crc kubenswrapper[4907]: I0313 15:44:34.371601 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1c868ec4-5e49-4be1-a1eb-c4748753aecb-scripts\") pod \"octavia-api-6db876fcbd-4wtkj\" (UID: \"1c868ec4-5e49-4be1-a1eb-c4748753aecb\") " pod="openstack/octavia-api-6db876fcbd-4wtkj" Mar 13 15:44:34 crc kubenswrapper[4907]: I0313 15:44:34.486983 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-api-6db876fcbd-4wtkj" Mar 13 15:44:35 crc kubenswrapper[4907]: I0313 15:44:35.137093 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-api-6db876fcbd-4wtkj"] Mar 13 15:44:35 crc kubenswrapper[4907]: I0313 15:44:35.174233 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-zb9k2-config-4n289" Mar 13 15:44:35 crc kubenswrapper[4907]: I0313 15:44:35.285103 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/ecd299e1-64f9-4ad1-9ff8-af388e993db9-var-run\") pod \"ecd299e1-64f9-4ad1-9ff8-af388e993db9\" (UID: \"ecd299e1-64f9-4ad1-9ff8-af388e993db9\") " Mar 13 15:44:35 crc kubenswrapper[4907]: I0313 15:44:35.285492 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ecd299e1-64f9-4ad1-9ff8-af388e993db9-scripts\") pod \"ecd299e1-64f9-4ad1-9ff8-af388e993db9\" (UID: \"ecd299e1-64f9-4ad1-9ff8-af388e993db9\") " Mar 13 15:44:35 crc kubenswrapper[4907]: I0313 15:44:35.285613 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/ecd299e1-64f9-4ad1-9ff8-af388e993db9-var-run-ovn\") pod \"ecd299e1-64f9-4ad1-9ff8-af388e993db9\" (UID: \"ecd299e1-64f9-4ad1-9ff8-af388e993db9\") " Mar 13 15:44:35 crc kubenswrapper[4907]: I0313 15:44:35.285737 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sf9zj\" (UniqueName: \"kubernetes.io/projected/ecd299e1-64f9-4ad1-9ff8-af388e993db9-kube-api-access-sf9zj\") pod \"ecd299e1-64f9-4ad1-9ff8-af388e993db9\" (UID: \"ecd299e1-64f9-4ad1-9ff8-af388e993db9\") " Mar 13 15:44:35 crc kubenswrapper[4907]: I0313 15:44:35.286023 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/ecd299e1-64f9-4ad1-9ff8-af388e993db9-additional-scripts\") pod \"ecd299e1-64f9-4ad1-9ff8-af388e993db9\" (UID: \"ecd299e1-64f9-4ad1-9ff8-af388e993db9\") " Mar 13 15:44:35 crc kubenswrapper[4907]: I0313 15:44:35.286275 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/ecd299e1-64f9-4ad1-9ff8-af388e993db9-var-log-ovn\") pod \"ecd299e1-64f9-4ad1-9ff8-af388e993db9\" (UID: \"ecd299e1-64f9-4ad1-9ff8-af388e993db9\") " Mar 13 15:44:35 crc kubenswrapper[4907]: I0313 15:44:35.286995 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ecd299e1-64f9-4ad1-9ff8-af388e993db9-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "ecd299e1-64f9-4ad1-9ff8-af388e993db9" (UID: "ecd299e1-64f9-4ad1-9ff8-af388e993db9"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 15:44:35 crc kubenswrapper[4907]: I0313 15:44:35.287075 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ecd299e1-64f9-4ad1-9ff8-af388e993db9-var-run" (OuterVolumeSpecName: "var-run") pod "ecd299e1-64f9-4ad1-9ff8-af388e993db9" (UID: "ecd299e1-64f9-4ad1-9ff8-af388e993db9"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 15:44:35 crc kubenswrapper[4907]: I0313 15:44:35.287046 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ecd299e1-64f9-4ad1-9ff8-af388e993db9-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "ecd299e1-64f9-4ad1-9ff8-af388e993db9" (UID: "ecd299e1-64f9-4ad1-9ff8-af388e993db9"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 15:44:35 crc kubenswrapper[4907]: I0313 15:44:35.287076 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ecd299e1-64f9-4ad1-9ff8-af388e993db9-scripts" (OuterVolumeSpecName: "scripts") pod "ecd299e1-64f9-4ad1-9ff8-af388e993db9" (UID: "ecd299e1-64f9-4ad1-9ff8-af388e993db9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:44:35 crc kubenswrapper[4907]: I0313 15:44:35.287625 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ecd299e1-64f9-4ad1-9ff8-af388e993db9-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "ecd299e1-64f9-4ad1-9ff8-af388e993db9" (UID: "ecd299e1-64f9-4ad1-9ff8-af388e993db9"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:44:35 crc kubenswrapper[4907]: I0313 15:44:35.293966 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ecd299e1-64f9-4ad1-9ff8-af388e993db9-kube-api-access-sf9zj" (OuterVolumeSpecName: "kube-api-access-sf9zj") pod "ecd299e1-64f9-4ad1-9ff8-af388e993db9" (UID: "ecd299e1-64f9-4ad1-9ff8-af388e993db9"). InnerVolumeSpecName "kube-api-access-sf9zj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:44:35 crc kubenswrapper[4907]: I0313 15:44:35.388404 4907 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/ecd299e1-64f9-4ad1-9ff8-af388e993db9-var-log-ovn\") on node \"crc\" DevicePath \"\"" Mar 13 15:44:35 crc kubenswrapper[4907]: I0313 15:44:35.388663 4907 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/ecd299e1-64f9-4ad1-9ff8-af388e993db9-var-run\") on node \"crc\" DevicePath \"\"" Mar 13 15:44:35 crc kubenswrapper[4907]: I0313 15:44:35.388723 4907 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ecd299e1-64f9-4ad1-9ff8-af388e993db9-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 15:44:35 crc kubenswrapper[4907]: I0313 15:44:35.388774 4907 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/ecd299e1-64f9-4ad1-9ff8-af388e993db9-var-run-ovn\") on node \"crc\" DevicePath \"\"" Mar 13 15:44:35 crc kubenswrapper[4907]: I0313 15:44:35.388837 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sf9zj\" (UniqueName: \"kubernetes.io/projected/ecd299e1-64f9-4ad1-9ff8-af388e993db9-kube-api-access-sf9zj\") on node \"crc\" DevicePath \"\"" Mar 13 15:44:35 crc kubenswrapper[4907]: I0313 15:44:35.388937 4907 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/ecd299e1-64f9-4ad1-9ff8-af388e993db9-additional-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 15:44:35 crc kubenswrapper[4907]: I0313 15:44:35.772514 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-api-6db876fcbd-4wtkj" event={"ID":"1c868ec4-5e49-4be1-a1eb-c4748753aecb","Type":"ContainerStarted","Data":"86b85bba126122ac436f8e2f67651035d64e2de892225b44d780ad2018449177"} Mar 13 15:44:35 crc kubenswrapper[4907]: I0313 15:44:35.773969 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-zb9k2-config-4n289" event={"ID":"ecd299e1-64f9-4ad1-9ff8-af388e993db9","Type":"ContainerDied","Data":"c4a92902f8ae073c889d6b0292d8ab146828a907c2ec89705db070f340ffd437"} Mar 13 15:44:35 crc kubenswrapper[4907]: I0313 15:44:35.773996 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c4a92902f8ae073c889d6b0292d8ab146828a907c2ec89705db070f340ffd437" Mar 13 15:44:35 crc kubenswrapper[4907]: I0313 15:44:35.774037 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-zb9k2-config-4n289" Mar 13 15:44:36 crc kubenswrapper[4907]: I0313 15:44:36.248744 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-zb9k2-config-4n289"] Mar 13 15:44:36 crc kubenswrapper[4907]: I0313 15:44:36.257663 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-zb9k2-config-4n289"] Mar 13 15:44:36 crc kubenswrapper[4907]: I0313 15:44:36.294473 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-zb9k2-config-jvnzd"] Mar 13 15:44:36 crc kubenswrapper[4907]: E0313 15:44:36.294916 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ecd299e1-64f9-4ad1-9ff8-af388e993db9" containerName="ovn-config" Mar 13 15:44:36 crc kubenswrapper[4907]: I0313 15:44:36.294932 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="ecd299e1-64f9-4ad1-9ff8-af388e993db9" containerName="ovn-config" Mar 13 15:44:36 crc kubenswrapper[4907]: I0313 15:44:36.295118 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="ecd299e1-64f9-4ad1-9ff8-af388e993db9" containerName="ovn-config" Mar 13 15:44:36 crc kubenswrapper[4907]: I0313 15:44:36.295775 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-zb9k2-config-jvnzd" Mar 13 15:44:36 crc kubenswrapper[4907]: I0313 15:44:36.299210 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Mar 13 15:44:36 crc kubenswrapper[4907]: I0313 15:44:36.309434 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-zb9k2-config-jvnzd"] Mar 13 15:44:36 crc kubenswrapper[4907]: I0313 15:44:36.409974 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/405d2caf-b667-4833-9e03-f6eaaff9f8c1-var-run-ovn\") pod \"ovn-controller-zb9k2-config-jvnzd\" (UID: \"405d2caf-b667-4833-9e03-f6eaaff9f8c1\") " pod="openstack/ovn-controller-zb9k2-config-jvnzd" Mar 13 15:44:36 crc kubenswrapper[4907]: I0313 15:44:36.410108 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/405d2caf-b667-4833-9e03-f6eaaff9f8c1-scripts\") pod \"ovn-controller-zb9k2-config-jvnzd\" (UID: \"405d2caf-b667-4833-9e03-f6eaaff9f8c1\") " pod="openstack/ovn-controller-zb9k2-config-jvnzd" Mar 13 15:44:36 crc kubenswrapper[4907]: I0313 15:44:36.410162 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/405d2caf-b667-4833-9e03-f6eaaff9f8c1-additional-scripts\") pod \"ovn-controller-zb9k2-config-jvnzd\" (UID: \"405d2caf-b667-4833-9e03-f6eaaff9f8c1\") " pod="openstack/ovn-controller-zb9k2-config-jvnzd" Mar 13 15:44:36 crc kubenswrapper[4907]: I0313 15:44:36.410191 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/405d2caf-b667-4833-9e03-f6eaaff9f8c1-var-run\") pod \"ovn-controller-zb9k2-config-jvnzd\" (UID: \"405d2caf-b667-4833-9e03-f6eaaff9f8c1\") " pod="openstack/ovn-controller-zb9k2-config-jvnzd" Mar 13 15:44:36 crc kubenswrapper[4907]: I0313 15:44:36.410410 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x4w5r\" (UniqueName: \"kubernetes.io/projected/405d2caf-b667-4833-9e03-f6eaaff9f8c1-kube-api-access-x4w5r\") pod \"ovn-controller-zb9k2-config-jvnzd\" (UID: \"405d2caf-b667-4833-9e03-f6eaaff9f8c1\") " pod="openstack/ovn-controller-zb9k2-config-jvnzd" Mar 13 15:44:36 crc kubenswrapper[4907]: I0313 15:44:36.410642 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/405d2caf-b667-4833-9e03-f6eaaff9f8c1-var-log-ovn\") pod \"ovn-controller-zb9k2-config-jvnzd\" (UID: \"405d2caf-b667-4833-9e03-f6eaaff9f8c1\") " pod="openstack/ovn-controller-zb9k2-config-jvnzd" Mar 13 15:44:36 crc kubenswrapper[4907]: I0313 15:44:36.512564 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x4w5r\" (UniqueName: \"kubernetes.io/projected/405d2caf-b667-4833-9e03-f6eaaff9f8c1-kube-api-access-x4w5r\") pod \"ovn-controller-zb9k2-config-jvnzd\" (UID: \"405d2caf-b667-4833-9e03-f6eaaff9f8c1\") " pod="openstack/ovn-controller-zb9k2-config-jvnzd" Mar 13 15:44:36 crc kubenswrapper[4907]: I0313 15:44:36.512706 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/405d2caf-b667-4833-9e03-f6eaaff9f8c1-var-log-ovn\") pod \"ovn-controller-zb9k2-config-jvnzd\" (UID: \"405d2caf-b667-4833-9e03-f6eaaff9f8c1\") " pod="openstack/ovn-controller-zb9k2-config-jvnzd" Mar 13 15:44:36 crc kubenswrapper[4907]: I0313 15:44:36.512800 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/405d2caf-b667-4833-9e03-f6eaaff9f8c1-var-run-ovn\") pod \"ovn-controller-zb9k2-config-jvnzd\" (UID: \"405d2caf-b667-4833-9e03-f6eaaff9f8c1\") " pod="openstack/ovn-controller-zb9k2-config-jvnzd" Mar 13 15:44:36 crc kubenswrapper[4907]: I0313 15:44:36.512913 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/405d2caf-b667-4833-9e03-f6eaaff9f8c1-scripts\") pod \"ovn-controller-zb9k2-config-jvnzd\" (UID: \"405d2caf-b667-4833-9e03-f6eaaff9f8c1\") " pod="openstack/ovn-controller-zb9k2-config-jvnzd" Mar 13 15:44:36 crc kubenswrapper[4907]: I0313 15:44:36.512958 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/405d2caf-b667-4833-9e03-f6eaaff9f8c1-additional-scripts\") pod \"ovn-controller-zb9k2-config-jvnzd\" (UID: \"405d2caf-b667-4833-9e03-f6eaaff9f8c1\") " pod="openstack/ovn-controller-zb9k2-config-jvnzd" Mar 13 15:44:36 crc kubenswrapper[4907]: I0313 15:44:36.512977 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/405d2caf-b667-4833-9e03-f6eaaff9f8c1-var-run\") pod \"ovn-controller-zb9k2-config-jvnzd\" (UID: \"405d2caf-b667-4833-9e03-f6eaaff9f8c1\") " pod="openstack/ovn-controller-zb9k2-config-jvnzd" Mar 13 15:44:36 crc kubenswrapper[4907]: I0313 15:44:36.513132 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/405d2caf-b667-4833-9e03-f6eaaff9f8c1-var-log-ovn\") pod \"ovn-controller-zb9k2-config-jvnzd\" (UID: \"405d2caf-b667-4833-9e03-f6eaaff9f8c1\") " pod="openstack/ovn-controller-zb9k2-config-jvnzd" Mar 13 15:44:36 crc kubenswrapper[4907]: I0313 15:44:36.513156 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/405d2caf-b667-4833-9e03-f6eaaff9f8c1-var-run-ovn\") pod \"ovn-controller-zb9k2-config-jvnzd\" (UID: \"405d2caf-b667-4833-9e03-f6eaaff9f8c1\") " pod="openstack/ovn-controller-zb9k2-config-jvnzd" Mar 13 15:44:36 crc kubenswrapper[4907]: I0313 15:44:36.513186 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/405d2caf-b667-4833-9e03-f6eaaff9f8c1-var-run\") pod \"ovn-controller-zb9k2-config-jvnzd\" (UID: \"405d2caf-b667-4833-9e03-f6eaaff9f8c1\") " pod="openstack/ovn-controller-zb9k2-config-jvnzd" Mar 13 15:44:36 crc kubenswrapper[4907]: I0313 15:44:36.514097 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/405d2caf-b667-4833-9e03-f6eaaff9f8c1-additional-scripts\") pod \"ovn-controller-zb9k2-config-jvnzd\" (UID: \"405d2caf-b667-4833-9e03-f6eaaff9f8c1\") " pod="openstack/ovn-controller-zb9k2-config-jvnzd" Mar 13 15:44:36 crc kubenswrapper[4907]: I0313 15:44:36.515990 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/405d2caf-b667-4833-9e03-f6eaaff9f8c1-scripts\") pod \"ovn-controller-zb9k2-config-jvnzd\" (UID: \"405d2caf-b667-4833-9e03-f6eaaff9f8c1\") " pod="openstack/ovn-controller-zb9k2-config-jvnzd" Mar 13 15:44:36 crc kubenswrapper[4907]: I0313 15:44:36.531897 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x4w5r\" (UniqueName: \"kubernetes.io/projected/405d2caf-b667-4833-9e03-f6eaaff9f8c1-kube-api-access-x4w5r\") pod \"ovn-controller-zb9k2-config-jvnzd\" (UID: \"405d2caf-b667-4833-9e03-f6eaaff9f8c1\") " pod="openstack/ovn-controller-zb9k2-config-jvnzd" Mar 13 15:44:36 crc kubenswrapper[4907]: I0313 15:44:36.622487 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-zb9k2-config-jvnzd" Mar 13 15:44:37 crc kubenswrapper[4907]: I0313 15:44:37.086351 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-zb9k2-config-jvnzd"] Mar 13 15:44:37 crc kubenswrapper[4907]: W0313 15:44:37.092064 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod405d2caf_b667_4833_9e03_f6eaaff9f8c1.slice/crio-69e762bd095eb220585503bedb5a31de1a473c47dfddf7b51f8a8e391d564369 WatchSource:0}: Error finding container 69e762bd095eb220585503bedb5a31de1a473c47dfddf7b51f8a8e391d564369: Status 404 returned error can't find the container with id 69e762bd095eb220585503bedb5a31de1a473c47dfddf7b51f8a8e391d564369 Mar 13 15:44:37 crc kubenswrapper[4907]: I0313 15:44:37.125578 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-zb9k2" Mar 13 15:44:37 crc kubenswrapper[4907]: I0313 15:44:37.799528 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ecd299e1-64f9-4ad1-9ff8-af388e993db9" path="/var/lib/kubelet/pods/ecd299e1-64f9-4ad1-9ff8-af388e993db9/volumes" Mar 13 15:44:37 crc kubenswrapper[4907]: I0313 15:44:37.805560 4907 generic.go:334] "Generic (PLEG): container finished" podID="405d2caf-b667-4833-9e03-f6eaaff9f8c1" containerID="fea279a9a11b2a38768139c4f75fb40e14428aafd4173f88229129062a569479" exitCode=0 Mar 13 15:44:37 crc kubenswrapper[4907]: I0313 15:44:37.805629 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-zb9k2-config-jvnzd" event={"ID":"405d2caf-b667-4833-9e03-f6eaaff9f8c1","Type":"ContainerDied","Data":"fea279a9a11b2a38768139c4f75fb40e14428aafd4173f88229129062a569479"} Mar 13 15:44:37 crc kubenswrapper[4907]: I0313 15:44:37.805686 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-zb9k2-config-jvnzd" event={"ID":"405d2caf-b667-4833-9e03-f6eaaff9f8c1","Type":"ContainerStarted","Data":"69e762bd095eb220585503bedb5a31de1a473c47dfddf7b51f8a8e391d564369"} Mar 13 15:44:44 crc kubenswrapper[4907]: I0313 15:44:44.559804 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-zb9k2-config-jvnzd" Mar 13 15:44:44 crc kubenswrapper[4907]: I0313 15:44:44.575700 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/405d2caf-b667-4833-9e03-f6eaaff9f8c1-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "405d2caf-b667-4833-9e03-f6eaaff9f8c1" (UID: "405d2caf-b667-4833-9e03-f6eaaff9f8c1"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 15:44:44 crc kubenswrapper[4907]: I0313 15:44:44.575646 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/405d2caf-b667-4833-9e03-f6eaaff9f8c1-var-run-ovn\") pod \"405d2caf-b667-4833-9e03-f6eaaff9f8c1\" (UID: \"405d2caf-b667-4833-9e03-f6eaaff9f8c1\") " Mar 13 15:44:44 crc kubenswrapper[4907]: I0313 15:44:44.575778 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/405d2caf-b667-4833-9e03-f6eaaff9f8c1-var-run\") pod \"405d2caf-b667-4833-9e03-f6eaaff9f8c1\" (UID: \"405d2caf-b667-4833-9e03-f6eaaff9f8c1\") " Mar 13 15:44:44 crc kubenswrapper[4907]: I0313 15:44:44.575892 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4w5r\" (UniqueName: \"kubernetes.io/projected/405d2caf-b667-4833-9e03-f6eaaff9f8c1-kube-api-access-x4w5r\") pod \"405d2caf-b667-4833-9e03-f6eaaff9f8c1\" (UID: \"405d2caf-b667-4833-9e03-f6eaaff9f8c1\") " Mar 13 15:44:44 crc kubenswrapper[4907]: I0313 15:44:44.575926 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/405d2caf-b667-4833-9e03-f6eaaff9f8c1-additional-scripts\") pod \"405d2caf-b667-4833-9e03-f6eaaff9f8c1\" (UID: \"405d2caf-b667-4833-9e03-f6eaaff9f8c1\") " Mar 13 15:44:44 crc kubenswrapper[4907]: I0313 15:44:44.575980 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/405d2caf-b667-4833-9e03-f6eaaff9f8c1-scripts\") pod \"405d2caf-b667-4833-9e03-f6eaaff9f8c1\" (UID: \"405d2caf-b667-4833-9e03-f6eaaff9f8c1\") " Mar 13 15:44:44 crc kubenswrapper[4907]: I0313 15:44:44.576069 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/405d2caf-b667-4833-9e03-f6eaaff9f8c1-var-log-ovn\") pod \"405d2caf-b667-4833-9e03-f6eaaff9f8c1\" (UID: \"405d2caf-b667-4833-9e03-f6eaaff9f8c1\") " Mar 13 15:44:44 crc kubenswrapper[4907]: I0313 15:44:44.576464 4907 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/405d2caf-b667-4833-9e03-f6eaaff9f8c1-var-run-ovn\") on node \"crc\" DevicePath \"\"" Mar 13 15:44:44 crc kubenswrapper[4907]: I0313 15:44:44.576505 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/405d2caf-b667-4833-9e03-f6eaaff9f8c1-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "405d2caf-b667-4833-9e03-f6eaaff9f8c1" (UID: "405d2caf-b667-4833-9e03-f6eaaff9f8c1"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 15:44:44 crc kubenswrapper[4907]: I0313 15:44:44.576532 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/405d2caf-b667-4833-9e03-f6eaaff9f8c1-var-run" (OuterVolumeSpecName: "var-run") pod "405d2caf-b667-4833-9e03-f6eaaff9f8c1" (UID: "405d2caf-b667-4833-9e03-f6eaaff9f8c1"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 15:44:44 crc kubenswrapper[4907]: I0313 15:44:44.578110 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/405d2caf-b667-4833-9e03-f6eaaff9f8c1-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "405d2caf-b667-4833-9e03-f6eaaff9f8c1" (UID: "405d2caf-b667-4833-9e03-f6eaaff9f8c1"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:44:44 crc kubenswrapper[4907]: I0313 15:44:44.579062 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/405d2caf-b667-4833-9e03-f6eaaff9f8c1-scripts" (OuterVolumeSpecName: "scripts") pod "405d2caf-b667-4833-9e03-f6eaaff9f8c1" (UID: "405d2caf-b667-4833-9e03-f6eaaff9f8c1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:44:44 crc kubenswrapper[4907]: I0313 15:44:44.582070 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/405d2caf-b667-4833-9e03-f6eaaff9f8c1-kube-api-access-x4w5r" (OuterVolumeSpecName: "kube-api-access-x4w5r") pod "405d2caf-b667-4833-9e03-f6eaaff9f8c1" (UID: "405d2caf-b667-4833-9e03-f6eaaff9f8c1"). InnerVolumeSpecName "kube-api-access-x4w5r". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:44:44 crc kubenswrapper[4907]: I0313 15:44:44.677692 4907 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/405d2caf-b667-4833-9e03-f6eaaff9f8c1-var-log-ovn\") on node \"crc\" DevicePath \"\"" Mar 13 15:44:44 crc kubenswrapper[4907]: I0313 15:44:44.677936 4907 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/405d2caf-b667-4833-9e03-f6eaaff9f8c1-var-run\") on node \"crc\" DevicePath \"\"" Mar 13 15:44:44 crc kubenswrapper[4907]: I0313 15:44:44.677949 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4w5r\" (UniqueName: \"kubernetes.io/projected/405d2caf-b667-4833-9e03-f6eaaff9f8c1-kube-api-access-x4w5r\") on node \"crc\" DevicePath \"\"" Mar 13 15:44:44 crc kubenswrapper[4907]: I0313 15:44:44.677959 4907 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/405d2caf-b667-4833-9e03-f6eaaff9f8c1-additional-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 15:44:44 crc kubenswrapper[4907]: I0313 15:44:44.677969 4907 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/405d2caf-b667-4833-9e03-f6eaaff9f8c1-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 15:44:44 crc kubenswrapper[4907]: I0313 15:44:44.873637 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-zb9k2-config-jvnzd" Mar 13 15:44:44 crc kubenswrapper[4907]: I0313 15:44:44.873782 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-zb9k2-config-jvnzd" event={"ID":"405d2caf-b667-4833-9e03-f6eaaff9f8c1","Type":"ContainerDied","Data":"69e762bd095eb220585503bedb5a31de1a473c47dfddf7b51f8a8e391d564369"} Mar 13 15:44:44 crc kubenswrapper[4907]: I0313 15:44:44.873819 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="69e762bd095eb220585503bedb5a31de1a473c47dfddf7b51f8a8e391d564369" Mar 13 15:44:44 crc kubenswrapper[4907]: I0313 15:44:44.875299 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-api-6db876fcbd-4wtkj" event={"ID":"1c868ec4-5e49-4be1-a1eb-c4748753aecb","Type":"ContainerStarted","Data":"55fb7423f1d00e837760507cfb3fc4502b1b604d521e1f69652f3853242bd445"} Mar 13 15:44:45 crc kubenswrapper[4907]: I0313 15:44:45.638521 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-zb9k2-config-jvnzd"] Mar 13 15:44:46 crc kubenswrapper[4907]: I0313 15:44:46.193791 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-zb9k2-config-jvnzd"] Mar 13 15:44:46 crc kubenswrapper[4907]: I0313 15:44:46.196935 4907 generic.go:334] "Generic (PLEG): container finished" podID="1c868ec4-5e49-4be1-a1eb-c4748753aecb" containerID="55fb7423f1d00e837760507cfb3fc4502b1b604d521e1f69652f3853242bd445" exitCode=0 Mar 13 15:44:46 crc kubenswrapper[4907]: I0313 15:44:46.196966 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-api-6db876fcbd-4wtkj" event={"ID":"1c868ec4-5e49-4be1-a1eb-c4748753aecb","Type":"ContainerDied","Data":"55fb7423f1d00e837760507cfb3fc4502b1b604d521e1f69652f3853242bd445"} Mar 13 15:44:46 crc kubenswrapper[4907]: I0313 15:44:46.199546 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-zb9k2-config-2d8dv"] Mar 13 15:44:46 crc kubenswrapper[4907]: E0313 15:44:46.200239 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="405d2caf-b667-4833-9e03-f6eaaff9f8c1" containerName="ovn-config" Mar 13 15:44:46 crc kubenswrapper[4907]: I0313 15:44:46.200276 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="405d2caf-b667-4833-9e03-f6eaaff9f8c1" containerName="ovn-config" Mar 13 15:44:46 crc kubenswrapper[4907]: I0313 15:44:46.200554 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="405d2caf-b667-4833-9e03-f6eaaff9f8c1" containerName="ovn-config" Mar 13 15:44:46 crc kubenswrapper[4907]: I0313 15:44:46.201453 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-zb9k2-config-2d8dv" Mar 13 15:44:46 crc kubenswrapper[4907]: I0313 15:44:46.205131 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Mar 13 15:44:46 crc kubenswrapper[4907]: I0313 15:44:46.210983 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-zb9k2-config-2d8dv"] Mar 13 15:44:46 crc kubenswrapper[4907]: I0313 15:44:46.337633 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/96a87d0f-b3d7-447b-a8db-1ad11b93f37a-var-run\") pod \"ovn-controller-zb9k2-config-2d8dv\" (UID: \"96a87d0f-b3d7-447b-a8db-1ad11b93f37a\") " pod="openstack/ovn-controller-zb9k2-config-2d8dv" Mar 13 15:44:46 crc kubenswrapper[4907]: I0313 15:44:46.337678 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/96a87d0f-b3d7-447b-a8db-1ad11b93f37a-scripts\") pod \"ovn-controller-zb9k2-config-2d8dv\" (UID: \"96a87d0f-b3d7-447b-a8db-1ad11b93f37a\") " pod="openstack/ovn-controller-zb9k2-config-2d8dv" Mar 13 15:44:46 crc kubenswrapper[4907]: I0313 15:44:46.337749 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d8zkc\" (UniqueName: \"kubernetes.io/projected/96a87d0f-b3d7-447b-a8db-1ad11b93f37a-kube-api-access-d8zkc\") pod \"ovn-controller-zb9k2-config-2d8dv\" (UID: \"96a87d0f-b3d7-447b-a8db-1ad11b93f37a\") " pod="openstack/ovn-controller-zb9k2-config-2d8dv" Mar 13 15:44:46 crc kubenswrapper[4907]: I0313 15:44:46.337779 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/96a87d0f-b3d7-447b-a8db-1ad11b93f37a-var-run-ovn\") pod \"ovn-controller-zb9k2-config-2d8dv\" (UID: \"96a87d0f-b3d7-447b-a8db-1ad11b93f37a\") " pod="openstack/ovn-controller-zb9k2-config-2d8dv" Mar 13 15:44:46 crc kubenswrapper[4907]: I0313 15:44:46.337831 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/96a87d0f-b3d7-447b-a8db-1ad11b93f37a-var-log-ovn\") pod \"ovn-controller-zb9k2-config-2d8dv\" (UID: \"96a87d0f-b3d7-447b-a8db-1ad11b93f37a\") " pod="openstack/ovn-controller-zb9k2-config-2d8dv" Mar 13 15:44:46 crc kubenswrapper[4907]: I0313 15:44:46.337907 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/96a87d0f-b3d7-447b-a8db-1ad11b93f37a-additional-scripts\") pod \"ovn-controller-zb9k2-config-2d8dv\" (UID: \"96a87d0f-b3d7-447b-a8db-1ad11b93f37a\") " pod="openstack/ovn-controller-zb9k2-config-2d8dv" Mar 13 15:44:46 crc kubenswrapper[4907]: I0313 15:44:46.439927 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/96a87d0f-b3d7-447b-a8db-1ad11b93f37a-var-run\") pod \"ovn-controller-zb9k2-config-2d8dv\" (UID: \"96a87d0f-b3d7-447b-a8db-1ad11b93f37a\") " pod="openstack/ovn-controller-zb9k2-config-2d8dv" Mar 13 15:44:46 crc kubenswrapper[4907]: I0313 15:44:46.439976 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/96a87d0f-b3d7-447b-a8db-1ad11b93f37a-scripts\") pod \"ovn-controller-zb9k2-config-2d8dv\" (UID: \"96a87d0f-b3d7-447b-a8db-1ad11b93f37a\") " pod="openstack/ovn-controller-zb9k2-config-2d8dv" Mar 13 15:44:46 crc kubenswrapper[4907]: I0313 15:44:46.440026 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d8zkc\" (UniqueName: \"kubernetes.io/projected/96a87d0f-b3d7-447b-a8db-1ad11b93f37a-kube-api-access-d8zkc\") pod \"ovn-controller-zb9k2-config-2d8dv\" (UID: \"96a87d0f-b3d7-447b-a8db-1ad11b93f37a\") " pod="openstack/ovn-controller-zb9k2-config-2d8dv" Mar 13 15:44:46 crc kubenswrapper[4907]: I0313 15:44:46.440047 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/96a87d0f-b3d7-447b-a8db-1ad11b93f37a-var-run-ovn\") pod \"ovn-controller-zb9k2-config-2d8dv\" (UID: \"96a87d0f-b3d7-447b-a8db-1ad11b93f37a\") " pod="openstack/ovn-controller-zb9k2-config-2d8dv" Mar 13 15:44:46 crc kubenswrapper[4907]: I0313 15:44:46.440112 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/96a87d0f-b3d7-447b-a8db-1ad11b93f37a-var-log-ovn\") pod \"ovn-controller-zb9k2-config-2d8dv\" (UID: \"96a87d0f-b3d7-447b-a8db-1ad11b93f37a\") " pod="openstack/ovn-controller-zb9k2-config-2d8dv" Mar 13 15:44:46 crc kubenswrapper[4907]: I0313 15:44:46.440155 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/96a87d0f-b3d7-447b-a8db-1ad11b93f37a-additional-scripts\") pod \"ovn-controller-zb9k2-config-2d8dv\" (UID: \"96a87d0f-b3d7-447b-a8db-1ad11b93f37a\") " pod="openstack/ovn-controller-zb9k2-config-2d8dv" Mar 13 15:44:46 crc kubenswrapper[4907]: I0313 15:44:46.440504 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/96a87d0f-b3d7-447b-a8db-1ad11b93f37a-var-run-ovn\") pod \"ovn-controller-zb9k2-config-2d8dv\" (UID: \"96a87d0f-b3d7-447b-a8db-1ad11b93f37a\") " pod="openstack/ovn-controller-zb9k2-config-2d8dv" Mar 13 15:44:46 crc kubenswrapper[4907]: I0313 15:44:46.440656 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/96a87d0f-b3d7-447b-a8db-1ad11b93f37a-var-log-ovn\") pod \"ovn-controller-zb9k2-config-2d8dv\" (UID: \"96a87d0f-b3d7-447b-a8db-1ad11b93f37a\") " pod="openstack/ovn-controller-zb9k2-config-2d8dv" Mar 13 15:44:46 crc kubenswrapper[4907]: I0313 15:44:46.440977 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/96a87d0f-b3d7-447b-a8db-1ad11b93f37a-additional-scripts\") pod \"ovn-controller-zb9k2-config-2d8dv\" (UID: \"96a87d0f-b3d7-447b-a8db-1ad11b93f37a\") " pod="openstack/ovn-controller-zb9k2-config-2d8dv" Mar 13 15:44:46 crc kubenswrapper[4907]: I0313 15:44:46.442642 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/96a87d0f-b3d7-447b-a8db-1ad11b93f37a-scripts\") pod \"ovn-controller-zb9k2-config-2d8dv\" (UID: \"96a87d0f-b3d7-447b-a8db-1ad11b93f37a\") " pod="openstack/ovn-controller-zb9k2-config-2d8dv" Mar 13 15:44:46 crc kubenswrapper[4907]: I0313 15:44:46.442735 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/96a87d0f-b3d7-447b-a8db-1ad11b93f37a-var-run\") pod \"ovn-controller-zb9k2-config-2d8dv\" (UID: \"96a87d0f-b3d7-447b-a8db-1ad11b93f37a\") " pod="openstack/ovn-controller-zb9k2-config-2d8dv" Mar 13 15:44:46 crc kubenswrapper[4907]: I0313 15:44:46.461486 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d8zkc\" (UniqueName: \"kubernetes.io/projected/96a87d0f-b3d7-447b-a8db-1ad11b93f37a-kube-api-access-d8zkc\") pod \"ovn-controller-zb9k2-config-2d8dv\" (UID: \"96a87d0f-b3d7-447b-a8db-1ad11b93f37a\") " pod="openstack/ovn-controller-zb9k2-config-2d8dv" Mar 13 15:44:46 crc kubenswrapper[4907]: I0313 15:44:46.695254 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-zb9k2-config-2d8dv" Mar 13 15:44:46 crc kubenswrapper[4907]: I0313 15:44:46.783280 4907 scope.go:117] "RemoveContainer" containerID="68ffe92dd03078633ec89d7ef29da9bbdc4c5f04c956a1e8487e3f2e01c5d1c5" Mar 13 15:44:46 crc kubenswrapper[4907]: E0313 15:44:46.783641 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:44:47 crc kubenswrapper[4907]: I0313 15:44:47.185570 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-zb9k2-config-2d8dv"] Mar 13 15:44:47 crc kubenswrapper[4907]: W0313 15:44:47.192849 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod96a87d0f_b3d7_447b_a8db_1ad11b93f37a.slice/crio-e01fd45fb886de621cf61d6ee0706af4f41676bc2037299d72cd797bdbf65768 WatchSource:0}: Error finding container e01fd45fb886de621cf61d6ee0706af4f41676bc2037299d72cd797bdbf65768: Status 404 returned error can't find the container with id e01fd45fb886de621cf61d6ee0706af4f41676bc2037299d72cd797bdbf65768 Mar 13 15:44:47 crc kubenswrapper[4907]: I0313 15:44:47.209208 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-zb9k2-config-2d8dv" event={"ID":"96a87d0f-b3d7-447b-a8db-1ad11b93f37a","Type":"ContainerStarted","Data":"e01fd45fb886de621cf61d6ee0706af4f41676bc2037299d72cd797bdbf65768"} Mar 13 15:44:47 crc kubenswrapper[4907]: I0313 15:44:47.211311 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-api-6db876fcbd-4wtkj" event={"ID":"1c868ec4-5e49-4be1-a1eb-c4748753aecb","Type":"ContainerStarted","Data":"432213e092050825f06d1f9ead3214f9559b8e4a02723df2cf0dd05d48859ce9"} Mar 13 15:44:47 crc kubenswrapper[4907]: I0313 15:44:47.211345 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-api-6db876fcbd-4wtkj" event={"ID":"1c868ec4-5e49-4be1-a1eb-c4748753aecb","Type":"ContainerStarted","Data":"1a0a2f6ef0a9cb2e75b38a9ad6ca30d1ad8381dc759967de44f409f3e35e7c40"} Mar 13 15:44:47 crc kubenswrapper[4907]: I0313 15:44:47.211594 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-api-6db876fcbd-4wtkj" Mar 13 15:44:47 crc kubenswrapper[4907]: I0313 15:44:47.250704 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-api-6db876fcbd-4wtkj" podStartSLOduration=3.88665379 podStartE2EDuration="13.250685772s" podCreationTimestamp="2026-03-13 15:44:34 +0000 UTC" firstStartedPulling="2026-03-13 15:44:35.168094025 +0000 UTC m=+5974.067881714" lastFinishedPulling="2026-03-13 15:44:44.532126007 +0000 UTC m=+5983.431913696" observedRunningTime="2026-03-13 15:44:47.243152357 +0000 UTC m=+5986.142940056" watchObservedRunningTime="2026-03-13 15:44:47.250685772 +0000 UTC m=+5986.150473461" Mar 13 15:44:47 crc kubenswrapper[4907]: I0313 15:44:47.792827 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="405d2caf-b667-4833-9e03-f6eaaff9f8c1" path="/var/lib/kubelet/pods/405d2caf-b667-4833-9e03-f6eaaff9f8c1/volumes" Mar 13 15:44:48 crc kubenswrapper[4907]: I0313 15:44:48.222649 4907 generic.go:334] "Generic (PLEG): container finished" podID="96a87d0f-b3d7-447b-a8db-1ad11b93f37a" containerID="37e65533c5fc864e730a3583c9b7118a1665874961e44f735e26e376e6b5f39d" exitCode=0 Mar 13 15:44:48 crc kubenswrapper[4907]: I0313 15:44:48.223027 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-zb9k2-config-2d8dv" event={"ID":"96a87d0f-b3d7-447b-a8db-1ad11b93f37a","Type":"ContainerDied","Data":"37e65533c5fc864e730a3583c9b7118a1665874961e44f735e26e376e6b5f39d"} Mar 13 15:44:48 crc kubenswrapper[4907]: I0313 15:44:48.223153 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-api-6db876fcbd-4wtkj" Mar 13 15:44:49 crc kubenswrapper[4907]: I0313 15:44:49.664387 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-zb9k2-config-2d8dv" Mar 13 15:44:49 crc kubenswrapper[4907]: I0313 15:44:49.805139 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/96a87d0f-b3d7-447b-a8db-1ad11b93f37a-var-run-ovn\") pod \"96a87d0f-b3d7-447b-a8db-1ad11b93f37a\" (UID: \"96a87d0f-b3d7-447b-a8db-1ad11b93f37a\") " Mar 13 15:44:49 crc kubenswrapper[4907]: I0313 15:44:49.805267 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/96a87d0f-b3d7-447b-a8db-1ad11b93f37a-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "96a87d0f-b3d7-447b-a8db-1ad11b93f37a" (UID: "96a87d0f-b3d7-447b-a8db-1ad11b93f37a"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 15:44:49 crc kubenswrapper[4907]: I0313 15:44:49.805319 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d8zkc\" (UniqueName: \"kubernetes.io/projected/96a87d0f-b3d7-447b-a8db-1ad11b93f37a-kube-api-access-d8zkc\") pod \"96a87d0f-b3d7-447b-a8db-1ad11b93f37a\" (UID: \"96a87d0f-b3d7-447b-a8db-1ad11b93f37a\") " Mar 13 15:44:49 crc kubenswrapper[4907]: I0313 15:44:49.805349 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/96a87d0f-b3d7-447b-a8db-1ad11b93f37a-additional-scripts\") pod \"96a87d0f-b3d7-447b-a8db-1ad11b93f37a\" (UID: \"96a87d0f-b3d7-447b-a8db-1ad11b93f37a\") " Mar 13 15:44:49 crc kubenswrapper[4907]: I0313 15:44:49.805380 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/96a87d0f-b3d7-447b-a8db-1ad11b93f37a-var-run\") pod \"96a87d0f-b3d7-447b-a8db-1ad11b93f37a\" (UID: \"96a87d0f-b3d7-447b-a8db-1ad11b93f37a\") " Mar 13 15:44:49 crc kubenswrapper[4907]: I0313 15:44:49.805468 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/96a87d0f-b3d7-447b-a8db-1ad11b93f37a-var-log-ovn\") pod \"96a87d0f-b3d7-447b-a8db-1ad11b93f37a\" (UID: \"96a87d0f-b3d7-447b-a8db-1ad11b93f37a\") " Mar 13 15:44:49 crc kubenswrapper[4907]: I0313 15:44:49.805519 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/96a87d0f-b3d7-447b-a8db-1ad11b93f37a-scripts\") pod \"96a87d0f-b3d7-447b-a8db-1ad11b93f37a\" (UID: \"96a87d0f-b3d7-447b-a8db-1ad11b93f37a\") " Mar 13 15:44:49 crc kubenswrapper[4907]: I0313 15:44:49.805611 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/96a87d0f-b3d7-447b-a8db-1ad11b93f37a-var-run" (OuterVolumeSpecName: "var-run") pod "96a87d0f-b3d7-447b-a8db-1ad11b93f37a" (UID: "96a87d0f-b3d7-447b-a8db-1ad11b93f37a"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 15:44:49 crc kubenswrapper[4907]: I0313 15:44:49.805643 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/96a87d0f-b3d7-447b-a8db-1ad11b93f37a-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "96a87d0f-b3d7-447b-a8db-1ad11b93f37a" (UID: "96a87d0f-b3d7-447b-a8db-1ad11b93f37a"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 15:44:49 crc kubenswrapper[4907]: I0313 15:44:49.806015 4907 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/96a87d0f-b3d7-447b-a8db-1ad11b93f37a-var-log-ovn\") on node \"crc\" DevicePath \"\"" Mar 13 15:44:49 crc kubenswrapper[4907]: I0313 15:44:49.806037 4907 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/96a87d0f-b3d7-447b-a8db-1ad11b93f37a-var-run-ovn\") on node \"crc\" DevicePath \"\"" Mar 13 15:44:49 crc kubenswrapper[4907]: I0313 15:44:49.806048 4907 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/96a87d0f-b3d7-447b-a8db-1ad11b93f37a-var-run\") on node \"crc\" DevicePath \"\"" Mar 13 15:44:49 crc kubenswrapper[4907]: I0313 15:44:49.818694 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96a87d0f-b3d7-447b-a8db-1ad11b93f37a-kube-api-access-d8zkc" (OuterVolumeSpecName: "kube-api-access-d8zkc") pod "96a87d0f-b3d7-447b-a8db-1ad11b93f37a" (UID: "96a87d0f-b3d7-447b-a8db-1ad11b93f37a"). InnerVolumeSpecName "kube-api-access-d8zkc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:44:49 crc kubenswrapper[4907]: I0313 15:44:49.856498 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/96a87d0f-b3d7-447b-a8db-1ad11b93f37a-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "96a87d0f-b3d7-447b-a8db-1ad11b93f37a" (UID: "96a87d0f-b3d7-447b-a8db-1ad11b93f37a"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:44:49 crc kubenswrapper[4907]: I0313 15:44:49.856711 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/96a87d0f-b3d7-447b-a8db-1ad11b93f37a-scripts" (OuterVolumeSpecName: "scripts") pod "96a87d0f-b3d7-447b-a8db-1ad11b93f37a" (UID: "96a87d0f-b3d7-447b-a8db-1ad11b93f37a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:44:49 crc kubenswrapper[4907]: I0313 15:44:49.908175 4907 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/96a87d0f-b3d7-447b-a8db-1ad11b93f37a-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 15:44:49 crc kubenswrapper[4907]: I0313 15:44:49.908217 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d8zkc\" (UniqueName: \"kubernetes.io/projected/96a87d0f-b3d7-447b-a8db-1ad11b93f37a-kube-api-access-d8zkc\") on node \"crc\" DevicePath \"\"" Mar 13 15:44:49 crc kubenswrapper[4907]: I0313 15:44:49.908231 4907 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/96a87d0f-b3d7-447b-a8db-1ad11b93f37a-additional-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 15:44:50 crc kubenswrapper[4907]: I0313 15:44:50.260472 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-zb9k2-config-2d8dv" event={"ID":"96a87d0f-b3d7-447b-a8db-1ad11b93f37a","Type":"ContainerDied","Data":"e01fd45fb886de621cf61d6ee0706af4f41676bc2037299d72cd797bdbf65768"} Mar 13 15:44:50 crc kubenswrapper[4907]: I0313 15:44:50.260550 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e01fd45fb886de621cf61d6ee0706af4f41676bc2037299d72cd797bdbf65768" Mar 13 15:44:50 crc kubenswrapper[4907]: I0313 15:44:50.260636 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-zb9k2-config-2d8dv" Mar 13 15:44:50 crc kubenswrapper[4907]: I0313 15:44:50.759613 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-zb9k2-config-2d8dv"] Mar 13 15:44:50 crc kubenswrapper[4907]: I0313 15:44:50.777551 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-zb9k2-config-2d8dv"] Mar 13 15:44:51 crc kubenswrapper[4907]: I0313 15:44:51.792459 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96a87d0f-b3d7-447b-a8db-1ad11b93f37a" path="/var/lib/kubelet/pods/96a87d0f-b3d7-447b-a8db-1ad11b93f37a/volumes" Mar 13 15:44:57 crc kubenswrapper[4907]: I0313 15:44:57.646461 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-rsyslog-zbs6z"] Mar 13 15:44:57 crc kubenswrapper[4907]: E0313 15:44:57.647431 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96a87d0f-b3d7-447b-a8db-1ad11b93f37a" containerName="ovn-config" Mar 13 15:44:57 crc kubenswrapper[4907]: I0313 15:44:57.647450 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="96a87d0f-b3d7-447b-a8db-1ad11b93f37a" containerName="ovn-config" Mar 13 15:44:57 crc kubenswrapper[4907]: I0313 15:44:57.647707 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="96a87d0f-b3d7-447b-a8db-1ad11b93f37a" containerName="ovn-config" Mar 13 15:44:57 crc kubenswrapper[4907]: I0313 15:44:57.648687 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-rsyslog-zbs6z" Mar 13 15:44:57 crc kubenswrapper[4907]: I0313 15:44:57.657154 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-rsyslog-config-data" Mar 13 15:44:57 crc kubenswrapper[4907]: I0313 15:44:57.657275 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"octavia-hmport-map" Mar 13 15:44:57 crc kubenswrapper[4907]: I0313 15:44:57.657606 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-rsyslog-scripts" Mar 13 15:44:57 crc kubenswrapper[4907]: I0313 15:44:57.674168 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-rsyslog-zbs6z"] Mar 13 15:44:57 crc kubenswrapper[4907]: I0313 15:44:57.750662 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/4df4068e-249f-47ea-a84a-3c178c27bdf3-hm-ports\") pod \"octavia-rsyslog-zbs6z\" (UID: \"4df4068e-249f-47ea-a84a-3c178c27bdf3\") " pod="openstack/octavia-rsyslog-zbs6z" Mar 13 15:44:57 crc kubenswrapper[4907]: I0313 15:44:57.751179 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4df4068e-249f-47ea-a84a-3c178c27bdf3-scripts\") pod \"octavia-rsyslog-zbs6z\" (UID: \"4df4068e-249f-47ea-a84a-3c178c27bdf3\") " pod="openstack/octavia-rsyslog-zbs6z" Mar 13 15:44:57 crc kubenswrapper[4907]: I0313 15:44:57.751242 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/4df4068e-249f-47ea-a84a-3c178c27bdf3-config-data-merged\") pod \"octavia-rsyslog-zbs6z\" (UID: \"4df4068e-249f-47ea-a84a-3c178c27bdf3\") " pod="openstack/octavia-rsyslog-zbs6z" Mar 13 15:44:57 crc kubenswrapper[4907]: I0313 15:44:57.751486 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4df4068e-249f-47ea-a84a-3c178c27bdf3-config-data\") pod \"octavia-rsyslog-zbs6z\" (UID: \"4df4068e-249f-47ea-a84a-3c178c27bdf3\") " pod="openstack/octavia-rsyslog-zbs6z" Mar 13 15:44:57 crc kubenswrapper[4907]: I0313 15:44:57.782852 4907 scope.go:117] "RemoveContainer" containerID="68ffe92dd03078633ec89d7ef29da9bbdc4c5f04c956a1e8487e3f2e01c5d1c5" Mar 13 15:44:57 crc kubenswrapper[4907]: E0313 15:44:57.783150 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:44:57 crc kubenswrapper[4907]: I0313 15:44:57.854033 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4df4068e-249f-47ea-a84a-3c178c27bdf3-config-data\") pod \"octavia-rsyslog-zbs6z\" (UID: \"4df4068e-249f-47ea-a84a-3c178c27bdf3\") " pod="openstack/octavia-rsyslog-zbs6z" Mar 13 15:44:57 crc kubenswrapper[4907]: I0313 15:44:57.854404 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/4df4068e-249f-47ea-a84a-3c178c27bdf3-hm-ports\") pod \"octavia-rsyslog-zbs6z\" (UID: \"4df4068e-249f-47ea-a84a-3c178c27bdf3\") " pod="openstack/octavia-rsyslog-zbs6z" Mar 13 15:44:57 crc kubenswrapper[4907]: I0313 15:44:57.854660 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4df4068e-249f-47ea-a84a-3c178c27bdf3-scripts\") pod \"octavia-rsyslog-zbs6z\" (UID: \"4df4068e-249f-47ea-a84a-3c178c27bdf3\") " pod="openstack/octavia-rsyslog-zbs6z" Mar 13 15:44:57 crc kubenswrapper[4907]: I0313 15:44:57.854745 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/4df4068e-249f-47ea-a84a-3c178c27bdf3-config-data-merged\") pod \"octavia-rsyslog-zbs6z\" (UID: \"4df4068e-249f-47ea-a84a-3c178c27bdf3\") " pod="openstack/octavia-rsyslog-zbs6z" Mar 13 15:44:57 crc kubenswrapper[4907]: I0313 15:44:57.855440 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/4df4068e-249f-47ea-a84a-3c178c27bdf3-config-data-merged\") pod \"octavia-rsyslog-zbs6z\" (UID: \"4df4068e-249f-47ea-a84a-3c178c27bdf3\") " pod="openstack/octavia-rsyslog-zbs6z" Mar 13 15:44:57 crc kubenswrapper[4907]: I0313 15:44:57.856994 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/4df4068e-249f-47ea-a84a-3c178c27bdf3-hm-ports\") pod \"octavia-rsyslog-zbs6z\" (UID: \"4df4068e-249f-47ea-a84a-3c178c27bdf3\") " pod="openstack/octavia-rsyslog-zbs6z" Mar 13 15:44:57 crc kubenswrapper[4907]: I0313 15:44:57.866459 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4df4068e-249f-47ea-a84a-3c178c27bdf3-config-data\") pod \"octavia-rsyslog-zbs6z\" (UID: \"4df4068e-249f-47ea-a84a-3c178c27bdf3\") " pod="openstack/octavia-rsyslog-zbs6z" Mar 13 15:44:57 crc kubenswrapper[4907]: I0313 15:44:57.866922 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4df4068e-249f-47ea-a84a-3c178c27bdf3-scripts\") pod \"octavia-rsyslog-zbs6z\" (UID: \"4df4068e-249f-47ea-a84a-3c178c27bdf3\") " pod="openstack/octavia-rsyslog-zbs6z" Mar 13 15:44:57 crc kubenswrapper[4907]: I0313 15:44:57.973414 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-rsyslog-zbs6z" Mar 13 15:44:58 crc kubenswrapper[4907]: I0313 15:44:58.596518 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-rsyslog-zbs6z"] Mar 13 15:44:58 crc kubenswrapper[4907]: W0313 15:44:58.597073 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4df4068e_249f_47ea_a84a_3c178c27bdf3.slice/crio-0d577fd9519d47b2c397ce97d78c1842b141c20b3852e2b82b1bddfd1f827fc5 WatchSource:0}: Error finding container 0d577fd9519d47b2c397ce97d78c1842b141c20b3852e2b82b1bddfd1f827fc5: Status 404 returned error can't find the container with id 0d577fd9519d47b2c397ce97d78c1842b141c20b3852e2b82b1bddfd1f827fc5 Mar 13 15:44:58 crc kubenswrapper[4907]: I0313 15:44:58.689209 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-image-upload-6f45c4fb85-gvk5g"] Mar 13 15:44:58 crc kubenswrapper[4907]: I0313 15:44:58.691002 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-image-upload-6f45c4fb85-gvk5g" Mar 13 15:44:58 crc kubenswrapper[4907]: I0313 15:44:58.697871 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-config-data" Mar 13 15:44:58 crc kubenswrapper[4907]: I0313 15:44:58.712947 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-image-upload-6f45c4fb85-gvk5g"] Mar 13 15:44:58 crc kubenswrapper[4907]: I0313 15:44:58.723588 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-rsyslog-zbs6z"] Mar 13 15:44:58 crc kubenswrapper[4907]: I0313 15:44:58.775439 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/7a6ac004-9828-4581-9147-88c96b1f5ef4-amphora-image\") pod \"octavia-image-upload-6f45c4fb85-gvk5g\" (UID: \"7a6ac004-9828-4581-9147-88c96b1f5ef4\") " pod="openstack/octavia-image-upload-6f45c4fb85-gvk5g" Mar 13 15:44:58 crc kubenswrapper[4907]: I0313 15:44:58.775742 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/7a6ac004-9828-4581-9147-88c96b1f5ef4-httpd-config\") pod \"octavia-image-upload-6f45c4fb85-gvk5g\" (UID: \"7a6ac004-9828-4581-9147-88c96b1f5ef4\") " pod="openstack/octavia-image-upload-6f45c4fb85-gvk5g" Mar 13 15:44:58 crc kubenswrapper[4907]: I0313 15:44:58.878764 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/7a6ac004-9828-4581-9147-88c96b1f5ef4-amphora-image\") pod \"octavia-image-upload-6f45c4fb85-gvk5g\" (UID: \"7a6ac004-9828-4581-9147-88c96b1f5ef4\") " pod="openstack/octavia-image-upload-6f45c4fb85-gvk5g" Mar 13 15:44:58 crc kubenswrapper[4907]: I0313 15:44:58.879241 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/7a6ac004-9828-4581-9147-88c96b1f5ef4-httpd-config\") pod \"octavia-image-upload-6f45c4fb85-gvk5g\" (UID: \"7a6ac004-9828-4581-9147-88c96b1f5ef4\") " pod="openstack/octavia-image-upload-6f45c4fb85-gvk5g" Mar 13 15:44:58 crc kubenswrapper[4907]: I0313 15:44:58.880525 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/7a6ac004-9828-4581-9147-88c96b1f5ef4-amphora-image\") pod \"octavia-image-upload-6f45c4fb85-gvk5g\" (UID: \"7a6ac004-9828-4581-9147-88c96b1f5ef4\") " pod="openstack/octavia-image-upload-6f45c4fb85-gvk5g" Mar 13 15:44:58 crc kubenswrapper[4907]: I0313 15:44:58.887339 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/7a6ac004-9828-4581-9147-88c96b1f5ef4-httpd-config\") pod \"octavia-image-upload-6f45c4fb85-gvk5g\" (UID: \"7a6ac004-9828-4581-9147-88c96b1f5ef4\") " pod="openstack/octavia-image-upload-6f45c4fb85-gvk5g" Mar 13 15:44:59 crc kubenswrapper[4907]: I0313 15:44:59.037228 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-image-upload-6f45c4fb85-gvk5g" Mar 13 15:44:59 crc kubenswrapper[4907]: I0313 15:44:59.361302 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-rsyslog-zbs6z" event={"ID":"4df4068e-249f-47ea-a84a-3c178c27bdf3","Type":"ContainerStarted","Data":"0d577fd9519d47b2c397ce97d78c1842b141c20b3852e2b82b1bddfd1f827fc5"} Mar 13 15:44:59 crc kubenswrapper[4907]: I0313 15:44:59.568144 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-image-upload-6f45c4fb85-gvk5g"] Mar 13 15:45:00 crc kubenswrapper[4907]: I0313 15:45:00.140196 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556945-vp2lp"] Mar 13 15:45:00 crc kubenswrapper[4907]: I0313 15:45:00.142308 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556945-vp2lp" Mar 13 15:45:00 crc kubenswrapper[4907]: I0313 15:45:00.145186 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 13 15:45:00 crc kubenswrapper[4907]: I0313 15:45:00.145388 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 13 15:45:00 crc kubenswrapper[4907]: I0313 15:45:00.163987 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556945-vp2lp"] Mar 13 15:45:00 crc kubenswrapper[4907]: I0313 15:45:00.203201 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/329cddab-6958-4dbb-a2d0-baef50885413-config-volume\") pod \"collect-profiles-29556945-vp2lp\" (UID: \"329cddab-6958-4dbb-a2d0-baef50885413\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556945-vp2lp" Mar 13 15:45:00 crc kubenswrapper[4907]: I0313 15:45:00.203263 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rbxfk\" (UniqueName: \"kubernetes.io/projected/329cddab-6958-4dbb-a2d0-baef50885413-kube-api-access-rbxfk\") pod \"collect-profiles-29556945-vp2lp\" (UID: \"329cddab-6958-4dbb-a2d0-baef50885413\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556945-vp2lp" Mar 13 15:45:00 crc kubenswrapper[4907]: I0313 15:45:00.203290 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/329cddab-6958-4dbb-a2d0-baef50885413-secret-volume\") pod \"collect-profiles-29556945-vp2lp\" (UID: \"329cddab-6958-4dbb-a2d0-baef50885413\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556945-vp2lp" Mar 13 15:45:00 crc kubenswrapper[4907]: I0313 15:45:00.305219 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/329cddab-6958-4dbb-a2d0-baef50885413-config-volume\") pod \"collect-profiles-29556945-vp2lp\" (UID: \"329cddab-6958-4dbb-a2d0-baef50885413\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556945-vp2lp" Mar 13 15:45:00 crc kubenswrapper[4907]: I0313 15:45:00.305273 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rbxfk\" (UniqueName: \"kubernetes.io/projected/329cddab-6958-4dbb-a2d0-baef50885413-kube-api-access-rbxfk\") pod \"collect-profiles-29556945-vp2lp\" (UID: \"329cddab-6958-4dbb-a2d0-baef50885413\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556945-vp2lp" Mar 13 15:45:00 crc kubenswrapper[4907]: I0313 15:45:00.305296 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/329cddab-6958-4dbb-a2d0-baef50885413-secret-volume\") pod \"collect-profiles-29556945-vp2lp\" (UID: \"329cddab-6958-4dbb-a2d0-baef50885413\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556945-vp2lp" Mar 13 15:45:00 crc kubenswrapper[4907]: I0313 15:45:00.306972 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/329cddab-6958-4dbb-a2d0-baef50885413-config-volume\") pod \"collect-profiles-29556945-vp2lp\" (UID: \"329cddab-6958-4dbb-a2d0-baef50885413\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556945-vp2lp" Mar 13 15:45:00 crc kubenswrapper[4907]: I0313 15:45:00.314500 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/329cddab-6958-4dbb-a2d0-baef50885413-secret-volume\") pod \"collect-profiles-29556945-vp2lp\" (UID: \"329cddab-6958-4dbb-a2d0-baef50885413\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556945-vp2lp" Mar 13 15:45:00 crc kubenswrapper[4907]: I0313 15:45:00.325819 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rbxfk\" (UniqueName: \"kubernetes.io/projected/329cddab-6958-4dbb-a2d0-baef50885413-kube-api-access-rbxfk\") pod \"collect-profiles-29556945-vp2lp\" (UID: \"329cddab-6958-4dbb-a2d0-baef50885413\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556945-vp2lp" Mar 13 15:45:00 crc kubenswrapper[4907]: I0313 15:45:00.391712 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-6f45c4fb85-gvk5g" event={"ID":"7a6ac004-9828-4581-9147-88c96b1f5ef4","Type":"ContainerStarted","Data":"779798f2bbcef27c1f90b8ec15e35048f6beabc9118118529e9bc26eeed24e32"} Mar 13 15:45:00 crc kubenswrapper[4907]: I0313 15:45:00.466977 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556945-vp2lp" Mar 13 15:45:01 crc kubenswrapper[4907]: I0313 15:45:01.022110 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556945-vp2lp"] Mar 13 15:45:01 crc kubenswrapper[4907]: I0313 15:45:01.412685 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-rsyslog-zbs6z" event={"ID":"4df4068e-249f-47ea-a84a-3c178c27bdf3","Type":"ContainerStarted","Data":"c19a72a707aaaf1cb3cc867c2a276800bcfd60b1ace8c673b72611d9326d6913"} Mar 13 15:45:02 crc kubenswrapper[4907]: I0313 15:45:02.424999 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556945-vp2lp" event={"ID":"329cddab-6958-4dbb-a2d0-baef50885413","Type":"ContainerStarted","Data":"114a7e04a442a61d9b64fe621624aaf4b09fa81b5a1bf440bf909e958cdc0ffe"} Mar 13 15:45:03 crc kubenswrapper[4907]: I0313 15:45:03.439673 4907 generic.go:334] "Generic (PLEG): container finished" podID="4df4068e-249f-47ea-a84a-3c178c27bdf3" containerID="c19a72a707aaaf1cb3cc867c2a276800bcfd60b1ace8c673b72611d9326d6913" exitCode=0 Mar 13 15:45:03 crc kubenswrapper[4907]: I0313 15:45:03.439859 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-rsyslog-zbs6z" event={"ID":"4df4068e-249f-47ea-a84a-3c178c27bdf3","Type":"ContainerDied","Data":"c19a72a707aaaf1cb3cc867c2a276800bcfd60b1ace8c673b72611d9326d6913"} Mar 13 15:45:03 crc kubenswrapper[4907]: I0313 15:45:03.449431 4907 generic.go:334] "Generic (PLEG): container finished" podID="329cddab-6958-4dbb-a2d0-baef50885413" containerID="65a8dfeb436104f422bc6ebeae3f1c135d2d01201103445d064625f2317603ad" exitCode=0 Mar 13 15:45:03 crc kubenswrapper[4907]: I0313 15:45:03.449506 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556945-vp2lp" event={"ID":"329cddab-6958-4dbb-a2d0-baef50885413","Type":"ContainerDied","Data":"65a8dfeb436104f422bc6ebeae3f1c135d2d01201103445d064625f2317603ad"} Mar 13 15:45:05 crc kubenswrapper[4907]: I0313 15:45:05.409452 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556945-vp2lp" Mar 13 15:45:05 crc kubenswrapper[4907]: I0313 15:45:05.491051 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556945-vp2lp" Mar 13 15:45:05 crc kubenswrapper[4907]: I0313 15:45:05.491527 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556945-vp2lp" event={"ID":"329cddab-6958-4dbb-a2d0-baef50885413","Type":"ContainerDied","Data":"114a7e04a442a61d9b64fe621624aaf4b09fa81b5a1bf440bf909e958cdc0ffe"} Mar 13 15:45:05 crc kubenswrapper[4907]: I0313 15:45:05.491552 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="114a7e04a442a61d9b64fe621624aaf4b09fa81b5a1bf440bf909e958cdc0ffe" Mar 13 15:45:05 crc kubenswrapper[4907]: I0313 15:45:05.550658 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/329cddab-6958-4dbb-a2d0-baef50885413-config-volume\") pod \"329cddab-6958-4dbb-a2d0-baef50885413\" (UID: \"329cddab-6958-4dbb-a2d0-baef50885413\") " Mar 13 15:45:05 crc kubenswrapper[4907]: I0313 15:45:05.550782 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rbxfk\" (UniqueName: \"kubernetes.io/projected/329cddab-6958-4dbb-a2d0-baef50885413-kube-api-access-rbxfk\") pod \"329cddab-6958-4dbb-a2d0-baef50885413\" (UID: \"329cddab-6958-4dbb-a2d0-baef50885413\") " Mar 13 15:45:05 crc kubenswrapper[4907]: I0313 15:45:05.550808 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/329cddab-6958-4dbb-a2d0-baef50885413-secret-volume\") pod \"329cddab-6958-4dbb-a2d0-baef50885413\" (UID: \"329cddab-6958-4dbb-a2d0-baef50885413\") " Mar 13 15:45:05 crc kubenswrapper[4907]: I0313 15:45:05.553620 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/329cddab-6958-4dbb-a2d0-baef50885413-config-volume" (OuterVolumeSpecName: "config-volume") pod "329cddab-6958-4dbb-a2d0-baef50885413" (UID: "329cddab-6958-4dbb-a2d0-baef50885413"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:45:05 crc kubenswrapper[4907]: I0313 15:45:05.562174 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/329cddab-6958-4dbb-a2d0-baef50885413-kube-api-access-rbxfk" (OuterVolumeSpecName: "kube-api-access-rbxfk") pod "329cddab-6958-4dbb-a2d0-baef50885413" (UID: "329cddab-6958-4dbb-a2d0-baef50885413"). InnerVolumeSpecName "kube-api-access-rbxfk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:45:05 crc kubenswrapper[4907]: I0313 15:45:05.572087 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/329cddab-6958-4dbb-a2d0-baef50885413-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "329cddab-6958-4dbb-a2d0-baef50885413" (UID: "329cddab-6958-4dbb-a2d0-baef50885413"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:45:05 crc kubenswrapper[4907]: I0313 15:45:05.652709 4907 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/329cddab-6958-4dbb-a2d0-baef50885413-config-volume\") on node \"crc\" DevicePath \"\"" Mar 13 15:45:05 crc kubenswrapper[4907]: I0313 15:45:05.652742 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rbxfk\" (UniqueName: \"kubernetes.io/projected/329cddab-6958-4dbb-a2d0-baef50885413-kube-api-access-rbxfk\") on node \"crc\" DevicePath \"\"" Mar 13 15:45:05 crc kubenswrapper[4907]: I0313 15:45:05.652753 4907 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/329cddab-6958-4dbb-a2d0-baef50885413-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 13 15:45:06 crc kubenswrapper[4907]: I0313 15:45:06.181148 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-db-sync-rjfnd"] Mar 13 15:45:06 crc kubenswrapper[4907]: E0313 15:45:06.182052 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="329cddab-6958-4dbb-a2d0-baef50885413" containerName="collect-profiles" Mar 13 15:45:06 crc kubenswrapper[4907]: I0313 15:45:06.182078 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="329cddab-6958-4dbb-a2d0-baef50885413" containerName="collect-profiles" Mar 13 15:45:06 crc kubenswrapper[4907]: I0313 15:45:06.182344 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="329cddab-6958-4dbb-a2d0-baef50885413" containerName="collect-profiles" Mar 13 15:45:06 crc kubenswrapper[4907]: I0313 15:45:06.183736 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-sync-rjfnd" Mar 13 15:45:06 crc kubenswrapper[4907]: I0313 15:45:06.188168 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-scripts" Mar 13 15:45:06 crc kubenswrapper[4907]: I0313 15:45:06.194828 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-db-sync-rjfnd"] Mar 13 15:45:06 crc kubenswrapper[4907]: I0313 15:45:06.264740 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/aefbf41f-fa5a-4763-af85-3c9e8df33b6a-config-data-merged\") pod \"octavia-db-sync-rjfnd\" (UID: \"aefbf41f-fa5a-4763-af85-3c9e8df33b6a\") " pod="openstack/octavia-db-sync-rjfnd" Mar 13 15:45:06 crc kubenswrapper[4907]: I0313 15:45:06.265072 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aefbf41f-fa5a-4763-af85-3c9e8df33b6a-combined-ca-bundle\") pod \"octavia-db-sync-rjfnd\" (UID: \"aefbf41f-fa5a-4763-af85-3c9e8df33b6a\") " pod="openstack/octavia-db-sync-rjfnd" Mar 13 15:45:06 crc kubenswrapper[4907]: I0313 15:45:06.265171 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aefbf41f-fa5a-4763-af85-3c9e8df33b6a-config-data\") pod \"octavia-db-sync-rjfnd\" (UID: \"aefbf41f-fa5a-4763-af85-3c9e8df33b6a\") " pod="openstack/octavia-db-sync-rjfnd" Mar 13 15:45:06 crc kubenswrapper[4907]: I0313 15:45:06.265342 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aefbf41f-fa5a-4763-af85-3c9e8df33b6a-scripts\") pod \"octavia-db-sync-rjfnd\" (UID: \"aefbf41f-fa5a-4763-af85-3c9e8df33b6a\") " pod="openstack/octavia-db-sync-rjfnd" Mar 13 15:45:06 crc kubenswrapper[4907]: I0313 15:45:06.366342 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aefbf41f-fa5a-4763-af85-3c9e8df33b6a-scripts\") pod \"octavia-db-sync-rjfnd\" (UID: \"aefbf41f-fa5a-4763-af85-3c9e8df33b6a\") " pod="openstack/octavia-db-sync-rjfnd" Mar 13 15:45:06 crc kubenswrapper[4907]: I0313 15:45:06.366392 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/aefbf41f-fa5a-4763-af85-3c9e8df33b6a-config-data-merged\") pod \"octavia-db-sync-rjfnd\" (UID: \"aefbf41f-fa5a-4763-af85-3c9e8df33b6a\") " pod="openstack/octavia-db-sync-rjfnd" Mar 13 15:45:06 crc kubenswrapper[4907]: I0313 15:45:06.366486 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aefbf41f-fa5a-4763-af85-3c9e8df33b6a-combined-ca-bundle\") pod \"octavia-db-sync-rjfnd\" (UID: \"aefbf41f-fa5a-4763-af85-3c9e8df33b6a\") " pod="openstack/octavia-db-sync-rjfnd" Mar 13 15:45:06 crc kubenswrapper[4907]: I0313 15:45:06.366525 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aefbf41f-fa5a-4763-af85-3c9e8df33b6a-config-data\") pod \"octavia-db-sync-rjfnd\" (UID: \"aefbf41f-fa5a-4763-af85-3c9e8df33b6a\") " pod="openstack/octavia-db-sync-rjfnd" Mar 13 15:45:06 crc kubenswrapper[4907]: I0313 15:45:06.367115 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/aefbf41f-fa5a-4763-af85-3c9e8df33b6a-config-data-merged\") pod \"octavia-db-sync-rjfnd\" (UID: \"aefbf41f-fa5a-4763-af85-3c9e8df33b6a\") " pod="openstack/octavia-db-sync-rjfnd" Mar 13 15:45:06 crc kubenswrapper[4907]: I0313 15:45:06.374191 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aefbf41f-fa5a-4763-af85-3c9e8df33b6a-combined-ca-bundle\") pod \"octavia-db-sync-rjfnd\" (UID: \"aefbf41f-fa5a-4763-af85-3c9e8df33b6a\") " pod="openstack/octavia-db-sync-rjfnd" Mar 13 15:45:06 crc kubenswrapper[4907]: I0313 15:45:06.376608 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aefbf41f-fa5a-4763-af85-3c9e8df33b6a-scripts\") pod \"octavia-db-sync-rjfnd\" (UID: \"aefbf41f-fa5a-4763-af85-3c9e8df33b6a\") " pod="openstack/octavia-db-sync-rjfnd" Mar 13 15:45:06 crc kubenswrapper[4907]: I0313 15:45:06.380739 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aefbf41f-fa5a-4763-af85-3c9e8df33b6a-config-data\") pod \"octavia-db-sync-rjfnd\" (UID: \"aefbf41f-fa5a-4763-af85-3c9e8df33b6a\") " pod="openstack/octavia-db-sync-rjfnd" Mar 13 15:45:06 crc kubenswrapper[4907]: I0313 15:45:06.505741 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556900-cxbzm"] Mar 13 15:45:06 crc kubenswrapper[4907]: I0313 15:45:06.512579 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-rsyslog-zbs6z" event={"ID":"4df4068e-249f-47ea-a84a-3c178c27bdf3","Type":"ContainerStarted","Data":"74a44fccf52afbb9c6ff10cceccc82caa3010599c5105999945e23c4bc3970d5"} Mar 13 15:45:06 crc kubenswrapper[4907]: I0313 15:45:06.512837 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-rsyslog-zbs6z" Mar 13 15:45:06 crc kubenswrapper[4907]: I0313 15:45:06.514805 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556900-cxbzm"] Mar 13 15:45:06 crc kubenswrapper[4907]: I0313 15:45:06.532366 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-sync-rjfnd" Mar 13 15:45:06 crc kubenswrapper[4907]: I0313 15:45:06.536398 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-rsyslog-zbs6z" podStartSLOduration=2.909128543 podStartE2EDuration="9.53637243s" podCreationTimestamp="2026-03-13 15:44:57 +0000 UTC" firstStartedPulling="2026-03-13 15:44:58.599486781 +0000 UTC m=+5997.499274470" lastFinishedPulling="2026-03-13 15:45:05.226730668 +0000 UTC m=+6004.126518357" observedRunningTime="2026-03-13 15:45:06.530014457 +0000 UTC m=+6005.429802156" watchObservedRunningTime="2026-03-13 15:45:06.53637243 +0000 UTC m=+6005.436160129" Mar 13 15:45:07 crc kubenswrapper[4907]: I0313 15:45:07.874250 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="48ae7008-a336-47d8-a3da-fd198db91cba" path="/var/lib/kubelet/pods/48ae7008-a336-47d8-a3da-fd198db91cba/volumes" Mar 13 15:45:08 crc kubenswrapper[4907]: I0313 15:45:08.783203 4907 scope.go:117] "RemoveContainer" containerID="68ffe92dd03078633ec89d7ef29da9bbdc4c5f04c956a1e8487e3f2e01c5d1c5" Mar 13 15:45:08 crc kubenswrapper[4907]: E0313 15:45:08.783767 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:45:09 crc kubenswrapper[4907]: I0313 15:45:09.448748 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-api-6db876fcbd-4wtkj" Mar 13 15:45:09 crc kubenswrapper[4907]: I0313 15:45:09.952180 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-api-6db876fcbd-4wtkj" Mar 13 15:45:10 crc kubenswrapper[4907]: I0313 15:45:10.520407 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-db-sync-rjfnd"] Mar 13 15:45:10 crc kubenswrapper[4907]: I0313 15:45:10.579693 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-6f45c4fb85-gvk5g" event={"ID":"7a6ac004-9828-4581-9147-88c96b1f5ef4","Type":"ContainerStarted","Data":"74f45f2ca03eeba83bae0697e69c57f95344fad1a6ad11c8a7d301ea69e6bdbc"} Mar 13 15:45:10 crc kubenswrapper[4907]: I0313 15:45:10.582699 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-sync-rjfnd" event={"ID":"aefbf41f-fa5a-4763-af85-3c9e8df33b6a","Type":"ContainerStarted","Data":"c0ee84ab0839ac1ec5f8ef593950c63bef20d37354985ce33bc2c5c08ed061a9"} Mar 13 15:45:11 crc kubenswrapper[4907]: I0313 15:45:11.591546 4907 generic.go:334] "Generic (PLEG): container finished" podID="7a6ac004-9828-4581-9147-88c96b1f5ef4" containerID="74f45f2ca03eeba83bae0697e69c57f95344fad1a6ad11c8a7d301ea69e6bdbc" exitCode=0 Mar 13 15:45:11 crc kubenswrapper[4907]: I0313 15:45:11.591723 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-6f45c4fb85-gvk5g" event={"ID":"7a6ac004-9828-4581-9147-88c96b1f5ef4","Type":"ContainerDied","Data":"74f45f2ca03eeba83bae0697e69c57f95344fad1a6ad11c8a7d301ea69e6bdbc"} Mar 13 15:45:11 crc kubenswrapper[4907]: I0313 15:45:11.594721 4907 generic.go:334] "Generic (PLEG): container finished" podID="aefbf41f-fa5a-4763-af85-3c9e8df33b6a" containerID="94c59567d9e99f3b6037554dff6847cedf9844e341b49a4a37077ea52d3428be" exitCode=0 Mar 13 15:45:11 crc kubenswrapper[4907]: I0313 15:45:11.594772 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-sync-rjfnd" event={"ID":"aefbf41f-fa5a-4763-af85-3c9e8df33b6a","Type":"ContainerDied","Data":"94c59567d9e99f3b6037554dff6847cedf9844e341b49a4a37077ea52d3428be"} Mar 13 15:45:12 crc kubenswrapper[4907]: I0313 15:45:12.604519 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-6f45c4fb85-gvk5g" event={"ID":"7a6ac004-9828-4581-9147-88c96b1f5ef4","Type":"ContainerStarted","Data":"8c362fd432faeca33e638f398523ea974dc977713306f90c2ab5a670b92e3cb1"} Mar 13 15:45:12 crc kubenswrapper[4907]: I0313 15:45:12.607225 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-sync-rjfnd" event={"ID":"aefbf41f-fa5a-4763-af85-3c9e8df33b6a","Type":"ContainerStarted","Data":"4586239eb0dcd4e94af6c8d52a6990fcfc7c67962591f11127b749cbaee23663"} Mar 13 15:45:12 crc kubenswrapper[4907]: I0313 15:45:12.627595 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-image-upload-6f45c4fb85-gvk5g" podStartSLOduration=4.037097773 podStartE2EDuration="14.627569314s" podCreationTimestamp="2026-03-13 15:44:58 +0000 UTC" firstStartedPulling="2026-03-13 15:44:59.579802235 +0000 UTC m=+5998.479589914" lastFinishedPulling="2026-03-13 15:45:10.170273766 +0000 UTC m=+6009.070061455" observedRunningTime="2026-03-13 15:45:12.616783659 +0000 UTC m=+6011.516571348" watchObservedRunningTime="2026-03-13 15:45:12.627569314 +0000 UTC m=+6011.527356993" Mar 13 15:45:12 crc kubenswrapper[4907]: I0313 15:45:12.633731 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-db-sync-rjfnd" podStartSLOduration=6.633712071 podStartE2EDuration="6.633712071s" podCreationTimestamp="2026-03-13 15:45:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:45:12.632037805 +0000 UTC m=+6011.531825494" watchObservedRunningTime="2026-03-13 15:45:12.633712071 +0000 UTC m=+6011.533499760" Mar 13 15:45:13 crc kubenswrapper[4907]: I0313 15:45:13.001398 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-rsyslog-zbs6z" Mar 13 15:45:14 crc kubenswrapper[4907]: I0313 15:45:14.626610 4907 generic.go:334] "Generic (PLEG): container finished" podID="aefbf41f-fa5a-4763-af85-3c9e8df33b6a" containerID="4586239eb0dcd4e94af6c8d52a6990fcfc7c67962591f11127b749cbaee23663" exitCode=0 Mar 13 15:45:14 crc kubenswrapper[4907]: I0313 15:45:14.626693 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-sync-rjfnd" event={"ID":"aefbf41f-fa5a-4763-af85-3c9e8df33b6a","Type":"ContainerDied","Data":"4586239eb0dcd4e94af6c8d52a6990fcfc7c67962591f11127b749cbaee23663"} Mar 13 15:45:16 crc kubenswrapper[4907]: I0313 15:45:16.010382 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-sync-rjfnd" Mar 13 15:45:16 crc kubenswrapper[4907]: I0313 15:45:16.187342 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aefbf41f-fa5a-4763-af85-3c9e8df33b6a-scripts\") pod \"aefbf41f-fa5a-4763-af85-3c9e8df33b6a\" (UID: \"aefbf41f-fa5a-4763-af85-3c9e8df33b6a\") " Mar 13 15:45:16 crc kubenswrapper[4907]: I0313 15:45:16.188179 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aefbf41f-fa5a-4763-af85-3c9e8df33b6a-config-data\") pod \"aefbf41f-fa5a-4763-af85-3c9e8df33b6a\" (UID: \"aefbf41f-fa5a-4763-af85-3c9e8df33b6a\") " Mar 13 15:45:16 crc kubenswrapper[4907]: I0313 15:45:16.188308 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aefbf41f-fa5a-4763-af85-3c9e8df33b6a-combined-ca-bundle\") pod \"aefbf41f-fa5a-4763-af85-3c9e8df33b6a\" (UID: \"aefbf41f-fa5a-4763-af85-3c9e8df33b6a\") " Mar 13 15:45:16 crc kubenswrapper[4907]: I0313 15:45:16.188500 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/aefbf41f-fa5a-4763-af85-3c9e8df33b6a-config-data-merged\") pod \"aefbf41f-fa5a-4763-af85-3c9e8df33b6a\" (UID: \"aefbf41f-fa5a-4763-af85-3c9e8df33b6a\") " Mar 13 15:45:16 crc kubenswrapper[4907]: I0313 15:45:16.192279 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aefbf41f-fa5a-4763-af85-3c9e8df33b6a-scripts" (OuterVolumeSpecName: "scripts") pod "aefbf41f-fa5a-4763-af85-3c9e8df33b6a" (UID: "aefbf41f-fa5a-4763-af85-3c9e8df33b6a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:45:16 crc kubenswrapper[4907]: I0313 15:45:16.193561 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aefbf41f-fa5a-4763-af85-3c9e8df33b6a-config-data" (OuterVolumeSpecName: "config-data") pod "aefbf41f-fa5a-4763-af85-3c9e8df33b6a" (UID: "aefbf41f-fa5a-4763-af85-3c9e8df33b6a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:45:16 crc kubenswrapper[4907]: I0313 15:45:16.218211 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aefbf41f-fa5a-4763-af85-3c9e8df33b6a-config-data-merged" (OuterVolumeSpecName: "config-data-merged") pod "aefbf41f-fa5a-4763-af85-3c9e8df33b6a" (UID: "aefbf41f-fa5a-4763-af85-3c9e8df33b6a"). InnerVolumeSpecName "config-data-merged". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:45:16 crc kubenswrapper[4907]: I0313 15:45:16.220382 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aefbf41f-fa5a-4763-af85-3c9e8df33b6a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "aefbf41f-fa5a-4763-af85-3c9e8df33b6a" (UID: "aefbf41f-fa5a-4763-af85-3c9e8df33b6a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:45:16 crc kubenswrapper[4907]: I0313 15:45:16.291017 4907 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aefbf41f-fa5a-4763-af85-3c9e8df33b6a-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 15:45:16 crc kubenswrapper[4907]: I0313 15:45:16.291051 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aefbf41f-fa5a-4763-af85-3c9e8df33b6a-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 15:45:16 crc kubenswrapper[4907]: I0313 15:45:16.291065 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aefbf41f-fa5a-4763-af85-3c9e8df33b6a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 15:45:16 crc kubenswrapper[4907]: I0313 15:45:16.291080 4907 reconciler_common.go:293] "Volume detached for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/aefbf41f-fa5a-4763-af85-3c9e8df33b6a-config-data-merged\") on node \"crc\" DevicePath \"\"" Mar 13 15:45:16 crc kubenswrapper[4907]: I0313 15:45:16.651298 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-sync-rjfnd" event={"ID":"aefbf41f-fa5a-4763-af85-3c9e8df33b6a","Type":"ContainerDied","Data":"c0ee84ab0839ac1ec5f8ef593950c63bef20d37354985ce33bc2c5c08ed061a9"} Mar 13 15:45:16 crc kubenswrapper[4907]: I0313 15:45:16.651343 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c0ee84ab0839ac1ec5f8ef593950c63bef20d37354985ce33bc2c5c08ed061a9" Mar 13 15:45:16 crc kubenswrapper[4907]: I0313 15:45:16.651385 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-sync-rjfnd" Mar 13 15:45:20 crc kubenswrapper[4907]: I0313 15:45:20.782436 4907 scope.go:117] "RemoveContainer" containerID="68ffe92dd03078633ec89d7ef29da9bbdc4c5f04c956a1e8487e3f2e01c5d1c5" Mar 13 15:45:20 crc kubenswrapper[4907]: E0313 15:45:20.784343 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:45:29 crc kubenswrapper[4907]: I0313 15:45:29.666216 4907 scope.go:117] "RemoveContainer" containerID="189d9f0966abd854ff0ae3d577047f2f411f08ff9ec8a2c05b10583fd1185431" Mar 13 15:45:33 crc kubenswrapper[4907]: I0313 15:45:33.784452 4907 scope.go:117] "RemoveContainer" containerID="68ffe92dd03078633ec89d7ef29da9bbdc4c5f04c956a1e8487e3f2e01c5d1c5" Mar 13 15:45:33 crc kubenswrapper[4907]: E0313 15:45:33.785257 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:45:38 crc kubenswrapper[4907]: I0313 15:45:38.806060 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-image-upload-6f45c4fb85-gvk5g"] Mar 13 15:45:38 crc kubenswrapper[4907]: I0313 15:45:38.806798 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/octavia-image-upload-6f45c4fb85-gvk5g" podUID="7a6ac004-9828-4581-9147-88c96b1f5ef4" containerName="octavia-amphora-httpd" containerID="cri-o://8c362fd432faeca33e638f398523ea974dc977713306f90c2ab5a670b92e3cb1" gracePeriod=30 Mar 13 15:45:39 crc kubenswrapper[4907]: I0313 15:45:39.334739 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-image-upload-6f45c4fb85-gvk5g" Mar 13 15:45:39 crc kubenswrapper[4907]: I0313 15:45:39.453385 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/7a6ac004-9828-4581-9147-88c96b1f5ef4-amphora-image\") pod \"7a6ac004-9828-4581-9147-88c96b1f5ef4\" (UID: \"7a6ac004-9828-4581-9147-88c96b1f5ef4\") " Mar 13 15:45:39 crc kubenswrapper[4907]: I0313 15:45:39.453923 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/7a6ac004-9828-4581-9147-88c96b1f5ef4-httpd-config\") pod \"7a6ac004-9828-4581-9147-88c96b1f5ef4\" (UID: \"7a6ac004-9828-4581-9147-88c96b1f5ef4\") " Mar 13 15:45:39 crc kubenswrapper[4907]: I0313 15:45:39.496830 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a6ac004-9828-4581-9147-88c96b1f5ef4-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "7a6ac004-9828-4581-9147-88c96b1f5ef4" (UID: "7a6ac004-9828-4581-9147-88c96b1f5ef4"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:45:39 crc kubenswrapper[4907]: I0313 15:45:39.509849 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7a6ac004-9828-4581-9147-88c96b1f5ef4-amphora-image" (OuterVolumeSpecName: "amphora-image") pod "7a6ac004-9828-4581-9147-88c96b1f5ef4" (UID: "7a6ac004-9828-4581-9147-88c96b1f5ef4"). InnerVolumeSpecName "amphora-image". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:45:39 crc kubenswrapper[4907]: I0313 15:45:39.556469 4907 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/7a6ac004-9828-4581-9147-88c96b1f5ef4-httpd-config\") on node \"crc\" DevicePath \"\"" Mar 13 15:45:39 crc kubenswrapper[4907]: I0313 15:45:39.556516 4907 reconciler_common.go:293] "Volume detached for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/7a6ac004-9828-4581-9147-88c96b1f5ef4-amphora-image\") on node \"crc\" DevicePath \"\"" Mar 13 15:45:39 crc kubenswrapper[4907]: I0313 15:45:39.847179 4907 generic.go:334] "Generic (PLEG): container finished" podID="7a6ac004-9828-4581-9147-88c96b1f5ef4" containerID="8c362fd432faeca33e638f398523ea974dc977713306f90c2ab5a670b92e3cb1" exitCode=0 Mar 13 15:45:39 crc kubenswrapper[4907]: I0313 15:45:39.847228 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-6f45c4fb85-gvk5g" event={"ID":"7a6ac004-9828-4581-9147-88c96b1f5ef4","Type":"ContainerDied","Data":"8c362fd432faeca33e638f398523ea974dc977713306f90c2ab5a670b92e3cb1"} Mar 13 15:45:39 crc kubenswrapper[4907]: I0313 15:45:39.847259 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-6f45c4fb85-gvk5g" event={"ID":"7a6ac004-9828-4581-9147-88c96b1f5ef4","Type":"ContainerDied","Data":"779798f2bbcef27c1f90b8ec15e35048f6beabc9118118529e9bc26eeed24e32"} Mar 13 15:45:39 crc kubenswrapper[4907]: I0313 15:45:39.847280 4907 scope.go:117] "RemoveContainer" containerID="8c362fd432faeca33e638f398523ea974dc977713306f90c2ab5a670b92e3cb1" Mar 13 15:45:39 crc kubenswrapper[4907]: I0313 15:45:39.847424 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-image-upload-6f45c4fb85-gvk5g" Mar 13 15:45:39 crc kubenswrapper[4907]: I0313 15:45:39.871543 4907 scope.go:117] "RemoveContainer" containerID="74f45f2ca03eeba83bae0697e69c57f95344fad1a6ad11c8a7d301ea69e6bdbc" Mar 13 15:45:39 crc kubenswrapper[4907]: I0313 15:45:39.876089 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-image-upload-6f45c4fb85-gvk5g"] Mar 13 15:45:39 crc kubenswrapper[4907]: I0313 15:45:39.889529 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/octavia-image-upload-6f45c4fb85-gvk5g"] Mar 13 15:45:39 crc kubenswrapper[4907]: I0313 15:45:39.893389 4907 scope.go:117] "RemoveContainer" containerID="8c362fd432faeca33e638f398523ea974dc977713306f90c2ab5a670b92e3cb1" Mar 13 15:45:39 crc kubenswrapper[4907]: E0313 15:45:39.895341 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8c362fd432faeca33e638f398523ea974dc977713306f90c2ab5a670b92e3cb1\": container with ID starting with 8c362fd432faeca33e638f398523ea974dc977713306f90c2ab5a670b92e3cb1 not found: ID does not exist" containerID="8c362fd432faeca33e638f398523ea974dc977713306f90c2ab5a670b92e3cb1" Mar 13 15:45:39 crc kubenswrapper[4907]: I0313 15:45:39.895369 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8c362fd432faeca33e638f398523ea974dc977713306f90c2ab5a670b92e3cb1"} err="failed to get container status \"8c362fd432faeca33e638f398523ea974dc977713306f90c2ab5a670b92e3cb1\": rpc error: code = NotFound desc = could not find container \"8c362fd432faeca33e638f398523ea974dc977713306f90c2ab5a670b92e3cb1\": container with ID starting with 8c362fd432faeca33e638f398523ea974dc977713306f90c2ab5a670b92e3cb1 not found: ID does not exist" Mar 13 15:45:39 crc kubenswrapper[4907]: I0313 15:45:39.895388 4907 scope.go:117] "RemoveContainer" containerID="74f45f2ca03eeba83bae0697e69c57f95344fad1a6ad11c8a7d301ea69e6bdbc" Mar 13 15:45:39 crc kubenswrapper[4907]: E0313 15:45:39.895697 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"74f45f2ca03eeba83bae0697e69c57f95344fad1a6ad11c8a7d301ea69e6bdbc\": container with ID starting with 74f45f2ca03eeba83bae0697e69c57f95344fad1a6ad11c8a7d301ea69e6bdbc not found: ID does not exist" containerID="74f45f2ca03eeba83bae0697e69c57f95344fad1a6ad11c8a7d301ea69e6bdbc" Mar 13 15:45:39 crc kubenswrapper[4907]: I0313 15:45:39.895737 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"74f45f2ca03eeba83bae0697e69c57f95344fad1a6ad11c8a7d301ea69e6bdbc"} err="failed to get container status \"74f45f2ca03eeba83bae0697e69c57f95344fad1a6ad11c8a7d301ea69e6bdbc\": rpc error: code = NotFound desc = could not find container \"74f45f2ca03eeba83bae0697e69c57f95344fad1a6ad11c8a7d301ea69e6bdbc\": container with ID starting with 74f45f2ca03eeba83bae0697e69c57f95344fad1a6ad11c8a7d301ea69e6bdbc not found: ID does not exist" Mar 13 15:45:41 crc kubenswrapper[4907]: I0313 15:45:41.792403 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7a6ac004-9828-4581-9147-88c96b1f5ef4" path="/var/lib/kubelet/pods/7a6ac004-9828-4581-9147-88c96b1f5ef4/volumes" Mar 13 15:45:46 crc kubenswrapper[4907]: I0313 15:45:46.782971 4907 scope.go:117] "RemoveContainer" containerID="68ffe92dd03078633ec89d7ef29da9bbdc4c5f04c956a1e8487e3f2e01c5d1c5" Mar 13 15:45:46 crc kubenswrapper[4907]: E0313 15:45:46.783695 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:45:58 crc kubenswrapper[4907]: I0313 15:45:58.782677 4907 scope.go:117] "RemoveContainer" containerID="68ffe92dd03078633ec89d7ef29da9bbdc4c5f04c956a1e8487e3f2e01c5d1c5" Mar 13 15:45:58 crc kubenswrapper[4907]: E0313 15:45:58.783404 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:46:00 crc kubenswrapper[4907]: I0313 15:46:00.143809 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556946-78rqg"] Mar 13 15:46:00 crc kubenswrapper[4907]: E0313 15:46:00.144537 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a6ac004-9828-4581-9147-88c96b1f5ef4" containerName="init" Mar 13 15:46:00 crc kubenswrapper[4907]: I0313 15:46:00.144550 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a6ac004-9828-4581-9147-88c96b1f5ef4" containerName="init" Mar 13 15:46:00 crc kubenswrapper[4907]: E0313 15:46:00.144557 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aefbf41f-fa5a-4763-af85-3c9e8df33b6a" containerName="octavia-db-sync" Mar 13 15:46:00 crc kubenswrapper[4907]: I0313 15:46:00.144563 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="aefbf41f-fa5a-4763-af85-3c9e8df33b6a" containerName="octavia-db-sync" Mar 13 15:46:00 crc kubenswrapper[4907]: E0313 15:46:00.144577 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aefbf41f-fa5a-4763-af85-3c9e8df33b6a" containerName="init" Mar 13 15:46:00 crc kubenswrapper[4907]: I0313 15:46:00.144583 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="aefbf41f-fa5a-4763-af85-3c9e8df33b6a" containerName="init" Mar 13 15:46:00 crc kubenswrapper[4907]: E0313 15:46:00.144608 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a6ac004-9828-4581-9147-88c96b1f5ef4" containerName="octavia-amphora-httpd" Mar 13 15:46:00 crc kubenswrapper[4907]: I0313 15:46:00.144615 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a6ac004-9828-4581-9147-88c96b1f5ef4" containerName="octavia-amphora-httpd" Mar 13 15:46:00 crc kubenswrapper[4907]: I0313 15:46:00.144827 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="aefbf41f-fa5a-4763-af85-3c9e8df33b6a" containerName="octavia-db-sync" Mar 13 15:46:00 crc kubenswrapper[4907]: I0313 15:46:00.144852 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a6ac004-9828-4581-9147-88c96b1f5ef4" containerName="octavia-amphora-httpd" Mar 13 15:46:00 crc kubenswrapper[4907]: I0313 15:46:00.145485 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556946-78rqg" Mar 13 15:46:00 crc kubenswrapper[4907]: I0313 15:46:00.148178 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 15:46:00 crc kubenswrapper[4907]: I0313 15:46:00.148317 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 15:46:00 crc kubenswrapper[4907]: I0313 15:46:00.151221 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 15:46:00 crc kubenswrapper[4907]: I0313 15:46:00.159468 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556946-78rqg"] Mar 13 15:46:00 crc kubenswrapper[4907]: I0313 15:46:00.283026 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lqghz\" (UniqueName: \"kubernetes.io/projected/2f0d1531-1d25-4331-a690-1fcbcb504071-kube-api-access-lqghz\") pod \"auto-csr-approver-29556946-78rqg\" (UID: \"2f0d1531-1d25-4331-a690-1fcbcb504071\") " pod="openshift-infra/auto-csr-approver-29556946-78rqg" Mar 13 15:46:00 crc kubenswrapper[4907]: I0313 15:46:00.384648 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lqghz\" (UniqueName: \"kubernetes.io/projected/2f0d1531-1d25-4331-a690-1fcbcb504071-kube-api-access-lqghz\") pod \"auto-csr-approver-29556946-78rqg\" (UID: \"2f0d1531-1d25-4331-a690-1fcbcb504071\") " pod="openshift-infra/auto-csr-approver-29556946-78rqg" Mar 13 15:46:00 crc kubenswrapper[4907]: I0313 15:46:00.412720 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lqghz\" (UniqueName: \"kubernetes.io/projected/2f0d1531-1d25-4331-a690-1fcbcb504071-kube-api-access-lqghz\") pod \"auto-csr-approver-29556946-78rqg\" (UID: \"2f0d1531-1d25-4331-a690-1fcbcb504071\") " pod="openshift-infra/auto-csr-approver-29556946-78rqg" Mar 13 15:46:00 crc kubenswrapper[4907]: I0313 15:46:00.464514 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556946-78rqg" Mar 13 15:46:00 crc kubenswrapper[4907]: I0313 15:46:00.923545 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556946-78rqg"] Mar 13 15:46:01 crc kubenswrapper[4907]: I0313 15:46:01.031924 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556946-78rqg" event={"ID":"2f0d1531-1d25-4331-a690-1fcbcb504071","Type":"ContainerStarted","Data":"9e7cadbe55b65797301703c241dfad204c1d7276015b0c5406cafc5fc5c850e2"} Mar 13 15:46:02 crc kubenswrapper[4907]: E0313 15:46:02.497075 4907 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2f0d1531_1d25_4331_a690_1fcbcb504071.slice/crio-conmon-971365269f08ef2266842a1c0ca18c6bb47126b67a1360b35d5a7dfd3acab380.scope\": RecentStats: unable to find data in memory cache]" Mar 13 15:46:02 crc kubenswrapper[4907]: I0313 15:46:02.563717 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-healthmanager-hbbw4"] Mar 13 15:46:02 crc kubenswrapper[4907]: I0313 15:46:02.566458 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-healthmanager-hbbw4" Mar 13 15:46:02 crc kubenswrapper[4907]: I0313 15:46:02.568272 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-healthmanager-scripts" Mar 13 15:46:02 crc kubenswrapper[4907]: I0313 15:46:02.568345 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-healthmanager-config-data" Mar 13 15:46:02 crc kubenswrapper[4907]: I0313 15:46:02.568669 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-certs-secret" Mar 13 15:46:02 crc kubenswrapper[4907]: I0313 15:46:02.580527 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-healthmanager-hbbw4"] Mar 13 15:46:02 crc kubenswrapper[4907]: I0313 15:46:02.737657 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/57610eac-b9eb-4628-bd99-b3ebb35b2772-amphora-certs\") pod \"octavia-healthmanager-hbbw4\" (UID: \"57610eac-b9eb-4628-bd99-b3ebb35b2772\") " pod="openstack/octavia-healthmanager-hbbw4" Mar 13 15:46:02 crc kubenswrapper[4907]: I0313 15:46:02.737706 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/57610eac-b9eb-4628-bd99-b3ebb35b2772-hm-ports\") pod \"octavia-healthmanager-hbbw4\" (UID: \"57610eac-b9eb-4628-bd99-b3ebb35b2772\") " pod="openstack/octavia-healthmanager-hbbw4" Mar 13 15:46:02 crc kubenswrapper[4907]: I0313 15:46:02.737795 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57610eac-b9eb-4628-bd99-b3ebb35b2772-combined-ca-bundle\") pod \"octavia-healthmanager-hbbw4\" (UID: \"57610eac-b9eb-4628-bd99-b3ebb35b2772\") " pod="openstack/octavia-healthmanager-hbbw4" Mar 13 15:46:02 crc kubenswrapper[4907]: I0313 15:46:02.737913 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/57610eac-b9eb-4628-bd99-b3ebb35b2772-scripts\") pod \"octavia-healthmanager-hbbw4\" (UID: \"57610eac-b9eb-4628-bd99-b3ebb35b2772\") " pod="openstack/octavia-healthmanager-hbbw4" Mar 13 15:46:02 crc kubenswrapper[4907]: I0313 15:46:02.737938 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57610eac-b9eb-4628-bd99-b3ebb35b2772-config-data\") pod \"octavia-healthmanager-hbbw4\" (UID: \"57610eac-b9eb-4628-bd99-b3ebb35b2772\") " pod="openstack/octavia-healthmanager-hbbw4" Mar 13 15:46:02 crc kubenswrapper[4907]: I0313 15:46:02.737986 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/57610eac-b9eb-4628-bd99-b3ebb35b2772-config-data-merged\") pod \"octavia-healthmanager-hbbw4\" (UID: \"57610eac-b9eb-4628-bd99-b3ebb35b2772\") " pod="openstack/octavia-healthmanager-hbbw4" Mar 13 15:46:02 crc kubenswrapper[4907]: I0313 15:46:02.839643 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/57610eac-b9eb-4628-bd99-b3ebb35b2772-amphora-certs\") pod \"octavia-healthmanager-hbbw4\" (UID: \"57610eac-b9eb-4628-bd99-b3ebb35b2772\") " pod="openstack/octavia-healthmanager-hbbw4" Mar 13 15:46:02 crc kubenswrapper[4907]: I0313 15:46:02.839702 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/57610eac-b9eb-4628-bd99-b3ebb35b2772-hm-ports\") pod \"octavia-healthmanager-hbbw4\" (UID: \"57610eac-b9eb-4628-bd99-b3ebb35b2772\") " pod="openstack/octavia-healthmanager-hbbw4" Mar 13 15:46:02 crc kubenswrapper[4907]: I0313 15:46:02.839760 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57610eac-b9eb-4628-bd99-b3ebb35b2772-combined-ca-bundle\") pod \"octavia-healthmanager-hbbw4\" (UID: \"57610eac-b9eb-4628-bd99-b3ebb35b2772\") " pod="openstack/octavia-healthmanager-hbbw4" Mar 13 15:46:02 crc kubenswrapper[4907]: I0313 15:46:02.839796 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/57610eac-b9eb-4628-bd99-b3ebb35b2772-scripts\") pod \"octavia-healthmanager-hbbw4\" (UID: \"57610eac-b9eb-4628-bd99-b3ebb35b2772\") " pod="openstack/octavia-healthmanager-hbbw4" Mar 13 15:46:02 crc kubenswrapper[4907]: I0313 15:46:02.839818 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57610eac-b9eb-4628-bd99-b3ebb35b2772-config-data\") pod \"octavia-healthmanager-hbbw4\" (UID: \"57610eac-b9eb-4628-bd99-b3ebb35b2772\") " pod="openstack/octavia-healthmanager-hbbw4" Mar 13 15:46:02 crc kubenswrapper[4907]: I0313 15:46:02.839859 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/57610eac-b9eb-4628-bd99-b3ebb35b2772-config-data-merged\") pod \"octavia-healthmanager-hbbw4\" (UID: \"57610eac-b9eb-4628-bd99-b3ebb35b2772\") " pod="openstack/octavia-healthmanager-hbbw4" Mar 13 15:46:02 crc kubenswrapper[4907]: I0313 15:46:02.840564 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/57610eac-b9eb-4628-bd99-b3ebb35b2772-config-data-merged\") pod \"octavia-healthmanager-hbbw4\" (UID: \"57610eac-b9eb-4628-bd99-b3ebb35b2772\") " pod="openstack/octavia-healthmanager-hbbw4" Mar 13 15:46:02 crc kubenswrapper[4907]: I0313 15:46:02.841109 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/57610eac-b9eb-4628-bd99-b3ebb35b2772-hm-ports\") pod \"octavia-healthmanager-hbbw4\" (UID: \"57610eac-b9eb-4628-bd99-b3ebb35b2772\") " pod="openstack/octavia-healthmanager-hbbw4" Mar 13 15:46:02 crc kubenswrapper[4907]: I0313 15:46:02.846032 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57610eac-b9eb-4628-bd99-b3ebb35b2772-config-data\") pod \"octavia-healthmanager-hbbw4\" (UID: \"57610eac-b9eb-4628-bd99-b3ebb35b2772\") " pod="openstack/octavia-healthmanager-hbbw4" Mar 13 15:46:02 crc kubenswrapper[4907]: I0313 15:46:02.846296 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/57610eac-b9eb-4628-bd99-b3ebb35b2772-scripts\") pod \"octavia-healthmanager-hbbw4\" (UID: \"57610eac-b9eb-4628-bd99-b3ebb35b2772\") " pod="openstack/octavia-healthmanager-hbbw4" Mar 13 15:46:02 crc kubenswrapper[4907]: I0313 15:46:02.860683 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57610eac-b9eb-4628-bd99-b3ebb35b2772-combined-ca-bundle\") pod \"octavia-healthmanager-hbbw4\" (UID: \"57610eac-b9eb-4628-bd99-b3ebb35b2772\") " pod="openstack/octavia-healthmanager-hbbw4" Mar 13 15:46:02 crc kubenswrapper[4907]: I0313 15:46:02.863785 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/57610eac-b9eb-4628-bd99-b3ebb35b2772-amphora-certs\") pod \"octavia-healthmanager-hbbw4\" (UID: \"57610eac-b9eb-4628-bd99-b3ebb35b2772\") " pod="openstack/octavia-healthmanager-hbbw4" Mar 13 15:46:02 crc kubenswrapper[4907]: I0313 15:46:02.885002 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-healthmanager-hbbw4" Mar 13 15:46:03 crc kubenswrapper[4907]: I0313 15:46:03.052854 4907 generic.go:334] "Generic (PLEG): container finished" podID="2f0d1531-1d25-4331-a690-1fcbcb504071" containerID="971365269f08ef2266842a1c0ca18c6bb47126b67a1360b35d5a7dfd3acab380" exitCode=0 Mar 13 15:46:03 crc kubenswrapper[4907]: I0313 15:46:03.053044 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556946-78rqg" event={"ID":"2f0d1531-1d25-4331-a690-1fcbcb504071","Type":"ContainerDied","Data":"971365269f08ef2266842a1c0ca18c6bb47126b67a1360b35d5a7dfd3acab380"} Mar 13 15:46:03 crc kubenswrapper[4907]: I0313 15:46:03.602994 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-healthmanager-hbbw4"] Mar 13 15:46:04 crc kubenswrapper[4907]: I0313 15:46:04.063907 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-healthmanager-hbbw4" event={"ID":"57610eac-b9eb-4628-bd99-b3ebb35b2772","Type":"ContainerStarted","Data":"27749e579eeb959a64f9197a7832237eb739e4daba11766edbfeb3f7aaf9ad37"} Mar 13 15:46:04 crc kubenswrapper[4907]: I0313 15:46:04.064218 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-healthmanager-hbbw4" event={"ID":"57610eac-b9eb-4628-bd99-b3ebb35b2772","Type":"ContainerStarted","Data":"b035a1861472eee7e65e830e7bb53c1fa16793b9e742ac953f22caab1ad82095"} Mar 13 15:46:04 crc kubenswrapper[4907]: I0313 15:46:04.231284 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-housekeeping-p69c5"] Mar 13 15:46:04 crc kubenswrapper[4907]: I0313 15:46:04.233230 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-housekeeping-p69c5" Mar 13 15:46:04 crc kubenswrapper[4907]: I0313 15:46:04.236975 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-housekeeping-config-data" Mar 13 15:46:04 crc kubenswrapper[4907]: I0313 15:46:04.237287 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-housekeeping-scripts" Mar 13 15:46:04 crc kubenswrapper[4907]: I0313 15:46:04.245078 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-housekeeping-p69c5"] Mar 13 15:46:04 crc kubenswrapper[4907]: I0313 15:46:04.374164 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/ab5868a8-f083-4417-b7c0-f01dad538927-hm-ports\") pod \"octavia-housekeeping-p69c5\" (UID: \"ab5868a8-f083-4417-b7c0-f01dad538927\") " pod="openstack/octavia-housekeeping-p69c5" Mar 13 15:46:04 crc kubenswrapper[4907]: I0313 15:46:04.374217 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/ab5868a8-f083-4417-b7c0-f01dad538927-amphora-certs\") pod \"octavia-housekeeping-p69c5\" (UID: \"ab5868a8-f083-4417-b7c0-f01dad538927\") " pod="openstack/octavia-housekeeping-p69c5" Mar 13 15:46:04 crc kubenswrapper[4907]: I0313 15:46:04.374286 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ab5868a8-f083-4417-b7c0-f01dad538927-scripts\") pod \"octavia-housekeeping-p69c5\" (UID: \"ab5868a8-f083-4417-b7c0-f01dad538927\") " pod="openstack/octavia-housekeeping-p69c5" Mar 13 15:46:04 crc kubenswrapper[4907]: I0313 15:46:04.374324 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab5868a8-f083-4417-b7c0-f01dad538927-combined-ca-bundle\") pod \"octavia-housekeeping-p69c5\" (UID: \"ab5868a8-f083-4417-b7c0-f01dad538927\") " pod="openstack/octavia-housekeeping-p69c5" Mar 13 15:46:04 crc kubenswrapper[4907]: I0313 15:46:04.374357 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab5868a8-f083-4417-b7c0-f01dad538927-config-data\") pod \"octavia-housekeeping-p69c5\" (UID: \"ab5868a8-f083-4417-b7c0-f01dad538927\") " pod="openstack/octavia-housekeeping-p69c5" Mar 13 15:46:04 crc kubenswrapper[4907]: I0313 15:46:04.374417 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/ab5868a8-f083-4417-b7c0-f01dad538927-config-data-merged\") pod \"octavia-housekeeping-p69c5\" (UID: \"ab5868a8-f083-4417-b7c0-f01dad538927\") " pod="openstack/octavia-housekeeping-p69c5" Mar 13 15:46:04 crc kubenswrapper[4907]: I0313 15:46:04.425791 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556946-78rqg" Mar 13 15:46:04 crc kubenswrapper[4907]: I0313 15:46:04.475656 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ab5868a8-f083-4417-b7c0-f01dad538927-scripts\") pod \"octavia-housekeeping-p69c5\" (UID: \"ab5868a8-f083-4417-b7c0-f01dad538927\") " pod="openstack/octavia-housekeeping-p69c5" Mar 13 15:46:04 crc kubenswrapper[4907]: I0313 15:46:04.475713 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab5868a8-f083-4417-b7c0-f01dad538927-combined-ca-bundle\") pod \"octavia-housekeeping-p69c5\" (UID: \"ab5868a8-f083-4417-b7c0-f01dad538927\") " pod="openstack/octavia-housekeeping-p69c5" Mar 13 15:46:04 crc kubenswrapper[4907]: I0313 15:46:04.475744 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab5868a8-f083-4417-b7c0-f01dad538927-config-data\") pod \"octavia-housekeeping-p69c5\" (UID: \"ab5868a8-f083-4417-b7c0-f01dad538927\") " pod="openstack/octavia-housekeeping-p69c5" Mar 13 15:46:04 crc kubenswrapper[4907]: I0313 15:46:04.475812 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/ab5868a8-f083-4417-b7c0-f01dad538927-config-data-merged\") pod \"octavia-housekeeping-p69c5\" (UID: \"ab5868a8-f083-4417-b7c0-f01dad538927\") " pod="openstack/octavia-housekeeping-p69c5" Mar 13 15:46:04 crc kubenswrapper[4907]: I0313 15:46:04.475847 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/ab5868a8-f083-4417-b7c0-f01dad538927-hm-ports\") pod \"octavia-housekeeping-p69c5\" (UID: \"ab5868a8-f083-4417-b7c0-f01dad538927\") " pod="openstack/octavia-housekeeping-p69c5" Mar 13 15:46:04 crc kubenswrapper[4907]: I0313 15:46:04.475873 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/ab5868a8-f083-4417-b7c0-f01dad538927-amphora-certs\") pod \"octavia-housekeeping-p69c5\" (UID: \"ab5868a8-f083-4417-b7c0-f01dad538927\") " pod="openstack/octavia-housekeeping-p69c5" Mar 13 15:46:04 crc kubenswrapper[4907]: I0313 15:46:04.476817 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/ab5868a8-f083-4417-b7c0-f01dad538927-config-data-merged\") pod \"octavia-housekeeping-p69c5\" (UID: \"ab5868a8-f083-4417-b7c0-f01dad538927\") " pod="openstack/octavia-housekeeping-p69c5" Mar 13 15:46:04 crc kubenswrapper[4907]: I0313 15:46:04.477953 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/ab5868a8-f083-4417-b7c0-f01dad538927-hm-ports\") pod \"octavia-housekeeping-p69c5\" (UID: \"ab5868a8-f083-4417-b7c0-f01dad538927\") " pod="openstack/octavia-housekeeping-p69c5" Mar 13 15:46:04 crc kubenswrapper[4907]: I0313 15:46:04.482025 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab5868a8-f083-4417-b7c0-f01dad538927-combined-ca-bundle\") pod \"octavia-housekeeping-p69c5\" (UID: \"ab5868a8-f083-4417-b7c0-f01dad538927\") " pod="openstack/octavia-housekeeping-p69c5" Mar 13 15:46:04 crc kubenswrapper[4907]: I0313 15:46:04.482248 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ab5868a8-f083-4417-b7c0-f01dad538927-scripts\") pod \"octavia-housekeeping-p69c5\" (UID: \"ab5868a8-f083-4417-b7c0-f01dad538927\") " pod="openstack/octavia-housekeeping-p69c5" Mar 13 15:46:04 crc kubenswrapper[4907]: I0313 15:46:04.485923 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab5868a8-f083-4417-b7c0-f01dad538927-config-data\") pod \"octavia-housekeeping-p69c5\" (UID: \"ab5868a8-f083-4417-b7c0-f01dad538927\") " pod="openstack/octavia-housekeeping-p69c5" Mar 13 15:46:04 crc kubenswrapper[4907]: I0313 15:46:04.486597 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/ab5868a8-f083-4417-b7c0-f01dad538927-amphora-certs\") pod \"octavia-housekeeping-p69c5\" (UID: \"ab5868a8-f083-4417-b7c0-f01dad538927\") " pod="openstack/octavia-housekeeping-p69c5" Mar 13 15:46:04 crc kubenswrapper[4907]: I0313 15:46:04.558082 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-housekeeping-p69c5" Mar 13 15:46:04 crc kubenswrapper[4907]: I0313 15:46:04.589187 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lqghz\" (UniqueName: \"kubernetes.io/projected/2f0d1531-1d25-4331-a690-1fcbcb504071-kube-api-access-lqghz\") pod \"2f0d1531-1d25-4331-a690-1fcbcb504071\" (UID: \"2f0d1531-1d25-4331-a690-1fcbcb504071\") " Mar 13 15:46:04 crc kubenswrapper[4907]: I0313 15:46:04.592334 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f0d1531-1d25-4331-a690-1fcbcb504071-kube-api-access-lqghz" (OuterVolumeSpecName: "kube-api-access-lqghz") pod "2f0d1531-1d25-4331-a690-1fcbcb504071" (UID: "2f0d1531-1d25-4331-a690-1fcbcb504071"). InnerVolumeSpecName "kube-api-access-lqghz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:46:04 crc kubenswrapper[4907]: I0313 15:46:04.691853 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lqghz\" (UniqueName: \"kubernetes.io/projected/2f0d1531-1d25-4331-a690-1fcbcb504071-kube-api-access-lqghz\") on node \"crc\" DevicePath \"\"" Mar 13 15:46:05 crc kubenswrapper[4907]: I0313 15:46:05.010870 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-worker-7pld2"] Mar 13 15:46:05 crc kubenswrapper[4907]: E0313 15:46:05.011943 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f0d1531-1d25-4331-a690-1fcbcb504071" containerName="oc" Mar 13 15:46:05 crc kubenswrapper[4907]: I0313 15:46:05.012019 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f0d1531-1d25-4331-a690-1fcbcb504071" containerName="oc" Mar 13 15:46:05 crc kubenswrapper[4907]: I0313 15:46:05.012292 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f0d1531-1d25-4331-a690-1fcbcb504071" containerName="oc" Mar 13 15:46:05 crc kubenswrapper[4907]: I0313 15:46:05.022154 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-worker-7pld2" Mar 13 15:46:05 crc kubenswrapper[4907]: I0313 15:46:05.032420 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-worker-scripts" Mar 13 15:46:05 crc kubenswrapper[4907]: I0313 15:46:05.032665 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-worker-config-data" Mar 13 15:46:05 crc kubenswrapper[4907]: I0313 15:46:05.050567 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-worker-7pld2"] Mar 13 15:46:05 crc kubenswrapper[4907]: I0313 15:46:05.086166 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556946-78rqg" Mar 13 15:46:05 crc kubenswrapper[4907]: I0313 15:46:05.098001 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556946-78rqg" event={"ID":"2f0d1531-1d25-4331-a690-1fcbcb504071","Type":"ContainerDied","Data":"9e7cadbe55b65797301703c241dfad204c1d7276015b0c5406cafc5fc5c850e2"} Mar 13 15:46:05 crc kubenswrapper[4907]: I0313 15:46:05.098067 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9e7cadbe55b65797301703c241dfad204c1d7276015b0c5406cafc5fc5c850e2" Mar 13 15:46:05 crc kubenswrapper[4907]: I0313 15:46:05.207947 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/f6c0a640-5177-4188-abc3-d934f75624bb-config-data-merged\") pod \"octavia-worker-7pld2\" (UID: \"f6c0a640-5177-4188-abc3-d934f75624bb\") " pod="openstack/octavia-worker-7pld2" Mar 13 15:46:05 crc kubenswrapper[4907]: I0313 15:46:05.208019 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/f6c0a640-5177-4188-abc3-d934f75624bb-amphora-certs\") pod \"octavia-worker-7pld2\" (UID: \"f6c0a640-5177-4188-abc3-d934f75624bb\") " pod="openstack/octavia-worker-7pld2" Mar 13 15:46:05 crc kubenswrapper[4907]: I0313 15:46:05.208109 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/f6c0a640-5177-4188-abc3-d934f75624bb-hm-ports\") pod \"octavia-worker-7pld2\" (UID: \"f6c0a640-5177-4188-abc3-d934f75624bb\") " pod="openstack/octavia-worker-7pld2" Mar 13 15:46:05 crc kubenswrapper[4907]: I0313 15:46:05.208155 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f6c0a640-5177-4188-abc3-d934f75624bb-scripts\") pod \"octavia-worker-7pld2\" (UID: \"f6c0a640-5177-4188-abc3-d934f75624bb\") " pod="openstack/octavia-worker-7pld2" Mar 13 15:46:05 crc kubenswrapper[4907]: I0313 15:46:05.208192 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6c0a640-5177-4188-abc3-d934f75624bb-config-data\") pod \"octavia-worker-7pld2\" (UID: \"f6c0a640-5177-4188-abc3-d934f75624bb\") " pod="openstack/octavia-worker-7pld2" Mar 13 15:46:05 crc kubenswrapper[4907]: I0313 15:46:05.208211 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6c0a640-5177-4188-abc3-d934f75624bb-combined-ca-bundle\") pod \"octavia-worker-7pld2\" (UID: \"f6c0a640-5177-4188-abc3-d934f75624bb\") " pod="openstack/octavia-worker-7pld2" Mar 13 15:46:05 crc kubenswrapper[4907]: I0313 15:46:05.254020 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-housekeeping-p69c5"] Mar 13 15:46:05 crc kubenswrapper[4907]: I0313 15:46:05.312205 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f6c0a640-5177-4188-abc3-d934f75624bb-scripts\") pod \"octavia-worker-7pld2\" (UID: \"f6c0a640-5177-4188-abc3-d934f75624bb\") " pod="openstack/octavia-worker-7pld2" Mar 13 15:46:05 crc kubenswrapper[4907]: I0313 15:46:05.312302 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6c0a640-5177-4188-abc3-d934f75624bb-config-data\") pod \"octavia-worker-7pld2\" (UID: \"f6c0a640-5177-4188-abc3-d934f75624bb\") " pod="openstack/octavia-worker-7pld2" Mar 13 15:46:05 crc kubenswrapper[4907]: I0313 15:46:05.312332 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6c0a640-5177-4188-abc3-d934f75624bb-combined-ca-bundle\") pod \"octavia-worker-7pld2\" (UID: \"f6c0a640-5177-4188-abc3-d934f75624bb\") " pod="openstack/octavia-worker-7pld2" Mar 13 15:46:05 crc kubenswrapper[4907]: I0313 15:46:05.312382 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/f6c0a640-5177-4188-abc3-d934f75624bb-config-data-merged\") pod \"octavia-worker-7pld2\" (UID: \"f6c0a640-5177-4188-abc3-d934f75624bb\") " pod="openstack/octavia-worker-7pld2" Mar 13 15:46:05 crc kubenswrapper[4907]: I0313 15:46:05.312434 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/f6c0a640-5177-4188-abc3-d934f75624bb-amphora-certs\") pod \"octavia-worker-7pld2\" (UID: \"f6c0a640-5177-4188-abc3-d934f75624bb\") " pod="openstack/octavia-worker-7pld2" Mar 13 15:46:05 crc kubenswrapper[4907]: I0313 15:46:05.312547 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/f6c0a640-5177-4188-abc3-d934f75624bb-hm-ports\") pod \"octavia-worker-7pld2\" (UID: \"f6c0a640-5177-4188-abc3-d934f75624bb\") " pod="openstack/octavia-worker-7pld2" Mar 13 15:46:05 crc kubenswrapper[4907]: I0313 15:46:05.313840 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/f6c0a640-5177-4188-abc3-d934f75624bb-hm-ports\") pod \"octavia-worker-7pld2\" (UID: \"f6c0a640-5177-4188-abc3-d934f75624bb\") " pod="openstack/octavia-worker-7pld2" Mar 13 15:46:05 crc kubenswrapper[4907]: I0313 15:46:05.314242 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/f6c0a640-5177-4188-abc3-d934f75624bb-config-data-merged\") pod \"octavia-worker-7pld2\" (UID: \"f6c0a640-5177-4188-abc3-d934f75624bb\") " pod="openstack/octavia-worker-7pld2" Mar 13 15:46:05 crc kubenswrapper[4907]: I0313 15:46:05.324312 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6c0a640-5177-4188-abc3-d934f75624bb-combined-ca-bundle\") pod \"octavia-worker-7pld2\" (UID: \"f6c0a640-5177-4188-abc3-d934f75624bb\") " pod="openstack/octavia-worker-7pld2" Mar 13 15:46:05 crc kubenswrapper[4907]: I0313 15:46:05.329057 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f6c0a640-5177-4188-abc3-d934f75624bb-scripts\") pod \"octavia-worker-7pld2\" (UID: \"f6c0a640-5177-4188-abc3-d934f75624bb\") " pod="openstack/octavia-worker-7pld2" Mar 13 15:46:05 crc kubenswrapper[4907]: I0313 15:46:05.333621 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/f6c0a640-5177-4188-abc3-d934f75624bb-amphora-certs\") pod \"octavia-worker-7pld2\" (UID: \"f6c0a640-5177-4188-abc3-d934f75624bb\") " pod="openstack/octavia-worker-7pld2" Mar 13 15:46:05 crc kubenswrapper[4907]: I0313 15:46:05.342308 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6c0a640-5177-4188-abc3-d934f75624bb-config-data\") pod \"octavia-worker-7pld2\" (UID: \"f6c0a640-5177-4188-abc3-d934f75624bb\") " pod="openstack/octavia-worker-7pld2" Mar 13 15:46:05 crc kubenswrapper[4907]: I0313 15:46:05.351508 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-worker-7pld2" Mar 13 15:46:05 crc kubenswrapper[4907]: I0313 15:46:05.509509 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556940-5rmt4"] Mar 13 15:46:05 crc kubenswrapper[4907]: I0313 15:46:05.522224 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556940-5rmt4"] Mar 13 15:46:05 crc kubenswrapper[4907]: I0313 15:46:05.793262 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="72d20fdf-a093-459f-9328-0549ef7b48c4" path="/var/lib/kubelet/pods/72d20fdf-a093-459f-9328-0549ef7b48c4/volumes" Mar 13 15:46:05 crc kubenswrapper[4907]: I0313 15:46:05.943601 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-worker-7pld2"] Mar 13 15:46:06 crc kubenswrapper[4907]: I0313 15:46:06.096644 4907 generic.go:334] "Generic (PLEG): container finished" podID="57610eac-b9eb-4628-bd99-b3ebb35b2772" containerID="27749e579eeb959a64f9197a7832237eb739e4daba11766edbfeb3f7aaf9ad37" exitCode=0 Mar 13 15:46:06 crc kubenswrapper[4907]: I0313 15:46:06.096747 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-healthmanager-hbbw4" event={"ID":"57610eac-b9eb-4628-bd99-b3ebb35b2772","Type":"ContainerDied","Data":"27749e579eeb959a64f9197a7832237eb739e4daba11766edbfeb3f7aaf9ad37"} Mar 13 15:46:06 crc kubenswrapper[4907]: I0313 15:46:06.101199 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-worker-7pld2" event={"ID":"f6c0a640-5177-4188-abc3-d934f75624bb","Type":"ContainerStarted","Data":"f76a35ffcd772eb2bf027b892918eded8ae0d90b81519998bea60cc5f8343949"} Mar 13 15:46:06 crc kubenswrapper[4907]: I0313 15:46:06.102842 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-housekeeping-p69c5" event={"ID":"ab5868a8-f083-4417-b7c0-f01dad538927","Type":"ContainerStarted","Data":"4400db9c9218563d608796b392e4cf7cc6279fc112b2adcdb10d47be37316c03"} Mar 13 15:46:06 crc kubenswrapper[4907]: I0313 15:46:06.858106 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-healthmanager-hbbw4"] Mar 13 15:46:07 crc kubenswrapper[4907]: I0313 15:46:07.115265 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-healthmanager-hbbw4" event={"ID":"57610eac-b9eb-4628-bd99-b3ebb35b2772","Type":"ContainerStarted","Data":"1158feea7d437e31315b4d5ffcb87d298513a93189b92fc5e0a54854516e5572"} Mar 13 15:46:07 crc kubenswrapper[4907]: I0313 15:46:07.115533 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-healthmanager-hbbw4" Mar 13 15:46:07 crc kubenswrapper[4907]: I0313 15:46:07.145145 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-healthmanager-hbbw4" podStartSLOduration=5.145126576 podStartE2EDuration="5.145126576s" podCreationTimestamp="2026-03-13 15:46:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:46:07.136434938 +0000 UTC m=+6066.036222637" watchObservedRunningTime="2026-03-13 15:46:07.145126576 +0000 UTC m=+6066.044914265" Mar 13 15:46:09 crc kubenswrapper[4907]: I0313 15:46:09.149822 4907 generic.go:334] "Generic (PLEG): container finished" podID="f6c0a640-5177-4188-abc3-d934f75624bb" containerID="f8136dbfa25da4d13b996f410c12861a563747986411d892b9b92b671cefafa6" exitCode=0 Mar 13 15:46:09 crc kubenswrapper[4907]: I0313 15:46:09.149887 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-worker-7pld2" event={"ID":"f6c0a640-5177-4188-abc3-d934f75624bb","Type":"ContainerDied","Data":"f8136dbfa25da4d13b996f410c12861a563747986411d892b9b92b671cefafa6"} Mar 13 15:46:09 crc kubenswrapper[4907]: I0313 15:46:09.152396 4907 generic.go:334] "Generic (PLEG): container finished" podID="ab5868a8-f083-4417-b7c0-f01dad538927" containerID="d83f400367245cd91cb0153fe29d0c6653531c070b4435e2eea12f7916668591" exitCode=0 Mar 13 15:46:09 crc kubenswrapper[4907]: I0313 15:46:09.152640 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-housekeeping-p69c5" event={"ID":"ab5868a8-f083-4417-b7c0-f01dad538927","Type":"ContainerDied","Data":"d83f400367245cd91cb0153fe29d0c6653531c070b4435e2eea12f7916668591"} Mar 13 15:46:09 crc kubenswrapper[4907]: I0313 15:46:09.782782 4907 scope.go:117] "RemoveContainer" containerID="68ffe92dd03078633ec89d7ef29da9bbdc4c5f04c956a1e8487e3f2e01c5d1c5" Mar 13 15:46:09 crc kubenswrapper[4907]: E0313 15:46:09.783569 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:46:10 crc kubenswrapper[4907]: I0313 15:46:10.163108 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-worker-7pld2" event={"ID":"f6c0a640-5177-4188-abc3-d934f75624bb","Type":"ContainerStarted","Data":"7d42d246473bf2b8e462bcf74b14d3e4997dd61f12af5cb13a8fb60e3bac48dd"} Mar 13 15:46:10 crc kubenswrapper[4907]: I0313 15:46:10.164367 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-worker-7pld2" Mar 13 15:46:10 crc kubenswrapper[4907]: I0313 15:46:10.167026 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-housekeeping-p69c5" event={"ID":"ab5868a8-f083-4417-b7c0-f01dad538927","Type":"ContainerStarted","Data":"4bcbc2bf02d639dacf282548340a376215a25f0b29e4612a986c495aa5db94fd"} Mar 13 15:46:10 crc kubenswrapper[4907]: I0313 15:46:10.167259 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-housekeeping-p69c5" Mar 13 15:46:10 crc kubenswrapper[4907]: I0313 15:46:10.187307 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-worker-7pld2" podStartSLOduration=4.42259988 podStartE2EDuration="6.187282843s" podCreationTimestamp="2026-03-13 15:46:04 +0000 UTC" firstStartedPulling="2026-03-13 15:46:05.968499869 +0000 UTC m=+6064.868287558" lastFinishedPulling="2026-03-13 15:46:07.733182832 +0000 UTC m=+6066.632970521" observedRunningTime="2026-03-13 15:46:10.18352672 +0000 UTC m=+6069.083314409" watchObservedRunningTime="2026-03-13 15:46:10.187282843 +0000 UTC m=+6069.087070532" Mar 13 15:46:10 crc kubenswrapper[4907]: I0313 15:46:10.210087 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-housekeeping-p69c5" podStartSLOduration=3.745419457 podStartE2EDuration="6.210068875s" podCreationTimestamp="2026-03-13 15:46:04 +0000 UTC" firstStartedPulling="2026-03-13 15:46:05.265751858 +0000 UTC m=+6064.165539547" lastFinishedPulling="2026-03-13 15:46:07.730401276 +0000 UTC m=+6066.630188965" observedRunningTime="2026-03-13 15:46:10.207867825 +0000 UTC m=+6069.107655524" watchObservedRunningTime="2026-03-13 15:46:10.210068875 +0000 UTC m=+6069.109856554" Mar 13 15:46:17 crc kubenswrapper[4907]: I0313 15:46:17.913726 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-healthmanager-hbbw4" Mar 13 15:46:19 crc kubenswrapper[4907]: I0313 15:46:19.713081 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-housekeeping-p69c5" Mar 13 15:46:20 crc kubenswrapper[4907]: I0313 15:46:20.384313 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-worker-7pld2" Mar 13 15:46:21 crc kubenswrapper[4907]: I0313 15:46:21.790206 4907 scope.go:117] "RemoveContainer" containerID="68ffe92dd03078633ec89d7ef29da9bbdc4c5f04c956a1e8487e3f2e01c5d1c5" Mar 13 15:46:22 crc kubenswrapper[4907]: I0313 15:46:22.278469 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerStarted","Data":"81d94e236cbdff3b86e8375b8abbaee97f49d844932ee0151cdfa9d97cb1c3b2"} Mar 13 15:46:26 crc kubenswrapper[4907]: I0313 15:46:26.272852 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-c7977bcd9-2vs8h"] Mar 13 15:46:26 crc kubenswrapper[4907]: I0313 15:46:26.275814 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-c7977bcd9-2vs8h" Mar 13 15:46:26 crc kubenswrapper[4907]: I0313 15:46:26.280996 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Mar 13 15:46:26 crc kubenswrapper[4907]: I0313 15:46:26.281027 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Mar 13 15:46:26 crc kubenswrapper[4907]: I0313 15:46:26.281318 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Mar 13 15:46:26 crc kubenswrapper[4907]: I0313 15:46:26.281394 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-tpjbz" Mar 13 15:46:26 crc kubenswrapper[4907]: I0313 15:46:26.289630 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-c7977bcd9-2vs8h"] Mar 13 15:46:26 crc kubenswrapper[4907]: I0313 15:46:26.336006 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/2d9fbab8-539c-48f6-8cae-3eda23300b28-horizon-secret-key\") pod \"horizon-c7977bcd9-2vs8h\" (UID: \"2d9fbab8-539c-48f6-8cae-3eda23300b28\") " pod="openstack/horizon-c7977bcd9-2vs8h" Mar 13 15:46:26 crc kubenswrapper[4907]: I0313 15:46:26.336056 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2d9fbab8-539c-48f6-8cae-3eda23300b28-logs\") pod \"horizon-c7977bcd9-2vs8h\" (UID: \"2d9fbab8-539c-48f6-8cae-3eda23300b28\") " pod="openstack/horizon-c7977bcd9-2vs8h" Mar 13 15:46:26 crc kubenswrapper[4907]: I0313 15:46:26.336148 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2d9fbab8-539c-48f6-8cae-3eda23300b28-config-data\") pod \"horizon-c7977bcd9-2vs8h\" (UID: \"2d9fbab8-539c-48f6-8cae-3eda23300b28\") " pod="openstack/horizon-c7977bcd9-2vs8h" Mar 13 15:46:26 crc kubenswrapper[4907]: I0313 15:46:26.336202 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bjhcq\" (UniqueName: \"kubernetes.io/projected/2d9fbab8-539c-48f6-8cae-3eda23300b28-kube-api-access-bjhcq\") pod \"horizon-c7977bcd9-2vs8h\" (UID: \"2d9fbab8-539c-48f6-8cae-3eda23300b28\") " pod="openstack/horizon-c7977bcd9-2vs8h" Mar 13 15:46:26 crc kubenswrapper[4907]: I0313 15:46:26.336254 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2d9fbab8-539c-48f6-8cae-3eda23300b28-scripts\") pod \"horizon-c7977bcd9-2vs8h\" (UID: \"2d9fbab8-539c-48f6-8cae-3eda23300b28\") " pod="openstack/horizon-c7977bcd9-2vs8h" Mar 13 15:46:26 crc kubenswrapper[4907]: I0313 15:46:26.342058 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 13 15:46:26 crc kubenswrapper[4907]: I0313 15:46:26.342313 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="86794caf-bf6d-4394-bc1c-3c516bdbada8" containerName="glance-log" containerID="cri-o://56ff9c5433a9cbee0386dea3d88b4e88f870fa5dcf67d5221e12245f174e8f85" gracePeriod=30 Mar 13 15:46:26 crc kubenswrapper[4907]: I0313 15:46:26.342732 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="86794caf-bf6d-4394-bc1c-3c516bdbada8" containerName="glance-httpd" containerID="cri-o://d5e6285fe8e2b22da125d5022a7924d7086ba68515ab90ca41e35d0607040a03" gracePeriod=30 Mar 13 15:46:26 crc kubenswrapper[4907]: I0313 15:46:26.411998 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 13 15:46:26 crc kubenswrapper[4907]: I0313 15:46:26.412549 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="096d3713-ea4f-43d3-bfb9-9170e8958ed0" containerName="glance-log" containerID="cri-o://02ef5ab885e1e984508608add8cbc5f616e0cf1ed8d42cf7ded2960a6fa681f8" gracePeriod=30 Mar 13 15:46:26 crc kubenswrapper[4907]: I0313 15:46:26.412711 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="096d3713-ea4f-43d3-bfb9-9170e8958ed0" containerName="glance-httpd" containerID="cri-o://27209680abb804b6c48ccdaa33b44ab4f54951d1df1401058af333f852827d0f" gracePeriod=30 Mar 13 15:46:26 crc kubenswrapper[4907]: I0313 15:46:26.437977 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2d9fbab8-539c-48f6-8cae-3eda23300b28-config-data\") pod \"horizon-c7977bcd9-2vs8h\" (UID: \"2d9fbab8-539c-48f6-8cae-3eda23300b28\") " pod="openstack/horizon-c7977bcd9-2vs8h" Mar 13 15:46:26 crc kubenswrapper[4907]: I0313 15:46:26.438052 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bjhcq\" (UniqueName: \"kubernetes.io/projected/2d9fbab8-539c-48f6-8cae-3eda23300b28-kube-api-access-bjhcq\") pod \"horizon-c7977bcd9-2vs8h\" (UID: \"2d9fbab8-539c-48f6-8cae-3eda23300b28\") " pod="openstack/horizon-c7977bcd9-2vs8h" Mar 13 15:46:26 crc kubenswrapper[4907]: I0313 15:46:26.438091 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2d9fbab8-539c-48f6-8cae-3eda23300b28-scripts\") pod \"horizon-c7977bcd9-2vs8h\" (UID: \"2d9fbab8-539c-48f6-8cae-3eda23300b28\") " pod="openstack/horizon-c7977bcd9-2vs8h" Mar 13 15:46:26 crc kubenswrapper[4907]: I0313 15:46:26.438178 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/2d9fbab8-539c-48f6-8cae-3eda23300b28-horizon-secret-key\") pod \"horizon-c7977bcd9-2vs8h\" (UID: \"2d9fbab8-539c-48f6-8cae-3eda23300b28\") " pod="openstack/horizon-c7977bcd9-2vs8h" Mar 13 15:46:26 crc kubenswrapper[4907]: I0313 15:46:26.438196 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2d9fbab8-539c-48f6-8cae-3eda23300b28-logs\") pod \"horizon-c7977bcd9-2vs8h\" (UID: \"2d9fbab8-539c-48f6-8cae-3eda23300b28\") " pod="openstack/horizon-c7977bcd9-2vs8h" Mar 13 15:46:26 crc kubenswrapper[4907]: I0313 15:46:26.438666 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2d9fbab8-539c-48f6-8cae-3eda23300b28-logs\") pod \"horizon-c7977bcd9-2vs8h\" (UID: \"2d9fbab8-539c-48f6-8cae-3eda23300b28\") " pod="openstack/horizon-c7977bcd9-2vs8h" Mar 13 15:46:26 crc kubenswrapper[4907]: I0313 15:46:26.439751 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2d9fbab8-539c-48f6-8cae-3eda23300b28-config-data\") pod \"horizon-c7977bcd9-2vs8h\" (UID: \"2d9fbab8-539c-48f6-8cae-3eda23300b28\") " pod="openstack/horizon-c7977bcd9-2vs8h" Mar 13 15:46:26 crc kubenswrapper[4907]: I0313 15:46:26.439801 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2d9fbab8-539c-48f6-8cae-3eda23300b28-scripts\") pod \"horizon-c7977bcd9-2vs8h\" (UID: \"2d9fbab8-539c-48f6-8cae-3eda23300b28\") " pod="openstack/horizon-c7977bcd9-2vs8h" Mar 13 15:46:26 crc kubenswrapper[4907]: I0313 15:46:26.450727 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/2d9fbab8-539c-48f6-8cae-3eda23300b28-horizon-secret-key\") pod \"horizon-c7977bcd9-2vs8h\" (UID: \"2d9fbab8-539c-48f6-8cae-3eda23300b28\") " pod="openstack/horizon-c7977bcd9-2vs8h" Mar 13 15:46:26 crc kubenswrapper[4907]: I0313 15:46:26.470493 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bjhcq\" (UniqueName: \"kubernetes.io/projected/2d9fbab8-539c-48f6-8cae-3eda23300b28-kube-api-access-bjhcq\") pod \"horizon-c7977bcd9-2vs8h\" (UID: \"2d9fbab8-539c-48f6-8cae-3eda23300b28\") " pod="openstack/horizon-c7977bcd9-2vs8h" Mar 13 15:46:26 crc kubenswrapper[4907]: I0313 15:46:26.480257 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-769b54bfcc-kbtds"] Mar 13 15:46:26 crc kubenswrapper[4907]: I0313 15:46:26.481982 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-769b54bfcc-kbtds" Mar 13 15:46:26 crc kubenswrapper[4907]: I0313 15:46:26.520159 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-769b54bfcc-kbtds"] Mar 13 15:46:26 crc kubenswrapper[4907]: I0313 15:46:26.540233 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/099b49c3-666e-4bf5-9c15-32cf9c40e1e7-logs\") pod \"horizon-769b54bfcc-kbtds\" (UID: \"099b49c3-666e-4bf5-9c15-32cf9c40e1e7\") " pod="openstack/horizon-769b54bfcc-kbtds" Mar 13 15:46:26 crc kubenswrapper[4907]: I0313 15:46:26.540357 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/099b49c3-666e-4bf5-9c15-32cf9c40e1e7-config-data\") pod \"horizon-769b54bfcc-kbtds\" (UID: \"099b49c3-666e-4bf5-9c15-32cf9c40e1e7\") " pod="openstack/horizon-769b54bfcc-kbtds" Mar 13 15:46:26 crc kubenswrapper[4907]: I0313 15:46:26.540417 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-scsbb\" (UniqueName: \"kubernetes.io/projected/099b49c3-666e-4bf5-9c15-32cf9c40e1e7-kube-api-access-scsbb\") pod \"horizon-769b54bfcc-kbtds\" (UID: \"099b49c3-666e-4bf5-9c15-32cf9c40e1e7\") " pod="openstack/horizon-769b54bfcc-kbtds" Mar 13 15:46:26 crc kubenswrapper[4907]: I0313 15:46:26.540471 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/099b49c3-666e-4bf5-9c15-32cf9c40e1e7-horizon-secret-key\") pod \"horizon-769b54bfcc-kbtds\" (UID: \"099b49c3-666e-4bf5-9c15-32cf9c40e1e7\") " pod="openstack/horizon-769b54bfcc-kbtds" Mar 13 15:46:26 crc kubenswrapper[4907]: I0313 15:46:26.540518 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/099b49c3-666e-4bf5-9c15-32cf9c40e1e7-scripts\") pod \"horizon-769b54bfcc-kbtds\" (UID: \"099b49c3-666e-4bf5-9c15-32cf9c40e1e7\") " pod="openstack/horizon-769b54bfcc-kbtds" Mar 13 15:46:26 crc kubenswrapper[4907]: I0313 15:46:26.605324 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-c7977bcd9-2vs8h" Mar 13 15:46:26 crc kubenswrapper[4907]: I0313 15:46:26.642612 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-scsbb\" (UniqueName: \"kubernetes.io/projected/099b49c3-666e-4bf5-9c15-32cf9c40e1e7-kube-api-access-scsbb\") pod \"horizon-769b54bfcc-kbtds\" (UID: \"099b49c3-666e-4bf5-9c15-32cf9c40e1e7\") " pod="openstack/horizon-769b54bfcc-kbtds" Mar 13 15:46:26 crc kubenswrapper[4907]: I0313 15:46:26.642714 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/099b49c3-666e-4bf5-9c15-32cf9c40e1e7-horizon-secret-key\") pod \"horizon-769b54bfcc-kbtds\" (UID: \"099b49c3-666e-4bf5-9c15-32cf9c40e1e7\") " pod="openstack/horizon-769b54bfcc-kbtds" Mar 13 15:46:26 crc kubenswrapper[4907]: I0313 15:46:26.642774 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/099b49c3-666e-4bf5-9c15-32cf9c40e1e7-scripts\") pod \"horizon-769b54bfcc-kbtds\" (UID: \"099b49c3-666e-4bf5-9c15-32cf9c40e1e7\") " pod="openstack/horizon-769b54bfcc-kbtds" Mar 13 15:46:26 crc kubenswrapper[4907]: I0313 15:46:26.642817 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/099b49c3-666e-4bf5-9c15-32cf9c40e1e7-logs\") pod \"horizon-769b54bfcc-kbtds\" (UID: \"099b49c3-666e-4bf5-9c15-32cf9c40e1e7\") " pod="openstack/horizon-769b54bfcc-kbtds" Mar 13 15:46:26 crc kubenswrapper[4907]: I0313 15:46:26.642927 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/099b49c3-666e-4bf5-9c15-32cf9c40e1e7-config-data\") pod \"horizon-769b54bfcc-kbtds\" (UID: \"099b49c3-666e-4bf5-9c15-32cf9c40e1e7\") " pod="openstack/horizon-769b54bfcc-kbtds" Mar 13 15:46:26 crc kubenswrapper[4907]: I0313 15:46:26.643717 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/099b49c3-666e-4bf5-9c15-32cf9c40e1e7-scripts\") pod \"horizon-769b54bfcc-kbtds\" (UID: \"099b49c3-666e-4bf5-9c15-32cf9c40e1e7\") " pod="openstack/horizon-769b54bfcc-kbtds" Mar 13 15:46:26 crc kubenswrapper[4907]: I0313 15:46:26.644392 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/099b49c3-666e-4bf5-9c15-32cf9c40e1e7-config-data\") pod \"horizon-769b54bfcc-kbtds\" (UID: \"099b49c3-666e-4bf5-9c15-32cf9c40e1e7\") " pod="openstack/horizon-769b54bfcc-kbtds" Mar 13 15:46:26 crc kubenswrapper[4907]: I0313 15:46:26.644481 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/099b49c3-666e-4bf5-9c15-32cf9c40e1e7-logs\") pod \"horizon-769b54bfcc-kbtds\" (UID: \"099b49c3-666e-4bf5-9c15-32cf9c40e1e7\") " pod="openstack/horizon-769b54bfcc-kbtds" Mar 13 15:46:26 crc kubenswrapper[4907]: I0313 15:46:26.647183 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/099b49c3-666e-4bf5-9c15-32cf9c40e1e7-horizon-secret-key\") pod \"horizon-769b54bfcc-kbtds\" (UID: \"099b49c3-666e-4bf5-9c15-32cf9c40e1e7\") " pod="openstack/horizon-769b54bfcc-kbtds" Mar 13 15:46:26 crc kubenswrapper[4907]: I0313 15:46:26.670729 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-scsbb\" (UniqueName: \"kubernetes.io/projected/099b49c3-666e-4bf5-9c15-32cf9c40e1e7-kube-api-access-scsbb\") pod \"horizon-769b54bfcc-kbtds\" (UID: \"099b49c3-666e-4bf5-9c15-32cf9c40e1e7\") " pod="openstack/horizon-769b54bfcc-kbtds" Mar 13 15:46:26 crc kubenswrapper[4907]: I0313 15:46:26.906198 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-769b54bfcc-kbtds" Mar 13 15:46:27 crc kubenswrapper[4907]: I0313 15:46:27.078179 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-769b54bfcc-kbtds"] Mar 13 15:46:27 crc kubenswrapper[4907]: I0313 15:46:27.123293 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-655cbf768f-dm9zz"] Mar 13 15:46:27 crc kubenswrapper[4907]: I0313 15:46:27.124912 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-655cbf768f-dm9zz" Mar 13 15:46:27 crc kubenswrapper[4907]: I0313 15:46:27.146239 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-c7977bcd9-2vs8h"] Mar 13 15:46:27 crc kubenswrapper[4907]: I0313 15:46:27.151850 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/afd4001d-40a5-4a09-ab05-36cbd4d06eb4-horizon-secret-key\") pod \"horizon-655cbf768f-dm9zz\" (UID: \"afd4001d-40a5-4a09-ab05-36cbd4d06eb4\") " pod="openstack/horizon-655cbf768f-dm9zz" Mar 13 15:46:27 crc kubenswrapper[4907]: I0313 15:46:27.152192 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/afd4001d-40a5-4a09-ab05-36cbd4d06eb4-logs\") pod \"horizon-655cbf768f-dm9zz\" (UID: \"afd4001d-40a5-4a09-ab05-36cbd4d06eb4\") " pod="openstack/horizon-655cbf768f-dm9zz" Mar 13 15:46:27 crc kubenswrapper[4907]: I0313 15:46:27.152579 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ml6dh\" (UniqueName: \"kubernetes.io/projected/afd4001d-40a5-4a09-ab05-36cbd4d06eb4-kube-api-access-ml6dh\") pod \"horizon-655cbf768f-dm9zz\" (UID: \"afd4001d-40a5-4a09-ab05-36cbd4d06eb4\") " pod="openstack/horizon-655cbf768f-dm9zz" Mar 13 15:46:27 crc kubenswrapper[4907]: I0313 15:46:27.152931 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/afd4001d-40a5-4a09-ab05-36cbd4d06eb4-scripts\") pod \"horizon-655cbf768f-dm9zz\" (UID: \"afd4001d-40a5-4a09-ab05-36cbd4d06eb4\") " pod="openstack/horizon-655cbf768f-dm9zz" Mar 13 15:46:27 crc kubenswrapper[4907]: I0313 15:46:27.153037 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/afd4001d-40a5-4a09-ab05-36cbd4d06eb4-config-data\") pod \"horizon-655cbf768f-dm9zz\" (UID: \"afd4001d-40a5-4a09-ab05-36cbd4d06eb4\") " pod="openstack/horizon-655cbf768f-dm9zz" Mar 13 15:46:27 crc kubenswrapper[4907]: W0313 15:46:27.159258 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2d9fbab8_539c_48f6_8cae_3eda23300b28.slice/crio-941b229bb7806138fc500a8ef3b0ed37c2227cce03ef0b5260159ccb8082332d WatchSource:0}: Error finding container 941b229bb7806138fc500a8ef3b0ed37c2227cce03ef0b5260159ccb8082332d: Status 404 returned error can't find the container with id 941b229bb7806138fc500a8ef3b0ed37c2227cce03ef0b5260159ccb8082332d Mar 13 15:46:27 crc kubenswrapper[4907]: I0313 15:46:27.180000 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-655cbf768f-dm9zz"] Mar 13 15:46:27 crc kubenswrapper[4907]: I0313 15:46:27.255358 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/afd4001d-40a5-4a09-ab05-36cbd4d06eb4-scripts\") pod \"horizon-655cbf768f-dm9zz\" (UID: \"afd4001d-40a5-4a09-ab05-36cbd4d06eb4\") " pod="openstack/horizon-655cbf768f-dm9zz" Mar 13 15:46:27 crc kubenswrapper[4907]: I0313 15:46:27.255430 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/afd4001d-40a5-4a09-ab05-36cbd4d06eb4-config-data\") pod \"horizon-655cbf768f-dm9zz\" (UID: \"afd4001d-40a5-4a09-ab05-36cbd4d06eb4\") " pod="openstack/horizon-655cbf768f-dm9zz" Mar 13 15:46:27 crc kubenswrapper[4907]: I0313 15:46:27.255462 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/afd4001d-40a5-4a09-ab05-36cbd4d06eb4-horizon-secret-key\") pod \"horizon-655cbf768f-dm9zz\" (UID: \"afd4001d-40a5-4a09-ab05-36cbd4d06eb4\") " pod="openstack/horizon-655cbf768f-dm9zz" Mar 13 15:46:27 crc kubenswrapper[4907]: I0313 15:46:27.255521 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/afd4001d-40a5-4a09-ab05-36cbd4d06eb4-logs\") pod \"horizon-655cbf768f-dm9zz\" (UID: \"afd4001d-40a5-4a09-ab05-36cbd4d06eb4\") " pod="openstack/horizon-655cbf768f-dm9zz" Mar 13 15:46:27 crc kubenswrapper[4907]: I0313 15:46:27.255590 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ml6dh\" (UniqueName: \"kubernetes.io/projected/afd4001d-40a5-4a09-ab05-36cbd4d06eb4-kube-api-access-ml6dh\") pod \"horizon-655cbf768f-dm9zz\" (UID: \"afd4001d-40a5-4a09-ab05-36cbd4d06eb4\") " pod="openstack/horizon-655cbf768f-dm9zz" Mar 13 15:46:27 crc kubenswrapper[4907]: I0313 15:46:27.257141 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/afd4001d-40a5-4a09-ab05-36cbd4d06eb4-config-data\") pod \"horizon-655cbf768f-dm9zz\" (UID: \"afd4001d-40a5-4a09-ab05-36cbd4d06eb4\") " pod="openstack/horizon-655cbf768f-dm9zz" Mar 13 15:46:27 crc kubenswrapper[4907]: I0313 15:46:27.258285 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/afd4001d-40a5-4a09-ab05-36cbd4d06eb4-logs\") pod \"horizon-655cbf768f-dm9zz\" (UID: \"afd4001d-40a5-4a09-ab05-36cbd4d06eb4\") " pod="openstack/horizon-655cbf768f-dm9zz" Mar 13 15:46:27 crc kubenswrapper[4907]: I0313 15:46:27.258459 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/afd4001d-40a5-4a09-ab05-36cbd4d06eb4-scripts\") pod \"horizon-655cbf768f-dm9zz\" (UID: \"afd4001d-40a5-4a09-ab05-36cbd4d06eb4\") " pod="openstack/horizon-655cbf768f-dm9zz" Mar 13 15:46:27 crc kubenswrapper[4907]: I0313 15:46:27.262408 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/afd4001d-40a5-4a09-ab05-36cbd4d06eb4-horizon-secret-key\") pod \"horizon-655cbf768f-dm9zz\" (UID: \"afd4001d-40a5-4a09-ab05-36cbd4d06eb4\") " pod="openstack/horizon-655cbf768f-dm9zz" Mar 13 15:46:27 crc kubenswrapper[4907]: I0313 15:46:27.276235 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ml6dh\" (UniqueName: \"kubernetes.io/projected/afd4001d-40a5-4a09-ab05-36cbd4d06eb4-kube-api-access-ml6dh\") pod \"horizon-655cbf768f-dm9zz\" (UID: \"afd4001d-40a5-4a09-ab05-36cbd4d06eb4\") " pod="openstack/horizon-655cbf768f-dm9zz" Mar 13 15:46:27 crc kubenswrapper[4907]: I0313 15:46:27.336044 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-c7977bcd9-2vs8h" event={"ID":"2d9fbab8-539c-48f6-8cae-3eda23300b28","Type":"ContainerStarted","Data":"941b229bb7806138fc500a8ef3b0ed37c2227cce03ef0b5260159ccb8082332d"} Mar 13 15:46:27 crc kubenswrapper[4907]: I0313 15:46:27.338730 4907 generic.go:334] "Generic (PLEG): container finished" podID="86794caf-bf6d-4394-bc1c-3c516bdbada8" containerID="56ff9c5433a9cbee0386dea3d88b4e88f870fa5dcf67d5221e12245f174e8f85" exitCode=143 Mar 13 15:46:27 crc kubenswrapper[4907]: I0313 15:46:27.338836 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"86794caf-bf6d-4394-bc1c-3c516bdbada8","Type":"ContainerDied","Data":"56ff9c5433a9cbee0386dea3d88b4e88f870fa5dcf67d5221e12245f174e8f85"} Mar 13 15:46:27 crc kubenswrapper[4907]: I0313 15:46:27.341556 4907 generic.go:334] "Generic (PLEG): container finished" podID="096d3713-ea4f-43d3-bfb9-9170e8958ed0" containerID="02ef5ab885e1e984508608add8cbc5f616e0cf1ed8d42cf7ded2960a6fa681f8" exitCode=143 Mar 13 15:46:27 crc kubenswrapper[4907]: I0313 15:46:27.341585 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"096d3713-ea4f-43d3-bfb9-9170e8958ed0","Type":"ContainerDied","Data":"02ef5ab885e1e984508608add8cbc5f616e0cf1ed8d42cf7ded2960a6fa681f8"} Mar 13 15:46:27 crc kubenswrapper[4907]: I0313 15:46:27.486452 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-655cbf768f-dm9zz" Mar 13 15:46:27 crc kubenswrapper[4907]: I0313 15:46:27.487729 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-769b54bfcc-kbtds"] Mar 13 15:46:27 crc kubenswrapper[4907]: W0313 15:46:27.489103 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod099b49c3_666e_4bf5_9c15_32cf9c40e1e7.slice/crio-e4caf9d5f22ffb193c93c87c06b4e64f7a29f8bf0582a1a8b7f47762f5ba55eb WatchSource:0}: Error finding container e4caf9d5f22ffb193c93c87c06b4e64f7a29f8bf0582a1a8b7f47762f5ba55eb: Status 404 returned error can't find the container with id e4caf9d5f22ffb193c93c87c06b4e64f7a29f8bf0582a1a8b7f47762f5ba55eb Mar 13 15:46:27 crc kubenswrapper[4907]: W0313 15:46:27.946233 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podafd4001d_40a5_4a09_ab05_36cbd4d06eb4.slice/crio-8eed0ad989f3d5a163602bd4e967ac325bbb242d625bd6a2d8dac4fb6b57f365 WatchSource:0}: Error finding container 8eed0ad989f3d5a163602bd4e967ac325bbb242d625bd6a2d8dac4fb6b57f365: Status 404 returned error can't find the container with id 8eed0ad989f3d5a163602bd4e967ac325bbb242d625bd6a2d8dac4fb6b57f365 Mar 13 15:46:27 crc kubenswrapper[4907]: I0313 15:46:27.948254 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-655cbf768f-dm9zz"] Mar 13 15:46:28 crc kubenswrapper[4907]: I0313 15:46:28.355980 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-769b54bfcc-kbtds" event={"ID":"099b49c3-666e-4bf5-9c15-32cf9c40e1e7","Type":"ContainerStarted","Data":"e4caf9d5f22ffb193c93c87c06b4e64f7a29f8bf0582a1a8b7f47762f5ba55eb"} Mar 13 15:46:28 crc kubenswrapper[4907]: I0313 15:46:28.358001 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-655cbf768f-dm9zz" event={"ID":"afd4001d-40a5-4a09-ab05-36cbd4d06eb4","Type":"ContainerStarted","Data":"8eed0ad989f3d5a163602bd4e967ac325bbb242d625bd6a2d8dac4fb6b57f365"} Mar 13 15:46:29 crc kubenswrapper[4907]: I0313 15:46:29.740267 4907 scope.go:117] "RemoveContainer" containerID="3f861dd2c3ff8fa5314ce9c35d45445320ad82d6381b4138837ad16c289319f5" Mar 13 15:46:30 crc kubenswrapper[4907]: I0313 15:46:30.382417 4907 generic.go:334] "Generic (PLEG): container finished" podID="096d3713-ea4f-43d3-bfb9-9170e8958ed0" containerID="27209680abb804b6c48ccdaa33b44ab4f54951d1df1401058af333f852827d0f" exitCode=0 Mar 13 15:46:30 crc kubenswrapper[4907]: I0313 15:46:30.382498 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"096d3713-ea4f-43d3-bfb9-9170e8958ed0","Type":"ContainerDied","Data":"27209680abb804b6c48ccdaa33b44ab4f54951d1df1401058af333f852827d0f"} Mar 13 15:46:30 crc kubenswrapper[4907]: I0313 15:46:30.387216 4907 generic.go:334] "Generic (PLEG): container finished" podID="86794caf-bf6d-4394-bc1c-3c516bdbada8" containerID="d5e6285fe8e2b22da125d5022a7924d7086ba68515ab90ca41e35d0607040a03" exitCode=0 Mar 13 15:46:30 crc kubenswrapper[4907]: I0313 15:46:30.387253 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"86794caf-bf6d-4394-bc1c-3c516bdbada8","Type":"ContainerDied","Data":"d5e6285fe8e2b22da125d5022a7924d7086ba68515ab90ca41e35d0607040a03"} Mar 13 15:46:34 crc kubenswrapper[4907]: I0313 15:46:34.649367 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 13 15:46:34 crc kubenswrapper[4907]: I0313 15:46:34.655225 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 13 15:46:34 crc kubenswrapper[4907]: I0313 15:46:34.697198 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/096d3713-ea4f-43d3-bfb9-9170e8958ed0-logs\") pod \"096d3713-ea4f-43d3-bfb9-9170e8958ed0\" (UID: \"096d3713-ea4f-43d3-bfb9-9170e8958ed0\") " Mar 13 15:46:34 crc kubenswrapper[4907]: I0313 15:46:34.697288 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h7zl5\" (UniqueName: \"kubernetes.io/projected/096d3713-ea4f-43d3-bfb9-9170e8958ed0-kube-api-access-h7zl5\") pod \"096d3713-ea4f-43d3-bfb9-9170e8958ed0\" (UID: \"096d3713-ea4f-43d3-bfb9-9170e8958ed0\") " Mar 13 15:46:34 crc kubenswrapper[4907]: I0313 15:46:34.697314 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/096d3713-ea4f-43d3-bfb9-9170e8958ed0-combined-ca-bundle\") pod \"096d3713-ea4f-43d3-bfb9-9170e8958ed0\" (UID: \"096d3713-ea4f-43d3-bfb9-9170e8958ed0\") " Mar 13 15:46:34 crc kubenswrapper[4907]: I0313 15:46:34.697347 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/096d3713-ea4f-43d3-bfb9-9170e8958ed0-scripts\") pod \"096d3713-ea4f-43d3-bfb9-9170e8958ed0\" (UID: \"096d3713-ea4f-43d3-bfb9-9170e8958ed0\") " Mar 13 15:46:34 crc kubenswrapper[4907]: I0313 15:46:34.697429 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/86794caf-bf6d-4394-bc1c-3c516bdbada8-logs\") pod \"86794caf-bf6d-4394-bc1c-3c516bdbada8\" (UID: \"86794caf-bf6d-4394-bc1c-3c516bdbada8\") " Mar 13 15:46:34 crc kubenswrapper[4907]: I0313 15:46:34.697487 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86794caf-bf6d-4394-bc1c-3c516bdbada8-combined-ca-bundle\") pod \"86794caf-bf6d-4394-bc1c-3c516bdbada8\" (UID: \"86794caf-bf6d-4394-bc1c-3c516bdbada8\") " Mar 13 15:46:34 crc kubenswrapper[4907]: I0313 15:46:34.697566 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/096d3713-ea4f-43d3-bfb9-9170e8958ed0-config-data\") pod \"096d3713-ea4f-43d3-bfb9-9170e8958ed0\" (UID: \"096d3713-ea4f-43d3-bfb9-9170e8958ed0\") " Mar 13 15:46:34 crc kubenswrapper[4907]: I0313 15:46:34.697600 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/86794caf-bf6d-4394-bc1c-3c516bdbada8-httpd-run\") pod \"86794caf-bf6d-4394-bc1c-3c516bdbada8\" (UID: \"86794caf-bf6d-4394-bc1c-3c516bdbada8\") " Mar 13 15:46:34 crc kubenswrapper[4907]: I0313 15:46:34.697682 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/096d3713-ea4f-43d3-bfb9-9170e8958ed0-httpd-run\") pod \"096d3713-ea4f-43d3-bfb9-9170e8958ed0\" (UID: \"096d3713-ea4f-43d3-bfb9-9170e8958ed0\") " Mar 13 15:46:34 crc kubenswrapper[4907]: I0313 15:46:34.697725 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/86794caf-bf6d-4394-bc1c-3c516bdbada8-scripts\") pod \"86794caf-bf6d-4394-bc1c-3c516bdbada8\" (UID: \"86794caf-bf6d-4394-bc1c-3c516bdbada8\") " Mar 13 15:46:34 crc kubenswrapper[4907]: I0313 15:46:34.697769 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/86794caf-bf6d-4394-bc1c-3c516bdbada8-config-data\") pod \"86794caf-bf6d-4394-bc1c-3c516bdbada8\" (UID: \"86794caf-bf6d-4394-bc1c-3c516bdbada8\") " Mar 13 15:46:34 crc kubenswrapper[4907]: I0313 15:46:34.697817 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/096d3713-ea4f-43d3-bfb9-9170e8958ed0-ceph\") pod \"096d3713-ea4f-43d3-bfb9-9170e8958ed0\" (UID: \"096d3713-ea4f-43d3-bfb9-9170e8958ed0\") " Mar 13 15:46:34 crc kubenswrapper[4907]: I0313 15:46:34.697870 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4srpn\" (UniqueName: \"kubernetes.io/projected/86794caf-bf6d-4394-bc1c-3c516bdbada8-kube-api-access-4srpn\") pod \"86794caf-bf6d-4394-bc1c-3c516bdbada8\" (UID: \"86794caf-bf6d-4394-bc1c-3c516bdbada8\") " Mar 13 15:46:34 crc kubenswrapper[4907]: I0313 15:46:34.697936 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/86794caf-bf6d-4394-bc1c-3c516bdbada8-ceph\") pod \"86794caf-bf6d-4394-bc1c-3c516bdbada8\" (UID: \"86794caf-bf6d-4394-bc1c-3c516bdbada8\") " Mar 13 15:46:34 crc kubenswrapper[4907]: I0313 15:46:34.708287 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/096d3713-ea4f-43d3-bfb9-9170e8958ed0-logs" (OuterVolumeSpecName: "logs") pod "096d3713-ea4f-43d3-bfb9-9170e8958ed0" (UID: "096d3713-ea4f-43d3-bfb9-9170e8958ed0"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:46:34 crc kubenswrapper[4907]: I0313 15:46:34.708823 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/86794caf-bf6d-4394-bc1c-3c516bdbada8-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "86794caf-bf6d-4394-bc1c-3c516bdbada8" (UID: "86794caf-bf6d-4394-bc1c-3c516bdbada8"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:46:34 crc kubenswrapper[4907]: I0313 15:46:34.711262 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86794caf-bf6d-4394-bc1c-3c516bdbada8-scripts" (OuterVolumeSpecName: "scripts") pod "86794caf-bf6d-4394-bc1c-3c516bdbada8" (UID: "86794caf-bf6d-4394-bc1c-3c516bdbada8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:46:34 crc kubenswrapper[4907]: I0313 15:46:34.711787 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/096d3713-ea4f-43d3-bfb9-9170e8958ed0-scripts" (OuterVolumeSpecName: "scripts") pod "096d3713-ea4f-43d3-bfb9-9170e8958ed0" (UID: "096d3713-ea4f-43d3-bfb9-9170e8958ed0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:46:34 crc kubenswrapper[4907]: I0313 15:46:34.711859 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/86794caf-bf6d-4394-bc1c-3c516bdbada8-logs" (OuterVolumeSpecName: "logs") pod "86794caf-bf6d-4394-bc1c-3c516bdbada8" (UID: "86794caf-bf6d-4394-bc1c-3c516bdbada8"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:46:34 crc kubenswrapper[4907]: I0313 15:46:34.712152 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/86794caf-bf6d-4394-bc1c-3c516bdbada8-ceph" (OuterVolumeSpecName: "ceph") pod "86794caf-bf6d-4394-bc1c-3c516bdbada8" (UID: "86794caf-bf6d-4394-bc1c-3c516bdbada8"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:46:34 crc kubenswrapper[4907]: I0313 15:46:34.718723 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/86794caf-bf6d-4394-bc1c-3c516bdbada8-kube-api-access-4srpn" (OuterVolumeSpecName: "kube-api-access-4srpn") pod "86794caf-bf6d-4394-bc1c-3c516bdbada8" (UID: "86794caf-bf6d-4394-bc1c-3c516bdbada8"). InnerVolumeSpecName "kube-api-access-4srpn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:46:34 crc kubenswrapper[4907]: I0313 15:46:34.719302 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/096d3713-ea4f-43d3-bfb9-9170e8958ed0-ceph" (OuterVolumeSpecName: "ceph") pod "096d3713-ea4f-43d3-bfb9-9170e8958ed0" (UID: "096d3713-ea4f-43d3-bfb9-9170e8958ed0"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:46:34 crc kubenswrapper[4907]: I0313 15:46:34.723431 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/096d3713-ea4f-43d3-bfb9-9170e8958ed0-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "096d3713-ea4f-43d3-bfb9-9170e8958ed0" (UID: "096d3713-ea4f-43d3-bfb9-9170e8958ed0"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:46:34 crc kubenswrapper[4907]: I0313 15:46:34.729601 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/096d3713-ea4f-43d3-bfb9-9170e8958ed0-kube-api-access-h7zl5" (OuterVolumeSpecName: "kube-api-access-h7zl5") pod "096d3713-ea4f-43d3-bfb9-9170e8958ed0" (UID: "096d3713-ea4f-43d3-bfb9-9170e8958ed0"). InnerVolumeSpecName "kube-api-access-h7zl5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:46:34 crc kubenswrapper[4907]: I0313 15:46:34.737316 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86794caf-bf6d-4394-bc1c-3c516bdbada8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "86794caf-bf6d-4394-bc1c-3c516bdbada8" (UID: "86794caf-bf6d-4394-bc1c-3c516bdbada8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:46:34 crc kubenswrapper[4907]: I0313 15:46:34.750450 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/096d3713-ea4f-43d3-bfb9-9170e8958ed0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "096d3713-ea4f-43d3-bfb9-9170e8958ed0" (UID: "096d3713-ea4f-43d3-bfb9-9170e8958ed0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:46:34 crc kubenswrapper[4907]: I0313 15:46:34.776996 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86794caf-bf6d-4394-bc1c-3c516bdbada8-config-data" (OuterVolumeSpecName: "config-data") pod "86794caf-bf6d-4394-bc1c-3c516bdbada8" (UID: "86794caf-bf6d-4394-bc1c-3c516bdbada8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:46:34 crc kubenswrapper[4907]: I0313 15:46:34.785630 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/096d3713-ea4f-43d3-bfb9-9170e8958ed0-config-data" (OuterVolumeSpecName: "config-data") pod "096d3713-ea4f-43d3-bfb9-9170e8958ed0" (UID: "096d3713-ea4f-43d3-bfb9-9170e8958ed0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:46:34 crc kubenswrapper[4907]: I0313 15:46:34.799729 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h7zl5\" (UniqueName: \"kubernetes.io/projected/096d3713-ea4f-43d3-bfb9-9170e8958ed0-kube-api-access-h7zl5\") on node \"crc\" DevicePath \"\"" Mar 13 15:46:34 crc kubenswrapper[4907]: I0313 15:46:34.799756 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/096d3713-ea4f-43d3-bfb9-9170e8958ed0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 15:46:34 crc kubenswrapper[4907]: I0313 15:46:34.799765 4907 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/096d3713-ea4f-43d3-bfb9-9170e8958ed0-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 15:46:34 crc kubenswrapper[4907]: I0313 15:46:34.799773 4907 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/86794caf-bf6d-4394-bc1c-3c516bdbada8-logs\") on node \"crc\" DevicePath \"\"" Mar 13 15:46:34 crc kubenswrapper[4907]: I0313 15:46:34.799783 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86794caf-bf6d-4394-bc1c-3c516bdbada8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 15:46:34 crc kubenswrapper[4907]: I0313 15:46:34.799791 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/096d3713-ea4f-43d3-bfb9-9170e8958ed0-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 15:46:34 crc kubenswrapper[4907]: I0313 15:46:34.799801 4907 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/86794caf-bf6d-4394-bc1c-3c516bdbada8-httpd-run\") on node \"crc\" DevicePath \"\"" Mar 13 15:46:34 crc kubenswrapper[4907]: I0313 15:46:34.799809 4907 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/096d3713-ea4f-43d3-bfb9-9170e8958ed0-httpd-run\") on node \"crc\" DevicePath \"\"" Mar 13 15:46:34 crc kubenswrapper[4907]: I0313 15:46:34.799817 4907 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/86794caf-bf6d-4394-bc1c-3c516bdbada8-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 15:46:34 crc kubenswrapper[4907]: I0313 15:46:34.799825 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/86794caf-bf6d-4394-bc1c-3c516bdbada8-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 15:46:34 crc kubenswrapper[4907]: I0313 15:46:34.799833 4907 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/096d3713-ea4f-43d3-bfb9-9170e8958ed0-ceph\") on node \"crc\" DevicePath \"\"" Mar 13 15:46:34 crc kubenswrapper[4907]: I0313 15:46:34.799842 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4srpn\" (UniqueName: \"kubernetes.io/projected/86794caf-bf6d-4394-bc1c-3c516bdbada8-kube-api-access-4srpn\") on node \"crc\" DevicePath \"\"" Mar 13 15:46:34 crc kubenswrapper[4907]: I0313 15:46:34.799850 4907 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/86794caf-bf6d-4394-bc1c-3c516bdbada8-ceph\") on node \"crc\" DevicePath \"\"" Mar 13 15:46:34 crc kubenswrapper[4907]: I0313 15:46:34.799858 4907 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/096d3713-ea4f-43d3-bfb9-9170e8958ed0-logs\") on node \"crc\" DevicePath \"\"" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.434964 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-c7977bcd9-2vs8h" event={"ID":"2d9fbab8-539c-48f6-8cae-3eda23300b28","Type":"ContainerStarted","Data":"64ab189d02c2d9e2e9123ec6de24b5d5f7db40d7ceaa4b16356eefe21fdc102c"} Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.435520 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-c7977bcd9-2vs8h" event={"ID":"2d9fbab8-539c-48f6-8cae-3eda23300b28","Type":"ContainerStarted","Data":"d405f8c27cb2c1dbfd5ce42b31129e474a50d6e8ddffc33a12dfa64551646e29"} Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.438229 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"86794caf-bf6d-4394-bc1c-3c516bdbada8","Type":"ContainerDied","Data":"e6d5e941f6575995d025c2bf565c8da48222d14bbb2fbb1710f7cd79e1394fae"} Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.438254 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.438328 4907 scope.go:117] "RemoveContainer" containerID="d5e6285fe8e2b22da125d5022a7924d7086ba68515ab90ca41e35d0607040a03" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.441961 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"096d3713-ea4f-43d3-bfb9-9170e8958ed0","Type":"ContainerDied","Data":"13bb6e3535193b961d70131bf3aebbe0d658a8a82ddb3d84795b101e8ae9b66a"} Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.442139 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.451546 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-769b54bfcc-kbtds" event={"ID":"099b49c3-666e-4bf5-9c15-32cf9c40e1e7","Type":"ContainerStarted","Data":"862291f2b2864cad1d39537f26d6a5251dddefb5e909a977e483fcdb34432151"} Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.451598 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-769b54bfcc-kbtds" event={"ID":"099b49c3-666e-4bf5-9c15-32cf9c40e1e7","Type":"ContainerStarted","Data":"fcce68c0980dbae069e759717a0e2db31c52e1f7a61e3feb3e5debebb904b5dd"} Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.451758 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-769b54bfcc-kbtds" podUID="099b49c3-666e-4bf5-9c15-32cf9c40e1e7" containerName="horizon-log" containerID="cri-o://fcce68c0980dbae069e759717a0e2db31c52e1f7a61e3feb3e5debebb904b5dd" gracePeriod=30 Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.452105 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-769b54bfcc-kbtds" podUID="099b49c3-666e-4bf5-9c15-32cf9c40e1e7" containerName="horizon" containerID="cri-o://862291f2b2864cad1d39537f26d6a5251dddefb5e909a977e483fcdb34432151" gracePeriod=30 Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.457832 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-655cbf768f-dm9zz" event={"ID":"afd4001d-40a5-4a09-ab05-36cbd4d06eb4","Type":"ContainerStarted","Data":"b42f8e33dfb6318b768e8c6b32af48621974ee97af5dab3f3e4bc582bbfc2931"} Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.457865 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-655cbf768f-dm9zz" event={"ID":"afd4001d-40a5-4a09-ab05-36cbd4d06eb4","Type":"ContainerStarted","Data":"4d9d087b4ad0675d46c48144b039251dc336275fe3320abbe51816f548542e9f"} Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.473405 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-c7977bcd9-2vs8h" podStartSLOduration=1.85554039 podStartE2EDuration="9.473387163s" podCreationTimestamp="2026-03-13 15:46:26 +0000 UTC" firstStartedPulling="2026-03-13 15:46:27.1744045 +0000 UTC m=+6086.074192189" lastFinishedPulling="2026-03-13 15:46:34.792251273 +0000 UTC m=+6093.692038962" observedRunningTime="2026-03-13 15:46:35.452509902 +0000 UTC m=+6094.352297591" watchObservedRunningTime="2026-03-13 15:46:35.473387163 +0000 UTC m=+6094.373174852" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.480274 4907 scope.go:117] "RemoveContainer" containerID="56ff9c5433a9cbee0386dea3d88b4e88f870fa5dcf67d5221e12245f174e8f85" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.491085 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-769b54bfcc-kbtds" podStartSLOduration=2.149175853 podStartE2EDuration="9.491046555s" podCreationTimestamp="2026-03-13 15:46:26 +0000 UTC" firstStartedPulling="2026-03-13 15:46:27.504369156 +0000 UTC m=+6086.404156845" lastFinishedPulling="2026-03-13 15:46:34.846239858 +0000 UTC m=+6093.746027547" observedRunningTime="2026-03-13 15:46:35.477732611 +0000 UTC m=+6094.377520340" watchObservedRunningTime="2026-03-13 15:46:35.491046555 +0000 UTC m=+6094.390834244" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.504671 4907 scope.go:117] "RemoveContainer" containerID="27209680abb804b6c48ccdaa33b44ab4f54951d1df1401058af333f852827d0f" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.504821 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.514522 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.533009 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 13 15:46:35 crc kubenswrapper[4907]: E0313 15:46:35.533967 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86794caf-bf6d-4394-bc1c-3c516bdbada8" containerName="glance-log" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.534113 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="86794caf-bf6d-4394-bc1c-3c516bdbada8" containerName="glance-log" Mar 13 15:46:35 crc kubenswrapper[4907]: E0313 15:46:35.534227 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="096d3713-ea4f-43d3-bfb9-9170e8958ed0" containerName="glance-httpd" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.534309 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="096d3713-ea4f-43d3-bfb9-9170e8958ed0" containerName="glance-httpd" Mar 13 15:46:35 crc kubenswrapper[4907]: E0313 15:46:35.534418 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="096d3713-ea4f-43d3-bfb9-9170e8958ed0" containerName="glance-log" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.534517 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="096d3713-ea4f-43d3-bfb9-9170e8958ed0" containerName="glance-log" Mar 13 15:46:35 crc kubenswrapper[4907]: E0313 15:46:35.534620 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86794caf-bf6d-4394-bc1c-3c516bdbada8" containerName="glance-httpd" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.534695 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="86794caf-bf6d-4394-bc1c-3c516bdbada8" containerName="glance-httpd" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.535072 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="86794caf-bf6d-4394-bc1c-3c516bdbada8" containerName="glance-log" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.535309 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="86794caf-bf6d-4394-bc1c-3c516bdbada8" containerName="glance-httpd" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.535419 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="096d3713-ea4f-43d3-bfb9-9170e8958ed0" containerName="glance-log" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.535499 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="096d3713-ea4f-43d3-bfb9-9170e8958ed0" containerName="glance-httpd" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.536002 4907 scope.go:117] "RemoveContainer" containerID="02ef5ab885e1e984508608add8cbc5f616e0cf1ed8d42cf7ded2960a6fa681f8" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.537249 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.537284 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-655cbf768f-dm9zz" podStartSLOduration=1.738425962 podStartE2EDuration="8.537263538s" podCreationTimestamp="2026-03-13 15:46:27 +0000 UTC" firstStartedPulling="2026-03-13 15:46:27.951089811 +0000 UTC m=+6086.850877500" lastFinishedPulling="2026-03-13 15:46:34.749927397 +0000 UTC m=+6093.649715076" observedRunningTime="2026-03-13 15:46:35.522764982 +0000 UTC m=+6094.422552671" watchObservedRunningTime="2026-03-13 15:46:35.537263538 +0000 UTC m=+6094.437051227" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.551839 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.552089 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.552547 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-5c589" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.622010 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qnn54\" (UniqueName: \"kubernetes.io/projected/70b550c7-5c28-416d-83c3-1ba63ecd5abd-kube-api-access-qnn54\") pod \"glance-default-internal-api-0\" (UID: \"70b550c7-5c28-416d-83c3-1ba63ecd5abd\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.623001 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/70b550c7-5c28-416d-83c3-1ba63ecd5abd-logs\") pod \"glance-default-internal-api-0\" (UID: \"70b550c7-5c28-416d-83c3-1ba63ecd5abd\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.623124 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70b550c7-5c28-416d-83c3-1ba63ecd5abd-config-data\") pod \"glance-default-internal-api-0\" (UID: \"70b550c7-5c28-416d-83c3-1ba63ecd5abd\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.623431 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/70b550c7-5c28-416d-83c3-1ba63ecd5abd-scripts\") pod \"glance-default-internal-api-0\" (UID: \"70b550c7-5c28-416d-83c3-1ba63ecd5abd\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.623562 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/70b550c7-5c28-416d-83c3-1ba63ecd5abd-ceph\") pod \"glance-default-internal-api-0\" (UID: \"70b550c7-5c28-416d-83c3-1ba63ecd5abd\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.623675 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70b550c7-5c28-416d-83c3-1ba63ecd5abd-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"70b550c7-5c28-416d-83c3-1ba63ecd5abd\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.623770 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/70b550c7-5c28-416d-83c3-1ba63ecd5abd-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"70b550c7-5c28-416d-83c3-1ba63ecd5abd\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.653382 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.664559 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.678596 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.695382 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.697495 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.703001 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.726011 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.727466 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qnn54\" (UniqueName: \"kubernetes.io/projected/70b550c7-5c28-416d-83c3-1ba63ecd5abd-kube-api-access-qnn54\") pod \"glance-default-internal-api-0\" (UID: \"70b550c7-5c28-416d-83c3-1ba63ecd5abd\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.727547 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/95435ffc-5322-4bd8-9558-421f953096eb-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"95435ffc-5322-4bd8-9558-421f953096eb\") " pod="openstack/glance-default-external-api-0" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.727624 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95435ffc-5322-4bd8-9558-421f953096eb-config-data\") pod \"glance-default-external-api-0\" (UID: \"95435ffc-5322-4bd8-9558-421f953096eb\") " pod="openstack/glance-default-external-api-0" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.727659 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/70b550c7-5c28-416d-83c3-1ba63ecd5abd-logs\") pod \"glance-default-internal-api-0\" (UID: \"70b550c7-5c28-416d-83c3-1ba63ecd5abd\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.727691 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70b550c7-5c28-416d-83c3-1ba63ecd5abd-config-data\") pod \"glance-default-internal-api-0\" (UID: \"70b550c7-5c28-416d-83c3-1ba63ecd5abd\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.727712 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/95435ffc-5322-4bd8-9558-421f953096eb-scripts\") pod \"glance-default-external-api-0\" (UID: \"95435ffc-5322-4bd8-9558-421f953096eb\") " pod="openstack/glance-default-external-api-0" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.728444 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/70b550c7-5c28-416d-83c3-1ba63ecd5abd-logs\") pod \"glance-default-internal-api-0\" (UID: \"70b550c7-5c28-416d-83c3-1ba63ecd5abd\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.727782 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/70b550c7-5c28-416d-83c3-1ba63ecd5abd-scripts\") pod \"glance-default-internal-api-0\" (UID: \"70b550c7-5c28-416d-83c3-1ba63ecd5abd\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.728565 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/95435ffc-5322-4bd8-9558-421f953096eb-ceph\") pod \"glance-default-external-api-0\" (UID: \"95435ffc-5322-4bd8-9558-421f953096eb\") " pod="openstack/glance-default-external-api-0" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.728595 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95435ffc-5322-4bd8-9558-421f953096eb-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"95435ffc-5322-4bd8-9558-421f953096eb\") " pod="openstack/glance-default-external-api-0" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.728621 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/70b550c7-5c28-416d-83c3-1ba63ecd5abd-ceph\") pod \"glance-default-internal-api-0\" (UID: \"70b550c7-5c28-416d-83c3-1ba63ecd5abd\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.728659 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70b550c7-5c28-416d-83c3-1ba63ecd5abd-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"70b550c7-5c28-416d-83c3-1ba63ecd5abd\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.728688 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/70b550c7-5c28-416d-83c3-1ba63ecd5abd-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"70b550c7-5c28-416d-83c3-1ba63ecd5abd\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.728711 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mvfdw\" (UniqueName: \"kubernetes.io/projected/95435ffc-5322-4bd8-9558-421f953096eb-kube-api-access-mvfdw\") pod \"glance-default-external-api-0\" (UID: \"95435ffc-5322-4bd8-9558-421f953096eb\") " pod="openstack/glance-default-external-api-0" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.728753 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/95435ffc-5322-4bd8-9558-421f953096eb-logs\") pod \"glance-default-external-api-0\" (UID: \"95435ffc-5322-4bd8-9558-421f953096eb\") " pod="openstack/glance-default-external-api-0" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.729290 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/70b550c7-5c28-416d-83c3-1ba63ecd5abd-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"70b550c7-5c28-416d-83c3-1ba63ecd5abd\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.734502 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/70b550c7-5c28-416d-83c3-1ba63ecd5abd-ceph\") pod \"glance-default-internal-api-0\" (UID: \"70b550c7-5c28-416d-83c3-1ba63ecd5abd\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.735795 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/70b550c7-5c28-416d-83c3-1ba63ecd5abd-scripts\") pod \"glance-default-internal-api-0\" (UID: \"70b550c7-5c28-416d-83c3-1ba63ecd5abd\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.745931 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70b550c7-5c28-416d-83c3-1ba63ecd5abd-config-data\") pod \"glance-default-internal-api-0\" (UID: \"70b550c7-5c28-416d-83c3-1ba63ecd5abd\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.747610 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70b550c7-5c28-416d-83c3-1ba63ecd5abd-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"70b550c7-5c28-416d-83c3-1ba63ecd5abd\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.751014 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qnn54\" (UniqueName: \"kubernetes.io/projected/70b550c7-5c28-416d-83c3-1ba63ecd5abd-kube-api-access-qnn54\") pod \"glance-default-internal-api-0\" (UID: \"70b550c7-5c28-416d-83c3-1ba63ecd5abd\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.794205 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="096d3713-ea4f-43d3-bfb9-9170e8958ed0" path="/var/lib/kubelet/pods/096d3713-ea4f-43d3-bfb9-9170e8958ed0/volumes" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.795318 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="86794caf-bf6d-4394-bc1c-3c516bdbada8" path="/var/lib/kubelet/pods/86794caf-bf6d-4394-bc1c-3c516bdbada8/volumes" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.830577 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95435ffc-5322-4bd8-9558-421f953096eb-config-data\") pod \"glance-default-external-api-0\" (UID: \"95435ffc-5322-4bd8-9558-421f953096eb\") " pod="openstack/glance-default-external-api-0" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.830645 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/95435ffc-5322-4bd8-9558-421f953096eb-scripts\") pod \"glance-default-external-api-0\" (UID: \"95435ffc-5322-4bd8-9558-421f953096eb\") " pod="openstack/glance-default-external-api-0" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.830732 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/95435ffc-5322-4bd8-9558-421f953096eb-ceph\") pod \"glance-default-external-api-0\" (UID: \"95435ffc-5322-4bd8-9558-421f953096eb\") " pod="openstack/glance-default-external-api-0" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.830750 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95435ffc-5322-4bd8-9558-421f953096eb-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"95435ffc-5322-4bd8-9558-421f953096eb\") " pod="openstack/glance-default-external-api-0" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.830793 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mvfdw\" (UniqueName: \"kubernetes.io/projected/95435ffc-5322-4bd8-9558-421f953096eb-kube-api-access-mvfdw\") pod \"glance-default-external-api-0\" (UID: \"95435ffc-5322-4bd8-9558-421f953096eb\") " pod="openstack/glance-default-external-api-0" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.830822 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/95435ffc-5322-4bd8-9558-421f953096eb-logs\") pod \"glance-default-external-api-0\" (UID: \"95435ffc-5322-4bd8-9558-421f953096eb\") " pod="openstack/glance-default-external-api-0" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.830871 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/95435ffc-5322-4bd8-9558-421f953096eb-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"95435ffc-5322-4bd8-9558-421f953096eb\") " pod="openstack/glance-default-external-api-0" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.831373 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/95435ffc-5322-4bd8-9558-421f953096eb-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"95435ffc-5322-4bd8-9558-421f953096eb\") " pod="openstack/glance-default-external-api-0" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.835630 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95435ffc-5322-4bd8-9558-421f953096eb-config-data\") pod \"glance-default-external-api-0\" (UID: \"95435ffc-5322-4bd8-9558-421f953096eb\") " pod="openstack/glance-default-external-api-0" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.836528 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/95435ffc-5322-4bd8-9558-421f953096eb-logs\") pod \"glance-default-external-api-0\" (UID: \"95435ffc-5322-4bd8-9558-421f953096eb\") " pod="openstack/glance-default-external-api-0" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.837161 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/95435ffc-5322-4bd8-9558-421f953096eb-scripts\") pod \"glance-default-external-api-0\" (UID: \"95435ffc-5322-4bd8-9558-421f953096eb\") " pod="openstack/glance-default-external-api-0" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.837373 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/95435ffc-5322-4bd8-9558-421f953096eb-ceph\") pod \"glance-default-external-api-0\" (UID: \"95435ffc-5322-4bd8-9558-421f953096eb\") " pod="openstack/glance-default-external-api-0" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.838314 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95435ffc-5322-4bd8-9558-421f953096eb-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"95435ffc-5322-4bd8-9558-421f953096eb\") " pod="openstack/glance-default-external-api-0" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.855948 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mvfdw\" (UniqueName: \"kubernetes.io/projected/95435ffc-5322-4bd8-9558-421f953096eb-kube-api-access-mvfdw\") pod \"glance-default-external-api-0\" (UID: \"95435ffc-5322-4bd8-9558-421f953096eb\") " pod="openstack/glance-default-external-api-0" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.949294 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 13 15:46:36 crc kubenswrapper[4907]: I0313 15:46:36.014423 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 13 15:46:36 crc kubenswrapper[4907]: I0313 15:46:36.531013 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 13 15:46:36 crc kubenswrapper[4907]: W0313 15:46:36.532107 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod70b550c7_5c28_416d_83c3_1ba63ecd5abd.slice/crio-1efed8d64e814a0df83522691cc4cb2688b670fd49736ac116a8018a6049120c WatchSource:0}: Error finding container 1efed8d64e814a0df83522691cc4cb2688b670fd49736ac116a8018a6049120c: Status 404 returned error can't find the container with id 1efed8d64e814a0df83522691cc4cb2688b670fd49736ac116a8018a6049120c Mar 13 15:46:36 crc kubenswrapper[4907]: I0313 15:46:36.606212 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-c7977bcd9-2vs8h" Mar 13 15:46:36 crc kubenswrapper[4907]: I0313 15:46:36.606266 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-c7977bcd9-2vs8h" Mar 13 15:46:36 crc kubenswrapper[4907]: I0313 15:46:36.677532 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 13 15:46:36 crc kubenswrapper[4907]: I0313 15:46:36.907392 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-769b54bfcc-kbtds" Mar 13 15:46:37 crc kubenswrapper[4907]: I0313 15:46:37.487015 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-655cbf768f-dm9zz" Mar 13 15:46:37 crc kubenswrapper[4907]: I0313 15:46:37.487333 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-655cbf768f-dm9zz" Mar 13 15:46:37 crc kubenswrapper[4907]: I0313 15:46:37.506700 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"70b550c7-5c28-416d-83c3-1ba63ecd5abd","Type":"ContainerStarted","Data":"36f29b0ba8237aac66775608213473b8361269473dc952d84e52449563485b76"} Mar 13 15:46:37 crc kubenswrapper[4907]: I0313 15:46:37.506755 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"70b550c7-5c28-416d-83c3-1ba63ecd5abd","Type":"ContainerStarted","Data":"1efed8d64e814a0df83522691cc4cb2688b670fd49736ac116a8018a6049120c"} Mar 13 15:46:37 crc kubenswrapper[4907]: I0313 15:46:37.509645 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"95435ffc-5322-4bd8-9558-421f953096eb","Type":"ContainerStarted","Data":"9db337a620f677125ecfbd67707c679d5ec92247b100824435ccaca12a038a56"} Mar 13 15:46:37 crc kubenswrapper[4907]: I0313 15:46:37.509801 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"95435ffc-5322-4bd8-9558-421f953096eb","Type":"ContainerStarted","Data":"9ad8613388673a254440d120f808de8b4cb2ead90a90acf7bbca778db8b02af1"} Mar 13 15:46:38 crc kubenswrapper[4907]: I0313 15:46:38.519736 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"95435ffc-5322-4bd8-9558-421f953096eb","Type":"ContainerStarted","Data":"8a4d44e895e491e861e19682edef50a39dd9abc89004a2a6cb88ed9b4b4b3e21"} Mar 13 15:46:38 crc kubenswrapper[4907]: I0313 15:46:38.522305 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"70b550c7-5c28-416d-83c3-1ba63ecd5abd","Type":"ContainerStarted","Data":"80d182356297060953e417203d7ce12f38371db0d77c321425cf2bb8d290165d"} Mar 13 15:46:38 crc kubenswrapper[4907]: I0313 15:46:38.577806 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.577783161 podStartE2EDuration="3.577783161s" podCreationTimestamp="2026-03-13 15:46:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:46:38.567684934 +0000 UTC m=+6097.467472623" watchObservedRunningTime="2026-03-13 15:46:38.577783161 +0000 UTC m=+6097.477570850" Mar 13 15:46:38 crc kubenswrapper[4907]: I0313 15:46:38.578269 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.578263393 podStartE2EDuration="3.578263393s" podCreationTimestamp="2026-03-13 15:46:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:46:38.550523745 +0000 UTC m=+6097.450311434" watchObservedRunningTime="2026-03-13 15:46:38.578263393 +0000 UTC m=+6097.478051082" Mar 13 15:46:44 crc kubenswrapper[4907]: I0313 15:46:44.041223 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-4jnct"] Mar 13 15:46:44 crc kubenswrapper[4907]: I0313 15:46:44.050003 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-b654-account-create-update-4jjkj"] Mar 13 15:46:44 crc kubenswrapper[4907]: I0313 15:46:44.058911 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-4jnct"] Mar 13 15:46:44 crc kubenswrapper[4907]: I0313 15:46:44.066714 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-b654-account-create-update-4jjkj"] Mar 13 15:46:45 crc kubenswrapper[4907]: I0313 15:46:45.798603 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1e73852b-fdd0-4cd2-a268-96f6ed551011" path="/var/lib/kubelet/pods/1e73852b-fdd0-4cd2-a268-96f6ed551011/volumes" Mar 13 15:46:45 crc kubenswrapper[4907]: I0313 15:46:45.800224 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="523d3b7c-7098-44bc-92a6-d48f3ae49b65" path="/var/lib/kubelet/pods/523d3b7c-7098-44bc-92a6-d48f3ae49b65/volumes" Mar 13 15:46:45 crc kubenswrapper[4907]: I0313 15:46:45.950483 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Mar 13 15:46:45 crc kubenswrapper[4907]: I0313 15:46:45.950551 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Mar 13 15:46:45 crc kubenswrapper[4907]: I0313 15:46:45.987410 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Mar 13 15:46:45 crc kubenswrapper[4907]: I0313 15:46:45.990954 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Mar 13 15:46:46 crc kubenswrapper[4907]: I0313 15:46:46.015285 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Mar 13 15:46:46 crc kubenswrapper[4907]: I0313 15:46:46.015353 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Mar 13 15:46:46 crc kubenswrapper[4907]: I0313 15:46:46.054287 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Mar 13 15:46:46 crc kubenswrapper[4907]: I0313 15:46:46.066742 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Mar 13 15:46:46 crc kubenswrapper[4907]: I0313 15:46:46.600332 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Mar 13 15:46:46 crc kubenswrapper[4907]: I0313 15:46:46.600381 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Mar 13 15:46:46 crc kubenswrapper[4907]: I0313 15:46:46.600397 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Mar 13 15:46:46 crc kubenswrapper[4907]: I0313 15:46:46.600408 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Mar 13 15:46:46 crc kubenswrapper[4907]: I0313 15:46:46.617829 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-c7977bcd9-2vs8h" podUID="2d9fbab8-539c-48f6-8cae-3eda23300b28" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.154:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.154:8080: connect: connection refused" Mar 13 15:46:47 crc kubenswrapper[4907]: I0313 15:46:47.490363 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-655cbf768f-dm9zz" podUID="afd4001d-40a5-4a09-ab05-36cbd4d06eb4" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.156:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.156:8080: connect: connection refused" Mar 13 15:46:48 crc kubenswrapper[4907]: I0313 15:46:48.930154 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Mar 13 15:46:48 crc kubenswrapper[4907]: I0313 15:46:48.930494 4907 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 13 15:46:48 crc kubenswrapper[4907]: I0313 15:46:48.934294 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Mar 13 15:46:48 crc kubenswrapper[4907]: I0313 15:46:48.934386 4907 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 13 15:46:48 crc kubenswrapper[4907]: I0313 15:46:48.934809 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Mar 13 15:46:48 crc kubenswrapper[4907]: I0313 15:46:48.953861 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Mar 13 15:46:50 crc kubenswrapper[4907]: I0313 15:46:50.034092 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-xrjlt"] Mar 13 15:46:50 crc kubenswrapper[4907]: I0313 15:46:50.044198 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-xrjlt"] Mar 13 15:46:51 crc kubenswrapper[4907]: I0313 15:46:51.824907 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="df0d68f5-c073-4dff-8f70-f2897aaef9dc" path="/var/lib/kubelet/pods/df0d68f5-c073-4dff-8f70-f2897aaef9dc/volumes" Mar 13 15:46:58 crc kubenswrapper[4907]: I0313 15:46:58.418937 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-c7977bcd9-2vs8h" Mar 13 15:46:59 crc kubenswrapper[4907]: I0313 15:46:59.335510 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-655cbf768f-dm9zz" Mar 13 15:47:00 crc kubenswrapper[4907]: I0313 15:47:00.030061 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-c7977bcd9-2vs8h" Mar 13 15:47:01 crc kubenswrapper[4907]: I0313 15:47:01.140310 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-655cbf768f-dm9zz" Mar 13 15:47:01 crc kubenswrapper[4907]: I0313 15:47:01.217549 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-c7977bcd9-2vs8h"] Mar 13 15:47:01 crc kubenswrapper[4907]: I0313 15:47:01.218101 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-c7977bcd9-2vs8h" podUID="2d9fbab8-539c-48f6-8cae-3eda23300b28" containerName="horizon-log" containerID="cri-o://d405f8c27cb2c1dbfd5ce42b31129e474a50d6e8ddffc33a12dfa64551646e29" gracePeriod=30 Mar 13 15:47:01 crc kubenswrapper[4907]: I0313 15:47:01.218200 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-c7977bcd9-2vs8h" podUID="2d9fbab8-539c-48f6-8cae-3eda23300b28" containerName="horizon" containerID="cri-o://64ab189d02c2d9e2e9123ec6de24b5d5f7db40d7ceaa4b16356eefe21fdc102c" gracePeriod=30 Mar 13 15:47:04 crc kubenswrapper[4907]: I0313 15:47:04.781706 4907 generic.go:334] "Generic (PLEG): container finished" podID="2d9fbab8-539c-48f6-8cae-3eda23300b28" containerID="64ab189d02c2d9e2e9123ec6de24b5d5f7db40d7ceaa4b16356eefe21fdc102c" exitCode=0 Mar 13 15:47:04 crc kubenswrapper[4907]: I0313 15:47:04.781768 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-c7977bcd9-2vs8h" event={"ID":"2d9fbab8-539c-48f6-8cae-3eda23300b28","Type":"ContainerDied","Data":"64ab189d02c2d9e2e9123ec6de24b5d5f7db40d7ceaa4b16356eefe21fdc102c"} Mar 13 15:47:05 crc kubenswrapper[4907]: I0313 15:47:05.795338 4907 generic.go:334] "Generic (PLEG): container finished" podID="099b49c3-666e-4bf5-9c15-32cf9c40e1e7" containerID="862291f2b2864cad1d39537f26d6a5251dddefb5e909a977e483fcdb34432151" exitCode=137 Mar 13 15:47:05 crc kubenswrapper[4907]: I0313 15:47:05.795629 4907 generic.go:334] "Generic (PLEG): container finished" podID="099b49c3-666e-4bf5-9c15-32cf9c40e1e7" containerID="fcce68c0980dbae069e759717a0e2db31c52e1f7a61e3feb3e5debebb904b5dd" exitCode=137 Mar 13 15:47:05 crc kubenswrapper[4907]: I0313 15:47:05.795625 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-769b54bfcc-kbtds" event={"ID":"099b49c3-666e-4bf5-9c15-32cf9c40e1e7","Type":"ContainerDied","Data":"862291f2b2864cad1d39537f26d6a5251dddefb5e909a977e483fcdb34432151"} Mar 13 15:47:05 crc kubenswrapper[4907]: I0313 15:47:05.795670 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-769b54bfcc-kbtds" event={"ID":"099b49c3-666e-4bf5-9c15-32cf9c40e1e7","Type":"ContainerDied","Data":"fcce68c0980dbae069e759717a0e2db31c52e1f7a61e3feb3e5debebb904b5dd"} Mar 13 15:47:06 crc kubenswrapper[4907]: I0313 15:47:06.062378 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-769b54bfcc-kbtds" Mar 13 15:47:06 crc kubenswrapper[4907]: I0313 15:47:06.159221 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/099b49c3-666e-4bf5-9c15-32cf9c40e1e7-scripts\") pod \"099b49c3-666e-4bf5-9c15-32cf9c40e1e7\" (UID: \"099b49c3-666e-4bf5-9c15-32cf9c40e1e7\") " Mar 13 15:47:06 crc kubenswrapper[4907]: I0313 15:47:06.159291 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/099b49c3-666e-4bf5-9c15-32cf9c40e1e7-horizon-secret-key\") pod \"099b49c3-666e-4bf5-9c15-32cf9c40e1e7\" (UID: \"099b49c3-666e-4bf5-9c15-32cf9c40e1e7\") " Mar 13 15:47:06 crc kubenswrapper[4907]: I0313 15:47:06.159562 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/099b49c3-666e-4bf5-9c15-32cf9c40e1e7-logs\") pod \"099b49c3-666e-4bf5-9c15-32cf9c40e1e7\" (UID: \"099b49c3-666e-4bf5-9c15-32cf9c40e1e7\") " Mar 13 15:47:06 crc kubenswrapper[4907]: I0313 15:47:06.159634 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-scsbb\" (UniqueName: \"kubernetes.io/projected/099b49c3-666e-4bf5-9c15-32cf9c40e1e7-kube-api-access-scsbb\") pod \"099b49c3-666e-4bf5-9c15-32cf9c40e1e7\" (UID: \"099b49c3-666e-4bf5-9c15-32cf9c40e1e7\") " Mar 13 15:47:06 crc kubenswrapper[4907]: I0313 15:47:06.159817 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/099b49c3-666e-4bf5-9c15-32cf9c40e1e7-config-data\") pod \"099b49c3-666e-4bf5-9c15-32cf9c40e1e7\" (UID: \"099b49c3-666e-4bf5-9c15-32cf9c40e1e7\") " Mar 13 15:47:06 crc kubenswrapper[4907]: I0313 15:47:06.159981 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/099b49c3-666e-4bf5-9c15-32cf9c40e1e7-logs" (OuterVolumeSpecName: "logs") pod "099b49c3-666e-4bf5-9c15-32cf9c40e1e7" (UID: "099b49c3-666e-4bf5-9c15-32cf9c40e1e7"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:47:06 crc kubenswrapper[4907]: I0313 15:47:06.160531 4907 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/099b49c3-666e-4bf5-9c15-32cf9c40e1e7-logs\") on node \"crc\" DevicePath \"\"" Mar 13 15:47:06 crc kubenswrapper[4907]: I0313 15:47:06.166252 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/099b49c3-666e-4bf5-9c15-32cf9c40e1e7-kube-api-access-scsbb" (OuterVolumeSpecName: "kube-api-access-scsbb") pod "099b49c3-666e-4bf5-9c15-32cf9c40e1e7" (UID: "099b49c3-666e-4bf5-9c15-32cf9c40e1e7"). InnerVolumeSpecName "kube-api-access-scsbb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:47:06 crc kubenswrapper[4907]: I0313 15:47:06.166263 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/099b49c3-666e-4bf5-9c15-32cf9c40e1e7-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "099b49c3-666e-4bf5-9c15-32cf9c40e1e7" (UID: "099b49c3-666e-4bf5-9c15-32cf9c40e1e7"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:47:06 crc kubenswrapper[4907]: I0313 15:47:06.203939 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/099b49c3-666e-4bf5-9c15-32cf9c40e1e7-config-data" (OuterVolumeSpecName: "config-data") pod "099b49c3-666e-4bf5-9c15-32cf9c40e1e7" (UID: "099b49c3-666e-4bf5-9c15-32cf9c40e1e7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:47:06 crc kubenswrapper[4907]: I0313 15:47:06.205402 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/099b49c3-666e-4bf5-9c15-32cf9c40e1e7-scripts" (OuterVolumeSpecName: "scripts") pod "099b49c3-666e-4bf5-9c15-32cf9c40e1e7" (UID: "099b49c3-666e-4bf5-9c15-32cf9c40e1e7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:47:06 crc kubenswrapper[4907]: I0313 15:47:06.262950 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/099b49c3-666e-4bf5-9c15-32cf9c40e1e7-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 15:47:06 crc kubenswrapper[4907]: I0313 15:47:06.262995 4907 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/099b49c3-666e-4bf5-9c15-32cf9c40e1e7-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 15:47:06 crc kubenswrapper[4907]: I0313 15:47:06.263007 4907 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/099b49c3-666e-4bf5-9c15-32cf9c40e1e7-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Mar 13 15:47:06 crc kubenswrapper[4907]: I0313 15:47:06.263020 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-scsbb\" (UniqueName: \"kubernetes.io/projected/099b49c3-666e-4bf5-9c15-32cf9c40e1e7-kube-api-access-scsbb\") on node \"crc\" DevicePath \"\"" Mar 13 15:47:06 crc kubenswrapper[4907]: I0313 15:47:06.606168 4907 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-c7977bcd9-2vs8h" podUID="2d9fbab8-539c-48f6-8cae-3eda23300b28" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.154:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.154:8080: connect: connection refused" Mar 13 15:47:06 crc kubenswrapper[4907]: I0313 15:47:06.804794 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-769b54bfcc-kbtds" event={"ID":"099b49c3-666e-4bf5-9c15-32cf9c40e1e7","Type":"ContainerDied","Data":"e4caf9d5f22ffb193c93c87c06b4e64f7a29f8bf0582a1a8b7f47762f5ba55eb"} Mar 13 15:47:06 crc kubenswrapper[4907]: I0313 15:47:06.804839 4907 scope.go:117] "RemoveContainer" containerID="862291f2b2864cad1d39537f26d6a5251dddefb5e909a977e483fcdb34432151" Mar 13 15:47:06 crc kubenswrapper[4907]: I0313 15:47:06.804963 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-769b54bfcc-kbtds" Mar 13 15:47:06 crc kubenswrapper[4907]: I0313 15:47:06.847674 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-769b54bfcc-kbtds"] Mar 13 15:47:06 crc kubenswrapper[4907]: I0313 15:47:06.855543 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-769b54bfcc-kbtds"] Mar 13 15:47:06 crc kubenswrapper[4907]: I0313 15:47:06.984711 4907 scope.go:117] "RemoveContainer" containerID="fcce68c0980dbae069e759717a0e2db31c52e1f7a61e3feb3e5debebb904b5dd" Mar 13 15:47:07 crc kubenswrapper[4907]: I0313 15:47:07.807311 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="099b49c3-666e-4bf5-9c15-32cf9c40e1e7" path="/var/lib/kubelet/pods/099b49c3-666e-4bf5-9c15-32cf9c40e1e7/volumes" Mar 13 15:47:08 crc kubenswrapper[4907]: I0313 15:47:08.546652 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-6d79dff969-ssj4t"] Mar 13 15:47:08 crc kubenswrapper[4907]: E0313 15:47:08.547425 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="099b49c3-666e-4bf5-9c15-32cf9c40e1e7" containerName="horizon" Mar 13 15:47:08 crc kubenswrapper[4907]: I0313 15:47:08.547447 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="099b49c3-666e-4bf5-9c15-32cf9c40e1e7" containerName="horizon" Mar 13 15:47:08 crc kubenswrapper[4907]: E0313 15:47:08.547492 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="099b49c3-666e-4bf5-9c15-32cf9c40e1e7" containerName="horizon-log" Mar 13 15:47:08 crc kubenswrapper[4907]: I0313 15:47:08.547502 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="099b49c3-666e-4bf5-9c15-32cf9c40e1e7" containerName="horizon-log" Mar 13 15:47:08 crc kubenswrapper[4907]: I0313 15:47:08.547737 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="099b49c3-666e-4bf5-9c15-32cf9c40e1e7" containerName="horizon-log" Mar 13 15:47:08 crc kubenswrapper[4907]: I0313 15:47:08.547774 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="099b49c3-666e-4bf5-9c15-32cf9c40e1e7" containerName="horizon" Mar 13 15:47:08 crc kubenswrapper[4907]: I0313 15:47:08.549238 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6d79dff969-ssj4t" Mar 13 15:47:08 crc kubenswrapper[4907]: I0313 15:47:08.562400 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6d79dff969-ssj4t"] Mar 13 15:47:08 crc kubenswrapper[4907]: I0313 15:47:08.615496 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8tsd8\" (UniqueName: \"kubernetes.io/projected/5315ef54-fd95-400c-a7ef-a356c0f867bc-kube-api-access-8tsd8\") pod \"horizon-6d79dff969-ssj4t\" (UID: \"5315ef54-fd95-400c-a7ef-a356c0f867bc\") " pod="openstack/horizon-6d79dff969-ssj4t" Mar 13 15:47:08 crc kubenswrapper[4907]: I0313 15:47:08.615579 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5315ef54-fd95-400c-a7ef-a356c0f867bc-scripts\") pod \"horizon-6d79dff969-ssj4t\" (UID: \"5315ef54-fd95-400c-a7ef-a356c0f867bc\") " pod="openstack/horizon-6d79dff969-ssj4t" Mar 13 15:47:08 crc kubenswrapper[4907]: I0313 15:47:08.615606 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5315ef54-fd95-400c-a7ef-a356c0f867bc-config-data\") pod \"horizon-6d79dff969-ssj4t\" (UID: \"5315ef54-fd95-400c-a7ef-a356c0f867bc\") " pod="openstack/horizon-6d79dff969-ssj4t" Mar 13 15:47:08 crc kubenswrapper[4907]: I0313 15:47:08.615629 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/5315ef54-fd95-400c-a7ef-a356c0f867bc-horizon-secret-key\") pod \"horizon-6d79dff969-ssj4t\" (UID: \"5315ef54-fd95-400c-a7ef-a356c0f867bc\") " pod="openstack/horizon-6d79dff969-ssj4t" Mar 13 15:47:08 crc kubenswrapper[4907]: I0313 15:47:08.615755 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5315ef54-fd95-400c-a7ef-a356c0f867bc-logs\") pod \"horizon-6d79dff969-ssj4t\" (UID: \"5315ef54-fd95-400c-a7ef-a356c0f867bc\") " pod="openstack/horizon-6d79dff969-ssj4t" Mar 13 15:47:08 crc kubenswrapper[4907]: I0313 15:47:08.717035 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5315ef54-fd95-400c-a7ef-a356c0f867bc-logs\") pod \"horizon-6d79dff969-ssj4t\" (UID: \"5315ef54-fd95-400c-a7ef-a356c0f867bc\") " pod="openstack/horizon-6d79dff969-ssj4t" Mar 13 15:47:08 crc kubenswrapper[4907]: I0313 15:47:08.717146 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8tsd8\" (UniqueName: \"kubernetes.io/projected/5315ef54-fd95-400c-a7ef-a356c0f867bc-kube-api-access-8tsd8\") pod \"horizon-6d79dff969-ssj4t\" (UID: \"5315ef54-fd95-400c-a7ef-a356c0f867bc\") " pod="openstack/horizon-6d79dff969-ssj4t" Mar 13 15:47:08 crc kubenswrapper[4907]: I0313 15:47:08.717188 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5315ef54-fd95-400c-a7ef-a356c0f867bc-scripts\") pod \"horizon-6d79dff969-ssj4t\" (UID: \"5315ef54-fd95-400c-a7ef-a356c0f867bc\") " pod="openstack/horizon-6d79dff969-ssj4t" Mar 13 15:47:08 crc kubenswrapper[4907]: I0313 15:47:08.717206 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5315ef54-fd95-400c-a7ef-a356c0f867bc-config-data\") pod \"horizon-6d79dff969-ssj4t\" (UID: \"5315ef54-fd95-400c-a7ef-a356c0f867bc\") " pod="openstack/horizon-6d79dff969-ssj4t" Mar 13 15:47:08 crc kubenswrapper[4907]: I0313 15:47:08.717224 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/5315ef54-fd95-400c-a7ef-a356c0f867bc-horizon-secret-key\") pod \"horizon-6d79dff969-ssj4t\" (UID: \"5315ef54-fd95-400c-a7ef-a356c0f867bc\") " pod="openstack/horizon-6d79dff969-ssj4t" Mar 13 15:47:08 crc kubenswrapper[4907]: I0313 15:47:08.717737 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5315ef54-fd95-400c-a7ef-a356c0f867bc-logs\") pod \"horizon-6d79dff969-ssj4t\" (UID: \"5315ef54-fd95-400c-a7ef-a356c0f867bc\") " pod="openstack/horizon-6d79dff969-ssj4t" Mar 13 15:47:08 crc kubenswrapper[4907]: I0313 15:47:08.718171 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5315ef54-fd95-400c-a7ef-a356c0f867bc-scripts\") pod \"horizon-6d79dff969-ssj4t\" (UID: \"5315ef54-fd95-400c-a7ef-a356c0f867bc\") " pod="openstack/horizon-6d79dff969-ssj4t" Mar 13 15:47:08 crc kubenswrapper[4907]: I0313 15:47:08.718983 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5315ef54-fd95-400c-a7ef-a356c0f867bc-config-data\") pod \"horizon-6d79dff969-ssj4t\" (UID: \"5315ef54-fd95-400c-a7ef-a356c0f867bc\") " pod="openstack/horizon-6d79dff969-ssj4t" Mar 13 15:47:08 crc kubenswrapper[4907]: I0313 15:47:08.722069 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/5315ef54-fd95-400c-a7ef-a356c0f867bc-horizon-secret-key\") pod \"horizon-6d79dff969-ssj4t\" (UID: \"5315ef54-fd95-400c-a7ef-a356c0f867bc\") " pod="openstack/horizon-6d79dff969-ssj4t" Mar 13 15:47:08 crc kubenswrapper[4907]: I0313 15:47:08.741262 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8tsd8\" (UniqueName: \"kubernetes.io/projected/5315ef54-fd95-400c-a7ef-a356c0f867bc-kube-api-access-8tsd8\") pod \"horizon-6d79dff969-ssj4t\" (UID: \"5315ef54-fd95-400c-a7ef-a356c0f867bc\") " pod="openstack/horizon-6d79dff969-ssj4t" Mar 13 15:47:08 crc kubenswrapper[4907]: I0313 15:47:08.870133 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6d79dff969-ssj4t" Mar 13 15:47:09 crc kubenswrapper[4907]: I0313 15:47:09.344547 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6d79dff969-ssj4t"] Mar 13 15:47:09 crc kubenswrapper[4907]: I0313 15:47:09.838126 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6d79dff969-ssj4t" event={"ID":"5315ef54-fd95-400c-a7ef-a356c0f867bc","Type":"ContainerStarted","Data":"d662824b3d4144a85a7254045e81fec495afd2d985d3e17becf10c307f110f3f"} Mar 13 15:47:09 crc kubenswrapper[4907]: I0313 15:47:09.838478 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6d79dff969-ssj4t" event={"ID":"5315ef54-fd95-400c-a7ef-a356c0f867bc","Type":"ContainerStarted","Data":"869c9c9925a981e0b81d87611070cdcfe490596c4315540b6e041d946c3a07dc"} Mar 13 15:47:09 crc kubenswrapper[4907]: I0313 15:47:09.838499 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6d79dff969-ssj4t" event={"ID":"5315ef54-fd95-400c-a7ef-a356c0f867bc","Type":"ContainerStarted","Data":"9c2349bc5006f4b36b5615d4e495b3a166fcd8e0f8d0962d4816a5e2a2c9a9a3"} Mar 13 15:47:09 crc kubenswrapper[4907]: I0313 15:47:09.859633 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-6d79dff969-ssj4t" podStartSLOduration=1.859611913 podStartE2EDuration="1.859611913s" podCreationTimestamp="2026-03-13 15:47:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:47:09.857254779 +0000 UTC m=+6128.757042488" watchObservedRunningTime="2026-03-13 15:47:09.859611913 +0000 UTC m=+6128.759399602" Mar 13 15:47:09 crc kubenswrapper[4907]: I0313 15:47:09.995160 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-create-lnsws"] Mar 13 15:47:09 crc kubenswrapper[4907]: I0313 15:47:09.997429 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-lnsws" Mar 13 15:47:10 crc kubenswrapper[4907]: I0313 15:47:10.017613 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-create-lnsws"] Mar 13 15:47:10 crc kubenswrapper[4907]: I0313 15:47:10.116643 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-019b-account-create-update-848sw"] Mar 13 15:47:10 crc kubenswrapper[4907]: I0313 15:47:10.131001 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-019b-account-create-update-848sw"] Mar 13 15:47:10 crc kubenswrapper[4907]: I0313 15:47:10.131333 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-019b-account-create-update-848sw" Mar 13 15:47:10 crc kubenswrapper[4907]: I0313 15:47:10.134590 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-db-secret" Mar 13 15:47:10 crc kubenswrapper[4907]: I0313 15:47:10.160584 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wx7rn\" (UniqueName: \"kubernetes.io/projected/490bd531-09f6-48e5-b194-b79e9b532131-kube-api-access-wx7rn\") pod \"heat-db-create-lnsws\" (UID: \"490bd531-09f6-48e5-b194-b79e9b532131\") " pod="openstack/heat-db-create-lnsws" Mar 13 15:47:10 crc kubenswrapper[4907]: I0313 15:47:10.161054 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/490bd531-09f6-48e5-b194-b79e9b532131-operator-scripts\") pod \"heat-db-create-lnsws\" (UID: \"490bd531-09f6-48e5-b194-b79e9b532131\") " pod="openstack/heat-db-create-lnsws" Mar 13 15:47:10 crc kubenswrapper[4907]: I0313 15:47:10.268764 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ce0b08b9-c4e1-4236-8f54-cf78b45d2d49-operator-scripts\") pod \"heat-019b-account-create-update-848sw\" (UID: \"ce0b08b9-c4e1-4236-8f54-cf78b45d2d49\") " pod="openstack/heat-019b-account-create-update-848sw" Mar 13 15:47:10 crc kubenswrapper[4907]: I0313 15:47:10.268871 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wx7rn\" (UniqueName: \"kubernetes.io/projected/490bd531-09f6-48e5-b194-b79e9b532131-kube-api-access-wx7rn\") pod \"heat-db-create-lnsws\" (UID: \"490bd531-09f6-48e5-b194-b79e9b532131\") " pod="openstack/heat-db-create-lnsws" Mar 13 15:47:10 crc kubenswrapper[4907]: I0313 15:47:10.268961 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/490bd531-09f6-48e5-b194-b79e9b532131-operator-scripts\") pod \"heat-db-create-lnsws\" (UID: \"490bd531-09f6-48e5-b194-b79e9b532131\") " pod="openstack/heat-db-create-lnsws" Mar 13 15:47:10 crc kubenswrapper[4907]: I0313 15:47:10.268987 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-khvms\" (UniqueName: \"kubernetes.io/projected/ce0b08b9-c4e1-4236-8f54-cf78b45d2d49-kube-api-access-khvms\") pod \"heat-019b-account-create-update-848sw\" (UID: \"ce0b08b9-c4e1-4236-8f54-cf78b45d2d49\") " pod="openstack/heat-019b-account-create-update-848sw" Mar 13 15:47:10 crc kubenswrapper[4907]: I0313 15:47:10.269743 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/490bd531-09f6-48e5-b194-b79e9b532131-operator-scripts\") pod \"heat-db-create-lnsws\" (UID: \"490bd531-09f6-48e5-b194-b79e9b532131\") " pod="openstack/heat-db-create-lnsws" Mar 13 15:47:10 crc kubenswrapper[4907]: I0313 15:47:10.298860 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wx7rn\" (UniqueName: \"kubernetes.io/projected/490bd531-09f6-48e5-b194-b79e9b532131-kube-api-access-wx7rn\") pod \"heat-db-create-lnsws\" (UID: \"490bd531-09f6-48e5-b194-b79e9b532131\") " pod="openstack/heat-db-create-lnsws" Mar 13 15:47:10 crc kubenswrapper[4907]: I0313 15:47:10.357764 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-lnsws" Mar 13 15:47:10 crc kubenswrapper[4907]: I0313 15:47:10.371100 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-khvms\" (UniqueName: \"kubernetes.io/projected/ce0b08b9-c4e1-4236-8f54-cf78b45d2d49-kube-api-access-khvms\") pod \"heat-019b-account-create-update-848sw\" (UID: \"ce0b08b9-c4e1-4236-8f54-cf78b45d2d49\") " pod="openstack/heat-019b-account-create-update-848sw" Mar 13 15:47:10 crc kubenswrapper[4907]: I0313 15:47:10.371412 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ce0b08b9-c4e1-4236-8f54-cf78b45d2d49-operator-scripts\") pod \"heat-019b-account-create-update-848sw\" (UID: \"ce0b08b9-c4e1-4236-8f54-cf78b45d2d49\") " pod="openstack/heat-019b-account-create-update-848sw" Mar 13 15:47:10 crc kubenswrapper[4907]: I0313 15:47:10.372122 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ce0b08b9-c4e1-4236-8f54-cf78b45d2d49-operator-scripts\") pod \"heat-019b-account-create-update-848sw\" (UID: \"ce0b08b9-c4e1-4236-8f54-cf78b45d2d49\") " pod="openstack/heat-019b-account-create-update-848sw" Mar 13 15:47:10 crc kubenswrapper[4907]: I0313 15:47:10.387250 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-khvms\" (UniqueName: \"kubernetes.io/projected/ce0b08b9-c4e1-4236-8f54-cf78b45d2d49-kube-api-access-khvms\") pod \"heat-019b-account-create-update-848sw\" (UID: \"ce0b08b9-c4e1-4236-8f54-cf78b45d2d49\") " pod="openstack/heat-019b-account-create-update-848sw" Mar 13 15:47:10 crc kubenswrapper[4907]: I0313 15:47:10.451998 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-019b-account-create-update-848sw" Mar 13 15:47:11 crc kubenswrapper[4907]: I0313 15:47:11.041025 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-create-lnsws"] Mar 13 15:47:11 crc kubenswrapper[4907]: I0313 15:47:11.137705 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-019b-account-create-update-848sw"] Mar 13 15:47:11 crc kubenswrapper[4907]: I0313 15:47:11.859320 4907 generic.go:334] "Generic (PLEG): container finished" podID="ce0b08b9-c4e1-4236-8f54-cf78b45d2d49" containerID="839f8683996fb111c90ced80fc315df41501ccd29aa3ce62746269e8e8d6d880" exitCode=0 Mar 13 15:47:11 crc kubenswrapper[4907]: I0313 15:47:11.859375 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-019b-account-create-update-848sw" event={"ID":"ce0b08b9-c4e1-4236-8f54-cf78b45d2d49","Type":"ContainerDied","Data":"839f8683996fb111c90ced80fc315df41501ccd29aa3ce62746269e8e8d6d880"} Mar 13 15:47:11 crc kubenswrapper[4907]: I0313 15:47:11.859424 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-019b-account-create-update-848sw" event={"ID":"ce0b08b9-c4e1-4236-8f54-cf78b45d2d49","Type":"ContainerStarted","Data":"d2ee4f22d00d1f9ceaf1d6575365b027f64e73a1473c8ba545cf5581cfc37e71"} Mar 13 15:47:11 crc kubenswrapper[4907]: I0313 15:47:11.861264 4907 generic.go:334] "Generic (PLEG): container finished" podID="490bd531-09f6-48e5-b194-b79e9b532131" containerID="daa82cb23c63e51c880db96744e0980e9ddf6f38f0cbcf2713496cba5c2d4486" exitCode=0 Mar 13 15:47:11 crc kubenswrapper[4907]: I0313 15:47:11.861314 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-lnsws" event={"ID":"490bd531-09f6-48e5-b194-b79e9b532131","Type":"ContainerDied","Data":"daa82cb23c63e51c880db96744e0980e9ddf6f38f0cbcf2713496cba5c2d4486"} Mar 13 15:47:11 crc kubenswrapper[4907]: I0313 15:47:11.861337 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-lnsws" event={"ID":"490bd531-09f6-48e5-b194-b79e9b532131","Type":"ContainerStarted","Data":"23b375d293ca9398feacd198bd47f6604e6d9e7963768ae62f0d245f7da436e3"} Mar 13 15:47:13 crc kubenswrapper[4907]: I0313 15:47:13.344617 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-lnsws" Mar 13 15:47:13 crc kubenswrapper[4907]: I0313 15:47:13.360989 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-019b-account-create-update-848sw" Mar 13 15:47:13 crc kubenswrapper[4907]: I0313 15:47:13.369499 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-khvms\" (UniqueName: \"kubernetes.io/projected/ce0b08b9-c4e1-4236-8f54-cf78b45d2d49-kube-api-access-khvms\") pod \"ce0b08b9-c4e1-4236-8f54-cf78b45d2d49\" (UID: \"ce0b08b9-c4e1-4236-8f54-cf78b45d2d49\") " Mar 13 15:47:13 crc kubenswrapper[4907]: I0313 15:47:13.370723 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/490bd531-09f6-48e5-b194-b79e9b532131-operator-scripts\") pod \"490bd531-09f6-48e5-b194-b79e9b532131\" (UID: \"490bd531-09f6-48e5-b194-b79e9b532131\") " Mar 13 15:47:13 crc kubenswrapper[4907]: I0313 15:47:13.370940 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wx7rn\" (UniqueName: \"kubernetes.io/projected/490bd531-09f6-48e5-b194-b79e9b532131-kube-api-access-wx7rn\") pod \"490bd531-09f6-48e5-b194-b79e9b532131\" (UID: \"490bd531-09f6-48e5-b194-b79e9b532131\") " Mar 13 15:47:13 crc kubenswrapper[4907]: I0313 15:47:13.371102 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ce0b08b9-c4e1-4236-8f54-cf78b45d2d49-operator-scripts\") pod \"ce0b08b9-c4e1-4236-8f54-cf78b45d2d49\" (UID: \"ce0b08b9-c4e1-4236-8f54-cf78b45d2d49\") " Mar 13 15:47:13 crc kubenswrapper[4907]: I0313 15:47:13.371774 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/490bd531-09f6-48e5-b194-b79e9b532131-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "490bd531-09f6-48e5-b194-b79e9b532131" (UID: "490bd531-09f6-48e5-b194-b79e9b532131"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:47:13 crc kubenswrapper[4907]: I0313 15:47:13.375320 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ce0b08b9-c4e1-4236-8f54-cf78b45d2d49-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ce0b08b9-c4e1-4236-8f54-cf78b45d2d49" (UID: "ce0b08b9-c4e1-4236-8f54-cf78b45d2d49"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:47:13 crc kubenswrapper[4907]: I0313 15:47:13.378783 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ce0b08b9-c4e1-4236-8f54-cf78b45d2d49-kube-api-access-khvms" (OuterVolumeSpecName: "kube-api-access-khvms") pod "ce0b08b9-c4e1-4236-8f54-cf78b45d2d49" (UID: "ce0b08b9-c4e1-4236-8f54-cf78b45d2d49"). InnerVolumeSpecName "kube-api-access-khvms". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:47:13 crc kubenswrapper[4907]: I0313 15:47:13.388589 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/490bd531-09f6-48e5-b194-b79e9b532131-kube-api-access-wx7rn" (OuterVolumeSpecName: "kube-api-access-wx7rn") pod "490bd531-09f6-48e5-b194-b79e9b532131" (UID: "490bd531-09f6-48e5-b194-b79e9b532131"). InnerVolumeSpecName "kube-api-access-wx7rn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:47:13 crc kubenswrapper[4907]: I0313 15:47:13.389774 4907 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ce0b08b9-c4e1-4236-8f54-cf78b45d2d49-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 15:47:13 crc kubenswrapper[4907]: I0313 15:47:13.389911 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-khvms\" (UniqueName: \"kubernetes.io/projected/ce0b08b9-c4e1-4236-8f54-cf78b45d2d49-kube-api-access-khvms\") on node \"crc\" DevicePath \"\"" Mar 13 15:47:13 crc kubenswrapper[4907]: I0313 15:47:13.389925 4907 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/490bd531-09f6-48e5-b194-b79e9b532131-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 15:47:13 crc kubenswrapper[4907]: I0313 15:47:13.390034 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wx7rn\" (UniqueName: \"kubernetes.io/projected/490bd531-09f6-48e5-b194-b79e9b532131-kube-api-access-wx7rn\") on node \"crc\" DevicePath \"\"" Mar 13 15:47:13 crc kubenswrapper[4907]: I0313 15:47:13.886553 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-019b-account-create-update-848sw" event={"ID":"ce0b08b9-c4e1-4236-8f54-cf78b45d2d49","Type":"ContainerDied","Data":"d2ee4f22d00d1f9ceaf1d6575365b027f64e73a1473c8ba545cf5581cfc37e71"} Mar 13 15:47:13 crc kubenswrapper[4907]: I0313 15:47:13.886601 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d2ee4f22d00d1f9ceaf1d6575365b027f64e73a1473c8ba545cf5581cfc37e71" Mar 13 15:47:13 crc kubenswrapper[4907]: I0313 15:47:13.886671 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-019b-account-create-update-848sw" Mar 13 15:47:13 crc kubenswrapper[4907]: I0313 15:47:13.889486 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-lnsws" event={"ID":"490bd531-09f6-48e5-b194-b79e9b532131","Type":"ContainerDied","Data":"23b375d293ca9398feacd198bd47f6604e6d9e7963768ae62f0d245f7da436e3"} Mar 13 15:47:13 crc kubenswrapper[4907]: I0313 15:47:13.889516 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="23b375d293ca9398feacd198bd47f6604e6d9e7963768ae62f0d245f7da436e3" Mar 13 15:47:13 crc kubenswrapper[4907]: I0313 15:47:13.889657 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-lnsws" Mar 13 15:47:15 crc kubenswrapper[4907]: I0313 15:47:15.326212 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-sync-mqnk8"] Mar 13 15:47:15 crc kubenswrapper[4907]: E0313 15:47:15.327135 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce0b08b9-c4e1-4236-8f54-cf78b45d2d49" containerName="mariadb-account-create-update" Mar 13 15:47:15 crc kubenswrapper[4907]: I0313 15:47:15.327153 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce0b08b9-c4e1-4236-8f54-cf78b45d2d49" containerName="mariadb-account-create-update" Mar 13 15:47:15 crc kubenswrapper[4907]: E0313 15:47:15.327221 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="490bd531-09f6-48e5-b194-b79e9b532131" containerName="mariadb-database-create" Mar 13 15:47:15 crc kubenswrapper[4907]: I0313 15:47:15.327231 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="490bd531-09f6-48e5-b194-b79e9b532131" containerName="mariadb-database-create" Mar 13 15:47:15 crc kubenswrapper[4907]: I0313 15:47:15.327484 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce0b08b9-c4e1-4236-8f54-cf78b45d2d49" containerName="mariadb-account-create-update" Mar 13 15:47:15 crc kubenswrapper[4907]: I0313 15:47:15.327505 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="490bd531-09f6-48e5-b194-b79e9b532131" containerName="mariadb-database-create" Mar 13 15:47:15 crc kubenswrapper[4907]: I0313 15:47:15.328439 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-mqnk8" Mar 13 15:47:15 crc kubenswrapper[4907]: I0313 15:47:15.332160 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-config-data" Mar 13 15:47:15 crc kubenswrapper[4907]: I0313 15:47:15.332653 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-heat-dockercfg-mcdkh" Mar 13 15:47:15 crc kubenswrapper[4907]: I0313 15:47:15.339027 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-mqnk8"] Mar 13 15:47:15 crc kubenswrapper[4907]: I0313 15:47:15.429016 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6df151b5-1c14-488c-9f23-9d6d629b493d-config-data\") pod \"heat-db-sync-mqnk8\" (UID: \"6df151b5-1c14-488c-9f23-9d6d629b493d\") " pod="openstack/heat-db-sync-mqnk8" Mar 13 15:47:15 crc kubenswrapper[4907]: I0313 15:47:15.429080 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jzrmm\" (UniqueName: \"kubernetes.io/projected/6df151b5-1c14-488c-9f23-9d6d629b493d-kube-api-access-jzrmm\") pod \"heat-db-sync-mqnk8\" (UID: \"6df151b5-1c14-488c-9f23-9d6d629b493d\") " pod="openstack/heat-db-sync-mqnk8" Mar 13 15:47:15 crc kubenswrapper[4907]: I0313 15:47:15.429449 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6df151b5-1c14-488c-9f23-9d6d629b493d-combined-ca-bundle\") pod \"heat-db-sync-mqnk8\" (UID: \"6df151b5-1c14-488c-9f23-9d6d629b493d\") " pod="openstack/heat-db-sync-mqnk8" Mar 13 15:47:15 crc kubenswrapper[4907]: I0313 15:47:15.531838 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6df151b5-1c14-488c-9f23-9d6d629b493d-config-data\") pod \"heat-db-sync-mqnk8\" (UID: \"6df151b5-1c14-488c-9f23-9d6d629b493d\") " pod="openstack/heat-db-sync-mqnk8" Mar 13 15:47:15 crc kubenswrapper[4907]: I0313 15:47:15.531903 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jzrmm\" (UniqueName: \"kubernetes.io/projected/6df151b5-1c14-488c-9f23-9d6d629b493d-kube-api-access-jzrmm\") pod \"heat-db-sync-mqnk8\" (UID: \"6df151b5-1c14-488c-9f23-9d6d629b493d\") " pod="openstack/heat-db-sync-mqnk8" Mar 13 15:47:15 crc kubenswrapper[4907]: I0313 15:47:15.531962 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6df151b5-1c14-488c-9f23-9d6d629b493d-combined-ca-bundle\") pod \"heat-db-sync-mqnk8\" (UID: \"6df151b5-1c14-488c-9f23-9d6d629b493d\") " pod="openstack/heat-db-sync-mqnk8" Mar 13 15:47:15 crc kubenswrapper[4907]: I0313 15:47:15.545922 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6df151b5-1c14-488c-9f23-9d6d629b493d-combined-ca-bundle\") pod \"heat-db-sync-mqnk8\" (UID: \"6df151b5-1c14-488c-9f23-9d6d629b493d\") " pod="openstack/heat-db-sync-mqnk8" Mar 13 15:47:15 crc kubenswrapper[4907]: I0313 15:47:15.546079 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6df151b5-1c14-488c-9f23-9d6d629b493d-config-data\") pod \"heat-db-sync-mqnk8\" (UID: \"6df151b5-1c14-488c-9f23-9d6d629b493d\") " pod="openstack/heat-db-sync-mqnk8" Mar 13 15:47:15 crc kubenswrapper[4907]: I0313 15:47:15.548287 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jzrmm\" (UniqueName: \"kubernetes.io/projected/6df151b5-1c14-488c-9f23-9d6d629b493d-kube-api-access-jzrmm\") pod \"heat-db-sync-mqnk8\" (UID: \"6df151b5-1c14-488c-9f23-9d6d629b493d\") " pod="openstack/heat-db-sync-mqnk8" Mar 13 15:47:15 crc kubenswrapper[4907]: I0313 15:47:15.649331 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-mqnk8" Mar 13 15:47:16 crc kubenswrapper[4907]: I0313 15:47:16.165412 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-mqnk8"] Mar 13 15:47:16 crc kubenswrapper[4907]: I0313 15:47:16.606996 4907 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-c7977bcd9-2vs8h" podUID="2d9fbab8-539c-48f6-8cae-3eda23300b28" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.154:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.154:8080: connect: connection refused" Mar 13 15:47:16 crc kubenswrapper[4907]: I0313 15:47:16.920864 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-mqnk8" event={"ID":"6df151b5-1c14-488c-9f23-9d6d629b493d","Type":"ContainerStarted","Data":"3d55337bfec3fbfeaf31255470fe615e54c37d18887cd6469654963c1c88fe12"} Mar 13 15:47:17 crc kubenswrapper[4907]: I0313 15:47:17.975420 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-tthk2"] Mar 13 15:47:17 crc kubenswrapper[4907]: I0313 15:47:17.979685 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tthk2" Mar 13 15:47:18 crc kubenswrapper[4907]: I0313 15:47:18.012166 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-tthk2"] Mar 13 15:47:18 crc kubenswrapper[4907]: I0313 15:47:18.055862 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-jh7h4"] Mar 13 15:47:18 crc kubenswrapper[4907]: I0313 15:47:18.066875 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-629b-account-create-update-rptv2"] Mar 13 15:47:18 crc kubenswrapper[4907]: I0313 15:47:18.076423 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-jh7h4"] Mar 13 15:47:18 crc kubenswrapper[4907]: I0313 15:47:18.085286 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-629b-account-create-update-rptv2"] Mar 13 15:47:18 crc kubenswrapper[4907]: I0313 15:47:18.086815 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4ecb3195-45a5-4ca4-b979-9a8c37ec4fbe-catalog-content\") pod \"community-operators-tthk2\" (UID: \"4ecb3195-45a5-4ca4-b979-9a8c37ec4fbe\") " pod="openshift-marketplace/community-operators-tthk2" Mar 13 15:47:18 crc kubenswrapper[4907]: I0313 15:47:18.086965 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6kbk5\" (UniqueName: \"kubernetes.io/projected/4ecb3195-45a5-4ca4-b979-9a8c37ec4fbe-kube-api-access-6kbk5\") pod \"community-operators-tthk2\" (UID: \"4ecb3195-45a5-4ca4-b979-9a8c37ec4fbe\") " pod="openshift-marketplace/community-operators-tthk2" Mar 13 15:47:18 crc kubenswrapper[4907]: I0313 15:47:18.087324 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4ecb3195-45a5-4ca4-b979-9a8c37ec4fbe-utilities\") pod \"community-operators-tthk2\" (UID: \"4ecb3195-45a5-4ca4-b979-9a8c37ec4fbe\") " pod="openshift-marketplace/community-operators-tthk2" Mar 13 15:47:18 crc kubenswrapper[4907]: I0313 15:47:18.189426 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4ecb3195-45a5-4ca4-b979-9a8c37ec4fbe-utilities\") pod \"community-operators-tthk2\" (UID: \"4ecb3195-45a5-4ca4-b979-9a8c37ec4fbe\") " pod="openshift-marketplace/community-operators-tthk2" Mar 13 15:47:18 crc kubenswrapper[4907]: I0313 15:47:18.189589 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4ecb3195-45a5-4ca4-b979-9a8c37ec4fbe-catalog-content\") pod \"community-operators-tthk2\" (UID: \"4ecb3195-45a5-4ca4-b979-9a8c37ec4fbe\") " pod="openshift-marketplace/community-operators-tthk2" Mar 13 15:47:18 crc kubenswrapper[4907]: I0313 15:47:18.189656 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6kbk5\" (UniqueName: \"kubernetes.io/projected/4ecb3195-45a5-4ca4-b979-9a8c37ec4fbe-kube-api-access-6kbk5\") pod \"community-operators-tthk2\" (UID: \"4ecb3195-45a5-4ca4-b979-9a8c37ec4fbe\") " pod="openshift-marketplace/community-operators-tthk2" Mar 13 15:47:18 crc kubenswrapper[4907]: I0313 15:47:18.190578 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4ecb3195-45a5-4ca4-b979-9a8c37ec4fbe-utilities\") pod \"community-operators-tthk2\" (UID: \"4ecb3195-45a5-4ca4-b979-9a8c37ec4fbe\") " pod="openshift-marketplace/community-operators-tthk2" Mar 13 15:47:18 crc kubenswrapper[4907]: I0313 15:47:18.190904 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4ecb3195-45a5-4ca4-b979-9a8c37ec4fbe-catalog-content\") pod \"community-operators-tthk2\" (UID: \"4ecb3195-45a5-4ca4-b979-9a8c37ec4fbe\") " pod="openshift-marketplace/community-operators-tthk2" Mar 13 15:47:18 crc kubenswrapper[4907]: I0313 15:47:18.222096 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6kbk5\" (UniqueName: \"kubernetes.io/projected/4ecb3195-45a5-4ca4-b979-9a8c37ec4fbe-kube-api-access-6kbk5\") pod \"community-operators-tthk2\" (UID: \"4ecb3195-45a5-4ca4-b979-9a8c37ec4fbe\") " pod="openshift-marketplace/community-operators-tthk2" Mar 13 15:47:18 crc kubenswrapper[4907]: I0313 15:47:18.317724 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tthk2" Mar 13 15:47:18 crc kubenswrapper[4907]: I0313 15:47:18.871205 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-6d79dff969-ssj4t" Mar 13 15:47:18 crc kubenswrapper[4907]: I0313 15:47:18.871484 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-6d79dff969-ssj4t" Mar 13 15:47:18 crc kubenswrapper[4907]: I0313 15:47:18.942160 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-tthk2"] Mar 13 15:47:19 crc kubenswrapper[4907]: I0313 15:47:19.799309 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cbbccb7-b0c0-49be-a969-ffa18b07bd13" path="/var/lib/kubelet/pods/3cbbccb7-b0c0-49be-a969-ffa18b07bd13/volumes" Mar 13 15:47:19 crc kubenswrapper[4907]: I0313 15:47:19.800726 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="608c9419-22d5-4915-b5c6-8be70eebdb23" path="/var/lib/kubelet/pods/608c9419-22d5-4915-b5c6-8be70eebdb23/volumes" Mar 13 15:47:19 crc kubenswrapper[4907]: I0313 15:47:19.963669 4907 generic.go:334] "Generic (PLEG): container finished" podID="4ecb3195-45a5-4ca4-b979-9a8c37ec4fbe" containerID="3fc3c676404f5dc909e1ad737ea131e306c394034b644f7d1fc0ccf0fd152916" exitCode=0 Mar 13 15:47:19 crc kubenswrapper[4907]: I0313 15:47:19.963762 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tthk2" event={"ID":"4ecb3195-45a5-4ca4-b979-9a8c37ec4fbe","Type":"ContainerDied","Data":"3fc3c676404f5dc909e1ad737ea131e306c394034b644f7d1fc0ccf0fd152916"} Mar 13 15:47:19 crc kubenswrapper[4907]: I0313 15:47:19.963798 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tthk2" event={"ID":"4ecb3195-45a5-4ca4-b979-9a8c37ec4fbe","Type":"ContainerStarted","Data":"ab17e9b088419fabf1ec951208897a9f96839403f2d2331918d09059ad41ce73"} Mar 13 15:47:25 crc kubenswrapper[4907]: I0313 15:47:25.017586 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tthk2" event={"ID":"4ecb3195-45a5-4ca4-b979-9a8c37ec4fbe","Type":"ContainerStarted","Data":"dfea42d9a630e3e1c6c04649bc5e93bd04f889efb044725230049bd0d1de35ff"} Mar 13 15:47:25 crc kubenswrapper[4907]: I0313 15:47:25.021017 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-mqnk8" event={"ID":"6df151b5-1c14-488c-9f23-9d6d629b493d","Type":"ContainerStarted","Data":"28274d3b65f2c9d6e1b8e4c19babc239c1199e687ba7bf22cb4ab37476252833"} Mar 13 15:47:25 crc kubenswrapper[4907]: I0313 15:47:25.072727 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-db-sync-mqnk8" podStartSLOduration=1.6402556719999999 podStartE2EDuration="10.072706601s" podCreationTimestamp="2026-03-13 15:47:15 +0000 UTC" firstStartedPulling="2026-03-13 15:47:16.165783898 +0000 UTC m=+6135.065571587" lastFinishedPulling="2026-03-13 15:47:24.598234827 +0000 UTC m=+6143.498022516" observedRunningTime="2026-03-13 15:47:25.063248522 +0000 UTC m=+6143.963036211" watchObservedRunningTime="2026-03-13 15:47:25.072706601 +0000 UTC m=+6143.972494280" Mar 13 15:47:26 crc kubenswrapper[4907]: I0313 15:47:26.033227 4907 generic.go:334] "Generic (PLEG): container finished" podID="4ecb3195-45a5-4ca4-b979-9a8c37ec4fbe" containerID="dfea42d9a630e3e1c6c04649bc5e93bd04f889efb044725230049bd0d1de35ff" exitCode=0 Mar 13 15:47:26 crc kubenswrapper[4907]: I0313 15:47:26.033271 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tthk2" event={"ID":"4ecb3195-45a5-4ca4-b979-9a8c37ec4fbe","Type":"ContainerDied","Data":"dfea42d9a630e3e1c6c04649bc5e93bd04f889efb044725230049bd0d1de35ff"} Mar 13 15:47:26 crc kubenswrapper[4907]: I0313 15:47:26.606825 4907 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-c7977bcd9-2vs8h" podUID="2d9fbab8-539c-48f6-8cae-3eda23300b28" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.154:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.154:8080: connect: connection refused" Mar 13 15:47:26 crc kubenswrapper[4907]: I0313 15:47:26.607049 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-c7977bcd9-2vs8h" Mar 13 15:47:27 crc kubenswrapper[4907]: I0313 15:47:27.037451 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-jtsxr"] Mar 13 15:47:27 crc kubenswrapper[4907]: I0313 15:47:27.048020 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tthk2" event={"ID":"4ecb3195-45a5-4ca4-b979-9a8c37ec4fbe","Type":"ContainerStarted","Data":"29ee2de0275600572b2cd057fb10e5905301ef9ab1753a5ad562223165c95348"} Mar 13 15:47:27 crc kubenswrapper[4907]: I0313 15:47:27.048726 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-jtsxr"] Mar 13 15:47:27 crc kubenswrapper[4907]: I0313 15:47:27.072539 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-tthk2" podStartSLOduration=6.958694294 podStartE2EDuration="10.072519259s" podCreationTimestamp="2026-03-13 15:47:17 +0000 UTC" firstStartedPulling="2026-03-13 15:47:23.666287335 +0000 UTC m=+6142.566075024" lastFinishedPulling="2026-03-13 15:47:26.7801123 +0000 UTC m=+6145.679899989" observedRunningTime="2026-03-13 15:47:27.065356883 +0000 UTC m=+6145.965144572" watchObservedRunningTime="2026-03-13 15:47:27.072519259 +0000 UTC m=+6145.972306948" Mar 13 15:47:27 crc kubenswrapper[4907]: I0313 15:47:27.793171 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1af3ec26-0a7c-44c4-8073-fc93093c0cf7" path="/var/lib/kubelet/pods/1af3ec26-0a7c-44c4-8073-fc93093c0cf7/volumes" Mar 13 15:47:28 crc kubenswrapper[4907]: I0313 15:47:28.058148 4907 generic.go:334] "Generic (PLEG): container finished" podID="6df151b5-1c14-488c-9f23-9d6d629b493d" containerID="28274d3b65f2c9d6e1b8e4c19babc239c1199e687ba7bf22cb4ab37476252833" exitCode=0 Mar 13 15:47:28 crc kubenswrapper[4907]: I0313 15:47:28.058232 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-mqnk8" event={"ID":"6df151b5-1c14-488c-9f23-9d6d629b493d","Type":"ContainerDied","Data":"28274d3b65f2c9d6e1b8e4c19babc239c1199e687ba7bf22cb4ab37476252833"} Mar 13 15:47:28 crc kubenswrapper[4907]: I0313 15:47:28.318276 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-tthk2" Mar 13 15:47:28 crc kubenswrapper[4907]: I0313 15:47:28.318323 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-tthk2" Mar 13 15:47:28 crc kubenswrapper[4907]: I0313 15:47:28.873665 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-6d79dff969-ssj4t" podUID="5315ef54-fd95-400c-a7ef-a356c0f867bc" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.159:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.159:8080: connect: connection refused" Mar 13 15:47:29 crc kubenswrapper[4907]: I0313 15:47:29.394435 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-tthk2" podUID="4ecb3195-45a5-4ca4-b979-9a8c37ec4fbe" containerName="registry-server" probeResult="failure" output=< Mar 13 15:47:29 crc kubenswrapper[4907]: timeout: failed to connect service ":50051" within 1s Mar 13 15:47:29 crc kubenswrapper[4907]: > Mar 13 15:47:29 crc kubenswrapper[4907]: I0313 15:47:29.425319 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-mqnk8" Mar 13 15:47:29 crc kubenswrapper[4907]: I0313 15:47:29.556466 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6df151b5-1c14-488c-9f23-9d6d629b493d-config-data\") pod \"6df151b5-1c14-488c-9f23-9d6d629b493d\" (UID: \"6df151b5-1c14-488c-9f23-9d6d629b493d\") " Mar 13 15:47:29 crc kubenswrapper[4907]: I0313 15:47:29.556625 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6df151b5-1c14-488c-9f23-9d6d629b493d-combined-ca-bundle\") pod \"6df151b5-1c14-488c-9f23-9d6d629b493d\" (UID: \"6df151b5-1c14-488c-9f23-9d6d629b493d\") " Mar 13 15:47:29 crc kubenswrapper[4907]: I0313 15:47:29.556670 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jzrmm\" (UniqueName: \"kubernetes.io/projected/6df151b5-1c14-488c-9f23-9d6d629b493d-kube-api-access-jzrmm\") pod \"6df151b5-1c14-488c-9f23-9d6d629b493d\" (UID: \"6df151b5-1c14-488c-9f23-9d6d629b493d\") " Mar 13 15:47:29 crc kubenswrapper[4907]: I0313 15:47:29.562335 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6df151b5-1c14-488c-9f23-9d6d629b493d-kube-api-access-jzrmm" (OuterVolumeSpecName: "kube-api-access-jzrmm") pod "6df151b5-1c14-488c-9f23-9d6d629b493d" (UID: "6df151b5-1c14-488c-9f23-9d6d629b493d"). InnerVolumeSpecName "kube-api-access-jzrmm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:47:29 crc kubenswrapper[4907]: I0313 15:47:29.595904 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6df151b5-1c14-488c-9f23-9d6d629b493d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6df151b5-1c14-488c-9f23-9d6d629b493d" (UID: "6df151b5-1c14-488c-9f23-9d6d629b493d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:47:29 crc kubenswrapper[4907]: I0313 15:47:29.639600 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6df151b5-1c14-488c-9f23-9d6d629b493d-config-data" (OuterVolumeSpecName: "config-data") pod "6df151b5-1c14-488c-9f23-9d6d629b493d" (UID: "6df151b5-1c14-488c-9f23-9d6d629b493d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:47:29 crc kubenswrapper[4907]: I0313 15:47:29.659153 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6df151b5-1c14-488c-9f23-9d6d629b493d-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 15:47:29 crc kubenswrapper[4907]: I0313 15:47:29.659191 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6df151b5-1c14-488c-9f23-9d6d629b493d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 15:47:29 crc kubenswrapper[4907]: I0313 15:47:29.659213 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jzrmm\" (UniqueName: \"kubernetes.io/projected/6df151b5-1c14-488c-9f23-9d6d629b493d-kube-api-access-jzrmm\") on node \"crc\" DevicePath \"\"" Mar 13 15:47:30 crc kubenswrapper[4907]: I0313 15:47:30.077300 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-mqnk8" event={"ID":"6df151b5-1c14-488c-9f23-9d6d629b493d","Type":"ContainerDied","Data":"3d55337bfec3fbfeaf31255470fe615e54c37d18887cd6469654963c1c88fe12"} Mar 13 15:47:30 crc kubenswrapper[4907]: I0313 15:47:30.077634 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3d55337bfec3fbfeaf31255470fe615e54c37d18887cd6469654963c1c88fe12" Mar 13 15:47:30 crc kubenswrapper[4907]: I0313 15:47:30.077376 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-mqnk8" Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.269935 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-engine-86cbbf5dd8-25c77"] Mar 13 15:47:31 crc kubenswrapper[4907]: E0313 15:47:31.270383 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6df151b5-1c14-488c-9f23-9d6d629b493d" containerName="heat-db-sync" Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.270394 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="6df151b5-1c14-488c-9f23-9d6d629b493d" containerName="heat-db-sync" Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.270596 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="6df151b5-1c14-488c-9f23-9d6d629b493d" containerName="heat-db-sync" Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.271283 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-86cbbf5dd8-25c77" Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.282015 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-engine-config-data" Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.282323 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-config-data" Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.282488 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-heat-dockercfg-mcdkh" Mar 13 15:47:31 crc kubenswrapper[4907]: W0313 15:47:31.293569 4907 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podce0b08b9_c4e1_4236_8f54_cf78b45d2d49.slice/crio-d2ee4f22d00d1f9ceaf1d6575365b027f64e73a1473c8ba545cf5581cfc37e71": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podce0b08b9_c4e1_4236_8f54_cf78b45d2d49.slice/crio-d2ee4f22d00d1f9ceaf1d6575365b027f64e73a1473c8ba545cf5581cfc37e71: no such file or directory Mar 13 15:47:31 crc kubenswrapper[4907]: W0313 15:47:31.293642 4907 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podce0b08b9_c4e1_4236_8f54_cf78b45d2d49.slice/crio-conmon-839f8683996fb111c90ced80fc315df41501ccd29aa3ce62746269e8e8d6d880.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podce0b08b9_c4e1_4236_8f54_cf78b45d2d49.slice/crio-conmon-839f8683996fb111c90ced80fc315df41501ccd29aa3ce62746269e8e8d6d880.scope: no such file or directory Mar 13 15:47:31 crc kubenswrapper[4907]: W0313 15:47:31.293665 4907 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod490bd531_09f6_48e5_b194_b79e9b532131.slice/crio-daa82cb23c63e51c880db96744e0980e9ddf6f38f0cbcf2713496cba5c2d4486.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod490bd531_09f6_48e5_b194_b79e9b532131.slice/crio-daa82cb23c63e51c880db96744e0980e9ddf6f38f0cbcf2713496cba5c2d4486.scope: no such file or directory Mar 13 15:47:31 crc kubenswrapper[4907]: W0313 15:47:31.293680 4907 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podce0b08b9_c4e1_4236_8f54_cf78b45d2d49.slice/crio-839f8683996fb111c90ced80fc315df41501ccd29aa3ce62746269e8e8d6d880.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podce0b08b9_c4e1_4236_8f54_cf78b45d2d49.slice/crio-839f8683996fb111c90ced80fc315df41501ccd29aa3ce62746269e8e8d6d880.scope: no such file or directory Mar 13 15:47:31 crc kubenswrapper[4907]: W0313 15:47:31.293754 4907 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6df151b5_1c14_488c_9f23_9d6d629b493d.slice": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6df151b5_1c14_488c_9f23_9d6d629b493d.slice: no such file or directory Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.321054 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-86cbbf5dd8-25c77"] Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.383856 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-7fd69496d-xglm5"] Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.390062 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3dd72cd1-7342-40aa-8c3c-9d14375b3930-config-data-custom\") pod \"heat-engine-86cbbf5dd8-25c77\" (UID: \"3dd72cd1-7342-40aa-8c3c-9d14375b3930\") " pod="openstack/heat-engine-86cbbf5dd8-25c77" Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.390163 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3dd72cd1-7342-40aa-8c3c-9d14375b3930-combined-ca-bundle\") pod \"heat-engine-86cbbf5dd8-25c77\" (UID: \"3dd72cd1-7342-40aa-8c3c-9d14375b3930\") " pod="openstack/heat-engine-86cbbf5dd8-25c77" Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.390226 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jfmqz\" (UniqueName: \"kubernetes.io/projected/3dd72cd1-7342-40aa-8c3c-9d14375b3930-kube-api-access-jfmqz\") pod \"heat-engine-86cbbf5dd8-25c77\" (UID: \"3dd72cd1-7342-40aa-8c3c-9d14375b3930\") " pod="openstack/heat-engine-86cbbf5dd8-25c77" Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.390430 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3dd72cd1-7342-40aa-8c3c-9d14375b3930-config-data\") pod \"heat-engine-86cbbf5dd8-25c77\" (UID: \"3dd72cd1-7342-40aa-8c3c-9d14375b3930\") " pod="openstack/heat-engine-86cbbf5dd8-25c77" Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.405243 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-7fd69496d-xglm5"] Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.405366 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-7fd69496d-xglm5" Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.412264 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-api-config-data" Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.426996 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-7dbbf4d5b8-846ft"] Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.428230 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-7dbbf4d5b8-846ft" Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.429935 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-cfnapi-config-data" Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.443549 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-7dbbf4d5b8-846ft"] Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.493286 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a76ba2c0-c077-44ad-a2ed-2ee4f1a0052a-config-data\") pod \"heat-api-7fd69496d-xglm5\" (UID: \"a76ba2c0-c077-44ad-a2ed-2ee4f1a0052a\") " pod="openstack/heat-api-7fd69496d-xglm5" Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.493380 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3dd72cd1-7342-40aa-8c3c-9d14375b3930-config-data\") pod \"heat-engine-86cbbf5dd8-25c77\" (UID: \"3dd72cd1-7342-40aa-8c3c-9d14375b3930\") " pod="openstack/heat-engine-86cbbf5dd8-25c77" Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.493418 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a76ba2c0-c077-44ad-a2ed-2ee4f1a0052a-combined-ca-bundle\") pod \"heat-api-7fd69496d-xglm5\" (UID: \"a76ba2c0-c077-44ad-a2ed-2ee4f1a0052a\") " pod="openstack/heat-api-7fd69496d-xglm5" Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.493435 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d7cc945-4d68-41c8-9ea0-359af3279cc8-combined-ca-bundle\") pod \"heat-cfnapi-7dbbf4d5b8-846ft\" (UID: \"1d7cc945-4d68-41c8-9ea0-359af3279cc8\") " pod="openstack/heat-cfnapi-7dbbf4d5b8-846ft" Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.493466 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3dd72cd1-7342-40aa-8c3c-9d14375b3930-config-data-custom\") pod \"heat-engine-86cbbf5dd8-25c77\" (UID: \"3dd72cd1-7342-40aa-8c3c-9d14375b3930\") " pod="openstack/heat-engine-86cbbf5dd8-25c77" Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.493487 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a76ba2c0-c077-44ad-a2ed-2ee4f1a0052a-config-data-custom\") pod \"heat-api-7fd69496d-xglm5\" (UID: \"a76ba2c0-c077-44ad-a2ed-2ee4f1a0052a\") " pod="openstack/heat-api-7fd69496d-xglm5" Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.493524 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3dd72cd1-7342-40aa-8c3c-9d14375b3930-combined-ca-bundle\") pod \"heat-engine-86cbbf5dd8-25c77\" (UID: \"3dd72cd1-7342-40aa-8c3c-9d14375b3930\") " pod="openstack/heat-engine-86cbbf5dd8-25c77" Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.493544 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d7cc945-4d68-41c8-9ea0-359af3279cc8-config-data\") pod \"heat-cfnapi-7dbbf4d5b8-846ft\" (UID: \"1d7cc945-4d68-41c8-9ea0-359af3279cc8\") " pod="openstack/heat-cfnapi-7dbbf4d5b8-846ft" Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.493565 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p8c5v\" (UniqueName: \"kubernetes.io/projected/1d7cc945-4d68-41c8-9ea0-359af3279cc8-kube-api-access-p8c5v\") pod \"heat-cfnapi-7dbbf4d5b8-846ft\" (UID: \"1d7cc945-4d68-41c8-9ea0-359af3279cc8\") " pod="openstack/heat-cfnapi-7dbbf4d5b8-846ft" Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.493593 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cw9pz\" (UniqueName: \"kubernetes.io/projected/a76ba2c0-c077-44ad-a2ed-2ee4f1a0052a-kube-api-access-cw9pz\") pod \"heat-api-7fd69496d-xglm5\" (UID: \"a76ba2c0-c077-44ad-a2ed-2ee4f1a0052a\") " pod="openstack/heat-api-7fd69496d-xglm5" Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.493616 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jfmqz\" (UniqueName: \"kubernetes.io/projected/3dd72cd1-7342-40aa-8c3c-9d14375b3930-kube-api-access-jfmqz\") pod \"heat-engine-86cbbf5dd8-25c77\" (UID: \"3dd72cd1-7342-40aa-8c3c-9d14375b3930\") " pod="openstack/heat-engine-86cbbf5dd8-25c77" Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.493638 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1d7cc945-4d68-41c8-9ea0-359af3279cc8-config-data-custom\") pod \"heat-cfnapi-7dbbf4d5b8-846ft\" (UID: \"1d7cc945-4d68-41c8-9ea0-359af3279cc8\") " pod="openstack/heat-cfnapi-7dbbf4d5b8-846ft" Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.514684 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3dd72cd1-7342-40aa-8c3c-9d14375b3930-config-data\") pod \"heat-engine-86cbbf5dd8-25c77\" (UID: \"3dd72cd1-7342-40aa-8c3c-9d14375b3930\") " pod="openstack/heat-engine-86cbbf5dd8-25c77" Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.515599 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3dd72cd1-7342-40aa-8c3c-9d14375b3930-config-data-custom\") pod \"heat-engine-86cbbf5dd8-25c77\" (UID: \"3dd72cd1-7342-40aa-8c3c-9d14375b3930\") " pod="openstack/heat-engine-86cbbf5dd8-25c77" Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.519680 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jfmqz\" (UniqueName: \"kubernetes.io/projected/3dd72cd1-7342-40aa-8c3c-9d14375b3930-kube-api-access-jfmqz\") pod \"heat-engine-86cbbf5dd8-25c77\" (UID: \"3dd72cd1-7342-40aa-8c3c-9d14375b3930\") " pod="openstack/heat-engine-86cbbf5dd8-25c77" Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.526661 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3dd72cd1-7342-40aa-8c3c-9d14375b3930-combined-ca-bundle\") pod \"heat-engine-86cbbf5dd8-25c77\" (UID: \"3dd72cd1-7342-40aa-8c3c-9d14375b3930\") " pod="openstack/heat-engine-86cbbf5dd8-25c77" Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.596856 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a76ba2c0-c077-44ad-a2ed-2ee4f1a0052a-config-data\") pod \"heat-api-7fd69496d-xglm5\" (UID: \"a76ba2c0-c077-44ad-a2ed-2ee4f1a0052a\") " pod="openstack/heat-api-7fd69496d-xglm5" Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.597059 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a76ba2c0-c077-44ad-a2ed-2ee4f1a0052a-combined-ca-bundle\") pod \"heat-api-7fd69496d-xglm5\" (UID: \"a76ba2c0-c077-44ad-a2ed-2ee4f1a0052a\") " pod="openstack/heat-api-7fd69496d-xglm5" Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.597085 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d7cc945-4d68-41c8-9ea0-359af3279cc8-combined-ca-bundle\") pod \"heat-cfnapi-7dbbf4d5b8-846ft\" (UID: \"1d7cc945-4d68-41c8-9ea0-359af3279cc8\") " pod="openstack/heat-cfnapi-7dbbf4d5b8-846ft" Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.597155 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a76ba2c0-c077-44ad-a2ed-2ee4f1a0052a-config-data-custom\") pod \"heat-api-7fd69496d-xglm5\" (UID: \"a76ba2c0-c077-44ad-a2ed-2ee4f1a0052a\") " pod="openstack/heat-api-7fd69496d-xglm5" Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.597238 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d7cc945-4d68-41c8-9ea0-359af3279cc8-config-data\") pod \"heat-cfnapi-7dbbf4d5b8-846ft\" (UID: \"1d7cc945-4d68-41c8-9ea0-359af3279cc8\") " pod="openstack/heat-cfnapi-7dbbf4d5b8-846ft" Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.597272 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p8c5v\" (UniqueName: \"kubernetes.io/projected/1d7cc945-4d68-41c8-9ea0-359af3279cc8-kube-api-access-p8c5v\") pod \"heat-cfnapi-7dbbf4d5b8-846ft\" (UID: \"1d7cc945-4d68-41c8-9ea0-359af3279cc8\") " pod="openstack/heat-cfnapi-7dbbf4d5b8-846ft" Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.597319 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cw9pz\" (UniqueName: \"kubernetes.io/projected/a76ba2c0-c077-44ad-a2ed-2ee4f1a0052a-kube-api-access-cw9pz\") pod \"heat-api-7fd69496d-xglm5\" (UID: \"a76ba2c0-c077-44ad-a2ed-2ee4f1a0052a\") " pod="openstack/heat-api-7fd69496d-xglm5" Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.597355 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1d7cc945-4d68-41c8-9ea0-359af3279cc8-config-data-custom\") pod \"heat-cfnapi-7dbbf4d5b8-846ft\" (UID: \"1d7cc945-4d68-41c8-9ea0-359af3279cc8\") " pod="openstack/heat-cfnapi-7dbbf4d5b8-846ft" Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.604786 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1d7cc945-4d68-41c8-9ea0-359af3279cc8-config-data-custom\") pod \"heat-cfnapi-7dbbf4d5b8-846ft\" (UID: \"1d7cc945-4d68-41c8-9ea0-359af3279cc8\") " pod="openstack/heat-cfnapi-7dbbf4d5b8-846ft" Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.604853 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a76ba2c0-c077-44ad-a2ed-2ee4f1a0052a-config-data\") pod \"heat-api-7fd69496d-xglm5\" (UID: \"a76ba2c0-c077-44ad-a2ed-2ee4f1a0052a\") " pod="openstack/heat-api-7fd69496d-xglm5" Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.608976 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a76ba2c0-c077-44ad-a2ed-2ee4f1a0052a-config-data-custom\") pod \"heat-api-7fd69496d-xglm5\" (UID: \"a76ba2c0-c077-44ad-a2ed-2ee4f1a0052a\") " pod="openstack/heat-api-7fd69496d-xglm5" Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.614707 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d7cc945-4d68-41c8-9ea0-359af3279cc8-config-data\") pod \"heat-cfnapi-7dbbf4d5b8-846ft\" (UID: \"1d7cc945-4d68-41c8-9ea0-359af3279cc8\") " pod="openstack/heat-cfnapi-7dbbf4d5b8-846ft" Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.616695 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p8c5v\" (UniqueName: \"kubernetes.io/projected/1d7cc945-4d68-41c8-9ea0-359af3279cc8-kube-api-access-p8c5v\") pod \"heat-cfnapi-7dbbf4d5b8-846ft\" (UID: \"1d7cc945-4d68-41c8-9ea0-359af3279cc8\") " pod="openstack/heat-cfnapi-7dbbf4d5b8-846ft" Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.623843 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a76ba2c0-c077-44ad-a2ed-2ee4f1a0052a-combined-ca-bundle\") pod \"heat-api-7fd69496d-xglm5\" (UID: \"a76ba2c0-c077-44ad-a2ed-2ee4f1a0052a\") " pod="openstack/heat-api-7fd69496d-xglm5" Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.623928 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d7cc945-4d68-41c8-9ea0-359af3279cc8-combined-ca-bundle\") pod \"heat-cfnapi-7dbbf4d5b8-846ft\" (UID: \"1d7cc945-4d68-41c8-9ea0-359af3279cc8\") " pod="openstack/heat-cfnapi-7dbbf4d5b8-846ft" Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.630723 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cw9pz\" (UniqueName: \"kubernetes.io/projected/a76ba2c0-c077-44ad-a2ed-2ee4f1a0052a-kube-api-access-cw9pz\") pod \"heat-api-7fd69496d-xglm5\" (UID: \"a76ba2c0-c077-44ad-a2ed-2ee4f1a0052a\") " pod="openstack/heat-api-7fd69496d-xglm5" Mar 13 15:47:31 crc kubenswrapper[4907]: E0313 15:47:31.641846 4907 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod099b49c3_666e_4bf5_9c15_32cf9c40e1e7.slice/crio-e4caf9d5f22ffb193c93c87c06b4e64f7a29f8bf0582a1a8b7f47762f5ba55eb\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2d9fbab8_539c_48f6_8cae_3eda23300b28.slice/crio-conmon-d405f8c27cb2c1dbfd5ce42b31129e474a50d6e8ddffc33a12dfa64551646e29.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2d9fbab8_539c_48f6_8cae_3eda23300b28.slice/crio-d405f8c27cb2c1dbfd5ce42b31129e474a50d6e8ddffc33a12dfa64551646e29.scope\": RecentStats: unable to find data in memory cache]" Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.726016 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-86cbbf5dd8-25c77" Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.754982 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-c7977bcd9-2vs8h" Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.767922 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-7fd69496d-xglm5" Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.776486 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-7dbbf4d5b8-846ft" Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.904002 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2d9fbab8-539c-48f6-8cae-3eda23300b28-logs\") pod \"2d9fbab8-539c-48f6-8cae-3eda23300b28\" (UID: \"2d9fbab8-539c-48f6-8cae-3eda23300b28\") " Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.904094 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/2d9fbab8-539c-48f6-8cae-3eda23300b28-horizon-secret-key\") pod \"2d9fbab8-539c-48f6-8cae-3eda23300b28\" (UID: \"2d9fbab8-539c-48f6-8cae-3eda23300b28\") " Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.904201 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2d9fbab8-539c-48f6-8cae-3eda23300b28-scripts\") pod \"2d9fbab8-539c-48f6-8cae-3eda23300b28\" (UID: \"2d9fbab8-539c-48f6-8cae-3eda23300b28\") " Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.904297 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2d9fbab8-539c-48f6-8cae-3eda23300b28-config-data\") pod \"2d9fbab8-539c-48f6-8cae-3eda23300b28\" (UID: \"2d9fbab8-539c-48f6-8cae-3eda23300b28\") " Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.904418 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bjhcq\" (UniqueName: \"kubernetes.io/projected/2d9fbab8-539c-48f6-8cae-3eda23300b28-kube-api-access-bjhcq\") pod \"2d9fbab8-539c-48f6-8cae-3eda23300b28\" (UID: \"2d9fbab8-539c-48f6-8cae-3eda23300b28\") " Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.904607 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2d9fbab8-539c-48f6-8cae-3eda23300b28-logs" (OuterVolumeSpecName: "logs") pod "2d9fbab8-539c-48f6-8cae-3eda23300b28" (UID: "2d9fbab8-539c-48f6-8cae-3eda23300b28"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.905015 4907 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2d9fbab8-539c-48f6-8cae-3eda23300b28-logs\") on node \"crc\" DevicePath \"\"" Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.910009 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d9fbab8-539c-48f6-8cae-3eda23300b28-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "2d9fbab8-539c-48f6-8cae-3eda23300b28" (UID: "2d9fbab8-539c-48f6-8cae-3eda23300b28"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.918778 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2d9fbab8-539c-48f6-8cae-3eda23300b28-kube-api-access-bjhcq" (OuterVolumeSpecName: "kube-api-access-bjhcq") pod "2d9fbab8-539c-48f6-8cae-3eda23300b28" (UID: "2d9fbab8-539c-48f6-8cae-3eda23300b28"). InnerVolumeSpecName "kube-api-access-bjhcq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.944909 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2d9fbab8-539c-48f6-8cae-3eda23300b28-scripts" (OuterVolumeSpecName: "scripts") pod "2d9fbab8-539c-48f6-8cae-3eda23300b28" (UID: "2d9fbab8-539c-48f6-8cae-3eda23300b28"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.979093 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2d9fbab8-539c-48f6-8cae-3eda23300b28-config-data" (OuterVolumeSpecName: "config-data") pod "2d9fbab8-539c-48f6-8cae-3eda23300b28" (UID: "2d9fbab8-539c-48f6-8cae-3eda23300b28"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:47:32 crc kubenswrapper[4907]: I0313 15:47:32.007400 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2d9fbab8-539c-48f6-8cae-3eda23300b28-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 15:47:32 crc kubenswrapper[4907]: I0313 15:47:32.007443 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bjhcq\" (UniqueName: \"kubernetes.io/projected/2d9fbab8-539c-48f6-8cae-3eda23300b28-kube-api-access-bjhcq\") on node \"crc\" DevicePath \"\"" Mar 13 15:47:32 crc kubenswrapper[4907]: I0313 15:47:32.007543 4907 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/2d9fbab8-539c-48f6-8cae-3eda23300b28-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Mar 13 15:47:32 crc kubenswrapper[4907]: I0313 15:47:32.007556 4907 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2d9fbab8-539c-48f6-8cae-3eda23300b28-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 15:47:32 crc kubenswrapper[4907]: I0313 15:47:32.099143 4907 generic.go:334] "Generic (PLEG): container finished" podID="2d9fbab8-539c-48f6-8cae-3eda23300b28" containerID="d405f8c27cb2c1dbfd5ce42b31129e474a50d6e8ddffc33a12dfa64551646e29" exitCode=137 Mar 13 15:47:32 crc kubenswrapper[4907]: I0313 15:47:32.099187 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-c7977bcd9-2vs8h" event={"ID":"2d9fbab8-539c-48f6-8cae-3eda23300b28","Type":"ContainerDied","Data":"d405f8c27cb2c1dbfd5ce42b31129e474a50d6e8ddffc33a12dfa64551646e29"} Mar 13 15:47:32 crc kubenswrapper[4907]: I0313 15:47:32.099211 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-c7977bcd9-2vs8h" event={"ID":"2d9fbab8-539c-48f6-8cae-3eda23300b28","Type":"ContainerDied","Data":"941b229bb7806138fc500a8ef3b0ed37c2227cce03ef0b5260159ccb8082332d"} Mar 13 15:47:32 crc kubenswrapper[4907]: I0313 15:47:32.099226 4907 scope.go:117] "RemoveContainer" containerID="64ab189d02c2d9e2e9123ec6de24b5d5f7db40d7ceaa4b16356eefe21fdc102c" Mar 13 15:47:32 crc kubenswrapper[4907]: I0313 15:47:32.099401 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-c7977bcd9-2vs8h" Mar 13 15:47:32 crc kubenswrapper[4907]: I0313 15:47:32.150869 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-c7977bcd9-2vs8h"] Mar 13 15:47:32 crc kubenswrapper[4907]: I0313 15:47:32.160384 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-c7977bcd9-2vs8h"] Mar 13 15:47:32 crc kubenswrapper[4907]: I0313 15:47:32.286699 4907 scope.go:117] "RemoveContainer" containerID="d405f8c27cb2c1dbfd5ce42b31129e474a50d6e8ddffc33a12dfa64551646e29" Mar 13 15:47:32 crc kubenswrapper[4907]: I0313 15:47:32.318174 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-86cbbf5dd8-25c77"] Mar 13 15:47:32 crc kubenswrapper[4907]: I0313 15:47:32.378841 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-7dbbf4d5b8-846ft"] Mar 13 15:47:32 crc kubenswrapper[4907]: I0313 15:47:32.467986 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-7fd69496d-xglm5"] Mar 13 15:47:32 crc kubenswrapper[4907]: I0313 15:47:32.532520 4907 scope.go:117] "RemoveContainer" containerID="64ab189d02c2d9e2e9123ec6de24b5d5f7db40d7ceaa4b16356eefe21fdc102c" Mar 13 15:47:32 crc kubenswrapper[4907]: E0313 15:47:32.533015 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"64ab189d02c2d9e2e9123ec6de24b5d5f7db40d7ceaa4b16356eefe21fdc102c\": container with ID starting with 64ab189d02c2d9e2e9123ec6de24b5d5f7db40d7ceaa4b16356eefe21fdc102c not found: ID does not exist" containerID="64ab189d02c2d9e2e9123ec6de24b5d5f7db40d7ceaa4b16356eefe21fdc102c" Mar 13 15:47:32 crc kubenswrapper[4907]: I0313 15:47:32.533062 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"64ab189d02c2d9e2e9123ec6de24b5d5f7db40d7ceaa4b16356eefe21fdc102c"} err="failed to get container status \"64ab189d02c2d9e2e9123ec6de24b5d5f7db40d7ceaa4b16356eefe21fdc102c\": rpc error: code = NotFound desc = could not find container \"64ab189d02c2d9e2e9123ec6de24b5d5f7db40d7ceaa4b16356eefe21fdc102c\": container with ID starting with 64ab189d02c2d9e2e9123ec6de24b5d5f7db40d7ceaa4b16356eefe21fdc102c not found: ID does not exist" Mar 13 15:47:32 crc kubenswrapper[4907]: I0313 15:47:32.533088 4907 scope.go:117] "RemoveContainer" containerID="d405f8c27cb2c1dbfd5ce42b31129e474a50d6e8ddffc33a12dfa64551646e29" Mar 13 15:47:32 crc kubenswrapper[4907]: E0313 15:47:32.533414 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d405f8c27cb2c1dbfd5ce42b31129e474a50d6e8ddffc33a12dfa64551646e29\": container with ID starting with d405f8c27cb2c1dbfd5ce42b31129e474a50d6e8ddffc33a12dfa64551646e29 not found: ID does not exist" containerID="d405f8c27cb2c1dbfd5ce42b31129e474a50d6e8ddffc33a12dfa64551646e29" Mar 13 15:47:32 crc kubenswrapper[4907]: I0313 15:47:32.533450 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d405f8c27cb2c1dbfd5ce42b31129e474a50d6e8ddffc33a12dfa64551646e29"} err="failed to get container status \"d405f8c27cb2c1dbfd5ce42b31129e474a50d6e8ddffc33a12dfa64551646e29\": rpc error: code = NotFound desc = could not find container \"d405f8c27cb2c1dbfd5ce42b31129e474a50d6e8ddffc33a12dfa64551646e29\": container with ID starting with d405f8c27cb2c1dbfd5ce42b31129e474a50d6e8ddffc33a12dfa64551646e29 not found: ID does not exist" Mar 13 15:47:33 crc kubenswrapper[4907]: I0313 15:47:33.121113 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-7dbbf4d5b8-846ft" event={"ID":"1d7cc945-4d68-41c8-9ea0-359af3279cc8","Type":"ContainerStarted","Data":"fea6e5248ff50df8d5cb262ab89780fb4759d674abef9321118b834ea41cbfcb"} Mar 13 15:47:33 crc kubenswrapper[4907]: I0313 15:47:33.130566 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-86cbbf5dd8-25c77" event={"ID":"3dd72cd1-7342-40aa-8c3c-9d14375b3930","Type":"ContainerStarted","Data":"d82e65a7846c7b8476e0e378027abc6125bb877534154e1e4f897fad13bd3dd2"} Mar 13 15:47:33 crc kubenswrapper[4907]: I0313 15:47:33.130605 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-86cbbf5dd8-25c77" event={"ID":"3dd72cd1-7342-40aa-8c3c-9d14375b3930","Type":"ContainerStarted","Data":"dacedb00970328528f11eabf2aca603e86d8a8f608fcd8855b8d709b6f89b644"} Mar 13 15:47:33 crc kubenswrapper[4907]: I0313 15:47:33.130700 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-engine-86cbbf5dd8-25c77" Mar 13 15:47:33 crc kubenswrapper[4907]: I0313 15:47:33.143288 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-7fd69496d-xglm5" event={"ID":"a76ba2c0-c077-44ad-a2ed-2ee4f1a0052a","Type":"ContainerStarted","Data":"2c4d11de023cb7ff9633f51fe1bcfa511e7ed99c116116d286b152f650b5ab8b"} Mar 13 15:47:33 crc kubenswrapper[4907]: I0313 15:47:33.159723 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-engine-86cbbf5dd8-25c77" podStartSLOduration=2.159703351 podStartE2EDuration="2.159703351s" podCreationTimestamp="2026-03-13 15:47:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:47:33.149610555 +0000 UTC m=+6152.049398244" watchObservedRunningTime="2026-03-13 15:47:33.159703351 +0000 UTC m=+6152.059491040" Mar 13 15:47:33 crc kubenswrapper[4907]: I0313 15:47:33.801943 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2d9fbab8-539c-48f6-8cae-3eda23300b28" path="/var/lib/kubelet/pods/2d9fbab8-539c-48f6-8cae-3eda23300b28/volumes" Mar 13 15:47:34 crc kubenswrapper[4907]: I0313 15:47:34.755706 4907 scope.go:117] "RemoveContainer" containerID="ca617bcf57734b16a051ae67c35004e1d491c4d4b7be39c47467fe81b31f47b6" Mar 13 15:47:34 crc kubenswrapper[4907]: I0313 15:47:34.797901 4907 scope.go:117] "RemoveContainer" containerID="29bfa2eaa41b750e3b4b9e54874f96fd9d36ce47ad55875b7c1d9bd6b63f1eef" Mar 13 15:47:34 crc kubenswrapper[4907]: I0313 15:47:34.871110 4907 scope.go:117] "RemoveContainer" containerID="eb5f297398bf730822a16b0adfb1835959f9d7d48e8a8a2e6d33b8a74b86db31" Mar 13 15:47:34 crc kubenswrapper[4907]: I0313 15:47:34.942979 4907 scope.go:117] "RemoveContainer" containerID="2d8f130696e7f9d5999a1090959bb47d9418a9bc337a9f062febc9b4224a5eb9" Mar 13 15:47:34 crc kubenswrapper[4907]: I0313 15:47:34.990189 4907 scope.go:117] "RemoveContainer" containerID="6b2aa6e8ccbdaa4de43a0268ebdb8841de04a3498e495de6ed723c183300253b" Mar 13 15:47:35 crc kubenswrapper[4907]: I0313 15:47:35.037006 4907 scope.go:117] "RemoveContainer" containerID="0e271cdd3f12ce6ad2a6fc124e7bdced8c6ac6c38d7c47a03deae9761bf3f57a" Mar 13 15:47:35 crc kubenswrapper[4907]: I0313 15:47:35.179667 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-7fd69496d-xglm5" event={"ID":"a76ba2c0-c077-44ad-a2ed-2ee4f1a0052a","Type":"ContainerStarted","Data":"0c5fc1a9c34f912d4a6872a0f9e97bab3fc66f28df1bf78d8be3f69d33aabd71"} Mar 13 15:47:35 crc kubenswrapper[4907]: I0313 15:47:35.179751 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-7fd69496d-xglm5" Mar 13 15:47:35 crc kubenswrapper[4907]: I0313 15:47:35.191298 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-7dbbf4d5b8-846ft" event={"ID":"1d7cc945-4d68-41c8-9ea0-359af3279cc8","Type":"ContainerStarted","Data":"b32704f8f08f4c5c467b88eaaa1d87c264c5f61cec547c047a8e6beb93271ea8"} Mar 13 15:47:35 crc kubenswrapper[4907]: I0313 15:47:35.192403 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-7dbbf4d5b8-846ft" Mar 13 15:47:35 crc kubenswrapper[4907]: I0313 15:47:35.222418 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-7fd69496d-xglm5" podStartSLOduration=2.537614197 podStartE2EDuration="4.222398938s" podCreationTimestamp="2026-03-13 15:47:31 +0000 UTC" firstStartedPulling="2026-03-13 15:47:32.461785113 +0000 UTC m=+6151.361572792" lastFinishedPulling="2026-03-13 15:47:34.146569844 +0000 UTC m=+6153.046357533" observedRunningTime="2026-03-13 15:47:35.204434067 +0000 UTC m=+6154.104221756" watchObservedRunningTime="2026-03-13 15:47:35.222398938 +0000 UTC m=+6154.122186647" Mar 13 15:47:35 crc kubenswrapper[4907]: I0313 15:47:35.227575 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-7dbbf4d5b8-846ft" podStartSLOduration=2.473599566 podStartE2EDuration="4.227558458s" podCreationTimestamp="2026-03-13 15:47:31 +0000 UTC" firstStartedPulling="2026-03-13 15:47:32.389048235 +0000 UTC m=+6151.288835914" lastFinishedPulling="2026-03-13 15:47:34.143007117 +0000 UTC m=+6153.042794806" observedRunningTime="2026-03-13 15:47:35.223931149 +0000 UTC m=+6154.123718858" watchObservedRunningTime="2026-03-13 15:47:35.227558458 +0000 UTC m=+6154.127346147" Mar 13 15:47:38 crc kubenswrapper[4907]: I0313 15:47:38.370029 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-tthk2" Mar 13 15:47:38 crc kubenswrapper[4907]: I0313 15:47:38.423974 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-tthk2" Mar 13 15:47:38 crc kubenswrapper[4907]: I0313 15:47:38.603955 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-tthk2"] Mar 13 15:47:40 crc kubenswrapper[4907]: I0313 15:47:40.233452 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-tthk2" podUID="4ecb3195-45a5-4ca4-b979-9a8c37ec4fbe" containerName="registry-server" containerID="cri-o://29ee2de0275600572b2cd057fb10e5905301ef9ab1753a5ad562223165c95348" gracePeriod=2 Mar 13 15:47:40 crc kubenswrapper[4907]: I0313 15:47:40.667214 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-6d79dff969-ssj4t" Mar 13 15:47:40 crc kubenswrapper[4907]: I0313 15:47:40.735874 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tthk2" Mar 13 15:47:40 crc kubenswrapper[4907]: I0313 15:47:40.823773 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4ecb3195-45a5-4ca4-b979-9a8c37ec4fbe-catalog-content\") pod \"4ecb3195-45a5-4ca4-b979-9a8c37ec4fbe\" (UID: \"4ecb3195-45a5-4ca4-b979-9a8c37ec4fbe\") " Mar 13 15:47:40 crc kubenswrapper[4907]: I0313 15:47:40.824482 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4ecb3195-45a5-4ca4-b979-9a8c37ec4fbe-utilities\") pod \"4ecb3195-45a5-4ca4-b979-9a8c37ec4fbe\" (UID: \"4ecb3195-45a5-4ca4-b979-9a8c37ec4fbe\") " Mar 13 15:47:40 crc kubenswrapper[4907]: I0313 15:47:40.824780 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6kbk5\" (UniqueName: \"kubernetes.io/projected/4ecb3195-45a5-4ca4-b979-9a8c37ec4fbe-kube-api-access-6kbk5\") pod \"4ecb3195-45a5-4ca4-b979-9a8c37ec4fbe\" (UID: \"4ecb3195-45a5-4ca4-b979-9a8c37ec4fbe\") " Mar 13 15:47:40 crc kubenswrapper[4907]: I0313 15:47:40.825268 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4ecb3195-45a5-4ca4-b979-9a8c37ec4fbe-utilities" (OuterVolumeSpecName: "utilities") pod "4ecb3195-45a5-4ca4-b979-9a8c37ec4fbe" (UID: "4ecb3195-45a5-4ca4-b979-9a8c37ec4fbe"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:47:40 crc kubenswrapper[4907]: I0313 15:47:40.825782 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4ecb3195-45a5-4ca4-b979-9a8c37ec4fbe-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 15:47:40 crc kubenswrapper[4907]: I0313 15:47:40.832545 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ecb3195-45a5-4ca4-b979-9a8c37ec4fbe-kube-api-access-6kbk5" (OuterVolumeSpecName: "kube-api-access-6kbk5") pod "4ecb3195-45a5-4ca4-b979-9a8c37ec4fbe" (UID: "4ecb3195-45a5-4ca4-b979-9a8c37ec4fbe"). InnerVolumeSpecName "kube-api-access-6kbk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:47:40 crc kubenswrapper[4907]: I0313 15:47:40.868670 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4ecb3195-45a5-4ca4-b979-9a8c37ec4fbe-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4ecb3195-45a5-4ca4-b979-9a8c37ec4fbe" (UID: "4ecb3195-45a5-4ca4-b979-9a8c37ec4fbe"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:47:40 crc kubenswrapper[4907]: I0313 15:47:40.928029 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6kbk5\" (UniqueName: \"kubernetes.io/projected/4ecb3195-45a5-4ca4-b979-9a8c37ec4fbe-kube-api-access-6kbk5\") on node \"crc\" DevicePath \"\"" Mar 13 15:47:40 crc kubenswrapper[4907]: I0313 15:47:40.928165 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4ecb3195-45a5-4ca4-b979-9a8c37ec4fbe-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 15:47:41 crc kubenswrapper[4907]: I0313 15:47:41.246907 4907 generic.go:334] "Generic (PLEG): container finished" podID="4ecb3195-45a5-4ca4-b979-9a8c37ec4fbe" containerID="29ee2de0275600572b2cd057fb10e5905301ef9ab1753a5ad562223165c95348" exitCode=0 Mar 13 15:47:41 crc kubenswrapper[4907]: I0313 15:47:41.246960 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tthk2" event={"ID":"4ecb3195-45a5-4ca4-b979-9a8c37ec4fbe","Type":"ContainerDied","Data":"29ee2de0275600572b2cd057fb10e5905301ef9ab1753a5ad562223165c95348"} Mar 13 15:47:41 crc kubenswrapper[4907]: I0313 15:47:41.247036 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tthk2" event={"ID":"4ecb3195-45a5-4ca4-b979-9a8c37ec4fbe","Type":"ContainerDied","Data":"ab17e9b088419fabf1ec951208897a9f96839403f2d2331918d09059ad41ce73"} Mar 13 15:47:41 crc kubenswrapper[4907]: I0313 15:47:41.247057 4907 scope.go:117] "RemoveContainer" containerID="29ee2de0275600572b2cd057fb10e5905301ef9ab1753a5ad562223165c95348" Mar 13 15:47:41 crc kubenswrapper[4907]: I0313 15:47:41.247015 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tthk2" Mar 13 15:47:41 crc kubenswrapper[4907]: I0313 15:47:41.276116 4907 scope.go:117] "RemoveContainer" containerID="dfea42d9a630e3e1c6c04649bc5e93bd04f889efb044725230049bd0d1de35ff" Mar 13 15:47:41 crc kubenswrapper[4907]: I0313 15:47:41.282778 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-tthk2"] Mar 13 15:47:41 crc kubenswrapper[4907]: I0313 15:47:41.291666 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-tthk2"] Mar 13 15:47:41 crc kubenswrapper[4907]: I0313 15:47:41.298274 4907 scope.go:117] "RemoveContainer" containerID="3fc3c676404f5dc909e1ad737ea131e306c394034b644f7d1fc0ccf0fd152916" Mar 13 15:47:41 crc kubenswrapper[4907]: I0313 15:47:41.349999 4907 scope.go:117] "RemoveContainer" containerID="29ee2de0275600572b2cd057fb10e5905301ef9ab1753a5ad562223165c95348" Mar 13 15:47:41 crc kubenswrapper[4907]: E0313 15:47:41.350555 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"29ee2de0275600572b2cd057fb10e5905301ef9ab1753a5ad562223165c95348\": container with ID starting with 29ee2de0275600572b2cd057fb10e5905301ef9ab1753a5ad562223165c95348 not found: ID does not exist" containerID="29ee2de0275600572b2cd057fb10e5905301ef9ab1753a5ad562223165c95348" Mar 13 15:47:41 crc kubenswrapper[4907]: I0313 15:47:41.350597 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"29ee2de0275600572b2cd057fb10e5905301ef9ab1753a5ad562223165c95348"} err="failed to get container status \"29ee2de0275600572b2cd057fb10e5905301ef9ab1753a5ad562223165c95348\": rpc error: code = NotFound desc = could not find container \"29ee2de0275600572b2cd057fb10e5905301ef9ab1753a5ad562223165c95348\": container with ID starting with 29ee2de0275600572b2cd057fb10e5905301ef9ab1753a5ad562223165c95348 not found: ID does not exist" Mar 13 15:47:41 crc kubenswrapper[4907]: I0313 15:47:41.350623 4907 scope.go:117] "RemoveContainer" containerID="dfea42d9a630e3e1c6c04649bc5e93bd04f889efb044725230049bd0d1de35ff" Mar 13 15:47:41 crc kubenswrapper[4907]: E0313 15:47:41.351264 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dfea42d9a630e3e1c6c04649bc5e93bd04f889efb044725230049bd0d1de35ff\": container with ID starting with dfea42d9a630e3e1c6c04649bc5e93bd04f889efb044725230049bd0d1de35ff not found: ID does not exist" containerID="dfea42d9a630e3e1c6c04649bc5e93bd04f889efb044725230049bd0d1de35ff" Mar 13 15:47:41 crc kubenswrapper[4907]: I0313 15:47:41.351315 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dfea42d9a630e3e1c6c04649bc5e93bd04f889efb044725230049bd0d1de35ff"} err="failed to get container status \"dfea42d9a630e3e1c6c04649bc5e93bd04f889efb044725230049bd0d1de35ff\": rpc error: code = NotFound desc = could not find container \"dfea42d9a630e3e1c6c04649bc5e93bd04f889efb044725230049bd0d1de35ff\": container with ID starting with dfea42d9a630e3e1c6c04649bc5e93bd04f889efb044725230049bd0d1de35ff not found: ID does not exist" Mar 13 15:47:41 crc kubenswrapper[4907]: I0313 15:47:41.351347 4907 scope.go:117] "RemoveContainer" containerID="3fc3c676404f5dc909e1ad737ea131e306c394034b644f7d1fc0ccf0fd152916" Mar 13 15:47:41 crc kubenswrapper[4907]: E0313 15:47:41.351689 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3fc3c676404f5dc909e1ad737ea131e306c394034b644f7d1fc0ccf0fd152916\": container with ID starting with 3fc3c676404f5dc909e1ad737ea131e306c394034b644f7d1fc0ccf0fd152916 not found: ID does not exist" containerID="3fc3c676404f5dc909e1ad737ea131e306c394034b644f7d1fc0ccf0fd152916" Mar 13 15:47:41 crc kubenswrapper[4907]: I0313 15:47:41.351720 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3fc3c676404f5dc909e1ad737ea131e306c394034b644f7d1fc0ccf0fd152916"} err="failed to get container status \"3fc3c676404f5dc909e1ad737ea131e306c394034b644f7d1fc0ccf0fd152916\": rpc error: code = NotFound desc = could not find container \"3fc3c676404f5dc909e1ad737ea131e306c394034b644f7d1fc0ccf0fd152916\": container with ID starting with 3fc3c676404f5dc909e1ad737ea131e306c394034b644f7d1fc0ccf0fd152916 not found: ID does not exist" Mar 13 15:47:41 crc kubenswrapper[4907]: I0313 15:47:41.806241 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4ecb3195-45a5-4ca4-b979-9a8c37ec4fbe" path="/var/lib/kubelet/pods/4ecb3195-45a5-4ca4-b979-9a8c37ec4fbe/volumes" Mar 13 15:47:41 crc kubenswrapper[4907]: E0313 15:47:41.888965 4907 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod099b49c3_666e_4bf5_9c15_32cf9c40e1e7.slice/crio-e4caf9d5f22ffb193c93c87c06b4e64f7a29f8bf0582a1a8b7f47762f5ba55eb\": RecentStats: unable to find data in memory cache]" Mar 13 15:47:42 crc kubenswrapper[4907]: I0313 15:47:42.241317 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-6d79dff969-ssj4t" Mar 13 15:47:42 crc kubenswrapper[4907]: I0313 15:47:42.351525 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-655cbf768f-dm9zz"] Mar 13 15:47:42 crc kubenswrapper[4907]: I0313 15:47:42.351815 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-655cbf768f-dm9zz" podUID="afd4001d-40a5-4a09-ab05-36cbd4d06eb4" containerName="horizon-log" containerID="cri-o://4d9d087b4ad0675d46c48144b039251dc336275fe3320abbe51816f548542e9f" gracePeriod=30 Mar 13 15:47:42 crc kubenswrapper[4907]: I0313 15:47:42.352404 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-655cbf768f-dm9zz" podUID="afd4001d-40a5-4a09-ab05-36cbd4d06eb4" containerName="horizon" containerID="cri-o://b42f8e33dfb6318b768e8c6b32af48621974ee97af5dab3f3e4bc582bbfc2931" gracePeriod=30 Mar 13 15:47:43 crc kubenswrapper[4907]: I0313 15:47:43.127301 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-api-7fd69496d-xglm5" Mar 13 15:47:43 crc kubenswrapper[4907]: I0313 15:47:43.203360 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-cfnapi-7dbbf4d5b8-846ft" Mar 13 15:47:46 crc kubenswrapper[4907]: I0313 15:47:46.294782 4907 generic.go:334] "Generic (PLEG): container finished" podID="afd4001d-40a5-4a09-ab05-36cbd4d06eb4" containerID="b42f8e33dfb6318b768e8c6b32af48621974ee97af5dab3f3e4bc582bbfc2931" exitCode=0 Mar 13 15:47:46 crc kubenswrapper[4907]: I0313 15:47:46.294861 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-655cbf768f-dm9zz" event={"ID":"afd4001d-40a5-4a09-ab05-36cbd4d06eb4","Type":"ContainerDied","Data":"b42f8e33dfb6318b768e8c6b32af48621974ee97af5dab3f3e4bc582bbfc2931"} Mar 13 15:47:47 crc kubenswrapper[4907]: I0313 15:47:47.488357 4907 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-655cbf768f-dm9zz" podUID="afd4001d-40a5-4a09-ab05-36cbd4d06eb4" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.156:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.156:8080: connect: connection refused" Mar 13 15:47:51 crc kubenswrapper[4907]: I0313 15:47:51.754354 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-engine-86cbbf5dd8-25c77" Mar 13 15:47:52 crc kubenswrapper[4907]: E0313 15:47:52.125160 4907 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod099b49c3_666e_4bf5_9c15_32cf9c40e1e7.slice/crio-e4caf9d5f22ffb193c93c87c06b4e64f7a29f8bf0582a1a8b7f47762f5ba55eb\": RecentStats: unable to find data in memory cache]" Mar 13 15:47:57 crc kubenswrapper[4907]: I0313 15:47:57.487934 4907 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-655cbf768f-dm9zz" podUID="afd4001d-40a5-4a09-ab05-36cbd4d06eb4" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.156:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.156:8080: connect: connection refused" Mar 13 15:48:00 crc kubenswrapper[4907]: I0313 15:48:00.206019 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556948-mmprl"] Mar 13 15:48:00 crc kubenswrapper[4907]: E0313 15:48:00.206959 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ecb3195-45a5-4ca4-b979-9a8c37ec4fbe" containerName="extract-content" Mar 13 15:48:00 crc kubenswrapper[4907]: I0313 15:48:00.206980 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ecb3195-45a5-4ca4-b979-9a8c37ec4fbe" containerName="extract-content" Mar 13 15:48:00 crc kubenswrapper[4907]: E0313 15:48:00.207007 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ecb3195-45a5-4ca4-b979-9a8c37ec4fbe" containerName="registry-server" Mar 13 15:48:00 crc kubenswrapper[4907]: I0313 15:48:00.207018 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ecb3195-45a5-4ca4-b979-9a8c37ec4fbe" containerName="registry-server" Mar 13 15:48:00 crc kubenswrapper[4907]: E0313 15:48:00.207050 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ecb3195-45a5-4ca4-b979-9a8c37ec4fbe" containerName="extract-utilities" Mar 13 15:48:00 crc kubenswrapper[4907]: I0313 15:48:00.207063 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ecb3195-45a5-4ca4-b979-9a8c37ec4fbe" containerName="extract-utilities" Mar 13 15:48:00 crc kubenswrapper[4907]: E0313 15:48:00.207094 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d9fbab8-539c-48f6-8cae-3eda23300b28" containerName="horizon" Mar 13 15:48:00 crc kubenswrapper[4907]: I0313 15:48:00.207107 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d9fbab8-539c-48f6-8cae-3eda23300b28" containerName="horizon" Mar 13 15:48:00 crc kubenswrapper[4907]: E0313 15:48:00.207125 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d9fbab8-539c-48f6-8cae-3eda23300b28" containerName="horizon-log" Mar 13 15:48:00 crc kubenswrapper[4907]: I0313 15:48:00.207135 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d9fbab8-539c-48f6-8cae-3eda23300b28" containerName="horizon-log" Mar 13 15:48:00 crc kubenswrapper[4907]: I0313 15:48:00.207444 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ecb3195-45a5-4ca4-b979-9a8c37ec4fbe" containerName="registry-server" Mar 13 15:48:00 crc kubenswrapper[4907]: I0313 15:48:00.207473 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d9fbab8-539c-48f6-8cae-3eda23300b28" containerName="horizon-log" Mar 13 15:48:00 crc kubenswrapper[4907]: I0313 15:48:00.207488 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d9fbab8-539c-48f6-8cae-3eda23300b28" containerName="horizon" Mar 13 15:48:00 crc kubenswrapper[4907]: I0313 15:48:00.208572 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556948-mmprl" Mar 13 15:48:00 crc kubenswrapper[4907]: I0313 15:48:00.211130 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 15:48:00 crc kubenswrapper[4907]: I0313 15:48:00.211376 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 15:48:00 crc kubenswrapper[4907]: I0313 15:48:00.211936 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 15:48:00 crc kubenswrapper[4907]: I0313 15:48:00.222265 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556948-mmprl"] Mar 13 15:48:00 crc kubenswrapper[4907]: I0313 15:48:00.270378 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-njncg\" (UniqueName: \"kubernetes.io/projected/673cadb2-82bd-4e26-8222-5a89c310f597-kube-api-access-njncg\") pod \"auto-csr-approver-29556948-mmprl\" (UID: \"673cadb2-82bd-4e26-8222-5a89c310f597\") " pod="openshift-infra/auto-csr-approver-29556948-mmprl" Mar 13 15:48:00 crc kubenswrapper[4907]: I0313 15:48:00.373049 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-njncg\" (UniqueName: \"kubernetes.io/projected/673cadb2-82bd-4e26-8222-5a89c310f597-kube-api-access-njncg\") pod \"auto-csr-approver-29556948-mmprl\" (UID: \"673cadb2-82bd-4e26-8222-5a89c310f597\") " pod="openshift-infra/auto-csr-approver-29556948-mmprl" Mar 13 15:48:00 crc kubenswrapper[4907]: I0313 15:48:00.395443 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-njncg\" (UniqueName: \"kubernetes.io/projected/673cadb2-82bd-4e26-8222-5a89c310f597-kube-api-access-njncg\") pod \"auto-csr-approver-29556948-mmprl\" (UID: \"673cadb2-82bd-4e26-8222-5a89c310f597\") " pod="openshift-infra/auto-csr-approver-29556948-mmprl" Mar 13 15:48:00 crc kubenswrapper[4907]: I0313 15:48:00.539022 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556948-mmprl" Mar 13 15:48:01 crc kubenswrapper[4907]: I0313 15:48:01.080129 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556948-mmprl"] Mar 13 15:48:01 crc kubenswrapper[4907]: I0313 15:48:01.083690 4907 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 13 15:48:01 crc kubenswrapper[4907]: I0313 15:48:01.444363 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556948-mmprl" event={"ID":"673cadb2-82bd-4e26-8222-5a89c310f597","Type":"ContainerStarted","Data":"965a5ffa68a0e452b4155d7f17ef1a1f0763c8d29a45b0bbc4ebb8845a50c5f4"} Mar 13 15:48:03 crc kubenswrapper[4907]: I0313 15:48:03.464558 4907 generic.go:334] "Generic (PLEG): container finished" podID="673cadb2-82bd-4e26-8222-5a89c310f597" containerID="48d98d79386ad3e2c5200a51cdf22425a3d47af926f18ea01f01295cd4a45c13" exitCode=0 Mar 13 15:48:03 crc kubenswrapper[4907]: I0313 15:48:03.464705 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556948-mmprl" event={"ID":"673cadb2-82bd-4e26-8222-5a89c310f597","Type":"ContainerDied","Data":"48d98d79386ad3e2c5200a51cdf22425a3d47af926f18ea01f01295cd4a45c13"} Mar 13 15:48:04 crc kubenswrapper[4907]: I0313 15:48:04.830766 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556948-mmprl" Mar 13 15:48:04 crc kubenswrapper[4907]: I0313 15:48:04.865056 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-njncg\" (UniqueName: \"kubernetes.io/projected/673cadb2-82bd-4e26-8222-5a89c310f597-kube-api-access-njncg\") pod \"673cadb2-82bd-4e26-8222-5a89c310f597\" (UID: \"673cadb2-82bd-4e26-8222-5a89c310f597\") " Mar 13 15:48:04 crc kubenswrapper[4907]: I0313 15:48:04.875581 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/673cadb2-82bd-4e26-8222-5a89c310f597-kube-api-access-njncg" (OuterVolumeSpecName: "kube-api-access-njncg") pod "673cadb2-82bd-4e26-8222-5a89c310f597" (UID: "673cadb2-82bd-4e26-8222-5a89c310f597"). InnerVolumeSpecName "kube-api-access-njncg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:48:04 crc kubenswrapper[4907]: I0313 15:48:04.968916 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-njncg\" (UniqueName: \"kubernetes.io/projected/673cadb2-82bd-4e26-8222-5a89c310f597-kube-api-access-njncg\") on node \"crc\" DevicePath \"\"" Mar 13 15:48:05 crc kubenswrapper[4907]: I0313 15:48:05.487545 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556948-mmprl" event={"ID":"673cadb2-82bd-4e26-8222-5a89c310f597","Type":"ContainerDied","Data":"965a5ffa68a0e452b4155d7f17ef1a1f0763c8d29a45b0bbc4ebb8845a50c5f4"} Mar 13 15:48:05 crc kubenswrapper[4907]: I0313 15:48:05.487912 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="965a5ffa68a0e452b4155d7f17ef1a1f0763c8d29a45b0bbc4ebb8845a50c5f4" Mar 13 15:48:05 crc kubenswrapper[4907]: I0313 15:48:05.487594 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556948-mmprl" Mar 13 15:48:05 crc kubenswrapper[4907]: I0313 15:48:05.903688 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556942-mdddr"] Mar 13 15:48:05 crc kubenswrapper[4907]: I0313 15:48:05.913740 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556942-mdddr"] Mar 13 15:48:07 crc kubenswrapper[4907]: I0313 15:48:07.487330 4907 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-655cbf768f-dm9zz" podUID="afd4001d-40a5-4a09-ab05-36cbd4d06eb4" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.156:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.156:8080: connect: connection refused" Mar 13 15:48:07 crc kubenswrapper[4907]: I0313 15:48:07.487910 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-655cbf768f-dm9zz" Mar 13 15:48:07 crc kubenswrapper[4907]: I0313 15:48:07.554305 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08m9qw2"] Mar 13 15:48:07 crc kubenswrapper[4907]: E0313 15:48:07.554747 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="673cadb2-82bd-4e26-8222-5a89c310f597" containerName="oc" Mar 13 15:48:07 crc kubenswrapper[4907]: I0313 15:48:07.554769 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="673cadb2-82bd-4e26-8222-5a89c310f597" containerName="oc" Mar 13 15:48:07 crc kubenswrapper[4907]: I0313 15:48:07.555619 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="673cadb2-82bd-4e26-8222-5a89c310f597" containerName="oc" Mar 13 15:48:07 crc kubenswrapper[4907]: I0313 15:48:07.557344 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08m9qw2" Mar 13 15:48:07 crc kubenswrapper[4907]: I0313 15:48:07.559432 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Mar 13 15:48:07 crc kubenswrapper[4907]: I0313 15:48:07.565200 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08m9qw2"] Mar 13 15:48:07 crc kubenswrapper[4907]: I0313 15:48:07.730379 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d91395e4-18e5-4da8-b482-093b71c472a3-bundle\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08m9qw2\" (UID: \"d91395e4-18e5-4da8-b482-093b71c472a3\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08m9qw2" Mar 13 15:48:07 crc kubenswrapper[4907]: I0313 15:48:07.730451 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sd9hv\" (UniqueName: \"kubernetes.io/projected/d91395e4-18e5-4da8-b482-093b71c472a3-kube-api-access-sd9hv\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08m9qw2\" (UID: \"d91395e4-18e5-4da8-b482-093b71c472a3\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08m9qw2" Mar 13 15:48:07 crc kubenswrapper[4907]: I0313 15:48:07.730510 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d91395e4-18e5-4da8-b482-093b71c472a3-util\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08m9qw2\" (UID: \"d91395e4-18e5-4da8-b482-093b71c472a3\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08m9qw2" Mar 13 15:48:07 crc kubenswrapper[4907]: I0313 15:48:07.798604 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ed0ed6c7-e5d4-4770-a9aa-308c0853573e" path="/var/lib/kubelet/pods/ed0ed6c7-e5d4-4770-a9aa-308c0853573e/volumes" Mar 13 15:48:07 crc kubenswrapper[4907]: I0313 15:48:07.832006 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d91395e4-18e5-4da8-b482-093b71c472a3-bundle\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08m9qw2\" (UID: \"d91395e4-18e5-4da8-b482-093b71c472a3\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08m9qw2" Mar 13 15:48:07 crc kubenswrapper[4907]: I0313 15:48:07.832068 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sd9hv\" (UniqueName: \"kubernetes.io/projected/d91395e4-18e5-4da8-b482-093b71c472a3-kube-api-access-sd9hv\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08m9qw2\" (UID: \"d91395e4-18e5-4da8-b482-093b71c472a3\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08m9qw2" Mar 13 15:48:07 crc kubenswrapper[4907]: I0313 15:48:07.832107 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d91395e4-18e5-4da8-b482-093b71c472a3-util\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08m9qw2\" (UID: \"d91395e4-18e5-4da8-b482-093b71c472a3\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08m9qw2" Mar 13 15:48:07 crc kubenswrapper[4907]: I0313 15:48:07.832640 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d91395e4-18e5-4da8-b482-093b71c472a3-bundle\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08m9qw2\" (UID: \"d91395e4-18e5-4da8-b482-093b71c472a3\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08m9qw2" Mar 13 15:48:07 crc kubenswrapper[4907]: I0313 15:48:07.832684 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d91395e4-18e5-4da8-b482-093b71c472a3-util\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08m9qw2\" (UID: \"d91395e4-18e5-4da8-b482-093b71c472a3\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08m9qw2" Mar 13 15:48:07 crc kubenswrapper[4907]: I0313 15:48:07.856065 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sd9hv\" (UniqueName: \"kubernetes.io/projected/d91395e4-18e5-4da8-b482-093b71c472a3-kube-api-access-sd9hv\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08m9qw2\" (UID: \"d91395e4-18e5-4da8-b482-093b71c472a3\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08m9qw2" Mar 13 15:48:07 crc kubenswrapper[4907]: I0313 15:48:07.883281 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08m9qw2" Mar 13 15:48:08 crc kubenswrapper[4907]: I0313 15:48:08.034211 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-kxfg2"] Mar 13 15:48:08 crc kubenswrapper[4907]: I0313 15:48:08.046215 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-fd12-account-create-update-kxbcs"] Mar 13 15:48:08 crc kubenswrapper[4907]: I0313 15:48:08.062544 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-fd12-account-create-update-kxbcs"] Mar 13 15:48:08 crc kubenswrapper[4907]: I0313 15:48:08.074412 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-kxfg2"] Mar 13 15:48:08 crc kubenswrapper[4907]: I0313 15:48:08.338609 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08m9qw2"] Mar 13 15:48:08 crc kubenswrapper[4907]: W0313 15:48:08.350871 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd91395e4_18e5_4da8_b482_093b71c472a3.slice/crio-6e048b993c2bad9cfa6a447e0cee7c6ba7ef45bc00d56e28386ad86b9c523334 WatchSource:0}: Error finding container 6e048b993c2bad9cfa6a447e0cee7c6ba7ef45bc00d56e28386ad86b9c523334: Status 404 returned error can't find the container with id 6e048b993c2bad9cfa6a447e0cee7c6ba7ef45bc00d56e28386ad86b9c523334 Mar 13 15:48:08 crc kubenswrapper[4907]: I0313 15:48:08.527178 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08m9qw2" event={"ID":"d91395e4-18e5-4da8-b482-093b71c472a3","Type":"ContainerStarted","Data":"6e048b993c2bad9cfa6a447e0cee7c6ba7ef45bc00d56e28386ad86b9c523334"} Mar 13 15:48:09 crc kubenswrapper[4907]: I0313 15:48:09.169606 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-7tp2h"] Mar 13 15:48:09 crc kubenswrapper[4907]: I0313 15:48:09.172455 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7tp2h" Mar 13 15:48:09 crc kubenswrapper[4907]: I0313 15:48:09.184216 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7tp2h"] Mar 13 15:48:09 crc kubenswrapper[4907]: I0313 15:48:09.365573 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ppphq\" (UniqueName: \"kubernetes.io/projected/e8c5a748-e7b7-43e0-82d3-c76673b17ea1-kube-api-access-ppphq\") pod \"redhat-operators-7tp2h\" (UID: \"e8c5a748-e7b7-43e0-82d3-c76673b17ea1\") " pod="openshift-marketplace/redhat-operators-7tp2h" Mar 13 15:48:09 crc kubenswrapper[4907]: I0313 15:48:09.365679 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e8c5a748-e7b7-43e0-82d3-c76673b17ea1-utilities\") pod \"redhat-operators-7tp2h\" (UID: \"e8c5a748-e7b7-43e0-82d3-c76673b17ea1\") " pod="openshift-marketplace/redhat-operators-7tp2h" Mar 13 15:48:09 crc kubenswrapper[4907]: I0313 15:48:09.365861 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e8c5a748-e7b7-43e0-82d3-c76673b17ea1-catalog-content\") pod \"redhat-operators-7tp2h\" (UID: \"e8c5a748-e7b7-43e0-82d3-c76673b17ea1\") " pod="openshift-marketplace/redhat-operators-7tp2h" Mar 13 15:48:09 crc kubenswrapper[4907]: I0313 15:48:09.467366 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ppphq\" (UniqueName: \"kubernetes.io/projected/e8c5a748-e7b7-43e0-82d3-c76673b17ea1-kube-api-access-ppphq\") pod \"redhat-operators-7tp2h\" (UID: \"e8c5a748-e7b7-43e0-82d3-c76673b17ea1\") " pod="openshift-marketplace/redhat-operators-7tp2h" Mar 13 15:48:09 crc kubenswrapper[4907]: I0313 15:48:09.467436 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e8c5a748-e7b7-43e0-82d3-c76673b17ea1-utilities\") pod \"redhat-operators-7tp2h\" (UID: \"e8c5a748-e7b7-43e0-82d3-c76673b17ea1\") " pod="openshift-marketplace/redhat-operators-7tp2h" Mar 13 15:48:09 crc kubenswrapper[4907]: I0313 15:48:09.467590 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e8c5a748-e7b7-43e0-82d3-c76673b17ea1-catalog-content\") pod \"redhat-operators-7tp2h\" (UID: \"e8c5a748-e7b7-43e0-82d3-c76673b17ea1\") " pod="openshift-marketplace/redhat-operators-7tp2h" Mar 13 15:48:09 crc kubenswrapper[4907]: I0313 15:48:09.468111 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e8c5a748-e7b7-43e0-82d3-c76673b17ea1-catalog-content\") pod \"redhat-operators-7tp2h\" (UID: \"e8c5a748-e7b7-43e0-82d3-c76673b17ea1\") " pod="openshift-marketplace/redhat-operators-7tp2h" Mar 13 15:48:09 crc kubenswrapper[4907]: I0313 15:48:09.468192 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e8c5a748-e7b7-43e0-82d3-c76673b17ea1-utilities\") pod \"redhat-operators-7tp2h\" (UID: \"e8c5a748-e7b7-43e0-82d3-c76673b17ea1\") " pod="openshift-marketplace/redhat-operators-7tp2h" Mar 13 15:48:09 crc kubenswrapper[4907]: I0313 15:48:09.488849 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ppphq\" (UniqueName: \"kubernetes.io/projected/e8c5a748-e7b7-43e0-82d3-c76673b17ea1-kube-api-access-ppphq\") pod \"redhat-operators-7tp2h\" (UID: \"e8c5a748-e7b7-43e0-82d3-c76673b17ea1\") " pod="openshift-marketplace/redhat-operators-7tp2h" Mar 13 15:48:09 crc kubenswrapper[4907]: I0313 15:48:09.500648 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7tp2h" Mar 13 15:48:09 crc kubenswrapper[4907]: I0313 15:48:09.544438 4907 generic.go:334] "Generic (PLEG): container finished" podID="d91395e4-18e5-4da8-b482-093b71c472a3" containerID="2cbc12de3757ad8a4d2b65395ca25daf3aeb5685c9886ea14b7ac0bfb3d45f3b" exitCode=0 Mar 13 15:48:09 crc kubenswrapper[4907]: I0313 15:48:09.544576 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08m9qw2" event={"ID":"d91395e4-18e5-4da8-b482-093b71c472a3","Type":"ContainerDied","Data":"2cbc12de3757ad8a4d2b65395ca25daf3aeb5685c9886ea14b7ac0bfb3d45f3b"} Mar 13 15:48:09 crc kubenswrapper[4907]: I0313 15:48:09.816969 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0d6a6021-7cc4-42f9-848b-5bbfc3650ee7" path="/var/lib/kubelet/pods/0d6a6021-7cc4-42f9-848b-5bbfc3650ee7/volumes" Mar 13 15:48:09 crc kubenswrapper[4907]: I0313 15:48:09.818179 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="440d4795-da82-461c-b460-bdd5c271de4d" path="/var/lib/kubelet/pods/440d4795-da82-461c-b460-bdd5c271de4d/volumes" Mar 13 15:48:10 crc kubenswrapper[4907]: I0313 15:48:10.003238 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7tp2h"] Mar 13 15:48:10 crc kubenswrapper[4907]: W0313 15:48:10.013276 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode8c5a748_e7b7_43e0_82d3_c76673b17ea1.slice/crio-a029d38e007ef1dd2747ad9ff6b5a9cef6fcfa563fa0dc97f6df3132418979f8 WatchSource:0}: Error finding container a029d38e007ef1dd2747ad9ff6b5a9cef6fcfa563fa0dc97f6df3132418979f8: Status 404 returned error can't find the container with id a029d38e007ef1dd2747ad9ff6b5a9cef6fcfa563fa0dc97f6df3132418979f8 Mar 13 15:48:10 crc kubenswrapper[4907]: I0313 15:48:10.564369 4907 generic.go:334] "Generic (PLEG): container finished" podID="e8c5a748-e7b7-43e0-82d3-c76673b17ea1" containerID="d49b21ae936fc12c5a09e43efc846d4e58d2f3ad32465957c3a2c3b2abeb9cfe" exitCode=0 Mar 13 15:48:10 crc kubenswrapper[4907]: I0313 15:48:10.564430 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7tp2h" event={"ID":"e8c5a748-e7b7-43e0-82d3-c76673b17ea1","Type":"ContainerDied","Data":"d49b21ae936fc12c5a09e43efc846d4e58d2f3ad32465957c3a2c3b2abeb9cfe"} Mar 13 15:48:10 crc kubenswrapper[4907]: I0313 15:48:10.564652 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7tp2h" event={"ID":"e8c5a748-e7b7-43e0-82d3-c76673b17ea1","Type":"ContainerStarted","Data":"a029d38e007ef1dd2747ad9ff6b5a9cef6fcfa563fa0dc97f6df3132418979f8"} Mar 13 15:48:11 crc kubenswrapper[4907]: I0313 15:48:11.578979 4907 generic.go:334] "Generic (PLEG): container finished" podID="d91395e4-18e5-4da8-b482-093b71c472a3" containerID="39803b5860304bb3627a169f6cc4cff88371b582077153cce63b25eb9b7f5995" exitCode=0 Mar 13 15:48:11 crc kubenswrapper[4907]: I0313 15:48:11.579077 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08m9qw2" event={"ID":"d91395e4-18e5-4da8-b482-093b71c472a3","Type":"ContainerDied","Data":"39803b5860304bb3627a169f6cc4cff88371b582077153cce63b25eb9b7f5995"} Mar 13 15:48:12 crc kubenswrapper[4907]: I0313 15:48:12.592805 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08m9qw2" event={"ID":"d91395e4-18e5-4da8-b482-093b71c472a3","Type":"ContainerStarted","Data":"97a66d09d1f3ef76b4030e5d8d68cadcb0f3a729374c9f8133ade6ef4e6bbf41"} Mar 13 15:48:12 crc kubenswrapper[4907]: I0313 15:48:12.595082 4907 generic.go:334] "Generic (PLEG): container finished" podID="afd4001d-40a5-4a09-ab05-36cbd4d06eb4" containerID="4d9d087b4ad0675d46c48144b039251dc336275fe3320abbe51816f548542e9f" exitCode=137 Mar 13 15:48:12 crc kubenswrapper[4907]: I0313 15:48:12.595149 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-655cbf768f-dm9zz" event={"ID":"afd4001d-40a5-4a09-ab05-36cbd4d06eb4","Type":"ContainerDied","Data":"4d9d087b4ad0675d46c48144b039251dc336275fe3320abbe51816f548542e9f"} Mar 13 15:48:12 crc kubenswrapper[4907]: I0313 15:48:12.597664 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7tp2h" event={"ID":"e8c5a748-e7b7-43e0-82d3-c76673b17ea1","Type":"ContainerStarted","Data":"2d7e94bcae9c69807004704893b2f10f4c520e505aa6a59bbe61e348ddae96df"} Mar 13 15:48:12 crc kubenswrapper[4907]: I0313 15:48:12.636182 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08m9qw2" podStartSLOduration=4.37281474 podStartE2EDuration="5.636158507s" podCreationTimestamp="2026-03-13 15:48:07 +0000 UTC" firstStartedPulling="2026-03-13 15:48:09.546382159 +0000 UTC m=+6188.446169848" lastFinishedPulling="2026-03-13 15:48:10.809725926 +0000 UTC m=+6189.709513615" observedRunningTime="2026-03-13 15:48:12.611654958 +0000 UTC m=+6191.511442677" watchObservedRunningTime="2026-03-13 15:48:12.636158507 +0000 UTC m=+6191.535946196" Mar 13 15:48:12 crc kubenswrapper[4907]: E0313 15:48:12.651503 4907 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podafd4001d_40a5_4a09_ab05_36cbd4d06eb4.slice/crio-conmon-4d9d087b4ad0675d46c48144b039251dc336275fe3320abbe51816f548542e9f.scope\": RecentStats: unable to find data in memory cache]" Mar 13 15:48:13 crc kubenswrapper[4907]: I0313 15:48:13.321406 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-655cbf768f-dm9zz" Mar 13 15:48:13 crc kubenswrapper[4907]: I0313 15:48:13.351629 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/afd4001d-40a5-4a09-ab05-36cbd4d06eb4-scripts\") pod \"afd4001d-40a5-4a09-ab05-36cbd4d06eb4\" (UID: \"afd4001d-40a5-4a09-ab05-36cbd4d06eb4\") " Mar 13 15:48:13 crc kubenswrapper[4907]: I0313 15:48:13.394625 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/afd4001d-40a5-4a09-ab05-36cbd4d06eb4-scripts" (OuterVolumeSpecName: "scripts") pod "afd4001d-40a5-4a09-ab05-36cbd4d06eb4" (UID: "afd4001d-40a5-4a09-ab05-36cbd4d06eb4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:48:13 crc kubenswrapper[4907]: I0313 15:48:13.454030 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/afd4001d-40a5-4a09-ab05-36cbd4d06eb4-horizon-secret-key\") pod \"afd4001d-40a5-4a09-ab05-36cbd4d06eb4\" (UID: \"afd4001d-40a5-4a09-ab05-36cbd4d06eb4\") " Mar 13 15:48:13 crc kubenswrapper[4907]: I0313 15:48:13.454207 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/afd4001d-40a5-4a09-ab05-36cbd4d06eb4-config-data\") pod \"afd4001d-40a5-4a09-ab05-36cbd4d06eb4\" (UID: \"afd4001d-40a5-4a09-ab05-36cbd4d06eb4\") " Mar 13 15:48:13 crc kubenswrapper[4907]: I0313 15:48:13.454301 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/afd4001d-40a5-4a09-ab05-36cbd4d06eb4-logs\") pod \"afd4001d-40a5-4a09-ab05-36cbd4d06eb4\" (UID: \"afd4001d-40a5-4a09-ab05-36cbd4d06eb4\") " Mar 13 15:48:13 crc kubenswrapper[4907]: I0313 15:48:13.454366 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ml6dh\" (UniqueName: \"kubernetes.io/projected/afd4001d-40a5-4a09-ab05-36cbd4d06eb4-kube-api-access-ml6dh\") pod \"afd4001d-40a5-4a09-ab05-36cbd4d06eb4\" (UID: \"afd4001d-40a5-4a09-ab05-36cbd4d06eb4\") " Mar 13 15:48:13 crc kubenswrapper[4907]: I0313 15:48:13.454792 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/afd4001d-40a5-4a09-ab05-36cbd4d06eb4-logs" (OuterVolumeSpecName: "logs") pod "afd4001d-40a5-4a09-ab05-36cbd4d06eb4" (UID: "afd4001d-40a5-4a09-ab05-36cbd4d06eb4"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:48:13 crc kubenswrapper[4907]: I0313 15:48:13.455069 4907 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/afd4001d-40a5-4a09-ab05-36cbd4d06eb4-logs\") on node \"crc\" DevicePath \"\"" Mar 13 15:48:13 crc kubenswrapper[4907]: I0313 15:48:13.455095 4907 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/afd4001d-40a5-4a09-ab05-36cbd4d06eb4-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 15:48:13 crc kubenswrapper[4907]: I0313 15:48:13.458036 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/afd4001d-40a5-4a09-ab05-36cbd4d06eb4-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "afd4001d-40a5-4a09-ab05-36cbd4d06eb4" (UID: "afd4001d-40a5-4a09-ab05-36cbd4d06eb4"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:48:13 crc kubenswrapper[4907]: I0313 15:48:13.458108 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/afd4001d-40a5-4a09-ab05-36cbd4d06eb4-kube-api-access-ml6dh" (OuterVolumeSpecName: "kube-api-access-ml6dh") pod "afd4001d-40a5-4a09-ab05-36cbd4d06eb4" (UID: "afd4001d-40a5-4a09-ab05-36cbd4d06eb4"). InnerVolumeSpecName "kube-api-access-ml6dh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:48:13 crc kubenswrapper[4907]: I0313 15:48:13.481348 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/afd4001d-40a5-4a09-ab05-36cbd4d06eb4-config-data" (OuterVolumeSpecName: "config-data") pod "afd4001d-40a5-4a09-ab05-36cbd4d06eb4" (UID: "afd4001d-40a5-4a09-ab05-36cbd4d06eb4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:48:13 crc kubenswrapper[4907]: I0313 15:48:13.556677 4907 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/afd4001d-40a5-4a09-ab05-36cbd4d06eb4-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Mar 13 15:48:13 crc kubenswrapper[4907]: I0313 15:48:13.556729 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/afd4001d-40a5-4a09-ab05-36cbd4d06eb4-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 15:48:13 crc kubenswrapper[4907]: I0313 15:48:13.556750 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ml6dh\" (UniqueName: \"kubernetes.io/projected/afd4001d-40a5-4a09-ab05-36cbd4d06eb4-kube-api-access-ml6dh\") on node \"crc\" DevicePath \"\"" Mar 13 15:48:13 crc kubenswrapper[4907]: I0313 15:48:13.628775 4907 generic.go:334] "Generic (PLEG): container finished" podID="d91395e4-18e5-4da8-b482-093b71c472a3" containerID="97a66d09d1f3ef76b4030e5d8d68cadcb0f3a729374c9f8133ade6ef4e6bbf41" exitCode=0 Mar 13 15:48:13 crc kubenswrapper[4907]: I0313 15:48:13.628850 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08m9qw2" event={"ID":"d91395e4-18e5-4da8-b482-093b71c472a3","Type":"ContainerDied","Data":"97a66d09d1f3ef76b4030e5d8d68cadcb0f3a729374c9f8133ade6ef4e6bbf41"} Mar 13 15:48:13 crc kubenswrapper[4907]: I0313 15:48:13.630544 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-655cbf768f-dm9zz" event={"ID":"afd4001d-40a5-4a09-ab05-36cbd4d06eb4","Type":"ContainerDied","Data":"8eed0ad989f3d5a163602bd4e967ac325bbb242d625bd6a2d8dac4fb6b57f365"} Mar 13 15:48:13 crc kubenswrapper[4907]: I0313 15:48:13.630588 4907 scope.go:117] "RemoveContainer" containerID="b42f8e33dfb6318b768e8c6b32af48621974ee97af5dab3f3e4bc582bbfc2931" Mar 13 15:48:13 crc kubenswrapper[4907]: I0313 15:48:13.630634 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-655cbf768f-dm9zz" Mar 13 15:48:13 crc kubenswrapper[4907]: I0313 15:48:13.689398 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-655cbf768f-dm9zz"] Mar 13 15:48:13 crc kubenswrapper[4907]: I0313 15:48:13.700511 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-655cbf768f-dm9zz"] Mar 13 15:48:13 crc kubenswrapper[4907]: I0313 15:48:13.792569 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="afd4001d-40a5-4a09-ab05-36cbd4d06eb4" path="/var/lib/kubelet/pods/afd4001d-40a5-4a09-ab05-36cbd4d06eb4/volumes" Mar 13 15:48:13 crc kubenswrapper[4907]: I0313 15:48:13.814920 4907 scope.go:117] "RemoveContainer" containerID="4d9d087b4ad0675d46c48144b039251dc336275fe3320abbe51816f548542e9f" Mar 13 15:48:14 crc kubenswrapper[4907]: I0313 15:48:14.641361 4907 generic.go:334] "Generic (PLEG): container finished" podID="e8c5a748-e7b7-43e0-82d3-c76673b17ea1" containerID="2d7e94bcae9c69807004704893b2f10f4c520e505aa6a59bbe61e348ddae96df" exitCode=0 Mar 13 15:48:14 crc kubenswrapper[4907]: I0313 15:48:14.641434 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7tp2h" event={"ID":"e8c5a748-e7b7-43e0-82d3-c76673b17ea1","Type":"ContainerDied","Data":"2d7e94bcae9c69807004704893b2f10f4c520e505aa6a59bbe61e348ddae96df"} Mar 13 15:48:15 crc kubenswrapper[4907]: I0313 15:48:15.057134 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08m9qw2" Mar 13 15:48:15 crc kubenswrapper[4907]: I0313 15:48:15.102681 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d91395e4-18e5-4da8-b482-093b71c472a3-bundle\") pod \"d91395e4-18e5-4da8-b482-093b71c472a3\" (UID: \"d91395e4-18e5-4da8-b482-093b71c472a3\") " Mar 13 15:48:15 crc kubenswrapper[4907]: I0313 15:48:15.102837 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d91395e4-18e5-4da8-b482-093b71c472a3-util\") pod \"d91395e4-18e5-4da8-b482-093b71c472a3\" (UID: \"d91395e4-18e5-4da8-b482-093b71c472a3\") " Mar 13 15:48:15 crc kubenswrapper[4907]: I0313 15:48:15.102891 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sd9hv\" (UniqueName: \"kubernetes.io/projected/d91395e4-18e5-4da8-b482-093b71c472a3-kube-api-access-sd9hv\") pod \"d91395e4-18e5-4da8-b482-093b71c472a3\" (UID: \"d91395e4-18e5-4da8-b482-093b71c472a3\") " Mar 13 15:48:15 crc kubenswrapper[4907]: I0313 15:48:15.104797 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d91395e4-18e5-4da8-b482-093b71c472a3-bundle" (OuterVolumeSpecName: "bundle") pod "d91395e4-18e5-4da8-b482-093b71c472a3" (UID: "d91395e4-18e5-4da8-b482-093b71c472a3"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:48:15 crc kubenswrapper[4907]: I0313 15:48:15.108851 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d91395e4-18e5-4da8-b482-093b71c472a3-kube-api-access-sd9hv" (OuterVolumeSpecName: "kube-api-access-sd9hv") pod "d91395e4-18e5-4da8-b482-093b71c472a3" (UID: "d91395e4-18e5-4da8-b482-093b71c472a3"). InnerVolumeSpecName "kube-api-access-sd9hv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:48:15 crc kubenswrapper[4907]: I0313 15:48:15.111571 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d91395e4-18e5-4da8-b482-093b71c472a3-util" (OuterVolumeSpecName: "util") pod "d91395e4-18e5-4da8-b482-093b71c472a3" (UID: "d91395e4-18e5-4da8-b482-093b71c472a3"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:48:15 crc kubenswrapper[4907]: I0313 15:48:15.207407 4907 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d91395e4-18e5-4da8-b482-093b71c472a3-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 15:48:15 crc kubenswrapper[4907]: I0313 15:48:15.207730 4907 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d91395e4-18e5-4da8-b482-093b71c472a3-util\") on node \"crc\" DevicePath \"\"" Mar 13 15:48:15 crc kubenswrapper[4907]: I0313 15:48:15.207746 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sd9hv\" (UniqueName: \"kubernetes.io/projected/d91395e4-18e5-4da8-b482-093b71c472a3-kube-api-access-sd9hv\") on node \"crc\" DevicePath \"\"" Mar 13 15:48:15 crc kubenswrapper[4907]: I0313 15:48:15.661533 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08m9qw2" event={"ID":"d91395e4-18e5-4da8-b482-093b71c472a3","Type":"ContainerDied","Data":"6e048b993c2bad9cfa6a447e0cee7c6ba7ef45bc00d56e28386ad86b9c523334"} Mar 13 15:48:15 crc kubenswrapper[4907]: I0313 15:48:15.661772 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6e048b993c2bad9cfa6a447e0cee7c6ba7ef45bc00d56e28386ad86b9c523334" Mar 13 15:48:15 crc kubenswrapper[4907]: I0313 15:48:15.661837 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08m9qw2" Mar 13 15:48:15 crc kubenswrapper[4907]: I0313 15:48:15.666832 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7tp2h" event={"ID":"e8c5a748-e7b7-43e0-82d3-c76673b17ea1","Type":"ContainerStarted","Data":"49ef110711e5f738221014e1ce44cead4c2950560589e797e31382bd54cac333"} Mar 13 15:48:15 crc kubenswrapper[4907]: I0313 15:48:15.687623 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-7tp2h" podStartSLOduration=2.392888227 podStartE2EDuration="6.687607617s" podCreationTimestamp="2026-03-13 15:48:09 +0000 UTC" firstStartedPulling="2026-03-13 15:48:10.747611058 +0000 UTC m=+6189.647398757" lastFinishedPulling="2026-03-13 15:48:15.042330458 +0000 UTC m=+6193.942118147" observedRunningTime="2026-03-13 15:48:15.682635822 +0000 UTC m=+6194.582423531" watchObservedRunningTime="2026-03-13 15:48:15.687607617 +0000 UTC m=+6194.587395306" Mar 13 15:48:17 crc kubenswrapper[4907]: I0313 15:48:17.049872 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-tlb57"] Mar 13 15:48:17 crc kubenswrapper[4907]: I0313 15:48:17.059023 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-tlb57"] Mar 13 15:48:17 crc kubenswrapper[4907]: I0313 15:48:17.795307 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="13c6b679-96d2-407c-8b94-cd908d61945b" path="/var/lib/kubelet/pods/13c6b679-96d2-407c-8b94-cd908d61945b/volumes" Mar 13 15:48:19 crc kubenswrapper[4907]: I0313 15:48:19.501257 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-7tp2h" Mar 13 15:48:19 crc kubenswrapper[4907]: I0313 15:48:19.501551 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-7tp2h" Mar 13 15:48:20 crc kubenswrapper[4907]: I0313 15:48:20.678368 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-7tp2h" podUID="e8c5a748-e7b7-43e0-82d3-c76673b17ea1" containerName="registry-server" probeResult="failure" output=< Mar 13 15:48:20 crc kubenswrapper[4907]: timeout: failed to connect service ":50051" within 1s Mar 13 15:48:20 crc kubenswrapper[4907]: > Mar 13 15:48:24 crc kubenswrapper[4907]: I0313 15:48:24.303677 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-68bc856cb9-w2ngb"] Mar 13 15:48:24 crc kubenswrapper[4907]: E0313 15:48:24.304633 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d91395e4-18e5-4da8-b482-093b71c472a3" containerName="pull" Mar 13 15:48:24 crc kubenswrapper[4907]: I0313 15:48:24.304649 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="d91395e4-18e5-4da8-b482-093b71c472a3" containerName="pull" Mar 13 15:48:24 crc kubenswrapper[4907]: E0313 15:48:24.304665 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="afd4001d-40a5-4a09-ab05-36cbd4d06eb4" containerName="horizon-log" Mar 13 15:48:24 crc kubenswrapper[4907]: I0313 15:48:24.304671 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="afd4001d-40a5-4a09-ab05-36cbd4d06eb4" containerName="horizon-log" Mar 13 15:48:24 crc kubenswrapper[4907]: E0313 15:48:24.310452 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d91395e4-18e5-4da8-b482-093b71c472a3" containerName="util" Mar 13 15:48:24 crc kubenswrapper[4907]: I0313 15:48:24.310522 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="d91395e4-18e5-4da8-b482-093b71c472a3" containerName="util" Mar 13 15:48:24 crc kubenswrapper[4907]: E0313 15:48:24.310564 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d91395e4-18e5-4da8-b482-093b71c472a3" containerName="extract" Mar 13 15:48:24 crc kubenswrapper[4907]: I0313 15:48:24.310574 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="d91395e4-18e5-4da8-b482-093b71c472a3" containerName="extract" Mar 13 15:48:24 crc kubenswrapper[4907]: E0313 15:48:24.310607 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="afd4001d-40a5-4a09-ab05-36cbd4d06eb4" containerName="horizon" Mar 13 15:48:24 crc kubenswrapper[4907]: I0313 15:48:24.310615 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="afd4001d-40a5-4a09-ab05-36cbd4d06eb4" containerName="horizon" Mar 13 15:48:24 crc kubenswrapper[4907]: I0313 15:48:24.311044 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="afd4001d-40a5-4a09-ab05-36cbd4d06eb4" containerName="horizon" Mar 13 15:48:24 crc kubenswrapper[4907]: I0313 15:48:24.311063 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="d91395e4-18e5-4da8-b482-093b71c472a3" containerName="extract" Mar 13 15:48:24 crc kubenswrapper[4907]: I0313 15:48:24.311083 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="afd4001d-40a5-4a09-ab05-36cbd4d06eb4" containerName="horizon-log" Mar 13 15:48:24 crc kubenswrapper[4907]: I0313 15:48:24.311813 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-w2ngb" Mar 13 15:48:24 crc kubenswrapper[4907]: I0313 15:48:24.316519 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"openshift-service-ca.crt" Mar 13 15:48:24 crc kubenswrapper[4907]: I0313 15:48:24.316763 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-dockercfg-kkxtc" Mar 13 15:48:24 crc kubenswrapper[4907]: I0313 15:48:24.316931 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"kube-root-ca.crt" Mar 13 15:48:24 crc kubenswrapper[4907]: I0313 15:48:24.330185 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-68bc856cb9-w2ngb"] Mar 13 15:48:24 crc kubenswrapper[4907]: I0313 15:48:24.415353 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9n5jn\" (UniqueName: \"kubernetes.io/projected/92caa667-d9ee-4043-95b1-6475bcc7439b-kube-api-access-9n5jn\") pod \"obo-prometheus-operator-68bc856cb9-w2ngb\" (UID: \"92caa667-d9ee-4043-95b1-6475bcc7439b\") " pod="openshift-operators/obo-prometheus-operator-68bc856cb9-w2ngb" Mar 13 15:48:24 crc kubenswrapper[4907]: I0313 15:48:24.434898 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-7dc9f6d87f-zlpcq"] Mar 13 15:48:24 crc kubenswrapper[4907]: I0313 15:48:24.436228 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7dc9f6d87f-zlpcq" Mar 13 15:48:24 crc kubenswrapper[4907]: I0313 15:48:24.440193 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-service-cert" Mar 13 15:48:24 crc kubenswrapper[4907]: I0313 15:48:24.440546 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-dockercfg-jm8k4" Mar 13 15:48:24 crc kubenswrapper[4907]: I0313 15:48:24.458238 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-7dc9f6d87f-zlpcq"] Mar 13 15:48:24 crc kubenswrapper[4907]: I0313 15:48:24.471593 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-7dc9f6d87f-j6dtz"] Mar 13 15:48:24 crc kubenswrapper[4907]: I0313 15:48:24.473336 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7dc9f6d87f-j6dtz" Mar 13 15:48:24 crc kubenswrapper[4907]: I0313 15:48:24.506141 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-7dc9f6d87f-j6dtz"] Mar 13 15:48:24 crc kubenswrapper[4907]: I0313 15:48:24.518077 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9n5jn\" (UniqueName: \"kubernetes.io/projected/92caa667-d9ee-4043-95b1-6475bcc7439b-kube-api-access-9n5jn\") pod \"obo-prometheus-operator-68bc856cb9-w2ngb\" (UID: \"92caa667-d9ee-4043-95b1-6475bcc7439b\") " pod="openshift-operators/obo-prometheus-operator-68bc856cb9-w2ngb" Mar 13 15:48:24 crc kubenswrapper[4907]: I0313 15:48:24.546427 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9n5jn\" (UniqueName: \"kubernetes.io/projected/92caa667-d9ee-4043-95b1-6475bcc7439b-kube-api-access-9n5jn\") pod \"obo-prometheus-operator-68bc856cb9-w2ngb\" (UID: \"92caa667-d9ee-4043-95b1-6475bcc7439b\") " pod="openshift-operators/obo-prometheus-operator-68bc856cb9-w2ngb" Mar 13 15:48:24 crc kubenswrapper[4907]: I0313 15:48:24.620135 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/d188fa9a-103d-49af-9626-90b7a3a5d70c-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-7dc9f6d87f-zlpcq\" (UID: \"d188fa9a-103d-49af-9626-90b7a3a5d70c\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7dc9f6d87f-zlpcq" Mar 13 15:48:24 crc kubenswrapper[4907]: I0313 15:48:24.620271 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/d188fa9a-103d-49af-9626-90b7a3a5d70c-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-7dc9f6d87f-zlpcq\" (UID: \"d188fa9a-103d-49af-9626-90b7a3a5d70c\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7dc9f6d87f-zlpcq" Mar 13 15:48:24 crc kubenswrapper[4907]: I0313 15:48:24.620304 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/683c61a5-c39a-400b-b673-19bbaf286482-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-7dc9f6d87f-j6dtz\" (UID: \"683c61a5-c39a-400b-b673-19bbaf286482\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7dc9f6d87f-j6dtz" Mar 13 15:48:24 crc kubenswrapper[4907]: I0313 15:48:24.620344 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/683c61a5-c39a-400b-b673-19bbaf286482-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-7dc9f6d87f-j6dtz\" (UID: \"683c61a5-c39a-400b-b673-19bbaf286482\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7dc9f6d87f-j6dtz" Mar 13 15:48:24 crc kubenswrapper[4907]: I0313 15:48:24.639348 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-w2ngb" Mar 13 15:48:24 crc kubenswrapper[4907]: I0313 15:48:24.650081 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/observability-operator-59bdc8b94-9hfqr"] Mar 13 15:48:24 crc kubenswrapper[4907]: I0313 15:48:24.670616 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-59bdc8b94-9hfqr" Mar 13 15:48:24 crc kubenswrapper[4907]: I0313 15:48:24.681150 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-sa-dockercfg-xctpc" Mar 13 15:48:24 crc kubenswrapper[4907]: I0313 15:48:24.682787 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-tls" Mar 13 15:48:24 crc kubenswrapper[4907]: I0313 15:48:24.709638 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-59bdc8b94-9hfqr"] Mar 13 15:48:24 crc kubenswrapper[4907]: I0313 15:48:24.726742 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/d188fa9a-103d-49af-9626-90b7a3a5d70c-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-7dc9f6d87f-zlpcq\" (UID: \"d188fa9a-103d-49af-9626-90b7a3a5d70c\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7dc9f6d87f-zlpcq" Mar 13 15:48:24 crc kubenswrapper[4907]: I0313 15:48:24.727123 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/d188fa9a-103d-49af-9626-90b7a3a5d70c-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-7dc9f6d87f-zlpcq\" (UID: \"d188fa9a-103d-49af-9626-90b7a3a5d70c\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7dc9f6d87f-zlpcq" Mar 13 15:48:24 crc kubenswrapper[4907]: I0313 15:48:24.727215 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/683c61a5-c39a-400b-b673-19bbaf286482-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-7dc9f6d87f-j6dtz\" (UID: \"683c61a5-c39a-400b-b673-19bbaf286482\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7dc9f6d87f-j6dtz" Mar 13 15:48:24 crc kubenswrapper[4907]: I0313 15:48:24.727325 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/683c61a5-c39a-400b-b673-19bbaf286482-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-7dc9f6d87f-j6dtz\" (UID: \"683c61a5-c39a-400b-b673-19bbaf286482\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7dc9f6d87f-j6dtz" Mar 13 15:48:24 crc kubenswrapper[4907]: I0313 15:48:24.731681 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/d188fa9a-103d-49af-9626-90b7a3a5d70c-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-7dc9f6d87f-zlpcq\" (UID: \"d188fa9a-103d-49af-9626-90b7a3a5d70c\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7dc9f6d87f-zlpcq" Mar 13 15:48:24 crc kubenswrapper[4907]: I0313 15:48:24.733446 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/d188fa9a-103d-49af-9626-90b7a3a5d70c-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-7dc9f6d87f-zlpcq\" (UID: \"d188fa9a-103d-49af-9626-90b7a3a5d70c\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7dc9f6d87f-zlpcq" Mar 13 15:48:24 crc kubenswrapper[4907]: I0313 15:48:24.733831 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/683c61a5-c39a-400b-b673-19bbaf286482-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-7dc9f6d87f-j6dtz\" (UID: \"683c61a5-c39a-400b-b673-19bbaf286482\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7dc9f6d87f-j6dtz" Mar 13 15:48:24 crc kubenswrapper[4907]: I0313 15:48:24.759360 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7dc9f6d87f-zlpcq" Mar 13 15:48:24 crc kubenswrapper[4907]: I0313 15:48:24.765138 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/683c61a5-c39a-400b-b673-19bbaf286482-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-7dc9f6d87f-j6dtz\" (UID: \"683c61a5-c39a-400b-b673-19bbaf286482\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7dc9f6d87f-j6dtz" Mar 13 15:48:24 crc kubenswrapper[4907]: I0313 15:48:24.801580 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7dc9f6d87f-j6dtz" Mar 13 15:48:24 crc kubenswrapper[4907]: I0313 15:48:24.830533 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fqnp6\" (UniqueName: \"kubernetes.io/projected/dccfe9f4-59d2-4606-8d8c-a29eb01dbb6d-kube-api-access-fqnp6\") pod \"observability-operator-59bdc8b94-9hfqr\" (UID: \"dccfe9f4-59d2-4606-8d8c-a29eb01dbb6d\") " pod="openshift-operators/observability-operator-59bdc8b94-9hfqr" Mar 13 15:48:24 crc kubenswrapper[4907]: I0313 15:48:24.830708 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/dccfe9f4-59d2-4606-8d8c-a29eb01dbb6d-observability-operator-tls\") pod \"observability-operator-59bdc8b94-9hfqr\" (UID: \"dccfe9f4-59d2-4606-8d8c-a29eb01dbb6d\") " pod="openshift-operators/observability-operator-59bdc8b94-9hfqr" Mar 13 15:48:24 crc kubenswrapper[4907]: I0313 15:48:24.878985 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/perses-operator-5bf474d74f-2n9b9"] Mar 13 15:48:24 crc kubenswrapper[4907]: I0313 15:48:24.881035 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5bf474d74f-2n9b9" Mar 13 15:48:24 crc kubenswrapper[4907]: I0313 15:48:24.883716 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"perses-operator-dockercfg-6699h" Mar 13 15:48:24 crc kubenswrapper[4907]: I0313 15:48:24.893446 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5bf474d74f-2n9b9"] Mar 13 15:48:24 crc kubenswrapper[4907]: I0313 15:48:24.935213 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fqnp6\" (UniqueName: \"kubernetes.io/projected/dccfe9f4-59d2-4606-8d8c-a29eb01dbb6d-kube-api-access-fqnp6\") pod \"observability-operator-59bdc8b94-9hfqr\" (UID: \"dccfe9f4-59d2-4606-8d8c-a29eb01dbb6d\") " pod="openshift-operators/observability-operator-59bdc8b94-9hfqr" Mar 13 15:48:24 crc kubenswrapper[4907]: I0313 15:48:24.935477 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/dccfe9f4-59d2-4606-8d8c-a29eb01dbb6d-observability-operator-tls\") pod \"observability-operator-59bdc8b94-9hfqr\" (UID: \"dccfe9f4-59d2-4606-8d8c-a29eb01dbb6d\") " pod="openshift-operators/observability-operator-59bdc8b94-9hfqr" Mar 13 15:48:24 crc kubenswrapper[4907]: I0313 15:48:24.943742 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/dccfe9f4-59d2-4606-8d8c-a29eb01dbb6d-observability-operator-tls\") pod \"observability-operator-59bdc8b94-9hfqr\" (UID: \"dccfe9f4-59d2-4606-8d8c-a29eb01dbb6d\") " pod="openshift-operators/observability-operator-59bdc8b94-9hfqr" Mar 13 15:48:24 crc kubenswrapper[4907]: I0313 15:48:24.973489 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fqnp6\" (UniqueName: \"kubernetes.io/projected/dccfe9f4-59d2-4606-8d8c-a29eb01dbb6d-kube-api-access-fqnp6\") pod \"observability-operator-59bdc8b94-9hfqr\" (UID: \"dccfe9f4-59d2-4606-8d8c-a29eb01dbb6d\") " pod="openshift-operators/observability-operator-59bdc8b94-9hfqr" Mar 13 15:48:25 crc kubenswrapper[4907]: I0313 15:48:25.036904 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nm24m\" (UniqueName: \"kubernetes.io/projected/c5a33f26-edc6-4cb0-b9d8-58ee6113fecc-kube-api-access-nm24m\") pod \"perses-operator-5bf474d74f-2n9b9\" (UID: \"c5a33f26-edc6-4cb0-b9d8-58ee6113fecc\") " pod="openshift-operators/perses-operator-5bf474d74f-2n9b9" Mar 13 15:48:25 crc kubenswrapper[4907]: I0313 15:48:25.036965 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/c5a33f26-edc6-4cb0-b9d8-58ee6113fecc-openshift-service-ca\") pod \"perses-operator-5bf474d74f-2n9b9\" (UID: \"c5a33f26-edc6-4cb0-b9d8-58ee6113fecc\") " pod="openshift-operators/perses-operator-5bf474d74f-2n9b9" Mar 13 15:48:25 crc kubenswrapper[4907]: I0313 15:48:25.139934 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nm24m\" (UniqueName: \"kubernetes.io/projected/c5a33f26-edc6-4cb0-b9d8-58ee6113fecc-kube-api-access-nm24m\") pod \"perses-operator-5bf474d74f-2n9b9\" (UID: \"c5a33f26-edc6-4cb0-b9d8-58ee6113fecc\") " pod="openshift-operators/perses-operator-5bf474d74f-2n9b9" Mar 13 15:48:25 crc kubenswrapper[4907]: I0313 15:48:25.140017 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/c5a33f26-edc6-4cb0-b9d8-58ee6113fecc-openshift-service-ca\") pod \"perses-operator-5bf474d74f-2n9b9\" (UID: \"c5a33f26-edc6-4cb0-b9d8-58ee6113fecc\") " pod="openshift-operators/perses-operator-5bf474d74f-2n9b9" Mar 13 15:48:25 crc kubenswrapper[4907]: I0313 15:48:25.141173 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/c5a33f26-edc6-4cb0-b9d8-58ee6113fecc-openshift-service-ca\") pod \"perses-operator-5bf474d74f-2n9b9\" (UID: \"c5a33f26-edc6-4cb0-b9d8-58ee6113fecc\") " pod="openshift-operators/perses-operator-5bf474d74f-2n9b9" Mar 13 15:48:25 crc kubenswrapper[4907]: I0313 15:48:25.165745 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nm24m\" (UniqueName: \"kubernetes.io/projected/c5a33f26-edc6-4cb0-b9d8-58ee6113fecc-kube-api-access-nm24m\") pod \"perses-operator-5bf474d74f-2n9b9\" (UID: \"c5a33f26-edc6-4cb0-b9d8-58ee6113fecc\") " pod="openshift-operators/perses-operator-5bf474d74f-2n9b9" Mar 13 15:48:25 crc kubenswrapper[4907]: I0313 15:48:25.239337 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-59bdc8b94-9hfqr" Mar 13 15:48:25 crc kubenswrapper[4907]: I0313 15:48:25.361038 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5bf474d74f-2n9b9" Mar 13 15:48:25 crc kubenswrapper[4907]: I0313 15:48:25.436311 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-68bc856cb9-w2ngb"] Mar 13 15:48:25 crc kubenswrapper[4907]: I0313 15:48:25.554106 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-7dc9f6d87f-zlpcq"] Mar 13 15:48:25 crc kubenswrapper[4907]: I0313 15:48:25.654680 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-7dc9f6d87f-j6dtz"] Mar 13 15:48:25 crc kubenswrapper[4907]: W0313 15:48:25.668052 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod683c61a5_c39a_400b_b673_19bbaf286482.slice/crio-c888be885a02727fc3c9a0bf84f48814d7f5af0fa589398b52218f373a6b1699 WatchSource:0}: Error finding container c888be885a02727fc3c9a0bf84f48814d7f5af0fa589398b52218f373a6b1699: Status 404 returned error can't find the container with id c888be885a02727fc3c9a0bf84f48814d7f5af0fa589398b52218f373a6b1699 Mar 13 15:48:25 crc kubenswrapper[4907]: I0313 15:48:25.897282 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-59bdc8b94-9hfqr"] Mar 13 15:48:25 crc kubenswrapper[4907]: I0313 15:48:25.914089 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7dc9f6d87f-zlpcq" event={"ID":"d188fa9a-103d-49af-9626-90b7a3a5d70c","Type":"ContainerStarted","Data":"f0d8af2bb6aee7cb78c99769c70b55465d6cb45654b154e2d62291621b414e7d"} Mar 13 15:48:25 crc kubenswrapper[4907]: I0313 15:48:25.925504 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7dc9f6d87f-j6dtz" event={"ID":"683c61a5-c39a-400b-b673-19bbaf286482","Type":"ContainerStarted","Data":"c888be885a02727fc3c9a0bf84f48814d7f5af0fa589398b52218f373a6b1699"} Mar 13 15:48:25 crc kubenswrapper[4907]: I0313 15:48:25.938053 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-w2ngb" event={"ID":"92caa667-d9ee-4043-95b1-6475bcc7439b","Type":"ContainerStarted","Data":"92c25da4a3fec1bfea2fada32dd1fb03ce8186eae22530574b27ef9d987a11a4"} Mar 13 15:48:26 crc kubenswrapper[4907]: I0313 15:48:26.025491 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5bf474d74f-2n9b9"] Mar 13 15:48:26 crc kubenswrapper[4907]: I0313 15:48:26.952897 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5bf474d74f-2n9b9" event={"ID":"c5a33f26-edc6-4cb0-b9d8-58ee6113fecc","Type":"ContainerStarted","Data":"33323ecd7b6da6cb3e421e9ee7b60e850ef23ab3122705fdfede0831368e2243"} Mar 13 15:48:26 crc kubenswrapper[4907]: I0313 15:48:26.957953 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-59bdc8b94-9hfqr" event={"ID":"dccfe9f4-59d2-4606-8d8c-a29eb01dbb6d","Type":"ContainerStarted","Data":"251bf2b12c5832c0144196c77dc1bb3b002c61ccb327435531f68fbea3f397fe"} Mar 13 15:48:30 crc kubenswrapper[4907]: I0313 15:48:30.566405 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-7tp2h" podUID="e8c5a748-e7b7-43e0-82d3-c76673b17ea1" containerName="registry-server" probeResult="failure" output=< Mar 13 15:48:30 crc kubenswrapper[4907]: timeout: failed to connect service ":50051" within 1s Mar 13 15:48:30 crc kubenswrapper[4907]: > Mar 13 15:48:35 crc kubenswrapper[4907]: I0313 15:48:35.323762 4907 scope.go:117] "RemoveContainer" containerID="7414c52bdeec9a9b8ecea00c6ccd0ca4cd0ecf4d6c333567e9430b7214c0182e" Mar 13 15:48:37 crc kubenswrapper[4907]: I0313 15:48:37.120158 4907 scope.go:117] "RemoveContainer" containerID="75dd2c3ec3d1e72efcbe86eb956d5d57de584abd904ef3f55fe25aad75bb5498" Mar 13 15:48:37 crc kubenswrapper[4907]: I0313 15:48:37.230541 4907 scope.go:117] "RemoveContainer" containerID="7c596675c8272c2dfe2507df7b83ab578498cef4ae7b5f27995638598d3ea046" Mar 13 15:48:37 crc kubenswrapper[4907]: I0313 15:48:37.396430 4907 scope.go:117] "RemoveContainer" containerID="38713faad1d72bf574ce7d6ae2bd740312997c935666c1fb81c567bd339dc0d0" Mar 13 15:48:38 crc kubenswrapper[4907]: I0313 15:48:38.116688 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-59bdc8b94-9hfqr" event={"ID":"dccfe9f4-59d2-4606-8d8c-a29eb01dbb6d","Type":"ContainerStarted","Data":"44b0e8afc6cb817bd60894134bccaf31cfd9deaca012003d688c414c82d33f25"} Mar 13 15:48:38 crc kubenswrapper[4907]: I0313 15:48:38.117340 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/observability-operator-59bdc8b94-9hfqr" Mar 13 15:48:38 crc kubenswrapper[4907]: I0313 15:48:38.121958 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-w2ngb" event={"ID":"92caa667-d9ee-4043-95b1-6475bcc7439b","Type":"ContainerStarted","Data":"2639c49c39197ee0e143615b69caabdfef7076a9f335b899f297bd55a60ed0bb"} Mar 13 15:48:38 crc kubenswrapper[4907]: I0313 15:48:38.123147 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/observability-operator-59bdc8b94-9hfqr" Mar 13 15:48:38 crc kubenswrapper[4907]: I0313 15:48:38.136441 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5bf474d74f-2n9b9" event={"ID":"c5a33f26-edc6-4cb0-b9d8-58ee6113fecc","Type":"ContainerStarted","Data":"fe3213f56f2b0e957e334972f5b7c9ae49f4208945edd4c62c8b0b5ee34da857"} Mar 13 15:48:38 crc kubenswrapper[4907]: I0313 15:48:38.136988 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/perses-operator-5bf474d74f-2n9b9" Mar 13 15:48:38 crc kubenswrapper[4907]: I0313 15:48:38.143792 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7dc9f6d87f-zlpcq" event={"ID":"d188fa9a-103d-49af-9626-90b7a3a5d70c","Type":"ContainerStarted","Data":"e9139275039d32450405a70812febf14bac9d1aa924b322f48f5b14bc30cc0a8"} Mar 13 15:48:38 crc kubenswrapper[4907]: I0313 15:48:38.149529 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7dc9f6d87f-j6dtz" event={"ID":"683c61a5-c39a-400b-b673-19bbaf286482","Type":"ContainerStarted","Data":"0b60abf6ed19147654aff611bdc1b1ae57ef3fc7348dd8abd597b16cee46e52f"} Mar 13 15:48:38 crc kubenswrapper[4907]: I0313 15:48:38.163266 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/observability-operator-59bdc8b94-9hfqr" podStartSLOduration=2.828113116 podStartE2EDuration="14.163247371s" podCreationTimestamp="2026-03-13 15:48:24 +0000 UTC" firstStartedPulling="2026-03-13 15:48:25.895928557 +0000 UTC m=+6204.795716236" lastFinishedPulling="2026-03-13 15:48:37.231062802 +0000 UTC m=+6216.130850491" observedRunningTime="2026-03-13 15:48:38.140343876 +0000 UTC m=+6217.040131585" watchObservedRunningTime="2026-03-13 15:48:38.163247371 +0000 UTC m=+6217.063035060" Mar 13 15:48:38 crc kubenswrapper[4907]: I0313 15:48:38.197381 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/perses-operator-5bf474d74f-2n9b9" podStartSLOduration=3.117126732 podStartE2EDuration="14.197361733s" podCreationTimestamp="2026-03-13 15:48:24 +0000 UTC" firstStartedPulling="2026-03-13 15:48:26.041064643 +0000 UTC m=+6204.940852332" lastFinishedPulling="2026-03-13 15:48:37.121299644 +0000 UTC m=+6216.021087333" observedRunningTime="2026-03-13 15:48:38.193316703 +0000 UTC m=+6217.093104392" watchObservedRunningTime="2026-03-13 15:48:38.197361733 +0000 UTC m=+6217.097149422" Mar 13 15:48:38 crc kubenswrapper[4907]: I0313 15:48:38.240425 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-w2ngb" podStartSLOduration=2.546116391 podStartE2EDuration="14.240400789s" podCreationTimestamp="2026-03-13 15:48:24 +0000 UTC" firstStartedPulling="2026-03-13 15:48:25.427457188 +0000 UTC m=+6204.327244877" lastFinishedPulling="2026-03-13 15:48:37.121741586 +0000 UTC m=+6216.021529275" observedRunningTime="2026-03-13 15:48:38.236752289 +0000 UTC m=+6217.136539998" watchObservedRunningTime="2026-03-13 15:48:38.240400789 +0000 UTC m=+6217.140188478" Mar 13 15:48:38 crc kubenswrapper[4907]: I0313 15:48:38.265429 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7dc9f6d87f-j6dtz" podStartSLOduration=2.842234182 podStartE2EDuration="14.265406803s" podCreationTimestamp="2026-03-13 15:48:24 +0000 UTC" firstStartedPulling="2026-03-13 15:48:25.697029523 +0000 UTC m=+6204.596817212" lastFinishedPulling="2026-03-13 15:48:37.120202144 +0000 UTC m=+6216.019989833" observedRunningTime="2026-03-13 15:48:38.260411946 +0000 UTC m=+6217.160199655" watchObservedRunningTime="2026-03-13 15:48:38.265406803 +0000 UTC m=+6217.165194492" Mar 13 15:48:38 crc kubenswrapper[4907]: I0313 15:48:38.318086 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7dc9f6d87f-zlpcq" podStartSLOduration=2.774009668 podStartE2EDuration="14.318061711s" podCreationTimestamp="2026-03-13 15:48:24 +0000 UTC" firstStartedPulling="2026-03-13 15:48:25.583130921 +0000 UTC m=+6204.482918610" lastFinishedPulling="2026-03-13 15:48:37.127182964 +0000 UTC m=+6216.026970653" observedRunningTime="2026-03-13 15:48:38.305170589 +0000 UTC m=+6217.204958288" watchObservedRunningTime="2026-03-13 15:48:38.318061711 +0000 UTC m=+6217.217849400" Mar 13 15:48:40 crc kubenswrapper[4907]: I0313 15:48:40.563556 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-7tp2h" podUID="e8c5a748-e7b7-43e0-82d3-c76673b17ea1" containerName="registry-server" probeResult="failure" output=< Mar 13 15:48:40 crc kubenswrapper[4907]: timeout: failed to connect service ":50051" within 1s Mar 13 15:48:40 crc kubenswrapper[4907]: > Mar 13 15:48:45 crc kubenswrapper[4907]: I0313 15:48:45.364874 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/perses-operator-5bf474d74f-2n9b9" Mar 13 15:48:48 crc kubenswrapper[4907]: I0313 15:48:48.041727 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 15:48:48 crc kubenswrapper[4907]: I0313 15:48:48.042270 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 15:48:48 crc kubenswrapper[4907]: I0313 15:48:48.334851 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Mar 13 15:48:48 crc kubenswrapper[4907]: I0313 15:48:48.335136 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstackclient" podUID="74a5c619-2dde-4f75-9106-cad6a39265ce" containerName="openstackclient" containerID="cri-o://e99b6a0dae893428515991f160f7db931ea2980ec0fe9b55c8e67e61a9c011ff" gracePeriod=2 Mar 13 15:48:48 crc kubenswrapper[4907]: I0313 15:48:48.349469 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Mar 13 15:48:48 crc kubenswrapper[4907]: I0313 15:48:48.396660 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Mar 13 15:48:48 crc kubenswrapper[4907]: E0313 15:48:48.397240 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74a5c619-2dde-4f75-9106-cad6a39265ce" containerName="openstackclient" Mar 13 15:48:48 crc kubenswrapper[4907]: I0313 15:48:48.397265 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="74a5c619-2dde-4f75-9106-cad6a39265ce" containerName="openstackclient" Mar 13 15:48:48 crc kubenswrapper[4907]: I0313 15:48:48.397495 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="74a5c619-2dde-4f75-9106-cad6a39265ce" containerName="openstackclient" Mar 13 15:48:48 crc kubenswrapper[4907]: I0313 15:48:48.398447 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Mar 13 15:48:48 crc kubenswrapper[4907]: I0313 15:48:48.407967 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Mar 13 15:48:48 crc kubenswrapper[4907]: I0313 15:48:48.428503 4907 status_manager.go:875] "Failed to update status for pod" pod="openstack/openstackclient" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc7436bb-91c1-4364-9ef4-dba14540c5c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T15:48:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T15:48:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T15:48:48Z\\\",\\\"message\\\":\\\"containers with unready status: [openstackclient]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T15:48:48Z\\\",\\\"message\\\":\\\"containers with unready status: [openstackclient]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/podified-antelope-centos9/openstack-openstackclient@sha256:d765b589a5f7bc8573b3b192ed265654699012e6342cc4829bd8ea65a7b239a5\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"openstackclient\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/home/cloud-admin/.config/openstack/clouds.yaml\\\",\\\"name\\\":\\\"openstack-config\\\"},{\\\"mountPath\\\":\\\"/home/cloud-admin/.config/openstack/secure.yaml\\\",\\\"name\\\":\\\"openstack-config-secret\\\"},{\\\"mountPath\\\":\\\"/home/cloud-admin/cloudrc\\\",\\\"name\\\":\\\"openstack-config-secret\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jj76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T15:48:48Z\\\"}}\" for pod \"openstack\"/\"openstackclient\": pods \"openstackclient\" not found" Mar 13 15:48:48 crc kubenswrapper[4907]: I0313 15:48:48.429478 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Mar 13 15:48:48 crc kubenswrapper[4907]: E0313 15:48:48.430141 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[kube-api-access-jj76j openstack-config openstack-config-secret], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/openstackclient" podUID="bc7436bb-91c1-4364-9ef4-dba14540c5c2" Mar 13 15:48:48 crc kubenswrapper[4907]: I0313 15:48:48.444444 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Mar 13 15:48:48 crc kubenswrapper[4907]: I0313 15:48:48.480548 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Mar 13 15:48:48 crc kubenswrapper[4907]: I0313 15:48:48.482445 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Mar 13 15:48:48 crc kubenswrapper[4907]: I0313 15:48:48.503510 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/bc7436bb-91c1-4364-9ef4-dba14540c5c2-openstack-config\") pod \"openstackclient\" (UID: \"bc7436bb-91c1-4364-9ef4-dba14540c5c2\") " pod="openstack/openstackclient" Mar 13 15:48:48 crc kubenswrapper[4907]: I0313 15:48:48.503592 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jj76j\" (UniqueName: \"kubernetes.io/projected/bc7436bb-91c1-4364-9ef4-dba14540c5c2-kube-api-access-jj76j\") pod \"openstackclient\" (UID: \"bc7436bb-91c1-4364-9ef4-dba14540c5c2\") " pod="openstack/openstackclient" Mar 13 15:48:48 crc kubenswrapper[4907]: I0313 15:48:48.503747 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/bc7436bb-91c1-4364-9ef4-dba14540c5c2-openstack-config-secret\") pod \"openstackclient\" (UID: \"bc7436bb-91c1-4364-9ef4-dba14540c5c2\") " pod="openstack/openstackclient" Mar 13 15:48:48 crc kubenswrapper[4907]: I0313 15:48:48.510297 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Mar 13 15:48:48 crc kubenswrapper[4907]: I0313 15:48:48.525204 4907 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="bc7436bb-91c1-4364-9ef4-dba14540c5c2" podUID="2316d199-eef1-4ac1-9f63-cc73cba8d4c7" Mar 13 15:48:48 crc kubenswrapper[4907]: I0313 15:48:48.539125 4907 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="74a5c619-2dde-4f75-9106-cad6a39265ce" podUID="2316d199-eef1-4ac1-9f63-cc73cba8d4c7" Mar 13 15:48:48 crc kubenswrapper[4907]: I0313 15:48:48.605896 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/2316d199-eef1-4ac1-9f63-cc73cba8d4c7-openstack-config-secret\") pod \"openstackclient\" (UID: \"2316d199-eef1-4ac1-9f63-cc73cba8d4c7\") " pod="openstack/openstackclient" Mar 13 15:48:48 crc kubenswrapper[4907]: I0313 15:48:48.606234 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/bc7436bb-91c1-4364-9ef4-dba14540c5c2-openstack-config-secret\") pod \"openstackclient\" (UID: \"bc7436bb-91c1-4364-9ef4-dba14540c5c2\") " pod="openstack/openstackclient" Mar 13 15:48:48 crc kubenswrapper[4907]: I0313 15:48:48.606312 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/bc7436bb-91c1-4364-9ef4-dba14540c5c2-openstack-config\") pod \"openstackclient\" (UID: \"bc7436bb-91c1-4364-9ef4-dba14540c5c2\") " pod="openstack/openstackclient" Mar 13 15:48:48 crc kubenswrapper[4907]: I0313 15:48:48.606343 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/2316d199-eef1-4ac1-9f63-cc73cba8d4c7-openstack-config\") pod \"openstackclient\" (UID: \"2316d199-eef1-4ac1-9f63-cc73cba8d4c7\") " pod="openstack/openstackclient" Mar 13 15:48:48 crc kubenswrapper[4907]: I0313 15:48:48.606422 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vxg26\" (UniqueName: \"kubernetes.io/projected/2316d199-eef1-4ac1-9f63-cc73cba8d4c7-kube-api-access-vxg26\") pod \"openstackclient\" (UID: \"2316d199-eef1-4ac1-9f63-cc73cba8d4c7\") " pod="openstack/openstackclient" Mar 13 15:48:48 crc kubenswrapper[4907]: I0313 15:48:48.606467 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jj76j\" (UniqueName: \"kubernetes.io/projected/bc7436bb-91c1-4364-9ef4-dba14540c5c2-kube-api-access-jj76j\") pod \"openstackclient\" (UID: \"bc7436bb-91c1-4364-9ef4-dba14540c5c2\") " pod="openstack/openstackclient" Mar 13 15:48:48 crc kubenswrapper[4907]: I0313 15:48:48.607589 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/bc7436bb-91c1-4364-9ef4-dba14540c5c2-openstack-config\") pod \"openstackclient\" (UID: \"bc7436bb-91c1-4364-9ef4-dba14540c5c2\") " pod="openstack/openstackclient" Mar 13 15:48:48 crc kubenswrapper[4907]: I0313 15:48:48.609391 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Mar 13 15:48:48 crc kubenswrapper[4907]: I0313 15:48:48.610868 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Mar 13 15:48:48 crc kubenswrapper[4907]: I0313 15:48:48.620795 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-bl4lv" Mar 13 15:48:48 crc kubenswrapper[4907]: I0313 15:48:48.620930 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/bc7436bb-91c1-4364-9ef4-dba14540c5c2-openstack-config-secret\") pod \"openstackclient\" (UID: \"bc7436bb-91c1-4364-9ef4-dba14540c5c2\") " pod="openstack/openstackclient" Mar 13 15:48:48 crc kubenswrapper[4907]: E0313 15:48:48.621116 4907 projected.go:194] Error preparing data for projected volume kube-api-access-jj76j for pod openstack/openstackclient: failed to fetch token: serviceaccounts "openstackclient-openstackclient" is forbidden: the UID in the bound object reference (bc7436bb-91c1-4364-9ef4-dba14540c5c2) does not match the UID in record. The object might have been deleted and then recreated Mar 13 15:48:48 crc kubenswrapper[4907]: E0313 15:48:48.621176 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/bc7436bb-91c1-4364-9ef4-dba14540c5c2-kube-api-access-jj76j podName:bc7436bb-91c1-4364-9ef4-dba14540c5c2 nodeName:}" failed. No retries permitted until 2026-03-13 15:48:49.121156849 +0000 UTC m=+6228.020944538 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-jj76j" (UniqueName: "kubernetes.io/projected/bc7436bb-91c1-4364-9ef4-dba14540c5c2-kube-api-access-jj76j") pod "openstackclient" (UID: "bc7436bb-91c1-4364-9ef4-dba14540c5c2") : failed to fetch token: serviceaccounts "openstackclient-openstackclient" is forbidden: the UID in the bound object reference (bc7436bb-91c1-4364-9ef4-dba14540c5c2) does not match the UID in record. The object might have been deleted and then recreated Mar 13 15:48:48 crc kubenswrapper[4907]: I0313 15:48:48.639221 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Mar 13 15:48:48 crc kubenswrapper[4907]: I0313 15:48:48.709086 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m4tqw\" (UniqueName: \"kubernetes.io/projected/53c58ef1-fd3c-4d26-9b00-5ce0bb0cc592-kube-api-access-m4tqw\") pod \"kube-state-metrics-0\" (UID: \"53c58ef1-fd3c-4d26-9b00-5ce0bb0cc592\") " pod="openstack/kube-state-metrics-0" Mar 13 15:48:48 crc kubenswrapper[4907]: I0313 15:48:48.709613 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/2316d199-eef1-4ac1-9f63-cc73cba8d4c7-openstack-config-secret\") pod \"openstackclient\" (UID: \"2316d199-eef1-4ac1-9f63-cc73cba8d4c7\") " pod="openstack/openstackclient" Mar 13 15:48:48 crc kubenswrapper[4907]: I0313 15:48:48.709725 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/2316d199-eef1-4ac1-9f63-cc73cba8d4c7-openstack-config\") pod \"openstackclient\" (UID: \"2316d199-eef1-4ac1-9f63-cc73cba8d4c7\") " pod="openstack/openstackclient" Mar 13 15:48:48 crc kubenswrapper[4907]: I0313 15:48:48.709765 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vxg26\" (UniqueName: \"kubernetes.io/projected/2316d199-eef1-4ac1-9f63-cc73cba8d4c7-kube-api-access-vxg26\") pod \"openstackclient\" (UID: \"2316d199-eef1-4ac1-9f63-cc73cba8d4c7\") " pod="openstack/openstackclient" Mar 13 15:48:48 crc kubenswrapper[4907]: I0313 15:48:48.714794 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/2316d199-eef1-4ac1-9f63-cc73cba8d4c7-openstack-config-secret\") pod \"openstackclient\" (UID: \"2316d199-eef1-4ac1-9f63-cc73cba8d4c7\") " pod="openstack/openstackclient" Mar 13 15:48:48 crc kubenswrapper[4907]: I0313 15:48:48.714837 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/2316d199-eef1-4ac1-9f63-cc73cba8d4c7-openstack-config\") pod \"openstackclient\" (UID: \"2316d199-eef1-4ac1-9f63-cc73cba8d4c7\") " pod="openstack/openstackclient" Mar 13 15:48:48 crc kubenswrapper[4907]: I0313 15:48:48.821261 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vxg26\" (UniqueName: \"kubernetes.io/projected/2316d199-eef1-4ac1-9f63-cc73cba8d4c7-kube-api-access-vxg26\") pod \"openstackclient\" (UID: \"2316d199-eef1-4ac1-9f63-cc73cba8d4c7\") " pod="openstack/openstackclient" Mar 13 15:48:48 crc kubenswrapper[4907]: I0313 15:48:48.823102 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Mar 13 15:48:48 crc kubenswrapper[4907]: I0313 15:48:48.868464 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m4tqw\" (UniqueName: \"kubernetes.io/projected/53c58ef1-fd3c-4d26-9b00-5ce0bb0cc592-kube-api-access-m4tqw\") pod \"kube-state-metrics-0\" (UID: \"53c58ef1-fd3c-4d26-9b00-5ce0bb0cc592\") " pod="openstack/kube-state-metrics-0" Mar 13 15:48:48 crc kubenswrapper[4907]: I0313 15:48:48.910715 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m4tqw\" (UniqueName: \"kubernetes.io/projected/53c58ef1-fd3c-4d26-9b00-5ce0bb0cc592-kube-api-access-m4tqw\") pod \"kube-state-metrics-0\" (UID: \"53c58ef1-fd3c-4d26-9b00-5ce0bb0cc592\") " pod="openstack/kube-state-metrics-0" Mar 13 15:48:49 crc kubenswrapper[4907]: I0313 15:48:49.006399 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Mar 13 15:48:49 crc kubenswrapper[4907]: I0313 15:48:49.184211 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jj76j\" (UniqueName: \"kubernetes.io/projected/bc7436bb-91c1-4364-9ef4-dba14540c5c2-kube-api-access-jj76j\") pod \"openstackclient\" (UID: \"bc7436bb-91c1-4364-9ef4-dba14540c5c2\") " pod="openstack/openstackclient" Mar 13 15:48:49 crc kubenswrapper[4907]: E0313 15:48:49.186694 4907 projected.go:194] Error preparing data for projected volume kube-api-access-jj76j for pod openstack/openstackclient: failed to fetch token: serviceaccounts "openstackclient-openstackclient" is forbidden: the UID in the bound object reference (bc7436bb-91c1-4364-9ef4-dba14540c5c2) does not match the UID in record. The object might have been deleted and then recreated Mar 13 15:48:49 crc kubenswrapper[4907]: E0313 15:48:49.186758 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/bc7436bb-91c1-4364-9ef4-dba14540c5c2-kube-api-access-jj76j podName:bc7436bb-91c1-4364-9ef4-dba14540c5c2 nodeName:}" failed. No retries permitted until 2026-03-13 15:48:50.186739182 +0000 UTC m=+6229.086526861 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-jj76j" (UniqueName: "kubernetes.io/projected/bc7436bb-91c1-4364-9ef4-dba14540c5c2-kube-api-access-jj76j") pod "openstackclient" (UID: "bc7436bb-91c1-4364-9ef4-dba14540c5c2") : failed to fetch token: serviceaccounts "openstackclient-openstackclient" is forbidden: the UID in the bound object reference (bc7436bb-91c1-4364-9ef4-dba14540c5c2) does not match the UID in record. The object might have been deleted and then recreated Mar 13 15:48:49 crc kubenswrapper[4907]: I0313 15:48:49.267816 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Mar 13 15:48:49 crc kubenswrapper[4907]: I0313 15:48:49.271804 4907 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="bc7436bb-91c1-4364-9ef4-dba14540c5c2" podUID="2316d199-eef1-4ac1-9f63-cc73cba8d4c7" Mar 13 15:48:49 crc kubenswrapper[4907]: I0313 15:48:49.295183 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Mar 13 15:48:49 crc kubenswrapper[4907]: I0313 15:48:49.300039 4907 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="bc7436bb-91c1-4364-9ef4-dba14540c5c2" podUID="2316d199-eef1-4ac1-9f63-cc73cba8d4c7" Mar 13 15:48:49 crc kubenswrapper[4907]: I0313 15:48:49.388471 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/alertmanager-metric-storage-0"] Mar 13 15:48:49 crc kubenswrapper[4907]: I0313 15:48:49.390973 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/alertmanager-metric-storage-0" Mar 13 15:48:49 crc kubenswrapper[4907]: I0313 15:48:49.409502 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-alertmanager-dockercfg-hc86w" Mar 13 15:48:49 crc kubenswrapper[4907]: I0313 15:48:49.409673 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-tls-assets-0" Mar 13 15:48:49 crc kubenswrapper[4907]: I0313 15:48:49.409793 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-generated" Mar 13 15:48:49 crc kubenswrapper[4907]: I0313 15:48:49.409907 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-web-config" Mar 13 15:48:49 crc kubenswrapper[4907]: I0313 15:48:49.410029 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-cluster-tls-config" Mar 13 15:48:49 crc kubenswrapper[4907]: I0313 15:48:49.423600 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/alertmanager-metric-storage-0"] Mar 13 15:48:49 crc kubenswrapper[4907]: I0313 15:48:49.490771 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/bc7436bb-91c1-4364-9ef4-dba14540c5c2-openstack-config\") pod \"bc7436bb-91c1-4364-9ef4-dba14540c5c2\" (UID: \"bc7436bb-91c1-4364-9ef4-dba14540c5c2\") " Mar 13 15:48:49 crc kubenswrapper[4907]: I0313 15:48:49.490864 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/bc7436bb-91c1-4364-9ef4-dba14540c5c2-openstack-config-secret\") pod \"bc7436bb-91c1-4364-9ef4-dba14540c5c2\" (UID: \"bc7436bb-91c1-4364-9ef4-dba14540c5c2\") " Mar 13 15:48:49 crc kubenswrapper[4907]: I0313 15:48:49.492434 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bc7436bb-91c1-4364-9ef4-dba14540c5c2-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "bc7436bb-91c1-4364-9ef4-dba14540c5c2" (UID: "bc7436bb-91c1-4364-9ef4-dba14540c5c2"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:48:49 crc kubenswrapper[4907]: I0313 15:48:49.497465 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/d29f6de0-3229-477c-86a6-550f2d465592-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"d29f6de0-3229-477c-86a6-550f2d465592\") " pod="openstack/alertmanager-metric-storage-0" Mar 13 15:48:49 crc kubenswrapper[4907]: I0313 15:48:49.497538 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mhctv\" (UniqueName: \"kubernetes.io/projected/d29f6de0-3229-477c-86a6-550f2d465592-kube-api-access-mhctv\") pod \"alertmanager-metric-storage-0\" (UID: \"d29f6de0-3229-477c-86a6-550f2d465592\") " pod="openstack/alertmanager-metric-storage-0" Mar 13 15:48:49 crc kubenswrapper[4907]: I0313 15:48:49.497776 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/d29f6de0-3229-477c-86a6-550f2d465592-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"d29f6de0-3229-477c-86a6-550f2d465592\") " pod="openstack/alertmanager-metric-storage-0" Mar 13 15:48:49 crc kubenswrapper[4907]: I0313 15:48:49.497818 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/d29f6de0-3229-477c-86a6-550f2d465592-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"d29f6de0-3229-477c-86a6-550f2d465592\") " pod="openstack/alertmanager-metric-storage-0" Mar 13 15:48:49 crc kubenswrapper[4907]: I0313 15:48:49.497845 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/d29f6de0-3229-477c-86a6-550f2d465592-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"d29f6de0-3229-477c-86a6-550f2d465592\") " pod="openstack/alertmanager-metric-storage-0" Mar 13 15:48:49 crc kubenswrapper[4907]: I0313 15:48:49.497896 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/d29f6de0-3229-477c-86a6-550f2d465592-cluster-tls-config\") pod \"alertmanager-metric-storage-0\" (UID: \"d29f6de0-3229-477c-86a6-550f2d465592\") " pod="openstack/alertmanager-metric-storage-0" Mar 13 15:48:49 crc kubenswrapper[4907]: I0313 15:48:49.498183 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/d29f6de0-3229-477c-86a6-550f2d465592-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"d29f6de0-3229-477c-86a6-550f2d465592\") " pod="openstack/alertmanager-metric-storage-0" Mar 13 15:48:49 crc kubenswrapper[4907]: I0313 15:48:49.498406 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jj76j\" (UniqueName: \"kubernetes.io/projected/bc7436bb-91c1-4364-9ef4-dba14540c5c2-kube-api-access-jj76j\") on node \"crc\" DevicePath \"\"" Mar 13 15:48:49 crc kubenswrapper[4907]: I0313 15:48:49.498426 4907 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/bc7436bb-91c1-4364-9ef4-dba14540c5c2-openstack-config\") on node \"crc\" DevicePath \"\"" Mar 13 15:48:49 crc kubenswrapper[4907]: I0313 15:48:49.512270 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc7436bb-91c1-4364-9ef4-dba14540c5c2-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "bc7436bb-91c1-4364-9ef4-dba14540c5c2" (UID: "bc7436bb-91c1-4364-9ef4-dba14540c5c2"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:48:49 crc kubenswrapper[4907]: I0313 15:48:49.602277 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/d29f6de0-3229-477c-86a6-550f2d465592-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"d29f6de0-3229-477c-86a6-550f2d465592\") " pod="openstack/alertmanager-metric-storage-0" Mar 13 15:48:49 crc kubenswrapper[4907]: I0313 15:48:49.602389 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/d29f6de0-3229-477c-86a6-550f2d465592-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"d29f6de0-3229-477c-86a6-550f2d465592\") " pod="openstack/alertmanager-metric-storage-0" Mar 13 15:48:49 crc kubenswrapper[4907]: I0313 15:48:49.602419 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mhctv\" (UniqueName: \"kubernetes.io/projected/d29f6de0-3229-477c-86a6-550f2d465592-kube-api-access-mhctv\") pod \"alertmanager-metric-storage-0\" (UID: \"d29f6de0-3229-477c-86a6-550f2d465592\") " pod="openstack/alertmanager-metric-storage-0" Mar 13 15:48:49 crc kubenswrapper[4907]: I0313 15:48:49.602491 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/d29f6de0-3229-477c-86a6-550f2d465592-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"d29f6de0-3229-477c-86a6-550f2d465592\") " pod="openstack/alertmanager-metric-storage-0" Mar 13 15:48:49 crc kubenswrapper[4907]: I0313 15:48:49.602513 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/d29f6de0-3229-477c-86a6-550f2d465592-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"d29f6de0-3229-477c-86a6-550f2d465592\") " pod="openstack/alertmanager-metric-storage-0" Mar 13 15:48:49 crc kubenswrapper[4907]: I0313 15:48:49.602529 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/d29f6de0-3229-477c-86a6-550f2d465592-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"d29f6de0-3229-477c-86a6-550f2d465592\") " pod="openstack/alertmanager-metric-storage-0" Mar 13 15:48:49 crc kubenswrapper[4907]: I0313 15:48:49.602546 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/d29f6de0-3229-477c-86a6-550f2d465592-cluster-tls-config\") pod \"alertmanager-metric-storage-0\" (UID: \"d29f6de0-3229-477c-86a6-550f2d465592\") " pod="openstack/alertmanager-metric-storage-0" Mar 13 15:48:49 crc kubenswrapper[4907]: I0313 15:48:49.602605 4907 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/bc7436bb-91c1-4364-9ef4-dba14540c5c2-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Mar 13 15:48:49 crc kubenswrapper[4907]: I0313 15:48:49.613137 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/d29f6de0-3229-477c-86a6-550f2d465592-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"d29f6de0-3229-477c-86a6-550f2d465592\") " pod="openstack/alertmanager-metric-storage-0" Mar 13 15:48:49 crc kubenswrapper[4907]: I0313 15:48:49.619220 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/d29f6de0-3229-477c-86a6-550f2d465592-cluster-tls-config\") pod \"alertmanager-metric-storage-0\" (UID: \"d29f6de0-3229-477c-86a6-550f2d465592\") " pod="openstack/alertmanager-metric-storage-0" Mar 13 15:48:49 crc kubenswrapper[4907]: I0313 15:48:49.623238 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/d29f6de0-3229-477c-86a6-550f2d465592-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"d29f6de0-3229-477c-86a6-550f2d465592\") " pod="openstack/alertmanager-metric-storage-0" Mar 13 15:48:49 crc kubenswrapper[4907]: I0313 15:48:49.623687 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/d29f6de0-3229-477c-86a6-550f2d465592-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"d29f6de0-3229-477c-86a6-550f2d465592\") " pod="openstack/alertmanager-metric-storage-0" Mar 13 15:48:49 crc kubenswrapper[4907]: I0313 15:48:49.636944 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/d29f6de0-3229-477c-86a6-550f2d465592-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"d29f6de0-3229-477c-86a6-550f2d465592\") " pod="openstack/alertmanager-metric-storage-0" Mar 13 15:48:49 crc kubenswrapper[4907]: I0313 15:48:49.641861 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/d29f6de0-3229-477c-86a6-550f2d465592-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"d29f6de0-3229-477c-86a6-550f2d465592\") " pod="openstack/alertmanager-metric-storage-0" Mar 13 15:48:49 crc kubenswrapper[4907]: I0313 15:48:49.647788 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Mar 13 15:48:49 crc kubenswrapper[4907]: I0313 15:48:49.718305 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mhctv\" (UniqueName: \"kubernetes.io/projected/d29f6de0-3229-477c-86a6-550f2d465592-kube-api-access-mhctv\") pod \"alertmanager-metric-storage-0\" (UID: \"d29f6de0-3229-477c-86a6-550f2d465592\") " pod="openstack/alertmanager-metric-storage-0" Mar 13 15:48:49 crc kubenswrapper[4907]: I0313 15:48:49.821399 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc7436bb-91c1-4364-9ef4-dba14540c5c2" path="/var/lib/kubelet/pods/bc7436bb-91c1-4364-9ef4-dba14540c5c2/volumes" Mar 13 15:48:49 crc kubenswrapper[4907]: I0313 15:48:49.925242 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Mar 13 15:48:50 crc kubenswrapper[4907]: I0313 15:48:50.002142 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Mar 13 15:48:50 crc kubenswrapper[4907]: I0313 15:48:50.004724 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Mar 13 15:48:50 crc kubenswrapper[4907]: I0313 15:48:50.019520 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/alertmanager-metric-storage-0" Mar 13 15:48:50 crc kubenswrapper[4907]: I0313 15:48:50.028008 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-1" Mar 13 15:48:50 crc kubenswrapper[4907]: I0313 15:48:50.028807 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Mar 13 15:48:50 crc kubenswrapper[4907]: I0313 15:48:50.028947 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-5kg6f" Mar 13 15:48:50 crc kubenswrapper[4907]: I0313 15:48:50.029042 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Mar 13 15:48:50 crc kubenswrapper[4907]: I0313 15:48:50.029140 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Mar 13 15:48:50 crc kubenswrapper[4907]: I0313 15:48:50.029232 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Mar 13 15:48:50 crc kubenswrapper[4907]: I0313 15:48:50.029325 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Mar 13 15:48:50 crc kubenswrapper[4907]: I0313 15:48:50.040283 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-2" Mar 13 15:48:50 crc kubenswrapper[4907]: I0313 15:48:50.079597 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Mar 13 15:48:50 crc kubenswrapper[4907]: I0313 15:48:50.117397 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-c036-account-create-update-26jl8"] Mar 13 15:48:50 crc kubenswrapper[4907]: I0313 15:48:50.122706 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9qphj\" (UniqueName: \"kubernetes.io/projected/3e5a8905-7f71-4979-aee3-f81e772ea46a-kube-api-access-9qphj\") pod \"prometheus-metric-storage-0\" (UID: \"3e5a8905-7f71-4979-aee3-f81e772ea46a\") " pod="openstack/prometheus-metric-storage-0" Mar 13 15:48:50 crc kubenswrapper[4907]: I0313 15:48:50.122768 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-1\" (UniqueName: \"kubernetes.io/configmap/3e5a8905-7f71-4979-aee3-f81e772ea46a-prometheus-metric-storage-rulefiles-1\") pod \"prometheus-metric-storage-0\" (UID: \"3e5a8905-7f71-4979-aee3-f81e772ea46a\") " pod="openstack/prometheus-metric-storage-0" Mar 13 15:48:50 crc kubenswrapper[4907]: I0313 15:48:50.122806 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/3e5a8905-7f71-4979-aee3-f81e772ea46a-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"3e5a8905-7f71-4979-aee3-f81e772ea46a\") " pod="openstack/prometheus-metric-storage-0" Mar 13 15:48:50 crc kubenswrapper[4907]: I0313 15:48:50.122856 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/3e5a8905-7f71-4979-aee3-f81e772ea46a-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"3e5a8905-7f71-4979-aee3-f81e772ea46a\") " pod="openstack/prometheus-metric-storage-0" Mar 13 15:48:50 crc kubenswrapper[4907]: I0313 15:48:50.135969 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/3e5a8905-7f71-4979-aee3-f81e772ea46a-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"3e5a8905-7f71-4979-aee3-f81e772ea46a\") " pod="openstack/prometheus-metric-storage-0" Mar 13 15:48:50 crc kubenswrapper[4907]: I0313 15:48:50.136105 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/3e5a8905-7f71-4979-aee3-f81e772ea46a-config\") pod \"prometheus-metric-storage-0\" (UID: \"3e5a8905-7f71-4979-aee3-f81e772ea46a\") " pod="openstack/prometheus-metric-storage-0" Mar 13 15:48:50 crc kubenswrapper[4907]: I0313 15:48:50.136188 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-ac09aaac-45f0-4ecf-8f08-3e715b4f4606\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ac09aaac-45f0-4ecf-8f08-3e715b4f4606\") pod \"prometheus-metric-storage-0\" (UID: \"3e5a8905-7f71-4979-aee3-f81e772ea46a\") " pod="openstack/prometheus-metric-storage-0" Mar 13 15:48:50 crc kubenswrapper[4907]: I0313 15:48:50.136312 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/3e5a8905-7f71-4979-aee3-f81e772ea46a-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"3e5a8905-7f71-4979-aee3-f81e772ea46a\") " pod="openstack/prometheus-metric-storage-0" Mar 13 15:48:50 crc kubenswrapper[4907]: I0313 15:48:50.136400 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-2\" (UniqueName: \"kubernetes.io/configmap/3e5a8905-7f71-4979-aee3-f81e772ea46a-prometheus-metric-storage-rulefiles-2\") pod \"prometheus-metric-storage-0\" (UID: \"3e5a8905-7f71-4979-aee3-f81e772ea46a\") " pod="openstack/prometheus-metric-storage-0" Mar 13 15:48:50 crc kubenswrapper[4907]: I0313 15:48:50.136447 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/3e5a8905-7f71-4979-aee3-f81e772ea46a-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"3e5a8905-7f71-4979-aee3-f81e772ea46a\") " pod="openstack/prometheus-metric-storage-0" Mar 13 15:48:50 crc kubenswrapper[4907]: I0313 15:48:50.152944 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-5svds"] Mar 13 15:48:50 crc kubenswrapper[4907]: I0313 15:48:50.180961 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-5svds"] Mar 13 15:48:50 crc kubenswrapper[4907]: I0313 15:48:50.201974 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-c036-account-create-update-26jl8"] Mar 13 15:48:50 crc kubenswrapper[4907]: I0313 15:48:50.238097 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9qphj\" (UniqueName: \"kubernetes.io/projected/3e5a8905-7f71-4979-aee3-f81e772ea46a-kube-api-access-9qphj\") pod \"prometheus-metric-storage-0\" (UID: \"3e5a8905-7f71-4979-aee3-f81e772ea46a\") " pod="openstack/prometheus-metric-storage-0" Mar 13 15:48:50 crc kubenswrapper[4907]: I0313 15:48:50.238157 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-1\" (UniqueName: \"kubernetes.io/configmap/3e5a8905-7f71-4979-aee3-f81e772ea46a-prometheus-metric-storage-rulefiles-1\") pod \"prometheus-metric-storage-0\" (UID: \"3e5a8905-7f71-4979-aee3-f81e772ea46a\") " pod="openstack/prometheus-metric-storage-0" Mar 13 15:48:50 crc kubenswrapper[4907]: I0313 15:48:50.238202 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/3e5a8905-7f71-4979-aee3-f81e772ea46a-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"3e5a8905-7f71-4979-aee3-f81e772ea46a\") " pod="openstack/prometheus-metric-storage-0" Mar 13 15:48:50 crc kubenswrapper[4907]: I0313 15:48:50.238238 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/3e5a8905-7f71-4979-aee3-f81e772ea46a-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"3e5a8905-7f71-4979-aee3-f81e772ea46a\") " pod="openstack/prometheus-metric-storage-0" Mar 13 15:48:50 crc kubenswrapper[4907]: I0313 15:48:50.238279 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/3e5a8905-7f71-4979-aee3-f81e772ea46a-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"3e5a8905-7f71-4979-aee3-f81e772ea46a\") " pod="openstack/prometheus-metric-storage-0" Mar 13 15:48:50 crc kubenswrapper[4907]: I0313 15:48:50.238304 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/3e5a8905-7f71-4979-aee3-f81e772ea46a-config\") pod \"prometheus-metric-storage-0\" (UID: \"3e5a8905-7f71-4979-aee3-f81e772ea46a\") " pod="openstack/prometheus-metric-storage-0" Mar 13 15:48:50 crc kubenswrapper[4907]: I0313 15:48:50.238333 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-ac09aaac-45f0-4ecf-8f08-3e715b4f4606\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ac09aaac-45f0-4ecf-8f08-3e715b4f4606\") pod \"prometheus-metric-storage-0\" (UID: \"3e5a8905-7f71-4979-aee3-f81e772ea46a\") " pod="openstack/prometheus-metric-storage-0" Mar 13 15:48:50 crc kubenswrapper[4907]: I0313 15:48:50.238370 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/3e5a8905-7f71-4979-aee3-f81e772ea46a-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"3e5a8905-7f71-4979-aee3-f81e772ea46a\") " pod="openstack/prometheus-metric-storage-0" Mar 13 15:48:50 crc kubenswrapper[4907]: I0313 15:48:50.238396 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-2\" (UniqueName: \"kubernetes.io/configmap/3e5a8905-7f71-4979-aee3-f81e772ea46a-prometheus-metric-storage-rulefiles-2\") pod \"prometheus-metric-storage-0\" (UID: \"3e5a8905-7f71-4979-aee3-f81e772ea46a\") " pod="openstack/prometheus-metric-storage-0" Mar 13 15:48:50 crc kubenswrapper[4907]: I0313 15:48:50.238418 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/3e5a8905-7f71-4979-aee3-f81e772ea46a-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"3e5a8905-7f71-4979-aee3-f81e772ea46a\") " pod="openstack/prometheus-metric-storage-0" Mar 13 15:48:50 crc kubenswrapper[4907]: I0313 15:48:50.242341 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-1\" (UniqueName: \"kubernetes.io/configmap/3e5a8905-7f71-4979-aee3-f81e772ea46a-prometheus-metric-storage-rulefiles-1\") pod \"prometheus-metric-storage-0\" (UID: \"3e5a8905-7f71-4979-aee3-f81e772ea46a\") " pod="openstack/prometheus-metric-storage-0" Mar 13 15:48:50 crc kubenswrapper[4907]: I0313 15:48:50.242805 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-2\" (UniqueName: \"kubernetes.io/configmap/3e5a8905-7f71-4979-aee3-f81e772ea46a-prometheus-metric-storage-rulefiles-2\") pod \"prometheus-metric-storage-0\" (UID: \"3e5a8905-7f71-4979-aee3-f81e772ea46a\") " pod="openstack/prometheus-metric-storage-0" Mar 13 15:48:50 crc kubenswrapper[4907]: I0313 15:48:50.243279 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/3e5a8905-7f71-4979-aee3-f81e772ea46a-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"3e5a8905-7f71-4979-aee3-f81e772ea46a\") " pod="openstack/prometheus-metric-storage-0" Mar 13 15:48:50 crc kubenswrapper[4907]: I0313 15:48:50.253935 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/3e5a8905-7f71-4979-aee3-f81e772ea46a-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"3e5a8905-7f71-4979-aee3-f81e772ea46a\") " pod="openstack/prometheus-metric-storage-0" Mar 13 15:48:50 crc kubenswrapper[4907]: I0313 15:48:50.258292 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/3e5a8905-7f71-4979-aee3-f81e772ea46a-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"3e5a8905-7f71-4979-aee3-f81e772ea46a\") " pod="openstack/prometheus-metric-storage-0" Mar 13 15:48:50 crc kubenswrapper[4907]: I0313 15:48:50.260720 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/3e5a8905-7f71-4979-aee3-f81e772ea46a-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"3e5a8905-7f71-4979-aee3-f81e772ea46a\") " pod="openstack/prometheus-metric-storage-0" Mar 13 15:48:50 crc kubenswrapper[4907]: I0313 15:48:50.267715 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/3e5a8905-7f71-4979-aee3-f81e772ea46a-config\") pod \"prometheus-metric-storage-0\" (UID: \"3e5a8905-7f71-4979-aee3-f81e772ea46a\") " pod="openstack/prometheus-metric-storage-0" Mar 13 15:48:50 crc kubenswrapper[4907]: I0313 15:48:50.267955 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/3e5a8905-7f71-4979-aee3-f81e772ea46a-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"3e5a8905-7f71-4979-aee3-f81e772ea46a\") " pod="openstack/prometheus-metric-storage-0" Mar 13 15:48:50 crc kubenswrapper[4907]: I0313 15:48:50.273612 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9qphj\" (UniqueName: \"kubernetes.io/projected/3e5a8905-7f71-4979-aee3-f81e772ea46a-kube-api-access-9qphj\") pod \"prometheus-metric-storage-0\" (UID: \"3e5a8905-7f71-4979-aee3-f81e772ea46a\") " pod="openstack/prometheus-metric-storage-0" Mar 13 15:48:50 crc kubenswrapper[4907]: I0313 15:48:50.312611 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"2316d199-eef1-4ac1-9f63-cc73cba8d4c7","Type":"ContainerStarted","Data":"e822b148bbadd75b0f6facc5f6ac13f0b1cab8ff45ef96f5417ba6cb764deb8a"} Mar 13 15:48:50 crc kubenswrapper[4907]: I0313 15:48:50.313864 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Mar 13 15:48:50 crc kubenswrapper[4907]: I0313 15:48:50.314585 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"53c58ef1-fd3c-4d26-9b00-5ce0bb0cc592","Type":"ContainerStarted","Data":"2356f8f220308508437db8c12932926e505f0ad6e95771702f34ffa0f6d6e8e1"} Mar 13 15:48:50 crc kubenswrapper[4907]: I0313 15:48:50.324232 4907 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="bc7436bb-91c1-4364-9ef4-dba14540c5c2" podUID="2316d199-eef1-4ac1-9f63-cc73cba8d4c7" Mar 13 15:48:50 crc kubenswrapper[4907]: I0313 15:48:50.379976 4907 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 13 15:48:50 crc kubenswrapper[4907]: I0313 15:48:50.380015 4907 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-ac09aaac-45f0-4ecf-8f08-3e715b4f4606\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ac09aaac-45f0-4ecf-8f08-3e715b4f4606\") pod \"prometheus-metric-storage-0\" (UID: \"3e5a8905-7f71-4979-aee3-f81e772ea46a\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/18cb5e38569b223e596eb1f57078b15c9d63dca5ccb98a8253238ad08ba4dd89/globalmount\"" pod="openstack/prometheus-metric-storage-0" Mar 13 15:48:50 crc kubenswrapper[4907]: I0313 15:48:50.534531 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-ac09aaac-45f0-4ecf-8f08-3e715b4f4606\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ac09aaac-45f0-4ecf-8f08-3e715b4f4606\") pod \"prometheus-metric-storage-0\" (UID: \"3e5a8905-7f71-4979-aee3-f81e772ea46a\") " pod="openstack/prometheus-metric-storage-0" Mar 13 15:48:50 crc kubenswrapper[4907]: I0313 15:48:50.594596 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-7tp2h" podUID="e8c5a748-e7b7-43e0-82d3-c76673b17ea1" containerName="registry-server" probeResult="failure" output=< Mar 13 15:48:50 crc kubenswrapper[4907]: timeout: failed to connect service ":50051" within 1s Mar 13 15:48:50 crc kubenswrapper[4907]: > Mar 13 15:48:50 crc kubenswrapper[4907]: I0313 15:48:50.678046 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Mar 13 15:48:50 crc kubenswrapper[4907]: I0313 15:48:50.915276 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/alertmanager-metric-storage-0"] Mar 13 15:48:50 crc kubenswrapper[4907]: I0313 15:48:50.929776 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Mar 13 15:48:51 crc kubenswrapper[4907]: I0313 15:48:51.063068 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/74a5c619-2dde-4f75-9106-cad6a39265ce-openstack-config\") pod \"74a5c619-2dde-4f75-9106-cad6a39265ce\" (UID: \"74a5c619-2dde-4f75-9106-cad6a39265ce\") " Mar 13 15:48:51 crc kubenswrapper[4907]: I0313 15:48:51.063930 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/74a5c619-2dde-4f75-9106-cad6a39265ce-openstack-config-secret\") pod \"74a5c619-2dde-4f75-9106-cad6a39265ce\" (UID: \"74a5c619-2dde-4f75-9106-cad6a39265ce\") " Mar 13 15:48:51 crc kubenswrapper[4907]: I0313 15:48:51.064165 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-85lwz\" (UniqueName: \"kubernetes.io/projected/74a5c619-2dde-4f75-9106-cad6a39265ce-kube-api-access-85lwz\") pod \"74a5c619-2dde-4f75-9106-cad6a39265ce\" (UID: \"74a5c619-2dde-4f75-9106-cad6a39265ce\") " Mar 13 15:48:51 crc kubenswrapper[4907]: I0313 15:48:51.072321 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/74a5c619-2dde-4f75-9106-cad6a39265ce-kube-api-access-85lwz" (OuterVolumeSpecName: "kube-api-access-85lwz") pod "74a5c619-2dde-4f75-9106-cad6a39265ce" (UID: "74a5c619-2dde-4f75-9106-cad6a39265ce"). InnerVolumeSpecName "kube-api-access-85lwz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:48:51 crc kubenswrapper[4907]: I0313 15:48:51.133787 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/74a5c619-2dde-4f75-9106-cad6a39265ce-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "74a5c619-2dde-4f75-9106-cad6a39265ce" (UID: "74a5c619-2dde-4f75-9106-cad6a39265ce"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:48:51 crc kubenswrapper[4907]: I0313 15:48:51.146285 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74a5c619-2dde-4f75-9106-cad6a39265ce-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "74a5c619-2dde-4f75-9106-cad6a39265ce" (UID: "74a5c619-2dde-4f75-9106-cad6a39265ce"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:48:51 crc kubenswrapper[4907]: I0313 15:48:51.168377 4907 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/74a5c619-2dde-4f75-9106-cad6a39265ce-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Mar 13 15:48:51 crc kubenswrapper[4907]: I0313 15:48:51.168413 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-85lwz\" (UniqueName: \"kubernetes.io/projected/74a5c619-2dde-4f75-9106-cad6a39265ce-kube-api-access-85lwz\") on node \"crc\" DevicePath \"\"" Mar 13 15:48:51 crc kubenswrapper[4907]: I0313 15:48:51.168423 4907 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/74a5c619-2dde-4f75-9106-cad6a39265ce-openstack-config\") on node \"crc\" DevicePath \"\"" Mar 13 15:48:51 crc kubenswrapper[4907]: I0313 15:48:51.325383 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Mar 13 15:48:51 crc kubenswrapper[4907]: I0313 15:48:51.328781 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"d29f6de0-3229-477c-86a6-550f2d465592","Type":"ContainerStarted","Data":"3e839fa1cb364797ef12f2e906290509ded0419ffb1acfd515495d8304c579b8"} Mar 13 15:48:51 crc kubenswrapper[4907]: I0313 15:48:51.330814 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"2316d199-eef1-4ac1-9f63-cc73cba8d4c7","Type":"ContainerStarted","Data":"8caed9aacdfdeb98c939313e87104e0a2c29efecca834afa9daa99b5055ebc8e"} Mar 13 15:48:51 crc kubenswrapper[4907]: I0313 15:48:51.337195 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"53c58ef1-fd3c-4d26-9b00-5ce0bb0cc592","Type":"ContainerStarted","Data":"77ad5fed4d3fec5023dd6cfa95b8071d1e0afddd56cc6b479d88fc6dfcc12714"} Mar 13 15:48:51 crc kubenswrapper[4907]: I0313 15:48:51.337610 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Mar 13 15:48:51 crc kubenswrapper[4907]: I0313 15:48:51.339279 4907 generic.go:334] "Generic (PLEG): container finished" podID="74a5c619-2dde-4f75-9106-cad6a39265ce" containerID="e99b6a0dae893428515991f160f7db931ea2980ec0fe9b55c8e67e61a9c011ff" exitCode=137 Mar 13 15:48:51 crc kubenswrapper[4907]: I0313 15:48:51.339320 4907 scope.go:117] "RemoveContainer" containerID="e99b6a0dae893428515991f160f7db931ea2980ec0fe9b55c8e67e61a9c011ff" Mar 13 15:48:51 crc kubenswrapper[4907]: I0313 15:48:51.339436 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Mar 13 15:48:51 crc kubenswrapper[4907]: I0313 15:48:51.351988 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=3.351964249 podStartE2EDuration="3.351964249s" podCreationTimestamp="2026-03-13 15:48:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:48:51.349527503 +0000 UTC m=+6230.249315192" watchObservedRunningTime="2026-03-13 15:48:51.351964249 +0000 UTC m=+6230.251751958" Mar 13 15:48:51 crc kubenswrapper[4907]: I0313 15:48:51.383617 4907 scope.go:117] "RemoveContainer" containerID="e99b6a0dae893428515991f160f7db931ea2980ec0fe9b55c8e67e61a9c011ff" Mar 13 15:48:51 crc kubenswrapper[4907]: E0313 15:48:51.385076 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e99b6a0dae893428515991f160f7db931ea2980ec0fe9b55c8e67e61a9c011ff\": container with ID starting with e99b6a0dae893428515991f160f7db931ea2980ec0fe9b55c8e67e61a9c011ff not found: ID does not exist" containerID="e99b6a0dae893428515991f160f7db931ea2980ec0fe9b55c8e67e61a9c011ff" Mar 13 15:48:51 crc kubenswrapper[4907]: I0313 15:48:51.385137 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e99b6a0dae893428515991f160f7db931ea2980ec0fe9b55c8e67e61a9c011ff"} err="failed to get container status \"e99b6a0dae893428515991f160f7db931ea2980ec0fe9b55c8e67e61a9c011ff\": rpc error: code = NotFound desc = could not find container \"e99b6a0dae893428515991f160f7db931ea2980ec0fe9b55c8e67e61a9c011ff\": container with ID starting with e99b6a0dae893428515991f160f7db931ea2980ec0fe9b55c8e67e61a9c011ff not found: ID does not exist" Mar 13 15:48:51 crc kubenswrapper[4907]: I0313 15:48:51.386009 4907 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="74a5c619-2dde-4f75-9106-cad6a39265ce" podUID="2316d199-eef1-4ac1-9f63-cc73cba8d4c7" Mar 13 15:48:51 crc kubenswrapper[4907]: I0313 15:48:51.390022 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.95926718 podStartE2EDuration="3.390006349s" podCreationTimestamp="2026-03-13 15:48:48 +0000 UTC" firstStartedPulling="2026-03-13 15:48:49.955482355 +0000 UTC m=+6228.855270044" lastFinishedPulling="2026-03-13 15:48:50.386221524 +0000 UTC m=+6229.286009213" observedRunningTime="2026-03-13 15:48:51.382323039 +0000 UTC m=+6230.282110728" watchObservedRunningTime="2026-03-13 15:48:51.390006349 +0000 UTC m=+6230.289794038" Mar 13 15:48:51 crc kubenswrapper[4907]: I0313 15:48:51.794211 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0e98ef6f-8765-4e4d-ab91-7dcdcb3f788d" path="/var/lib/kubelet/pods/0e98ef6f-8765-4e4d-ab91-7dcdcb3f788d/volumes" Mar 13 15:48:51 crc kubenswrapper[4907]: I0313 15:48:51.795464 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="74a5c619-2dde-4f75-9106-cad6a39265ce" path="/var/lib/kubelet/pods/74a5c619-2dde-4f75-9106-cad6a39265ce/volumes" Mar 13 15:48:51 crc kubenswrapper[4907]: I0313 15:48:51.796156 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="978c19e3-b11d-44ef-86d5-6b0674e488d1" path="/var/lib/kubelet/pods/978c19e3-b11d-44ef-86d5-6b0674e488d1/volumes" Mar 13 15:48:52 crc kubenswrapper[4907]: I0313 15:48:52.353450 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"3e5a8905-7f71-4979-aee3-f81e772ea46a","Type":"ContainerStarted","Data":"f98c017c2a67861fb920f9dadf2d97df3ff6075477cb06f3b46a33ec5ad21c7b"} Mar 13 15:48:57 crc kubenswrapper[4907]: I0313 15:48:57.033519 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-z7wl4"] Mar 13 15:48:57 crc kubenswrapper[4907]: I0313 15:48:57.044637 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-z7wl4"] Mar 13 15:48:57 crc kubenswrapper[4907]: I0313 15:48:57.413264 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"d29f6de0-3229-477c-86a6-550f2d465592","Type":"ContainerStarted","Data":"998196e6dbaae3637cbc8ca152cc5014c622bb70e9857a79cc0522fca4602736"} Mar 13 15:48:57 crc kubenswrapper[4907]: I0313 15:48:57.418986 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"3e5a8905-7f71-4979-aee3-f81e772ea46a","Type":"ContainerStarted","Data":"a3daab1d7da324d8c75c9e53d1520c30cb91a375d6d03e39cbc921749c0bb23d"} Mar 13 15:48:57 crc kubenswrapper[4907]: I0313 15:48:57.801397 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a43a8ffb-4f15-4524-b750-517442c1f561" path="/var/lib/kubelet/pods/a43a8ffb-4f15-4524-b750-517442c1f561/volumes" Mar 13 15:48:59 crc kubenswrapper[4907]: I0313 15:48:59.011519 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Mar 13 15:49:00 crc kubenswrapper[4907]: I0313 15:49:00.557550 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-7tp2h" podUID="e8c5a748-e7b7-43e0-82d3-c76673b17ea1" containerName="registry-server" probeResult="failure" output=< Mar 13 15:49:00 crc kubenswrapper[4907]: timeout: failed to connect service ":50051" within 1s Mar 13 15:49:00 crc kubenswrapper[4907]: > Mar 13 15:49:03 crc kubenswrapper[4907]: I0313 15:49:03.473864 4907 generic.go:334] "Generic (PLEG): container finished" podID="3e5a8905-7f71-4979-aee3-f81e772ea46a" containerID="a3daab1d7da324d8c75c9e53d1520c30cb91a375d6d03e39cbc921749c0bb23d" exitCode=0 Mar 13 15:49:03 crc kubenswrapper[4907]: I0313 15:49:03.474021 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"3e5a8905-7f71-4979-aee3-f81e772ea46a","Type":"ContainerDied","Data":"a3daab1d7da324d8c75c9e53d1520c30cb91a375d6d03e39cbc921749c0bb23d"} Mar 13 15:49:03 crc kubenswrapper[4907]: I0313 15:49:03.476076 4907 generic.go:334] "Generic (PLEG): container finished" podID="d29f6de0-3229-477c-86a6-550f2d465592" containerID="998196e6dbaae3637cbc8ca152cc5014c622bb70e9857a79cc0522fca4602736" exitCode=0 Mar 13 15:49:03 crc kubenswrapper[4907]: I0313 15:49:03.476115 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"d29f6de0-3229-477c-86a6-550f2d465592","Type":"ContainerDied","Data":"998196e6dbaae3637cbc8ca152cc5014c622bb70e9857a79cc0522fca4602736"} Mar 13 15:49:06 crc kubenswrapper[4907]: I0313 15:49:06.515044 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"d29f6de0-3229-477c-86a6-550f2d465592","Type":"ContainerStarted","Data":"ed52a0746ec744f915cfe385f5b3d69db47e9fac5239ec3264f15b6665ddca8f"} Mar 13 15:49:09 crc kubenswrapper[4907]: I0313 15:49:09.558050 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-7tp2h" Mar 13 15:49:09 crc kubenswrapper[4907]: I0313 15:49:09.566911 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"d29f6de0-3229-477c-86a6-550f2d465592","Type":"ContainerStarted","Data":"494105b0e22ae748f4d76099d291149db9c96acc20858558d71b2c1c8372bb0c"} Mar 13 15:49:09 crc kubenswrapper[4907]: I0313 15:49:09.567295 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/alertmanager-metric-storage-0" Mar 13 15:49:09 crc kubenswrapper[4907]: I0313 15:49:09.570986 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/alertmanager-metric-storage-0" Mar 13 15:49:09 crc kubenswrapper[4907]: I0313 15:49:09.611648 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/alertmanager-metric-storage-0" podStartSLOduration=5.368136396 podStartE2EDuration="20.611629335s" podCreationTimestamp="2026-03-13 15:48:49 +0000 UTC" firstStartedPulling="2026-03-13 15:48:50.977982331 +0000 UTC m=+6229.877770020" lastFinishedPulling="2026-03-13 15:49:06.22147526 +0000 UTC m=+6245.121262959" observedRunningTime="2026-03-13 15:49:09.595809172 +0000 UTC m=+6248.495596861" watchObservedRunningTime="2026-03-13 15:49:09.611629335 +0000 UTC m=+6248.511417024" Mar 13 15:49:09 crc kubenswrapper[4907]: I0313 15:49:09.640482 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-7tp2h" Mar 13 15:49:09 crc kubenswrapper[4907]: I0313 15:49:09.813317 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7tp2h"] Mar 13 15:49:11 crc kubenswrapper[4907]: I0313 15:49:11.586732 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"3e5a8905-7f71-4979-aee3-f81e772ea46a","Type":"ContainerStarted","Data":"38f43333273c3672449a529b41205de80c2ce59a4d18582d8c5417bdd810b24a"} Mar 13 15:49:11 crc kubenswrapper[4907]: I0313 15:49:11.586919 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-7tp2h" podUID="e8c5a748-e7b7-43e0-82d3-c76673b17ea1" containerName="registry-server" containerID="cri-o://49ef110711e5f738221014e1ce44cead4c2950560589e797e31382bd54cac333" gracePeriod=2 Mar 13 15:49:12 crc kubenswrapper[4907]: I0313 15:49:12.061068 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7tp2h" Mar 13 15:49:12 crc kubenswrapper[4907]: I0313 15:49:12.146805 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ppphq\" (UniqueName: \"kubernetes.io/projected/e8c5a748-e7b7-43e0-82d3-c76673b17ea1-kube-api-access-ppphq\") pod \"e8c5a748-e7b7-43e0-82d3-c76673b17ea1\" (UID: \"e8c5a748-e7b7-43e0-82d3-c76673b17ea1\") " Mar 13 15:49:12 crc kubenswrapper[4907]: I0313 15:49:12.147106 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e8c5a748-e7b7-43e0-82d3-c76673b17ea1-utilities\") pod \"e8c5a748-e7b7-43e0-82d3-c76673b17ea1\" (UID: \"e8c5a748-e7b7-43e0-82d3-c76673b17ea1\") " Mar 13 15:49:12 crc kubenswrapper[4907]: I0313 15:49:12.147435 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e8c5a748-e7b7-43e0-82d3-c76673b17ea1-catalog-content\") pod \"e8c5a748-e7b7-43e0-82d3-c76673b17ea1\" (UID: \"e8c5a748-e7b7-43e0-82d3-c76673b17ea1\") " Mar 13 15:49:12 crc kubenswrapper[4907]: I0313 15:49:12.147763 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e8c5a748-e7b7-43e0-82d3-c76673b17ea1-utilities" (OuterVolumeSpecName: "utilities") pod "e8c5a748-e7b7-43e0-82d3-c76673b17ea1" (UID: "e8c5a748-e7b7-43e0-82d3-c76673b17ea1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:49:12 crc kubenswrapper[4907]: I0313 15:49:12.148230 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e8c5a748-e7b7-43e0-82d3-c76673b17ea1-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 15:49:12 crc kubenswrapper[4907]: I0313 15:49:12.152773 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e8c5a748-e7b7-43e0-82d3-c76673b17ea1-kube-api-access-ppphq" (OuterVolumeSpecName: "kube-api-access-ppphq") pod "e8c5a748-e7b7-43e0-82d3-c76673b17ea1" (UID: "e8c5a748-e7b7-43e0-82d3-c76673b17ea1"). InnerVolumeSpecName "kube-api-access-ppphq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:49:12 crc kubenswrapper[4907]: I0313 15:49:12.250783 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ppphq\" (UniqueName: \"kubernetes.io/projected/e8c5a748-e7b7-43e0-82d3-c76673b17ea1-kube-api-access-ppphq\") on node \"crc\" DevicePath \"\"" Mar 13 15:49:12 crc kubenswrapper[4907]: I0313 15:49:12.277750 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e8c5a748-e7b7-43e0-82d3-c76673b17ea1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e8c5a748-e7b7-43e0-82d3-c76673b17ea1" (UID: "e8c5a748-e7b7-43e0-82d3-c76673b17ea1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:49:12 crc kubenswrapper[4907]: I0313 15:49:12.352531 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e8c5a748-e7b7-43e0-82d3-c76673b17ea1-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 15:49:12 crc kubenswrapper[4907]: I0313 15:49:12.606985 4907 generic.go:334] "Generic (PLEG): container finished" podID="e8c5a748-e7b7-43e0-82d3-c76673b17ea1" containerID="49ef110711e5f738221014e1ce44cead4c2950560589e797e31382bd54cac333" exitCode=0 Mar 13 15:49:12 crc kubenswrapper[4907]: I0313 15:49:12.607047 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7tp2h" event={"ID":"e8c5a748-e7b7-43e0-82d3-c76673b17ea1","Type":"ContainerDied","Data":"49ef110711e5f738221014e1ce44cead4c2950560589e797e31382bd54cac333"} Mar 13 15:49:12 crc kubenswrapper[4907]: I0313 15:49:12.607080 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7tp2h" event={"ID":"e8c5a748-e7b7-43e0-82d3-c76673b17ea1","Type":"ContainerDied","Data":"a029d38e007ef1dd2747ad9ff6b5a9cef6fcfa563fa0dc97f6df3132418979f8"} Mar 13 15:49:12 crc kubenswrapper[4907]: I0313 15:49:12.607100 4907 scope.go:117] "RemoveContainer" containerID="49ef110711e5f738221014e1ce44cead4c2950560589e797e31382bd54cac333" Mar 13 15:49:12 crc kubenswrapper[4907]: I0313 15:49:12.607137 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7tp2h" Mar 13 15:49:12 crc kubenswrapper[4907]: I0313 15:49:12.632055 4907 scope.go:117] "RemoveContainer" containerID="2d7e94bcae9c69807004704893b2f10f4c520e505aa6a59bbe61e348ddae96df" Mar 13 15:49:12 crc kubenswrapper[4907]: I0313 15:49:12.646492 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7tp2h"] Mar 13 15:49:12 crc kubenswrapper[4907]: I0313 15:49:12.658968 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-7tp2h"] Mar 13 15:49:12 crc kubenswrapper[4907]: I0313 15:49:12.687673 4907 scope.go:117] "RemoveContainer" containerID="d49b21ae936fc12c5a09e43efc846d4e58d2f3ad32465957c3a2c3b2abeb9cfe" Mar 13 15:49:12 crc kubenswrapper[4907]: I0313 15:49:12.709559 4907 scope.go:117] "RemoveContainer" containerID="49ef110711e5f738221014e1ce44cead4c2950560589e797e31382bd54cac333" Mar 13 15:49:12 crc kubenswrapper[4907]: E0313 15:49:12.710174 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"49ef110711e5f738221014e1ce44cead4c2950560589e797e31382bd54cac333\": container with ID starting with 49ef110711e5f738221014e1ce44cead4c2950560589e797e31382bd54cac333 not found: ID does not exist" containerID="49ef110711e5f738221014e1ce44cead4c2950560589e797e31382bd54cac333" Mar 13 15:49:12 crc kubenswrapper[4907]: I0313 15:49:12.710283 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"49ef110711e5f738221014e1ce44cead4c2950560589e797e31382bd54cac333"} err="failed to get container status \"49ef110711e5f738221014e1ce44cead4c2950560589e797e31382bd54cac333\": rpc error: code = NotFound desc = could not find container \"49ef110711e5f738221014e1ce44cead4c2950560589e797e31382bd54cac333\": container with ID starting with 49ef110711e5f738221014e1ce44cead4c2950560589e797e31382bd54cac333 not found: ID does not exist" Mar 13 15:49:12 crc kubenswrapper[4907]: I0313 15:49:12.710358 4907 scope.go:117] "RemoveContainer" containerID="2d7e94bcae9c69807004704893b2f10f4c520e505aa6a59bbe61e348ddae96df" Mar 13 15:49:12 crc kubenswrapper[4907]: E0313 15:49:12.710853 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2d7e94bcae9c69807004704893b2f10f4c520e505aa6a59bbe61e348ddae96df\": container with ID starting with 2d7e94bcae9c69807004704893b2f10f4c520e505aa6a59bbe61e348ddae96df not found: ID does not exist" containerID="2d7e94bcae9c69807004704893b2f10f4c520e505aa6a59bbe61e348ddae96df" Mar 13 15:49:12 crc kubenswrapper[4907]: I0313 15:49:12.710981 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2d7e94bcae9c69807004704893b2f10f4c520e505aa6a59bbe61e348ddae96df"} err="failed to get container status \"2d7e94bcae9c69807004704893b2f10f4c520e505aa6a59bbe61e348ddae96df\": rpc error: code = NotFound desc = could not find container \"2d7e94bcae9c69807004704893b2f10f4c520e505aa6a59bbe61e348ddae96df\": container with ID starting with 2d7e94bcae9c69807004704893b2f10f4c520e505aa6a59bbe61e348ddae96df not found: ID does not exist" Mar 13 15:49:12 crc kubenswrapper[4907]: I0313 15:49:12.711011 4907 scope.go:117] "RemoveContainer" containerID="d49b21ae936fc12c5a09e43efc846d4e58d2f3ad32465957c3a2c3b2abeb9cfe" Mar 13 15:49:12 crc kubenswrapper[4907]: E0313 15:49:12.711498 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d49b21ae936fc12c5a09e43efc846d4e58d2f3ad32465957c3a2c3b2abeb9cfe\": container with ID starting with d49b21ae936fc12c5a09e43efc846d4e58d2f3ad32465957c3a2c3b2abeb9cfe not found: ID does not exist" containerID="d49b21ae936fc12c5a09e43efc846d4e58d2f3ad32465957c3a2c3b2abeb9cfe" Mar 13 15:49:12 crc kubenswrapper[4907]: I0313 15:49:12.711569 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d49b21ae936fc12c5a09e43efc846d4e58d2f3ad32465957c3a2c3b2abeb9cfe"} err="failed to get container status \"d49b21ae936fc12c5a09e43efc846d4e58d2f3ad32465957c3a2c3b2abeb9cfe\": rpc error: code = NotFound desc = could not find container \"d49b21ae936fc12c5a09e43efc846d4e58d2f3ad32465957c3a2c3b2abeb9cfe\": container with ID starting with d49b21ae936fc12c5a09e43efc846d4e58d2f3ad32465957c3a2c3b2abeb9cfe not found: ID does not exist" Mar 13 15:49:13 crc kubenswrapper[4907]: I0313 15:49:13.797679 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e8c5a748-e7b7-43e0-82d3-c76673b17ea1" path="/var/lib/kubelet/pods/e8c5a748-e7b7-43e0-82d3-c76673b17ea1/volumes" Mar 13 15:49:15 crc kubenswrapper[4907]: I0313 15:49:15.645168 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"3e5a8905-7f71-4979-aee3-f81e772ea46a","Type":"ContainerStarted","Data":"a7f3d5e1b9c30bf5160801062d9ffb06f51eec5b64919c41a4a982e147424b23"} Mar 13 15:49:18 crc kubenswrapper[4907]: I0313 15:49:18.041796 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 15:49:18 crc kubenswrapper[4907]: I0313 15:49:18.042157 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 15:49:18 crc kubenswrapper[4907]: I0313 15:49:18.675383 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"3e5a8905-7f71-4979-aee3-f81e772ea46a","Type":"ContainerStarted","Data":"22eedd14ce51e6e1c8b08afa6ad93b877c9db3053519763c8f6b052b58e2dddc"} Mar 13 15:49:18 crc kubenswrapper[4907]: I0313 15:49:18.707169 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=4.163441901 podStartE2EDuration="30.70714764s" podCreationTimestamp="2026-03-13 15:48:48 +0000 UTC" firstStartedPulling="2026-03-13 15:48:51.324819928 +0000 UTC m=+6230.224607617" lastFinishedPulling="2026-03-13 15:49:17.868525667 +0000 UTC m=+6256.768313356" observedRunningTime="2026-03-13 15:49:18.7009332 +0000 UTC m=+6257.600720899" watchObservedRunningTime="2026-03-13 15:49:18.70714764 +0000 UTC m=+6257.606935329" Mar 13 15:49:20 crc kubenswrapper[4907]: I0313 15:49:20.678490 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Mar 13 15:49:20 crc kubenswrapper[4907]: I0313 15:49:20.679091 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Mar 13 15:49:20 crc kubenswrapper[4907]: I0313 15:49:20.680831 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Mar 13 15:49:20 crc kubenswrapper[4907]: I0313 15:49:20.701466 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Mar 13 15:49:22 crc kubenswrapper[4907]: I0313 15:49:22.194800 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 13 15:49:22 crc kubenswrapper[4907]: E0313 15:49:22.195626 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8c5a748-e7b7-43e0-82d3-c76673b17ea1" containerName="registry-server" Mar 13 15:49:22 crc kubenswrapper[4907]: I0313 15:49:22.195646 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8c5a748-e7b7-43e0-82d3-c76673b17ea1" containerName="registry-server" Mar 13 15:49:22 crc kubenswrapper[4907]: E0313 15:49:22.195671 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8c5a748-e7b7-43e0-82d3-c76673b17ea1" containerName="extract-utilities" Mar 13 15:49:22 crc kubenswrapper[4907]: I0313 15:49:22.195679 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8c5a748-e7b7-43e0-82d3-c76673b17ea1" containerName="extract-utilities" Mar 13 15:49:22 crc kubenswrapper[4907]: E0313 15:49:22.195710 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8c5a748-e7b7-43e0-82d3-c76673b17ea1" containerName="extract-content" Mar 13 15:49:22 crc kubenswrapper[4907]: I0313 15:49:22.195719 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8c5a748-e7b7-43e0-82d3-c76673b17ea1" containerName="extract-content" Mar 13 15:49:22 crc kubenswrapper[4907]: I0313 15:49:22.195997 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="e8c5a748-e7b7-43e0-82d3-c76673b17ea1" containerName="registry-server" Mar 13 15:49:22 crc kubenswrapper[4907]: I0313 15:49:22.198463 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 15:49:22 crc kubenswrapper[4907]: I0313 15:49:22.204686 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 13 15:49:22 crc kubenswrapper[4907]: I0313 15:49:22.205051 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 13 15:49:22 crc kubenswrapper[4907]: I0313 15:49:22.208126 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 13 15:49:22 crc kubenswrapper[4907]: I0313 15:49:22.260458 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eaec5907-c808-4ac2-b131-608b228885e6-log-httpd\") pod \"ceilometer-0\" (UID: \"eaec5907-c808-4ac2-b131-608b228885e6\") " pod="openstack/ceilometer-0" Mar 13 15:49:22 crc kubenswrapper[4907]: I0313 15:49:22.260536 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x6gpx\" (UniqueName: \"kubernetes.io/projected/eaec5907-c808-4ac2-b131-608b228885e6-kube-api-access-x6gpx\") pod \"ceilometer-0\" (UID: \"eaec5907-c808-4ac2-b131-608b228885e6\") " pod="openstack/ceilometer-0" Mar 13 15:49:22 crc kubenswrapper[4907]: I0313 15:49:22.260627 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eaec5907-c808-4ac2-b131-608b228885e6-scripts\") pod \"ceilometer-0\" (UID: \"eaec5907-c808-4ac2-b131-608b228885e6\") " pod="openstack/ceilometer-0" Mar 13 15:49:22 crc kubenswrapper[4907]: I0313 15:49:22.260840 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eaec5907-c808-4ac2-b131-608b228885e6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"eaec5907-c808-4ac2-b131-608b228885e6\") " pod="openstack/ceilometer-0" Mar 13 15:49:22 crc kubenswrapper[4907]: I0313 15:49:22.261167 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eaec5907-c808-4ac2-b131-608b228885e6-run-httpd\") pod \"ceilometer-0\" (UID: \"eaec5907-c808-4ac2-b131-608b228885e6\") " pod="openstack/ceilometer-0" Mar 13 15:49:22 crc kubenswrapper[4907]: I0313 15:49:22.261221 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/eaec5907-c808-4ac2-b131-608b228885e6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"eaec5907-c808-4ac2-b131-608b228885e6\") " pod="openstack/ceilometer-0" Mar 13 15:49:22 crc kubenswrapper[4907]: I0313 15:49:22.261307 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eaec5907-c808-4ac2-b131-608b228885e6-config-data\") pod \"ceilometer-0\" (UID: \"eaec5907-c808-4ac2-b131-608b228885e6\") " pod="openstack/ceilometer-0" Mar 13 15:49:22 crc kubenswrapper[4907]: I0313 15:49:22.362812 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eaec5907-c808-4ac2-b131-608b228885e6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"eaec5907-c808-4ac2-b131-608b228885e6\") " pod="openstack/ceilometer-0" Mar 13 15:49:22 crc kubenswrapper[4907]: I0313 15:49:22.362933 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eaec5907-c808-4ac2-b131-608b228885e6-run-httpd\") pod \"ceilometer-0\" (UID: \"eaec5907-c808-4ac2-b131-608b228885e6\") " pod="openstack/ceilometer-0" Mar 13 15:49:22 crc kubenswrapper[4907]: I0313 15:49:22.362958 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/eaec5907-c808-4ac2-b131-608b228885e6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"eaec5907-c808-4ac2-b131-608b228885e6\") " pod="openstack/ceilometer-0" Mar 13 15:49:22 crc kubenswrapper[4907]: I0313 15:49:22.362994 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eaec5907-c808-4ac2-b131-608b228885e6-config-data\") pod \"ceilometer-0\" (UID: \"eaec5907-c808-4ac2-b131-608b228885e6\") " pod="openstack/ceilometer-0" Mar 13 15:49:22 crc kubenswrapper[4907]: I0313 15:49:22.363021 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eaec5907-c808-4ac2-b131-608b228885e6-log-httpd\") pod \"ceilometer-0\" (UID: \"eaec5907-c808-4ac2-b131-608b228885e6\") " pod="openstack/ceilometer-0" Mar 13 15:49:22 crc kubenswrapper[4907]: I0313 15:49:22.363044 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x6gpx\" (UniqueName: \"kubernetes.io/projected/eaec5907-c808-4ac2-b131-608b228885e6-kube-api-access-x6gpx\") pod \"ceilometer-0\" (UID: \"eaec5907-c808-4ac2-b131-608b228885e6\") " pod="openstack/ceilometer-0" Mar 13 15:49:22 crc kubenswrapper[4907]: I0313 15:49:22.363096 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eaec5907-c808-4ac2-b131-608b228885e6-scripts\") pod \"ceilometer-0\" (UID: \"eaec5907-c808-4ac2-b131-608b228885e6\") " pod="openstack/ceilometer-0" Mar 13 15:49:22 crc kubenswrapper[4907]: I0313 15:49:22.363467 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eaec5907-c808-4ac2-b131-608b228885e6-run-httpd\") pod \"ceilometer-0\" (UID: \"eaec5907-c808-4ac2-b131-608b228885e6\") " pod="openstack/ceilometer-0" Mar 13 15:49:22 crc kubenswrapper[4907]: I0313 15:49:22.363571 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eaec5907-c808-4ac2-b131-608b228885e6-log-httpd\") pod \"ceilometer-0\" (UID: \"eaec5907-c808-4ac2-b131-608b228885e6\") " pod="openstack/ceilometer-0" Mar 13 15:49:22 crc kubenswrapper[4907]: I0313 15:49:22.368913 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eaec5907-c808-4ac2-b131-608b228885e6-config-data\") pod \"ceilometer-0\" (UID: \"eaec5907-c808-4ac2-b131-608b228885e6\") " pod="openstack/ceilometer-0" Mar 13 15:49:22 crc kubenswrapper[4907]: I0313 15:49:22.370594 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/eaec5907-c808-4ac2-b131-608b228885e6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"eaec5907-c808-4ac2-b131-608b228885e6\") " pod="openstack/ceilometer-0" Mar 13 15:49:22 crc kubenswrapper[4907]: I0313 15:49:22.377584 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eaec5907-c808-4ac2-b131-608b228885e6-scripts\") pod \"ceilometer-0\" (UID: \"eaec5907-c808-4ac2-b131-608b228885e6\") " pod="openstack/ceilometer-0" Mar 13 15:49:22 crc kubenswrapper[4907]: I0313 15:49:22.377639 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eaec5907-c808-4ac2-b131-608b228885e6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"eaec5907-c808-4ac2-b131-608b228885e6\") " pod="openstack/ceilometer-0" Mar 13 15:49:22 crc kubenswrapper[4907]: I0313 15:49:22.383635 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x6gpx\" (UniqueName: \"kubernetes.io/projected/eaec5907-c808-4ac2-b131-608b228885e6-kube-api-access-x6gpx\") pod \"ceilometer-0\" (UID: \"eaec5907-c808-4ac2-b131-608b228885e6\") " pod="openstack/ceilometer-0" Mar 13 15:49:22 crc kubenswrapper[4907]: I0313 15:49:22.536622 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 15:49:23 crc kubenswrapper[4907]: I0313 15:49:23.020620 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 13 15:49:23 crc kubenswrapper[4907]: I0313 15:49:23.741084 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eaec5907-c808-4ac2-b131-608b228885e6","Type":"ContainerStarted","Data":"ae9b3358826e66742e51bbbc204d8b30a566e7e0b936fca1df7ae895ee7043e9"} Mar 13 15:49:24 crc kubenswrapper[4907]: I0313 15:49:24.759862 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eaec5907-c808-4ac2-b131-608b228885e6","Type":"ContainerStarted","Data":"f8bccaf5228a1db9069e313bf3466a6f8cbd61fae45f55bda0a236da15bba7f0"} Mar 13 15:49:24 crc kubenswrapper[4907]: I0313 15:49:24.760517 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eaec5907-c808-4ac2-b131-608b228885e6","Type":"ContainerStarted","Data":"b3ab302efa65117dbfc68c5b768e1ee7a624666e299637f13e417c20da84fcf0"} Mar 13 15:49:25 crc kubenswrapper[4907]: I0313 15:49:25.771419 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eaec5907-c808-4ac2-b131-608b228885e6","Type":"ContainerStarted","Data":"3b5b22ec27be798569c983dbf41141eab318a4ae784c5efd3348ea56c4f000f7"} Mar 13 15:49:27 crc kubenswrapper[4907]: I0313 15:49:27.800040 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eaec5907-c808-4ac2-b131-608b228885e6","Type":"ContainerStarted","Data":"5a450b86aabb5b84d29c1e3252b0f33c4167256fbc37a7560f1e23aae3ba5797"} Mar 13 15:49:27 crc kubenswrapper[4907]: I0313 15:49:27.800606 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Mar 13 15:49:27 crc kubenswrapper[4907]: I0313 15:49:27.829619 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.415617243 podStartE2EDuration="5.82959926s" podCreationTimestamp="2026-03-13 15:49:22 +0000 UTC" firstStartedPulling="2026-03-13 15:49:23.027813188 +0000 UTC m=+6261.927600877" lastFinishedPulling="2026-03-13 15:49:27.441795205 +0000 UTC m=+6266.341582894" observedRunningTime="2026-03-13 15:49:27.82117969 +0000 UTC m=+6266.720967379" watchObservedRunningTime="2026-03-13 15:49:27.82959926 +0000 UTC m=+6266.729386949" Mar 13 15:49:33 crc kubenswrapper[4907]: I0313 15:49:33.386563 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-db-create-6qbgl"] Mar 13 15:49:33 crc kubenswrapper[4907]: I0313 15:49:33.388448 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-6qbgl" Mar 13 15:49:33 crc kubenswrapper[4907]: I0313 15:49:33.397612 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-create-6qbgl"] Mar 13 15:49:33 crc kubenswrapper[4907]: I0313 15:49:33.417030 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r7s7z\" (UniqueName: \"kubernetes.io/projected/689fb651-0113-48ba-b6e4-29c41c5fa21f-kube-api-access-r7s7z\") pod \"aodh-db-create-6qbgl\" (UID: \"689fb651-0113-48ba-b6e4-29c41c5fa21f\") " pod="openstack/aodh-db-create-6qbgl" Mar 13 15:49:33 crc kubenswrapper[4907]: I0313 15:49:33.417475 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/689fb651-0113-48ba-b6e4-29c41c5fa21f-operator-scripts\") pod \"aodh-db-create-6qbgl\" (UID: \"689fb651-0113-48ba-b6e4-29c41c5fa21f\") " pod="openstack/aodh-db-create-6qbgl" Mar 13 15:49:33 crc kubenswrapper[4907]: I0313 15:49:33.519729 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/689fb651-0113-48ba-b6e4-29c41c5fa21f-operator-scripts\") pod \"aodh-db-create-6qbgl\" (UID: \"689fb651-0113-48ba-b6e4-29c41c5fa21f\") " pod="openstack/aodh-db-create-6qbgl" Mar 13 15:49:33 crc kubenswrapper[4907]: I0313 15:49:33.519901 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r7s7z\" (UniqueName: \"kubernetes.io/projected/689fb651-0113-48ba-b6e4-29c41c5fa21f-kube-api-access-r7s7z\") pod \"aodh-db-create-6qbgl\" (UID: \"689fb651-0113-48ba-b6e4-29c41c5fa21f\") " pod="openstack/aodh-db-create-6qbgl" Mar 13 15:49:33 crc kubenswrapper[4907]: I0313 15:49:33.520741 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/689fb651-0113-48ba-b6e4-29c41c5fa21f-operator-scripts\") pod \"aodh-db-create-6qbgl\" (UID: \"689fb651-0113-48ba-b6e4-29c41c5fa21f\") " pod="openstack/aodh-db-create-6qbgl" Mar 13 15:49:33 crc kubenswrapper[4907]: I0313 15:49:33.545523 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r7s7z\" (UniqueName: \"kubernetes.io/projected/689fb651-0113-48ba-b6e4-29c41c5fa21f-kube-api-access-r7s7z\") pod \"aodh-db-create-6qbgl\" (UID: \"689fb651-0113-48ba-b6e4-29c41c5fa21f\") " pod="openstack/aodh-db-create-6qbgl" Mar 13 15:49:33 crc kubenswrapper[4907]: I0313 15:49:33.595684 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-cc97-account-create-update-fc5r7"] Mar 13 15:49:33 crc kubenswrapper[4907]: I0313 15:49:33.597249 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-cc97-account-create-update-fc5r7" Mar 13 15:49:33 crc kubenswrapper[4907]: I0313 15:49:33.602375 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-db-secret" Mar 13 15:49:33 crc kubenswrapper[4907]: I0313 15:49:33.611649 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-cc97-account-create-update-fc5r7"] Mar 13 15:49:33 crc kubenswrapper[4907]: I0313 15:49:33.711760 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-6qbgl" Mar 13 15:49:33 crc kubenswrapper[4907]: I0313 15:49:33.723912 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/442daa10-2973-4304-9e1c-039d35226686-operator-scripts\") pod \"aodh-cc97-account-create-update-fc5r7\" (UID: \"442daa10-2973-4304-9e1c-039d35226686\") " pod="openstack/aodh-cc97-account-create-update-fc5r7" Mar 13 15:49:33 crc kubenswrapper[4907]: I0313 15:49:33.724007 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-74mmw\" (UniqueName: \"kubernetes.io/projected/442daa10-2973-4304-9e1c-039d35226686-kube-api-access-74mmw\") pod \"aodh-cc97-account-create-update-fc5r7\" (UID: \"442daa10-2973-4304-9e1c-039d35226686\") " pod="openstack/aodh-cc97-account-create-update-fc5r7" Mar 13 15:49:33 crc kubenswrapper[4907]: I0313 15:49:33.832830 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/442daa10-2973-4304-9e1c-039d35226686-operator-scripts\") pod \"aodh-cc97-account-create-update-fc5r7\" (UID: \"442daa10-2973-4304-9e1c-039d35226686\") " pod="openstack/aodh-cc97-account-create-update-fc5r7" Mar 13 15:49:33 crc kubenswrapper[4907]: I0313 15:49:33.832968 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-74mmw\" (UniqueName: \"kubernetes.io/projected/442daa10-2973-4304-9e1c-039d35226686-kube-api-access-74mmw\") pod \"aodh-cc97-account-create-update-fc5r7\" (UID: \"442daa10-2973-4304-9e1c-039d35226686\") " pod="openstack/aodh-cc97-account-create-update-fc5r7" Mar 13 15:49:33 crc kubenswrapper[4907]: I0313 15:49:33.840830 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/442daa10-2973-4304-9e1c-039d35226686-operator-scripts\") pod \"aodh-cc97-account-create-update-fc5r7\" (UID: \"442daa10-2973-4304-9e1c-039d35226686\") " pod="openstack/aodh-cc97-account-create-update-fc5r7" Mar 13 15:49:33 crc kubenswrapper[4907]: I0313 15:49:33.880169 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-74mmw\" (UniqueName: \"kubernetes.io/projected/442daa10-2973-4304-9e1c-039d35226686-kube-api-access-74mmw\") pod \"aodh-cc97-account-create-update-fc5r7\" (UID: \"442daa10-2973-4304-9e1c-039d35226686\") " pod="openstack/aodh-cc97-account-create-update-fc5r7" Mar 13 15:49:33 crc kubenswrapper[4907]: I0313 15:49:33.920481 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-cc97-account-create-update-fc5r7" Mar 13 15:49:34 crc kubenswrapper[4907]: I0313 15:49:34.303209 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-create-6qbgl"] Mar 13 15:49:34 crc kubenswrapper[4907]: W0313 15:49:34.663319 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod442daa10_2973_4304_9e1c_039d35226686.slice/crio-42a5129d078b177eff9fafc5af8936456bf22055b57fe17bc2a77ae226405e6d WatchSource:0}: Error finding container 42a5129d078b177eff9fafc5af8936456bf22055b57fe17bc2a77ae226405e6d: Status 404 returned error can't find the container with id 42a5129d078b177eff9fafc5af8936456bf22055b57fe17bc2a77ae226405e6d Mar 13 15:49:34 crc kubenswrapper[4907]: I0313 15:49:34.677778 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-cc97-account-create-update-fc5r7"] Mar 13 15:49:34 crc kubenswrapper[4907]: I0313 15:49:34.880667 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-cc97-account-create-update-fc5r7" event={"ID":"442daa10-2973-4304-9e1c-039d35226686","Type":"ContainerStarted","Data":"42a5129d078b177eff9fafc5af8936456bf22055b57fe17bc2a77ae226405e6d"} Mar 13 15:49:34 crc kubenswrapper[4907]: I0313 15:49:34.882186 4907 generic.go:334] "Generic (PLEG): container finished" podID="689fb651-0113-48ba-b6e4-29c41c5fa21f" containerID="7eeac1862f5788609b87b66dfacb73f343af69811adf4dfc0c22454194586bf3" exitCode=0 Mar 13 15:49:34 crc kubenswrapper[4907]: I0313 15:49:34.882256 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-6qbgl" event={"ID":"689fb651-0113-48ba-b6e4-29c41c5fa21f","Type":"ContainerDied","Data":"7eeac1862f5788609b87b66dfacb73f343af69811adf4dfc0c22454194586bf3"} Mar 13 15:49:34 crc kubenswrapper[4907]: I0313 15:49:34.882352 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-6qbgl" event={"ID":"689fb651-0113-48ba-b6e4-29c41c5fa21f","Type":"ContainerStarted","Data":"36da2d4a680e719ad1fe919746dbf7987f40a93fe7fec96a6662addf2879d7df"} Mar 13 15:49:35 crc kubenswrapper[4907]: I0313 15:49:35.892135 4907 generic.go:334] "Generic (PLEG): container finished" podID="442daa10-2973-4304-9e1c-039d35226686" containerID="d38dedf8f9062f909f22038364138fce749f1c07518101c6bfdc3e1d5f365369" exitCode=0 Mar 13 15:49:35 crc kubenswrapper[4907]: I0313 15:49:35.892182 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-cc97-account-create-update-fc5r7" event={"ID":"442daa10-2973-4304-9e1c-039d35226686","Type":"ContainerDied","Data":"d38dedf8f9062f909f22038364138fce749f1c07518101c6bfdc3e1d5f365369"} Mar 13 15:49:36 crc kubenswrapper[4907]: I0313 15:49:36.309817 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-6qbgl" Mar 13 15:49:36 crc kubenswrapper[4907]: I0313 15:49:36.393610 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/689fb651-0113-48ba-b6e4-29c41c5fa21f-operator-scripts\") pod \"689fb651-0113-48ba-b6e4-29c41c5fa21f\" (UID: \"689fb651-0113-48ba-b6e4-29c41c5fa21f\") " Mar 13 15:49:36 crc kubenswrapper[4907]: I0313 15:49:36.393728 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r7s7z\" (UniqueName: \"kubernetes.io/projected/689fb651-0113-48ba-b6e4-29c41c5fa21f-kube-api-access-r7s7z\") pod \"689fb651-0113-48ba-b6e4-29c41c5fa21f\" (UID: \"689fb651-0113-48ba-b6e4-29c41c5fa21f\") " Mar 13 15:49:36 crc kubenswrapper[4907]: I0313 15:49:36.394403 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/689fb651-0113-48ba-b6e4-29c41c5fa21f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "689fb651-0113-48ba-b6e4-29c41c5fa21f" (UID: "689fb651-0113-48ba-b6e4-29c41c5fa21f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:49:36 crc kubenswrapper[4907]: I0313 15:49:36.394599 4907 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/689fb651-0113-48ba-b6e4-29c41c5fa21f-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 15:49:36 crc kubenswrapper[4907]: I0313 15:49:36.399958 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/689fb651-0113-48ba-b6e4-29c41c5fa21f-kube-api-access-r7s7z" (OuterVolumeSpecName: "kube-api-access-r7s7z") pod "689fb651-0113-48ba-b6e4-29c41c5fa21f" (UID: "689fb651-0113-48ba-b6e4-29c41c5fa21f"). InnerVolumeSpecName "kube-api-access-r7s7z". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:49:36 crc kubenswrapper[4907]: I0313 15:49:36.496973 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r7s7z\" (UniqueName: \"kubernetes.io/projected/689fb651-0113-48ba-b6e4-29c41c5fa21f-kube-api-access-r7s7z\") on node \"crc\" DevicePath \"\"" Mar 13 15:49:36 crc kubenswrapper[4907]: I0313 15:49:36.906007 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-6qbgl" event={"ID":"689fb651-0113-48ba-b6e4-29c41c5fa21f","Type":"ContainerDied","Data":"36da2d4a680e719ad1fe919746dbf7987f40a93fe7fec96a6662addf2879d7df"} Mar 13 15:49:36 crc kubenswrapper[4907]: I0313 15:49:36.906069 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="36da2d4a680e719ad1fe919746dbf7987f40a93fe7fec96a6662addf2879d7df" Mar 13 15:49:36 crc kubenswrapper[4907]: I0313 15:49:36.906131 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-6qbgl" Mar 13 15:49:37 crc kubenswrapper[4907]: I0313 15:49:37.328666 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-cc97-account-create-update-fc5r7" Mar 13 15:49:37 crc kubenswrapper[4907]: I0313 15:49:37.418213 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/442daa10-2973-4304-9e1c-039d35226686-operator-scripts\") pod \"442daa10-2973-4304-9e1c-039d35226686\" (UID: \"442daa10-2973-4304-9e1c-039d35226686\") " Mar 13 15:49:37 crc kubenswrapper[4907]: I0313 15:49:37.418345 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-74mmw\" (UniqueName: \"kubernetes.io/projected/442daa10-2973-4304-9e1c-039d35226686-kube-api-access-74mmw\") pod \"442daa10-2973-4304-9e1c-039d35226686\" (UID: \"442daa10-2973-4304-9e1c-039d35226686\") " Mar 13 15:49:37 crc kubenswrapper[4907]: I0313 15:49:37.418845 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/442daa10-2973-4304-9e1c-039d35226686-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "442daa10-2973-4304-9e1c-039d35226686" (UID: "442daa10-2973-4304-9e1c-039d35226686"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:49:37 crc kubenswrapper[4907]: I0313 15:49:37.419490 4907 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/442daa10-2973-4304-9e1c-039d35226686-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 15:49:37 crc kubenswrapper[4907]: I0313 15:49:37.443649 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/442daa10-2973-4304-9e1c-039d35226686-kube-api-access-74mmw" (OuterVolumeSpecName: "kube-api-access-74mmw") pod "442daa10-2973-4304-9e1c-039d35226686" (UID: "442daa10-2973-4304-9e1c-039d35226686"). InnerVolumeSpecName "kube-api-access-74mmw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:49:37 crc kubenswrapper[4907]: I0313 15:49:37.522603 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-74mmw\" (UniqueName: \"kubernetes.io/projected/442daa10-2973-4304-9e1c-039d35226686-kube-api-access-74mmw\") on node \"crc\" DevicePath \"\"" Mar 13 15:49:37 crc kubenswrapper[4907]: I0313 15:49:37.578752 4907 scope.go:117] "RemoveContainer" containerID="10110c9c9bf7e0c30409af7a457a19b276708bc01363735a9e6dff6935d07eb0" Mar 13 15:49:37 crc kubenswrapper[4907]: I0313 15:49:37.620170 4907 scope.go:117] "RemoveContainer" containerID="48b0748e550ebe6b33142e46ea370e9b43cfd8df0c8482529bda2ee099868df0" Mar 13 15:49:37 crc kubenswrapper[4907]: I0313 15:49:37.658649 4907 scope.go:117] "RemoveContainer" containerID="dcfa157fb4134a5c5ee1865cf6f3b6a035a94e4eeb4f7a317b410b36548c37aa" Mar 13 15:49:37 crc kubenswrapper[4907]: I0313 15:49:37.917392 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-cc97-account-create-update-fc5r7" event={"ID":"442daa10-2973-4304-9e1c-039d35226686","Type":"ContainerDied","Data":"42a5129d078b177eff9fafc5af8936456bf22055b57fe17bc2a77ae226405e6d"} Mar 13 15:49:37 crc kubenswrapper[4907]: I0313 15:49:37.917699 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="42a5129d078b177eff9fafc5af8936456bf22055b57fe17bc2a77ae226405e6d" Mar 13 15:49:37 crc kubenswrapper[4907]: I0313 15:49:37.917587 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-cc97-account-create-update-fc5r7" Mar 13 15:49:38 crc kubenswrapper[4907]: I0313 15:49:38.880442 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-db-sync-nwwc8"] Mar 13 15:49:38 crc kubenswrapper[4907]: E0313 15:49:38.880866 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="689fb651-0113-48ba-b6e4-29c41c5fa21f" containerName="mariadb-database-create" Mar 13 15:49:38 crc kubenswrapper[4907]: I0313 15:49:38.880899 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="689fb651-0113-48ba-b6e4-29c41c5fa21f" containerName="mariadb-database-create" Mar 13 15:49:38 crc kubenswrapper[4907]: E0313 15:49:38.880914 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="442daa10-2973-4304-9e1c-039d35226686" containerName="mariadb-account-create-update" Mar 13 15:49:38 crc kubenswrapper[4907]: I0313 15:49:38.880920 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="442daa10-2973-4304-9e1c-039d35226686" containerName="mariadb-account-create-update" Mar 13 15:49:38 crc kubenswrapper[4907]: I0313 15:49:38.881293 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="442daa10-2973-4304-9e1c-039d35226686" containerName="mariadb-account-create-update" Mar 13 15:49:38 crc kubenswrapper[4907]: I0313 15:49:38.881310 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="689fb651-0113-48ba-b6e4-29c41c5fa21f" containerName="mariadb-database-create" Mar 13 15:49:38 crc kubenswrapper[4907]: I0313 15:49:38.882039 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-nwwc8" Mar 13 15:49:38 crc kubenswrapper[4907]: I0313 15:49:38.886124 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Mar 13 15:49:38 crc kubenswrapper[4907]: I0313 15:49:38.886300 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-j6c7x" Mar 13 15:49:38 crc kubenswrapper[4907]: I0313 15:49:38.886329 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Mar 13 15:49:38 crc kubenswrapper[4907]: I0313 15:49:38.886440 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Mar 13 15:49:38 crc kubenswrapper[4907]: I0313 15:49:38.896080 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-nwwc8"] Mar 13 15:49:38 crc kubenswrapper[4907]: I0313 15:49:38.951433 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c5a77890-eb92-4d55-acf8-481f70c9277c-scripts\") pod \"aodh-db-sync-nwwc8\" (UID: \"c5a77890-eb92-4d55-acf8-481f70c9277c\") " pod="openstack/aodh-db-sync-nwwc8" Mar 13 15:49:38 crc kubenswrapper[4907]: I0313 15:49:38.951642 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rt9sh\" (UniqueName: \"kubernetes.io/projected/c5a77890-eb92-4d55-acf8-481f70c9277c-kube-api-access-rt9sh\") pod \"aodh-db-sync-nwwc8\" (UID: \"c5a77890-eb92-4d55-acf8-481f70c9277c\") " pod="openstack/aodh-db-sync-nwwc8" Mar 13 15:49:38 crc kubenswrapper[4907]: I0313 15:49:38.951735 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5a77890-eb92-4d55-acf8-481f70c9277c-combined-ca-bundle\") pod \"aodh-db-sync-nwwc8\" (UID: \"c5a77890-eb92-4d55-acf8-481f70c9277c\") " pod="openstack/aodh-db-sync-nwwc8" Mar 13 15:49:38 crc kubenswrapper[4907]: I0313 15:49:38.952024 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5a77890-eb92-4d55-acf8-481f70c9277c-config-data\") pod \"aodh-db-sync-nwwc8\" (UID: \"c5a77890-eb92-4d55-acf8-481f70c9277c\") " pod="openstack/aodh-db-sync-nwwc8" Mar 13 15:49:39 crc kubenswrapper[4907]: I0313 15:49:39.053953 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5a77890-eb92-4d55-acf8-481f70c9277c-config-data\") pod \"aodh-db-sync-nwwc8\" (UID: \"c5a77890-eb92-4d55-acf8-481f70c9277c\") " pod="openstack/aodh-db-sync-nwwc8" Mar 13 15:49:39 crc kubenswrapper[4907]: I0313 15:49:39.054071 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c5a77890-eb92-4d55-acf8-481f70c9277c-scripts\") pod \"aodh-db-sync-nwwc8\" (UID: \"c5a77890-eb92-4d55-acf8-481f70c9277c\") " pod="openstack/aodh-db-sync-nwwc8" Mar 13 15:49:39 crc kubenswrapper[4907]: I0313 15:49:39.054172 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rt9sh\" (UniqueName: \"kubernetes.io/projected/c5a77890-eb92-4d55-acf8-481f70c9277c-kube-api-access-rt9sh\") pod \"aodh-db-sync-nwwc8\" (UID: \"c5a77890-eb92-4d55-acf8-481f70c9277c\") " pod="openstack/aodh-db-sync-nwwc8" Mar 13 15:49:39 crc kubenswrapper[4907]: I0313 15:49:39.054224 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5a77890-eb92-4d55-acf8-481f70c9277c-combined-ca-bundle\") pod \"aodh-db-sync-nwwc8\" (UID: \"c5a77890-eb92-4d55-acf8-481f70c9277c\") " pod="openstack/aodh-db-sync-nwwc8" Mar 13 15:49:39 crc kubenswrapper[4907]: I0313 15:49:39.060528 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5a77890-eb92-4d55-acf8-481f70c9277c-config-data\") pod \"aodh-db-sync-nwwc8\" (UID: \"c5a77890-eb92-4d55-acf8-481f70c9277c\") " pod="openstack/aodh-db-sync-nwwc8" Mar 13 15:49:39 crc kubenswrapper[4907]: I0313 15:49:39.062343 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c5a77890-eb92-4d55-acf8-481f70c9277c-scripts\") pod \"aodh-db-sync-nwwc8\" (UID: \"c5a77890-eb92-4d55-acf8-481f70c9277c\") " pod="openstack/aodh-db-sync-nwwc8" Mar 13 15:49:39 crc kubenswrapper[4907]: I0313 15:49:39.066566 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5a77890-eb92-4d55-acf8-481f70c9277c-combined-ca-bundle\") pod \"aodh-db-sync-nwwc8\" (UID: \"c5a77890-eb92-4d55-acf8-481f70c9277c\") " pod="openstack/aodh-db-sync-nwwc8" Mar 13 15:49:39 crc kubenswrapper[4907]: I0313 15:49:39.081992 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rt9sh\" (UniqueName: \"kubernetes.io/projected/c5a77890-eb92-4d55-acf8-481f70c9277c-kube-api-access-rt9sh\") pod \"aodh-db-sync-nwwc8\" (UID: \"c5a77890-eb92-4d55-acf8-481f70c9277c\") " pod="openstack/aodh-db-sync-nwwc8" Mar 13 15:49:39 crc kubenswrapper[4907]: I0313 15:49:39.202558 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-nwwc8" Mar 13 15:49:39 crc kubenswrapper[4907]: I0313 15:49:39.677869 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-nwwc8"] Mar 13 15:49:39 crc kubenswrapper[4907]: I0313 15:49:39.948968 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-nwwc8" event={"ID":"c5a77890-eb92-4d55-acf8-481f70c9277c","Type":"ContainerStarted","Data":"5d0309bb66d24476037e34dfde5f9f67db1d48d62b2ca5025abb9b5b8f5d7cb3"} Mar 13 15:49:45 crc kubenswrapper[4907]: I0313 15:49:45.110656 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-nwwc8" event={"ID":"c5a77890-eb92-4d55-acf8-481f70c9277c","Type":"ContainerStarted","Data":"84256d7f66a6d2a7c62b2540cda3765257adc44483f0deff6d7044e7a488955a"} Mar 13 15:49:45 crc kubenswrapper[4907]: I0313 15:49:45.144519 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-db-sync-nwwc8" podStartSLOduration=2.779150826 podStartE2EDuration="7.144496635s" podCreationTimestamp="2026-03-13 15:49:38 +0000 UTC" firstStartedPulling="2026-03-13 15:49:39.693870674 +0000 UTC m=+6278.593658353" lastFinishedPulling="2026-03-13 15:49:44.059216473 +0000 UTC m=+6282.959004162" observedRunningTime="2026-03-13 15:49:45.139740874 +0000 UTC m=+6284.039528573" watchObservedRunningTime="2026-03-13 15:49:45.144496635 +0000 UTC m=+6284.044284324" Mar 13 15:49:45 crc kubenswrapper[4907]: I0313 15:49:45.735635 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-marketplace/community-operators-bwznz" podUID="e3a3b609-3625-4670-b669-d7fd07386be6" containerName="registry-server" probeResult="failure" output=< Mar 13 15:49:45 crc kubenswrapper[4907]: timeout: failed to connect service ":50051" within 1s Mar 13 15:49:45 crc kubenswrapper[4907]: > Mar 13 15:49:45 crc kubenswrapper[4907]: I0313 15:49:45.735931 4907 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/community-operators-bwznz" podUID="e3a3b609-3625-4670-b669-d7fd07386be6" containerName="registry-server" probeResult="failure" output=< Mar 13 15:49:45 crc kubenswrapper[4907]: timeout: failed to connect service ":50051" within 1s Mar 13 15:49:45 crc kubenswrapper[4907]: > Mar 13 15:49:48 crc kubenswrapper[4907]: I0313 15:49:48.041470 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 15:49:48 crc kubenswrapper[4907]: I0313 15:49:48.042832 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 15:49:48 crc kubenswrapper[4907]: I0313 15:49:48.042983 4907 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" Mar 13 15:49:48 crc kubenswrapper[4907]: I0313 15:49:48.044005 4907 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"81d94e236cbdff3b86e8375b8abbaee97f49d844932ee0151cdfa9d97cb1c3b2"} pod="openshift-machine-config-operator/machine-config-daemon-hm56j" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 15:49:48 crc kubenswrapper[4907]: I0313 15:49:48.044144 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" containerID="cri-o://81d94e236cbdff3b86e8375b8abbaee97f49d844932ee0151cdfa9d97cb1c3b2" gracePeriod=600 Mar 13 15:49:48 crc kubenswrapper[4907]: I0313 15:49:48.143350 4907 generic.go:334] "Generic (PLEG): container finished" podID="c5a77890-eb92-4d55-acf8-481f70c9277c" containerID="84256d7f66a6d2a7c62b2540cda3765257adc44483f0deff6d7044e7a488955a" exitCode=0 Mar 13 15:49:48 crc kubenswrapper[4907]: I0313 15:49:48.143393 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-nwwc8" event={"ID":"c5a77890-eb92-4d55-acf8-481f70c9277c","Type":"ContainerDied","Data":"84256d7f66a6d2a7c62b2540cda3765257adc44483f0deff6d7044e7a488955a"} Mar 13 15:49:49 crc kubenswrapper[4907]: I0313 15:49:49.156388 4907 generic.go:334] "Generic (PLEG): container finished" podID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerID="81d94e236cbdff3b86e8375b8abbaee97f49d844932ee0151cdfa9d97cb1c3b2" exitCode=0 Mar 13 15:49:49 crc kubenswrapper[4907]: I0313 15:49:49.157127 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerDied","Data":"81d94e236cbdff3b86e8375b8abbaee97f49d844932ee0151cdfa9d97cb1c3b2"} Mar 13 15:49:49 crc kubenswrapper[4907]: I0313 15:49:49.157155 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerStarted","Data":"33f8e315523143b157797243610ad7b82c7815f9e551d1169f2c036cfcadf792"} Mar 13 15:49:49 crc kubenswrapper[4907]: I0313 15:49:49.157170 4907 scope.go:117] "RemoveContainer" containerID="68ffe92dd03078633ec89d7ef29da9bbdc4c5f04c956a1e8487e3f2e01c5d1c5" Mar 13 15:49:49 crc kubenswrapper[4907]: I0313 15:49:49.677359 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-nwwc8" Mar 13 15:49:49 crc kubenswrapper[4907]: I0313 15:49:49.746329 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rt9sh\" (UniqueName: \"kubernetes.io/projected/c5a77890-eb92-4d55-acf8-481f70c9277c-kube-api-access-rt9sh\") pod \"c5a77890-eb92-4d55-acf8-481f70c9277c\" (UID: \"c5a77890-eb92-4d55-acf8-481f70c9277c\") " Mar 13 15:49:49 crc kubenswrapper[4907]: I0313 15:49:49.746679 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c5a77890-eb92-4d55-acf8-481f70c9277c-scripts\") pod \"c5a77890-eb92-4d55-acf8-481f70c9277c\" (UID: \"c5a77890-eb92-4d55-acf8-481f70c9277c\") " Mar 13 15:49:49 crc kubenswrapper[4907]: I0313 15:49:49.746778 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5a77890-eb92-4d55-acf8-481f70c9277c-config-data\") pod \"c5a77890-eb92-4d55-acf8-481f70c9277c\" (UID: \"c5a77890-eb92-4d55-acf8-481f70c9277c\") " Mar 13 15:49:49 crc kubenswrapper[4907]: I0313 15:49:49.746823 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5a77890-eb92-4d55-acf8-481f70c9277c-combined-ca-bundle\") pod \"c5a77890-eb92-4d55-acf8-481f70c9277c\" (UID: \"c5a77890-eb92-4d55-acf8-481f70c9277c\") " Mar 13 15:49:49 crc kubenswrapper[4907]: I0313 15:49:49.751496 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c5a77890-eb92-4d55-acf8-481f70c9277c-kube-api-access-rt9sh" (OuterVolumeSpecName: "kube-api-access-rt9sh") pod "c5a77890-eb92-4d55-acf8-481f70c9277c" (UID: "c5a77890-eb92-4d55-acf8-481f70c9277c"). InnerVolumeSpecName "kube-api-access-rt9sh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:49:49 crc kubenswrapper[4907]: I0313 15:49:49.753153 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5a77890-eb92-4d55-acf8-481f70c9277c-scripts" (OuterVolumeSpecName: "scripts") pod "c5a77890-eb92-4d55-acf8-481f70c9277c" (UID: "c5a77890-eb92-4d55-acf8-481f70c9277c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:49:49 crc kubenswrapper[4907]: I0313 15:49:49.773787 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5a77890-eb92-4d55-acf8-481f70c9277c-config-data" (OuterVolumeSpecName: "config-data") pod "c5a77890-eb92-4d55-acf8-481f70c9277c" (UID: "c5a77890-eb92-4d55-acf8-481f70c9277c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:49:49 crc kubenswrapper[4907]: I0313 15:49:49.774969 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5a77890-eb92-4d55-acf8-481f70c9277c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c5a77890-eb92-4d55-acf8-481f70c9277c" (UID: "c5a77890-eb92-4d55-acf8-481f70c9277c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:49:49 crc kubenswrapper[4907]: I0313 15:49:49.849273 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5a77890-eb92-4d55-acf8-481f70c9277c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 15:49:49 crc kubenswrapper[4907]: I0313 15:49:49.849303 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5a77890-eb92-4d55-acf8-481f70c9277c-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 15:49:49 crc kubenswrapper[4907]: I0313 15:49:49.849313 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rt9sh\" (UniqueName: \"kubernetes.io/projected/c5a77890-eb92-4d55-acf8-481f70c9277c-kube-api-access-rt9sh\") on node \"crc\" DevicePath \"\"" Mar 13 15:49:49 crc kubenswrapper[4907]: I0313 15:49:49.849321 4907 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c5a77890-eb92-4d55-acf8-481f70c9277c-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 15:49:50 crc kubenswrapper[4907]: I0313 15:49:50.171511 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-nwwc8" event={"ID":"c5a77890-eb92-4d55-acf8-481f70c9277c","Type":"ContainerDied","Data":"5d0309bb66d24476037e34dfde5f9f67db1d48d62b2ca5025abb9b5b8f5d7cb3"} Mar 13 15:49:50 crc kubenswrapper[4907]: I0313 15:49:50.171834 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5d0309bb66d24476037e34dfde5f9f67db1d48d62b2ca5025abb9b5b8f5d7cb3" Mar 13 15:49:50 crc kubenswrapper[4907]: I0313 15:49:50.171574 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-nwwc8" Mar 13 15:49:50 crc kubenswrapper[4907]: I0313 15:49:50.767564 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-jj9xr"] Mar 13 15:49:50 crc kubenswrapper[4907]: E0313 15:49:50.768516 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5a77890-eb92-4d55-acf8-481f70c9277c" containerName="aodh-db-sync" Mar 13 15:49:50 crc kubenswrapper[4907]: I0313 15:49:50.768541 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5a77890-eb92-4d55-acf8-481f70c9277c" containerName="aodh-db-sync" Mar 13 15:49:50 crc kubenswrapper[4907]: I0313 15:49:50.768777 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="c5a77890-eb92-4d55-acf8-481f70c9277c" containerName="aodh-db-sync" Mar 13 15:49:50 crc kubenswrapper[4907]: I0313 15:49:50.770324 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jj9xr" Mar 13 15:49:50 crc kubenswrapper[4907]: I0313 15:49:50.791032 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jj9xr"] Mar 13 15:49:50 crc kubenswrapper[4907]: I0313 15:49:50.867113 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/951bf99b-9f5e-4609-bff0-0c594f6a5b2e-catalog-content\") pod \"certified-operators-jj9xr\" (UID: \"951bf99b-9f5e-4609-bff0-0c594f6a5b2e\") " pod="openshift-marketplace/certified-operators-jj9xr" Mar 13 15:49:50 crc kubenswrapper[4907]: I0313 15:49:50.867165 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/951bf99b-9f5e-4609-bff0-0c594f6a5b2e-utilities\") pod \"certified-operators-jj9xr\" (UID: \"951bf99b-9f5e-4609-bff0-0c594f6a5b2e\") " pod="openshift-marketplace/certified-operators-jj9xr" Mar 13 15:49:50 crc kubenswrapper[4907]: I0313 15:49:50.867209 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h9rf2\" (UniqueName: \"kubernetes.io/projected/951bf99b-9f5e-4609-bff0-0c594f6a5b2e-kube-api-access-h9rf2\") pod \"certified-operators-jj9xr\" (UID: \"951bf99b-9f5e-4609-bff0-0c594f6a5b2e\") " pod="openshift-marketplace/certified-operators-jj9xr" Mar 13 15:49:50 crc kubenswrapper[4907]: I0313 15:49:50.969984 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/951bf99b-9f5e-4609-bff0-0c594f6a5b2e-catalog-content\") pod \"certified-operators-jj9xr\" (UID: \"951bf99b-9f5e-4609-bff0-0c594f6a5b2e\") " pod="openshift-marketplace/certified-operators-jj9xr" Mar 13 15:49:50 crc kubenswrapper[4907]: I0313 15:49:50.970027 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/951bf99b-9f5e-4609-bff0-0c594f6a5b2e-utilities\") pod \"certified-operators-jj9xr\" (UID: \"951bf99b-9f5e-4609-bff0-0c594f6a5b2e\") " pod="openshift-marketplace/certified-operators-jj9xr" Mar 13 15:49:50 crc kubenswrapper[4907]: I0313 15:49:50.970078 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h9rf2\" (UniqueName: \"kubernetes.io/projected/951bf99b-9f5e-4609-bff0-0c594f6a5b2e-kube-api-access-h9rf2\") pod \"certified-operators-jj9xr\" (UID: \"951bf99b-9f5e-4609-bff0-0c594f6a5b2e\") " pod="openshift-marketplace/certified-operators-jj9xr" Mar 13 15:49:50 crc kubenswrapper[4907]: I0313 15:49:50.971249 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/951bf99b-9f5e-4609-bff0-0c594f6a5b2e-catalog-content\") pod \"certified-operators-jj9xr\" (UID: \"951bf99b-9f5e-4609-bff0-0c594f6a5b2e\") " pod="openshift-marketplace/certified-operators-jj9xr" Mar 13 15:49:50 crc kubenswrapper[4907]: I0313 15:49:50.971262 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/951bf99b-9f5e-4609-bff0-0c594f6a5b2e-utilities\") pod \"certified-operators-jj9xr\" (UID: \"951bf99b-9f5e-4609-bff0-0c594f6a5b2e\") " pod="openshift-marketplace/certified-operators-jj9xr" Mar 13 15:49:50 crc kubenswrapper[4907]: I0313 15:49:50.994655 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h9rf2\" (UniqueName: \"kubernetes.io/projected/951bf99b-9f5e-4609-bff0-0c594f6a5b2e-kube-api-access-h9rf2\") pod \"certified-operators-jj9xr\" (UID: \"951bf99b-9f5e-4609-bff0-0c594f6a5b2e\") " pod="openshift-marketplace/certified-operators-jj9xr" Mar 13 15:49:51 crc kubenswrapper[4907]: I0313 15:49:51.096161 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jj9xr" Mar 13 15:49:51 crc kubenswrapper[4907]: I0313 15:49:51.599927 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jj9xr"] Mar 13 15:49:52 crc kubenswrapper[4907]: I0313 15:49:52.199462 4907 generic.go:334] "Generic (PLEG): container finished" podID="951bf99b-9f5e-4609-bff0-0c594f6a5b2e" containerID="240a7e1dba69382ab9e9bd056e343991036dcabf16f06d6bad39768a4324c999" exitCode=0 Mar 13 15:49:52 crc kubenswrapper[4907]: I0313 15:49:52.199523 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jj9xr" event={"ID":"951bf99b-9f5e-4609-bff0-0c594f6a5b2e","Type":"ContainerDied","Data":"240a7e1dba69382ab9e9bd056e343991036dcabf16f06d6bad39768a4324c999"} Mar 13 15:49:52 crc kubenswrapper[4907]: I0313 15:49:52.199807 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jj9xr" event={"ID":"951bf99b-9f5e-4609-bff0-0c594f6a5b2e","Type":"ContainerStarted","Data":"02bff9741d914eb2d1510a0ab7003a1cb31cbd80634618a50cf32118561743ca"} Mar 13 15:49:52 crc kubenswrapper[4907]: I0313 15:49:52.546148 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Mar 13 15:49:53 crc kubenswrapper[4907]: I0313 15:49:53.212304 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jj9xr" event={"ID":"951bf99b-9f5e-4609-bff0-0c594f6a5b2e","Type":"ContainerStarted","Data":"d2ea7f99cadaa4f0d0bfca645932300ec0e331fafcb1e08d6511678f159fdd05"} Mar 13 15:49:54 crc kubenswrapper[4907]: I0313 15:49:54.040194 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-0"] Mar 13 15:49:54 crc kubenswrapper[4907]: I0313 15:49:54.043446 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Mar 13 15:49:54 crc kubenswrapper[4907]: I0313 15:49:54.046275 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-j6c7x" Mar 13 15:49:54 crc kubenswrapper[4907]: I0313 15:49:54.046950 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Mar 13 15:49:54 crc kubenswrapper[4907]: I0313 15:49:54.047649 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Mar 13 15:49:54 crc kubenswrapper[4907]: I0313 15:49:54.051780 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Mar 13 15:49:54 crc kubenswrapper[4907]: I0313 15:49:54.139345 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wdszn\" (UniqueName: \"kubernetes.io/projected/8ce53a96-2566-4d5c-b4cd-bbfd2516975a-kube-api-access-wdszn\") pod \"aodh-0\" (UID: \"8ce53a96-2566-4d5c-b4cd-bbfd2516975a\") " pod="openstack/aodh-0" Mar 13 15:49:54 crc kubenswrapper[4907]: I0313 15:49:54.139417 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8ce53a96-2566-4d5c-b4cd-bbfd2516975a-scripts\") pod \"aodh-0\" (UID: \"8ce53a96-2566-4d5c-b4cd-bbfd2516975a\") " pod="openstack/aodh-0" Mar 13 15:49:54 crc kubenswrapper[4907]: I0313 15:49:54.139486 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ce53a96-2566-4d5c-b4cd-bbfd2516975a-config-data\") pod \"aodh-0\" (UID: \"8ce53a96-2566-4d5c-b4cd-bbfd2516975a\") " pod="openstack/aodh-0" Mar 13 15:49:54 crc kubenswrapper[4907]: I0313 15:49:54.139567 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ce53a96-2566-4d5c-b4cd-bbfd2516975a-combined-ca-bundle\") pod \"aodh-0\" (UID: \"8ce53a96-2566-4d5c-b4cd-bbfd2516975a\") " pod="openstack/aodh-0" Mar 13 15:49:54 crc kubenswrapper[4907]: I0313 15:49:54.241374 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wdszn\" (UniqueName: \"kubernetes.io/projected/8ce53a96-2566-4d5c-b4cd-bbfd2516975a-kube-api-access-wdszn\") pod \"aodh-0\" (UID: \"8ce53a96-2566-4d5c-b4cd-bbfd2516975a\") " pod="openstack/aodh-0" Mar 13 15:49:54 crc kubenswrapper[4907]: I0313 15:49:54.241438 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8ce53a96-2566-4d5c-b4cd-bbfd2516975a-scripts\") pod \"aodh-0\" (UID: \"8ce53a96-2566-4d5c-b4cd-bbfd2516975a\") " pod="openstack/aodh-0" Mar 13 15:49:54 crc kubenswrapper[4907]: I0313 15:49:54.241505 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ce53a96-2566-4d5c-b4cd-bbfd2516975a-config-data\") pod \"aodh-0\" (UID: \"8ce53a96-2566-4d5c-b4cd-bbfd2516975a\") " pod="openstack/aodh-0" Mar 13 15:49:54 crc kubenswrapper[4907]: I0313 15:49:54.241581 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ce53a96-2566-4d5c-b4cd-bbfd2516975a-combined-ca-bundle\") pod \"aodh-0\" (UID: \"8ce53a96-2566-4d5c-b4cd-bbfd2516975a\") " pod="openstack/aodh-0" Mar 13 15:49:54 crc kubenswrapper[4907]: I0313 15:49:54.253671 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8ce53a96-2566-4d5c-b4cd-bbfd2516975a-scripts\") pod \"aodh-0\" (UID: \"8ce53a96-2566-4d5c-b4cd-bbfd2516975a\") " pod="openstack/aodh-0" Mar 13 15:49:54 crc kubenswrapper[4907]: I0313 15:49:54.253950 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ce53a96-2566-4d5c-b4cd-bbfd2516975a-combined-ca-bundle\") pod \"aodh-0\" (UID: \"8ce53a96-2566-4d5c-b4cd-bbfd2516975a\") " pod="openstack/aodh-0" Mar 13 15:49:54 crc kubenswrapper[4907]: I0313 15:49:54.254698 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ce53a96-2566-4d5c-b4cd-bbfd2516975a-config-data\") pod \"aodh-0\" (UID: \"8ce53a96-2566-4d5c-b4cd-bbfd2516975a\") " pod="openstack/aodh-0" Mar 13 15:49:54 crc kubenswrapper[4907]: I0313 15:49:54.268354 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wdszn\" (UniqueName: \"kubernetes.io/projected/8ce53a96-2566-4d5c-b4cd-bbfd2516975a-kube-api-access-wdszn\") pod \"aodh-0\" (UID: \"8ce53a96-2566-4d5c-b4cd-bbfd2516975a\") " pod="openstack/aodh-0" Mar 13 15:49:54 crc kubenswrapper[4907]: I0313 15:49:54.364538 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Mar 13 15:49:55 crc kubenswrapper[4907]: W0313 15:49:54.956069 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8ce53a96_2566_4d5c_b4cd_bbfd2516975a.slice/crio-643f6e0412f2540662588d14df1c54b0fff1ba0d761c6475ac478360aa911c13 WatchSource:0}: Error finding container 643f6e0412f2540662588d14df1c54b0fff1ba0d761c6475ac478360aa911c13: Status 404 returned error can't find the container with id 643f6e0412f2540662588d14df1c54b0fff1ba0d761c6475ac478360aa911c13 Mar 13 15:49:55 crc kubenswrapper[4907]: I0313 15:49:54.958421 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Mar 13 15:49:55 crc kubenswrapper[4907]: I0313 15:49:55.048655 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-5r4tb"] Mar 13 15:49:55 crc kubenswrapper[4907]: I0313 15:49:55.063112 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-5r4tb"] Mar 13 15:49:55 crc kubenswrapper[4907]: I0313 15:49:55.237747 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"8ce53a96-2566-4d5c-b4cd-bbfd2516975a","Type":"ContainerStarted","Data":"643f6e0412f2540662588d14df1c54b0fff1ba0d761c6475ac478360aa911c13"} Mar 13 15:49:55 crc kubenswrapper[4907]: I0313 15:49:55.795307 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e5aab76f-4bcc-47be-989d-7d9d6ecf371b" path="/var/lib/kubelet/pods/e5aab76f-4bcc-47be-989d-7d9d6ecf371b/volumes" Mar 13 15:49:56 crc kubenswrapper[4907]: I0313 15:49:56.047679 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-7djpm"] Mar 13 15:49:56 crc kubenswrapper[4907]: I0313 15:49:56.064397 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-62d0-account-create-update-vf8sr"] Mar 13 15:49:56 crc kubenswrapper[4907]: I0313 15:49:56.074394 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-5a99-account-create-update-nkg4x"] Mar 13 15:49:56 crc kubenswrapper[4907]: I0313 15:49:56.086867 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-7djpm"] Mar 13 15:49:56 crc kubenswrapper[4907]: I0313 15:49:56.095761 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-brbzj"] Mar 13 15:49:56 crc kubenswrapper[4907]: I0313 15:49:56.105287 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-62d0-account-create-update-vf8sr"] Mar 13 15:49:56 crc kubenswrapper[4907]: I0313 15:49:56.115159 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-5a99-account-create-update-nkg4x"] Mar 13 15:49:56 crc kubenswrapper[4907]: I0313 15:49:56.124522 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-356d-account-create-update-kb8lr"] Mar 13 15:49:56 crc kubenswrapper[4907]: I0313 15:49:56.133302 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-brbzj"] Mar 13 15:49:56 crc kubenswrapper[4907]: I0313 15:49:56.141249 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-356d-account-create-update-kb8lr"] Mar 13 15:49:56 crc kubenswrapper[4907]: I0313 15:49:56.247862 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"8ce53a96-2566-4d5c-b4cd-bbfd2516975a","Type":"ContainerStarted","Data":"f540978f1e9f16494cf5a8dc9d6288ee493f15ca0f0d76fc59aee4c243f42647"} Mar 13 15:49:56 crc kubenswrapper[4907]: I0313 15:49:56.338903 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 13 15:49:56 crc kubenswrapper[4907]: I0313 15:49:56.339308 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="eaec5907-c808-4ac2-b131-608b228885e6" containerName="ceilometer-central-agent" containerID="cri-o://b3ab302efa65117dbfc68c5b768e1ee7a624666e299637f13e417c20da84fcf0" gracePeriod=30 Mar 13 15:49:56 crc kubenswrapper[4907]: I0313 15:49:56.339758 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="eaec5907-c808-4ac2-b131-608b228885e6" containerName="proxy-httpd" containerID="cri-o://5a450b86aabb5b84d29c1e3252b0f33c4167256fbc37a7560f1e23aae3ba5797" gracePeriod=30 Mar 13 15:49:56 crc kubenswrapper[4907]: I0313 15:49:56.339834 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="eaec5907-c808-4ac2-b131-608b228885e6" containerName="ceilometer-notification-agent" containerID="cri-o://f8bccaf5228a1db9069e313bf3466a6f8cbd61fae45f55bda0a236da15bba7f0" gracePeriod=30 Mar 13 15:49:56 crc kubenswrapper[4907]: I0313 15:49:56.340081 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="eaec5907-c808-4ac2-b131-608b228885e6" containerName="sg-core" containerID="cri-o://3b5b22ec27be798569c983dbf41141eab318a4ae784c5efd3348ea56c4f000f7" gracePeriod=30 Mar 13 15:49:57 crc kubenswrapper[4907]: I0313 15:49:57.288406 4907 generic.go:334] "Generic (PLEG): container finished" podID="951bf99b-9f5e-4609-bff0-0c594f6a5b2e" containerID="d2ea7f99cadaa4f0d0bfca645932300ec0e331fafcb1e08d6511678f159fdd05" exitCode=0 Mar 13 15:49:57 crc kubenswrapper[4907]: I0313 15:49:57.288485 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jj9xr" event={"ID":"951bf99b-9f5e-4609-bff0-0c594f6a5b2e","Type":"ContainerDied","Data":"d2ea7f99cadaa4f0d0bfca645932300ec0e331fafcb1e08d6511678f159fdd05"} Mar 13 15:49:57 crc kubenswrapper[4907]: I0313 15:49:57.305409 4907 generic.go:334] "Generic (PLEG): container finished" podID="eaec5907-c808-4ac2-b131-608b228885e6" containerID="5a450b86aabb5b84d29c1e3252b0f33c4167256fbc37a7560f1e23aae3ba5797" exitCode=0 Mar 13 15:49:57 crc kubenswrapper[4907]: I0313 15:49:57.305440 4907 generic.go:334] "Generic (PLEG): container finished" podID="eaec5907-c808-4ac2-b131-608b228885e6" containerID="3b5b22ec27be798569c983dbf41141eab318a4ae784c5efd3348ea56c4f000f7" exitCode=2 Mar 13 15:49:57 crc kubenswrapper[4907]: I0313 15:49:57.305448 4907 generic.go:334] "Generic (PLEG): container finished" podID="eaec5907-c808-4ac2-b131-608b228885e6" containerID="f8bccaf5228a1db9069e313bf3466a6f8cbd61fae45f55bda0a236da15bba7f0" exitCode=0 Mar 13 15:49:57 crc kubenswrapper[4907]: I0313 15:49:57.305461 4907 generic.go:334] "Generic (PLEG): container finished" podID="eaec5907-c808-4ac2-b131-608b228885e6" containerID="b3ab302efa65117dbfc68c5b768e1ee7a624666e299637f13e417c20da84fcf0" exitCode=0 Mar 13 15:49:57 crc kubenswrapper[4907]: I0313 15:49:57.305482 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eaec5907-c808-4ac2-b131-608b228885e6","Type":"ContainerDied","Data":"5a450b86aabb5b84d29c1e3252b0f33c4167256fbc37a7560f1e23aae3ba5797"} Mar 13 15:49:57 crc kubenswrapper[4907]: I0313 15:49:57.305510 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eaec5907-c808-4ac2-b131-608b228885e6","Type":"ContainerDied","Data":"3b5b22ec27be798569c983dbf41141eab318a4ae784c5efd3348ea56c4f000f7"} Mar 13 15:49:57 crc kubenswrapper[4907]: I0313 15:49:57.305519 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eaec5907-c808-4ac2-b131-608b228885e6","Type":"ContainerDied","Data":"f8bccaf5228a1db9069e313bf3466a6f8cbd61fae45f55bda0a236da15bba7f0"} Mar 13 15:49:57 crc kubenswrapper[4907]: I0313 15:49:57.305529 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eaec5907-c808-4ac2-b131-608b228885e6","Type":"ContainerDied","Data":"b3ab302efa65117dbfc68c5b768e1ee7a624666e299637f13e417c20da84fcf0"} Mar 13 15:49:57 crc kubenswrapper[4907]: I0313 15:49:57.442192 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 15:49:57 crc kubenswrapper[4907]: I0313 15:49:57.513515 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eaec5907-c808-4ac2-b131-608b228885e6-config-data\") pod \"eaec5907-c808-4ac2-b131-608b228885e6\" (UID: \"eaec5907-c808-4ac2-b131-608b228885e6\") " Mar 13 15:49:57 crc kubenswrapper[4907]: I0313 15:49:57.513600 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/eaec5907-c808-4ac2-b131-608b228885e6-sg-core-conf-yaml\") pod \"eaec5907-c808-4ac2-b131-608b228885e6\" (UID: \"eaec5907-c808-4ac2-b131-608b228885e6\") " Mar 13 15:49:57 crc kubenswrapper[4907]: I0313 15:49:57.513737 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eaec5907-c808-4ac2-b131-608b228885e6-combined-ca-bundle\") pod \"eaec5907-c808-4ac2-b131-608b228885e6\" (UID: \"eaec5907-c808-4ac2-b131-608b228885e6\") " Mar 13 15:49:57 crc kubenswrapper[4907]: I0313 15:49:57.513801 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x6gpx\" (UniqueName: \"kubernetes.io/projected/eaec5907-c808-4ac2-b131-608b228885e6-kube-api-access-x6gpx\") pod \"eaec5907-c808-4ac2-b131-608b228885e6\" (UID: \"eaec5907-c808-4ac2-b131-608b228885e6\") " Mar 13 15:49:57 crc kubenswrapper[4907]: I0313 15:49:57.513868 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eaec5907-c808-4ac2-b131-608b228885e6-run-httpd\") pod \"eaec5907-c808-4ac2-b131-608b228885e6\" (UID: \"eaec5907-c808-4ac2-b131-608b228885e6\") " Mar 13 15:49:57 crc kubenswrapper[4907]: I0313 15:49:57.513970 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eaec5907-c808-4ac2-b131-608b228885e6-log-httpd\") pod \"eaec5907-c808-4ac2-b131-608b228885e6\" (UID: \"eaec5907-c808-4ac2-b131-608b228885e6\") " Mar 13 15:49:57 crc kubenswrapper[4907]: I0313 15:49:57.513998 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eaec5907-c808-4ac2-b131-608b228885e6-scripts\") pod \"eaec5907-c808-4ac2-b131-608b228885e6\" (UID: \"eaec5907-c808-4ac2-b131-608b228885e6\") " Mar 13 15:49:57 crc kubenswrapper[4907]: I0313 15:49:57.517475 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eaec5907-c808-4ac2-b131-608b228885e6-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "eaec5907-c808-4ac2-b131-608b228885e6" (UID: "eaec5907-c808-4ac2-b131-608b228885e6"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:49:57 crc kubenswrapper[4907]: I0313 15:49:57.517792 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eaec5907-c808-4ac2-b131-608b228885e6-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "eaec5907-c808-4ac2-b131-608b228885e6" (UID: "eaec5907-c808-4ac2-b131-608b228885e6"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:49:57 crc kubenswrapper[4907]: I0313 15:49:57.536765 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eaec5907-c808-4ac2-b131-608b228885e6-scripts" (OuterVolumeSpecName: "scripts") pod "eaec5907-c808-4ac2-b131-608b228885e6" (UID: "eaec5907-c808-4ac2-b131-608b228885e6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:49:57 crc kubenswrapper[4907]: I0313 15:49:57.536876 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eaec5907-c808-4ac2-b131-608b228885e6-kube-api-access-x6gpx" (OuterVolumeSpecName: "kube-api-access-x6gpx") pod "eaec5907-c808-4ac2-b131-608b228885e6" (UID: "eaec5907-c808-4ac2-b131-608b228885e6"). InnerVolumeSpecName "kube-api-access-x6gpx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:49:57 crc kubenswrapper[4907]: I0313 15:49:57.547468 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eaec5907-c808-4ac2-b131-608b228885e6-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "eaec5907-c808-4ac2-b131-608b228885e6" (UID: "eaec5907-c808-4ac2-b131-608b228885e6"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:49:57 crc kubenswrapper[4907]: I0313 15:49:57.616958 4907 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/eaec5907-c808-4ac2-b131-608b228885e6-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 13 15:49:57 crc kubenswrapper[4907]: I0313 15:49:57.616996 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x6gpx\" (UniqueName: \"kubernetes.io/projected/eaec5907-c808-4ac2-b131-608b228885e6-kube-api-access-x6gpx\") on node \"crc\" DevicePath \"\"" Mar 13 15:49:57 crc kubenswrapper[4907]: I0313 15:49:57.617011 4907 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eaec5907-c808-4ac2-b131-608b228885e6-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 13 15:49:57 crc kubenswrapper[4907]: I0313 15:49:57.617021 4907 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eaec5907-c808-4ac2-b131-608b228885e6-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 13 15:49:57 crc kubenswrapper[4907]: I0313 15:49:57.617031 4907 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eaec5907-c808-4ac2-b131-608b228885e6-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 15:49:57 crc kubenswrapper[4907]: I0313 15:49:57.648963 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eaec5907-c808-4ac2-b131-608b228885e6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "eaec5907-c808-4ac2-b131-608b228885e6" (UID: "eaec5907-c808-4ac2-b131-608b228885e6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:49:57 crc kubenswrapper[4907]: I0313 15:49:57.651075 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eaec5907-c808-4ac2-b131-608b228885e6-config-data" (OuterVolumeSpecName: "config-data") pod "eaec5907-c808-4ac2-b131-608b228885e6" (UID: "eaec5907-c808-4ac2-b131-608b228885e6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:49:57 crc kubenswrapper[4907]: I0313 15:49:57.718322 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eaec5907-c808-4ac2-b131-608b228885e6-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 15:49:57 crc kubenswrapper[4907]: I0313 15:49:57.718350 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eaec5907-c808-4ac2-b131-608b228885e6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 15:49:57 crc kubenswrapper[4907]: I0313 15:49:57.799664 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="18fa9304-50a2-45cf-b983-9d459f0dcb5f" path="/var/lib/kubelet/pods/18fa9304-50a2-45cf-b983-9d459f0dcb5f/volumes" Mar 13 15:49:57 crc kubenswrapper[4907]: I0313 15:49:57.809348 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="40763271-36d9-4bd2-8ae8-82140648fcf4" path="/var/lib/kubelet/pods/40763271-36d9-4bd2-8ae8-82140648fcf4/volumes" Mar 13 15:49:57 crc kubenswrapper[4907]: I0313 15:49:57.810444 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="94ba887c-f88a-48e7-b4bc-2f171d21fa47" path="/var/lib/kubelet/pods/94ba887c-f88a-48e7-b4bc-2f171d21fa47/volumes" Mar 13 15:49:57 crc kubenswrapper[4907]: I0313 15:49:57.811001 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b66093c1-e494-4975-b8d1-217844060204" path="/var/lib/kubelet/pods/b66093c1-e494-4975-b8d1-217844060204/volumes" Mar 13 15:49:57 crc kubenswrapper[4907]: I0313 15:49:57.812036 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c2c52fb5-c4ba-4c24-95af-ce75cd609126" path="/var/lib/kubelet/pods/c2c52fb5-c4ba-4c24-95af-ce75cd609126/volumes" Mar 13 15:49:58 crc kubenswrapper[4907]: I0313 15:49:58.317450 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jj9xr" event={"ID":"951bf99b-9f5e-4609-bff0-0c594f6a5b2e","Type":"ContainerStarted","Data":"ad22a18bbef271426a4077aab78c5dbcaeed67525c562d753181de83cefb8f37"} Mar 13 15:49:58 crc kubenswrapper[4907]: I0313 15:49:58.322792 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"8ce53a96-2566-4d5c-b4cd-bbfd2516975a","Type":"ContainerStarted","Data":"646195359af961146eb2ff6e07795dd84d4ca2e44a087864e132385c5bbf36c8"} Mar 13 15:49:58 crc kubenswrapper[4907]: I0313 15:49:58.326529 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eaec5907-c808-4ac2-b131-608b228885e6","Type":"ContainerDied","Data":"ae9b3358826e66742e51bbbc204d8b30a566e7e0b936fca1df7ae895ee7043e9"} Mar 13 15:49:58 crc kubenswrapper[4907]: I0313 15:49:58.326582 4907 scope.go:117] "RemoveContainer" containerID="5a450b86aabb5b84d29c1e3252b0f33c4167256fbc37a7560f1e23aae3ba5797" Mar 13 15:49:58 crc kubenswrapper[4907]: I0313 15:49:58.326765 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 15:49:58 crc kubenswrapper[4907]: I0313 15:49:58.347777 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-jj9xr" podStartSLOduration=2.58015895 podStartE2EDuration="8.34774993s" podCreationTimestamp="2026-03-13 15:49:50 +0000 UTC" firstStartedPulling="2026-03-13 15:49:52.20284506 +0000 UTC m=+6291.102632749" lastFinishedPulling="2026-03-13 15:49:57.97043604 +0000 UTC m=+6296.870223729" observedRunningTime="2026-03-13 15:49:58.333314825 +0000 UTC m=+6297.233102514" watchObservedRunningTime="2026-03-13 15:49:58.34774993 +0000 UTC m=+6297.247537619" Mar 13 15:49:58 crc kubenswrapper[4907]: I0313 15:49:58.367000 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 13 15:49:58 crc kubenswrapper[4907]: I0313 15:49:58.367963 4907 scope.go:117] "RemoveContainer" containerID="3b5b22ec27be798569c983dbf41141eab318a4ae784c5efd3348ea56c4f000f7" Mar 13 15:49:58 crc kubenswrapper[4907]: I0313 15:49:58.388673 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Mar 13 15:49:58 crc kubenswrapper[4907]: I0313 15:49:58.401296 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 13 15:49:58 crc kubenswrapper[4907]: E0313 15:49:58.402028 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eaec5907-c808-4ac2-b131-608b228885e6" containerName="proxy-httpd" Mar 13 15:49:58 crc kubenswrapper[4907]: I0313 15:49:58.402053 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="eaec5907-c808-4ac2-b131-608b228885e6" containerName="proxy-httpd" Mar 13 15:49:58 crc kubenswrapper[4907]: E0313 15:49:58.402079 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eaec5907-c808-4ac2-b131-608b228885e6" containerName="ceilometer-notification-agent" Mar 13 15:49:58 crc kubenswrapper[4907]: I0313 15:49:58.402088 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="eaec5907-c808-4ac2-b131-608b228885e6" containerName="ceilometer-notification-agent" Mar 13 15:49:58 crc kubenswrapper[4907]: E0313 15:49:58.402104 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eaec5907-c808-4ac2-b131-608b228885e6" containerName="ceilometer-central-agent" Mar 13 15:49:58 crc kubenswrapper[4907]: I0313 15:49:58.402113 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="eaec5907-c808-4ac2-b131-608b228885e6" containerName="ceilometer-central-agent" Mar 13 15:49:58 crc kubenswrapper[4907]: E0313 15:49:58.402137 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eaec5907-c808-4ac2-b131-608b228885e6" containerName="sg-core" Mar 13 15:49:58 crc kubenswrapper[4907]: I0313 15:49:58.402146 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="eaec5907-c808-4ac2-b131-608b228885e6" containerName="sg-core" Mar 13 15:49:58 crc kubenswrapper[4907]: I0313 15:49:58.402430 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="eaec5907-c808-4ac2-b131-608b228885e6" containerName="sg-core" Mar 13 15:49:58 crc kubenswrapper[4907]: I0313 15:49:58.402455 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="eaec5907-c808-4ac2-b131-608b228885e6" containerName="ceilometer-central-agent" Mar 13 15:49:58 crc kubenswrapper[4907]: I0313 15:49:58.402478 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="eaec5907-c808-4ac2-b131-608b228885e6" containerName="ceilometer-notification-agent" Mar 13 15:49:58 crc kubenswrapper[4907]: I0313 15:49:58.402491 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="eaec5907-c808-4ac2-b131-608b228885e6" containerName="proxy-httpd" Mar 13 15:49:58 crc kubenswrapper[4907]: I0313 15:49:58.405135 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 15:49:58 crc kubenswrapper[4907]: I0313 15:49:58.406865 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 13 15:49:58 crc kubenswrapper[4907]: I0313 15:49:58.408092 4907 scope.go:117] "RemoveContainer" containerID="f8bccaf5228a1db9069e313bf3466a6f8cbd61fae45f55bda0a236da15bba7f0" Mar 13 15:49:58 crc kubenswrapper[4907]: I0313 15:49:58.408263 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 13 15:49:58 crc kubenswrapper[4907]: I0313 15:49:58.408324 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 13 15:49:58 crc kubenswrapper[4907]: I0313 15:49:58.471221 4907 scope.go:117] "RemoveContainer" containerID="b3ab302efa65117dbfc68c5b768e1ee7a624666e299637f13e417c20da84fcf0" Mar 13 15:49:58 crc kubenswrapper[4907]: I0313 15:49:58.536211 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14aa8b3a-f849-42c4-9248-2329fd28c6c3-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"14aa8b3a-f849-42c4-9248-2329fd28c6c3\") " pod="openstack/ceilometer-0" Mar 13 15:49:58 crc kubenswrapper[4907]: I0313 15:49:58.536261 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/14aa8b3a-f849-42c4-9248-2329fd28c6c3-run-httpd\") pod \"ceilometer-0\" (UID: \"14aa8b3a-f849-42c4-9248-2329fd28c6c3\") " pod="openstack/ceilometer-0" Mar 13 15:49:58 crc kubenswrapper[4907]: I0313 15:49:58.536284 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/14aa8b3a-f849-42c4-9248-2329fd28c6c3-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"14aa8b3a-f849-42c4-9248-2329fd28c6c3\") " pod="openstack/ceilometer-0" Mar 13 15:49:58 crc kubenswrapper[4907]: I0313 15:49:58.536378 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ttn2w\" (UniqueName: \"kubernetes.io/projected/14aa8b3a-f849-42c4-9248-2329fd28c6c3-kube-api-access-ttn2w\") pod \"ceilometer-0\" (UID: \"14aa8b3a-f849-42c4-9248-2329fd28c6c3\") " pod="openstack/ceilometer-0" Mar 13 15:49:58 crc kubenswrapper[4907]: I0313 15:49:58.536415 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/14aa8b3a-f849-42c4-9248-2329fd28c6c3-log-httpd\") pod \"ceilometer-0\" (UID: \"14aa8b3a-f849-42c4-9248-2329fd28c6c3\") " pod="openstack/ceilometer-0" Mar 13 15:49:58 crc kubenswrapper[4907]: I0313 15:49:58.536449 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/14aa8b3a-f849-42c4-9248-2329fd28c6c3-scripts\") pod \"ceilometer-0\" (UID: \"14aa8b3a-f849-42c4-9248-2329fd28c6c3\") " pod="openstack/ceilometer-0" Mar 13 15:49:58 crc kubenswrapper[4907]: I0313 15:49:58.536508 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14aa8b3a-f849-42c4-9248-2329fd28c6c3-config-data\") pod \"ceilometer-0\" (UID: \"14aa8b3a-f849-42c4-9248-2329fd28c6c3\") " pod="openstack/ceilometer-0" Mar 13 15:49:58 crc kubenswrapper[4907]: I0313 15:49:58.638099 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14aa8b3a-f849-42c4-9248-2329fd28c6c3-config-data\") pod \"ceilometer-0\" (UID: \"14aa8b3a-f849-42c4-9248-2329fd28c6c3\") " pod="openstack/ceilometer-0" Mar 13 15:49:58 crc kubenswrapper[4907]: I0313 15:49:58.638235 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14aa8b3a-f849-42c4-9248-2329fd28c6c3-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"14aa8b3a-f849-42c4-9248-2329fd28c6c3\") " pod="openstack/ceilometer-0" Mar 13 15:49:58 crc kubenswrapper[4907]: I0313 15:49:58.638265 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/14aa8b3a-f849-42c4-9248-2329fd28c6c3-run-httpd\") pod \"ceilometer-0\" (UID: \"14aa8b3a-f849-42c4-9248-2329fd28c6c3\") " pod="openstack/ceilometer-0" Mar 13 15:49:58 crc kubenswrapper[4907]: I0313 15:49:58.638292 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/14aa8b3a-f849-42c4-9248-2329fd28c6c3-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"14aa8b3a-f849-42c4-9248-2329fd28c6c3\") " pod="openstack/ceilometer-0" Mar 13 15:49:58 crc kubenswrapper[4907]: I0313 15:49:58.638356 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ttn2w\" (UniqueName: \"kubernetes.io/projected/14aa8b3a-f849-42c4-9248-2329fd28c6c3-kube-api-access-ttn2w\") pod \"ceilometer-0\" (UID: \"14aa8b3a-f849-42c4-9248-2329fd28c6c3\") " pod="openstack/ceilometer-0" Mar 13 15:49:58 crc kubenswrapper[4907]: I0313 15:49:58.638389 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/14aa8b3a-f849-42c4-9248-2329fd28c6c3-log-httpd\") pod \"ceilometer-0\" (UID: \"14aa8b3a-f849-42c4-9248-2329fd28c6c3\") " pod="openstack/ceilometer-0" Mar 13 15:49:58 crc kubenswrapper[4907]: I0313 15:49:58.638436 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/14aa8b3a-f849-42c4-9248-2329fd28c6c3-scripts\") pod \"ceilometer-0\" (UID: \"14aa8b3a-f849-42c4-9248-2329fd28c6c3\") " pod="openstack/ceilometer-0" Mar 13 15:49:58 crc kubenswrapper[4907]: I0313 15:49:58.638971 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/14aa8b3a-f849-42c4-9248-2329fd28c6c3-log-httpd\") pod \"ceilometer-0\" (UID: \"14aa8b3a-f849-42c4-9248-2329fd28c6c3\") " pod="openstack/ceilometer-0" Mar 13 15:49:58 crc kubenswrapper[4907]: I0313 15:49:58.639008 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/14aa8b3a-f849-42c4-9248-2329fd28c6c3-run-httpd\") pod \"ceilometer-0\" (UID: \"14aa8b3a-f849-42c4-9248-2329fd28c6c3\") " pod="openstack/ceilometer-0" Mar 13 15:49:58 crc kubenswrapper[4907]: I0313 15:49:58.645975 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/14aa8b3a-f849-42c4-9248-2329fd28c6c3-scripts\") pod \"ceilometer-0\" (UID: \"14aa8b3a-f849-42c4-9248-2329fd28c6c3\") " pod="openstack/ceilometer-0" Mar 13 15:49:58 crc kubenswrapper[4907]: I0313 15:49:58.647536 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14aa8b3a-f849-42c4-9248-2329fd28c6c3-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"14aa8b3a-f849-42c4-9248-2329fd28c6c3\") " pod="openstack/ceilometer-0" Mar 13 15:49:58 crc kubenswrapper[4907]: I0313 15:49:58.648986 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14aa8b3a-f849-42c4-9248-2329fd28c6c3-config-data\") pod \"ceilometer-0\" (UID: \"14aa8b3a-f849-42c4-9248-2329fd28c6c3\") " pod="openstack/ceilometer-0" Mar 13 15:49:58 crc kubenswrapper[4907]: I0313 15:49:58.654493 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/14aa8b3a-f849-42c4-9248-2329fd28c6c3-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"14aa8b3a-f849-42c4-9248-2329fd28c6c3\") " pod="openstack/ceilometer-0" Mar 13 15:49:58 crc kubenswrapper[4907]: I0313 15:49:58.658523 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ttn2w\" (UniqueName: \"kubernetes.io/projected/14aa8b3a-f849-42c4-9248-2329fd28c6c3-kube-api-access-ttn2w\") pod \"ceilometer-0\" (UID: \"14aa8b3a-f849-42c4-9248-2329fd28c6c3\") " pod="openstack/ceilometer-0" Mar 13 15:49:58 crc kubenswrapper[4907]: I0313 15:49:58.728916 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 15:49:59 crc kubenswrapper[4907]: I0313 15:49:59.360716 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 13 15:49:59 crc kubenswrapper[4907]: I0313 15:49:59.794002 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eaec5907-c808-4ac2-b131-608b228885e6" path="/var/lib/kubelet/pods/eaec5907-c808-4ac2-b131-608b228885e6/volumes" Mar 13 15:50:00 crc kubenswrapper[4907]: I0313 15:50:00.134494 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556950-pqwgk"] Mar 13 15:50:00 crc kubenswrapper[4907]: I0313 15:50:00.136372 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556950-pqwgk" Mar 13 15:50:00 crc kubenswrapper[4907]: I0313 15:50:00.139218 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 15:50:00 crc kubenswrapper[4907]: I0313 15:50:00.139490 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 15:50:00 crc kubenswrapper[4907]: I0313 15:50:00.144362 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 15:50:00 crc kubenswrapper[4907]: I0313 15:50:00.152696 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556950-pqwgk"] Mar 13 15:50:00 crc kubenswrapper[4907]: I0313 15:50:00.172301 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z6m5k\" (UniqueName: \"kubernetes.io/projected/63d124ed-9ee2-437b-8ebd-b38a53e499af-kube-api-access-z6m5k\") pod \"auto-csr-approver-29556950-pqwgk\" (UID: \"63d124ed-9ee2-437b-8ebd-b38a53e499af\") " pod="openshift-infra/auto-csr-approver-29556950-pqwgk" Mar 13 15:50:00 crc kubenswrapper[4907]: I0313 15:50:00.274612 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z6m5k\" (UniqueName: \"kubernetes.io/projected/63d124ed-9ee2-437b-8ebd-b38a53e499af-kube-api-access-z6m5k\") pod \"auto-csr-approver-29556950-pqwgk\" (UID: \"63d124ed-9ee2-437b-8ebd-b38a53e499af\") " pod="openshift-infra/auto-csr-approver-29556950-pqwgk" Mar 13 15:50:00 crc kubenswrapper[4907]: I0313 15:50:00.299062 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z6m5k\" (UniqueName: \"kubernetes.io/projected/63d124ed-9ee2-437b-8ebd-b38a53e499af-kube-api-access-z6m5k\") pod \"auto-csr-approver-29556950-pqwgk\" (UID: \"63d124ed-9ee2-437b-8ebd-b38a53e499af\") " pod="openshift-infra/auto-csr-approver-29556950-pqwgk" Mar 13 15:50:00 crc kubenswrapper[4907]: I0313 15:50:00.357205 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"8ce53a96-2566-4d5c-b4cd-bbfd2516975a","Type":"ContainerStarted","Data":"179402b0df76b90aa1275176df179af7bc2ec062262c7c78fb13ace6f27c6838"} Mar 13 15:50:00 crc kubenswrapper[4907]: I0313 15:50:00.363071 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"14aa8b3a-f849-42c4-9248-2329fd28c6c3","Type":"ContainerStarted","Data":"327a65f960da4ed89404c82b3f62aa43c7df17d2c7fdfd45430eda4484d631d9"} Mar 13 15:50:00 crc kubenswrapper[4907]: I0313 15:50:00.363106 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"14aa8b3a-f849-42c4-9248-2329fd28c6c3","Type":"ContainerStarted","Data":"9ba05f967433eff9725a6dcc1ca7f6cab5e912da0e5ca7e888033a6d0b137522"} Mar 13 15:50:00 crc kubenswrapper[4907]: I0313 15:50:00.474203 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556950-pqwgk" Mar 13 15:50:00 crc kubenswrapper[4907]: I0313 15:50:00.975074 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556950-pqwgk"] Mar 13 15:50:01 crc kubenswrapper[4907]: I0313 15:50:01.096607 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-jj9xr" Mar 13 15:50:01 crc kubenswrapper[4907]: I0313 15:50:01.096992 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-jj9xr" Mar 13 15:50:01 crc kubenswrapper[4907]: I0313 15:50:01.161868 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-jj9xr" Mar 13 15:50:01 crc kubenswrapper[4907]: I0313 15:50:01.377543 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556950-pqwgk" event={"ID":"63d124ed-9ee2-437b-8ebd-b38a53e499af","Type":"ContainerStarted","Data":"df5d57d55555efa41307fa8944531c055731a32dc5db86d222d79f540686b70e"} Mar 13 15:50:01 crc kubenswrapper[4907]: I0313 15:50:01.379489 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"14aa8b3a-f849-42c4-9248-2329fd28c6c3","Type":"ContainerStarted","Data":"34cd3ac68b3f061e9ab0b1f70d520ca0eebcf6f1d8afd0348c0574488177a64d"} Mar 13 15:50:02 crc kubenswrapper[4907]: I0313 15:50:02.392573 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556950-pqwgk" event={"ID":"63d124ed-9ee2-437b-8ebd-b38a53e499af","Type":"ContainerStarted","Data":"fa4769654b9ebff08198d548ba01872a5dbbbcb75936088b49f7d253d89a76c0"} Mar 13 15:50:02 crc kubenswrapper[4907]: I0313 15:50:02.398194 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"8ce53a96-2566-4d5c-b4cd-bbfd2516975a","Type":"ContainerStarted","Data":"75b45c6cebfb5cd8199369e1018195ba07a407a3aa02a55e64b34a452c02f794"} Mar 13 15:50:02 crc kubenswrapper[4907]: I0313 15:50:02.400412 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"14aa8b3a-f849-42c4-9248-2329fd28c6c3","Type":"ContainerStarted","Data":"d5a88d5e47093bf38e94a5d416f23f1547843a36e587d1615bb482fff4d0a092"} Mar 13 15:50:02 crc kubenswrapper[4907]: I0313 15:50:02.418656 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29556950-pqwgk" podStartSLOduration=1.423557527 podStartE2EDuration="2.418640263s" podCreationTimestamp="2026-03-13 15:50:00 +0000 UTC" firstStartedPulling="2026-03-13 15:50:01.015037555 +0000 UTC m=+6299.914825244" lastFinishedPulling="2026-03-13 15:50:02.010120291 +0000 UTC m=+6300.909907980" observedRunningTime="2026-03-13 15:50:02.41742792 +0000 UTC m=+6301.317215609" watchObservedRunningTime="2026-03-13 15:50:02.418640263 +0000 UTC m=+6301.318427972" Mar 13 15:50:04 crc kubenswrapper[4907]: I0313 15:50:04.426790 4907 generic.go:334] "Generic (PLEG): container finished" podID="63d124ed-9ee2-437b-8ebd-b38a53e499af" containerID="fa4769654b9ebff08198d548ba01872a5dbbbcb75936088b49f7d253d89a76c0" exitCode=0 Mar 13 15:50:04 crc kubenswrapper[4907]: I0313 15:50:04.426890 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556950-pqwgk" event={"ID":"63d124ed-9ee2-437b-8ebd-b38a53e499af","Type":"ContainerDied","Data":"fa4769654b9ebff08198d548ba01872a5dbbbcb75936088b49f7d253d89a76c0"} Mar 13 15:50:04 crc kubenswrapper[4907]: I0313 15:50:04.444802 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-0" podStartSLOduration=3.4641411570000002 podStartE2EDuration="10.44478282s" podCreationTimestamp="2026-03-13 15:49:54 +0000 UTC" firstStartedPulling="2026-03-13 15:49:54.958841099 +0000 UTC m=+6293.858628788" lastFinishedPulling="2026-03-13 15:50:01.939482762 +0000 UTC m=+6300.839270451" observedRunningTime="2026-03-13 15:50:02.443082941 +0000 UTC m=+6301.342870630" watchObservedRunningTime="2026-03-13 15:50:04.44478282 +0000 UTC m=+6303.344570509" Mar 13 15:50:05 crc kubenswrapper[4907]: I0313 15:50:05.035721 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-mwt47"] Mar 13 15:50:05 crc kubenswrapper[4907]: I0313 15:50:05.045297 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-mwt47"] Mar 13 15:50:05 crc kubenswrapper[4907]: I0313 15:50:05.438405 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"14aa8b3a-f849-42c4-9248-2329fd28c6c3","Type":"ContainerStarted","Data":"ea745f91341a94f03f73269b8a99b71eff9cf760adb53889a4160dbc700df09a"} Mar 13 15:50:05 crc kubenswrapper[4907]: I0313 15:50:05.438903 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Mar 13 15:50:05 crc kubenswrapper[4907]: I0313 15:50:05.464221 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.584673066 podStartE2EDuration="7.464197563s" podCreationTimestamp="2026-03-13 15:49:58 +0000 UTC" firstStartedPulling="2026-03-13 15:49:59.38505795 +0000 UTC m=+6298.284845639" lastFinishedPulling="2026-03-13 15:50:04.264582457 +0000 UTC m=+6303.164370136" observedRunningTime="2026-03-13 15:50:05.462577018 +0000 UTC m=+6304.362364717" watchObservedRunningTime="2026-03-13 15:50:05.464197563 +0000 UTC m=+6304.363985252" Mar 13 15:50:05 crc kubenswrapper[4907]: I0313 15:50:05.800787 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0336c3c0-cc18-4bb3-afe3-7e2a0232c889" path="/var/lib/kubelet/pods/0336c3c0-cc18-4bb3-afe3-7e2a0232c889/volumes" Mar 13 15:50:05 crc kubenswrapper[4907]: I0313 15:50:05.954159 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556950-pqwgk" Mar 13 15:50:06 crc kubenswrapper[4907]: I0313 15:50:06.009953 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z6m5k\" (UniqueName: \"kubernetes.io/projected/63d124ed-9ee2-437b-8ebd-b38a53e499af-kube-api-access-z6m5k\") pod \"63d124ed-9ee2-437b-8ebd-b38a53e499af\" (UID: \"63d124ed-9ee2-437b-8ebd-b38a53e499af\") " Mar 13 15:50:06 crc kubenswrapper[4907]: I0313 15:50:06.018262 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/63d124ed-9ee2-437b-8ebd-b38a53e499af-kube-api-access-z6m5k" (OuterVolumeSpecName: "kube-api-access-z6m5k") pod "63d124ed-9ee2-437b-8ebd-b38a53e499af" (UID: "63d124ed-9ee2-437b-8ebd-b38a53e499af"). InnerVolumeSpecName "kube-api-access-z6m5k". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:50:06 crc kubenswrapper[4907]: I0313 15:50:06.112427 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z6m5k\" (UniqueName: \"kubernetes.io/projected/63d124ed-9ee2-437b-8ebd-b38a53e499af-kube-api-access-z6m5k\") on node \"crc\" DevicePath \"\"" Mar 13 15:50:06 crc kubenswrapper[4907]: I0313 15:50:06.453141 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556950-pqwgk" Mar 13 15:50:06 crc kubenswrapper[4907]: I0313 15:50:06.453220 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556950-pqwgk" event={"ID":"63d124ed-9ee2-437b-8ebd-b38a53e499af","Type":"ContainerDied","Data":"df5d57d55555efa41307fa8944531c055731a32dc5db86d222d79f540686b70e"} Mar 13 15:50:06 crc kubenswrapper[4907]: I0313 15:50:06.453775 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="df5d57d55555efa41307fa8944531c055731a32dc5db86d222d79f540686b70e" Mar 13 15:50:06 crc kubenswrapper[4907]: I0313 15:50:06.503310 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556944-gcp6x"] Mar 13 15:50:06 crc kubenswrapper[4907]: I0313 15:50:06.512717 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556944-gcp6x"] Mar 13 15:50:07 crc kubenswrapper[4907]: I0313 15:50:07.797923 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="04ca577b-ab8d-44a6-a598-1297bd5ab664" path="/var/lib/kubelet/pods/04ca577b-ab8d-44a6-a598-1297bd5ab664/volumes" Mar 13 15:50:08 crc kubenswrapper[4907]: I0313 15:50:08.977660 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-db-create-mzct7"] Mar 13 15:50:08 crc kubenswrapper[4907]: E0313 15:50:08.978296 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63d124ed-9ee2-437b-8ebd-b38a53e499af" containerName="oc" Mar 13 15:50:08 crc kubenswrapper[4907]: I0313 15:50:08.978316 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="63d124ed-9ee2-437b-8ebd-b38a53e499af" containerName="oc" Mar 13 15:50:08 crc kubenswrapper[4907]: I0313 15:50:08.978564 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="63d124ed-9ee2-437b-8ebd-b38a53e499af" containerName="oc" Mar 13 15:50:08 crc kubenswrapper[4907]: I0313 15:50:08.979521 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-mzct7" Mar 13 15:50:08 crc kubenswrapper[4907]: I0313 15:50:08.987618 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-create-mzct7"] Mar 13 15:50:09 crc kubenswrapper[4907]: I0313 15:50:09.075526 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2bvvd\" (UniqueName: \"kubernetes.io/projected/b2a34745-a555-4029-a203-ee17762929e2-kube-api-access-2bvvd\") pod \"manila-db-create-mzct7\" (UID: \"b2a34745-a555-4029-a203-ee17762929e2\") " pod="openstack/manila-db-create-mzct7" Mar 13 15:50:09 crc kubenswrapper[4907]: I0313 15:50:09.075800 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b2a34745-a555-4029-a203-ee17762929e2-operator-scripts\") pod \"manila-db-create-mzct7\" (UID: \"b2a34745-a555-4029-a203-ee17762929e2\") " pod="openstack/manila-db-create-mzct7" Mar 13 15:50:09 crc kubenswrapper[4907]: I0313 15:50:09.178370 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2bvvd\" (UniqueName: \"kubernetes.io/projected/b2a34745-a555-4029-a203-ee17762929e2-kube-api-access-2bvvd\") pod \"manila-db-create-mzct7\" (UID: \"b2a34745-a555-4029-a203-ee17762929e2\") " pod="openstack/manila-db-create-mzct7" Mar 13 15:50:09 crc kubenswrapper[4907]: I0313 15:50:09.178541 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b2a34745-a555-4029-a203-ee17762929e2-operator-scripts\") pod \"manila-db-create-mzct7\" (UID: \"b2a34745-a555-4029-a203-ee17762929e2\") " pod="openstack/manila-db-create-mzct7" Mar 13 15:50:09 crc kubenswrapper[4907]: I0313 15:50:09.179356 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b2a34745-a555-4029-a203-ee17762929e2-operator-scripts\") pod \"manila-db-create-mzct7\" (UID: \"b2a34745-a555-4029-a203-ee17762929e2\") " pod="openstack/manila-db-create-mzct7" Mar 13 15:50:09 crc kubenswrapper[4907]: I0313 15:50:09.184615 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-e749-account-create-update-gsxrn"] Mar 13 15:50:09 crc kubenswrapper[4907]: I0313 15:50:09.185984 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-e749-account-create-update-gsxrn" Mar 13 15:50:09 crc kubenswrapper[4907]: I0313 15:50:09.188717 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-db-secret" Mar 13 15:50:09 crc kubenswrapper[4907]: I0313 15:50:09.200639 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2bvvd\" (UniqueName: \"kubernetes.io/projected/b2a34745-a555-4029-a203-ee17762929e2-kube-api-access-2bvvd\") pod \"manila-db-create-mzct7\" (UID: \"b2a34745-a555-4029-a203-ee17762929e2\") " pod="openstack/manila-db-create-mzct7" Mar 13 15:50:09 crc kubenswrapper[4907]: I0313 15:50:09.203059 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-e749-account-create-update-gsxrn"] Mar 13 15:50:09 crc kubenswrapper[4907]: I0313 15:50:09.284342 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xw8m6\" (UniqueName: \"kubernetes.io/projected/0498e113-86c3-4718-a580-13bfdb277ff2-kube-api-access-xw8m6\") pod \"manila-e749-account-create-update-gsxrn\" (UID: \"0498e113-86c3-4718-a580-13bfdb277ff2\") " pod="openstack/manila-e749-account-create-update-gsxrn" Mar 13 15:50:09 crc kubenswrapper[4907]: I0313 15:50:09.284482 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0498e113-86c3-4718-a580-13bfdb277ff2-operator-scripts\") pod \"manila-e749-account-create-update-gsxrn\" (UID: \"0498e113-86c3-4718-a580-13bfdb277ff2\") " pod="openstack/manila-e749-account-create-update-gsxrn" Mar 13 15:50:09 crc kubenswrapper[4907]: I0313 15:50:09.301445 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-mzct7" Mar 13 15:50:09 crc kubenswrapper[4907]: I0313 15:50:09.386212 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0498e113-86c3-4718-a580-13bfdb277ff2-operator-scripts\") pod \"manila-e749-account-create-update-gsxrn\" (UID: \"0498e113-86c3-4718-a580-13bfdb277ff2\") " pod="openstack/manila-e749-account-create-update-gsxrn" Mar 13 15:50:09 crc kubenswrapper[4907]: I0313 15:50:09.386630 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xw8m6\" (UniqueName: \"kubernetes.io/projected/0498e113-86c3-4718-a580-13bfdb277ff2-kube-api-access-xw8m6\") pod \"manila-e749-account-create-update-gsxrn\" (UID: \"0498e113-86c3-4718-a580-13bfdb277ff2\") " pod="openstack/manila-e749-account-create-update-gsxrn" Mar 13 15:50:09 crc kubenswrapper[4907]: I0313 15:50:09.386990 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0498e113-86c3-4718-a580-13bfdb277ff2-operator-scripts\") pod \"manila-e749-account-create-update-gsxrn\" (UID: \"0498e113-86c3-4718-a580-13bfdb277ff2\") " pod="openstack/manila-e749-account-create-update-gsxrn" Mar 13 15:50:09 crc kubenswrapper[4907]: I0313 15:50:09.406471 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xw8m6\" (UniqueName: \"kubernetes.io/projected/0498e113-86c3-4718-a580-13bfdb277ff2-kube-api-access-xw8m6\") pod \"manila-e749-account-create-update-gsxrn\" (UID: \"0498e113-86c3-4718-a580-13bfdb277ff2\") " pod="openstack/manila-e749-account-create-update-gsxrn" Mar 13 15:50:09 crc kubenswrapper[4907]: I0313 15:50:09.568283 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-e749-account-create-update-gsxrn" Mar 13 15:50:09 crc kubenswrapper[4907]: I0313 15:50:09.867464 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-create-mzct7"] Mar 13 15:50:10 crc kubenswrapper[4907]: W0313 15:50:10.082485 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0498e113_86c3_4718_a580_13bfdb277ff2.slice/crio-89c63ae209e5b88e5155699498019037c249d937a18a0b93309137427255d070 WatchSource:0}: Error finding container 89c63ae209e5b88e5155699498019037c249d937a18a0b93309137427255d070: Status 404 returned error can't find the container with id 89c63ae209e5b88e5155699498019037c249d937a18a0b93309137427255d070 Mar 13 15:50:10 crc kubenswrapper[4907]: I0313 15:50:10.084506 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-e749-account-create-update-gsxrn"] Mar 13 15:50:10 crc kubenswrapper[4907]: I0313 15:50:10.507688 4907 generic.go:334] "Generic (PLEG): container finished" podID="b2a34745-a555-4029-a203-ee17762929e2" containerID="7fe743a125369b9e5af6f2cb7a4bc8d2add93450e84c6dadc626912538eac565" exitCode=0 Mar 13 15:50:10 crc kubenswrapper[4907]: I0313 15:50:10.507752 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-mzct7" event={"ID":"b2a34745-a555-4029-a203-ee17762929e2","Type":"ContainerDied","Data":"7fe743a125369b9e5af6f2cb7a4bc8d2add93450e84c6dadc626912538eac565"} Mar 13 15:50:10 crc kubenswrapper[4907]: I0313 15:50:10.508101 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-mzct7" event={"ID":"b2a34745-a555-4029-a203-ee17762929e2","Type":"ContainerStarted","Data":"1b3b4efdf5f0221b588e062f2e104134c1f7d2650739a9addb8c08e1e1a79b65"} Mar 13 15:50:10 crc kubenswrapper[4907]: I0313 15:50:10.509935 4907 generic.go:334] "Generic (PLEG): container finished" podID="0498e113-86c3-4718-a580-13bfdb277ff2" containerID="3da3591c8cb3c5a70194cd0e1a361e544712615fd8402b78ddd2f90160064c36" exitCode=0 Mar 13 15:50:10 crc kubenswrapper[4907]: I0313 15:50:10.509986 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-e749-account-create-update-gsxrn" event={"ID":"0498e113-86c3-4718-a580-13bfdb277ff2","Type":"ContainerDied","Data":"3da3591c8cb3c5a70194cd0e1a361e544712615fd8402b78ddd2f90160064c36"} Mar 13 15:50:10 crc kubenswrapper[4907]: I0313 15:50:10.510013 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-e749-account-create-update-gsxrn" event={"ID":"0498e113-86c3-4718-a580-13bfdb277ff2","Type":"ContainerStarted","Data":"89c63ae209e5b88e5155699498019037c249d937a18a0b93309137427255d070"} Mar 13 15:50:11 crc kubenswrapper[4907]: I0313 15:50:11.153568 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-jj9xr" Mar 13 15:50:11 crc kubenswrapper[4907]: I0313 15:50:11.208708 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-jj9xr"] Mar 13 15:50:11 crc kubenswrapper[4907]: I0313 15:50:11.518576 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-jj9xr" podUID="951bf99b-9f5e-4609-bff0-0c594f6a5b2e" containerName="registry-server" containerID="cri-o://ad22a18bbef271426a4077aab78c5dbcaeed67525c562d753181de83cefb8f37" gracePeriod=2 Mar 13 15:50:12 crc kubenswrapper[4907]: I0313 15:50:12.092661 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-e749-account-create-update-gsxrn" Mar 13 15:50:12 crc kubenswrapper[4907]: I0313 15:50:12.147814 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0498e113-86c3-4718-a580-13bfdb277ff2-operator-scripts\") pod \"0498e113-86c3-4718-a580-13bfdb277ff2\" (UID: \"0498e113-86c3-4718-a580-13bfdb277ff2\") " Mar 13 15:50:12 crc kubenswrapper[4907]: I0313 15:50:12.147952 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xw8m6\" (UniqueName: \"kubernetes.io/projected/0498e113-86c3-4718-a580-13bfdb277ff2-kube-api-access-xw8m6\") pod \"0498e113-86c3-4718-a580-13bfdb277ff2\" (UID: \"0498e113-86c3-4718-a580-13bfdb277ff2\") " Mar 13 15:50:12 crc kubenswrapper[4907]: I0313 15:50:12.149310 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0498e113-86c3-4718-a580-13bfdb277ff2-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "0498e113-86c3-4718-a580-13bfdb277ff2" (UID: "0498e113-86c3-4718-a580-13bfdb277ff2"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:50:12 crc kubenswrapper[4907]: I0313 15:50:12.155080 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0498e113-86c3-4718-a580-13bfdb277ff2-kube-api-access-xw8m6" (OuterVolumeSpecName: "kube-api-access-xw8m6") pod "0498e113-86c3-4718-a580-13bfdb277ff2" (UID: "0498e113-86c3-4718-a580-13bfdb277ff2"). InnerVolumeSpecName "kube-api-access-xw8m6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:50:12 crc kubenswrapper[4907]: I0313 15:50:12.244818 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jj9xr" Mar 13 15:50:12 crc kubenswrapper[4907]: I0313 15:50:12.247373 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-mzct7" Mar 13 15:50:12 crc kubenswrapper[4907]: I0313 15:50:12.251869 4907 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0498e113-86c3-4718-a580-13bfdb277ff2-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 15:50:12 crc kubenswrapper[4907]: I0313 15:50:12.252040 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xw8m6\" (UniqueName: \"kubernetes.io/projected/0498e113-86c3-4718-a580-13bfdb277ff2-kube-api-access-xw8m6\") on node \"crc\" DevicePath \"\"" Mar 13 15:50:12 crc kubenswrapper[4907]: I0313 15:50:12.353145 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b2a34745-a555-4029-a203-ee17762929e2-operator-scripts\") pod \"b2a34745-a555-4029-a203-ee17762929e2\" (UID: \"b2a34745-a555-4029-a203-ee17762929e2\") " Mar 13 15:50:12 crc kubenswrapper[4907]: I0313 15:50:12.353185 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2bvvd\" (UniqueName: \"kubernetes.io/projected/b2a34745-a555-4029-a203-ee17762929e2-kube-api-access-2bvvd\") pod \"b2a34745-a555-4029-a203-ee17762929e2\" (UID: \"b2a34745-a555-4029-a203-ee17762929e2\") " Mar 13 15:50:12 crc kubenswrapper[4907]: I0313 15:50:12.353230 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/951bf99b-9f5e-4609-bff0-0c594f6a5b2e-utilities\") pod \"951bf99b-9f5e-4609-bff0-0c594f6a5b2e\" (UID: \"951bf99b-9f5e-4609-bff0-0c594f6a5b2e\") " Mar 13 15:50:12 crc kubenswrapper[4907]: I0313 15:50:12.353518 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/951bf99b-9f5e-4609-bff0-0c594f6a5b2e-catalog-content\") pod \"951bf99b-9f5e-4609-bff0-0c594f6a5b2e\" (UID: \"951bf99b-9f5e-4609-bff0-0c594f6a5b2e\") " Mar 13 15:50:12 crc kubenswrapper[4907]: I0313 15:50:12.353544 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h9rf2\" (UniqueName: \"kubernetes.io/projected/951bf99b-9f5e-4609-bff0-0c594f6a5b2e-kube-api-access-h9rf2\") pod \"951bf99b-9f5e-4609-bff0-0c594f6a5b2e\" (UID: \"951bf99b-9f5e-4609-bff0-0c594f6a5b2e\") " Mar 13 15:50:12 crc kubenswrapper[4907]: I0313 15:50:12.354393 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b2a34745-a555-4029-a203-ee17762929e2-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b2a34745-a555-4029-a203-ee17762929e2" (UID: "b2a34745-a555-4029-a203-ee17762929e2"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:50:12 crc kubenswrapper[4907]: I0313 15:50:12.354534 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/951bf99b-9f5e-4609-bff0-0c594f6a5b2e-utilities" (OuterVolumeSpecName: "utilities") pod "951bf99b-9f5e-4609-bff0-0c594f6a5b2e" (UID: "951bf99b-9f5e-4609-bff0-0c594f6a5b2e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:50:12 crc kubenswrapper[4907]: I0313 15:50:12.357627 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/951bf99b-9f5e-4609-bff0-0c594f6a5b2e-kube-api-access-h9rf2" (OuterVolumeSpecName: "kube-api-access-h9rf2") pod "951bf99b-9f5e-4609-bff0-0c594f6a5b2e" (UID: "951bf99b-9f5e-4609-bff0-0c594f6a5b2e"). InnerVolumeSpecName "kube-api-access-h9rf2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:50:12 crc kubenswrapper[4907]: I0313 15:50:12.359130 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b2a34745-a555-4029-a203-ee17762929e2-kube-api-access-2bvvd" (OuterVolumeSpecName: "kube-api-access-2bvvd") pod "b2a34745-a555-4029-a203-ee17762929e2" (UID: "b2a34745-a555-4029-a203-ee17762929e2"). InnerVolumeSpecName "kube-api-access-2bvvd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:50:12 crc kubenswrapper[4907]: I0313 15:50:12.405413 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/951bf99b-9f5e-4609-bff0-0c594f6a5b2e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "951bf99b-9f5e-4609-bff0-0c594f6a5b2e" (UID: "951bf99b-9f5e-4609-bff0-0c594f6a5b2e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:50:12 crc kubenswrapper[4907]: I0313 15:50:12.455895 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/951bf99b-9f5e-4609-bff0-0c594f6a5b2e-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 15:50:12 crc kubenswrapper[4907]: I0313 15:50:12.455926 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/951bf99b-9f5e-4609-bff0-0c594f6a5b2e-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 15:50:12 crc kubenswrapper[4907]: I0313 15:50:12.455937 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h9rf2\" (UniqueName: \"kubernetes.io/projected/951bf99b-9f5e-4609-bff0-0c594f6a5b2e-kube-api-access-h9rf2\") on node \"crc\" DevicePath \"\"" Mar 13 15:50:12 crc kubenswrapper[4907]: I0313 15:50:12.455946 4907 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b2a34745-a555-4029-a203-ee17762929e2-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 15:50:12 crc kubenswrapper[4907]: I0313 15:50:12.455955 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2bvvd\" (UniqueName: \"kubernetes.io/projected/b2a34745-a555-4029-a203-ee17762929e2-kube-api-access-2bvvd\") on node \"crc\" DevicePath \"\"" Mar 13 15:50:12 crc kubenswrapper[4907]: I0313 15:50:12.539183 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-mzct7" Mar 13 15:50:12 crc kubenswrapper[4907]: I0313 15:50:12.539235 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-mzct7" event={"ID":"b2a34745-a555-4029-a203-ee17762929e2","Type":"ContainerDied","Data":"1b3b4efdf5f0221b588e062f2e104134c1f7d2650739a9addb8c08e1e1a79b65"} Mar 13 15:50:12 crc kubenswrapper[4907]: I0313 15:50:12.539630 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1b3b4efdf5f0221b588e062f2e104134c1f7d2650739a9addb8c08e1e1a79b65" Mar 13 15:50:12 crc kubenswrapper[4907]: I0313 15:50:12.542567 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-e749-account-create-update-gsxrn" Mar 13 15:50:12 crc kubenswrapper[4907]: I0313 15:50:12.542607 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-e749-account-create-update-gsxrn" event={"ID":"0498e113-86c3-4718-a580-13bfdb277ff2","Type":"ContainerDied","Data":"89c63ae209e5b88e5155699498019037c249d937a18a0b93309137427255d070"} Mar 13 15:50:12 crc kubenswrapper[4907]: I0313 15:50:12.542659 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="89c63ae209e5b88e5155699498019037c249d937a18a0b93309137427255d070" Mar 13 15:50:12 crc kubenswrapper[4907]: I0313 15:50:12.545444 4907 generic.go:334] "Generic (PLEG): container finished" podID="951bf99b-9f5e-4609-bff0-0c594f6a5b2e" containerID="ad22a18bbef271426a4077aab78c5dbcaeed67525c562d753181de83cefb8f37" exitCode=0 Mar 13 15:50:12 crc kubenswrapper[4907]: I0313 15:50:12.545490 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jj9xr" event={"ID":"951bf99b-9f5e-4609-bff0-0c594f6a5b2e","Type":"ContainerDied","Data":"ad22a18bbef271426a4077aab78c5dbcaeed67525c562d753181de83cefb8f37"} Mar 13 15:50:12 crc kubenswrapper[4907]: I0313 15:50:12.545521 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jj9xr" event={"ID":"951bf99b-9f5e-4609-bff0-0c594f6a5b2e","Type":"ContainerDied","Data":"02bff9741d914eb2d1510a0ab7003a1cb31cbd80634618a50cf32118561743ca"} Mar 13 15:50:12 crc kubenswrapper[4907]: I0313 15:50:12.545540 4907 scope.go:117] "RemoveContainer" containerID="ad22a18bbef271426a4077aab78c5dbcaeed67525c562d753181de83cefb8f37" Mar 13 15:50:12 crc kubenswrapper[4907]: I0313 15:50:12.545536 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jj9xr" Mar 13 15:50:12 crc kubenswrapper[4907]: I0313 15:50:12.579155 4907 scope.go:117] "RemoveContainer" containerID="d2ea7f99cadaa4f0d0bfca645932300ec0e331fafcb1e08d6511678f159fdd05" Mar 13 15:50:12 crc kubenswrapper[4907]: I0313 15:50:12.602553 4907 scope.go:117] "RemoveContainer" containerID="240a7e1dba69382ab9e9bd056e343991036dcabf16f06d6bad39768a4324c999" Mar 13 15:50:12 crc kubenswrapper[4907]: I0313 15:50:12.603693 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-jj9xr"] Mar 13 15:50:12 crc kubenswrapper[4907]: I0313 15:50:12.626792 4907 scope.go:117] "RemoveContainer" containerID="ad22a18bbef271426a4077aab78c5dbcaeed67525c562d753181de83cefb8f37" Mar 13 15:50:12 crc kubenswrapper[4907]: E0313 15:50:12.627438 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ad22a18bbef271426a4077aab78c5dbcaeed67525c562d753181de83cefb8f37\": container with ID starting with ad22a18bbef271426a4077aab78c5dbcaeed67525c562d753181de83cefb8f37 not found: ID does not exist" containerID="ad22a18bbef271426a4077aab78c5dbcaeed67525c562d753181de83cefb8f37" Mar 13 15:50:12 crc kubenswrapper[4907]: I0313 15:50:12.627501 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ad22a18bbef271426a4077aab78c5dbcaeed67525c562d753181de83cefb8f37"} err="failed to get container status \"ad22a18bbef271426a4077aab78c5dbcaeed67525c562d753181de83cefb8f37\": rpc error: code = NotFound desc = could not find container \"ad22a18bbef271426a4077aab78c5dbcaeed67525c562d753181de83cefb8f37\": container with ID starting with ad22a18bbef271426a4077aab78c5dbcaeed67525c562d753181de83cefb8f37 not found: ID does not exist" Mar 13 15:50:12 crc kubenswrapper[4907]: I0313 15:50:12.627533 4907 scope.go:117] "RemoveContainer" containerID="d2ea7f99cadaa4f0d0bfca645932300ec0e331fafcb1e08d6511678f159fdd05" Mar 13 15:50:12 crc kubenswrapper[4907]: E0313 15:50:12.628127 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d2ea7f99cadaa4f0d0bfca645932300ec0e331fafcb1e08d6511678f159fdd05\": container with ID starting with d2ea7f99cadaa4f0d0bfca645932300ec0e331fafcb1e08d6511678f159fdd05 not found: ID does not exist" containerID="d2ea7f99cadaa4f0d0bfca645932300ec0e331fafcb1e08d6511678f159fdd05" Mar 13 15:50:12 crc kubenswrapper[4907]: I0313 15:50:12.628185 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d2ea7f99cadaa4f0d0bfca645932300ec0e331fafcb1e08d6511678f159fdd05"} err="failed to get container status \"d2ea7f99cadaa4f0d0bfca645932300ec0e331fafcb1e08d6511678f159fdd05\": rpc error: code = NotFound desc = could not find container \"d2ea7f99cadaa4f0d0bfca645932300ec0e331fafcb1e08d6511678f159fdd05\": container with ID starting with d2ea7f99cadaa4f0d0bfca645932300ec0e331fafcb1e08d6511678f159fdd05 not found: ID does not exist" Mar 13 15:50:12 crc kubenswrapper[4907]: I0313 15:50:12.628217 4907 scope.go:117] "RemoveContainer" containerID="240a7e1dba69382ab9e9bd056e343991036dcabf16f06d6bad39768a4324c999" Mar 13 15:50:12 crc kubenswrapper[4907]: E0313 15:50:12.628548 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"240a7e1dba69382ab9e9bd056e343991036dcabf16f06d6bad39768a4324c999\": container with ID starting with 240a7e1dba69382ab9e9bd056e343991036dcabf16f06d6bad39768a4324c999 not found: ID does not exist" containerID="240a7e1dba69382ab9e9bd056e343991036dcabf16f06d6bad39768a4324c999" Mar 13 15:50:12 crc kubenswrapper[4907]: I0313 15:50:12.628580 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"240a7e1dba69382ab9e9bd056e343991036dcabf16f06d6bad39768a4324c999"} err="failed to get container status \"240a7e1dba69382ab9e9bd056e343991036dcabf16f06d6bad39768a4324c999\": rpc error: code = NotFound desc = could not find container \"240a7e1dba69382ab9e9bd056e343991036dcabf16f06d6bad39768a4324c999\": container with ID starting with 240a7e1dba69382ab9e9bd056e343991036dcabf16f06d6bad39768a4324c999 not found: ID does not exist" Mar 13 15:50:12 crc kubenswrapper[4907]: I0313 15:50:12.634644 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-jj9xr"] Mar 13 15:50:13 crc kubenswrapper[4907]: I0313 15:50:13.795339 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="951bf99b-9f5e-4609-bff0-0c594f6a5b2e" path="/var/lib/kubelet/pods/951bf99b-9f5e-4609-bff0-0c594f6a5b2e/volumes" Mar 13 15:50:14 crc kubenswrapper[4907]: I0313 15:50:14.525781 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-db-sync-nhx6q"] Mar 13 15:50:14 crc kubenswrapper[4907]: E0313 15:50:14.526523 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="951bf99b-9f5e-4609-bff0-0c594f6a5b2e" containerName="registry-server" Mar 13 15:50:14 crc kubenswrapper[4907]: I0313 15:50:14.526543 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="951bf99b-9f5e-4609-bff0-0c594f6a5b2e" containerName="registry-server" Mar 13 15:50:14 crc kubenswrapper[4907]: E0313 15:50:14.526554 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0498e113-86c3-4718-a580-13bfdb277ff2" containerName="mariadb-account-create-update" Mar 13 15:50:14 crc kubenswrapper[4907]: I0313 15:50:14.526564 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="0498e113-86c3-4718-a580-13bfdb277ff2" containerName="mariadb-account-create-update" Mar 13 15:50:14 crc kubenswrapper[4907]: E0313 15:50:14.526583 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2a34745-a555-4029-a203-ee17762929e2" containerName="mariadb-database-create" Mar 13 15:50:14 crc kubenswrapper[4907]: I0313 15:50:14.526594 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2a34745-a555-4029-a203-ee17762929e2" containerName="mariadb-database-create" Mar 13 15:50:14 crc kubenswrapper[4907]: E0313 15:50:14.526624 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="951bf99b-9f5e-4609-bff0-0c594f6a5b2e" containerName="extract-content" Mar 13 15:50:14 crc kubenswrapper[4907]: I0313 15:50:14.526632 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="951bf99b-9f5e-4609-bff0-0c594f6a5b2e" containerName="extract-content" Mar 13 15:50:14 crc kubenswrapper[4907]: E0313 15:50:14.526668 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="951bf99b-9f5e-4609-bff0-0c594f6a5b2e" containerName="extract-utilities" Mar 13 15:50:14 crc kubenswrapper[4907]: I0313 15:50:14.526678 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="951bf99b-9f5e-4609-bff0-0c594f6a5b2e" containerName="extract-utilities" Mar 13 15:50:14 crc kubenswrapper[4907]: I0313 15:50:14.526923 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="b2a34745-a555-4029-a203-ee17762929e2" containerName="mariadb-database-create" Mar 13 15:50:14 crc kubenswrapper[4907]: I0313 15:50:14.526962 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="0498e113-86c3-4718-a580-13bfdb277ff2" containerName="mariadb-account-create-update" Mar 13 15:50:14 crc kubenswrapper[4907]: I0313 15:50:14.526976 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="951bf99b-9f5e-4609-bff0-0c594f6a5b2e" containerName="registry-server" Mar 13 15:50:14 crc kubenswrapper[4907]: I0313 15:50:14.527827 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-nhx6q" Mar 13 15:50:14 crc kubenswrapper[4907]: I0313 15:50:14.531414 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-config-data" Mar 13 15:50:14 crc kubenswrapper[4907]: I0313 15:50:14.531632 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-manila-dockercfg-97kc6" Mar 13 15:50:14 crc kubenswrapper[4907]: I0313 15:50:14.550405 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-sync-nhx6q"] Mar 13 15:50:14 crc kubenswrapper[4907]: I0313 15:50:14.598828 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6hjjj\" (UniqueName: \"kubernetes.io/projected/2df4eaaf-eb5c-4e04-b987-5aedb6561f7e-kube-api-access-6hjjj\") pod \"manila-db-sync-nhx6q\" (UID: \"2df4eaaf-eb5c-4e04-b987-5aedb6561f7e\") " pod="openstack/manila-db-sync-nhx6q" Mar 13 15:50:14 crc kubenswrapper[4907]: I0313 15:50:14.598877 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/2df4eaaf-eb5c-4e04-b987-5aedb6561f7e-job-config-data\") pod \"manila-db-sync-nhx6q\" (UID: \"2df4eaaf-eb5c-4e04-b987-5aedb6561f7e\") " pod="openstack/manila-db-sync-nhx6q" Mar 13 15:50:14 crc kubenswrapper[4907]: I0313 15:50:14.598978 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2df4eaaf-eb5c-4e04-b987-5aedb6561f7e-config-data\") pod \"manila-db-sync-nhx6q\" (UID: \"2df4eaaf-eb5c-4e04-b987-5aedb6561f7e\") " pod="openstack/manila-db-sync-nhx6q" Mar 13 15:50:14 crc kubenswrapper[4907]: I0313 15:50:14.599085 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2df4eaaf-eb5c-4e04-b987-5aedb6561f7e-combined-ca-bundle\") pod \"manila-db-sync-nhx6q\" (UID: \"2df4eaaf-eb5c-4e04-b987-5aedb6561f7e\") " pod="openstack/manila-db-sync-nhx6q" Mar 13 15:50:14 crc kubenswrapper[4907]: I0313 15:50:14.701463 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2df4eaaf-eb5c-4e04-b987-5aedb6561f7e-config-data\") pod \"manila-db-sync-nhx6q\" (UID: \"2df4eaaf-eb5c-4e04-b987-5aedb6561f7e\") " pod="openstack/manila-db-sync-nhx6q" Mar 13 15:50:14 crc kubenswrapper[4907]: I0313 15:50:14.702196 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2df4eaaf-eb5c-4e04-b987-5aedb6561f7e-combined-ca-bundle\") pod \"manila-db-sync-nhx6q\" (UID: \"2df4eaaf-eb5c-4e04-b987-5aedb6561f7e\") " pod="openstack/manila-db-sync-nhx6q" Mar 13 15:50:14 crc kubenswrapper[4907]: I0313 15:50:14.702296 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6hjjj\" (UniqueName: \"kubernetes.io/projected/2df4eaaf-eb5c-4e04-b987-5aedb6561f7e-kube-api-access-6hjjj\") pod \"manila-db-sync-nhx6q\" (UID: \"2df4eaaf-eb5c-4e04-b987-5aedb6561f7e\") " pod="openstack/manila-db-sync-nhx6q" Mar 13 15:50:14 crc kubenswrapper[4907]: I0313 15:50:14.702325 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/2df4eaaf-eb5c-4e04-b987-5aedb6561f7e-job-config-data\") pod \"manila-db-sync-nhx6q\" (UID: \"2df4eaaf-eb5c-4e04-b987-5aedb6561f7e\") " pod="openstack/manila-db-sync-nhx6q" Mar 13 15:50:14 crc kubenswrapper[4907]: I0313 15:50:14.708765 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2df4eaaf-eb5c-4e04-b987-5aedb6561f7e-combined-ca-bundle\") pod \"manila-db-sync-nhx6q\" (UID: \"2df4eaaf-eb5c-4e04-b987-5aedb6561f7e\") " pod="openstack/manila-db-sync-nhx6q" Mar 13 15:50:14 crc kubenswrapper[4907]: I0313 15:50:14.708891 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2df4eaaf-eb5c-4e04-b987-5aedb6561f7e-config-data\") pod \"manila-db-sync-nhx6q\" (UID: \"2df4eaaf-eb5c-4e04-b987-5aedb6561f7e\") " pod="openstack/manila-db-sync-nhx6q" Mar 13 15:50:14 crc kubenswrapper[4907]: I0313 15:50:14.709590 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/2df4eaaf-eb5c-4e04-b987-5aedb6561f7e-job-config-data\") pod \"manila-db-sync-nhx6q\" (UID: \"2df4eaaf-eb5c-4e04-b987-5aedb6561f7e\") " pod="openstack/manila-db-sync-nhx6q" Mar 13 15:50:14 crc kubenswrapper[4907]: I0313 15:50:14.718746 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6hjjj\" (UniqueName: \"kubernetes.io/projected/2df4eaaf-eb5c-4e04-b987-5aedb6561f7e-kube-api-access-6hjjj\") pod \"manila-db-sync-nhx6q\" (UID: \"2df4eaaf-eb5c-4e04-b987-5aedb6561f7e\") " pod="openstack/manila-db-sync-nhx6q" Mar 13 15:50:14 crc kubenswrapper[4907]: I0313 15:50:14.862404 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-nhx6q" Mar 13 15:50:15 crc kubenswrapper[4907]: W0313 15:50:15.591276 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2df4eaaf_eb5c_4e04_b987_5aedb6561f7e.slice/crio-b2c4c3a6559d1515e062025b3cd0e22f753423b868d9846b459513b583ca8b0a WatchSource:0}: Error finding container b2c4c3a6559d1515e062025b3cd0e22f753423b868d9846b459513b583ca8b0a: Status 404 returned error can't find the container with id b2c4c3a6559d1515e062025b3cd0e22f753423b868d9846b459513b583ca8b0a Mar 13 15:50:15 crc kubenswrapper[4907]: I0313 15:50:15.591483 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-sync-nhx6q"] Mar 13 15:50:16 crc kubenswrapper[4907]: I0313 15:50:16.593479 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-nhx6q" event={"ID":"2df4eaaf-eb5c-4e04-b987-5aedb6561f7e","Type":"ContainerStarted","Data":"b2c4c3a6559d1515e062025b3cd0e22f753423b868d9846b459513b583ca8b0a"} Mar 13 15:50:18 crc kubenswrapper[4907]: I0313 15:50:18.046968 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-c2mhb"] Mar 13 15:50:18 crc kubenswrapper[4907]: I0313 15:50:18.061079 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-c2mhb"] Mar 13 15:50:19 crc kubenswrapper[4907]: I0313 15:50:19.045925 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-lvbcr"] Mar 13 15:50:19 crc kubenswrapper[4907]: I0313 15:50:19.062348 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-lvbcr"] Mar 13 15:50:19 crc kubenswrapper[4907]: I0313 15:50:19.793541 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="89840ecf-ddb9-4a66-855b-0a1ae524bd6e" path="/var/lib/kubelet/pods/89840ecf-ddb9-4a66-855b-0a1ae524bd6e/volumes" Mar 13 15:50:19 crc kubenswrapper[4907]: I0313 15:50:19.877168 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e935140e-65a6-48ff-9dda-fdf57edb9a4d" path="/var/lib/kubelet/pods/e935140e-65a6-48ff-9dda-fdf57edb9a4d/volumes" Mar 13 15:50:21 crc kubenswrapper[4907]: I0313 15:50:21.646957 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-nhx6q" event={"ID":"2df4eaaf-eb5c-4e04-b987-5aedb6561f7e","Type":"ContainerStarted","Data":"8cc12017f30ada4d93b2676752544b7d25e61944ae3bcbe64687e02b090ffb18"} Mar 13 15:50:21 crc kubenswrapper[4907]: I0313 15:50:21.671596 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-db-sync-nhx6q" podStartSLOduration=2.719396925 podStartE2EDuration="7.671578957s" podCreationTimestamp="2026-03-13 15:50:14 +0000 UTC" firstStartedPulling="2026-03-13 15:50:15.593100602 +0000 UTC m=+6314.492888291" lastFinishedPulling="2026-03-13 15:50:20.545282634 +0000 UTC m=+6319.445070323" observedRunningTime="2026-03-13 15:50:21.665091169 +0000 UTC m=+6320.564878858" watchObservedRunningTime="2026-03-13 15:50:21.671578957 +0000 UTC m=+6320.571366646" Mar 13 15:50:23 crc kubenswrapper[4907]: I0313 15:50:23.664777 4907 generic.go:334] "Generic (PLEG): container finished" podID="2df4eaaf-eb5c-4e04-b987-5aedb6561f7e" containerID="8cc12017f30ada4d93b2676752544b7d25e61944ae3bcbe64687e02b090ffb18" exitCode=0 Mar 13 15:50:23 crc kubenswrapper[4907]: I0313 15:50:23.664861 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-nhx6q" event={"ID":"2df4eaaf-eb5c-4e04-b987-5aedb6561f7e","Type":"ContainerDied","Data":"8cc12017f30ada4d93b2676752544b7d25e61944ae3bcbe64687e02b090ffb18"} Mar 13 15:50:25 crc kubenswrapper[4907]: I0313 15:50:25.155648 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-nhx6q" Mar 13 15:50:25 crc kubenswrapper[4907]: I0313 15:50:25.305116 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2df4eaaf-eb5c-4e04-b987-5aedb6561f7e-config-data\") pod \"2df4eaaf-eb5c-4e04-b987-5aedb6561f7e\" (UID: \"2df4eaaf-eb5c-4e04-b987-5aedb6561f7e\") " Mar 13 15:50:25 crc kubenswrapper[4907]: I0313 15:50:25.305277 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/2df4eaaf-eb5c-4e04-b987-5aedb6561f7e-job-config-data\") pod \"2df4eaaf-eb5c-4e04-b987-5aedb6561f7e\" (UID: \"2df4eaaf-eb5c-4e04-b987-5aedb6561f7e\") " Mar 13 15:50:25 crc kubenswrapper[4907]: I0313 15:50:25.305447 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6hjjj\" (UniqueName: \"kubernetes.io/projected/2df4eaaf-eb5c-4e04-b987-5aedb6561f7e-kube-api-access-6hjjj\") pod \"2df4eaaf-eb5c-4e04-b987-5aedb6561f7e\" (UID: \"2df4eaaf-eb5c-4e04-b987-5aedb6561f7e\") " Mar 13 15:50:25 crc kubenswrapper[4907]: I0313 15:50:25.305539 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2df4eaaf-eb5c-4e04-b987-5aedb6561f7e-combined-ca-bundle\") pod \"2df4eaaf-eb5c-4e04-b987-5aedb6561f7e\" (UID: \"2df4eaaf-eb5c-4e04-b987-5aedb6561f7e\") " Mar 13 15:50:25 crc kubenswrapper[4907]: I0313 15:50:25.313189 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2df4eaaf-eb5c-4e04-b987-5aedb6561f7e-job-config-data" (OuterVolumeSpecName: "job-config-data") pod "2df4eaaf-eb5c-4e04-b987-5aedb6561f7e" (UID: "2df4eaaf-eb5c-4e04-b987-5aedb6561f7e"). InnerVolumeSpecName "job-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:50:25 crc kubenswrapper[4907]: I0313 15:50:25.314091 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2df4eaaf-eb5c-4e04-b987-5aedb6561f7e-kube-api-access-6hjjj" (OuterVolumeSpecName: "kube-api-access-6hjjj") pod "2df4eaaf-eb5c-4e04-b987-5aedb6561f7e" (UID: "2df4eaaf-eb5c-4e04-b987-5aedb6561f7e"). InnerVolumeSpecName "kube-api-access-6hjjj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:50:25 crc kubenswrapper[4907]: I0313 15:50:25.315523 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2df4eaaf-eb5c-4e04-b987-5aedb6561f7e-config-data" (OuterVolumeSpecName: "config-data") pod "2df4eaaf-eb5c-4e04-b987-5aedb6561f7e" (UID: "2df4eaaf-eb5c-4e04-b987-5aedb6561f7e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:50:25 crc kubenswrapper[4907]: I0313 15:50:25.337964 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2df4eaaf-eb5c-4e04-b987-5aedb6561f7e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2df4eaaf-eb5c-4e04-b987-5aedb6561f7e" (UID: "2df4eaaf-eb5c-4e04-b987-5aedb6561f7e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:50:25 crc kubenswrapper[4907]: I0313 15:50:25.407795 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6hjjj\" (UniqueName: \"kubernetes.io/projected/2df4eaaf-eb5c-4e04-b987-5aedb6561f7e-kube-api-access-6hjjj\") on node \"crc\" DevicePath \"\"" Mar 13 15:50:25 crc kubenswrapper[4907]: I0313 15:50:25.408248 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2df4eaaf-eb5c-4e04-b987-5aedb6561f7e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 15:50:25 crc kubenswrapper[4907]: I0313 15:50:25.408327 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2df4eaaf-eb5c-4e04-b987-5aedb6561f7e-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 15:50:25 crc kubenswrapper[4907]: I0313 15:50:25.408402 4907 reconciler_common.go:293] "Volume detached for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/2df4eaaf-eb5c-4e04-b987-5aedb6561f7e-job-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 15:50:25 crc kubenswrapper[4907]: I0313 15:50:25.702067 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-nhx6q" event={"ID":"2df4eaaf-eb5c-4e04-b987-5aedb6561f7e","Type":"ContainerDied","Data":"b2c4c3a6559d1515e062025b3cd0e22f753423b868d9846b459513b583ca8b0a"} Mar 13 15:50:25 crc kubenswrapper[4907]: I0313 15:50:25.702125 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b2c4c3a6559d1515e062025b3cd0e22f753423b868d9846b459513b583ca8b0a" Mar 13 15:50:25 crc kubenswrapper[4907]: I0313 15:50:25.702245 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-nhx6q" Mar 13 15:50:25 crc kubenswrapper[4907]: I0313 15:50:25.988516 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-scheduler-0"] Mar 13 15:50:25 crc kubenswrapper[4907]: E0313 15:50:25.989125 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2df4eaaf-eb5c-4e04-b987-5aedb6561f7e" containerName="manila-db-sync" Mar 13 15:50:25 crc kubenswrapper[4907]: I0313 15:50:25.989145 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="2df4eaaf-eb5c-4e04-b987-5aedb6561f7e" containerName="manila-db-sync" Mar 13 15:50:25 crc kubenswrapper[4907]: I0313 15:50:25.989374 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="2df4eaaf-eb5c-4e04-b987-5aedb6561f7e" containerName="manila-db-sync" Mar 13 15:50:25 crc kubenswrapper[4907]: I0313 15:50:25.990545 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Mar 13 15:50:25 crc kubenswrapper[4907]: I0313 15:50:25.993852 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-scheduler-config-data" Mar 13 15:50:25 crc kubenswrapper[4907]: I0313 15:50:25.994104 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-config-data" Mar 13 15:50:25 crc kubenswrapper[4907]: I0313 15:50:25.994176 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-manila-dockercfg-97kc6" Mar 13 15:50:25 crc kubenswrapper[4907]: I0313 15:50:25.995559 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-scripts" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.001122 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.068574 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-share-share1-0"] Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.070964 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.073535 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-share-share1-config-data" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.101180 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.127935 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7tkg6\" (UniqueName: \"kubernetes.io/projected/2e179ecf-0fc5-4675-89a6-287d3398dc2d-kube-api-access-7tkg6\") pod \"manila-scheduler-0\" (UID: \"2e179ecf-0fc5-4675-89a6-287d3398dc2d\") " pod="openstack/manila-scheduler-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.128192 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2e179ecf-0fc5-4675-89a6-287d3398dc2d-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"2e179ecf-0fc5-4675-89a6-287d3398dc2d\") " pod="openstack/manila-scheduler-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.128231 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e179ecf-0fc5-4675-89a6-287d3398dc2d-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"2e179ecf-0fc5-4675-89a6-287d3398dc2d\") " pod="openstack/manila-scheduler-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.128279 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e179ecf-0fc5-4675-89a6-287d3398dc2d-config-data\") pod \"manila-scheduler-0\" (UID: \"2e179ecf-0fc5-4675-89a6-287d3398dc2d\") " pod="openstack/manila-scheduler-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.128306 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2e179ecf-0fc5-4675-89a6-287d3398dc2d-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"2e179ecf-0fc5-4675-89a6-287d3398dc2d\") " pod="openstack/manila-scheduler-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.128374 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e179ecf-0fc5-4675-89a6-287d3398dc2d-scripts\") pod \"manila-scheduler-0\" (UID: \"2e179ecf-0fc5-4675-89a6-287d3398dc2d\") " pod="openstack/manila-scheduler-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.183940 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7559d477c5-jggzg"] Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.186172 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7559d477c5-jggzg" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.198936 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7559d477c5-jggzg"] Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.235373 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93911045-021f-489b-9817-f0a24066c906-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"93911045-021f-489b-9817-f0a24066c906\") " pod="openstack/manila-share-share1-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.235448 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tbkz8\" (UniqueName: \"kubernetes.io/projected/93911045-021f-489b-9817-f0a24066c906-kube-api-access-tbkz8\") pod \"manila-share-share1-0\" (UID: \"93911045-021f-489b-9817-f0a24066c906\") " pod="openstack/manila-share-share1-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.235532 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/93911045-021f-489b-9817-f0a24066c906-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"93911045-021f-489b-9817-f0a24066c906\") " pod="openstack/manila-share-share1-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.235570 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/93911045-021f-489b-9817-f0a24066c906-scripts\") pod \"manila-share-share1-0\" (UID: \"93911045-021f-489b-9817-f0a24066c906\") " pod="openstack/manila-share-share1-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.235652 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/93911045-021f-489b-9817-f0a24066c906-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"93911045-021f-489b-9817-f0a24066c906\") " pod="openstack/manila-share-share1-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.235849 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7tkg6\" (UniqueName: \"kubernetes.io/projected/2e179ecf-0fc5-4675-89a6-287d3398dc2d-kube-api-access-7tkg6\") pod \"manila-scheduler-0\" (UID: \"2e179ecf-0fc5-4675-89a6-287d3398dc2d\") " pod="openstack/manila-scheduler-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.235940 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2e179ecf-0fc5-4675-89a6-287d3398dc2d-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"2e179ecf-0fc5-4675-89a6-287d3398dc2d\") " pod="openstack/manila-scheduler-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.235982 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e179ecf-0fc5-4675-89a6-287d3398dc2d-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"2e179ecf-0fc5-4675-89a6-287d3398dc2d\") " pod="openstack/manila-scheduler-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.236043 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e179ecf-0fc5-4675-89a6-287d3398dc2d-config-data\") pod \"manila-scheduler-0\" (UID: \"2e179ecf-0fc5-4675-89a6-287d3398dc2d\") " pod="openstack/manila-scheduler-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.236076 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2e179ecf-0fc5-4675-89a6-287d3398dc2d-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"2e179ecf-0fc5-4675-89a6-287d3398dc2d\") " pod="openstack/manila-scheduler-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.236136 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93911045-021f-489b-9817-f0a24066c906-config-data\") pod \"manila-share-share1-0\" (UID: \"93911045-021f-489b-9817-f0a24066c906\") " pod="openstack/manila-share-share1-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.236187 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e179ecf-0fc5-4675-89a6-287d3398dc2d-scripts\") pod \"manila-scheduler-0\" (UID: \"2e179ecf-0fc5-4675-89a6-287d3398dc2d\") " pod="openstack/manila-scheduler-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.236227 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/93911045-021f-489b-9817-f0a24066c906-ceph\") pod \"manila-share-share1-0\" (UID: \"93911045-021f-489b-9817-f0a24066c906\") " pod="openstack/manila-share-share1-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.236253 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/93911045-021f-489b-9817-f0a24066c906-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"93911045-021f-489b-9817-f0a24066c906\") " pod="openstack/manila-share-share1-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.238660 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2e179ecf-0fc5-4675-89a6-287d3398dc2d-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"2e179ecf-0fc5-4675-89a6-287d3398dc2d\") " pod="openstack/manila-scheduler-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.242521 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e179ecf-0fc5-4675-89a6-287d3398dc2d-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"2e179ecf-0fc5-4675-89a6-287d3398dc2d\") " pod="openstack/manila-scheduler-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.243048 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2e179ecf-0fc5-4675-89a6-287d3398dc2d-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"2e179ecf-0fc5-4675-89a6-287d3398dc2d\") " pod="openstack/manila-scheduler-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.251603 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e179ecf-0fc5-4675-89a6-287d3398dc2d-config-data\") pod \"manila-scheduler-0\" (UID: \"2e179ecf-0fc5-4675-89a6-287d3398dc2d\") " pod="openstack/manila-scheduler-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.258432 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e179ecf-0fc5-4675-89a6-287d3398dc2d-scripts\") pod \"manila-scheduler-0\" (UID: \"2e179ecf-0fc5-4675-89a6-287d3398dc2d\") " pod="openstack/manila-scheduler-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.258754 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7tkg6\" (UniqueName: \"kubernetes.io/projected/2e179ecf-0fc5-4675-89a6-287d3398dc2d-kube-api-access-7tkg6\") pod \"manila-scheduler-0\" (UID: \"2e179ecf-0fc5-4675-89a6-287d3398dc2d\") " pod="openstack/manila-scheduler-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.322090 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.337799 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/93911045-021f-489b-9817-f0a24066c906-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"93911045-021f-489b-9817-f0a24066c906\") " pod="openstack/manila-share-share1-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.338281 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/93911045-021f-489b-9817-f0a24066c906-scripts\") pod \"manila-share-share1-0\" (UID: \"93911045-021f-489b-9817-f0a24066c906\") " pod="openstack/manila-share-share1-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.338331 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/93911045-021f-489b-9817-f0a24066c906-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"93911045-021f-489b-9817-f0a24066c906\") " pod="openstack/manila-share-share1-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.338362 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a01ed54c-beec-437b-a4fb-c4c62a407d76-config\") pod \"dnsmasq-dns-7559d477c5-jggzg\" (UID: \"a01ed54c-beec-437b-a4fb-c4c62a407d76\") " pod="openstack/dnsmasq-dns-7559d477c5-jggzg" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.338392 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a01ed54c-beec-437b-a4fb-c4c62a407d76-ovsdbserver-nb\") pod \"dnsmasq-dns-7559d477c5-jggzg\" (UID: \"a01ed54c-beec-437b-a4fb-c4c62a407d76\") " pod="openstack/dnsmasq-dns-7559d477c5-jggzg" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.338497 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a01ed54c-beec-437b-a4fb-c4c62a407d76-dns-svc\") pod \"dnsmasq-dns-7559d477c5-jggzg\" (UID: \"a01ed54c-beec-437b-a4fb-c4c62a407d76\") " pod="openstack/dnsmasq-dns-7559d477c5-jggzg" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.338579 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93911045-021f-489b-9817-f0a24066c906-config-data\") pod \"manila-share-share1-0\" (UID: \"93911045-021f-489b-9817-f0a24066c906\") " pod="openstack/manila-share-share1-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.338658 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/93911045-021f-489b-9817-f0a24066c906-ceph\") pod \"manila-share-share1-0\" (UID: \"93911045-021f-489b-9817-f0a24066c906\") " pod="openstack/manila-share-share1-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.338694 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/93911045-021f-489b-9817-f0a24066c906-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"93911045-021f-489b-9817-f0a24066c906\") " pod="openstack/manila-share-share1-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.338763 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93911045-021f-489b-9817-f0a24066c906-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"93911045-021f-489b-9817-f0a24066c906\") " pod="openstack/manila-share-share1-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.338828 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tbkz8\" (UniqueName: \"kubernetes.io/projected/93911045-021f-489b-9817-f0a24066c906-kube-api-access-tbkz8\") pod \"manila-share-share1-0\" (UID: \"93911045-021f-489b-9817-f0a24066c906\") " pod="openstack/manila-share-share1-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.338874 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a01ed54c-beec-437b-a4fb-c4c62a407d76-ovsdbserver-sb\") pod \"dnsmasq-dns-7559d477c5-jggzg\" (UID: \"a01ed54c-beec-437b-a4fb-c4c62a407d76\") " pod="openstack/dnsmasq-dns-7559d477c5-jggzg" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.338946 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bnxr5\" (UniqueName: \"kubernetes.io/projected/a01ed54c-beec-437b-a4fb-c4c62a407d76-kube-api-access-bnxr5\") pod \"dnsmasq-dns-7559d477c5-jggzg\" (UID: \"a01ed54c-beec-437b-a4fb-c4c62a407d76\") " pod="openstack/dnsmasq-dns-7559d477c5-jggzg" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.339118 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/93911045-021f-489b-9817-f0a24066c906-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"93911045-021f-489b-9817-f0a24066c906\") " pod="openstack/manila-share-share1-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.341918 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/93911045-021f-489b-9817-f0a24066c906-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"93911045-021f-489b-9817-f0a24066c906\") " pod="openstack/manila-share-share1-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.348059 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/93911045-021f-489b-9817-f0a24066c906-ceph\") pod \"manila-share-share1-0\" (UID: \"93911045-021f-489b-9817-f0a24066c906\") " pod="openstack/manila-share-share1-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.354072 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93911045-021f-489b-9817-f0a24066c906-config-data\") pod \"manila-share-share1-0\" (UID: \"93911045-021f-489b-9817-f0a24066c906\") " pod="openstack/manila-share-share1-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.356331 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/93911045-021f-489b-9817-f0a24066c906-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"93911045-021f-489b-9817-f0a24066c906\") " pod="openstack/manila-share-share1-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.367167 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-api-0"] Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.369915 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.378085 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.379542 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-api-config-data" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.381227 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93911045-021f-489b-9817-f0a24066c906-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"93911045-021f-489b-9817-f0a24066c906\") " pod="openstack/manila-share-share1-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.381798 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/93911045-021f-489b-9817-f0a24066c906-scripts\") pod \"manila-share-share1-0\" (UID: \"93911045-021f-489b-9817-f0a24066c906\") " pod="openstack/manila-share-share1-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.395152 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tbkz8\" (UniqueName: \"kubernetes.io/projected/93911045-021f-489b-9817-f0a24066c906-kube-api-access-tbkz8\") pod \"manila-share-share1-0\" (UID: \"93911045-021f-489b-9817-f0a24066c906\") " pod="openstack/manila-share-share1-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.440509 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a01ed54c-beec-437b-a4fb-c4c62a407d76-ovsdbserver-sb\") pod \"dnsmasq-dns-7559d477c5-jggzg\" (UID: \"a01ed54c-beec-437b-a4fb-c4c62a407d76\") " pod="openstack/dnsmasq-dns-7559d477c5-jggzg" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.440747 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bnxr5\" (UniqueName: \"kubernetes.io/projected/a01ed54c-beec-437b-a4fb-c4c62a407d76-kube-api-access-bnxr5\") pod \"dnsmasq-dns-7559d477c5-jggzg\" (UID: \"a01ed54c-beec-437b-a4fb-c4c62a407d76\") " pod="openstack/dnsmasq-dns-7559d477c5-jggzg" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.440870 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a01ed54c-beec-437b-a4fb-c4c62a407d76-config\") pod \"dnsmasq-dns-7559d477c5-jggzg\" (UID: \"a01ed54c-beec-437b-a4fb-c4c62a407d76\") " pod="openstack/dnsmasq-dns-7559d477c5-jggzg" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.440996 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a01ed54c-beec-437b-a4fb-c4c62a407d76-ovsdbserver-nb\") pod \"dnsmasq-dns-7559d477c5-jggzg\" (UID: \"a01ed54c-beec-437b-a4fb-c4c62a407d76\") " pod="openstack/dnsmasq-dns-7559d477c5-jggzg" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.441110 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a01ed54c-beec-437b-a4fb-c4c62a407d76-dns-svc\") pod \"dnsmasq-dns-7559d477c5-jggzg\" (UID: \"a01ed54c-beec-437b-a4fb-c4c62a407d76\") " pod="openstack/dnsmasq-dns-7559d477c5-jggzg" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.442520 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a01ed54c-beec-437b-a4fb-c4c62a407d76-ovsdbserver-nb\") pod \"dnsmasq-dns-7559d477c5-jggzg\" (UID: \"a01ed54c-beec-437b-a4fb-c4c62a407d76\") " pod="openstack/dnsmasq-dns-7559d477c5-jggzg" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.442805 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a01ed54c-beec-437b-a4fb-c4c62a407d76-ovsdbserver-sb\") pod \"dnsmasq-dns-7559d477c5-jggzg\" (UID: \"a01ed54c-beec-437b-a4fb-c4c62a407d76\") " pod="openstack/dnsmasq-dns-7559d477c5-jggzg" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.443608 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a01ed54c-beec-437b-a4fb-c4c62a407d76-config\") pod \"dnsmasq-dns-7559d477c5-jggzg\" (UID: \"a01ed54c-beec-437b-a4fb-c4c62a407d76\") " pod="openstack/dnsmasq-dns-7559d477c5-jggzg" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.444224 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a01ed54c-beec-437b-a4fb-c4c62a407d76-dns-svc\") pod \"dnsmasq-dns-7559d477c5-jggzg\" (UID: \"a01ed54c-beec-437b-a4fb-c4c62a407d76\") " pod="openstack/dnsmasq-dns-7559d477c5-jggzg" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.469030 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bnxr5\" (UniqueName: \"kubernetes.io/projected/a01ed54c-beec-437b-a4fb-c4c62a407d76-kube-api-access-bnxr5\") pod \"dnsmasq-dns-7559d477c5-jggzg\" (UID: \"a01ed54c-beec-437b-a4fb-c4c62a407d76\") " pod="openstack/dnsmasq-dns-7559d477c5-jggzg" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.507589 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7559d477c5-jggzg" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.543393 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2518181-dd4d-4d35-9377-061bbd91c811-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"f2518181-dd4d-4d35-9377-061bbd91c811\") " pod="openstack/manila-api-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.543432 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f2518181-dd4d-4d35-9377-061bbd91c811-etc-machine-id\") pod \"manila-api-0\" (UID: \"f2518181-dd4d-4d35-9377-061bbd91c811\") " pod="openstack/manila-api-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.543468 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wtp6p\" (UniqueName: \"kubernetes.io/projected/f2518181-dd4d-4d35-9377-061bbd91c811-kube-api-access-wtp6p\") pod \"manila-api-0\" (UID: \"f2518181-dd4d-4d35-9377-061bbd91c811\") " pod="openstack/manila-api-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.543595 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f2518181-dd4d-4d35-9377-061bbd91c811-scripts\") pod \"manila-api-0\" (UID: \"f2518181-dd4d-4d35-9377-061bbd91c811\") " pod="openstack/manila-api-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.543663 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f2518181-dd4d-4d35-9377-061bbd91c811-config-data-custom\") pod \"manila-api-0\" (UID: \"f2518181-dd4d-4d35-9377-061bbd91c811\") " pod="openstack/manila-api-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.543694 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f2518181-dd4d-4d35-9377-061bbd91c811-config-data\") pod \"manila-api-0\" (UID: \"f2518181-dd4d-4d35-9377-061bbd91c811\") " pod="openstack/manila-api-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.544047 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f2518181-dd4d-4d35-9377-061bbd91c811-logs\") pod \"manila-api-0\" (UID: \"f2518181-dd4d-4d35-9377-061bbd91c811\") " pod="openstack/manila-api-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.645547 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f2518181-dd4d-4d35-9377-061bbd91c811-logs\") pod \"manila-api-0\" (UID: \"f2518181-dd4d-4d35-9377-061bbd91c811\") " pod="openstack/manila-api-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.645782 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2518181-dd4d-4d35-9377-061bbd91c811-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"f2518181-dd4d-4d35-9377-061bbd91c811\") " pod="openstack/manila-api-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.645802 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f2518181-dd4d-4d35-9377-061bbd91c811-etc-machine-id\") pod \"manila-api-0\" (UID: \"f2518181-dd4d-4d35-9377-061bbd91c811\") " pod="openstack/manila-api-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.645830 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wtp6p\" (UniqueName: \"kubernetes.io/projected/f2518181-dd4d-4d35-9377-061bbd91c811-kube-api-access-wtp6p\") pod \"manila-api-0\" (UID: \"f2518181-dd4d-4d35-9377-061bbd91c811\") " pod="openstack/manila-api-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.645954 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f2518181-dd4d-4d35-9377-061bbd91c811-scripts\") pod \"manila-api-0\" (UID: \"f2518181-dd4d-4d35-9377-061bbd91c811\") " pod="openstack/manila-api-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.645992 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f2518181-dd4d-4d35-9377-061bbd91c811-config-data-custom\") pod \"manila-api-0\" (UID: \"f2518181-dd4d-4d35-9377-061bbd91c811\") " pod="openstack/manila-api-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.646009 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f2518181-dd4d-4d35-9377-061bbd91c811-config-data\") pod \"manila-api-0\" (UID: \"f2518181-dd4d-4d35-9377-061bbd91c811\") " pod="openstack/manila-api-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.646896 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f2518181-dd4d-4d35-9377-061bbd91c811-etc-machine-id\") pod \"manila-api-0\" (UID: \"f2518181-dd4d-4d35-9377-061bbd91c811\") " pod="openstack/manila-api-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.647353 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f2518181-dd4d-4d35-9377-061bbd91c811-logs\") pod \"manila-api-0\" (UID: \"f2518181-dd4d-4d35-9377-061bbd91c811\") " pod="openstack/manila-api-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.653174 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f2518181-dd4d-4d35-9377-061bbd91c811-config-data\") pod \"manila-api-0\" (UID: \"f2518181-dd4d-4d35-9377-061bbd91c811\") " pod="openstack/manila-api-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.655242 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f2518181-dd4d-4d35-9377-061bbd91c811-scripts\") pod \"manila-api-0\" (UID: \"f2518181-dd4d-4d35-9377-061bbd91c811\") " pod="openstack/manila-api-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.656793 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2518181-dd4d-4d35-9377-061bbd91c811-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"f2518181-dd4d-4d35-9377-061bbd91c811\") " pod="openstack/manila-api-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.659975 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f2518181-dd4d-4d35-9377-061bbd91c811-config-data-custom\") pod \"manila-api-0\" (UID: \"f2518181-dd4d-4d35-9377-061bbd91c811\") " pod="openstack/manila-api-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.668895 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wtp6p\" (UniqueName: \"kubernetes.io/projected/f2518181-dd4d-4d35-9377-061bbd91c811-kube-api-access-wtp6p\") pod \"manila-api-0\" (UID: \"f2518181-dd4d-4d35-9377-061bbd91c811\") " pod="openstack/manila-api-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.695225 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.832870 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Mar 13 15:50:27 crc kubenswrapper[4907]: I0313 15:50:27.049904 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Mar 13 15:50:27 crc kubenswrapper[4907]: W0313 15:50:27.076420 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2e179ecf_0fc5_4675_89a6_287d3398dc2d.slice/crio-1f1c2255067bd6fc9165910e84d3cb05d7d77bbad2543937a8f65a853ae24c2a WatchSource:0}: Error finding container 1f1c2255067bd6fc9165910e84d3cb05d7d77bbad2543937a8f65a853ae24c2a: Status 404 returned error can't find the container with id 1f1c2255067bd6fc9165910e84d3cb05d7d77bbad2543937a8f65a853ae24c2a Mar 13 15:50:27 crc kubenswrapper[4907]: I0313 15:50:27.238861 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7559d477c5-jggzg"] Mar 13 15:50:27 crc kubenswrapper[4907]: W0313 15:50:27.243194 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda01ed54c_beec_437b_a4fb_c4c62a407d76.slice/crio-9777020954fe435146b223f1e22d2da8657fcd1b5305e36c2273a7c3f9c349b1 WatchSource:0}: Error finding container 9777020954fe435146b223f1e22d2da8657fcd1b5305e36c2273a7c3f9c349b1: Status 404 returned error can't find the container with id 9777020954fe435146b223f1e22d2da8657fcd1b5305e36c2273a7c3f9c349b1 Mar 13 15:50:27 crc kubenswrapper[4907]: W0313 15:50:27.488533 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod93911045_021f_489b_9817_f0a24066c906.slice/crio-03324e1e5a626682ca7e1985f811fb94aa7f345d5a58a62c8ac49866dcee12fc WatchSource:0}: Error finding container 03324e1e5a626682ca7e1985f811fb94aa7f345d5a58a62c8ac49866dcee12fc: Status 404 returned error can't find the container with id 03324e1e5a626682ca7e1985f811fb94aa7f345d5a58a62c8ac49866dcee12fc Mar 13 15:50:27 crc kubenswrapper[4907]: I0313 15:50:27.495452 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Mar 13 15:50:27 crc kubenswrapper[4907]: I0313 15:50:27.653068 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Mar 13 15:50:27 crc kubenswrapper[4907]: W0313 15:50:27.673324 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf2518181_dd4d_4d35_9377_061bbd91c811.slice/crio-ab6f686c99c98e885820c86563b65060a971372055d87706c79e25a3c7cb42ad WatchSource:0}: Error finding container ab6f686c99c98e885820c86563b65060a971372055d87706c79e25a3c7cb42ad: Status 404 returned error can't find the container with id ab6f686c99c98e885820c86563b65060a971372055d87706c79e25a3c7cb42ad Mar 13 15:50:27 crc kubenswrapper[4907]: I0313 15:50:27.747989 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"93911045-021f-489b-9817-f0a24066c906","Type":"ContainerStarted","Data":"03324e1e5a626682ca7e1985f811fb94aa7f345d5a58a62c8ac49866dcee12fc"} Mar 13 15:50:27 crc kubenswrapper[4907]: I0313 15:50:27.751808 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"f2518181-dd4d-4d35-9377-061bbd91c811","Type":"ContainerStarted","Data":"ab6f686c99c98e885820c86563b65060a971372055d87706c79e25a3c7cb42ad"} Mar 13 15:50:27 crc kubenswrapper[4907]: I0313 15:50:27.755328 4907 generic.go:334] "Generic (PLEG): container finished" podID="a01ed54c-beec-437b-a4fb-c4c62a407d76" containerID="6a9689f3bb28f496b72b2c47ce88ac358b381c9e951286ab8db719b6cd0bd636" exitCode=0 Mar 13 15:50:27 crc kubenswrapper[4907]: I0313 15:50:27.755389 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7559d477c5-jggzg" event={"ID":"a01ed54c-beec-437b-a4fb-c4c62a407d76","Type":"ContainerDied","Data":"6a9689f3bb28f496b72b2c47ce88ac358b381c9e951286ab8db719b6cd0bd636"} Mar 13 15:50:27 crc kubenswrapper[4907]: I0313 15:50:27.755410 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7559d477c5-jggzg" event={"ID":"a01ed54c-beec-437b-a4fb-c4c62a407d76","Type":"ContainerStarted","Data":"9777020954fe435146b223f1e22d2da8657fcd1b5305e36c2273a7c3f9c349b1"} Mar 13 15:50:27 crc kubenswrapper[4907]: I0313 15:50:27.761334 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"2e179ecf-0fc5-4675-89a6-287d3398dc2d","Type":"ContainerStarted","Data":"1f1c2255067bd6fc9165910e84d3cb05d7d77bbad2543937a8f65a853ae24c2a"} Mar 13 15:50:28 crc kubenswrapper[4907]: I0313 15:50:28.738727 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Mar 13 15:50:28 crc kubenswrapper[4907]: I0313 15:50:28.808535 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"f2518181-dd4d-4d35-9377-061bbd91c811","Type":"ContainerStarted","Data":"5f1ed1ab0466cc73b3b8f1643b218bd51370eb20dc17ff9fab6a25c347de95be"} Mar 13 15:50:28 crc kubenswrapper[4907]: I0313 15:50:28.808614 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"f2518181-dd4d-4d35-9377-061bbd91c811","Type":"ContainerStarted","Data":"5b2926df5bae903b244bc30e7e4d11c8e0b8953f321de39acc3a4dd0facc6050"} Mar 13 15:50:28 crc kubenswrapper[4907]: I0313 15:50:28.808656 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/manila-api-0" Mar 13 15:50:28 crc kubenswrapper[4907]: I0313 15:50:28.813221 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7559d477c5-jggzg" event={"ID":"a01ed54c-beec-437b-a4fb-c4c62a407d76","Type":"ContainerStarted","Data":"f427c6c6d4cea6920aec5e3d13f7c210e91731dd6a8240743203f4fd1462ee2f"} Mar 13 15:50:28 crc kubenswrapper[4907]: I0313 15:50:28.836974 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"2e179ecf-0fc5-4675-89a6-287d3398dc2d","Type":"ContainerStarted","Data":"46b91298287e1c80e41badb65f95f6de2c660f1da51593026a7395f0692ceb8e"} Mar 13 15:50:28 crc kubenswrapper[4907]: I0313 15:50:28.839051 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-api-0" podStartSLOduration=2.839035584 podStartE2EDuration="2.839035584s" podCreationTimestamp="2026-03-13 15:50:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:50:28.836487434 +0000 UTC m=+6327.736275133" watchObservedRunningTime="2026-03-13 15:50:28.839035584 +0000 UTC m=+6327.738823273" Mar 13 15:50:28 crc kubenswrapper[4907]: I0313 15:50:28.876625 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7559d477c5-jggzg" podStartSLOduration=2.8766038099999998 podStartE2EDuration="2.87660381s" podCreationTimestamp="2026-03-13 15:50:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:50:28.859561734 +0000 UTC m=+6327.759349423" watchObservedRunningTime="2026-03-13 15:50:28.87660381 +0000 UTC m=+6327.776391499" Mar 13 15:50:29 crc kubenswrapper[4907]: I0313 15:50:29.850701 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"2e179ecf-0fc5-4675-89a6-287d3398dc2d","Type":"ContainerStarted","Data":"1a5fd5ea066c835cbc4a2587d84d4226edb7adf4e37526ac446862123a7cc55a"} Mar 13 15:50:29 crc kubenswrapper[4907]: I0313 15:50:29.851116 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7559d477c5-jggzg" Mar 13 15:50:29 crc kubenswrapper[4907]: I0313 15:50:29.874474 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-scheduler-0" podStartSLOduration=3.8626676189999998 podStartE2EDuration="4.874456533s" podCreationTimestamp="2026-03-13 15:50:25 +0000 UTC" firstStartedPulling="2026-03-13 15:50:27.079849419 +0000 UTC m=+6325.979637108" lastFinishedPulling="2026-03-13 15:50:28.091638333 +0000 UTC m=+6326.991426022" observedRunningTime="2026-03-13 15:50:29.87178299 +0000 UTC m=+6328.771570689" watchObservedRunningTime="2026-03-13 15:50:29.874456533 +0000 UTC m=+6328.774244222" Mar 13 15:50:33 crc kubenswrapper[4907]: I0313 15:50:33.030756 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-d9b97"] Mar 13 15:50:33 crc kubenswrapper[4907]: I0313 15:50:33.039632 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-d9b97"] Mar 13 15:50:33 crc kubenswrapper[4907]: I0313 15:50:33.805975 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f6e87a99-1744-4002-a08b-dabf8a97e48b" path="/var/lib/kubelet/pods/f6e87a99-1744-4002-a08b-dabf8a97e48b/volumes" Mar 13 15:50:34 crc kubenswrapper[4907]: I0313 15:50:34.905298 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"93911045-021f-489b-9817-f0a24066c906","Type":"ContainerStarted","Data":"99ecdb015c2877cd4051dfe0e106413112d1b3457e2928edbcb32905a4c80b35"} Mar 13 15:50:35 crc kubenswrapper[4907]: I0313 15:50:35.917713 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"93911045-021f-489b-9817-f0a24066c906","Type":"ContainerStarted","Data":"5a0ea1386dc8ac86951f18d7e6d31e118f0d651a291679d095103cc0e5e2b43e"} Mar 13 15:50:35 crc kubenswrapper[4907]: I0313 15:50:35.942688 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-share-share1-0" podStartSLOduration=2.9867187680000002 podStartE2EDuration="9.942455971s" podCreationTimestamp="2026-03-13 15:50:26 +0000 UTC" firstStartedPulling="2026-03-13 15:50:27.490488369 +0000 UTC m=+6326.390276058" lastFinishedPulling="2026-03-13 15:50:34.446225572 +0000 UTC m=+6333.346013261" observedRunningTime="2026-03-13 15:50:35.942446661 +0000 UTC m=+6334.842234360" watchObservedRunningTime="2026-03-13 15:50:35.942455971 +0000 UTC m=+6334.842243670" Mar 13 15:50:36 crc kubenswrapper[4907]: I0313 15:50:36.323751 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-scheduler-0" Mar 13 15:50:36 crc kubenswrapper[4907]: I0313 15:50:36.510024 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7559d477c5-jggzg" Mar 13 15:50:36 crc kubenswrapper[4907]: I0313 15:50:36.593784 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6d59c8ff75-hm7rf"] Mar 13 15:50:36 crc kubenswrapper[4907]: I0313 15:50:36.594026 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6d59c8ff75-hm7rf" podUID="bb294da7-acf1-4984-a401-bbfc15ee97b7" containerName="dnsmasq-dns" containerID="cri-o://04739d4af66a61e29f63286c048d55b12a764eb2ad7e3456b1fff0c9a19ff45a" gracePeriod=10 Mar 13 15:50:36 crc kubenswrapper[4907]: I0313 15:50:36.697973 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-share-share1-0" Mar 13 15:50:36 crc kubenswrapper[4907]: I0313 15:50:36.938553 4907 generic.go:334] "Generic (PLEG): container finished" podID="bb294da7-acf1-4984-a401-bbfc15ee97b7" containerID="04739d4af66a61e29f63286c048d55b12a764eb2ad7e3456b1fff0c9a19ff45a" exitCode=0 Mar 13 15:50:36 crc kubenswrapper[4907]: I0313 15:50:36.938615 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d59c8ff75-hm7rf" event={"ID":"bb294da7-acf1-4984-a401-bbfc15ee97b7","Type":"ContainerDied","Data":"04739d4af66a61e29f63286c048d55b12a764eb2ad7e3456b1fff0c9a19ff45a"} Mar 13 15:50:37 crc kubenswrapper[4907]: I0313 15:50:37.273121 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d59c8ff75-hm7rf" Mar 13 15:50:37 crc kubenswrapper[4907]: I0313 15:50:37.399523 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bb294da7-acf1-4984-a401-bbfc15ee97b7-ovsdbserver-nb\") pod \"bb294da7-acf1-4984-a401-bbfc15ee97b7\" (UID: \"bb294da7-acf1-4984-a401-bbfc15ee97b7\") " Mar 13 15:50:37 crc kubenswrapper[4907]: I0313 15:50:37.399732 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bb294da7-acf1-4984-a401-bbfc15ee97b7-dns-svc\") pod \"bb294da7-acf1-4984-a401-bbfc15ee97b7\" (UID: \"bb294da7-acf1-4984-a401-bbfc15ee97b7\") " Mar 13 15:50:37 crc kubenswrapper[4907]: I0313 15:50:37.399794 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jrkzt\" (UniqueName: \"kubernetes.io/projected/bb294da7-acf1-4984-a401-bbfc15ee97b7-kube-api-access-jrkzt\") pod \"bb294da7-acf1-4984-a401-bbfc15ee97b7\" (UID: \"bb294da7-acf1-4984-a401-bbfc15ee97b7\") " Mar 13 15:50:37 crc kubenswrapper[4907]: I0313 15:50:37.399849 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bb294da7-acf1-4984-a401-bbfc15ee97b7-ovsdbserver-sb\") pod \"bb294da7-acf1-4984-a401-bbfc15ee97b7\" (UID: \"bb294da7-acf1-4984-a401-bbfc15ee97b7\") " Mar 13 15:50:37 crc kubenswrapper[4907]: I0313 15:50:37.399906 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bb294da7-acf1-4984-a401-bbfc15ee97b7-config\") pod \"bb294da7-acf1-4984-a401-bbfc15ee97b7\" (UID: \"bb294da7-acf1-4984-a401-bbfc15ee97b7\") " Mar 13 15:50:37 crc kubenswrapper[4907]: I0313 15:50:37.414124 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb294da7-acf1-4984-a401-bbfc15ee97b7-kube-api-access-jrkzt" (OuterVolumeSpecName: "kube-api-access-jrkzt") pod "bb294da7-acf1-4984-a401-bbfc15ee97b7" (UID: "bb294da7-acf1-4984-a401-bbfc15ee97b7"). InnerVolumeSpecName "kube-api-access-jrkzt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:50:37 crc kubenswrapper[4907]: I0313 15:50:37.464761 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bb294da7-acf1-4984-a401-bbfc15ee97b7-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "bb294da7-acf1-4984-a401-bbfc15ee97b7" (UID: "bb294da7-acf1-4984-a401-bbfc15ee97b7"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:50:37 crc kubenswrapper[4907]: I0313 15:50:37.475615 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bb294da7-acf1-4984-a401-bbfc15ee97b7-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "bb294da7-acf1-4984-a401-bbfc15ee97b7" (UID: "bb294da7-acf1-4984-a401-bbfc15ee97b7"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:50:37 crc kubenswrapper[4907]: I0313 15:50:37.479276 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bb294da7-acf1-4984-a401-bbfc15ee97b7-config" (OuterVolumeSpecName: "config") pod "bb294da7-acf1-4984-a401-bbfc15ee97b7" (UID: "bb294da7-acf1-4984-a401-bbfc15ee97b7"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:50:37 crc kubenswrapper[4907]: I0313 15:50:37.505021 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bb294da7-acf1-4984-a401-bbfc15ee97b7-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "bb294da7-acf1-4984-a401-bbfc15ee97b7" (UID: "bb294da7-acf1-4984-a401-bbfc15ee97b7"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:50:37 crc kubenswrapper[4907]: I0313 15:50:37.505910 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bb294da7-acf1-4984-a401-bbfc15ee97b7-config\") on node \"crc\" DevicePath \"\"" Mar 13 15:50:37 crc kubenswrapper[4907]: I0313 15:50:37.505940 4907 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bb294da7-acf1-4984-a401-bbfc15ee97b7-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 13 15:50:37 crc kubenswrapper[4907]: I0313 15:50:37.505953 4907 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bb294da7-acf1-4984-a401-bbfc15ee97b7-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 13 15:50:37 crc kubenswrapper[4907]: I0313 15:50:37.505965 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jrkzt\" (UniqueName: \"kubernetes.io/projected/bb294da7-acf1-4984-a401-bbfc15ee97b7-kube-api-access-jrkzt\") on node \"crc\" DevicePath \"\"" Mar 13 15:50:37 crc kubenswrapper[4907]: I0313 15:50:37.505977 4907 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bb294da7-acf1-4984-a401-bbfc15ee97b7-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 13 15:50:37 crc kubenswrapper[4907]: I0313 15:50:37.797664 4907 scope.go:117] "RemoveContainer" containerID="3d3465a0228af3b611e5ca1d2d283d5739801f17bcfa78429f88dbe78da94add" Mar 13 15:50:37 crc kubenswrapper[4907]: I0313 15:50:37.831861 4907 scope.go:117] "RemoveContainer" containerID="219cbe1415b39a75ce8171bb61ada0538ae87eb0c8b470a088b03fdc9289f0b1" Mar 13 15:50:37 crc kubenswrapper[4907]: I0313 15:50:37.885532 4907 scope.go:117] "RemoveContainer" containerID="0948541af6cc99af21bd6ee88e49185bc604c42dbf31f8661316b429337a2c7f" Mar 13 15:50:37 crc kubenswrapper[4907]: I0313 15:50:37.928438 4907 scope.go:117] "RemoveContainer" containerID="7157d1dc97b867b0493ef78bf3ed0dea2af629c4ae1bace22c4fa35387dfbe4f" Mar 13 15:50:37 crc kubenswrapper[4907]: I0313 15:50:37.955738 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d59c8ff75-hm7rf" Mar 13 15:50:37 crc kubenswrapper[4907]: I0313 15:50:37.956954 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d59c8ff75-hm7rf" event={"ID":"bb294da7-acf1-4984-a401-bbfc15ee97b7","Type":"ContainerDied","Data":"062befe692770a942fb75e8e6fd540c55d77667654c81f8c1cedf1f5d48f5908"} Mar 13 15:50:37 crc kubenswrapper[4907]: I0313 15:50:37.957030 4907 scope.go:117] "RemoveContainer" containerID="04739d4af66a61e29f63286c048d55b12a764eb2ad7e3456b1fff0c9a19ff45a" Mar 13 15:50:38 crc kubenswrapper[4907]: I0313 15:50:38.000787 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6d59c8ff75-hm7rf"] Mar 13 15:50:38 crc kubenswrapper[4907]: I0313 15:50:38.025897 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6d59c8ff75-hm7rf"] Mar 13 15:50:38 crc kubenswrapper[4907]: I0313 15:50:38.091201 4907 scope.go:117] "RemoveContainer" containerID="04739d4af66a61e29f63286c048d55b12a764eb2ad7e3456b1fff0c9a19ff45a" Mar 13 15:50:38 crc kubenswrapper[4907]: E0313 15:50:38.107055 4907 log.go:32] "RemoveContainer from runtime service failed" err="rpc error: code = Unknown desc = failed to delete container k8s_dnsmasq-dns_dnsmasq-dns-6d59c8ff75-hm7rf_openstack_bb294da7-acf1-4984-a401-bbfc15ee97b7_0 in pod sandbox 062befe692770a942fb75e8e6fd540c55d77667654c81f8c1cedf1f5d48f5908 from index: no such id: '04739d4af66a61e29f63286c048d55b12a764eb2ad7e3456b1fff0c9a19ff45a'" containerID="04739d4af66a61e29f63286c048d55b12a764eb2ad7e3456b1fff0c9a19ff45a" Mar 13 15:50:38 crc kubenswrapper[4907]: E0313 15:50:38.107111 4907 kuberuntime_gc.go:150] "Failed to remove container" err="rpc error: code = Unknown desc = failed to delete container k8s_dnsmasq-dns_dnsmasq-dns-6d59c8ff75-hm7rf_openstack_bb294da7-acf1-4984-a401-bbfc15ee97b7_0 in pod sandbox 062befe692770a942fb75e8e6fd540c55d77667654c81f8c1cedf1f5d48f5908 from index: no such id: '04739d4af66a61e29f63286c048d55b12a764eb2ad7e3456b1fff0c9a19ff45a'" containerID="04739d4af66a61e29f63286c048d55b12a764eb2ad7e3456b1fff0c9a19ff45a" Mar 13 15:50:38 crc kubenswrapper[4907]: I0313 15:50:38.107140 4907 scope.go:117] "RemoveContainer" containerID="e0422eb37278e23b8d6a6f631b14d7a520f0f17ab9fd9d3ee373391ac6276186" Mar 13 15:50:38 crc kubenswrapper[4907]: I0313 15:50:38.155696 4907 scope.go:117] "RemoveContainer" containerID="c3aa022d7c9308a16e2c36ea9708c7b13f760c092c5e791147b871724f8b2c96" Mar 13 15:50:38 crc kubenswrapper[4907]: I0313 15:50:38.184002 4907 scope.go:117] "RemoveContainer" containerID="fea279a9a11b2a38768139c4f75fb40e14428aafd4173f88229129062a569479" Mar 13 15:50:38 crc kubenswrapper[4907]: I0313 15:50:38.209082 4907 scope.go:117] "RemoveContainer" containerID="d3e4c590d97414f576974cec65543e931ac60a3bcf12387dbe9df0a1a39e2c28" Mar 13 15:50:38 crc kubenswrapper[4907]: I0313 15:50:38.237766 4907 scope.go:117] "RemoveContainer" containerID="9d15efc6444fb5cd991e1e807468f40b653afac37426c92c9f3e8218e37197cf" Mar 13 15:50:38 crc kubenswrapper[4907]: I0313 15:50:38.265388 4907 scope.go:117] "RemoveContainer" containerID="0bdd4c179fd6402b48e39e0498ccf620cd42daa3f95e10bf8dc41cdb2673e660" Mar 13 15:50:38 crc kubenswrapper[4907]: I0313 15:50:38.313343 4907 scope.go:117] "RemoveContainer" containerID="2e59f7eb348a0c98bd44ebc16e7f2b9007773ae7074558512509321e58dadcfc" Mar 13 15:50:38 crc kubenswrapper[4907]: I0313 15:50:38.372054 4907 scope.go:117] "RemoveContainer" containerID="04b92be87927c1c66e4f4d3a575acbb3afb714960c495c7fe45aa93f70f86d7e" Mar 13 15:50:38 crc kubenswrapper[4907]: I0313 15:50:38.433295 4907 scope.go:117] "RemoveContainer" containerID="c18542c3b44a6f6eb28a26dfdcfcff6b4ceb92d4f5c1a2a581f9880ab136aa90" Mar 13 15:50:38 crc kubenswrapper[4907]: I0313 15:50:38.488249 4907 scope.go:117] "RemoveContainer" containerID="f18dc1660972c456737f08b933ea252844a2bf45813b5a1dae54e428ebc166a5" Mar 13 15:50:39 crc kubenswrapper[4907]: I0313 15:50:39.273619 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 13 15:50:39 crc kubenswrapper[4907]: I0313 15:50:39.274229 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="14aa8b3a-f849-42c4-9248-2329fd28c6c3" containerName="ceilometer-central-agent" containerID="cri-o://327a65f960da4ed89404c82b3f62aa43c7df17d2c7fdfd45430eda4484d631d9" gracePeriod=30 Mar 13 15:50:39 crc kubenswrapper[4907]: I0313 15:50:39.274363 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="14aa8b3a-f849-42c4-9248-2329fd28c6c3" containerName="proxy-httpd" containerID="cri-o://ea745f91341a94f03f73269b8a99b71eff9cf760adb53889a4160dbc700df09a" gracePeriod=30 Mar 13 15:50:39 crc kubenswrapper[4907]: I0313 15:50:39.274410 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="14aa8b3a-f849-42c4-9248-2329fd28c6c3" containerName="sg-core" containerID="cri-o://d5a88d5e47093bf38e94a5d416f23f1547843a36e587d1615bb482fff4d0a092" gracePeriod=30 Mar 13 15:50:39 crc kubenswrapper[4907]: I0313 15:50:39.274443 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="14aa8b3a-f849-42c4-9248-2329fd28c6c3" containerName="ceilometer-notification-agent" containerID="cri-o://34cd3ac68b3f061e9ab0b1f70d520ca0eebcf6f1d8afd0348c0574488177a64d" gracePeriod=30 Mar 13 15:50:39 crc kubenswrapper[4907]: I0313 15:50:39.798413 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bb294da7-acf1-4984-a401-bbfc15ee97b7" path="/var/lib/kubelet/pods/bb294da7-acf1-4984-a401-bbfc15ee97b7/volumes" Mar 13 15:50:39 crc kubenswrapper[4907]: I0313 15:50:39.994437 4907 generic.go:334] "Generic (PLEG): container finished" podID="14aa8b3a-f849-42c4-9248-2329fd28c6c3" containerID="ea745f91341a94f03f73269b8a99b71eff9cf760adb53889a4160dbc700df09a" exitCode=0 Mar 13 15:50:39 crc kubenswrapper[4907]: I0313 15:50:39.994471 4907 generic.go:334] "Generic (PLEG): container finished" podID="14aa8b3a-f849-42c4-9248-2329fd28c6c3" containerID="d5a88d5e47093bf38e94a5d416f23f1547843a36e587d1615bb482fff4d0a092" exitCode=2 Mar 13 15:50:39 crc kubenswrapper[4907]: I0313 15:50:39.994480 4907 generic.go:334] "Generic (PLEG): container finished" podID="14aa8b3a-f849-42c4-9248-2329fd28c6c3" containerID="34cd3ac68b3f061e9ab0b1f70d520ca0eebcf6f1d8afd0348c0574488177a64d" exitCode=0 Mar 13 15:50:39 crc kubenswrapper[4907]: I0313 15:50:39.994488 4907 generic.go:334] "Generic (PLEG): container finished" podID="14aa8b3a-f849-42c4-9248-2329fd28c6c3" containerID="327a65f960da4ed89404c82b3f62aa43c7df17d2c7fdfd45430eda4484d631d9" exitCode=0 Mar 13 15:50:39 crc kubenswrapper[4907]: I0313 15:50:39.994508 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"14aa8b3a-f849-42c4-9248-2329fd28c6c3","Type":"ContainerDied","Data":"ea745f91341a94f03f73269b8a99b71eff9cf760adb53889a4160dbc700df09a"} Mar 13 15:50:39 crc kubenswrapper[4907]: I0313 15:50:39.994535 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"14aa8b3a-f849-42c4-9248-2329fd28c6c3","Type":"ContainerDied","Data":"d5a88d5e47093bf38e94a5d416f23f1547843a36e587d1615bb482fff4d0a092"} Mar 13 15:50:39 crc kubenswrapper[4907]: I0313 15:50:39.994545 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"14aa8b3a-f849-42c4-9248-2329fd28c6c3","Type":"ContainerDied","Data":"34cd3ac68b3f061e9ab0b1f70d520ca0eebcf6f1d8afd0348c0574488177a64d"} Mar 13 15:50:39 crc kubenswrapper[4907]: I0313 15:50:39.994554 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"14aa8b3a-f849-42c4-9248-2329fd28c6c3","Type":"ContainerDied","Data":"327a65f960da4ed89404c82b3f62aa43c7df17d2c7fdfd45430eda4484d631d9"} Mar 13 15:50:40 crc kubenswrapper[4907]: I0313 15:50:40.242349 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 15:50:40 crc kubenswrapper[4907]: I0313 15:50:40.365048 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14aa8b3a-f849-42c4-9248-2329fd28c6c3-combined-ca-bundle\") pod \"14aa8b3a-f849-42c4-9248-2329fd28c6c3\" (UID: \"14aa8b3a-f849-42c4-9248-2329fd28c6c3\") " Mar 13 15:50:40 crc kubenswrapper[4907]: I0313 15:50:40.365149 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/14aa8b3a-f849-42c4-9248-2329fd28c6c3-scripts\") pod \"14aa8b3a-f849-42c4-9248-2329fd28c6c3\" (UID: \"14aa8b3a-f849-42c4-9248-2329fd28c6c3\") " Mar 13 15:50:40 crc kubenswrapper[4907]: I0313 15:50:40.365237 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/14aa8b3a-f849-42c4-9248-2329fd28c6c3-run-httpd\") pod \"14aa8b3a-f849-42c4-9248-2329fd28c6c3\" (UID: \"14aa8b3a-f849-42c4-9248-2329fd28c6c3\") " Mar 13 15:50:40 crc kubenswrapper[4907]: I0313 15:50:40.365272 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ttn2w\" (UniqueName: \"kubernetes.io/projected/14aa8b3a-f849-42c4-9248-2329fd28c6c3-kube-api-access-ttn2w\") pod \"14aa8b3a-f849-42c4-9248-2329fd28c6c3\" (UID: \"14aa8b3a-f849-42c4-9248-2329fd28c6c3\") " Mar 13 15:50:40 crc kubenswrapper[4907]: I0313 15:50:40.365303 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14aa8b3a-f849-42c4-9248-2329fd28c6c3-config-data\") pod \"14aa8b3a-f849-42c4-9248-2329fd28c6c3\" (UID: \"14aa8b3a-f849-42c4-9248-2329fd28c6c3\") " Mar 13 15:50:40 crc kubenswrapper[4907]: I0313 15:50:40.365326 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/14aa8b3a-f849-42c4-9248-2329fd28c6c3-log-httpd\") pod \"14aa8b3a-f849-42c4-9248-2329fd28c6c3\" (UID: \"14aa8b3a-f849-42c4-9248-2329fd28c6c3\") " Mar 13 15:50:40 crc kubenswrapper[4907]: I0313 15:50:40.365396 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/14aa8b3a-f849-42c4-9248-2329fd28c6c3-sg-core-conf-yaml\") pod \"14aa8b3a-f849-42c4-9248-2329fd28c6c3\" (UID: \"14aa8b3a-f849-42c4-9248-2329fd28c6c3\") " Mar 13 15:50:40 crc kubenswrapper[4907]: I0313 15:50:40.365655 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/14aa8b3a-f849-42c4-9248-2329fd28c6c3-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "14aa8b3a-f849-42c4-9248-2329fd28c6c3" (UID: "14aa8b3a-f849-42c4-9248-2329fd28c6c3"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:50:40 crc kubenswrapper[4907]: I0313 15:50:40.365902 4907 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/14aa8b3a-f849-42c4-9248-2329fd28c6c3-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 13 15:50:40 crc kubenswrapper[4907]: I0313 15:50:40.366297 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/14aa8b3a-f849-42c4-9248-2329fd28c6c3-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "14aa8b3a-f849-42c4-9248-2329fd28c6c3" (UID: "14aa8b3a-f849-42c4-9248-2329fd28c6c3"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:50:40 crc kubenswrapper[4907]: I0313 15:50:40.371542 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/14aa8b3a-f849-42c4-9248-2329fd28c6c3-kube-api-access-ttn2w" (OuterVolumeSpecName: "kube-api-access-ttn2w") pod "14aa8b3a-f849-42c4-9248-2329fd28c6c3" (UID: "14aa8b3a-f849-42c4-9248-2329fd28c6c3"). InnerVolumeSpecName "kube-api-access-ttn2w". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:50:40 crc kubenswrapper[4907]: I0313 15:50:40.374552 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14aa8b3a-f849-42c4-9248-2329fd28c6c3-scripts" (OuterVolumeSpecName: "scripts") pod "14aa8b3a-f849-42c4-9248-2329fd28c6c3" (UID: "14aa8b3a-f849-42c4-9248-2329fd28c6c3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:50:40 crc kubenswrapper[4907]: I0313 15:50:40.417235 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14aa8b3a-f849-42c4-9248-2329fd28c6c3-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "14aa8b3a-f849-42c4-9248-2329fd28c6c3" (UID: "14aa8b3a-f849-42c4-9248-2329fd28c6c3"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:50:40 crc kubenswrapper[4907]: I0313 15:50:40.466433 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14aa8b3a-f849-42c4-9248-2329fd28c6c3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "14aa8b3a-f849-42c4-9248-2329fd28c6c3" (UID: "14aa8b3a-f849-42c4-9248-2329fd28c6c3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:50:40 crc kubenswrapper[4907]: I0313 15:50:40.467367 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14aa8b3a-f849-42c4-9248-2329fd28c6c3-combined-ca-bundle\") pod \"14aa8b3a-f849-42c4-9248-2329fd28c6c3\" (UID: \"14aa8b3a-f849-42c4-9248-2329fd28c6c3\") " Mar 13 15:50:40 crc kubenswrapper[4907]: I0313 15:50:40.467988 4907 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/14aa8b3a-f849-42c4-9248-2329fd28c6c3-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 15:50:40 crc kubenswrapper[4907]: I0313 15:50:40.468012 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ttn2w\" (UniqueName: \"kubernetes.io/projected/14aa8b3a-f849-42c4-9248-2329fd28c6c3-kube-api-access-ttn2w\") on node \"crc\" DevicePath \"\"" Mar 13 15:50:40 crc kubenswrapper[4907]: I0313 15:50:40.468025 4907 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/14aa8b3a-f849-42c4-9248-2329fd28c6c3-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 13 15:50:40 crc kubenswrapper[4907]: I0313 15:50:40.468037 4907 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/14aa8b3a-f849-42c4-9248-2329fd28c6c3-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 13 15:50:40 crc kubenswrapper[4907]: W0313 15:50:40.468134 4907 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/14aa8b3a-f849-42c4-9248-2329fd28c6c3/volumes/kubernetes.io~secret/combined-ca-bundle Mar 13 15:50:40 crc kubenswrapper[4907]: I0313 15:50:40.468149 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14aa8b3a-f849-42c4-9248-2329fd28c6c3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "14aa8b3a-f849-42c4-9248-2329fd28c6c3" (UID: "14aa8b3a-f849-42c4-9248-2329fd28c6c3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:50:40 crc kubenswrapper[4907]: I0313 15:50:40.475688 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14aa8b3a-f849-42c4-9248-2329fd28c6c3-config-data" (OuterVolumeSpecName: "config-data") pod "14aa8b3a-f849-42c4-9248-2329fd28c6c3" (UID: "14aa8b3a-f849-42c4-9248-2329fd28c6c3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:50:40 crc kubenswrapper[4907]: I0313 15:50:40.570158 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14aa8b3a-f849-42c4-9248-2329fd28c6c3-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 15:50:40 crc kubenswrapper[4907]: I0313 15:50:40.570205 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14aa8b3a-f849-42c4-9248-2329fd28c6c3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 15:50:41 crc kubenswrapper[4907]: I0313 15:50:41.007057 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"14aa8b3a-f849-42c4-9248-2329fd28c6c3","Type":"ContainerDied","Data":"9ba05f967433eff9725a6dcc1ca7f6cab5e912da0e5ca7e888033a6d0b137522"} Mar 13 15:50:41 crc kubenswrapper[4907]: I0313 15:50:41.007377 4907 scope.go:117] "RemoveContainer" containerID="ea745f91341a94f03f73269b8a99b71eff9cf760adb53889a4160dbc700df09a" Mar 13 15:50:41 crc kubenswrapper[4907]: I0313 15:50:41.007241 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 15:50:41 crc kubenswrapper[4907]: I0313 15:50:41.035140 4907 scope.go:117] "RemoveContainer" containerID="d5a88d5e47093bf38e94a5d416f23f1547843a36e587d1615bb482fff4d0a092" Mar 13 15:50:41 crc kubenswrapper[4907]: I0313 15:50:41.054051 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 13 15:50:41 crc kubenswrapper[4907]: I0313 15:50:41.065001 4907 scope.go:117] "RemoveContainer" containerID="34cd3ac68b3f061e9ab0b1f70d520ca0eebcf6f1d8afd0348c0574488177a64d" Mar 13 15:50:41 crc kubenswrapper[4907]: I0313 15:50:41.065071 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Mar 13 15:50:41 crc kubenswrapper[4907]: I0313 15:50:41.083904 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 13 15:50:41 crc kubenswrapper[4907]: E0313 15:50:41.084336 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14aa8b3a-f849-42c4-9248-2329fd28c6c3" containerName="ceilometer-central-agent" Mar 13 15:50:41 crc kubenswrapper[4907]: I0313 15:50:41.084353 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="14aa8b3a-f849-42c4-9248-2329fd28c6c3" containerName="ceilometer-central-agent" Mar 13 15:50:41 crc kubenswrapper[4907]: E0313 15:50:41.084414 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14aa8b3a-f849-42c4-9248-2329fd28c6c3" containerName="sg-core" Mar 13 15:50:41 crc kubenswrapper[4907]: I0313 15:50:41.084427 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="14aa8b3a-f849-42c4-9248-2329fd28c6c3" containerName="sg-core" Mar 13 15:50:41 crc kubenswrapper[4907]: E0313 15:50:41.084438 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14aa8b3a-f849-42c4-9248-2329fd28c6c3" containerName="ceilometer-notification-agent" Mar 13 15:50:41 crc kubenswrapper[4907]: I0313 15:50:41.084448 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="14aa8b3a-f849-42c4-9248-2329fd28c6c3" containerName="ceilometer-notification-agent" Mar 13 15:50:41 crc kubenswrapper[4907]: E0313 15:50:41.084464 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14aa8b3a-f849-42c4-9248-2329fd28c6c3" containerName="proxy-httpd" Mar 13 15:50:41 crc kubenswrapper[4907]: I0313 15:50:41.084470 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="14aa8b3a-f849-42c4-9248-2329fd28c6c3" containerName="proxy-httpd" Mar 13 15:50:41 crc kubenswrapper[4907]: E0313 15:50:41.084488 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb294da7-acf1-4984-a401-bbfc15ee97b7" containerName="dnsmasq-dns" Mar 13 15:50:41 crc kubenswrapper[4907]: I0313 15:50:41.084494 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb294da7-acf1-4984-a401-bbfc15ee97b7" containerName="dnsmasq-dns" Mar 13 15:50:41 crc kubenswrapper[4907]: E0313 15:50:41.084525 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb294da7-acf1-4984-a401-bbfc15ee97b7" containerName="init" Mar 13 15:50:41 crc kubenswrapper[4907]: I0313 15:50:41.084532 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb294da7-acf1-4984-a401-bbfc15ee97b7" containerName="init" Mar 13 15:50:41 crc kubenswrapper[4907]: I0313 15:50:41.084758 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="14aa8b3a-f849-42c4-9248-2329fd28c6c3" containerName="ceilometer-notification-agent" Mar 13 15:50:41 crc kubenswrapper[4907]: I0313 15:50:41.084771 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="14aa8b3a-f849-42c4-9248-2329fd28c6c3" containerName="proxy-httpd" Mar 13 15:50:41 crc kubenswrapper[4907]: I0313 15:50:41.084788 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="14aa8b3a-f849-42c4-9248-2329fd28c6c3" containerName="sg-core" Mar 13 15:50:41 crc kubenswrapper[4907]: I0313 15:50:41.084801 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb294da7-acf1-4984-a401-bbfc15ee97b7" containerName="dnsmasq-dns" Mar 13 15:50:41 crc kubenswrapper[4907]: I0313 15:50:41.084810 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="14aa8b3a-f849-42c4-9248-2329fd28c6c3" containerName="ceilometer-central-agent" Mar 13 15:50:41 crc kubenswrapper[4907]: I0313 15:50:41.088001 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 15:50:41 crc kubenswrapper[4907]: I0313 15:50:41.090097 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 13 15:50:41 crc kubenswrapper[4907]: I0313 15:50:41.092850 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 13 15:50:41 crc kubenswrapper[4907]: I0313 15:50:41.097424 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 13 15:50:41 crc kubenswrapper[4907]: I0313 15:50:41.103642 4907 scope.go:117] "RemoveContainer" containerID="327a65f960da4ed89404c82b3f62aa43c7df17d2c7fdfd45430eda4484d631d9" Mar 13 15:50:41 crc kubenswrapper[4907]: I0313 15:50:41.185511 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/59fb2cd9-5414-4356-9438-d33f90ad084c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"59fb2cd9-5414-4356-9438-d33f90ad084c\") " pod="openstack/ceilometer-0" Mar 13 15:50:41 crc kubenswrapper[4907]: I0313 15:50:41.185559 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59fb2cd9-5414-4356-9438-d33f90ad084c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"59fb2cd9-5414-4356-9438-d33f90ad084c\") " pod="openstack/ceilometer-0" Mar 13 15:50:41 crc kubenswrapper[4907]: I0313 15:50:41.185660 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/59fb2cd9-5414-4356-9438-d33f90ad084c-scripts\") pod \"ceilometer-0\" (UID: \"59fb2cd9-5414-4356-9438-d33f90ad084c\") " pod="openstack/ceilometer-0" Mar 13 15:50:41 crc kubenswrapper[4907]: I0313 15:50:41.185683 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/59fb2cd9-5414-4356-9438-d33f90ad084c-log-httpd\") pod \"ceilometer-0\" (UID: \"59fb2cd9-5414-4356-9438-d33f90ad084c\") " pod="openstack/ceilometer-0" Mar 13 15:50:41 crc kubenswrapper[4907]: I0313 15:50:41.185699 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/59fb2cd9-5414-4356-9438-d33f90ad084c-config-data\") pod \"ceilometer-0\" (UID: \"59fb2cd9-5414-4356-9438-d33f90ad084c\") " pod="openstack/ceilometer-0" Mar 13 15:50:41 crc kubenswrapper[4907]: I0313 15:50:41.185789 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/59fb2cd9-5414-4356-9438-d33f90ad084c-run-httpd\") pod \"ceilometer-0\" (UID: \"59fb2cd9-5414-4356-9438-d33f90ad084c\") " pod="openstack/ceilometer-0" Mar 13 15:50:41 crc kubenswrapper[4907]: I0313 15:50:41.185828 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-svq8j\" (UniqueName: \"kubernetes.io/projected/59fb2cd9-5414-4356-9438-d33f90ad084c-kube-api-access-svq8j\") pod \"ceilometer-0\" (UID: \"59fb2cd9-5414-4356-9438-d33f90ad084c\") " pod="openstack/ceilometer-0" Mar 13 15:50:41 crc kubenswrapper[4907]: I0313 15:50:41.287607 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/59fb2cd9-5414-4356-9438-d33f90ad084c-run-httpd\") pod \"ceilometer-0\" (UID: \"59fb2cd9-5414-4356-9438-d33f90ad084c\") " pod="openstack/ceilometer-0" Mar 13 15:50:41 crc kubenswrapper[4907]: I0313 15:50:41.287667 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-svq8j\" (UniqueName: \"kubernetes.io/projected/59fb2cd9-5414-4356-9438-d33f90ad084c-kube-api-access-svq8j\") pod \"ceilometer-0\" (UID: \"59fb2cd9-5414-4356-9438-d33f90ad084c\") " pod="openstack/ceilometer-0" Mar 13 15:50:41 crc kubenswrapper[4907]: I0313 15:50:41.287801 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/59fb2cd9-5414-4356-9438-d33f90ad084c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"59fb2cd9-5414-4356-9438-d33f90ad084c\") " pod="openstack/ceilometer-0" Mar 13 15:50:41 crc kubenswrapper[4907]: I0313 15:50:41.287824 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59fb2cd9-5414-4356-9438-d33f90ad084c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"59fb2cd9-5414-4356-9438-d33f90ad084c\") " pod="openstack/ceilometer-0" Mar 13 15:50:41 crc kubenswrapper[4907]: I0313 15:50:41.287921 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/59fb2cd9-5414-4356-9438-d33f90ad084c-scripts\") pod \"ceilometer-0\" (UID: \"59fb2cd9-5414-4356-9438-d33f90ad084c\") " pod="openstack/ceilometer-0" Mar 13 15:50:41 crc kubenswrapper[4907]: I0313 15:50:41.287951 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/59fb2cd9-5414-4356-9438-d33f90ad084c-log-httpd\") pod \"ceilometer-0\" (UID: \"59fb2cd9-5414-4356-9438-d33f90ad084c\") " pod="openstack/ceilometer-0" Mar 13 15:50:41 crc kubenswrapper[4907]: I0313 15:50:41.287973 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/59fb2cd9-5414-4356-9438-d33f90ad084c-config-data\") pod \"ceilometer-0\" (UID: \"59fb2cd9-5414-4356-9438-d33f90ad084c\") " pod="openstack/ceilometer-0" Mar 13 15:50:41 crc kubenswrapper[4907]: I0313 15:50:41.288868 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/59fb2cd9-5414-4356-9438-d33f90ad084c-run-httpd\") pod \"ceilometer-0\" (UID: \"59fb2cd9-5414-4356-9438-d33f90ad084c\") " pod="openstack/ceilometer-0" Mar 13 15:50:41 crc kubenswrapper[4907]: I0313 15:50:41.289309 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/59fb2cd9-5414-4356-9438-d33f90ad084c-log-httpd\") pod \"ceilometer-0\" (UID: \"59fb2cd9-5414-4356-9438-d33f90ad084c\") " pod="openstack/ceilometer-0" Mar 13 15:50:41 crc kubenswrapper[4907]: I0313 15:50:41.293861 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/59fb2cd9-5414-4356-9438-d33f90ad084c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"59fb2cd9-5414-4356-9438-d33f90ad084c\") " pod="openstack/ceilometer-0" Mar 13 15:50:41 crc kubenswrapper[4907]: I0313 15:50:41.293977 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59fb2cd9-5414-4356-9438-d33f90ad084c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"59fb2cd9-5414-4356-9438-d33f90ad084c\") " pod="openstack/ceilometer-0" Mar 13 15:50:41 crc kubenswrapper[4907]: I0313 15:50:41.294384 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/59fb2cd9-5414-4356-9438-d33f90ad084c-scripts\") pod \"ceilometer-0\" (UID: \"59fb2cd9-5414-4356-9438-d33f90ad084c\") " pod="openstack/ceilometer-0" Mar 13 15:50:41 crc kubenswrapper[4907]: I0313 15:50:41.298077 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/59fb2cd9-5414-4356-9438-d33f90ad084c-config-data\") pod \"ceilometer-0\" (UID: \"59fb2cd9-5414-4356-9438-d33f90ad084c\") " pod="openstack/ceilometer-0" Mar 13 15:50:41 crc kubenswrapper[4907]: I0313 15:50:41.307421 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-svq8j\" (UniqueName: \"kubernetes.io/projected/59fb2cd9-5414-4356-9438-d33f90ad084c-kube-api-access-svq8j\") pod \"ceilometer-0\" (UID: \"59fb2cd9-5414-4356-9438-d33f90ad084c\") " pod="openstack/ceilometer-0" Mar 13 15:50:41 crc kubenswrapper[4907]: I0313 15:50:41.406470 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 15:50:41 crc kubenswrapper[4907]: I0313 15:50:41.794489 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="14aa8b3a-f849-42c4-9248-2329fd28c6c3" path="/var/lib/kubelet/pods/14aa8b3a-f849-42c4-9248-2329fd28c6c3/volumes" Mar 13 15:50:41 crc kubenswrapper[4907]: I0313 15:50:41.981682 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 13 15:50:42 crc kubenswrapper[4907]: I0313 15:50:42.017578 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"59fb2cd9-5414-4356-9438-d33f90ad084c","Type":"ContainerStarted","Data":"0df3568657a2601e9a878c56a11954c34f6ef4dcfe1b9a5d4f12772552e8c314"} Mar 13 15:50:43 crc kubenswrapper[4907]: I0313 15:50:43.028190 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"59fb2cd9-5414-4356-9438-d33f90ad084c","Type":"ContainerStarted","Data":"7eea587aa87adf598d468a3ecc98d2f4ada4de324f142b4337f94eaa09b7c6ad"} Mar 13 15:50:44 crc kubenswrapper[4907]: I0313 15:50:44.049831 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"59fb2cd9-5414-4356-9438-d33f90ad084c","Type":"ContainerStarted","Data":"71de4d49169f6bb2dd395e4b26e3d015c4d648d12080086a842e391f8eef4e66"} Mar 13 15:50:45 crc kubenswrapper[4907]: I0313 15:50:45.062667 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"59fb2cd9-5414-4356-9438-d33f90ad084c","Type":"ContainerStarted","Data":"351dde2349b18437c873dc5ab10acd990453b7071ad53b6cbdc2cd3f577a83c6"} Mar 13 15:50:47 crc kubenswrapper[4907]: I0313 15:50:47.123568 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"59fb2cd9-5414-4356-9438-d33f90ad084c","Type":"ContainerStarted","Data":"6df1468393b247353225c6e4ccd83f9a2b2fb80610a9be665a4cea857f2ce364"} Mar 13 15:50:47 crc kubenswrapper[4907]: I0313 15:50:47.125209 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Mar 13 15:50:47 crc kubenswrapper[4907]: I0313 15:50:47.155815 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.097447519 podStartE2EDuration="6.155794699s" podCreationTimestamp="2026-03-13 15:50:41 +0000 UTC" firstStartedPulling="2026-03-13 15:50:41.973807868 +0000 UTC m=+6340.873595557" lastFinishedPulling="2026-03-13 15:50:46.032155048 +0000 UTC m=+6344.931942737" observedRunningTime="2026-03-13 15:50:47.152827778 +0000 UTC m=+6346.052615467" watchObservedRunningTime="2026-03-13 15:50:47.155794699 +0000 UTC m=+6346.055582398" Mar 13 15:50:48 crc kubenswrapper[4907]: I0313 15:50:48.049269 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-scheduler-0" Mar 13 15:50:48 crc kubenswrapper[4907]: I0313 15:50:48.597577 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/manila-api-0" Mar 13 15:50:48 crc kubenswrapper[4907]: I0313 15:50:48.723556 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-share-share1-0" Mar 13 15:51:11 crc kubenswrapper[4907]: I0313 15:51:11.414322 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Mar 13 15:51:16 crc kubenswrapper[4907]: I0313 15:51:16.055611 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-9290-account-create-update-qws42"] Mar 13 15:51:16 crc kubenswrapper[4907]: I0313 15:51:16.069123 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-9290-account-create-update-qws42"] Mar 13 15:51:17 crc kubenswrapper[4907]: I0313 15:51:17.029814 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-pvpwm"] Mar 13 15:51:17 crc kubenswrapper[4907]: I0313 15:51:17.043383 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-pvpwm"] Mar 13 15:51:17 crc kubenswrapper[4907]: I0313 15:51:17.795569 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0ab8c7fa-9565-4b6c-b8f8-033cfdf444db" path="/var/lib/kubelet/pods/0ab8c7fa-9565-4b6c-b8f8-033cfdf444db/volumes" Mar 13 15:51:17 crc kubenswrapper[4907]: I0313 15:51:17.796846 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="499bc96f-8fa9-4900-bd92-2a507215791f" path="/var/lib/kubelet/pods/499bc96f-8fa9-4900-bd92-2a507215791f/volumes" Mar 13 15:51:24 crc kubenswrapper[4907]: I0313 15:51:24.046265 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-hv2mf"] Mar 13 15:51:24 crc kubenswrapper[4907]: I0313 15:51:24.063745 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-hv2mf"] Mar 13 15:51:25 crc kubenswrapper[4907]: I0313 15:51:25.793178 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ef105bad-6763-4234-b52b-6d2820d48b02" path="/var/lib/kubelet/pods/ef105bad-6763-4234-b52b-6d2820d48b02/volumes" Mar 13 15:51:32 crc kubenswrapper[4907]: I0313 15:51:32.302145 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5ff7fb8bcc-45q4w"] Mar 13 15:51:32 crc kubenswrapper[4907]: I0313 15:51:32.330610 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5ff7fb8bcc-45q4w"] Mar 13 15:51:32 crc kubenswrapper[4907]: I0313 15:51:32.330756 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ff7fb8bcc-45q4w" Mar 13 15:51:32 crc kubenswrapper[4907]: I0313 15:51:32.337734 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1" Mar 13 15:51:32 crc kubenswrapper[4907]: I0313 15:51:32.431108 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/68d11f2a-37b6-4a6c-b877-ae32e177761d-ovsdbserver-sb\") pod \"dnsmasq-dns-5ff7fb8bcc-45q4w\" (UID: \"68d11f2a-37b6-4a6c-b877-ae32e177761d\") " pod="openstack/dnsmasq-dns-5ff7fb8bcc-45q4w" Mar 13 15:51:32 crc kubenswrapper[4907]: I0313 15:51:32.431633 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/68d11f2a-37b6-4a6c-b877-ae32e177761d-dns-svc\") pod \"dnsmasq-dns-5ff7fb8bcc-45q4w\" (UID: \"68d11f2a-37b6-4a6c-b877-ae32e177761d\") " pod="openstack/dnsmasq-dns-5ff7fb8bcc-45q4w" Mar 13 15:51:32 crc kubenswrapper[4907]: I0313 15:51:32.431861 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vww7s\" (UniqueName: \"kubernetes.io/projected/68d11f2a-37b6-4a6c-b877-ae32e177761d-kube-api-access-vww7s\") pod \"dnsmasq-dns-5ff7fb8bcc-45q4w\" (UID: \"68d11f2a-37b6-4a6c-b877-ae32e177761d\") " pod="openstack/dnsmasq-dns-5ff7fb8bcc-45q4w" Mar 13 15:51:32 crc kubenswrapper[4907]: I0313 15:51:32.431974 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/68d11f2a-37b6-4a6c-b877-ae32e177761d-config\") pod \"dnsmasq-dns-5ff7fb8bcc-45q4w\" (UID: \"68d11f2a-37b6-4a6c-b877-ae32e177761d\") " pod="openstack/dnsmasq-dns-5ff7fb8bcc-45q4w" Mar 13 15:51:32 crc kubenswrapper[4907]: I0313 15:51:32.432054 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/68d11f2a-37b6-4a6c-b877-ae32e177761d-ovsdbserver-nb\") pod \"dnsmasq-dns-5ff7fb8bcc-45q4w\" (UID: \"68d11f2a-37b6-4a6c-b877-ae32e177761d\") " pod="openstack/dnsmasq-dns-5ff7fb8bcc-45q4w" Mar 13 15:51:32 crc kubenswrapper[4907]: I0313 15:51:32.432159 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/68d11f2a-37b6-4a6c-b877-ae32e177761d-openstack-cell1\") pod \"dnsmasq-dns-5ff7fb8bcc-45q4w\" (UID: \"68d11f2a-37b6-4a6c-b877-ae32e177761d\") " pod="openstack/dnsmasq-dns-5ff7fb8bcc-45q4w" Mar 13 15:51:32 crc kubenswrapper[4907]: I0313 15:51:32.534983 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vww7s\" (UniqueName: \"kubernetes.io/projected/68d11f2a-37b6-4a6c-b877-ae32e177761d-kube-api-access-vww7s\") pod \"dnsmasq-dns-5ff7fb8bcc-45q4w\" (UID: \"68d11f2a-37b6-4a6c-b877-ae32e177761d\") " pod="openstack/dnsmasq-dns-5ff7fb8bcc-45q4w" Mar 13 15:51:32 crc kubenswrapper[4907]: I0313 15:51:32.535162 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/68d11f2a-37b6-4a6c-b877-ae32e177761d-config\") pod \"dnsmasq-dns-5ff7fb8bcc-45q4w\" (UID: \"68d11f2a-37b6-4a6c-b877-ae32e177761d\") " pod="openstack/dnsmasq-dns-5ff7fb8bcc-45q4w" Mar 13 15:51:32 crc kubenswrapper[4907]: I0313 15:51:32.535239 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/68d11f2a-37b6-4a6c-b877-ae32e177761d-ovsdbserver-nb\") pod \"dnsmasq-dns-5ff7fb8bcc-45q4w\" (UID: \"68d11f2a-37b6-4a6c-b877-ae32e177761d\") " pod="openstack/dnsmasq-dns-5ff7fb8bcc-45q4w" Mar 13 15:51:32 crc kubenswrapper[4907]: I0313 15:51:32.535335 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/68d11f2a-37b6-4a6c-b877-ae32e177761d-openstack-cell1\") pod \"dnsmasq-dns-5ff7fb8bcc-45q4w\" (UID: \"68d11f2a-37b6-4a6c-b877-ae32e177761d\") " pod="openstack/dnsmasq-dns-5ff7fb8bcc-45q4w" Mar 13 15:51:32 crc kubenswrapper[4907]: I0313 15:51:32.535394 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/68d11f2a-37b6-4a6c-b877-ae32e177761d-ovsdbserver-sb\") pod \"dnsmasq-dns-5ff7fb8bcc-45q4w\" (UID: \"68d11f2a-37b6-4a6c-b877-ae32e177761d\") " pod="openstack/dnsmasq-dns-5ff7fb8bcc-45q4w" Mar 13 15:51:32 crc kubenswrapper[4907]: I0313 15:51:32.535602 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/68d11f2a-37b6-4a6c-b877-ae32e177761d-dns-svc\") pod \"dnsmasq-dns-5ff7fb8bcc-45q4w\" (UID: \"68d11f2a-37b6-4a6c-b877-ae32e177761d\") " pod="openstack/dnsmasq-dns-5ff7fb8bcc-45q4w" Mar 13 15:51:32 crc kubenswrapper[4907]: I0313 15:51:32.536340 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/68d11f2a-37b6-4a6c-b877-ae32e177761d-ovsdbserver-nb\") pod \"dnsmasq-dns-5ff7fb8bcc-45q4w\" (UID: \"68d11f2a-37b6-4a6c-b877-ae32e177761d\") " pod="openstack/dnsmasq-dns-5ff7fb8bcc-45q4w" Mar 13 15:51:32 crc kubenswrapper[4907]: I0313 15:51:32.536455 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/68d11f2a-37b6-4a6c-b877-ae32e177761d-openstack-cell1\") pod \"dnsmasq-dns-5ff7fb8bcc-45q4w\" (UID: \"68d11f2a-37b6-4a6c-b877-ae32e177761d\") " pod="openstack/dnsmasq-dns-5ff7fb8bcc-45q4w" Mar 13 15:51:32 crc kubenswrapper[4907]: I0313 15:51:32.536491 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/68d11f2a-37b6-4a6c-b877-ae32e177761d-ovsdbserver-sb\") pod \"dnsmasq-dns-5ff7fb8bcc-45q4w\" (UID: \"68d11f2a-37b6-4a6c-b877-ae32e177761d\") " pod="openstack/dnsmasq-dns-5ff7fb8bcc-45q4w" Mar 13 15:51:32 crc kubenswrapper[4907]: I0313 15:51:32.536741 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/68d11f2a-37b6-4a6c-b877-ae32e177761d-dns-svc\") pod \"dnsmasq-dns-5ff7fb8bcc-45q4w\" (UID: \"68d11f2a-37b6-4a6c-b877-ae32e177761d\") " pod="openstack/dnsmasq-dns-5ff7fb8bcc-45q4w" Mar 13 15:51:32 crc kubenswrapper[4907]: I0313 15:51:32.537094 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/68d11f2a-37b6-4a6c-b877-ae32e177761d-config\") pod \"dnsmasq-dns-5ff7fb8bcc-45q4w\" (UID: \"68d11f2a-37b6-4a6c-b877-ae32e177761d\") " pod="openstack/dnsmasq-dns-5ff7fb8bcc-45q4w" Mar 13 15:51:32 crc kubenswrapper[4907]: I0313 15:51:32.569430 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vww7s\" (UniqueName: \"kubernetes.io/projected/68d11f2a-37b6-4a6c-b877-ae32e177761d-kube-api-access-vww7s\") pod \"dnsmasq-dns-5ff7fb8bcc-45q4w\" (UID: \"68d11f2a-37b6-4a6c-b877-ae32e177761d\") " pod="openstack/dnsmasq-dns-5ff7fb8bcc-45q4w" Mar 13 15:51:32 crc kubenswrapper[4907]: I0313 15:51:32.649666 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ff7fb8bcc-45q4w" Mar 13 15:51:33 crc kubenswrapper[4907]: I0313 15:51:33.121247 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5ff7fb8bcc-45q4w"] Mar 13 15:51:33 crc kubenswrapper[4907]: I0313 15:51:33.564227 4907 generic.go:334] "Generic (PLEG): container finished" podID="68d11f2a-37b6-4a6c-b877-ae32e177761d" containerID="bb89c1e3b4d030f795279c7977f6bc141b8f573b1af14ac2704526b6da21d04c" exitCode=0 Mar 13 15:51:33 crc kubenswrapper[4907]: I0313 15:51:33.564293 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ff7fb8bcc-45q4w" event={"ID":"68d11f2a-37b6-4a6c-b877-ae32e177761d","Type":"ContainerDied","Data":"bb89c1e3b4d030f795279c7977f6bc141b8f573b1af14ac2704526b6da21d04c"} Mar 13 15:51:33 crc kubenswrapper[4907]: I0313 15:51:33.564351 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ff7fb8bcc-45q4w" event={"ID":"68d11f2a-37b6-4a6c-b877-ae32e177761d","Type":"ContainerStarted","Data":"5b3c0f84402a8ec16bbf16451e1bc1400975871789f821c116960d27711f8335"} Mar 13 15:51:34 crc kubenswrapper[4907]: I0313 15:51:34.577247 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ff7fb8bcc-45q4w" event={"ID":"68d11f2a-37b6-4a6c-b877-ae32e177761d","Type":"ContainerStarted","Data":"997f7ae74c7025a1a37f82e6b1e89fdb84b9a3fd4054caac0d2b4e2d822d549a"} Mar 13 15:51:34 crc kubenswrapper[4907]: I0313 15:51:34.577810 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5ff7fb8bcc-45q4w" Mar 13 15:51:34 crc kubenswrapper[4907]: I0313 15:51:34.594589 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5ff7fb8bcc-45q4w" podStartSLOduration=2.594569538 podStartE2EDuration="2.594569538s" podCreationTimestamp="2026-03-13 15:51:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:51:34.592232055 +0000 UTC m=+6393.492019764" watchObservedRunningTime="2026-03-13 15:51:34.594569538 +0000 UTC m=+6393.494357227" Mar 13 15:51:38 crc kubenswrapper[4907]: I0313 15:51:38.840676 4907 scope.go:117] "RemoveContainer" containerID="5140f9e51d2f4e22927da88ea13b8b76a7a069356e41a43262ca41c166a3779c" Mar 13 15:51:38 crc kubenswrapper[4907]: I0313 15:51:38.875557 4907 scope.go:117] "RemoveContainer" containerID="37e65533c5fc864e730a3583c9b7118a1665874961e44f735e26e376e6b5f39d" Mar 13 15:51:38 crc kubenswrapper[4907]: I0313 15:51:38.929815 4907 scope.go:117] "RemoveContainer" containerID="0559fde0f7d2daa5d36de2042211ef456904ceea0804232b86a43826c351c91c" Mar 13 15:51:38 crc kubenswrapper[4907]: I0313 15:51:38.971598 4907 scope.go:117] "RemoveContainer" containerID="cfb2b66bd790bad49fda00e0dee115de187df8f601026658c19c9100dc1ee877" Mar 13 15:51:42 crc kubenswrapper[4907]: I0313 15:51:42.651101 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5ff7fb8bcc-45q4w" Mar 13 15:51:42 crc kubenswrapper[4907]: I0313 15:51:42.708042 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7559d477c5-jggzg"] Mar 13 15:51:42 crc kubenswrapper[4907]: I0313 15:51:42.708290 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7559d477c5-jggzg" podUID="a01ed54c-beec-437b-a4fb-c4c62a407d76" containerName="dnsmasq-dns" containerID="cri-o://f427c6c6d4cea6920aec5e3d13f7c210e91731dd6a8240743203f4fd1462ee2f" gracePeriod=10 Mar 13 15:51:42 crc kubenswrapper[4907]: I0313 15:51:42.901215 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-8644fdb9df-gm62m"] Mar 13 15:51:42 crc kubenswrapper[4907]: I0313 15:51:42.903073 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8644fdb9df-gm62m" Mar 13 15:51:42 crc kubenswrapper[4907]: I0313 15:51:42.915896 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8644fdb9df-gm62m"] Mar 13 15:51:43 crc kubenswrapper[4907]: I0313 15:51:43.069813 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8v9g6\" (UniqueName: \"kubernetes.io/projected/44fcf282-3688-4a6f-b390-6b321e89b158-kube-api-access-8v9g6\") pod \"dnsmasq-dns-8644fdb9df-gm62m\" (UID: \"44fcf282-3688-4a6f-b390-6b321e89b158\") " pod="openstack/dnsmasq-dns-8644fdb9df-gm62m" Mar 13 15:51:43 crc kubenswrapper[4907]: I0313 15:51:43.069908 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/44fcf282-3688-4a6f-b390-6b321e89b158-ovsdbserver-nb\") pod \"dnsmasq-dns-8644fdb9df-gm62m\" (UID: \"44fcf282-3688-4a6f-b390-6b321e89b158\") " pod="openstack/dnsmasq-dns-8644fdb9df-gm62m" Mar 13 15:51:43 crc kubenswrapper[4907]: I0313 15:51:43.069940 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/44fcf282-3688-4a6f-b390-6b321e89b158-openstack-cell1\") pod \"dnsmasq-dns-8644fdb9df-gm62m\" (UID: \"44fcf282-3688-4a6f-b390-6b321e89b158\") " pod="openstack/dnsmasq-dns-8644fdb9df-gm62m" Mar 13 15:51:43 crc kubenswrapper[4907]: I0313 15:51:43.069962 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/44fcf282-3688-4a6f-b390-6b321e89b158-ovsdbserver-sb\") pod \"dnsmasq-dns-8644fdb9df-gm62m\" (UID: \"44fcf282-3688-4a6f-b390-6b321e89b158\") " pod="openstack/dnsmasq-dns-8644fdb9df-gm62m" Mar 13 15:51:43 crc kubenswrapper[4907]: I0313 15:51:43.070004 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/44fcf282-3688-4a6f-b390-6b321e89b158-dns-svc\") pod \"dnsmasq-dns-8644fdb9df-gm62m\" (UID: \"44fcf282-3688-4a6f-b390-6b321e89b158\") " pod="openstack/dnsmasq-dns-8644fdb9df-gm62m" Mar 13 15:51:43 crc kubenswrapper[4907]: I0313 15:51:43.070053 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/44fcf282-3688-4a6f-b390-6b321e89b158-config\") pod \"dnsmasq-dns-8644fdb9df-gm62m\" (UID: \"44fcf282-3688-4a6f-b390-6b321e89b158\") " pod="openstack/dnsmasq-dns-8644fdb9df-gm62m" Mar 13 15:51:43 crc kubenswrapper[4907]: I0313 15:51:43.172284 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8v9g6\" (UniqueName: \"kubernetes.io/projected/44fcf282-3688-4a6f-b390-6b321e89b158-kube-api-access-8v9g6\") pod \"dnsmasq-dns-8644fdb9df-gm62m\" (UID: \"44fcf282-3688-4a6f-b390-6b321e89b158\") " pod="openstack/dnsmasq-dns-8644fdb9df-gm62m" Mar 13 15:51:43 crc kubenswrapper[4907]: I0313 15:51:43.172665 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/44fcf282-3688-4a6f-b390-6b321e89b158-ovsdbserver-nb\") pod \"dnsmasq-dns-8644fdb9df-gm62m\" (UID: \"44fcf282-3688-4a6f-b390-6b321e89b158\") " pod="openstack/dnsmasq-dns-8644fdb9df-gm62m" Mar 13 15:51:43 crc kubenswrapper[4907]: I0313 15:51:43.172704 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/44fcf282-3688-4a6f-b390-6b321e89b158-openstack-cell1\") pod \"dnsmasq-dns-8644fdb9df-gm62m\" (UID: \"44fcf282-3688-4a6f-b390-6b321e89b158\") " pod="openstack/dnsmasq-dns-8644fdb9df-gm62m" Mar 13 15:51:43 crc kubenswrapper[4907]: I0313 15:51:43.172733 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/44fcf282-3688-4a6f-b390-6b321e89b158-ovsdbserver-sb\") pod \"dnsmasq-dns-8644fdb9df-gm62m\" (UID: \"44fcf282-3688-4a6f-b390-6b321e89b158\") " pod="openstack/dnsmasq-dns-8644fdb9df-gm62m" Mar 13 15:51:43 crc kubenswrapper[4907]: I0313 15:51:43.172782 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/44fcf282-3688-4a6f-b390-6b321e89b158-dns-svc\") pod \"dnsmasq-dns-8644fdb9df-gm62m\" (UID: \"44fcf282-3688-4a6f-b390-6b321e89b158\") " pod="openstack/dnsmasq-dns-8644fdb9df-gm62m" Mar 13 15:51:43 crc kubenswrapper[4907]: I0313 15:51:43.172844 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/44fcf282-3688-4a6f-b390-6b321e89b158-config\") pod \"dnsmasq-dns-8644fdb9df-gm62m\" (UID: \"44fcf282-3688-4a6f-b390-6b321e89b158\") " pod="openstack/dnsmasq-dns-8644fdb9df-gm62m" Mar 13 15:51:43 crc kubenswrapper[4907]: I0313 15:51:43.174004 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/44fcf282-3688-4a6f-b390-6b321e89b158-config\") pod \"dnsmasq-dns-8644fdb9df-gm62m\" (UID: \"44fcf282-3688-4a6f-b390-6b321e89b158\") " pod="openstack/dnsmasq-dns-8644fdb9df-gm62m" Mar 13 15:51:43 crc kubenswrapper[4907]: I0313 15:51:43.175974 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/44fcf282-3688-4a6f-b390-6b321e89b158-ovsdbserver-sb\") pod \"dnsmasq-dns-8644fdb9df-gm62m\" (UID: \"44fcf282-3688-4a6f-b390-6b321e89b158\") " pod="openstack/dnsmasq-dns-8644fdb9df-gm62m" Mar 13 15:51:43 crc kubenswrapper[4907]: I0313 15:51:43.176696 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/44fcf282-3688-4a6f-b390-6b321e89b158-openstack-cell1\") pod \"dnsmasq-dns-8644fdb9df-gm62m\" (UID: \"44fcf282-3688-4a6f-b390-6b321e89b158\") " pod="openstack/dnsmasq-dns-8644fdb9df-gm62m" Mar 13 15:51:43 crc kubenswrapper[4907]: I0313 15:51:43.177491 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/44fcf282-3688-4a6f-b390-6b321e89b158-dns-svc\") pod \"dnsmasq-dns-8644fdb9df-gm62m\" (UID: \"44fcf282-3688-4a6f-b390-6b321e89b158\") " pod="openstack/dnsmasq-dns-8644fdb9df-gm62m" Mar 13 15:51:43 crc kubenswrapper[4907]: I0313 15:51:43.179942 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/44fcf282-3688-4a6f-b390-6b321e89b158-ovsdbserver-nb\") pod \"dnsmasq-dns-8644fdb9df-gm62m\" (UID: \"44fcf282-3688-4a6f-b390-6b321e89b158\") " pod="openstack/dnsmasq-dns-8644fdb9df-gm62m" Mar 13 15:51:43 crc kubenswrapper[4907]: I0313 15:51:43.193759 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8v9g6\" (UniqueName: \"kubernetes.io/projected/44fcf282-3688-4a6f-b390-6b321e89b158-kube-api-access-8v9g6\") pod \"dnsmasq-dns-8644fdb9df-gm62m\" (UID: \"44fcf282-3688-4a6f-b390-6b321e89b158\") " pod="openstack/dnsmasq-dns-8644fdb9df-gm62m" Mar 13 15:51:43 crc kubenswrapper[4907]: I0313 15:51:43.279126 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8644fdb9df-gm62m" Mar 13 15:51:43 crc kubenswrapper[4907]: I0313 15:51:43.402561 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7559d477c5-jggzg" Mar 13 15:51:43 crc kubenswrapper[4907]: I0313 15:51:43.481708 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a01ed54c-beec-437b-a4fb-c4c62a407d76-dns-svc\") pod \"a01ed54c-beec-437b-a4fb-c4c62a407d76\" (UID: \"a01ed54c-beec-437b-a4fb-c4c62a407d76\") " Mar 13 15:51:43 crc kubenswrapper[4907]: I0313 15:51:43.481777 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bnxr5\" (UniqueName: \"kubernetes.io/projected/a01ed54c-beec-437b-a4fb-c4c62a407d76-kube-api-access-bnxr5\") pod \"a01ed54c-beec-437b-a4fb-c4c62a407d76\" (UID: \"a01ed54c-beec-437b-a4fb-c4c62a407d76\") " Mar 13 15:51:43 crc kubenswrapper[4907]: I0313 15:51:43.482526 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a01ed54c-beec-437b-a4fb-c4c62a407d76-config\") pod \"a01ed54c-beec-437b-a4fb-c4c62a407d76\" (UID: \"a01ed54c-beec-437b-a4fb-c4c62a407d76\") " Mar 13 15:51:43 crc kubenswrapper[4907]: I0313 15:51:43.482624 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a01ed54c-beec-437b-a4fb-c4c62a407d76-ovsdbserver-sb\") pod \"a01ed54c-beec-437b-a4fb-c4c62a407d76\" (UID: \"a01ed54c-beec-437b-a4fb-c4c62a407d76\") " Mar 13 15:51:43 crc kubenswrapper[4907]: I0313 15:51:43.483095 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a01ed54c-beec-437b-a4fb-c4c62a407d76-ovsdbserver-nb\") pod \"a01ed54c-beec-437b-a4fb-c4c62a407d76\" (UID: \"a01ed54c-beec-437b-a4fb-c4c62a407d76\") " Mar 13 15:51:43 crc kubenswrapper[4907]: I0313 15:51:43.664939 4907 generic.go:334] "Generic (PLEG): container finished" podID="a01ed54c-beec-437b-a4fb-c4c62a407d76" containerID="f427c6c6d4cea6920aec5e3d13f7c210e91731dd6a8240743203f4fd1462ee2f" exitCode=0 Mar 13 15:51:43 crc kubenswrapper[4907]: I0313 15:51:43.664983 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7559d477c5-jggzg" Mar 13 15:51:43 crc kubenswrapper[4907]: I0313 15:51:43.664986 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7559d477c5-jggzg" event={"ID":"a01ed54c-beec-437b-a4fb-c4c62a407d76","Type":"ContainerDied","Data":"f427c6c6d4cea6920aec5e3d13f7c210e91731dd6a8240743203f4fd1462ee2f"} Mar 13 15:51:43 crc kubenswrapper[4907]: I0313 15:51:43.665017 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7559d477c5-jggzg" event={"ID":"a01ed54c-beec-437b-a4fb-c4c62a407d76","Type":"ContainerDied","Data":"9777020954fe435146b223f1e22d2da8657fcd1b5305e36c2273a7c3f9c349b1"} Mar 13 15:51:43 crc kubenswrapper[4907]: I0313 15:51:43.665040 4907 scope.go:117] "RemoveContainer" containerID="f427c6c6d4cea6920aec5e3d13f7c210e91731dd6a8240743203f4fd1462ee2f" Mar 13 15:51:44 crc kubenswrapper[4907]: I0313 15:51:44.093608 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a01ed54c-beec-437b-a4fb-c4c62a407d76-kube-api-access-bnxr5" (OuterVolumeSpecName: "kube-api-access-bnxr5") pod "a01ed54c-beec-437b-a4fb-c4c62a407d76" (UID: "a01ed54c-beec-437b-a4fb-c4c62a407d76"). InnerVolumeSpecName "kube-api-access-bnxr5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:51:44 crc kubenswrapper[4907]: I0313 15:51:44.098764 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bnxr5\" (UniqueName: \"kubernetes.io/projected/a01ed54c-beec-437b-a4fb-c4c62a407d76-kube-api-access-bnxr5\") on node \"crc\" DevicePath \"\"" Mar 13 15:51:44 crc kubenswrapper[4907]: I0313 15:51:44.147296 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a01ed54c-beec-437b-a4fb-c4c62a407d76-config" (OuterVolumeSpecName: "config") pod "a01ed54c-beec-437b-a4fb-c4c62a407d76" (UID: "a01ed54c-beec-437b-a4fb-c4c62a407d76"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:51:44 crc kubenswrapper[4907]: I0313 15:51:44.149419 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a01ed54c-beec-437b-a4fb-c4c62a407d76-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "a01ed54c-beec-437b-a4fb-c4c62a407d76" (UID: "a01ed54c-beec-437b-a4fb-c4c62a407d76"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:51:44 crc kubenswrapper[4907]: I0313 15:51:44.165280 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a01ed54c-beec-437b-a4fb-c4c62a407d76-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a01ed54c-beec-437b-a4fb-c4c62a407d76" (UID: "a01ed54c-beec-437b-a4fb-c4c62a407d76"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:51:44 crc kubenswrapper[4907]: I0313 15:51:44.176674 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a01ed54c-beec-437b-a4fb-c4c62a407d76-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "a01ed54c-beec-437b-a4fb-c4c62a407d76" (UID: "a01ed54c-beec-437b-a4fb-c4c62a407d76"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:51:44 crc kubenswrapper[4907]: I0313 15:51:44.200653 4907 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a01ed54c-beec-437b-a4fb-c4c62a407d76-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 13 15:51:44 crc kubenswrapper[4907]: I0313 15:51:44.200698 4907 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a01ed54c-beec-437b-a4fb-c4c62a407d76-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 13 15:51:44 crc kubenswrapper[4907]: I0313 15:51:44.200708 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a01ed54c-beec-437b-a4fb-c4c62a407d76-config\") on node \"crc\" DevicePath \"\"" Mar 13 15:51:44 crc kubenswrapper[4907]: I0313 15:51:44.200719 4907 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a01ed54c-beec-437b-a4fb-c4c62a407d76-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 13 15:51:44 crc kubenswrapper[4907]: I0313 15:51:44.305541 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8644fdb9df-gm62m"] Mar 13 15:51:44 crc kubenswrapper[4907]: I0313 15:51:44.334644 4907 scope.go:117] "RemoveContainer" containerID="6a9689f3bb28f496b72b2c47ce88ac358b381c9e951286ab8db719b6cd0bd636" Mar 13 15:51:44 crc kubenswrapper[4907]: I0313 15:51:44.344562 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7559d477c5-jggzg"] Mar 13 15:51:44 crc kubenswrapper[4907]: I0313 15:51:44.356554 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7559d477c5-jggzg"] Mar 13 15:51:44 crc kubenswrapper[4907]: I0313 15:51:44.374079 4907 scope.go:117] "RemoveContainer" containerID="f427c6c6d4cea6920aec5e3d13f7c210e91731dd6a8240743203f4fd1462ee2f" Mar 13 15:51:44 crc kubenswrapper[4907]: E0313 15:51:44.374590 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f427c6c6d4cea6920aec5e3d13f7c210e91731dd6a8240743203f4fd1462ee2f\": container with ID starting with f427c6c6d4cea6920aec5e3d13f7c210e91731dd6a8240743203f4fd1462ee2f not found: ID does not exist" containerID="f427c6c6d4cea6920aec5e3d13f7c210e91731dd6a8240743203f4fd1462ee2f" Mar 13 15:51:44 crc kubenswrapper[4907]: I0313 15:51:44.374626 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f427c6c6d4cea6920aec5e3d13f7c210e91731dd6a8240743203f4fd1462ee2f"} err="failed to get container status \"f427c6c6d4cea6920aec5e3d13f7c210e91731dd6a8240743203f4fd1462ee2f\": rpc error: code = NotFound desc = could not find container \"f427c6c6d4cea6920aec5e3d13f7c210e91731dd6a8240743203f4fd1462ee2f\": container with ID starting with f427c6c6d4cea6920aec5e3d13f7c210e91731dd6a8240743203f4fd1462ee2f not found: ID does not exist" Mar 13 15:51:44 crc kubenswrapper[4907]: I0313 15:51:44.374646 4907 scope.go:117] "RemoveContainer" containerID="6a9689f3bb28f496b72b2c47ce88ac358b381c9e951286ab8db719b6cd0bd636" Mar 13 15:51:44 crc kubenswrapper[4907]: E0313 15:51:44.375002 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6a9689f3bb28f496b72b2c47ce88ac358b381c9e951286ab8db719b6cd0bd636\": container with ID starting with 6a9689f3bb28f496b72b2c47ce88ac358b381c9e951286ab8db719b6cd0bd636 not found: ID does not exist" containerID="6a9689f3bb28f496b72b2c47ce88ac358b381c9e951286ab8db719b6cd0bd636" Mar 13 15:51:44 crc kubenswrapper[4907]: I0313 15:51:44.375026 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6a9689f3bb28f496b72b2c47ce88ac358b381c9e951286ab8db719b6cd0bd636"} err="failed to get container status \"6a9689f3bb28f496b72b2c47ce88ac358b381c9e951286ab8db719b6cd0bd636\": rpc error: code = NotFound desc = could not find container \"6a9689f3bb28f496b72b2c47ce88ac358b381c9e951286ab8db719b6cd0bd636\": container with ID starting with 6a9689f3bb28f496b72b2c47ce88ac358b381c9e951286ab8db719b6cd0bd636 not found: ID does not exist" Mar 13 15:51:44 crc kubenswrapper[4907]: I0313 15:51:44.676522 4907 generic.go:334] "Generic (PLEG): container finished" podID="44fcf282-3688-4a6f-b390-6b321e89b158" containerID="287d6d7b61fca2739c775824ecadcb069c5c0686f2843b53995e1578a705d92d" exitCode=0 Mar 13 15:51:44 crc kubenswrapper[4907]: I0313 15:51:44.676560 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8644fdb9df-gm62m" event={"ID":"44fcf282-3688-4a6f-b390-6b321e89b158","Type":"ContainerDied","Data":"287d6d7b61fca2739c775824ecadcb069c5c0686f2843b53995e1578a705d92d"} Mar 13 15:51:44 crc kubenswrapper[4907]: I0313 15:51:44.676583 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8644fdb9df-gm62m" event={"ID":"44fcf282-3688-4a6f-b390-6b321e89b158","Type":"ContainerStarted","Data":"21b3b48967ffccd6fd70cace9989429623d571f5726ff15b165aad899cca0333"} Mar 13 15:51:45 crc kubenswrapper[4907]: I0313 15:51:45.687999 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8644fdb9df-gm62m" event={"ID":"44fcf282-3688-4a6f-b390-6b321e89b158","Type":"ContainerStarted","Data":"115cd9adc28b0ba5e541b1c130a3291150179c7d1f979fd6c223d918a88300b5"} Mar 13 15:51:45 crc kubenswrapper[4907]: I0313 15:51:45.689094 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-8644fdb9df-gm62m" Mar 13 15:51:45 crc kubenswrapper[4907]: I0313 15:51:45.711051 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-8644fdb9df-gm62m" podStartSLOduration=3.711034649 podStartE2EDuration="3.711034649s" podCreationTimestamp="2026-03-13 15:51:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:51:45.705509028 +0000 UTC m=+6404.605296717" watchObservedRunningTime="2026-03-13 15:51:45.711034649 +0000 UTC m=+6404.610822338" Mar 13 15:51:45 crc kubenswrapper[4907]: I0313 15:51:45.793554 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a01ed54c-beec-437b-a4fb-c4c62a407d76" path="/var/lib/kubelet/pods/a01ed54c-beec-437b-a4fb-c4c62a407d76/volumes" Mar 13 15:51:48 crc kubenswrapper[4907]: I0313 15:51:48.041623 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 15:51:48 crc kubenswrapper[4907]: I0313 15:51:48.042188 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 15:51:53 crc kubenswrapper[4907]: I0313 15:51:53.280045 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-8644fdb9df-gm62m" Mar 13 15:51:53 crc kubenswrapper[4907]: I0313 15:51:53.348680 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5ff7fb8bcc-45q4w"] Mar 13 15:51:53 crc kubenswrapper[4907]: I0313 15:51:53.348959 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5ff7fb8bcc-45q4w" podUID="68d11f2a-37b6-4a6c-b877-ae32e177761d" containerName="dnsmasq-dns" containerID="cri-o://997f7ae74c7025a1a37f82e6b1e89fdb84b9a3fd4054caac0d2b4e2d822d549a" gracePeriod=10 Mar 13 15:51:53 crc kubenswrapper[4907]: I0313 15:51:53.794592 4907 generic.go:334] "Generic (PLEG): container finished" podID="68d11f2a-37b6-4a6c-b877-ae32e177761d" containerID="997f7ae74c7025a1a37f82e6b1e89fdb84b9a3fd4054caac0d2b4e2d822d549a" exitCode=0 Mar 13 15:51:53 crc kubenswrapper[4907]: I0313 15:51:53.797819 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ff7fb8bcc-45q4w" event={"ID":"68d11f2a-37b6-4a6c-b877-ae32e177761d","Type":"ContainerDied","Data":"997f7ae74c7025a1a37f82e6b1e89fdb84b9a3fd4054caac0d2b4e2d822d549a"} Mar 13 15:51:53 crc kubenswrapper[4907]: I0313 15:51:53.902373 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ff7fb8bcc-45q4w" Mar 13 15:51:54 crc kubenswrapper[4907]: I0313 15:51:54.026979 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vww7s\" (UniqueName: \"kubernetes.io/projected/68d11f2a-37b6-4a6c-b877-ae32e177761d-kube-api-access-vww7s\") pod \"68d11f2a-37b6-4a6c-b877-ae32e177761d\" (UID: \"68d11f2a-37b6-4a6c-b877-ae32e177761d\") " Mar 13 15:51:54 crc kubenswrapper[4907]: I0313 15:51:54.027032 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/68d11f2a-37b6-4a6c-b877-ae32e177761d-ovsdbserver-nb\") pod \"68d11f2a-37b6-4a6c-b877-ae32e177761d\" (UID: \"68d11f2a-37b6-4a6c-b877-ae32e177761d\") " Mar 13 15:51:54 crc kubenswrapper[4907]: I0313 15:51:54.027175 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/68d11f2a-37b6-4a6c-b877-ae32e177761d-ovsdbserver-sb\") pod \"68d11f2a-37b6-4a6c-b877-ae32e177761d\" (UID: \"68d11f2a-37b6-4a6c-b877-ae32e177761d\") " Mar 13 15:51:54 crc kubenswrapper[4907]: I0313 15:51:54.027300 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/68d11f2a-37b6-4a6c-b877-ae32e177761d-config\") pod \"68d11f2a-37b6-4a6c-b877-ae32e177761d\" (UID: \"68d11f2a-37b6-4a6c-b877-ae32e177761d\") " Mar 13 15:51:54 crc kubenswrapper[4907]: I0313 15:51:54.027340 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/68d11f2a-37b6-4a6c-b877-ae32e177761d-openstack-cell1\") pod \"68d11f2a-37b6-4a6c-b877-ae32e177761d\" (UID: \"68d11f2a-37b6-4a6c-b877-ae32e177761d\") " Mar 13 15:51:54 crc kubenswrapper[4907]: I0313 15:51:54.027476 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/68d11f2a-37b6-4a6c-b877-ae32e177761d-dns-svc\") pod \"68d11f2a-37b6-4a6c-b877-ae32e177761d\" (UID: \"68d11f2a-37b6-4a6c-b877-ae32e177761d\") " Mar 13 15:51:54 crc kubenswrapper[4907]: I0313 15:51:54.032981 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/68d11f2a-37b6-4a6c-b877-ae32e177761d-kube-api-access-vww7s" (OuterVolumeSpecName: "kube-api-access-vww7s") pod "68d11f2a-37b6-4a6c-b877-ae32e177761d" (UID: "68d11f2a-37b6-4a6c-b877-ae32e177761d"). InnerVolumeSpecName "kube-api-access-vww7s". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:51:54 crc kubenswrapper[4907]: I0313 15:51:54.085467 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/68d11f2a-37b6-4a6c-b877-ae32e177761d-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "68d11f2a-37b6-4a6c-b877-ae32e177761d" (UID: "68d11f2a-37b6-4a6c-b877-ae32e177761d"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:51:54 crc kubenswrapper[4907]: I0313 15:51:54.093105 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/68d11f2a-37b6-4a6c-b877-ae32e177761d-openstack-cell1" (OuterVolumeSpecName: "openstack-cell1") pod "68d11f2a-37b6-4a6c-b877-ae32e177761d" (UID: "68d11f2a-37b6-4a6c-b877-ae32e177761d"). InnerVolumeSpecName "openstack-cell1". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:51:54 crc kubenswrapper[4907]: I0313 15:51:54.094537 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/68d11f2a-37b6-4a6c-b877-ae32e177761d-config" (OuterVolumeSpecName: "config") pod "68d11f2a-37b6-4a6c-b877-ae32e177761d" (UID: "68d11f2a-37b6-4a6c-b877-ae32e177761d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:51:54 crc kubenswrapper[4907]: I0313 15:51:54.095456 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/68d11f2a-37b6-4a6c-b877-ae32e177761d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "68d11f2a-37b6-4a6c-b877-ae32e177761d" (UID: "68d11f2a-37b6-4a6c-b877-ae32e177761d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:51:54 crc kubenswrapper[4907]: I0313 15:51:54.121357 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/68d11f2a-37b6-4a6c-b877-ae32e177761d-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "68d11f2a-37b6-4a6c-b877-ae32e177761d" (UID: "68d11f2a-37b6-4a6c-b877-ae32e177761d"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:51:54 crc kubenswrapper[4907]: I0313 15:51:54.130222 4907 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/68d11f2a-37b6-4a6c-b877-ae32e177761d-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 13 15:51:54 crc kubenswrapper[4907]: I0313 15:51:54.130271 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/68d11f2a-37b6-4a6c-b877-ae32e177761d-config\") on node \"crc\" DevicePath \"\"" Mar 13 15:51:54 crc kubenswrapper[4907]: I0313 15:51:54.130283 4907 reconciler_common.go:293] "Volume detached for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/68d11f2a-37b6-4a6c-b877-ae32e177761d-openstack-cell1\") on node \"crc\" DevicePath \"\"" Mar 13 15:51:54 crc kubenswrapper[4907]: I0313 15:51:54.130294 4907 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/68d11f2a-37b6-4a6c-b877-ae32e177761d-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 13 15:51:54 crc kubenswrapper[4907]: I0313 15:51:54.130309 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vww7s\" (UniqueName: \"kubernetes.io/projected/68d11f2a-37b6-4a6c-b877-ae32e177761d-kube-api-access-vww7s\") on node \"crc\" DevicePath \"\"" Mar 13 15:51:54 crc kubenswrapper[4907]: I0313 15:51:54.130324 4907 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/68d11f2a-37b6-4a6c-b877-ae32e177761d-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 13 15:51:54 crc kubenswrapper[4907]: I0313 15:51:54.805364 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ff7fb8bcc-45q4w" event={"ID":"68d11f2a-37b6-4a6c-b877-ae32e177761d","Type":"ContainerDied","Data":"5b3c0f84402a8ec16bbf16451e1bc1400975871789f821c116960d27711f8335"} Mar 13 15:51:54 crc kubenswrapper[4907]: I0313 15:51:54.805416 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ff7fb8bcc-45q4w" Mar 13 15:51:54 crc kubenswrapper[4907]: I0313 15:51:54.805677 4907 scope.go:117] "RemoveContainer" containerID="997f7ae74c7025a1a37f82e6b1e89fdb84b9a3fd4054caac0d2b4e2d822d549a" Mar 13 15:51:54 crc kubenswrapper[4907]: I0313 15:51:54.840548 4907 scope.go:117] "RemoveContainer" containerID="bb89c1e3b4d030f795279c7977f6bc141b8f573b1af14ac2704526b6da21d04c" Mar 13 15:51:54 crc kubenswrapper[4907]: I0313 15:51:54.850320 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5ff7fb8bcc-45q4w"] Mar 13 15:51:54 crc kubenswrapper[4907]: I0313 15:51:54.872028 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5ff7fb8bcc-45q4w"] Mar 13 15:51:55 crc kubenswrapper[4907]: I0313 15:51:55.793598 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="68d11f2a-37b6-4a6c-b877-ae32e177761d" path="/var/lib/kubelet/pods/68d11f2a-37b6-4a6c-b877-ae32e177761d/volumes" Mar 13 15:52:00 crc kubenswrapper[4907]: I0313 15:52:00.142346 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556952-b69cd"] Mar 13 15:52:00 crc kubenswrapper[4907]: E0313 15:52:00.143035 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68d11f2a-37b6-4a6c-b877-ae32e177761d" containerName="dnsmasq-dns" Mar 13 15:52:00 crc kubenswrapper[4907]: I0313 15:52:00.143048 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="68d11f2a-37b6-4a6c-b877-ae32e177761d" containerName="dnsmasq-dns" Mar 13 15:52:00 crc kubenswrapper[4907]: E0313 15:52:00.143067 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68d11f2a-37b6-4a6c-b877-ae32e177761d" containerName="init" Mar 13 15:52:00 crc kubenswrapper[4907]: I0313 15:52:00.143073 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="68d11f2a-37b6-4a6c-b877-ae32e177761d" containerName="init" Mar 13 15:52:00 crc kubenswrapper[4907]: E0313 15:52:00.143097 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a01ed54c-beec-437b-a4fb-c4c62a407d76" containerName="init" Mar 13 15:52:00 crc kubenswrapper[4907]: I0313 15:52:00.143103 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="a01ed54c-beec-437b-a4fb-c4c62a407d76" containerName="init" Mar 13 15:52:00 crc kubenswrapper[4907]: E0313 15:52:00.143116 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a01ed54c-beec-437b-a4fb-c4c62a407d76" containerName="dnsmasq-dns" Mar 13 15:52:00 crc kubenswrapper[4907]: I0313 15:52:00.143122 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="a01ed54c-beec-437b-a4fb-c4c62a407d76" containerName="dnsmasq-dns" Mar 13 15:52:00 crc kubenswrapper[4907]: I0313 15:52:00.143296 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="a01ed54c-beec-437b-a4fb-c4c62a407d76" containerName="dnsmasq-dns" Mar 13 15:52:00 crc kubenswrapper[4907]: I0313 15:52:00.143306 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="68d11f2a-37b6-4a6c-b877-ae32e177761d" containerName="dnsmasq-dns" Mar 13 15:52:00 crc kubenswrapper[4907]: I0313 15:52:00.143997 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556952-b69cd" Mar 13 15:52:00 crc kubenswrapper[4907]: I0313 15:52:00.154069 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556952-b69cd"] Mar 13 15:52:00 crc kubenswrapper[4907]: I0313 15:52:00.199503 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 15:52:00 crc kubenswrapper[4907]: I0313 15:52:00.199706 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 15:52:00 crc kubenswrapper[4907]: I0313 15:52:00.199859 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 15:52:00 crc kubenswrapper[4907]: I0313 15:52:00.302055 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q6vc7\" (UniqueName: \"kubernetes.io/projected/d2915500-b9a0-45a4-960e-5325f53b445c-kube-api-access-q6vc7\") pod \"auto-csr-approver-29556952-b69cd\" (UID: \"d2915500-b9a0-45a4-960e-5325f53b445c\") " pod="openshift-infra/auto-csr-approver-29556952-b69cd" Mar 13 15:52:00 crc kubenswrapper[4907]: I0313 15:52:00.404801 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q6vc7\" (UniqueName: \"kubernetes.io/projected/d2915500-b9a0-45a4-960e-5325f53b445c-kube-api-access-q6vc7\") pod \"auto-csr-approver-29556952-b69cd\" (UID: \"d2915500-b9a0-45a4-960e-5325f53b445c\") " pod="openshift-infra/auto-csr-approver-29556952-b69cd" Mar 13 15:52:00 crc kubenswrapper[4907]: I0313 15:52:00.427025 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q6vc7\" (UniqueName: \"kubernetes.io/projected/d2915500-b9a0-45a4-960e-5325f53b445c-kube-api-access-q6vc7\") pod \"auto-csr-approver-29556952-b69cd\" (UID: \"d2915500-b9a0-45a4-960e-5325f53b445c\") " pod="openshift-infra/auto-csr-approver-29556952-b69cd" Mar 13 15:52:00 crc kubenswrapper[4907]: I0313 15:52:00.519082 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556952-b69cd" Mar 13 15:52:00 crc kubenswrapper[4907]: I0313 15:52:00.972482 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556952-b69cd"] Mar 13 15:52:01 crc kubenswrapper[4907]: I0313 15:52:01.876365 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556952-b69cd" event={"ID":"d2915500-b9a0-45a4-960e-5325f53b445c","Type":"ContainerStarted","Data":"427c240f7d555a5e5b1cbc79551c7beea9222cc99ef5a82bd5d85f606af5e2b4"} Mar 13 15:52:04 crc kubenswrapper[4907]: I0313 15:52:04.261448 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/pre-adoption-validation-openstack-pre-adoption-openstack-ckdmlw"] Mar 13 15:52:04 crc kubenswrapper[4907]: I0313 15:52:04.263271 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-ckdmlw" Mar 13 15:52:04 crc kubenswrapper[4907]: I0313 15:52:04.265219 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-44kxw" Mar 13 15:52:04 crc kubenswrapper[4907]: I0313 15:52:04.267390 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Mar 13 15:52:04 crc kubenswrapper[4907]: I0313 15:52:04.267659 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Mar 13 15:52:04 crc kubenswrapper[4907]: I0313 15:52:04.267809 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 13 15:52:04 crc kubenswrapper[4907]: I0313 15:52:04.285241 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/pre-adoption-validation-openstack-pre-adoption-openstack-ckdmlw"] Mar 13 15:52:04 crc kubenswrapper[4907]: I0313 15:52:04.388578 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/044c1d1f-5bd4-4359-9262-e08daf71a4ae-inventory\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-ckdmlw\" (UID: \"044c1d1f-5bd4-4359-9262-e08daf71a4ae\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-ckdmlw" Mar 13 15:52:04 crc kubenswrapper[4907]: I0313 15:52:04.389099 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/044c1d1f-5bd4-4359-9262-e08daf71a4ae-ssh-key-openstack-cell1\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-ckdmlw\" (UID: \"044c1d1f-5bd4-4359-9262-e08daf71a4ae\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-ckdmlw" Mar 13 15:52:04 crc kubenswrapper[4907]: I0313 15:52:04.389212 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/044c1d1f-5bd4-4359-9262-e08daf71a4ae-ceph\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-ckdmlw\" (UID: \"044c1d1f-5bd4-4359-9262-e08daf71a4ae\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-ckdmlw" Mar 13 15:52:04 crc kubenswrapper[4907]: I0313 15:52:04.389469 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-khw28\" (UniqueName: \"kubernetes.io/projected/044c1d1f-5bd4-4359-9262-e08daf71a4ae-kube-api-access-khw28\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-ckdmlw\" (UID: \"044c1d1f-5bd4-4359-9262-e08daf71a4ae\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-ckdmlw" Mar 13 15:52:04 crc kubenswrapper[4907]: I0313 15:52:04.389564 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/044c1d1f-5bd4-4359-9262-e08daf71a4ae-pre-adoption-validation-combined-ca-bundle\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-ckdmlw\" (UID: \"044c1d1f-5bd4-4359-9262-e08daf71a4ae\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-ckdmlw" Mar 13 15:52:04 crc kubenswrapper[4907]: I0313 15:52:04.491191 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-khw28\" (UniqueName: \"kubernetes.io/projected/044c1d1f-5bd4-4359-9262-e08daf71a4ae-kube-api-access-khw28\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-ckdmlw\" (UID: \"044c1d1f-5bd4-4359-9262-e08daf71a4ae\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-ckdmlw" Mar 13 15:52:04 crc kubenswrapper[4907]: I0313 15:52:04.491247 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/044c1d1f-5bd4-4359-9262-e08daf71a4ae-pre-adoption-validation-combined-ca-bundle\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-ckdmlw\" (UID: \"044c1d1f-5bd4-4359-9262-e08daf71a4ae\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-ckdmlw" Mar 13 15:52:04 crc kubenswrapper[4907]: I0313 15:52:04.491285 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/044c1d1f-5bd4-4359-9262-e08daf71a4ae-inventory\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-ckdmlw\" (UID: \"044c1d1f-5bd4-4359-9262-e08daf71a4ae\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-ckdmlw" Mar 13 15:52:04 crc kubenswrapper[4907]: I0313 15:52:04.491410 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/044c1d1f-5bd4-4359-9262-e08daf71a4ae-ssh-key-openstack-cell1\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-ckdmlw\" (UID: \"044c1d1f-5bd4-4359-9262-e08daf71a4ae\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-ckdmlw" Mar 13 15:52:04 crc kubenswrapper[4907]: I0313 15:52:04.491444 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/044c1d1f-5bd4-4359-9262-e08daf71a4ae-ceph\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-ckdmlw\" (UID: \"044c1d1f-5bd4-4359-9262-e08daf71a4ae\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-ckdmlw" Mar 13 15:52:04 crc kubenswrapper[4907]: I0313 15:52:04.497607 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/044c1d1f-5bd4-4359-9262-e08daf71a4ae-ceph\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-ckdmlw\" (UID: \"044c1d1f-5bd4-4359-9262-e08daf71a4ae\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-ckdmlw" Mar 13 15:52:04 crc kubenswrapper[4907]: I0313 15:52:04.498510 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/044c1d1f-5bd4-4359-9262-e08daf71a4ae-ssh-key-openstack-cell1\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-ckdmlw\" (UID: \"044c1d1f-5bd4-4359-9262-e08daf71a4ae\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-ckdmlw" Mar 13 15:52:04 crc kubenswrapper[4907]: I0313 15:52:04.499587 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/044c1d1f-5bd4-4359-9262-e08daf71a4ae-pre-adoption-validation-combined-ca-bundle\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-ckdmlw\" (UID: \"044c1d1f-5bd4-4359-9262-e08daf71a4ae\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-ckdmlw" Mar 13 15:52:04 crc kubenswrapper[4907]: I0313 15:52:04.501192 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/044c1d1f-5bd4-4359-9262-e08daf71a4ae-inventory\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-ckdmlw\" (UID: \"044c1d1f-5bd4-4359-9262-e08daf71a4ae\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-ckdmlw" Mar 13 15:52:04 crc kubenswrapper[4907]: I0313 15:52:04.513497 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-khw28\" (UniqueName: \"kubernetes.io/projected/044c1d1f-5bd4-4359-9262-e08daf71a4ae-kube-api-access-khw28\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-ckdmlw\" (UID: \"044c1d1f-5bd4-4359-9262-e08daf71a4ae\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-ckdmlw" Mar 13 15:52:04 crc kubenswrapper[4907]: I0313 15:52:04.591749 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-ckdmlw" Mar 13 15:52:04 crc kubenswrapper[4907]: I0313 15:52:04.903770 4907 generic.go:334] "Generic (PLEG): container finished" podID="d2915500-b9a0-45a4-960e-5325f53b445c" containerID="ae08e1f5613f0f2169631235f70419712208d7aa173e1b4d3da70be8496080b9" exitCode=0 Mar 13 15:52:04 crc kubenswrapper[4907]: I0313 15:52:04.904131 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556952-b69cd" event={"ID":"d2915500-b9a0-45a4-960e-5325f53b445c","Type":"ContainerDied","Data":"ae08e1f5613f0f2169631235f70419712208d7aa173e1b4d3da70be8496080b9"} Mar 13 15:52:05 crc kubenswrapper[4907]: I0313 15:52:05.194526 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/pre-adoption-validation-openstack-pre-adoption-openstack-ckdmlw"] Mar 13 15:52:05 crc kubenswrapper[4907]: W0313 15:52:05.201355 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod044c1d1f_5bd4_4359_9262_e08daf71a4ae.slice/crio-40e9313e8c34a1e8fd1dc1058b1ee1e7934bed65552e6326335a68a5db3bf4f8 WatchSource:0}: Error finding container 40e9313e8c34a1e8fd1dc1058b1ee1e7934bed65552e6326335a68a5db3bf4f8: Status 404 returned error can't find the container with id 40e9313e8c34a1e8fd1dc1058b1ee1e7934bed65552e6326335a68a5db3bf4f8 Mar 13 15:52:05 crc kubenswrapper[4907]: I0313 15:52:05.922794 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-ckdmlw" event={"ID":"044c1d1f-5bd4-4359-9262-e08daf71a4ae","Type":"ContainerStarted","Data":"40e9313e8c34a1e8fd1dc1058b1ee1e7934bed65552e6326335a68a5db3bf4f8"} Mar 13 15:52:06 crc kubenswrapper[4907]: I0313 15:52:06.300114 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556952-b69cd" Mar 13 15:52:06 crc kubenswrapper[4907]: I0313 15:52:06.437894 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q6vc7\" (UniqueName: \"kubernetes.io/projected/d2915500-b9a0-45a4-960e-5325f53b445c-kube-api-access-q6vc7\") pod \"d2915500-b9a0-45a4-960e-5325f53b445c\" (UID: \"d2915500-b9a0-45a4-960e-5325f53b445c\") " Mar 13 15:52:06 crc kubenswrapper[4907]: I0313 15:52:06.446621 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d2915500-b9a0-45a4-960e-5325f53b445c-kube-api-access-q6vc7" (OuterVolumeSpecName: "kube-api-access-q6vc7") pod "d2915500-b9a0-45a4-960e-5325f53b445c" (UID: "d2915500-b9a0-45a4-960e-5325f53b445c"). InnerVolumeSpecName "kube-api-access-q6vc7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:52:06 crc kubenswrapper[4907]: I0313 15:52:06.540954 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q6vc7\" (UniqueName: \"kubernetes.io/projected/d2915500-b9a0-45a4-960e-5325f53b445c-kube-api-access-q6vc7\") on node \"crc\" DevicePath \"\"" Mar 13 15:52:06 crc kubenswrapper[4907]: I0313 15:52:06.956690 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556952-b69cd" event={"ID":"d2915500-b9a0-45a4-960e-5325f53b445c","Type":"ContainerDied","Data":"427c240f7d555a5e5b1cbc79551c7beea9222cc99ef5a82bd5d85f606af5e2b4"} Mar 13 15:52:06 crc kubenswrapper[4907]: I0313 15:52:06.957010 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="427c240f7d555a5e5b1cbc79551c7beea9222cc99ef5a82bd5d85f606af5e2b4" Mar 13 15:52:06 crc kubenswrapper[4907]: I0313 15:52:06.956917 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556952-b69cd" Mar 13 15:52:07 crc kubenswrapper[4907]: I0313 15:52:07.375856 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556946-78rqg"] Mar 13 15:52:07 crc kubenswrapper[4907]: I0313 15:52:07.386359 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556946-78rqg"] Mar 13 15:52:07 crc kubenswrapper[4907]: I0313 15:52:07.798076 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2f0d1531-1d25-4331-a690-1fcbcb504071" path="/var/lib/kubelet/pods/2f0d1531-1d25-4331-a690-1fcbcb504071/volumes" Mar 13 15:52:15 crc kubenswrapper[4907]: I0313 15:52:15.035736 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-ckdmlw" event={"ID":"044c1d1f-5bd4-4359-9262-e08daf71a4ae","Type":"ContainerStarted","Data":"a98ef489a033787cdd255f40580b6674e782d47c8d6a6991c57a495e6e6e8aad"} Mar 13 15:52:15 crc kubenswrapper[4907]: I0313 15:52:15.067069 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-ckdmlw" podStartSLOduration=2.208048563 podStartE2EDuration="11.067034696s" podCreationTimestamp="2026-03-13 15:52:04 +0000 UTC" firstStartedPulling="2026-03-13 15:52:05.209144362 +0000 UTC m=+6424.108932051" lastFinishedPulling="2026-03-13 15:52:14.068130495 +0000 UTC m=+6432.967918184" observedRunningTime="2026-03-13 15:52:15.062509323 +0000 UTC m=+6433.962297032" watchObservedRunningTime="2026-03-13 15:52:15.067034696 +0000 UTC m=+6433.966822375" Mar 13 15:52:18 crc kubenswrapper[4907]: I0313 15:52:18.041380 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 15:52:18 crc kubenswrapper[4907]: I0313 15:52:18.041930 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 15:52:28 crc kubenswrapper[4907]: I0313 15:52:28.154028 4907 generic.go:334] "Generic (PLEG): container finished" podID="044c1d1f-5bd4-4359-9262-e08daf71a4ae" containerID="a98ef489a033787cdd255f40580b6674e782d47c8d6a6991c57a495e6e6e8aad" exitCode=0 Mar 13 15:52:28 crc kubenswrapper[4907]: I0313 15:52:28.154115 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-ckdmlw" event={"ID":"044c1d1f-5bd4-4359-9262-e08daf71a4ae","Type":"ContainerDied","Data":"a98ef489a033787cdd255f40580b6674e782d47c8d6a6991c57a495e6e6e8aad"} Mar 13 15:52:29 crc kubenswrapper[4907]: I0313 15:52:29.658013 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-ckdmlw" Mar 13 15:52:29 crc kubenswrapper[4907]: I0313 15:52:29.781107 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/044c1d1f-5bd4-4359-9262-e08daf71a4ae-ssh-key-openstack-cell1\") pod \"044c1d1f-5bd4-4359-9262-e08daf71a4ae\" (UID: \"044c1d1f-5bd4-4359-9262-e08daf71a4ae\") " Mar 13 15:52:29 crc kubenswrapper[4907]: I0313 15:52:29.781196 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-khw28\" (UniqueName: \"kubernetes.io/projected/044c1d1f-5bd4-4359-9262-e08daf71a4ae-kube-api-access-khw28\") pod \"044c1d1f-5bd4-4359-9262-e08daf71a4ae\" (UID: \"044c1d1f-5bd4-4359-9262-e08daf71a4ae\") " Mar 13 15:52:29 crc kubenswrapper[4907]: I0313 15:52:29.781691 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/044c1d1f-5bd4-4359-9262-e08daf71a4ae-inventory\") pod \"044c1d1f-5bd4-4359-9262-e08daf71a4ae\" (UID: \"044c1d1f-5bd4-4359-9262-e08daf71a4ae\") " Mar 13 15:52:29 crc kubenswrapper[4907]: I0313 15:52:29.781843 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/044c1d1f-5bd4-4359-9262-e08daf71a4ae-pre-adoption-validation-combined-ca-bundle\") pod \"044c1d1f-5bd4-4359-9262-e08daf71a4ae\" (UID: \"044c1d1f-5bd4-4359-9262-e08daf71a4ae\") " Mar 13 15:52:29 crc kubenswrapper[4907]: I0313 15:52:29.781917 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/044c1d1f-5bd4-4359-9262-e08daf71a4ae-ceph\") pod \"044c1d1f-5bd4-4359-9262-e08daf71a4ae\" (UID: \"044c1d1f-5bd4-4359-9262-e08daf71a4ae\") " Mar 13 15:52:29 crc kubenswrapper[4907]: I0313 15:52:29.787442 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/044c1d1f-5bd4-4359-9262-e08daf71a4ae-kube-api-access-khw28" (OuterVolumeSpecName: "kube-api-access-khw28") pod "044c1d1f-5bd4-4359-9262-e08daf71a4ae" (UID: "044c1d1f-5bd4-4359-9262-e08daf71a4ae"). InnerVolumeSpecName "kube-api-access-khw28". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:52:29 crc kubenswrapper[4907]: I0313 15:52:29.787737 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/044c1d1f-5bd4-4359-9262-e08daf71a4ae-pre-adoption-validation-combined-ca-bundle" (OuterVolumeSpecName: "pre-adoption-validation-combined-ca-bundle") pod "044c1d1f-5bd4-4359-9262-e08daf71a4ae" (UID: "044c1d1f-5bd4-4359-9262-e08daf71a4ae"). InnerVolumeSpecName "pre-adoption-validation-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:52:29 crc kubenswrapper[4907]: I0313 15:52:29.789207 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/044c1d1f-5bd4-4359-9262-e08daf71a4ae-ceph" (OuterVolumeSpecName: "ceph") pod "044c1d1f-5bd4-4359-9262-e08daf71a4ae" (UID: "044c1d1f-5bd4-4359-9262-e08daf71a4ae"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:52:29 crc kubenswrapper[4907]: I0313 15:52:29.816733 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/044c1d1f-5bd4-4359-9262-e08daf71a4ae-ssh-key-openstack-cell1" (OuterVolumeSpecName: "ssh-key-openstack-cell1") pod "044c1d1f-5bd4-4359-9262-e08daf71a4ae" (UID: "044c1d1f-5bd4-4359-9262-e08daf71a4ae"). InnerVolumeSpecName "ssh-key-openstack-cell1". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:52:29 crc kubenswrapper[4907]: I0313 15:52:29.841080 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/044c1d1f-5bd4-4359-9262-e08daf71a4ae-inventory" (OuterVolumeSpecName: "inventory") pod "044c1d1f-5bd4-4359-9262-e08daf71a4ae" (UID: "044c1d1f-5bd4-4359-9262-e08daf71a4ae"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:52:29 crc kubenswrapper[4907]: I0313 15:52:29.886315 4907 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/044c1d1f-5bd4-4359-9262-e08daf71a4ae-ssh-key-openstack-cell1\") on node \"crc\" DevicePath \"\"" Mar 13 15:52:29 crc kubenswrapper[4907]: I0313 15:52:29.886370 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-khw28\" (UniqueName: \"kubernetes.io/projected/044c1d1f-5bd4-4359-9262-e08daf71a4ae-kube-api-access-khw28\") on node \"crc\" DevicePath \"\"" Mar 13 15:52:29 crc kubenswrapper[4907]: I0313 15:52:29.886386 4907 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/044c1d1f-5bd4-4359-9262-e08daf71a4ae-inventory\") on node \"crc\" DevicePath \"\"" Mar 13 15:52:29 crc kubenswrapper[4907]: I0313 15:52:29.886399 4907 reconciler_common.go:293] "Volume detached for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/044c1d1f-5bd4-4359-9262-e08daf71a4ae-pre-adoption-validation-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 15:52:29 crc kubenswrapper[4907]: I0313 15:52:29.886415 4907 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/044c1d1f-5bd4-4359-9262-e08daf71a4ae-ceph\") on node \"crc\" DevicePath \"\"" Mar 13 15:52:30 crc kubenswrapper[4907]: I0313 15:52:30.177665 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-ckdmlw" event={"ID":"044c1d1f-5bd4-4359-9262-e08daf71a4ae","Type":"ContainerDied","Data":"40e9313e8c34a1e8fd1dc1058b1ee1e7934bed65552e6326335a68a5db3bf4f8"} Mar 13 15:52:30 crc kubenswrapper[4907]: I0313 15:52:30.178108 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="40e9313e8c34a1e8fd1dc1058b1ee1e7934bed65552e6326335a68a5db3bf4f8" Mar 13 15:52:30 crc kubenswrapper[4907]: I0313 15:52:30.178057 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-ckdmlw" Mar 13 15:52:36 crc kubenswrapper[4907]: I0313 15:52:36.992524 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-fj6qz"] Mar 13 15:52:36 crc kubenswrapper[4907]: E0313 15:52:36.994387 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="044c1d1f-5bd4-4359-9262-e08daf71a4ae" containerName="pre-adoption-validation-openstack-pre-adoption-openstack-cell1" Mar 13 15:52:36 crc kubenswrapper[4907]: I0313 15:52:36.994492 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="044c1d1f-5bd4-4359-9262-e08daf71a4ae" containerName="pre-adoption-validation-openstack-pre-adoption-openstack-cell1" Mar 13 15:52:36 crc kubenswrapper[4907]: E0313 15:52:36.994563 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2915500-b9a0-45a4-960e-5325f53b445c" containerName="oc" Mar 13 15:52:36 crc kubenswrapper[4907]: I0313 15:52:36.994619 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2915500-b9a0-45a4-960e-5325f53b445c" containerName="oc" Mar 13 15:52:36 crc kubenswrapper[4907]: I0313 15:52:36.995061 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="d2915500-b9a0-45a4-960e-5325f53b445c" containerName="oc" Mar 13 15:52:36 crc kubenswrapper[4907]: I0313 15:52:36.995146 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="044c1d1f-5bd4-4359-9262-e08daf71a4ae" containerName="pre-adoption-validation-openstack-pre-adoption-openstack-cell1" Mar 13 15:52:36 crc kubenswrapper[4907]: I0313 15:52:36.996072 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-fj6qz" Mar 13 15:52:36 crc kubenswrapper[4907]: I0313 15:52:36.999038 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Mar 13 15:52:37 crc kubenswrapper[4907]: I0313 15:52:37.004480 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-44kxw" Mar 13 15:52:37 crc kubenswrapper[4907]: I0313 15:52:37.004707 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Mar 13 15:52:37 crc kubenswrapper[4907]: I0313 15:52:37.004722 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 13 15:52:37 crc kubenswrapper[4907]: I0313 15:52:37.005831 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-fj6qz"] Mar 13 15:52:37 crc kubenswrapper[4907]: I0313 15:52:37.147770 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02ef8f0f-4ae4-478a-a729-3bde4adade43-tripleo-cleanup-combined-ca-bundle\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-fj6qz\" (UID: \"02ef8f0f-4ae4-478a-a729-3bde4adade43\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-fj6qz" Mar 13 15:52:37 crc kubenswrapper[4907]: I0313 15:52:37.148758 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/02ef8f0f-4ae4-478a-a729-3bde4adade43-ssh-key-openstack-cell1\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-fj6qz\" (UID: \"02ef8f0f-4ae4-478a-a729-3bde4adade43\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-fj6qz" Mar 13 15:52:37 crc kubenswrapper[4907]: I0313 15:52:37.149060 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/02ef8f0f-4ae4-478a-a729-3bde4adade43-inventory\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-fj6qz\" (UID: \"02ef8f0f-4ae4-478a-a729-3bde4adade43\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-fj6qz" Mar 13 15:52:37 crc kubenswrapper[4907]: I0313 15:52:37.149144 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gsfkk\" (UniqueName: \"kubernetes.io/projected/02ef8f0f-4ae4-478a-a729-3bde4adade43-kube-api-access-gsfkk\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-fj6qz\" (UID: \"02ef8f0f-4ae4-478a-a729-3bde4adade43\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-fj6qz" Mar 13 15:52:37 crc kubenswrapper[4907]: I0313 15:52:37.149217 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/02ef8f0f-4ae4-478a-a729-3bde4adade43-ceph\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-fj6qz\" (UID: \"02ef8f0f-4ae4-478a-a729-3bde4adade43\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-fj6qz" Mar 13 15:52:37 crc kubenswrapper[4907]: I0313 15:52:37.251215 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/02ef8f0f-4ae4-478a-a729-3bde4adade43-ceph\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-fj6qz\" (UID: \"02ef8f0f-4ae4-478a-a729-3bde4adade43\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-fj6qz" Mar 13 15:52:37 crc kubenswrapper[4907]: I0313 15:52:37.251336 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02ef8f0f-4ae4-478a-a729-3bde4adade43-tripleo-cleanup-combined-ca-bundle\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-fj6qz\" (UID: \"02ef8f0f-4ae4-478a-a729-3bde4adade43\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-fj6qz" Mar 13 15:52:37 crc kubenswrapper[4907]: I0313 15:52:37.251358 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/02ef8f0f-4ae4-478a-a729-3bde4adade43-ssh-key-openstack-cell1\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-fj6qz\" (UID: \"02ef8f0f-4ae4-478a-a729-3bde4adade43\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-fj6qz" Mar 13 15:52:37 crc kubenswrapper[4907]: I0313 15:52:37.251435 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/02ef8f0f-4ae4-478a-a729-3bde4adade43-inventory\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-fj6qz\" (UID: \"02ef8f0f-4ae4-478a-a729-3bde4adade43\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-fj6qz" Mar 13 15:52:37 crc kubenswrapper[4907]: I0313 15:52:37.251480 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gsfkk\" (UniqueName: \"kubernetes.io/projected/02ef8f0f-4ae4-478a-a729-3bde4adade43-kube-api-access-gsfkk\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-fj6qz\" (UID: \"02ef8f0f-4ae4-478a-a729-3bde4adade43\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-fj6qz" Mar 13 15:52:37 crc kubenswrapper[4907]: I0313 15:52:37.257687 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/02ef8f0f-4ae4-478a-a729-3bde4adade43-inventory\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-fj6qz\" (UID: \"02ef8f0f-4ae4-478a-a729-3bde4adade43\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-fj6qz" Mar 13 15:52:37 crc kubenswrapper[4907]: I0313 15:52:37.257721 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/02ef8f0f-4ae4-478a-a729-3bde4adade43-ceph\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-fj6qz\" (UID: \"02ef8f0f-4ae4-478a-a729-3bde4adade43\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-fj6qz" Mar 13 15:52:37 crc kubenswrapper[4907]: I0313 15:52:37.257854 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02ef8f0f-4ae4-478a-a729-3bde4adade43-tripleo-cleanup-combined-ca-bundle\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-fj6qz\" (UID: \"02ef8f0f-4ae4-478a-a729-3bde4adade43\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-fj6qz" Mar 13 15:52:37 crc kubenswrapper[4907]: I0313 15:52:37.262692 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/02ef8f0f-4ae4-478a-a729-3bde4adade43-ssh-key-openstack-cell1\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-fj6qz\" (UID: \"02ef8f0f-4ae4-478a-a729-3bde4adade43\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-fj6qz" Mar 13 15:52:37 crc kubenswrapper[4907]: I0313 15:52:37.269475 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gsfkk\" (UniqueName: \"kubernetes.io/projected/02ef8f0f-4ae4-478a-a729-3bde4adade43-kube-api-access-gsfkk\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-fj6qz\" (UID: \"02ef8f0f-4ae4-478a-a729-3bde4adade43\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-fj6qz" Mar 13 15:52:37 crc kubenswrapper[4907]: I0313 15:52:37.323331 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-fj6qz" Mar 13 15:52:37 crc kubenswrapper[4907]: I0313 15:52:37.840357 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-fj6qz"] Mar 13 15:52:38 crc kubenswrapper[4907]: I0313 15:52:38.250679 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-fj6qz" event={"ID":"02ef8f0f-4ae4-478a-a729-3bde4adade43","Type":"ContainerStarted","Data":"4e306d67a326c5a22924c14c80f0301c152e3eb9bef16aa1cc486b8e33a33cc7"} Mar 13 15:52:39 crc kubenswrapper[4907]: I0313 15:52:39.188945 4907 scope.go:117] "RemoveContainer" containerID="971365269f08ef2266842a1c0ca18c6bb47126b67a1360b35d5a7dfd3acab380" Mar 13 15:52:39 crc kubenswrapper[4907]: I0313 15:52:39.264042 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-fj6qz" event={"ID":"02ef8f0f-4ae4-478a-a729-3bde4adade43","Type":"ContainerStarted","Data":"384d76deb615542ee9afab32cbd864c5a2ef0e524a3354232d127d85ac034766"} Mar 13 15:52:39 crc kubenswrapper[4907]: I0313 15:52:39.310276 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-fj6qz" podStartSLOduration=2.524722311 podStartE2EDuration="3.310247263s" podCreationTimestamp="2026-03-13 15:52:36 +0000 UTC" firstStartedPulling="2026-03-13 15:52:37.853966894 +0000 UTC m=+6456.753754583" lastFinishedPulling="2026-03-13 15:52:38.639491846 +0000 UTC m=+6457.539279535" observedRunningTime="2026-03-13 15:52:39.284281743 +0000 UTC m=+6458.184069472" watchObservedRunningTime="2026-03-13 15:52:39.310247263 +0000 UTC m=+6458.210034982" Mar 13 15:52:48 crc kubenswrapper[4907]: I0313 15:52:48.041330 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 15:52:48 crc kubenswrapper[4907]: I0313 15:52:48.042088 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 15:52:48 crc kubenswrapper[4907]: I0313 15:52:48.042151 4907 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" Mar 13 15:52:48 crc kubenswrapper[4907]: I0313 15:52:48.043329 4907 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"33f8e315523143b157797243610ad7b82c7815f9e551d1169f2c036cfcadf792"} pod="openshift-machine-config-operator/machine-config-daemon-hm56j" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 15:52:48 crc kubenswrapper[4907]: I0313 15:52:48.043418 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" containerID="cri-o://33f8e315523143b157797243610ad7b82c7815f9e551d1169f2c036cfcadf792" gracePeriod=600 Mar 13 15:52:48 crc kubenswrapper[4907]: E0313 15:52:48.184056 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:52:48 crc kubenswrapper[4907]: I0313 15:52:48.350440 4907 generic.go:334] "Generic (PLEG): container finished" podID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerID="33f8e315523143b157797243610ad7b82c7815f9e551d1169f2c036cfcadf792" exitCode=0 Mar 13 15:52:48 crc kubenswrapper[4907]: I0313 15:52:48.350486 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerDied","Data":"33f8e315523143b157797243610ad7b82c7815f9e551d1169f2c036cfcadf792"} Mar 13 15:52:48 crc kubenswrapper[4907]: I0313 15:52:48.350521 4907 scope.go:117] "RemoveContainer" containerID="81d94e236cbdff3b86e8375b8abbaee97f49d844932ee0151cdfa9d97cb1c3b2" Mar 13 15:52:48 crc kubenswrapper[4907]: I0313 15:52:48.351071 4907 scope.go:117] "RemoveContainer" containerID="33f8e315523143b157797243610ad7b82c7815f9e551d1169f2c036cfcadf792" Mar 13 15:52:48 crc kubenswrapper[4907]: E0313 15:52:48.351388 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:52:54 crc kubenswrapper[4907]: I0313 15:52:54.626048 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-6fnjv"] Mar 13 15:52:54 crc kubenswrapper[4907]: I0313 15:52:54.630248 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6fnjv" Mar 13 15:52:54 crc kubenswrapper[4907]: I0313 15:52:54.631832 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-6fnjv"] Mar 13 15:52:54 crc kubenswrapper[4907]: I0313 15:52:54.714550 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/63bc7557-0465-42dc-873c-a8eb54e03a45-catalog-content\") pod \"redhat-marketplace-6fnjv\" (UID: \"63bc7557-0465-42dc-873c-a8eb54e03a45\") " pod="openshift-marketplace/redhat-marketplace-6fnjv" Mar 13 15:52:54 crc kubenswrapper[4907]: I0313 15:52:54.714839 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/63bc7557-0465-42dc-873c-a8eb54e03a45-utilities\") pod \"redhat-marketplace-6fnjv\" (UID: \"63bc7557-0465-42dc-873c-a8eb54e03a45\") " pod="openshift-marketplace/redhat-marketplace-6fnjv" Mar 13 15:52:54 crc kubenswrapper[4907]: I0313 15:52:54.715113 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d9qjl\" (UniqueName: \"kubernetes.io/projected/63bc7557-0465-42dc-873c-a8eb54e03a45-kube-api-access-d9qjl\") pod \"redhat-marketplace-6fnjv\" (UID: \"63bc7557-0465-42dc-873c-a8eb54e03a45\") " pod="openshift-marketplace/redhat-marketplace-6fnjv" Mar 13 15:52:54 crc kubenswrapper[4907]: I0313 15:52:54.816779 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/63bc7557-0465-42dc-873c-a8eb54e03a45-catalog-content\") pod \"redhat-marketplace-6fnjv\" (UID: \"63bc7557-0465-42dc-873c-a8eb54e03a45\") " pod="openshift-marketplace/redhat-marketplace-6fnjv" Mar 13 15:52:54 crc kubenswrapper[4907]: I0313 15:52:54.816845 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/63bc7557-0465-42dc-873c-a8eb54e03a45-utilities\") pod \"redhat-marketplace-6fnjv\" (UID: \"63bc7557-0465-42dc-873c-a8eb54e03a45\") " pod="openshift-marketplace/redhat-marketplace-6fnjv" Mar 13 15:52:54 crc kubenswrapper[4907]: I0313 15:52:54.816919 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d9qjl\" (UniqueName: \"kubernetes.io/projected/63bc7557-0465-42dc-873c-a8eb54e03a45-kube-api-access-d9qjl\") pod \"redhat-marketplace-6fnjv\" (UID: \"63bc7557-0465-42dc-873c-a8eb54e03a45\") " pod="openshift-marketplace/redhat-marketplace-6fnjv" Mar 13 15:52:54 crc kubenswrapper[4907]: I0313 15:52:54.817814 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/63bc7557-0465-42dc-873c-a8eb54e03a45-catalog-content\") pod \"redhat-marketplace-6fnjv\" (UID: \"63bc7557-0465-42dc-873c-a8eb54e03a45\") " pod="openshift-marketplace/redhat-marketplace-6fnjv" Mar 13 15:52:54 crc kubenswrapper[4907]: I0313 15:52:54.817872 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/63bc7557-0465-42dc-873c-a8eb54e03a45-utilities\") pod \"redhat-marketplace-6fnjv\" (UID: \"63bc7557-0465-42dc-873c-a8eb54e03a45\") " pod="openshift-marketplace/redhat-marketplace-6fnjv" Mar 13 15:52:54 crc kubenswrapper[4907]: I0313 15:52:54.837143 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d9qjl\" (UniqueName: \"kubernetes.io/projected/63bc7557-0465-42dc-873c-a8eb54e03a45-kube-api-access-d9qjl\") pod \"redhat-marketplace-6fnjv\" (UID: \"63bc7557-0465-42dc-873c-a8eb54e03a45\") " pod="openshift-marketplace/redhat-marketplace-6fnjv" Mar 13 15:52:54 crc kubenswrapper[4907]: I0313 15:52:54.958255 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6fnjv" Mar 13 15:52:55 crc kubenswrapper[4907]: I0313 15:52:55.430644 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-6fnjv"] Mar 13 15:52:55 crc kubenswrapper[4907]: W0313 15:52:55.440497 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod63bc7557_0465_42dc_873c_a8eb54e03a45.slice/crio-ebcd18dfd1b4c9a8759ca1e5be43efea680a80538844ba60fe37c5d9da7366ed WatchSource:0}: Error finding container ebcd18dfd1b4c9a8759ca1e5be43efea680a80538844ba60fe37c5d9da7366ed: Status 404 returned error can't find the container with id ebcd18dfd1b4c9a8759ca1e5be43efea680a80538844ba60fe37c5d9da7366ed Mar 13 15:52:56 crc kubenswrapper[4907]: I0313 15:52:56.437869 4907 generic.go:334] "Generic (PLEG): container finished" podID="63bc7557-0465-42dc-873c-a8eb54e03a45" containerID="80503b331ede81d6354f232f75d1fa04ac94047782d538502272582284122726" exitCode=0 Mar 13 15:52:56 crc kubenswrapper[4907]: I0313 15:52:56.438099 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6fnjv" event={"ID":"63bc7557-0465-42dc-873c-a8eb54e03a45","Type":"ContainerDied","Data":"80503b331ede81d6354f232f75d1fa04ac94047782d538502272582284122726"} Mar 13 15:52:56 crc kubenswrapper[4907]: I0313 15:52:56.438181 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6fnjv" event={"ID":"63bc7557-0465-42dc-873c-a8eb54e03a45","Type":"ContainerStarted","Data":"ebcd18dfd1b4c9a8759ca1e5be43efea680a80538844ba60fe37c5d9da7366ed"} Mar 13 15:52:57 crc kubenswrapper[4907]: I0313 15:52:57.452000 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6fnjv" event={"ID":"63bc7557-0465-42dc-873c-a8eb54e03a45","Type":"ContainerStarted","Data":"321294b102a32d438b57983725dd613a9c65287a746583d505990a3f7122ab85"} Mar 13 15:52:58 crc kubenswrapper[4907]: I0313 15:52:58.465419 4907 generic.go:334] "Generic (PLEG): container finished" podID="63bc7557-0465-42dc-873c-a8eb54e03a45" containerID="321294b102a32d438b57983725dd613a9c65287a746583d505990a3f7122ab85" exitCode=0 Mar 13 15:52:58 crc kubenswrapper[4907]: I0313 15:52:58.465672 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6fnjv" event={"ID":"63bc7557-0465-42dc-873c-a8eb54e03a45","Type":"ContainerDied","Data":"321294b102a32d438b57983725dd613a9c65287a746583d505990a3f7122ab85"} Mar 13 15:52:59 crc kubenswrapper[4907]: I0313 15:52:59.477421 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6fnjv" event={"ID":"63bc7557-0465-42dc-873c-a8eb54e03a45","Type":"ContainerStarted","Data":"7a058c3c97b1e6d74dc7c5eefe16beabcf70d2ab01c388a4525e76f026793229"} Mar 13 15:52:59 crc kubenswrapper[4907]: I0313 15:52:59.506003 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-6fnjv" podStartSLOduration=3.013929319 podStartE2EDuration="5.505980356s" podCreationTimestamp="2026-03-13 15:52:54 +0000 UTC" firstStartedPulling="2026-03-13 15:52:56.443332329 +0000 UTC m=+6475.343120018" lastFinishedPulling="2026-03-13 15:52:58.935383366 +0000 UTC m=+6477.835171055" observedRunningTime="2026-03-13 15:52:59.495791508 +0000 UTC m=+6478.395579217" watchObservedRunningTime="2026-03-13 15:52:59.505980356 +0000 UTC m=+6478.405768055" Mar 13 15:52:59 crc kubenswrapper[4907]: I0313 15:52:59.782469 4907 scope.go:117] "RemoveContainer" containerID="33f8e315523143b157797243610ad7b82c7815f9e551d1169f2c036cfcadf792" Mar 13 15:52:59 crc kubenswrapper[4907]: E0313 15:52:59.782864 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:53:04 crc kubenswrapper[4907]: I0313 15:53:04.959102 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-6fnjv" Mar 13 15:53:04 crc kubenswrapper[4907]: I0313 15:53:04.959593 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-6fnjv" Mar 13 15:53:05 crc kubenswrapper[4907]: I0313 15:53:05.012760 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-6fnjv" Mar 13 15:53:05 crc kubenswrapper[4907]: I0313 15:53:05.608354 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-6fnjv" Mar 13 15:53:05 crc kubenswrapper[4907]: I0313 15:53:05.663237 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-6fnjv"] Mar 13 15:53:07 crc kubenswrapper[4907]: I0313 15:53:07.575174 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-6fnjv" podUID="63bc7557-0465-42dc-873c-a8eb54e03a45" containerName="registry-server" containerID="cri-o://7a058c3c97b1e6d74dc7c5eefe16beabcf70d2ab01c388a4525e76f026793229" gracePeriod=2 Mar 13 15:53:08 crc kubenswrapper[4907]: I0313 15:53:08.046145 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6fnjv" Mar 13 15:53:08 crc kubenswrapper[4907]: I0313 15:53:08.107155 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/63bc7557-0465-42dc-873c-a8eb54e03a45-catalog-content\") pod \"63bc7557-0465-42dc-873c-a8eb54e03a45\" (UID: \"63bc7557-0465-42dc-873c-a8eb54e03a45\") " Mar 13 15:53:08 crc kubenswrapper[4907]: I0313 15:53:08.107520 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/63bc7557-0465-42dc-873c-a8eb54e03a45-utilities\") pod \"63bc7557-0465-42dc-873c-a8eb54e03a45\" (UID: \"63bc7557-0465-42dc-873c-a8eb54e03a45\") " Mar 13 15:53:08 crc kubenswrapper[4907]: I0313 15:53:08.107556 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d9qjl\" (UniqueName: \"kubernetes.io/projected/63bc7557-0465-42dc-873c-a8eb54e03a45-kube-api-access-d9qjl\") pod \"63bc7557-0465-42dc-873c-a8eb54e03a45\" (UID: \"63bc7557-0465-42dc-873c-a8eb54e03a45\") " Mar 13 15:53:08 crc kubenswrapper[4907]: I0313 15:53:08.108311 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/63bc7557-0465-42dc-873c-a8eb54e03a45-utilities" (OuterVolumeSpecName: "utilities") pod "63bc7557-0465-42dc-873c-a8eb54e03a45" (UID: "63bc7557-0465-42dc-873c-a8eb54e03a45"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:53:08 crc kubenswrapper[4907]: I0313 15:53:08.115212 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/63bc7557-0465-42dc-873c-a8eb54e03a45-kube-api-access-d9qjl" (OuterVolumeSpecName: "kube-api-access-d9qjl") pod "63bc7557-0465-42dc-873c-a8eb54e03a45" (UID: "63bc7557-0465-42dc-873c-a8eb54e03a45"). InnerVolumeSpecName "kube-api-access-d9qjl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:53:08 crc kubenswrapper[4907]: I0313 15:53:08.139029 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/63bc7557-0465-42dc-873c-a8eb54e03a45-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "63bc7557-0465-42dc-873c-a8eb54e03a45" (UID: "63bc7557-0465-42dc-873c-a8eb54e03a45"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:53:08 crc kubenswrapper[4907]: I0313 15:53:08.210045 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/63bc7557-0465-42dc-873c-a8eb54e03a45-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 15:53:08 crc kubenswrapper[4907]: I0313 15:53:08.210080 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/63bc7557-0465-42dc-873c-a8eb54e03a45-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 15:53:08 crc kubenswrapper[4907]: I0313 15:53:08.210093 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d9qjl\" (UniqueName: \"kubernetes.io/projected/63bc7557-0465-42dc-873c-a8eb54e03a45-kube-api-access-d9qjl\") on node \"crc\" DevicePath \"\"" Mar 13 15:53:08 crc kubenswrapper[4907]: I0313 15:53:08.586961 4907 generic.go:334] "Generic (PLEG): container finished" podID="63bc7557-0465-42dc-873c-a8eb54e03a45" containerID="7a058c3c97b1e6d74dc7c5eefe16beabcf70d2ab01c388a4525e76f026793229" exitCode=0 Mar 13 15:53:08 crc kubenswrapper[4907]: I0313 15:53:08.587046 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6fnjv" Mar 13 15:53:08 crc kubenswrapper[4907]: I0313 15:53:08.587043 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6fnjv" event={"ID":"63bc7557-0465-42dc-873c-a8eb54e03a45","Type":"ContainerDied","Data":"7a058c3c97b1e6d74dc7c5eefe16beabcf70d2ab01c388a4525e76f026793229"} Mar 13 15:53:08 crc kubenswrapper[4907]: I0313 15:53:08.587171 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6fnjv" event={"ID":"63bc7557-0465-42dc-873c-a8eb54e03a45","Type":"ContainerDied","Data":"ebcd18dfd1b4c9a8759ca1e5be43efea680a80538844ba60fe37c5d9da7366ed"} Mar 13 15:53:08 crc kubenswrapper[4907]: I0313 15:53:08.587196 4907 scope.go:117] "RemoveContainer" containerID="7a058c3c97b1e6d74dc7c5eefe16beabcf70d2ab01c388a4525e76f026793229" Mar 13 15:53:08 crc kubenswrapper[4907]: I0313 15:53:08.614443 4907 scope.go:117] "RemoveContainer" containerID="321294b102a32d438b57983725dd613a9c65287a746583d505990a3f7122ab85" Mar 13 15:53:08 crc kubenswrapper[4907]: I0313 15:53:08.624384 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-6fnjv"] Mar 13 15:53:08 crc kubenswrapper[4907]: I0313 15:53:08.635602 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-6fnjv"] Mar 13 15:53:08 crc kubenswrapper[4907]: I0313 15:53:08.642839 4907 scope.go:117] "RemoveContainer" containerID="80503b331ede81d6354f232f75d1fa04ac94047782d538502272582284122726" Mar 13 15:53:08 crc kubenswrapper[4907]: I0313 15:53:08.684945 4907 scope.go:117] "RemoveContainer" containerID="7a058c3c97b1e6d74dc7c5eefe16beabcf70d2ab01c388a4525e76f026793229" Mar 13 15:53:08 crc kubenswrapper[4907]: E0313 15:53:08.685465 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7a058c3c97b1e6d74dc7c5eefe16beabcf70d2ab01c388a4525e76f026793229\": container with ID starting with 7a058c3c97b1e6d74dc7c5eefe16beabcf70d2ab01c388a4525e76f026793229 not found: ID does not exist" containerID="7a058c3c97b1e6d74dc7c5eefe16beabcf70d2ab01c388a4525e76f026793229" Mar 13 15:53:08 crc kubenswrapper[4907]: I0313 15:53:08.685506 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7a058c3c97b1e6d74dc7c5eefe16beabcf70d2ab01c388a4525e76f026793229"} err="failed to get container status \"7a058c3c97b1e6d74dc7c5eefe16beabcf70d2ab01c388a4525e76f026793229\": rpc error: code = NotFound desc = could not find container \"7a058c3c97b1e6d74dc7c5eefe16beabcf70d2ab01c388a4525e76f026793229\": container with ID starting with 7a058c3c97b1e6d74dc7c5eefe16beabcf70d2ab01c388a4525e76f026793229 not found: ID does not exist" Mar 13 15:53:08 crc kubenswrapper[4907]: I0313 15:53:08.685533 4907 scope.go:117] "RemoveContainer" containerID="321294b102a32d438b57983725dd613a9c65287a746583d505990a3f7122ab85" Mar 13 15:53:08 crc kubenswrapper[4907]: E0313 15:53:08.685936 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"321294b102a32d438b57983725dd613a9c65287a746583d505990a3f7122ab85\": container with ID starting with 321294b102a32d438b57983725dd613a9c65287a746583d505990a3f7122ab85 not found: ID does not exist" containerID="321294b102a32d438b57983725dd613a9c65287a746583d505990a3f7122ab85" Mar 13 15:53:08 crc kubenswrapper[4907]: I0313 15:53:08.685971 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"321294b102a32d438b57983725dd613a9c65287a746583d505990a3f7122ab85"} err="failed to get container status \"321294b102a32d438b57983725dd613a9c65287a746583d505990a3f7122ab85\": rpc error: code = NotFound desc = could not find container \"321294b102a32d438b57983725dd613a9c65287a746583d505990a3f7122ab85\": container with ID starting with 321294b102a32d438b57983725dd613a9c65287a746583d505990a3f7122ab85 not found: ID does not exist" Mar 13 15:53:08 crc kubenswrapper[4907]: I0313 15:53:08.685988 4907 scope.go:117] "RemoveContainer" containerID="80503b331ede81d6354f232f75d1fa04ac94047782d538502272582284122726" Mar 13 15:53:08 crc kubenswrapper[4907]: E0313 15:53:08.686276 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"80503b331ede81d6354f232f75d1fa04ac94047782d538502272582284122726\": container with ID starting with 80503b331ede81d6354f232f75d1fa04ac94047782d538502272582284122726 not found: ID does not exist" containerID="80503b331ede81d6354f232f75d1fa04ac94047782d538502272582284122726" Mar 13 15:53:08 crc kubenswrapper[4907]: I0313 15:53:08.686304 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80503b331ede81d6354f232f75d1fa04ac94047782d538502272582284122726"} err="failed to get container status \"80503b331ede81d6354f232f75d1fa04ac94047782d538502272582284122726\": rpc error: code = NotFound desc = could not find container \"80503b331ede81d6354f232f75d1fa04ac94047782d538502272582284122726\": container with ID starting with 80503b331ede81d6354f232f75d1fa04ac94047782d538502272582284122726 not found: ID does not exist" Mar 13 15:53:09 crc kubenswrapper[4907]: I0313 15:53:09.796389 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="63bc7557-0465-42dc-873c-a8eb54e03a45" path="/var/lib/kubelet/pods/63bc7557-0465-42dc-873c-a8eb54e03a45/volumes" Mar 13 15:53:14 crc kubenswrapper[4907]: I0313 15:53:14.783227 4907 scope.go:117] "RemoveContainer" containerID="33f8e315523143b157797243610ad7b82c7815f9e551d1169f2c036cfcadf792" Mar 13 15:53:14 crc kubenswrapper[4907]: E0313 15:53:14.783980 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:53:25 crc kubenswrapper[4907]: I0313 15:53:25.782344 4907 scope.go:117] "RemoveContainer" containerID="33f8e315523143b157797243610ad7b82c7815f9e551d1169f2c036cfcadf792" Mar 13 15:53:25 crc kubenswrapper[4907]: E0313 15:53:25.783143 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:53:38 crc kubenswrapper[4907]: I0313 15:53:38.783446 4907 scope.go:117] "RemoveContainer" containerID="33f8e315523143b157797243610ad7b82c7815f9e551d1169f2c036cfcadf792" Mar 13 15:53:38 crc kubenswrapper[4907]: E0313 15:53:38.784174 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:53:53 crc kubenswrapper[4907]: I0313 15:53:53.783648 4907 scope.go:117] "RemoveContainer" containerID="33f8e315523143b157797243610ad7b82c7815f9e551d1169f2c036cfcadf792" Mar 13 15:53:53 crc kubenswrapper[4907]: E0313 15:53:53.784374 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:54:00 crc kubenswrapper[4907]: I0313 15:54:00.171298 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556954-kn875"] Mar 13 15:54:00 crc kubenswrapper[4907]: E0313 15:54:00.173195 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63bc7557-0465-42dc-873c-a8eb54e03a45" containerName="extract-content" Mar 13 15:54:00 crc kubenswrapper[4907]: I0313 15:54:00.173600 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="63bc7557-0465-42dc-873c-a8eb54e03a45" containerName="extract-content" Mar 13 15:54:00 crc kubenswrapper[4907]: E0313 15:54:00.175212 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63bc7557-0465-42dc-873c-a8eb54e03a45" containerName="extract-utilities" Mar 13 15:54:00 crc kubenswrapper[4907]: I0313 15:54:00.175243 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="63bc7557-0465-42dc-873c-a8eb54e03a45" containerName="extract-utilities" Mar 13 15:54:00 crc kubenswrapper[4907]: E0313 15:54:00.175280 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63bc7557-0465-42dc-873c-a8eb54e03a45" containerName="registry-server" Mar 13 15:54:00 crc kubenswrapper[4907]: I0313 15:54:00.175290 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="63bc7557-0465-42dc-873c-a8eb54e03a45" containerName="registry-server" Mar 13 15:54:00 crc kubenswrapper[4907]: I0313 15:54:00.175740 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="63bc7557-0465-42dc-873c-a8eb54e03a45" containerName="registry-server" Mar 13 15:54:00 crc kubenswrapper[4907]: I0313 15:54:00.176725 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556954-kn875" Mar 13 15:54:00 crc kubenswrapper[4907]: I0313 15:54:00.178942 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 15:54:00 crc kubenswrapper[4907]: I0313 15:54:00.179077 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 15:54:00 crc kubenswrapper[4907]: I0313 15:54:00.180386 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 15:54:00 crc kubenswrapper[4907]: I0313 15:54:00.184272 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556954-kn875"] Mar 13 15:54:00 crc kubenswrapper[4907]: I0313 15:54:00.273505 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m79fh\" (UniqueName: \"kubernetes.io/projected/97bac75d-c6b8-4d2e-8a24-d0c6279347b5-kube-api-access-m79fh\") pod \"auto-csr-approver-29556954-kn875\" (UID: \"97bac75d-c6b8-4d2e-8a24-d0c6279347b5\") " pod="openshift-infra/auto-csr-approver-29556954-kn875" Mar 13 15:54:00 crc kubenswrapper[4907]: I0313 15:54:00.376460 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m79fh\" (UniqueName: \"kubernetes.io/projected/97bac75d-c6b8-4d2e-8a24-d0c6279347b5-kube-api-access-m79fh\") pod \"auto-csr-approver-29556954-kn875\" (UID: \"97bac75d-c6b8-4d2e-8a24-d0c6279347b5\") " pod="openshift-infra/auto-csr-approver-29556954-kn875" Mar 13 15:54:00 crc kubenswrapper[4907]: I0313 15:54:00.394853 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m79fh\" (UniqueName: \"kubernetes.io/projected/97bac75d-c6b8-4d2e-8a24-d0c6279347b5-kube-api-access-m79fh\") pod \"auto-csr-approver-29556954-kn875\" (UID: \"97bac75d-c6b8-4d2e-8a24-d0c6279347b5\") " pod="openshift-infra/auto-csr-approver-29556954-kn875" Mar 13 15:54:00 crc kubenswrapper[4907]: I0313 15:54:00.509514 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556954-kn875" Mar 13 15:54:00 crc kubenswrapper[4907]: I0313 15:54:00.842437 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556954-kn875"] Mar 13 15:54:00 crc kubenswrapper[4907]: I0313 15:54:00.842656 4907 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 13 15:54:01 crc kubenswrapper[4907]: I0313 15:54:01.094832 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556954-kn875" event={"ID":"97bac75d-c6b8-4d2e-8a24-d0c6279347b5","Type":"ContainerStarted","Data":"cc96e61a9c12fab292fda588e891fee3cfcbebd049975f4172aaba1347b1eabd"} Mar 13 15:54:02 crc kubenswrapper[4907]: E0313 15:54:02.640627 4907 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod97bac75d_c6b8_4d2e_8a24_d0c6279347b5.slice/crio-4dac6251d0188e6b289ee01e6cd109fb2dacb259e342a008dc8836c090408baa.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod97bac75d_c6b8_4d2e_8a24_d0c6279347b5.slice/crio-conmon-4dac6251d0188e6b289ee01e6cd109fb2dacb259e342a008dc8836c090408baa.scope\": RecentStats: unable to find data in memory cache]" Mar 13 15:54:03 crc kubenswrapper[4907]: I0313 15:54:03.114038 4907 generic.go:334] "Generic (PLEG): container finished" podID="97bac75d-c6b8-4d2e-8a24-d0c6279347b5" containerID="4dac6251d0188e6b289ee01e6cd109fb2dacb259e342a008dc8836c090408baa" exitCode=0 Mar 13 15:54:03 crc kubenswrapper[4907]: I0313 15:54:03.114109 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556954-kn875" event={"ID":"97bac75d-c6b8-4d2e-8a24-d0c6279347b5","Type":"ContainerDied","Data":"4dac6251d0188e6b289ee01e6cd109fb2dacb259e342a008dc8836c090408baa"} Mar 13 15:54:04 crc kubenswrapper[4907]: I0313 15:54:04.493463 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556954-kn875" Mar 13 15:54:04 crc kubenswrapper[4907]: I0313 15:54:04.572332 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m79fh\" (UniqueName: \"kubernetes.io/projected/97bac75d-c6b8-4d2e-8a24-d0c6279347b5-kube-api-access-m79fh\") pod \"97bac75d-c6b8-4d2e-8a24-d0c6279347b5\" (UID: \"97bac75d-c6b8-4d2e-8a24-d0c6279347b5\") " Mar 13 15:54:04 crc kubenswrapper[4907]: I0313 15:54:04.580312 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/97bac75d-c6b8-4d2e-8a24-d0c6279347b5-kube-api-access-m79fh" (OuterVolumeSpecName: "kube-api-access-m79fh") pod "97bac75d-c6b8-4d2e-8a24-d0c6279347b5" (UID: "97bac75d-c6b8-4d2e-8a24-d0c6279347b5"). InnerVolumeSpecName "kube-api-access-m79fh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:54:04 crc kubenswrapper[4907]: I0313 15:54:04.675730 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m79fh\" (UniqueName: \"kubernetes.io/projected/97bac75d-c6b8-4d2e-8a24-d0c6279347b5-kube-api-access-m79fh\") on node \"crc\" DevicePath \"\"" Mar 13 15:54:05 crc kubenswrapper[4907]: I0313 15:54:05.135683 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556954-kn875" event={"ID":"97bac75d-c6b8-4d2e-8a24-d0c6279347b5","Type":"ContainerDied","Data":"cc96e61a9c12fab292fda588e891fee3cfcbebd049975f4172aaba1347b1eabd"} Mar 13 15:54:05 crc kubenswrapper[4907]: I0313 15:54:05.135722 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cc96e61a9c12fab292fda588e891fee3cfcbebd049975f4172aaba1347b1eabd" Mar 13 15:54:05 crc kubenswrapper[4907]: I0313 15:54:05.135771 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556954-kn875" Mar 13 15:54:05 crc kubenswrapper[4907]: I0313 15:54:05.551593 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556948-mmprl"] Mar 13 15:54:05 crc kubenswrapper[4907]: I0313 15:54:05.560177 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556948-mmprl"] Mar 13 15:54:05 crc kubenswrapper[4907]: I0313 15:54:05.795088 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="673cadb2-82bd-4e26-8222-5a89c310f597" path="/var/lib/kubelet/pods/673cadb2-82bd-4e26-8222-5a89c310f597/volumes" Mar 13 15:54:08 crc kubenswrapper[4907]: I0313 15:54:08.783011 4907 scope.go:117] "RemoveContainer" containerID="33f8e315523143b157797243610ad7b82c7815f9e551d1169f2c036cfcadf792" Mar 13 15:54:08 crc kubenswrapper[4907]: E0313 15:54:08.783855 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:54:21 crc kubenswrapper[4907]: I0313 15:54:21.788554 4907 scope.go:117] "RemoveContainer" containerID="33f8e315523143b157797243610ad7b82c7815f9e551d1169f2c036cfcadf792" Mar 13 15:54:21 crc kubenswrapper[4907]: E0313 15:54:21.789360 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:54:24 crc kubenswrapper[4907]: I0313 15:54:24.043395 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-db-create-ghczh"] Mar 13 15:54:24 crc kubenswrapper[4907]: I0313 15:54:24.052372 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/octavia-db-create-ghczh"] Mar 13 15:54:25 crc kubenswrapper[4907]: I0313 15:54:25.798324 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0aa6e501-c175-44be-af00-02b8aac76bab" path="/var/lib/kubelet/pods/0aa6e501-c175-44be-af00-02b8aac76bab/volumes" Mar 13 15:54:26 crc kubenswrapper[4907]: I0313 15:54:26.026759 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-6234-account-create-update-wjdx9"] Mar 13 15:54:26 crc kubenswrapper[4907]: I0313 15:54:26.035415 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/octavia-6234-account-create-update-wjdx9"] Mar 13 15:54:27 crc kubenswrapper[4907]: I0313 15:54:27.810202 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c420aeff-7a0b-4efa-b9c0-a93791eb0378" path="/var/lib/kubelet/pods/c420aeff-7a0b-4efa-b9c0-a93791eb0378/volumes" Mar 13 15:54:32 crc kubenswrapper[4907]: I0313 15:54:32.027504 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-persistence-db-create-p9hf2"] Mar 13 15:54:32 crc kubenswrapper[4907]: I0313 15:54:32.037450 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/octavia-persistence-db-create-p9hf2"] Mar 13 15:54:33 crc kubenswrapper[4907]: I0313 15:54:33.030580 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-1da8-account-create-update-82kpd"] Mar 13 15:54:33 crc kubenswrapper[4907]: I0313 15:54:33.039453 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/octavia-1da8-account-create-update-82kpd"] Mar 13 15:54:33 crc kubenswrapper[4907]: I0313 15:54:33.796128 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3b46488e-1fef-479f-94b6-1958e5174219" path="/var/lib/kubelet/pods/3b46488e-1fef-479f-94b6-1958e5174219/volumes" Mar 13 15:54:33 crc kubenswrapper[4907]: I0313 15:54:33.798278 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d218d11d-2afc-4f21-add2-af15cd8c50ad" path="/var/lib/kubelet/pods/d218d11d-2afc-4f21-add2-af15cd8c50ad/volumes" Mar 13 15:54:36 crc kubenswrapper[4907]: I0313 15:54:36.781966 4907 scope.go:117] "RemoveContainer" containerID="33f8e315523143b157797243610ad7b82c7815f9e551d1169f2c036cfcadf792" Mar 13 15:54:36 crc kubenswrapper[4907]: E0313 15:54:36.783445 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:54:39 crc kubenswrapper[4907]: I0313 15:54:39.360263 4907 scope.go:117] "RemoveContainer" containerID="1d71acbd6f06389aec2ad567011e53ff359b678fb6ad5c86cc822259fbdd55b9" Mar 13 15:54:39 crc kubenswrapper[4907]: I0313 15:54:39.384895 4907 scope.go:117] "RemoveContainer" containerID="ef160b1245008062ab4f675d742f68b73acd696e5db3bbb49ff009cadb324f56" Mar 13 15:54:39 crc kubenswrapper[4907]: I0313 15:54:39.436969 4907 scope.go:117] "RemoveContainer" containerID="48d98d79386ad3e2c5200a51cdf22425a3d47af926f18ea01f01295cd4a45c13" Mar 13 15:54:39 crc kubenswrapper[4907]: I0313 15:54:39.493430 4907 scope.go:117] "RemoveContainer" containerID="f56058903e847c31c47f45cadeb76ef300892a82a5c55a4fd28b47f0d9e1eeec" Mar 13 15:54:39 crc kubenswrapper[4907]: I0313 15:54:39.534997 4907 scope.go:117] "RemoveContainer" containerID="e63bf49f2a3afc218c4b126f220439d3e3bcaa22e7a77cab733b2ce5385658bb" Mar 13 15:54:51 crc kubenswrapper[4907]: I0313 15:54:51.790019 4907 scope.go:117] "RemoveContainer" containerID="33f8e315523143b157797243610ad7b82c7815f9e551d1169f2c036cfcadf792" Mar 13 15:54:51 crc kubenswrapper[4907]: E0313 15:54:51.791042 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:55:04 crc kubenswrapper[4907]: I0313 15:55:04.782813 4907 scope.go:117] "RemoveContainer" containerID="33f8e315523143b157797243610ad7b82c7815f9e551d1169f2c036cfcadf792" Mar 13 15:55:04 crc kubenswrapper[4907]: E0313 15:55:04.784586 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:55:17 crc kubenswrapper[4907]: I0313 15:55:17.046382 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-db-sync-rjfnd"] Mar 13 15:55:17 crc kubenswrapper[4907]: I0313 15:55:17.057513 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/octavia-db-sync-rjfnd"] Mar 13 15:55:17 crc kubenswrapper[4907]: I0313 15:55:17.795010 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aefbf41f-fa5a-4763-af85-3c9e8df33b6a" path="/var/lib/kubelet/pods/aefbf41f-fa5a-4763-af85-3c9e8df33b6a/volumes" Mar 13 15:55:19 crc kubenswrapper[4907]: I0313 15:55:19.782536 4907 scope.go:117] "RemoveContainer" containerID="33f8e315523143b157797243610ad7b82c7815f9e551d1169f2c036cfcadf792" Mar 13 15:55:19 crc kubenswrapper[4907]: E0313 15:55:19.784021 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:55:32 crc kubenswrapper[4907]: I0313 15:55:32.782925 4907 scope.go:117] "RemoveContainer" containerID="33f8e315523143b157797243610ad7b82c7815f9e551d1169f2c036cfcadf792" Mar 13 15:55:32 crc kubenswrapper[4907]: E0313 15:55:32.783821 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:55:39 crc kubenswrapper[4907]: I0313 15:55:39.663851 4907 scope.go:117] "RemoveContainer" containerID="94c59567d9e99f3b6037554dff6847cedf9844e341b49a4a37077ea52d3428be" Mar 13 15:55:39 crc kubenswrapper[4907]: I0313 15:55:39.692297 4907 scope.go:117] "RemoveContainer" containerID="4586239eb0dcd4e94af6c8d52a6990fcfc7c67962591f11127b749cbaee23663" Mar 13 15:55:44 crc kubenswrapper[4907]: I0313 15:55:44.783122 4907 scope.go:117] "RemoveContainer" containerID="33f8e315523143b157797243610ad7b82c7815f9e551d1169f2c036cfcadf792" Mar 13 15:55:44 crc kubenswrapper[4907]: E0313 15:55:44.784007 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:55:58 crc kubenswrapper[4907]: I0313 15:55:58.783111 4907 scope.go:117] "RemoveContainer" containerID="33f8e315523143b157797243610ad7b82c7815f9e551d1169f2c036cfcadf792" Mar 13 15:55:58 crc kubenswrapper[4907]: E0313 15:55:58.784118 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:56:00 crc kubenswrapper[4907]: I0313 15:56:00.158937 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556956-txd4m"] Mar 13 15:56:00 crc kubenswrapper[4907]: E0313 15:56:00.159865 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97bac75d-c6b8-4d2e-8a24-d0c6279347b5" containerName="oc" Mar 13 15:56:00 crc kubenswrapper[4907]: I0313 15:56:00.159902 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="97bac75d-c6b8-4d2e-8a24-d0c6279347b5" containerName="oc" Mar 13 15:56:00 crc kubenswrapper[4907]: I0313 15:56:00.160189 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="97bac75d-c6b8-4d2e-8a24-d0c6279347b5" containerName="oc" Mar 13 15:56:00 crc kubenswrapper[4907]: I0313 15:56:00.161213 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556956-txd4m" Mar 13 15:56:00 crc kubenswrapper[4907]: I0313 15:56:00.163431 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 15:56:00 crc kubenswrapper[4907]: I0313 15:56:00.163726 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 15:56:00 crc kubenswrapper[4907]: I0313 15:56:00.167259 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rg9kx\" (UniqueName: \"kubernetes.io/projected/a22eedb1-70a8-4232-9a54-8625a4dd7022-kube-api-access-rg9kx\") pod \"auto-csr-approver-29556956-txd4m\" (UID: \"a22eedb1-70a8-4232-9a54-8625a4dd7022\") " pod="openshift-infra/auto-csr-approver-29556956-txd4m" Mar 13 15:56:00 crc kubenswrapper[4907]: I0313 15:56:00.167487 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 15:56:00 crc kubenswrapper[4907]: I0313 15:56:00.184870 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556956-txd4m"] Mar 13 15:56:00 crc kubenswrapper[4907]: I0313 15:56:00.269436 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rg9kx\" (UniqueName: \"kubernetes.io/projected/a22eedb1-70a8-4232-9a54-8625a4dd7022-kube-api-access-rg9kx\") pod \"auto-csr-approver-29556956-txd4m\" (UID: \"a22eedb1-70a8-4232-9a54-8625a4dd7022\") " pod="openshift-infra/auto-csr-approver-29556956-txd4m" Mar 13 15:56:00 crc kubenswrapper[4907]: I0313 15:56:00.291124 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rg9kx\" (UniqueName: \"kubernetes.io/projected/a22eedb1-70a8-4232-9a54-8625a4dd7022-kube-api-access-rg9kx\") pod \"auto-csr-approver-29556956-txd4m\" (UID: \"a22eedb1-70a8-4232-9a54-8625a4dd7022\") " pod="openshift-infra/auto-csr-approver-29556956-txd4m" Mar 13 15:56:00 crc kubenswrapper[4907]: I0313 15:56:00.488583 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556956-txd4m" Mar 13 15:56:00 crc kubenswrapper[4907]: I0313 15:56:00.970362 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556956-txd4m"] Mar 13 15:56:01 crc kubenswrapper[4907]: I0313 15:56:01.252657 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556956-txd4m" event={"ID":"a22eedb1-70a8-4232-9a54-8625a4dd7022","Type":"ContainerStarted","Data":"2493883c79e986e8d864cbb5316247394c34dc9c3fd5ce98c48828fdd9942fc7"} Mar 13 15:56:03 crc kubenswrapper[4907]: I0313 15:56:03.279491 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556956-txd4m" event={"ID":"a22eedb1-70a8-4232-9a54-8625a4dd7022","Type":"ContainerDied","Data":"76fd8f67a17de0068ab3b4548c472b268fc93309cdb749646d71dbe2fb8aef2e"} Mar 13 15:56:03 crc kubenswrapper[4907]: I0313 15:56:03.279507 4907 generic.go:334] "Generic (PLEG): container finished" podID="a22eedb1-70a8-4232-9a54-8625a4dd7022" containerID="76fd8f67a17de0068ab3b4548c472b268fc93309cdb749646d71dbe2fb8aef2e" exitCode=0 Mar 13 15:56:04 crc kubenswrapper[4907]: I0313 15:56:04.681166 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556956-txd4m" Mar 13 15:56:04 crc kubenswrapper[4907]: I0313 15:56:04.884136 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rg9kx\" (UniqueName: \"kubernetes.io/projected/a22eedb1-70a8-4232-9a54-8625a4dd7022-kube-api-access-rg9kx\") pod \"a22eedb1-70a8-4232-9a54-8625a4dd7022\" (UID: \"a22eedb1-70a8-4232-9a54-8625a4dd7022\") " Mar 13 15:56:04 crc kubenswrapper[4907]: I0313 15:56:04.893228 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a22eedb1-70a8-4232-9a54-8625a4dd7022-kube-api-access-rg9kx" (OuterVolumeSpecName: "kube-api-access-rg9kx") pod "a22eedb1-70a8-4232-9a54-8625a4dd7022" (UID: "a22eedb1-70a8-4232-9a54-8625a4dd7022"). InnerVolumeSpecName "kube-api-access-rg9kx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:56:04 crc kubenswrapper[4907]: I0313 15:56:04.987927 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rg9kx\" (UniqueName: \"kubernetes.io/projected/a22eedb1-70a8-4232-9a54-8625a4dd7022-kube-api-access-rg9kx\") on node \"crc\" DevicePath \"\"" Mar 13 15:56:05 crc kubenswrapper[4907]: I0313 15:56:05.303713 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556956-txd4m" event={"ID":"a22eedb1-70a8-4232-9a54-8625a4dd7022","Type":"ContainerDied","Data":"2493883c79e986e8d864cbb5316247394c34dc9c3fd5ce98c48828fdd9942fc7"} Mar 13 15:56:05 crc kubenswrapper[4907]: I0313 15:56:05.304335 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2493883c79e986e8d864cbb5316247394c34dc9c3fd5ce98c48828fdd9942fc7" Mar 13 15:56:05 crc kubenswrapper[4907]: I0313 15:56:05.303767 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556956-txd4m" Mar 13 15:56:05 crc kubenswrapper[4907]: I0313 15:56:05.774942 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556950-pqwgk"] Mar 13 15:56:05 crc kubenswrapper[4907]: I0313 15:56:05.798431 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556950-pqwgk"] Mar 13 15:56:07 crc kubenswrapper[4907]: I0313 15:56:07.795675 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="63d124ed-9ee2-437b-8ebd-b38a53e499af" path="/var/lib/kubelet/pods/63d124ed-9ee2-437b-8ebd-b38a53e499af/volumes" Mar 13 15:56:10 crc kubenswrapper[4907]: I0313 15:56:10.783684 4907 scope.go:117] "RemoveContainer" containerID="33f8e315523143b157797243610ad7b82c7815f9e551d1169f2c036cfcadf792" Mar 13 15:56:10 crc kubenswrapper[4907]: E0313 15:56:10.785369 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:56:23 crc kubenswrapper[4907]: I0313 15:56:23.782640 4907 scope.go:117] "RemoveContainer" containerID="33f8e315523143b157797243610ad7b82c7815f9e551d1169f2c036cfcadf792" Mar 13 15:56:23 crc kubenswrapper[4907]: E0313 15:56:23.783323 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:56:37 crc kubenswrapper[4907]: I0313 15:56:37.782754 4907 scope.go:117] "RemoveContainer" containerID="33f8e315523143b157797243610ad7b82c7815f9e551d1169f2c036cfcadf792" Mar 13 15:56:37 crc kubenswrapper[4907]: E0313 15:56:37.783585 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:56:39 crc kubenswrapper[4907]: I0313 15:56:39.792565 4907 scope.go:117] "RemoveContainer" containerID="fa4769654b9ebff08198d548ba01872a5dbbbcb75936088b49f7d253d89a76c0" Mar 13 15:56:49 crc kubenswrapper[4907]: I0313 15:56:49.783820 4907 scope.go:117] "RemoveContainer" containerID="33f8e315523143b157797243610ad7b82c7815f9e551d1169f2c036cfcadf792" Mar 13 15:56:49 crc kubenswrapper[4907]: E0313 15:56:49.784554 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:57:02 crc kubenswrapper[4907]: I0313 15:57:02.783519 4907 scope.go:117] "RemoveContainer" containerID="33f8e315523143b157797243610ad7b82c7815f9e551d1169f2c036cfcadf792" Mar 13 15:57:02 crc kubenswrapper[4907]: E0313 15:57:02.786387 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:57:14 crc kubenswrapper[4907]: I0313 15:57:14.062843 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-019b-account-create-update-848sw"] Mar 13 15:57:14 crc kubenswrapper[4907]: I0313 15:57:14.080330 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-create-lnsws"] Mar 13 15:57:14 crc kubenswrapper[4907]: I0313 15:57:14.089104 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-create-lnsws"] Mar 13 15:57:14 crc kubenswrapper[4907]: I0313 15:57:14.096960 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-019b-account-create-update-848sw"] Mar 13 15:57:15 crc kubenswrapper[4907]: I0313 15:57:15.796459 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="490bd531-09f6-48e5-b194-b79e9b532131" path="/var/lib/kubelet/pods/490bd531-09f6-48e5-b194-b79e9b532131/volumes" Mar 13 15:57:15 crc kubenswrapper[4907]: I0313 15:57:15.797463 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ce0b08b9-c4e1-4236-8f54-cf78b45d2d49" path="/var/lib/kubelet/pods/ce0b08b9-c4e1-4236-8f54-cf78b45d2d49/volumes" Mar 13 15:57:17 crc kubenswrapper[4907]: I0313 15:57:17.782306 4907 scope.go:117] "RemoveContainer" containerID="33f8e315523143b157797243610ad7b82c7815f9e551d1169f2c036cfcadf792" Mar 13 15:57:17 crc kubenswrapper[4907]: E0313 15:57:17.782826 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:57:30 crc kubenswrapper[4907]: I0313 15:57:30.038118 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-sync-mqnk8"] Mar 13 15:57:30 crc kubenswrapper[4907]: I0313 15:57:30.054117 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-sync-mqnk8"] Mar 13 15:57:30 crc kubenswrapper[4907]: I0313 15:57:30.782589 4907 scope.go:117] "RemoveContainer" containerID="33f8e315523143b157797243610ad7b82c7815f9e551d1169f2c036cfcadf792" Mar 13 15:57:30 crc kubenswrapper[4907]: E0313 15:57:30.782849 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:57:31 crc kubenswrapper[4907]: I0313 15:57:31.793174 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6df151b5-1c14-488c-9f23-9d6d629b493d" path="/var/lib/kubelet/pods/6df151b5-1c14-488c-9f23-9d6d629b493d/volumes" Mar 13 15:57:39 crc kubenswrapper[4907]: I0313 15:57:39.879768 4907 scope.go:117] "RemoveContainer" containerID="daa82cb23c63e51c880db96744e0980e9ddf6f38f0cbcf2713496cba5c2d4486" Mar 13 15:57:39 crc kubenswrapper[4907]: I0313 15:57:39.913422 4907 scope.go:117] "RemoveContainer" containerID="839f8683996fb111c90ced80fc315df41501ccd29aa3ce62746269e8e8d6d880" Mar 13 15:57:39 crc kubenswrapper[4907]: I0313 15:57:39.959639 4907 scope.go:117] "RemoveContainer" containerID="28274d3b65f2c9d6e1b8e4c19babc239c1199e687ba7bf22cb4ab37476252833" Mar 13 15:57:41 crc kubenswrapper[4907]: I0313 15:57:41.794786 4907 scope.go:117] "RemoveContainer" containerID="33f8e315523143b157797243610ad7b82c7815f9e551d1169f2c036cfcadf792" Mar 13 15:57:41 crc kubenswrapper[4907]: E0313 15:57:41.795247 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:57:52 crc kubenswrapper[4907]: I0313 15:57:52.782299 4907 scope.go:117] "RemoveContainer" containerID="33f8e315523143b157797243610ad7b82c7815f9e551d1169f2c036cfcadf792" Mar 13 15:57:53 crc kubenswrapper[4907]: I0313 15:57:53.350619 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerStarted","Data":"2a3c556bf9f1ef9e4af9dec2c5d043c5ea60bb781b3c04fb7c2299179fb58a19"} Mar 13 15:58:00 crc kubenswrapper[4907]: I0313 15:58:00.152207 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556958-dk8dz"] Mar 13 15:58:00 crc kubenswrapper[4907]: E0313 15:58:00.154683 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a22eedb1-70a8-4232-9a54-8625a4dd7022" containerName="oc" Mar 13 15:58:00 crc kubenswrapper[4907]: I0313 15:58:00.154814 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="a22eedb1-70a8-4232-9a54-8625a4dd7022" containerName="oc" Mar 13 15:58:00 crc kubenswrapper[4907]: I0313 15:58:00.155312 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="a22eedb1-70a8-4232-9a54-8625a4dd7022" containerName="oc" Mar 13 15:58:00 crc kubenswrapper[4907]: I0313 15:58:00.156654 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556958-dk8dz" Mar 13 15:58:00 crc kubenswrapper[4907]: I0313 15:58:00.159076 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 15:58:00 crc kubenswrapper[4907]: I0313 15:58:00.159716 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 15:58:00 crc kubenswrapper[4907]: I0313 15:58:00.160136 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 15:58:00 crc kubenswrapper[4907]: I0313 15:58:00.167185 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556958-dk8dz"] Mar 13 15:58:00 crc kubenswrapper[4907]: I0313 15:58:00.314114 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4ncj8\" (UniqueName: \"kubernetes.io/projected/2d38cd8e-5a08-4c56-b35e-88617df1d3a8-kube-api-access-4ncj8\") pod \"auto-csr-approver-29556958-dk8dz\" (UID: \"2d38cd8e-5a08-4c56-b35e-88617df1d3a8\") " pod="openshift-infra/auto-csr-approver-29556958-dk8dz" Mar 13 15:58:00 crc kubenswrapper[4907]: I0313 15:58:00.416202 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4ncj8\" (UniqueName: \"kubernetes.io/projected/2d38cd8e-5a08-4c56-b35e-88617df1d3a8-kube-api-access-4ncj8\") pod \"auto-csr-approver-29556958-dk8dz\" (UID: \"2d38cd8e-5a08-4c56-b35e-88617df1d3a8\") " pod="openshift-infra/auto-csr-approver-29556958-dk8dz" Mar 13 15:58:00 crc kubenswrapper[4907]: I0313 15:58:00.437861 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4ncj8\" (UniqueName: \"kubernetes.io/projected/2d38cd8e-5a08-4c56-b35e-88617df1d3a8-kube-api-access-4ncj8\") pod \"auto-csr-approver-29556958-dk8dz\" (UID: \"2d38cd8e-5a08-4c56-b35e-88617df1d3a8\") " pod="openshift-infra/auto-csr-approver-29556958-dk8dz" Mar 13 15:58:00 crc kubenswrapper[4907]: I0313 15:58:00.490162 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556958-dk8dz" Mar 13 15:58:00 crc kubenswrapper[4907]: I0313 15:58:00.962665 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556958-dk8dz"] Mar 13 15:58:01 crc kubenswrapper[4907]: I0313 15:58:01.427571 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556958-dk8dz" event={"ID":"2d38cd8e-5a08-4c56-b35e-88617df1d3a8","Type":"ContainerStarted","Data":"bb6038124e3db92e462b38fae6dbf334e4cb83a55a475138ec56e8d71be7fae6"} Mar 13 15:58:02 crc kubenswrapper[4907]: I0313 15:58:02.439170 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556958-dk8dz" event={"ID":"2d38cd8e-5a08-4c56-b35e-88617df1d3a8","Type":"ContainerStarted","Data":"d5d998084a23eab3592589afda6539d62cf5a3559ead1f830c6d581d1f9410c7"} Mar 13 15:58:02 crc kubenswrapper[4907]: I0313 15:58:02.460515 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29556958-dk8dz" podStartSLOduration=1.320104216 podStartE2EDuration="2.460496856s" podCreationTimestamp="2026-03-13 15:58:00 +0000 UTC" firstStartedPulling="2026-03-13 15:58:00.966859772 +0000 UTC m=+6779.866647451" lastFinishedPulling="2026-03-13 15:58:02.107252382 +0000 UTC m=+6781.007040091" observedRunningTime="2026-03-13 15:58:02.450678037 +0000 UTC m=+6781.350465726" watchObservedRunningTime="2026-03-13 15:58:02.460496856 +0000 UTC m=+6781.360284545" Mar 13 15:58:03 crc kubenswrapper[4907]: I0313 15:58:03.449041 4907 generic.go:334] "Generic (PLEG): container finished" podID="2d38cd8e-5a08-4c56-b35e-88617df1d3a8" containerID="d5d998084a23eab3592589afda6539d62cf5a3559ead1f830c6d581d1f9410c7" exitCode=0 Mar 13 15:58:03 crc kubenswrapper[4907]: I0313 15:58:03.449150 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556958-dk8dz" event={"ID":"2d38cd8e-5a08-4c56-b35e-88617df1d3a8","Type":"ContainerDied","Data":"d5d998084a23eab3592589afda6539d62cf5a3559ead1f830c6d581d1f9410c7"} Mar 13 15:58:04 crc kubenswrapper[4907]: I0313 15:58:04.828297 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556958-dk8dz" Mar 13 15:58:04 crc kubenswrapper[4907]: I0313 15:58:04.911970 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4ncj8\" (UniqueName: \"kubernetes.io/projected/2d38cd8e-5a08-4c56-b35e-88617df1d3a8-kube-api-access-4ncj8\") pod \"2d38cd8e-5a08-4c56-b35e-88617df1d3a8\" (UID: \"2d38cd8e-5a08-4c56-b35e-88617df1d3a8\") " Mar 13 15:58:04 crc kubenswrapper[4907]: I0313 15:58:04.920388 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2d38cd8e-5a08-4c56-b35e-88617df1d3a8-kube-api-access-4ncj8" (OuterVolumeSpecName: "kube-api-access-4ncj8") pod "2d38cd8e-5a08-4c56-b35e-88617df1d3a8" (UID: "2d38cd8e-5a08-4c56-b35e-88617df1d3a8"). InnerVolumeSpecName "kube-api-access-4ncj8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:58:05 crc kubenswrapper[4907]: I0313 15:58:05.014754 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4ncj8\" (UniqueName: \"kubernetes.io/projected/2d38cd8e-5a08-4c56-b35e-88617df1d3a8-kube-api-access-4ncj8\") on node \"crc\" DevicePath \"\"" Mar 13 15:58:05 crc kubenswrapper[4907]: I0313 15:58:05.469814 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556958-dk8dz" event={"ID":"2d38cd8e-5a08-4c56-b35e-88617df1d3a8","Type":"ContainerDied","Data":"bb6038124e3db92e462b38fae6dbf334e4cb83a55a475138ec56e8d71be7fae6"} Mar 13 15:58:05 crc kubenswrapper[4907]: I0313 15:58:05.469852 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bb6038124e3db92e462b38fae6dbf334e4cb83a55a475138ec56e8d71be7fae6" Mar 13 15:58:05 crc kubenswrapper[4907]: I0313 15:58:05.469869 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556958-dk8dz" Mar 13 15:58:05 crc kubenswrapper[4907]: I0313 15:58:05.903818 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556952-b69cd"] Mar 13 15:58:05 crc kubenswrapper[4907]: I0313 15:58:05.912515 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556952-b69cd"] Mar 13 15:58:07 crc kubenswrapper[4907]: I0313 15:58:07.794448 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d2915500-b9a0-45a4-960e-5325f53b445c" path="/var/lib/kubelet/pods/d2915500-b9a0-45a4-960e-5325f53b445c/volumes" Mar 13 15:58:16 crc kubenswrapper[4907]: I0313 15:58:16.421834 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-pk7rd"] Mar 13 15:58:16 crc kubenswrapper[4907]: E0313 15:58:16.422929 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d38cd8e-5a08-4c56-b35e-88617df1d3a8" containerName="oc" Mar 13 15:58:16 crc kubenswrapper[4907]: I0313 15:58:16.422946 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d38cd8e-5a08-4c56-b35e-88617df1d3a8" containerName="oc" Mar 13 15:58:16 crc kubenswrapper[4907]: I0313 15:58:16.423232 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d38cd8e-5a08-4c56-b35e-88617df1d3a8" containerName="oc" Mar 13 15:58:16 crc kubenswrapper[4907]: I0313 15:58:16.425182 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pk7rd" Mar 13 15:58:16 crc kubenswrapper[4907]: I0313 15:58:16.439117 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-pk7rd"] Mar 13 15:58:16 crc kubenswrapper[4907]: I0313 15:58:16.583204 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac68b7dd-54d2-4691-ac90-4f6f5ff0bc43-catalog-content\") pod \"community-operators-pk7rd\" (UID: \"ac68b7dd-54d2-4691-ac90-4f6f5ff0bc43\") " pod="openshift-marketplace/community-operators-pk7rd" Mar 13 15:58:16 crc kubenswrapper[4907]: I0313 15:58:16.583291 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rrgtc\" (UniqueName: \"kubernetes.io/projected/ac68b7dd-54d2-4691-ac90-4f6f5ff0bc43-kube-api-access-rrgtc\") pod \"community-operators-pk7rd\" (UID: \"ac68b7dd-54d2-4691-ac90-4f6f5ff0bc43\") " pod="openshift-marketplace/community-operators-pk7rd" Mar 13 15:58:16 crc kubenswrapper[4907]: I0313 15:58:16.583595 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac68b7dd-54d2-4691-ac90-4f6f5ff0bc43-utilities\") pod \"community-operators-pk7rd\" (UID: \"ac68b7dd-54d2-4691-ac90-4f6f5ff0bc43\") " pod="openshift-marketplace/community-operators-pk7rd" Mar 13 15:58:16 crc kubenswrapper[4907]: I0313 15:58:16.685822 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac68b7dd-54d2-4691-ac90-4f6f5ff0bc43-catalog-content\") pod \"community-operators-pk7rd\" (UID: \"ac68b7dd-54d2-4691-ac90-4f6f5ff0bc43\") " pod="openshift-marketplace/community-operators-pk7rd" Mar 13 15:58:16 crc kubenswrapper[4907]: I0313 15:58:16.685972 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rrgtc\" (UniqueName: \"kubernetes.io/projected/ac68b7dd-54d2-4691-ac90-4f6f5ff0bc43-kube-api-access-rrgtc\") pod \"community-operators-pk7rd\" (UID: \"ac68b7dd-54d2-4691-ac90-4f6f5ff0bc43\") " pod="openshift-marketplace/community-operators-pk7rd" Mar 13 15:58:16 crc kubenswrapper[4907]: I0313 15:58:16.686107 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac68b7dd-54d2-4691-ac90-4f6f5ff0bc43-utilities\") pod \"community-operators-pk7rd\" (UID: \"ac68b7dd-54d2-4691-ac90-4f6f5ff0bc43\") " pod="openshift-marketplace/community-operators-pk7rd" Mar 13 15:58:16 crc kubenswrapper[4907]: I0313 15:58:16.686352 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac68b7dd-54d2-4691-ac90-4f6f5ff0bc43-catalog-content\") pod \"community-operators-pk7rd\" (UID: \"ac68b7dd-54d2-4691-ac90-4f6f5ff0bc43\") " pod="openshift-marketplace/community-operators-pk7rd" Mar 13 15:58:16 crc kubenswrapper[4907]: I0313 15:58:16.686599 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac68b7dd-54d2-4691-ac90-4f6f5ff0bc43-utilities\") pod \"community-operators-pk7rd\" (UID: \"ac68b7dd-54d2-4691-ac90-4f6f5ff0bc43\") " pod="openshift-marketplace/community-operators-pk7rd" Mar 13 15:58:16 crc kubenswrapper[4907]: I0313 15:58:16.707435 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rrgtc\" (UniqueName: \"kubernetes.io/projected/ac68b7dd-54d2-4691-ac90-4f6f5ff0bc43-kube-api-access-rrgtc\") pod \"community-operators-pk7rd\" (UID: \"ac68b7dd-54d2-4691-ac90-4f6f5ff0bc43\") " pod="openshift-marketplace/community-operators-pk7rd" Mar 13 15:58:16 crc kubenswrapper[4907]: I0313 15:58:16.803663 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pk7rd" Mar 13 15:58:17 crc kubenswrapper[4907]: I0313 15:58:17.387500 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-pk7rd"] Mar 13 15:58:17 crc kubenswrapper[4907]: I0313 15:58:17.611469 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pk7rd" event={"ID":"ac68b7dd-54d2-4691-ac90-4f6f5ff0bc43","Type":"ContainerStarted","Data":"53b83ec7d94b2543162fe15e2a0e9e0d2a45294fa4e4b0c3681b0c95abfa7080"} Mar 13 15:58:17 crc kubenswrapper[4907]: I0313 15:58:17.611531 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pk7rd" event={"ID":"ac68b7dd-54d2-4691-ac90-4f6f5ff0bc43","Type":"ContainerStarted","Data":"31bb1f1b4403f006239e8d040d3abbb5c7a93523d9ec4c3f298d90363dac8352"} Mar 13 15:58:18 crc kubenswrapper[4907]: I0313 15:58:18.622465 4907 generic.go:334] "Generic (PLEG): container finished" podID="ac68b7dd-54d2-4691-ac90-4f6f5ff0bc43" containerID="53b83ec7d94b2543162fe15e2a0e9e0d2a45294fa4e4b0c3681b0c95abfa7080" exitCode=0 Mar 13 15:58:18 crc kubenswrapper[4907]: I0313 15:58:18.622657 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pk7rd" event={"ID":"ac68b7dd-54d2-4691-ac90-4f6f5ff0bc43","Type":"ContainerDied","Data":"53b83ec7d94b2543162fe15e2a0e9e0d2a45294fa4e4b0c3681b0c95abfa7080"} Mar 13 15:58:19 crc kubenswrapper[4907]: I0313 15:58:19.632262 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pk7rd" event={"ID":"ac68b7dd-54d2-4691-ac90-4f6f5ff0bc43","Type":"ContainerStarted","Data":"db5c85295b78a586f9a9d79869cb88866e67ead5e0fb9606d34998d9455a4702"} Mar 13 15:58:21 crc kubenswrapper[4907]: I0313 15:58:21.651818 4907 generic.go:334] "Generic (PLEG): container finished" podID="ac68b7dd-54d2-4691-ac90-4f6f5ff0bc43" containerID="db5c85295b78a586f9a9d79869cb88866e67ead5e0fb9606d34998d9455a4702" exitCode=0 Mar 13 15:58:21 crc kubenswrapper[4907]: I0313 15:58:21.651916 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pk7rd" event={"ID":"ac68b7dd-54d2-4691-ac90-4f6f5ff0bc43","Type":"ContainerDied","Data":"db5c85295b78a586f9a9d79869cb88866e67ead5e0fb9606d34998d9455a4702"} Mar 13 15:58:22 crc kubenswrapper[4907]: I0313 15:58:22.663144 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pk7rd" event={"ID":"ac68b7dd-54d2-4691-ac90-4f6f5ff0bc43","Type":"ContainerStarted","Data":"ffa10747043fef78a41d5b63379aa594a58eda0c580b0fdb3ed18e4cc5b00c07"} Mar 13 15:58:22 crc kubenswrapper[4907]: I0313 15:58:22.680565 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-pk7rd" podStartSLOduration=2.931828724 podStartE2EDuration="6.680548074s" podCreationTimestamp="2026-03-13 15:58:16 +0000 UTC" firstStartedPulling="2026-03-13 15:58:18.6263783 +0000 UTC m=+6797.526165989" lastFinishedPulling="2026-03-13 15:58:22.37509764 +0000 UTC m=+6801.274885339" observedRunningTime="2026-03-13 15:58:22.677209833 +0000 UTC m=+6801.576997542" watchObservedRunningTime="2026-03-13 15:58:22.680548074 +0000 UTC m=+6801.580335773" Mar 13 15:58:26 crc kubenswrapper[4907]: I0313 15:58:26.803928 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-pk7rd" Mar 13 15:58:26 crc kubenswrapper[4907]: I0313 15:58:26.804580 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-pk7rd" Mar 13 15:58:26 crc kubenswrapper[4907]: I0313 15:58:26.853282 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-pk7rd" Mar 13 15:58:27 crc kubenswrapper[4907]: I0313 15:58:27.764628 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-pk7rd" Mar 13 15:58:27 crc kubenswrapper[4907]: I0313 15:58:27.816603 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-pk7rd"] Mar 13 15:58:29 crc kubenswrapper[4907]: I0313 15:58:29.737390 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-pk7rd" podUID="ac68b7dd-54d2-4691-ac90-4f6f5ff0bc43" containerName="registry-server" containerID="cri-o://ffa10747043fef78a41d5b63379aa594a58eda0c580b0fdb3ed18e4cc5b00c07" gracePeriod=2 Mar 13 15:58:29 crc kubenswrapper[4907]: E0313 15:58:29.799798 4907 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podac68b7dd_54d2_4691_ac90_4f6f5ff0bc43.slice/crio-ffa10747043fef78a41d5b63379aa594a58eda0c580b0fdb3ed18e4cc5b00c07.scope\": RecentStats: unable to find data in memory cache]" Mar 13 15:58:30 crc kubenswrapper[4907]: I0313 15:58:30.240668 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pk7rd" Mar 13 15:58:30 crc kubenswrapper[4907]: I0313 15:58:30.400587 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rrgtc\" (UniqueName: \"kubernetes.io/projected/ac68b7dd-54d2-4691-ac90-4f6f5ff0bc43-kube-api-access-rrgtc\") pod \"ac68b7dd-54d2-4691-ac90-4f6f5ff0bc43\" (UID: \"ac68b7dd-54d2-4691-ac90-4f6f5ff0bc43\") " Mar 13 15:58:30 crc kubenswrapper[4907]: I0313 15:58:30.400840 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac68b7dd-54d2-4691-ac90-4f6f5ff0bc43-utilities\") pod \"ac68b7dd-54d2-4691-ac90-4f6f5ff0bc43\" (UID: \"ac68b7dd-54d2-4691-ac90-4f6f5ff0bc43\") " Mar 13 15:58:30 crc kubenswrapper[4907]: I0313 15:58:30.400917 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac68b7dd-54d2-4691-ac90-4f6f5ff0bc43-catalog-content\") pod \"ac68b7dd-54d2-4691-ac90-4f6f5ff0bc43\" (UID: \"ac68b7dd-54d2-4691-ac90-4f6f5ff0bc43\") " Mar 13 15:58:30 crc kubenswrapper[4907]: I0313 15:58:30.402386 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ac68b7dd-54d2-4691-ac90-4f6f5ff0bc43-utilities" (OuterVolumeSpecName: "utilities") pod "ac68b7dd-54d2-4691-ac90-4f6f5ff0bc43" (UID: "ac68b7dd-54d2-4691-ac90-4f6f5ff0bc43"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:58:30 crc kubenswrapper[4907]: I0313 15:58:30.406080 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ac68b7dd-54d2-4691-ac90-4f6f5ff0bc43-kube-api-access-rrgtc" (OuterVolumeSpecName: "kube-api-access-rrgtc") pod "ac68b7dd-54d2-4691-ac90-4f6f5ff0bc43" (UID: "ac68b7dd-54d2-4691-ac90-4f6f5ff0bc43"). InnerVolumeSpecName "kube-api-access-rrgtc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:58:30 crc kubenswrapper[4907]: I0313 15:58:30.469679 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ac68b7dd-54d2-4691-ac90-4f6f5ff0bc43-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ac68b7dd-54d2-4691-ac90-4f6f5ff0bc43" (UID: "ac68b7dd-54d2-4691-ac90-4f6f5ff0bc43"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:58:30 crc kubenswrapper[4907]: I0313 15:58:30.503495 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac68b7dd-54d2-4691-ac90-4f6f5ff0bc43-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 15:58:30 crc kubenswrapper[4907]: I0313 15:58:30.503528 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rrgtc\" (UniqueName: \"kubernetes.io/projected/ac68b7dd-54d2-4691-ac90-4f6f5ff0bc43-kube-api-access-rrgtc\") on node \"crc\" DevicePath \"\"" Mar 13 15:58:30 crc kubenswrapper[4907]: I0313 15:58:30.503539 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac68b7dd-54d2-4691-ac90-4f6f5ff0bc43-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 15:58:30 crc kubenswrapper[4907]: I0313 15:58:30.751218 4907 generic.go:334] "Generic (PLEG): container finished" podID="ac68b7dd-54d2-4691-ac90-4f6f5ff0bc43" containerID="ffa10747043fef78a41d5b63379aa594a58eda0c580b0fdb3ed18e4cc5b00c07" exitCode=0 Mar 13 15:58:30 crc kubenswrapper[4907]: I0313 15:58:30.751259 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pk7rd" event={"ID":"ac68b7dd-54d2-4691-ac90-4f6f5ff0bc43","Type":"ContainerDied","Data":"ffa10747043fef78a41d5b63379aa594a58eda0c580b0fdb3ed18e4cc5b00c07"} Mar 13 15:58:30 crc kubenswrapper[4907]: I0313 15:58:30.751285 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pk7rd" event={"ID":"ac68b7dd-54d2-4691-ac90-4f6f5ff0bc43","Type":"ContainerDied","Data":"31bb1f1b4403f006239e8d040d3abbb5c7a93523d9ec4c3f298d90363dac8352"} Mar 13 15:58:30 crc kubenswrapper[4907]: I0313 15:58:30.751303 4907 scope.go:117] "RemoveContainer" containerID="ffa10747043fef78a41d5b63379aa594a58eda0c580b0fdb3ed18e4cc5b00c07" Mar 13 15:58:30 crc kubenswrapper[4907]: I0313 15:58:30.751352 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pk7rd" Mar 13 15:58:30 crc kubenswrapper[4907]: I0313 15:58:30.793735 4907 scope.go:117] "RemoveContainer" containerID="db5c85295b78a586f9a9d79869cb88866e67ead5e0fb9606d34998d9455a4702" Mar 13 15:58:30 crc kubenswrapper[4907]: I0313 15:58:30.793891 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-pk7rd"] Mar 13 15:58:30 crc kubenswrapper[4907]: I0313 15:58:30.816553 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-pk7rd"] Mar 13 15:58:30 crc kubenswrapper[4907]: I0313 15:58:30.830345 4907 scope.go:117] "RemoveContainer" containerID="53b83ec7d94b2543162fe15e2a0e9e0d2a45294fa4e4b0c3681b0c95abfa7080" Mar 13 15:58:30 crc kubenswrapper[4907]: I0313 15:58:30.905192 4907 scope.go:117] "RemoveContainer" containerID="ffa10747043fef78a41d5b63379aa594a58eda0c580b0fdb3ed18e4cc5b00c07" Mar 13 15:58:30 crc kubenswrapper[4907]: E0313 15:58:30.906130 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ffa10747043fef78a41d5b63379aa594a58eda0c580b0fdb3ed18e4cc5b00c07\": container with ID starting with ffa10747043fef78a41d5b63379aa594a58eda0c580b0fdb3ed18e4cc5b00c07 not found: ID does not exist" containerID="ffa10747043fef78a41d5b63379aa594a58eda0c580b0fdb3ed18e4cc5b00c07" Mar 13 15:58:30 crc kubenswrapper[4907]: I0313 15:58:30.906176 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ffa10747043fef78a41d5b63379aa594a58eda0c580b0fdb3ed18e4cc5b00c07"} err="failed to get container status \"ffa10747043fef78a41d5b63379aa594a58eda0c580b0fdb3ed18e4cc5b00c07\": rpc error: code = NotFound desc = could not find container \"ffa10747043fef78a41d5b63379aa594a58eda0c580b0fdb3ed18e4cc5b00c07\": container with ID starting with ffa10747043fef78a41d5b63379aa594a58eda0c580b0fdb3ed18e4cc5b00c07 not found: ID does not exist" Mar 13 15:58:30 crc kubenswrapper[4907]: I0313 15:58:30.906201 4907 scope.go:117] "RemoveContainer" containerID="db5c85295b78a586f9a9d79869cb88866e67ead5e0fb9606d34998d9455a4702" Mar 13 15:58:30 crc kubenswrapper[4907]: E0313 15:58:30.906693 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"db5c85295b78a586f9a9d79869cb88866e67ead5e0fb9606d34998d9455a4702\": container with ID starting with db5c85295b78a586f9a9d79869cb88866e67ead5e0fb9606d34998d9455a4702 not found: ID does not exist" containerID="db5c85295b78a586f9a9d79869cb88866e67ead5e0fb9606d34998d9455a4702" Mar 13 15:58:30 crc kubenswrapper[4907]: I0313 15:58:30.906730 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db5c85295b78a586f9a9d79869cb88866e67ead5e0fb9606d34998d9455a4702"} err="failed to get container status \"db5c85295b78a586f9a9d79869cb88866e67ead5e0fb9606d34998d9455a4702\": rpc error: code = NotFound desc = could not find container \"db5c85295b78a586f9a9d79869cb88866e67ead5e0fb9606d34998d9455a4702\": container with ID starting with db5c85295b78a586f9a9d79869cb88866e67ead5e0fb9606d34998d9455a4702 not found: ID does not exist" Mar 13 15:58:30 crc kubenswrapper[4907]: I0313 15:58:30.906751 4907 scope.go:117] "RemoveContainer" containerID="53b83ec7d94b2543162fe15e2a0e9e0d2a45294fa4e4b0c3681b0c95abfa7080" Mar 13 15:58:30 crc kubenswrapper[4907]: E0313 15:58:30.907191 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"53b83ec7d94b2543162fe15e2a0e9e0d2a45294fa4e4b0c3681b0c95abfa7080\": container with ID starting with 53b83ec7d94b2543162fe15e2a0e9e0d2a45294fa4e4b0c3681b0c95abfa7080 not found: ID does not exist" containerID="53b83ec7d94b2543162fe15e2a0e9e0d2a45294fa4e4b0c3681b0c95abfa7080" Mar 13 15:58:30 crc kubenswrapper[4907]: I0313 15:58:30.907224 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"53b83ec7d94b2543162fe15e2a0e9e0d2a45294fa4e4b0c3681b0c95abfa7080"} err="failed to get container status \"53b83ec7d94b2543162fe15e2a0e9e0d2a45294fa4e4b0c3681b0c95abfa7080\": rpc error: code = NotFound desc = could not find container \"53b83ec7d94b2543162fe15e2a0e9e0d2a45294fa4e4b0c3681b0c95abfa7080\": container with ID starting with 53b83ec7d94b2543162fe15e2a0e9e0d2a45294fa4e4b0c3681b0c95abfa7080 not found: ID does not exist" Mar 13 15:58:31 crc kubenswrapper[4907]: I0313 15:58:31.804384 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ac68b7dd-54d2-4691-ac90-4f6f5ff0bc43" path="/var/lib/kubelet/pods/ac68b7dd-54d2-4691-ac90-4f6f5ff0bc43/volumes" Mar 13 15:58:38 crc kubenswrapper[4907]: I0313 15:58:38.304689 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-kvhlf"] Mar 13 15:58:38 crc kubenswrapper[4907]: E0313 15:58:38.305538 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac68b7dd-54d2-4691-ac90-4f6f5ff0bc43" containerName="extract-content" Mar 13 15:58:38 crc kubenswrapper[4907]: I0313 15:58:38.305555 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac68b7dd-54d2-4691-ac90-4f6f5ff0bc43" containerName="extract-content" Mar 13 15:58:38 crc kubenswrapper[4907]: E0313 15:58:38.305599 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac68b7dd-54d2-4691-ac90-4f6f5ff0bc43" containerName="registry-server" Mar 13 15:58:38 crc kubenswrapper[4907]: I0313 15:58:38.305608 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac68b7dd-54d2-4691-ac90-4f6f5ff0bc43" containerName="registry-server" Mar 13 15:58:38 crc kubenswrapper[4907]: E0313 15:58:38.305625 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac68b7dd-54d2-4691-ac90-4f6f5ff0bc43" containerName="extract-utilities" Mar 13 15:58:38 crc kubenswrapper[4907]: I0313 15:58:38.305631 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac68b7dd-54d2-4691-ac90-4f6f5ff0bc43" containerName="extract-utilities" Mar 13 15:58:38 crc kubenswrapper[4907]: I0313 15:58:38.305898 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac68b7dd-54d2-4691-ac90-4f6f5ff0bc43" containerName="registry-server" Mar 13 15:58:38 crc kubenswrapper[4907]: I0313 15:58:38.307797 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kvhlf" Mar 13 15:58:38 crc kubenswrapper[4907]: I0313 15:58:38.329257 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kvhlf"] Mar 13 15:58:38 crc kubenswrapper[4907]: I0313 15:58:38.416173 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l5p4j\" (UniqueName: \"kubernetes.io/projected/9739d592-69ae-4c44-ba8a-f4a66d3dd37e-kube-api-access-l5p4j\") pod \"redhat-operators-kvhlf\" (UID: \"9739d592-69ae-4c44-ba8a-f4a66d3dd37e\") " pod="openshift-marketplace/redhat-operators-kvhlf" Mar 13 15:58:38 crc kubenswrapper[4907]: I0313 15:58:38.416222 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9739d592-69ae-4c44-ba8a-f4a66d3dd37e-catalog-content\") pod \"redhat-operators-kvhlf\" (UID: \"9739d592-69ae-4c44-ba8a-f4a66d3dd37e\") " pod="openshift-marketplace/redhat-operators-kvhlf" Mar 13 15:58:38 crc kubenswrapper[4907]: I0313 15:58:38.416340 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9739d592-69ae-4c44-ba8a-f4a66d3dd37e-utilities\") pod \"redhat-operators-kvhlf\" (UID: \"9739d592-69ae-4c44-ba8a-f4a66d3dd37e\") " pod="openshift-marketplace/redhat-operators-kvhlf" Mar 13 15:58:38 crc kubenswrapper[4907]: I0313 15:58:38.518697 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l5p4j\" (UniqueName: \"kubernetes.io/projected/9739d592-69ae-4c44-ba8a-f4a66d3dd37e-kube-api-access-l5p4j\") pod \"redhat-operators-kvhlf\" (UID: \"9739d592-69ae-4c44-ba8a-f4a66d3dd37e\") " pod="openshift-marketplace/redhat-operators-kvhlf" Mar 13 15:58:38 crc kubenswrapper[4907]: I0313 15:58:38.518750 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9739d592-69ae-4c44-ba8a-f4a66d3dd37e-catalog-content\") pod \"redhat-operators-kvhlf\" (UID: \"9739d592-69ae-4c44-ba8a-f4a66d3dd37e\") " pod="openshift-marketplace/redhat-operators-kvhlf" Mar 13 15:58:38 crc kubenswrapper[4907]: I0313 15:58:38.518947 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9739d592-69ae-4c44-ba8a-f4a66d3dd37e-utilities\") pod \"redhat-operators-kvhlf\" (UID: \"9739d592-69ae-4c44-ba8a-f4a66d3dd37e\") " pod="openshift-marketplace/redhat-operators-kvhlf" Mar 13 15:58:38 crc kubenswrapper[4907]: I0313 15:58:38.519474 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9739d592-69ae-4c44-ba8a-f4a66d3dd37e-catalog-content\") pod \"redhat-operators-kvhlf\" (UID: \"9739d592-69ae-4c44-ba8a-f4a66d3dd37e\") " pod="openshift-marketplace/redhat-operators-kvhlf" Mar 13 15:58:38 crc kubenswrapper[4907]: I0313 15:58:38.519606 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9739d592-69ae-4c44-ba8a-f4a66d3dd37e-utilities\") pod \"redhat-operators-kvhlf\" (UID: \"9739d592-69ae-4c44-ba8a-f4a66d3dd37e\") " pod="openshift-marketplace/redhat-operators-kvhlf" Mar 13 15:58:38 crc kubenswrapper[4907]: I0313 15:58:38.545802 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l5p4j\" (UniqueName: \"kubernetes.io/projected/9739d592-69ae-4c44-ba8a-f4a66d3dd37e-kube-api-access-l5p4j\") pod \"redhat-operators-kvhlf\" (UID: \"9739d592-69ae-4c44-ba8a-f4a66d3dd37e\") " pod="openshift-marketplace/redhat-operators-kvhlf" Mar 13 15:58:38 crc kubenswrapper[4907]: I0313 15:58:38.648704 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kvhlf" Mar 13 15:58:39 crc kubenswrapper[4907]: I0313 15:58:39.196775 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kvhlf"] Mar 13 15:58:39 crc kubenswrapper[4907]: I0313 15:58:39.853275 4907 generic.go:334] "Generic (PLEG): container finished" podID="9739d592-69ae-4c44-ba8a-f4a66d3dd37e" containerID="fe843f7ce1ebfd01e87874065c8fa24f6b69455e96ff16a1e7272446b070a978" exitCode=0 Mar 13 15:58:39 crc kubenswrapper[4907]: I0313 15:58:39.853530 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kvhlf" event={"ID":"9739d592-69ae-4c44-ba8a-f4a66d3dd37e","Type":"ContainerDied","Data":"fe843f7ce1ebfd01e87874065c8fa24f6b69455e96ff16a1e7272446b070a978"} Mar 13 15:58:39 crc kubenswrapper[4907]: I0313 15:58:39.853555 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kvhlf" event={"ID":"9739d592-69ae-4c44-ba8a-f4a66d3dd37e","Type":"ContainerStarted","Data":"a8505ebf077b55d207686198c5233286e1d24c14f8a53397bbeaddc40331ff03"} Mar 13 15:58:40 crc kubenswrapper[4907]: I0313 15:58:40.091012 4907 scope.go:117] "RemoveContainer" containerID="ae08e1f5613f0f2169631235f70419712208d7aa173e1b4d3da70be8496080b9" Mar 13 15:58:40 crc kubenswrapper[4907]: I0313 15:58:40.866025 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kvhlf" event={"ID":"9739d592-69ae-4c44-ba8a-f4a66d3dd37e","Type":"ContainerStarted","Data":"72383279f5aa6921bbcd47b4d1969465c8d8b98982eda484eee6233bb1706a35"} Mar 13 15:58:45 crc kubenswrapper[4907]: I0313 15:58:45.936691 4907 generic.go:334] "Generic (PLEG): container finished" podID="9739d592-69ae-4c44-ba8a-f4a66d3dd37e" containerID="72383279f5aa6921bbcd47b4d1969465c8d8b98982eda484eee6233bb1706a35" exitCode=0 Mar 13 15:58:45 crc kubenswrapper[4907]: I0313 15:58:45.936743 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kvhlf" event={"ID":"9739d592-69ae-4c44-ba8a-f4a66d3dd37e","Type":"ContainerDied","Data":"72383279f5aa6921bbcd47b4d1969465c8d8b98982eda484eee6233bb1706a35"} Mar 13 15:58:46 crc kubenswrapper[4907]: I0313 15:58:46.951062 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kvhlf" event={"ID":"9739d592-69ae-4c44-ba8a-f4a66d3dd37e","Type":"ContainerStarted","Data":"bf318b47b118afa14ad5c1f8dd0ece551b6717fd4d445868fc8150982fe2ef70"} Mar 13 15:58:46 crc kubenswrapper[4907]: I0313 15:58:46.986928 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-kvhlf" podStartSLOduration=2.479931635 podStartE2EDuration="8.986900599s" podCreationTimestamp="2026-03-13 15:58:38 +0000 UTC" firstStartedPulling="2026-03-13 15:58:39.855469184 +0000 UTC m=+6818.755256873" lastFinishedPulling="2026-03-13 15:58:46.362438138 +0000 UTC m=+6825.262225837" observedRunningTime="2026-03-13 15:58:46.972177485 +0000 UTC m=+6825.871965184" watchObservedRunningTime="2026-03-13 15:58:46.986900599 +0000 UTC m=+6825.886688318" Mar 13 15:58:48 crc kubenswrapper[4907]: I0313 15:58:48.649051 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-kvhlf" Mar 13 15:58:48 crc kubenswrapper[4907]: I0313 15:58:48.649374 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-kvhlf" Mar 13 15:58:49 crc kubenswrapper[4907]: I0313 15:58:49.702084 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-kvhlf" podUID="9739d592-69ae-4c44-ba8a-f4a66d3dd37e" containerName="registry-server" probeResult="failure" output=< Mar 13 15:58:49 crc kubenswrapper[4907]: timeout: failed to connect service ":50051" within 1s Mar 13 15:58:49 crc kubenswrapper[4907]: > Mar 13 15:58:58 crc kubenswrapper[4907]: I0313 15:58:58.703947 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-kvhlf" Mar 13 15:58:58 crc kubenswrapper[4907]: I0313 15:58:58.757623 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-kvhlf" Mar 13 15:58:58 crc kubenswrapper[4907]: I0313 15:58:58.944054 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-kvhlf"] Mar 13 15:59:00 crc kubenswrapper[4907]: I0313 15:59:00.103029 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-kvhlf" podUID="9739d592-69ae-4c44-ba8a-f4a66d3dd37e" containerName="registry-server" containerID="cri-o://bf318b47b118afa14ad5c1f8dd0ece551b6717fd4d445868fc8150982fe2ef70" gracePeriod=2 Mar 13 15:59:00 crc kubenswrapper[4907]: I0313 15:59:00.713529 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kvhlf" Mar 13 15:59:00 crc kubenswrapper[4907]: I0313 15:59:00.873221 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9739d592-69ae-4c44-ba8a-f4a66d3dd37e-catalog-content\") pod \"9739d592-69ae-4c44-ba8a-f4a66d3dd37e\" (UID: \"9739d592-69ae-4c44-ba8a-f4a66d3dd37e\") " Mar 13 15:59:00 crc kubenswrapper[4907]: I0313 15:59:00.873519 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9739d592-69ae-4c44-ba8a-f4a66d3dd37e-utilities\") pod \"9739d592-69ae-4c44-ba8a-f4a66d3dd37e\" (UID: \"9739d592-69ae-4c44-ba8a-f4a66d3dd37e\") " Mar 13 15:59:00 crc kubenswrapper[4907]: I0313 15:59:00.873610 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l5p4j\" (UniqueName: \"kubernetes.io/projected/9739d592-69ae-4c44-ba8a-f4a66d3dd37e-kube-api-access-l5p4j\") pod \"9739d592-69ae-4c44-ba8a-f4a66d3dd37e\" (UID: \"9739d592-69ae-4c44-ba8a-f4a66d3dd37e\") " Mar 13 15:59:00 crc kubenswrapper[4907]: I0313 15:59:00.874311 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9739d592-69ae-4c44-ba8a-f4a66d3dd37e-utilities" (OuterVolumeSpecName: "utilities") pod "9739d592-69ae-4c44-ba8a-f4a66d3dd37e" (UID: "9739d592-69ae-4c44-ba8a-f4a66d3dd37e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:59:00 crc kubenswrapper[4907]: I0313 15:59:00.875542 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9739d592-69ae-4c44-ba8a-f4a66d3dd37e-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 15:59:00 crc kubenswrapper[4907]: I0313 15:59:00.880831 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9739d592-69ae-4c44-ba8a-f4a66d3dd37e-kube-api-access-l5p4j" (OuterVolumeSpecName: "kube-api-access-l5p4j") pod "9739d592-69ae-4c44-ba8a-f4a66d3dd37e" (UID: "9739d592-69ae-4c44-ba8a-f4a66d3dd37e"). InnerVolumeSpecName "kube-api-access-l5p4j". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:59:00 crc kubenswrapper[4907]: I0313 15:59:00.977774 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l5p4j\" (UniqueName: \"kubernetes.io/projected/9739d592-69ae-4c44-ba8a-f4a66d3dd37e-kube-api-access-l5p4j\") on node \"crc\" DevicePath \"\"" Mar 13 15:59:01 crc kubenswrapper[4907]: I0313 15:59:01.026358 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9739d592-69ae-4c44-ba8a-f4a66d3dd37e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9739d592-69ae-4c44-ba8a-f4a66d3dd37e" (UID: "9739d592-69ae-4c44-ba8a-f4a66d3dd37e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:59:01 crc kubenswrapper[4907]: I0313 15:59:01.079946 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9739d592-69ae-4c44-ba8a-f4a66d3dd37e-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 15:59:01 crc kubenswrapper[4907]: I0313 15:59:01.113849 4907 generic.go:334] "Generic (PLEG): container finished" podID="9739d592-69ae-4c44-ba8a-f4a66d3dd37e" containerID="bf318b47b118afa14ad5c1f8dd0ece551b6717fd4d445868fc8150982fe2ef70" exitCode=0 Mar 13 15:59:01 crc kubenswrapper[4907]: I0313 15:59:01.113919 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kvhlf" event={"ID":"9739d592-69ae-4c44-ba8a-f4a66d3dd37e","Type":"ContainerDied","Data":"bf318b47b118afa14ad5c1f8dd0ece551b6717fd4d445868fc8150982fe2ef70"} Mar 13 15:59:01 crc kubenswrapper[4907]: I0313 15:59:01.113951 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kvhlf" event={"ID":"9739d592-69ae-4c44-ba8a-f4a66d3dd37e","Type":"ContainerDied","Data":"a8505ebf077b55d207686198c5233286e1d24c14f8a53397bbeaddc40331ff03"} Mar 13 15:59:01 crc kubenswrapper[4907]: I0313 15:59:01.113973 4907 scope.go:117] "RemoveContainer" containerID="bf318b47b118afa14ad5c1f8dd0ece551b6717fd4d445868fc8150982fe2ef70" Mar 13 15:59:01 crc kubenswrapper[4907]: I0313 15:59:01.115454 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kvhlf" Mar 13 15:59:01 crc kubenswrapper[4907]: I0313 15:59:01.154011 4907 scope.go:117] "RemoveContainer" containerID="72383279f5aa6921bbcd47b4d1969465c8d8b98982eda484eee6233bb1706a35" Mar 13 15:59:01 crc kubenswrapper[4907]: I0313 15:59:01.162505 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-kvhlf"] Mar 13 15:59:01 crc kubenswrapper[4907]: I0313 15:59:01.178869 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-kvhlf"] Mar 13 15:59:01 crc kubenswrapper[4907]: I0313 15:59:01.187209 4907 scope.go:117] "RemoveContainer" containerID="fe843f7ce1ebfd01e87874065c8fa24f6b69455e96ff16a1e7272446b070a978" Mar 13 15:59:01 crc kubenswrapper[4907]: I0313 15:59:01.245184 4907 scope.go:117] "RemoveContainer" containerID="bf318b47b118afa14ad5c1f8dd0ece551b6717fd4d445868fc8150982fe2ef70" Mar 13 15:59:01 crc kubenswrapper[4907]: E0313 15:59:01.246053 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bf318b47b118afa14ad5c1f8dd0ece551b6717fd4d445868fc8150982fe2ef70\": container with ID starting with bf318b47b118afa14ad5c1f8dd0ece551b6717fd4d445868fc8150982fe2ef70 not found: ID does not exist" containerID="bf318b47b118afa14ad5c1f8dd0ece551b6717fd4d445868fc8150982fe2ef70" Mar 13 15:59:01 crc kubenswrapper[4907]: I0313 15:59:01.246105 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bf318b47b118afa14ad5c1f8dd0ece551b6717fd4d445868fc8150982fe2ef70"} err="failed to get container status \"bf318b47b118afa14ad5c1f8dd0ece551b6717fd4d445868fc8150982fe2ef70\": rpc error: code = NotFound desc = could not find container \"bf318b47b118afa14ad5c1f8dd0ece551b6717fd4d445868fc8150982fe2ef70\": container with ID starting with bf318b47b118afa14ad5c1f8dd0ece551b6717fd4d445868fc8150982fe2ef70 not found: ID does not exist" Mar 13 15:59:01 crc kubenswrapper[4907]: I0313 15:59:01.246140 4907 scope.go:117] "RemoveContainer" containerID="72383279f5aa6921bbcd47b4d1969465c8d8b98982eda484eee6233bb1706a35" Mar 13 15:59:01 crc kubenswrapper[4907]: E0313 15:59:01.246934 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"72383279f5aa6921bbcd47b4d1969465c8d8b98982eda484eee6233bb1706a35\": container with ID starting with 72383279f5aa6921bbcd47b4d1969465c8d8b98982eda484eee6233bb1706a35 not found: ID does not exist" containerID="72383279f5aa6921bbcd47b4d1969465c8d8b98982eda484eee6233bb1706a35" Mar 13 15:59:01 crc kubenswrapper[4907]: I0313 15:59:01.247004 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"72383279f5aa6921bbcd47b4d1969465c8d8b98982eda484eee6233bb1706a35"} err="failed to get container status \"72383279f5aa6921bbcd47b4d1969465c8d8b98982eda484eee6233bb1706a35\": rpc error: code = NotFound desc = could not find container \"72383279f5aa6921bbcd47b4d1969465c8d8b98982eda484eee6233bb1706a35\": container with ID starting with 72383279f5aa6921bbcd47b4d1969465c8d8b98982eda484eee6233bb1706a35 not found: ID does not exist" Mar 13 15:59:01 crc kubenswrapper[4907]: I0313 15:59:01.247046 4907 scope.go:117] "RemoveContainer" containerID="fe843f7ce1ebfd01e87874065c8fa24f6b69455e96ff16a1e7272446b070a978" Mar 13 15:59:01 crc kubenswrapper[4907]: E0313 15:59:01.247709 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fe843f7ce1ebfd01e87874065c8fa24f6b69455e96ff16a1e7272446b070a978\": container with ID starting with fe843f7ce1ebfd01e87874065c8fa24f6b69455e96ff16a1e7272446b070a978 not found: ID does not exist" containerID="fe843f7ce1ebfd01e87874065c8fa24f6b69455e96ff16a1e7272446b070a978" Mar 13 15:59:01 crc kubenswrapper[4907]: I0313 15:59:01.247755 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fe843f7ce1ebfd01e87874065c8fa24f6b69455e96ff16a1e7272446b070a978"} err="failed to get container status \"fe843f7ce1ebfd01e87874065c8fa24f6b69455e96ff16a1e7272446b070a978\": rpc error: code = NotFound desc = could not find container \"fe843f7ce1ebfd01e87874065c8fa24f6b69455e96ff16a1e7272446b070a978\": container with ID starting with fe843f7ce1ebfd01e87874065c8fa24f6b69455e96ff16a1e7272446b070a978 not found: ID does not exist" Mar 13 15:59:01 crc kubenswrapper[4907]: I0313 15:59:01.797841 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9739d592-69ae-4c44-ba8a-f4a66d3dd37e" path="/var/lib/kubelet/pods/9739d592-69ae-4c44-ba8a-f4a66d3dd37e/volumes" Mar 13 15:59:37 crc kubenswrapper[4907]: I0313 15:59:37.037832 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-db-create-6qbgl"] Mar 13 15:59:37 crc kubenswrapper[4907]: I0313 15:59:37.047144 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-db-create-6qbgl"] Mar 13 15:59:37 crc kubenswrapper[4907]: I0313 15:59:37.799713 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="689fb651-0113-48ba-b6e4-29c41c5fa21f" path="/var/lib/kubelet/pods/689fb651-0113-48ba-b6e4-29c41c5fa21f/volumes" Mar 13 15:59:38 crc kubenswrapper[4907]: I0313 15:59:38.044165 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-cc97-account-create-update-fc5r7"] Mar 13 15:59:38 crc kubenswrapper[4907]: I0313 15:59:38.056497 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-cc97-account-create-update-fc5r7"] Mar 13 15:59:39 crc kubenswrapper[4907]: I0313 15:59:39.810569 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="442daa10-2973-4304-9e1c-039d35226686" path="/var/lib/kubelet/pods/442daa10-2973-4304-9e1c-039d35226686/volumes" Mar 13 15:59:40 crc kubenswrapper[4907]: I0313 15:59:40.177188 4907 scope.go:117] "RemoveContainer" containerID="d38dedf8f9062f909f22038364138fce749f1c07518101c6bfdc3e1d5f365369" Mar 13 15:59:40 crc kubenswrapper[4907]: I0313 15:59:40.218277 4907 scope.go:117] "RemoveContainer" containerID="7eeac1862f5788609b87b66dfacb73f343af69811adf4dfc0c22454194586bf3" Mar 13 15:59:50 crc kubenswrapper[4907]: I0313 15:59:50.050184 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-db-sync-nwwc8"] Mar 13 15:59:50 crc kubenswrapper[4907]: I0313 15:59:50.066374 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-db-sync-nwwc8"] Mar 13 15:59:51 crc kubenswrapper[4907]: I0313 15:59:51.807741 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c5a77890-eb92-4d55-acf8-481f70c9277c" path="/var/lib/kubelet/pods/c5a77890-eb92-4d55-acf8-481f70c9277c/volumes" Mar 13 16:00:00 crc kubenswrapper[4907]: I0313 16:00:00.153034 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556960-8xc84"] Mar 13 16:00:00 crc kubenswrapper[4907]: E0313 16:00:00.154055 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9739d592-69ae-4c44-ba8a-f4a66d3dd37e" containerName="registry-server" Mar 13 16:00:00 crc kubenswrapper[4907]: I0313 16:00:00.154068 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="9739d592-69ae-4c44-ba8a-f4a66d3dd37e" containerName="registry-server" Mar 13 16:00:00 crc kubenswrapper[4907]: E0313 16:00:00.154088 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9739d592-69ae-4c44-ba8a-f4a66d3dd37e" containerName="extract-content" Mar 13 16:00:00 crc kubenswrapper[4907]: I0313 16:00:00.154094 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="9739d592-69ae-4c44-ba8a-f4a66d3dd37e" containerName="extract-content" Mar 13 16:00:00 crc kubenswrapper[4907]: E0313 16:00:00.154114 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9739d592-69ae-4c44-ba8a-f4a66d3dd37e" containerName="extract-utilities" Mar 13 16:00:00 crc kubenswrapper[4907]: I0313 16:00:00.154121 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="9739d592-69ae-4c44-ba8a-f4a66d3dd37e" containerName="extract-utilities" Mar 13 16:00:00 crc kubenswrapper[4907]: I0313 16:00:00.154325 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="9739d592-69ae-4c44-ba8a-f4a66d3dd37e" containerName="registry-server" Mar 13 16:00:00 crc kubenswrapper[4907]: I0313 16:00:00.155108 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556960-8xc84" Mar 13 16:00:00 crc kubenswrapper[4907]: I0313 16:00:00.157625 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556960-8xc84"] Mar 13 16:00:00 crc kubenswrapper[4907]: I0313 16:00:00.159099 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 16:00:00 crc kubenswrapper[4907]: I0313 16:00:00.159210 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 16:00:00 crc kubenswrapper[4907]: I0313 16:00:00.163028 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 16:00:00 crc kubenswrapper[4907]: I0313 16:00:00.230157 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xgvkq\" (UniqueName: \"kubernetes.io/projected/acb3140e-a115-443a-a39e-5f180a196b13-kube-api-access-xgvkq\") pod \"auto-csr-approver-29556960-8xc84\" (UID: \"acb3140e-a115-443a-a39e-5f180a196b13\") " pod="openshift-infra/auto-csr-approver-29556960-8xc84" Mar 13 16:00:00 crc kubenswrapper[4907]: I0313 16:00:00.250463 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556960-mkbxd"] Mar 13 16:00:00 crc kubenswrapper[4907]: I0313 16:00:00.252709 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556960-mkbxd" Mar 13 16:00:00 crc kubenswrapper[4907]: I0313 16:00:00.255020 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 13 16:00:00 crc kubenswrapper[4907]: I0313 16:00:00.255368 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 13 16:00:00 crc kubenswrapper[4907]: I0313 16:00:00.265779 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556960-mkbxd"] Mar 13 16:00:00 crc kubenswrapper[4907]: I0313 16:00:00.331959 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mpbb6\" (UniqueName: \"kubernetes.io/projected/e1001d4c-6984-4a3c-8d1f-167a6a48ea8a-kube-api-access-mpbb6\") pod \"collect-profiles-29556960-mkbxd\" (UID: \"e1001d4c-6984-4a3c-8d1f-167a6a48ea8a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556960-mkbxd" Mar 13 16:00:00 crc kubenswrapper[4907]: I0313 16:00:00.332102 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e1001d4c-6984-4a3c-8d1f-167a6a48ea8a-secret-volume\") pod \"collect-profiles-29556960-mkbxd\" (UID: \"e1001d4c-6984-4a3c-8d1f-167a6a48ea8a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556960-mkbxd" Mar 13 16:00:00 crc kubenswrapper[4907]: I0313 16:00:00.332139 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e1001d4c-6984-4a3c-8d1f-167a6a48ea8a-config-volume\") pod \"collect-profiles-29556960-mkbxd\" (UID: \"e1001d4c-6984-4a3c-8d1f-167a6a48ea8a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556960-mkbxd" Mar 13 16:00:00 crc kubenswrapper[4907]: I0313 16:00:00.332196 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xgvkq\" (UniqueName: \"kubernetes.io/projected/acb3140e-a115-443a-a39e-5f180a196b13-kube-api-access-xgvkq\") pod \"auto-csr-approver-29556960-8xc84\" (UID: \"acb3140e-a115-443a-a39e-5f180a196b13\") " pod="openshift-infra/auto-csr-approver-29556960-8xc84" Mar 13 16:00:00 crc kubenswrapper[4907]: I0313 16:00:00.350704 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xgvkq\" (UniqueName: \"kubernetes.io/projected/acb3140e-a115-443a-a39e-5f180a196b13-kube-api-access-xgvkq\") pod \"auto-csr-approver-29556960-8xc84\" (UID: \"acb3140e-a115-443a-a39e-5f180a196b13\") " pod="openshift-infra/auto-csr-approver-29556960-8xc84" Mar 13 16:00:00 crc kubenswrapper[4907]: I0313 16:00:00.433468 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mpbb6\" (UniqueName: \"kubernetes.io/projected/e1001d4c-6984-4a3c-8d1f-167a6a48ea8a-kube-api-access-mpbb6\") pod \"collect-profiles-29556960-mkbxd\" (UID: \"e1001d4c-6984-4a3c-8d1f-167a6a48ea8a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556960-mkbxd" Mar 13 16:00:00 crc kubenswrapper[4907]: I0313 16:00:00.433581 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e1001d4c-6984-4a3c-8d1f-167a6a48ea8a-secret-volume\") pod \"collect-profiles-29556960-mkbxd\" (UID: \"e1001d4c-6984-4a3c-8d1f-167a6a48ea8a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556960-mkbxd" Mar 13 16:00:00 crc kubenswrapper[4907]: I0313 16:00:00.434375 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e1001d4c-6984-4a3c-8d1f-167a6a48ea8a-config-volume\") pod \"collect-profiles-29556960-mkbxd\" (UID: \"e1001d4c-6984-4a3c-8d1f-167a6a48ea8a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556960-mkbxd" Mar 13 16:00:00 crc kubenswrapper[4907]: I0313 16:00:00.435457 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e1001d4c-6984-4a3c-8d1f-167a6a48ea8a-config-volume\") pod \"collect-profiles-29556960-mkbxd\" (UID: \"e1001d4c-6984-4a3c-8d1f-167a6a48ea8a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556960-mkbxd" Mar 13 16:00:00 crc kubenswrapper[4907]: I0313 16:00:00.436869 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e1001d4c-6984-4a3c-8d1f-167a6a48ea8a-secret-volume\") pod \"collect-profiles-29556960-mkbxd\" (UID: \"e1001d4c-6984-4a3c-8d1f-167a6a48ea8a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556960-mkbxd" Mar 13 16:00:00 crc kubenswrapper[4907]: I0313 16:00:00.448209 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mpbb6\" (UniqueName: \"kubernetes.io/projected/e1001d4c-6984-4a3c-8d1f-167a6a48ea8a-kube-api-access-mpbb6\") pod \"collect-profiles-29556960-mkbxd\" (UID: \"e1001d4c-6984-4a3c-8d1f-167a6a48ea8a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556960-mkbxd" Mar 13 16:00:00 crc kubenswrapper[4907]: I0313 16:00:00.480090 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556960-8xc84" Mar 13 16:00:00 crc kubenswrapper[4907]: I0313 16:00:00.572750 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556960-mkbxd" Mar 13 16:00:00 crc kubenswrapper[4907]: I0313 16:00:00.945657 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556960-8xc84"] Mar 13 16:00:00 crc kubenswrapper[4907]: I0313 16:00:00.948964 4907 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 13 16:00:01 crc kubenswrapper[4907]: W0313 16:00:01.060481 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode1001d4c_6984_4a3c_8d1f_167a6a48ea8a.slice/crio-961c155ac80d225a208f0f959fac939f4afb3a793830f43150c57ec2c6d02d2a WatchSource:0}: Error finding container 961c155ac80d225a208f0f959fac939f4afb3a793830f43150c57ec2c6d02d2a: Status 404 returned error can't find the container with id 961c155ac80d225a208f0f959fac939f4afb3a793830f43150c57ec2c6d02d2a Mar 13 16:00:01 crc kubenswrapper[4907]: I0313 16:00:01.068232 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556960-mkbxd"] Mar 13 16:00:01 crc kubenswrapper[4907]: I0313 16:00:01.796562 4907 generic.go:334] "Generic (PLEG): container finished" podID="e1001d4c-6984-4a3c-8d1f-167a6a48ea8a" containerID="0c0e8b2fe415ff36180c0f0cbc79e145510c4e8b607c452be50ff3ed2394d856" exitCode=0 Mar 13 16:00:01 crc kubenswrapper[4907]: I0313 16:00:01.796641 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556960-mkbxd" event={"ID":"e1001d4c-6984-4a3c-8d1f-167a6a48ea8a","Type":"ContainerDied","Data":"0c0e8b2fe415ff36180c0f0cbc79e145510c4e8b607c452be50ff3ed2394d856"} Mar 13 16:00:01 crc kubenswrapper[4907]: I0313 16:00:01.796864 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556960-mkbxd" event={"ID":"e1001d4c-6984-4a3c-8d1f-167a6a48ea8a","Type":"ContainerStarted","Data":"961c155ac80d225a208f0f959fac939f4afb3a793830f43150c57ec2c6d02d2a"} Mar 13 16:00:01 crc kubenswrapper[4907]: I0313 16:00:01.797972 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556960-8xc84" event={"ID":"acb3140e-a115-443a-a39e-5f180a196b13","Type":"ContainerStarted","Data":"4a57c716b0e24bb904434f950bb9538fc19745e51f498ae299351e13dbb5f5e4"} Mar 13 16:00:03 crc kubenswrapper[4907]: I0313 16:00:03.179738 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556960-mkbxd" Mar 13 16:00:03 crc kubenswrapper[4907]: I0313 16:00:03.294935 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e1001d4c-6984-4a3c-8d1f-167a6a48ea8a-config-volume\") pod \"e1001d4c-6984-4a3c-8d1f-167a6a48ea8a\" (UID: \"e1001d4c-6984-4a3c-8d1f-167a6a48ea8a\") " Mar 13 16:00:03 crc kubenswrapper[4907]: I0313 16:00:03.295028 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mpbb6\" (UniqueName: \"kubernetes.io/projected/e1001d4c-6984-4a3c-8d1f-167a6a48ea8a-kube-api-access-mpbb6\") pod \"e1001d4c-6984-4a3c-8d1f-167a6a48ea8a\" (UID: \"e1001d4c-6984-4a3c-8d1f-167a6a48ea8a\") " Mar 13 16:00:03 crc kubenswrapper[4907]: I0313 16:00:03.295073 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e1001d4c-6984-4a3c-8d1f-167a6a48ea8a-secret-volume\") pod \"e1001d4c-6984-4a3c-8d1f-167a6a48ea8a\" (UID: \"e1001d4c-6984-4a3c-8d1f-167a6a48ea8a\") " Mar 13 16:00:03 crc kubenswrapper[4907]: I0313 16:00:03.295735 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e1001d4c-6984-4a3c-8d1f-167a6a48ea8a-config-volume" (OuterVolumeSpecName: "config-volume") pod "e1001d4c-6984-4a3c-8d1f-167a6a48ea8a" (UID: "e1001d4c-6984-4a3c-8d1f-167a6a48ea8a"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 16:00:03 crc kubenswrapper[4907]: I0313 16:00:03.303168 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1001d4c-6984-4a3c-8d1f-167a6a48ea8a-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "e1001d4c-6984-4a3c-8d1f-167a6a48ea8a" (UID: "e1001d4c-6984-4a3c-8d1f-167a6a48ea8a"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:00:03 crc kubenswrapper[4907]: I0313 16:00:03.303177 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e1001d4c-6984-4a3c-8d1f-167a6a48ea8a-kube-api-access-mpbb6" (OuterVolumeSpecName: "kube-api-access-mpbb6") pod "e1001d4c-6984-4a3c-8d1f-167a6a48ea8a" (UID: "e1001d4c-6984-4a3c-8d1f-167a6a48ea8a"). InnerVolumeSpecName "kube-api-access-mpbb6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:00:03 crc kubenswrapper[4907]: I0313 16:00:03.398731 4907 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e1001d4c-6984-4a3c-8d1f-167a6a48ea8a-config-volume\") on node \"crc\" DevicePath \"\"" Mar 13 16:00:03 crc kubenswrapper[4907]: I0313 16:00:03.398769 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mpbb6\" (UniqueName: \"kubernetes.io/projected/e1001d4c-6984-4a3c-8d1f-167a6a48ea8a-kube-api-access-mpbb6\") on node \"crc\" DevicePath \"\"" Mar 13 16:00:03 crc kubenswrapper[4907]: I0313 16:00:03.398788 4907 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e1001d4c-6984-4a3c-8d1f-167a6a48ea8a-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 13 16:00:03 crc kubenswrapper[4907]: I0313 16:00:03.821587 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556960-mkbxd" event={"ID":"e1001d4c-6984-4a3c-8d1f-167a6a48ea8a","Type":"ContainerDied","Data":"961c155ac80d225a208f0f959fac939f4afb3a793830f43150c57ec2c6d02d2a"} Mar 13 16:00:03 crc kubenswrapper[4907]: I0313 16:00:03.821640 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="961c155ac80d225a208f0f959fac939f4afb3a793830f43150c57ec2c6d02d2a" Mar 13 16:00:03 crc kubenswrapper[4907]: I0313 16:00:03.821718 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556960-mkbxd" Mar 13 16:00:04 crc kubenswrapper[4907]: I0313 16:00:04.248316 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556915-jtzgq"] Mar 13 16:00:04 crc kubenswrapper[4907]: I0313 16:00:04.260686 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556915-jtzgq"] Mar 13 16:00:05 crc kubenswrapper[4907]: I0313 16:00:05.799079 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1fb6a3c2-e8df-47a4-9ae7-2ee38842051a" path="/var/lib/kubelet/pods/1fb6a3c2-e8df-47a4-9ae7-2ee38842051a/volumes" Mar 13 16:00:10 crc kubenswrapper[4907]: I0313 16:00:10.890458 4907 generic.go:334] "Generic (PLEG): container finished" podID="acb3140e-a115-443a-a39e-5f180a196b13" containerID="90c713255623c6a0cd4f0904c875e6fe6995c5ac18245db4d00a77bd7e6d5d8c" exitCode=0 Mar 13 16:00:10 crc kubenswrapper[4907]: I0313 16:00:10.890615 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556960-8xc84" event={"ID":"acb3140e-a115-443a-a39e-5f180a196b13","Type":"ContainerDied","Data":"90c713255623c6a0cd4f0904c875e6fe6995c5ac18245db4d00a77bd7e6d5d8c"} Mar 13 16:00:12 crc kubenswrapper[4907]: I0313 16:00:12.357506 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556960-8xc84" Mar 13 16:00:12 crc kubenswrapper[4907]: I0313 16:00:12.508047 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xgvkq\" (UniqueName: \"kubernetes.io/projected/acb3140e-a115-443a-a39e-5f180a196b13-kube-api-access-xgvkq\") pod \"acb3140e-a115-443a-a39e-5f180a196b13\" (UID: \"acb3140e-a115-443a-a39e-5f180a196b13\") " Mar 13 16:00:12 crc kubenswrapper[4907]: I0313 16:00:12.524205 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/acb3140e-a115-443a-a39e-5f180a196b13-kube-api-access-xgvkq" (OuterVolumeSpecName: "kube-api-access-xgvkq") pod "acb3140e-a115-443a-a39e-5f180a196b13" (UID: "acb3140e-a115-443a-a39e-5f180a196b13"). InnerVolumeSpecName "kube-api-access-xgvkq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:00:12 crc kubenswrapper[4907]: I0313 16:00:12.610865 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xgvkq\" (UniqueName: \"kubernetes.io/projected/acb3140e-a115-443a-a39e-5f180a196b13-kube-api-access-xgvkq\") on node \"crc\" DevicePath \"\"" Mar 13 16:00:12 crc kubenswrapper[4907]: I0313 16:00:12.916438 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556960-8xc84" event={"ID":"acb3140e-a115-443a-a39e-5f180a196b13","Type":"ContainerDied","Data":"4a57c716b0e24bb904434f950bb9538fc19745e51f498ae299351e13dbb5f5e4"} Mar 13 16:00:12 crc kubenswrapper[4907]: I0313 16:00:12.916789 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4a57c716b0e24bb904434f950bb9538fc19745e51f498ae299351e13dbb5f5e4" Mar 13 16:00:12 crc kubenswrapper[4907]: I0313 16:00:12.916524 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556960-8xc84" Mar 13 16:00:13 crc kubenswrapper[4907]: I0313 16:00:13.066971 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-e749-account-create-update-gsxrn"] Mar 13 16:00:13 crc kubenswrapper[4907]: I0313 16:00:13.083971 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-db-create-mzct7"] Mar 13 16:00:13 crc kubenswrapper[4907]: I0313 16:00:13.095683 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-e749-account-create-update-gsxrn"] Mar 13 16:00:13 crc kubenswrapper[4907]: I0313 16:00:13.106725 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-db-create-mzct7"] Mar 13 16:00:13 crc kubenswrapper[4907]: I0313 16:00:13.417943 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556954-kn875"] Mar 13 16:00:13 crc kubenswrapper[4907]: I0313 16:00:13.426550 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556954-kn875"] Mar 13 16:00:13 crc kubenswrapper[4907]: I0313 16:00:13.796576 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0498e113-86c3-4718-a580-13bfdb277ff2" path="/var/lib/kubelet/pods/0498e113-86c3-4718-a580-13bfdb277ff2/volumes" Mar 13 16:00:13 crc kubenswrapper[4907]: I0313 16:00:13.797472 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="97bac75d-c6b8-4d2e-8a24-d0c6279347b5" path="/var/lib/kubelet/pods/97bac75d-c6b8-4d2e-8a24-d0c6279347b5/volumes" Mar 13 16:00:13 crc kubenswrapper[4907]: I0313 16:00:13.798281 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b2a34745-a555-4029-a203-ee17762929e2" path="/var/lib/kubelet/pods/b2a34745-a555-4029-a203-ee17762929e2/volumes" Mar 13 16:00:18 crc kubenswrapper[4907]: I0313 16:00:18.042213 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 16:00:18 crc kubenswrapper[4907]: I0313 16:00:18.042830 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 16:00:25 crc kubenswrapper[4907]: I0313 16:00:25.042058 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-db-sync-nhx6q"] Mar 13 16:00:25 crc kubenswrapper[4907]: I0313 16:00:25.055193 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-db-sync-nhx6q"] Mar 13 16:00:25 crc kubenswrapper[4907]: I0313 16:00:25.793960 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2df4eaaf-eb5c-4e04-b987-5aedb6561f7e" path="/var/lib/kubelet/pods/2df4eaaf-eb5c-4e04-b987-5aedb6561f7e/volumes" Mar 13 16:00:40 crc kubenswrapper[4907]: I0313 16:00:40.355303 4907 scope.go:117] "RemoveContainer" containerID="7fe743a125369b9e5af6f2cb7a4bc8d2add93450e84c6dadc626912538eac565" Mar 13 16:00:40 crc kubenswrapper[4907]: I0313 16:00:40.380870 4907 scope.go:117] "RemoveContainer" containerID="4dac6251d0188e6b289ee01e6cd109fb2dacb259e342a008dc8836c090408baa" Mar 13 16:00:40 crc kubenswrapper[4907]: I0313 16:00:40.453159 4907 scope.go:117] "RemoveContainer" containerID="f8d1778bda06ccfd24371ed557383017005b23e6277f212ccf01bbaf98e64b94" Mar 13 16:00:40 crc kubenswrapper[4907]: I0313 16:00:40.514868 4907 scope.go:117] "RemoveContainer" containerID="3da3591c8cb3c5a70194cd0e1a361e544712615fd8402b78ddd2f90160064c36" Mar 13 16:00:40 crc kubenswrapper[4907]: I0313 16:00:40.539444 4907 scope.go:117] "RemoveContainer" containerID="8cc12017f30ada4d93b2676752544b7d25e61944ae3bcbe64687e02b090ffb18" Mar 13 16:00:40 crc kubenswrapper[4907]: I0313 16:00:40.592031 4907 scope.go:117] "RemoveContainer" containerID="84256d7f66a6d2a7c62b2540cda3765257adc44483f0deff6d7044e7a488955a" Mar 13 16:00:48 crc kubenswrapper[4907]: I0313 16:00:48.041666 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 16:00:48 crc kubenswrapper[4907]: I0313 16:00:48.042563 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 16:00:52 crc kubenswrapper[4907]: I0313 16:00:52.756594 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-7nbgp"] Mar 13 16:00:52 crc kubenswrapper[4907]: E0313 16:00:52.757698 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1001d4c-6984-4a3c-8d1f-167a6a48ea8a" containerName="collect-profiles" Mar 13 16:00:52 crc kubenswrapper[4907]: I0313 16:00:52.757714 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1001d4c-6984-4a3c-8d1f-167a6a48ea8a" containerName="collect-profiles" Mar 13 16:00:52 crc kubenswrapper[4907]: E0313 16:00:52.757734 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="acb3140e-a115-443a-a39e-5f180a196b13" containerName="oc" Mar 13 16:00:52 crc kubenswrapper[4907]: I0313 16:00:52.757742 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="acb3140e-a115-443a-a39e-5f180a196b13" containerName="oc" Mar 13 16:00:52 crc kubenswrapper[4907]: I0313 16:00:52.758005 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="acb3140e-a115-443a-a39e-5f180a196b13" containerName="oc" Mar 13 16:00:52 crc kubenswrapper[4907]: I0313 16:00:52.758021 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1001d4c-6984-4a3c-8d1f-167a6a48ea8a" containerName="collect-profiles" Mar 13 16:00:52 crc kubenswrapper[4907]: I0313 16:00:52.760110 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7nbgp" Mar 13 16:00:52 crc kubenswrapper[4907]: I0313 16:00:52.777687 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7nbgp"] Mar 13 16:00:52 crc kubenswrapper[4907]: I0313 16:00:52.952615 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x5stk\" (UniqueName: \"kubernetes.io/projected/a13cade5-fe42-46a8-9c42-d1733bde2b78-kube-api-access-x5stk\") pod \"certified-operators-7nbgp\" (UID: \"a13cade5-fe42-46a8-9c42-d1733bde2b78\") " pod="openshift-marketplace/certified-operators-7nbgp" Mar 13 16:00:52 crc kubenswrapper[4907]: I0313 16:00:52.954081 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a13cade5-fe42-46a8-9c42-d1733bde2b78-utilities\") pod \"certified-operators-7nbgp\" (UID: \"a13cade5-fe42-46a8-9c42-d1733bde2b78\") " pod="openshift-marketplace/certified-operators-7nbgp" Mar 13 16:00:52 crc kubenswrapper[4907]: I0313 16:00:52.954263 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a13cade5-fe42-46a8-9c42-d1733bde2b78-catalog-content\") pod \"certified-operators-7nbgp\" (UID: \"a13cade5-fe42-46a8-9c42-d1733bde2b78\") " pod="openshift-marketplace/certified-operators-7nbgp" Mar 13 16:00:53 crc kubenswrapper[4907]: I0313 16:00:53.057481 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x5stk\" (UniqueName: \"kubernetes.io/projected/a13cade5-fe42-46a8-9c42-d1733bde2b78-kube-api-access-x5stk\") pod \"certified-operators-7nbgp\" (UID: \"a13cade5-fe42-46a8-9c42-d1733bde2b78\") " pod="openshift-marketplace/certified-operators-7nbgp" Mar 13 16:00:53 crc kubenswrapper[4907]: I0313 16:00:53.057601 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a13cade5-fe42-46a8-9c42-d1733bde2b78-utilities\") pod \"certified-operators-7nbgp\" (UID: \"a13cade5-fe42-46a8-9c42-d1733bde2b78\") " pod="openshift-marketplace/certified-operators-7nbgp" Mar 13 16:00:53 crc kubenswrapper[4907]: I0313 16:00:53.057635 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a13cade5-fe42-46a8-9c42-d1733bde2b78-catalog-content\") pod \"certified-operators-7nbgp\" (UID: \"a13cade5-fe42-46a8-9c42-d1733bde2b78\") " pod="openshift-marketplace/certified-operators-7nbgp" Mar 13 16:00:53 crc kubenswrapper[4907]: I0313 16:00:53.058145 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a13cade5-fe42-46a8-9c42-d1733bde2b78-utilities\") pod \"certified-operators-7nbgp\" (UID: \"a13cade5-fe42-46a8-9c42-d1733bde2b78\") " pod="openshift-marketplace/certified-operators-7nbgp" Mar 13 16:00:53 crc kubenswrapper[4907]: I0313 16:00:53.058223 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a13cade5-fe42-46a8-9c42-d1733bde2b78-catalog-content\") pod \"certified-operators-7nbgp\" (UID: \"a13cade5-fe42-46a8-9c42-d1733bde2b78\") " pod="openshift-marketplace/certified-operators-7nbgp" Mar 13 16:00:53 crc kubenswrapper[4907]: I0313 16:00:53.078051 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x5stk\" (UniqueName: \"kubernetes.io/projected/a13cade5-fe42-46a8-9c42-d1733bde2b78-kube-api-access-x5stk\") pod \"certified-operators-7nbgp\" (UID: \"a13cade5-fe42-46a8-9c42-d1733bde2b78\") " pod="openshift-marketplace/certified-operators-7nbgp" Mar 13 16:00:53 crc kubenswrapper[4907]: I0313 16:00:53.115688 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7nbgp" Mar 13 16:00:53 crc kubenswrapper[4907]: I0313 16:00:53.664369 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7nbgp"] Mar 13 16:00:54 crc kubenswrapper[4907]: I0313 16:00:54.353802 4907 generic.go:334] "Generic (PLEG): container finished" podID="a13cade5-fe42-46a8-9c42-d1733bde2b78" containerID="d774a5d1c6b13e107977d37001f9f575455851b3923de08704f840ff5ea48c41" exitCode=0 Mar 13 16:00:54 crc kubenswrapper[4907]: I0313 16:00:54.353945 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7nbgp" event={"ID":"a13cade5-fe42-46a8-9c42-d1733bde2b78","Type":"ContainerDied","Data":"d774a5d1c6b13e107977d37001f9f575455851b3923de08704f840ff5ea48c41"} Mar 13 16:00:54 crc kubenswrapper[4907]: I0313 16:00:54.354150 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7nbgp" event={"ID":"a13cade5-fe42-46a8-9c42-d1733bde2b78","Type":"ContainerStarted","Data":"673546a70a0ddd984900b7492ee18632405c8a55ae70039f43b9cdb5ae87218b"} Mar 13 16:00:56 crc kubenswrapper[4907]: I0313 16:00:56.387956 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7nbgp" event={"ID":"a13cade5-fe42-46a8-9c42-d1733bde2b78","Type":"ContainerStarted","Data":"526804910f6c587f26fe53ee4b1003b6d89146391dbb3182ce1f442e7cc49213"} Mar 13 16:00:57 crc kubenswrapper[4907]: I0313 16:00:57.401933 4907 generic.go:334] "Generic (PLEG): container finished" podID="a13cade5-fe42-46a8-9c42-d1733bde2b78" containerID="526804910f6c587f26fe53ee4b1003b6d89146391dbb3182ce1f442e7cc49213" exitCode=0 Mar 13 16:00:57 crc kubenswrapper[4907]: I0313 16:00:57.402009 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7nbgp" event={"ID":"a13cade5-fe42-46a8-9c42-d1733bde2b78","Type":"ContainerDied","Data":"526804910f6c587f26fe53ee4b1003b6d89146391dbb3182ce1f442e7cc49213"} Mar 13 16:00:59 crc kubenswrapper[4907]: I0313 16:00:59.432600 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7nbgp" event={"ID":"a13cade5-fe42-46a8-9c42-d1733bde2b78","Type":"ContainerStarted","Data":"5770c77601b4306c58f1d10a7f3d071183506a543de6bfd4b22de746737a8480"} Mar 13 16:00:59 crc kubenswrapper[4907]: I0313 16:00:59.468757 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-7nbgp" podStartSLOduration=3.459966921 podStartE2EDuration="7.468733582s" podCreationTimestamp="2026-03-13 16:00:52 +0000 UTC" firstStartedPulling="2026-03-13 16:00:54.355803613 +0000 UTC m=+6953.255591302" lastFinishedPulling="2026-03-13 16:00:58.364570244 +0000 UTC m=+6957.264357963" observedRunningTime="2026-03-13 16:00:59.45369479 +0000 UTC m=+6958.353482519" watchObservedRunningTime="2026-03-13 16:00:59.468733582 +0000 UTC m=+6958.368521281" Mar 13 16:01:00 crc kubenswrapper[4907]: I0313 16:01:00.168039 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29556961-gshbz"] Mar 13 16:01:00 crc kubenswrapper[4907]: I0313 16:01:00.170956 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29556961-gshbz" Mar 13 16:01:00 crc kubenswrapper[4907]: I0313 16:01:00.186096 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29556961-gshbz"] Mar 13 16:01:00 crc kubenswrapper[4907]: I0313 16:01:00.335130 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e548a313-fc14-4ce9-b268-bf37c79eed83-config-data\") pod \"keystone-cron-29556961-gshbz\" (UID: \"e548a313-fc14-4ce9-b268-bf37c79eed83\") " pod="openstack/keystone-cron-29556961-gshbz" Mar 13 16:01:00 crc kubenswrapper[4907]: I0313 16:01:00.335181 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e548a313-fc14-4ce9-b268-bf37c79eed83-combined-ca-bundle\") pod \"keystone-cron-29556961-gshbz\" (UID: \"e548a313-fc14-4ce9-b268-bf37c79eed83\") " pod="openstack/keystone-cron-29556961-gshbz" Mar 13 16:01:00 crc kubenswrapper[4907]: I0313 16:01:00.335358 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5qmmz\" (UniqueName: \"kubernetes.io/projected/e548a313-fc14-4ce9-b268-bf37c79eed83-kube-api-access-5qmmz\") pod \"keystone-cron-29556961-gshbz\" (UID: \"e548a313-fc14-4ce9-b268-bf37c79eed83\") " pod="openstack/keystone-cron-29556961-gshbz" Mar 13 16:01:00 crc kubenswrapper[4907]: I0313 16:01:00.335673 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e548a313-fc14-4ce9-b268-bf37c79eed83-fernet-keys\") pod \"keystone-cron-29556961-gshbz\" (UID: \"e548a313-fc14-4ce9-b268-bf37c79eed83\") " pod="openstack/keystone-cron-29556961-gshbz" Mar 13 16:01:00 crc kubenswrapper[4907]: I0313 16:01:00.437428 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5qmmz\" (UniqueName: \"kubernetes.io/projected/e548a313-fc14-4ce9-b268-bf37c79eed83-kube-api-access-5qmmz\") pod \"keystone-cron-29556961-gshbz\" (UID: \"e548a313-fc14-4ce9-b268-bf37c79eed83\") " pod="openstack/keystone-cron-29556961-gshbz" Mar 13 16:01:00 crc kubenswrapper[4907]: I0313 16:01:00.437625 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e548a313-fc14-4ce9-b268-bf37c79eed83-fernet-keys\") pod \"keystone-cron-29556961-gshbz\" (UID: \"e548a313-fc14-4ce9-b268-bf37c79eed83\") " pod="openstack/keystone-cron-29556961-gshbz" Mar 13 16:01:00 crc kubenswrapper[4907]: I0313 16:01:00.437744 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e548a313-fc14-4ce9-b268-bf37c79eed83-config-data\") pod \"keystone-cron-29556961-gshbz\" (UID: \"e548a313-fc14-4ce9-b268-bf37c79eed83\") " pod="openstack/keystone-cron-29556961-gshbz" Mar 13 16:01:00 crc kubenswrapper[4907]: I0313 16:01:00.437777 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e548a313-fc14-4ce9-b268-bf37c79eed83-combined-ca-bundle\") pod \"keystone-cron-29556961-gshbz\" (UID: \"e548a313-fc14-4ce9-b268-bf37c79eed83\") " pod="openstack/keystone-cron-29556961-gshbz" Mar 13 16:01:00 crc kubenswrapper[4907]: I0313 16:01:00.443791 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e548a313-fc14-4ce9-b268-bf37c79eed83-config-data\") pod \"keystone-cron-29556961-gshbz\" (UID: \"e548a313-fc14-4ce9-b268-bf37c79eed83\") " pod="openstack/keystone-cron-29556961-gshbz" Mar 13 16:01:00 crc kubenswrapper[4907]: I0313 16:01:00.448673 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e548a313-fc14-4ce9-b268-bf37c79eed83-fernet-keys\") pod \"keystone-cron-29556961-gshbz\" (UID: \"e548a313-fc14-4ce9-b268-bf37c79eed83\") " pod="openstack/keystone-cron-29556961-gshbz" Mar 13 16:01:00 crc kubenswrapper[4907]: I0313 16:01:00.450173 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e548a313-fc14-4ce9-b268-bf37c79eed83-combined-ca-bundle\") pod \"keystone-cron-29556961-gshbz\" (UID: \"e548a313-fc14-4ce9-b268-bf37c79eed83\") " pod="openstack/keystone-cron-29556961-gshbz" Mar 13 16:01:00 crc kubenswrapper[4907]: I0313 16:01:00.460065 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5qmmz\" (UniqueName: \"kubernetes.io/projected/e548a313-fc14-4ce9-b268-bf37c79eed83-kube-api-access-5qmmz\") pod \"keystone-cron-29556961-gshbz\" (UID: \"e548a313-fc14-4ce9-b268-bf37c79eed83\") " pod="openstack/keystone-cron-29556961-gshbz" Mar 13 16:01:00 crc kubenswrapper[4907]: I0313 16:01:00.505490 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29556961-gshbz" Mar 13 16:01:00 crc kubenswrapper[4907]: I0313 16:01:00.981168 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29556961-gshbz"] Mar 13 16:01:00 crc kubenswrapper[4907]: W0313 16:01:00.984457 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode548a313_fc14_4ce9_b268_bf37c79eed83.slice/crio-b8e1101812a621dd17914318034354e00ada6a7f97ba9cc9819509f6432c55e1 WatchSource:0}: Error finding container b8e1101812a621dd17914318034354e00ada6a7f97ba9cc9819509f6432c55e1: Status 404 returned error can't find the container with id b8e1101812a621dd17914318034354e00ada6a7f97ba9cc9819509f6432c55e1 Mar 13 16:01:01 crc kubenswrapper[4907]: I0313 16:01:01.453249 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29556961-gshbz" event={"ID":"e548a313-fc14-4ce9-b268-bf37c79eed83","Type":"ContainerStarted","Data":"4ff9398b9dec5963cf443cc668dffeb42de87e641ad4f61c8cd02708682d9979"} Mar 13 16:01:01 crc kubenswrapper[4907]: I0313 16:01:01.453327 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29556961-gshbz" event={"ID":"e548a313-fc14-4ce9-b268-bf37c79eed83","Type":"ContainerStarted","Data":"b8e1101812a621dd17914318034354e00ada6a7f97ba9cc9819509f6432c55e1"} Mar 13 16:01:01 crc kubenswrapper[4907]: I0313 16:01:01.486058 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29556961-gshbz" podStartSLOduration=1.4860244360000001 podStartE2EDuration="1.486024436s" podCreationTimestamp="2026-03-13 16:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 16:01:01.480968997 +0000 UTC m=+6960.380756686" watchObservedRunningTime="2026-03-13 16:01:01.486024436 +0000 UTC m=+6960.385812185" Mar 13 16:01:03 crc kubenswrapper[4907]: I0313 16:01:03.115842 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-7nbgp" Mar 13 16:01:03 crc kubenswrapper[4907]: I0313 16:01:03.116488 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-7nbgp" Mar 13 16:01:03 crc kubenswrapper[4907]: I0313 16:01:03.209069 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-7nbgp" Mar 13 16:01:03 crc kubenswrapper[4907]: I0313 16:01:03.546833 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-7nbgp" Mar 13 16:01:03 crc kubenswrapper[4907]: I0313 16:01:03.600925 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7nbgp"] Mar 13 16:01:04 crc kubenswrapper[4907]: E0313 16:01:04.040533 4907 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode548a313_fc14_4ce9_b268_bf37c79eed83.slice/crio-4ff9398b9dec5963cf443cc668dffeb42de87e641ad4f61c8cd02708682d9979.scope\": RecentStats: unable to find data in memory cache]" Mar 13 16:01:04 crc kubenswrapper[4907]: I0313 16:01:04.508538 4907 generic.go:334] "Generic (PLEG): container finished" podID="e548a313-fc14-4ce9-b268-bf37c79eed83" containerID="4ff9398b9dec5963cf443cc668dffeb42de87e641ad4f61c8cd02708682d9979" exitCode=0 Mar 13 16:01:04 crc kubenswrapper[4907]: I0313 16:01:04.508959 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29556961-gshbz" event={"ID":"e548a313-fc14-4ce9-b268-bf37c79eed83","Type":"ContainerDied","Data":"4ff9398b9dec5963cf443cc668dffeb42de87e641ad4f61c8cd02708682d9979"} Mar 13 16:01:05 crc kubenswrapper[4907]: I0313 16:01:05.516214 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-7nbgp" podUID="a13cade5-fe42-46a8-9c42-d1733bde2b78" containerName="registry-server" containerID="cri-o://5770c77601b4306c58f1d10a7f3d071183506a543de6bfd4b22de746737a8480" gracePeriod=2 Mar 13 16:01:05 crc kubenswrapper[4907]: I0313 16:01:05.893545 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29556961-gshbz" Mar 13 16:01:06 crc kubenswrapper[4907]: I0313 16:01:06.006709 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7nbgp" Mar 13 16:01:06 crc kubenswrapper[4907]: I0313 16:01:06.068337 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e548a313-fc14-4ce9-b268-bf37c79eed83-combined-ca-bundle\") pod \"e548a313-fc14-4ce9-b268-bf37c79eed83\" (UID: \"e548a313-fc14-4ce9-b268-bf37c79eed83\") " Mar 13 16:01:06 crc kubenswrapper[4907]: I0313 16:01:06.068516 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e548a313-fc14-4ce9-b268-bf37c79eed83-fernet-keys\") pod \"e548a313-fc14-4ce9-b268-bf37c79eed83\" (UID: \"e548a313-fc14-4ce9-b268-bf37c79eed83\") " Mar 13 16:01:06 crc kubenswrapper[4907]: I0313 16:01:06.068710 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5qmmz\" (UniqueName: \"kubernetes.io/projected/e548a313-fc14-4ce9-b268-bf37c79eed83-kube-api-access-5qmmz\") pod \"e548a313-fc14-4ce9-b268-bf37c79eed83\" (UID: \"e548a313-fc14-4ce9-b268-bf37c79eed83\") " Mar 13 16:01:06 crc kubenswrapper[4907]: I0313 16:01:06.068813 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e548a313-fc14-4ce9-b268-bf37c79eed83-config-data\") pod \"e548a313-fc14-4ce9-b268-bf37c79eed83\" (UID: \"e548a313-fc14-4ce9-b268-bf37c79eed83\") " Mar 13 16:01:06 crc kubenswrapper[4907]: I0313 16:01:06.073722 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e548a313-fc14-4ce9-b268-bf37c79eed83-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "e548a313-fc14-4ce9-b268-bf37c79eed83" (UID: "e548a313-fc14-4ce9-b268-bf37c79eed83"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:01:06 crc kubenswrapper[4907]: I0313 16:01:06.073786 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e548a313-fc14-4ce9-b268-bf37c79eed83-kube-api-access-5qmmz" (OuterVolumeSpecName: "kube-api-access-5qmmz") pod "e548a313-fc14-4ce9-b268-bf37c79eed83" (UID: "e548a313-fc14-4ce9-b268-bf37c79eed83"). InnerVolumeSpecName "kube-api-access-5qmmz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:01:06 crc kubenswrapper[4907]: I0313 16:01:06.096107 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e548a313-fc14-4ce9-b268-bf37c79eed83-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e548a313-fc14-4ce9-b268-bf37c79eed83" (UID: "e548a313-fc14-4ce9-b268-bf37c79eed83"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:01:06 crc kubenswrapper[4907]: I0313 16:01:06.121636 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e548a313-fc14-4ce9-b268-bf37c79eed83-config-data" (OuterVolumeSpecName: "config-data") pod "e548a313-fc14-4ce9-b268-bf37c79eed83" (UID: "e548a313-fc14-4ce9-b268-bf37c79eed83"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:01:06 crc kubenswrapper[4907]: I0313 16:01:06.170784 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a13cade5-fe42-46a8-9c42-d1733bde2b78-catalog-content\") pod \"a13cade5-fe42-46a8-9c42-d1733bde2b78\" (UID: \"a13cade5-fe42-46a8-9c42-d1733bde2b78\") " Mar 13 16:01:06 crc kubenswrapper[4907]: I0313 16:01:06.171089 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x5stk\" (UniqueName: \"kubernetes.io/projected/a13cade5-fe42-46a8-9c42-d1733bde2b78-kube-api-access-x5stk\") pod \"a13cade5-fe42-46a8-9c42-d1733bde2b78\" (UID: \"a13cade5-fe42-46a8-9c42-d1733bde2b78\") " Mar 13 16:01:06 crc kubenswrapper[4907]: I0313 16:01:06.171148 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a13cade5-fe42-46a8-9c42-d1733bde2b78-utilities\") pod \"a13cade5-fe42-46a8-9c42-d1733bde2b78\" (UID: \"a13cade5-fe42-46a8-9c42-d1733bde2b78\") " Mar 13 16:01:06 crc kubenswrapper[4907]: I0313 16:01:06.171728 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5qmmz\" (UniqueName: \"kubernetes.io/projected/e548a313-fc14-4ce9-b268-bf37c79eed83-kube-api-access-5qmmz\") on node \"crc\" DevicePath \"\"" Mar 13 16:01:06 crc kubenswrapper[4907]: I0313 16:01:06.171747 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e548a313-fc14-4ce9-b268-bf37c79eed83-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 16:01:06 crc kubenswrapper[4907]: I0313 16:01:06.171758 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e548a313-fc14-4ce9-b268-bf37c79eed83-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 16:01:06 crc kubenswrapper[4907]: I0313 16:01:06.171780 4907 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e548a313-fc14-4ce9-b268-bf37c79eed83-fernet-keys\") on node \"crc\" DevicePath \"\"" Mar 13 16:01:06 crc kubenswrapper[4907]: I0313 16:01:06.171968 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a13cade5-fe42-46a8-9c42-d1733bde2b78-utilities" (OuterVolumeSpecName: "utilities") pod "a13cade5-fe42-46a8-9c42-d1733bde2b78" (UID: "a13cade5-fe42-46a8-9c42-d1733bde2b78"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 16:01:06 crc kubenswrapper[4907]: I0313 16:01:06.174476 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a13cade5-fe42-46a8-9c42-d1733bde2b78-kube-api-access-x5stk" (OuterVolumeSpecName: "kube-api-access-x5stk") pod "a13cade5-fe42-46a8-9c42-d1733bde2b78" (UID: "a13cade5-fe42-46a8-9c42-d1733bde2b78"). InnerVolumeSpecName "kube-api-access-x5stk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:01:06 crc kubenswrapper[4907]: I0313 16:01:06.273508 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x5stk\" (UniqueName: \"kubernetes.io/projected/a13cade5-fe42-46a8-9c42-d1733bde2b78-kube-api-access-x5stk\") on node \"crc\" DevicePath \"\"" Mar 13 16:01:06 crc kubenswrapper[4907]: I0313 16:01:06.273543 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a13cade5-fe42-46a8-9c42-d1733bde2b78-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 16:01:06 crc kubenswrapper[4907]: I0313 16:01:06.528013 4907 generic.go:334] "Generic (PLEG): container finished" podID="a13cade5-fe42-46a8-9c42-d1733bde2b78" containerID="5770c77601b4306c58f1d10a7f3d071183506a543de6bfd4b22de746737a8480" exitCode=0 Mar 13 16:01:06 crc kubenswrapper[4907]: I0313 16:01:06.528055 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7nbgp" event={"ID":"a13cade5-fe42-46a8-9c42-d1733bde2b78","Type":"ContainerDied","Data":"5770c77601b4306c58f1d10a7f3d071183506a543de6bfd4b22de746737a8480"} Mar 13 16:01:06 crc kubenswrapper[4907]: I0313 16:01:06.528080 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7nbgp" Mar 13 16:01:06 crc kubenswrapper[4907]: I0313 16:01:06.528485 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7nbgp" event={"ID":"a13cade5-fe42-46a8-9c42-d1733bde2b78","Type":"ContainerDied","Data":"673546a70a0ddd984900b7492ee18632405c8a55ae70039f43b9cdb5ae87218b"} Mar 13 16:01:06 crc kubenswrapper[4907]: I0313 16:01:06.528515 4907 scope.go:117] "RemoveContainer" containerID="5770c77601b4306c58f1d10a7f3d071183506a543de6bfd4b22de746737a8480" Mar 13 16:01:06 crc kubenswrapper[4907]: I0313 16:01:06.531156 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29556961-gshbz" event={"ID":"e548a313-fc14-4ce9-b268-bf37c79eed83","Type":"ContainerDied","Data":"b8e1101812a621dd17914318034354e00ada6a7f97ba9cc9819509f6432c55e1"} Mar 13 16:01:06 crc kubenswrapper[4907]: I0313 16:01:06.531192 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b8e1101812a621dd17914318034354e00ada6a7f97ba9cc9819509f6432c55e1" Mar 13 16:01:06 crc kubenswrapper[4907]: I0313 16:01:06.531218 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29556961-gshbz" Mar 13 16:01:06 crc kubenswrapper[4907]: I0313 16:01:06.553281 4907 scope.go:117] "RemoveContainer" containerID="526804910f6c587f26fe53ee4b1003b6d89146391dbb3182ce1f442e7cc49213" Mar 13 16:01:06 crc kubenswrapper[4907]: I0313 16:01:06.587255 4907 scope.go:117] "RemoveContainer" containerID="d774a5d1c6b13e107977d37001f9f575455851b3923de08704f840ff5ea48c41" Mar 13 16:01:06 crc kubenswrapper[4907]: I0313 16:01:06.612605 4907 scope.go:117] "RemoveContainer" containerID="5770c77601b4306c58f1d10a7f3d071183506a543de6bfd4b22de746737a8480" Mar 13 16:01:06 crc kubenswrapper[4907]: E0313 16:01:06.613166 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5770c77601b4306c58f1d10a7f3d071183506a543de6bfd4b22de746737a8480\": container with ID starting with 5770c77601b4306c58f1d10a7f3d071183506a543de6bfd4b22de746737a8480 not found: ID does not exist" containerID="5770c77601b4306c58f1d10a7f3d071183506a543de6bfd4b22de746737a8480" Mar 13 16:01:06 crc kubenswrapper[4907]: I0313 16:01:06.613306 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5770c77601b4306c58f1d10a7f3d071183506a543de6bfd4b22de746737a8480"} err="failed to get container status \"5770c77601b4306c58f1d10a7f3d071183506a543de6bfd4b22de746737a8480\": rpc error: code = NotFound desc = could not find container \"5770c77601b4306c58f1d10a7f3d071183506a543de6bfd4b22de746737a8480\": container with ID starting with 5770c77601b4306c58f1d10a7f3d071183506a543de6bfd4b22de746737a8480 not found: ID does not exist" Mar 13 16:01:06 crc kubenswrapper[4907]: I0313 16:01:06.613388 4907 scope.go:117] "RemoveContainer" containerID="526804910f6c587f26fe53ee4b1003b6d89146391dbb3182ce1f442e7cc49213" Mar 13 16:01:06 crc kubenswrapper[4907]: E0313 16:01:06.613768 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"526804910f6c587f26fe53ee4b1003b6d89146391dbb3182ce1f442e7cc49213\": container with ID starting with 526804910f6c587f26fe53ee4b1003b6d89146391dbb3182ce1f442e7cc49213 not found: ID does not exist" containerID="526804910f6c587f26fe53ee4b1003b6d89146391dbb3182ce1f442e7cc49213" Mar 13 16:01:06 crc kubenswrapper[4907]: I0313 16:01:06.613841 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"526804910f6c587f26fe53ee4b1003b6d89146391dbb3182ce1f442e7cc49213"} err="failed to get container status \"526804910f6c587f26fe53ee4b1003b6d89146391dbb3182ce1f442e7cc49213\": rpc error: code = NotFound desc = could not find container \"526804910f6c587f26fe53ee4b1003b6d89146391dbb3182ce1f442e7cc49213\": container with ID starting with 526804910f6c587f26fe53ee4b1003b6d89146391dbb3182ce1f442e7cc49213 not found: ID does not exist" Mar 13 16:01:06 crc kubenswrapper[4907]: I0313 16:01:06.613922 4907 scope.go:117] "RemoveContainer" containerID="d774a5d1c6b13e107977d37001f9f575455851b3923de08704f840ff5ea48c41" Mar 13 16:01:06 crc kubenswrapper[4907]: E0313 16:01:06.614465 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d774a5d1c6b13e107977d37001f9f575455851b3923de08704f840ff5ea48c41\": container with ID starting with d774a5d1c6b13e107977d37001f9f575455851b3923de08704f840ff5ea48c41 not found: ID does not exist" containerID="d774a5d1c6b13e107977d37001f9f575455851b3923de08704f840ff5ea48c41" Mar 13 16:01:06 crc kubenswrapper[4907]: I0313 16:01:06.614512 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d774a5d1c6b13e107977d37001f9f575455851b3923de08704f840ff5ea48c41"} err="failed to get container status \"d774a5d1c6b13e107977d37001f9f575455851b3923de08704f840ff5ea48c41\": rpc error: code = NotFound desc = could not find container \"d774a5d1c6b13e107977d37001f9f575455851b3923de08704f840ff5ea48c41\": container with ID starting with d774a5d1c6b13e107977d37001f9f575455851b3923de08704f840ff5ea48c41 not found: ID does not exist" Mar 13 16:01:07 crc kubenswrapper[4907]: I0313 16:01:07.234165 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a13cade5-fe42-46a8-9c42-d1733bde2b78-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a13cade5-fe42-46a8-9c42-d1733bde2b78" (UID: "a13cade5-fe42-46a8-9c42-d1733bde2b78"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 16:01:07 crc kubenswrapper[4907]: I0313 16:01:07.295275 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a13cade5-fe42-46a8-9c42-d1733bde2b78-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 16:01:07 crc kubenswrapper[4907]: I0313 16:01:07.474944 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7nbgp"] Mar 13 16:01:07 crc kubenswrapper[4907]: I0313 16:01:07.491132 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-7nbgp"] Mar 13 16:01:07 crc kubenswrapper[4907]: I0313 16:01:07.797377 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a13cade5-fe42-46a8-9c42-d1733bde2b78" path="/var/lib/kubelet/pods/a13cade5-fe42-46a8-9c42-d1733bde2b78/volumes" Mar 13 16:01:18 crc kubenswrapper[4907]: I0313 16:01:18.041790 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 16:01:18 crc kubenswrapper[4907]: I0313 16:01:18.042349 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 16:01:18 crc kubenswrapper[4907]: I0313 16:01:18.042408 4907 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" Mar 13 16:01:18 crc kubenswrapper[4907]: I0313 16:01:18.043299 4907 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2a3c556bf9f1ef9e4af9dec2c5d043c5ea60bb781b3c04fb7c2299179fb58a19"} pod="openshift-machine-config-operator/machine-config-daemon-hm56j" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 16:01:18 crc kubenswrapper[4907]: I0313 16:01:18.043351 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" containerID="cri-o://2a3c556bf9f1ef9e4af9dec2c5d043c5ea60bb781b3c04fb7c2299179fb58a19" gracePeriod=600 Mar 13 16:01:18 crc kubenswrapper[4907]: I0313 16:01:18.674862 4907 generic.go:334] "Generic (PLEG): container finished" podID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerID="2a3c556bf9f1ef9e4af9dec2c5d043c5ea60bb781b3c04fb7c2299179fb58a19" exitCode=0 Mar 13 16:01:18 crc kubenswrapper[4907]: I0313 16:01:18.675223 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerDied","Data":"2a3c556bf9f1ef9e4af9dec2c5d043c5ea60bb781b3c04fb7c2299179fb58a19"} Mar 13 16:01:18 crc kubenswrapper[4907]: I0313 16:01:18.675251 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerStarted","Data":"97ad1b399368adb5c1fc3c4cf2810fa240ec320625c36e99f246d6171b951964"} Mar 13 16:01:18 crc kubenswrapper[4907]: I0313 16:01:18.675268 4907 scope.go:117] "RemoveContainer" containerID="33f8e315523143b157797243610ad7b82c7815f9e551d1169f2c036cfcadf792" Mar 13 16:02:00 crc kubenswrapper[4907]: I0313 16:02:00.143833 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556962-d4bk5"] Mar 13 16:02:00 crc kubenswrapper[4907]: E0313 16:02:00.144791 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a13cade5-fe42-46a8-9c42-d1733bde2b78" containerName="registry-server" Mar 13 16:02:00 crc kubenswrapper[4907]: I0313 16:02:00.144806 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="a13cade5-fe42-46a8-9c42-d1733bde2b78" containerName="registry-server" Mar 13 16:02:00 crc kubenswrapper[4907]: E0313 16:02:00.144827 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a13cade5-fe42-46a8-9c42-d1733bde2b78" containerName="extract-content" Mar 13 16:02:00 crc kubenswrapper[4907]: I0313 16:02:00.144833 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="a13cade5-fe42-46a8-9c42-d1733bde2b78" containerName="extract-content" Mar 13 16:02:00 crc kubenswrapper[4907]: E0313 16:02:00.144859 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a13cade5-fe42-46a8-9c42-d1733bde2b78" containerName="extract-utilities" Mar 13 16:02:00 crc kubenswrapper[4907]: I0313 16:02:00.144867 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="a13cade5-fe42-46a8-9c42-d1733bde2b78" containerName="extract-utilities" Mar 13 16:02:00 crc kubenswrapper[4907]: E0313 16:02:00.144917 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e548a313-fc14-4ce9-b268-bf37c79eed83" containerName="keystone-cron" Mar 13 16:02:00 crc kubenswrapper[4907]: I0313 16:02:00.144926 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="e548a313-fc14-4ce9-b268-bf37c79eed83" containerName="keystone-cron" Mar 13 16:02:00 crc kubenswrapper[4907]: I0313 16:02:00.145140 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="a13cade5-fe42-46a8-9c42-d1733bde2b78" containerName="registry-server" Mar 13 16:02:00 crc kubenswrapper[4907]: I0313 16:02:00.145170 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="e548a313-fc14-4ce9-b268-bf37c79eed83" containerName="keystone-cron" Mar 13 16:02:00 crc kubenswrapper[4907]: I0313 16:02:00.146014 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556962-d4bk5" Mar 13 16:02:00 crc kubenswrapper[4907]: I0313 16:02:00.148912 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 16:02:00 crc kubenswrapper[4907]: I0313 16:02:00.149100 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 16:02:00 crc kubenswrapper[4907]: I0313 16:02:00.149369 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 16:02:00 crc kubenswrapper[4907]: I0313 16:02:00.155196 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556962-d4bk5"] Mar 13 16:02:00 crc kubenswrapper[4907]: I0313 16:02:00.215408 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7wgf8\" (UniqueName: \"kubernetes.io/projected/927835ce-92e1-4a29-b064-0e08a5288442-kube-api-access-7wgf8\") pod \"auto-csr-approver-29556962-d4bk5\" (UID: \"927835ce-92e1-4a29-b064-0e08a5288442\") " pod="openshift-infra/auto-csr-approver-29556962-d4bk5" Mar 13 16:02:00 crc kubenswrapper[4907]: I0313 16:02:00.318080 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7wgf8\" (UniqueName: \"kubernetes.io/projected/927835ce-92e1-4a29-b064-0e08a5288442-kube-api-access-7wgf8\") pod \"auto-csr-approver-29556962-d4bk5\" (UID: \"927835ce-92e1-4a29-b064-0e08a5288442\") " pod="openshift-infra/auto-csr-approver-29556962-d4bk5" Mar 13 16:02:00 crc kubenswrapper[4907]: I0313 16:02:00.336353 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7wgf8\" (UniqueName: \"kubernetes.io/projected/927835ce-92e1-4a29-b064-0e08a5288442-kube-api-access-7wgf8\") pod \"auto-csr-approver-29556962-d4bk5\" (UID: \"927835ce-92e1-4a29-b064-0e08a5288442\") " pod="openshift-infra/auto-csr-approver-29556962-d4bk5" Mar 13 16:02:00 crc kubenswrapper[4907]: I0313 16:02:00.466979 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556962-d4bk5" Mar 13 16:02:00 crc kubenswrapper[4907]: I0313 16:02:00.938286 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556962-d4bk5"] Mar 13 16:02:01 crc kubenswrapper[4907]: I0313 16:02:01.047290 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556962-d4bk5" event={"ID":"927835ce-92e1-4a29-b064-0e08a5288442","Type":"ContainerStarted","Data":"59176c1fbdb0435bb688c727d47f358b21ee89c1dba9385298bbb961b4960ce1"} Mar 13 16:02:03 crc kubenswrapper[4907]: I0313 16:02:03.066648 4907 generic.go:334] "Generic (PLEG): container finished" podID="927835ce-92e1-4a29-b064-0e08a5288442" containerID="fd8d416a2c11ed375f8397b31e95379985b54fa1a07f700d79a805d7106a6c4e" exitCode=0 Mar 13 16:02:03 crc kubenswrapper[4907]: I0313 16:02:03.066767 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556962-d4bk5" event={"ID":"927835ce-92e1-4a29-b064-0e08a5288442","Type":"ContainerDied","Data":"fd8d416a2c11ed375f8397b31e95379985b54fa1a07f700d79a805d7106a6c4e"} Mar 13 16:02:04 crc kubenswrapper[4907]: I0313 16:02:04.452952 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556962-d4bk5" Mar 13 16:02:04 crc kubenswrapper[4907]: I0313 16:02:04.623664 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7wgf8\" (UniqueName: \"kubernetes.io/projected/927835ce-92e1-4a29-b064-0e08a5288442-kube-api-access-7wgf8\") pod \"927835ce-92e1-4a29-b064-0e08a5288442\" (UID: \"927835ce-92e1-4a29-b064-0e08a5288442\") " Mar 13 16:02:04 crc kubenswrapper[4907]: I0313 16:02:04.629233 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/927835ce-92e1-4a29-b064-0e08a5288442-kube-api-access-7wgf8" (OuterVolumeSpecName: "kube-api-access-7wgf8") pod "927835ce-92e1-4a29-b064-0e08a5288442" (UID: "927835ce-92e1-4a29-b064-0e08a5288442"). InnerVolumeSpecName "kube-api-access-7wgf8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:02:04 crc kubenswrapper[4907]: I0313 16:02:04.726707 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7wgf8\" (UniqueName: \"kubernetes.io/projected/927835ce-92e1-4a29-b064-0e08a5288442-kube-api-access-7wgf8\") on node \"crc\" DevicePath \"\"" Mar 13 16:02:05 crc kubenswrapper[4907]: I0313 16:02:05.084539 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556962-d4bk5" event={"ID":"927835ce-92e1-4a29-b064-0e08a5288442","Type":"ContainerDied","Data":"59176c1fbdb0435bb688c727d47f358b21ee89c1dba9385298bbb961b4960ce1"} Mar 13 16:02:05 crc kubenswrapper[4907]: I0313 16:02:05.084578 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="59176c1fbdb0435bb688c727d47f358b21ee89c1dba9385298bbb961b4960ce1" Mar 13 16:02:05 crc kubenswrapper[4907]: I0313 16:02:05.084638 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556962-d4bk5" Mar 13 16:02:05 crc kubenswrapper[4907]: I0313 16:02:05.527560 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556956-txd4m"] Mar 13 16:02:05 crc kubenswrapper[4907]: I0313 16:02:05.537656 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556956-txd4m"] Mar 13 16:02:05 crc kubenswrapper[4907]: I0313 16:02:05.797256 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a22eedb1-70a8-4232-9a54-8625a4dd7022" path="/var/lib/kubelet/pods/a22eedb1-70a8-4232-9a54-8625a4dd7022/volumes" Mar 13 16:02:40 crc kubenswrapper[4907]: I0313 16:02:40.789789 4907 scope.go:117] "RemoveContainer" containerID="76fd8f67a17de0068ab3b4548c472b268fc93309cdb749646d71dbe2fb8aef2e" Mar 13 16:03:18 crc kubenswrapper[4907]: I0313 16:03:18.042496 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 16:03:18 crc kubenswrapper[4907]: I0313 16:03:18.043414 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 16:03:29 crc kubenswrapper[4907]: I0313 16:03:29.038535 4907 generic.go:334] "Generic (PLEG): container finished" podID="02ef8f0f-4ae4-478a-a729-3bde4adade43" containerID="384d76deb615542ee9afab32cbd864c5a2ef0e524a3354232d127d85ac034766" exitCode=0 Mar 13 16:03:29 crc kubenswrapper[4907]: I0313 16:03:29.038646 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-fj6qz" event={"ID":"02ef8f0f-4ae4-478a-a729-3bde4adade43","Type":"ContainerDied","Data":"384d76deb615542ee9afab32cbd864c5a2ef0e524a3354232d127d85ac034766"} Mar 13 16:03:30 crc kubenswrapper[4907]: I0313 16:03:30.708986 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-fj6qz" Mar 13 16:03:30 crc kubenswrapper[4907]: I0313 16:03:30.830688 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/02ef8f0f-4ae4-478a-a729-3bde4adade43-ssh-key-openstack-cell1\") pod \"02ef8f0f-4ae4-478a-a729-3bde4adade43\" (UID: \"02ef8f0f-4ae4-478a-a729-3bde4adade43\") " Mar 13 16:03:30 crc kubenswrapper[4907]: I0313 16:03:30.831252 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gsfkk\" (UniqueName: \"kubernetes.io/projected/02ef8f0f-4ae4-478a-a729-3bde4adade43-kube-api-access-gsfkk\") pod \"02ef8f0f-4ae4-478a-a729-3bde4adade43\" (UID: \"02ef8f0f-4ae4-478a-a729-3bde4adade43\") " Mar 13 16:03:30 crc kubenswrapper[4907]: I0313 16:03:30.831440 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02ef8f0f-4ae4-478a-a729-3bde4adade43-tripleo-cleanup-combined-ca-bundle\") pod \"02ef8f0f-4ae4-478a-a729-3bde4adade43\" (UID: \"02ef8f0f-4ae4-478a-a729-3bde4adade43\") " Mar 13 16:03:30 crc kubenswrapper[4907]: I0313 16:03:30.831570 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/02ef8f0f-4ae4-478a-a729-3bde4adade43-ceph\") pod \"02ef8f0f-4ae4-478a-a729-3bde4adade43\" (UID: \"02ef8f0f-4ae4-478a-a729-3bde4adade43\") " Mar 13 16:03:30 crc kubenswrapper[4907]: I0313 16:03:30.831674 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/02ef8f0f-4ae4-478a-a729-3bde4adade43-inventory\") pod \"02ef8f0f-4ae4-478a-a729-3bde4adade43\" (UID: \"02ef8f0f-4ae4-478a-a729-3bde4adade43\") " Mar 13 16:03:30 crc kubenswrapper[4907]: I0313 16:03:30.837256 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02ef8f0f-4ae4-478a-a729-3bde4adade43-ceph" (OuterVolumeSpecName: "ceph") pod "02ef8f0f-4ae4-478a-a729-3bde4adade43" (UID: "02ef8f0f-4ae4-478a-a729-3bde4adade43"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:03:30 crc kubenswrapper[4907]: I0313 16:03:30.837611 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02ef8f0f-4ae4-478a-a729-3bde4adade43-tripleo-cleanup-combined-ca-bundle" (OuterVolumeSpecName: "tripleo-cleanup-combined-ca-bundle") pod "02ef8f0f-4ae4-478a-a729-3bde4adade43" (UID: "02ef8f0f-4ae4-478a-a729-3bde4adade43"). InnerVolumeSpecName "tripleo-cleanup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:03:30 crc kubenswrapper[4907]: I0313 16:03:30.838740 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/02ef8f0f-4ae4-478a-a729-3bde4adade43-kube-api-access-gsfkk" (OuterVolumeSpecName: "kube-api-access-gsfkk") pod "02ef8f0f-4ae4-478a-a729-3bde4adade43" (UID: "02ef8f0f-4ae4-478a-a729-3bde4adade43"). InnerVolumeSpecName "kube-api-access-gsfkk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:03:30 crc kubenswrapper[4907]: I0313 16:03:30.872304 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02ef8f0f-4ae4-478a-a729-3bde4adade43-inventory" (OuterVolumeSpecName: "inventory") pod "02ef8f0f-4ae4-478a-a729-3bde4adade43" (UID: "02ef8f0f-4ae4-478a-a729-3bde4adade43"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:03:30 crc kubenswrapper[4907]: I0313 16:03:30.872394 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02ef8f0f-4ae4-478a-a729-3bde4adade43-ssh-key-openstack-cell1" (OuterVolumeSpecName: "ssh-key-openstack-cell1") pod "02ef8f0f-4ae4-478a-a729-3bde4adade43" (UID: "02ef8f0f-4ae4-478a-a729-3bde4adade43"). InnerVolumeSpecName "ssh-key-openstack-cell1". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:03:30 crc kubenswrapper[4907]: I0313 16:03:30.935154 4907 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/02ef8f0f-4ae4-478a-a729-3bde4adade43-ssh-key-openstack-cell1\") on node \"crc\" DevicePath \"\"" Mar 13 16:03:30 crc kubenswrapper[4907]: I0313 16:03:30.935199 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gsfkk\" (UniqueName: \"kubernetes.io/projected/02ef8f0f-4ae4-478a-a729-3bde4adade43-kube-api-access-gsfkk\") on node \"crc\" DevicePath \"\"" Mar 13 16:03:30 crc kubenswrapper[4907]: I0313 16:03:30.935220 4907 reconciler_common.go:293] "Volume detached for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02ef8f0f-4ae4-478a-a729-3bde4adade43-tripleo-cleanup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 16:03:30 crc kubenswrapper[4907]: I0313 16:03:30.935239 4907 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/02ef8f0f-4ae4-478a-a729-3bde4adade43-ceph\") on node \"crc\" DevicePath \"\"" Mar 13 16:03:30 crc kubenswrapper[4907]: I0313 16:03:30.935261 4907 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/02ef8f0f-4ae4-478a-a729-3bde4adade43-inventory\") on node \"crc\" DevicePath \"\"" Mar 13 16:03:31 crc kubenswrapper[4907]: I0313 16:03:31.061698 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-fj6qz" event={"ID":"02ef8f0f-4ae4-478a-a729-3bde4adade43","Type":"ContainerDied","Data":"4e306d67a326c5a22924c14c80f0301c152e3eb9bef16aa1cc486b8e33a33cc7"} Mar 13 16:03:31 crc kubenswrapper[4907]: I0313 16:03:31.061947 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4e306d67a326c5a22924c14c80f0301c152e3eb9bef16aa1cc486b8e33a33cc7" Mar 13 16:03:31 crc kubenswrapper[4907]: I0313 16:03:31.061728 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-fj6qz" Mar 13 16:03:36 crc kubenswrapper[4907]: I0313 16:03:36.712957 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-openstack-openstack-cell1-tb9sw"] Mar 13 16:03:36 crc kubenswrapper[4907]: E0313 16:03:36.713840 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02ef8f0f-4ae4-478a-a729-3bde4adade43" containerName="tripleo-cleanup-tripleo-cleanup-openstack-cell1" Mar 13 16:03:36 crc kubenswrapper[4907]: I0313 16:03:36.713853 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="02ef8f0f-4ae4-478a-a729-3bde4adade43" containerName="tripleo-cleanup-tripleo-cleanup-openstack-cell1" Mar 13 16:03:36 crc kubenswrapper[4907]: E0313 16:03:36.713869 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="927835ce-92e1-4a29-b064-0e08a5288442" containerName="oc" Mar 13 16:03:36 crc kubenswrapper[4907]: I0313 16:03:36.713889 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="927835ce-92e1-4a29-b064-0e08a5288442" containerName="oc" Mar 13 16:03:36 crc kubenswrapper[4907]: I0313 16:03:36.714165 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="02ef8f0f-4ae4-478a-a729-3bde4adade43" containerName="tripleo-cleanup-tripleo-cleanup-openstack-cell1" Mar 13 16:03:36 crc kubenswrapper[4907]: I0313 16:03:36.714183 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="927835ce-92e1-4a29-b064-0e08a5288442" containerName="oc" Mar 13 16:03:36 crc kubenswrapper[4907]: I0313 16:03:36.714926 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-cell1-tb9sw" Mar 13 16:03:36 crc kubenswrapper[4907]: I0313 16:03:36.719779 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Mar 13 16:03:36 crc kubenswrapper[4907]: I0313 16:03:36.719796 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-44kxw" Mar 13 16:03:36 crc kubenswrapper[4907]: I0313 16:03:36.721313 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 13 16:03:36 crc kubenswrapper[4907]: I0313 16:03:36.721345 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Mar 13 16:03:36 crc kubenswrapper[4907]: I0313 16:03:36.732379 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-openstack-openstack-cell1-tb9sw"] Mar 13 16:03:36 crc kubenswrapper[4907]: I0313 16:03:36.870120 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d7346ccb-6154-42a9-9097-bad966681d81-ceph\") pod \"bootstrap-openstack-openstack-cell1-tb9sw\" (UID: \"d7346ccb-6154-42a9-9097-bad966681d81\") " pod="openstack/bootstrap-openstack-openstack-cell1-tb9sw" Mar 13 16:03:36 crc kubenswrapper[4907]: I0313 16:03:36.871050 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d7346ccb-6154-42a9-9097-bad966681d81-inventory\") pod \"bootstrap-openstack-openstack-cell1-tb9sw\" (UID: \"d7346ccb-6154-42a9-9097-bad966681d81\") " pod="openstack/bootstrap-openstack-openstack-cell1-tb9sw" Mar 13 16:03:36 crc kubenswrapper[4907]: I0313 16:03:36.871118 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8bpfb\" (UniqueName: \"kubernetes.io/projected/d7346ccb-6154-42a9-9097-bad966681d81-kube-api-access-8bpfb\") pod \"bootstrap-openstack-openstack-cell1-tb9sw\" (UID: \"d7346ccb-6154-42a9-9097-bad966681d81\") " pod="openstack/bootstrap-openstack-openstack-cell1-tb9sw" Mar 13 16:03:36 crc kubenswrapper[4907]: I0313 16:03:36.872118 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7346ccb-6154-42a9-9097-bad966681d81-bootstrap-combined-ca-bundle\") pod \"bootstrap-openstack-openstack-cell1-tb9sw\" (UID: \"d7346ccb-6154-42a9-9097-bad966681d81\") " pod="openstack/bootstrap-openstack-openstack-cell1-tb9sw" Mar 13 16:03:36 crc kubenswrapper[4907]: I0313 16:03:36.872404 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/d7346ccb-6154-42a9-9097-bad966681d81-ssh-key-openstack-cell1\") pod \"bootstrap-openstack-openstack-cell1-tb9sw\" (UID: \"d7346ccb-6154-42a9-9097-bad966681d81\") " pod="openstack/bootstrap-openstack-openstack-cell1-tb9sw" Mar 13 16:03:36 crc kubenswrapper[4907]: I0313 16:03:36.974259 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7346ccb-6154-42a9-9097-bad966681d81-bootstrap-combined-ca-bundle\") pod \"bootstrap-openstack-openstack-cell1-tb9sw\" (UID: \"d7346ccb-6154-42a9-9097-bad966681d81\") " pod="openstack/bootstrap-openstack-openstack-cell1-tb9sw" Mar 13 16:03:36 crc kubenswrapper[4907]: I0313 16:03:36.974389 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/d7346ccb-6154-42a9-9097-bad966681d81-ssh-key-openstack-cell1\") pod \"bootstrap-openstack-openstack-cell1-tb9sw\" (UID: \"d7346ccb-6154-42a9-9097-bad966681d81\") " pod="openstack/bootstrap-openstack-openstack-cell1-tb9sw" Mar 13 16:03:36 crc kubenswrapper[4907]: I0313 16:03:36.974433 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d7346ccb-6154-42a9-9097-bad966681d81-ceph\") pod \"bootstrap-openstack-openstack-cell1-tb9sw\" (UID: \"d7346ccb-6154-42a9-9097-bad966681d81\") " pod="openstack/bootstrap-openstack-openstack-cell1-tb9sw" Mar 13 16:03:36 crc kubenswrapper[4907]: I0313 16:03:36.974471 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d7346ccb-6154-42a9-9097-bad966681d81-inventory\") pod \"bootstrap-openstack-openstack-cell1-tb9sw\" (UID: \"d7346ccb-6154-42a9-9097-bad966681d81\") " pod="openstack/bootstrap-openstack-openstack-cell1-tb9sw" Mar 13 16:03:36 crc kubenswrapper[4907]: I0313 16:03:36.974497 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8bpfb\" (UniqueName: \"kubernetes.io/projected/d7346ccb-6154-42a9-9097-bad966681d81-kube-api-access-8bpfb\") pod \"bootstrap-openstack-openstack-cell1-tb9sw\" (UID: \"d7346ccb-6154-42a9-9097-bad966681d81\") " pod="openstack/bootstrap-openstack-openstack-cell1-tb9sw" Mar 13 16:03:36 crc kubenswrapper[4907]: I0313 16:03:36.980710 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d7346ccb-6154-42a9-9097-bad966681d81-inventory\") pod \"bootstrap-openstack-openstack-cell1-tb9sw\" (UID: \"d7346ccb-6154-42a9-9097-bad966681d81\") " pod="openstack/bootstrap-openstack-openstack-cell1-tb9sw" Mar 13 16:03:36 crc kubenswrapper[4907]: I0313 16:03:36.981384 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/d7346ccb-6154-42a9-9097-bad966681d81-ssh-key-openstack-cell1\") pod \"bootstrap-openstack-openstack-cell1-tb9sw\" (UID: \"d7346ccb-6154-42a9-9097-bad966681d81\") " pod="openstack/bootstrap-openstack-openstack-cell1-tb9sw" Mar 13 16:03:36 crc kubenswrapper[4907]: I0313 16:03:36.982342 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7346ccb-6154-42a9-9097-bad966681d81-bootstrap-combined-ca-bundle\") pod \"bootstrap-openstack-openstack-cell1-tb9sw\" (UID: \"d7346ccb-6154-42a9-9097-bad966681d81\") " pod="openstack/bootstrap-openstack-openstack-cell1-tb9sw" Mar 13 16:03:36 crc kubenswrapper[4907]: I0313 16:03:36.982554 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d7346ccb-6154-42a9-9097-bad966681d81-ceph\") pod \"bootstrap-openstack-openstack-cell1-tb9sw\" (UID: \"d7346ccb-6154-42a9-9097-bad966681d81\") " pod="openstack/bootstrap-openstack-openstack-cell1-tb9sw" Mar 13 16:03:37 crc kubenswrapper[4907]: I0313 16:03:37.002779 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8bpfb\" (UniqueName: \"kubernetes.io/projected/d7346ccb-6154-42a9-9097-bad966681d81-kube-api-access-8bpfb\") pod \"bootstrap-openstack-openstack-cell1-tb9sw\" (UID: \"d7346ccb-6154-42a9-9097-bad966681d81\") " pod="openstack/bootstrap-openstack-openstack-cell1-tb9sw" Mar 13 16:03:37 crc kubenswrapper[4907]: I0313 16:03:37.035677 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-cell1-tb9sw" Mar 13 16:03:37 crc kubenswrapper[4907]: I0313 16:03:37.605183 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-openstack-openstack-cell1-tb9sw"] Mar 13 16:03:38 crc kubenswrapper[4907]: I0313 16:03:38.166074 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-cell1-tb9sw" event={"ID":"d7346ccb-6154-42a9-9097-bad966681d81","Type":"ContainerStarted","Data":"239301abbc130c5a894d3a44690a47910fda2c931b2896b41f7ec97e0db3d777"} Mar 13 16:03:38 crc kubenswrapper[4907]: I0313 16:03:38.596151 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-frdhh"] Mar 13 16:03:38 crc kubenswrapper[4907]: I0313 16:03:38.598791 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-frdhh" Mar 13 16:03:38 crc kubenswrapper[4907]: I0313 16:03:38.608475 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-frdhh"] Mar 13 16:03:38 crc kubenswrapper[4907]: I0313 16:03:38.706992 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8371fb8e-84f3-4e9a-a810-9ab7ed345cbc-catalog-content\") pod \"redhat-marketplace-frdhh\" (UID: \"8371fb8e-84f3-4e9a-a810-9ab7ed345cbc\") " pod="openshift-marketplace/redhat-marketplace-frdhh" Mar 13 16:03:38 crc kubenswrapper[4907]: I0313 16:03:38.707061 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8371fb8e-84f3-4e9a-a810-9ab7ed345cbc-utilities\") pod \"redhat-marketplace-frdhh\" (UID: \"8371fb8e-84f3-4e9a-a810-9ab7ed345cbc\") " pod="openshift-marketplace/redhat-marketplace-frdhh" Mar 13 16:03:38 crc kubenswrapper[4907]: I0313 16:03:38.707275 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g57nf\" (UniqueName: \"kubernetes.io/projected/8371fb8e-84f3-4e9a-a810-9ab7ed345cbc-kube-api-access-g57nf\") pod \"redhat-marketplace-frdhh\" (UID: \"8371fb8e-84f3-4e9a-a810-9ab7ed345cbc\") " pod="openshift-marketplace/redhat-marketplace-frdhh" Mar 13 16:03:38 crc kubenswrapper[4907]: I0313 16:03:38.809249 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8371fb8e-84f3-4e9a-a810-9ab7ed345cbc-catalog-content\") pod \"redhat-marketplace-frdhh\" (UID: \"8371fb8e-84f3-4e9a-a810-9ab7ed345cbc\") " pod="openshift-marketplace/redhat-marketplace-frdhh" Mar 13 16:03:38 crc kubenswrapper[4907]: I0313 16:03:38.809322 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8371fb8e-84f3-4e9a-a810-9ab7ed345cbc-utilities\") pod \"redhat-marketplace-frdhh\" (UID: \"8371fb8e-84f3-4e9a-a810-9ab7ed345cbc\") " pod="openshift-marketplace/redhat-marketplace-frdhh" Mar 13 16:03:38 crc kubenswrapper[4907]: I0313 16:03:38.809384 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g57nf\" (UniqueName: \"kubernetes.io/projected/8371fb8e-84f3-4e9a-a810-9ab7ed345cbc-kube-api-access-g57nf\") pod \"redhat-marketplace-frdhh\" (UID: \"8371fb8e-84f3-4e9a-a810-9ab7ed345cbc\") " pod="openshift-marketplace/redhat-marketplace-frdhh" Mar 13 16:03:38 crc kubenswrapper[4907]: I0313 16:03:38.810084 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8371fb8e-84f3-4e9a-a810-9ab7ed345cbc-catalog-content\") pod \"redhat-marketplace-frdhh\" (UID: \"8371fb8e-84f3-4e9a-a810-9ab7ed345cbc\") " pod="openshift-marketplace/redhat-marketplace-frdhh" Mar 13 16:03:38 crc kubenswrapper[4907]: I0313 16:03:38.810340 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8371fb8e-84f3-4e9a-a810-9ab7ed345cbc-utilities\") pod \"redhat-marketplace-frdhh\" (UID: \"8371fb8e-84f3-4e9a-a810-9ab7ed345cbc\") " pod="openshift-marketplace/redhat-marketplace-frdhh" Mar 13 16:03:38 crc kubenswrapper[4907]: I0313 16:03:38.835166 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g57nf\" (UniqueName: \"kubernetes.io/projected/8371fb8e-84f3-4e9a-a810-9ab7ed345cbc-kube-api-access-g57nf\") pod \"redhat-marketplace-frdhh\" (UID: \"8371fb8e-84f3-4e9a-a810-9ab7ed345cbc\") " pod="openshift-marketplace/redhat-marketplace-frdhh" Mar 13 16:03:38 crc kubenswrapper[4907]: I0313 16:03:38.923035 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-frdhh" Mar 13 16:03:39 crc kubenswrapper[4907]: I0313 16:03:39.176394 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-cell1-tb9sw" event={"ID":"d7346ccb-6154-42a9-9097-bad966681d81","Type":"ContainerStarted","Data":"709051d88f8ea62f2c110585bd6b47f69dfdd1063e53bf548cb2db80a34cddd9"} Mar 13 16:03:39 crc kubenswrapper[4907]: I0313 16:03:39.199042 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-openstack-openstack-cell1-tb9sw" podStartSLOduration=2.799800523 podStartE2EDuration="3.199024336s" podCreationTimestamp="2026-03-13 16:03:36 +0000 UTC" firstStartedPulling="2026-03-13 16:03:37.616083567 +0000 UTC m=+7116.515871256" lastFinishedPulling="2026-03-13 16:03:38.01530736 +0000 UTC m=+7116.915095069" observedRunningTime="2026-03-13 16:03:39.193914256 +0000 UTC m=+7118.093701975" watchObservedRunningTime="2026-03-13 16:03:39.199024336 +0000 UTC m=+7118.098812015" Mar 13 16:03:39 crc kubenswrapper[4907]: I0313 16:03:39.411019 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-frdhh"] Mar 13 16:03:40 crc kubenswrapper[4907]: I0313 16:03:40.188762 4907 generic.go:334] "Generic (PLEG): container finished" podID="8371fb8e-84f3-4e9a-a810-9ab7ed345cbc" containerID="fa32b7f26cf0790f4f10f343dafcb6fe343d4f8cfbb9a8065a1c17955425e202" exitCode=0 Mar 13 16:03:40 crc kubenswrapper[4907]: I0313 16:03:40.189697 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-frdhh" event={"ID":"8371fb8e-84f3-4e9a-a810-9ab7ed345cbc","Type":"ContainerDied","Data":"fa32b7f26cf0790f4f10f343dafcb6fe343d4f8cfbb9a8065a1c17955425e202"} Mar 13 16:03:40 crc kubenswrapper[4907]: I0313 16:03:40.189759 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-frdhh" event={"ID":"8371fb8e-84f3-4e9a-a810-9ab7ed345cbc","Type":"ContainerStarted","Data":"f1c6800484e1fbe1a1b02926ef17c56558da8199ea2caad3fd5b7d3de1859a09"} Mar 13 16:03:43 crc kubenswrapper[4907]: I0313 16:03:43.238930 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-frdhh" event={"ID":"8371fb8e-84f3-4e9a-a810-9ab7ed345cbc","Type":"ContainerStarted","Data":"ec8675a76aec53278dabecd965f5899500f4f28f9bc79e544593e636a482ce77"} Mar 13 16:03:44 crc kubenswrapper[4907]: I0313 16:03:44.248926 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-frdhh" event={"ID":"8371fb8e-84f3-4e9a-a810-9ab7ed345cbc","Type":"ContainerDied","Data":"ec8675a76aec53278dabecd965f5899500f4f28f9bc79e544593e636a482ce77"} Mar 13 16:03:44 crc kubenswrapper[4907]: I0313 16:03:44.248868 4907 generic.go:334] "Generic (PLEG): container finished" podID="8371fb8e-84f3-4e9a-a810-9ab7ed345cbc" containerID="ec8675a76aec53278dabecd965f5899500f4f28f9bc79e544593e636a482ce77" exitCode=0 Mar 13 16:03:45 crc kubenswrapper[4907]: I0313 16:03:45.261459 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-frdhh" event={"ID":"8371fb8e-84f3-4e9a-a810-9ab7ed345cbc","Type":"ContainerStarted","Data":"f0fb1ab8a39a940de2067640271d97e3a6053bce6138d1bb4cd8baad8354c46c"} Mar 13 16:03:45 crc kubenswrapper[4907]: I0313 16:03:45.283190 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-frdhh" podStartSLOduration=2.7292045910000002 podStartE2EDuration="7.283170972s" podCreationTimestamp="2026-03-13 16:03:38 +0000 UTC" firstStartedPulling="2026-03-13 16:03:40.19131186 +0000 UTC m=+7119.091099549" lastFinishedPulling="2026-03-13 16:03:44.745278241 +0000 UTC m=+7123.645065930" observedRunningTime="2026-03-13 16:03:45.278893765 +0000 UTC m=+7124.178681474" watchObservedRunningTime="2026-03-13 16:03:45.283170972 +0000 UTC m=+7124.182958671" Mar 13 16:03:48 crc kubenswrapper[4907]: I0313 16:03:48.042401 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 16:03:48 crc kubenswrapper[4907]: I0313 16:03:48.044995 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 16:03:48 crc kubenswrapper[4907]: I0313 16:03:48.924485 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-frdhh" Mar 13 16:03:48 crc kubenswrapper[4907]: I0313 16:03:48.924797 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-frdhh" Mar 13 16:03:48 crc kubenswrapper[4907]: I0313 16:03:48.973340 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-frdhh" Mar 13 16:03:49 crc kubenswrapper[4907]: I0313 16:03:49.348651 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-frdhh" Mar 13 16:03:52 crc kubenswrapper[4907]: I0313 16:03:52.063731 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-frdhh"] Mar 13 16:03:52 crc kubenswrapper[4907]: I0313 16:03:52.064470 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-frdhh" podUID="8371fb8e-84f3-4e9a-a810-9ab7ed345cbc" containerName="registry-server" containerID="cri-o://f0fb1ab8a39a940de2067640271d97e3a6053bce6138d1bb4cd8baad8354c46c" gracePeriod=2 Mar 13 16:03:52 crc kubenswrapper[4907]: I0313 16:03:52.364640 4907 generic.go:334] "Generic (PLEG): container finished" podID="8371fb8e-84f3-4e9a-a810-9ab7ed345cbc" containerID="f0fb1ab8a39a940de2067640271d97e3a6053bce6138d1bb4cd8baad8354c46c" exitCode=0 Mar 13 16:03:52 crc kubenswrapper[4907]: I0313 16:03:52.364730 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-frdhh" event={"ID":"8371fb8e-84f3-4e9a-a810-9ab7ed345cbc","Type":"ContainerDied","Data":"f0fb1ab8a39a940de2067640271d97e3a6053bce6138d1bb4cd8baad8354c46c"} Mar 13 16:03:52 crc kubenswrapper[4907]: I0313 16:03:52.527502 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-frdhh" Mar 13 16:03:52 crc kubenswrapper[4907]: I0313 16:03:52.603518 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8371fb8e-84f3-4e9a-a810-9ab7ed345cbc-catalog-content\") pod \"8371fb8e-84f3-4e9a-a810-9ab7ed345cbc\" (UID: \"8371fb8e-84f3-4e9a-a810-9ab7ed345cbc\") " Mar 13 16:03:52 crc kubenswrapper[4907]: I0313 16:03:52.603779 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8371fb8e-84f3-4e9a-a810-9ab7ed345cbc-utilities\") pod \"8371fb8e-84f3-4e9a-a810-9ab7ed345cbc\" (UID: \"8371fb8e-84f3-4e9a-a810-9ab7ed345cbc\") " Mar 13 16:03:52 crc kubenswrapper[4907]: I0313 16:03:52.603939 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g57nf\" (UniqueName: \"kubernetes.io/projected/8371fb8e-84f3-4e9a-a810-9ab7ed345cbc-kube-api-access-g57nf\") pod \"8371fb8e-84f3-4e9a-a810-9ab7ed345cbc\" (UID: \"8371fb8e-84f3-4e9a-a810-9ab7ed345cbc\") " Mar 13 16:03:52 crc kubenswrapper[4907]: I0313 16:03:52.604818 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8371fb8e-84f3-4e9a-a810-9ab7ed345cbc-utilities" (OuterVolumeSpecName: "utilities") pod "8371fb8e-84f3-4e9a-a810-9ab7ed345cbc" (UID: "8371fb8e-84f3-4e9a-a810-9ab7ed345cbc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 16:03:52 crc kubenswrapper[4907]: I0313 16:03:52.614135 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8371fb8e-84f3-4e9a-a810-9ab7ed345cbc-kube-api-access-g57nf" (OuterVolumeSpecName: "kube-api-access-g57nf") pod "8371fb8e-84f3-4e9a-a810-9ab7ed345cbc" (UID: "8371fb8e-84f3-4e9a-a810-9ab7ed345cbc"). InnerVolumeSpecName "kube-api-access-g57nf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:03:52 crc kubenswrapper[4907]: I0313 16:03:52.650966 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8371fb8e-84f3-4e9a-a810-9ab7ed345cbc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8371fb8e-84f3-4e9a-a810-9ab7ed345cbc" (UID: "8371fb8e-84f3-4e9a-a810-9ab7ed345cbc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 16:03:52 crc kubenswrapper[4907]: I0313 16:03:52.707272 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g57nf\" (UniqueName: \"kubernetes.io/projected/8371fb8e-84f3-4e9a-a810-9ab7ed345cbc-kube-api-access-g57nf\") on node \"crc\" DevicePath \"\"" Mar 13 16:03:52 crc kubenswrapper[4907]: I0313 16:03:52.707317 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8371fb8e-84f3-4e9a-a810-9ab7ed345cbc-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 16:03:52 crc kubenswrapper[4907]: I0313 16:03:52.707330 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8371fb8e-84f3-4e9a-a810-9ab7ed345cbc-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 16:03:53 crc kubenswrapper[4907]: I0313 16:03:53.377478 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-frdhh" event={"ID":"8371fb8e-84f3-4e9a-a810-9ab7ed345cbc","Type":"ContainerDied","Data":"f1c6800484e1fbe1a1b02926ef17c56558da8199ea2caad3fd5b7d3de1859a09"} Mar 13 16:03:53 crc kubenswrapper[4907]: I0313 16:03:53.377911 4907 scope.go:117] "RemoveContainer" containerID="f0fb1ab8a39a940de2067640271d97e3a6053bce6138d1bb4cd8baad8354c46c" Mar 13 16:03:53 crc kubenswrapper[4907]: I0313 16:03:53.377527 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-frdhh" Mar 13 16:03:53 crc kubenswrapper[4907]: I0313 16:03:53.414322 4907 scope.go:117] "RemoveContainer" containerID="ec8675a76aec53278dabecd965f5899500f4f28f9bc79e544593e636a482ce77" Mar 13 16:03:53 crc kubenswrapper[4907]: I0313 16:03:53.433735 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-frdhh"] Mar 13 16:03:53 crc kubenswrapper[4907]: I0313 16:03:53.441400 4907 scope.go:117] "RemoveContainer" containerID="fa32b7f26cf0790f4f10f343dafcb6fe343d4f8cfbb9a8065a1c17955425e202" Mar 13 16:03:53 crc kubenswrapper[4907]: I0313 16:03:53.447184 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-frdhh"] Mar 13 16:03:53 crc kubenswrapper[4907]: I0313 16:03:53.797340 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8371fb8e-84f3-4e9a-a810-9ab7ed345cbc" path="/var/lib/kubelet/pods/8371fb8e-84f3-4e9a-a810-9ab7ed345cbc/volumes" Mar 13 16:04:00 crc kubenswrapper[4907]: I0313 16:04:00.163954 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556964-gfm8d"] Mar 13 16:04:00 crc kubenswrapper[4907]: E0313 16:04:00.165785 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8371fb8e-84f3-4e9a-a810-9ab7ed345cbc" containerName="registry-server" Mar 13 16:04:00 crc kubenswrapper[4907]: I0313 16:04:00.165822 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="8371fb8e-84f3-4e9a-a810-9ab7ed345cbc" containerName="registry-server" Mar 13 16:04:00 crc kubenswrapper[4907]: E0313 16:04:00.165850 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8371fb8e-84f3-4e9a-a810-9ab7ed345cbc" containerName="extract-utilities" Mar 13 16:04:00 crc kubenswrapper[4907]: I0313 16:04:00.165867 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="8371fb8e-84f3-4e9a-a810-9ab7ed345cbc" containerName="extract-utilities" Mar 13 16:04:00 crc kubenswrapper[4907]: E0313 16:04:00.165946 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8371fb8e-84f3-4e9a-a810-9ab7ed345cbc" containerName="extract-content" Mar 13 16:04:00 crc kubenswrapper[4907]: I0313 16:04:00.165968 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="8371fb8e-84f3-4e9a-a810-9ab7ed345cbc" containerName="extract-content" Mar 13 16:04:00 crc kubenswrapper[4907]: I0313 16:04:00.166472 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="8371fb8e-84f3-4e9a-a810-9ab7ed345cbc" containerName="registry-server" Mar 13 16:04:00 crc kubenswrapper[4907]: I0313 16:04:00.167916 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556964-gfm8d" Mar 13 16:04:00 crc kubenswrapper[4907]: I0313 16:04:00.172334 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 16:04:00 crc kubenswrapper[4907]: I0313 16:04:00.172492 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 16:04:00 crc kubenswrapper[4907]: I0313 16:04:00.172566 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 16:04:00 crc kubenswrapper[4907]: I0313 16:04:00.181611 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556964-gfm8d"] Mar 13 16:04:00 crc kubenswrapper[4907]: I0313 16:04:00.205874 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lqn6p\" (UniqueName: \"kubernetes.io/projected/969a7b07-f302-48b0-b48e-730836ea94a4-kube-api-access-lqn6p\") pod \"auto-csr-approver-29556964-gfm8d\" (UID: \"969a7b07-f302-48b0-b48e-730836ea94a4\") " pod="openshift-infra/auto-csr-approver-29556964-gfm8d" Mar 13 16:04:00 crc kubenswrapper[4907]: I0313 16:04:00.308369 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lqn6p\" (UniqueName: \"kubernetes.io/projected/969a7b07-f302-48b0-b48e-730836ea94a4-kube-api-access-lqn6p\") pod \"auto-csr-approver-29556964-gfm8d\" (UID: \"969a7b07-f302-48b0-b48e-730836ea94a4\") " pod="openshift-infra/auto-csr-approver-29556964-gfm8d" Mar 13 16:04:00 crc kubenswrapper[4907]: I0313 16:04:00.327208 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lqn6p\" (UniqueName: \"kubernetes.io/projected/969a7b07-f302-48b0-b48e-730836ea94a4-kube-api-access-lqn6p\") pod \"auto-csr-approver-29556964-gfm8d\" (UID: \"969a7b07-f302-48b0-b48e-730836ea94a4\") " pod="openshift-infra/auto-csr-approver-29556964-gfm8d" Mar 13 16:04:00 crc kubenswrapper[4907]: I0313 16:04:00.506475 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556964-gfm8d" Mar 13 16:04:01 crc kubenswrapper[4907]: I0313 16:04:01.046074 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556964-gfm8d"] Mar 13 16:04:01 crc kubenswrapper[4907]: W0313 16:04:01.057132 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod969a7b07_f302_48b0_b48e_730836ea94a4.slice/crio-997b3fe277f17cc809d2ed0464d2c10cc99ea1e615bd392ec2c3ba14ec9dd0f7 WatchSource:0}: Error finding container 997b3fe277f17cc809d2ed0464d2c10cc99ea1e615bd392ec2c3ba14ec9dd0f7: Status 404 returned error can't find the container with id 997b3fe277f17cc809d2ed0464d2c10cc99ea1e615bd392ec2c3ba14ec9dd0f7 Mar 13 16:04:01 crc kubenswrapper[4907]: I0313 16:04:01.480228 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556964-gfm8d" event={"ID":"969a7b07-f302-48b0-b48e-730836ea94a4","Type":"ContainerStarted","Data":"997b3fe277f17cc809d2ed0464d2c10cc99ea1e615bd392ec2c3ba14ec9dd0f7"} Mar 13 16:04:03 crc kubenswrapper[4907]: I0313 16:04:03.506293 4907 generic.go:334] "Generic (PLEG): container finished" podID="969a7b07-f302-48b0-b48e-730836ea94a4" containerID="e87bfe7ae00d8363f28ac3bbec93999ba43f3034b744a941e08c8397ba4ba459" exitCode=0 Mar 13 16:04:03 crc kubenswrapper[4907]: I0313 16:04:03.506357 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556964-gfm8d" event={"ID":"969a7b07-f302-48b0-b48e-730836ea94a4","Type":"ContainerDied","Data":"e87bfe7ae00d8363f28ac3bbec93999ba43f3034b744a941e08c8397ba4ba459"} Mar 13 16:04:04 crc kubenswrapper[4907]: I0313 16:04:04.918925 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556964-gfm8d" Mar 13 16:04:05 crc kubenswrapper[4907]: I0313 16:04:05.026833 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lqn6p\" (UniqueName: \"kubernetes.io/projected/969a7b07-f302-48b0-b48e-730836ea94a4-kube-api-access-lqn6p\") pod \"969a7b07-f302-48b0-b48e-730836ea94a4\" (UID: \"969a7b07-f302-48b0-b48e-730836ea94a4\") " Mar 13 16:04:05 crc kubenswrapper[4907]: I0313 16:04:05.033684 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/969a7b07-f302-48b0-b48e-730836ea94a4-kube-api-access-lqn6p" (OuterVolumeSpecName: "kube-api-access-lqn6p") pod "969a7b07-f302-48b0-b48e-730836ea94a4" (UID: "969a7b07-f302-48b0-b48e-730836ea94a4"). InnerVolumeSpecName "kube-api-access-lqn6p". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:04:05 crc kubenswrapper[4907]: I0313 16:04:05.129435 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lqn6p\" (UniqueName: \"kubernetes.io/projected/969a7b07-f302-48b0-b48e-730836ea94a4-kube-api-access-lqn6p\") on node \"crc\" DevicePath \"\"" Mar 13 16:04:05 crc kubenswrapper[4907]: I0313 16:04:05.531179 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556964-gfm8d" event={"ID":"969a7b07-f302-48b0-b48e-730836ea94a4","Type":"ContainerDied","Data":"997b3fe277f17cc809d2ed0464d2c10cc99ea1e615bd392ec2c3ba14ec9dd0f7"} Mar 13 16:04:05 crc kubenswrapper[4907]: I0313 16:04:05.531255 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="997b3fe277f17cc809d2ed0464d2c10cc99ea1e615bd392ec2c3ba14ec9dd0f7" Mar 13 16:04:05 crc kubenswrapper[4907]: I0313 16:04:05.531304 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556964-gfm8d" Mar 13 16:04:05 crc kubenswrapper[4907]: I0313 16:04:05.992205 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556958-dk8dz"] Mar 13 16:04:06 crc kubenswrapper[4907]: I0313 16:04:06.001259 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556958-dk8dz"] Mar 13 16:04:07 crc kubenswrapper[4907]: I0313 16:04:07.793948 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2d38cd8e-5a08-4c56-b35e-88617df1d3a8" path="/var/lib/kubelet/pods/2d38cd8e-5a08-4c56-b35e-88617df1d3a8/volumes" Mar 13 16:04:18 crc kubenswrapper[4907]: I0313 16:04:18.041186 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 16:04:18 crc kubenswrapper[4907]: I0313 16:04:18.041823 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 16:04:18 crc kubenswrapper[4907]: I0313 16:04:18.041911 4907 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" Mar 13 16:04:18 crc kubenswrapper[4907]: I0313 16:04:18.042724 4907 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"97ad1b399368adb5c1fc3c4cf2810fa240ec320625c36e99f246d6171b951964"} pod="openshift-machine-config-operator/machine-config-daemon-hm56j" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 16:04:18 crc kubenswrapper[4907]: I0313 16:04:18.042844 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" containerID="cri-o://97ad1b399368adb5c1fc3c4cf2810fa240ec320625c36e99f246d6171b951964" gracePeriod=600 Mar 13 16:04:18 crc kubenswrapper[4907]: E0313 16:04:18.173978 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:04:18 crc kubenswrapper[4907]: I0313 16:04:18.657452 4907 generic.go:334] "Generic (PLEG): container finished" podID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerID="97ad1b399368adb5c1fc3c4cf2810fa240ec320625c36e99f246d6171b951964" exitCode=0 Mar 13 16:04:18 crc kubenswrapper[4907]: I0313 16:04:18.657580 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerDied","Data":"97ad1b399368adb5c1fc3c4cf2810fa240ec320625c36e99f246d6171b951964"} Mar 13 16:04:18 crc kubenswrapper[4907]: I0313 16:04:18.658135 4907 scope.go:117] "RemoveContainer" containerID="2a3c556bf9f1ef9e4af9dec2c5d043c5ea60bb781b3c04fb7c2299179fb58a19" Mar 13 16:04:18 crc kubenswrapper[4907]: I0313 16:04:18.659014 4907 scope.go:117] "RemoveContainer" containerID="97ad1b399368adb5c1fc3c4cf2810fa240ec320625c36e99f246d6171b951964" Mar 13 16:04:18 crc kubenswrapper[4907]: E0313 16:04:18.659678 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:04:29 crc kubenswrapper[4907]: I0313 16:04:29.782809 4907 scope.go:117] "RemoveContainer" containerID="97ad1b399368adb5c1fc3c4cf2810fa240ec320625c36e99f246d6171b951964" Mar 13 16:04:29 crc kubenswrapper[4907]: E0313 16:04:29.783790 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:04:40 crc kubenswrapper[4907]: I0313 16:04:40.901229 4907 scope.go:117] "RemoveContainer" containerID="d5d998084a23eab3592589afda6539d62cf5a3559ead1f830c6d581d1f9410c7" Mar 13 16:04:42 crc kubenswrapper[4907]: I0313 16:04:42.782577 4907 scope.go:117] "RemoveContainer" containerID="97ad1b399368adb5c1fc3c4cf2810fa240ec320625c36e99f246d6171b951964" Mar 13 16:04:42 crc kubenswrapper[4907]: E0313 16:04:42.784149 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:04:57 crc kubenswrapper[4907]: I0313 16:04:57.782348 4907 scope.go:117] "RemoveContainer" containerID="97ad1b399368adb5c1fc3c4cf2810fa240ec320625c36e99f246d6171b951964" Mar 13 16:04:57 crc kubenswrapper[4907]: E0313 16:04:57.783161 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:05:10 crc kubenswrapper[4907]: I0313 16:05:10.782165 4907 scope.go:117] "RemoveContainer" containerID="97ad1b399368adb5c1fc3c4cf2810fa240ec320625c36e99f246d6171b951964" Mar 13 16:05:10 crc kubenswrapper[4907]: E0313 16:05:10.783216 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:05:23 crc kubenswrapper[4907]: I0313 16:05:23.785366 4907 scope.go:117] "RemoveContainer" containerID="97ad1b399368adb5c1fc3c4cf2810fa240ec320625c36e99f246d6171b951964" Mar 13 16:05:23 crc kubenswrapper[4907]: E0313 16:05:23.786255 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:05:38 crc kubenswrapper[4907]: I0313 16:05:38.782507 4907 scope.go:117] "RemoveContainer" containerID="97ad1b399368adb5c1fc3c4cf2810fa240ec320625c36e99f246d6171b951964" Mar 13 16:05:38 crc kubenswrapper[4907]: E0313 16:05:38.783502 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:05:53 crc kubenswrapper[4907]: I0313 16:05:53.783766 4907 scope.go:117] "RemoveContainer" containerID="97ad1b399368adb5c1fc3c4cf2810fa240ec320625c36e99f246d6171b951964" Mar 13 16:05:53 crc kubenswrapper[4907]: E0313 16:05:53.784712 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:06:00 crc kubenswrapper[4907]: I0313 16:06:00.164764 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556966-drz8k"] Mar 13 16:06:00 crc kubenswrapper[4907]: E0313 16:06:00.166214 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="969a7b07-f302-48b0-b48e-730836ea94a4" containerName="oc" Mar 13 16:06:00 crc kubenswrapper[4907]: I0313 16:06:00.166237 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="969a7b07-f302-48b0-b48e-730836ea94a4" containerName="oc" Mar 13 16:06:00 crc kubenswrapper[4907]: I0313 16:06:00.166701 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="969a7b07-f302-48b0-b48e-730836ea94a4" containerName="oc" Mar 13 16:06:00 crc kubenswrapper[4907]: I0313 16:06:00.168256 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556966-drz8k" Mar 13 16:06:00 crc kubenswrapper[4907]: I0313 16:06:00.170671 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 16:06:00 crc kubenswrapper[4907]: I0313 16:06:00.171081 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 16:06:00 crc kubenswrapper[4907]: I0313 16:06:00.171098 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 16:06:00 crc kubenswrapper[4907]: I0313 16:06:00.178345 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556966-drz8k"] Mar 13 16:06:00 crc kubenswrapper[4907]: I0313 16:06:00.199362 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ttfcm\" (UniqueName: \"kubernetes.io/projected/9e6bd258-54f3-49e8-8b92-d101e6b19872-kube-api-access-ttfcm\") pod \"auto-csr-approver-29556966-drz8k\" (UID: \"9e6bd258-54f3-49e8-8b92-d101e6b19872\") " pod="openshift-infra/auto-csr-approver-29556966-drz8k" Mar 13 16:06:00 crc kubenswrapper[4907]: I0313 16:06:00.301793 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ttfcm\" (UniqueName: \"kubernetes.io/projected/9e6bd258-54f3-49e8-8b92-d101e6b19872-kube-api-access-ttfcm\") pod \"auto-csr-approver-29556966-drz8k\" (UID: \"9e6bd258-54f3-49e8-8b92-d101e6b19872\") " pod="openshift-infra/auto-csr-approver-29556966-drz8k" Mar 13 16:06:00 crc kubenswrapper[4907]: I0313 16:06:00.321411 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ttfcm\" (UniqueName: \"kubernetes.io/projected/9e6bd258-54f3-49e8-8b92-d101e6b19872-kube-api-access-ttfcm\") pod \"auto-csr-approver-29556966-drz8k\" (UID: \"9e6bd258-54f3-49e8-8b92-d101e6b19872\") " pod="openshift-infra/auto-csr-approver-29556966-drz8k" Mar 13 16:06:00 crc kubenswrapper[4907]: I0313 16:06:00.496345 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556966-drz8k" Mar 13 16:06:00 crc kubenswrapper[4907]: W0313 16:06:00.990363 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9e6bd258_54f3_49e8_8b92_d101e6b19872.slice/crio-a293a2a950339887882ecd7a9d1f5edab077c3fee5d7b369f1ab1d20a5402f93 WatchSource:0}: Error finding container a293a2a950339887882ecd7a9d1f5edab077c3fee5d7b369f1ab1d20a5402f93: Status 404 returned error can't find the container with id a293a2a950339887882ecd7a9d1f5edab077c3fee5d7b369f1ab1d20a5402f93 Mar 13 16:06:00 crc kubenswrapper[4907]: I0313 16:06:00.990534 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556966-drz8k"] Mar 13 16:06:00 crc kubenswrapper[4907]: I0313 16:06:00.994008 4907 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 13 16:06:01 crc kubenswrapper[4907]: I0313 16:06:01.847757 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556966-drz8k" event={"ID":"9e6bd258-54f3-49e8-8b92-d101e6b19872","Type":"ContainerStarted","Data":"a293a2a950339887882ecd7a9d1f5edab077c3fee5d7b369f1ab1d20a5402f93"} Mar 13 16:06:02 crc kubenswrapper[4907]: I0313 16:06:02.860700 4907 generic.go:334] "Generic (PLEG): container finished" podID="9e6bd258-54f3-49e8-8b92-d101e6b19872" containerID="fba4ffaa28135dbeabe5c056323b0e99fffec19d264cb5778eaf0cb6c3cc2d40" exitCode=0 Mar 13 16:06:02 crc kubenswrapper[4907]: I0313 16:06:02.860788 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556966-drz8k" event={"ID":"9e6bd258-54f3-49e8-8b92-d101e6b19872","Type":"ContainerDied","Data":"fba4ffaa28135dbeabe5c056323b0e99fffec19d264cb5778eaf0cb6c3cc2d40"} Mar 13 16:06:04 crc kubenswrapper[4907]: I0313 16:06:04.347316 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556966-drz8k" Mar 13 16:06:04 crc kubenswrapper[4907]: I0313 16:06:04.387159 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ttfcm\" (UniqueName: \"kubernetes.io/projected/9e6bd258-54f3-49e8-8b92-d101e6b19872-kube-api-access-ttfcm\") pod \"9e6bd258-54f3-49e8-8b92-d101e6b19872\" (UID: \"9e6bd258-54f3-49e8-8b92-d101e6b19872\") " Mar 13 16:06:04 crc kubenswrapper[4907]: I0313 16:06:04.395138 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9e6bd258-54f3-49e8-8b92-d101e6b19872-kube-api-access-ttfcm" (OuterVolumeSpecName: "kube-api-access-ttfcm") pod "9e6bd258-54f3-49e8-8b92-d101e6b19872" (UID: "9e6bd258-54f3-49e8-8b92-d101e6b19872"). InnerVolumeSpecName "kube-api-access-ttfcm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:06:04 crc kubenswrapper[4907]: I0313 16:06:04.489442 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ttfcm\" (UniqueName: \"kubernetes.io/projected/9e6bd258-54f3-49e8-8b92-d101e6b19872-kube-api-access-ttfcm\") on node \"crc\" DevicePath \"\"" Mar 13 16:06:04 crc kubenswrapper[4907]: I0313 16:06:04.911006 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556966-drz8k" event={"ID":"9e6bd258-54f3-49e8-8b92-d101e6b19872","Type":"ContainerDied","Data":"a293a2a950339887882ecd7a9d1f5edab077c3fee5d7b369f1ab1d20a5402f93"} Mar 13 16:06:04 crc kubenswrapper[4907]: I0313 16:06:04.911077 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a293a2a950339887882ecd7a9d1f5edab077c3fee5d7b369f1ab1d20a5402f93" Mar 13 16:06:04 crc kubenswrapper[4907]: I0313 16:06:04.911178 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556966-drz8k" Mar 13 16:06:05 crc kubenswrapper[4907]: I0313 16:06:05.447327 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556960-8xc84"] Mar 13 16:06:05 crc kubenswrapper[4907]: I0313 16:06:05.455324 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556960-8xc84"] Mar 13 16:06:05 crc kubenswrapper[4907]: I0313 16:06:05.797677 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="acb3140e-a115-443a-a39e-5f180a196b13" path="/var/lib/kubelet/pods/acb3140e-a115-443a-a39e-5f180a196b13/volumes" Mar 13 16:06:08 crc kubenswrapper[4907]: I0313 16:06:08.782508 4907 scope.go:117] "RemoveContainer" containerID="97ad1b399368adb5c1fc3c4cf2810fa240ec320625c36e99f246d6171b951964" Mar 13 16:06:08 crc kubenswrapper[4907]: E0313 16:06:08.783268 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:06:22 crc kubenswrapper[4907]: I0313 16:06:22.782172 4907 scope.go:117] "RemoveContainer" containerID="97ad1b399368adb5c1fc3c4cf2810fa240ec320625c36e99f246d6171b951964" Mar 13 16:06:22 crc kubenswrapper[4907]: E0313 16:06:22.783015 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:06:37 crc kubenswrapper[4907]: I0313 16:06:37.782553 4907 scope.go:117] "RemoveContainer" containerID="97ad1b399368adb5c1fc3c4cf2810fa240ec320625c36e99f246d6171b951964" Mar 13 16:06:37 crc kubenswrapper[4907]: E0313 16:06:37.783581 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:06:41 crc kubenswrapper[4907]: I0313 16:06:41.030460 4907 scope.go:117] "RemoveContainer" containerID="90c713255623c6a0cd4f0904c875e6fe6995c5ac18245db4d00a77bd7e6d5d8c" Mar 13 16:06:51 crc kubenswrapper[4907]: I0313 16:06:51.377567 4907 generic.go:334] "Generic (PLEG): container finished" podID="d7346ccb-6154-42a9-9097-bad966681d81" containerID="709051d88f8ea62f2c110585bd6b47f69dfdd1063e53bf548cb2db80a34cddd9" exitCode=0 Mar 13 16:06:51 crc kubenswrapper[4907]: I0313 16:06:51.377700 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-cell1-tb9sw" event={"ID":"d7346ccb-6154-42a9-9097-bad966681d81","Type":"ContainerDied","Data":"709051d88f8ea62f2c110585bd6b47f69dfdd1063e53bf548cb2db80a34cddd9"} Mar 13 16:06:51 crc kubenswrapper[4907]: I0313 16:06:51.788336 4907 scope.go:117] "RemoveContainer" containerID="97ad1b399368adb5c1fc3c4cf2810fa240ec320625c36e99f246d6171b951964" Mar 13 16:06:51 crc kubenswrapper[4907]: E0313 16:06:51.789817 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:06:52 crc kubenswrapper[4907]: I0313 16:06:52.879459 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-cell1-tb9sw" Mar 13 16:06:52 crc kubenswrapper[4907]: I0313 16:06:52.990541 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7346ccb-6154-42a9-9097-bad966681d81-bootstrap-combined-ca-bundle\") pod \"d7346ccb-6154-42a9-9097-bad966681d81\" (UID: \"d7346ccb-6154-42a9-9097-bad966681d81\") " Mar 13 16:06:52 crc kubenswrapper[4907]: I0313 16:06:52.990611 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8bpfb\" (UniqueName: \"kubernetes.io/projected/d7346ccb-6154-42a9-9097-bad966681d81-kube-api-access-8bpfb\") pod \"d7346ccb-6154-42a9-9097-bad966681d81\" (UID: \"d7346ccb-6154-42a9-9097-bad966681d81\") " Mar 13 16:06:52 crc kubenswrapper[4907]: I0313 16:06:52.990922 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/d7346ccb-6154-42a9-9097-bad966681d81-ssh-key-openstack-cell1\") pod \"d7346ccb-6154-42a9-9097-bad966681d81\" (UID: \"d7346ccb-6154-42a9-9097-bad966681d81\") " Mar 13 16:06:52 crc kubenswrapper[4907]: I0313 16:06:52.991043 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d7346ccb-6154-42a9-9097-bad966681d81-ceph\") pod \"d7346ccb-6154-42a9-9097-bad966681d81\" (UID: \"d7346ccb-6154-42a9-9097-bad966681d81\") " Mar 13 16:06:52 crc kubenswrapper[4907]: I0313 16:06:52.991175 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d7346ccb-6154-42a9-9097-bad966681d81-inventory\") pod \"d7346ccb-6154-42a9-9097-bad966681d81\" (UID: \"d7346ccb-6154-42a9-9097-bad966681d81\") " Mar 13 16:06:52 crc kubenswrapper[4907]: I0313 16:06:52.996500 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7346ccb-6154-42a9-9097-bad966681d81-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "d7346ccb-6154-42a9-9097-bad966681d81" (UID: "d7346ccb-6154-42a9-9097-bad966681d81"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:06:52 crc kubenswrapper[4907]: I0313 16:06:52.998047 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7346ccb-6154-42a9-9097-bad966681d81-ceph" (OuterVolumeSpecName: "ceph") pod "d7346ccb-6154-42a9-9097-bad966681d81" (UID: "d7346ccb-6154-42a9-9097-bad966681d81"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:06:53 crc kubenswrapper[4907]: I0313 16:06:53.001255 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d7346ccb-6154-42a9-9097-bad966681d81-kube-api-access-8bpfb" (OuterVolumeSpecName: "kube-api-access-8bpfb") pod "d7346ccb-6154-42a9-9097-bad966681d81" (UID: "d7346ccb-6154-42a9-9097-bad966681d81"). InnerVolumeSpecName "kube-api-access-8bpfb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:06:53 crc kubenswrapper[4907]: I0313 16:06:53.023287 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7346ccb-6154-42a9-9097-bad966681d81-inventory" (OuterVolumeSpecName: "inventory") pod "d7346ccb-6154-42a9-9097-bad966681d81" (UID: "d7346ccb-6154-42a9-9097-bad966681d81"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:06:53 crc kubenswrapper[4907]: I0313 16:06:53.029706 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7346ccb-6154-42a9-9097-bad966681d81-ssh-key-openstack-cell1" (OuterVolumeSpecName: "ssh-key-openstack-cell1") pod "d7346ccb-6154-42a9-9097-bad966681d81" (UID: "d7346ccb-6154-42a9-9097-bad966681d81"). InnerVolumeSpecName "ssh-key-openstack-cell1". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:06:53 crc kubenswrapper[4907]: I0313 16:06:53.097463 4907 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/d7346ccb-6154-42a9-9097-bad966681d81-ssh-key-openstack-cell1\") on node \"crc\" DevicePath \"\"" Mar 13 16:06:53 crc kubenswrapper[4907]: I0313 16:06:53.097500 4907 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d7346ccb-6154-42a9-9097-bad966681d81-ceph\") on node \"crc\" DevicePath \"\"" Mar 13 16:06:53 crc kubenswrapper[4907]: I0313 16:06:53.097512 4907 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d7346ccb-6154-42a9-9097-bad966681d81-inventory\") on node \"crc\" DevicePath \"\"" Mar 13 16:06:53 crc kubenswrapper[4907]: I0313 16:06:53.097522 4907 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7346ccb-6154-42a9-9097-bad966681d81-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 16:06:53 crc kubenswrapper[4907]: I0313 16:06:53.097532 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8bpfb\" (UniqueName: \"kubernetes.io/projected/d7346ccb-6154-42a9-9097-bad966681d81-kube-api-access-8bpfb\") on node \"crc\" DevicePath \"\"" Mar 13 16:06:53 crc kubenswrapper[4907]: I0313 16:06:53.404013 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-cell1-tb9sw" event={"ID":"d7346ccb-6154-42a9-9097-bad966681d81","Type":"ContainerDied","Data":"239301abbc130c5a894d3a44690a47910fda2c931b2896b41f7ec97e0db3d777"} Mar 13 16:06:53 crc kubenswrapper[4907]: I0313 16:06:53.404060 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-cell1-tb9sw" Mar 13 16:06:53 crc kubenswrapper[4907]: I0313 16:06:53.404081 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="239301abbc130c5a894d3a44690a47910fda2c931b2896b41f7ec97e0db3d777" Mar 13 16:06:53 crc kubenswrapper[4907]: I0313 16:06:53.543967 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-openstack-openstack-cell1-84zm8"] Mar 13 16:06:53 crc kubenswrapper[4907]: E0313 16:06:53.544502 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e6bd258-54f3-49e8-8b92-d101e6b19872" containerName="oc" Mar 13 16:06:53 crc kubenswrapper[4907]: I0313 16:06:53.544523 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e6bd258-54f3-49e8-8b92-d101e6b19872" containerName="oc" Mar 13 16:06:53 crc kubenswrapper[4907]: E0313 16:06:53.544558 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7346ccb-6154-42a9-9097-bad966681d81" containerName="bootstrap-openstack-openstack-cell1" Mar 13 16:06:53 crc kubenswrapper[4907]: I0313 16:06:53.544566 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7346ccb-6154-42a9-9097-bad966681d81" containerName="bootstrap-openstack-openstack-cell1" Mar 13 16:06:53 crc kubenswrapper[4907]: I0313 16:06:53.544842 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7346ccb-6154-42a9-9097-bad966681d81" containerName="bootstrap-openstack-openstack-cell1" Mar 13 16:06:53 crc kubenswrapper[4907]: I0313 16:06:53.544897 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e6bd258-54f3-49e8-8b92-d101e6b19872" containerName="oc" Mar 13 16:06:53 crc kubenswrapper[4907]: I0313 16:06:53.545789 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-cell1-84zm8" Mar 13 16:06:53 crc kubenswrapper[4907]: I0313 16:06:53.551499 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-44kxw" Mar 13 16:06:53 crc kubenswrapper[4907]: I0313 16:06:53.551810 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Mar 13 16:06:53 crc kubenswrapper[4907]: I0313 16:06:53.552041 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 13 16:06:53 crc kubenswrapper[4907]: I0313 16:06:53.552493 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Mar 13 16:06:53 crc kubenswrapper[4907]: I0313 16:06:53.574948 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-openstack-openstack-cell1-84zm8"] Mar 13 16:06:53 crc kubenswrapper[4907]: E0313 16:06:53.580211 4907 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd7346ccb_6154_42a9_9097_bad966681d81.slice/crio-239301abbc130c5a894d3a44690a47910fda2c931b2896b41f7ec97e0db3d777\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd7346ccb_6154_42a9_9097_bad966681d81.slice\": RecentStats: unable to find data in memory cache]" Mar 13 16:06:53 crc kubenswrapper[4907]: I0313 16:06:53.616680 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/441cfe9f-03ed-4a4f-b3f6-58ff77c45cc9-inventory\") pod \"download-cache-openstack-openstack-cell1-84zm8\" (UID: \"441cfe9f-03ed-4a4f-b3f6-58ff77c45cc9\") " pod="openstack/download-cache-openstack-openstack-cell1-84zm8" Mar 13 16:06:53 crc kubenswrapper[4907]: I0313 16:06:53.616786 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/441cfe9f-03ed-4a4f-b3f6-58ff77c45cc9-ceph\") pod \"download-cache-openstack-openstack-cell1-84zm8\" (UID: \"441cfe9f-03ed-4a4f-b3f6-58ff77c45cc9\") " pod="openstack/download-cache-openstack-openstack-cell1-84zm8" Mar 13 16:06:53 crc kubenswrapper[4907]: I0313 16:06:53.616984 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2jvln\" (UniqueName: \"kubernetes.io/projected/441cfe9f-03ed-4a4f-b3f6-58ff77c45cc9-kube-api-access-2jvln\") pod \"download-cache-openstack-openstack-cell1-84zm8\" (UID: \"441cfe9f-03ed-4a4f-b3f6-58ff77c45cc9\") " pod="openstack/download-cache-openstack-openstack-cell1-84zm8" Mar 13 16:06:53 crc kubenswrapper[4907]: I0313 16:06:53.617086 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/441cfe9f-03ed-4a4f-b3f6-58ff77c45cc9-ssh-key-openstack-cell1\") pod \"download-cache-openstack-openstack-cell1-84zm8\" (UID: \"441cfe9f-03ed-4a4f-b3f6-58ff77c45cc9\") " pod="openstack/download-cache-openstack-openstack-cell1-84zm8" Mar 13 16:06:53 crc kubenswrapper[4907]: I0313 16:06:53.719245 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2jvln\" (UniqueName: \"kubernetes.io/projected/441cfe9f-03ed-4a4f-b3f6-58ff77c45cc9-kube-api-access-2jvln\") pod \"download-cache-openstack-openstack-cell1-84zm8\" (UID: \"441cfe9f-03ed-4a4f-b3f6-58ff77c45cc9\") " pod="openstack/download-cache-openstack-openstack-cell1-84zm8" Mar 13 16:06:53 crc kubenswrapper[4907]: I0313 16:06:53.719334 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/441cfe9f-03ed-4a4f-b3f6-58ff77c45cc9-ssh-key-openstack-cell1\") pod \"download-cache-openstack-openstack-cell1-84zm8\" (UID: \"441cfe9f-03ed-4a4f-b3f6-58ff77c45cc9\") " pod="openstack/download-cache-openstack-openstack-cell1-84zm8" Mar 13 16:06:53 crc kubenswrapper[4907]: I0313 16:06:53.719398 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/441cfe9f-03ed-4a4f-b3f6-58ff77c45cc9-inventory\") pod \"download-cache-openstack-openstack-cell1-84zm8\" (UID: \"441cfe9f-03ed-4a4f-b3f6-58ff77c45cc9\") " pod="openstack/download-cache-openstack-openstack-cell1-84zm8" Mar 13 16:06:53 crc kubenswrapper[4907]: I0313 16:06:53.719450 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/441cfe9f-03ed-4a4f-b3f6-58ff77c45cc9-ceph\") pod \"download-cache-openstack-openstack-cell1-84zm8\" (UID: \"441cfe9f-03ed-4a4f-b3f6-58ff77c45cc9\") " pod="openstack/download-cache-openstack-openstack-cell1-84zm8" Mar 13 16:06:53 crc kubenswrapper[4907]: I0313 16:06:53.732668 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/441cfe9f-03ed-4a4f-b3f6-58ff77c45cc9-ceph\") pod \"download-cache-openstack-openstack-cell1-84zm8\" (UID: \"441cfe9f-03ed-4a4f-b3f6-58ff77c45cc9\") " pod="openstack/download-cache-openstack-openstack-cell1-84zm8" Mar 13 16:06:53 crc kubenswrapper[4907]: I0313 16:06:53.732964 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/441cfe9f-03ed-4a4f-b3f6-58ff77c45cc9-ssh-key-openstack-cell1\") pod \"download-cache-openstack-openstack-cell1-84zm8\" (UID: \"441cfe9f-03ed-4a4f-b3f6-58ff77c45cc9\") " pod="openstack/download-cache-openstack-openstack-cell1-84zm8" Mar 13 16:06:53 crc kubenswrapper[4907]: I0313 16:06:53.739462 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/441cfe9f-03ed-4a4f-b3f6-58ff77c45cc9-inventory\") pod \"download-cache-openstack-openstack-cell1-84zm8\" (UID: \"441cfe9f-03ed-4a4f-b3f6-58ff77c45cc9\") " pod="openstack/download-cache-openstack-openstack-cell1-84zm8" Mar 13 16:06:53 crc kubenswrapper[4907]: I0313 16:06:53.755941 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2jvln\" (UniqueName: \"kubernetes.io/projected/441cfe9f-03ed-4a4f-b3f6-58ff77c45cc9-kube-api-access-2jvln\") pod \"download-cache-openstack-openstack-cell1-84zm8\" (UID: \"441cfe9f-03ed-4a4f-b3f6-58ff77c45cc9\") " pod="openstack/download-cache-openstack-openstack-cell1-84zm8" Mar 13 16:06:53 crc kubenswrapper[4907]: I0313 16:06:53.870124 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-cell1-84zm8" Mar 13 16:06:54 crc kubenswrapper[4907]: I0313 16:06:54.437077 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-openstack-openstack-cell1-84zm8"] Mar 13 16:06:55 crc kubenswrapper[4907]: I0313 16:06:55.422614 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-cell1-84zm8" event={"ID":"441cfe9f-03ed-4a4f-b3f6-58ff77c45cc9","Type":"ContainerStarted","Data":"1047ba81465e9116e35201b4b46e027b0e8aa7f144dbbe56c1ec6cbfadc3afa4"} Mar 13 16:06:55 crc kubenswrapper[4907]: I0313 16:06:55.423058 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-cell1-84zm8" event={"ID":"441cfe9f-03ed-4a4f-b3f6-58ff77c45cc9","Type":"ContainerStarted","Data":"a2c0ec2fb6e0a54b73a9abdf461f21aec42d36ae856307b5fc9e10ec95882a6b"} Mar 13 16:06:55 crc kubenswrapper[4907]: I0313 16:06:55.454997 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-openstack-openstack-cell1-84zm8" podStartSLOduration=1.9359263119999999 podStartE2EDuration="2.454978757s" podCreationTimestamp="2026-03-13 16:06:53 +0000 UTC" firstStartedPulling="2026-03-13 16:06:54.447318523 +0000 UTC m=+7313.347106212" lastFinishedPulling="2026-03-13 16:06:54.966370958 +0000 UTC m=+7313.866158657" observedRunningTime="2026-03-13 16:06:55.448566022 +0000 UTC m=+7314.348353741" watchObservedRunningTime="2026-03-13 16:06:55.454978757 +0000 UTC m=+7314.354766446" Mar 13 16:07:06 crc kubenswrapper[4907]: I0313 16:07:06.782748 4907 scope.go:117] "RemoveContainer" containerID="97ad1b399368adb5c1fc3c4cf2810fa240ec320625c36e99f246d6171b951964" Mar 13 16:07:06 crc kubenswrapper[4907]: E0313 16:07:06.783602 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:07:19 crc kubenswrapper[4907]: I0313 16:07:19.783489 4907 scope.go:117] "RemoveContainer" containerID="97ad1b399368adb5c1fc3c4cf2810fa240ec320625c36e99f246d6171b951964" Mar 13 16:07:19 crc kubenswrapper[4907]: E0313 16:07:19.784235 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:07:33 crc kubenswrapper[4907]: I0313 16:07:33.783402 4907 scope.go:117] "RemoveContainer" containerID="97ad1b399368adb5c1fc3c4cf2810fa240ec320625c36e99f246d6171b951964" Mar 13 16:07:33 crc kubenswrapper[4907]: E0313 16:07:33.784308 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:07:44 crc kubenswrapper[4907]: I0313 16:07:44.782058 4907 scope.go:117] "RemoveContainer" containerID="97ad1b399368adb5c1fc3c4cf2810fa240ec320625c36e99f246d6171b951964" Mar 13 16:07:44 crc kubenswrapper[4907]: E0313 16:07:44.782860 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:07:57 crc kubenswrapper[4907]: I0313 16:07:57.784103 4907 scope.go:117] "RemoveContainer" containerID="97ad1b399368adb5c1fc3c4cf2810fa240ec320625c36e99f246d6171b951964" Mar 13 16:07:57 crc kubenswrapper[4907]: E0313 16:07:57.785235 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:08:00 crc kubenswrapper[4907]: I0313 16:08:00.176491 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556968-xbmhk"] Mar 13 16:08:00 crc kubenswrapper[4907]: I0313 16:08:00.179513 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556968-xbmhk" Mar 13 16:08:00 crc kubenswrapper[4907]: I0313 16:08:00.183673 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 16:08:00 crc kubenswrapper[4907]: I0313 16:08:00.184421 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 16:08:00 crc kubenswrapper[4907]: I0313 16:08:00.185359 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 16:08:00 crc kubenswrapper[4907]: I0313 16:08:00.195104 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556968-xbmhk"] Mar 13 16:08:00 crc kubenswrapper[4907]: I0313 16:08:00.260496 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vks5f\" (UniqueName: \"kubernetes.io/projected/cfdfc46b-b215-491c-ab2f-2a108c2c0882-kube-api-access-vks5f\") pod \"auto-csr-approver-29556968-xbmhk\" (UID: \"cfdfc46b-b215-491c-ab2f-2a108c2c0882\") " pod="openshift-infra/auto-csr-approver-29556968-xbmhk" Mar 13 16:08:00 crc kubenswrapper[4907]: I0313 16:08:00.363105 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vks5f\" (UniqueName: \"kubernetes.io/projected/cfdfc46b-b215-491c-ab2f-2a108c2c0882-kube-api-access-vks5f\") pod \"auto-csr-approver-29556968-xbmhk\" (UID: \"cfdfc46b-b215-491c-ab2f-2a108c2c0882\") " pod="openshift-infra/auto-csr-approver-29556968-xbmhk" Mar 13 16:08:00 crc kubenswrapper[4907]: I0313 16:08:00.393204 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vks5f\" (UniqueName: \"kubernetes.io/projected/cfdfc46b-b215-491c-ab2f-2a108c2c0882-kube-api-access-vks5f\") pod \"auto-csr-approver-29556968-xbmhk\" (UID: \"cfdfc46b-b215-491c-ab2f-2a108c2c0882\") " pod="openshift-infra/auto-csr-approver-29556968-xbmhk" Mar 13 16:08:00 crc kubenswrapper[4907]: I0313 16:08:00.514284 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556968-xbmhk" Mar 13 16:08:01 crc kubenswrapper[4907]: I0313 16:08:01.010492 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556968-xbmhk"] Mar 13 16:08:01 crc kubenswrapper[4907]: W0313 16:08:01.018786 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcfdfc46b_b215_491c_ab2f_2a108c2c0882.slice/crio-2c4008c2c653fef7d8f4f2a44b0be2fffd84b61991c523324d0a0429b4aa4490 WatchSource:0}: Error finding container 2c4008c2c653fef7d8f4f2a44b0be2fffd84b61991c523324d0a0429b4aa4490: Status 404 returned error can't find the container with id 2c4008c2c653fef7d8f4f2a44b0be2fffd84b61991c523324d0a0429b4aa4490 Mar 13 16:08:01 crc kubenswrapper[4907]: I0313 16:08:01.125300 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556968-xbmhk" event={"ID":"cfdfc46b-b215-491c-ab2f-2a108c2c0882","Type":"ContainerStarted","Data":"2c4008c2c653fef7d8f4f2a44b0be2fffd84b61991c523324d0a0429b4aa4490"} Mar 13 16:08:03 crc kubenswrapper[4907]: I0313 16:08:03.152090 4907 generic.go:334] "Generic (PLEG): container finished" podID="cfdfc46b-b215-491c-ab2f-2a108c2c0882" containerID="6f4aac0052e053a0aefd7363aac839de6a2ba5c8481dbf090c19d76f2294e95d" exitCode=0 Mar 13 16:08:03 crc kubenswrapper[4907]: I0313 16:08:03.152245 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556968-xbmhk" event={"ID":"cfdfc46b-b215-491c-ab2f-2a108c2c0882","Type":"ContainerDied","Data":"6f4aac0052e053a0aefd7363aac839de6a2ba5c8481dbf090c19d76f2294e95d"} Mar 13 16:08:04 crc kubenswrapper[4907]: I0313 16:08:04.627774 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556968-xbmhk" Mar 13 16:08:04 crc kubenswrapper[4907]: I0313 16:08:04.681313 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vks5f\" (UniqueName: \"kubernetes.io/projected/cfdfc46b-b215-491c-ab2f-2a108c2c0882-kube-api-access-vks5f\") pod \"cfdfc46b-b215-491c-ab2f-2a108c2c0882\" (UID: \"cfdfc46b-b215-491c-ab2f-2a108c2c0882\") " Mar 13 16:08:04 crc kubenswrapper[4907]: I0313 16:08:04.687991 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cfdfc46b-b215-491c-ab2f-2a108c2c0882-kube-api-access-vks5f" (OuterVolumeSpecName: "kube-api-access-vks5f") pod "cfdfc46b-b215-491c-ab2f-2a108c2c0882" (UID: "cfdfc46b-b215-491c-ab2f-2a108c2c0882"). InnerVolumeSpecName "kube-api-access-vks5f". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:08:04 crc kubenswrapper[4907]: I0313 16:08:04.783967 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vks5f\" (UniqueName: \"kubernetes.io/projected/cfdfc46b-b215-491c-ab2f-2a108c2c0882-kube-api-access-vks5f\") on node \"crc\" DevicePath \"\"" Mar 13 16:08:05 crc kubenswrapper[4907]: I0313 16:08:05.176005 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556968-xbmhk" event={"ID":"cfdfc46b-b215-491c-ab2f-2a108c2c0882","Type":"ContainerDied","Data":"2c4008c2c653fef7d8f4f2a44b0be2fffd84b61991c523324d0a0429b4aa4490"} Mar 13 16:08:05 crc kubenswrapper[4907]: I0313 16:08:05.176426 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2c4008c2c653fef7d8f4f2a44b0be2fffd84b61991c523324d0a0429b4aa4490" Mar 13 16:08:05 crc kubenswrapper[4907]: I0313 16:08:05.176049 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556968-xbmhk" Mar 13 16:08:05 crc kubenswrapper[4907]: I0313 16:08:05.704428 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556962-d4bk5"] Mar 13 16:08:05 crc kubenswrapper[4907]: I0313 16:08:05.716412 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556962-d4bk5"] Mar 13 16:08:05 crc kubenswrapper[4907]: I0313 16:08:05.814594 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="927835ce-92e1-4a29-b064-0e08a5288442" path="/var/lib/kubelet/pods/927835ce-92e1-4a29-b064-0e08a5288442/volumes" Mar 13 16:08:11 crc kubenswrapper[4907]: I0313 16:08:11.800651 4907 scope.go:117] "RemoveContainer" containerID="97ad1b399368adb5c1fc3c4cf2810fa240ec320625c36e99f246d6171b951964" Mar 13 16:08:11 crc kubenswrapper[4907]: E0313 16:08:11.801957 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:08:24 crc kubenswrapper[4907]: I0313 16:08:24.782306 4907 scope.go:117] "RemoveContainer" containerID="97ad1b399368adb5c1fc3c4cf2810fa240ec320625c36e99f246d6171b951964" Mar 13 16:08:24 crc kubenswrapper[4907]: E0313 16:08:24.783222 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:08:36 crc kubenswrapper[4907]: I0313 16:08:36.783526 4907 scope.go:117] "RemoveContainer" containerID="97ad1b399368adb5c1fc3c4cf2810fa240ec320625c36e99f246d6171b951964" Mar 13 16:08:36 crc kubenswrapper[4907]: E0313 16:08:36.784287 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:08:41 crc kubenswrapper[4907]: I0313 16:08:41.130166 4907 scope.go:117] "RemoveContainer" containerID="fd8d416a2c11ed375f8397b31e95379985b54fa1a07f700d79a805d7106a6c4e" Mar 13 16:08:50 crc kubenswrapper[4907]: I0313 16:08:50.784173 4907 scope.go:117] "RemoveContainer" containerID="97ad1b399368adb5c1fc3c4cf2810fa240ec320625c36e99f246d6171b951964" Mar 13 16:08:50 crc kubenswrapper[4907]: E0313 16:08:50.784912 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:08:54 crc kubenswrapper[4907]: I0313 16:08:54.882833 4907 generic.go:334] "Generic (PLEG): container finished" podID="441cfe9f-03ed-4a4f-b3f6-58ff77c45cc9" containerID="1047ba81465e9116e35201b4b46e027b0e8aa7f144dbbe56c1ec6cbfadc3afa4" exitCode=0 Mar 13 16:08:54 crc kubenswrapper[4907]: I0313 16:08:54.882917 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-cell1-84zm8" event={"ID":"441cfe9f-03ed-4a4f-b3f6-58ff77c45cc9","Type":"ContainerDied","Data":"1047ba81465e9116e35201b4b46e027b0e8aa7f144dbbe56c1ec6cbfadc3afa4"} Mar 13 16:08:56 crc kubenswrapper[4907]: I0313 16:08:56.323703 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-cell1-84zm8" Mar 13 16:08:56 crc kubenswrapper[4907]: I0313 16:08:56.497654 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2jvln\" (UniqueName: \"kubernetes.io/projected/441cfe9f-03ed-4a4f-b3f6-58ff77c45cc9-kube-api-access-2jvln\") pod \"441cfe9f-03ed-4a4f-b3f6-58ff77c45cc9\" (UID: \"441cfe9f-03ed-4a4f-b3f6-58ff77c45cc9\") " Mar 13 16:08:56 crc kubenswrapper[4907]: I0313 16:08:56.497832 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/441cfe9f-03ed-4a4f-b3f6-58ff77c45cc9-inventory\") pod \"441cfe9f-03ed-4a4f-b3f6-58ff77c45cc9\" (UID: \"441cfe9f-03ed-4a4f-b3f6-58ff77c45cc9\") " Mar 13 16:08:56 crc kubenswrapper[4907]: I0313 16:08:56.498018 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/441cfe9f-03ed-4a4f-b3f6-58ff77c45cc9-ceph\") pod \"441cfe9f-03ed-4a4f-b3f6-58ff77c45cc9\" (UID: \"441cfe9f-03ed-4a4f-b3f6-58ff77c45cc9\") " Mar 13 16:08:56 crc kubenswrapper[4907]: I0313 16:08:56.498204 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/441cfe9f-03ed-4a4f-b3f6-58ff77c45cc9-ssh-key-openstack-cell1\") pod \"441cfe9f-03ed-4a4f-b3f6-58ff77c45cc9\" (UID: \"441cfe9f-03ed-4a4f-b3f6-58ff77c45cc9\") " Mar 13 16:08:56 crc kubenswrapper[4907]: I0313 16:08:56.506559 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/441cfe9f-03ed-4a4f-b3f6-58ff77c45cc9-ceph" (OuterVolumeSpecName: "ceph") pod "441cfe9f-03ed-4a4f-b3f6-58ff77c45cc9" (UID: "441cfe9f-03ed-4a4f-b3f6-58ff77c45cc9"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:08:56 crc kubenswrapper[4907]: I0313 16:08:56.506747 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/441cfe9f-03ed-4a4f-b3f6-58ff77c45cc9-kube-api-access-2jvln" (OuterVolumeSpecName: "kube-api-access-2jvln") pod "441cfe9f-03ed-4a4f-b3f6-58ff77c45cc9" (UID: "441cfe9f-03ed-4a4f-b3f6-58ff77c45cc9"). InnerVolumeSpecName "kube-api-access-2jvln". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:08:56 crc kubenswrapper[4907]: I0313 16:08:56.527294 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/441cfe9f-03ed-4a4f-b3f6-58ff77c45cc9-ssh-key-openstack-cell1" (OuterVolumeSpecName: "ssh-key-openstack-cell1") pod "441cfe9f-03ed-4a4f-b3f6-58ff77c45cc9" (UID: "441cfe9f-03ed-4a4f-b3f6-58ff77c45cc9"). InnerVolumeSpecName "ssh-key-openstack-cell1". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:08:56 crc kubenswrapper[4907]: I0313 16:08:56.532972 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/441cfe9f-03ed-4a4f-b3f6-58ff77c45cc9-inventory" (OuterVolumeSpecName: "inventory") pod "441cfe9f-03ed-4a4f-b3f6-58ff77c45cc9" (UID: "441cfe9f-03ed-4a4f-b3f6-58ff77c45cc9"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:08:56 crc kubenswrapper[4907]: I0313 16:08:56.600824 4907 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/441cfe9f-03ed-4a4f-b3f6-58ff77c45cc9-ceph\") on node \"crc\" DevicePath \"\"" Mar 13 16:08:56 crc kubenswrapper[4907]: I0313 16:08:56.600855 4907 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/441cfe9f-03ed-4a4f-b3f6-58ff77c45cc9-ssh-key-openstack-cell1\") on node \"crc\" DevicePath \"\"" Mar 13 16:08:56 crc kubenswrapper[4907]: I0313 16:08:56.600867 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2jvln\" (UniqueName: \"kubernetes.io/projected/441cfe9f-03ed-4a4f-b3f6-58ff77c45cc9-kube-api-access-2jvln\") on node \"crc\" DevicePath \"\"" Mar 13 16:08:56 crc kubenswrapper[4907]: I0313 16:08:56.600875 4907 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/441cfe9f-03ed-4a4f-b3f6-58ff77c45cc9-inventory\") on node \"crc\" DevicePath \"\"" Mar 13 16:08:56 crc kubenswrapper[4907]: I0313 16:08:56.901762 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-cell1-84zm8" event={"ID":"441cfe9f-03ed-4a4f-b3f6-58ff77c45cc9","Type":"ContainerDied","Data":"a2c0ec2fb6e0a54b73a9abdf461f21aec42d36ae856307b5fc9e10ec95882a6b"} Mar 13 16:08:56 crc kubenswrapper[4907]: I0313 16:08:56.901799 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a2c0ec2fb6e0a54b73a9abdf461f21aec42d36ae856307b5fc9e10ec95882a6b" Mar 13 16:08:56 crc kubenswrapper[4907]: I0313 16:08:56.901806 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-cell1-84zm8" Mar 13 16:08:57 crc kubenswrapper[4907]: I0313 16:08:57.000440 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-openstack-openstack-cell1-wrvtc"] Mar 13 16:08:57 crc kubenswrapper[4907]: E0313 16:08:57.000913 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="441cfe9f-03ed-4a4f-b3f6-58ff77c45cc9" containerName="download-cache-openstack-openstack-cell1" Mar 13 16:08:57 crc kubenswrapper[4907]: I0313 16:08:57.000929 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="441cfe9f-03ed-4a4f-b3f6-58ff77c45cc9" containerName="download-cache-openstack-openstack-cell1" Mar 13 16:08:57 crc kubenswrapper[4907]: E0313 16:08:57.000955 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cfdfc46b-b215-491c-ab2f-2a108c2c0882" containerName="oc" Mar 13 16:08:57 crc kubenswrapper[4907]: I0313 16:08:57.000961 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="cfdfc46b-b215-491c-ab2f-2a108c2c0882" containerName="oc" Mar 13 16:08:57 crc kubenswrapper[4907]: I0313 16:08:57.001164 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="441cfe9f-03ed-4a4f-b3f6-58ff77c45cc9" containerName="download-cache-openstack-openstack-cell1" Mar 13 16:08:57 crc kubenswrapper[4907]: I0313 16:08:57.001201 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="cfdfc46b-b215-491c-ab2f-2a108c2c0882" containerName="oc" Mar 13 16:08:57 crc kubenswrapper[4907]: I0313 16:08:57.001942 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-cell1-wrvtc" Mar 13 16:08:57 crc kubenswrapper[4907]: I0313 16:08:57.006011 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-44kxw" Mar 13 16:08:57 crc kubenswrapper[4907]: I0313 16:08:57.006064 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Mar 13 16:08:57 crc kubenswrapper[4907]: I0313 16:08:57.006730 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 13 16:08:57 crc kubenswrapper[4907]: I0313 16:08:57.008202 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Mar 13 16:08:57 crc kubenswrapper[4907]: I0313 16:08:57.019278 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-openstack-openstack-cell1-wrvtc"] Mar 13 16:08:57 crc kubenswrapper[4907]: I0313 16:08:57.111466 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bfe0612b-2d05-4625-b14e-8b9ea02a8c06-inventory\") pod \"configure-network-openstack-openstack-cell1-wrvtc\" (UID: \"bfe0612b-2d05-4625-b14e-8b9ea02a8c06\") " pod="openstack/configure-network-openstack-openstack-cell1-wrvtc" Mar 13 16:08:57 crc kubenswrapper[4907]: I0313 16:08:57.111794 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/bfe0612b-2d05-4625-b14e-8b9ea02a8c06-ssh-key-openstack-cell1\") pod \"configure-network-openstack-openstack-cell1-wrvtc\" (UID: \"bfe0612b-2d05-4625-b14e-8b9ea02a8c06\") " pod="openstack/configure-network-openstack-openstack-cell1-wrvtc" Mar 13 16:08:57 crc kubenswrapper[4907]: I0313 16:08:57.111903 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gcn6p\" (UniqueName: \"kubernetes.io/projected/bfe0612b-2d05-4625-b14e-8b9ea02a8c06-kube-api-access-gcn6p\") pod \"configure-network-openstack-openstack-cell1-wrvtc\" (UID: \"bfe0612b-2d05-4625-b14e-8b9ea02a8c06\") " pod="openstack/configure-network-openstack-openstack-cell1-wrvtc" Mar 13 16:08:57 crc kubenswrapper[4907]: I0313 16:08:57.112032 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/bfe0612b-2d05-4625-b14e-8b9ea02a8c06-ceph\") pod \"configure-network-openstack-openstack-cell1-wrvtc\" (UID: \"bfe0612b-2d05-4625-b14e-8b9ea02a8c06\") " pod="openstack/configure-network-openstack-openstack-cell1-wrvtc" Mar 13 16:08:57 crc kubenswrapper[4907]: I0313 16:08:57.214397 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gcn6p\" (UniqueName: \"kubernetes.io/projected/bfe0612b-2d05-4625-b14e-8b9ea02a8c06-kube-api-access-gcn6p\") pod \"configure-network-openstack-openstack-cell1-wrvtc\" (UID: \"bfe0612b-2d05-4625-b14e-8b9ea02a8c06\") " pod="openstack/configure-network-openstack-openstack-cell1-wrvtc" Mar 13 16:08:57 crc kubenswrapper[4907]: I0313 16:08:57.214548 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/bfe0612b-2d05-4625-b14e-8b9ea02a8c06-ceph\") pod \"configure-network-openstack-openstack-cell1-wrvtc\" (UID: \"bfe0612b-2d05-4625-b14e-8b9ea02a8c06\") " pod="openstack/configure-network-openstack-openstack-cell1-wrvtc" Mar 13 16:08:57 crc kubenswrapper[4907]: I0313 16:08:57.214669 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bfe0612b-2d05-4625-b14e-8b9ea02a8c06-inventory\") pod \"configure-network-openstack-openstack-cell1-wrvtc\" (UID: \"bfe0612b-2d05-4625-b14e-8b9ea02a8c06\") " pod="openstack/configure-network-openstack-openstack-cell1-wrvtc" Mar 13 16:08:57 crc kubenswrapper[4907]: I0313 16:08:57.214698 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/bfe0612b-2d05-4625-b14e-8b9ea02a8c06-ssh-key-openstack-cell1\") pod \"configure-network-openstack-openstack-cell1-wrvtc\" (UID: \"bfe0612b-2d05-4625-b14e-8b9ea02a8c06\") " pod="openstack/configure-network-openstack-openstack-cell1-wrvtc" Mar 13 16:08:57 crc kubenswrapper[4907]: I0313 16:08:57.220722 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bfe0612b-2d05-4625-b14e-8b9ea02a8c06-inventory\") pod \"configure-network-openstack-openstack-cell1-wrvtc\" (UID: \"bfe0612b-2d05-4625-b14e-8b9ea02a8c06\") " pod="openstack/configure-network-openstack-openstack-cell1-wrvtc" Mar 13 16:08:57 crc kubenswrapper[4907]: I0313 16:08:57.220938 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/bfe0612b-2d05-4625-b14e-8b9ea02a8c06-ceph\") pod \"configure-network-openstack-openstack-cell1-wrvtc\" (UID: \"bfe0612b-2d05-4625-b14e-8b9ea02a8c06\") " pod="openstack/configure-network-openstack-openstack-cell1-wrvtc" Mar 13 16:08:57 crc kubenswrapper[4907]: I0313 16:08:57.226012 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/bfe0612b-2d05-4625-b14e-8b9ea02a8c06-ssh-key-openstack-cell1\") pod \"configure-network-openstack-openstack-cell1-wrvtc\" (UID: \"bfe0612b-2d05-4625-b14e-8b9ea02a8c06\") " pod="openstack/configure-network-openstack-openstack-cell1-wrvtc" Mar 13 16:08:57 crc kubenswrapper[4907]: I0313 16:08:57.232533 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gcn6p\" (UniqueName: \"kubernetes.io/projected/bfe0612b-2d05-4625-b14e-8b9ea02a8c06-kube-api-access-gcn6p\") pod \"configure-network-openstack-openstack-cell1-wrvtc\" (UID: \"bfe0612b-2d05-4625-b14e-8b9ea02a8c06\") " pod="openstack/configure-network-openstack-openstack-cell1-wrvtc" Mar 13 16:08:57 crc kubenswrapper[4907]: I0313 16:08:57.373597 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-cell1-wrvtc" Mar 13 16:08:57 crc kubenswrapper[4907]: I0313 16:08:57.924736 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-openstack-openstack-cell1-wrvtc"] Mar 13 16:08:58 crc kubenswrapper[4907]: I0313 16:08:58.925441 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-cell1-wrvtc" event={"ID":"bfe0612b-2d05-4625-b14e-8b9ea02a8c06","Type":"ContainerStarted","Data":"af77d9f29bb5e409e89eed0618c3878a4fc70669c940d590206ec36dfdf086f6"} Mar 13 16:08:58 crc kubenswrapper[4907]: I0313 16:08:58.925857 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-cell1-wrvtc" event={"ID":"bfe0612b-2d05-4625-b14e-8b9ea02a8c06","Type":"ContainerStarted","Data":"ff5e634a1d2fde12b87f2739aa06c6e49877afb509500e105fbf347bc7380990"} Mar 13 16:08:58 crc kubenswrapper[4907]: I0313 16:08:58.953381 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-openstack-openstack-cell1-wrvtc" podStartSLOduration=2.46910909 podStartE2EDuration="2.953362101s" podCreationTimestamp="2026-03-13 16:08:56 +0000 UTC" firstStartedPulling="2026-03-13 16:08:57.936360451 +0000 UTC m=+7436.836148150" lastFinishedPulling="2026-03-13 16:08:58.420613472 +0000 UTC m=+7437.320401161" observedRunningTime="2026-03-13 16:08:58.94744873 +0000 UTC m=+7437.847236509" watchObservedRunningTime="2026-03-13 16:08:58.953362101 +0000 UTC m=+7437.853149790" Mar 13 16:09:02 crc kubenswrapper[4907]: I0313 16:09:02.783311 4907 scope.go:117] "RemoveContainer" containerID="97ad1b399368adb5c1fc3c4cf2810fa240ec320625c36e99f246d6171b951964" Mar 13 16:09:02 crc kubenswrapper[4907]: E0313 16:09:02.784576 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:09:15 crc kubenswrapper[4907]: I0313 16:09:15.783484 4907 scope.go:117] "RemoveContainer" containerID="97ad1b399368adb5c1fc3c4cf2810fa240ec320625c36e99f246d6171b951964" Mar 13 16:09:15 crc kubenswrapper[4907]: E0313 16:09:15.785987 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:09:29 crc kubenswrapper[4907]: I0313 16:09:29.783786 4907 scope.go:117] "RemoveContainer" containerID="97ad1b399368adb5c1fc3c4cf2810fa240ec320625c36e99f246d6171b951964" Mar 13 16:09:30 crc kubenswrapper[4907]: I0313 16:09:30.270012 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerStarted","Data":"c38b8b19c660dee13d11b404db7a1f6a1a7c1615cf40ea14a2875b4e3648d560"} Mar 13 16:09:54 crc kubenswrapper[4907]: I0313 16:09:54.238818 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-4vfxw"] Mar 13 16:09:54 crc kubenswrapper[4907]: I0313 16:09:54.246547 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4vfxw" Mar 13 16:09:54 crc kubenswrapper[4907]: I0313 16:09:54.253432 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4vfxw"] Mar 13 16:09:54 crc kubenswrapper[4907]: I0313 16:09:54.447794 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3b8b3456-61ff-4e08-9b0b-079a4e8fa42e-catalog-content\") pod \"redhat-operators-4vfxw\" (UID: \"3b8b3456-61ff-4e08-9b0b-079a4e8fa42e\") " pod="openshift-marketplace/redhat-operators-4vfxw" Mar 13 16:09:54 crc kubenswrapper[4907]: I0313 16:09:54.448029 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ck5cx\" (UniqueName: \"kubernetes.io/projected/3b8b3456-61ff-4e08-9b0b-079a4e8fa42e-kube-api-access-ck5cx\") pod \"redhat-operators-4vfxw\" (UID: \"3b8b3456-61ff-4e08-9b0b-079a4e8fa42e\") " pod="openshift-marketplace/redhat-operators-4vfxw" Mar 13 16:09:54 crc kubenswrapper[4907]: I0313 16:09:54.448092 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3b8b3456-61ff-4e08-9b0b-079a4e8fa42e-utilities\") pod \"redhat-operators-4vfxw\" (UID: \"3b8b3456-61ff-4e08-9b0b-079a4e8fa42e\") " pod="openshift-marketplace/redhat-operators-4vfxw" Mar 13 16:09:54 crc kubenswrapper[4907]: I0313 16:09:54.552461 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3b8b3456-61ff-4e08-9b0b-079a4e8fa42e-catalog-content\") pod \"redhat-operators-4vfxw\" (UID: \"3b8b3456-61ff-4e08-9b0b-079a4e8fa42e\") " pod="openshift-marketplace/redhat-operators-4vfxw" Mar 13 16:09:54 crc kubenswrapper[4907]: I0313 16:09:54.552873 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ck5cx\" (UniqueName: \"kubernetes.io/projected/3b8b3456-61ff-4e08-9b0b-079a4e8fa42e-kube-api-access-ck5cx\") pod \"redhat-operators-4vfxw\" (UID: \"3b8b3456-61ff-4e08-9b0b-079a4e8fa42e\") " pod="openshift-marketplace/redhat-operators-4vfxw" Mar 13 16:09:54 crc kubenswrapper[4907]: I0313 16:09:54.553317 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3b8b3456-61ff-4e08-9b0b-079a4e8fa42e-utilities\") pod \"redhat-operators-4vfxw\" (UID: \"3b8b3456-61ff-4e08-9b0b-079a4e8fa42e\") " pod="openshift-marketplace/redhat-operators-4vfxw" Mar 13 16:09:54 crc kubenswrapper[4907]: I0313 16:09:54.555926 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3b8b3456-61ff-4e08-9b0b-079a4e8fa42e-utilities\") pod \"redhat-operators-4vfxw\" (UID: \"3b8b3456-61ff-4e08-9b0b-079a4e8fa42e\") " pod="openshift-marketplace/redhat-operators-4vfxw" Mar 13 16:09:54 crc kubenswrapper[4907]: I0313 16:09:54.556406 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3b8b3456-61ff-4e08-9b0b-079a4e8fa42e-catalog-content\") pod \"redhat-operators-4vfxw\" (UID: \"3b8b3456-61ff-4e08-9b0b-079a4e8fa42e\") " pod="openshift-marketplace/redhat-operators-4vfxw" Mar 13 16:09:54 crc kubenswrapper[4907]: I0313 16:09:54.587698 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ck5cx\" (UniqueName: \"kubernetes.io/projected/3b8b3456-61ff-4e08-9b0b-079a4e8fa42e-kube-api-access-ck5cx\") pod \"redhat-operators-4vfxw\" (UID: \"3b8b3456-61ff-4e08-9b0b-079a4e8fa42e\") " pod="openshift-marketplace/redhat-operators-4vfxw" Mar 13 16:09:54 crc kubenswrapper[4907]: I0313 16:09:54.595271 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4vfxw" Mar 13 16:09:55 crc kubenswrapper[4907]: I0313 16:09:55.085181 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4vfxw"] Mar 13 16:09:55 crc kubenswrapper[4907]: I0313 16:09:55.567652 4907 generic.go:334] "Generic (PLEG): container finished" podID="3b8b3456-61ff-4e08-9b0b-079a4e8fa42e" containerID="47f0591583c6efa846bfde7cb8df75e77e3b66b020180593108050ef792d97f5" exitCode=0 Mar 13 16:09:55 crc kubenswrapper[4907]: I0313 16:09:55.567704 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4vfxw" event={"ID":"3b8b3456-61ff-4e08-9b0b-079a4e8fa42e","Type":"ContainerDied","Data":"47f0591583c6efa846bfde7cb8df75e77e3b66b020180593108050ef792d97f5"} Mar 13 16:09:55 crc kubenswrapper[4907]: I0313 16:09:55.567965 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4vfxw" event={"ID":"3b8b3456-61ff-4e08-9b0b-079a4e8fa42e","Type":"ContainerStarted","Data":"c899082ad7772f673db6c6ad02b7f849f4e3f906cbceceed2c5b27f98250c37d"} Mar 13 16:09:57 crc kubenswrapper[4907]: I0313 16:09:57.590592 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4vfxw" event={"ID":"3b8b3456-61ff-4e08-9b0b-079a4e8fa42e","Type":"ContainerStarted","Data":"01aba326819da9d4d1dbe1d2fabc474c41e125a31756f7caf14aac90810e3e60"} Mar 13 16:09:58 crc kubenswrapper[4907]: I0313 16:09:58.598974 4907 generic.go:334] "Generic (PLEG): container finished" podID="3b8b3456-61ff-4e08-9b0b-079a4e8fa42e" containerID="01aba326819da9d4d1dbe1d2fabc474c41e125a31756f7caf14aac90810e3e60" exitCode=0 Mar 13 16:09:58 crc kubenswrapper[4907]: I0313 16:09:58.600120 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4vfxw" event={"ID":"3b8b3456-61ff-4e08-9b0b-079a4e8fa42e","Type":"ContainerDied","Data":"01aba326819da9d4d1dbe1d2fabc474c41e125a31756f7caf14aac90810e3e60"} Mar 13 16:09:59 crc kubenswrapper[4907]: I0313 16:09:59.614354 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4vfxw" event={"ID":"3b8b3456-61ff-4e08-9b0b-079a4e8fa42e","Type":"ContainerStarted","Data":"422bb86eade127734f05b830061308f31ef3c58377ce0de2a7e534efc87721b8"} Mar 13 16:09:59 crc kubenswrapper[4907]: I0313 16:09:59.645406 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-4vfxw" podStartSLOduration=2.182975002 podStartE2EDuration="5.64538447s" podCreationTimestamp="2026-03-13 16:09:54 +0000 UTC" firstStartedPulling="2026-03-13 16:09:55.571195908 +0000 UTC m=+7494.470983587" lastFinishedPulling="2026-03-13 16:09:59.033605366 +0000 UTC m=+7497.933393055" observedRunningTime="2026-03-13 16:09:59.636365203 +0000 UTC m=+7498.536152962" watchObservedRunningTime="2026-03-13 16:09:59.64538447 +0000 UTC m=+7498.545172169" Mar 13 16:10:00 crc kubenswrapper[4907]: I0313 16:10:00.148039 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556970-lbxp9"] Mar 13 16:10:00 crc kubenswrapper[4907]: I0313 16:10:00.149560 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556970-lbxp9" Mar 13 16:10:00 crc kubenswrapper[4907]: I0313 16:10:00.152567 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 16:10:00 crc kubenswrapper[4907]: I0313 16:10:00.153040 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 16:10:00 crc kubenswrapper[4907]: I0313 16:10:00.154322 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 16:10:00 crc kubenswrapper[4907]: I0313 16:10:00.157963 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556970-lbxp9"] Mar 13 16:10:00 crc kubenswrapper[4907]: I0313 16:10:00.196484 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m4mmw\" (UniqueName: \"kubernetes.io/projected/2b3a466a-0c1c-4c86-897f-0a3375a9fb53-kube-api-access-m4mmw\") pod \"auto-csr-approver-29556970-lbxp9\" (UID: \"2b3a466a-0c1c-4c86-897f-0a3375a9fb53\") " pod="openshift-infra/auto-csr-approver-29556970-lbxp9" Mar 13 16:10:00 crc kubenswrapper[4907]: I0313 16:10:00.297914 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m4mmw\" (UniqueName: \"kubernetes.io/projected/2b3a466a-0c1c-4c86-897f-0a3375a9fb53-kube-api-access-m4mmw\") pod \"auto-csr-approver-29556970-lbxp9\" (UID: \"2b3a466a-0c1c-4c86-897f-0a3375a9fb53\") " pod="openshift-infra/auto-csr-approver-29556970-lbxp9" Mar 13 16:10:00 crc kubenswrapper[4907]: I0313 16:10:00.317623 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m4mmw\" (UniqueName: \"kubernetes.io/projected/2b3a466a-0c1c-4c86-897f-0a3375a9fb53-kube-api-access-m4mmw\") pod \"auto-csr-approver-29556970-lbxp9\" (UID: \"2b3a466a-0c1c-4c86-897f-0a3375a9fb53\") " pod="openshift-infra/auto-csr-approver-29556970-lbxp9" Mar 13 16:10:00 crc kubenswrapper[4907]: I0313 16:10:00.479922 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556970-lbxp9" Mar 13 16:10:00 crc kubenswrapper[4907]: I0313 16:10:00.981326 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556970-lbxp9"] Mar 13 16:10:01 crc kubenswrapper[4907]: I0313 16:10:01.673852 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556970-lbxp9" event={"ID":"2b3a466a-0c1c-4c86-897f-0a3375a9fb53","Type":"ContainerStarted","Data":"6da23a5e86abc141883193c299ad8e19cb1959fa6ac9e21e18427d7c1e087244"} Mar 13 16:10:03 crc kubenswrapper[4907]: I0313 16:10:03.701584 4907 generic.go:334] "Generic (PLEG): container finished" podID="2b3a466a-0c1c-4c86-897f-0a3375a9fb53" containerID="1ffb094832567d23bf593fb1872edfb355e12059011cdaa6ddadd04821d723cc" exitCode=0 Mar 13 16:10:03 crc kubenswrapper[4907]: I0313 16:10:03.701667 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556970-lbxp9" event={"ID":"2b3a466a-0c1c-4c86-897f-0a3375a9fb53","Type":"ContainerDied","Data":"1ffb094832567d23bf593fb1872edfb355e12059011cdaa6ddadd04821d723cc"} Mar 13 16:10:04 crc kubenswrapper[4907]: I0313 16:10:04.595829 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-4vfxw" Mar 13 16:10:04 crc kubenswrapper[4907]: I0313 16:10:04.596163 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-4vfxw" Mar 13 16:10:05 crc kubenswrapper[4907]: I0313 16:10:05.136233 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556970-lbxp9" Mar 13 16:10:05 crc kubenswrapper[4907]: I0313 16:10:05.214690 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m4mmw\" (UniqueName: \"kubernetes.io/projected/2b3a466a-0c1c-4c86-897f-0a3375a9fb53-kube-api-access-m4mmw\") pod \"2b3a466a-0c1c-4c86-897f-0a3375a9fb53\" (UID: \"2b3a466a-0c1c-4c86-897f-0a3375a9fb53\") " Mar 13 16:10:05 crc kubenswrapper[4907]: I0313 16:10:05.223244 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2b3a466a-0c1c-4c86-897f-0a3375a9fb53-kube-api-access-m4mmw" (OuterVolumeSpecName: "kube-api-access-m4mmw") pod "2b3a466a-0c1c-4c86-897f-0a3375a9fb53" (UID: "2b3a466a-0c1c-4c86-897f-0a3375a9fb53"). InnerVolumeSpecName "kube-api-access-m4mmw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:10:05 crc kubenswrapper[4907]: I0313 16:10:05.317627 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m4mmw\" (UniqueName: \"kubernetes.io/projected/2b3a466a-0c1c-4c86-897f-0a3375a9fb53-kube-api-access-m4mmw\") on node \"crc\" DevicePath \"\"" Mar 13 16:10:05 crc kubenswrapper[4907]: I0313 16:10:05.643266 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-4vfxw" podUID="3b8b3456-61ff-4e08-9b0b-079a4e8fa42e" containerName="registry-server" probeResult="failure" output=< Mar 13 16:10:05 crc kubenswrapper[4907]: timeout: failed to connect service ":50051" within 1s Mar 13 16:10:05 crc kubenswrapper[4907]: > Mar 13 16:10:05 crc kubenswrapper[4907]: I0313 16:10:05.721352 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556970-lbxp9" event={"ID":"2b3a466a-0c1c-4c86-897f-0a3375a9fb53","Type":"ContainerDied","Data":"6da23a5e86abc141883193c299ad8e19cb1959fa6ac9e21e18427d7c1e087244"} Mar 13 16:10:05 crc kubenswrapper[4907]: I0313 16:10:05.721402 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6da23a5e86abc141883193c299ad8e19cb1959fa6ac9e21e18427d7c1e087244" Mar 13 16:10:05 crc kubenswrapper[4907]: I0313 16:10:05.721489 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556970-lbxp9" Mar 13 16:10:06 crc kubenswrapper[4907]: I0313 16:10:06.209932 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556964-gfm8d"] Mar 13 16:10:06 crc kubenswrapper[4907]: I0313 16:10:06.221743 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556964-gfm8d"] Mar 13 16:10:07 crc kubenswrapper[4907]: I0313 16:10:07.797521 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="969a7b07-f302-48b0-b48e-730836ea94a4" path="/var/lib/kubelet/pods/969a7b07-f302-48b0-b48e-730836ea94a4/volumes" Mar 13 16:10:14 crc kubenswrapper[4907]: I0313 16:10:14.643443 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-4vfxw" Mar 13 16:10:14 crc kubenswrapper[4907]: I0313 16:10:14.703607 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-4vfxw" Mar 13 16:10:14 crc kubenswrapper[4907]: I0313 16:10:14.879804 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-4vfxw"] Mar 13 16:10:15 crc kubenswrapper[4907]: I0313 16:10:15.823340 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-4vfxw" podUID="3b8b3456-61ff-4e08-9b0b-079a4e8fa42e" containerName="registry-server" containerID="cri-o://422bb86eade127734f05b830061308f31ef3c58377ce0de2a7e534efc87721b8" gracePeriod=2 Mar 13 16:10:16 crc kubenswrapper[4907]: I0313 16:10:16.308513 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4vfxw" Mar 13 16:10:16 crc kubenswrapper[4907]: I0313 16:10:16.403791 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3b8b3456-61ff-4e08-9b0b-079a4e8fa42e-utilities\") pod \"3b8b3456-61ff-4e08-9b0b-079a4e8fa42e\" (UID: \"3b8b3456-61ff-4e08-9b0b-079a4e8fa42e\") " Mar 13 16:10:16 crc kubenswrapper[4907]: I0313 16:10:16.404186 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3b8b3456-61ff-4e08-9b0b-079a4e8fa42e-catalog-content\") pod \"3b8b3456-61ff-4e08-9b0b-079a4e8fa42e\" (UID: \"3b8b3456-61ff-4e08-9b0b-079a4e8fa42e\") " Mar 13 16:10:16 crc kubenswrapper[4907]: I0313 16:10:16.404304 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ck5cx\" (UniqueName: \"kubernetes.io/projected/3b8b3456-61ff-4e08-9b0b-079a4e8fa42e-kube-api-access-ck5cx\") pod \"3b8b3456-61ff-4e08-9b0b-079a4e8fa42e\" (UID: \"3b8b3456-61ff-4e08-9b0b-079a4e8fa42e\") " Mar 13 16:10:16 crc kubenswrapper[4907]: I0313 16:10:16.404708 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3b8b3456-61ff-4e08-9b0b-079a4e8fa42e-utilities" (OuterVolumeSpecName: "utilities") pod "3b8b3456-61ff-4e08-9b0b-079a4e8fa42e" (UID: "3b8b3456-61ff-4e08-9b0b-079a4e8fa42e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 16:10:16 crc kubenswrapper[4907]: I0313 16:10:16.405013 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3b8b3456-61ff-4e08-9b0b-079a4e8fa42e-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 16:10:16 crc kubenswrapper[4907]: I0313 16:10:16.410277 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3b8b3456-61ff-4e08-9b0b-079a4e8fa42e-kube-api-access-ck5cx" (OuterVolumeSpecName: "kube-api-access-ck5cx") pod "3b8b3456-61ff-4e08-9b0b-079a4e8fa42e" (UID: "3b8b3456-61ff-4e08-9b0b-079a4e8fa42e"). InnerVolumeSpecName "kube-api-access-ck5cx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:10:16 crc kubenswrapper[4907]: I0313 16:10:16.506758 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ck5cx\" (UniqueName: \"kubernetes.io/projected/3b8b3456-61ff-4e08-9b0b-079a4e8fa42e-kube-api-access-ck5cx\") on node \"crc\" DevicePath \"\"" Mar 13 16:10:16 crc kubenswrapper[4907]: I0313 16:10:16.538281 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3b8b3456-61ff-4e08-9b0b-079a4e8fa42e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3b8b3456-61ff-4e08-9b0b-079a4e8fa42e" (UID: "3b8b3456-61ff-4e08-9b0b-079a4e8fa42e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 16:10:16 crc kubenswrapper[4907]: I0313 16:10:16.608072 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3b8b3456-61ff-4e08-9b0b-079a4e8fa42e-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 16:10:16 crc kubenswrapper[4907]: I0313 16:10:16.833592 4907 generic.go:334] "Generic (PLEG): container finished" podID="3b8b3456-61ff-4e08-9b0b-079a4e8fa42e" containerID="422bb86eade127734f05b830061308f31ef3c58377ce0de2a7e534efc87721b8" exitCode=0 Mar 13 16:10:16 crc kubenswrapper[4907]: I0313 16:10:16.833651 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4vfxw" event={"ID":"3b8b3456-61ff-4e08-9b0b-079a4e8fa42e","Type":"ContainerDied","Data":"422bb86eade127734f05b830061308f31ef3c58377ce0de2a7e534efc87721b8"} Mar 13 16:10:16 crc kubenswrapper[4907]: I0313 16:10:16.833693 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4vfxw" Mar 13 16:10:16 crc kubenswrapper[4907]: I0313 16:10:16.833715 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4vfxw" event={"ID":"3b8b3456-61ff-4e08-9b0b-079a4e8fa42e","Type":"ContainerDied","Data":"c899082ad7772f673db6c6ad02b7f849f4e3f906cbceceed2c5b27f98250c37d"} Mar 13 16:10:16 crc kubenswrapper[4907]: I0313 16:10:16.833739 4907 scope.go:117] "RemoveContainer" containerID="422bb86eade127734f05b830061308f31ef3c58377ce0de2a7e534efc87721b8" Mar 13 16:10:16 crc kubenswrapper[4907]: I0313 16:10:16.872925 4907 scope.go:117] "RemoveContainer" containerID="01aba326819da9d4d1dbe1d2fabc474c41e125a31756f7caf14aac90810e3e60" Mar 13 16:10:16 crc kubenswrapper[4907]: I0313 16:10:16.876740 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-4vfxw"] Mar 13 16:10:16 crc kubenswrapper[4907]: I0313 16:10:16.892228 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-4vfxw"] Mar 13 16:10:16 crc kubenswrapper[4907]: I0313 16:10:16.910832 4907 scope.go:117] "RemoveContainer" containerID="47f0591583c6efa846bfde7cb8df75e77e3b66b020180593108050ef792d97f5" Mar 13 16:10:16 crc kubenswrapper[4907]: I0313 16:10:16.954292 4907 scope.go:117] "RemoveContainer" containerID="422bb86eade127734f05b830061308f31ef3c58377ce0de2a7e534efc87721b8" Mar 13 16:10:16 crc kubenswrapper[4907]: E0313 16:10:16.954764 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"422bb86eade127734f05b830061308f31ef3c58377ce0de2a7e534efc87721b8\": container with ID starting with 422bb86eade127734f05b830061308f31ef3c58377ce0de2a7e534efc87721b8 not found: ID does not exist" containerID="422bb86eade127734f05b830061308f31ef3c58377ce0de2a7e534efc87721b8" Mar 13 16:10:16 crc kubenswrapper[4907]: I0313 16:10:16.954815 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"422bb86eade127734f05b830061308f31ef3c58377ce0de2a7e534efc87721b8"} err="failed to get container status \"422bb86eade127734f05b830061308f31ef3c58377ce0de2a7e534efc87721b8\": rpc error: code = NotFound desc = could not find container \"422bb86eade127734f05b830061308f31ef3c58377ce0de2a7e534efc87721b8\": container with ID starting with 422bb86eade127734f05b830061308f31ef3c58377ce0de2a7e534efc87721b8 not found: ID does not exist" Mar 13 16:10:16 crc kubenswrapper[4907]: I0313 16:10:16.954846 4907 scope.go:117] "RemoveContainer" containerID="01aba326819da9d4d1dbe1d2fabc474c41e125a31756f7caf14aac90810e3e60" Mar 13 16:10:16 crc kubenswrapper[4907]: E0313 16:10:16.955184 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"01aba326819da9d4d1dbe1d2fabc474c41e125a31756f7caf14aac90810e3e60\": container with ID starting with 01aba326819da9d4d1dbe1d2fabc474c41e125a31756f7caf14aac90810e3e60 not found: ID does not exist" containerID="01aba326819da9d4d1dbe1d2fabc474c41e125a31756f7caf14aac90810e3e60" Mar 13 16:10:16 crc kubenswrapper[4907]: I0313 16:10:16.955231 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"01aba326819da9d4d1dbe1d2fabc474c41e125a31756f7caf14aac90810e3e60"} err="failed to get container status \"01aba326819da9d4d1dbe1d2fabc474c41e125a31756f7caf14aac90810e3e60\": rpc error: code = NotFound desc = could not find container \"01aba326819da9d4d1dbe1d2fabc474c41e125a31756f7caf14aac90810e3e60\": container with ID starting with 01aba326819da9d4d1dbe1d2fabc474c41e125a31756f7caf14aac90810e3e60 not found: ID does not exist" Mar 13 16:10:16 crc kubenswrapper[4907]: I0313 16:10:16.955262 4907 scope.go:117] "RemoveContainer" containerID="47f0591583c6efa846bfde7cb8df75e77e3b66b020180593108050ef792d97f5" Mar 13 16:10:16 crc kubenswrapper[4907]: E0313 16:10:16.955518 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"47f0591583c6efa846bfde7cb8df75e77e3b66b020180593108050ef792d97f5\": container with ID starting with 47f0591583c6efa846bfde7cb8df75e77e3b66b020180593108050ef792d97f5 not found: ID does not exist" containerID="47f0591583c6efa846bfde7cb8df75e77e3b66b020180593108050ef792d97f5" Mar 13 16:10:16 crc kubenswrapper[4907]: I0313 16:10:16.955548 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"47f0591583c6efa846bfde7cb8df75e77e3b66b020180593108050ef792d97f5"} err="failed to get container status \"47f0591583c6efa846bfde7cb8df75e77e3b66b020180593108050ef792d97f5\": rpc error: code = NotFound desc = could not find container \"47f0591583c6efa846bfde7cb8df75e77e3b66b020180593108050ef792d97f5\": container with ID starting with 47f0591583c6efa846bfde7cb8df75e77e3b66b020180593108050ef792d97f5 not found: ID does not exist" Mar 13 16:10:17 crc kubenswrapper[4907]: I0313 16:10:17.804485 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3b8b3456-61ff-4e08-9b0b-079a4e8fa42e" path="/var/lib/kubelet/pods/3b8b3456-61ff-4e08-9b0b-079a4e8fa42e/volumes" Mar 13 16:10:18 crc kubenswrapper[4907]: I0313 16:10:18.860847 4907 generic.go:334] "Generic (PLEG): container finished" podID="bfe0612b-2d05-4625-b14e-8b9ea02a8c06" containerID="af77d9f29bb5e409e89eed0618c3878a4fc70669c940d590206ec36dfdf086f6" exitCode=0 Mar 13 16:10:18 crc kubenswrapper[4907]: I0313 16:10:18.860967 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-cell1-wrvtc" event={"ID":"bfe0612b-2d05-4625-b14e-8b9ea02a8c06","Type":"ContainerDied","Data":"af77d9f29bb5e409e89eed0618c3878a4fc70669c940d590206ec36dfdf086f6"} Mar 13 16:10:20 crc kubenswrapper[4907]: I0313 16:10:20.351908 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-cell1-wrvtc" Mar 13 16:10:20 crc kubenswrapper[4907]: I0313 16:10:20.394207 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/bfe0612b-2d05-4625-b14e-8b9ea02a8c06-ssh-key-openstack-cell1\") pod \"bfe0612b-2d05-4625-b14e-8b9ea02a8c06\" (UID: \"bfe0612b-2d05-4625-b14e-8b9ea02a8c06\") " Mar 13 16:10:20 crc kubenswrapper[4907]: I0313 16:10:20.394308 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bfe0612b-2d05-4625-b14e-8b9ea02a8c06-inventory\") pod \"bfe0612b-2d05-4625-b14e-8b9ea02a8c06\" (UID: \"bfe0612b-2d05-4625-b14e-8b9ea02a8c06\") " Mar 13 16:10:20 crc kubenswrapper[4907]: I0313 16:10:20.394386 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/bfe0612b-2d05-4625-b14e-8b9ea02a8c06-ceph\") pod \"bfe0612b-2d05-4625-b14e-8b9ea02a8c06\" (UID: \"bfe0612b-2d05-4625-b14e-8b9ea02a8c06\") " Mar 13 16:10:20 crc kubenswrapper[4907]: I0313 16:10:20.394446 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gcn6p\" (UniqueName: \"kubernetes.io/projected/bfe0612b-2d05-4625-b14e-8b9ea02a8c06-kube-api-access-gcn6p\") pod \"bfe0612b-2d05-4625-b14e-8b9ea02a8c06\" (UID: \"bfe0612b-2d05-4625-b14e-8b9ea02a8c06\") " Mar 13 16:10:20 crc kubenswrapper[4907]: I0313 16:10:20.400752 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bfe0612b-2d05-4625-b14e-8b9ea02a8c06-kube-api-access-gcn6p" (OuterVolumeSpecName: "kube-api-access-gcn6p") pod "bfe0612b-2d05-4625-b14e-8b9ea02a8c06" (UID: "bfe0612b-2d05-4625-b14e-8b9ea02a8c06"). InnerVolumeSpecName "kube-api-access-gcn6p". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:10:20 crc kubenswrapper[4907]: I0313 16:10:20.400972 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bfe0612b-2d05-4625-b14e-8b9ea02a8c06-ceph" (OuterVolumeSpecName: "ceph") pod "bfe0612b-2d05-4625-b14e-8b9ea02a8c06" (UID: "bfe0612b-2d05-4625-b14e-8b9ea02a8c06"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:10:20 crc kubenswrapper[4907]: I0313 16:10:20.427164 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bfe0612b-2d05-4625-b14e-8b9ea02a8c06-inventory" (OuterVolumeSpecName: "inventory") pod "bfe0612b-2d05-4625-b14e-8b9ea02a8c06" (UID: "bfe0612b-2d05-4625-b14e-8b9ea02a8c06"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:10:20 crc kubenswrapper[4907]: I0313 16:10:20.438629 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bfe0612b-2d05-4625-b14e-8b9ea02a8c06-ssh-key-openstack-cell1" (OuterVolumeSpecName: "ssh-key-openstack-cell1") pod "bfe0612b-2d05-4625-b14e-8b9ea02a8c06" (UID: "bfe0612b-2d05-4625-b14e-8b9ea02a8c06"). InnerVolumeSpecName "ssh-key-openstack-cell1". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:10:20 crc kubenswrapper[4907]: I0313 16:10:20.497074 4907 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/bfe0612b-2d05-4625-b14e-8b9ea02a8c06-ssh-key-openstack-cell1\") on node \"crc\" DevicePath \"\"" Mar 13 16:10:20 crc kubenswrapper[4907]: I0313 16:10:20.497206 4907 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bfe0612b-2d05-4625-b14e-8b9ea02a8c06-inventory\") on node \"crc\" DevicePath \"\"" Mar 13 16:10:20 crc kubenswrapper[4907]: I0313 16:10:20.497290 4907 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/bfe0612b-2d05-4625-b14e-8b9ea02a8c06-ceph\") on node \"crc\" DevicePath \"\"" Mar 13 16:10:20 crc kubenswrapper[4907]: I0313 16:10:20.497350 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gcn6p\" (UniqueName: \"kubernetes.io/projected/bfe0612b-2d05-4625-b14e-8b9ea02a8c06-kube-api-access-gcn6p\") on node \"crc\" DevicePath \"\"" Mar 13 16:10:20 crc kubenswrapper[4907]: I0313 16:10:20.882160 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-cell1-wrvtc" event={"ID":"bfe0612b-2d05-4625-b14e-8b9ea02a8c06","Type":"ContainerDied","Data":"ff5e634a1d2fde12b87f2739aa06c6e49877afb509500e105fbf347bc7380990"} Mar 13 16:10:20 crc kubenswrapper[4907]: I0313 16:10:20.882234 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ff5e634a1d2fde12b87f2739aa06c6e49877afb509500e105fbf347bc7380990" Mar 13 16:10:20 crc kubenswrapper[4907]: I0313 16:10:20.882323 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-cell1-wrvtc" Mar 13 16:10:21 crc kubenswrapper[4907]: I0313 16:10:21.016639 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-openstack-openstack-cell1-rrpt2"] Mar 13 16:10:21 crc kubenswrapper[4907]: E0313 16:10:21.017460 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b8b3456-61ff-4e08-9b0b-079a4e8fa42e" containerName="extract-utilities" Mar 13 16:10:21 crc kubenswrapper[4907]: I0313 16:10:21.017480 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b8b3456-61ff-4e08-9b0b-079a4e8fa42e" containerName="extract-utilities" Mar 13 16:10:21 crc kubenswrapper[4907]: E0313 16:10:21.017516 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b8b3456-61ff-4e08-9b0b-079a4e8fa42e" containerName="registry-server" Mar 13 16:10:21 crc kubenswrapper[4907]: I0313 16:10:21.017524 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b8b3456-61ff-4e08-9b0b-079a4e8fa42e" containerName="registry-server" Mar 13 16:10:21 crc kubenswrapper[4907]: E0313 16:10:21.017558 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b8b3456-61ff-4e08-9b0b-079a4e8fa42e" containerName="extract-content" Mar 13 16:10:21 crc kubenswrapper[4907]: I0313 16:10:21.017566 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b8b3456-61ff-4e08-9b0b-079a4e8fa42e" containerName="extract-content" Mar 13 16:10:21 crc kubenswrapper[4907]: E0313 16:10:21.017584 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b3a466a-0c1c-4c86-897f-0a3375a9fb53" containerName="oc" Mar 13 16:10:21 crc kubenswrapper[4907]: I0313 16:10:21.017593 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b3a466a-0c1c-4c86-897f-0a3375a9fb53" containerName="oc" Mar 13 16:10:21 crc kubenswrapper[4907]: E0313 16:10:21.017606 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bfe0612b-2d05-4625-b14e-8b9ea02a8c06" containerName="configure-network-openstack-openstack-cell1" Mar 13 16:10:21 crc kubenswrapper[4907]: I0313 16:10:21.017614 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="bfe0612b-2d05-4625-b14e-8b9ea02a8c06" containerName="configure-network-openstack-openstack-cell1" Mar 13 16:10:21 crc kubenswrapper[4907]: I0313 16:10:21.017867 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b8b3456-61ff-4e08-9b0b-079a4e8fa42e" containerName="registry-server" Mar 13 16:10:21 crc kubenswrapper[4907]: I0313 16:10:21.017905 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="2b3a466a-0c1c-4c86-897f-0a3375a9fb53" containerName="oc" Mar 13 16:10:21 crc kubenswrapper[4907]: I0313 16:10:21.017919 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="bfe0612b-2d05-4625-b14e-8b9ea02a8c06" containerName="configure-network-openstack-openstack-cell1" Mar 13 16:10:21 crc kubenswrapper[4907]: I0313 16:10:21.018755 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-cell1-rrpt2" Mar 13 16:10:21 crc kubenswrapper[4907]: I0313 16:10:21.021867 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Mar 13 16:10:21 crc kubenswrapper[4907]: I0313 16:10:21.022584 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Mar 13 16:10:21 crc kubenswrapper[4907]: I0313 16:10:21.022849 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 13 16:10:21 crc kubenswrapper[4907]: I0313 16:10:21.023110 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-44kxw" Mar 13 16:10:21 crc kubenswrapper[4907]: I0313 16:10:21.027911 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9a4f6803-5919-4737-acd0-5b25e12a7c55-inventory\") pod \"validate-network-openstack-openstack-cell1-rrpt2\" (UID: \"9a4f6803-5919-4737-acd0-5b25e12a7c55\") " pod="openstack/validate-network-openstack-openstack-cell1-rrpt2" Mar 13 16:10:21 crc kubenswrapper[4907]: I0313 16:10:21.027963 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9a4f6803-5919-4737-acd0-5b25e12a7c55-ceph\") pod \"validate-network-openstack-openstack-cell1-rrpt2\" (UID: \"9a4f6803-5919-4737-acd0-5b25e12a7c55\") " pod="openstack/validate-network-openstack-openstack-cell1-rrpt2" Mar 13 16:10:21 crc kubenswrapper[4907]: I0313 16:10:21.028008 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/9a4f6803-5919-4737-acd0-5b25e12a7c55-ssh-key-openstack-cell1\") pod \"validate-network-openstack-openstack-cell1-rrpt2\" (UID: \"9a4f6803-5919-4737-acd0-5b25e12a7c55\") " pod="openstack/validate-network-openstack-openstack-cell1-rrpt2" Mar 13 16:10:21 crc kubenswrapper[4907]: I0313 16:10:21.028075 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nb44t\" (UniqueName: \"kubernetes.io/projected/9a4f6803-5919-4737-acd0-5b25e12a7c55-kube-api-access-nb44t\") pod \"validate-network-openstack-openstack-cell1-rrpt2\" (UID: \"9a4f6803-5919-4737-acd0-5b25e12a7c55\") " pod="openstack/validate-network-openstack-openstack-cell1-rrpt2" Mar 13 16:10:21 crc kubenswrapper[4907]: I0313 16:10:21.029303 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-openstack-openstack-cell1-rrpt2"] Mar 13 16:10:21 crc kubenswrapper[4907]: I0313 16:10:21.129482 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/9a4f6803-5919-4737-acd0-5b25e12a7c55-ssh-key-openstack-cell1\") pod \"validate-network-openstack-openstack-cell1-rrpt2\" (UID: \"9a4f6803-5919-4737-acd0-5b25e12a7c55\") " pod="openstack/validate-network-openstack-openstack-cell1-rrpt2" Mar 13 16:10:21 crc kubenswrapper[4907]: I0313 16:10:21.129574 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nb44t\" (UniqueName: \"kubernetes.io/projected/9a4f6803-5919-4737-acd0-5b25e12a7c55-kube-api-access-nb44t\") pod \"validate-network-openstack-openstack-cell1-rrpt2\" (UID: \"9a4f6803-5919-4737-acd0-5b25e12a7c55\") " pod="openstack/validate-network-openstack-openstack-cell1-rrpt2" Mar 13 16:10:21 crc kubenswrapper[4907]: I0313 16:10:21.130137 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9a4f6803-5919-4737-acd0-5b25e12a7c55-inventory\") pod \"validate-network-openstack-openstack-cell1-rrpt2\" (UID: \"9a4f6803-5919-4737-acd0-5b25e12a7c55\") " pod="openstack/validate-network-openstack-openstack-cell1-rrpt2" Mar 13 16:10:21 crc kubenswrapper[4907]: I0313 16:10:21.130257 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9a4f6803-5919-4737-acd0-5b25e12a7c55-ceph\") pod \"validate-network-openstack-openstack-cell1-rrpt2\" (UID: \"9a4f6803-5919-4737-acd0-5b25e12a7c55\") " pod="openstack/validate-network-openstack-openstack-cell1-rrpt2" Mar 13 16:10:21 crc kubenswrapper[4907]: I0313 16:10:21.135037 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/9a4f6803-5919-4737-acd0-5b25e12a7c55-ssh-key-openstack-cell1\") pod \"validate-network-openstack-openstack-cell1-rrpt2\" (UID: \"9a4f6803-5919-4737-acd0-5b25e12a7c55\") " pod="openstack/validate-network-openstack-openstack-cell1-rrpt2" Mar 13 16:10:21 crc kubenswrapper[4907]: I0313 16:10:21.135579 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9a4f6803-5919-4737-acd0-5b25e12a7c55-inventory\") pod \"validate-network-openstack-openstack-cell1-rrpt2\" (UID: \"9a4f6803-5919-4737-acd0-5b25e12a7c55\") " pod="openstack/validate-network-openstack-openstack-cell1-rrpt2" Mar 13 16:10:21 crc kubenswrapper[4907]: I0313 16:10:21.136073 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9a4f6803-5919-4737-acd0-5b25e12a7c55-ceph\") pod \"validate-network-openstack-openstack-cell1-rrpt2\" (UID: \"9a4f6803-5919-4737-acd0-5b25e12a7c55\") " pod="openstack/validate-network-openstack-openstack-cell1-rrpt2" Mar 13 16:10:21 crc kubenswrapper[4907]: I0313 16:10:21.146847 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nb44t\" (UniqueName: \"kubernetes.io/projected/9a4f6803-5919-4737-acd0-5b25e12a7c55-kube-api-access-nb44t\") pod \"validate-network-openstack-openstack-cell1-rrpt2\" (UID: \"9a4f6803-5919-4737-acd0-5b25e12a7c55\") " pod="openstack/validate-network-openstack-openstack-cell1-rrpt2" Mar 13 16:10:21 crc kubenswrapper[4907]: I0313 16:10:21.337135 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-cell1-rrpt2" Mar 13 16:10:21 crc kubenswrapper[4907]: I0313 16:10:21.931736 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-openstack-openstack-cell1-rrpt2"] Mar 13 16:10:21 crc kubenswrapper[4907]: W0313 16:10:21.941038 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9a4f6803_5919_4737_acd0_5b25e12a7c55.slice/crio-f6512817fa65e9fa0d8596c44a85970e3319de57228fe970260d0992907856d6 WatchSource:0}: Error finding container f6512817fa65e9fa0d8596c44a85970e3319de57228fe970260d0992907856d6: Status 404 returned error can't find the container with id f6512817fa65e9fa0d8596c44a85970e3319de57228fe970260d0992907856d6 Mar 13 16:10:22 crc kubenswrapper[4907]: I0313 16:10:22.909562 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-cell1-rrpt2" event={"ID":"9a4f6803-5919-4737-acd0-5b25e12a7c55","Type":"ContainerStarted","Data":"f6512817fa65e9fa0d8596c44a85970e3319de57228fe970260d0992907856d6"} Mar 13 16:10:23 crc kubenswrapper[4907]: I0313 16:10:23.919239 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-cell1-rrpt2" event={"ID":"9a4f6803-5919-4737-acd0-5b25e12a7c55","Type":"ContainerStarted","Data":"3982867b847d25c43ce67d6447d7efe49307a957e163fd43946487b24dfbdbe8"} Mar 13 16:10:23 crc kubenswrapper[4907]: I0313 16:10:23.942728 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-openstack-openstack-cell1-rrpt2" podStartSLOduration=3.241210455 podStartE2EDuration="3.942708365s" podCreationTimestamp="2026-03-13 16:10:20 +0000 UTC" firstStartedPulling="2026-03-13 16:10:21.942910181 +0000 UTC m=+7520.842697860" lastFinishedPulling="2026-03-13 16:10:22.644408071 +0000 UTC m=+7521.544195770" observedRunningTime="2026-03-13 16:10:23.933970046 +0000 UTC m=+7522.833757735" watchObservedRunningTime="2026-03-13 16:10:23.942708365 +0000 UTC m=+7522.842496054" Mar 13 16:10:28 crc kubenswrapper[4907]: I0313 16:10:28.967153 4907 generic.go:334] "Generic (PLEG): container finished" podID="9a4f6803-5919-4737-acd0-5b25e12a7c55" containerID="3982867b847d25c43ce67d6447d7efe49307a957e163fd43946487b24dfbdbe8" exitCode=0 Mar 13 16:10:28 crc kubenswrapper[4907]: I0313 16:10:28.967264 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-cell1-rrpt2" event={"ID":"9a4f6803-5919-4737-acd0-5b25e12a7c55","Type":"ContainerDied","Data":"3982867b847d25c43ce67d6447d7efe49307a957e163fd43946487b24dfbdbe8"} Mar 13 16:10:30 crc kubenswrapper[4907]: I0313 16:10:30.442475 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-cell1-rrpt2" Mar 13 16:10:30 crc kubenswrapper[4907]: I0313 16:10:30.551436 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/9a4f6803-5919-4737-acd0-5b25e12a7c55-ssh-key-openstack-cell1\") pod \"9a4f6803-5919-4737-acd0-5b25e12a7c55\" (UID: \"9a4f6803-5919-4737-acd0-5b25e12a7c55\") " Mar 13 16:10:30 crc kubenswrapper[4907]: I0313 16:10:30.551510 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9a4f6803-5919-4737-acd0-5b25e12a7c55-inventory\") pod \"9a4f6803-5919-4737-acd0-5b25e12a7c55\" (UID: \"9a4f6803-5919-4737-acd0-5b25e12a7c55\") " Mar 13 16:10:30 crc kubenswrapper[4907]: I0313 16:10:30.551624 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9a4f6803-5919-4737-acd0-5b25e12a7c55-ceph\") pod \"9a4f6803-5919-4737-acd0-5b25e12a7c55\" (UID: \"9a4f6803-5919-4737-acd0-5b25e12a7c55\") " Mar 13 16:10:30 crc kubenswrapper[4907]: I0313 16:10:30.551666 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nb44t\" (UniqueName: \"kubernetes.io/projected/9a4f6803-5919-4737-acd0-5b25e12a7c55-kube-api-access-nb44t\") pod \"9a4f6803-5919-4737-acd0-5b25e12a7c55\" (UID: \"9a4f6803-5919-4737-acd0-5b25e12a7c55\") " Mar 13 16:10:30 crc kubenswrapper[4907]: I0313 16:10:30.556390 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9a4f6803-5919-4737-acd0-5b25e12a7c55-kube-api-access-nb44t" (OuterVolumeSpecName: "kube-api-access-nb44t") pod "9a4f6803-5919-4737-acd0-5b25e12a7c55" (UID: "9a4f6803-5919-4737-acd0-5b25e12a7c55"). InnerVolumeSpecName "kube-api-access-nb44t". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:10:30 crc kubenswrapper[4907]: I0313 16:10:30.558482 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a4f6803-5919-4737-acd0-5b25e12a7c55-ceph" (OuterVolumeSpecName: "ceph") pod "9a4f6803-5919-4737-acd0-5b25e12a7c55" (UID: "9a4f6803-5919-4737-acd0-5b25e12a7c55"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:10:30 crc kubenswrapper[4907]: I0313 16:10:30.586286 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a4f6803-5919-4737-acd0-5b25e12a7c55-inventory" (OuterVolumeSpecName: "inventory") pod "9a4f6803-5919-4737-acd0-5b25e12a7c55" (UID: "9a4f6803-5919-4737-acd0-5b25e12a7c55"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:10:30 crc kubenswrapper[4907]: I0313 16:10:30.591020 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a4f6803-5919-4737-acd0-5b25e12a7c55-ssh-key-openstack-cell1" (OuterVolumeSpecName: "ssh-key-openstack-cell1") pod "9a4f6803-5919-4737-acd0-5b25e12a7c55" (UID: "9a4f6803-5919-4737-acd0-5b25e12a7c55"). InnerVolumeSpecName "ssh-key-openstack-cell1". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:10:30 crc kubenswrapper[4907]: I0313 16:10:30.654272 4907 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/9a4f6803-5919-4737-acd0-5b25e12a7c55-ssh-key-openstack-cell1\") on node \"crc\" DevicePath \"\"" Mar 13 16:10:30 crc kubenswrapper[4907]: I0313 16:10:30.654306 4907 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9a4f6803-5919-4737-acd0-5b25e12a7c55-inventory\") on node \"crc\" DevicePath \"\"" Mar 13 16:10:30 crc kubenswrapper[4907]: I0313 16:10:30.654316 4907 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9a4f6803-5919-4737-acd0-5b25e12a7c55-ceph\") on node \"crc\" DevicePath \"\"" Mar 13 16:10:30 crc kubenswrapper[4907]: I0313 16:10:30.654325 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nb44t\" (UniqueName: \"kubernetes.io/projected/9a4f6803-5919-4737-acd0-5b25e12a7c55-kube-api-access-nb44t\") on node \"crc\" DevicePath \"\"" Mar 13 16:10:31 crc kubenswrapper[4907]: I0313 16:10:31.024391 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-cell1-rrpt2" event={"ID":"9a4f6803-5919-4737-acd0-5b25e12a7c55","Type":"ContainerDied","Data":"f6512817fa65e9fa0d8596c44a85970e3319de57228fe970260d0992907856d6"} Mar 13 16:10:31 crc kubenswrapper[4907]: I0313 16:10:31.024435 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f6512817fa65e9fa0d8596c44a85970e3319de57228fe970260d0992907856d6" Mar 13 16:10:31 crc kubenswrapper[4907]: I0313 16:10:31.024465 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-cell1-rrpt2" Mar 13 16:10:31 crc kubenswrapper[4907]: I0313 16:10:31.099575 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-openstack-openstack-cell1-vw9h5"] Mar 13 16:10:31 crc kubenswrapper[4907]: E0313 16:10:31.100090 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a4f6803-5919-4737-acd0-5b25e12a7c55" containerName="validate-network-openstack-openstack-cell1" Mar 13 16:10:31 crc kubenswrapper[4907]: I0313 16:10:31.100111 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a4f6803-5919-4737-acd0-5b25e12a7c55" containerName="validate-network-openstack-openstack-cell1" Mar 13 16:10:31 crc kubenswrapper[4907]: I0313 16:10:31.100311 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a4f6803-5919-4737-acd0-5b25e12a7c55" containerName="validate-network-openstack-openstack-cell1" Mar 13 16:10:31 crc kubenswrapper[4907]: I0313 16:10:31.101072 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-cell1-vw9h5" Mar 13 16:10:31 crc kubenswrapper[4907]: I0313 16:10:31.105799 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Mar 13 16:10:31 crc kubenswrapper[4907]: I0313 16:10:31.105794 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 13 16:10:31 crc kubenswrapper[4907]: I0313 16:10:31.106282 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-44kxw" Mar 13 16:10:31 crc kubenswrapper[4907]: I0313 16:10:31.106401 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Mar 13 16:10:31 crc kubenswrapper[4907]: I0313 16:10:31.156996 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-openstack-openstack-cell1-vw9h5"] Mar 13 16:10:31 crc kubenswrapper[4907]: I0313 16:10:31.267999 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p22zf\" (UniqueName: \"kubernetes.io/projected/cc30c8d7-5f7e-47c8-b89f-6a638b886769-kube-api-access-p22zf\") pod \"install-os-openstack-openstack-cell1-vw9h5\" (UID: \"cc30c8d7-5f7e-47c8-b89f-6a638b886769\") " pod="openstack/install-os-openstack-openstack-cell1-vw9h5" Mar 13 16:10:31 crc kubenswrapper[4907]: I0313 16:10:31.268392 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cc30c8d7-5f7e-47c8-b89f-6a638b886769-inventory\") pod \"install-os-openstack-openstack-cell1-vw9h5\" (UID: \"cc30c8d7-5f7e-47c8-b89f-6a638b886769\") " pod="openstack/install-os-openstack-openstack-cell1-vw9h5" Mar 13 16:10:31 crc kubenswrapper[4907]: I0313 16:10:31.268497 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/cc30c8d7-5f7e-47c8-b89f-6a638b886769-ssh-key-openstack-cell1\") pod \"install-os-openstack-openstack-cell1-vw9h5\" (UID: \"cc30c8d7-5f7e-47c8-b89f-6a638b886769\") " pod="openstack/install-os-openstack-openstack-cell1-vw9h5" Mar 13 16:10:31 crc kubenswrapper[4907]: I0313 16:10:31.268634 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/cc30c8d7-5f7e-47c8-b89f-6a638b886769-ceph\") pod \"install-os-openstack-openstack-cell1-vw9h5\" (UID: \"cc30c8d7-5f7e-47c8-b89f-6a638b886769\") " pod="openstack/install-os-openstack-openstack-cell1-vw9h5" Mar 13 16:10:31 crc kubenswrapper[4907]: I0313 16:10:31.370899 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cc30c8d7-5f7e-47c8-b89f-6a638b886769-inventory\") pod \"install-os-openstack-openstack-cell1-vw9h5\" (UID: \"cc30c8d7-5f7e-47c8-b89f-6a638b886769\") " pod="openstack/install-os-openstack-openstack-cell1-vw9h5" Mar 13 16:10:31 crc kubenswrapper[4907]: I0313 16:10:31.371448 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/cc30c8d7-5f7e-47c8-b89f-6a638b886769-ssh-key-openstack-cell1\") pod \"install-os-openstack-openstack-cell1-vw9h5\" (UID: \"cc30c8d7-5f7e-47c8-b89f-6a638b886769\") " pod="openstack/install-os-openstack-openstack-cell1-vw9h5" Mar 13 16:10:31 crc kubenswrapper[4907]: I0313 16:10:31.371515 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/cc30c8d7-5f7e-47c8-b89f-6a638b886769-ceph\") pod \"install-os-openstack-openstack-cell1-vw9h5\" (UID: \"cc30c8d7-5f7e-47c8-b89f-6a638b886769\") " pod="openstack/install-os-openstack-openstack-cell1-vw9h5" Mar 13 16:10:31 crc kubenswrapper[4907]: I0313 16:10:31.371563 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p22zf\" (UniqueName: \"kubernetes.io/projected/cc30c8d7-5f7e-47c8-b89f-6a638b886769-kube-api-access-p22zf\") pod \"install-os-openstack-openstack-cell1-vw9h5\" (UID: \"cc30c8d7-5f7e-47c8-b89f-6a638b886769\") " pod="openstack/install-os-openstack-openstack-cell1-vw9h5" Mar 13 16:10:31 crc kubenswrapper[4907]: I0313 16:10:31.383932 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/cc30c8d7-5f7e-47c8-b89f-6a638b886769-ssh-key-openstack-cell1\") pod \"install-os-openstack-openstack-cell1-vw9h5\" (UID: \"cc30c8d7-5f7e-47c8-b89f-6a638b886769\") " pod="openstack/install-os-openstack-openstack-cell1-vw9h5" Mar 13 16:10:31 crc kubenswrapper[4907]: I0313 16:10:31.384292 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cc30c8d7-5f7e-47c8-b89f-6a638b886769-inventory\") pod \"install-os-openstack-openstack-cell1-vw9h5\" (UID: \"cc30c8d7-5f7e-47c8-b89f-6a638b886769\") " pod="openstack/install-os-openstack-openstack-cell1-vw9h5" Mar 13 16:10:31 crc kubenswrapper[4907]: I0313 16:10:31.384825 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/cc30c8d7-5f7e-47c8-b89f-6a638b886769-ceph\") pod \"install-os-openstack-openstack-cell1-vw9h5\" (UID: \"cc30c8d7-5f7e-47c8-b89f-6a638b886769\") " pod="openstack/install-os-openstack-openstack-cell1-vw9h5" Mar 13 16:10:31 crc kubenswrapper[4907]: I0313 16:10:31.396786 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p22zf\" (UniqueName: \"kubernetes.io/projected/cc30c8d7-5f7e-47c8-b89f-6a638b886769-kube-api-access-p22zf\") pod \"install-os-openstack-openstack-cell1-vw9h5\" (UID: \"cc30c8d7-5f7e-47c8-b89f-6a638b886769\") " pod="openstack/install-os-openstack-openstack-cell1-vw9h5" Mar 13 16:10:31 crc kubenswrapper[4907]: I0313 16:10:31.424836 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-cell1-vw9h5" Mar 13 16:10:31 crc kubenswrapper[4907]: I0313 16:10:31.999097 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-openstack-openstack-cell1-vw9h5"] Mar 13 16:10:32 crc kubenswrapper[4907]: I0313 16:10:32.052137 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-cell1-vw9h5" event={"ID":"cc30c8d7-5f7e-47c8-b89f-6a638b886769","Type":"ContainerStarted","Data":"4eeb9f6355cdd98a6c410784295be2591cbbb0e5ae4d88b6b06b43662075461d"} Mar 13 16:10:33 crc kubenswrapper[4907]: I0313 16:10:33.062776 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-cell1-vw9h5" event={"ID":"cc30c8d7-5f7e-47c8-b89f-6a638b886769","Type":"ContainerStarted","Data":"86fb48771b3fd8ed723bd1ee960c478c1f6aca7779807b8ba2adcc20af416927"} Mar 13 16:10:33 crc kubenswrapper[4907]: I0313 16:10:33.086950 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-openstack-openstack-cell1-vw9h5" podStartSLOduration=1.601636751 podStartE2EDuration="2.086932172s" podCreationTimestamp="2026-03-13 16:10:31 +0000 UTC" firstStartedPulling="2026-03-13 16:10:32.005521816 +0000 UTC m=+7530.905309495" lastFinishedPulling="2026-03-13 16:10:32.490817237 +0000 UTC m=+7531.390604916" observedRunningTime="2026-03-13 16:10:33.079742435 +0000 UTC m=+7531.979530134" watchObservedRunningTime="2026-03-13 16:10:33.086932172 +0000 UTC m=+7531.986719861" Mar 13 16:10:41 crc kubenswrapper[4907]: I0313 16:10:41.281575 4907 scope.go:117] "RemoveContainer" containerID="e87bfe7ae00d8363f28ac3bbec93999ba43f3034b744a941e08c8397ba4ba459" Mar 13 16:11:19 crc kubenswrapper[4907]: I0313 16:11:19.565108 4907 generic.go:334] "Generic (PLEG): container finished" podID="cc30c8d7-5f7e-47c8-b89f-6a638b886769" containerID="86fb48771b3fd8ed723bd1ee960c478c1f6aca7779807b8ba2adcc20af416927" exitCode=0 Mar 13 16:11:19 crc kubenswrapper[4907]: I0313 16:11:19.565181 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-cell1-vw9h5" event={"ID":"cc30c8d7-5f7e-47c8-b89f-6a638b886769","Type":"ContainerDied","Data":"86fb48771b3fd8ed723bd1ee960c478c1f6aca7779807b8ba2adcc20af416927"} Mar 13 16:11:21 crc kubenswrapper[4907]: I0313 16:11:21.115157 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-cell1-vw9h5" Mar 13 16:11:21 crc kubenswrapper[4907]: I0313 16:11:21.219797 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p22zf\" (UniqueName: \"kubernetes.io/projected/cc30c8d7-5f7e-47c8-b89f-6a638b886769-kube-api-access-p22zf\") pod \"cc30c8d7-5f7e-47c8-b89f-6a638b886769\" (UID: \"cc30c8d7-5f7e-47c8-b89f-6a638b886769\") " Mar 13 16:11:21 crc kubenswrapper[4907]: I0313 16:11:21.219845 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cc30c8d7-5f7e-47c8-b89f-6a638b886769-inventory\") pod \"cc30c8d7-5f7e-47c8-b89f-6a638b886769\" (UID: \"cc30c8d7-5f7e-47c8-b89f-6a638b886769\") " Mar 13 16:11:21 crc kubenswrapper[4907]: I0313 16:11:21.219958 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/cc30c8d7-5f7e-47c8-b89f-6a638b886769-ceph\") pod \"cc30c8d7-5f7e-47c8-b89f-6a638b886769\" (UID: \"cc30c8d7-5f7e-47c8-b89f-6a638b886769\") " Mar 13 16:11:21 crc kubenswrapper[4907]: I0313 16:11:21.219994 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/cc30c8d7-5f7e-47c8-b89f-6a638b886769-ssh-key-openstack-cell1\") pod \"cc30c8d7-5f7e-47c8-b89f-6a638b886769\" (UID: \"cc30c8d7-5f7e-47c8-b89f-6a638b886769\") " Mar 13 16:11:21 crc kubenswrapper[4907]: I0313 16:11:21.228073 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc30c8d7-5f7e-47c8-b89f-6a638b886769-kube-api-access-p22zf" (OuterVolumeSpecName: "kube-api-access-p22zf") pod "cc30c8d7-5f7e-47c8-b89f-6a638b886769" (UID: "cc30c8d7-5f7e-47c8-b89f-6a638b886769"). InnerVolumeSpecName "kube-api-access-p22zf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:11:21 crc kubenswrapper[4907]: I0313 16:11:21.234566 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc30c8d7-5f7e-47c8-b89f-6a638b886769-ceph" (OuterVolumeSpecName: "ceph") pod "cc30c8d7-5f7e-47c8-b89f-6a638b886769" (UID: "cc30c8d7-5f7e-47c8-b89f-6a638b886769"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:11:21 crc kubenswrapper[4907]: I0313 16:11:21.251334 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc30c8d7-5f7e-47c8-b89f-6a638b886769-ssh-key-openstack-cell1" (OuterVolumeSpecName: "ssh-key-openstack-cell1") pod "cc30c8d7-5f7e-47c8-b89f-6a638b886769" (UID: "cc30c8d7-5f7e-47c8-b89f-6a638b886769"). InnerVolumeSpecName "ssh-key-openstack-cell1". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:11:21 crc kubenswrapper[4907]: I0313 16:11:21.275968 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc30c8d7-5f7e-47c8-b89f-6a638b886769-inventory" (OuterVolumeSpecName: "inventory") pod "cc30c8d7-5f7e-47c8-b89f-6a638b886769" (UID: "cc30c8d7-5f7e-47c8-b89f-6a638b886769"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:11:21 crc kubenswrapper[4907]: I0313 16:11:21.323558 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p22zf\" (UniqueName: \"kubernetes.io/projected/cc30c8d7-5f7e-47c8-b89f-6a638b886769-kube-api-access-p22zf\") on node \"crc\" DevicePath \"\"" Mar 13 16:11:21 crc kubenswrapper[4907]: I0313 16:11:21.323621 4907 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cc30c8d7-5f7e-47c8-b89f-6a638b886769-inventory\") on node \"crc\" DevicePath \"\"" Mar 13 16:11:21 crc kubenswrapper[4907]: I0313 16:11:21.323640 4907 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/cc30c8d7-5f7e-47c8-b89f-6a638b886769-ceph\") on node \"crc\" DevicePath \"\"" Mar 13 16:11:21 crc kubenswrapper[4907]: I0313 16:11:21.323659 4907 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/cc30c8d7-5f7e-47c8-b89f-6a638b886769-ssh-key-openstack-cell1\") on node \"crc\" DevicePath \"\"" Mar 13 16:11:21 crc kubenswrapper[4907]: I0313 16:11:21.588390 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-cell1-vw9h5" event={"ID":"cc30c8d7-5f7e-47c8-b89f-6a638b886769","Type":"ContainerDied","Data":"4eeb9f6355cdd98a6c410784295be2591cbbb0e5ae4d88b6b06b43662075461d"} Mar 13 16:11:21 crc kubenswrapper[4907]: I0313 16:11:21.588748 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4eeb9f6355cdd98a6c410784295be2591cbbb0e5ae4d88b6b06b43662075461d" Mar 13 16:11:21 crc kubenswrapper[4907]: I0313 16:11:21.588542 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-cell1-vw9h5" Mar 13 16:11:21 crc kubenswrapper[4907]: I0313 16:11:21.690756 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-openstack-openstack-cell1-s9llf"] Mar 13 16:11:21 crc kubenswrapper[4907]: E0313 16:11:21.691703 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc30c8d7-5f7e-47c8-b89f-6a638b886769" containerName="install-os-openstack-openstack-cell1" Mar 13 16:11:21 crc kubenswrapper[4907]: I0313 16:11:21.691825 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc30c8d7-5f7e-47c8-b89f-6a638b886769" containerName="install-os-openstack-openstack-cell1" Mar 13 16:11:21 crc kubenswrapper[4907]: I0313 16:11:21.692265 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc30c8d7-5f7e-47c8-b89f-6a638b886769" containerName="install-os-openstack-openstack-cell1" Mar 13 16:11:21 crc kubenswrapper[4907]: I0313 16:11:21.693617 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-cell1-s9llf" Mar 13 16:11:21 crc kubenswrapper[4907]: I0313 16:11:21.696755 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 13 16:11:21 crc kubenswrapper[4907]: I0313 16:11:21.697753 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Mar 13 16:11:21 crc kubenswrapper[4907]: I0313 16:11:21.698151 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-44kxw" Mar 13 16:11:21 crc kubenswrapper[4907]: I0313 16:11:21.702847 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-openstack-openstack-cell1-s9llf"] Mar 13 16:11:21 crc kubenswrapper[4907]: I0313 16:11:21.707503 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Mar 13 16:11:21 crc kubenswrapper[4907]: I0313 16:11:21.834771 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/173f7de2-ef0c-4750-a8b3-23e078ab9bde-inventory\") pod \"configure-os-openstack-openstack-cell1-s9llf\" (UID: \"173f7de2-ef0c-4750-a8b3-23e078ab9bde\") " pod="openstack/configure-os-openstack-openstack-cell1-s9llf" Mar 13 16:11:21 crc kubenswrapper[4907]: I0313 16:11:21.835052 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/173f7de2-ef0c-4750-a8b3-23e078ab9bde-ceph\") pod \"configure-os-openstack-openstack-cell1-s9llf\" (UID: \"173f7de2-ef0c-4750-a8b3-23e078ab9bde\") " pod="openstack/configure-os-openstack-openstack-cell1-s9llf" Mar 13 16:11:21 crc kubenswrapper[4907]: I0313 16:11:21.835094 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/173f7de2-ef0c-4750-a8b3-23e078ab9bde-ssh-key-openstack-cell1\") pod \"configure-os-openstack-openstack-cell1-s9llf\" (UID: \"173f7de2-ef0c-4750-a8b3-23e078ab9bde\") " pod="openstack/configure-os-openstack-openstack-cell1-s9llf" Mar 13 16:11:21 crc kubenswrapper[4907]: I0313 16:11:21.835370 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mt98f\" (UniqueName: \"kubernetes.io/projected/173f7de2-ef0c-4750-a8b3-23e078ab9bde-kube-api-access-mt98f\") pod \"configure-os-openstack-openstack-cell1-s9llf\" (UID: \"173f7de2-ef0c-4750-a8b3-23e078ab9bde\") " pod="openstack/configure-os-openstack-openstack-cell1-s9llf" Mar 13 16:11:21 crc kubenswrapper[4907]: I0313 16:11:21.938077 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/173f7de2-ef0c-4750-a8b3-23e078ab9bde-inventory\") pod \"configure-os-openstack-openstack-cell1-s9llf\" (UID: \"173f7de2-ef0c-4750-a8b3-23e078ab9bde\") " pod="openstack/configure-os-openstack-openstack-cell1-s9llf" Mar 13 16:11:21 crc kubenswrapper[4907]: I0313 16:11:21.938209 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/173f7de2-ef0c-4750-a8b3-23e078ab9bde-ceph\") pod \"configure-os-openstack-openstack-cell1-s9llf\" (UID: \"173f7de2-ef0c-4750-a8b3-23e078ab9bde\") " pod="openstack/configure-os-openstack-openstack-cell1-s9llf" Mar 13 16:11:21 crc kubenswrapper[4907]: I0313 16:11:21.938227 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/173f7de2-ef0c-4750-a8b3-23e078ab9bde-ssh-key-openstack-cell1\") pod \"configure-os-openstack-openstack-cell1-s9llf\" (UID: \"173f7de2-ef0c-4750-a8b3-23e078ab9bde\") " pod="openstack/configure-os-openstack-openstack-cell1-s9llf" Mar 13 16:11:21 crc kubenswrapper[4907]: I0313 16:11:21.938365 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mt98f\" (UniqueName: \"kubernetes.io/projected/173f7de2-ef0c-4750-a8b3-23e078ab9bde-kube-api-access-mt98f\") pod \"configure-os-openstack-openstack-cell1-s9llf\" (UID: \"173f7de2-ef0c-4750-a8b3-23e078ab9bde\") " pod="openstack/configure-os-openstack-openstack-cell1-s9llf" Mar 13 16:11:21 crc kubenswrapper[4907]: I0313 16:11:21.943962 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/173f7de2-ef0c-4750-a8b3-23e078ab9bde-inventory\") pod \"configure-os-openstack-openstack-cell1-s9llf\" (UID: \"173f7de2-ef0c-4750-a8b3-23e078ab9bde\") " pod="openstack/configure-os-openstack-openstack-cell1-s9llf" Mar 13 16:11:21 crc kubenswrapper[4907]: I0313 16:11:21.944030 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/173f7de2-ef0c-4750-a8b3-23e078ab9bde-ceph\") pod \"configure-os-openstack-openstack-cell1-s9llf\" (UID: \"173f7de2-ef0c-4750-a8b3-23e078ab9bde\") " pod="openstack/configure-os-openstack-openstack-cell1-s9llf" Mar 13 16:11:21 crc kubenswrapper[4907]: I0313 16:11:21.945045 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/173f7de2-ef0c-4750-a8b3-23e078ab9bde-ssh-key-openstack-cell1\") pod \"configure-os-openstack-openstack-cell1-s9llf\" (UID: \"173f7de2-ef0c-4750-a8b3-23e078ab9bde\") " pod="openstack/configure-os-openstack-openstack-cell1-s9llf" Mar 13 16:11:21 crc kubenswrapper[4907]: I0313 16:11:21.960561 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mt98f\" (UniqueName: \"kubernetes.io/projected/173f7de2-ef0c-4750-a8b3-23e078ab9bde-kube-api-access-mt98f\") pod \"configure-os-openstack-openstack-cell1-s9llf\" (UID: \"173f7de2-ef0c-4750-a8b3-23e078ab9bde\") " pod="openstack/configure-os-openstack-openstack-cell1-s9llf" Mar 13 16:11:22 crc kubenswrapper[4907]: I0313 16:11:22.011176 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-cell1-s9llf" Mar 13 16:11:22 crc kubenswrapper[4907]: I0313 16:11:22.600017 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-openstack-openstack-cell1-s9llf"] Mar 13 16:11:22 crc kubenswrapper[4907]: I0313 16:11:22.609134 4907 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 13 16:11:22 crc kubenswrapper[4907]: I0313 16:11:22.717604 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-rq8fl"] Mar 13 16:11:22 crc kubenswrapper[4907]: I0313 16:11:22.720810 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rq8fl" Mar 13 16:11:22 crc kubenswrapper[4907]: I0313 16:11:22.741796 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rq8fl"] Mar 13 16:11:22 crc kubenswrapper[4907]: I0313 16:11:22.868840 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a822760f-41c5-40ab-9b8a-64f4a465c25f-utilities\") pod \"certified-operators-rq8fl\" (UID: \"a822760f-41c5-40ab-9b8a-64f4a465c25f\") " pod="openshift-marketplace/certified-operators-rq8fl" Mar 13 16:11:22 crc kubenswrapper[4907]: I0313 16:11:22.869604 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a822760f-41c5-40ab-9b8a-64f4a465c25f-catalog-content\") pod \"certified-operators-rq8fl\" (UID: \"a822760f-41c5-40ab-9b8a-64f4a465c25f\") " pod="openshift-marketplace/certified-operators-rq8fl" Mar 13 16:11:22 crc kubenswrapper[4907]: I0313 16:11:22.869646 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g2jm6\" (UniqueName: \"kubernetes.io/projected/a822760f-41c5-40ab-9b8a-64f4a465c25f-kube-api-access-g2jm6\") pod \"certified-operators-rq8fl\" (UID: \"a822760f-41c5-40ab-9b8a-64f4a465c25f\") " pod="openshift-marketplace/certified-operators-rq8fl" Mar 13 16:11:22 crc kubenswrapper[4907]: I0313 16:11:22.972474 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a822760f-41c5-40ab-9b8a-64f4a465c25f-utilities\") pod \"certified-operators-rq8fl\" (UID: \"a822760f-41c5-40ab-9b8a-64f4a465c25f\") " pod="openshift-marketplace/certified-operators-rq8fl" Mar 13 16:11:22 crc kubenswrapper[4907]: I0313 16:11:22.972573 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a822760f-41c5-40ab-9b8a-64f4a465c25f-catalog-content\") pod \"certified-operators-rq8fl\" (UID: \"a822760f-41c5-40ab-9b8a-64f4a465c25f\") " pod="openshift-marketplace/certified-operators-rq8fl" Mar 13 16:11:22 crc kubenswrapper[4907]: I0313 16:11:22.972599 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g2jm6\" (UniqueName: \"kubernetes.io/projected/a822760f-41c5-40ab-9b8a-64f4a465c25f-kube-api-access-g2jm6\") pod \"certified-operators-rq8fl\" (UID: \"a822760f-41c5-40ab-9b8a-64f4a465c25f\") " pod="openshift-marketplace/certified-operators-rq8fl" Mar 13 16:11:22 crc kubenswrapper[4907]: I0313 16:11:22.972909 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a822760f-41c5-40ab-9b8a-64f4a465c25f-utilities\") pod \"certified-operators-rq8fl\" (UID: \"a822760f-41c5-40ab-9b8a-64f4a465c25f\") " pod="openshift-marketplace/certified-operators-rq8fl" Mar 13 16:11:22 crc kubenswrapper[4907]: I0313 16:11:22.973520 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a822760f-41c5-40ab-9b8a-64f4a465c25f-catalog-content\") pod \"certified-operators-rq8fl\" (UID: \"a822760f-41c5-40ab-9b8a-64f4a465c25f\") " pod="openshift-marketplace/certified-operators-rq8fl" Mar 13 16:11:22 crc kubenswrapper[4907]: I0313 16:11:22.999636 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g2jm6\" (UniqueName: \"kubernetes.io/projected/a822760f-41c5-40ab-9b8a-64f4a465c25f-kube-api-access-g2jm6\") pod \"certified-operators-rq8fl\" (UID: \"a822760f-41c5-40ab-9b8a-64f4a465c25f\") " pod="openshift-marketplace/certified-operators-rq8fl" Mar 13 16:11:23 crc kubenswrapper[4907]: I0313 16:11:23.045475 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rq8fl" Mar 13 16:11:23 crc kubenswrapper[4907]: I0313 16:11:23.553400 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rq8fl"] Mar 13 16:11:23 crc kubenswrapper[4907]: W0313 16:11:23.557767 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda822760f_41c5_40ab_9b8a_64f4a465c25f.slice/crio-bdc94ca7df4e674f0317453fa5febaf6716d3284c442d035f275bc10a9e3bbde WatchSource:0}: Error finding container bdc94ca7df4e674f0317453fa5febaf6716d3284c442d035f275bc10a9e3bbde: Status 404 returned error can't find the container with id bdc94ca7df4e674f0317453fa5febaf6716d3284c442d035f275bc10a9e3bbde Mar 13 16:11:23 crc kubenswrapper[4907]: I0313 16:11:23.616806 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-cell1-s9llf" event={"ID":"173f7de2-ef0c-4750-a8b3-23e078ab9bde","Type":"ContainerStarted","Data":"5ea72312e8819f02cc06ddf2ae911b43e3f99fe396d78ccee0fc42da52e0192d"} Mar 13 16:11:23 crc kubenswrapper[4907]: I0313 16:11:23.616854 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-cell1-s9llf" event={"ID":"173f7de2-ef0c-4750-a8b3-23e078ab9bde","Type":"ContainerStarted","Data":"463d170ab3d7d3c1c1ed4d54bd7ad5aae875c69e0edf649f17b5303e58e8fd44"} Mar 13 16:11:23 crc kubenswrapper[4907]: I0313 16:11:23.619323 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rq8fl" event={"ID":"a822760f-41c5-40ab-9b8a-64f4a465c25f","Type":"ContainerStarted","Data":"bdc94ca7df4e674f0317453fa5febaf6716d3284c442d035f275bc10a9e3bbde"} Mar 13 16:11:23 crc kubenswrapper[4907]: I0313 16:11:23.647347 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-openstack-openstack-cell1-s9llf" podStartSLOduration=1.87948698 podStartE2EDuration="2.647327813s" podCreationTimestamp="2026-03-13 16:11:21 +0000 UTC" firstStartedPulling="2026-03-13 16:11:22.608849102 +0000 UTC m=+7581.508636791" lastFinishedPulling="2026-03-13 16:11:23.376689945 +0000 UTC m=+7582.276477624" observedRunningTime="2026-03-13 16:11:23.640607359 +0000 UTC m=+7582.540395048" watchObservedRunningTime="2026-03-13 16:11:23.647327813 +0000 UTC m=+7582.547115502" Mar 13 16:11:24 crc kubenswrapper[4907]: I0313 16:11:24.629983 4907 generic.go:334] "Generic (PLEG): container finished" podID="a822760f-41c5-40ab-9b8a-64f4a465c25f" containerID="eb08050c707ce49c267ab4b19c8d14852dd33b822d7c83e2aa9e2003f02b3fd9" exitCode=0 Mar 13 16:11:24 crc kubenswrapper[4907]: I0313 16:11:24.630043 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rq8fl" event={"ID":"a822760f-41c5-40ab-9b8a-64f4a465c25f","Type":"ContainerDied","Data":"eb08050c707ce49c267ab4b19c8d14852dd33b822d7c83e2aa9e2003f02b3fd9"} Mar 13 16:11:26 crc kubenswrapper[4907]: I0313 16:11:26.653083 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rq8fl" event={"ID":"a822760f-41c5-40ab-9b8a-64f4a465c25f","Type":"ContainerStarted","Data":"1d61ea08412ba1da32fdc2116795bc5b58af53fd12a07f6d91a43af634811d58"} Mar 13 16:11:28 crc kubenswrapper[4907]: I0313 16:11:28.670124 4907 generic.go:334] "Generic (PLEG): container finished" podID="a822760f-41c5-40ab-9b8a-64f4a465c25f" containerID="1d61ea08412ba1da32fdc2116795bc5b58af53fd12a07f6d91a43af634811d58" exitCode=0 Mar 13 16:11:28 crc kubenswrapper[4907]: I0313 16:11:28.670186 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rq8fl" event={"ID":"a822760f-41c5-40ab-9b8a-64f4a465c25f","Type":"ContainerDied","Data":"1d61ea08412ba1da32fdc2116795bc5b58af53fd12a07f6d91a43af634811d58"} Mar 13 16:11:29 crc kubenswrapper[4907]: I0313 16:11:29.681614 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rq8fl" event={"ID":"a822760f-41c5-40ab-9b8a-64f4a465c25f","Type":"ContainerStarted","Data":"c2f9dccdffbee706cf1dccbed6bcf5f01e43c63bb6ec1faaef0a1728b7246db3"} Mar 13 16:11:29 crc kubenswrapper[4907]: I0313 16:11:29.711501 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-rq8fl" podStartSLOduration=3.225665568 podStartE2EDuration="7.711482796s" podCreationTimestamp="2026-03-13 16:11:22 +0000 UTC" firstStartedPulling="2026-03-13 16:11:24.632508704 +0000 UTC m=+7583.532296393" lastFinishedPulling="2026-03-13 16:11:29.118325932 +0000 UTC m=+7588.018113621" observedRunningTime="2026-03-13 16:11:29.707292281 +0000 UTC m=+7588.607079970" watchObservedRunningTime="2026-03-13 16:11:29.711482796 +0000 UTC m=+7588.611270485" Mar 13 16:11:33 crc kubenswrapper[4907]: I0313 16:11:33.046242 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-rq8fl" Mar 13 16:11:33 crc kubenswrapper[4907]: I0313 16:11:33.046768 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-rq8fl" Mar 13 16:11:33 crc kubenswrapper[4907]: I0313 16:11:33.124026 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-rq8fl" Mar 13 16:11:43 crc kubenswrapper[4907]: I0313 16:11:43.155640 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-rq8fl" Mar 13 16:11:43 crc kubenswrapper[4907]: I0313 16:11:43.224188 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rq8fl"] Mar 13 16:11:43 crc kubenswrapper[4907]: I0313 16:11:43.811728 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-rq8fl" podUID="a822760f-41c5-40ab-9b8a-64f4a465c25f" containerName="registry-server" containerID="cri-o://c2f9dccdffbee706cf1dccbed6bcf5f01e43c63bb6ec1faaef0a1728b7246db3" gracePeriod=2 Mar 13 16:11:44 crc kubenswrapper[4907]: I0313 16:11:44.755583 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rq8fl" Mar 13 16:11:44 crc kubenswrapper[4907]: I0313 16:11:44.827657 4907 generic.go:334] "Generic (PLEG): container finished" podID="a822760f-41c5-40ab-9b8a-64f4a465c25f" containerID="c2f9dccdffbee706cf1dccbed6bcf5f01e43c63bb6ec1faaef0a1728b7246db3" exitCode=0 Mar 13 16:11:44 crc kubenswrapper[4907]: I0313 16:11:44.827723 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rq8fl" event={"ID":"a822760f-41c5-40ab-9b8a-64f4a465c25f","Type":"ContainerDied","Data":"c2f9dccdffbee706cf1dccbed6bcf5f01e43c63bb6ec1faaef0a1728b7246db3"} Mar 13 16:11:44 crc kubenswrapper[4907]: I0313 16:11:44.827820 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rq8fl" event={"ID":"a822760f-41c5-40ab-9b8a-64f4a465c25f","Type":"ContainerDied","Data":"bdc94ca7df4e674f0317453fa5febaf6716d3284c442d035f275bc10a9e3bbde"} Mar 13 16:11:44 crc kubenswrapper[4907]: I0313 16:11:44.827835 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rq8fl" Mar 13 16:11:44 crc kubenswrapper[4907]: I0313 16:11:44.827854 4907 scope.go:117] "RemoveContainer" containerID="c2f9dccdffbee706cf1dccbed6bcf5f01e43c63bb6ec1faaef0a1728b7246db3" Mar 13 16:11:44 crc kubenswrapper[4907]: I0313 16:11:44.861979 4907 scope.go:117] "RemoveContainer" containerID="1d61ea08412ba1da32fdc2116795bc5b58af53fd12a07f6d91a43af634811d58" Mar 13 16:11:44 crc kubenswrapper[4907]: I0313 16:11:44.870361 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g2jm6\" (UniqueName: \"kubernetes.io/projected/a822760f-41c5-40ab-9b8a-64f4a465c25f-kube-api-access-g2jm6\") pod \"a822760f-41c5-40ab-9b8a-64f4a465c25f\" (UID: \"a822760f-41c5-40ab-9b8a-64f4a465c25f\") " Mar 13 16:11:44 crc kubenswrapper[4907]: I0313 16:11:44.870676 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a822760f-41c5-40ab-9b8a-64f4a465c25f-catalog-content\") pod \"a822760f-41c5-40ab-9b8a-64f4a465c25f\" (UID: \"a822760f-41c5-40ab-9b8a-64f4a465c25f\") " Mar 13 16:11:44 crc kubenswrapper[4907]: I0313 16:11:44.870728 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a822760f-41c5-40ab-9b8a-64f4a465c25f-utilities\") pod \"a822760f-41c5-40ab-9b8a-64f4a465c25f\" (UID: \"a822760f-41c5-40ab-9b8a-64f4a465c25f\") " Mar 13 16:11:44 crc kubenswrapper[4907]: I0313 16:11:44.871725 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a822760f-41c5-40ab-9b8a-64f4a465c25f-utilities" (OuterVolumeSpecName: "utilities") pod "a822760f-41c5-40ab-9b8a-64f4a465c25f" (UID: "a822760f-41c5-40ab-9b8a-64f4a465c25f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 16:11:44 crc kubenswrapper[4907]: I0313 16:11:44.881734 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a822760f-41c5-40ab-9b8a-64f4a465c25f-kube-api-access-g2jm6" (OuterVolumeSpecName: "kube-api-access-g2jm6") pod "a822760f-41c5-40ab-9b8a-64f4a465c25f" (UID: "a822760f-41c5-40ab-9b8a-64f4a465c25f"). InnerVolumeSpecName "kube-api-access-g2jm6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:11:44 crc kubenswrapper[4907]: I0313 16:11:44.890014 4907 scope.go:117] "RemoveContainer" containerID="eb08050c707ce49c267ab4b19c8d14852dd33b822d7c83e2aa9e2003f02b3fd9" Mar 13 16:11:44 crc kubenswrapper[4907]: I0313 16:11:44.972764 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a822760f-41c5-40ab-9b8a-64f4a465c25f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a822760f-41c5-40ab-9b8a-64f4a465c25f" (UID: "a822760f-41c5-40ab-9b8a-64f4a465c25f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 16:11:44 crc kubenswrapper[4907]: I0313 16:11:44.973475 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g2jm6\" (UniqueName: \"kubernetes.io/projected/a822760f-41c5-40ab-9b8a-64f4a465c25f-kube-api-access-g2jm6\") on node \"crc\" DevicePath \"\"" Mar 13 16:11:44 crc kubenswrapper[4907]: I0313 16:11:44.973510 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a822760f-41c5-40ab-9b8a-64f4a465c25f-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 16:11:44 crc kubenswrapper[4907]: I0313 16:11:44.973520 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a822760f-41c5-40ab-9b8a-64f4a465c25f-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 16:11:45 crc kubenswrapper[4907]: I0313 16:11:45.009655 4907 scope.go:117] "RemoveContainer" containerID="c2f9dccdffbee706cf1dccbed6bcf5f01e43c63bb6ec1faaef0a1728b7246db3" Mar 13 16:11:45 crc kubenswrapper[4907]: E0313 16:11:45.010316 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c2f9dccdffbee706cf1dccbed6bcf5f01e43c63bb6ec1faaef0a1728b7246db3\": container with ID starting with c2f9dccdffbee706cf1dccbed6bcf5f01e43c63bb6ec1faaef0a1728b7246db3 not found: ID does not exist" containerID="c2f9dccdffbee706cf1dccbed6bcf5f01e43c63bb6ec1faaef0a1728b7246db3" Mar 13 16:11:45 crc kubenswrapper[4907]: I0313 16:11:45.010361 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c2f9dccdffbee706cf1dccbed6bcf5f01e43c63bb6ec1faaef0a1728b7246db3"} err="failed to get container status \"c2f9dccdffbee706cf1dccbed6bcf5f01e43c63bb6ec1faaef0a1728b7246db3\": rpc error: code = NotFound desc = could not find container \"c2f9dccdffbee706cf1dccbed6bcf5f01e43c63bb6ec1faaef0a1728b7246db3\": container with ID starting with c2f9dccdffbee706cf1dccbed6bcf5f01e43c63bb6ec1faaef0a1728b7246db3 not found: ID does not exist" Mar 13 16:11:45 crc kubenswrapper[4907]: I0313 16:11:45.010395 4907 scope.go:117] "RemoveContainer" containerID="1d61ea08412ba1da32fdc2116795bc5b58af53fd12a07f6d91a43af634811d58" Mar 13 16:11:45 crc kubenswrapper[4907]: E0313 16:11:45.010797 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1d61ea08412ba1da32fdc2116795bc5b58af53fd12a07f6d91a43af634811d58\": container with ID starting with 1d61ea08412ba1da32fdc2116795bc5b58af53fd12a07f6d91a43af634811d58 not found: ID does not exist" containerID="1d61ea08412ba1da32fdc2116795bc5b58af53fd12a07f6d91a43af634811d58" Mar 13 16:11:45 crc kubenswrapper[4907]: I0313 16:11:45.010826 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d61ea08412ba1da32fdc2116795bc5b58af53fd12a07f6d91a43af634811d58"} err="failed to get container status \"1d61ea08412ba1da32fdc2116795bc5b58af53fd12a07f6d91a43af634811d58\": rpc error: code = NotFound desc = could not find container \"1d61ea08412ba1da32fdc2116795bc5b58af53fd12a07f6d91a43af634811d58\": container with ID starting with 1d61ea08412ba1da32fdc2116795bc5b58af53fd12a07f6d91a43af634811d58 not found: ID does not exist" Mar 13 16:11:45 crc kubenswrapper[4907]: I0313 16:11:45.010863 4907 scope.go:117] "RemoveContainer" containerID="eb08050c707ce49c267ab4b19c8d14852dd33b822d7c83e2aa9e2003f02b3fd9" Mar 13 16:11:45 crc kubenswrapper[4907]: E0313 16:11:45.011268 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eb08050c707ce49c267ab4b19c8d14852dd33b822d7c83e2aa9e2003f02b3fd9\": container with ID starting with eb08050c707ce49c267ab4b19c8d14852dd33b822d7c83e2aa9e2003f02b3fd9 not found: ID does not exist" containerID="eb08050c707ce49c267ab4b19c8d14852dd33b822d7c83e2aa9e2003f02b3fd9" Mar 13 16:11:45 crc kubenswrapper[4907]: I0313 16:11:45.011303 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eb08050c707ce49c267ab4b19c8d14852dd33b822d7c83e2aa9e2003f02b3fd9"} err="failed to get container status \"eb08050c707ce49c267ab4b19c8d14852dd33b822d7c83e2aa9e2003f02b3fd9\": rpc error: code = NotFound desc = could not find container \"eb08050c707ce49c267ab4b19c8d14852dd33b822d7c83e2aa9e2003f02b3fd9\": container with ID starting with eb08050c707ce49c267ab4b19c8d14852dd33b822d7c83e2aa9e2003f02b3fd9 not found: ID does not exist" Mar 13 16:11:45 crc kubenswrapper[4907]: I0313 16:11:45.165923 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rq8fl"] Mar 13 16:11:45 crc kubenswrapper[4907]: I0313 16:11:45.175377 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-rq8fl"] Mar 13 16:11:45 crc kubenswrapper[4907]: I0313 16:11:45.824364 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a822760f-41c5-40ab-9b8a-64f4a465c25f" path="/var/lib/kubelet/pods/a822760f-41c5-40ab-9b8a-64f4a465c25f/volumes" Mar 13 16:11:48 crc kubenswrapper[4907]: I0313 16:11:48.041265 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 16:11:48 crc kubenswrapper[4907]: I0313 16:11:48.041839 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 16:12:00 crc kubenswrapper[4907]: I0313 16:12:00.177355 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556972-llpxn"] Mar 13 16:12:00 crc kubenswrapper[4907]: E0313 16:12:00.178591 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a822760f-41c5-40ab-9b8a-64f4a465c25f" containerName="extract-utilities" Mar 13 16:12:00 crc kubenswrapper[4907]: I0313 16:12:00.178605 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="a822760f-41c5-40ab-9b8a-64f4a465c25f" containerName="extract-utilities" Mar 13 16:12:00 crc kubenswrapper[4907]: E0313 16:12:00.178640 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a822760f-41c5-40ab-9b8a-64f4a465c25f" containerName="extract-content" Mar 13 16:12:00 crc kubenswrapper[4907]: I0313 16:12:00.178648 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="a822760f-41c5-40ab-9b8a-64f4a465c25f" containerName="extract-content" Mar 13 16:12:00 crc kubenswrapper[4907]: E0313 16:12:00.178677 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a822760f-41c5-40ab-9b8a-64f4a465c25f" containerName="registry-server" Mar 13 16:12:00 crc kubenswrapper[4907]: I0313 16:12:00.178684 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="a822760f-41c5-40ab-9b8a-64f4a465c25f" containerName="registry-server" Mar 13 16:12:00 crc kubenswrapper[4907]: I0313 16:12:00.179245 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="a822760f-41c5-40ab-9b8a-64f4a465c25f" containerName="registry-server" Mar 13 16:12:00 crc kubenswrapper[4907]: I0313 16:12:00.184740 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556972-llpxn" Mar 13 16:12:00 crc kubenswrapper[4907]: I0313 16:12:00.186922 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 16:12:00 crc kubenswrapper[4907]: I0313 16:12:00.187413 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 16:12:00 crc kubenswrapper[4907]: I0313 16:12:00.188457 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 16:12:00 crc kubenswrapper[4907]: I0313 16:12:00.192319 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556972-llpxn"] Mar 13 16:12:00 crc kubenswrapper[4907]: I0313 16:12:00.353854 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rmjn8\" (UniqueName: \"kubernetes.io/projected/85500c41-93c1-44ae-b418-d7f125c04621-kube-api-access-rmjn8\") pod \"auto-csr-approver-29556972-llpxn\" (UID: \"85500c41-93c1-44ae-b418-d7f125c04621\") " pod="openshift-infra/auto-csr-approver-29556972-llpxn" Mar 13 16:12:00 crc kubenswrapper[4907]: I0313 16:12:00.455622 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rmjn8\" (UniqueName: \"kubernetes.io/projected/85500c41-93c1-44ae-b418-d7f125c04621-kube-api-access-rmjn8\") pod \"auto-csr-approver-29556972-llpxn\" (UID: \"85500c41-93c1-44ae-b418-d7f125c04621\") " pod="openshift-infra/auto-csr-approver-29556972-llpxn" Mar 13 16:12:00 crc kubenswrapper[4907]: I0313 16:12:00.475293 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rmjn8\" (UniqueName: \"kubernetes.io/projected/85500c41-93c1-44ae-b418-d7f125c04621-kube-api-access-rmjn8\") pod \"auto-csr-approver-29556972-llpxn\" (UID: \"85500c41-93c1-44ae-b418-d7f125c04621\") " pod="openshift-infra/auto-csr-approver-29556972-llpxn" Mar 13 16:12:00 crc kubenswrapper[4907]: I0313 16:12:00.506647 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556972-llpxn" Mar 13 16:12:00 crc kubenswrapper[4907]: I0313 16:12:00.977774 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556972-llpxn"] Mar 13 16:12:01 crc kubenswrapper[4907]: I0313 16:12:01.988635 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556972-llpxn" event={"ID":"85500c41-93c1-44ae-b418-d7f125c04621","Type":"ContainerStarted","Data":"aa6902af445c9119dffb6494cc2d7852bdb913b5b91041562d50c8af3f525493"} Mar 13 16:12:02 crc kubenswrapper[4907]: I0313 16:12:02.998853 4907 generic.go:334] "Generic (PLEG): container finished" podID="85500c41-93c1-44ae-b418-d7f125c04621" containerID="1597ac303cc302238cff89d1286c78c824124bc23ca555a00a5dd03f41ac4104" exitCode=0 Mar 13 16:12:02 crc kubenswrapper[4907]: I0313 16:12:02.998933 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556972-llpxn" event={"ID":"85500c41-93c1-44ae-b418-d7f125c04621","Type":"ContainerDied","Data":"1597ac303cc302238cff89d1286c78c824124bc23ca555a00a5dd03f41ac4104"} Mar 13 16:12:04 crc kubenswrapper[4907]: I0313 16:12:04.471116 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556972-llpxn" Mar 13 16:12:04 crc kubenswrapper[4907]: I0313 16:12:04.560180 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rmjn8\" (UniqueName: \"kubernetes.io/projected/85500c41-93c1-44ae-b418-d7f125c04621-kube-api-access-rmjn8\") pod \"85500c41-93c1-44ae-b418-d7f125c04621\" (UID: \"85500c41-93c1-44ae-b418-d7f125c04621\") " Mar 13 16:12:04 crc kubenswrapper[4907]: I0313 16:12:04.566677 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/85500c41-93c1-44ae-b418-d7f125c04621-kube-api-access-rmjn8" (OuterVolumeSpecName: "kube-api-access-rmjn8") pod "85500c41-93c1-44ae-b418-d7f125c04621" (UID: "85500c41-93c1-44ae-b418-d7f125c04621"). InnerVolumeSpecName "kube-api-access-rmjn8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:12:04 crc kubenswrapper[4907]: I0313 16:12:04.663610 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rmjn8\" (UniqueName: \"kubernetes.io/projected/85500c41-93c1-44ae-b418-d7f125c04621-kube-api-access-rmjn8\") on node \"crc\" DevicePath \"\"" Mar 13 16:12:05 crc kubenswrapper[4907]: I0313 16:12:05.023935 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556972-llpxn" event={"ID":"85500c41-93c1-44ae-b418-d7f125c04621","Type":"ContainerDied","Data":"aa6902af445c9119dffb6494cc2d7852bdb913b5b91041562d50c8af3f525493"} Mar 13 16:12:05 crc kubenswrapper[4907]: I0313 16:12:05.023973 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aa6902af445c9119dffb6494cc2d7852bdb913b5b91041562d50c8af3f525493" Mar 13 16:12:05 crc kubenswrapper[4907]: I0313 16:12:05.023979 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556972-llpxn" Mar 13 16:12:05 crc kubenswrapper[4907]: I0313 16:12:05.546718 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556966-drz8k"] Mar 13 16:12:05 crc kubenswrapper[4907]: I0313 16:12:05.560828 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556966-drz8k"] Mar 13 16:12:05 crc kubenswrapper[4907]: I0313 16:12:05.796848 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9e6bd258-54f3-49e8-8b92-d101e6b19872" path="/var/lib/kubelet/pods/9e6bd258-54f3-49e8-8b92-d101e6b19872/volumes" Mar 13 16:12:09 crc kubenswrapper[4907]: I0313 16:12:09.065111 4907 generic.go:334] "Generic (PLEG): container finished" podID="173f7de2-ef0c-4750-a8b3-23e078ab9bde" containerID="5ea72312e8819f02cc06ddf2ae911b43e3f99fe396d78ccee0fc42da52e0192d" exitCode=0 Mar 13 16:12:09 crc kubenswrapper[4907]: I0313 16:12:09.065230 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-cell1-s9llf" event={"ID":"173f7de2-ef0c-4750-a8b3-23e078ab9bde","Type":"ContainerDied","Data":"5ea72312e8819f02cc06ddf2ae911b43e3f99fe396d78ccee0fc42da52e0192d"} Mar 13 16:12:10 crc kubenswrapper[4907]: I0313 16:12:10.625555 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-cell1-s9llf" Mar 13 16:12:10 crc kubenswrapper[4907]: I0313 16:12:10.818472 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/173f7de2-ef0c-4750-a8b3-23e078ab9bde-inventory\") pod \"173f7de2-ef0c-4750-a8b3-23e078ab9bde\" (UID: \"173f7de2-ef0c-4750-a8b3-23e078ab9bde\") " Mar 13 16:12:10 crc kubenswrapper[4907]: I0313 16:12:10.818536 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/173f7de2-ef0c-4750-a8b3-23e078ab9bde-ceph\") pod \"173f7de2-ef0c-4750-a8b3-23e078ab9bde\" (UID: \"173f7de2-ef0c-4750-a8b3-23e078ab9bde\") " Mar 13 16:12:10 crc kubenswrapper[4907]: I0313 16:12:10.818691 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mt98f\" (UniqueName: \"kubernetes.io/projected/173f7de2-ef0c-4750-a8b3-23e078ab9bde-kube-api-access-mt98f\") pod \"173f7de2-ef0c-4750-a8b3-23e078ab9bde\" (UID: \"173f7de2-ef0c-4750-a8b3-23e078ab9bde\") " Mar 13 16:12:10 crc kubenswrapper[4907]: I0313 16:12:10.818727 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/173f7de2-ef0c-4750-a8b3-23e078ab9bde-ssh-key-openstack-cell1\") pod \"173f7de2-ef0c-4750-a8b3-23e078ab9bde\" (UID: \"173f7de2-ef0c-4750-a8b3-23e078ab9bde\") " Mar 13 16:12:10 crc kubenswrapper[4907]: I0313 16:12:10.833031 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/173f7de2-ef0c-4750-a8b3-23e078ab9bde-ceph" (OuterVolumeSpecName: "ceph") pod "173f7de2-ef0c-4750-a8b3-23e078ab9bde" (UID: "173f7de2-ef0c-4750-a8b3-23e078ab9bde"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:12:10 crc kubenswrapper[4907]: I0313 16:12:10.834412 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/173f7de2-ef0c-4750-a8b3-23e078ab9bde-kube-api-access-mt98f" (OuterVolumeSpecName: "kube-api-access-mt98f") pod "173f7de2-ef0c-4750-a8b3-23e078ab9bde" (UID: "173f7de2-ef0c-4750-a8b3-23e078ab9bde"). InnerVolumeSpecName "kube-api-access-mt98f". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:12:10 crc kubenswrapper[4907]: I0313 16:12:10.846343 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/173f7de2-ef0c-4750-a8b3-23e078ab9bde-inventory" (OuterVolumeSpecName: "inventory") pod "173f7de2-ef0c-4750-a8b3-23e078ab9bde" (UID: "173f7de2-ef0c-4750-a8b3-23e078ab9bde"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:12:10 crc kubenswrapper[4907]: I0313 16:12:10.862466 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/173f7de2-ef0c-4750-a8b3-23e078ab9bde-ssh-key-openstack-cell1" (OuterVolumeSpecName: "ssh-key-openstack-cell1") pod "173f7de2-ef0c-4750-a8b3-23e078ab9bde" (UID: "173f7de2-ef0c-4750-a8b3-23e078ab9bde"). InnerVolumeSpecName "ssh-key-openstack-cell1". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:12:10 crc kubenswrapper[4907]: I0313 16:12:10.921419 4907 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/173f7de2-ef0c-4750-a8b3-23e078ab9bde-inventory\") on node \"crc\" DevicePath \"\"" Mar 13 16:12:10 crc kubenswrapper[4907]: I0313 16:12:10.921451 4907 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/173f7de2-ef0c-4750-a8b3-23e078ab9bde-ceph\") on node \"crc\" DevicePath \"\"" Mar 13 16:12:10 crc kubenswrapper[4907]: I0313 16:12:10.921460 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mt98f\" (UniqueName: \"kubernetes.io/projected/173f7de2-ef0c-4750-a8b3-23e078ab9bde-kube-api-access-mt98f\") on node \"crc\" DevicePath \"\"" Mar 13 16:12:10 crc kubenswrapper[4907]: I0313 16:12:10.921473 4907 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/173f7de2-ef0c-4750-a8b3-23e078ab9bde-ssh-key-openstack-cell1\") on node \"crc\" DevicePath \"\"" Mar 13 16:12:11 crc kubenswrapper[4907]: I0313 16:12:11.085383 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-cell1-s9llf" event={"ID":"173f7de2-ef0c-4750-a8b3-23e078ab9bde","Type":"ContainerDied","Data":"463d170ab3d7d3c1c1ed4d54bd7ad5aae875c69e0edf649f17b5303e58e8fd44"} Mar 13 16:12:11 crc kubenswrapper[4907]: I0313 16:12:11.085432 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="463d170ab3d7d3c1c1ed4d54bd7ad5aae875c69e0edf649f17b5303e58e8fd44" Mar 13 16:12:11 crc kubenswrapper[4907]: I0313 16:12:11.085432 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-cell1-s9llf" Mar 13 16:12:11 crc kubenswrapper[4907]: I0313 16:12:11.178297 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-openstack-phzr9"] Mar 13 16:12:11 crc kubenswrapper[4907]: E0313 16:12:11.179045 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85500c41-93c1-44ae-b418-d7f125c04621" containerName="oc" Mar 13 16:12:11 crc kubenswrapper[4907]: I0313 16:12:11.179074 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="85500c41-93c1-44ae-b418-d7f125c04621" containerName="oc" Mar 13 16:12:11 crc kubenswrapper[4907]: E0313 16:12:11.179093 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="173f7de2-ef0c-4750-a8b3-23e078ab9bde" containerName="configure-os-openstack-openstack-cell1" Mar 13 16:12:11 crc kubenswrapper[4907]: I0313 16:12:11.179101 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="173f7de2-ef0c-4750-a8b3-23e078ab9bde" containerName="configure-os-openstack-openstack-cell1" Mar 13 16:12:11 crc kubenswrapper[4907]: I0313 16:12:11.179309 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="85500c41-93c1-44ae-b418-d7f125c04621" containerName="oc" Mar 13 16:12:11 crc kubenswrapper[4907]: I0313 16:12:11.179330 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="173f7de2-ef0c-4750-a8b3-23e078ab9bde" containerName="configure-os-openstack-openstack-cell1" Mar 13 16:12:11 crc kubenswrapper[4907]: I0313 16:12:11.180181 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-openstack-phzr9" Mar 13 16:12:11 crc kubenswrapper[4907]: I0313 16:12:11.182586 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 13 16:12:11 crc kubenswrapper[4907]: I0313 16:12:11.182785 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Mar 13 16:12:11 crc kubenswrapper[4907]: I0313 16:12:11.183239 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-44kxw" Mar 13 16:12:11 crc kubenswrapper[4907]: I0313 16:12:11.184279 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Mar 13 16:12:11 crc kubenswrapper[4907]: I0313 16:12:11.219157 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-openstack-phzr9"] Mar 13 16:12:11 crc kubenswrapper[4907]: I0313 16:12:11.331189 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/4aef751c-592b-403e-b581-92c88cf65785-ssh-key-openstack-cell1\") pod \"ssh-known-hosts-openstack-phzr9\" (UID: \"4aef751c-592b-403e-b581-92c88cf65785\") " pod="openstack/ssh-known-hosts-openstack-phzr9" Mar 13 16:12:11 crc kubenswrapper[4907]: I0313 16:12:11.331380 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4aef751c-592b-403e-b581-92c88cf65785-ceph\") pod \"ssh-known-hosts-openstack-phzr9\" (UID: \"4aef751c-592b-403e-b581-92c88cf65785\") " pod="openstack/ssh-known-hosts-openstack-phzr9" Mar 13 16:12:11 crc kubenswrapper[4907]: I0313 16:12:11.331582 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/4aef751c-592b-403e-b581-92c88cf65785-inventory-0\") pod \"ssh-known-hosts-openstack-phzr9\" (UID: \"4aef751c-592b-403e-b581-92c88cf65785\") " pod="openstack/ssh-known-hosts-openstack-phzr9" Mar 13 16:12:11 crc kubenswrapper[4907]: I0313 16:12:11.331758 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tg6tq\" (UniqueName: \"kubernetes.io/projected/4aef751c-592b-403e-b581-92c88cf65785-kube-api-access-tg6tq\") pod \"ssh-known-hosts-openstack-phzr9\" (UID: \"4aef751c-592b-403e-b581-92c88cf65785\") " pod="openstack/ssh-known-hosts-openstack-phzr9" Mar 13 16:12:11 crc kubenswrapper[4907]: I0313 16:12:11.434582 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/4aef751c-592b-403e-b581-92c88cf65785-ssh-key-openstack-cell1\") pod \"ssh-known-hosts-openstack-phzr9\" (UID: \"4aef751c-592b-403e-b581-92c88cf65785\") " pod="openstack/ssh-known-hosts-openstack-phzr9" Mar 13 16:12:11 crc kubenswrapper[4907]: I0313 16:12:11.434723 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4aef751c-592b-403e-b581-92c88cf65785-ceph\") pod \"ssh-known-hosts-openstack-phzr9\" (UID: \"4aef751c-592b-403e-b581-92c88cf65785\") " pod="openstack/ssh-known-hosts-openstack-phzr9" Mar 13 16:12:11 crc kubenswrapper[4907]: I0313 16:12:11.434918 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/4aef751c-592b-403e-b581-92c88cf65785-inventory-0\") pod \"ssh-known-hosts-openstack-phzr9\" (UID: \"4aef751c-592b-403e-b581-92c88cf65785\") " pod="openstack/ssh-known-hosts-openstack-phzr9" Mar 13 16:12:11 crc kubenswrapper[4907]: I0313 16:12:11.435035 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tg6tq\" (UniqueName: \"kubernetes.io/projected/4aef751c-592b-403e-b581-92c88cf65785-kube-api-access-tg6tq\") pod \"ssh-known-hosts-openstack-phzr9\" (UID: \"4aef751c-592b-403e-b581-92c88cf65785\") " pod="openstack/ssh-known-hosts-openstack-phzr9" Mar 13 16:12:11 crc kubenswrapper[4907]: I0313 16:12:11.444174 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4aef751c-592b-403e-b581-92c88cf65785-ceph\") pod \"ssh-known-hosts-openstack-phzr9\" (UID: \"4aef751c-592b-403e-b581-92c88cf65785\") " pod="openstack/ssh-known-hosts-openstack-phzr9" Mar 13 16:12:11 crc kubenswrapper[4907]: I0313 16:12:11.444415 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/4aef751c-592b-403e-b581-92c88cf65785-ssh-key-openstack-cell1\") pod \"ssh-known-hosts-openstack-phzr9\" (UID: \"4aef751c-592b-403e-b581-92c88cf65785\") " pod="openstack/ssh-known-hosts-openstack-phzr9" Mar 13 16:12:11 crc kubenswrapper[4907]: I0313 16:12:11.458805 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tg6tq\" (UniqueName: \"kubernetes.io/projected/4aef751c-592b-403e-b581-92c88cf65785-kube-api-access-tg6tq\") pod \"ssh-known-hosts-openstack-phzr9\" (UID: \"4aef751c-592b-403e-b581-92c88cf65785\") " pod="openstack/ssh-known-hosts-openstack-phzr9" Mar 13 16:12:11 crc kubenswrapper[4907]: I0313 16:12:11.468044 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/4aef751c-592b-403e-b581-92c88cf65785-inventory-0\") pod \"ssh-known-hosts-openstack-phzr9\" (UID: \"4aef751c-592b-403e-b581-92c88cf65785\") " pod="openstack/ssh-known-hosts-openstack-phzr9" Mar 13 16:12:11 crc kubenswrapper[4907]: I0313 16:12:11.501640 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-openstack-phzr9" Mar 13 16:12:12 crc kubenswrapper[4907]: I0313 16:12:12.140164 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-openstack-phzr9"] Mar 13 16:12:13 crc kubenswrapper[4907]: I0313 16:12:13.127681 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-openstack-phzr9" event={"ID":"4aef751c-592b-403e-b581-92c88cf65785","Type":"ContainerStarted","Data":"fae553dd1e8789a4971d11997504d70757ec24d649cdf5a39afeb67e1286709d"} Mar 13 16:12:13 crc kubenswrapper[4907]: I0313 16:12:13.128267 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-openstack-phzr9" event={"ID":"4aef751c-592b-403e-b581-92c88cf65785","Type":"ContainerStarted","Data":"3606c840833edcd74a624c6afbba128aa52c01cc3569fec6423293e6c8311b53"} Mar 13 16:12:13 crc kubenswrapper[4907]: I0313 16:12:13.158499 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-openstack-phzr9" podStartSLOduration=1.679763903 podStartE2EDuration="2.158478044s" podCreationTimestamp="2026-03-13 16:12:11 +0000 UTC" firstStartedPulling="2026-03-13 16:12:12.161353065 +0000 UTC m=+7631.061140774" lastFinishedPulling="2026-03-13 16:12:12.640067216 +0000 UTC m=+7631.539854915" observedRunningTime="2026-03-13 16:12:13.144360077 +0000 UTC m=+7632.044147766" watchObservedRunningTime="2026-03-13 16:12:13.158478044 +0000 UTC m=+7632.058265733" Mar 13 16:12:15 crc kubenswrapper[4907]: I0313 16:12:15.672519 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-9q59x"] Mar 13 16:12:15 crc kubenswrapper[4907]: I0313 16:12:15.675236 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9q59x" Mar 13 16:12:15 crc kubenswrapper[4907]: I0313 16:12:15.716038 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9q59x"] Mar 13 16:12:15 crc kubenswrapper[4907]: I0313 16:12:15.730954 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aabfe118-3ad6-4625-bdab-6a1f2bc50890-utilities\") pod \"community-operators-9q59x\" (UID: \"aabfe118-3ad6-4625-bdab-6a1f2bc50890\") " pod="openshift-marketplace/community-operators-9q59x" Mar 13 16:12:15 crc kubenswrapper[4907]: I0313 16:12:15.731017 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aabfe118-3ad6-4625-bdab-6a1f2bc50890-catalog-content\") pod \"community-operators-9q59x\" (UID: \"aabfe118-3ad6-4625-bdab-6a1f2bc50890\") " pod="openshift-marketplace/community-operators-9q59x" Mar 13 16:12:15 crc kubenswrapper[4907]: I0313 16:12:15.731374 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9xr6b\" (UniqueName: \"kubernetes.io/projected/aabfe118-3ad6-4625-bdab-6a1f2bc50890-kube-api-access-9xr6b\") pod \"community-operators-9q59x\" (UID: \"aabfe118-3ad6-4625-bdab-6a1f2bc50890\") " pod="openshift-marketplace/community-operators-9q59x" Mar 13 16:12:15 crc kubenswrapper[4907]: I0313 16:12:15.832355 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aabfe118-3ad6-4625-bdab-6a1f2bc50890-utilities\") pod \"community-operators-9q59x\" (UID: \"aabfe118-3ad6-4625-bdab-6a1f2bc50890\") " pod="openshift-marketplace/community-operators-9q59x" Mar 13 16:12:15 crc kubenswrapper[4907]: I0313 16:12:15.832416 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aabfe118-3ad6-4625-bdab-6a1f2bc50890-catalog-content\") pod \"community-operators-9q59x\" (UID: \"aabfe118-3ad6-4625-bdab-6a1f2bc50890\") " pod="openshift-marketplace/community-operators-9q59x" Mar 13 16:12:15 crc kubenswrapper[4907]: I0313 16:12:15.832538 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9xr6b\" (UniqueName: \"kubernetes.io/projected/aabfe118-3ad6-4625-bdab-6a1f2bc50890-kube-api-access-9xr6b\") pod \"community-operators-9q59x\" (UID: \"aabfe118-3ad6-4625-bdab-6a1f2bc50890\") " pod="openshift-marketplace/community-operators-9q59x" Mar 13 16:12:15 crc kubenswrapper[4907]: I0313 16:12:15.833000 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aabfe118-3ad6-4625-bdab-6a1f2bc50890-catalog-content\") pod \"community-operators-9q59x\" (UID: \"aabfe118-3ad6-4625-bdab-6a1f2bc50890\") " pod="openshift-marketplace/community-operators-9q59x" Mar 13 16:12:15 crc kubenswrapper[4907]: I0313 16:12:15.833136 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aabfe118-3ad6-4625-bdab-6a1f2bc50890-utilities\") pod \"community-operators-9q59x\" (UID: \"aabfe118-3ad6-4625-bdab-6a1f2bc50890\") " pod="openshift-marketplace/community-operators-9q59x" Mar 13 16:12:15 crc kubenswrapper[4907]: I0313 16:12:15.851519 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9xr6b\" (UniqueName: \"kubernetes.io/projected/aabfe118-3ad6-4625-bdab-6a1f2bc50890-kube-api-access-9xr6b\") pod \"community-operators-9q59x\" (UID: \"aabfe118-3ad6-4625-bdab-6a1f2bc50890\") " pod="openshift-marketplace/community-operators-9q59x" Mar 13 16:12:16 crc kubenswrapper[4907]: I0313 16:12:16.018645 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9q59x" Mar 13 16:12:16 crc kubenswrapper[4907]: I0313 16:12:16.571496 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9q59x"] Mar 13 16:12:17 crc kubenswrapper[4907]: I0313 16:12:17.165739 4907 generic.go:334] "Generic (PLEG): container finished" podID="aabfe118-3ad6-4625-bdab-6a1f2bc50890" containerID="4ec6ccaa677ccbbb66f21ece73e47ce24307c31346e8c97bf83dbdeb0ab4c535" exitCode=0 Mar 13 16:12:17 crc kubenswrapper[4907]: I0313 16:12:17.165784 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9q59x" event={"ID":"aabfe118-3ad6-4625-bdab-6a1f2bc50890","Type":"ContainerDied","Data":"4ec6ccaa677ccbbb66f21ece73e47ce24307c31346e8c97bf83dbdeb0ab4c535"} Mar 13 16:12:17 crc kubenswrapper[4907]: I0313 16:12:17.166016 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9q59x" event={"ID":"aabfe118-3ad6-4625-bdab-6a1f2bc50890","Type":"ContainerStarted","Data":"e562e09fb98e987deb99f7ca2ebea739da366ebb50fdb52113a88f14af75db5b"} Mar 13 16:12:18 crc kubenswrapper[4907]: I0313 16:12:18.042014 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 16:12:18 crc kubenswrapper[4907]: I0313 16:12:18.042711 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 16:12:18 crc kubenswrapper[4907]: I0313 16:12:18.194557 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9q59x" event={"ID":"aabfe118-3ad6-4625-bdab-6a1f2bc50890","Type":"ContainerStarted","Data":"c706458164f7c6287d29e34e31517532fdb1cc00947568aaa9db8fa03c2c9c02"} Mar 13 16:12:20 crc kubenswrapper[4907]: I0313 16:12:20.218763 4907 generic.go:334] "Generic (PLEG): container finished" podID="aabfe118-3ad6-4625-bdab-6a1f2bc50890" containerID="c706458164f7c6287d29e34e31517532fdb1cc00947568aaa9db8fa03c2c9c02" exitCode=0 Mar 13 16:12:20 crc kubenswrapper[4907]: I0313 16:12:20.218836 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9q59x" event={"ID":"aabfe118-3ad6-4625-bdab-6a1f2bc50890","Type":"ContainerDied","Data":"c706458164f7c6287d29e34e31517532fdb1cc00947568aaa9db8fa03c2c9c02"} Mar 13 16:12:21 crc kubenswrapper[4907]: I0313 16:12:21.231249 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9q59x" event={"ID":"aabfe118-3ad6-4625-bdab-6a1f2bc50890","Type":"ContainerStarted","Data":"d01c41ac19b3db32d5d13701af11e1e2508fcdc64db4c92f02188f251ddd2445"} Mar 13 16:12:21 crc kubenswrapper[4907]: I0313 16:12:21.266370 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-9q59x" podStartSLOduration=2.477429075 podStartE2EDuration="6.266354991s" podCreationTimestamp="2026-03-13 16:12:15 +0000 UTC" firstStartedPulling="2026-03-13 16:12:17.169039535 +0000 UTC m=+7636.068827224" lastFinishedPulling="2026-03-13 16:12:20.957965441 +0000 UTC m=+7639.857753140" observedRunningTime="2026-03-13 16:12:21.2597576 +0000 UTC m=+7640.159545289" watchObservedRunningTime="2026-03-13 16:12:21.266354991 +0000 UTC m=+7640.166142680" Mar 13 16:12:22 crc kubenswrapper[4907]: I0313 16:12:22.239762 4907 generic.go:334] "Generic (PLEG): container finished" podID="4aef751c-592b-403e-b581-92c88cf65785" containerID="fae553dd1e8789a4971d11997504d70757ec24d649cdf5a39afeb67e1286709d" exitCode=0 Mar 13 16:12:22 crc kubenswrapper[4907]: I0313 16:12:22.239875 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-openstack-phzr9" event={"ID":"4aef751c-592b-403e-b581-92c88cf65785","Type":"ContainerDied","Data":"fae553dd1e8789a4971d11997504d70757ec24d649cdf5a39afeb67e1286709d"} Mar 13 16:12:23 crc kubenswrapper[4907]: I0313 16:12:23.845314 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-openstack-phzr9" Mar 13 16:12:23 crc kubenswrapper[4907]: I0313 16:12:23.921525 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tg6tq\" (UniqueName: \"kubernetes.io/projected/4aef751c-592b-403e-b581-92c88cf65785-kube-api-access-tg6tq\") pod \"4aef751c-592b-403e-b581-92c88cf65785\" (UID: \"4aef751c-592b-403e-b581-92c88cf65785\") " Mar 13 16:12:23 crc kubenswrapper[4907]: I0313 16:12:23.921617 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4aef751c-592b-403e-b581-92c88cf65785-ceph\") pod \"4aef751c-592b-403e-b581-92c88cf65785\" (UID: \"4aef751c-592b-403e-b581-92c88cf65785\") " Mar 13 16:12:23 crc kubenswrapper[4907]: I0313 16:12:23.921675 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/4aef751c-592b-403e-b581-92c88cf65785-ssh-key-openstack-cell1\") pod \"4aef751c-592b-403e-b581-92c88cf65785\" (UID: \"4aef751c-592b-403e-b581-92c88cf65785\") " Mar 13 16:12:23 crc kubenswrapper[4907]: I0313 16:12:23.921711 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/4aef751c-592b-403e-b581-92c88cf65785-inventory-0\") pod \"4aef751c-592b-403e-b581-92c88cf65785\" (UID: \"4aef751c-592b-403e-b581-92c88cf65785\") " Mar 13 16:12:23 crc kubenswrapper[4907]: I0313 16:12:23.930583 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4aef751c-592b-403e-b581-92c88cf65785-kube-api-access-tg6tq" (OuterVolumeSpecName: "kube-api-access-tg6tq") pod "4aef751c-592b-403e-b581-92c88cf65785" (UID: "4aef751c-592b-403e-b581-92c88cf65785"). InnerVolumeSpecName "kube-api-access-tg6tq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:12:23 crc kubenswrapper[4907]: I0313 16:12:23.931188 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4aef751c-592b-403e-b581-92c88cf65785-ceph" (OuterVolumeSpecName: "ceph") pod "4aef751c-592b-403e-b581-92c88cf65785" (UID: "4aef751c-592b-403e-b581-92c88cf65785"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:12:23 crc kubenswrapper[4907]: I0313 16:12:23.949856 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4aef751c-592b-403e-b581-92c88cf65785-ssh-key-openstack-cell1" (OuterVolumeSpecName: "ssh-key-openstack-cell1") pod "4aef751c-592b-403e-b581-92c88cf65785" (UID: "4aef751c-592b-403e-b581-92c88cf65785"). InnerVolumeSpecName "ssh-key-openstack-cell1". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:12:23 crc kubenswrapper[4907]: I0313 16:12:23.950268 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4aef751c-592b-403e-b581-92c88cf65785-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "4aef751c-592b-403e-b581-92c88cf65785" (UID: "4aef751c-592b-403e-b581-92c88cf65785"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:12:24 crc kubenswrapper[4907]: I0313 16:12:24.024066 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tg6tq\" (UniqueName: \"kubernetes.io/projected/4aef751c-592b-403e-b581-92c88cf65785-kube-api-access-tg6tq\") on node \"crc\" DevicePath \"\"" Mar 13 16:12:24 crc kubenswrapper[4907]: I0313 16:12:24.024096 4907 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4aef751c-592b-403e-b581-92c88cf65785-ceph\") on node \"crc\" DevicePath \"\"" Mar 13 16:12:24 crc kubenswrapper[4907]: I0313 16:12:24.024106 4907 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/4aef751c-592b-403e-b581-92c88cf65785-ssh-key-openstack-cell1\") on node \"crc\" DevicePath \"\"" Mar 13 16:12:24 crc kubenswrapper[4907]: I0313 16:12:24.024116 4907 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/4aef751c-592b-403e-b581-92c88cf65785-inventory-0\") on node \"crc\" DevicePath \"\"" Mar 13 16:12:24 crc kubenswrapper[4907]: I0313 16:12:24.260800 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-openstack-phzr9" event={"ID":"4aef751c-592b-403e-b581-92c88cf65785","Type":"ContainerDied","Data":"3606c840833edcd74a624c6afbba128aa52c01cc3569fec6423293e6c8311b53"} Mar 13 16:12:24 crc kubenswrapper[4907]: I0313 16:12:24.260840 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3606c840833edcd74a624c6afbba128aa52c01cc3569fec6423293e6c8311b53" Mar 13 16:12:24 crc kubenswrapper[4907]: I0313 16:12:24.260916 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-openstack-phzr9" Mar 13 16:12:24 crc kubenswrapper[4907]: I0313 16:12:24.364673 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-openstack-openstack-cell1-hml5k"] Mar 13 16:12:24 crc kubenswrapper[4907]: E0313 16:12:24.365209 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4aef751c-592b-403e-b581-92c88cf65785" containerName="ssh-known-hosts-openstack" Mar 13 16:12:24 crc kubenswrapper[4907]: I0313 16:12:24.365227 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="4aef751c-592b-403e-b581-92c88cf65785" containerName="ssh-known-hosts-openstack" Mar 13 16:12:24 crc kubenswrapper[4907]: I0313 16:12:24.365443 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="4aef751c-592b-403e-b581-92c88cf65785" containerName="ssh-known-hosts-openstack" Mar 13 16:12:24 crc kubenswrapper[4907]: I0313 16:12:24.366239 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-cell1-hml5k" Mar 13 16:12:24 crc kubenswrapper[4907]: I0313 16:12:24.370252 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Mar 13 16:12:24 crc kubenswrapper[4907]: I0313 16:12:24.372735 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 13 16:12:24 crc kubenswrapper[4907]: I0313 16:12:24.372810 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-44kxw" Mar 13 16:12:24 crc kubenswrapper[4907]: I0313 16:12:24.372820 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Mar 13 16:12:24 crc kubenswrapper[4907]: I0313 16:12:24.378947 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-openstack-openstack-cell1-hml5k"] Mar 13 16:12:24 crc kubenswrapper[4907]: I0313 16:12:24.434097 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/34db6153-7522-4aae-9712-eb87cc9aa543-inventory\") pod \"run-os-openstack-openstack-cell1-hml5k\" (UID: \"34db6153-7522-4aae-9712-eb87cc9aa543\") " pod="openstack/run-os-openstack-openstack-cell1-hml5k" Mar 13 16:12:24 crc kubenswrapper[4907]: I0313 16:12:24.434302 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/34db6153-7522-4aae-9712-eb87cc9aa543-ssh-key-openstack-cell1\") pod \"run-os-openstack-openstack-cell1-hml5k\" (UID: \"34db6153-7522-4aae-9712-eb87cc9aa543\") " pod="openstack/run-os-openstack-openstack-cell1-hml5k" Mar 13 16:12:24 crc kubenswrapper[4907]: I0313 16:12:24.434339 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/34db6153-7522-4aae-9712-eb87cc9aa543-ceph\") pod \"run-os-openstack-openstack-cell1-hml5k\" (UID: \"34db6153-7522-4aae-9712-eb87cc9aa543\") " pod="openstack/run-os-openstack-openstack-cell1-hml5k" Mar 13 16:12:24 crc kubenswrapper[4907]: I0313 16:12:24.434549 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7dstf\" (UniqueName: \"kubernetes.io/projected/34db6153-7522-4aae-9712-eb87cc9aa543-kube-api-access-7dstf\") pod \"run-os-openstack-openstack-cell1-hml5k\" (UID: \"34db6153-7522-4aae-9712-eb87cc9aa543\") " pod="openstack/run-os-openstack-openstack-cell1-hml5k" Mar 13 16:12:24 crc kubenswrapper[4907]: I0313 16:12:24.536520 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/34db6153-7522-4aae-9712-eb87cc9aa543-ssh-key-openstack-cell1\") pod \"run-os-openstack-openstack-cell1-hml5k\" (UID: \"34db6153-7522-4aae-9712-eb87cc9aa543\") " pod="openstack/run-os-openstack-openstack-cell1-hml5k" Mar 13 16:12:24 crc kubenswrapper[4907]: I0313 16:12:24.536559 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/34db6153-7522-4aae-9712-eb87cc9aa543-ceph\") pod \"run-os-openstack-openstack-cell1-hml5k\" (UID: \"34db6153-7522-4aae-9712-eb87cc9aa543\") " pod="openstack/run-os-openstack-openstack-cell1-hml5k" Mar 13 16:12:24 crc kubenswrapper[4907]: I0313 16:12:24.536626 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7dstf\" (UniqueName: \"kubernetes.io/projected/34db6153-7522-4aae-9712-eb87cc9aa543-kube-api-access-7dstf\") pod \"run-os-openstack-openstack-cell1-hml5k\" (UID: \"34db6153-7522-4aae-9712-eb87cc9aa543\") " pod="openstack/run-os-openstack-openstack-cell1-hml5k" Mar 13 16:12:24 crc kubenswrapper[4907]: I0313 16:12:24.536686 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/34db6153-7522-4aae-9712-eb87cc9aa543-inventory\") pod \"run-os-openstack-openstack-cell1-hml5k\" (UID: \"34db6153-7522-4aae-9712-eb87cc9aa543\") " pod="openstack/run-os-openstack-openstack-cell1-hml5k" Mar 13 16:12:24 crc kubenswrapper[4907]: I0313 16:12:24.541541 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/34db6153-7522-4aae-9712-eb87cc9aa543-ssh-key-openstack-cell1\") pod \"run-os-openstack-openstack-cell1-hml5k\" (UID: \"34db6153-7522-4aae-9712-eb87cc9aa543\") " pod="openstack/run-os-openstack-openstack-cell1-hml5k" Mar 13 16:12:24 crc kubenswrapper[4907]: I0313 16:12:24.546569 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/34db6153-7522-4aae-9712-eb87cc9aa543-inventory\") pod \"run-os-openstack-openstack-cell1-hml5k\" (UID: \"34db6153-7522-4aae-9712-eb87cc9aa543\") " pod="openstack/run-os-openstack-openstack-cell1-hml5k" Mar 13 16:12:24 crc kubenswrapper[4907]: I0313 16:12:24.554016 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/34db6153-7522-4aae-9712-eb87cc9aa543-ceph\") pod \"run-os-openstack-openstack-cell1-hml5k\" (UID: \"34db6153-7522-4aae-9712-eb87cc9aa543\") " pod="openstack/run-os-openstack-openstack-cell1-hml5k" Mar 13 16:12:24 crc kubenswrapper[4907]: I0313 16:12:24.558627 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7dstf\" (UniqueName: \"kubernetes.io/projected/34db6153-7522-4aae-9712-eb87cc9aa543-kube-api-access-7dstf\") pod \"run-os-openstack-openstack-cell1-hml5k\" (UID: \"34db6153-7522-4aae-9712-eb87cc9aa543\") " pod="openstack/run-os-openstack-openstack-cell1-hml5k" Mar 13 16:12:24 crc kubenswrapper[4907]: I0313 16:12:24.700298 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-cell1-hml5k" Mar 13 16:12:25 crc kubenswrapper[4907]: I0313 16:12:25.249993 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-openstack-openstack-cell1-hml5k"] Mar 13 16:12:25 crc kubenswrapper[4907]: W0313 16:12:25.250060 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod34db6153_7522_4aae_9712_eb87cc9aa543.slice/crio-1120479ce59841724e323313ca8d2a10c18000da39d4ac71ea46d45f1179803e WatchSource:0}: Error finding container 1120479ce59841724e323313ca8d2a10c18000da39d4ac71ea46d45f1179803e: Status 404 returned error can't find the container with id 1120479ce59841724e323313ca8d2a10c18000da39d4ac71ea46d45f1179803e Mar 13 16:12:25 crc kubenswrapper[4907]: I0313 16:12:25.278245 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-cell1-hml5k" event={"ID":"34db6153-7522-4aae-9712-eb87cc9aa543","Type":"ContainerStarted","Data":"1120479ce59841724e323313ca8d2a10c18000da39d4ac71ea46d45f1179803e"} Mar 13 16:12:26 crc kubenswrapper[4907]: I0313 16:12:26.019855 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-9q59x" Mar 13 16:12:26 crc kubenswrapper[4907]: I0313 16:12:26.020178 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-9q59x" Mar 13 16:12:26 crc kubenswrapper[4907]: I0313 16:12:26.068530 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-9q59x" Mar 13 16:12:26 crc kubenswrapper[4907]: I0313 16:12:26.291050 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-cell1-hml5k" event={"ID":"34db6153-7522-4aae-9712-eb87cc9aa543","Type":"ContainerStarted","Data":"1a4edf4adc052232eadad9bc1d0620ff53938e2a203a59f9c95cd4e2f267081a"} Mar 13 16:12:26 crc kubenswrapper[4907]: I0313 16:12:26.336007 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-9q59x" Mar 13 16:12:26 crc kubenswrapper[4907]: I0313 16:12:26.365167 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-openstack-openstack-cell1-hml5k" podStartSLOduration=1.864461102 podStartE2EDuration="2.365144924s" podCreationTimestamp="2026-03-13 16:12:24 +0000 UTC" firstStartedPulling="2026-03-13 16:12:25.252453403 +0000 UTC m=+7644.152241102" lastFinishedPulling="2026-03-13 16:12:25.753137195 +0000 UTC m=+7644.652924924" observedRunningTime="2026-03-13 16:12:26.313428789 +0000 UTC m=+7645.213216488" watchObservedRunningTime="2026-03-13 16:12:26.365144924 +0000 UTC m=+7645.264932623" Mar 13 16:12:26 crc kubenswrapper[4907]: I0313 16:12:26.386311 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9q59x"] Mar 13 16:12:28 crc kubenswrapper[4907]: I0313 16:12:28.307693 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-9q59x" podUID="aabfe118-3ad6-4625-bdab-6a1f2bc50890" containerName="registry-server" containerID="cri-o://d01c41ac19b3db32d5d13701af11e1e2508fcdc64db4c92f02188f251ddd2445" gracePeriod=2 Mar 13 16:12:28 crc kubenswrapper[4907]: I0313 16:12:28.899351 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9q59x" Mar 13 16:12:29 crc kubenswrapper[4907]: I0313 16:12:29.039521 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aabfe118-3ad6-4625-bdab-6a1f2bc50890-catalog-content\") pod \"aabfe118-3ad6-4625-bdab-6a1f2bc50890\" (UID: \"aabfe118-3ad6-4625-bdab-6a1f2bc50890\") " Mar 13 16:12:29 crc kubenswrapper[4907]: I0313 16:12:29.039598 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xr6b\" (UniqueName: \"kubernetes.io/projected/aabfe118-3ad6-4625-bdab-6a1f2bc50890-kube-api-access-9xr6b\") pod \"aabfe118-3ad6-4625-bdab-6a1f2bc50890\" (UID: \"aabfe118-3ad6-4625-bdab-6a1f2bc50890\") " Mar 13 16:12:29 crc kubenswrapper[4907]: I0313 16:12:29.039649 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aabfe118-3ad6-4625-bdab-6a1f2bc50890-utilities\") pod \"aabfe118-3ad6-4625-bdab-6a1f2bc50890\" (UID: \"aabfe118-3ad6-4625-bdab-6a1f2bc50890\") " Mar 13 16:12:29 crc kubenswrapper[4907]: I0313 16:12:29.040899 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aabfe118-3ad6-4625-bdab-6a1f2bc50890-utilities" (OuterVolumeSpecName: "utilities") pod "aabfe118-3ad6-4625-bdab-6a1f2bc50890" (UID: "aabfe118-3ad6-4625-bdab-6a1f2bc50890"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 16:12:29 crc kubenswrapper[4907]: I0313 16:12:29.054229 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aabfe118-3ad6-4625-bdab-6a1f2bc50890-kube-api-access-9xr6b" (OuterVolumeSpecName: "kube-api-access-9xr6b") pod "aabfe118-3ad6-4625-bdab-6a1f2bc50890" (UID: "aabfe118-3ad6-4625-bdab-6a1f2bc50890"). InnerVolumeSpecName "kube-api-access-9xr6b". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:12:29 crc kubenswrapper[4907]: I0313 16:12:29.092629 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aabfe118-3ad6-4625-bdab-6a1f2bc50890-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "aabfe118-3ad6-4625-bdab-6a1f2bc50890" (UID: "aabfe118-3ad6-4625-bdab-6a1f2bc50890"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 16:12:29 crc kubenswrapper[4907]: I0313 16:12:29.142053 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aabfe118-3ad6-4625-bdab-6a1f2bc50890-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 16:12:29 crc kubenswrapper[4907]: I0313 16:12:29.142094 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xr6b\" (UniqueName: \"kubernetes.io/projected/aabfe118-3ad6-4625-bdab-6a1f2bc50890-kube-api-access-9xr6b\") on node \"crc\" DevicePath \"\"" Mar 13 16:12:29 crc kubenswrapper[4907]: I0313 16:12:29.142108 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aabfe118-3ad6-4625-bdab-6a1f2bc50890-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 16:12:29 crc kubenswrapper[4907]: I0313 16:12:29.318326 4907 generic.go:334] "Generic (PLEG): container finished" podID="aabfe118-3ad6-4625-bdab-6a1f2bc50890" containerID="d01c41ac19b3db32d5d13701af11e1e2508fcdc64db4c92f02188f251ddd2445" exitCode=0 Mar 13 16:12:29 crc kubenswrapper[4907]: I0313 16:12:29.318384 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9q59x" event={"ID":"aabfe118-3ad6-4625-bdab-6a1f2bc50890","Type":"ContainerDied","Data":"d01c41ac19b3db32d5d13701af11e1e2508fcdc64db4c92f02188f251ddd2445"} Mar 13 16:12:29 crc kubenswrapper[4907]: I0313 16:12:29.318415 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9q59x" event={"ID":"aabfe118-3ad6-4625-bdab-6a1f2bc50890","Type":"ContainerDied","Data":"e562e09fb98e987deb99f7ca2ebea739da366ebb50fdb52113a88f14af75db5b"} Mar 13 16:12:29 crc kubenswrapper[4907]: I0313 16:12:29.318434 4907 scope.go:117] "RemoveContainer" containerID="d01c41ac19b3db32d5d13701af11e1e2508fcdc64db4c92f02188f251ddd2445" Mar 13 16:12:29 crc kubenswrapper[4907]: I0313 16:12:29.318426 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9q59x" Mar 13 16:12:29 crc kubenswrapper[4907]: I0313 16:12:29.350832 4907 scope.go:117] "RemoveContainer" containerID="c706458164f7c6287d29e34e31517532fdb1cc00947568aaa9db8fa03c2c9c02" Mar 13 16:12:29 crc kubenswrapper[4907]: I0313 16:12:29.351487 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9q59x"] Mar 13 16:12:29 crc kubenswrapper[4907]: I0313 16:12:29.363065 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-9q59x"] Mar 13 16:12:29 crc kubenswrapper[4907]: I0313 16:12:29.375650 4907 scope.go:117] "RemoveContainer" containerID="4ec6ccaa677ccbbb66f21ece73e47ce24307c31346e8c97bf83dbdeb0ab4c535" Mar 13 16:12:29 crc kubenswrapper[4907]: I0313 16:12:29.437717 4907 scope.go:117] "RemoveContainer" containerID="d01c41ac19b3db32d5d13701af11e1e2508fcdc64db4c92f02188f251ddd2445" Mar 13 16:12:29 crc kubenswrapper[4907]: E0313 16:12:29.438984 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d01c41ac19b3db32d5d13701af11e1e2508fcdc64db4c92f02188f251ddd2445\": container with ID starting with d01c41ac19b3db32d5d13701af11e1e2508fcdc64db4c92f02188f251ddd2445 not found: ID does not exist" containerID="d01c41ac19b3db32d5d13701af11e1e2508fcdc64db4c92f02188f251ddd2445" Mar 13 16:12:29 crc kubenswrapper[4907]: I0313 16:12:29.439045 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d01c41ac19b3db32d5d13701af11e1e2508fcdc64db4c92f02188f251ddd2445"} err="failed to get container status \"d01c41ac19b3db32d5d13701af11e1e2508fcdc64db4c92f02188f251ddd2445\": rpc error: code = NotFound desc = could not find container \"d01c41ac19b3db32d5d13701af11e1e2508fcdc64db4c92f02188f251ddd2445\": container with ID starting with d01c41ac19b3db32d5d13701af11e1e2508fcdc64db4c92f02188f251ddd2445 not found: ID does not exist" Mar 13 16:12:29 crc kubenswrapper[4907]: I0313 16:12:29.439088 4907 scope.go:117] "RemoveContainer" containerID="c706458164f7c6287d29e34e31517532fdb1cc00947568aaa9db8fa03c2c9c02" Mar 13 16:12:29 crc kubenswrapper[4907]: E0313 16:12:29.439947 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c706458164f7c6287d29e34e31517532fdb1cc00947568aaa9db8fa03c2c9c02\": container with ID starting with c706458164f7c6287d29e34e31517532fdb1cc00947568aaa9db8fa03c2c9c02 not found: ID does not exist" containerID="c706458164f7c6287d29e34e31517532fdb1cc00947568aaa9db8fa03c2c9c02" Mar 13 16:12:29 crc kubenswrapper[4907]: I0313 16:12:29.440000 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c706458164f7c6287d29e34e31517532fdb1cc00947568aaa9db8fa03c2c9c02"} err="failed to get container status \"c706458164f7c6287d29e34e31517532fdb1cc00947568aaa9db8fa03c2c9c02\": rpc error: code = NotFound desc = could not find container \"c706458164f7c6287d29e34e31517532fdb1cc00947568aaa9db8fa03c2c9c02\": container with ID starting with c706458164f7c6287d29e34e31517532fdb1cc00947568aaa9db8fa03c2c9c02 not found: ID does not exist" Mar 13 16:12:29 crc kubenswrapper[4907]: I0313 16:12:29.440018 4907 scope.go:117] "RemoveContainer" containerID="4ec6ccaa677ccbbb66f21ece73e47ce24307c31346e8c97bf83dbdeb0ab4c535" Mar 13 16:12:29 crc kubenswrapper[4907]: E0313 16:12:29.440613 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4ec6ccaa677ccbbb66f21ece73e47ce24307c31346e8c97bf83dbdeb0ab4c535\": container with ID starting with 4ec6ccaa677ccbbb66f21ece73e47ce24307c31346e8c97bf83dbdeb0ab4c535 not found: ID does not exist" containerID="4ec6ccaa677ccbbb66f21ece73e47ce24307c31346e8c97bf83dbdeb0ab4c535" Mar 13 16:12:29 crc kubenswrapper[4907]: I0313 16:12:29.440644 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4ec6ccaa677ccbbb66f21ece73e47ce24307c31346e8c97bf83dbdeb0ab4c535"} err="failed to get container status \"4ec6ccaa677ccbbb66f21ece73e47ce24307c31346e8c97bf83dbdeb0ab4c535\": rpc error: code = NotFound desc = could not find container \"4ec6ccaa677ccbbb66f21ece73e47ce24307c31346e8c97bf83dbdeb0ab4c535\": container with ID starting with 4ec6ccaa677ccbbb66f21ece73e47ce24307c31346e8c97bf83dbdeb0ab4c535 not found: ID does not exist" Mar 13 16:12:29 crc kubenswrapper[4907]: I0313 16:12:29.793459 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aabfe118-3ad6-4625-bdab-6a1f2bc50890" path="/var/lib/kubelet/pods/aabfe118-3ad6-4625-bdab-6a1f2bc50890/volumes" Mar 13 16:12:36 crc kubenswrapper[4907]: I0313 16:12:36.390980 4907 generic.go:334] "Generic (PLEG): container finished" podID="34db6153-7522-4aae-9712-eb87cc9aa543" containerID="1a4edf4adc052232eadad9bc1d0620ff53938e2a203a59f9c95cd4e2f267081a" exitCode=0 Mar 13 16:12:36 crc kubenswrapper[4907]: I0313 16:12:36.391050 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-cell1-hml5k" event={"ID":"34db6153-7522-4aae-9712-eb87cc9aa543","Type":"ContainerDied","Data":"1a4edf4adc052232eadad9bc1d0620ff53938e2a203a59f9c95cd4e2f267081a"} Mar 13 16:12:37 crc kubenswrapper[4907]: I0313 16:12:37.926535 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-cell1-hml5k" Mar 13 16:12:38 crc kubenswrapper[4907]: I0313 16:12:38.042902 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/34db6153-7522-4aae-9712-eb87cc9aa543-inventory\") pod \"34db6153-7522-4aae-9712-eb87cc9aa543\" (UID: \"34db6153-7522-4aae-9712-eb87cc9aa543\") " Mar 13 16:12:38 crc kubenswrapper[4907]: I0313 16:12:38.043070 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/34db6153-7522-4aae-9712-eb87cc9aa543-ceph\") pod \"34db6153-7522-4aae-9712-eb87cc9aa543\" (UID: \"34db6153-7522-4aae-9712-eb87cc9aa543\") " Mar 13 16:12:38 crc kubenswrapper[4907]: I0313 16:12:38.043155 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7dstf\" (UniqueName: \"kubernetes.io/projected/34db6153-7522-4aae-9712-eb87cc9aa543-kube-api-access-7dstf\") pod \"34db6153-7522-4aae-9712-eb87cc9aa543\" (UID: \"34db6153-7522-4aae-9712-eb87cc9aa543\") " Mar 13 16:12:38 crc kubenswrapper[4907]: I0313 16:12:38.043186 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/34db6153-7522-4aae-9712-eb87cc9aa543-ssh-key-openstack-cell1\") pod \"34db6153-7522-4aae-9712-eb87cc9aa543\" (UID: \"34db6153-7522-4aae-9712-eb87cc9aa543\") " Mar 13 16:12:38 crc kubenswrapper[4907]: I0313 16:12:38.052166 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/34db6153-7522-4aae-9712-eb87cc9aa543-kube-api-access-7dstf" (OuterVolumeSpecName: "kube-api-access-7dstf") pod "34db6153-7522-4aae-9712-eb87cc9aa543" (UID: "34db6153-7522-4aae-9712-eb87cc9aa543"). InnerVolumeSpecName "kube-api-access-7dstf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:12:38 crc kubenswrapper[4907]: I0313 16:12:38.065168 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34db6153-7522-4aae-9712-eb87cc9aa543-ceph" (OuterVolumeSpecName: "ceph") pod "34db6153-7522-4aae-9712-eb87cc9aa543" (UID: "34db6153-7522-4aae-9712-eb87cc9aa543"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:12:38 crc kubenswrapper[4907]: I0313 16:12:38.080294 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34db6153-7522-4aae-9712-eb87cc9aa543-inventory" (OuterVolumeSpecName: "inventory") pod "34db6153-7522-4aae-9712-eb87cc9aa543" (UID: "34db6153-7522-4aae-9712-eb87cc9aa543"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:12:38 crc kubenswrapper[4907]: I0313 16:12:38.087212 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34db6153-7522-4aae-9712-eb87cc9aa543-ssh-key-openstack-cell1" (OuterVolumeSpecName: "ssh-key-openstack-cell1") pod "34db6153-7522-4aae-9712-eb87cc9aa543" (UID: "34db6153-7522-4aae-9712-eb87cc9aa543"). InnerVolumeSpecName "ssh-key-openstack-cell1". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:12:38 crc kubenswrapper[4907]: I0313 16:12:38.146153 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7dstf\" (UniqueName: \"kubernetes.io/projected/34db6153-7522-4aae-9712-eb87cc9aa543-kube-api-access-7dstf\") on node \"crc\" DevicePath \"\"" Mar 13 16:12:38 crc kubenswrapper[4907]: I0313 16:12:38.146554 4907 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/34db6153-7522-4aae-9712-eb87cc9aa543-ssh-key-openstack-cell1\") on node \"crc\" DevicePath \"\"" Mar 13 16:12:38 crc kubenswrapper[4907]: I0313 16:12:38.146575 4907 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/34db6153-7522-4aae-9712-eb87cc9aa543-inventory\") on node \"crc\" DevicePath \"\"" Mar 13 16:12:38 crc kubenswrapper[4907]: I0313 16:12:38.146596 4907 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/34db6153-7522-4aae-9712-eb87cc9aa543-ceph\") on node \"crc\" DevicePath \"\"" Mar 13 16:12:38 crc kubenswrapper[4907]: I0313 16:12:38.419098 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-cell1-hml5k" event={"ID":"34db6153-7522-4aae-9712-eb87cc9aa543","Type":"ContainerDied","Data":"1120479ce59841724e323313ca8d2a10c18000da39d4ac71ea46d45f1179803e"} Mar 13 16:12:38 crc kubenswrapper[4907]: I0313 16:12:38.419161 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1120479ce59841724e323313ca8d2a10c18000da39d4ac71ea46d45f1179803e" Mar 13 16:12:38 crc kubenswrapper[4907]: I0313 16:12:38.419195 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-cell1-hml5k" Mar 13 16:12:38 crc kubenswrapper[4907]: I0313 16:12:38.482642 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-openstack-openstack-cell1-x7hvj"] Mar 13 16:12:38 crc kubenswrapper[4907]: E0313 16:12:38.483110 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aabfe118-3ad6-4625-bdab-6a1f2bc50890" containerName="extract-content" Mar 13 16:12:38 crc kubenswrapper[4907]: I0313 16:12:38.483124 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="aabfe118-3ad6-4625-bdab-6a1f2bc50890" containerName="extract-content" Mar 13 16:12:38 crc kubenswrapper[4907]: E0313 16:12:38.483146 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aabfe118-3ad6-4625-bdab-6a1f2bc50890" containerName="extract-utilities" Mar 13 16:12:38 crc kubenswrapper[4907]: I0313 16:12:38.483151 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="aabfe118-3ad6-4625-bdab-6a1f2bc50890" containerName="extract-utilities" Mar 13 16:12:38 crc kubenswrapper[4907]: E0313 16:12:38.483176 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34db6153-7522-4aae-9712-eb87cc9aa543" containerName="run-os-openstack-openstack-cell1" Mar 13 16:12:38 crc kubenswrapper[4907]: I0313 16:12:38.483181 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="34db6153-7522-4aae-9712-eb87cc9aa543" containerName="run-os-openstack-openstack-cell1" Mar 13 16:12:38 crc kubenswrapper[4907]: E0313 16:12:38.483193 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aabfe118-3ad6-4625-bdab-6a1f2bc50890" containerName="registry-server" Mar 13 16:12:38 crc kubenswrapper[4907]: I0313 16:12:38.483199 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="aabfe118-3ad6-4625-bdab-6a1f2bc50890" containerName="registry-server" Mar 13 16:12:38 crc kubenswrapper[4907]: I0313 16:12:38.483388 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="aabfe118-3ad6-4625-bdab-6a1f2bc50890" containerName="registry-server" Mar 13 16:12:38 crc kubenswrapper[4907]: I0313 16:12:38.483406 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="34db6153-7522-4aae-9712-eb87cc9aa543" containerName="run-os-openstack-openstack-cell1" Mar 13 16:12:38 crc kubenswrapper[4907]: I0313 16:12:38.484177 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-cell1-x7hvj" Mar 13 16:12:38 crc kubenswrapper[4907]: I0313 16:12:38.489479 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Mar 13 16:12:38 crc kubenswrapper[4907]: I0313 16:12:38.489723 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-44kxw" Mar 13 16:12:38 crc kubenswrapper[4907]: I0313 16:12:38.489851 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 13 16:12:38 crc kubenswrapper[4907]: I0313 16:12:38.490246 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Mar 13 16:12:38 crc kubenswrapper[4907]: I0313 16:12:38.495106 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-openstack-openstack-cell1-x7hvj"] Mar 13 16:12:38 crc kubenswrapper[4907]: I0313 16:12:38.554038 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/cf92c1e7-1021-4405-840e-f2286adea31d-ceph\") pod \"reboot-os-openstack-openstack-cell1-x7hvj\" (UID: \"cf92c1e7-1021-4405-840e-f2286adea31d\") " pod="openstack/reboot-os-openstack-openstack-cell1-x7hvj" Mar 13 16:12:38 crc kubenswrapper[4907]: I0313 16:12:38.554114 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cf92c1e7-1021-4405-840e-f2286adea31d-inventory\") pod \"reboot-os-openstack-openstack-cell1-x7hvj\" (UID: \"cf92c1e7-1021-4405-840e-f2286adea31d\") " pod="openstack/reboot-os-openstack-openstack-cell1-x7hvj" Mar 13 16:12:38 crc kubenswrapper[4907]: I0313 16:12:38.554208 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fsql6\" (UniqueName: \"kubernetes.io/projected/cf92c1e7-1021-4405-840e-f2286adea31d-kube-api-access-fsql6\") pod \"reboot-os-openstack-openstack-cell1-x7hvj\" (UID: \"cf92c1e7-1021-4405-840e-f2286adea31d\") " pod="openstack/reboot-os-openstack-openstack-cell1-x7hvj" Mar 13 16:12:38 crc kubenswrapper[4907]: I0313 16:12:38.554721 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/cf92c1e7-1021-4405-840e-f2286adea31d-ssh-key-openstack-cell1\") pod \"reboot-os-openstack-openstack-cell1-x7hvj\" (UID: \"cf92c1e7-1021-4405-840e-f2286adea31d\") " pod="openstack/reboot-os-openstack-openstack-cell1-x7hvj" Mar 13 16:12:38 crc kubenswrapper[4907]: I0313 16:12:38.657668 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fsql6\" (UniqueName: \"kubernetes.io/projected/cf92c1e7-1021-4405-840e-f2286adea31d-kube-api-access-fsql6\") pod \"reboot-os-openstack-openstack-cell1-x7hvj\" (UID: \"cf92c1e7-1021-4405-840e-f2286adea31d\") " pod="openstack/reboot-os-openstack-openstack-cell1-x7hvj" Mar 13 16:12:38 crc kubenswrapper[4907]: I0313 16:12:38.658192 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/cf92c1e7-1021-4405-840e-f2286adea31d-ssh-key-openstack-cell1\") pod \"reboot-os-openstack-openstack-cell1-x7hvj\" (UID: \"cf92c1e7-1021-4405-840e-f2286adea31d\") " pod="openstack/reboot-os-openstack-openstack-cell1-x7hvj" Mar 13 16:12:38 crc kubenswrapper[4907]: I0313 16:12:38.658322 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/cf92c1e7-1021-4405-840e-f2286adea31d-ceph\") pod \"reboot-os-openstack-openstack-cell1-x7hvj\" (UID: \"cf92c1e7-1021-4405-840e-f2286adea31d\") " pod="openstack/reboot-os-openstack-openstack-cell1-x7hvj" Mar 13 16:12:38 crc kubenswrapper[4907]: I0313 16:12:38.658419 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cf92c1e7-1021-4405-840e-f2286adea31d-inventory\") pod \"reboot-os-openstack-openstack-cell1-x7hvj\" (UID: \"cf92c1e7-1021-4405-840e-f2286adea31d\") " pod="openstack/reboot-os-openstack-openstack-cell1-x7hvj" Mar 13 16:12:38 crc kubenswrapper[4907]: I0313 16:12:38.664210 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cf92c1e7-1021-4405-840e-f2286adea31d-inventory\") pod \"reboot-os-openstack-openstack-cell1-x7hvj\" (UID: \"cf92c1e7-1021-4405-840e-f2286adea31d\") " pod="openstack/reboot-os-openstack-openstack-cell1-x7hvj" Mar 13 16:12:38 crc kubenswrapper[4907]: I0313 16:12:38.664611 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/cf92c1e7-1021-4405-840e-f2286adea31d-ssh-key-openstack-cell1\") pod \"reboot-os-openstack-openstack-cell1-x7hvj\" (UID: \"cf92c1e7-1021-4405-840e-f2286adea31d\") " pod="openstack/reboot-os-openstack-openstack-cell1-x7hvj" Mar 13 16:12:38 crc kubenswrapper[4907]: I0313 16:12:38.666687 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/cf92c1e7-1021-4405-840e-f2286adea31d-ceph\") pod \"reboot-os-openstack-openstack-cell1-x7hvj\" (UID: \"cf92c1e7-1021-4405-840e-f2286adea31d\") " pod="openstack/reboot-os-openstack-openstack-cell1-x7hvj" Mar 13 16:12:38 crc kubenswrapper[4907]: I0313 16:12:38.681653 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fsql6\" (UniqueName: \"kubernetes.io/projected/cf92c1e7-1021-4405-840e-f2286adea31d-kube-api-access-fsql6\") pod \"reboot-os-openstack-openstack-cell1-x7hvj\" (UID: \"cf92c1e7-1021-4405-840e-f2286adea31d\") " pod="openstack/reboot-os-openstack-openstack-cell1-x7hvj" Mar 13 16:12:38 crc kubenswrapper[4907]: I0313 16:12:38.803176 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-cell1-x7hvj" Mar 13 16:12:39 crc kubenswrapper[4907]: I0313 16:12:39.326493 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-openstack-openstack-cell1-x7hvj"] Mar 13 16:12:39 crc kubenswrapper[4907]: I0313 16:12:39.428753 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-cell1-x7hvj" event={"ID":"cf92c1e7-1021-4405-840e-f2286adea31d","Type":"ContainerStarted","Data":"895e8f2c18c5de20a20d4cb067c8d1021996b85a83d2443bfe64b2080b187ce0"} Mar 13 16:12:40 crc kubenswrapper[4907]: I0313 16:12:40.443257 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-cell1-x7hvj" event={"ID":"cf92c1e7-1021-4405-840e-f2286adea31d","Type":"ContainerStarted","Data":"84b83daed0821729d3e57629749b6c8b5286ebc80f25b88784c7bd96792460bd"} Mar 13 16:12:41 crc kubenswrapper[4907]: I0313 16:12:41.427066 4907 scope.go:117] "RemoveContainer" containerID="fba4ffaa28135dbeabe5c056323b0e99fffec19d264cb5778eaf0cb6c3cc2d40" Mar 13 16:12:48 crc kubenswrapper[4907]: I0313 16:12:48.041516 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 16:12:48 crc kubenswrapper[4907]: I0313 16:12:48.041961 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 16:12:48 crc kubenswrapper[4907]: I0313 16:12:48.042007 4907 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" Mar 13 16:12:48 crc kubenswrapper[4907]: I0313 16:12:48.042789 4907 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c38b8b19c660dee13d11b404db7a1f6a1a7c1615cf40ea14a2875b4e3648d560"} pod="openshift-machine-config-operator/machine-config-daemon-hm56j" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 16:12:48 crc kubenswrapper[4907]: I0313 16:12:48.042843 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" containerID="cri-o://c38b8b19c660dee13d11b404db7a1f6a1a7c1615cf40ea14a2875b4e3648d560" gracePeriod=600 Mar 13 16:12:48 crc kubenswrapper[4907]: I0313 16:12:48.522093 4907 generic.go:334] "Generic (PLEG): container finished" podID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerID="c38b8b19c660dee13d11b404db7a1f6a1a7c1615cf40ea14a2875b4e3648d560" exitCode=0 Mar 13 16:12:48 crc kubenswrapper[4907]: I0313 16:12:48.522580 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerDied","Data":"c38b8b19c660dee13d11b404db7a1f6a1a7c1615cf40ea14a2875b4e3648d560"} Mar 13 16:12:48 crc kubenswrapper[4907]: I0313 16:12:48.522679 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerStarted","Data":"300e429cce24fa4eb9aefff9494fe9905ee7b0767504365170f1f9645a3c8b5f"} Mar 13 16:12:48 crc kubenswrapper[4907]: I0313 16:12:48.522708 4907 scope.go:117] "RemoveContainer" containerID="97ad1b399368adb5c1fc3c4cf2810fa240ec320625c36e99f246d6171b951964" Mar 13 16:12:48 crc kubenswrapper[4907]: I0313 16:12:48.543194 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-openstack-openstack-cell1-x7hvj" podStartSLOduration=10.052340781 podStartE2EDuration="10.543178325s" podCreationTimestamp="2026-03-13 16:12:38 +0000 UTC" firstStartedPulling="2026-03-13 16:12:39.331122958 +0000 UTC m=+7658.230910657" lastFinishedPulling="2026-03-13 16:12:39.821960512 +0000 UTC m=+7658.721748201" observedRunningTime="2026-03-13 16:12:40.469095794 +0000 UTC m=+7659.368883513" watchObservedRunningTime="2026-03-13 16:12:48.543178325 +0000 UTC m=+7667.442966014" Mar 13 16:12:56 crc kubenswrapper[4907]: I0313 16:12:56.633838 4907 generic.go:334] "Generic (PLEG): container finished" podID="cf92c1e7-1021-4405-840e-f2286adea31d" containerID="84b83daed0821729d3e57629749b6c8b5286ebc80f25b88784c7bd96792460bd" exitCode=0 Mar 13 16:12:56 crc kubenswrapper[4907]: I0313 16:12:56.633941 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-cell1-x7hvj" event={"ID":"cf92c1e7-1021-4405-840e-f2286adea31d","Type":"ContainerDied","Data":"84b83daed0821729d3e57629749b6c8b5286ebc80f25b88784c7bd96792460bd"} Mar 13 16:12:58 crc kubenswrapper[4907]: I0313 16:12:58.196337 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-cell1-x7hvj" Mar 13 16:12:58 crc kubenswrapper[4907]: I0313 16:12:58.286395 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/cf92c1e7-1021-4405-840e-f2286adea31d-ceph\") pod \"cf92c1e7-1021-4405-840e-f2286adea31d\" (UID: \"cf92c1e7-1021-4405-840e-f2286adea31d\") " Mar 13 16:12:58 crc kubenswrapper[4907]: I0313 16:12:58.286525 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cf92c1e7-1021-4405-840e-f2286adea31d-inventory\") pod \"cf92c1e7-1021-4405-840e-f2286adea31d\" (UID: \"cf92c1e7-1021-4405-840e-f2286adea31d\") " Mar 13 16:12:58 crc kubenswrapper[4907]: I0313 16:12:58.286670 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/cf92c1e7-1021-4405-840e-f2286adea31d-ssh-key-openstack-cell1\") pod \"cf92c1e7-1021-4405-840e-f2286adea31d\" (UID: \"cf92c1e7-1021-4405-840e-f2286adea31d\") " Mar 13 16:12:58 crc kubenswrapper[4907]: I0313 16:12:58.286702 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fsql6\" (UniqueName: \"kubernetes.io/projected/cf92c1e7-1021-4405-840e-f2286adea31d-kube-api-access-fsql6\") pod \"cf92c1e7-1021-4405-840e-f2286adea31d\" (UID: \"cf92c1e7-1021-4405-840e-f2286adea31d\") " Mar 13 16:12:58 crc kubenswrapper[4907]: I0313 16:12:58.293651 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cf92c1e7-1021-4405-840e-f2286adea31d-kube-api-access-fsql6" (OuterVolumeSpecName: "kube-api-access-fsql6") pod "cf92c1e7-1021-4405-840e-f2286adea31d" (UID: "cf92c1e7-1021-4405-840e-f2286adea31d"). InnerVolumeSpecName "kube-api-access-fsql6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:12:58 crc kubenswrapper[4907]: I0313 16:12:58.294779 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cf92c1e7-1021-4405-840e-f2286adea31d-ceph" (OuterVolumeSpecName: "ceph") pod "cf92c1e7-1021-4405-840e-f2286adea31d" (UID: "cf92c1e7-1021-4405-840e-f2286adea31d"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:12:58 crc kubenswrapper[4907]: I0313 16:12:58.317104 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cf92c1e7-1021-4405-840e-f2286adea31d-inventory" (OuterVolumeSpecName: "inventory") pod "cf92c1e7-1021-4405-840e-f2286adea31d" (UID: "cf92c1e7-1021-4405-840e-f2286adea31d"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:12:58 crc kubenswrapper[4907]: I0313 16:12:58.321829 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cf92c1e7-1021-4405-840e-f2286adea31d-ssh-key-openstack-cell1" (OuterVolumeSpecName: "ssh-key-openstack-cell1") pod "cf92c1e7-1021-4405-840e-f2286adea31d" (UID: "cf92c1e7-1021-4405-840e-f2286adea31d"). InnerVolumeSpecName "ssh-key-openstack-cell1". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:12:58 crc kubenswrapper[4907]: I0313 16:12:58.390251 4907 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/cf92c1e7-1021-4405-840e-f2286adea31d-ssh-key-openstack-cell1\") on node \"crc\" DevicePath \"\"" Mar 13 16:12:58 crc kubenswrapper[4907]: I0313 16:12:58.390472 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fsql6\" (UniqueName: \"kubernetes.io/projected/cf92c1e7-1021-4405-840e-f2286adea31d-kube-api-access-fsql6\") on node \"crc\" DevicePath \"\"" Mar 13 16:12:58 crc kubenswrapper[4907]: I0313 16:12:58.390547 4907 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/cf92c1e7-1021-4405-840e-f2286adea31d-ceph\") on node \"crc\" DevicePath \"\"" Mar 13 16:12:58 crc kubenswrapper[4907]: I0313 16:12:58.390621 4907 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cf92c1e7-1021-4405-840e-f2286adea31d-inventory\") on node \"crc\" DevicePath \"\"" Mar 13 16:12:58 crc kubenswrapper[4907]: I0313 16:12:58.656189 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-cell1-x7hvj" event={"ID":"cf92c1e7-1021-4405-840e-f2286adea31d","Type":"ContainerDied","Data":"895e8f2c18c5de20a20d4cb067c8d1021996b85a83d2443bfe64b2080b187ce0"} Mar 13 16:12:58 crc kubenswrapper[4907]: I0313 16:12:58.656439 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="895e8f2c18c5de20a20d4cb067c8d1021996b85a83d2443bfe64b2080b187ce0" Mar 13 16:12:58 crc kubenswrapper[4907]: I0313 16:12:58.656457 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-cell1-x7hvj" Mar 13 16:12:58 crc kubenswrapper[4907]: I0313 16:12:58.753418 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-openstack-openstack-cell1-55pht"] Mar 13 16:12:58 crc kubenswrapper[4907]: E0313 16:12:58.754097 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf92c1e7-1021-4405-840e-f2286adea31d" containerName="reboot-os-openstack-openstack-cell1" Mar 13 16:12:58 crc kubenswrapper[4907]: I0313 16:12:58.754122 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf92c1e7-1021-4405-840e-f2286adea31d" containerName="reboot-os-openstack-openstack-cell1" Mar 13 16:12:58 crc kubenswrapper[4907]: I0313 16:12:58.754317 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="cf92c1e7-1021-4405-840e-f2286adea31d" containerName="reboot-os-openstack-openstack-cell1" Mar 13 16:12:58 crc kubenswrapper[4907]: I0313 16:12:58.755161 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-cell1-55pht" Mar 13 16:12:58 crc kubenswrapper[4907]: I0313 16:12:58.757417 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 13 16:12:58 crc kubenswrapper[4907]: I0313 16:12:58.757780 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Mar 13 16:12:58 crc kubenswrapper[4907]: I0313 16:12:58.760902 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Mar 13 16:12:58 crc kubenswrapper[4907]: I0313 16:12:58.763546 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-44kxw" Mar 13 16:12:58 crc kubenswrapper[4907]: I0313 16:12:58.771358 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-openstack-openstack-cell1-55pht"] Mar 13 16:12:58 crc kubenswrapper[4907]: I0313 16:12:58.900084 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-libvirt-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-55pht\" (UID: \"c32eadac-12ec-49b8-930e-bc6af99631e8\") " pod="openstack/install-certs-openstack-openstack-cell1-55pht" Mar 13 16:12:58 crc kubenswrapper[4907]: I0313 16:12:58.900130 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-neutron-dhcp-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-55pht\" (UID: \"c32eadac-12ec-49b8-930e-bc6af99631e8\") " pod="openstack/install-certs-openstack-openstack-cell1-55pht" Mar 13 16:12:58 crc kubenswrapper[4907]: I0313 16:12:58.900174 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-ceph\") pod \"install-certs-openstack-openstack-cell1-55pht\" (UID: \"c32eadac-12ec-49b8-930e-bc6af99631e8\") " pod="openstack/install-certs-openstack-openstack-cell1-55pht" Mar 13 16:12:58 crc kubenswrapper[4907]: I0313 16:12:58.900278 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-neutron-metadata-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-55pht\" (UID: \"c32eadac-12ec-49b8-930e-bc6af99631e8\") " pod="openstack/install-certs-openstack-openstack-cell1-55pht" Mar 13 16:12:58 crc kubenswrapper[4907]: I0313 16:12:58.900306 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-inventory\") pod \"install-certs-openstack-openstack-cell1-55pht\" (UID: \"c32eadac-12ec-49b8-930e-bc6af99631e8\") " pod="openstack/install-certs-openstack-openstack-cell1-55pht" Mar 13 16:12:58 crc kubenswrapper[4907]: I0313 16:12:58.900399 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-ssh-key-openstack-cell1\") pod \"install-certs-openstack-openstack-cell1-55pht\" (UID: \"c32eadac-12ec-49b8-930e-bc6af99631e8\") " pod="openstack/install-certs-openstack-openstack-cell1-55pht" Mar 13 16:12:58 crc kubenswrapper[4907]: I0313 16:12:58.900462 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qrltz\" (UniqueName: \"kubernetes.io/projected/c32eadac-12ec-49b8-930e-bc6af99631e8-kube-api-access-qrltz\") pod \"install-certs-openstack-openstack-cell1-55pht\" (UID: \"c32eadac-12ec-49b8-930e-bc6af99631e8\") " pod="openstack/install-certs-openstack-openstack-cell1-55pht" Mar 13 16:12:58 crc kubenswrapper[4907]: I0313 16:12:58.900483 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-neutron-sriov-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-55pht\" (UID: \"c32eadac-12ec-49b8-930e-bc6af99631e8\") " pod="openstack/install-certs-openstack-openstack-cell1-55pht" Mar 13 16:12:58 crc kubenswrapper[4907]: I0313 16:12:58.900514 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-ovn-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-55pht\" (UID: \"c32eadac-12ec-49b8-930e-bc6af99631e8\") " pod="openstack/install-certs-openstack-openstack-cell1-55pht" Mar 13 16:12:58 crc kubenswrapper[4907]: I0313 16:12:58.900542 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-bootstrap-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-55pht\" (UID: \"c32eadac-12ec-49b8-930e-bc6af99631e8\") " pod="openstack/install-certs-openstack-openstack-cell1-55pht" Mar 13 16:12:58 crc kubenswrapper[4907]: I0313 16:12:58.900595 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-nova-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-55pht\" (UID: \"c32eadac-12ec-49b8-930e-bc6af99631e8\") " pod="openstack/install-certs-openstack-openstack-cell1-55pht" Mar 13 16:12:58 crc kubenswrapper[4907]: I0313 16:12:58.900614 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-telemetry-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-55pht\" (UID: \"c32eadac-12ec-49b8-930e-bc6af99631e8\") " pod="openstack/install-certs-openstack-openstack-cell1-55pht" Mar 13 16:12:59 crc kubenswrapper[4907]: I0313 16:12:59.002540 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qrltz\" (UniqueName: \"kubernetes.io/projected/c32eadac-12ec-49b8-930e-bc6af99631e8-kube-api-access-qrltz\") pod \"install-certs-openstack-openstack-cell1-55pht\" (UID: \"c32eadac-12ec-49b8-930e-bc6af99631e8\") " pod="openstack/install-certs-openstack-openstack-cell1-55pht" Mar 13 16:12:59 crc kubenswrapper[4907]: I0313 16:12:59.002585 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-neutron-sriov-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-55pht\" (UID: \"c32eadac-12ec-49b8-930e-bc6af99631e8\") " pod="openstack/install-certs-openstack-openstack-cell1-55pht" Mar 13 16:12:59 crc kubenswrapper[4907]: I0313 16:12:59.002614 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-ovn-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-55pht\" (UID: \"c32eadac-12ec-49b8-930e-bc6af99631e8\") " pod="openstack/install-certs-openstack-openstack-cell1-55pht" Mar 13 16:12:59 crc kubenswrapper[4907]: I0313 16:12:59.002641 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-bootstrap-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-55pht\" (UID: \"c32eadac-12ec-49b8-930e-bc6af99631e8\") " pod="openstack/install-certs-openstack-openstack-cell1-55pht" Mar 13 16:12:59 crc kubenswrapper[4907]: I0313 16:12:59.002671 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-nova-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-55pht\" (UID: \"c32eadac-12ec-49b8-930e-bc6af99631e8\") " pod="openstack/install-certs-openstack-openstack-cell1-55pht" Mar 13 16:12:59 crc kubenswrapper[4907]: I0313 16:12:59.002694 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-telemetry-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-55pht\" (UID: \"c32eadac-12ec-49b8-930e-bc6af99631e8\") " pod="openstack/install-certs-openstack-openstack-cell1-55pht" Mar 13 16:12:59 crc kubenswrapper[4907]: I0313 16:12:59.002738 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-libvirt-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-55pht\" (UID: \"c32eadac-12ec-49b8-930e-bc6af99631e8\") " pod="openstack/install-certs-openstack-openstack-cell1-55pht" Mar 13 16:12:59 crc kubenswrapper[4907]: I0313 16:12:59.002758 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-neutron-dhcp-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-55pht\" (UID: \"c32eadac-12ec-49b8-930e-bc6af99631e8\") " pod="openstack/install-certs-openstack-openstack-cell1-55pht" Mar 13 16:12:59 crc kubenswrapper[4907]: I0313 16:12:59.002780 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-ceph\") pod \"install-certs-openstack-openstack-cell1-55pht\" (UID: \"c32eadac-12ec-49b8-930e-bc6af99631e8\") " pod="openstack/install-certs-openstack-openstack-cell1-55pht" Mar 13 16:12:59 crc kubenswrapper[4907]: I0313 16:12:59.002839 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-neutron-metadata-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-55pht\" (UID: \"c32eadac-12ec-49b8-930e-bc6af99631e8\") " pod="openstack/install-certs-openstack-openstack-cell1-55pht" Mar 13 16:12:59 crc kubenswrapper[4907]: I0313 16:12:59.002867 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-inventory\") pod \"install-certs-openstack-openstack-cell1-55pht\" (UID: \"c32eadac-12ec-49b8-930e-bc6af99631e8\") " pod="openstack/install-certs-openstack-openstack-cell1-55pht" Mar 13 16:12:59 crc kubenswrapper[4907]: I0313 16:12:59.002941 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-ssh-key-openstack-cell1\") pod \"install-certs-openstack-openstack-cell1-55pht\" (UID: \"c32eadac-12ec-49b8-930e-bc6af99631e8\") " pod="openstack/install-certs-openstack-openstack-cell1-55pht" Mar 13 16:12:59 crc kubenswrapper[4907]: I0313 16:12:59.007656 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-ssh-key-openstack-cell1\") pod \"install-certs-openstack-openstack-cell1-55pht\" (UID: \"c32eadac-12ec-49b8-930e-bc6af99631e8\") " pod="openstack/install-certs-openstack-openstack-cell1-55pht" Mar 13 16:12:59 crc kubenswrapper[4907]: I0313 16:12:59.008697 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-telemetry-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-55pht\" (UID: \"c32eadac-12ec-49b8-930e-bc6af99631e8\") " pod="openstack/install-certs-openstack-openstack-cell1-55pht" Mar 13 16:12:59 crc kubenswrapper[4907]: I0313 16:12:59.009045 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-bootstrap-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-55pht\" (UID: \"c32eadac-12ec-49b8-930e-bc6af99631e8\") " pod="openstack/install-certs-openstack-openstack-cell1-55pht" Mar 13 16:12:59 crc kubenswrapper[4907]: I0313 16:12:59.009548 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-neutron-metadata-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-55pht\" (UID: \"c32eadac-12ec-49b8-930e-bc6af99631e8\") " pod="openstack/install-certs-openstack-openstack-cell1-55pht" Mar 13 16:12:59 crc kubenswrapper[4907]: I0313 16:12:59.009559 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-nova-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-55pht\" (UID: \"c32eadac-12ec-49b8-930e-bc6af99631e8\") " pod="openstack/install-certs-openstack-openstack-cell1-55pht" Mar 13 16:12:59 crc kubenswrapper[4907]: I0313 16:12:59.010559 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-neutron-sriov-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-55pht\" (UID: \"c32eadac-12ec-49b8-930e-bc6af99631e8\") " pod="openstack/install-certs-openstack-openstack-cell1-55pht" Mar 13 16:12:59 crc kubenswrapper[4907]: I0313 16:12:59.011290 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-libvirt-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-55pht\" (UID: \"c32eadac-12ec-49b8-930e-bc6af99631e8\") " pod="openstack/install-certs-openstack-openstack-cell1-55pht" Mar 13 16:12:59 crc kubenswrapper[4907]: I0313 16:12:59.011628 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-inventory\") pod \"install-certs-openstack-openstack-cell1-55pht\" (UID: \"c32eadac-12ec-49b8-930e-bc6af99631e8\") " pod="openstack/install-certs-openstack-openstack-cell1-55pht" Mar 13 16:12:59 crc kubenswrapper[4907]: I0313 16:12:59.011659 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-ovn-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-55pht\" (UID: \"c32eadac-12ec-49b8-930e-bc6af99631e8\") " pod="openstack/install-certs-openstack-openstack-cell1-55pht" Mar 13 16:12:59 crc kubenswrapper[4907]: I0313 16:12:59.012119 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-neutron-dhcp-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-55pht\" (UID: \"c32eadac-12ec-49b8-930e-bc6af99631e8\") " pod="openstack/install-certs-openstack-openstack-cell1-55pht" Mar 13 16:12:59 crc kubenswrapper[4907]: I0313 16:12:59.012488 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-ceph\") pod \"install-certs-openstack-openstack-cell1-55pht\" (UID: \"c32eadac-12ec-49b8-930e-bc6af99631e8\") " pod="openstack/install-certs-openstack-openstack-cell1-55pht" Mar 13 16:12:59 crc kubenswrapper[4907]: I0313 16:12:59.021762 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qrltz\" (UniqueName: \"kubernetes.io/projected/c32eadac-12ec-49b8-930e-bc6af99631e8-kube-api-access-qrltz\") pod \"install-certs-openstack-openstack-cell1-55pht\" (UID: \"c32eadac-12ec-49b8-930e-bc6af99631e8\") " pod="openstack/install-certs-openstack-openstack-cell1-55pht" Mar 13 16:12:59 crc kubenswrapper[4907]: I0313 16:12:59.071970 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-cell1-55pht" Mar 13 16:12:59 crc kubenswrapper[4907]: I0313 16:12:59.697514 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-openstack-openstack-cell1-55pht"] Mar 13 16:13:00 crc kubenswrapper[4907]: I0313 16:13:00.674300 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-cell1-55pht" event={"ID":"c32eadac-12ec-49b8-930e-bc6af99631e8","Type":"ContainerStarted","Data":"c3d0a8f0dcc500de771fd7594059297f23bc149de65101bf8d50df48f2d537c6"} Mar 13 16:13:01 crc kubenswrapper[4907]: I0313 16:13:01.683544 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-cell1-55pht" event={"ID":"c32eadac-12ec-49b8-930e-bc6af99631e8","Type":"ContainerStarted","Data":"3281f97dcc70691f05829d00e2c76abe637c47e042b459a53a28abfa20ebaf42"} Mar 13 16:13:01 crc kubenswrapper[4907]: I0313 16:13:01.710098 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-openstack-openstack-cell1-55pht" podStartSLOduration=2.907388228 podStartE2EDuration="3.710077637s" podCreationTimestamp="2026-03-13 16:12:58 +0000 UTC" firstStartedPulling="2026-03-13 16:12:59.706589795 +0000 UTC m=+7678.606377484" lastFinishedPulling="2026-03-13 16:13:00.509279204 +0000 UTC m=+7679.409066893" observedRunningTime="2026-03-13 16:13:01.70580821 +0000 UTC m=+7680.605595909" watchObservedRunningTime="2026-03-13 16:13:01.710077637 +0000 UTC m=+7680.609865326" Mar 13 16:13:19 crc kubenswrapper[4907]: I0313 16:13:19.887085 4907 generic.go:334] "Generic (PLEG): container finished" podID="c32eadac-12ec-49b8-930e-bc6af99631e8" containerID="3281f97dcc70691f05829d00e2c76abe637c47e042b459a53a28abfa20ebaf42" exitCode=0 Mar 13 16:13:19 crc kubenswrapper[4907]: I0313 16:13:19.887186 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-cell1-55pht" event={"ID":"c32eadac-12ec-49b8-930e-bc6af99631e8","Type":"ContainerDied","Data":"3281f97dcc70691f05829d00e2c76abe637c47e042b459a53a28abfa20ebaf42"} Mar 13 16:13:21 crc kubenswrapper[4907]: I0313 16:13:21.456432 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-cell1-55pht" Mar 13 16:13:21 crc kubenswrapper[4907]: I0313 16:13:21.486640 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-bootstrap-combined-ca-bundle\") pod \"c32eadac-12ec-49b8-930e-bc6af99631e8\" (UID: \"c32eadac-12ec-49b8-930e-bc6af99631e8\") " Mar 13 16:13:21 crc kubenswrapper[4907]: I0313 16:13:21.486704 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-ssh-key-openstack-cell1\") pod \"c32eadac-12ec-49b8-930e-bc6af99631e8\" (UID: \"c32eadac-12ec-49b8-930e-bc6af99631e8\") " Mar 13 16:13:21 crc kubenswrapper[4907]: I0313 16:13:21.486769 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-neutron-dhcp-combined-ca-bundle\") pod \"c32eadac-12ec-49b8-930e-bc6af99631e8\" (UID: \"c32eadac-12ec-49b8-930e-bc6af99631e8\") " Mar 13 16:13:21 crc kubenswrapper[4907]: I0313 16:13:21.486795 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-inventory\") pod \"c32eadac-12ec-49b8-930e-bc6af99631e8\" (UID: \"c32eadac-12ec-49b8-930e-bc6af99631e8\") " Mar 13 16:13:21 crc kubenswrapper[4907]: I0313 16:13:21.486820 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qrltz\" (UniqueName: \"kubernetes.io/projected/c32eadac-12ec-49b8-930e-bc6af99631e8-kube-api-access-qrltz\") pod \"c32eadac-12ec-49b8-930e-bc6af99631e8\" (UID: \"c32eadac-12ec-49b8-930e-bc6af99631e8\") " Mar 13 16:13:21 crc kubenswrapper[4907]: I0313 16:13:21.486892 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-libvirt-combined-ca-bundle\") pod \"c32eadac-12ec-49b8-930e-bc6af99631e8\" (UID: \"c32eadac-12ec-49b8-930e-bc6af99631e8\") " Mar 13 16:13:21 crc kubenswrapper[4907]: I0313 16:13:21.492833 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "c32eadac-12ec-49b8-930e-bc6af99631e8" (UID: "c32eadac-12ec-49b8-930e-bc6af99631e8"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:13:21 crc kubenswrapper[4907]: I0313 16:13:21.496311 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c32eadac-12ec-49b8-930e-bc6af99631e8-kube-api-access-qrltz" (OuterVolumeSpecName: "kube-api-access-qrltz") pod "c32eadac-12ec-49b8-930e-bc6af99631e8" (UID: "c32eadac-12ec-49b8-930e-bc6af99631e8"). InnerVolumeSpecName "kube-api-access-qrltz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:13:21 crc kubenswrapper[4907]: I0313 16:13:21.502203 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "c32eadac-12ec-49b8-930e-bc6af99631e8" (UID: "c32eadac-12ec-49b8-930e-bc6af99631e8"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:13:21 crc kubenswrapper[4907]: I0313 16:13:21.518035 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-neutron-dhcp-combined-ca-bundle" (OuterVolumeSpecName: "neutron-dhcp-combined-ca-bundle") pod "c32eadac-12ec-49b8-930e-bc6af99631e8" (UID: "c32eadac-12ec-49b8-930e-bc6af99631e8"). InnerVolumeSpecName "neutron-dhcp-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:13:21 crc kubenswrapper[4907]: I0313 16:13:21.520608 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-ssh-key-openstack-cell1" (OuterVolumeSpecName: "ssh-key-openstack-cell1") pod "c32eadac-12ec-49b8-930e-bc6af99631e8" (UID: "c32eadac-12ec-49b8-930e-bc6af99631e8"). InnerVolumeSpecName "ssh-key-openstack-cell1". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:13:21 crc kubenswrapper[4907]: I0313 16:13:21.542658 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-inventory" (OuterVolumeSpecName: "inventory") pod "c32eadac-12ec-49b8-930e-bc6af99631e8" (UID: "c32eadac-12ec-49b8-930e-bc6af99631e8"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:13:21 crc kubenswrapper[4907]: I0313 16:13:21.588649 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-ceph\") pod \"c32eadac-12ec-49b8-930e-bc6af99631e8\" (UID: \"c32eadac-12ec-49b8-930e-bc6af99631e8\") " Mar 13 16:13:21 crc kubenswrapper[4907]: I0313 16:13:21.588962 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-telemetry-combined-ca-bundle\") pod \"c32eadac-12ec-49b8-930e-bc6af99631e8\" (UID: \"c32eadac-12ec-49b8-930e-bc6af99631e8\") " Mar 13 16:13:21 crc kubenswrapper[4907]: I0313 16:13:21.589009 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-neutron-metadata-combined-ca-bundle\") pod \"c32eadac-12ec-49b8-930e-bc6af99631e8\" (UID: \"c32eadac-12ec-49b8-930e-bc6af99631e8\") " Mar 13 16:13:21 crc kubenswrapper[4907]: I0313 16:13:21.589052 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-neutron-sriov-combined-ca-bundle\") pod \"c32eadac-12ec-49b8-930e-bc6af99631e8\" (UID: \"c32eadac-12ec-49b8-930e-bc6af99631e8\") " Mar 13 16:13:21 crc kubenswrapper[4907]: I0313 16:13:21.589092 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-ovn-combined-ca-bundle\") pod \"c32eadac-12ec-49b8-930e-bc6af99631e8\" (UID: \"c32eadac-12ec-49b8-930e-bc6af99631e8\") " Mar 13 16:13:21 crc kubenswrapper[4907]: I0313 16:13:21.589176 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-nova-combined-ca-bundle\") pod \"c32eadac-12ec-49b8-930e-bc6af99631e8\" (UID: \"c32eadac-12ec-49b8-930e-bc6af99631e8\") " Mar 13 16:13:21 crc kubenswrapper[4907]: I0313 16:13:21.589854 4907 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 16:13:21 crc kubenswrapper[4907]: I0313 16:13:21.589873 4907 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-ssh-key-openstack-cell1\") on node \"crc\" DevicePath \"\"" Mar 13 16:13:21 crc kubenswrapper[4907]: I0313 16:13:21.589899 4907 reconciler_common.go:293] "Volume detached for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-neutron-dhcp-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 16:13:21 crc kubenswrapper[4907]: I0313 16:13:21.589910 4907 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-inventory\") on node \"crc\" DevicePath \"\"" Mar 13 16:13:21 crc kubenswrapper[4907]: I0313 16:13:21.589919 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qrltz\" (UniqueName: \"kubernetes.io/projected/c32eadac-12ec-49b8-930e-bc6af99631e8-kube-api-access-qrltz\") on node \"crc\" DevicePath \"\"" Mar 13 16:13:21 crc kubenswrapper[4907]: I0313 16:13:21.589928 4907 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 16:13:21 crc kubenswrapper[4907]: I0313 16:13:21.592452 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "c32eadac-12ec-49b8-930e-bc6af99631e8" (UID: "c32eadac-12ec-49b8-930e-bc6af99631e8"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:13:21 crc kubenswrapper[4907]: I0313 16:13:21.592661 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-ceph" (OuterVolumeSpecName: "ceph") pod "c32eadac-12ec-49b8-930e-bc6af99631e8" (UID: "c32eadac-12ec-49b8-930e-bc6af99631e8"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:13:21 crc kubenswrapper[4907]: I0313 16:13:21.593121 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "c32eadac-12ec-49b8-930e-bc6af99631e8" (UID: "c32eadac-12ec-49b8-930e-bc6af99631e8"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:13:21 crc kubenswrapper[4907]: I0313 16:13:21.593233 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "c32eadac-12ec-49b8-930e-bc6af99631e8" (UID: "c32eadac-12ec-49b8-930e-bc6af99631e8"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:13:21 crc kubenswrapper[4907]: I0313 16:13:21.593859 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-neutron-sriov-combined-ca-bundle" (OuterVolumeSpecName: "neutron-sriov-combined-ca-bundle") pod "c32eadac-12ec-49b8-930e-bc6af99631e8" (UID: "c32eadac-12ec-49b8-930e-bc6af99631e8"). InnerVolumeSpecName "neutron-sriov-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:13:21 crc kubenswrapper[4907]: I0313 16:13:21.595141 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "c32eadac-12ec-49b8-930e-bc6af99631e8" (UID: "c32eadac-12ec-49b8-930e-bc6af99631e8"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:13:21 crc kubenswrapper[4907]: I0313 16:13:21.692241 4907 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 16:13:21 crc kubenswrapper[4907]: I0313 16:13:21.692307 4907 reconciler_common.go:293] "Volume detached for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-neutron-sriov-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 16:13:21 crc kubenswrapper[4907]: I0313 16:13:21.692321 4907 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 16:13:21 crc kubenswrapper[4907]: I0313 16:13:21.692333 4907 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 16:13:21 crc kubenswrapper[4907]: I0313 16:13:21.692344 4907 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-ceph\") on node \"crc\" DevicePath \"\"" Mar 13 16:13:21 crc kubenswrapper[4907]: I0313 16:13:21.692353 4907 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 16:13:21 crc kubenswrapper[4907]: I0313 16:13:21.909642 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-cell1-55pht" event={"ID":"c32eadac-12ec-49b8-930e-bc6af99631e8","Type":"ContainerDied","Data":"c3d0a8f0dcc500de771fd7594059297f23bc149de65101bf8d50df48f2d537c6"} Mar 13 16:13:21 crc kubenswrapper[4907]: I0313 16:13:21.909688 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c3d0a8f0dcc500de771fd7594059297f23bc149de65101bf8d50df48f2d537c6" Mar 13 16:13:21 crc kubenswrapper[4907]: I0313 16:13:21.909719 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-cell1-55pht" Mar 13 16:13:22 crc kubenswrapper[4907]: I0313 16:13:22.004061 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceph-client-openstack-openstack-cell1-tcvvp"] Mar 13 16:13:22 crc kubenswrapper[4907]: E0313 16:13:22.004715 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c32eadac-12ec-49b8-930e-bc6af99631e8" containerName="install-certs-openstack-openstack-cell1" Mar 13 16:13:22 crc kubenswrapper[4907]: I0313 16:13:22.004740 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="c32eadac-12ec-49b8-930e-bc6af99631e8" containerName="install-certs-openstack-openstack-cell1" Mar 13 16:13:22 crc kubenswrapper[4907]: I0313 16:13:22.005066 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="c32eadac-12ec-49b8-930e-bc6af99631e8" containerName="install-certs-openstack-openstack-cell1" Mar 13 16:13:22 crc kubenswrapper[4907]: I0313 16:13:22.006080 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-openstack-openstack-cell1-tcvvp" Mar 13 16:13:22 crc kubenswrapper[4907]: I0313 16:13:22.010776 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-44kxw" Mar 13 16:13:22 crc kubenswrapper[4907]: I0313 16:13:22.011474 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 13 16:13:22 crc kubenswrapper[4907]: I0313 16:13:22.012138 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Mar 13 16:13:22 crc kubenswrapper[4907]: I0313 16:13:22.013820 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Mar 13 16:13:22 crc kubenswrapper[4907]: I0313 16:13:22.021623 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-client-openstack-openstack-cell1-tcvvp"] Mar 13 16:13:22 crc kubenswrapper[4907]: I0313 16:13:22.201674 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/ef037fca-09ee-48ed-b2f7-fc65aea9155a-ssh-key-openstack-cell1\") pod \"ceph-client-openstack-openstack-cell1-tcvvp\" (UID: \"ef037fca-09ee-48ed-b2f7-fc65aea9155a\") " pod="openstack/ceph-client-openstack-openstack-cell1-tcvvp" Mar 13 16:13:22 crc kubenswrapper[4907]: I0313 16:13:22.201757 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ef037fca-09ee-48ed-b2f7-fc65aea9155a-ceph\") pod \"ceph-client-openstack-openstack-cell1-tcvvp\" (UID: \"ef037fca-09ee-48ed-b2f7-fc65aea9155a\") " pod="openstack/ceph-client-openstack-openstack-cell1-tcvvp" Mar 13 16:13:22 crc kubenswrapper[4907]: I0313 16:13:22.201838 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ef037fca-09ee-48ed-b2f7-fc65aea9155a-inventory\") pod \"ceph-client-openstack-openstack-cell1-tcvvp\" (UID: \"ef037fca-09ee-48ed-b2f7-fc65aea9155a\") " pod="openstack/ceph-client-openstack-openstack-cell1-tcvvp" Mar 13 16:13:22 crc kubenswrapper[4907]: I0313 16:13:22.201946 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hvtsj\" (UniqueName: \"kubernetes.io/projected/ef037fca-09ee-48ed-b2f7-fc65aea9155a-kube-api-access-hvtsj\") pod \"ceph-client-openstack-openstack-cell1-tcvvp\" (UID: \"ef037fca-09ee-48ed-b2f7-fc65aea9155a\") " pod="openstack/ceph-client-openstack-openstack-cell1-tcvvp" Mar 13 16:13:22 crc kubenswrapper[4907]: I0313 16:13:22.304636 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hvtsj\" (UniqueName: \"kubernetes.io/projected/ef037fca-09ee-48ed-b2f7-fc65aea9155a-kube-api-access-hvtsj\") pod \"ceph-client-openstack-openstack-cell1-tcvvp\" (UID: \"ef037fca-09ee-48ed-b2f7-fc65aea9155a\") " pod="openstack/ceph-client-openstack-openstack-cell1-tcvvp" Mar 13 16:13:22 crc kubenswrapper[4907]: I0313 16:13:22.304845 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/ef037fca-09ee-48ed-b2f7-fc65aea9155a-ssh-key-openstack-cell1\") pod \"ceph-client-openstack-openstack-cell1-tcvvp\" (UID: \"ef037fca-09ee-48ed-b2f7-fc65aea9155a\") " pod="openstack/ceph-client-openstack-openstack-cell1-tcvvp" Mar 13 16:13:22 crc kubenswrapper[4907]: I0313 16:13:22.305025 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ef037fca-09ee-48ed-b2f7-fc65aea9155a-ceph\") pod \"ceph-client-openstack-openstack-cell1-tcvvp\" (UID: \"ef037fca-09ee-48ed-b2f7-fc65aea9155a\") " pod="openstack/ceph-client-openstack-openstack-cell1-tcvvp" Mar 13 16:13:22 crc kubenswrapper[4907]: I0313 16:13:22.305158 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ef037fca-09ee-48ed-b2f7-fc65aea9155a-inventory\") pod \"ceph-client-openstack-openstack-cell1-tcvvp\" (UID: \"ef037fca-09ee-48ed-b2f7-fc65aea9155a\") " pod="openstack/ceph-client-openstack-openstack-cell1-tcvvp" Mar 13 16:13:22 crc kubenswrapper[4907]: I0313 16:13:22.309107 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/ef037fca-09ee-48ed-b2f7-fc65aea9155a-ssh-key-openstack-cell1\") pod \"ceph-client-openstack-openstack-cell1-tcvvp\" (UID: \"ef037fca-09ee-48ed-b2f7-fc65aea9155a\") " pod="openstack/ceph-client-openstack-openstack-cell1-tcvvp" Mar 13 16:13:22 crc kubenswrapper[4907]: I0313 16:13:22.309142 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ef037fca-09ee-48ed-b2f7-fc65aea9155a-inventory\") pod \"ceph-client-openstack-openstack-cell1-tcvvp\" (UID: \"ef037fca-09ee-48ed-b2f7-fc65aea9155a\") " pod="openstack/ceph-client-openstack-openstack-cell1-tcvvp" Mar 13 16:13:22 crc kubenswrapper[4907]: I0313 16:13:22.309839 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ef037fca-09ee-48ed-b2f7-fc65aea9155a-ceph\") pod \"ceph-client-openstack-openstack-cell1-tcvvp\" (UID: \"ef037fca-09ee-48ed-b2f7-fc65aea9155a\") " pod="openstack/ceph-client-openstack-openstack-cell1-tcvvp" Mar 13 16:13:22 crc kubenswrapper[4907]: I0313 16:13:22.326099 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hvtsj\" (UniqueName: \"kubernetes.io/projected/ef037fca-09ee-48ed-b2f7-fc65aea9155a-kube-api-access-hvtsj\") pod \"ceph-client-openstack-openstack-cell1-tcvvp\" (UID: \"ef037fca-09ee-48ed-b2f7-fc65aea9155a\") " pod="openstack/ceph-client-openstack-openstack-cell1-tcvvp" Mar 13 16:13:22 crc kubenswrapper[4907]: I0313 16:13:22.337389 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-openstack-openstack-cell1-tcvvp" Mar 13 16:13:22 crc kubenswrapper[4907]: I0313 16:13:22.926390 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-client-openstack-openstack-cell1-tcvvp"] Mar 13 16:13:23 crc kubenswrapper[4907]: I0313 16:13:23.941369 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-openstack-openstack-cell1-tcvvp" event={"ID":"ef037fca-09ee-48ed-b2f7-fc65aea9155a","Type":"ContainerStarted","Data":"023c9098cc761ac6f84ee782fabb6a2460877d7ae1e899af2870592e5799025d"} Mar 13 16:13:23 crc kubenswrapper[4907]: I0313 16:13:23.941779 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-openstack-openstack-cell1-tcvvp" event={"ID":"ef037fca-09ee-48ed-b2f7-fc65aea9155a","Type":"ContainerStarted","Data":"2510c8b37338fdfac88a3c5515d0d7e6350abffa7916555d9078ef3f9701bbfd"} Mar 13 16:13:23 crc kubenswrapper[4907]: I0313 16:13:23.966045 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceph-client-openstack-openstack-cell1-tcvvp" podStartSLOduration=2.462028175 podStartE2EDuration="2.966025778s" podCreationTimestamp="2026-03-13 16:13:21 +0000 UTC" firstStartedPulling="2026-03-13 16:13:22.936953155 +0000 UTC m=+7701.836740844" lastFinishedPulling="2026-03-13 16:13:23.440950728 +0000 UTC m=+7702.340738447" observedRunningTime="2026-03-13 16:13:23.956669342 +0000 UTC m=+7702.856457071" watchObservedRunningTime="2026-03-13 16:13:23.966025778 +0000 UTC m=+7702.865813467" Mar 13 16:13:29 crc kubenswrapper[4907]: I0313 16:13:29.007442 4907 generic.go:334] "Generic (PLEG): container finished" podID="ef037fca-09ee-48ed-b2f7-fc65aea9155a" containerID="023c9098cc761ac6f84ee782fabb6a2460877d7ae1e899af2870592e5799025d" exitCode=0 Mar 13 16:13:29 crc kubenswrapper[4907]: I0313 16:13:29.007546 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-openstack-openstack-cell1-tcvvp" event={"ID":"ef037fca-09ee-48ed-b2f7-fc65aea9155a","Type":"ContainerDied","Data":"023c9098cc761ac6f84ee782fabb6a2460877d7ae1e899af2870592e5799025d"} Mar 13 16:13:30 crc kubenswrapper[4907]: I0313 16:13:30.499672 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-openstack-openstack-cell1-tcvvp" Mar 13 16:13:30 crc kubenswrapper[4907]: I0313 16:13:30.613741 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ef037fca-09ee-48ed-b2f7-fc65aea9155a-inventory\") pod \"ef037fca-09ee-48ed-b2f7-fc65aea9155a\" (UID: \"ef037fca-09ee-48ed-b2f7-fc65aea9155a\") " Mar 13 16:13:30 crc kubenswrapper[4907]: I0313 16:13:30.613869 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ef037fca-09ee-48ed-b2f7-fc65aea9155a-ceph\") pod \"ef037fca-09ee-48ed-b2f7-fc65aea9155a\" (UID: \"ef037fca-09ee-48ed-b2f7-fc65aea9155a\") " Mar 13 16:13:30 crc kubenswrapper[4907]: I0313 16:13:30.613954 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/ef037fca-09ee-48ed-b2f7-fc65aea9155a-ssh-key-openstack-cell1\") pod \"ef037fca-09ee-48ed-b2f7-fc65aea9155a\" (UID: \"ef037fca-09ee-48ed-b2f7-fc65aea9155a\") " Mar 13 16:13:30 crc kubenswrapper[4907]: I0313 16:13:30.613996 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hvtsj\" (UniqueName: \"kubernetes.io/projected/ef037fca-09ee-48ed-b2f7-fc65aea9155a-kube-api-access-hvtsj\") pod \"ef037fca-09ee-48ed-b2f7-fc65aea9155a\" (UID: \"ef037fca-09ee-48ed-b2f7-fc65aea9155a\") " Mar 13 16:13:30 crc kubenswrapper[4907]: I0313 16:13:30.627023 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef037fca-09ee-48ed-b2f7-fc65aea9155a-ceph" (OuterVolumeSpecName: "ceph") pod "ef037fca-09ee-48ed-b2f7-fc65aea9155a" (UID: "ef037fca-09ee-48ed-b2f7-fc65aea9155a"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:13:30 crc kubenswrapper[4907]: I0313 16:13:30.627172 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef037fca-09ee-48ed-b2f7-fc65aea9155a-kube-api-access-hvtsj" (OuterVolumeSpecName: "kube-api-access-hvtsj") pod "ef037fca-09ee-48ed-b2f7-fc65aea9155a" (UID: "ef037fca-09ee-48ed-b2f7-fc65aea9155a"). InnerVolumeSpecName "kube-api-access-hvtsj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:13:30 crc kubenswrapper[4907]: I0313 16:13:30.647534 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef037fca-09ee-48ed-b2f7-fc65aea9155a-inventory" (OuterVolumeSpecName: "inventory") pod "ef037fca-09ee-48ed-b2f7-fc65aea9155a" (UID: "ef037fca-09ee-48ed-b2f7-fc65aea9155a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:13:30 crc kubenswrapper[4907]: I0313 16:13:30.658456 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef037fca-09ee-48ed-b2f7-fc65aea9155a-ssh-key-openstack-cell1" (OuterVolumeSpecName: "ssh-key-openstack-cell1") pod "ef037fca-09ee-48ed-b2f7-fc65aea9155a" (UID: "ef037fca-09ee-48ed-b2f7-fc65aea9155a"). InnerVolumeSpecName "ssh-key-openstack-cell1". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:13:30 crc kubenswrapper[4907]: I0313 16:13:30.717826 4907 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/ef037fca-09ee-48ed-b2f7-fc65aea9155a-ssh-key-openstack-cell1\") on node \"crc\" DevicePath \"\"" Mar 13 16:13:30 crc kubenswrapper[4907]: I0313 16:13:30.718149 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hvtsj\" (UniqueName: \"kubernetes.io/projected/ef037fca-09ee-48ed-b2f7-fc65aea9155a-kube-api-access-hvtsj\") on node \"crc\" DevicePath \"\"" Mar 13 16:13:30 crc kubenswrapper[4907]: I0313 16:13:30.718163 4907 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ef037fca-09ee-48ed-b2f7-fc65aea9155a-inventory\") on node \"crc\" DevicePath \"\"" Mar 13 16:13:30 crc kubenswrapper[4907]: I0313 16:13:30.718204 4907 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ef037fca-09ee-48ed-b2f7-fc65aea9155a-ceph\") on node \"crc\" DevicePath \"\"" Mar 13 16:13:31 crc kubenswrapper[4907]: I0313 16:13:31.033161 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-openstack-openstack-cell1-tcvvp" event={"ID":"ef037fca-09ee-48ed-b2f7-fc65aea9155a","Type":"ContainerDied","Data":"2510c8b37338fdfac88a3c5515d0d7e6350abffa7916555d9078ef3f9701bbfd"} Mar 13 16:13:31 crc kubenswrapper[4907]: I0313 16:13:31.033207 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2510c8b37338fdfac88a3c5515d0d7e6350abffa7916555d9078ef3f9701bbfd" Mar 13 16:13:31 crc kubenswrapper[4907]: I0313 16:13:31.033179 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-openstack-openstack-cell1-tcvvp" Mar 13 16:13:31 crc kubenswrapper[4907]: I0313 16:13:31.120026 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-openstack-openstack-cell1-8dwgb"] Mar 13 16:13:31 crc kubenswrapper[4907]: E0313 16:13:31.121028 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef037fca-09ee-48ed-b2f7-fc65aea9155a" containerName="ceph-client-openstack-openstack-cell1" Mar 13 16:13:31 crc kubenswrapper[4907]: I0313 16:13:31.121105 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef037fca-09ee-48ed-b2f7-fc65aea9155a" containerName="ceph-client-openstack-openstack-cell1" Mar 13 16:13:31 crc kubenswrapper[4907]: I0313 16:13:31.121692 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef037fca-09ee-48ed-b2f7-fc65aea9155a" containerName="ceph-client-openstack-openstack-cell1" Mar 13 16:13:31 crc kubenswrapper[4907]: I0313 16:13:31.123565 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-cell1-8dwgb" Mar 13 16:13:31 crc kubenswrapper[4907]: I0313 16:13:31.126453 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 13 16:13:31 crc kubenswrapper[4907]: I0313 16:13:31.126811 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Mar 13 16:13:31 crc kubenswrapper[4907]: I0313 16:13:31.126831 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Mar 13 16:13:31 crc kubenswrapper[4907]: I0313 16:13:31.127052 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-44kxw" Mar 13 16:13:31 crc kubenswrapper[4907]: I0313 16:13:31.127148 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Mar 13 16:13:31 crc kubenswrapper[4907]: I0313 16:13:31.136950 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-openstack-openstack-cell1-8dwgb"] Mar 13 16:13:31 crc kubenswrapper[4907]: I0313 16:13:31.228301 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/47d984fd-dfa5-4a10-91c3-3c9a13edc91c-ceph\") pod \"ovn-openstack-openstack-cell1-8dwgb\" (UID: \"47d984fd-dfa5-4a10-91c3-3c9a13edc91c\") " pod="openstack/ovn-openstack-openstack-cell1-8dwgb" Mar 13 16:13:31 crc kubenswrapper[4907]: I0313 16:13:31.228444 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47d984fd-dfa5-4a10-91c3-3c9a13edc91c-ovn-combined-ca-bundle\") pod \"ovn-openstack-openstack-cell1-8dwgb\" (UID: \"47d984fd-dfa5-4a10-91c3-3c9a13edc91c\") " pod="openstack/ovn-openstack-openstack-cell1-8dwgb" Mar 13 16:13:31 crc kubenswrapper[4907]: I0313 16:13:31.228529 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/47d984fd-dfa5-4a10-91c3-3c9a13edc91c-ssh-key-openstack-cell1\") pod \"ovn-openstack-openstack-cell1-8dwgb\" (UID: \"47d984fd-dfa5-4a10-91c3-3c9a13edc91c\") " pod="openstack/ovn-openstack-openstack-cell1-8dwgb" Mar 13 16:13:31 crc kubenswrapper[4907]: I0313 16:13:31.228707 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/47d984fd-dfa5-4a10-91c3-3c9a13edc91c-ovncontroller-config-0\") pod \"ovn-openstack-openstack-cell1-8dwgb\" (UID: \"47d984fd-dfa5-4a10-91c3-3c9a13edc91c\") " pod="openstack/ovn-openstack-openstack-cell1-8dwgb" Mar 13 16:13:31 crc kubenswrapper[4907]: I0313 16:13:31.228811 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/47d984fd-dfa5-4a10-91c3-3c9a13edc91c-inventory\") pod \"ovn-openstack-openstack-cell1-8dwgb\" (UID: \"47d984fd-dfa5-4a10-91c3-3c9a13edc91c\") " pod="openstack/ovn-openstack-openstack-cell1-8dwgb" Mar 13 16:13:31 crc kubenswrapper[4907]: I0313 16:13:31.229324 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4zlnv\" (UniqueName: \"kubernetes.io/projected/47d984fd-dfa5-4a10-91c3-3c9a13edc91c-kube-api-access-4zlnv\") pod \"ovn-openstack-openstack-cell1-8dwgb\" (UID: \"47d984fd-dfa5-4a10-91c3-3c9a13edc91c\") " pod="openstack/ovn-openstack-openstack-cell1-8dwgb" Mar 13 16:13:31 crc kubenswrapper[4907]: I0313 16:13:31.331898 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/47d984fd-dfa5-4a10-91c3-3c9a13edc91c-ovncontroller-config-0\") pod \"ovn-openstack-openstack-cell1-8dwgb\" (UID: \"47d984fd-dfa5-4a10-91c3-3c9a13edc91c\") " pod="openstack/ovn-openstack-openstack-cell1-8dwgb" Mar 13 16:13:31 crc kubenswrapper[4907]: I0313 16:13:31.331972 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/47d984fd-dfa5-4a10-91c3-3c9a13edc91c-inventory\") pod \"ovn-openstack-openstack-cell1-8dwgb\" (UID: \"47d984fd-dfa5-4a10-91c3-3c9a13edc91c\") " pod="openstack/ovn-openstack-openstack-cell1-8dwgb" Mar 13 16:13:31 crc kubenswrapper[4907]: I0313 16:13:31.332092 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4zlnv\" (UniqueName: \"kubernetes.io/projected/47d984fd-dfa5-4a10-91c3-3c9a13edc91c-kube-api-access-4zlnv\") pod \"ovn-openstack-openstack-cell1-8dwgb\" (UID: \"47d984fd-dfa5-4a10-91c3-3c9a13edc91c\") " pod="openstack/ovn-openstack-openstack-cell1-8dwgb" Mar 13 16:13:31 crc kubenswrapper[4907]: I0313 16:13:31.332122 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/47d984fd-dfa5-4a10-91c3-3c9a13edc91c-ceph\") pod \"ovn-openstack-openstack-cell1-8dwgb\" (UID: \"47d984fd-dfa5-4a10-91c3-3c9a13edc91c\") " pod="openstack/ovn-openstack-openstack-cell1-8dwgb" Mar 13 16:13:31 crc kubenswrapper[4907]: I0313 16:13:31.332175 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47d984fd-dfa5-4a10-91c3-3c9a13edc91c-ovn-combined-ca-bundle\") pod \"ovn-openstack-openstack-cell1-8dwgb\" (UID: \"47d984fd-dfa5-4a10-91c3-3c9a13edc91c\") " pod="openstack/ovn-openstack-openstack-cell1-8dwgb" Mar 13 16:13:31 crc kubenswrapper[4907]: I0313 16:13:31.332223 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/47d984fd-dfa5-4a10-91c3-3c9a13edc91c-ssh-key-openstack-cell1\") pod \"ovn-openstack-openstack-cell1-8dwgb\" (UID: \"47d984fd-dfa5-4a10-91c3-3c9a13edc91c\") " pod="openstack/ovn-openstack-openstack-cell1-8dwgb" Mar 13 16:13:31 crc kubenswrapper[4907]: I0313 16:13:31.333138 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/47d984fd-dfa5-4a10-91c3-3c9a13edc91c-ovncontroller-config-0\") pod \"ovn-openstack-openstack-cell1-8dwgb\" (UID: \"47d984fd-dfa5-4a10-91c3-3c9a13edc91c\") " pod="openstack/ovn-openstack-openstack-cell1-8dwgb" Mar 13 16:13:31 crc kubenswrapper[4907]: I0313 16:13:31.336325 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/47d984fd-dfa5-4a10-91c3-3c9a13edc91c-ssh-key-openstack-cell1\") pod \"ovn-openstack-openstack-cell1-8dwgb\" (UID: \"47d984fd-dfa5-4a10-91c3-3c9a13edc91c\") " pod="openstack/ovn-openstack-openstack-cell1-8dwgb" Mar 13 16:13:31 crc kubenswrapper[4907]: I0313 16:13:31.336427 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/47d984fd-dfa5-4a10-91c3-3c9a13edc91c-ceph\") pod \"ovn-openstack-openstack-cell1-8dwgb\" (UID: \"47d984fd-dfa5-4a10-91c3-3c9a13edc91c\") " pod="openstack/ovn-openstack-openstack-cell1-8dwgb" Mar 13 16:13:31 crc kubenswrapper[4907]: I0313 16:13:31.338348 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/47d984fd-dfa5-4a10-91c3-3c9a13edc91c-inventory\") pod \"ovn-openstack-openstack-cell1-8dwgb\" (UID: \"47d984fd-dfa5-4a10-91c3-3c9a13edc91c\") " pod="openstack/ovn-openstack-openstack-cell1-8dwgb" Mar 13 16:13:31 crc kubenswrapper[4907]: I0313 16:13:31.338963 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47d984fd-dfa5-4a10-91c3-3c9a13edc91c-ovn-combined-ca-bundle\") pod \"ovn-openstack-openstack-cell1-8dwgb\" (UID: \"47d984fd-dfa5-4a10-91c3-3c9a13edc91c\") " pod="openstack/ovn-openstack-openstack-cell1-8dwgb" Mar 13 16:13:31 crc kubenswrapper[4907]: I0313 16:13:31.353592 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4zlnv\" (UniqueName: \"kubernetes.io/projected/47d984fd-dfa5-4a10-91c3-3c9a13edc91c-kube-api-access-4zlnv\") pod \"ovn-openstack-openstack-cell1-8dwgb\" (UID: \"47d984fd-dfa5-4a10-91c3-3c9a13edc91c\") " pod="openstack/ovn-openstack-openstack-cell1-8dwgb" Mar 13 16:13:31 crc kubenswrapper[4907]: I0313 16:13:31.498123 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-cell1-8dwgb" Mar 13 16:13:32 crc kubenswrapper[4907]: I0313 16:13:32.041072 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-openstack-openstack-cell1-8dwgb"] Mar 13 16:13:33 crc kubenswrapper[4907]: I0313 16:13:33.058287 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-cell1-8dwgb" event={"ID":"47d984fd-dfa5-4a10-91c3-3c9a13edc91c","Type":"ContainerStarted","Data":"d1f808535be426e781215cf8fdab292d9eab6619345966de095be637979de839"} Mar 13 16:13:33 crc kubenswrapper[4907]: I0313 16:13:33.059057 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-cell1-8dwgb" event={"ID":"47d984fd-dfa5-4a10-91c3-3c9a13edc91c","Type":"ContainerStarted","Data":"3da1f54adfbf7e7555f40a0016c9c1e8f607c6f36a3249f1f9d7c62e57e1380d"} Mar 13 16:13:33 crc kubenswrapper[4907]: I0313 16:13:33.089730 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-openstack-openstack-cell1-8dwgb" podStartSLOduration=1.666269276 podStartE2EDuration="2.089704936s" podCreationTimestamp="2026-03-13 16:13:31 +0000 UTC" firstStartedPulling="2026-03-13 16:13:32.040355886 +0000 UTC m=+7710.940143595" lastFinishedPulling="2026-03-13 16:13:32.463791566 +0000 UTC m=+7711.363579255" observedRunningTime="2026-03-13 16:13:33.085037638 +0000 UTC m=+7711.984825337" watchObservedRunningTime="2026-03-13 16:13:33.089704936 +0000 UTC m=+7711.989492665" Mar 13 16:14:00 crc kubenswrapper[4907]: I0313 16:14:00.148539 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556974-8sntm"] Mar 13 16:14:00 crc kubenswrapper[4907]: I0313 16:14:00.150763 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556974-8sntm" Mar 13 16:14:00 crc kubenswrapper[4907]: I0313 16:14:00.153873 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 16:14:00 crc kubenswrapper[4907]: I0313 16:14:00.154304 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 16:14:00 crc kubenswrapper[4907]: I0313 16:14:00.156574 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 16:14:00 crc kubenswrapper[4907]: I0313 16:14:00.158364 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556974-8sntm"] Mar 13 16:14:00 crc kubenswrapper[4907]: I0313 16:14:00.255849 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q7kw6\" (UniqueName: \"kubernetes.io/projected/5d0046fb-05a0-47a0-912c-a9495e2a0835-kube-api-access-q7kw6\") pod \"auto-csr-approver-29556974-8sntm\" (UID: \"5d0046fb-05a0-47a0-912c-a9495e2a0835\") " pod="openshift-infra/auto-csr-approver-29556974-8sntm" Mar 13 16:14:00 crc kubenswrapper[4907]: I0313 16:14:00.366444 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q7kw6\" (UniqueName: \"kubernetes.io/projected/5d0046fb-05a0-47a0-912c-a9495e2a0835-kube-api-access-q7kw6\") pod \"auto-csr-approver-29556974-8sntm\" (UID: \"5d0046fb-05a0-47a0-912c-a9495e2a0835\") " pod="openshift-infra/auto-csr-approver-29556974-8sntm" Mar 13 16:14:00 crc kubenswrapper[4907]: I0313 16:14:00.405834 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q7kw6\" (UniqueName: \"kubernetes.io/projected/5d0046fb-05a0-47a0-912c-a9495e2a0835-kube-api-access-q7kw6\") pod \"auto-csr-approver-29556974-8sntm\" (UID: \"5d0046fb-05a0-47a0-912c-a9495e2a0835\") " pod="openshift-infra/auto-csr-approver-29556974-8sntm" Mar 13 16:14:00 crc kubenswrapper[4907]: I0313 16:14:00.481634 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556974-8sntm" Mar 13 16:14:01 crc kubenswrapper[4907]: I0313 16:14:01.532161 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556974-8sntm"] Mar 13 16:14:02 crc kubenswrapper[4907]: I0313 16:14:02.455844 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556974-8sntm" event={"ID":"5d0046fb-05a0-47a0-912c-a9495e2a0835","Type":"ContainerStarted","Data":"a72b1649802cb4c03ecb12ae633e70af70234cd1f39e5f25f67919475783e4b7"} Mar 13 16:14:04 crc kubenswrapper[4907]: I0313 16:14:04.484747 4907 generic.go:334] "Generic (PLEG): container finished" podID="5d0046fb-05a0-47a0-912c-a9495e2a0835" containerID="048a0a595dbc6f4c8cb2226742f2beb5b639bec2395f93e0935301c145de8608" exitCode=0 Mar 13 16:14:04 crc kubenswrapper[4907]: I0313 16:14:04.484834 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556974-8sntm" event={"ID":"5d0046fb-05a0-47a0-912c-a9495e2a0835","Type":"ContainerDied","Data":"048a0a595dbc6f4c8cb2226742f2beb5b639bec2395f93e0935301c145de8608"} Mar 13 16:14:05 crc kubenswrapper[4907]: I0313 16:14:05.868674 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556974-8sntm" Mar 13 16:14:06 crc kubenswrapper[4907]: I0313 16:14:06.009864 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q7kw6\" (UniqueName: \"kubernetes.io/projected/5d0046fb-05a0-47a0-912c-a9495e2a0835-kube-api-access-q7kw6\") pod \"5d0046fb-05a0-47a0-912c-a9495e2a0835\" (UID: \"5d0046fb-05a0-47a0-912c-a9495e2a0835\") " Mar 13 16:14:06 crc kubenswrapper[4907]: I0313 16:14:06.020823 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5d0046fb-05a0-47a0-912c-a9495e2a0835-kube-api-access-q7kw6" (OuterVolumeSpecName: "kube-api-access-q7kw6") pod "5d0046fb-05a0-47a0-912c-a9495e2a0835" (UID: "5d0046fb-05a0-47a0-912c-a9495e2a0835"). InnerVolumeSpecName "kube-api-access-q7kw6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:14:06 crc kubenswrapper[4907]: I0313 16:14:06.113040 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q7kw6\" (UniqueName: \"kubernetes.io/projected/5d0046fb-05a0-47a0-912c-a9495e2a0835-kube-api-access-q7kw6\") on node \"crc\" DevicePath \"\"" Mar 13 16:14:06 crc kubenswrapper[4907]: I0313 16:14:06.508863 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556974-8sntm" event={"ID":"5d0046fb-05a0-47a0-912c-a9495e2a0835","Type":"ContainerDied","Data":"a72b1649802cb4c03ecb12ae633e70af70234cd1f39e5f25f67919475783e4b7"} Mar 13 16:14:06 crc kubenswrapper[4907]: I0313 16:14:06.508932 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a72b1649802cb4c03ecb12ae633e70af70234cd1f39e5f25f67919475783e4b7" Mar 13 16:14:06 crc kubenswrapper[4907]: I0313 16:14:06.508970 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556974-8sntm" Mar 13 16:14:06 crc kubenswrapper[4907]: I0313 16:14:06.963460 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556968-xbmhk"] Mar 13 16:14:06 crc kubenswrapper[4907]: I0313 16:14:06.972208 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556968-xbmhk"] Mar 13 16:14:07 crc kubenswrapper[4907]: I0313 16:14:07.795213 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cfdfc46b-b215-491c-ab2f-2a108c2c0882" path="/var/lib/kubelet/pods/cfdfc46b-b215-491c-ab2f-2a108c2c0882/volumes" Mar 13 16:14:37 crc kubenswrapper[4907]: I0313 16:14:37.844771 4907 generic.go:334] "Generic (PLEG): container finished" podID="47d984fd-dfa5-4a10-91c3-3c9a13edc91c" containerID="d1f808535be426e781215cf8fdab292d9eab6619345966de095be637979de839" exitCode=0 Mar 13 16:14:37 crc kubenswrapper[4907]: I0313 16:14:37.844872 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-cell1-8dwgb" event={"ID":"47d984fd-dfa5-4a10-91c3-3c9a13edc91c","Type":"ContainerDied","Data":"d1f808535be426e781215cf8fdab292d9eab6619345966de095be637979de839"} Mar 13 16:14:39 crc kubenswrapper[4907]: I0313 16:14:39.259990 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-cell1-8dwgb" Mar 13 16:14:39 crc kubenswrapper[4907]: I0313 16:14:39.408812 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47d984fd-dfa5-4a10-91c3-3c9a13edc91c-ovn-combined-ca-bundle\") pod \"47d984fd-dfa5-4a10-91c3-3c9a13edc91c\" (UID: \"47d984fd-dfa5-4a10-91c3-3c9a13edc91c\") " Mar 13 16:14:39 crc kubenswrapper[4907]: I0313 16:14:39.408873 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/47d984fd-dfa5-4a10-91c3-3c9a13edc91c-ceph\") pod \"47d984fd-dfa5-4a10-91c3-3c9a13edc91c\" (UID: \"47d984fd-dfa5-4a10-91c3-3c9a13edc91c\") " Mar 13 16:14:39 crc kubenswrapper[4907]: I0313 16:14:39.408955 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/47d984fd-dfa5-4a10-91c3-3c9a13edc91c-ssh-key-openstack-cell1\") pod \"47d984fd-dfa5-4a10-91c3-3c9a13edc91c\" (UID: \"47d984fd-dfa5-4a10-91c3-3c9a13edc91c\") " Mar 13 16:14:39 crc kubenswrapper[4907]: I0313 16:14:39.408989 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4zlnv\" (UniqueName: \"kubernetes.io/projected/47d984fd-dfa5-4a10-91c3-3c9a13edc91c-kube-api-access-4zlnv\") pod \"47d984fd-dfa5-4a10-91c3-3c9a13edc91c\" (UID: \"47d984fd-dfa5-4a10-91c3-3c9a13edc91c\") " Mar 13 16:14:39 crc kubenswrapper[4907]: I0313 16:14:39.409012 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/47d984fd-dfa5-4a10-91c3-3c9a13edc91c-ovncontroller-config-0\") pod \"47d984fd-dfa5-4a10-91c3-3c9a13edc91c\" (UID: \"47d984fd-dfa5-4a10-91c3-3c9a13edc91c\") " Mar 13 16:14:39 crc kubenswrapper[4907]: I0313 16:14:39.409119 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/47d984fd-dfa5-4a10-91c3-3c9a13edc91c-inventory\") pod \"47d984fd-dfa5-4a10-91c3-3c9a13edc91c\" (UID: \"47d984fd-dfa5-4a10-91c3-3c9a13edc91c\") " Mar 13 16:14:39 crc kubenswrapper[4907]: I0313 16:14:39.414120 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47d984fd-dfa5-4a10-91c3-3c9a13edc91c-ceph" (OuterVolumeSpecName: "ceph") pod "47d984fd-dfa5-4a10-91c3-3c9a13edc91c" (UID: "47d984fd-dfa5-4a10-91c3-3c9a13edc91c"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:14:39 crc kubenswrapper[4907]: I0313 16:14:39.420419 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/47d984fd-dfa5-4a10-91c3-3c9a13edc91c-kube-api-access-4zlnv" (OuterVolumeSpecName: "kube-api-access-4zlnv") pod "47d984fd-dfa5-4a10-91c3-3c9a13edc91c" (UID: "47d984fd-dfa5-4a10-91c3-3c9a13edc91c"). InnerVolumeSpecName "kube-api-access-4zlnv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:14:39 crc kubenswrapper[4907]: I0313 16:14:39.432147 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47d984fd-dfa5-4a10-91c3-3c9a13edc91c-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "47d984fd-dfa5-4a10-91c3-3c9a13edc91c" (UID: "47d984fd-dfa5-4a10-91c3-3c9a13edc91c"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:14:39 crc kubenswrapper[4907]: I0313 16:14:39.446041 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47d984fd-dfa5-4a10-91c3-3c9a13edc91c-inventory" (OuterVolumeSpecName: "inventory") pod "47d984fd-dfa5-4a10-91c3-3c9a13edc91c" (UID: "47d984fd-dfa5-4a10-91c3-3c9a13edc91c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:14:39 crc kubenswrapper[4907]: I0313 16:14:39.447497 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47d984fd-dfa5-4a10-91c3-3c9a13edc91c-ssh-key-openstack-cell1" (OuterVolumeSpecName: "ssh-key-openstack-cell1") pod "47d984fd-dfa5-4a10-91c3-3c9a13edc91c" (UID: "47d984fd-dfa5-4a10-91c3-3c9a13edc91c"). InnerVolumeSpecName "ssh-key-openstack-cell1". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:14:39 crc kubenswrapper[4907]: I0313 16:14:39.455530 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/47d984fd-dfa5-4a10-91c3-3c9a13edc91c-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "47d984fd-dfa5-4a10-91c3-3c9a13edc91c" (UID: "47d984fd-dfa5-4a10-91c3-3c9a13edc91c"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 16:14:39 crc kubenswrapper[4907]: I0313 16:14:39.511801 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4zlnv\" (UniqueName: \"kubernetes.io/projected/47d984fd-dfa5-4a10-91c3-3c9a13edc91c-kube-api-access-4zlnv\") on node \"crc\" DevicePath \"\"" Mar 13 16:14:39 crc kubenswrapper[4907]: I0313 16:14:39.511844 4907 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/47d984fd-dfa5-4a10-91c3-3c9a13edc91c-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Mar 13 16:14:39 crc kubenswrapper[4907]: I0313 16:14:39.511861 4907 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/47d984fd-dfa5-4a10-91c3-3c9a13edc91c-inventory\") on node \"crc\" DevicePath \"\"" Mar 13 16:14:39 crc kubenswrapper[4907]: I0313 16:14:39.511874 4907 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47d984fd-dfa5-4a10-91c3-3c9a13edc91c-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 16:14:39 crc kubenswrapper[4907]: I0313 16:14:39.511905 4907 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/47d984fd-dfa5-4a10-91c3-3c9a13edc91c-ceph\") on node \"crc\" DevicePath \"\"" Mar 13 16:14:39 crc kubenswrapper[4907]: I0313 16:14:39.511941 4907 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/47d984fd-dfa5-4a10-91c3-3c9a13edc91c-ssh-key-openstack-cell1\") on node \"crc\" DevicePath \"\"" Mar 13 16:14:39 crc kubenswrapper[4907]: I0313 16:14:39.870038 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-cell1-8dwgb" event={"ID":"47d984fd-dfa5-4a10-91c3-3c9a13edc91c","Type":"ContainerDied","Data":"3da1f54adfbf7e7555f40a0016c9c1e8f607c6f36a3249f1f9d7c62e57e1380d"} Mar 13 16:14:39 crc kubenswrapper[4907]: I0313 16:14:39.870362 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-cell1-8dwgb" Mar 13 16:14:39 crc kubenswrapper[4907]: I0313 16:14:39.870376 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3da1f54adfbf7e7555f40a0016c9c1e8f607c6f36a3249f1f9d7c62e57e1380d" Mar 13 16:14:40 crc kubenswrapper[4907]: I0313 16:14:40.002483 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-openstack-openstack-cell1-pwqj6"] Mar 13 16:14:40 crc kubenswrapper[4907]: E0313 16:14:40.003089 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47d984fd-dfa5-4a10-91c3-3c9a13edc91c" containerName="ovn-openstack-openstack-cell1" Mar 13 16:14:40 crc kubenswrapper[4907]: I0313 16:14:40.003114 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="47d984fd-dfa5-4a10-91c3-3c9a13edc91c" containerName="ovn-openstack-openstack-cell1" Mar 13 16:14:40 crc kubenswrapper[4907]: E0313 16:14:40.003152 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d0046fb-05a0-47a0-912c-a9495e2a0835" containerName="oc" Mar 13 16:14:40 crc kubenswrapper[4907]: I0313 16:14:40.003166 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d0046fb-05a0-47a0-912c-a9495e2a0835" containerName="oc" Mar 13 16:14:40 crc kubenswrapper[4907]: I0313 16:14:40.003526 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="47d984fd-dfa5-4a10-91c3-3c9a13edc91c" containerName="ovn-openstack-openstack-cell1" Mar 13 16:14:40 crc kubenswrapper[4907]: I0313 16:14:40.003564 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="5d0046fb-05a0-47a0-912c-a9495e2a0835" containerName="oc" Mar 13 16:14:40 crc kubenswrapper[4907]: I0313 16:14:40.004871 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-cell1-pwqj6" Mar 13 16:14:40 crc kubenswrapper[4907]: I0313 16:14:40.012489 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-44kxw" Mar 13 16:14:40 crc kubenswrapper[4907]: I0313 16:14:40.012682 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Mar 13 16:14:40 crc kubenswrapper[4907]: I0313 16:14:40.012794 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Mar 13 16:14:40 crc kubenswrapper[4907]: I0313 16:14:40.012924 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Mar 13 16:14:40 crc kubenswrapper[4907]: I0313 16:14:40.013021 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 13 16:14:40 crc kubenswrapper[4907]: I0313 16:14:40.013402 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Mar 13 16:14:40 crc kubenswrapper[4907]: I0313 16:14:40.041273 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-openstack-openstack-cell1-pwqj6"] Mar 13 16:14:40 crc kubenswrapper[4907]: I0313 16:14:40.126567 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bcsp6\" (UniqueName: \"kubernetes.io/projected/3dc18cfe-d63a-4960-8c7d-10913cc78896-kube-api-access-bcsp6\") pod \"neutron-metadata-openstack-openstack-cell1-pwqj6\" (UID: \"3dc18cfe-d63a-4960-8c7d-10913cc78896\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-pwqj6" Mar 13 16:14:40 crc kubenswrapper[4907]: I0313 16:14:40.127113 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3dc18cfe-d63a-4960-8c7d-10913cc78896-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-openstack-openstack-cell1-pwqj6\" (UID: \"3dc18cfe-d63a-4960-8c7d-10913cc78896\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-pwqj6" Mar 13 16:14:40 crc kubenswrapper[4907]: I0313 16:14:40.127233 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/3dc18cfe-d63a-4960-8c7d-10913cc78896-nova-metadata-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-pwqj6\" (UID: \"3dc18cfe-d63a-4960-8c7d-10913cc78896\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-pwqj6" Mar 13 16:14:40 crc kubenswrapper[4907]: I0313 16:14:40.127407 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3dc18cfe-d63a-4960-8c7d-10913cc78896-ceph\") pod \"neutron-metadata-openstack-openstack-cell1-pwqj6\" (UID: \"3dc18cfe-d63a-4960-8c7d-10913cc78896\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-pwqj6" Mar 13 16:14:40 crc kubenswrapper[4907]: I0313 16:14:40.127614 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/3dc18cfe-d63a-4960-8c7d-10913cc78896-ssh-key-openstack-cell1\") pod \"neutron-metadata-openstack-openstack-cell1-pwqj6\" (UID: \"3dc18cfe-d63a-4960-8c7d-10913cc78896\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-pwqj6" Mar 13 16:14:40 crc kubenswrapper[4907]: I0313 16:14:40.127825 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/3dc18cfe-d63a-4960-8c7d-10913cc78896-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-pwqj6\" (UID: \"3dc18cfe-d63a-4960-8c7d-10913cc78896\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-pwqj6" Mar 13 16:14:40 crc kubenswrapper[4907]: I0313 16:14:40.128005 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3dc18cfe-d63a-4960-8c7d-10913cc78896-inventory\") pod \"neutron-metadata-openstack-openstack-cell1-pwqj6\" (UID: \"3dc18cfe-d63a-4960-8c7d-10913cc78896\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-pwqj6" Mar 13 16:14:40 crc kubenswrapper[4907]: I0313 16:14:40.230417 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/3dc18cfe-d63a-4960-8c7d-10913cc78896-ssh-key-openstack-cell1\") pod \"neutron-metadata-openstack-openstack-cell1-pwqj6\" (UID: \"3dc18cfe-d63a-4960-8c7d-10913cc78896\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-pwqj6" Mar 13 16:14:40 crc kubenswrapper[4907]: I0313 16:14:40.230696 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/3dc18cfe-d63a-4960-8c7d-10913cc78896-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-pwqj6\" (UID: \"3dc18cfe-d63a-4960-8c7d-10913cc78896\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-pwqj6" Mar 13 16:14:40 crc kubenswrapper[4907]: I0313 16:14:40.230800 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3dc18cfe-d63a-4960-8c7d-10913cc78896-inventory\") pod \"neutron-metadata-openstack-openstack-cell1-pwqj6\" (UID: \"3dc18cfe-d63a-4960-8c7d-10913cc78896\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-pwqj6" Mar 13 16:14:40 crc kubenswrapper[4907]: I0313 16:14:40.230986 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bcsp6\" (UniqueName: \"kubernetes.io/projected/3dc18cfe-d63a-4960-8c7d-10913cc78896-kube-api-access-bcsp6\") pod \"neutron-metadata-openstack-openstack-cell1-pwqj6\" (UID: \"3dc18cfe-d63a-4960-8c7d-10913cc78896\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-pwqj6" Mar 13 16:14:40 crc kubenswrapper[4907]: I0313 16:14:40.231154 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3dc18cfe-d63a-4960-8c7d-10913cc78896-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-openstack-openstack-cell1-pwqj6\" (UID: \"3dc18cfe-d63a-4960-8c7d-10913cc78896\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-pwqj6" Mar 13 16:14:40 crc kubenswrapper[4907]: I0313 16:14:40.231237 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/3dc18cfe-d63a-4960-8c7d-10913cc78896-nova-metadata-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-pwqj6\" (UID: \"3dc18cfe-d63a-4960-8c7d-10913cc78896\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-pwqj6" Mar 13 16:14:40 crc kubenswrapper[4907]: I0313 16:14:40.231353 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3dc18cfe-d63a-4960-8c7d-10913cc78896-ceph\") pod \"neutron-metadata-openstack-openstack-cell1-pwqj6\" (UID: \"3dc18cfe-d63a-4960-8c7d-10913cc78896\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-pwqj6" Mar 13 16:14:40 crc kubenswrapper[4907]: I0313 16:14:40.237432 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/3dc18cfe-d63a-4960-8c7d-10913cc78896-ssh-key-openstack-cell1\") pod \"neutron-metadata-openstack-openstack-cell1-pwqj6\" (UID: \"3dc18cfe-d63a-4960-8c7d-10913cc78896\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-pwqj6" Mar 13 16:14:40 crc kubenswrapper[4907]: I0313 16:14:40.237478 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/3dc18cfe-d63a-4960-8c7d-10913cc78896-nova-metadata-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-pwqj6\" (UID: \"3dc18cfe-d63a-4960-8c7d-10913cc78896\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-pwqj6" Mar 13 16:14:40 crc kubenswrapper[4907]: I0313 16:14:40.237499 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3dc18cfe-d63a-4960-8c7d-10913cc78896-inventory\") pod \"neutron-metadata-openstack-openstack-cell1-pwqj6\" (UID: \"3dc18cfe-d63a-4960-8c7d-10913cc78896\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-pwqj6" Mar 13 16:14:40 crc kubenswrapper[4907]: I0313 16:14:40.237999 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/3dc18cfe-d63a-4960-8c7d-10913cc78896-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-pwqj6\" (UID: \"3dc18cfe-d63a-4960-8c7d-10913cc78896\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-pwqj6" Mar 13 16:14:40 crc kubenswrapper[4907]: I0313 16:14:40.239593 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3dc18cfe-d63a-4960-8c7d-10913cc78896-ceph\") pod \"neutron-metadata-openstack-openstack-cell1-pwqj6\" (UID: \"3dc18cfe-d63a-4960-8c7d-10913cc78896\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-pwqj6" Mar 13 16:14:40 crc kubenswrapper[4907]: I0313 16:14:40.242956 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3dc18cfe-d63a-4960-8c7d-10913cc78896-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-openstack-openstack-cell1-pwqj6\" (UID: \"3dc18cfe-d63a-4960-8c7d-10913cc78896\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-pwqj6" Mar 13 16:14:40 crc kubenswrapper[4907]: I0313 16:14:40.249663 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bcsp6\" (UniqueName: \"kubernetes.io/projected/3dc18cfe-d63a-4960-8c7d-10913cc78896-kube-api-access-bcsp6\") pod \"neutron-metadata-openstack-openstack-cell1-pwqj6\" (UID: \"3dc18cfe-d63a-4960-8c7d-10913cc78896\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-pwqj6" Mar 13 16:14:40 crc kubenswrapper[4907]: I0313 16:14:40.343526 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-cell1-pwqj6" Mar 13 16:14:40 crc kubenswrapper[4907]: I0313 16:14:40.950085 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-openstack-openstack-cell1-pwqj6"] Mar 13 16:14:41 crc kubenswrapper[4907]: I0313 16:14:41.599064 4907 scope.go:117] "RemoveContainer" containerID="6f4aac0052e053a0aefd7363aac839de6a2ba5c8481dbf090c19d76f2294e95d" Mar 13 16:14:41 crc kubenswrapper[4907]: I0313 16:14:41.905511 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-cell1-pwqj6" event={"ID":"3dc18cfe-d63a-4960-8c7d-10913cc78896","Type":"ContainerStarted","Data":"974721c911609f40407a7889cdfed0cb7412642260334ed010d57dbf5bb59c51"} Mar 13 16:14:41 crc kubenswrapper[4907]: I0313 16:14:41.905927 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-cell1-pwqj6" event={"ID":"3dc18cfe-d63a-4960-8c7d-10913cc78896","Type":"ContainerStarted","Data":"f4a4333655b463ae652caae8d8ffaa8b9979eafbf031d280da29757bdd65f7a4"} Mar 13 16:14:41 crc kubenswrapper[4907]: I0313 16:14:41.935555 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-openstack-openstack-cell1-pwqj6" podStartSLOduration=2.391647765 podStartE2EDuration="2.93550783s" podCreationTimestamp="2026-03-13 16:14:39 +0000 UTC" firstStartedPulling="2026-03-13 16:14:40.959663003 +0000 UTC m=+7779.859450732" lastFinishedPulling="2026-03-13 16:14:41.503523098 +0000 UTC m=+7780.403310797" observedRunningTime="2026-03-13 16:14:41.921312401 +0000 UTC m=+7780.821100100" watchObservedRunningTime="2026-03-13 16:14:41.93550783 +0000 UTC m=+7780.835295519" Mar 13 16:14:48 crc kubenswrapper[4907]: I0313 16:14:48.041288 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 16:14:48 crc kubenswrapper[4907]: I0313 16:14:48.041713 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 16:15:00 crc kubenswrapper[4907]: I0313 16:15:00.160391 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556975-rmwsm"] Mar 13 16:15:00 crc kubenswrapper[4907]: I0313 16:15:00.162194 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556975-rmwsm" Mar 13 16:15:00 crc kubenswrapper[4907]: I0313 16:15:00.168063 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 13 16:15:00 crc kubenswrapper[4907]: I0313 16:15:00.168173 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 13 16:15:00 crc kubenswrapper[4907]: I0313 16:15:00.176250 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556975-rmwsm"] Mar 13 16:15:00 crc kubenswrapper[4907]: I0313 16:15:00.346191 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cw7x4\" (UniqueName: \"kubernetes.io/projected/e1d3954c-686f-4c87-bebd-95a286df5781-kube-api-access-cw7x4\") pod \"collect-profiles-29556975-rmwsm\" (UID: \"e1d3954c-686f-4c87-bebd-95a286df5781\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556975-rmwsm" Mar 13 16:15:00 crc kubenswrapper[4907]: I0313 16:15:00.346405 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e1d3954c-686f-4c87-bebd-95a286df5781-config-volume\") pod \"collect-profiles-29556975-rmwsm\" (UID: \"e1d3954c-686f-4c87-bebd-95a286df5781\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556975-rmwsm" Mar 13 16:15:00 crc kubenswrapper[4907]: I0313 16:15:00.346587 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e1d3954c-686f-4c87-bebd-95a286df5781-secret-volume\") pod \"collect-profiles-29556975-rmwsm\" (UID: \"e1d3954c-686f-4c87-bebd-95a286df5781\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556975-rmwsm" Mar 13 16:15:00 crc kubenswrapper[4907]: I0313 16:15:00.449613 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e1d3954c-686f-4c87-bebd-95a286df5781-config-volume\") pod \"collect-profiles-29556975-rmwsm\" (UID: \"e1d3954c-686f-4c87-bebd-95a286df5781\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556975-rmwsm" Mar 13 16:15:00 crc kubenswrapper[4907]: I0313 16:15:00.449813 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e1d3954c-686f-4c87-bebd-95a286df5781-secret-volume\") pod \"collect-profiles-29556975-rmwsm\" (UID: \"e1d3954c-686f-4c87-bebd-95a286df5781\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556975-rmwsm" Mar 13 16:15:00 crc kubenswrapper[4907]: I0313 16:15:00.450153 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cw7x4\" (UniqueName: \"kubernetes.io/projected/e1d3954c-686f-4c87-bebd-95a286df5781-kube-api-access-cw7x4\") pod \"collect-profiles-29556975-rmwsm\" (UID: \"e1d3954c-686f-4c87-bebd-95a286df5781\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556975-rmwsm" Mar 13 16:15:00 crc kubenswrapper[4907]: I0313 16:15:00.453127 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e1d3954c-686f-4c87-bebd-95a286df5781-config-volume\") pod \"collect-profiles-29556975-rmwsm\" (UID: \"e1d3954c-686f-4c87-bebd-95a286df5781\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556975-rmwsm" Mar 13 16:15:00 crc kubenswrapper[4907]: I0313 16:15:00.463333 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e1d3954c-686f-4c87-bebd-95a286df5781-secret-volume\") pod \"collect-profiles-29556975-rmwsm\" (UID: \"e1d3954c-686f-4c87-bebd-95a286df5781\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556975-rmwsm" Mar 13 16:15:00 crc kubenswrapper[4907]: I0313 16:15:00.481101 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cw7x4\" (UniqueName: \"kubernetes.io/projected/e1d3954c-686f-4c87-bebd-95a286df5781-kube-api-access-cw7x4\") pod \"collect-profiles-29556975-rmwsm\" (UID: \"e1d3954c-686f-4c87-bebd-95a286df5781\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556975-rmwsm" Mar 13 16:15:00 crc kubenswrapper[4907]: I0313 16:15:00.494726 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556975-rmwsm" Mar 13 16:15:00 crc kubenswrapper[4907]: I0313 16:15:00.941576 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556975-rmwsm"] Mar 13 16:15:00 crc kubenswrapper[4907]: W0313 16:15:00.943735 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode1d3954c_686f_4c87_bebd_95a286df5781.slice/crio-7110db02ed258c363241254fa3b18e476c70ce370c9163d88c74a6f6dca09ec3 WatchSource:0}: Error finding container 7110db02ed258c363241254fa3b18e476c70ce370c9163d88c74a6f6dca09ec3: Status 404 returned error can't find the container with id 7110db02ed258c363241254fa3b18e476c70ce370c9163d88c74a6f6dca09ec3 Mar 13 16:15:01 crc kubenswrapper[4907]: I0313 16:15:01.111890 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556975-rmwsm" event={"ID":"e1d3954c-686f-4c87-bebd-95a286df5781","Type":"ContainerStarted","Data":"7110db02ed258c363241254fa3b18e476c70ce370c9163d88c74a6f6dca09ec3"} Mar 13 16:15:02 crc kubenswrapper[4907]: I0313 16:15:02.128651 4907 generic.go:334] "Generic (PLEG): container finished" podID="e1d3954c-686f-4c87-bebd-95a286df5781" containerID="5d7f3378e64b40f003c74c11f875b9d214af6de777aed6922c0bc466b32e0063" exitCode=0 Mar 13 16:15:02 crc kubenswrapper[4907]: I0313 16:15:02.128892 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556975-rmwsm" event={"ID":"e1d3954c-686f-4c87-bebd-95a286df5781","Type":"ContainerDied","Data":"5d7f3378e64b40f003c74c11f875b9d214af6de777aed6922c0bc466b32e0063"} Mar 13 16:15:03 crc kubenswrapper[4907]: I0313 16:15:03.521807 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556975-rmwsm" Mar 13 16:15:03 crc kubenswrapper[4907]: I0313 16:15:03.714317 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e1d3954c-686f-4c87-bebd-95a286df5781-config-volume\") pod \"e1d3954c-686f-4c87-bebd-95a286df5781\" (UID: \"e1d3954c-686f-4c87-bebd-95a286df5781\") " Mar 13 16:15:03 crc kubenswrapper[4907]: I0313 16:15:03.714539 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e1d3954c-686f-4c87-bebd-95a286df5781-secret-volume\") pod \"e1d3954c-686f-4c87-bebd-95a286df5781\" (UID: \"e1d3954c-686f-4c87-bebd-95a286df5781\") " Mar 13 16:15:03 crc kubenswrapper[4907]: I0313 16:15:03.714567 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cw7x4\" (UniqueName: \"kubernetes.io/projected/e1d3954c-686f-4c87-bebd-95a286df5781-kube-api-access-cw7x4\") pod \"e1d3954c-686f-4c87-bebd-95a286df5781\" (UID: \"e1d3954c-686f-4c87-bebd-95a286df5781\") " Mar 13 16:15:03 crc kubenswrapper[4907]: I0313 16:15:03.715049 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e1d3954c-686f-4c87-bebd-95a286df5781-config-volume" (OuterVolumeSpecName: "config-volume") pod "e1d3954c-686f-4c87-bebd-95a286df5781" (UID: "e1d3954c-686f-4c87-bebd-95a286df5781"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 16:15:03 crc kubenswrapper[4907]: I0313 16:15:03.716330 4907 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e1d3954c-686f-4c87-bebd-95a286df5781-config-volume\") on node \"crc\" DevicePath \"\"" Mar 13 16:15:03 crc kubenswrapper[4907]: I0313 16:15:03.720615 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1d3954c-686f-4c87-bebd-95a286df5781-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "e1d3954c-686f-4c87-bebd-95a286df5781" (UID: "e1d3954c-686f-4c87-bebd-95a286df5781"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:15:03 crc kubenswrapper[4907]: I0313 16:15:03.720779 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e1d3954c-686f-4c87-bebd-95a286df5781-kube-api-access-cw7x4" (OuterVolumeSpecName: "kube-api-access-cw7x4") pod "e1d3954c-686f-4c87-bebd-95a286df5781" (UID: "e1d3954c-686f-4c87-bebd-95a286df5781"). InnerVolumeSpecName "kube-api-access-cw7x4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:15:03 crc kubenswrapper[4907]: I0313 16:15:03.817385 4907 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e1d3954c-686f-4c87-bebd-95a286df5781-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 13 16:15:03 crc kubenswrapper[4907]: I0313 16:15:03.817725 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cw7x4\" (UniqueName: \"kubernetes.io/projected/e1d3954c-686f-4c87-bebd-95a286df5781-kube-api-access-cw7x4\") on node \"crc\" DevicePath \"\"" Mar 13 16:15:04 crc kubenswrapper[4907]: I0313 16:15:04.154002 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556975-rmwsm" event={"ID":"e1d3954c-686f-4c87-bebd-95a286df5781","Type":"ContainerDied","Data":"7110db02ed258c363241254fa3b18e476c70ce370c9163d88c74a6f6dca09ec3"} Mar 13 16:15:04 crc kubenswrapper[4907]: I0313 16:15:04.154054 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556975-rmwsm" Mar 13 16:15:04 crc kubenswrapper[4907]: I0313 16:15:04.154061 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7110db02ed258c363241254fa3b18e476c70ce370c9163d88c74a6f6dca09ec3" Mar 13 16:15:04 crc kubenswrapper[4907]: I0313 16:15:04.602077 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556930-f852z"] Mar 13 16:15:04 crc kubenswrapper[4907]: I0313 16:15:04.611643 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556930-f852z"] Mar 13 16:15:05 crc kubenswrapper[4907]: I0313 16:15:05.794788 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2983956e-5d8d-4595-9fda-91759741010a" path="/var/lib/kubelet/pods/2983956e-5d8d-4595-9fda-91759741010a/volumes" Mar 13 16:15:18 crc kubenswrapper[4907]: I0313 16:15:18.042201 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 16:15:18 crc kubenswrapper[4907]: I0313 16:15:18.042723 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 16:15:34 crc kubenswrapper[4907]: I0313 16:15:34.464470 4907 generic.go:334] "Generic (PLEG): container finished" podID="3dc18cfe-d63a-4960-8c7d-10913cc78896" containerID="974721c911609f40407a7889cdfed0cb7412642260334ed010d57dbf5bb59c51" exitCode=0 Mar 13 16:15:34 crc kubenswrapper[4907]: I0313 16:15:34.464556 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-cell1-pwqj6" event={"ID":"3dc18cfe-d63a-4960-8c7d-10913cc78896","Type":"ContainerDied","Data":"974721c911609f40407a7889cdfed0cb7412642260334ed010d57dbf5bb59c51"} Mar 13 16:15:35 crc kubenswrapper[4907]: I0313 16:15:35.955103 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-cell1-pwqj6" Mar 13 16:15:36 crc kubenswrapper[4907]: I0313 16:15:36.135978 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3dc18cfe-d63a-4960-8c7d-10913cc78896-ceph\") pod \"3dc18cfe-d63a-4960-8c7d-10913cc78896\" (UID: \"3dc18cfe-d63a-4960-8c7d-10913cc78896\") " Mar 13 16:15:36 crc kubenswrapper[4907]: I0313 16:15:36.136214 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/3dc18cfe-d63a-4960-8c7d-10913cc78896-ssh-key-openstack-cell1\") pod \"3dc18cfe-d63a-4960-8c7d-10913cc78896\" (UID: \"3dc18cfe-d63a-4960-8c7d-10913cc78896\") " Mar 13 16:15:36 crc kubenswrapper[4907]: I0313 16:15:36.136272 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3dc18cfe-d63a-4960-8c7d-10913cc78896-inventory\") pod \"3dc18cfe-d63a-4960-8c7d-10913cc78896\" (UID: \"3dc18cfe-d63a-4960-8c7d-10913cc78896\") " Mar 13 16:15:36 crc kubenswrapper[4907]: I0313 16:15:36.136549 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/3dc18cfe-d63a-4960-8c7d-10913cc78896-nova-metadata-neutron-config-0\") pod \"3dc18cfe-d63a-4960-8c7d-10913cc78896\" (UID: \"3dc18cfe-d63a-4960-8c7d-10913cc78896\") " Mar 13 16:15:36 crc kubenswrapper[4907]: I0313 16:15:36.136601 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/3dc18cfe-d63a-4960-8c7d-10913cc78896-neutron-ovn-metadata-agent-neutron-config-0\") pod \"3dc18cfe-d63a-4960-8c7d-10913cc78896\" (UID: \"3dc18cfe-d63a-4960-8c7d-10913cc78896\") " Mar 13 16:15:36 crc kubenswrapper[4907]: I0313 16:15:36.136700 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3dc18cfe-d63a-4960-8c7d-10913cc78896-neutron-metadata-combined-ca-bundle\") pod \"3dc18cfe-d63a-4960-8c7d-10913cc78896\" (UID: \"3dc18cfe-d63a-4960-8c7d-10913cc78896\") " Mar 13 16:15:36 crc kubenswrapper[4907]: I0313 16:15:36.136758 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bcsp6\" (UniqueName: \"kubernetes.io/projected/3dc18cfe-d63a-4960-8c7d-10913cc78896-kube-api-access-bcsp6\") pod \"3dc18cfe-d63a-4960-8c7d-10913cc78896\" (UID: \"3dc18cfe-d63a-4960-8c7d-10913cc78896\") " Mar 13 16:15:36 crc kubenswrapper[4907]: I0313 16:15:36.145481 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3dc18cfe-d63a-4960-8c7d-10913cc78896-kube-api-access-bcsp6" (OuterVolumeSpecName: "kube-api-access-bcsp6") pod "3dc18cfe-d63a-4960-8c7d-10913cc78896" (UID: "3dc18cfe-d63a-4960-8c7d-10913cc78896"). InnerVolumeSpecName "kube-api-access-bcsp6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:15:36 crc kubenswrapper[4907]: I0313 16:15:36.146231 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3dc18cfe-d63a-4960-8c7d-10913cc78896-ceph" (OuterVolumeSpecName: "ceph") pod "3dc18cfe-d63a-4960-8c7d-10913cc78896" (UID: "3dc18cfe-d63a-4960-8c7d-10913cc78896"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:15:36 crc kubenswrapper[4907]: I0313 16:15:36.147896 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3dc18cfe-d63a-4960-8c7d-10913cc78896-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "3dc18cfe-d63a-4960-8c7d-10913cc78896" (UID: "3dc18cfe-d63a-4960-8c7d-10913cc78896"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:15:36 crc kubenswrapper[4907]: I0313 16:15:36.172368 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3dc18cfe-d63a-4960-8c7d-10913cc78896-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "3dc18cfe-d63a-4960-8c7d-10913cc78896" (UID: "3dc18cfe-d63a-4960-8c7d-10913cc78896"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:15:36 crc kubenswrapper[4907]: I0313 16:15:36.174183 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3dc18cfe-d63a-4960-8c7d-10913cc78896-ssh-key-openstack-cell1" (OuterVolumeSpecName: "ssh-key-openstack-cell1") pod "3dc18cfe-d63a-4960-8c7d-10913cc78896" (UID: "3dc18cfe-d63a-4960-8c7d-10913cc78896"). InnerVolumeSpecName "ssh-key-openstack-cell1". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:15:36 crc kubenswrapper[4907]: I0313 16:15:36.174556 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3dc18cfe-d63a-4960-8c7d-10913cc78896-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "3dc18cfe-d63a-4960-8c7d-10913cc78896" (UID: "3dc18cfe-d63a-4960-8c7d-10913cc78896"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:15:36 crc kubenswrapper[4907]: I0313 16:15:36.175981 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3dc18cfe-d63a-4960-8c7d-10913cc78896-inventory" (OuterVolumeSpecName: "inventory") pod "3dc18cfe-d63a-4960-8c7d-10913cc78896" (UID: "3dc18cfe-d63a-4960-8c7d-10913cc78896"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:15:36 crc kubenswrapper[4907]: I0313 16:15:36.241751 4907 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/3dc18cfe-d63a-4960-8c7d-10913cc78896-ssh-key-openstack-cell1\") on node \"crc\" DevicePath \"\"" Mar 13 16:15:36 crc kubenswrapper[4907]: I0313 16:15:36.241809 4907 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3dc18cfe-d63a-4960-8c7d-10913cc78896-inventory\") on node \"crc\" DevicePath \"\"" Mar 13 16:15:36 crc kubenswrapper[4907]: I0313 16:15:36.241822 4907 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/3dc18cfe-d63a-4960-8c7d-10913cc78896-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Mar 13 16:15:36 crc kubenswrapper[4907]: I0313 16:15:36.241864 4907 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/3dc18cfe-d63a-4960-8c7d-10913cc78896-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Mar 13 16:15:36 crc kubenswrapper[4907]: I0313 16:15:36.241874 4907 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3dc18cfe-d63a-4960-8c7d-10913cc78896-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 16:15:36 crc kubenswrapper[4907]: I0313 16:15:36.241903 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bcsp6\" (UniqueName: \"kubernetes.io/projected/3dc18cfe-d63a-4960-8c7d-10913cc78896-kube-api-access-bcsp6\") on node \"crc\" DevicePath \"\"" Mar 13 16:15:36 crc kubenswrapper[4907]: I0313 16:15:36.241915 4907 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3dc18cfe-d63a-4960-8c7d-10913cc78896-ceph\") on node \"crc\" DevicePath \"\"" Mar 13 16:15:36 crc kubenswrapper[4907]: I0313 16:15:36.485805 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-cell1-pwqj6" event={"ID":"3dc18cfe-d63a-4960-8c7d-10913cc78896","Type":"ContainerDied","Data":"f4a4333655b463ae652caae8d8ffaa8b9979eafbf031d280da29757bdd65f7a4"} Mar 13 16:15:36 crc kubenswrapper[4907]: I0313 16:15:36.486215 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f4a4333655b463ae652caae8d8ffaa8b9979eafbf031d280da29757bdd65f7a4" Mar 13 16:15:36 crc kubenswrapper[4907]: I0313 16:15:36.485861 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-cell1-pwqj6" Mar 13 16:15:36 crc kubenswrapper[4907]: I0313 16:15:36.654328 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-openstack-openstack-cell1-bdwbh"] Mar 13 16:15:36 crc kubenswrapper[4907]: E0313 16:15:36.654800 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1d3954c-686f-4c87-bebd-95a286df5781" containerName="collect-profiles" Mar 13 16:15:36 crc kubenswrapper[4907]: I0313 16:15:36.654821 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1d3954c-686f-4c87-bebd-95a286df5781" containerName="collect-profiles" Mar 13 16:15:36 crc kubenswrapper[4907]: E0313 16:15:36.654843 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3dc18cfe-d63a-4960-8c7d-10913cc78896" containerName="neutron-metadata-openstack-openstack-cell1" Mar 13 16:15:36 crc kubenswrapper[4907]: I0313 16:15:36.654854 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="3dc18cfe-d63a-4960-8c7d-10913cc78896" containerName="neutron-metadata-openstack-openstack-cell1" Mar 13 16:15:36 crc kubenswrapper[4907]: I0313 16:15:36.655155 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1d3954c-686f-4c87-bebd-95a286df5781" containerName="collect-profiles" Mar 13 16:15:36 crc kubenswrapper[4907]: I0313 16:15:36.655176 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="3dc18cfe-d63a-4960-8c7d-10913cc78896" containerName="neutron-metadata-openstack-openstack-cell1" Mar 13 16:15:36 crc kubenswrapper[4907]: I0313 16:15:36.656017 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-openstack-openstack-cell1-bdwbh" Mar 13 16:15:36 crc kubenswrapper[4907]: I0313 16:15:36.660137 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 13 16:15:36 crc kubenswrapper[4907]: I0313 16:15:36.660285 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Mar 13 16:15:36 crc kubenswrapper[4907]: I0313 16:15:36.660653 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Mar 13 16:15:36 crc kubenswrapper[4907]: I0313 16:15:36.661351 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-44kxw" Mar 13 16:15:36 crc kubenswrapper[4907]: I0313 16:15:36.661508 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Mar 13 16:15:36 crc kubenswrapper[4907]: I0313 16:15:36.676518 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-openstack-openstack-cell1-bdwbh"] Mar 13 16:15:36 crc kubenswrapper[4907]: I0313 16:15:36.751146 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/82c7c5cb-4f6b-4d38-a009-263025faa45b-ssh-key-openstack-cell1\") pod \"libvirt-openstack-openstack-cell1-bdwbh\" (UID: \"82c7c5cb-4f6b-4d38-a009-263025faa45b\") " pod="openstack/libvirt-openstack-openstack-cell1-bdwbh" Mar 13 16:15:36 crc kubenswrapper[4907]: I0313 16:15:36.751190 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82c7c5cb-4f6b-4d38-a009-263025faa45b-libvirt-combined-ca-bundle\") pod \"libvirt-openstack-openstack-cell1-bdwbh\" (UID: \"82c7c5cb-4f6b-4d38-a009-263025faa45b\") " pod="openstack/libvirt-openstack-openstack-cell1-bdwbh" Mar 13 16:15:36 crc kubenswrapper[4907]: I0313 16:15:36.751221 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/82c7c5cb-4f6b-4d38-a009-263025faa45b-ceph\") pod \"libvirt-openstack-openstack-cell1-bdwbh\" (UID: \"82c7c5cb-4f6b-4d38-a009-263025faa45b\") " pod="openstack/libvirt-openstack-openstack-cell1-bdwbh" Mar 13 16:15:36 crc kubenswrapper[4907]: I0313 16:15:36.751383 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/82c7c5cb-4f6b-4d38-a009-263025faa45b-libvirt-secret-0\") pod \"libvirt-openstack-openstack-cell1-bdwbh\" (UID: \"82c7c5cb-4f6b-4d38-a009-263025faa45b\") " pod="openstack/libvirt-openstack-openstack-cell1-bdwbh" Mar 13 16:15:36 crc kubenswrapper[4907]: I0313 16:15:36.751575 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h7hx5\" (UniqueName: \"kubernetes.io/projected/82c7c5cb-4f6b-4d38-a009-263025faa45b-kube-api-access-h7hx5\") pod \"libvirt-openstack-openstack-cell1-bdwbh\" (UID: \"82c7c5cb-4f6b-4d38-a009-263025faa45b\") " pod="openstack/libvirt-openstack-openstack-cell1-bdwbh" Mar 13 16:15:36 crc kubenswrapper[4907]: I0313 16:15:36.751703 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/82c7c5cb-4f6b-4d38-a009-263025faa45b-inventory\") pod \"libvirt-openstack-openstack-cell1-bdwbh\" (UID: \"82c7c5cb-4f6b-4d38-a009-263025faa45b\") " pod="openstack/libvirt-openstack-openstack-cell1-bdwbh" Mar 13 16:15:36 crc kubenswrapper[4907]: I0313 16:15:36.853517 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/82c7c5cb-4f6b-4d38-a009-263025faa45b-ssh-key-openstack-cell1\") pod \"libvirt-openstack-openstack-cell1-bdwbh\" (UID: \"82c7c5cb-4f6b-4d38-a009-263025faa45b\") " pod="openstack/libvirt-openstack-openstack-cell1-bdwbh" Mar 13 16:15:36 crc kubenswrapper[4907]: I0313 16:15:36.853598 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82c7c5cb-4f6b-4d38-a009-263025faa45b-libvirt-combined-ca-bundle\") pod \"libvirt-openstack-openstack-cell1-bdwbh\" (UID: \"82c7c5cb-4f6b-4d38-a009-263025faa45b\") " pod="openstack/libvirt-openstack-openstack-cell1-bdwbh" Mar 13 16:15:36 crc kubenswrapper[4907]: I0313 16:15:36.853647 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/82c7c5cb-4f6b-4d38-a009-263025faa45b-ceph\") pod \"libvirt-openstack-openstack-cell1-bdwbh\" (UID: \"82c7c5cb-4f6b-4d38-a009-263025faa45b\") " pod="openstack/libvirt-openstack-openstack-cell1-bdwbh" Mar 13 16:15:36 crc kubenswrapper[4907]: I0313 16:15:36.853699 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/82c7c5cb-4f6b-4d38-a009-263025faa45b-libvirt-secret-0\") pod \"libvirt-openstack-openstack-cell1-bdwbh\" (UID: \"82c7c5cb-4f6b-4d38-a009-263025faa45b\") " pod="openstack/libvirt-openstack-openstack-cell1-bdwbh" Mar 13 16:15:36 crc kubenswrapper[4907]: I0313 16:15:36.853765 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h7hx5\" (UniqueName: \"kubernetes.io/projected/82c7c5cb-4f6b-4d38-a009-263025faa45b-kube-api-access-h7hx5\") pod \"libvirt-openstack-openstack-cell1-bdwbh\" (UID: \"82c7c5cb-4f6b-4d38-a009-263025faa45b\") " pod="openstack/libvirt-openstack-openstack-cell1-bdwbh" Mar 13 16:15:36 crc kubenswrapper[4907]: I0313 16:15:36.853828 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/82c7c5cb-4f6b-4d38-a009-263025faa45b-inventory\") pod \"libvirt-openstack-openstack-cell1-bdwbh\" (UID: \"82c7c5cb-4f6b-4d38-a009-263025faa45b\") " pod="openstack/libvirt-openstack-openstack-cell1-bdwbh" Mar 13 16:15:36 crc kubenswrapper[4907]: I0313 16:15:36.857863 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/82c7c5cb-4f6b-4d38-a009-263025faa45b-inventory\") pod \"libvirt-openstack-openstack-cell1-bdwbh\" (UID: \"82c7c5cb-4f6b-4d38-a009-263025faa45b\") " pod="openstack/libvirt-openstack-openstack-cell1-bdwbh" Mar 13 16:15:36 crc kubenswrapper[4907]: I0313 16:15:36.858339 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82c7c5cb-4f6b-4d38-a009-263025faa45b-libvirt-combined-ca-bundle\") pod \"libvirt-openstack-openstack-cell1-bdwbh\" (UID: \"82c7c5cb-4f6b-4d38-a009-263025faa45b\") " pod="openstack/libvirt-openstack-openstack-cell1-bdwbh" Mar 13 16:15:36 crc kubenswrapper[4907]: I0313 16:15:36.859650 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/82c7c5cb-4f6b-4d38-a009-263025faa45b-ceph\") pod \"libvirt-openstack-openstack-cell1-bdwbh\" (UID: \"82c7c5cb-4f6b-4d38-a009-263025faa45b\") " pod="openstack/libvirt-openstack-openstack-cell1-bdwbh" Mar 13 16:15:36 crc kubenswrapper[4907]: I0313 16:15:36.860356 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/82c7c5cb-4f6b-4d38-a009-263025faa45b-ssh-key-openstack-cell1\") pod \"libvirt-openstack-openstack-cell1-bdwbh\" (UID: \"82c7c5cb-4f6b-4d38-a009-263025faa45b\") " pod="openstack/libvirt-openstack-openstack-cell1-bdwbh" Mar 13 16:15:36 crc kubenswrapper[4907]: I0313 16:15:36.862771 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/82c7c5cb-4f6b-4d38-a009-263025faa45b-libvirt-secret-0\") pod \"libvirt-openstack-openstack-cell1-bdwbh\" (UID: \"82c7c5cb-4f6b-4d38-a009-263025faa45b\") " pod="openstack/libvirt-openstack-openstack-cell1-bdwbh" Mar 13 16:15:36 crc kubenswrapper[4907]: I0313 16:15:36.871534 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h7hx5\" (UniqueName: \"kubernetes.io/projected/82c7c5cb-4f6b-4d38-a009-263025faa45b-kube-api-access-h7hx5\") pod \"libvirt-openstack-openstack-cell1-bdwbh\" (UID: \"82c7c5cb-4f6b-4d38-a009-263025faa45b\") " pod="openstack/libvirt-openstack-openstack-cell1-bdwbh" Mar 13 16:15:37 crc kubenswrapper[4907]: I0313 16:15:37.036057 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-openstack-openstack-cell1-bdwbh" Mar 13 16:15:37 crc kubenswrapper[4907]: I0313 16:15:37.588160 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-openstack-openstack-cell1-bdwbh"] Mar 13 16:15:38 crc kubenswrapper[4907]: I0313 16:15:38.505874 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-openstack-openstack-cell1-bdwbh" event={"ID":"82c7c5cb-4f6b-4d38-a009-263025faa45b","Type":"ContainerStarted","Data":"9ec4cbc3bc8367d5cb8d52b834231c90815a53ebf59eb9499587563be45dd71a"} Mar 13 16:15:38 crc kubenswrapper[4907]: I0313 16:15:38.506562 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-openstack-openstack-cell1-bdwbh" event={"ID":"82c7c5cb-4f6b-4d38-a009-263025faa45b","Type":"ContainerStarted","Data":"e4c7c890b0bd5338464cbbe531bda4c9b8efa360ef93c0753d3a22a75540214d"} Mar 13 16:15:38 crc kubenswrapper[4907]: I0313 16:15:38.532839 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-openstack-openstack-cell1-bdwbh" podStartSLOduration=2.051412183 podStartE2EDuration="2.532821719s" podCreationTimestamp="2026-03-13 16:15:36 +0000 UTC" firstStartedPulling="2026-03-13 16:15:37.597014747 +0000 UTC m=+7836.496802436" lastFinishedPulling="2026-03-13 16:15:38.078424283 +0000 UTC m=+7836.978211972" observedRunningTime="2026-03-13 16:15:38.520227314 +0000 UTC m=+7837.420015013" watchObservedRunningTime="2026-03-13 16:15:38.532821719 +0000 UTC m=+7837.432609408" Mar 13 16:15:41 crc kubenswrapper[4907]: I0313 16:15:41.775721 4907 scope.go:117] "RemoveContainer" containerID="74d889f05e8d8fa3b0cfc7fed167cff28426e276bf0f0c9f759dd1496f3c1058" Mar 13 16:15:48 crc kubenswrapper[4907]: I0313 16:15:48.042308 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 16:15:48 crc kubenswrapper[4907]: I0313 16:15:48.042966 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 16:15:48 crc kubenswrapper[4907]: I0313 16:15:48.043015 4907 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" Mar 13 16:15:48 crc kubenswrapper[4907]: I0313 16:15:48.043913 4907 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"300e429cce24fa4eb9aefff9494fe9905ee7b0767504365170f1f9645a3c8b5f"} pod="openshift-machine-config-operator/machine-config-daemon-hm56j" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 16:15:48 crc kubenswrapper[4907]: I0313 16:15:48.043980 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" containerID="cri-o://300e429cce24fa4eb9aefff9494fe9905ee7b0767504365170f1f9645a3c8b5f" gracePeriod=600 Mar 13 16:15:48 crc kubenswrapper[4907]: E0313 16:15:48.198077 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:15:48 crc kubenswrapper[4907]: I0313 16:15:48.661946 4907 generic.go:334] "Generic (PLEG): container finished" podID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerID="300e429cce24fa4eb9aefff9494fe9905ee7b0767504365170f1f9645a3c8b5f" exitCode=0 Mar 13 16:15:48 crc kubenswrapper[4907]: I0313 16:15:48.661998 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerDied","Data":"300e429cce24fa4eb9aefff9494fe9905ee7b0767504365170f1f9645a3c8b5f"} Mar 13 16:15:48 crc kubenswrapper[4907]: I0313 16:15:48.662035 4907 scope.go:117] "RemoveContainer" containerID="c38b8b19c660dee13d11b404db7a1f6a1a7c1615cf40ea14a2875b4e3648d560" Mar 13 16:15:48 crc kubenswrapper[4907]: I0313 16:15:48.663965 4907 scope.go:117] "RemoveContainer" containerID="300e429cce24fa4eb9aefff9494fe9905ee7b0767504365170f1f9645a3c8b5f" Mar 13 16:15:48 crc kubenswrapper[4907]: E0313 16:15:48.664397 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:16:00 crc kubenswrapper[4907]: I0313 16:16:00.148576 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556976-twcx4"] Mar 13 16:16:00 crc kubenswrapper[4907]: I0313 16:16:00.151113 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556976-twcx4" Mar 13 16:16:00 crc kubenswrapper[4907]: I0313 16:16:00.154660 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 16:16:00 crc kubenswrapper[4907]: I0313 16:16:00.155178 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 16:16:00 crc kubenswrapper[4907]: I0313 16:16:00.160064 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 16:16:00 crc kubenswrapper[4907]: I0313 16:16:00.162933 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556976-twcx4"] Mar 13 16:16:00 crc kubenswrapper[4907]: I0313 16:16:00.257310 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5vkvw\" (UniqueName: \"kubernetes.io/projected/362f626f-a4dd-45a7-ad88-4c539732eebb-kube-api-access-5vkvw\") pod \"auto-csr-approver-29556976-twcx4\" (UID: \"362f626f-a4dd-45a7-ad88-4c539732eebb\") " pod="openshift-infra/auto-csr-approver-29556976-twcx4" Mar 13 16:16:00 crc kubenswrapper[4907]: I0313 16:16:00.359257 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5vkvw\" (UniqueName: \"kubernetes.io/projected/362f626f-a4dd-45a7-ad88-4c539732eebb-kube-api-access-5vkvw\") pod \"auto-csr-approver-29556976-twcx4\" (UID: \"362f626f-a4dd-45a7-ad88-4c539732eebb\") " pod="openshift-infra/auto-csr-approver-29556976-twcx4" Mar 13 16:16:00 crc kubenswrapper[4907]: I0313 16:16:00.378200 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5vkvw\" (UniqueName: \"kubernetes.io/projected/362f626f-a4dd-45a7-ad88-4c539732eebb-kube-api-access-5vkvw\") pod \"auto-csr-approver-29556976-twcx4\" (UID: \"362f626f-a4dd-45a7-ad88-4c539732eebb\") " pod="openshift-infra/auto-csr-approver-29556976-twcx4" Mar 13 16:16:00 crc kubenswrapper[4907]: I0313 16:16:00.473491 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556976-twcx4" Mar 13 16:16:00 crc kubenswrapper[4907]: I0313 16:16:00.941709 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556976-twcx4"] Mar 13 16:16:00 crc kubenswrapper[4907]: W0313 16:16:00.950064 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod362f626f_a4dd_45a7_ad88_4c539732eebb.slice/crio-bda227a271f38ea0cca40341cba0484c8a2a657702d708161303a4775090dbbe WatchSource:0}: Error finding container bda227a271f38ea0cca40341cba0484c8a2a657702d708161303a4775090dbbe: Status 404 returned error can't find the container with id bda227a271f38ea0cca40341cba0484c8a2a657702d708161303a4775090dbbe Mar 13 16:16:01 crc kubenswrapper[4907]: I0313 16:16:01.802127 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556976-twcx4" event={"ID":"362f626f-a4dd-45a7-ad88-4c539732eebb","Type":"ContainerStarted","Data":"bda227a271f38ea0cca40341cba0484c8a2a657702d708161303a4775090dbbe"} Mar 13 16:16:01 crc kubenswrapper[4907]: I0313 16:16:01.802636 4907 scope.go:117] "RemoveContainer" containerID="300e429cce24fa4eb9aefff9494fe9905ee7b0767504365170f1f9645a3c8b5f" Mar 13 16:16:01 crc kubenswrapper[4907]: E0313 16:16:01.803158 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:16:02 crc kubenswrapper[4907]: I0313 16:16:02.818020 4907 generic.go:334] "Generic (PLEG): container finished" podID="362f626f-a4dd-45a7-ad88-4c539732eebb" containerID="7e0af7bb09c83c72fd3ad1550fb6bc5bec78d9ec9c71037ff20efe437fc967c9" exitCode=0 Mar 13 16:16:02 crc kubenswrapper[4907]: I0313 16:16:02.818119 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556976-twcx4" event={"ID":"362f626f-a4dd-45a7-ad88-4c539732eebb","Type":"ContainerDied","Data":"7e0af7bb09c83c72fd3ad1550fb6bc5bec78d9ec9c71037ff20efe437fc967c9"} Mar 13 16:16:04 crc kubenswrapper[4907]: I0313 16:16:04.192694 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556976-twcx4" Mar 13 16:16:04 crc kubenswrapper[4907]: I0313 16:16:04.345014 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5vkvw\" (UniqueName: \"kubernetes.io/projected/362f626f-a4dd-45a7-ad88-4c539732eebb-kube-api-access-5vkvw\") pod \"362f626f-a4dd-45a7-ad88-4c539732eebb\" (UID: \"362f626f-a4dd-45a7-ad88-4c539732eebb\") " Mar 13 16:16:04 crc kubenswrapper[4907]: I0313 16:16:04.350467 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/362f626f-a4dd-45a7-ad88-4c539732eebb-kube-api-access-5vkvw" (OuterVolumeSpecName: "kube-api-access-5vkvw") pod "362f626f-a4dd-45a7-ad88-4c539732eebb" (UID: "362f626f-a4dd-45a7-ad88-4c539732eebb"). InnerVolumeSpecName "kube-api-access-5vkvw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:16:04 crc kubenswrapper[4907]: I0313 16:16:04.450579 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5vkvw\" (UniqueName: \"kubernetes.io/projected/362f626f-a4dd-45a7-ad88-4c539732eebb-kube-api-access-5vkvw\") on node \"crc\" DevicePath \"\"" Mar 13 16:16:04 crc kubenswrapper[4907]: I0313 16:16:04.857349 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556976-twcx4" event={"ID":"362f626f-a4dd-45a7-ad88-4c539732eebb","Type":"ContainerDied","Data":"bda227a271f38ea0cca40341cba0484c8a2a657702d708161303a4775090dbbe"} Mar 13 16:16:04 crc kubenswrapper[4907]: I0313 16:16:04.857976 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bda227a271f38ea0cca40341cba0484c8a2a657702d708161303a4775090dbbe" Mar 13 16:16:04 crc kubenswrapper[4907]: I0313 16:16:04.857413 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556976-twcx4" Mar 13 16:16:05 crc kubenswrapper[4907]: I0313 16:16:05.291125 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556970-lbxp9"] Mar 13 16:16:05 crc kubenswrapper[4907]: I0313 16:16:05.302838 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556970-lbxp9"] Mar 13 16:16:05 crc kubenswrapper[4907]: I0313 16:16:05.796231 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2b3a466a-0c1c-4c86-897f-0a3375a9fb53" path="/var/lib/kubelet/pods/2b3a466a-0c1c-4c86-897f-0a3375a9fb53/volumes" Mar 13 16:16:14 crc kubenswrapper[4907]: I0313 16:16:14.782824 4907 scope.go:117] "RemoveContainer" containerID="300e429cce24fa4eb9aefff9494fe9905ee7b0767504365170f1f9645a3c8b5f" Mar 13 16:16:14 crc kubenswrapper[4907]: E0313 16:16:14.783786 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:16:29 crc kubenswrapper[4907]: I0313 16:16:29.782966 4907 scope.go:117] "RemoveContainer" containerID="300e429cce24fa4eb9aefff9494fe9905ee7b0767504365170f1f9645a3c8b5f" Mar 13 16:16:29 crc kubenswrapper[4907]: E0313 16:16:29.784070 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:16:41 crc kubenswrapper[4907]: I0313 16:16:41.862222 4907 scope.go:117] "RemoveContainer" containerID="1ffb094832567d23bf593fb1872edfb355e12059011cdaa6ddadd04821d723cc" Mar 13 16:16:44 crc kubenswrapper[4907]: I0313 16:16:44.783282 4907 scope.go:117] "RemoveContainer" containerID="300e429cce24fa4eb9aefff9494fe9905ee7b0767504365170f1f9645a3c8b5f" Mar 13 16:16:44 crc kubenswrapper[4907]: E0313 16:16:44.784329 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:16:59 crc kubenswrapper[4907]: I0313 16:16:59.782737 4907 scope.go:117] "RemoveContainer" containerID="300e429cce24fa4eb9aefff9494fe9905ee7b0767504365170f1f9645a3c8b5f" Mar 13 16:16:59 crc kubenswrapper[4907]: E0313 16:16:59.785234 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:17:14 crc kubenswrapper[4907]: I0313 16:17:14.781978 4907 scope.go:117] "RemoveContainer" containerID="300e429cce24fa4eb9aefff9494fe9905ee7b0767504365170f1f9645a3c8b5f" Mar 13 16:17:14 crc kubenswrapper[4907]: E0313 16:17:14.782788 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:17:29 crc kubenswrapper[4907]: I0313 16:17:29.784023 4907 scope.go:117] "RemoveContainer" containerID="300e429cce24fa4eb9aefff9494fe9905ee7b0767504365170f1f9645a3c8b5f" Mar 13 16:17:29 crc kubenswrapper[4907]: E0313 16:17:29.787066 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:17:40 crc kubenswrapper[4907]: I0313 16:17:40.782697 4907 scope.go:117] "RemoveContainer" containerID="300e429cce24fa4eb9aefff9494fe9905ee7b0767504365170f1f9645a3c8b5f" Mar 13 16:17:40 crc kubenswrapper[4907]: E0313 16:17:40.783535 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:17:48 crc kubenswrapper[4907]: I0313 16:17:48.787184 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-4cppb"] Mar 13 16:17:48 crc kubenswrapper[4907]: E0313 16:17:48.788581 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="362f626f-a4dd-45a7-ad88-4c539732eebb" containerName="oc" Mar 13 16:17:48 crc kubenswrapper[4907]: I0313 16:17:48.788602 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="362f626f-a4dd-45a7-ad88-4c539732eebb" containerName="oc" Mar 13 16:17:48 crc kubenswrapper[4907]: I0313 16:17:48.788912 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="362f626f-a4dd-45a7-ad88-4c539732eebb" containerName="oc" Mar 13 16:17:48 crc kubenswrapper[4907]: I0313 16:17:48.791004 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4cppb" Mar 13 16:17:48 crc kubenswrapper[4907]: I0313 16:17:48.824136 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4cppb"] Mar 13 16:17:48 crc kubenswrapper[4907]: I0313 16:17:48.870702 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d417de61-9b72-4c15-bfae-6f1b6ee03849-catalog-content\") pod \"redhat-marketplace-4cppb\" (UID: \"d417de61-9b72-4c15-bfae-6f1b6ee03849\") " pod="openshift-marketplace/redhat-marketplace-4cppb" Mar 13 16:17:48 crc kubenswrapper[4907]: I0313 16:17:48.870803 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vf7bk\" (UniqueName: \"kubernetes.io/projected/d417de61-9b72-4c15-bfae-6f1b6ee03849-kube-api-access-vf7bk\") pod \"redhat-marketplace-4cppb\" (UID: \"d417de61-9b72-4c15-bfae-6f1b6ee03849\") " pod="openshift-marketplace/redhat-marketplace-4cppb" Mar 13 16:17:48 crc kubenswrapper[4907]: I0313 16:17:48.871309 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d417de61-9b72-4c15-bfae-6f1b6ee03849-utilities\") pod \"redhat-marketplace-4cppb\" (UID: \"d417de61-9b72-4c15-bfae-6f1b6ee03849\") " pod="openshift-marketplace/redhat-marketplace-4cppb" Mar 13 16:17:48 crc kubenswrapper[4907]: I0313 16:17:48.973909 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vf7bk\" (UniqueName: \"kubernetes.io/projected/d417de61-9b72-4c15-bfae-6f1b6ee03849-kube-api-access-vf7bk\") pod \"redhat-marketplace-4cppb\" (UID: \"d417de61-9b72-4c15-bfae-6f1b6ee03849\") " pod="openshift-marketplace/redhat-marketplace-4cppb" Mar 13 16:17:48 crc kubenswrapper[4907]: I0313 16:17:48.973998 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d417de61-9b72-4c15-bfae-6f1b6ee03849-utilities\") pod \"redhat-marketplace-4cppb\" (UID: \"d417de61-9b72-4c15-bfae-6f1b6ee03849\") " pod="openshift-marketplace/redhat-marketplace-4cppb" Mar 13 16:17:48 crc kubenswrapper[4907]: I0313 16:17:48.974133 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d417de61-9b72-4c15-bfae-6f1b6ee03849-catalog-content\") pod \"redhat-marketplace-4cppb\" (UID: \"d417de61-9b72-4c15-bfae-6f1b6ee03849\") " pod="openshift-marketplace/redhat-marketplace-4cppb" Mar 13 16:17:48 crc kubenswrapper[4907]: I0313 16:17:48.974680 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d417de61-9b72-4c15-bfae-6f1b6ee03849-catalog-content\") pod \"redhat-marketplace-4cppb\" (UID: \"d417de61-9b72-4c15-bfae-6f1b6ee03849\") " pod="openshift-marketplace/redhat-marketplace-4cppb" Mar 13 16:17:48 crc kubenswrapper[4907]: I0313 16:17:48.975001 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d417de61-9b72-4c15-bfae-6f1b6ee03849-utilities\") pod \"redhat-marketplace-4cppb\" (UID: \"d417de61-9b72-4c15-bfae-6f1b6ee03849\") " pod="openshift-marketplace/redhat-marketplace-4cppb" Mar 13 16:17:49 crc kubenswrapper[4907]: I0313 16:17:49.006958 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vf7bk\" (UniqueName: \"kubernetes.io/projected/d417de61-9b72-4c15-bfae-6f1b6ee03849-kube-api-access-vf7bk\") pod \"redhat-marketplace-4cppb\" (UID: \"d417de61-9b72-4c15-bfae-6f1b6ee03849\") " pod="openshift-marketplace/redhat-marketplace-4cppb" Mar 13 16:17:49 crc kubenswrapper[4907]: I0313 16:17:49.131786 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4cppb" Mar 13 16:17:49 crc kubenswrapper[4907]: I0313 16:17:49.598564 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4cppb"] Mar 13 16:17:50 crc kubenswrapper[4907]: I0313 16:17:50.080461 4907 generic.go:334] "Generic (PLEG): container finished" podID="d417de61-9b72-4c15-bfae-6f1b6ee03849" containerID="e3c0406c3eaa17bc2c3ae343da02515c95a6e17e38c2bc85e8e82bdf569d1c03" exitCode=0 Mar 13 16:17:50 crc kubenswrapper[4907]: I0313 16:17:50.080512 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4cppb" event={"ID":"d417de61-9b72-4c15-bfae-6f1b6ee03849","Type":"ContainerDied","Data":"e3c0406c3eaa17bc2c3ae343da02515c95a6e17e38c2bc85e8e82bdf569d1c03"} Mar 13 16:17:50 crc kubenswrapper[4907]: I0313 16:17:50.080563 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4cppb" event={"ID":"d417de61-9b72-4c15-bfae-6f1b6ee03849","Type":"ContainerStarted","Data":"5a312c55abb195fc381c69253b66e281eba78efb1b39b9a51a72abfc897857dc"} Mar 13 16:17:50 crc kubenswrapper[4907]: I0313 16:17:50.085521 4907 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 13 16:17:51 crc kubenswrapper[4907]: I0313 16:17:51.091325 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4cppb" event={"ID":"d417de61-9b72-4c15-bfae-6f1b6ee03849","Type":"ContainerStarted","Data":"cc8ae91cc5b08e53b842e2f7a69a2fbd79580167c23866df62a867359907d98f"} Mar 13 16:17:52 crc kubenswrapper[4907]: I0313 16:17:52.105096 4907 generic.go:334] "Generic (PLEG): container finished" podID="d417de61-9b72-4c15-bfae-6f1b6ee03849" containerID="cc8ae91cc5b08e53b842e2f7a69a2fbd79580167c23866df62a867359907d98f" exitCode=0 Mar 13 16:17:52 crc kubenswrapper[4907]: I0313 16:17:52.105150 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4cppb" event={"ID":"d417de61-9b72-4c15-bfae-6f1b6ee03849","Type":"ContainerDied","Data":"cc8ae91cc5b08e53b842e2f7a69a2fbd79580167c23866df62a867359907d98f"} Mar 13 16:17:53 crc kubenswrapper[4907]: I0313 16:17:53.116116 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4cppb" event={"ID":"d417de61-9b72-4c15-bfae-6f1b6ee03849","Type":"ContainerStarted","Data":"5bcc136eb40ce02d1e8aec26066c465f5d2b04e99adae87f346f7d3227130ae6"} Mar 13 16:17:53 crc kubenswrapper[4907]: I0313 16:17:53.137262 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-4cppb" podStartSLOduration=2.604871424 podStartE2EDuration="5.137244139s" podCreationTimestamp="2026-03-13 16:17:48 +0000 UTC" firstStartedPulling="2026-03-13 16:17:50.084665006 +0000 UTC m=+7968.984452705" lastFinishedPulling="2026-03-13 16:17:52.617037731 +0000 UTC m=+7971.516825420" observedRunningTime="2026-03-13 16:17:53.132621932 +0000 UTC m=+7972.032409621" watchObservedRunningTime="2026-03-13 16:17:53.137244139 +0000 UTC m=+7972.037031828" Mar 13 16:17:55 crc kubenswrapper[4907]: I0313 16:17:55.782616 4907 scope.go:117] "RemoveContainer" containerID="300e429cce24fa4eb9aefff9494fe9905ee7b0767504365170f1f9645a3c8b5f" Mar 13 16:17:55 crc kubenswrapper[4907]: E0313 16:17:55.783467 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:17:59 crc kubenswrapper[4907]: I0313 16:17:59.132637 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-4cppb" Mar 13 16:17:59 crc kubenswrapper[4907]: I0313 16:17:59.134092 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-4cppb" Mar 13 16:17:59 crc kubenswrapper[4907]: I0313 16:17:59.189682 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-4cppb" Mar 13 16:17:59 crc kubenswrapper[4907]: I0313 16:17:59.260862 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-4cppb" Mar 13 16:17:59 crc kubenswrapper[4907]: I0313 16:17:59.442679 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4cppb"] Mar 13 16:18:00 crc kubenswrapper[4907]: I0313 16:18:00.154332 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556978-m5zzz"] Mar 13 16:18:00 crc kubenswrapper[4907]: I0313 16:18:00.157192 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556978-m5zzz" Mar 13 16:18:00 crc kubenswrapper[4907]: I0313 16:18:00.160813 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 16:18:00 crc kubenswrapper[4907]: I0313 16:18:00.161167 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 16:18:00 crc kubenswrapper[4907]: I0313 16:18:00.161414 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 16:18:00 crc kubenswrapper[4907]: I0313 16:18:00.167205 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556978-m5zzz"] Mar 13 16:18:00 crc kubenswrapper[4907]: I0313 16:18:00.225477 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rhvqr\" (UniqueName: \"kubernetes.io/projected/b4a302e0-25a2-43d0-b5c0-5047d7a2c5a1-kube-api-access-rhvqr\") pod \"auto-csr-approver-29556978-m5zzz\" (UID: \"b4a302e0-25a2-43d0-b5c0-5047d7a2c5a1\") " pod="openshift-infra/auto-csr-approver-29556978-m5zzz" Mar 13 16:18:00 crc kubenswrapper[4907]: I0313 16:18:00.327835 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rhvqr\" (UniqueName: \"kubernetes.io/projected/b4a302e0-25a2-43d0-b5c0-5047d7a2c5a1-kube-api-access-rhvqr\") pod \"auto-csr-approver-29556978-m5zzz\" (UID: \"b4a302e0-25a2-43d0-b5c0-5047d7a2c5a1\") " pod="openshift-infra/auto-csr-approver-29556978-m5zzz" Mar 13 16:18:00 crc kubenswrapper[4907]: I0313 16:18:00.360212 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rhvqr\" (UniqueName: \"kubernetes.io/projected/b4a302e0-25a2-43d0-b5c0-5047d7a2c5a1-kube-api-access-rhvqr\") pod \"auto-csr-approver-29556978-m5zzz\" (UID: \"b4a302e0-25a2-43d0-b5c0-5047d7a2c5a1\") " pod="openshift-infra/auto-csr-approver-29556978-m5zzz" Mar 13 16:18:00 crc kubenswrapper[4907]: I0313 16:18:00.486707 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556978-m5zzz" Mar 13 16:18:00 crc kubenswrapper[4907]: I0313 16:18:00.977999 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556978-m5zzz"] Mar 13 16:18:01 crc kubenswrapper[4907]: I0313 16:18:01.203498 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556978-m5zzz" event={"ID":"b4a302e0-25a2-43d0-b5c0-5047d7a2c5a1","Type":"ContainerStarted","Data":"432517852f22b2935c9497d6a30eeb142c617505442bf2e407018010babc7523"} Mar 13 16:18:01 crc kubenswrapper[4907]: I0313 16:18:01.203655 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-4cppb" podUID="d417de61-9b72-4c15-bfae-6f1b6ee03849" containerName="registry-server" containerID="cri-o://5bcc136eb40ce02d1e8aec26066c465f5d2b04e99adae87f346f7d3227130ae6" gracePeriod=2 Mar 13 16:18:01 crc kubenswrapper[4907]: I0313 16:18:01.714804 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4cppb" Mar 13 16:18:01 crc kubenswrapper[4907]: I0313 16:18:01.885028 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d417de61-9b72-4c15-bfae-6f1b6ee03849-catalog-content\") pod \"d417de61-9b72-4c15-bfae-6f1b6ee03849\" (UID: \"d417de61-9b72-4c15-bfae-6f1b6ee03849\") " Mar 13 16:18:01 crc kubenswrapper[4907]: I0313 16:18:01.885114 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vf7bk\" (UniqueName: \"kubernetes.io/projected/d417de61-9b72-4c15-bfae-6f1b6ee03849-kube-api-access-vf7bk\") pod \"d417de61-9b72-4c15-bfae-6f1b6ee03849\" (UID: \"d417de61-9b72-4c15-bfae-6f1b6ee03849\") " Mar 13 16:18:01 crc kubenswrapper[4907]: I0313 16:18:01.885362 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d417de61-9b72-4c15-bfae-6f1b6ee03849-utilities\") pod \"d417de61-9b72-4c15-bfae-6f1b6ee03849\" (UID: \"d417de61-9b72-4c15-bfae-6f1b6ee03849\") " Mar 13 16:18:01 crc kubenswrapper[4907]: I0313 16:18:01.886963 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d417de61-9b72-4c15-bfae-6f1b6ee03849-utilities" (OuterVolumeSpecName: "utilities") pod "d417de61-9b72-4c15-bfae-6f1b6ee03849" (UID: "d417de61-9b72-4c15-bfae-6f1b6ee03849"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 16:18:01 crc kubenswrapper[4907]: I0313 16:18:01.892765 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d417de61-9b72-4c15-bfae-6f1b6ee03849-kube-api-access-vf7bk" (OuterVolumeSpecName: "kube-api-access-vf7bk") pod "d417de61-9b72-4c15-bfae-6f1b6ee03849" (UID: "d417de61-9b72-4c15-bfae-6f1b6ee03849"). InnerVolumeSpecName "kube-api-access-vf7bk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:18:01 crc kubenswrapper[4907]: I0313 16:18:01.926851 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d417de61-9b72-4c15-bfae-6f1b6ee03849-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d417de61-9b72-4c15-bfae-6f1b6ee03849" (UID: "d417de61-9b72-4c15-bfae-6f1b6ee03849"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 16:18:01 crc kubenswrapper[4907]: I0313 16:18:01.987628 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d417de61-9b72-4c15-bfae-6f1b6ee03849-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 16:18:01 crc kubenswrapper[4907]: I0313 16:18:01.987672 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d417de61-9b72-4c15-bfae-6f1b6ee03849-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 16:18:01 crc kubenswrapper[4907]: I0313 16:18:01.987683 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vf7bk\" (UniqueName: \"kubernetes.io/projected/d417de61-9b72-4c15-bfae-6f1b6ee03849-kube-api-access-vf7bk\") on node \"crc\" DevicePath \"\"" Mar 13 16:18:02 crc kubenswrapper[4907]: I0313 16:18:02.213665 4907 generic.go:334] "Generic (PLEG): container finished" podID="d417de61-9b72-4c15-bfae-6f1b6ee03849" containerID="5bcc136eb40ce02d1e8aec26066c465f5d2b04e99adae87f346f7d3227130ae6" exitCode=0 Mar 13 16:18:02 crc kubenswrapper[4907]: I0313 16:18:02.214030 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4cppb" event={"ID":"d417de61-9b72-4c15-bfae-6f1b6ee03849","Type":"ContainerDied","Data":"5bcc136eb40ce02d1e8aec26066c465f5d2b04e99adae87f346f7d3227130ae6"} Mar 13 16:18:02 crc kubenswrapper[4907]: I0313 16:18:02.214059 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4cppb" event={"ID":"d417de61-9b72-4c15-bfae-6f1b6ee03849","Type":"ContainerDied","Data":"5a312c55abb195fc381c69253b66e281eba78efb1b39b9a51a72abfc897857dc"} Mar 13 16:18:02 crc kubenswrapper[4907]: I0313 16:18:02.214077 4907 scope.go:117] "RemoveContainer" containerID="5bcc136eb40ce02d1e8aec26066c465f5d2b04e99adae87f346f7d3227130ae6" Mar 13 16:18:02 crc kubenswrapper[4907]: I0313 16:18:02.214219 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4cppb" Mar 13 16:18:02 crc kubenswrapper[4907]: I0313 16:18:02.242894 4907 scope.go:117] "RemoveContainer" containerID="cc8ae91cc5b08e53b842e2f7a69a2fbd79580167c23866df62a867359907d98f" Mar 13 16:18:02 crc kubenswrapper[4907]: I0313 16:18:02.256554 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4cppb"] Mar 13 16:18:02 crc kubenswrapper[4907]: I0313 16:18:02.277375 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-4cppb"] Mar 13 16:18:02 crc kubenswrapper[4907]: I0313 16:18:02.281476 4907 scope.go:117] "RemoveContainer" containerID="e3c0406c3eaa17bc2c3ae343da02515c95a6e17e38c2bc85e8e82bdf569d1c03" Mar 13 16:18:02 crc kubenswrapper[4907]: I0313 16:18:02.325168 4907 scope.go:117] "RemoveContainer" containerID="5bcc136eb40ce02d1e8aec26066c465f5d2b04e99adae87f346f7d3227130ae6" Mar 13 16:18:02 crc kubenswrapper[4907]: E0313 16:18:02.325838 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5bcc136eb40ce02d1e8aec26066c465f5d2b04e99adae87f346f7d3227130ae6\": container with ID starting with 5bcc136eb40ce02d1e8aec26066c465f5d2b04e99adae87f346f7d3227130ae6 not found: ID does not exist" containerID="5bcc136eb40ce02d1e8aec26066c465f5d2b04e99adae87f346f7d3227130ae6" Mar 13 16:18:02 crc kubenswrapper[4907]: I0313 16:18:02.325973 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5bcc136eb40ce02d1e8aec26066c465f5d2b04e99adae87f346f7d3227130ae6"} err="failed to get container status \"5bcc136eb40ce02d1e8aec26066c465f5d2b04e99adae87f346f7d3227130ae6\": rpc error: code = NotFound desc = could not find container \"5bcc136eb40ce02d1e8aec26066c465f5d2b04e99adae87f346f7d3227130ae6\": container with ID starting with 5bcc136eb40ce02d1e8aec26066c465f5d2b04e99adae87f346f7d3227130ae6 not found: ID does not exist" Mar 13 16:18:02 crc kubenswrapper[4907]: I0313 16:18:02.326018 4907 scope.go:117] "RemoveContainer" containerID="cc8ae91cc5b08e53b842e2f7a69a2fbd79580167c23866df62a867359907d98f" Mar 13 16:18:02 crc kubenswrapper[4907]: E0313 16:18:02.327030 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cc8ae91cc5b08e53b842e2f7a69a2fbd79580167c23866df62a867359907d98f\": container with ID starting with cc8ae91cc5b08e53b842e2f7a69a2fbd79580167c23866df62a867359907d98f not found: ID does not exist" containerID="cc8ae91cc5b08e53b842e2f7a69a2fbd79580167c23866df62a867359907d98f" Mar 13 16:18:02 crc kubenswrapper[4907]: I0313 16:18:02.327068 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cc8ae91cc5b08e53b842e2f7a69a2fbd79580167c23866df62a867359907d98f"} err="failed to get container status \"cc8ae91cc5b08e53b842e2f7a69a2fbd79580167c23866df62a867359907d98f\": rpc error: code = NotFound desc = could not find container \"cc8ae91cc5b08e53b842e2f7a69a2fbd79580167c23866df62a867359907d98f\": container with ID starting with cc8ae91cc5b08e53b842e2f7a69a2fbd79580167c23866df62a867359907d98f not found: ID does not exist" Mar 13 16:18:02 crc kubenswrapper[4907]: I0313 16:18:02.327089 4907 scope.go:117] "RemoveContainer" containerID="e3c0406c3eaa17bc2c3ae343da02515c95a6e17e38c2bc85e8e82bdf569d1c03" Mar 13 16:18:02 crc kubenswrapper[4907]: E0313 16:18:02.327331 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e3c0406c3eaa17bc2c3ae343da02515c95a6e17e38c2bc85e8e82bdf569d1c03\": container with ID starting with e3c0406c3eaa17bc2c3ae343da02515c95a6e17e38c2bc85e8e82bdf569d1c03 not found: ID does not exist" containerID="e3c0406c3eaa17bc2c3ae343da02515c95a6e17e38c2bc85e8e82bdf569d1c03" Mar 13 16:18:02 crc kubenswrapper[4907]: I0313 16:18:02.327362 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e3c0406c3eaa17bc2c3ae343da02515c95a6e17e38c2bc85e8e82bdf569d1c03"} err="failed to get container status \"e3c0406c3eaa17bc2c3ae343da02515c95a6e17e38c2bc85e8e82bdf569d1c03\": rpc error: code = NotFound desc = could not find container \"e3c0406c3eaa17bc2c3ae343da02515c95a6e17e38c2bc85e8e82bdf569d1c03\": container with ID starting with e3c0406c3eaa17bc2c3ae343da02515c95a6e17e38c2bc85e8e82bdf569d1c03 not found: ID does not exist" Mar 13 16:18:03 crc kubenswrapper[4907]: I0313 16:18:03.231461 4907 generic.go:334] "Generic (PLEG): container finished" podID="b4a302e0-25a2-43d0-b5c0-5047d7a2c5a1" containerID="f49981cfe3caeb1a9f80b1991ed24d10adb510db888dce88031b1e2cc8045b1b" exitCode=0 Mar 13 16:18:03 crc kubenswrapper[4907]: I0313 16:18:03.231765 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556978-m5zzz" event={"ID":"b4a302e0-25a2-43d0-b5c0-5047d7a2c5a1","Type":"ContainerDied","Data":"f49981cfe3caeb1a9f80b1991ed24d10adb510db888dce88031b1e2cc8045b1b"} Mar 13 16:18:03 crc kubenswrapper[4907]: I0313 16:18:03.798729 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d417de61-9b72-4c15-bfae-6f1b6ee03849" path="/var/lib/kubelet/pods/d417de61-9b72-4c15-bfae-6f1b6ee03849/volumes" Mar 13 16:18:04 crc kubenswrapper[4907]: I0313 16:18:04.694515 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556978-m5zzz" Mar 13 16:18:04 crc kubenswrapper[4907]: I0313 16:18:04.860686 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rhvqr\" (UniqueName: \"kubernetes.io/projected/b4a302e0-25a2-43d0-b5c0-5047d7a2c5a1-kube-api-access-rhvqr\") pod \"b4a302e0-25a2-43d0-b5c0-5047d7a2c5a1\" (UID: \"b4a302e0-25a2-43d0-b5c0-5047d7a2c5a1\") " Mar 13 16:18:04 crc kubenswrapper[4907]: I0313 16:18:04.867289 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b4a302e0-25a2-43d0-b5c0-5047d7a2c5a1-kube-api-access-rhvqr" (OuterVolumeSpecName: "kube-api-access-rhvqr") pod "b4a302e0-25a2-43d0-b5c0-5047d7a2c5a1" (UID: "b4a302e0-25a2-43d0-b5c0-5047d7a2c5a1"). InnerVolumeSpecName "kube-api-access-rhvqr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:18:04 crc kubenswrapper[4907]: I0313 16:18:04.963711 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rhvqr\" (UniqueName: \"kubernetes.io/projected/b4a302e0-25a2-43d0-b5c0-5047d7a2c5a1-kube-api-access-rhvqr\") on node \"crc\" DevicePath \"\"" Mar 13 16:18:05 crc kubenswrapper[4907]: I0313 16:18:05.255954 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556978-m5zzz" event={"ID":"b4a302e0-25a2-43d0-b5c0-5047d7a2c5a1","Type":"ContainerDied","Data":"432517852f22b2935c9497d6a30eeb142c617505442bf2e407018010babc7523"} Mar 13 16:18:05 crc kubenswrapper[4907]: I0313 16:18:05.255998 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="432517852f22b2935c9497d6a30eeb142c617505442bf2e407018010babc7523" Mar 13 16:18:05 crc kubenswrapper[4907]: I0313 16:18:05.256025 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556978-m5zzz" Mar 13 16:18:05 crc kubenswrapper[4907]: I0313 16:18:05.755397 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556972-llpxn"] Mar 13 16:18:05 crc kubenswrapper[4907]: I0313 16:18:05.763980 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556972-llpxn"] Mar 13 16:18:05 crc kubenswrapper[4907]: I0313 16:18:05.794331 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="85500c41-93c1-44ae-b418-d7f125c04621" path="/var/lib/kubelet/pods/85500c41-93c1-44ae-b418-d7f125c04621/volumes" Mar 13 16:18:07 crc kubenswrapper[4907]: I0313 16:18:07.785912 4907 scope.go:117] "RemoveContainer" containerID="300e429cce24fa4eb9aefff9494fe9905ee7b0767504365170f1f9645a3c8b5f" Mar 13 16:18:07 crc kubenswrapper[4907]: E0313 16:18:07.786578 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:18:21 crc kubenswrapper[4907]: I0313 16:18:21.792937 4907 scope.go:117] "RemoveContainer" containerID="300e429cce24fa4eb9aefff9494fe9905ee7b0767504365170f1f9645a3c8b5f" Mar 13 16:18:21 crc kubenswrapper[4907]: E0313 16:18:21.794101 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:18:32 crc kubenswrapper[4907]: I0313 16:18:32.782583 4907 scope.go:117] "RemoveContainer" containerID="300e429cce24fa4eb9aefff9494fe9905ee7b0767504365170f1f9645a3c8b5f" Mar 13 16:18:32 crc kubenswrapper[4907]: E0313 16:18:32.783395 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:18:42 crc kubenswrapper[4907]: I0313 16:18:42.035904 4907 scope.go:117] "RemoveContainer" containerID="1597ac303cc302238cff89d1286c78c824124bc23ca555a00a5dd03f41ac4104" Mar 13 16:18:43 crc kubenswrapper[4907]: I0313 16:18:43.784988 4907 scope.go:117] "RemoveContainer" containerID="300e429cce24fa4eb9aefff9494fe9905ee7b0767504365170f1f9645a3c8b5f" Mar 13 16:18:43 crc kubenswrapper[4907]: E0313 16:18:43.786003 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:18:55 crc kubenswrapper[4907]: I0313 16:18:55.784044 4907 scope.go:117] "RemoveContainer" containerID="300e429cce24fa4eb9aefff9494fe9905ee7b0767504365170f1f9645a3c8b5f" Mar 13 16:18:55 crc kubenswrapper[4907]: E0313 16:18:55.784876 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:19:08 crc kubenswrapper[4907]: I0313 16:19:08.782671 4907 scope.go:117] "RemoveContainer" containerID="300e429cce24fa4eb9aefff9494fe9905ee7b0767504365170f1f9645a3c8b5f" Mar 13 16:19:08 crc kubenswrapper[4907]: E0313 16:19:08.783476 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:19:22 crc kubenswrapper[4907]: I0313 16:19:22.783422 4907 scope.go:117] "RemoveContainer" containerID="300e429cce24fa4eb9aefff9494fe9905ee7b0767504365170f1f9645a3c8b5f" Mar 13 16:19:22 crc kubenswrapper[4907]: E0313 16:19:22.784404 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:19:36 crc kubenswrapper[4907]: I0313 16:19:36.783137 4907 scope.go:117] "RemoveContainer" containerID="300e429cce24fa4eb9aefff9494fe9905ee7b0767504365170f1f9645a3c8b5f" Mar 13 16:19:36 crc kubenswrapper[4907]: E0313 16:19:36.784158 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:19:50 crc kubenswrapper[4907]: I0313 16:19:50.782669 4907 scope.go:117] "RemoveContainer" containerID="300e429cce24fa4eb9aefff9494fe9905ee7b0767504365170f1f9645a3c8b5f" Mar 13 16:19:50 crc kubenswrapper[4907]: E0313 16:19:50.783596 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:20:00 crc kubenswrapper[4907]: I0313 16:20:00.156934 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556980-5b2q4"] Mar 13 16:20:00 crc kubenswrapper[4907]: E0313 16:20:00.157872 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d417de61-9b72-4c15-bfae-6f1b6ee03849" containerName="extract-utilities" Mar 13 16:20:00 crc kubenswrapper[4907]: I0313 16:20:00.157905 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="d417de61-9b72-4c15-bfae-6f1b6ee03849" containerName="extract-utilities" Mar 13 16:20:00 crc kubenswrapper[4907]: E0313 16:20:00.157913 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d417de61-9b72-4c15-bfae-6f1b6ee03849" containerName="registry-server" Mar 13 16:20:00 crc kubenswrapper[4907]: I0313 16:20:00.157921 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="d417de61-9b72-4c15-bfae-6f1b6ee03849" containerName="registry-server" Mar 13 16:20:00 crc kubenswrapper[4907]: E0313 16:20:00.157952 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d417de61-9b72-4c15-bfae-6f1b6ee03849" containerName="extract-content" Mar 13 16:20:00 crc kubenswrapper[4907]: I0313 16:20:00.157959 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="d417de61-9b72-4c15-bfae-6f1b6ee03849" containerName="extract-content" Mar 13 16:20:00 crc kubenswrapper[4907]: E0313 16:20:00.157967 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4a302e0-25a2-43d0-b5c0-5047d7a2c5a1" containerName="oc" Mar 13 16:20:00 crc kubenswrapper[4907]: I0313 16:20:00.157973 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4a302e0-25a2-43d0-b5c0-5047d7a2c5a1" containerName="oc" Mar 13 16:20:00 crc kubenswrapper[4907]: I0313 16:20:00.158221 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="d417de61-9b72-4c15-bfae-6f1b6ee03849" containerName="registry-server" Mar 13 16:20:00 crc kubenswrapper[4907]: I0313 16:20:00.158237 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="b4a302e0-25a2-43d0-b5c0-5047d7a2c5a1" containerName="oc" Mar 13 16:20:00 crc kubenswrapper[4907]: I0313 16:20:00.158979 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556980-5b2q4" Mar 13 16:20:00 crc kubenswrapper[4907]: I0313 16:20:00.162324 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 16:20:00 crc kubenswrapper[4907]: I0313 16:20:00.169270 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 16:20:00 crc kubenswrapper[4907]: I0313 16:20:00.170994 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556980-5b2q4"] Mar 13 16:20:00 crc kubenswrapper[4907]: I0313 16:20:00.172134 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 16:20:00 crc kubenswrapper[4907]: I0313 16:20:00.226148 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cg2d6\" (UniqueName: \"kubernetes.io/projected/0b9615a4-0270-4660-bd34-5efded647a80-kube-api-access-cg2d6\") pod \"auto-csr-approver-29556980-5b2q4\" (UID: \"0b9615a4-0270-4660-bd34-5efded647a80\") " pod="openshift-infra/auto-csr-approver-29556980-5b2q4" Mar 13 16:20:00 crc kubenswrapper[4907]: I0313 16:20:00.328704 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cg2d6\" (UniqueName: \"kubernetes.io/projected/0b9615a4-0270-4660-bd34-5efded647a80-kube-api-access-cg2d6\") pod \"auto-csr-approver-29556980-5b2q4\" (UID: \"0b9615a4-0270-4660-bd34-5efded647a80\") " pod="openshift-infra/auto-csr-approver-29556980-5b2q4" Mar 13 16:20:00 crc kubenswrapper[4907]: I0313 16:20:00.348644 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cg2d6\" (UniqueName: \"kubernetes.io/projected/0b9615a4-0270-4660-bd34-5efded647a80-kube-api-access-cg2d6\") pod \"auto-csr-approver-29556980-5b2q4\" (UID: \"0b9615a4-0270-4660-bd34-5efded647a80\") " pod="openshift-infra/auto-csr-approver-29556980-5b2q4" Mar 13 16:20:00 crc kubenswrapper[4907]: I0313 16:20:00.493090 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556980-5b2q4" Mar 13 16:20:01 crc kubenswrapper[4907]: W0313 16:20:01.045798 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0b9615a4_0270_4660_bd34_5efded647a80.slice/crio-1e0bbe7efbc6537e7c285a15ed6f9873236deb0bf8e950cf4296237e90a80601 WatchSource:0}: Error finding container 1e0bbe7efbc6537e7c285a15ed6f9873236deb0bf8e950cf4296237e90a80601: Status 404 returned error can't find the container with id 1e0bbe7efbc6537e7c285a15ed6f9873236deb0bf8e950cf4296237e90a80601 Mar 13 16:20:01 crc kubenswrapper[4907]: I0313 16:20:01.049257 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556980-5b2q4"] Mar 13 16:20:01 crc kubenswrapper[4907]: I0313 16:20:01.642504 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556980-5b2q4" event={"ID":"0b9615a4-0270-4660-bd34-5efded647a80","Type":"ContainerStarted","Data":"1e0bbe7efbc6537e7c285a15ed6f9873236deb0bf8e950cf4296237e90a80601"} Mar 13 16:20:03 crc kubenswrapper[4907]: I0313 16:20:03.674521 4907 generic.go:334] "Generic (PLEG): container finished" podID="0b9615a4-0270-4660-bd34-5efded647a80" containerID="d626b98830070c30af12d3ea7ba374e64df5c680e5d7d98055b77c7c0b8bf280" exitCode=0 Mar 13 16:20:03 crc kubenswrapper[4907]: I0313 16:20:03.674640 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556980-5b2q4" event={"ID":"0b9615a4-0270-4660-bd34-5efded647a80","Type":"ContainerDied","Data":"d626b98830070c30af12d3ea7ba374e64df5c680e5d7d98055b77c7c0b8bf280"} Mar 13 16:20:05 crc kubenswrapper[4907]: I0313 16:20:05.043240 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556980-5b2q4" Mar 13 16:20:05 crc kubenswrapper[4907]: I0313 16:20:05.149446 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cg2d6\" (UniqueName: \"kubernetes.io/projected/0b9615a4-0270-4660-bd34-5efded647a80-kube-api-access-cg2d6\") pod \"0b9615a4-0270-4660-bd34-5efded647a80\" (UID: \"0b9615a4-0270-4660-bd34-5efded647a80\") " Mar 13 16:20:05 crc kubenswrapper[4907]: I0313 16:20:05.156473 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b9615a4-0270-4660-bd34-5efded647a80-kube-api-access-cg2d6" (OuterVolumeSpecName: "kube-api-access-cg2d6") pod "0b9615a4-0270-4660-bd34-5efded647a80" (UID: "0b9615a4-0270-4660-bd34-5efded647a80"). InnerVolumeSpecName "kube-api-access-cg2d6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:20:05 crc kubenswrapper[4907]: I0313 16:20:05.253607 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cg2d6\" (UniqueName: \"kubernetes.io/projected/0b9615a4-0270-4660-bd34-5efded647a80-kube-api-access-cg2d6\") on node \"crc\" DevicePath \"\"" Mar 13 16:20:05 crc kubenswrapper[4907]: I0313 16:20:05.707641 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556980-5b2q4" event={"ID":"0b9615a4-0270-4660-bd34-5efded647a80","Type":"ContainerDied","Data":"1e0bbe7efbc6537e7c285a15ed6f9873236deb0bf8e950cf4296237e90a80601"} Mar 13 16:20:05 crc kubenswrapper[4907]: I0313 16:20:05.707710 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1e0bbe7efbc6537e7c285a15ed6f9873236deb0bf8e950cf4296237e90a80601" Mar 13 16:20:05 crc kubenswrapper[4907]: I0313 16:20:05.707801 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556980-5b2q4" Mar 13 16:20:05 crc kubenswrapper[4907]: I0313 16:20:05.785227 4907 scope.go:117] "RemoveContainer" containerID="300e429cce24fa4eb9aefff9494fe9905ee7b0767504365170f1f9645a3c8b5f" Mar 13 16:20:05 crc kubenswrapper[4907]: E0313 16:20:05.785636 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:20:06 crc kubenswrapper[4907]: I0313 16:20:06.135966 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556974-8sntm"] Mar 13 16:20:06 crc kubenswrapper[4907]: I0313 16:20:06.149596 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556974-8sntm"] Mar 13 16:20:07 crc kubenswrapper[4907]: I0313 16:20:07.561590 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-p5chz"] Mar 13 16:20:07 crc kubenswrapper[4907]: E0313 16:20:07.562469 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b9615a4-0270-4660-bd34-5efded647a80" containerName="oc" Mar 13 16:20:07 crc kubenswrapper[4907]: I0313 16:20:07.562487 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b9615a4-0270-4660-bd34-5efded647a80" containerName="oc" Mar 13 16:20:07 crc kubenswrapper[4907]: I0313 16:20:07.562786 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="0b9615a4-0270-4660-bd34-5efded647a80" containerName="oc" Mar 13 16:20:07 crc kubenswrapper[4907]: I0313 16:20:07.564872 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-p5chz" Mar 13 16:20:07 crc kubenswrapper[4907]: I0313 16:20:07.580785 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-p5chz"] Mar 13 16:20:07 crc kubenswrapper[4907]: I0313 16:20:07.719863 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e879dfd4-a757-496a-9255-3905537d4c2f-utilities\") pod \"redhat-operators-p5chz\" (UID: \"e879dfd4-a757-496a-9255-3905537d4c2f\") " pod="openshift-marketplace/redhat-operators-p5chz" Mar 13 16:20:07 crc kubenswrapper[4907]: I0313 16:20:07.720055 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e879dfd4-a757-496a-9255-3905537d4c2f-catalog-content\") pod \"redhat-operators-p5chz\" (UID: \"e879dfd4-a757-496a-9255-3905537d4c2f\") " pod="openshift-marketplace/redhat-operators-p5chz" Mar 13 16:20:07 crc kubenswrapper[4907]: I0313 16:20:07.720587 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-48v99\" (UniqueName: \"kubernetes.io/projected/e879dfd4-a757-496a-9255-3905537d4c2f-kube-api-access-48v99\") pod \"redhat-operators-p5chz\" (UID: \"e879dfd4-a757-496a-9255-3905537d4c2f\") " pod="openshift-marketplace/redhat-operators-p5chz" Mar 13 16:20:07 crc kubenswrapper[4907]: I0313 16:20:07.728598 4907 generic.go:334] "Generic (PLEG): container finished" podID="82c7c5cb-4f6b-4d38-a009-263025faa45b" containerID="9ec4cbc3bc8367d5cb8d52b834231c90815a53ebf59eb9499587563be45dd71a" exitCode=0 Mar 13 16:20:07 crc kubenswrapper[4907]: I0313 16:20:07.728640 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-openstack-openstack-cell1-bdwbh" event={"ID":"82c7c5cb-4f6b-4d38-a009-263025faa45b","Type":"ContainerDied","Data":"9ec4cbc3bc8367d5cb8d52b834231c90815a53ebf59eb9499587563be45dd71a"} Mar 13 16:20:07 crc kubenswrapper[4907]: I0313 16:20:07.795459 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5d0046fb-05a0-47a0-912c-a9495e2a0835" path="/var/lib/kubelet/pods/5d0046fb-05a0-47a0-912c-a9495e2a0835/volumes" Mar 13 16:20:07 crc kubenswrapper[4907]: I0313 16:20:07.822320 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e879dfd4-a757-496a-9255-3905537d4c2f-catalog-content\") pod \"redhat-operators-p5chz\" (UID: \"e879dfd4-a757-496a-9255-3905537d4c2f\") " pod="openshift-marketplace/redhat-operators-p5chz" Mar 13 16:20:07 crc kubenswrapper[4907]: I0313 16:20:07.822477 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-48v99\" (UniqueName: \"kubernetes.io/projected/e879dfd4-a757-496a-9255-3905537d4c2f-kube-api-access-48v99\") pod \"redhat-operators-p5chz\" (UID: \"e879dfd4-a757-496a-9255-3905537d4c2f\") " pod="openshift-marketplace/redhat-operators-p5chz" Mar 13 16:20:07 crc kubenswrapper[4907]: I0313 16:20:07.822513 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e879dfd4-a757-496a-9255-3905537d4c2f-utilities\") pod \"redhat-operators-p5chz\" (UID: \"e879dfd4-a757-496a-9255-3905537d4c2f\") " pod="openshift-marketplace/redhat-operators-p5chz" Mar 13 16:20:07 crc kubenswrapper[4907]: I0313 16:20:07.822984 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e879dfd4-a757-496a-9255-3905537d4c2f-catalog-content\") pod \"redhat-operators-p5chz\" (UID: \"e879dfd4-a757-496a-9255-3905537d4c2f\") " pod="openshift-marketplace/redhat-operators-p5chz" Mar 13 16:20:07 crc kubenswrapper[4907]: I0313 16:20:07.822999 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e879dfd4-a757-496a-9255-3905537d4c2f-utilities\") pod \"redhat-operators-p5chz\" (UID: \"e879dfd4-a757-496a-9255-3905537d4c2f\") " pod="openshift-marketplace/redhat-operators-p5chz" Mar 13 16:20:07 crc kubenswrapper[4907]: I0313 16:20:07.851284 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-48v99\" (UniqueName: \"kubernetes.io/projected/e879dfd4-a757-496a-9255-3905537d4c2f-kube-api-access-48v99\") pod \"redhat-operators-p5chz\" (UID: \"e879dfd4-a757-496a-9255-3905537d4c2f\") " pod="openshift-marketplace/redhat-operators-p5chz" Mar 13 16:20:07 crc kubenswrapper[4907]: I0313 16:20:07.898270 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-p5chz" Mar 13 16:20:08 crc kubenswrapper[4907]: I0313 16:20:08.394044 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-p5chz"] Mar 13 16:20:08 crc kubenswrapper[4907]: I0313 16:20:08.754807 4907 generic.go:334] "Generic (PLEG): container finished" podID="e879dfd4-a757-496a-9255-3905537d4c2f" containerID="b5e33e84c8e3054fb9999d59d99976f5ddce20795949dbe8c0de722094ed3f85" exitCode=0 Mar 13 16:20:08 crc kubenswrapper[4907]: I0313 16:20:08.755247 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p5chz" event={"ID":"e879dfd4-a757-496a-9255-3905537d4c2f","Type":"ContainerDied","Data":"b5e33e84c8e3054fb9999d59d99976f5ddce20795949dbe8c0de722094ed3f85"} Mar 13 16:20:08 crc kubenswrapper[4907]: I0313 16:20:08.755308 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p5chz" event={"ID":"e879dfd4-a757-496a-9255-3905537d4c2f","Type":"ContainerStarted","Data":"b255de30c32bfb6d5b4d9175c23c14378544689fd3ec3a2b3dc3a46fb608cce3"} Mar 13 16:20:09 crc kubenswrapper[4907]: I0313 16:20:09.197979 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-openstack-openstack-cell1-bdwbh" Mar 13 16:20:09 crc kubenswrapper[4907]: I0313 16:20:09.369142 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/82c7c5cb-4f6b-4d38-a009-263025faa45b-libvirt-secret-0\") pod \"82c7c5cb-4f6b-4d38-a009-263025faa45b\" (UID: \"82c7c5cb-4f6b-4d38-a009-263025faa45b\") " Mar 13 16:20:09 crc kubenswrapper[4907]: I0313 16:20:09.369278 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h7hx5\" (UniqueName: \"kubernetes.io/projected/82c7c5cb-4f6b-4d38-a009-263025faa45b-kube-api-access-h7hx5\") pod \"82c7c5cb-4f6b-4d38-a009-263025faa45b\" (UID: \"82c7c5cb-4f6b-4d38-a009-263025faa45b\") " Mar 13 16:20:09 crc kubenswrapper[4907]: I0313 16:20:09.369319 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/82c7c5cb-4f6b-4d38-a009-263025faa45b-inventory\") pod \"82c7c5cb-4f6b-4d38-a009-263025faa45b\" (UID: \"82c7c5cb-4f6b-4d38-a009-263025faa45b\") " Mar 13 16:20:09 crc kubenswrapper[4907]: I0313 16:20:09.369391 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/82c7c5cb-4f6b-4d38-a009-263025faa45b-ceph\") pod \"82c7c5cb-4f6b-4d38-a009-263025faa45b\" (UID: \"82c7c5cb-4f6b-4d38-a009-263025faa45b\") " Mar 13 16:20:09 crc kubenswrapper[4907]: I0313 16:20:09.369689 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/82c7c5cb-4f6b-4d38-a009-263025faa45b-ssh-key-openstack-cell1\") pod \"82c7c5cb-4f6b-4d38-a009-263025faa45b\" (UID: \"82c7c5cb-4f6b-4d38-a009-263025faa45b\") " Mar 13 16:20:09 crc kubenswrapper[4907]: I0313 16:20:09.370274 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82c7c5cb-4f6b-4d38-a009-263025faa45b-libvirt-combined-ca-bundle\") pod \"82c7c5cb-4f6b-4d38-a009-263025faa45b\" (UID: \"82c7c5cb-4f6b-4d38-a009-263025faa45b\") " Mar 13 16:20:09 crc kubenswrapper[4907]: I0313 16:20:09.374575 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82c7c5cb-4f6b-4d38-a009-263025faa45b-ceph" (OuterVolumeSpecName: "ceph") pod "82c7c5cb-4f6b-4d38-a009-263025faa45b" (UID: "82c7c5cb-4f6b-4d38-a009-263025faa45b"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:20:09 crc kubenswrapper[4907]: I0313 16:20:09.375652 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82c7c5cb-4f6b-4d38-a009-263025faa45b-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "82c7c5cb-4f6b-4d38-a009-263025faa45b" (UID: "82c7c5cb-4f6b-4d38-a009-263025faa45b"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:20:09 crc kubenswrapper[4907]: I0313 16:20:09.375753 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/82c7c5cb-4f6b-4d38-a009-263025faa45b-kube-api-access-h7hx5" (OuterVolumeSpecName: "kube-api-access-h7hx5") pod "82c7c5cb-4f6b-4d38-a009-263025faa45b" (UID: "82c7c5cb-4f6b-4d38-a009-263025faa45b"). InnerVolumeSpecName "kube-api-access-h7hx5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:20:09 crc kubenswrapper[4907]: I0313 16:20:09.398656 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82c7c5cb-4f6b-4d38-a009-263025faa45b-ssh-key-openstack-cell1" (OuterVolumeSpecName: "ssh-key-openstack-cell1") pod "82c7c5cb-4f6b-4d38-a009-263025faa45b" (UID: "82c7c5cb-4f6b-4d38-a009-263025faa45b"). InnerVolumeSpecName "ssh-key-openstack-cell1". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:20:09 crc kubenswrapper[4907]: I0313 16:20:09.399302 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82c7c5cb-4f6b-4d38-a009-263025faa45b-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "82c7c5cb-4f6b-4d38-a009-263025faa45b" (UID: "82c7c5cb-4f6b-4d38-a009-263025faa45b"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:20:09 crc kubenswrapper[4907]: I0313 16:20:09.400413 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82c7c5cb-4f6b-4d38-a009-263025faa45b-inventory" (OuterVolumeSpecName: "inventory") pod "82c7c5cb-4f6b-4d38-a009-263025faa45b" (UID: "82c7c5cb-4f6b-4d38-a009-263025faa45b"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:20:09 crc kubenswrapper[4907]: I0313 16:20:09.473134 4907 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/82c7c5cb-4f6b-4d38-a009-263025faa45b-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Mar 13 16:20:09 crc kubenswrapper[4907]: I0313 16:20:09.473174 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h7hx5\" (UniqueName: \"kubernetes.io/projected/82c7c5cb-4f6b-4d38-a009-263025faa45b-kube-api-access-h7hx5\") on node \"crc\" DevicePath \"\"" Mar 13 16:20:09 crc kubenswrapper[4907]: I0313 16:20:09.473188 4907 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/82c7c5cb-4f6b-4d38-a009-263025faa45b-inventory\") on node \"crc\" DevicePath \"\"" Mar 13 16:20:09 crc kubenswrapper[4907]: I0313 16:20:09.473197 4907 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/82c7c5cb-4f6b-4d38-a009-263025faa45b-ceph\") on node \"crc\" DevicePath \"\"" Mar 13 16:20:09 crc kubenswrapper[4907]: I0313 16:20:09.473206 4907 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/82c7c5cb-4f6b-4d38-a009-263025faa45b-ssh-key-openstack-cell1\") on node \"crc\" DevicePath \"\"" Mar 13 16:20:09 crc kubenswrapper[4907]: I0313 16:20:09.473215 4907 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82c7c5cb-4f6b-4d38-a009-263025faa45b-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 16:20:09 crc kubenswrapper[4907]: I0313 16:20:09.775260 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-openstack-openstack-cell1-bdwbh" event={"ID":"82c7c5cb-4f6b-4d38-a009-263025faa45b","Type":"ContainerDied","Data":"e4c7c890b0bd5338464cbbe531bda4c9b8efa360ef93c0753d3a22a75540214d"} Mar 13 16:20:09 crc kubenswrapper[4907]: I0313 16:20:09.775355 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e4c7c890b0bd5338464cbbe531bda4c9b8efa360ef93c0753d3a22a75540214d" Mar 13 16:20:09 crc kubenswrapper[4907]: I0313 16:20:09.775324 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-openstack-openstack-cell1-bdwbh" Mar 13 16:20:09 crc kubenswrapper[4907]: I0313 16:20:09.858782 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-openstack-openstack-cell1-2swxs"] Mar 13 16:20:09 crc kubenswrapper[4907]: E0313 16:20:09.859699 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82c7c5cb-4f6b-4d38-a009-263025faa45b" containerName="libvirt-openstack-openstack-cell1" Mar 13 16:20:09 crc kubenswrapper[4907]: I0313 16:20:09.859834 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="82c7c5cb-4f6b-4d38-a009-263025faa45b" containerName="libvirt-openstack-openstack-cell1" Mar 13 16:20:09 crc kubenswrapper[4907]: I0313 16:20:09.860408 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="82c7c5cb-4f6b-4d38-a009-263025faa45b" containerName="libvirt-openstack-openstack-cell1" Mar 13 16:20:09 crc kubenswrapper[4907]: I0313 16:20:09.862194 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-openstack-cell1-2swxs" Mar 13 16:20:09 crc kubenswrapper[4907]: I0313 16:20:09.864852 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Mar 13 16:20:09 crc kubenswrapper[4907]: I0313 16:20:09.865070 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Mar 13 16:20:09 crc kubenswrapper[4907]: I0313 16:20:09.865306 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 13 16:20:09 crc kubenswrapper[4907]: I0313 16:20:09.865597 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Mar 13 16:20:09 crc kubenswrapper[4907]: I0313 16:20:09.865815 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Mar 13 16:20:09 crc kubenswrapper[4907]: I0313 16:20:09.865647 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-cells-global-config" Mar 13 16:20:09 crc kubenswrapper[4907]: I0313 16:20:09.865662 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-44kxw" Mar 13 16:20:09 crc kubenswrapper[4907]: I0313 16:20:09.898664 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-openstack-openstack-cell1-2swxs"] Mar 13 16:20:09 crc kubenswrapper[4907]: I0313 16:20:09.982902 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/563c85ca-38c8-4853-a278-2a46fd0347a3-ceph\") pod \"nova-cell1-openstack-openstack-cell1-2swxs\" (UID: \"563c85ca-38c8-4853-a278-2a46fd0347a3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-2swxs" Mar 13 16:20:09 crc kubenswrapper[4907]: I0313 16:20:09.982968 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/563c85ca-38c8-4853-a278-2a46fd0347a3-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-openstack-cell1-2swxs\" (UID: \"563c85ca-38c8-4853-a278-2a46fd0347a3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-2swxs" Mar 13 16:20:09 crc kubenswrapper[4907]: I0313 16:20:09.983038 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/563c85ca-38c8-4853-a278-2a46fd0347a3-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-openstack-cell1-2swxs\" (UID: \"563c85ca-38c8-4853-a278-2a46fd0347a3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-2swxs" Mar 13 16:20:09 crc kubenswrapper[4907]: I0313 16:20:09.983079 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/563c85ca-38c8-4853-a278-2a46fd0347a3-ssh-key-openstack-cell1\") pod \"nova-cell1-openstack-openstack-cell1-2swxs\" (UID: \"563c85ca-38c8-4853-a278-2a46fd0347a3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-2swxs" Mar 13 16:20:09 crc kubenswrapper[4907]: I0313 16:20:09.983098 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/563c85ca-38c8-4853-a278-2a46fd0347a3-nova-cell1-compute-config-3\") pod \"nova-cell1-openstack-openstack-cell1-2swxs\" (UID: \"563c85ca-38c8-4853-a278-2a46fd0347a3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-2swxs" Mar 13 16:20:09 crc kubenswrapper[4907]: I0313 16:20:09.983130 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/563c85ca-38c8-4853-a278-2a46fd0347a3-nova-cell1-compute-config-2\") pod \"nova-cell1-openstack-openstack-cell1-2swxs\" (UID: \"563c85ca-38c8-4853-a278-2a46fd0347a3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-2swxs" Mar 13 16:20:09 crc kubenswrapper[4907]: I0313 16:20:09.983156 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d896k\" (UniqueName: \"kubernetes.io/projected/563c85ca-38c8-4853-a278-2a46fd0347a3-kube-api-access-d896k\") pod \"nova-cell1-openstack-openstack-cell1-2swxs\" (UID: \"563c85ca-38c8-4853-a278-2a46fd0347a3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-2swxs" Mar 13 16:20:09 crc kubenswrapper[4907]: I0313 16:20:09.983178 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/563c85ca-38c8-4853-a278-2a46fd0347a3-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-openstack-cell1-2swxs\" (UID: \"563c85ca-38c8-4853-a278-2a46fd0347a3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-2swxs" Mar 13 16:20:09 crc kubenswrapper[4907]: I0313 16:20:09.983210 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/563c85ca-38c8-4853-a278-2a46fd0347a3-nova-cells-global-config-0\") pod \"nova-cell1-openstack-openstack-cell1-2swxs\" (UID: \"563c85ca-38c8-4853-a278-2a46fd0347a3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-2swxs" Mar 13 16:20:09 crc kubenswrapper[4907]: I0313 16:20:09.983235 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/563c85ca-38c8-4853-a278-2a46fd0347a3-inventory\") pod \"nova-cell1-openstack-openstack-cell1-2swxs\" (UID: \"563c85ca-38c8-4853-a278-2a46fd0347a3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-2swxs" Mar 13 16:20:09 crc kubenswrapper[4907]: I0313 16:20:09.983282 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/563c85ca-38c8-4853-a278-2a46fd0347a3-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-openstack-cell1-2swxs\" (UID: \"563c85ca-38c8-4853-a278-2a46fd0347a3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-2swxs" Mar 13 16:20:09 crc kubenswrapper[4907]: I0313 16:20:09.983303 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/563c85ca-38c8-4853-a278-2a46fd0347a3-nova-cells-global-config-1\") pod \"nova-cell1-openstack-openstack-cell1-2swxs\" (UID: \"563c85ca-38c8-4853-a278-2a46fd0347a3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-2swxs" Mar 13 16:20:09 crc kubenswrapper[4907]: I0313 16:20:09.983351 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/563c85ca-38c8-4853-a278-2a46fd0347a3-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-openstack-cell1-2swxs\" (UID: \"563c85ca-38c8-4853-a278-2a46fd0347a3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-2swxs" Mar 13 16:20:10 crc kubenswrapper[4907]: I0313 16:20:10.084745 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/563c85ca-38c8-4853-a278-2a46fd0347a3-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-openstack-cell1-2swxs\" (UID: \"563c85ca-38c8-4853-a278-2a46fd0347a3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-2swxs" Mar 13 16:20:10 crc kubenswrapper[4907]: I0313 16:20:10.084795 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/563c85ca-38c8-4853-a278-2a46fd0347a3-ceph\") pod \"nova-cell1-openstack-openstack-cell1-2swxs\" (UID: \"563c85ca-38c8-4853-a278-2a46fd0347a3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-2swxs" Mar 13 16:20:10 crc kubenswrapper[4907]: I0313 16:20:10.084832 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/563c85ca-38c8-4853-a278-2a46fd0347a3-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-openstack-cell1-2swxs\" (UID: \"563c85ca-38c8-4853-a278-2a46fd0347a3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-2swxs" Mar 13 16:20:10 crc kubenswrapper[4907]: I0313 16:20:10.084866 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/563c85ca-38c8-4853-a278-2a46fd0347a3-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-openstack-cell1-2swxs\" (UID: \"563c85ca-38c8-4853-a278-2a46fd0347a3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-2swxs" Mar 13 16:20:10 crc kubenswrapper[4907]: I0313 16:20:10.084936 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/563c85ca-38c8-4853-a278-2a46fd0347a3-ssh-key-openstack-cell1\") pod \"nova-cell1-openstack-openstack-cell1-2swxs\" (UID: \"563c85ca-38c8-4853-a278-2a46fd0347a3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-2swxs" Mar 13 16:20:10 crc kubenswrapper[4907]: I0313 16:20:10.084960 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/563c85ca-38c8-4853-a278-2a46fd0347a3-nova-cell1-compute-config-3\") pod \"nova-cell1-openstack-openstack-cell1-2swxs\" (UID: \"563c85ca-38c8-4853-a278-2a46fd0347a3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-2swxs" Mar 13 16:20:10 crc kubenswrapper[4907]: I0313 16:20:10.084995 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/563c85ca-38c8-4853-a278-2a46fd0347a3-nova-cell1-compute-config-2\") pod \"nova-cell1-openstack-openstack-cell1-2swxs\" (UID: \"563c85ca-38c8-4853-a278-2a46fd0347a3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-2swxs" Mar 13 16:20:10 crc kubenswrapper[4907]: I0313 16:20:10.085024 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d896k\" (UniqueName: \"kubernetes.io/projected/563c85ca-38c8-4853-a278-2a46fd0347a3-kube-api-access-d896k\") pod \"nova-cell1-openstack-openstack-cell1-2swxs\" (UID: \"563c85ca-38c8-4853-a278-2a46fd0347a3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-2swxs" Mar 13 16:20:10 crc kubenswrapper[4907]: I0313 16:20:10.085063 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/563c85ca-38c8-4853-a278-2a46fd0347a3-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-openstack-cell1-2swxs\" (UID: \"563c85ca-38c8-4853-a278-2a46fd0347a3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-2swxs" Mar 13 16:20:10 crc kubenswrapper[4907]: I0313 16:20:10.085100 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/563c85ca-38c8-4853-a278-2a46fd0347a3-nova-cells-global-config-0\") pod \"nova-cell1-openstack-openstack-cell1-2swxs\" (UID: \"563c85ca-38c8-4853-a278-2a46fd0347a3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-2swxs" Mar 13 16:20:10 crc kubenswrapper[4907]: I0313 16:20:10.085123 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/563c85ca-38c8-4853-a278-2a46fd0347a3-inventory\") pod \"nova-cell1-openstack-openstack-cell1-2swxs\" (UID: \"563c85ca-38c8-4853-a278-2a46fd0347a3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-2swxs" Mar 13 16:20:10 crc kubenswrapper[4907]: I0313 16:20:10.085174 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/563c85ca-38c8-4853-a278-2a46fd0347a3-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-openstack-cell1-2swxs\" (UID: \"563c85ca-38c8-4853-a278-2a46fd0347a3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-2swxs" Mar 13 16:20:10 crc kubenswrapper[4907]: I0313 16:20:10.085195 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/563c85ca-38c8-4853-a278-2a46fd0347a3-nova-cells-global-config-1\") pod \"nova-cell1-openstack-openstack-cell1-2swxs\" (UID: \"563c85ca-38c8-4853-a278-2a46fd0347a3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-2swxs" Mar 13 16:20:10 crc kubenswrapper[4907]: I0313 16:20:10.086229 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/563c85ca-38c8-4853-a278-2a46fd0347a3-nova-cells-global-config-1\") pod \"nova-cell1-openstack-openstack-cell1-2swxs\" (UID: \"563c85ca-38c8-4853-a278-2a46fd0347a3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-2swxs" Mar 13 16:20:10 crc kubenswrapper[4907]: I0313 16:20:10.087667 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/563c85ca-38c8-4853-a278-2a46fd0347a3-nova-cells-global-config-0\") pod \"nova-cell1-openstack-openstack-cell1-2swxs\" (UID: \"563c85ca-38c8-4853-a278-2a46fd0347a3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-2swxs" Mar 13 16:20:10 crc kubenswrapper[4907]: I0313 16:20:10.089071 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/563c85ca-38c8-4853-a278-2a46fd0347a3-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-openstack-cell1-2swxs\" (UID: \"563c85ca-38c8-4853-a278-2a46fd0347a3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-2swxs" Mar 13 16:20:10 crc kubenswrapper[4907]: I0313 16:20:10.090543 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/563c85ca-38c8-4853-a278-2a46fd0347a3-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-openstack-cell1-2swxs\" (UID: \"563c85ca-38c8-4853-a278-2a46fd0347a3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-2swxs" Mar 13 16:20:10 crc kubenswrapper[4907]: I0313 16:20:10.090655 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/563c85ca-38c8-4853-a278-2a46fd0347a3-ssh-key-openstack-cell1\") pod \"nova-cell1-openstack-openstack-cell1-2swxs\" (UID: \"563c85ca-38c8-4853-a278-2a46fd0347a3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-2swxs" Mar 13 16:20:10 crc kubenswrapper[4907]: I0313 16:20:10.090742 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/563c85ca-38c8-4853-a278-2a46fd0347a3-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-openstack-cell1-2swxs\" (UID: \"563c85ca-38c8-4853-a278-2a46fd0347a3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-2swxs" Mar 13 16:20:10 crc kubenswrapper[4907]: I0313 16:20:10.091085 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/563c85ca-38c8-4853-a278-2a46fd0347a3-ceph\") pod \"nova-cell1-openstack-openstack-cell1-2swxs\" (UID: \"563c85ca-38c8-4853-a278-2a46fd0347a3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-2swxs" Mar 13 16:20:10 crc kubenswrapper[4907]: I0313 16:20:10.091479 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/563c85ca-38c8-4853-a278-2a46fd0347a3-nova-cell1-compute-config-3\") pod \"nova-cell1-openstack-openstack-cell1-2swxs\" (UID: \"563c85ca-38c8-4853-a278-2a46fd0347a3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-2swxs" Mar 13 16:20:10 crc kubenswrapper[4907]: I0313 16:20:10.091642 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/563c85ca-38c8-4853-a278-2a46fd0347a3-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-openstack-cell1-2swxs\" (UID: \"563c85ca-38c8-4853-a278-2a46fd0347a3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-2swxs" Mar 13 16:20:10 crc kubenswrapper[4907]: I0313 16:20:10.092128 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/563c85ca-38c8-4853-a278-2a46fd0347a3-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-openstack-cell1-2swxs\" (UID: \"563c85ca-38c8-4853-a278-2a46fd0347a3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-2swxs" Mar 13 16:20:10 crc kubenswrapper[4907]: I0313 16:20:10.093584 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/563c85ca-38c8-4853-a278-2a46fd0347a3-nova-cell1-compute-config-2\") pod \"nova-cell1-openstack-openstack-cell1-2swxs\" (UID: \"563c85ca-38c8-4853-a278-2a46fd0347a3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-2swxs" Mar 13 16:20:10 crc kubenswrapper[4907]: I0313 16:20:10.098176 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/563c85ca-38c8-4853-a278-2a46fd0347a3-inventory\") pod \"nova-cell1-openstack-openstack-cell1-2swxs\" (UID: \"563c85ca-38c8-4853-a278-2a46fd0347a3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-2swxs" Mar 13 16:20:10 crc kubenswrapper[4907]: I0313 16:20:10.113597 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d896k\" (UniqueName: \"kubernetes.io/projected/563c85ca-38c8-4853-a278-2a46fd0347a3-kube-api-access-d896k\") pod \"nova-cell1-openstack-openstack-cell1-2swxs\" (UID: \"563c85ca-38c8-4853-a278-2a46fd0347a3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-2swxs" Mar 13 16:20:10 crc kubenswrapper[4907]: I0313 16:20:10.183066 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-openstack-cell1-2swxs" Mar 13 16:20:10 crc kubenswrapper[4907]: I0313 16:20:10.718388 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-openstack-openstack-cell1-2swxs"] Mar 13 16:20:10 crc kubenswrapper[4907]: I0313 16:20:10.786548 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-openstack-cell1-2swxs" event={"ID":"563c85ca-38c8-4853-a278-2a46fd0347a3","Type":"ContainerStarted","Data":"f3963807c8d98e582dff2a1a7b85c9b8817c6b50cae040491cc0e2d933ebfd1c"} Mar 13 16:20:10 crc kubenswrapper[4907]: I0313 16:20:10.789305 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p5chz" event={"ID":"e879dfd4-a757-496a-9255-3905537d4c2f","Type":"ContainerStarted","Data":"35231ee78b021fd72e84af197de71532352f303bb83514e6ef94ec4d98cb8594"} Mar 13 16:20:11 crc kubenswrapper[4907]: I0313 16:20:11.802108 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-openstack-cell1-2swxs" event={"ID":"563c85ca-38c8-4853-a278-2a46fd0347a3","Type":"ContainerStarted","Data":"24ef5c64c1022adaf88de5eb0ff8963c4a40b5475aaa5f2ebc31c41febbdb0f6"} Mar 13 16:20:11 crc kubenswrapper[4907]: I0313 16:20:11.840419 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-openstack-openstack-cell1-2swxs" podStartSLOduration=2.216571227 podStartE2EDuration="2.84039971s" podCreationTimestamp="2026-03-13 16:20:09 +0000 UTC" firstStartedPulling="2026-03-13 16:20:10.72265324 +0000 UTC m=+8109.622440929" lastFinishedPulling="2026-03-13 16:20:11.346481723 +0000 UTC m=+8110.246269412" observedRunningTime="2026-03-13 16:20:11.83855202 +0000 UTC m=+8110.738339729" watchObservedRunningTime="2026-03-13 16:20:11.84039971 +0000 UTC m=+8110.740187409" Mar 13 16:20:14 crc kubenswrapper[4907]: I0313 16:20:14.835848 4907 generic.go:334] "Generic (PLEG): container finished" podID="e879dfd4-a757-496a-9255-3905537d4c2f" containerID="35231ee78b021fd72e84af197de71532352f303bb83514e6ef94ec4d98cb8594" exitCode=0 Mar 13 16:20:14 crc kubenswrapper[4907]: I0313 16:20:14.835928 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p5chz" event={"ID":"e879dfd4-a757-496a-9255-3905537d4c2f","Type":"ContainerDied","Data":"35231ee78b021fd72e84af197de71532352f303bb83514e6ef94ec4d98cb8594"} Mar 13 16:20:15 crc kubenswrapper[4907]: I0313 16:20:15.848586 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p5chz" event={"ID":"e879dfd4-a757-496a-9255-3905537d4c2f","Type":"ContainerStarted","Data":"77ec1446661bc5f367d029f3cae8705d509f9b1e3cacfc9bceeb4dc3c70ed20e"} Mar 13 16:20:15 crc kubenswrapper[4907]: I0313 16:20:15.869510 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-p5chz" podStartSLOduration=2.294996328 podStartE2EDuration="8.869494769s" podCreationTimestamp="2026-03-13 16:20:07 +0000 UTC" firstStartedPulling="2026-03-13 16:20:08.761747144 +0000 UTC m=+8107.661534833" lastFinishedPulling="2026-03-13 16:20:15.336245595 +0000 UTC m=+8114.236033274" observedRunningTime="2026-03-13 16:20:15.867219887 +0000 UTC m=+8114.767007576" watchObservedRunningTime="2026-03-13 16:20:15.869494769 +0000 UTC m=+8114.769282458" Mar 13 16:20:16 crc kubenswrapper[4907]: I0313 16:20:16.782029 4907 scope.go:117] "RemoveContainer" containerID="300e429cce24fa4eb9aefff9494fe9905ee7b0767504365170f1f9645a3c8b5f" Mar 13 16:20:16 crc kubenswrapper[4907]: E0313 16:20:16.782348 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:20:17 crc kubenswrapper[4907]: I0313 16:20:17.899528 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-p5chz" Mar 13 16:20:17 crc kubenswrapper[4907]: I0313 16:20:17.899988 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-p5chz" Mar 13 16:20:18 crc kubenswrapper[4907]: I0313 16:20:18.955588 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-p5chz" podUID="e879dfd4-a757-496a-9255-3905537d4c2f" containerName="registry-server" probeResult="failure" output=< Mar 13 16:20:18 crc kubenswrapper[4907]: timeout: failed to connect service ":50051" within 1s Mar 13 16:20:18 crc kubenswrapper[4907]: > Mar 13 16:20:27 crc kubenswrapper[4907]: I0313 16:20:27.965705 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-p5chz" Mar 13 16:20:28 crc kubenswrapper[4907]: I0313 16:20:28.041491 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-p5chz" Mar 13 16:20:28 crc kubenswrapper[4907]: I0313 16:20:28.216414 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-p5chz"] Mar 13 16:20:29 crc kubenswrapper[4907]: I0313 16:20:29.041271 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-p5chz" podUID="e879dfd4-a757-496a-9255-3905537d4c2f" containerName="registry-server" containerID="cri-o://77ec1446661bc5f367d029f3cae8705d509f9b1e3cacfc9bceeb4dc3c70ed20e" gracePeriod=2 Mar 13 16:20:29 crc kubenswrapper[4907]: I0313 16:20:29.567228 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-p5chz" Mar 13 16:20:29 crc kubenswrapper[4907]: I0313 16:20:29.641976 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-48v99\" (UniqueName: \"kubernetes.io/projected/e879dfd4-a757-496a-9255-3905537d4c2f-kube-api-access-48v99\") pod \"e879dfd4-a757-496a-9255-3905537d4c2f\" (UID: \"e879dfd4-a757-496a-9255-3905537d4c2f\") " Mar 13 16:20:29 crc kubenswrapper[4907]: I0313 16:20:29.642245 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e879dfd4-a757-496a-9255-3905537d4c2f-utilities\") pod \"e879dfd4-a757-496a-9255-3905537d4c2f\" (UID: \"e879dfd4-a757-496a-9255-3905537d4c2f\") " Mar 13 16:20:29 crc kubenswrapper[4907]: I0313 16:20:29.642519 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e879dfd4-a757-496a-9255-3905537d4c2f-catalog-content\") pod \"e879dfd4-a757-496a-9255-3905537d4c2f\" (UID: \"e879dfd4-a757-496a-9255-3905537d4c2f\") " Mar 13 16:20:29 crc kubenswrapper[4907]: I0313 16:20:29.643290 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e879dfd4-a757-496a-9255-3905537d4c2f-utilities" (OuterVolumeSpecName: "utilities") pod "e879dfd4-a757-496a-9255-3905537d4c2f" (UID: "e879dfd4-a757-496a-9255-3905537d4c2f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 16:20:29 crc kubenswrapper[4907]: I0313 16:20:29.647731 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e879dfd4-a757-496a-9255-3905537d4c2f-kube-api-access-48v99" (OuterVolumeSpecName: "kube-api-access-48v99") pod "e879dfd4-a757-496a-9255-3905537d4c2f" (UID: "e879dfd4-a757-496a-9255-3905537d4c2f"). InnerVolumeSpecName "kube-api-access-48v99". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:20:29 crc kubenswrapper[4907]: I0313 16:20:29.744768 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e879dfd4-a757-496a-9255-3905537d4c2f-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 16:20:29 crc kubenswrapper[4907]: I0313 16:20:29.744796 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-48v99\" (UniqueName: \"kubernetes.io/projected/e879dfd4-a757-496a-9255-3905537d4c2f-kube-api-access-48v99\") on node \"crc\" DevicePath \"\"" Mar 13 16:20:29 crc kubenswrapper[4907]: I0313 16:20:29.793132 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e879dfd4-a757-496a-9255-3905537d4c2f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e879dfd4-a757-496a-9255-3905537d4c2f" (UID: "e879dfd4-a757-496a-9255-3905537d4c2f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 16:20:29 crc kubenswrapper[4907]: I0313 16:20:29.848170 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e879dfd4-a757-496a-9255-3905537d4c2f-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 16:20:30 crc kubenswrapper[4907]: I0313 16:20:30.059305 4907 generic.go:334] "Generic (PLEG): container finished" podID="e879dfd4-a757-496a-9255-3905537d4c2f" containerID="77ec1446661bc5f367d029f3cae8705d509f9b1e3cacfc9bceeb4dc3c70ed20e" exitCode=0 Mar 13 16:20:30 crc kubenswrapper[4907]: I0313 16:20:30.059338 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-p5chz" Mar 13 16:20:30 crc kubenswrapper[4907]: I0313 16:20:30.059397 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p5chz" event={"ID":"e879dfd4-a757-496a-9255-3905537d4c2f","Type":"ContainerDied","Data":"77ec1446661bc5f367d029f3cae8705d509f9b1e3cacfc9bceeb4dc3c70ed20e"} Mar 13 16:20:30 crc kubenswrapper[4907]: I0313 16:20:30.059480 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p5chz" event={"ID":"e879dfd4-a757-496a-9255-3905537d4c2f","Type":"ContainerDied","Data":"b255de30c32bfb6d5b4d9175c23c14378544689fd3ec3a2b3dc3a46fb608cce3"} Mar 13 16:20:30 crc kubenswrapper[4907]: I0313 16:20:30.059516 4907 scope.go:117] "RemoveContainer" containerID="77ec1446661bc5f367d029f3cae8705d509f9b1e3cacfc9bceeb4dc3c70ed20e" Mar 13 16:20:30 crc kubenswrapper[4907]: I0313 16:20:30.090752 4907 scope.go:117] "RemoveContainer" containerID="35231ee78b021fd72e84af197de71532352f303bb83514e6ef94ec4d98cb8594" Mar 13 16:20:30 crc kubenswrapper[4907]: I0313 16:20:30.098783 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-p5chz"] Mar 13 16:20:30 crc kubenswrapper[4907]: I0313 16:20:30.112589 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-p5chz"] Mar 13 16:20:30 crc kubenswrapper[4907]: I0313 16:20:30.121671 4907 scope.go:117] "RemoveContainer" containerID="b5e33e84c8e3054fb9999d59d99976f5ddce20795949dbe8c0de722094ed3f85" Mar 13 16:20:30 crc kubenswrapper[4907]: I0313 16:20:30.190099 4907 scope.go:117] "RemoveContainer" containerID="77ec1446661bc5f367d029f3cae8705d509f9b1e3cacfc9bceeb4dc3c70ed20e" Mar 13 16:20:30 crc kubenswrapper[4907]: E0313 16:20:30.190653 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"77ec1446661bc5f367d029f3cae8705d509f9b1e3cacfc9bceeb4dc3c70ed20e\": container with ID starting with 77ec1446661bc5f367d029f3cae8705d509f9b1e3cacfc9bceeb4dc3c70ed20e not found: ID does not exist" containerID="77ec1446661bc5f367d029f3cae8705d509f9b1e3cacfc9bceeb4dc3c70ed20e" Mar 13 16:20:30 crc kubenswrapper[4907]: I0313 16:20:30.190729 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"77ec1446661bc5f367d029f3cae8705d509f9b1e3cacfc9bceeb4dc3c70ed20e"} err="failed to get container status \"77ec1446661bc5f367d029f3cae8705d509f9b1e3cacfc9bceeb4dc3c70ed20e\": rpc error: code = NotFound desc = could not find container \"77ec1446661bc5f367d029f3cae8705d509f9b1e3cacfc9bceeb4dc3c70ed20e\": container with ID starting with 77ec1446661bc5f367d029f3cae8705d509f9b1e3cacfc9bceeb4dc3c70ed20e not found: ID does not exist" Mar 13 16:20:30 crc kubenswrapper[4907]: I0313 16:20:30.190772 4907 scope.go:117] "RemoveContainer" containerID="35231ee78b021fd72e84af197de71532352f303bb83514e6ef94ec4d98cb8594" Mar 13 16:20:30 crc kubenswrapper[4907]: E0313 16:20:30.191368 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"35231ee78b021fd72e84af197de71532352f303bb83514e6ef94ec4d98cb8594\": container with ID starting with 35231ee78b021fd72e84af197de71532352f303bb83514e6ef94ec4d98cb8594 not found: ID does not exist" containerID="35231ee78b021fd72e84af197de71532352f303bb83514e6ef94ec4d98cb8594" Mar 13 16:20:30 crc kubenswrapper[4907]: I0313 16:20:30.191409 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"35231ee78b021fd72e84af197de71532352f303bb83514e6ef94ec4d98cb8594"} err="failed to get container status \"35231ee78b021fd72e84af197de71532352f303bb83514e6ef94ec4d98cb8594\": rpc error: code = NotFound desc = could not find container \"35231ee78b021fd72e84af197de71532352f303bb83514e6ef94ec4d98cb8594\": container with ID starting with 35231ee78b021fd72e84af197de71532352f303bb83514e6ef94ec4d98cb8594 not found: ID does not exist" Mar 13 16:20:30 crc kubenswrapper[4907]: I0313 16:20:30.191437 4907 scope.go:117] "RemoveContainer" containerID="b5e33e84c8e3054fb9999d59d99976f5ddce20795949dbe8c0de722094ed3f85" Mar 13 16:20:30 crc kubenswrapper[4907]: E0313 16:20:30.191762 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b5e33e84c8e3054fb9999d59d99976f5ddce20795949dbe8c0de722094ed3f85\": container with ID starting with b5e33e84c8e3054fb9999d59d99976f5ddce20795949dbe8c0de722094ed3f85 not found: ID does not exist" containerID="b5e33e84c8e3054fb9999d59d99976f5ddce20795949dbe8c0de722094ed3f85" Mar 13 16:20:30 crc kubenswrapper[4907]: I0313 16:20:30.191797 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b5e33e84c8e3054fb9999d59d99976f5ddce20795949dbe8c0de722094ed3f85"} err="failed to get container status \"b5e33e84c8e3054fb9999d59d99976f5ddce20795949dbe8c0de722094ed3f85\": rpc error: code = NotFound desc = could not find container \"b5e33e84c8e3054fb9999d59d99976f5ddce20795949dbe8c0de722094ed3f85\": container with ID starting with b5e33e84c8e3054fb9999d59d99976f5ddce20795949dbe8c0de722094ed3f85 not found: ID does not exist" Mar 13 16:20:31 crc kubenswrapper[4907]: I0313 16:20:31.791707 4907 scope.go:117] "RemoveContainer" containerID="300e429cce24fa4eb9aefff9494fe9905ee7b0767504365170f1f9645a3c8b5f" Mar 13 16:20:31 crc kubenswrapper[4907]: E0313 16:20:31.792114 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:20:31 crc kubenswrapper[4907]: I0313 16:20:31.798720 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e879dfd4-a757-496a-9255-3905537d4c2f" path="/var/lib/kubelet/pods/e879dfd4-a757-496a-9255-3905537d4c2f/volumes" Mar 13 16:20:42 crc kubenswrapper[4907]: I0313 16:20:42.181860 4907 scope.go:117] "RemoveContainer" containerID="048a0a595dbc6f4c8cb2226742f2beb5b639bec2395f93e0935301c145de8608" Mar 13 16:20:42 crc kubenswrapper[4907]: I0313 16:20:42.783315 4907 scope.go:117] "RemoveContainer" containerID="300e429cce24fa4eb9aefff9494fe9905ee7b0767504365170f1f9645a3c8b5f" Mar 13 16:20:42 crc kubenswrapper[4907]: E0313 16:20:42.784316 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:20:53 crc kubenswrapper[4907]: I0313 16:20:53.782648 4907 scope.go:117] "RemoveContainer" containerID="300e429cce24fa4eb9aefff9494fe9905ee7b0767504365170f1f9645a3c8b5f" Mar 13 16:20:54 crc kubenswrapper[4907]: I0313 16:20:54.325860 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerStarted","Data":"50d5a864941e0a2162babd980c37981cbcd4e9bfc7e1598617888cc1ec5d9b31"} Mar 13 16:21:57 crc kubenswrapper[4907]: I0313 16:21:57.671722 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/neutron-564986d4bf-zhh4c" podUID="3ce663fe-b525-44ea-a6d4-33f6b2366f46" containerName="neutron-api" probeResult="failure" output="HTTP probe failed with statuscode: 502" Mar 13 16:22:00 crc kubenswrapper[4907]: I0313 16:22:00.158063 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556982-t6wrd"] Mar 13 16:22:00 crc kubenswrapper[4907]: E0313 16:22:00.159130 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e879dfd4-a757-496a-9255-3905537d4c2f" containerName="registry-server" Mar 13 16:22:00 crc kubenswrapper[4907]: I0313 16:22:00.159146 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="e879dfd4-a757-496a-9255-3905537d4c2f" containerName="registry-server" Mar 13 16:22:00 crc kubenswrapper[4907]: E0313 16:22:00.159180 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e879dfd4-a757-496a-9255-3905537d4c2f" containerName="extract-utilities" Mar 13 16:22:00 crc kubenswrapper[4907]: I0313 16:22:00.159190 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="e879dfd4-a757-496a-9255-3905537d4c2f" containerName="extract-utilities" Mar 13 16:22:00 crc kubenswrapper[4907]: E0313 16:22:00.159225 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e879dfd4-a757-496a-9255-3905537d4c2f" containerName="extract-content" Mar 13 16:22:00 crc kubenswrapper[4907]: I0313 16:22:00.159234 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="e879dfd4-a757-496a-9255-3905537d4c2f" containerName="extract-content" Mar 13 16:22:00 crc kubenswrapper[4907]: I0313 16:22:00.159516 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="e879dfd4-a757-496a-9255-3905537d4c2f" containerName="registry-server" Mar 13 16:22:00 crc kubenswrapper[4907]: I0313 16:22:00.160428 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556982-t6wrd" Mar 13 16:22:00 crc kubenswrapper[4907]: I0313 16:22:00.162595 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 16:22:00 crc kubenswrapper[4907]: I0313 16:22:00.162819 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 16:22:00 crc kubenswrapper[4907]: I0313 16:22:00.163865 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 16:22:00 crc kubenswrapper[4907]: I0313 16:22:00.167471 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556982-t6wrd"] Mar 13 16:22:00 crc kubenswrapper[4907]: I0313 16:22:00.264348 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mxsfw\" (UniqueName: \"kubernetes.io/projected/2364a186-f894-44a5-9c69-101cbfa14a82-kube-api-access-mxsfw\") pod \"auto-csr-approver-29556982-t6wrd\" (UID: \"2364a186-f894-44a5-9c69-101cbfa14a82\") " pod="openshift-infra/auto-csr-approver-29556982-t6wrd" Mar 13 16:22:00 crc kubenswrapper[4907]: I0313 16:22:00.366114 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mxsfw\" (UniqueName: \"kubernetes.io/projected/2364a186-f894-44a5-9c69-101cbfa14a82-kube-api-access-mxsfw\") pod \"auto-csr-approver-29556982-t6wrd\" (UID: \"2364a186-f894-44a5-9c69-101cbfa14a82\") " pod="openshift-infra/auto-csr-approver-29556982-t6wrd" Mar 13 16:22:00 crc kubenswrapper[4907]: I0313 16:22:00.390584 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mxsfw\" (UniqueName: \"kubernetes.io/projected/2364a186-f894-44a5-9c69-101cbfa14a82-kube-api-access-mxsfw\") pod \"auto-csr-approver-29556982-t6wrd\" (UID: \"2364a186-f894-44a5-9c69-101cbfa14a82\") " pod="openshift-infra/auto-csr-approver-29556982-t6wrd" Mar 13 16:22:00 crc kubenswrapper[4907]: I0313 16:22:00.492787 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556982-t6wrd" Mar 13 16:22:00 crc kubenswrapper[4907]: I0313 16:22:00.970084 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556982-t6wrd"] Mar 13 16:22:00 crc kubenswrapper[4907]: W0313 16:22:00.981851 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2364a186_f894_44a5_9c69_101cbfa14a82.slice/crio-86827b0005e293ad98b929277b369ca063a389d985875e7d08c7e28a85e0163a WatchSource:0}: Error finding container 86827b0005e293ad98b929277b369ca063a389d985875e7d08c7e28a85e0163a: Status 404 returned error can't find the container with id 86827b0005e293ad98b929277b369ca063a389d985875e7d08c7e28a85e0163a Mar 13 16:22:01 crc kubenswrapper[4907]: I0313 16:22:01.073674 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556982-t6wrd" event={"ID":"2364a186-f894-44a5-9c69-101cbfa14a82","Type":"ContainerStarted","Data":"86827b0005e293ad98b929277b369ca063a389d985875e7d08c7e28a85e0163a"} Mar 13 16:22:03 crc kubenswrapper[4907]: I0313 16:22:03.099076 4907 generic.go:334] "Generic (PLEG): container finished" podID="2364a186-f894-44a5-9c69-101cbfa14a82" containerID="48a114e32689f3100e68121daa48f2b4ba681c9f23fd2dfae93832f7ab1cf591" exitCode=0 Mar 13 16:22:03 crc kubenswrapper[4907]: I0313 16:22:03.099599 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556982-t6wrd" event={"ID":"2364a186-f894-44a5-9c69-101cbfa14a82","Type":"ContainerDied","Data":"48a114e32689f3100e68121daa48f2b4ba681c9f23fd2dfae93832f7ab1cf591"} Mar 13 16:22:04 crc kubenswrapper[4907]: I0313 16:22:04.596297 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556982-t6wrd" Mar 13 16:22:04 crc kubenswrapper[4907]: I0313 16:22:04.669310 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mxsfw\" (UniqueName: \"kubernetes.io/projected/2364a186-f894-44a5-9c69-101cbfa14a82-kube-api-access-mxsfw\") pod \"2364a186-f894-44a5-9c69-101cbfa14a82\" (UID: \"2364a186-f894-44a5-9c69-101cbfa14a82\") " Mar 13 16:22:04 crc kubenswrapper[4907]: I0313 16:22:04.675784 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2364a186-f894-44a5-9c69-101cbfa14a82-kube-api-access-mxsfw" (OuterVolumeSpecName: "kube-api-access-mxsfw") pod "2364a186-f894-44a5-9c69-101cbfa14a82" (UID: "2364a186-f894-44a5-9c69-101cbfa14a82"). InnerVolumeSpecName "kube-api-access-mxsfw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:22:04 crc kubenswrapper[4907]: I0313 16:22:04.772460 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mxsfw\" (UniqueName: \"kubernetes.io/projected/2364a186-f894-44a5-9c69-101cbfa14a82-kube-api-access-mxsfw\") on node \"crc\" DevicePath \"\"" Mar 13 16:22:05 crc kubenswrapper[4907]: I0313 16:22:05.124193 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556982-t6wrd" event={"ID":"2364a186-f894-44a5-9c69-101cbfa14a82","Type":"ContainerDied","Data":"86827b0005e293ad98b929277b369ca063a389d985875e7d08c7e28a85e0163a"} Mar 13 16:22:05 crc kubenswrapper[4907]: I0313 16:22:05.124221 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556982-t6wrd" Mar 13 16:22:05 crc kubenswrapper[4907]: I0313 16:22:05.124238 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="86827b0005e293ad98b929277b369ca063a389d985875e7d08c7e28a85e0163a" Mar 13 16:22:05 crc kubenswrapper[4907]: I0313 16:22:05.673237 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556976-twcx4"] Mar 13 16:22:05 crc kubenswrapper[4907]: I0313 16:22:05.680979 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556976-twcx4"] Mar 13 16:22:05 crc kubenswrapper[4907]: I0313 16:22:05.803775 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="362f626f-a4dd-45a7-ad88-4c539732eebb" path="/var/lib/kubelet/pods/362f626f-a4dd-45a7-ad88-4c539732eebb/volumes" Mar 13 16:22:42 crc kubenswrapper[4907]: I0313 16:22:42.329462 4907 scope.go:117] "RemoveContainer" containerID="7e0af7bb09c83c72fd3ad1550fb6bc5bec78d9ec9c71037ff20efe437fc967c9" Mar 13 16:22:50 crc kubenswrapper[4907]: I0313 16:22:50.607343 4907 generic.go:334] "Generic (PLEG): container finished" podID="563c85ca-38c8-4853-a278-2a46fd0347a3" containerID="24ef5c64c1022adaf88de5eb0ff8963c4a40b5475aaa5f2ebc31c41febbdb0f6" exitCode=0 Mar 13 16:22:50 crc kubenswrapper[4907]: I0313 16:22:50.607421 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-openstack-cell1-2swxs" event={"ID":"563c85ca-38c8-4853-a278-2a46fd0347a3","Type":"ContainerDied","Data":"24ef5c64c1022adaf88de5eb0ff8963c4a40b5475aaa5f2ebc31c41febbdb0f6"} Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.198814 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-openstack-cell1-2swxs" Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.331072 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/563c85ca-38c8-4853-a278-2a46fd0347a3-inventory\") pod \"563c85ca-38c8-4853-a278-2a46fd0347a3\" (UID: \"563c85ca-38c8-4853-a278-2a46fd0347a3\") " Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.331132 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/563c85ca-38c8-4853-a278-2a46fd0347a3-nova-cell1-compute-config-2\") pod \"563c85ca-38c8-4853-a278-2a46fd0347a3\" (UID: \"563c85ca-38c8-4853-a278-2a46fd0347a3\") " Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.331217 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/563c85ca-38c8-4853-a278-2a46fd0347a3-nova-cell1-compute-config-1\") pod \"563c85ca-38c8-4853-a278-2a46fd0347a3\" (UID: \"563c85ca-38c8-4853-a278-2a46fd0347a3\") " Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.331249 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/563c85ca-38c8-4853-a278-2a46fd0347a3-nova-cells-global-config-0\") pod \"563c85ca-38c8-4853-a278-2a46fd0347a3\" (UID: \"563c85ca-38c8-4853-a278-2a46fd0347a3\") " Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.331322 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/563c85ca-38c8-4853-a278-2a46fd0347a3-ceph\") pod \"563c85ca-38c8-4853-a278-2a46fd0347a3\" (UID: \"563c85ca-38c8-4853-a278-2a46fd0347a3\") " Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.331344 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/563c85ca-38c8-4853-a278-2a46fd0347a3-nova-cell1-combined-ca-bundle\") pod \"563c85ca-38c8-4853-a278-2a46fd0347a3\" (UID: \"563c85ca-38c8-4853-a278-2a46fd0347a3\") " Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.331365 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/563c85ca-38c8-4853-a278-2a46fd0347a3-nova-cells-global-config-1\") pod \"563c85ca-38c8-4853-a278-2a46fd0347a3\" (UID: \"563c85ca-38c8-4853-a278-2a46fd0347a3\") " Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.331382 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/563c85ca-38c8-4853-a278-2a46fd0347a3-ssh-key-openstack-cell1\") pod \"563c85ca-38c8-4853-a278-2a46fd0347a3\" (UID: \"563c85ca-38c8-4853-a278-2a46fd0347a3\") " Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.331420 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/563c85ca-38c8-4853-a278-2a46fd0347a3-nova-migration-ssh-key-1\") pod \"563c85ca-38c8-4853-a278-2a46fd0347a3\" (UID: \"563c85ca-38c8-4853-a278-2a46fd0347a3\") " Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.331470 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/563c85ca-38c8-4853-a278-2a46fd0347a3-nova-migration-ssh-key-0\") pod \"563c85ca-38c8-4853-a278-2a46fd0347a3\" (UID: \"563c85ca-38c8-4853-a278-2a46fd0347a3\") " Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.331502 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d896k\" (UniqueName: \"kubernetes.io/projected/563c85ca-38c8-4853-a278-2a46fd0347a3-kube-api-access-d896k\") pod \"563c85ca-38c8-4853-a278-2a46fd0347a3\" (UID: \"563c85ca-38c8-4853-a278-2a46fd0347a3\") " Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.331554 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/563c85ca-38c8-4853-a278-2a46fd0347a3-nova-cell1-compute-config-0\") pod \"563c85ca-38c8-4853-a278-2a46fd0347a3\" (UID: \"563c85ca-38c8-4853-a278-2a46fd0347a3\") " Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.331579 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/563c85ca-38c8-4853-a278-2a46fd0347a3-nova-cell1-compute-config-3\") pod \"563c85ca-38c8-4853-a278-2a46fd0347a3\" (UID: \"563c85ca-38c8-4853-a278-2a46fd0347a3\") " Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.341645 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/563c85ca-38c8-4853-a278-2a46fd0347a3-ceph" (OuterVolumeSpecName: "ceph") pod "563c85ca-38c8-4853-a278-2a46fd0347a3" (UID: "563c85ca-38c8-4853-a278-2a46fd0347a3"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.346615 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/563c85ca-38c8-4853-a278-2a46fd0347a3-nova-cell1-combined-ca-bundle" (OuterVolumeSpecName: "nova-cell1-combined-ca-bundle") pod "563c85ca-38c8-4853-a278-2a46fd0347a3" (UID: "563c85ca-38c8-4853-a278-2a46fd0347a3"). InnerVolumeSpecName "nova-cell1-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.347480 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/563c85ca-38c8-4853-a278-2a46fd0347a3-kube-api-access-d896k" (OuterVolumeSpecName: "kube-api-access-d896k") pod "563c85ca-38c8-4853-a278-2a46fd0347a3" (UID: "563c85ca-38c8-4853-a278-2a46fd0347a3"). InnerVolumeSpecName "kube-api-access-d896k". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.367673 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/563c85ca-38c8-4853-a278-2a46fd0347a3-nova-cells-global-config-1" (OuterVolumeSpecName: "nova-cells-global-config-1") pod "563c85ca-38c8-4853-a278-2a46fd0347a3" (UID: "563c85ca-38c8-4853-a278-2a46fd0347a3"). InnerVolumeSpecName "nova-cells-global-config-1". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.367829 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/563c85ca-38c8-4853-a278-2a46fd0347a3-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "563c85ca-38c8-4853-a278-2a46fd0347a3" (UID: "563c85ca-38c8-4853-a278-2a46fd0347a3"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.371926 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/563c85ca-38c8-4853-a278-2a46fd0347a3-nova-cells-global-config-0" (OuterVolumeSpecName: "nova-cells-global-config-0") pod "563c85ca-38c8-4853-a278-2a46fd0347a3" (UID: "563c85ca-38c8-4853-a278-2a46fd0347a3"). InnerVolumeSpecName "nova-cells-global-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.372334 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/563c85ca-38c8-4853-a278-2a46fd0347a3-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "563c85ca-38c8-4853-a278-2a46fd0347a3" (UID: "563c85ca-38c8-4853-a278-2a46fd0347a3"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.376262 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/563c85ca-38c8-4853-a278-2a46fd0347a3-nova-cell1-compute-config-2" (OuterVolumeSpecName: "nova-cell1-compute-config-2") pod "563c85ca-38c8-4853-a278-2a46fd0347a3" (UID: "563c85ca-38c8-4853-a278-2a46fd0347a3"). InnerVolumeSpecName "nova-cell1-compute-config-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.385461 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/563c85ca-38c8-4853-a278-2a46fd0347a3-nova-cell1-compute-config-3" (OuterVolumeSpecName: "nova-cell1-compute-config-3") pod "563c85ca-38c8-4853-a278-2a46fd0347a3" (UID: "563c85ca-38c8-4853-a278-2a46fd0347a3"). InnerVolumeSpecName "nova-cell1-compute-config-3". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.386462 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/563c85ca-38c8-4853-a278-2a46fd0347a3-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "563c85ca-38c8-4853-a278-2a46fd0347a3" (UID: "563c85ca-38c8-4853-a278-2a46fd0347a3"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.406088 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/563c85ca-38c8-4853-a278-2a46fd0347a3-inventory" (OuterVolumeSpecName: "inventory") pod "563c85ca-38c8-4853-a278-2a46fd0347a3" (UID: "563c85ca-38c8-4853-a278-2a46fd0347a3"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.407956 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/563c85ca-38c8-4853-a278-2a46fd0347a3-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "563c85ca-38c8-4853-a278-2a46fd0347a3" (UID: "563c85ca-38c8-4853-a278-2a46fd0347a3"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.418747 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/563c85ca-38c8-4853-a278-2a46fd0347a3-ssh-key-openstack-cell1" (OuterVolumeSpecName: "ssh-key-openstack-cell1") pod "563c85ca-38c8-4853-a278-2a46fd0347a3" (UID: "563c85ca-38c8-4853-a278-2a46fd0347a3"). InnerVolumeSpecName "ssh-key-openstack-cell1". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.439808 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d896k\" (UniqueName: \"kubernetes.io/projected/563c85ca-38c8-4853-a278-2a46fd0347a3-kube-api-access-d896k\") on node \"crc\" DevicePath \"\"" Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.439853 4907 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/563c85ca-38c8-4853-a278-2a46fd0347a3-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.439872 4907 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/563c85ca-38c8-4853-a278-2a46fd0347a3-nova-cell1-compute-config-3\") on node \"crc\" DevicePath \"\"" Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.440173 4907 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/563c85ca-38c8-4853-a278-2a46fd0347a3-inventory\") on node \"crc\" DevicePath \"\"" Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.440203 4907 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/563c85ca-38c8-4853-a278-2a46fd0347a3-nova-cell1-compute-config-2\") on node \"crc\" DevicePath \"\"" Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.440213 4907 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/563c85ca-38c8-4853-a278-2a46fd0347a3-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.440225 4907 reconciler_common.go:293] "Volume detached for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/563c85ca-38c8-4853-a278-2a46fd0347a3-nova-cells-global-config-0\") on node \"crc\" DevicePath \"\"" Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.440237 4907 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/563c85ca-38c8-4853-a278-2a46fd0347a3-ceph\") on node \"crc\" DevicePath \"\"" Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.440250 4907 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/563c85ca-38c8-4853-a278-2a46fd0347a3-nova-cell1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.440626 4907 reconciler_common.go:293] "Volume detached for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/563c85ca-38c8-4853-a278-2a46fd0347a3-nova-cells-global-config-1\") on node \"crc\" DevicePath \"\"" Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.440785 4907 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/563c85ca-38c8-4853-a278-2a46fd0347a3-ssh-key-openstack-cell1\") on node \"crc\" DevicePath \"\"" Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.440809 4907 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/563c85ca-38c8-4853-a278-2a46fd0347a3-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.440822 4907 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/563c85ca-38c8-4853-a278-2a46fd0347a3-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.635642 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-openstack-cell1-2swxs" event={"ID":"563c85ca-38c8-4853-a278-2a46fd0347a3","Type":"ContainerDied","Data":"f3963807c8d98e582dff2a1a7b85c9b8817c6b50cae040491cc0e2d933ebfd1c"} Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.635699 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f3963807c8d98e582dff2a1a7b85c9b8817c6b50cae040491cc0e2d933ebfd1c" Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.635838 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-openstack-cell1-2swxs" Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.813341 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-openstack-openstack-cell1-gmnh8"] Mar 13 16:22:52 crc kubenswrapper[4907]: E0313 16:22:52.813950 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="563c85ca-38c8-4853-a278-2a46fd0347a3" containerName="nova-cell1-openstack-openstack-cell1" Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.813974 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="563c85ca-38c8-4853-a278-2a46fd0347a3" containerName="nova-cell1-openstack-openstack-cell1" Mar 13 16:22:52 crc kubenswrapper[4907]: E0313 16:22:52.814010 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2364a186-f894-44a5-9c69-101cbfa14a82" containerName="oc" Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.814018 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="2364a186-f894-44a5-9c69-101cbfa14a82" containerName="oc" Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.814284 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="563c85ca-38c8-4853-a278-2a46fd0347a3" containerName="nova-cell1-openstack-openstack-cell1" Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.814331 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="2364a186-f894-44a5-9c69-101cbfa14a82" containerName="oc" Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.815929 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-openstack-openstack-cell1-gmnh8" Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.837130 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-openstack-openstack-cell1-gmnh8"] Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.872381 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.872439 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.872613 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.872447 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-44kxw" Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.873743 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.977379 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/b5718b26-a6b1-4b64-812b-a85da3d6f0a4-ssh-key-openstack-cell1\") pod \"telemetry-openstack-openstack-cell1-gmnh8\" (UID: \"b5718b26-a6b1-4b64-812b-a85da3d6f0a4\") " pod="openstack/telemetry-openstack-openstack-cell1-gmnh8" Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.977712 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/b5718b26-a6b1-4b64-812b-a85da3d6f0a4-ceilometer-compute-config-data-0\") pod \"telemetry-openstack-openstack-cell1-gmnh8\" (UID: \"b5718b26-a6b1-4b64-812b-a85da3d6f0a4\") " pod="openstack/telemetry-openstack-openstack-cell1-gmnh8" Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.977800 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/b5718b26-a6b1-4b64-812b-a85da3d6f0a4-ceilometer-compute-config-data-1\") pod \"telemetry-openstack-openstack-cell1-gmnh8\" (UID: \"b5718b26-a6b1-4b64-812b-a85da3d6f0a4\") " pod="openstack/telemetry-openstack-openstack-cell1-gmnh8" Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.977823 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5718b26-a6b1-4b64-812b-a85da3d6f0a4-telemetry-combined-ca-bundle\") pod \"telemetry-openstack-openstack-cell1-gmnh8\" (UID: \"b5718b26-a6b1-4b64-812b-a85da3d6f0a4\") " pod="openstack/telemetry-openstack-openstack-cell1-gmnh8" Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.977864 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b5718b26-a6b1-4b64-812b-a85da3d6f0a4-ceph\") pod \"telemetry-openstack-openstack-cell1-gmnh8\" (UID: \"b5718b26-a6b1-4b64-812b-a85da3d6f0a4\") " pod="openstack/telemetry-openstack-openstack-cell1-gmnh8" Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.977907 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b5718b26-a6b1-4b64-812b-a85da3d6f0a4-inventory\") pod \"telemetry-openstack-openstack-cell1-gmnh8\" (UID: \"b5718b26-a6b1-4b64-812b-a85da3d6f0a4\") " pod="openstack/telemetry-openstack-openstack-cell1-gmnh8" Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.977929 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zkd9z\" (UniqueName: \"kubernetes.io/projected/b5718b26-a6b1-4b64-812b-a85da3d6f0a4-kube-api-access-zkd9z\") pod \"telemetry-openstack-openstack-cell1-gmnh8\" (UID: \"b5718b26-a6b1-4b64-812b-a85da3d6f0a4\") " pod="openstack/telemetry-openstack-openstack-cell1-gmnh8" Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.977971 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/b5718b26-a6b1-4b64-812b-a85da3d6f0a4-ceilometer-compute-config-data-2\") pod \"telemetry-openstack-openstack-cell1-gmnh8\" (UID: \"b5718b26-a6b1-4b64-812b-a85da3d6f0a4\") " pod="openstack/telemetry-openstack-openstack-cell1-gmnh8" Mar 13 16:22:53 crc kubenswrapper[4907]: I0313 16:22:53.083064 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/b5718b26-a6b1-4b64-812b-a85da3d6f0a4-ceilometer-compute-config-data-1\") pod \"telemetry-openstack-openstack-cell1-gmnh8\" (UID: \"b5718b26-a6b1-4b64-812b-a85da3d6f0a4\") " pod="openstack/telemetry-openstack-openstack-cell1-gmnh8" Mar 13 16:22:53 crc kubenswrapper[4907]: I0313 16:22:53.083345 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5718b26-a6b1-4b64-812b-a85da3d6f0a4-telemetry-combined-ca-bundle\") pod \"telemetry-openstack-openstack-cell1-gmnh8\" (UID: \"b5718b26-a6b1-4b64-812b-a85da3d6f0a4\") " pod="openstack/telemetry-openstack-openstack-cell1-gmnh8" Mar 13 16:22:53 crc kubenswrapper[4907]: I0313 16:22:53.083455 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b5718b26-a6b1-4b64-812b-a85da3d6f0a4-ceph\") pod \"telemetry-openstack-openstack-cell1-gmnh8\" (UID: \"b5718b26-a6b1-4b64-812b-a85da3d6f0a4\") " pod="openstack/telemetry-openstack-openstack-cell1-gmnh8" Mar 13 16:22:53 crc kubenswrapper[4907]: I0313 16:22:53.083533 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b5718b26-a6b1-4b64-812b-a85da3d6f0a4-inventory\") pod \"telemetry-openstack-openstack-cell1-gmnh8\" (UID: \"b5718b26-a6b1-4b64-812b-a85da3d6f0a4\") " pod="openstack/telemetry-openstack-openstack-cell1-gmnh8" Mar 13 16:22:53 crc kubenswrapper[4907]: I0313 16:22:53.083599 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zkd9z\" (UniqueName: \"kubernetes.io/projected/b5718b26-a6b1-4b64-812b-a85da3d6f0a4-kube-api-access-zkd9z\") pod \"telemetry-openstack-openstack-cell1-gmnh8\" (UID: \"b5718b26-a6b1-4b64-812b-a85da3d6f0a4\") " pod="openstack/telemetry-openstack-openstack-cell1-gmnh8" Mar 13 16:22:53 crc kubenswrapper[4907]: I0313 16:22:53.083694 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/b5718b26-a6b1-4b64-812b-a85da3d6f0a4-ceilometer-compute-config-data-2\") pod \"telemetry-openstack-openstack-cell1-gmnh8\" (UID: \"b5718b26-a6b1-4b64-812b-a85da3d6f0a4\") " pod="openstack/telemetry-openstack-openstack-cell1-gmnh8" Mar 13 16:22:53 crc kubenswrapper[4907]: I0313 16:22:53.083832 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/b5718b26-a6b1-4b64-812b-a85da3d6f0a4-ssh-key-openstack-cell1\") pod \"telemetry-openstack-openstack-cell1-gmnh8\" (UID: \"b5718b26-a6b1-4b64-812b-a85da3d6f0a4\") " pod="openstack/telemetry-openstack-openstack-cell1-gmnh8" Mar 13 16:22:53 crc kubenswrapper[4907]: I0313 16:22:53.083920 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/b5718b26-a6b1-4b64-812b-a85da3d6f0a4-ceilometer-compute-config-data-0\") pod \"telemetry-openstack-openstack-cell1-gmnh8\" (UID: \"b5718b26-a6b1-4b64-812b-a85da3d6f0a4\") " pod="openstack/telemetry-openstack-openstack-cell1-gmnh8" Mar 13 16:22:53 crc kubenswrapper[4907]: I0313 16:22:53.096787 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/b5718b26-a6b1-4b64-812b-a85da3d6f0a4-ceilometer-compute-config-data-2\") pod \"telemetry-openstack-openstack-cell1-gmnh8\" (UID: \"b5718b26-a6b1-4b64-812b-a85da3d6f0a4\") " pod="openstack/telemetry-openstack-openstack-cell1-gmnh8" Mar 13 16:22:53 crc kubenswrapper[4907]: I0313 16:22:53.102475 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/b5718b26-a6b1-4b64-812b-a85da3d6f0a4-ceilometer-compute-config-data-0\") pod \"telemetry-openstack-openstack-cell1-gmnh8\" (UID: \"b5718b26-a6b1-4b64-812b-a85da3d6f0a4\") " pod="openstack/telemetry-openstack-openstack-cell1-gmnh8" Mar 13 16:22:53 crc kubenswrapper[4907]: I0313 16:22:53.104671 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/b5718b26-a6b1-4b64-812b-a85da3d6f0a4-ceilometer-compute-config-data-1\") pod \"telemetry-openstack-openstack-cell1-gmnh8\" (UID: \"b5718b26-a6b1-4b64-812b-a85da3d6f0a4\") " pod="openstack/telemetry-openstack-openstack-cell1-gmnh8" Mar 13 16:22:53 crc kubenswrapper[4907]: I0313 16:22:53.105261 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/b5718b26-a6b1-4b64-812b-a85da3d6f0a4-ssh-key-openstack-cell1\") pod \"telemetry-openstack-openstack-cell1-gmnh8\" (UID: \"b5718b26-a6b1-4b64-812b-a85da3d6f0a4\") " pod="openstack/telemetry-openstack-openstack-cell1-gmnh8" Mar 13 16:22:53 crc kubenswrapper[4907]: I0313 16:22:53.105268 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b5718b26-a6b1-4b64-812b-a85da3d6f0a4-inventory\") pod \"telemetry-openstack-openstack-cell1-gmnh8\" (UID: \"b5718b26-a6b1-4b64-812b-a85da3d6f0a4\") " pod="openstack/telemetry-openstack-openstack-cell1-gmnh8" Mar 13 16:22:53 crc kubenswrapper[4907]: I0313 16:22:53.108554 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b5718b26-a6b1-4b64-812b-a85da3d6f0a4-ceph\") pod \"telemetry-openstack-openstack-cell1-gmnh8\" (UID: \"b5718b26-a6b1-4b64-812b-a85da3d6f0a4\") " pod="openstack/telemetry-openstack-openstack-cell1-gmnh8" Mar 13 16:22:53 crc kubenswrapper[4907]: I0313 16:22:53.114648 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5718b26-a6b1-4b64-812b-a85da3d6f0a4-telemetry-combined-ca-bundle\") pod \"telemetry-openstack-openstack-cell1-gmnh8\" (UID: \"b5718b26-a6b1-4b64-812b-a85da3d6f0a4\") " pod="openstack/telemetry-openstack-openstack-cell1-gmnh8" Mar 13 16:22:53 crc kubenswrapper[4907]: I0313 16:22:53.124760 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zkd9z\" (UniqueName: \"kubernetes.io/projected/b5718b26-a6b1-4b64-812b-a85da3d6f0a4-kube-api-access-zkd9z\") pod \"telemetry-openstack-openstack-cell1-gmnh8\" (UID: \"b5718b26-a6b1-4b64-812b-a85da3d6f0a4\") " pod="openstack/telemetry-openstack-openstack-cell1-gmnh8" Mar 13 16:22:53 crc kubenswrapper[4907]: I0313 16:22:53.209856 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-openstack-openstack-cell1-gmnh8" Mar 13 16:22:53 crc kubenswrapper[4907]: I0313 16:22:53.636251 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-openstack-openstack-cell1-gmnh8"] Mar 13 16:22:53 crc kubenswrapper[4907]: I0313 16:22:53.639355 4907 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 13 16:22:53 crc kubenswrapper[4907]: I0313 16:22:53.648673 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-openstack-openstack-cell1-gmnh8" event={"ID":"b5718b26-a6b1-4b64-812b-a85da3d6f0a4","Type":"ContainerStarted","Data":"7dfdeb24873be617cea682f0960f2b304c8f052d988caccff6de24f90a7a69e8"} Mar 13 16:22:54 crc kubenswrapper[4907]: I0313 16:22:54.659383 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-openstack-openstack-cell1-gmnh8" event={"ID":"b5718b26-a6b1-4b64-812b-a85da3d6f0a4","Type":"ContainerStarted","Data":"f1b9311ea4a0fb969131c4a9d82756d2d971316a7c3c03b8fb8bdcf2c37cd315"} Mar 13 16:22:54 crc kubenswrapper[4907]: I0313 16:22:54.686334 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-openstack-openstack-cell1-gmnh8" podStartSLOduration=2.107624687 podStartE2EDuration="2.686315194s" podCreationTimestamp="2026-03-13 16:22:52 +0000 UTC" firstStartedPulling="2026-03-13 16:22:53.639155655 +0000 UTC m=+8272.538943344" lastFinishedPulling="2026-03-13 16:22:54.217846142 +0000 UTC m=+8273.117633851" observedRunningTime="2026-03-13 16:22:54.681346177 +0000 UTC m=+8273.581133866" watchObservedRunningTime="2026-03-13 16:22:54.686315194 +0000 UTC m=+8273.586102883" Mar 13 16:23:05 crc kubenswrapper[4907]: I0313 16:23:05.185447 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-xgc75"] Mar 13 16:23:05 crc kubenswrapper[4907]: I0313 16:23:05.196660 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xgc75" Mar 13 16:23:05 crc kubenswrapper[4907]: I0313 16:23:05.256893 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xgc75"] Mar 13 16:23:05 crc kubenswrapper[4907]: I0313 16:23:05.287058 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dde3cb5c-6ddf-48c3-b6e1-5bb6a317076f-catalog-content\") pod \"certified-operators-xgc75\" (UID: \"dde3cb5c-6ddf-48c3-b6e1-5bb6a317076f\") " pod="openshift-marketplace/certified-operators-xgc75" Mar 13 16:23:05 crc kubenswrapper[4907]: I0313 16:23:05.287161 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dde3cb5c-6ddf-48c3-b6e1-5bb6a317076f-utilities\") pod \"certified-operators-xgc75\" (UID: \"dde3cb5c-6ddf-48c3-b6e1-5bb6a317076f\") " pod="openshift-marketplace/certified-operators-xgc75" Mar 13 16:23:05 crc kubenswrapper[4907]: I0313 16:23:05.287199 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pqvp5\" (UniqueName: \"kubernetes.io/projected/dde3cb5c-6ddf-48c3-b6e1-5bb6a317076f-kube-api-access-pqvp5\") pod \"certified-operators-xgc75\" (UID: \"dde3cb5c-6ddf-48c3-b6e1-5bb6a317076f\") " pod="openshift-marketplace/certified-operators-xgc75" Mar 13 16:23:05 crc kubenswrapper[4907]: I0313 16:23:05.388422 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dde3cb5c-6ddf-48c3-b6e1-5bb6a317076f-catalog-content\") pod \"certified-operators-xgc75\" (UID: \"dde3cb5c-6ddf-48c3-b6e1-5bb6a317076f\") " pod="openshift-marketplace/certified-operators-xgc75" Mar 13 16:23:05 crc kubenswrapper[4907]: I0313 16:23:05.388524 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dde3cb5c-6ddf-48c3-b6e1-5bb6a317076f-utilities\") pod \"certified-operators-xgc75\" (UID: \"dde3cb5c-6ddf-48c3-b6e1-5bb6a317076f\") " pod="openshift-marketplace/certified-operators-xgc75" Mar 13 16:23:05 crc kubenswrapper[4907]: I0313 16:23:05.388555 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pqvp5\" (UniqueName: \"kubernetes.io/projected/dde3cb5c-6ddf-48c3-b6e1-5bb6a317076f-kube-api-access-pqvp5\") pod \"certified-operators-xgc75\" (UID: \"dde3cb5c-6ddf-48c3-b6e1-5bb6a317076f\") " pod="openshift-marketplace/certified-operators-xgc75" Mar 13 16:23:05 crc kubenswrapper[4907]: I0313 16:23:05.388963 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dde3cb5c-6ddf-48c3-b6e1-5bb6a317076f-catalog-content\") pod \"certified-operators-xgc75\" (UID: \"dde3cb5c-6ddf-48c3-b6e1-5bb6a317076f\") " pod="openshift-marketplace/certified-operators-xgc75" Mar 13 16:23:05 crc kubenswrapper[4907]: I0313 16:23:05.389100 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dde3cb5c-6ddf-48c3-b6e1-5bb6a317076f-utilities\") pod \"certified-operators-xgc75\" (UID: \"dde3cb5c-6ddf-48c3-b6e1-5bb6a317076f\") " pod="openshift-marketplace/certified-operators-xgc75" Mar 13 16:23:05 crc kubenswrapper[4907]: I0313 16:23:05.436593 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pqvp5\" (UniqueName: \"kubernetes.io/projected/dde3cb5c-6ddf-48c3-b6e1-5bb6a317076f-kube-api-access-pqvp5\") pod \"certified-operators-xgc75\" (UID: \"dde3cb5c-6ddf-48c3-b6e1-5bb6a317076f\") " pod="openshift-marketplace/certified-operators-xgc75" Mar 13 16:23:05 crc kubenswrapper[4907]: I0313 16:23:05.546663 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xgc75" Mar 13 16:23:06 crc kubenswrapper[4907]: I0313 16:23:06.090535 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xgc75"] Mar 13 16:23:06 crc kubenswrapper[4907]: I0313 16:23:06.798036 4907 generic.go:334] "Generic (PLEG): container finished" podID="dde3cb5c-6ddf-48c3-b6e1-5bb6a317076f" containerID="28de211c3a6fefc72bee431d7f78d30aea1198e921e5c1b38cddb8faf2098070" exitCode=0 Mar 13 16:23:06 crc kubenswrapper[4907]: I0313 16:23:06.798186 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xgc75" event={"ID":"dde3cb5c-6ddf-48c3-b6e1-5bb6a317076f","Type":"ContainerDied","Data":"28de211c3a6fefc72bee431d7f78d30aea1198e921e5c1b38cddb8faf2098070"} Mar 13 16:23:06 crc kubenswrapper[4907]: I0313 16:23:06.798498 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xgc75" event={"ID":"dde3cb5c-6ddf-48c3-b6e1-5bb6a317076f","Type":"ContainerStarted","Data":"9469a335dd59f1115dde7a5de517c050a6b96d5a4cf517b7778fd82b0e770bc9"} Mar 13 16:23:06 crc kubenswrapper[4907]: I0313 16:23:06.987094 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-65t7j"] Mar 13 16:23:06 crc kubenswrapper[4907]: I0313 16:23:06.990668 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-65t7j" Mar 13 16:23:07 crc kubenswrapper[4907]: I0313 16:23:07.015949 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-65t7j"] Mar 13 16:23:07 crc kubenswrapper[4907]: I0313 16:23:07.127539 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/07d0b87a-9167-4d87-8e03-ce6d6a804339-utilities\") pod \"community-operators-65t7j\" (UID: \"07d0b87a-9167-4d87-8e03-ce6d6a804339\") " pod="openshift-marketplace/community-operators-65t7j" Mar 13 16:23:07 crc kubenswrapper[4907]: I0313 16:23:07.127682 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/07d0b87a-9167-4d87-8e03-ce6d6a804339-catalog-content\") pod \"community-operators-65t7j\" (UID: \"07d0b87a-9167-4d87-8e03-ce6d6a804339\") " pod="openshift-marketplace/community-operators-65t7j" Mar 13 16:23:07 crc kubenswrapper[4907]: I0313 16:23:07.127764 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xz4lr\" (UniqueName: \"kubernetes.io/projected/07d0b87a-9167-4d87-8e03-ce6d6a804339-kube-api-access-xz4lr\") pod \"community-operators-65t7j\" (UID: \"07d0b87a-9167-4d87-8e03-ce6d6a804339\") " pod="openshift-marketplace/community-operators-65t7j" Mar 13 16:23:07 crc kubenswrapper[4907]: I0313 16:23:07.229757 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xz4lr\" (UniqueName: \"kubernetes.io/projected/07d0b87a-9167-4d87-8e03-ce6d6a804339-kube-api-access-xz4lr\") pod \"community-operators-65t7j\" (UID: \"07d0b87a-9167-4d87-8e03-ce6d6a804339\") " pod="openshift-marketplace/community-operators-65t7j" Mar 13 16:23:07 crc kubenswrapper[4907]: I0313 16:23:07.229965 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/07d0b87a-9167-4d87-8e03-ce6d6a804339-utilities\") pod \"community-operators-65t7j\" (UID: \"07d0b87a-9167-4d87-8e03-ce6d6a804339\") " pod="openshift-marketplace/community-operators-65t7j" Mar 13 16:23:07 crc kubenswrapper[4907]: I0313 16:23:07.230063 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/07d0b87a-9167-4d87-8e03-ce6d6a804339-catalog-content\") pod \"community-operators-65t7j\" (UID: \"07d0b87a-9167-4d87-8e03-ce6d6a804339\") " pod="openshift-marketplace/community-operators-65t7j" Mar 13 16:23:07 crc kubenswrapper[4907]: I0313 16:23:07.230466 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/07d0b87a-9167-4d87-8e03-ce6d6a804339-utilities\") pod \"community-operators-65t7j\" (UID: \"07d0b87a-9167-4d87-8e03-ce6d6a804339\") " pod="openshift-marketplace/community-operators-65t7j" Mar 13 16:23:07 crc kubenswrapper[4907]: I0313 16:23:07.230558 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/07d0b87a-9167-4d87-8e03-ce6d6a804339-catalog-content\") pod \"community-operators-65t7j\" (UID: \"07d0b87a-9167-4d87-8e03-ce6d6a804339\") " pod="openshift-marketplace/community-operators-65t7j" Mar 13 16:23:07 crc kubenswrapper[4907]: I0313 16:23:07.253044 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xz4lr\" (UniqueName: \"kubernetes.io/projected/07d0b87a-9167-4d87-8e03-ce6d6a804339-kube-api-access-xz4lr\") pod \"community-operators-65t7j\" (UID: \"07d0b87a-9167-4d87-8e03-ce6d6a804339\") " pod="openshift-marketplace/community-operators-65t7j" Mar 13 16:23:07 crc kubenswrapper[4907]: I0313 16:23:07.314014 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-65t7j" Mar 13 16:23:07 crc kubenswrapper[4907]: I0313 16:23:07.852556 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-65t7j"] Mar 13 16:23:08 crc kubenswrapper[4907]: I0313 16:23:08.821111 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xgc75" event={"ID":"dde3cb5c-6ddf-48c3-b6e1-5bb6a317076f","Type":"ContainerStarted","Data":"11f2c82c12dc7835d9df366ec5b9750f88b9f2e5542cae61b51ff42bea344468"} Mar 13 16:23:08 crc kubenswrapper[4907]: I0313 16:23:08.823873 4907 generic.go:334] "Generic (PLEG): container finished" podID="07d0b87a-9167-4d87-8e03-ce6d6a804339" containerID="ae19ff70cc9978ff9421f0487f5d5f5b18831382d5fa4d8e47971fb1fcf6259c" exitCode=0 Mar 13 16:23:08 crc kubenswrapper[4907]: I0313 16:23:08.824039 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-65t7j" event={"ID":"07d0b87a-9167-4d87-8e03-ce6d6a804339","Type":"ContainerDied","Data":"ae19ff70cc9978ff9421f0487f5d5f5b18831382d5fa4d8e47971fb1fcf6259c"} Mar 13 16:23:08 crc kubenswrapper[4907]: I0313 16:23:08.824147 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-65t7j" event={"ID":"07d0b87a-9167-4d87-8e03-ce6d6a804339","Type":"ContainerStarted","Data":"8c53a2d922cabc54204bd5744de5a1de782425903b460f4444aa5c0d4c525729"} Mar 13 16:23:09 crc kubenswrapper[4907]: I0313 16:23:09.839057 4907 generic.go:334] "Generic (PLEG): container finished" podID="dde3cb5c-6ddf-48c3-b6e1-5bb6a317076f" containerID="11f2c82c12dc7835d9df366ec5b9750f88b9f2e5542cae61b51ff42bea344468" exitCode=0 Mar 13 16:23:09 crc kubenswrapper[4907]: I0313 16:23:09.839172 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xgc75" event={"ID":"dde3cb5c-6ddf-48c3-b6e1-5bb6a317076f","Type":"ContainerDied","Data":"11f2c82c12dc7835d9df366ec5b9750f88b9f2e5542cae61b51ff42bea344468"} Mar 13 16:23:10 crc kubenswrapper[4907]: I0313 16:23:10.863225 4907 generic.go:334] "Generic (PLEG): container finished" podID="07d0b87a-9167-4d87-8e03-ce6d6a804339" containerID="ae44c65486353b068afdd09b8a8dcfc34f3a1a06614480a64352aea46afeaa55" exitCode=0 Mar 13 16:23:10 crc kubenswrapper[4907]: I0313 16:23:10.863761 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-65t7j" event={"ID":"07d0b87a-9167-4d87-8e03-ce6d6a804339","Type":"ContainerDied","Data":"ae44c65486353b068afdd09b8a8dcfc34f3a1a06614480a64352aea46afeaa55"} Mar 13 16:23:10 crc kubenswrapper[4907]: I0313 16:23:10.867253 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xgc75" event={"ID":"dde3cb5c-6ddf-48c3-b6e1-5bb6a317076f","Type":"ContainerStarted","Data":"0dd6c65f3084d0e1c74b304e1362948ff901486662743a1e17191174f6c000f8"} Mar 13 16:23:10 crc kubenswrapper[4907]: I0313 16:23:10.923335 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-xgc75" podStartSLOduration=2.303136071 podStartE2EDuration="5.923317688s" podCreationTimestamp="2026-03-13 16:23:05 +0000 UTC" firstStartedPulling="2026-03-13 16:23:06.80032306 +0000 UTC m=+8285.700110769" lastFinishedPulling="2026-03-13 16:23:10.420504667 +0000 UTC m=+8289.320292386" observedRunningTime="2026-03-13 16:23:10.910080996 +0000 UTC m=+8289.809868685" watchObservedRunningTime="2026-03-13 16:23:10.923317688 +0000 UTC m=+8289.823105377" Mar 13 16:23:11 crc kubenswrapper[4907]: I0313 16:23:11.880773 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-65t7j" event={"ID":"07d0b87a-9167-4d87-8e03-ce6d6a804339","Type":"ContainerStarted","Data":"be78711dbfea89545df6cec6b7c88e5079fe5c9b3bea035012fa58e156104ad0"} Mar 13 16:23:11 crc kubenswrapper[4907]: I0313 16:23:11.901459 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-65t7j" podStartSLOduration=3.233586485 podStartE2EDuration="5.901439458s" podCreationTimestamp="2026-03-13 16:23:06 +0000 UTC" firstStartedPulling="2026-03-13 16:23:08.82600973 +0000 UTC m=+8287.725797419" lastFinishedPulling="2026-03-13 16:23:11.493862693 +0000 UTC m=+8290.393650392" observedRunningTime="2026-03-13 16:23:11.899571427 +0000 UTC m=+8290.799359126" watchObservedRunningTime="2026-03-13 16:23:11.901439458 +0000 UTC m=+8290.801227157" Mar 13 16:23:15 crc kubenswrapper[4907]: I0313 16:23:15.547147 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-xgc75" Mar 13 16:23:15 crc kubenswrapper[4907]: I0313 16:23:15.547751 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-xgc75" Mar 13 16:23:15 crc kubenswrapper[4907]: I0313 16:23:15.639974 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-xgc75" Mar 13 16:23:16 crc kubenswrapper[4907]: I0313 16:23:16.011586 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-xgc75" Mar 13 16:23:17 crc kubenswrapper[4907]: I0313 16:23:17.314191 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-65t7j" Mar 13 16:23:17 crc kubenswrapper[4907]: I0313 16:23:17.314244 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-65t7j" Mar 13 16:23:17 crc kubenswrapper[4907]: I0313 16:23:17.373968 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-65t7j" Mar 13 16:23:18 crc kubenswrapper[4907]: I0313 16:23:18.017549 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-65t7j" Mar 13 16:23:18 crc kubenswrapper[4907]: I0313 16:23:18.041355 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 16:23:18 crc kubenswrapper[4907]: I0313 16:23:18.041424 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 16:23:20 crc kubenswrapper[4907]: I0313 16:23:20.403785 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xgc75"] Mar 13 16:23:20 crc kubenswrapper[4907]: I0313 16:23:20.404309 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-xgc75" podUID="dde3cb5c-6ddf-48c3-b6e1-5bb6a317076f" containerName="registry-server" containerID="cri-o://0dd6c65f3084d0e1c74b304e1362948ff901486662743a1e17191174f6c000f8" gracePeriod=2 Mar 13 16:23:20 crc kubenswrapper[4907]: I0313 16:23:20.921547 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xgc75" Mar 13 16:23:20 crc kubenswrapper[4907]: I0313 16:23:20.981918 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-65t7j"] Mar 13 16:23:20 crc kubenswrapper[4907]: I0313 16:23:20.982136 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-65t7j" podUID="07d0b87a-9167-4d87-8e03-ce6d6a804339" containerName="registry-server" containerID="cri-o://be78711dbfea89545df6cec6b7c88e5079fe5c9b3bea035012fa58e156104ad0" gracePeriod=2 Mar 13 16:23:21 crc kubenswrapper[4907]: I0313 16:23:20.999850 4907 generic.go:334] "Generic (PLEG): container finished" podID="dde3cb5c-6ddf-48c3-b6e1-5bb6a317076f" containerID="0dd6c65f3084d0e1c74b304e1362948ff901486662743a1e17191174f6c000f8" exitCode=0 Mar 13 16:23:21 crc kubenswrapper[4907]: I0313 16:23:20.999972 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xgc75" Mar 13 16:23:21 crc kubenswrapper[4907]: I0313 16:23:20.999994 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xgc75" event={"ID":"dde3cb5c-6ddf-48c3-b6e1-5bb6a317076f","Type":"ContainerDied","Data":"0dd6c65f3084d0e1c74b304e1362948ff901486662743a1e17191174f6c000f8"} Mar 13 16:23:21 crc kubenswrapper[4907]: I0313 16:23:21.000280 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xgc75" event={"ID":"dde3cb5c-6ddf-48c3-b6e1-5bb6a317076f","Type":"ContainerDied","Data":"9469a335dd59f1115dde7a5de517c050a6b96d5a4cf517b7778fd82b0e770bc9"} Mar 13 16:23:21 crc kubenswrapper[4907]: I0313 16:23:21.000297 4907 scope.go:117] "RemoveContainer" containerID="0dd6c65f3084d0e1c74b304e1362948ff901486662743a1e17191174f6c000f8" Mar 13 16:23:21 crc kubenswrapper[4907]: I0313 16:23:21.020709 4907 scope.go:117] "RemoveContainer" containerID="11f2c82c12dc7835d9df366ec5b9750f88b9f2e5542cae61b51ff42bea344468" Mar 13 16:23:21 crc kubenswrapper[4907]: I0313 16:23:21.036111 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pqvp5\" (UniqueName: \"kubernetes.io/projected/dde3cb5c-6ddf-48c3-b6e1-5bb6a317076f-kube-api-access-pqvp5\") pod \"dde3cb5c-6ddf-48c3-b6e1-5bb6a317076f\" (UID: \"dde3cb5c-6ddf-48c3-b6e1-5bb6a317076f\") " Mar 13 16:23:21 crc kubenswrapper[4907]: I0313 16:23:21.036259 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dde3cb5c-6ddf-48c3-b6e1-5bb6a317076f-utilities\") pod \"dde3cb5c-6ddf-48c3-b6e1-5bb6a317076f\" (UID: \"dde3cb5c-6ddf-48c3-b6e1-5bb6a317076f\") " Mar 13 16:23:21 crc kubenswrapper[4907]: I0313 16:23:21.036453 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dde3cb5c-6ddf-48c3-b6e1-5bb6a317076f-catalog-content\") pod \"dde3cb5c-6ddf-48c3-b6e1-5bb6a317076f\" (UID: \"dde3cb5c-6ddf-48c3-b6e1-5bb6a317076f\") " Mar 13 16:23:21 crc kubenswrapper[4907]: I0313 16:23:21.042935 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dde3cb5c-6ddf-48c3-b6e1-5bb6a317076f-utilities" (OuterVolumeSpecName: "utilities") pod "dde3cb5c-6ddf-48c3-b6e1-5bb6a317076f" (UID: "dde3cb5c-6ddf-48c3-b6e1-5bb6a317076f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 16:23:21 crc kubenswrapper[4907]: I0313 16:23:21.050231 4907 scope.go:117] "RemoveContainer" containerID="28de211c3a6fefc72bee431d7f78d30aea1198e921e5c1b38cddb8faf2098070" Mar 13 16:23:21 crc kubenswrapper[4907]: I0313 16:23:21.050315 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dde3cb5c-6ddf-48c3-b6e1-5bb6a317076f-kube-api-access-pqvp5" (OuterVolumeSpecName: "kube-api-access-pqvp5") pod "dde3cb5c-6ddf-48c3-b6e1-5bb6a317076f" (UID: "dde3cb5c-6ddf-48c3-b6e1-5bb6a317076f"). InnerVolumeSpecName "kube-api-access-pqvp5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:23:21 crc kubenswrapper[4907]: I0313 16:23:21.089587 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dde3cb5c-6ddf-48c3-b6e1-5bb6a317076f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "dde3cb5c-6ddf-48c3-b6e1-5bb6a317076f" (UID: "dde3cb5c-6ddf-48c3-b6e1-5bb6a317076f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 16:23:21 crc kubenswrapper[4907]: I0313 16:23:21.139104 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dde3cb5c-6ddf-48c3-b6e1-5bb6a317076f-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 16:23:21 crc kubenswrapper[4907]: I0313 16:23:21.139141 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pqvp5\" (UniqueName: \"kubernetes.io/projected/dde3cb5c-6ddf-48c3-b6e1-5bb6a317076f-kube-api-access-pqvp5\") on node \"crc\" DevicePath \"\"" Mar 13 16:23:21 crc kubenswrapper[4907]: I0313 16:23:21.139154 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dde3cb5c-6ddf-48c3-b6e1-5bb6a317076f-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 16:23:21 crc kubenswrapper[4907]: I0313 16:23:21.210641 4907 scope.go:117] "RemoveContainer" containerID="0dd6c65f3084d0e1c74b304e1362948ff901486662743a1e17191174f6c000f8" Mar 13 16:23:21 crc kubenswrapper[4907]: E0313 16:23:21.211293 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0dd6c65f3084d0e1c74b304e1362948ff901486662743a1e17191174f6c000f8\": container with ID starting with 0dd6c65f3084d0e1c74b304e1362948ff901486662743a1e17191174f6c000f8 not found: ID does not exist" containerID="0dd6c65f3084d0e1c74b304e1362948ff901486662743a1e17191174f6c000f8" Mar 13 16:23:21 crc kubenswrapper[4907]: I0313 16:23:21.211336 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0dd6c65f3084d0e1c74b304e1362948ff901486662743a1e17191174f6c000f8"} err="failed to get container status \"0dd6c65f3084d0e1c74b304e1362948ff901486662743a1e17191174f6c000f8\": rpc error: code = NotFound desc = could not find container \"0dd6c65f3084d0e1c74b304e1362948ff901486662743a1e17191174f6c000f8\": container with ID starting with 0dd6c65f3084d0e1c74b304e1362948ff901486662743a1e17191174f6c000f8 not found: ID does not exist" Mar 13 16:23:21 crc kubenswrapper[4907]: I0313 16:23:21.211365 4907 scope.go:117] "RemoveContainer" containerID="11f2c82c12dc7835d9df366ec5b9750f88b9f2e5542cae61b51ff42bea344468" Mar 13 16:23:21 crc kubenswrapper[4907]: E0313 16:23:21.212932 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"11f2c82c12dc7835d9df366ec5b9750f88b9f2e5542cae61b51ff42bea344468\": container with ID starting with 11f2c82c12dc7835d9df366ec5b9750f88b9f2e5542cae61b51ff42bea344468 not found: ID does not exist" containerID="11f2c82c12dc7835d9df366ec5b9750f88b9f2e5542cae61b51ff42bea344468" Mar 13 16:23:21 crc kubenswrapper[4907]: I0313 16:23:21.212964 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"11f2c82c12dc7835d9df366ec5b9750f88b9f2e5542cae61b51ff42bea344468"} err="failed to get container status \"11f2c82c12dc7835d9df366ec5b9750f88b9f2e5542cae61b51ff42bea344468\": rpc error: code = NotFound desc = could not find container \"11f2c82c12dc7835d9df366ec5b9750f88b9f2e5542cae61b51ff42bea344468\": container with ID starting with 11f2c82c12dc7835d9df366ec5b9750f88b9f2e5542cae61b51ff42bea344468 not found: ID does not exist" Mar 13 16:23:21 crc kubenswrapper[4907]: I0313 16:23:21.212983 4907 scope.go:117] "RemoveContainer" containerID="28de211c3a6fefc72bee431d7f78d30aea1198e921e5c1b38cddb8faf2098070" Mar 13 16:23:21 crc kubenswrapper[4907]: E0313 16:23:21.213248 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"28de211c3a6fefc72bee431d7f78d30aea1198e921e5c1b38cddb8faf2098070\": container with ID starting with 28de211c3a6fefc72bee431d7f78d30aea1198e921e5c1b38cddb8faf2098070 not found: ID does not exist" containerID="28de211c3a6fefc72bee431d7f78d30aea1198e921e5c1b38cddb8faf2098070" Mar 13 16:23:21 crc kubenswrapper[4907]: I0313 16:23:21.213278 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"28de211c3a6fefc72bee431d7f78d30aea1198e921e5c1b38cddb8faf2098070"} err="failed to get container status \"28de211c3a6fefc72bee431d7f78d30aea1198e921e5c1b38cddb8faf2098070\": rpc error: code = NotFound desc = could not find container \"28de211c3a6fefc72bee431d7f78d30aea1198e921e5c1b38cddb8faf2098070\": container with ID starting with 28de211c3a6fefc72bee431d7f78d30aea1198e921e5c1b38cddb8faf2098070 not found: ID does not exist" Mar 13 16:23:21 crc kubenswrapper[4907]: E0313 16:23:21.247324 4907 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod07d0b87a_9167_4d87_8e03_ce6d6a804339.slice/crio-be78711dbfea89545df6cec6b7c88e5079fe5c9b3bea035012fa58e156104ad0.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod07d0b87a_9167_4d87_8e03_ce6d6a804339.slice/crio-conmon-be78711dbfea89545df6cec6b7c88e5079fe5c9b3bea035012fa58e156104ad0.scope\": RecentStats: unable to find data in memory cache]" Mar 13 16:23:21 crc kubenswrapper[4907]: I0313 16:23:21.340692 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xgc75"] Mar 13 16:23:21 crc kubenswrapper[4907]: I0313 16:23:21.350222 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-xgc75"] Mar 13 16:23:21 crc kubenswrapper[4907]: I0313 16:23:21.414302 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-65t7j" Mar 13 16:23:21 crc kubenswrapper[4907]: I0313 16:23:21.546938 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xz4lr\" (UniqueName: \"kubernetes.io/projected/07d0b87a-9167-4d87-8e03-ce6d6a804339-kube-api-access-xz4lr\") pod \"07d0b87a-9167-4d87-8e03-ce6d6a804339\" (UID: \"07d0b87a-9167-4d87-8e03-ce6d6a804339\") " Mar 13 16:23:21 crc kubenswrapper[4907]: I0313 16:23:21.547262 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/07d0b87a-9167-4d87-8e03-ce6d6a804339-catalog-content\") pod \"07d0b87a-9167-4d87-8e03-ce6d6a804339\" (UID: \"07d0b87a-9167-4d87-8e03-ce6d6a804339\") " Mar 13 16:23:21 crc kubenswrapper[4907]: I0313 16:23:21.547428 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/07d0b87a-9167-4d87-8e03-ce6d6a804339-utilities\") pod \"07d0b87a-9167-4d87-8e03-ce6d6a804339\" (UID: \"07d0b87a-9167-4d87-8e03-ce6d6a804339\") " Mar 13 16:23:21 crc kubenswrapper[4907]: I0313 16:23:21.548230 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/07d0b87a-9167-4d87-8e03-ce6d6a804339-utilities" (OuterVolumeSpecName: "utilities") pod "07d0b87a-9167-4d87-8e03-ce6d6a804339" (UID: "07d0b87a-9167-4d87-8e03-ce6d6a804339"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 16:23:21 crc kubenswrapper[4907]: I0313 16:23:21.551289 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/07d0b87a-9167-4d87-8e03-ce6d6a804339-kube-api-access-xz4lr" (OuterVolumeSpecName: "kube-api-access-xz4lr") pod "07d0b87a-9167-4d87-8e03-ce6d6a804339" (UID: "07d0b87a-9167-4d87-8e03-ce6d6a804339"). InnerVolumeSpecName "kube-api-access-xz4lr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:23:21 crc kubenswrapper[4907]: I0313 16:23:21.599280 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/07d0b87a-9167-4d87-8e03-ce6d6a804339-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "07d0b87a-9167-4d87-8e03-ce6d6a804339" (UID: "07d0b87a-9167-4d87-8e03-ce6d6a804339"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 16:23:21 crc kubenswrapper[4907]: I0313 16:23:21.650042 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xz4lr\" (UniqueName: \"kubernetes.io/projected/07d0b87a-9167-4d87-8e03-ce6d6a804339-kube-api-access-xz4lr\") on node \"crc\" DevicePath \"\"" Mar 13 16:23:21 crc kubenswrapper[4907]: I0313 16:23:21.650075 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/07d0b87a-9167-4d87-8e03-ce6d6a804339-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 16:23:21 crc kubenswrapper[4907]: I0313 16:23:21.650086 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/07d0b87a-9167-4d87-8e03-ce6d6a804339-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 16:23:21 crc kubenswrapper[4907]: I0313 16:23:21.795026 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dde3cb5c-6ddf-48c3-b6e1-5bb6a317076f" path="/var/lib/kubelet/pods/dde3cb5c-6ddf-48c3-b6e1-5bb6a317076f/volumes" Mar 13 16:23:22 crc kubenswrapper[4907]: I0313 16:23:22.017464 4907 generic.go:334] "Generic (PLEG): container finished" podID="07d0b87a-9167-4d87-8e03-ce6d6a804339" containerID="be78711dbfea89545df6cec6b7c88e5079fe5c9b3bea035012fa58e156104ad0" exitCode=0 Mar 13 16:23:22 crc kubenswrapper[4907]: I0313 16:23:22.017517 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-65t7j" Mar 13 16:23:22 crc kubenswrapper[4907]: I0313 16:23:22.017535 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-65t7j" event={"ID":"07d0b87a-9167-4d87-8e03-ce6d6a804339","Type":"ContainerDied","Data":"be78711dbfea89545df6cec6b7c88e5079fe5c9b3bea035012fa58e156104ad0"} Mar 13 16:23:22 crc kubenswrapper[4907]: I0313 16:23:22.018043 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-65t7j" event={"ID":"07d0b87a-9167-4d87-8e03-ce6d6a804339","Type":"ContainerDied","Data":"8c53a2d922cabc54204bd5744de5a1de782425903b460f4444aa5c0d4c525729"} Mar 13 16:23:22 crc kubenswrapper[4907]: I0313 16:23:22.018070 4907 scope.go:117] "RemoveContainer" containerID="be78711dbfea89545df6cec6b7c88e5079fe5c9b3bea035012fa58e156104ad0" Mar 13 16:23:22 crc kubenswrapper[4907]: I0313 16:23:22.052218 4907 scope.go:117] "RemoveContainer" containerID="ae44c65486353b068afdd09b8a8dcfc34f3a1a06614480a64352aea46afeaa55" Mar 13 16:23:22 crc kubenswrapper[4907]: I0313 16:23:22.060294 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-65t7j"] Mar 13 16:23:22 crc kubenswrapper[4907]: I0313 16:23:22.073873 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-65t7j"] Mar 13 16:23:22 crc kubenswrapper[4907]: I0313 16:23:22.084358 4907 scope.go:117] "RemoveContainer" containerID="ae19ff70cc9978ff9421f0487f5d5f5b18831382d5fa4d8e47971fb1fcf6259c" Mar 13 16:23:22 crc kubenswrapper[4907]: I0313 16:23:22.145282 4907 scope.go:117] "RemoveContainer" containerID="be78711dbfea89545df6cec6b7c88e5079fe5c9b3bea035012fa58e156104ad0" Mar 13 16:23:22 crc kubenswrapper[4907]: E0313 16:23:22.145781 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"be78711dbfea89545df6cec6b7c88e5079fe5c9b3bea035012fa58e156104ad0\": container with ID starting with be78711dbfea89545df6cec6b7c88e5079fe5c9b3bea035012fa58e156104ad0 not found: ID does not exist" containerID="be78711dbfea89545df6cec6b7c88e5079fe5c9b3bea035012fa58e156104ad0" Mar 13 16:23:22 crc kubenswrapper[4907]: I0313 16:23:22.145832 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"be78711dbfea89545df6cec6b7c88e5079fe5c9b3bea035012fa58e156104ad0"} err="failed to get container status \"be78711dbfea89545df6cec6b7c88e5079fe5c9b3bea035012fa58e156104ad0\": rpc error: code = NotFound desc = could not find container \"be78711dbfea89545df6cec6b7c88e5079fe5c9b3bea035012fa58e156104ad0\": container with ID starting with be78711dbfea89545df6cec6b7c88e5079fe5c9b3bea035012fa58e156104ad0 not found: ID does not exist" Mar 13 16:23:22 crc kubenswrapper[4907]: I0313 16:23:22.145864 4907 scope.go:117] "RemoveContainer" containerID="ae44c65486353b068afdd09b8a8dcfc34f3a1a06614480a64352aea46afeaa55" Mar 13 16:23:22 crc kubenswrapper[4907]: E0313 16:23:22.146467 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ae44c65486353b068afdd09b8a8dcfc34f3a1a06614480a64352aea46afeaa55\": container with ID starting with ae44c65486353b068afdd09b8a8dcfc34f3a1a06614480a64352aea46afeaa55 not found: ID does not exist" containerID="ae44c65486353b068afdd09b8a8dcfc34f3a1a06614480a64352aea46afeaa55" Mar 13 16:23:22 crc kubenswrapper[4907]: I0313 16:23:22.146497 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ae44c65486353b068afdd09b8a8dcfc34f3a1a06614480a64352aea46afeaa55"} err="failed to get container status \"ae44c65486353b068afdd09b8a8dcfc34f3a1a06614480a64352aea46afeaa55\": rpc error: code = NotFound desc = could not find container \"ae44c65486353b068afdd09b8a8dcfc34f3a1a06614480a64352aea46afeaa55\": container with ID starting with ae44c65486353b068afdd09b8a8dcfc34f3a1a06614480a64352aea46afeaa55 not found: ID does not exist" Mar 13 16:23:22 crc kubenswrapper[4907]: I0313 16:23:22.146514 4907 scope.go:117] "RemoveContainer" containerID="ae19ff70cc9978ff9421f0487f5d5f5b18831382d5fa4d8e47971fb1fcf6259c" Mar 13 16:23:22 crc kubenswrapper[4907]: E0313 16:23:22.147116 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ae19ff70cc9978ff9421f0487f5d5f5b18831382d5fa4d8e47971fb1fcf6259c\": container with ID starting with ae19ff70cc9978ff9421f0487f5d5f5b18831382d5fa4d8e47971fb1fcf6259c not found: ID does not exist" containerID="ae19ff70cc9978ff9421f0487f5d5f5b18831382d5fa4d8e47971fb1fcf6259c" Mar 13 16:23:22 crc kubenswrapper[4907]: I0313 16:23:22.147177 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ae19ff70cc9978ff9421f0487f5d5f5b18831382d5fa4d8e47971fb1fcf6259c"} err="failed to get container status \"ae19ff70cc9978ff9421f0487f5d5f5b18831382d5fa4d8e47971fb1fcf6259c\": rpc error: code = NotFound desc = could not find container \"ae19ff70cc9978ff9421f0487f5d5f5b18831382d5fa4d8e47971fb1fcf6259c\": container with ID starting with ae19ff70cc9978ff9421f0487f5d5f5b18831382d5fa4d8e47971fb1fcf6259c not found: ID does not exist" Mar 13 16:23:23 crc kubenswrapper[4907]: I0313 16:23:23.812424 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="07d0b87a-9167-4d87-8e03-ce6d6a804339" path="/var/lib/kubelet/pods/07d0b87a-9167-4d87-8e03-ce6d6a804339/volumes" Mar 13 16:23:48 crc kubenswrapper[4907]: I0313 16:23:48.041731 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 16:23:48 crc kubenswrapper[4907]: I0313 16:23:48.042439 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 16:24:00 crc kubenswrapper[4907]: I0313 16:24:00.163542 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556984-2lfjr"] Mar 13 16:24:00 crc kubenswrapper[4907]: E0313 16:24:00.164332 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07d0b87a-9167-4d87-8e03-ce6d6a804339" containerName="registry-server" Mar 13 16:24:00 crc kubenswrapper[4907]: I0313 16:24:00.164343 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="07d0b87a-9167-4d87-8e03-ce6d6a804339" containerName="registry-server" Mar 13 16:24:00 crc kubenswrapper[4907]: E0313 16:24:00.164374 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07d0b87a-9167-4d87-8e03-ce6d6a804339" containerName="extract-utilities" Mar 13 16:24:00 crc kubenswrapper[4907]: I0313 16:24:00.164380 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="07d0b87a-9167-4d87-8e03-ce6d6a804339" containerName="extract-utilities" Mar 13 16:24:00 crc kubenswrapper[4907]: E0313 16:24:00.164389 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dde3cb5c-6ddf-48c3-b6e1-5bb6a317076f" containerName="extract-content" Mar 13 16:24:00 crc kubenswrapper[4907]: I0313 16:24:00.164395 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="dde3cb5c-6ddf-48c3-b6e1-5bb6a317076f" containerName="extract-content" Mar 13 16:24:00 crc kubenswrapper[4907]: E0313 16:24:00.164409 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dde3cb5c-6ddf-48c3-b6e1-5bb6a317076f" containerName="extract-utilities" Mar 13 16:24:00 crc kubenswrapper[4907]: I0313 16:24:00.164415 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="dde3cb5c-6ddf-48c3-b6e1-5bb6a317076f" containerName="extract-utilities" Mar 13 16:24:00 crc kubenswrapper[4907]: E0313 16:24:00.164429 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dde3cb5c-6ddf-48c3-b6e1-5bb6a317076f" containerName="registry-server" Mar 13 16:24:00 crc kubenswrapper[4907]: I0313 16:24:00.164434 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="dde3cb5c-6ddf-48c3-b6e1-5bb6a317076f" containerName="registry-server" Mar 13 16:24:00 crc kubenswrapper[4907]: E0313 16:24:00.164448 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07d0b87a-9167-4d87-8e03-ce6d6a804339" containerName="extract-content" Mar 13 16:24:00 crc kubenswrapper[4907]: I0313 16:24:00.164454 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="07d0b87a-9167-4d87-8e03-ce6d6a804339" containerName="extract-content" Mar 13 16:24:00 crc kubenswrapper[4907]: I0313 16:24:00.164629 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="07d0b87a-9167-4d87-8e03-ce6d6a804339" containerName="registry-server" Mar 13 16:24:00 crc kubenswrapper[4907]: I0313 16:24:00.164653 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="dde3cb5c-6ddf-48c3-b6e1-5bb6a317076f" containerName="registry-server" Mar 13 16:24:00 crc kubenswrapper[4907]: I0313 16:24:00.165323 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556984-2lfjr" Mar 13 16:24:00 crc kubenswrapper[4907]: I0313 16:24:00.167294 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 16:24:00 crc kubenswrapper[4907]: I0313 16:24:00.167407 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 16:24:00 crc kubenswrapper[4907]: I0313 16:24:00.171609 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 16:24:00 crc kubenswrapper[4907]: I0313 16:24:00.185296 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556984-2lfjr"] Mar 13 16:24:00 crc kubenswrapper[4907]: I0313 16:24:00.258118 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-655b6\" (UniqueName: \"kubernetes.io/projected/d0090a6a-15d8-4445-9830-9cab1aec7fdb-kube-api-access-655b6\") pod \"auto-csr-approver-29556984-2lfjr\" (UID: \"d0090a6a-15d8-4445-9830-9cab1aec7fdb\") " pod="openshift-infra/auto-csr-approver-29556984-2lfjr" Mar 13 16:24:00 crc kubenswrapper[4907]: I0313 16:24:00.360249 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-655b6\" (UniqueName: \"kubernetes.io/projected/d0090a6a-15d8-4445-9830-9cab1aec7fdb-kube-api-access-655b6\") pod \"auto-csr-approver-29556984-2lfjr\" (UID: \"d0090a6a-15d8-4445-9830-9cab1aec7fdb\") " pod="openshift-infra/auto-csr-approver-29556984-2lfjr" Mar 13 16:24:00 crc kubenswrapper[4907]: I0313 16:24:00.377742 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-655b6\" (UniqueName: \"kubernetes.io/projected/d0090a6a-15d8-4445-9830-9cab1aec7fdb-kube-api-access-655b6\") pod \"auto-csr-approver-29556984-2lfjr\" (UID: \"d0090a6a-15d8-4445-9830-9cab1aec7fdb\") " pod="openshift-infra/auto-csr-approver-29556984-2lfjr" Mar 13 16:24:00 crc kubenswrapper[4907]: I0313 16:24:00.482971 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556984-2lfjr" Mar 13 16:24:00 crc kubenswrapper[4907]: I0313 16:24:00.972213 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556984-2lfjr"] Mar 13 16:24:01 crc kubenswrapper[4907]: I0313 16:24:01.476091 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556984-2lfjr" event={"ID":"d0090a6a-15d8-4445-9830-9cab1aec7fdb","Type":"ContainerStarted","Data":"9efd579447758f2672ed2e2d8b1e7abfd2ad683604b38a7047ec29a50d072a9f"} Mar 13 16:24:02 crc kubenswrapper[4907]: I0313 16:24:02.489658 4907 generic.go:334] "Generic (PLEG): container finished" podID="d0090a6a-15d8-4445-9830-9cab1aec7fdb" containerID="5fb59125e06d01b85c52baad33f05fd7f12c902ae1f36e7af1e62fdbcb0056b5" exitCode=0 Mar 13 16:24:02 crc kubenswrapper[4907]: I0313 16:24:02.489718 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556984-2lfjr" event={"ID":"d0090a6a-15d8-4445-9830-9cab1aec7fdb","Type":"ContainerDied","Data":"5fb59125e06d01b85c52baad33f05fd7f12c902ae1f36e7af1e62fdbcb0056b5"} Mar 13 16:24:03 crc kubenswrapper[4907]: I0313 16:24:03.916290 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556984-2lfjr" Mar 13 16:24:03 crc kubenswrapper[4907]: I0313 16:24:03.955989 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-655b6\" (UniqueName: \"kubernetes.io/projected/d0090a6a-15d8-4445-9830-9cab1aec7fdb-kube-api-access-655b6\") pod \"d0090a6a-15d8-4445-9830-9cab1aec7fdb\" (UID: \"d0090a6a-15d8-4445-9830-9cab1aec7fdb\") " Mar 13 16:24:03 crc kubenswrapper[4907]: I0313 16:24:03.964904 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d0090a6a-15d8-4445-9830-9cab1aec7fdb-kube-api-access-655b6" (OuterVolumeSpecName: "kube-api-access-655b6") pod "d0090a6a-15d8-4445-9830-9cab1aec7fdb" (UID: "d0090a6a-15d8-4445-9830-9cab1aec7fdb"). InnerVolumeSpecName "kube-api-access-655b6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:24:04 crc kubenswrapper[4907]: I0313 16:24:04.058856 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-655b6\" (UniqueName: \"kubernetes.io/projected/d0090a6a-15d8-4445-9830-9cab1aec7fdb-kube-api-access-655b6\") on node \"crc\" DevicePath \"\"" Mar 13 16:24:04 crc kubenswrapper[4907]: I0313 16:24:04.517411 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556984-2lfjr" event={"ID":"d0090a6a-15d8-4445-9830-9cab1aec7fdb","Type":"ContainerDied","Data":"9efd579447758f2672ed2e2d8b1e7abfd2ad683604b38a7047ec29a50d072a9f"} Mar 13 16:24:04 crc kubenswrapper[4907]: I0313 16:24:04.517454 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9efd579447758f2672ed2e2d8b1e7abfd2ad683604b38a7047ec29a50d072a9f" Mar 13 16:24:04 crc kubenswrapper[4907]: I0313 16:24:04.517475 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556984-2lfjr" Mar 13 16:24:05 crc kubenswrapper[4907]: I0313 16:24:05.005594 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556978-m5zzz"] Mar 13 16:24:05 crc kubenswrapper[4907]: I0313 16:24:05.014848 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556978-m5zzz"] Mar 13 16:24:05 crc kubenswrapper[4907]: I0313 16:24:05.829446 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b4a302e0-25a2-43d0-b5c0-5047d7a2c5a1" path="/var/lib/kubelet/pods/b4a302e0-25a2-43d0-b5c0-5047d7a2c5a1/volumes" Mar 13 16:24:18 crc kubenswrapper[4907]: I0313 16:24:18.042407 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 16:24:18 crc kubenswrapper[4907]: I0313 16:24:18.043205 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 16:24:18 crc kubenswrapper[4907]: I0313 16:24:18.043268 4907 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" Mar 13 16:24:18 crc kubenswrapper[4907]: I0313 16:24:18.044518 4907 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"50d5a864941e0a2162babd980c37981cbcd4e9bfc7e1598617888cc1ec5d9b31"} pod="openshift-machine-config-operator/machine-config-daemon-hm56j" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 16:24:18 crc kubenswrapper[4907]: I0313 16:24:18.044601 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" containerID="cri-o://50d5a864941e0a2162babd980c37981cbcd4e9bfc7e1598617888cc1ec5d9b31" gracePeriod=600 Mar 13 16:24:18 crc kubenswrapper[4907]: I0313 16:24:18.673920 4907 generic.go:334] "Generic (PLEG): container finished" podID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerID="50d5a864941e0a2162babd980c37981cbcd4e9bfc7e1598617888cc1ec5d9b31" exitCode=0 Mar 13 16:24:18 crc kubenswrapper[4907]: I0313 16:24:18.674140 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerDied","Data":"50d5a864941e0a2162babd980c37981cbcd4e9bfc7e1598617888cc1ec5d9b31"} Mar 13 16:24:18 crc kubenswrapper[4907]: I0313 16:24:18.674376 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerStarted","Data":"dfb327c310937cae28a687e8c5c0256a2ab50e59a756669f2586112310e1dc69"} Mar 13 16:24:18 crc kubenswrapper[4907]: I0313 16:24:18.674411 4907 scope.go:117] "RemoveContainer" containerID="300e429cce24fa4eb9aefff9494fe9905ee7b0767504365170f1f9645a3c8b5f" Mar 13 16:24:42 crc kubenswrapper[4907]: I0313 16:24:42.493791 4907 scope.go:117] "RemoveContainer" containerID="f49981cfe3caeb1a9f80b1991ed24d10adb510db888dce88031b1e2cc8045b1b" Mar 13 16:26:00 crc kubenswrapper[4907]: I0313 16:26:00.168240 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556986-nlfck"] Mar 13 16:26:00 crc kubenswrapper[4907]: E0313 16:26:00.169546 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0090a6a-15d8-4445-9830-9cab1aec7fdb" containerName="oc" Mar 13 16:26:00 crc kubenswrapper[4907]: I0313 16:26:00.169569 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0090a6a-15d8-4445-9830-9cab1aec7fdb" containerName="oc" Mar 13 16:26:00 crc kubenswrapper[4907]: I0313 16:26:00.169930 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0090a6a-15d8-4445-9830-9cab1aec7fdb" containerName="oc" Mar 13 16:26:00 crc kubenswrapper[4907]: I0313 16:26:00.171293 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556986-nlfck" Mar 13 16:26:00 crc kubenswrapper[4907]: I0313 16:26:00.174226 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 16:26:00 crc kubenswrapper[4907]: I0313 16:26:00.179777 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 16:26:00 crc kubenswrapper[4907]: I0313 16:26:00.180045 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 16:26:00 crc kubenswrapper[4907]: I0313 16:26:00.212093 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556986-nlfck"] Mar 13 16:26:00 crc kubenswrapper[4907]: I0313 16:26:00.213695 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zbssm\" (UniqueName: \"kubernetes.io/projected/1bfa4f11-b5c4-4767-bb8e-3940faa2a5f4-kube-api-access-zbssm\") pod \"auto-csr-approver-29556986-nlfck\" (UID: \"1bfa4f11-b5c4-4767-bb8e-3940faa2a5f4\") " pod="openshift-infra/auto-csr-approver-29556986-nlfck" Mar 13 16:26:00 crc kubenswrapper[4907]: I0313 16:26:00.315765 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zbssm\" (UniqueName: \"kubernetes.io/projected/1bfa4f11-b5c4-4767-bb8e-3940faa2a5f4-kube-api-access-zbssm\") pod \"auto-csr-approver-29556986-nlfck\" (UID: \"1bfa4f11-b5c4-4767-bb8e-3940faa2a5f4\") " pod="openshift-infra/auto-csr-approver-29556986-nlfck" Mar 13 16:26:00 crc kubenswrapper[4907]: I0313 16:26:00.342531 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zbssm\" (UniqueName: \"kubernetes.io/projected/1bfa4f11-b5c4-4767-bb8e-3940faa2a5f4-kube-api-access-zbssm\") pod \"auto-csr-approver-29556986-nlfck\" (UID: \"1bfa4f11-b5c4-4767-bb8e-3940faa2a5f4\") " pod="openshift-infra/auto-csr-approver-29556986-nlfck" Mar 13 16:26:00 crc kubenswrapper[4907]: I0313 16:26:00.497323 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556986-nlfck" Mar 13 16:26:01 crc kubenswrapper[4907]: I0313 16:26:01.025715 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556986-nlfck"] Mar 13 16:26:01 crc kubenswrapper[4907]: I0313 16:26:01.974847 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556986-nlfck" event={"ID":"1bfa4f11-b5c4-4767-bb8e-3940faa2a5f4","Type":"ContainerStarted","Data":"802d4a1e6437ddc0b99366dc5755b1fe26de2d7d7692ce9bb52745e8afad2ea6"} Mar 13 16:26:02 crc kubenswrapper[4907]: I0313 16:26:02.990051 4907 generic.go:334] "Generic (PLEG): container finished" podID="1bfa4f11-b5c4-4767-bb8e-3940faa2a5f4" containerID="4fc188ddf095682c93e7e1f3f3249b322a5fe91043eda372761cb774c37d43db" exitCode=0 Mar 13 16:26:02 crc kubenswrapper[4907]: I0313 16:26:02.990136 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556986-nlfck" event={"ID":"1bfa4f11-b5c4-4767-bb8e-3940faa2a5f4","Type":"ContainerDied","Data":"4fc188ddf095682c93e7e1f3f3249b322a5fe91043eda372761cb774c37d43db"} Mar 13 16:26:04 crc kubenswrapper[4907]: I0313 16:26:04.489997 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556986-nlfck" Mar 13 16:26:04 crc kubenswrapper[4907]: I0313 16:26:04.530005 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zbssm\" (UniqueName: \"kubernetes.io/projected/1bfa4f11-b5c4-4767-bb8e-3940faa2a5f4-kube-api-access-zbssm\") pod \"1bfa4f11-b5c4-4767-bb8e-3940faa2a5f4\" (UID: \"1bfa4f11-b5c4-4767-bb8e-3940faa2a5f4\") " Mar 13 16:26:04 crc kubenswrapper[4907]: I0313 16:26:04.547196 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bfa4f11-b5c4-4767-bb8e-3940faa2a5f4-kube-api-access-zbssm" (OuterVolumeSpecName: "kube-api-access-zbssm") pod "1bfa4f11-b5c4-4767-bb8e-3940faa2a5f4" (UID: "1bfa4f11-b5c4-4767-bb8e-3940faa2a5f4"). InnerVolumeSpecName "kube-api-access-zbssm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:26:04 crc kubenswrapper[4907]: I0313 16:26:04.631787 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zbssm\" (UniqueName: \"kubernetes.io/projected/1bfa4f11-b5c4-4767-bb8e-3940faa2a5f4-kube-api-access-zbssm\") on node \"crc\" DevicePath \"\"" Mar 13 16:26:05 crc kubenswrapper[4907]: I0313 16:26:05.023995 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556986-nlfck" event={"ID":"1bfa4f11-b5c4-4767-bb8e-3940faa2a5f4","Type":"ContainerDied","Data":"802d4a1e6437ddc0b99366dc5755b1fe26de2d7d7692ce9bb52745e8afad2ea6"} Mar 13 16:26:05 crc kubenswrapper[4907]: I0313 16:26:05.024044 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="802d4a1e6437ddc0b99366dc5755b1fe26de2d7d7692ce9bb52745e8afad2ea6" Mar 13 16:26:05 crc kubenswrapper[4907]: I0313 16:26:05.024113 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556986-nlfck" Mar 13 16:26:05 crc kubenswrapper[4907]: I0313 16:26:05.574973 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556980-5b2q4"] Mar 13 16:26:05 crc kubenswrapper[4907]: I0313 16:26:05.586184 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556980-5b2q4"] Mar 13 16:26:05 crc kubenswrapper[4907]: I0313 16:26:05.797251 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b9615a4-0270-4660-bd34-5efded647a80" path="/var/lib/kubelet/pods/0b9615a4-0270-4660-bd34-5efded647a80/volumes" Mar 13 16:26:18 crc kubenswrapper[4907]: I0313 16:26:18.041722 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 16:26:18 crc kubenswrapper[4907]: I0313 16:26:18.042404 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 16:26:27 crc kubenswrapper[4907]: I0313 16:26:27.322134 4907 generic.go:334] "Generic (PLEG): container finished" podID="b5718b26-a6b1-4b64-812b-a85da3d6f0a4" containerID="f1b9311ea4a0fb969131c4a9d82756d2d971316a7c3c03b8fb8bdcf2c37cd315" exitCode=0 Mar 13 16:26:27 crc kubenswrapper[4907]: I0313 16:26:27.322223 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-openstack-openstack-cell1-gmnh8" event={"ID":"b5718b26-a6b1-4b64-812b-a85da3d6f0a4","Type":"ContainerDied","Data":"f1b9311ea4a0fb969131c4a9d82756d2d971316a7c3c03b8fb8bdcf2c37cd315"} Mar 13 16:26:28 crc kubenswrapper[4907]: I0313 16:26:28.851693 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-openstack-openstack-cell1-gmnh8" Mar 13 16:26:28 crc kubenswrapper[4907]: I0313 16:26:28.869742 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/b5718b26-a6b1-4b64-812b-a85da3d6f0a4-ssh-key-openstack-cell1\") pod \"b5718b26-a6b1-4b64-812b-a85da3d6f0a4\" (UID: \"b5718b26-a6b1-4b64-812b-a85da3d6f0a4\") " Mar 13 16:26:28 crc kubenswrapper[4907]: I0313 16:26:28.872577 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/b5718b26-a6b1-4b64-812b-a85da3d6f0a4-ceilometer-compute-config-data-2\") pod \"b5718b26-a6b1-4b64-812b-a85da3d6f0a4\" (UID: \"b5718b26-a6b1-4b64-812b-a85da3d6f0a4\") " Mar 13 16:26:28 crc kubenswrapper[4907]: I0313 16:26:28.872753 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b5718b26-a6b1-4b64-812b-a85da3d6f0a4-inventory\") pod \"b5718b26-a6b1-4b64-812b-a85da3d6f0a4\" (UID: \"b5718b26-a6b1-4b64-812b-a85da3d6f0a4\") " Mar 13 16:26:28 crc kubenswrapper[4907]: I0313 16:26:28.872863 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/b5718b26-a6b1-4b64-812b-a85da3d6f0a4-ceilometer-compute-config-data-1\") pod \"b5718b26-a6b1-4b64-812b-a85da3d6f0a4\" (UID: \"b5718b26-a6b1-4b64-812b-a85da3d6f0a4\") " Mar 13 16:26:28 crc kubenswrapper[4907]: I0313 16:26:28.873029 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/b5718b26-a6b1-4b64-812b-a85da3d6f0a4-ceilometer-compute-config-data-0\") pod \"b5718b26-a6b1-4b64-812b-a85da3d6f0a4\" (UID: \"b5718b26-a6b1-4b64-812b-a85da3d6f0a4\") " Mar 13 16:26:28 crc kubenswrapper[4907]: I0313 16:26:28.873242 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5718b26-a6b1-4b64-812b-a85da3d6f0a4-telemetry-combined-ca-bundle\") pod \"b5718b26-a6b1-4b64-812b-a85da3d6f0a4\" (UID: \"b5718b26-a6b1-4b64-812b-a85da3d6f0a4\") " Mar 13 16:26:28 crc kubenswrapper[4907]: I0313 16:26:28.873837 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkd9z\" (UniqueName: \"kubernetes.io/projected/b5718b26-a6b1-4b64-812b-a85da3d6f0a4-kube-api-access-zkd9z\") pod \"b5718b26-a6b1-4b64-812b-a85da3d6f0a4\" (UID: \"b5718b26-a6b1-4b64-812b-a85da3d6f0a4\") " Mar 13 16:26:28 crc kubenswrapper[4907]: I0313 16:26:28.874260 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b5718b26-a6b1-4b64-812b-a85da3d6f0a4-ceph\") pod \"b5718b26-a6b1-4b64-812b-a85da3d6f0a4\" (UID: \"b5718b26-a6b1-4b64-812b-a85da3d6f0a4\") " Mar 13 16:26:28 crc kubenswrapper[4907]: I0313 16:26:28.886859 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5718b26-a6b1-4b64-812b-a85da3d6f0a4-ceph" (OuterVolumeSpecName: "ceph") pod "b5718b26-a6b1-4b64-812b-a85da3d6f0a4" (UID: "b5718b26-a6b1-4b64-812b-a85da3d6f0a4"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:26:28 crc kubenswrapper[4907]: I0313 16:26:28.884936 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5718b26-a6b1-4b64-812b-a85da3d6f0a4-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "b5718b26-a6b1-4b64-812b-a85da3d6f0a4" (UID: "b5718b26-a6b1-4b64-812b-a85da3d6f0a4"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:26:28 crc kubenswrapper[4907]: I0313 16:26:28.918499 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5718b26-a6b1-4b64-812b-a85da3d6f0a4-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "b5718b26-a6b1-4b64-812b-a85da3d6f0a4" (UID: "b5718b26-a6b1-4b64-812b-a85da3d6f0a4"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:26:28 crc kubenswrapper[4907]: I0313 16:26:28.929031 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b5718b26-a6b1-4b64-812b-a85da3d6f0a4-kube-api-access-zkd9z" (OuterVolumeSpecName: "kube-api-access-zkd9z") pod "b5718b26-a6b1-4b64-812b-a85da3d6f0a4" (UID: "b5718b26-a6b1-4b64-812b-a85da3d6f0a4"). InnerVolumeSpecName "kube-api-access-zkd9z". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:26:28 crc kubenswrapper[4907]: I0313 16:26:28.932805 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5718b26-a6b1-4b64-812b-a85da3d6f0a4-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "b5718b26-a6b1-4b64-812b-a85da3d6f0a4" (UID: "b5718b26-a6b1-4b64-812b-a85da3d6f0a4"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:26:28 crc kubenswrapper[4907]: I0313 16:26:28.934296 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5718b26-a6b1-4b64-812b-a85da3d6f0a4-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "b5718b26-a6b1-4b64-812b-a85da3d6f0a4" (UID: "b5718b26-a6b1-4b64-812b-a85da3d6f0a4"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:26:28 crc kubenswrapper[4907]: I0313 16:26:28.935056 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5718b26-a6b1-4b64-812b-a85da3d6f0a4-ssh-key-openstack-cell1" (OuterVolumeSpecName: "ssh-key-openstack-cell1") pod "b5718b26-a6b1-4b64-812b-a85da3d6f0a4" (UID: "b5718b26-a6b1-4b64-812b-a85da3d6f0a4"). InnerVolumeSpecName "ssh-key-openstack-cell1". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:26:28 crc kubenswrapper[4907]: I0313 16:26:28.949644 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5718b26-a6b1-4b64-812b-a85da3d6f0a4-inventory" (OuterVolumeSpecName: "inventory") pod "b5718b26-a6b1-4b64-812b-a85da3d6f0a4" (UID: "b5718b26-a6b1-4b64-812b-a85da3d6f0a4"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:26:28 crc kubenswrapper[4907]: I0313 16:26:28.978982 4907 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b5718b26-a6b1-4b64-812b-a85da3d6f0a4-ceph\") on node \"crc\" DevicePath \"\"" Mar 13 16:26:28 crc kubenswrapper[4907]: I0313 16:26:28.979223 4907 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/b5718b26-a6b1-4b64-812b-a85da3d6f0a4-ssh-key-openstack-cell1\") on node \"crc\" DevicePath \"\"" Mar 13 16:26:28 crc kubenswrapper[4907]: I0313 16:26:28.979359 4907 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/b5718b26-a6b1-4b64-812b-a85da3d6f0a4-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Mar 13 16:26:28 crc kubenswrapper[4907]: I0313 16:26:28.979471 4907 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b5718b26-a6b1-4b64-812b-a85da3d6f0a4-inventory\") on node \"crc\" DevicePath \"\"" Mar 13 16:26:28 crc kubenswrapper[4907]: I0313 16:26:28.979580 4907 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/b5718b26-a6b1-4b64-812b-a85da3d6f0a4-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Mar 13 16:26:28 crc kubenswrapper[4907]: I0313 16:26:28.979688 4907 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/b5718b26-a6b1-4b64-812b-a85da3d6f0a4-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Mar 13 16:26:28 crc kubenswrapper[4907]: I0313 16:26:28.979949 4907 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5718b26-a6b1-4b64-812b-a85da3d6f0a4-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 16:26:28 crc kubenswrapper[4907]: I0313 16:26:28.980251 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkd9z\" (UniqueName: \"kubernetes.io/projected/b5718b26-a6b1-4b64-812b-a85da3d6f0a4-kube-api-access-zkd9z\") on node \"crc\" DevicePath \"\"" Mar 13 16:26:29 crc kubenswrapper[4907]: I0313 16:26:29.353371 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-openstack-openstack-cell1-gmnh8" event={"ID":"b5718b26-a6b1-4b64-812b-a85da3d6f0a4","Type":"ContainerDied","Data":"7dfdeb24873be617cea682f0960f2b304c8f052d988caccff6de24f90a7a69e8"} Mar 13 16:26:29 crc kubenswrapper[4907]: I0313 16:26:29.353414 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7dfdeb24873be617cea682f0960f2b304c8f052d988caccff6de24f90a7a69e8" Mar 13 16:26:29 crc kubenswrapper[4907]: I0313 16:26:29.353477 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-openstack-openstack-cell1-gmnh8" Mar 13 16:26:29 crc kubenswrapper[4907]: I0313 16:26:29.459415 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-sriov-openstack-openstack-cell1-sqxhp"] Mar 13 16:26:29 crc kubenswrapper[4907]: E0313 16:26:29.460005 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5718b26-a6b1-4b64-812b-a85da3d6f0a4" containerName="telemetry-openstack-openstack-cell1" Mar 13 16:26:29 crc kubenswrapper[4907]: I0313 16:26:29.460034 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5718b26-a6b1-4b64-812b-a85da3d6f0a4" containerName="telemetry-openstack-openstack-cell1" Mar 13 16:26:29 crc kubenswrapper[4907]: E0313 16:26:29.460069 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bfa4f11-b5c4-4767-bb8e-3940faa2a5f4" containerName="oc" Mar 13 16:26:29 crc kubenswrapper[4907]: I0313 16:26:29.460078 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bfa4f11-b5c4-4767-bb8e-3940faa2a5f4" containerName="oc" Mar 13 16:26:29 crc kubenswrapper[4907]: I0313 16:26:29.460327 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="1bfa4f11-b5c4-4767-bb8e-3940faa2a5f4" containerName="oc" Mar 13 16:26:29 crc kubenswrapper[4907]: I0313 16:26:29.460772 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="b5718b26-a6b1-4b64-812b-a85da3d6f0a4" containerName="telemetry-openstack-openstack-cell1" Mar 13 16:26:29 crc kubenswrapper[4907]: I0313 16:26:29.464934 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-sriov-openstack-openstack-cell1-sqxhp" Mar 13 16:26:29 crc kubenswrapper[4907]: I0313 16:26:29.466796 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-sriov-agent-neutron-config" Mar 13 16:26:29 crc kubenswrapper[4907]: I0313 16:26:29.467364 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Mar 13 16:26:29 crc kubenswrapper[4907]: I0313 16:26:29.467382 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Mar 13 16:26:29 crc kubenswrapper[4907]: I0313 16:26:29.469202 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-44kxw" Mar 13 16:26:29 crc kubenswrapper[4907]: I0313 16:26:29.469847 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 13 16:26:29 crc kubenswrapper[4907]: I0313 16:26:29.490049 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/05d384d7-9635-4d68-8eaf-48fea8bed0b8-neutron-sriov-agent-neutron-config-0\") pod \"neutron-sriov-openstack-openstack-cell1-sqxhp\" (UID: \"05d384d7-9635-4d68-8eaf-48fea8bed0b8\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-sqxhp" Mar 13 16:26:29 crc kubenswrapper[4907]: I0313 16:26:29.490118 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/05d384d7-9635-4d68-8eaf-48fea8bed0b8-ceph\") pod \"neutron-sriov-openstack-openstack-cell1-sqxhp\" (UID: \"05d384d7-9635-4d68-8eaf-48fea8bed0b8\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-sqxhp" Mar 13 16:26:29 crc kubenswrapper[4907]: I0313 16:26:29.490141 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jl9p6\" (UniqueName: \"kubernetes.io/projected/05d384d7-9635-4d68-8eaf-48fea8bed0b8-kube-api-access-jl9p6\") pod \"neutron-sriov-openstack-openstack-cell1-sqxhp\" (UID: \"05d384d7-9635-4d68-8eaf-48fea8bed0b8\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-sqxhp" Mar 13 16:26:29 crc kubenswrapper[4907]: I0313 16:26:29.490167 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/05d384d7-9635-4d68-8eaf-48fea8bed0b8-ssh-key-openstack-cell1\") pod \"neutron-sriov-openstack-openstack-cell1-sqxhp\" (UID: \"05d384d7-9635-4d68-8eaf-48fea8bed0b8\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-sqxhp" Mar 13 16:26:29 crc kubenswrapper[4907]: I0313 16:26:29.490392 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05d384d7-9635-4d68-8eaf-48fea8bed0b8-neutron-sriov-combined-ca-bundle\") pod \"neutron-sriov-openstack-openstack-cell1-sqxhp\" (UID: \"05d384d7-9635-4d68-8eaf-48fea8bed0b8\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-sqxhp" Mar 13 16:26:29 crc kubenswrapper[4907]: I0313 16:26:29.490529 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/05d384d7-9635-4d68-8eaf-48fea8bed0b8-inventory\") pod \"neutron-sriov-openstack-openstack-cell1-sqxhp\" (UID: \"05d384d7-9635-4d68-8eaf-48fea8bed0b8\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-sqxhp" Mar 13 16:26:29 crc kubenswrapper[4907]: I0313 16:26:29.497747 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-sriov-openstack-openstack-cell1-sqxhp"] Mar 13 16:26:29 crc kubenswrapper[4907]: I0313 16:26:29.592584 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05d384d7-9635-4d68-8eaf-48fea8bed0b8-neutron-sriov-combined-ca-bundle\") pod \"neutron-sriov-openstack-openstack-cell1-sqxhp\" (UID: \"05d384d7-9635-4d68-8eaf-48fea8bed0b8\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-sqxhp" Mar 13 16:26:29 crc kubenswrapper[4907]: I0313 16:26:29.592674 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/05d384d7-9635-4d68-8eaf-48fea8bed0b8-inventory\") pod \"neutron-sriov-openstack-openstack-cell1-sqxhp\" (UID: \"05d384d7-9635-4d68-8eaf-48fea8bed0b8\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-sqxhp" Mar 13 16:26:29 crc kubenswrapper[4907]: I0313 16:26:29.592791 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/05d384d7-9635-4d68-8eaf-48fea8bed0b8-neutron-sriov-agent-neutron-config-0\") pod \"neutron-sriov-openstack-openstack-cell1-sqxhp\" (UID: \"05d384d7-9635-4d68-8eaf-48fea8bed0b8\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-sqxhp" Mar 13 16:26:29 crc kubenswrapper[4907]: I0313 16:26:29.592824 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/05d384d7-9635-4d68-8eaf-48fea8bed0b8-ceph\") pod \"neutron-sriov-openstack-openstack-cell1-sqxhp\" (UID: \"05d384d7-9635-4d68-8eaf-48fea8bed0b8\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-sqxhp" Mar 13 16:26:29 crc kubenswrapper[4907]: I0313 16:26:29.592846 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jl9p6\" (UniqueName: \"kubernetes.io/projected/05d384d7-9635-4d68-8eaf-48fea8bed0b8-kube-api-access-jl9p6\") pod \"neutron-sriov-openstack-openstack-cell1-sqxhp\" (UID: \"05d384d7-9635-4d68-8eaf-48fea8bed0b8\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-sqxhp" Mar 13 16:26:29 crc kubenswrapper[4907]: I0313 16:26:29.592874 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/05d384d7-9635-4d68-8eaf-48fea8bed0b8-ssh-key-openstack-cell1\") pod \"neutron-sriov-openstack-openstack-cell1-sqxhp\" (UID: \"05d384d7-9635-4d68-8eaf-48fea8bed0b8\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-sqxhp" Mar 13 16:26:29 crc kubenswrapper[4907]: I0313 16:26:29.596432 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/05d384d7-9635-4d68-8eaf-48fea8bed0b8-neutron-sriov-agent-neutron-config-0\") pod \"neutron-sriov-openstack-openstack-cell1-sqxhp\" (UID: \"05d384d7-9635-4d68-8eaf-48fea8bed0b8\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-sqxhp" Mar 13 16:26:29 crc kubenswrapper[4907]: I0313 16:26:29.597018 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/05d384d7-9635-4d68-8eaf-48fea8bed0b8-inventory\") pod \"neutron-sriov-openstack-openstack-cell1-sqxhp\" (UID: \"05d384d7-9635-4d68-8eaf-48fea8bed0b8\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-sqxhp" Mar 13 16:26:29 crc kubenswrapper[4907]: I0313 16:26:29.597797 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/05d384d7-9635-4d68-8eaf-48fea8bed0b8-ssh-key-openstack-cell1\") pod \"neutron-sriov-openstack-openstack-cell1-sqxhp\" (UID: \"05d384d7-9635-4d68-8eaf-48fea8bed0b8\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-sqxhp" Mar 13 16:26:29 crc kubenswrapper[4907]: I0313 16:26:29.598058 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05d384d7-9635-4d68-8eaf-48fea8bed0b8-neutron-sriov-combined-ca-bundle\") pod \"neutron-sriov-openstack-openstack-cell1-sqxhp\" (UID: \"05d384d7-9635-4d68-8eaf-48fea8bed0b8\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-sqxhp" Mar 13 16:26:29 crc kubenswrapper[4907]: I0313 16:26:29.598512 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/05d384d7-9635-4d68-8eaf-48fea8bed0b8-ceph\") pod \"neutron-sriov-openstack-openstack-cell1-sqxhp\" (UID: \"05d384d7-9635-4d68-8eaf-48fea8bed0b8\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-sqxhp" Mar 13 16:26:29 crc kubenswrapper[4907]: I0313 16:26:29.612260 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jl9p6\" (UniqueName: \"kubernetes.io/projected/05d384d7-9635-4d68-8eaf-48fea8bed0b8-kube-api-access-jl9p6\") pod \"neutron-sriov-openstack-openstack-cell1-sqxhp\" (UID: \"05d384d7-9635-4d68-8eaf-48fea8bed0b8\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-sqxhp" Mar 13 16:26:29 crc kubenswrapper[4907]: I0313 16:26:29.783468 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-sriov-openstack-openstack-cell1-sqxhp" Mar 13 16:26:30 crc kubenswrapper[4907]: I0313 16:26:30.316542 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-sriov-openstack-openstack-cell1-sqxhp"] Mar 13 16:26:30 crc kubenswrapper[4907]: I0313 16:26:30.369760 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-sriov-openstack-openstack-cell1-sqxhp" event={"ID":"05d384d7-9635-4d68-8eaf-48fea8bed0b8","Type":"ContainerStarted","Data":"762fb1ddd2d8fd925d29b186df5f88568978b2b1ec0e68aa4f06e19c4e53f86e"} Mar 13 16:26:31 crc kubenswrapper[4907]: I0313 16:26:31.387602 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-sriov-openstack-openstack-cell1-sqxhp" event={"ID":"05d384d7-9635-4d68-8eaf-48fea8bed0b8","Type":"ContainerStarted","Data":"259cfb9d82d9b9b7b45bab20327a4d8c2e39c52001f26a91569b8c3052ad5700"} Mar 13 16:26:31 crc kubenswrapper[4907]: I0313 16:26:31.425771 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-sriov-openstack-openstack-cell1-sqxhp" podStartSLOduration=1.8257704860000001 podStartE2EDuration="2.425743676s" podCreationTimestamp="2026-03-13 16:26:29 +0000 UTC" firstStartedPulling="2026-03-13 16:26:30.325658588 +0000 UTC m=+8489.225446277" lastFinishedPulling="2026-03-13 16:26:30.925631748 +0000 UTC m=+8489.825419467" observedRunningTime="2026-03-13 16:26:31.406685244 +0000 UTC m=+8490.306472973" watchObservedRunningTime="2026-03-13 16:26:31.425743676 +0000 UTC m=+8490.325531405" Mar 13 16:26:42 crc kubenswrapper[4907]: I0313 16:26:42.661215 4907 scope.go:117] "RemoveContainer" containerID="d626b98830070c30af12d3ea7ba374e64df5c680e5d7d98055b77c7c0b8bf280" Mar 13 16:26:48 crc kubenswrapper[4907]: I0313 16:26:48.041985 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 16:26:48 crc kubenswrapper[4907]: I0313 16:26:48.042934 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 16:27:18 crc kubenswrapper[4907]: I0313 16:27:18.043018 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 16:27:18 crc kubenswrapper[4907]: I0313 16:27:18.043768 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 16:27:18 crc kubenswrapper[4907]: I0313 16:27:18.043839 4907 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" Mar 13 16:27:18 crc kubenswrapper[4907]: I0313 16:27:18.045244 4907 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"dfb327c310937cae28a687e8c5c0256a2ab50e59a756669f2586112310e1dc69"} pod="openshift-machine-config-operator/machine-config-daemon-hm56j" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 16:27:18 crc kubenswrapper[4907]: I0313 16:27:18.045349 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" containerID="cri-o://dfb327c310937cae28a687e8c5c0256a2ab50e59a756669f2586112310e1dc69" gracePeriod=600 Mar 13 16:27:18 crc kubenswrapper[4907]: E0313 16:27:18.169454 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:27:18 crc kubenswrapper[4907]: I0313 16:27:18.948802 4907 generic.go:334] "Generic (PLEG): container finished" podID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerID="dfb327c310937cae28a687e8c5c0256a2ab50e59a756669f2586112310e1dc69" exitCode=0 Mar 13 16:27:18 crc kubenswrapper[4907]: I0313 16:27:18.948898 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerDied","Data":"dfb327c310937cae28a687e8c5c0256a2ab50e59a756669f2586112310e1dc69"} Mar 13 16:27:18 crc kubenswrapper[4907]: I0313 16:27:18.949215 4907 scope.go:117] "RemoveContainer" containerID="50d5a864941e0a2162babd980c37981cbcd4e9bfc7e1598617888cc1ec5d9b31" Mar 13 16:27:18 crc kubenswrapper[4907]: I0313 16:27:18.950018 4907 scope.go:117] "RemoveContainer" containerID="dfb327c310937cae28a687e8c5c0256a2ab50e59a756669f2586112310e1dc69" Mar 13 16:27:18 crc kubenswrapper[4907]: E0313 16:27:18.950444 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:27:33 crc kubenswrapper[4907]: I0313 16:27:33.783759 4907 scope.go:117] "RemoveContainer" containerID="dfb327c310937cae28a687e8c5c0256a2ab50e59a756669f2586112310e1dc69" Mar 13 16:27:33 crc kubenswrapper[4907]: E0313 16:27:33.785055 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:27:40 crc kubenswrapper[4907]: I0313 16:27:40.198097 4907 generic.go:334] "Generic (PLEG): container finished" podID="05d384d7-9635-4d68-8eaf-48fea8bed0b8" containerID="259cfb9d82d9b9b7b45bab20327a4d8c2e39c52001f26a91569b8c3052ad5700" exitCode=0 Mar 13 16:27:40 crc kubenswrapper[4907]: I0313 16:27:40.198239 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-sriov-openstack-openstack-cell1-sqxhp" event={"ID":"05d384d7-9635-4d68-8eaf-48fea8bed0b8","Type":"ContainerDied","Data":"259cfb9d82d9b9b7b45bab20327a4d8c2e39c52001f26a91569b8c3052ad5700"} Mar 13 16:27:41 crc kubenswrapper[4907]: I0313 16:27:41.724518 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-sriov-openstack-openstack-cell1-sqxhp" Mar 13 16:27:41 crc kubenswrapper[4907]: I0313 16:27:41.736543 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/05d384d7-9635-4d68-8eaf-48fea8bed0b8-neutron-sriov-agent-neutron-config-0\") pod \"05d384d7-9635-4d68-8eaf-48fea8bed0b8\" (UID: \"05d384d7-9635-4d68-8eaf-48fea8bed0b8\") " Mar 13 16:27:41 crc kubenswrapper[4907]: I0313 16:27:41.736701 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05d384d7-9635-4d68-8eaf-48fea8bed0b8-neutron-sriov-combined-ca-bundle\") pod \"05d384d7-9635-4d68-8eaf-48fea8bed0b8\" (UID: \"05d384d7-9635-4d68-8eaf-48fea8bed0b8\") " Mar 13 16:27:41 crc kubenswrapper[4907]: I0313 16:27:41.736786 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/05d384d7-9635-4d68-8eaf-48fea8bed0b8-ssh-key-openstack-cell1\") pod \"05d384d7-9635-4d68-8eaf-48fea8bed0b8\" (UID: \"05d384d7-9635-4d68-8eaf-48fea8bed0b8\") " Mar 13 16:27:41 crc kubenswrapper[4907]: I0313 16:27:41.736856 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/05d384d7-9635-4d68-8eaf-48fea8bed0b8-ceph\") pod \"05d384d7-9635-4d68-8eaf-48fea8bed0b8\" (UID: \"05d384d7-9635-4d68-8eaf-48fea8bed0b8\") " Mar 13 16:27:41 crc kubenswrapper[4907]: I0313 16:27:41.737063 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/05d384d7-9635-4d68-8eaf-48fea8bed0b8-inventory\") pod \"05d384d7-9635-4d68-8eaf-48fea8bed0b8\" (UID: \"05d384d7-9635-4d68-8eaf-48fea8bed0b8\") " Mar 13 16:27:41 crc kubenswrapper[4907]: I0313 16:27:41.737169 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jl9p6\" (UniqueName: \"kubernetes.io/projected/05d384d7-9635-4d68-8eaf-48fea8bed0b8-kube-api-access-jl9p6\") pod \"05d384d7-9635-4d68-8eaf-48fea8bed0b8\" (UID: \"05d384d7-9635-4d68-8eaf-48fea8bed0b8\") " Mar 13 16:27:41 crc kubenswrapper[4907]: I0313 16:27:41.745986 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/05d384d7-9635-4d68-8eaf-48fea8bed0b8-neutron-sriov-combined-ca-bundle" (OuterVolumeSpecName: "neutron-sriov-combined-ca-bundle") pod "05d384d7-9635-4d68-8eaf-48fea8bed0b8" (UID: "05d384d7-9635-4d68-8eaf-48fea8bed0b8"). InnerVolumeSpecName "neutron-sriov-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:27:41 crc kubenswrapper[4907]: I0313 16:27:41.747047 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/05d384d7-9635-4d68-8eaf-48fea8bed0b8-kube-api-access-jl9p6" (OuterVolumeSpecName: "kube-api-access-jl9p6") pod "05d384d7-9635-4d68-8eaf-48fea8bed0b8" (UID: "05d384d7-9635-4d68-8eaf-48fea8bed0b8"). InnerVolumeSpecName "kube-api-access-jl9p6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:27:41 crc kubenswrapper[4907]: I0313 16:27:41.762449 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/05d384d7-9635-4d68-8eaf-48fea8bed0b8-ceph" (OuterVolumeSpecName: "ceph") pod "05d384d7-9635-4d68-8eaf-48fea8bed0b8" (UID: "05d384d7-9635-4d68-8eaf-48fea8bed0b8"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:27:41 crc kubenswrapper[4907]: I0313 16:27:41.793013 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/05d384d7-9635-4d68-8eaf-48fea8bed0b8-neutron-sriov-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-sriov-agent-neutron-config-0") pod "05d384d7-9635-4d68-8eaf-48fea8bed0b8" (UID: "05d384d7-9635-4d68-8eaf-48fea8bed0b8"). InnerVolumeSpecName "neutron-sriov-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:27:41 crc kubenswrapper[4907]: I0313 16:27:41.796348 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/05d384d7-9635-4d68-8eaf-48fea8bed0b8-inventory" (OuterVolumeSpecName: "inventory") pod "05d384d7-9635-4d68-8eaf-48fea8bed0b8" (UID: "05d384d7-9635-4d68-8eaf-48fea8bed0b8"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:27:41 crc kubenswrapper[4907]: I0313 16:27:41.814815 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/05d384d7-9635-4d68-8eaf-48fea8bed0b8-ssh-key-openstack-cell1" (OuterVolumeSpecName: "ssh-key-openstack-cell1") pod "05d384d7-9635-4d68-8eaf-48fea8bed0b8" (UID: "05d384d7-9635-4d68-8eaf-48fea8bed0b8"). InnerVolumeSpecName "ssh-key-openstack-cell1". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:27:41 crc kubenswrapper[4907]: I0313 16:27:41.849474 4907 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/05d384d7-9635-4d68-8eaf-48fea8bed0b8-inventory\") on node \"crc\" DevicePath \"\"" Mar 13 16:27:41 crc kubenswrapper[4907]: I0313 16:27:41.849538 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jl9p6\" (UniqueName: \"kubernetes.io/projected/05d384d7-9635-4d68-8eaf-48fea8bed0b8-kube-api-access-jl9p6\") on node \"crc\" DevicePath \"\"" Mar 13 16:27:41 crc kubenswrapper[4907]: I0313 16:27:41.849555 4907 reconciler_common.go:293] "Volume detached for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/05d384d7-9635-4d68-8eaf-48fea8bed0b8-neutron-sriov-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Mar 13 16:27:41 crc kubenswrapper[4907]: I0313 16:27:41.849567 4907 reconciler_common.go:293] "Volume detached for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05d384d7-9635-4d68-8eaf-48fea8bed0b8-neutron-sriov-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 16:27:41 crc kubenswrapper[4907]: I0313 16:27:41.849579 4907 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/05d384d7-9635-4d68-8eaf-48fea8bed0b8-ssh-key-openstack-cell1\") on node \"crc\" DevicePath \"\"" Mar 13 16:27:41 crc kubenswrapper[4907]: I0313 16:27:41.849614 4907 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/05d384d7-9635-4d68-8eaf-48fea8bed0b8-ceph\") on node \"crc\" DevicePath \"\"" Mar 13 16:27:42 crc kubenswrapper[4907]: I0313 16:27:42.228140 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-sriov-openstack-openstack-cell1-sqxhp" event={"ID":"05d384d7-9635-4d68-8eaf-48fea8bed0b8","Type":"ContainerDied","Data":"762fb1ddd2d8fd925d29b186df5f88568978b2b1ec0e68aa4f06e19c4e53f86e"} Mar 13 16:27:42 crc kubenswrapper[4907]: I0313 16:27:42.228176 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-sriov-openstack-openstack-cell1-sqxhp" Mar 13 16:27:42 crc kubenswrapper[4907]: I0313 16:27:42.228182 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="762fb1ddd2d8fd925d29b186df5f88568978b2b1ec0e68aa4f06e19c4e53f86e" Mar 13 16:27:42 crc kubenswrapper[4907]: I0313 16:27:42.373174 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-dhcp-openstack-openstack-cell1-jmbms"] Mar 13 16:27:42 crc kubenswrapper[4907]: E0313 16:27:42.376598 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05d384d7-9635-4d68-8eaf-48fea8bed0b8" containerName="neutron-sriov-openstack-openstack-cell1" Mar 13 16:27:42 crc kubenswrapper[4907]: I0313 16:27:42.376755 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="05d384d7-9635-4d68-8eaf-48fea8bed0b8" containerName="neutron-sriov-openstack-openstack-cell1" Mar 13 16:27:42 crc kubenswrapper[4907]: I0313 16:27:42.377149 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="05d384d7-9635-4d68-8eaf-48fea8bed0b8" containerName="neutron-sriov-openstack-openstack-cell1" Mar 13 16:27:42 crc kubenswrapper[4907]: I0313 16:27:42.378164 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dhcp-openstack-openstack-cell1-jmbms" Mar 13 16:27:42 crc kubenswrapper[4907]: I0313 16:27:42.381954 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Mar 13 16:27:42 crc kubenswrapper[4907]: I0313 16:27:42.382240 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-dhcp-agent-neutron-config" Mar 13 16:27:42 crc kubenswrapper[4907]: I0313 16:27:42.385435 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Mar 13 16:27:42 crc kubenswrapper[4907]: I0313 16:27:42.385440 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 13 16:27:42 crc kubenswrapper[4907]: I0313 16:27:42.385976 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-44kxw" Mar 13 16:27:42 crc kubenswrapper[4907]: I0313 16:27:42.395705 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-dhcp-openstack-openstack-cell1-jmbms"] Mar 13 16:27:42 crc kubenswrapper[4907]: I0313 16:27:42.464422 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5764492-ad4a-4e43-8a98-99ea7cfd9248-neutron-dhcp-combined-ca-bundle\") pod \"neutron-dhcp-openstack-openstack-cell1-jmbms\" (UID: \"b5764492-ad4a-4e43-8a98-99ea7cfd9248\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-jmbms" Mar 13 16:27:42 crc kubenswrapper[4907]: I0313 16:27:42.464640 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/b5764492-ad4a-4e43-8a98-99ea7cfd9248-ssh-key-openstack-cell1\") pod \"neutron-dhcp-openstack-openstack-cell1-jmbms\" (UID: \"b5764492-ad4a-4e43-8a98-99ea7cfd9248\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-jmbms" Mar 13 16:27:42 crc kubenswrapper[4907]: I0313 16:27:42.464846 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dw6mz\" (UniqueName: \"kubernetes.io/projected/b5764492-ad4a-4e43-8a98-99ea7cfd9248-kube-api-access-dw6mz\") pod \"neutron-dhcp-openstack-openstack-cell1-jmbms\" (UID: \"b5764492-ad4a-4e43-8a98-99ea7cfd9248\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-jmbms" Mar 13 16:27:42 crc kubenswrapper[4907]: I0313 16:27:42.465004 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b5764492-ad4a-4e43-8a98-99ea7cfd9248-inventory\") pod \"neutron-dhcp-openstack-openstack-cell1-jmbms\" (UID: \"b5764492-ad4a-4e43-8a98-99ea7cfd9248\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-jmbms" Mar 13 16:27:42 crc kubenswrapper[4907]: I0313 16:27:42.465146 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b5764492-ad4a-4e43-8a98-99ea7cfd9248-ceph\") pod \"neutron-dhcp-openstack-openstack-cell1-jmbms\" (UID: \"b5764492-ad4a-4e43-8a98-99ea7cfd9248\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-jmbms" Mar 13 16:27:42 crc kubenswrapper[4907]: I0313 16:27:42.465207 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/b5764492-ad4a-4e43-8a98-99ea7cfd9248-neutron-dhcp-agent-neutron-config-0\") pod \"neutron-dhcp-openstack-openstack-cell1-jmbms\" (UID: \"b5764492-ad4a-4e43-8a98-99ea7cfd9248\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-jmbms" Mar 13 16:27:42 crc kubenswrapper[4907]: I0313 16:27:42.566878 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5764492-ad4a-4e43-8a98-99ea7cfd9248-neutron-dhcp-combined-ca-bundle\") pod \"neutron-dhcp-openstack-openstack-cell1-jmbms\" (UID: \"b5764492-ad4a-4e43-8a98-99ea7cfd9248\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-jmbms" Mar 13 16:27:42 crc kubenswrapper[4907]: I0313 16:27:42.566997 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/b5764492-ad4a-4e43-8a98-99ea7cfd9248-ssh-key-openstack-cell1\") pod \"neutron-dhcp-openstack-openstack-cell1-jmbms\" (UID: \"b5764492-ad4a-4e43-8a98-99ea7cfd9248\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-jmbms" Mar 13 16:27:42 crc kubenswrapper[4907]: I0313 16:27:42.567042 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dw6mz\" (UniqueName: \"kubernetes.io/projected/b5764492-ad4a-4e43-8a98-99ea7cfd9248-kube-api-access-dw6mz\") pod \"neutron-dhcp-openstack-openstack-cell1-jmbms\" (UID: \"b5764492-ad4a-4e43-8a98-99ea7cfd9248\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-jmbms" Mar 13 16:27:42 crc kubenswrapper[4907]: I0313 16:27:42.567070 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b5764492-ad4a-4e43-8a98-99ea7cfd9248-inventory\") pod \"neutron-dhcp-openstack-openstack-cell1-jmbms\" (UID: \"b5764492-ad4a-4e43-8a98-99ea7cfd9248\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-jmbms" Mar 13 16:27:42 crc kubenswrapper[4907]: I0313 16:27:42.567124 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b5764492-ad4a-4e43-8a98-99ea7cfd9248-ceph\") pod \"neutron-dhcp-openstack-openstack-cell1-jmbms\" (UID: \"b5764492-ad4a-4e43-8a98-99ea7cfd9248\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-jmbms" Mar 13 16:27:42 crc kubenswrapper[4907]: I0313 16:27:42.567150 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/b5764492-ad4a-4e43-8a98-99ea7cfd9248-neutron-dhcp-agent-neutron-config-0\") pod \"neutron-dhcp-openstack-openstack-cell1-jmbms\" (UID: \"b5764492-ad4a-4e43-8a98-99ea7cfd9248\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-jmbms" Mar 13 16:27:42 crc kubenswrapper[4907]: I0313 16:27:42.572074 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/b5764492-ad4a-4e43-8a98-99ea7cfd9248-neutron-dhcp-agent-neutron-config-0\") pod \"neutron-dhcp-openstack-openstack-cell1-jmbms\" (UID: \"b5764492-ad4a-4e43-8a98-99ea7cfd9248\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-jmbms" Mar 13 16:27:42 crc kubenswrapper[4907]: I0313 16:27:42.572622 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b5764492-ad4a-4e43-8a98-99ea7cfd9248-ceph\") pod \"neutron-dhcp-openstack-openstack-cell1-jmbms\" (UID: \"b5764492-ad4a-4e43-8a98-99ea7cfd9248\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-jmbms" Mar 13 16:27:42 crc kubenswrapper[4907]: I0313 16:27:42.573628 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b5764492-ad4a-4e43-8a98-99ea7cfd9248-inventory\") pod \"neutron-dhcp-openstack-openstack-cell1-jmbms\" (UID: \"b5764492-ad4a-4e43-8a98-99ea7cfd9248\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-jmbms" Mar 13 16:27:42 crc kubenswrapper[4907]: I0313 16:27:42.579267 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/b5764492-ad4a-4e43-8a98-99ea7cfd9248-ssh-key-openstack-cell1\") pod \"neutron-dhcp-openstack-openstack-cell1-jmbms\" (UID: \"b5764492-ad4a-4e43-8a98-99ea7cfd9248\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-jmbms" Mar 13 16:27:42 crc kubenswrapper[4907]: I0313 16:27:42.580008 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5764492-ad4a-4e43-8a98-99ea7cfd9248-neutron-dhcp-combined-ca-bundle\") pod \"neutron-dhcp-openstack-openstack-cell1-jmbms\" (UID: \"b5764492-ad4a-4e43-8a98-99ea7cfd9248\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-jmbms" Mar 13 16:27:42 crc kubenswrapper[4907]: I0313 16:27:42.600514 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dw6mz\" (UniqueName: \"kubernetes.io/projected/b5764492-ad4a-4e43-8a98-99ea7cfd9248-kube-api-access-dw6mz\") pod \"neutron-dhcp-openstack-openstack-cell1-jmbms\" (UID: \"b5764492-ad4a-4e43-8a98-99ea7cfd9248\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-jmbms" Mar 13 16:27:42 crc kubenswrapper[4907]: I0313 16:27:42.710570 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dhcp-openstack-openstack-cell1-jmbms" Mar 13 16:27:43 crc kubenswrapper[4907]: I0313 16:27:43.362405 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-dhcp-openstack-openstack-cell1-jmbms"] Mar 13 16:27:43 crc kubenswrapper[4907]: W0313 16:27:43.365916 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb5764492_ad4a_4e43_8a98_99ea7cfd9248.slice/crio-ba830f46ef1e05ea0e668ef0942e0f24774e231102347a25077be9ae94ffda32 WatchSource:0}: Error finding container ba830f46ef1e05ea0e668ef0942e0f24774e231102347a25077be9ae94ffda32: Status 404 returned error can't find the container with id ba830f46ef1e05ea0e668ef0942e0f24774e231102347a25077be9ae94ffda32 Mar 13 16:27:44 crc kubenswrapper[4907]: I0313 16:27:44.256238 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dhcp-openstack-openstack-cell1-jmbms" event={"ID":"b5764492-ad4a-4e43-8a98-99ea7cfd9248","Type":"ContainerStarted","Data":"0e4b9902269aa35ede42d672cbbcad84fd72d75aa3d82966fa7bf7b01617b050"} Mar 13 16:27:44 crc kubenswrapper[4907]: I0313 16:27:44.256923 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dhcp-openstack-openstack-cell1-jmbms" event={"ID":"b5764492-ad4a-4e43-8a98-99ea7cfd9248","Type":"ContainerStarted","Data":"ba830f46ef1e05ea0e668ef0942e0f24774e231102347a25077be9ae94ffda32"} Mar 13 16:27:44 crc kubenswrapper[4907]: I0313 16:27:44.284023 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-dhcp-openstack-openstack-cell1-jmbms" podStartSLOduration=1.6693486210000001 podStartE2EDuration="2.284006223s" podCreationTimestamp="2026-03-13 16:27:42 +0000 UTC" firstStartedPulling="2026-03-13 16:27:43.369038892 +0000 UTC m=+8562.268826581" lastFinishedPulling="2026-03-13 16:27:43.983696464 +0000 UTC m=+8562.883484183" observedRunningTime="2026-03-13 16:27:44.279127119 +0000 UTC m=+8563.178914828" watchObservedRunningTime="2026-03-13 16:27:44.284006223 +0000 UTC m=+8563.183793912" Mar 13 16:27:45 crc kubenswrapper[4907]: I0313 16:27:45.782621 4907 scope.go:117] "RemoveContainer" containerID="dfb327c310937cae28a687e8c5c0256a2ab50e59a756669f2586112310e1dc69" Mar 13 16:27:45 crc kubenswrapper[4907]: E0313 16:27:45.783291 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:27:59 crc kubenswrapper[4907]: I0313 16:27:59.782847 4907 scope.go:117] "RemoveContainer" containerID="dfb327c310937cae28a687e8c5c0256a2ab50e59a756669f2586112310e1dc69" Mar 13 16:27:59 crc kubenswrapper[4907]: E0313 16:27:59.783870 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:28:00 crc kubenswrapper[4907]: I0313 16:28:00.156619 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556988-t5tgt"] Mar 13 16:28:00 crc kubenswrapper[4907]: I0313 16:28:00.158229 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556988-t5tgt" Mar 13 16:28:00 crc kubenswrapper[4907]: I0313 16:28:00.160404 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 16:28:00 crc kubenswrapper[4907]: I0313 16:28:00.161774 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 16:28:00 crc kubenswrapper[4907]: I0313 16:28:00.162012 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 16:28:00 crc kubenswrapper[4907]: I0313 16:28:00.183126 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556988-t5tgt"] Mar 13 16:28:00 crc kubenswrapper[4907]: I0313 16:28:00.292921 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2p6nd\" (UniqueName: \"kubernetes.io/projected/6aa9f963-9a2b-4101-88e5-9553783a9963-kube-api-access-2p6nd\") pod \"auto-csr-approver-29556988-t5tgt\" (UID: \"6aa9f963-9a2b-4101-88e5-9553783a9963\") " pod="openshift-infra/auto-csr-approver-29556988-t5tgt" Mar 13 16:28:00 crc kubenswrapper[4907]: I0313 16:28:00.394628 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2p6nd\" (UniqueName: \"kubernetes.io/projected/6aa9f963-9a2b-4101-88e5-9553783a9963-kube-api-access-2p6nd\") pod \"auto-csr-approver-29556988-t5tgt\" (UID: \"6aa9f963-9a2b-4101-88e5-9553783a9963\") " pod="openshift-infra/auto-csr-approver-29556988-t5tgt" Mar 13 16:28:00 crc kubenswrapper[4907]: I0313 16:28:00.425552 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2p6nd\" (UniqueName: \"kubernetes.io/projected/6aa9f963-9a2b-4101-88e5-9553783a9963-kube-api-access-2p6nd\") pod \"auto-csr-approver-29556988-t5tgt\" (UID: \"6aa9f963-9a2b-4101-88e5-9553783a9963\") " pod="openshift-infra/auto-csr-approver-29556988-t5tgt" Mar 13 16:28:00 crc kubenswrapper[4907]: I0313 16:28:00.485534 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556988-t5tgt" Mar 13 16:28:00 crc kubenswrapper[4907]: I0313 16:28:00.971090 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556988-t5tgt"] Mar 13 16:28:00 crc kubenswrapper[4907]: I0313 16:28:00.985565 4907 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 13 16:28:01 crc kubenswrapper[4907]: I0313 16:28:01.478195 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556988-t5tgt" event={"ID":"6aa9f963-9a2b-4101-88e5-9553783a9963","Type":"ContainerStarted","Data":"f9ca69cff36c9453cd1eb3650007fe40309486ddb121cbafcc1bfcc8520c31ba"} Mar 13 16:28:02 crc kubenswrapper[4907]: I0313 16:28:02.490401 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556988-t5tgt" event={"ID":"6aa9f963-9a2b-4101-88e5-9553783a9963","Type":"ContainerStarted","Data":"68b2fa8f8bf43cf605cf2bff20ff9f2bc9eb89c3d74246f3836b57e3ae2b5d1c"} Mar 13 16:28:02 crc kubenswrapper[4907]: I0313 16:28:02.514834 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29556988-t5tgt" podStartSLOduration=1.494269263 podStartE2EDuration="2.514815854s" podCreationTimestamp="2026-03-13 16:28:00 +0000 UTC" firstStartedPulling="2026-03-13 16:28:00.984854603 +0000 UTC m=+8579.884642332" lastFinishedPulling="2026-03-13 16:28:02.005401234 +0000 UTC m=+8580.905188923" observedRunningTime="2026-03-13 16:28:02.505434708 +0000 UTC m=+8581.405222397" watchObservedRunningTime="2026-03-13 16:28:02.514815854 +0000 UTC m=+8581.414603543" Mar 13 16:28:03 crc kubenswrapper[4907]: I0313 16:28:03.506479 4907 generic.go:334] "Generic (PLEG): container finished" podID="6aa9f963-9a2b-4101-88e5-9553783a9963" containerID="68b2fa8f8bf43cf605cf2bff20ff9f2bc9eb89c3d74246f3836b57e3ae2b5d1c" exitCode=0 Mar 13 16:28:03 crc kubenswrapper[4907]: I0313 16:28:03.506573 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556988-t5tgt" event={"ID":"6aa9f963-9a2b-4101-88e5-9553783a9963","Type":"ContainerDied","Data":"68b2fa8f8bf43cf605cf2bff20ff9f2bc9eb89c3d74246f3836b57e3ae2b5d1c"} Mar 13 16:28:04 crc kubenswrapper[4907]: I0313 16:28:04.956299 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556988-t5tgt" Mar 13 16:28:05 crc kubenswrapper[4907]: I0313 16:28:05.028865 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2p6nd\" (UniqueName: \"kubernetes.io/projected/6aa9f963-9a2b-4101-88e5-9553783a9963-kube-api-access-2p6nd\") pod \"6aa9f963-9a2b-4101-88e5-9553783a9963\" (UID: \"6aa9f963-9a2b-4101-88e5-9553783a9963\") " Mar 13 16:28:05 crc kubenswrapper[4907]: I0313 16:28:05.051178 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6aa9f963-9a2b-4101-88e5-9553783a9963-kube-api-access-2p6nd" (OuterVolumeSpecName: "kube-api-access-2p6nd") pod "6aa9f963-9a2b-4101-88e5-9553783a9963" (UID: "6aa9f963-9a2b-4101-88e5-9553783a9963"). InnerVolumeSpecName "kube-api-access-2p6nd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:28:05 crc kubenswrapper[4907]: I0313 16:28:05.131116 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2p6nd\" (UniqueName: \"kubernetes.io/projected/6aa9f963-9a2b-4101-88e5-9553783a9963-kube-api-access-2p6nd\") on node \"crc\" DevicePath \"\"" Mar 13 16:28:05 crc kubenswrapper[4907]: I0313 16:28:05.530816 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556988-t5tgt" event={"ID":"6aa9f963-9a2b-4101-88e5-9553783a9963","Type":"ContainerDied","Data":"f9ca69cff36c9453cd1eb3650007fe40309486ddb121cbafcc1bfcc8520c31ba"} Mar 13 16:28:05 crc kubenswrapper[4907]: I0313 16:28:05.530852 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f9ca69cff36c9453cd1eb3650007fe40309486ddb121cbafcc1bfcc8520c31ba" Mar 13 16:28:05 crc kubenswrapper[4907]: I0313 16:28:05.530938 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556988-t5tgt" Mar 13 16:28:06 crc kubenswrapper[4907]: I0313 16:28:06.048426 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556982-t6wrd"] Mar 13 16:28:06 crc kubenswrapper[4907]: I0313 16:28:06.062301 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556982-t6wrd"] Mar 13 16:28:07 crc kubenswrapper[4907]: I0313 16:28:07.806026 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2364a186-f894-44a5-9c69-101cbfa14a82" path="/var/lib/kubelet/pods/2364a186-f894-44a5-9c69-101cbfa14a82/volumes" Mar 13 16:28:13 crc kubenswrapper[4907]: I0313 16:28:13.783290 4907 scope.go:117] "RemoveContainer" containerID="dfb327c310937cae28a687e8c5c0256a2ab50e59a756669f2586112310e1dc69" Mar 13 16:28:13 crc kubenswrapper[4907]: E0313 16:28:13.784541 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:28:26 crc kubenswrapper[4907]: I0313 16:28:26.782381 4907 scope.go:117] "RemoveContainer" containerID="dfb327c310937cae28a687e8c5c0256a2ab50e59a756669f2586112310e1dc69" Mar 13 16:28:26 crc kubenswrapper[4907]: E0313 16:28:26.783289 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:28:33 crc kubenswrapper[4907]: I0313 16:28:33.818157 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-96kxt"] Mar 13 16:28:33 crc kubenswrapper[4907]: E0313 16:28:33.819026 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6aa9f963-9a2b-4101-88e5-9553783a9963" containerName="oc" Mar 13 16:28:33 crc kubenswrapper[4907]: I0313 16:28:33.819038 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="6aa9f963-9a2b-4101-88e5-9553783a9963" containerName="oc" Mar 13 16:28:33 crc kubenswrapper[4907]: I0313 16:28:33.819235 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="6aa9f963-9a2b-4101-88e5-9553783a9963" containerName="oc" Mar 13 16:28:33 crc kubenswrapper[4907]: I0313 16:28:33.820700 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-96kxt" Mar 13 16:28:33 crc kubenswrapper[4907]: I0313 16:28:33.839365 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-96kxt"] Mar 13 16:28:33 crc kubenswrapper[4907]: I0313 16:28:33.958948 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1ba75cd1-2e4a-4695-b33c-8ed0cbbbb748-catalog-content\") pod \"redhat-marketplace-96kxt\" (UID: \"1ba75cd1-2e4a-4695-b33c-8ed0cbbbb748\") " pod="openshift-marketplace/redhat-marketplace-96kxt" Mar 13 16:28:33 crc kubenswrapper[4907]: I0313 16:28:33.958992 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1ba75cd1-2e4a-4695-b33c-8ed0cbbbb748-utilities\") pod \"redhat-marketplace-96kxt\" (UID: \"1ba75cd1-2e4a-4695-b33c-8ed0cbbbb748\") " pod="openshift-marketplace/redhat-marketplace-96kxt" Mar 13 16:28:33 crc kubenswrapper[4907]: I0313 16:28:33.959376 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zg9rp\" (UniqueName: \"kubernetes.io/projected/1ba75cd1-2e4a-4695-b33c-8ed0cbbbb748-kube-api-access-zg9rp\") pod \"redhat-marketplace-96kxt\" (UID: \"1ba75cd1-2e4a-4695-b33c-8ed0cbbbb748\") " pod="openshift-marketplace/redhat-marketplace-96kxt" Mar 13 16:28:34 crc kubenswrapper[4907]: I0313 16:28:34.061445 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1ba75cd1-2e4a-4695-b33c-8ed0cbbbb748-catalog-content\") pod \"redhat-marketplace-96kxt\" (UID: \"1ba75cd1-2e4a-4695-b33c-8ed0cbbbb748\") " pod="openshift-marketplace/redhat-marketplace-96kxt" Mar 13 16:28:34 crc kubenswrapper[4907]: I0313 16:28:34.061537 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1ba75cd1-2e4a-4695-b33c-8ed0cbbbb748-utilities\") pod \"redhat-marketplace-96kxt\" (UID: \"1ba75cd1-2e4a-4695-b33c-8ed0cbbbb748\") " pod="openshift-marketplace/redhat-marketplace-96kxt" Mar 13 16:28:34 crc kubenswrapper[4907]: I0313 16:28:34.061752 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zg9rp\" (UniqueName: \"kubernetes.io/projected/1ba75cd1-2e4a-4695-b33c-8ed0cbbbb748-kube-api-access-zg9rp\") pod \"redhat-marketplace-96kxt\" (UID: \"1ba75cd1-2e4a-4695-b33c-8ed0cbbbb748\") " pod="openshift-marketplace/redhat-marketplace-96kxt" Mar 13 16:28:34 crc kubenswrapper[4907]: I0313 16:28:34.062023 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1ba75cd1-2e4a-4695-b33c-8ed0cbbbb748-catalog-content\") pod \"redhat-marketplace-96kxt\" (UID: \"1ba75cd1-2e4a-4695-b33c-8ed0cbbbb748\") " pod="openshift-marketplace/redhat-marketplace-96kxt" Mar 13 16:28:34 crc kubenswrapper[4907]: I0313 16:28:34.062210 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1ba75cd1-2e4a-4695-b33c-8ed0cbbbb748-utilities\") pod \"redhat-marketplace-96kxt\" (UID: \"1ba75cd1-2e4a-4695-b33c-8ed0cbbbb748\") " pod="openshift-marketplace/redhat-marketplace-96kxt" Mar 13 16:28:34 crc kubenswrapper[4907]: I0313 16:28:34.089147 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zg9rp\" (UniqueName: \"kubernetes.io/projected/1ba75cd1-2e4a-4695-b33c-8ed0cbbbb748-kube-api-access-zg9rp\") pod \"redhat-marketplace-96kxt\" (UID: \"1ba75cd1-2e4a-4695-b33c-8ed0cbbbb748\") " pod="openshift-marketplace/redhat-marketplace-96kxt" Mar 13 16:28:34 crc kubenswrapper[4907]: I0313 16:28:34.139440 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-96kxt" Mar 13 16:28:34 crc kubenswrapper[4907]: I0313 16:28:34.692325 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-96kxt"] Mar 13 16:28:34 crc kubenswrapper[4907]: I0313 16:28:34.898194 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-96kxt" event={"ID":"1ba75cd1-2e4a-4695-b33c-8ed0cbbbb748","Type":"ContainerStarted","Data":"0dafac436ece1b2ddb45a68c7f74b1a3663a57e9df7fd821ecccf3252eb8e180"} Mar 13 16:28:35 crc kubenswrapper[4907]: I0313 16:28:35.907522 4907 generic.go:334] "Generic (PLEG): container finished" podID="1ba75cd1-2e4a-4695-b33c-8ed0cbbbb748" containerID="fc7c8ee006b2576371976490d6643daa901bc4a5c44c8411005168ae914f4170" exitCode=0 Mar 13 16:28:35 crc kubenswrapper[4907]: I0313 16:28:35.907734 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-96kxt" event={"ID":"1ba75cd1-2e4a-4695-b33c-8ed0cbbbb748","Type":"ContainerDied","Data":"fc7c8ee006b2576371976490d6643daa901bc4a5c44c8411005168ae914f4170"} Mar 13 16:28:37 crc kubenswrapper[4907]: I0313 16:28:37.936019 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-96kxt" event={"ID":"1ba75cd1-2e4a-4695-b33c-8ed0cbbbb748","Type":"ContainerStarted","Data":"9c0f31afbd3f516e339a816053788dfb8af39539535629431e6e575f2db70864"} Mar 13 16:28:38 crc kubenswrapper[4907]: I0313 16:28:38.949452 4907 generic.go:334] "Generic (PLEG): container finished" podID="1ba75cd1-2e4a-4695-b33c-8ed0cbbbb748" containerID="9c0f31afbd3f516e339a816053788dfb8af39539535629431e6e575f2db70864" exitCode=0 Mar 13 16:28:38 crc kubenswrapper[4907]: I0313 16:28:38.949551 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-96kxt" event={"ID":"1ba75cd1-2e4a-4695-b33c-8ed0cbbbb748","Type":"ContainerDied","Data":"9c0f31afbd3f516e339a816053788dfb8af39539535629431e6e575f2db70864"} Mar 13 16:28:40 crc kubenswrapper[4907]: I0313 16:28:40.783311 4907 scope.go:117] "RemoveContainer" containerID="dfb327c310937cae28a687e8c5c0256a2ab50e59a756669f2586112310e1dc69" Mar 13 16:28:40 crc kubenswrapper[4907]: E0313 16:28:40.784109 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:28:41 crc kubenswrapper[4907]: I0313 16:28:41.011778 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-96kxt" event={"ID":"1ba75cd1-2e4a-4695-b33c-8ed0cbbbb748","Type":"ContainerStarted","Data":"15dbe6019db9afdf340463b89d9f2149f147e6017e7c5bec357846162e1071e6"} Mar 13 16:28:41 crc kubenswrapper[4907]: I0313 16:28:41.047570 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-96kxt" podStartSLOduration=4.519417311 podStartE2EDuration="8.047537196s" podCreationTimestamp="2026-03-13 16:28:33 +0000 UTC" firstStartedPulling="2026-03-13 16:28:35.909557751 +0000 UTC m=+8614.809345440" lastFinishedPulling="2026-03-13 16:28:39.437677626 +0000 UTC m=+8618.337465325" observedRunningTime="2026-03-13 16:28:41.039721745 +0000 UTC m=+8619.939509474" watchObservedRunningTime="2026-03-13 16:28:41.047537196 +0000 UTC m=+8619.947324925" Mar 13 16:28:42 crc kubenswrapper[4907]: I0313 16:28:42.781020 4907 scope.go:117] "RemoveContainer" containerID="48a114e32689f3100e68121daa48f2b4ba681c9f23fd2dfae93832f7ab1cf591" Mar 13 16:28:44 crc kubenswrapper[4907]: I0313 16:28:44.140115 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-96kxt" Mar 13 16:28:44 crc kubenswrapper[4907]: I0313 16:28:44.140545 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-96kxt" Mar 13 16:28:44 crc kubenswrapper[4907]: I0313 16:28:44.204611 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-96kxt" Mar 13 16:28:45 crc kubenswrapper[4907]: I0313 16:28:45.117603 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-96kxt" Mar 13 16:28:45 crc kubenswrapper[4907]: I0313 16:28:45.176589 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-96kxt"] Mar 13 16:28:47 crc kubenswrapper[4907]: I0313 16:28:47.081225 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-96kxt" podUID="1ba75cd1-2e4a-4695-b33c-8ed0cbbbb748" containerName="registry-server" containerID="cri-o://15dbe6019db9afdf340463b89d9f2149f147e6017e7c5bec357846162e1071e6" gracePeriod=2 Mar 13 16:28:47 crc kubenswrapper[4907]: I0313 16:28:47.603759 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-96kxt" Mar 13 16:28:47 crc kubenswrapper[4907]: I0313 16:28:47.623621 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zg9rp\" (UniqueName: \"kubernetes.io/projected/1ba75cd1-2e4a-4695-b33c-8ed0cbbbb748-kube-api-access-zg9rp\") pod \"1ba75cd1-2e4a-4695-b33c-8ed0cbbbb748\" (UID: \"1ba75cd1-2e4a-4695-b33c-8ed0cbbbb748\") " Mar 13 16:28:47 crc kubenswrapper[4907]: I0313 16:28:47.623687 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1ba75cd1-2e4a-4695-b33c-8ed0cbbbb748-utilities\") pod \"1ba75cd1-2e4a-4695-b33c-8ed0cbbbb748\" (UID: \"1ba75cd1-2e4a-4695-b33c-8ed0cbbbb748\") " Mar 13 16:28:47 crc kubenswrapper[4907]: I0313 16:28:47.623733 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1ba75cd1-2e4a-4695-b33c-8ed0cbbbb748-catalog-content\") pod \"1ba75cd1-2e4a-4695-b33c-8ed0cbbbb748\" (UID: \"1ba75cd1-2e4a-4695-b33c-8ed0cbbbb748\") " Mar 13 16:28:47 crc kubenswrapper[4907]: I0313 16:28:47.624738 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1ba75cd1-2e4a-4695-b33c-8ed0cbbbb748-utilities" (OuterVolumeSpecName: "utilities") pod "1ba75cd1-2e4a-4695-b33c-8ed0cbbbb748" (UID: "1ba75cd1-2e4a-4695-b33c-8ed0cbbbb748"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 16:28:47 crc kubenswrapper[4907]: I0313 16:28:47.633620 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1ba75cd1-2e4a-4695-b33c-8ed0cbbbb748-kube-api-access-zg9rp" (OuterVolumeSpecName: "kube-api-access-zg9rp") pod "1ba75cd1-2e4a-4695-b33c-8ed0cbbbb748" (UID: "1ba75cd1-2e4a-4695-b33c-8ed0cbbbb748"). InnerVolumeSpecName "kube-api-access-zg9rp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:28:47 crc kubenswrapper[4907]: I0313 16:28:47.668248 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1ba75cd1-2e4a-4695-b33c-8ed0cbbbb748-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1ba75cd1-2e4a-4695-b33c-8ed0cbbbb748" (UID: "1ba75cd1-2e4a-4695-b33c-8ed0cbbbb748"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 16:28:47 crc kubenswrapper[4907]: I0313 16:28:47.726291 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zg9rp\" (UniqueName: \"kubernetes.io/projected/1ba75cd1-2e4a-4695-b33c-8ed0cbbbb748-kube-api-access-zg9rp\") on node \"crc\" DevicePath \"\"" Mar 13 16:28:47 crc kubenswrapper[4907]: I0313 16:28:47.726345 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1ba75cd1-2e4a-4695-b33c-8ed0cbbbb748-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 16:28:47 crc kubenswrapper[4907]: I0313 16:28:47.726366 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1ba75cd1-2e4a-4695-b33c-8ed0cbbbb748-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 16:28:48 crc kubenswrapper[4907]: I0313 16:28:48.127088 4907 generic.go:334] "Generic (PLEG): container finished" podID="1ba75cd1-2e4a-4695-b33c-8ed0cbbbb748" containerID="15dbe6019db9afdf340463b89d9f2149f147e6017e7c5bec357846162e1071e6" exitCode=0 Mar 13 16:28:48 crc kubenswrapper[4907]: I0313 16:28:48.127459 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-96kxt" Mar 13 16:28:48 crc kubenswrapper[4907]: I0313 16:28:48.127544 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-96kxt" event={"ID":"1ba75cd1-2e4a-4695-b33c-8ed0cbbbb748","Type":"ContainerDied","Data":"15dbe6019db9afdf340463b89d9f2149f147e6017e7c5bec357846162e1071e6"} Mar 13 16:28:48 crc kubenswrapper[4907]: I0313 16:28:48.127579 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-96kxt" event={"ID":"1ba75cd1-2e4a-4695-b33c-8ed0cbbbb748","Type":"ContainerDied","Data":"0dafac436ece1b2ddb45a68c7f74b1a3663a57e9df7fd821ecccf3252eb8e180"} Mar 13 16:28:48 crc kubenswrapper[4907]: I0313 16:28:48.127601 4907 scope.go:117] "RemoveContainer" containerID="15dbe6019db9afdf340463b89d9f2149f147e6017e7c5bec357846162e1071e6" Mar 13 16:28:48 crc kubenswrapper[4907]: I0313 16:28:48.164456 4907 scope.go:117] "RemoveContainer" containerID="9c0f31afbd3f516e339a816053788dfb8af39539535629431e6e575f2db70864" Mar 13 16:28:48 crc kubenswrapper[4907]: I0313 16:28:48.170028 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-96kxt"] Mar 13 16:28:48 crc kubenswrapper[4907]: I0313 16:28:48.178851 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-96kxt"] Mar 13 16:28:48 crc kubenswrapper[4907]: I0313 16:28:48.197461 4907 scope.go:117] "RemoveContainer" containerID="fc7c8ee006b2576371976490d6643daa901bc4a5c44c8411005168ae914f4170" Mar 13 16:28:48 crc kubenswrapper[4907]: I0313 16:28:48.239390 4907 scope.go:117] "RemoveContainer" containerID="15dbe6019db9afdf340463b89d9f2149f147e6017e7c5bec357846162e1071e6" Mar 13 16:28:48 crc kubenswrapper[4907]: E0313 16:28:48.239904 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"15dbe6019db9afdf340463b89d9f2149f147e6017e7c5bec357846162e1071e6\": container with ID starting with 15dbe6019db9afdf340463b89d9f2149f147e6017e7c5bec357846162e1071e6 not found: ID does not exist" containerID="15dbe6019db9afdf340463b89d9f2149f147e6017e7c5bec357846162e1071e6" Mar 13 16:28:48 crc kubenswrapper[4907]: I0313 16:28:48.239989 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"15dbe6019db9afdf340463b89d9f2149f147e6017e7c5bec357846162e1071e6"} err="failed to get container status \"15dbe6019db9afdf340463b89d9f2149f147e6017e7c5bec357846162e1071e6\": rpc error: code = NotFound desc = could not find container \"15dbe6019db9afdf340463b89d9f2149f147e6017e7c5bec357846162e1071e6\": container with ID starting with 15dbe6019db9afdf340463b89d9f2149f147e6017e7c5bec357846162e1071e6 not found: ID does not exist" Mar 13 16:28:48 crc kubenswrapper[4907]: I0313 16:28:48.240021 4907 scope.go:117] "RemoveContainer" containerID="9c0f31afbd3f516e339a816053788dfb8af39539535629431e6e575f2db70864" Mar 13 16:28:48 crc kubenswrapper[4907]: E0313 16:28:48.240543 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9c0f31afbd3f516e339a816053788dfb8af39539535629431e6e575f2db70864\": container with ID starting with 9c0f31afbd3f516e339a816053788dfb8af39539535629431e6e575f2db70864 not found: ID does not exist" containerID="9c0f31afbd3f516e339a816053788dfb8af39539535629431e6e575f2db70864" Mar 13 16:28:48 crc kubenswrapper[4907]: I0313 16:28:48.240635 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9c0f31afbd3f516e339a816053788dfb8af39539535629431e6e575f2db70864"} err="failed to get container status \"9c0f31afbd3f516e339a816053788dfb8af39539535629431e6e575f2db70864\": rpc error: code = NotFound desc = could not find container \"9c0f31afbd3f516e339a816053788dfb8af39539535629431e6e575f2db70864\": container with ID starting with 9c0f31afbd3f516e339a816053788dfb8af39539535629431e6e575f2db70864 not found: ID does not exist" Mar 13 16:28:48 crc kubenswrapper[4907]: I0313 16:28:48.240669 4907 scope.go:117] "RemoveContainer" containerID="fc7c8ee006b2576371976490d6643daa901bc4a5c44c8411005168ae914f4170" Mar 13 16:28:48 crc kubenswrapper[4907]: E0313 16:28:48.241204 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fc7c8ee006b2576371976490d6643daa901bc4a5c44c8411005168ae914f4170\": container with ID starting with fc7c8ee006b2576371976490d6643daa901bc4a5c44c8411005168ae914f4170 not found: ID does not exist" containerID="fc7c8ee006b2576371976490d6643daa901bc4a5c44c8411005168ae914f4170" Mar 13 16:28:48 crc kubenswrapper[4907]: I0313 16:28:48.241251 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc7c8ee006b2576371976490d6643daa901bc4a5c44c8411005168ae914f4170"} err="failed to get container status \"fc7c8ee006b2576371976490d6643daa901bc4a5c44c8411005168ae914f4170\": rpc error: code = NotFound desc = could not find container \"fc7c8ee006b2576371976490d6643daa901bc4a5c44c8411005168ae914f4170\": container with ID starting with fc7c8ee006b2576371976490d6643daa901bc4a5c44c8411005168ae914f4170 not found: ID does not exist" Mar 13 16:28:49 crc kubenswrapper[4907]: I0313 16:28:49.794153 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1ba75cd1-2e4a-4695-b33c-8ed0cbbbb748" path="/var/lib/kubelet/pods/1ba75cd1-2e4a-4695-b33c-8ed0cbbbb748/volumes" Mar 13 16:28:51 crc kubenswrapper[4907]: I0313 16:28:51.796631 4907 scope.go:117] "RemoveContainer" containerID="dfb327c310937cae28a687e8c5c0256a2ab50e59a756669f2586112310e1dc69" Mar 13 16:28:51 crc kubenswrapper[4907]: E0313 16:28:51.797665 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:29:01 crc kubenswrapper[4907]: I0313 16:29:01.303450 4907 generic.go:334] "Generic (PLEG): container finished" podID="b5764492-ad4a-4e43-8a98-99ea7cfd9248" containerID="0e4b9902269aa35ede42d672cbbcad84fd72d75aa3d82966fa7bf7b01617b050" exitCode=0 Mar 13 16:29:01 crc kubenswrapper[4907]: I0313 16:29:01.303657 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dhcp-openstack-openstack-cell1-jmbms" event={"ID":"b5764492-ad4a-4e43-8a98-99ea7cfd9248","Type":"ContainerDied","Data":"0e4b9902269aa35ede42d672cbbcad84fd72d75aa3d82966fa7bf7b01617b050"} Mar 13 16:29:02 crc kubenswrapper[4907]: I0313 16:29:02.783854 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dhcp-openstack-openstack-cell1-jmbms" Mar 13 16:29:02 crc kubenswrapper[4907]: I0313 16:29:02.887724 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b5764492-ad4a-4e43-8a98-99ea7cfd9248-ceph\") pod \"b5764492-ad4a-4e43-8a98-99ea7cfd9248\" (UID: \"b5764492-ad4a-4e43-8a98-99ea7cfd9248\") " Mar 13 16:29:02 crc kubenswrapper[4907]: I0313 16:29:02.888133 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b5764492-ad4a-4e43-8a98-99ea7cfd9248-inventory\") pod \"b5764492-ad4a-4e43-8a98-99ea7cfd9248\" (UID: \"b5764492-ad4a-4e43-8a98-99ea7cfd9248\") " Mar 13 16:29:02 crc kubenswrapper[4907]: I0313 16:29:02.888222 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dw6mz\" (UniqueName: \"kubernetes.io/projected/b5764492-ad4a-4e43-8a98-99ea7cfd9248-kube-api-access-dw6mz\") pod \"b5764492-ad4a-4e43-8a98-99ea7cfd9248\" (UID: \"b5764492-ad4a-4e43-8a98-99ea7cfd9248\") " Mar 13 16:29:02 crc kubenswrapper[4907]: I0313 16:29:02.888315 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/b5764492-ad4a-4e43-8a98-99ea7cfd9248-neutron-dhcp-agent-neutron-config-0\") pod \"b5764492-ad4a-4e43-8a98-99ea7cfd9248\" (UID: \"b5764492-ad4a-4e43-8a98-99ea7cfd9248\") " Mar 13 16:29:02 crc kubenswrapper[4907]: I0313 16:29:02.888358 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5764492-ad4a-4e43-8a98-99ea7cfd9248-neutron-dhcp-combined-ca-bundle\") pod \"b5764492-ad4a-4e43-8a98-99ea7cfd9248\" (UID: \"b5764492-ad4a-4e43-8a98-99ea7cfd9248\") " Mar 13 16:29:02 crc kubenswrapper[4907]: I0313 16:29:02.888394 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/b5764492-ad4a-4e43-8a98-99ea7cfd9248-ssh-key-openstack-cell1\") pod \"b5764492-ad4a-4e43-8a98-99ea7cfd9248\" (UID: \"b5764492-ad4a-4e43-8a98-99ea7cfd9248\") " Mar 13 16:29:02 crc kubenswrapper[4907]: I0313 16:29:02.893758 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b5764492-ad4a-4e43-8a98-99ea7cfd9248-kube-api-access-dw6mz" (OuterVolumeSpecName: "kube-api-access-dw6mz") pod "b5764492-ad4a-4e43-8a98-99ea7cfd9248" (UID: "b5764492-ad4a-4e43-8a98-99ea7cfd9248"). InnerVolumeSpecName "kube-api-access-dw6mz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:29:02 crc kubenswrapper[4907]: I0313 16:29:02.894047 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5764492-ad4a-4e43-8a98-99ea7cfd9248-neutron-dhcp-combined-ca-bundle" (OuterVolumeSpecName: "neutron-dhcp-combined-ca-bundle") pod "b5764492-ad4a-4e43-8a98-99ea7cfd9248" (UID: "b5764492-ad4a-4e43-8a98-99ea7cfd9248"). InnerVolumeSpecName "neutron-dhcp-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:29:02 crc kubenswrapper[4907]: I0313 16:29:02.899748 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5764492-ad4a-4e43-8a98-99ea7cfd9248-ceph" (OuterVolumeSpecName: "ceph") pod "b5764492-ad4a-4e43-8a98-99ea7cfd9248" (UID: "b5764492-ad4a-4e43-8a98-99ea7cfd9248"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:29:02 crc kubenswrapper[4907]: I0313 16:29:02.916717 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5764492-ad4a-4e43-8a98-99ea7cfd9248-inventory" (OuterVolumeSpecName: "inventory") pod "b5764492-ad4a-4e43-8a98-99ea7cfd9248" (UID: "b5764492-ad4a-4e43-8a98-99ea7cfd9248"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:29:02 crc kubenswrapper[4907]: I0313 16:29:02.929066 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5764492-ad4a-4e43-8a98-99ea7cfd9248-ssh-key-openstack-cell1" (OuterVolumeSpecName: "ssh-key-openstack-cell1") pod "b5764492-ad4a-4e43-8a98-99ea7cfd9248" (UID: "b5764492-ad4a-4e43-8a98-99ea7cfd9248"). InnerVolumeSpecName "ssh-key-openstack-cell1". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:29:02 crc kubenswrapper[4907]: I0313 16:29:02.940341 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5764492-ad4a-4e43-8a98-99ea7cfd9248-neutron-dhcp-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-dhcp-agent-neutron-config-0") pod "b5764492-ad4a-4e43-8a98-99ea7cfd9248" (UID: "b5764492-ad4a-4e43-8a98-99ea7cfd9248"). InnerVolumeSpecName "neutron-dhcp-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:29:02 crc kubenswrapper[4907]: I0313 16:29:02.990779 4907 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b5764492-ad4a-4e43-8a98-99ea7cfd9248-ceph\") on node \"crc\" DevicePath \"\"" Mar 13 16:29:02 crc kubenswrapper[4907]: I0313 16:29:02.990935 4907 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b5764492-ad4a-4e43-8a98-99ea7cfd9248-inventory\") on node \"crc\" DevicePath \"\"" Mar 13 16:29:02 crc kubenswrapper[4907]: I0313 16:29:02.990999 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dw6mz\" (UniqueName: \"kubernetes.io/projected/b5764492-ad4a-4e43-8a98-99ea7cfd9248-kube-api-access-dw6mz\") on node \"crc\" DevicePath \"\"" Mar 13 16:29:02 crc kubenswrapper[4907]: I0313 16:29:02.991068 4907 reconciler_common.go:293] "Volume detached for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/b5764492-ad4a-4e43-8a98-99ea7cfd9248-neutron-dhcp-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Mar 13 16:29:02 crc kubenswrapper[4907]: I0313 16:29:02.991129 4907 reconciler_common.go:293] "Volume detached for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5764492-ad4a-4e43-8a98-99ea7cfd9248-neutron-dhcp-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 16:29:02 crc kubenswrapper[4907]: I0313 16:29:02.991186 4907 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/b5764492-ad4a-4e43-8a98-99ea7cfd9248-ssh-key-openstack-cell1\") on node \"crc\" DevicePath \"\"" Mar 13 16:29:03 crc kubenswrapper[4907]: I0313 16:29:03.335975 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dhcp-openstack-openstack-cell1-jmbms" event={"ID":"b5764492-ad4a-4e43-8a98-99ea7cfd9248","Type":"ContainerDied","Data":"ba830f46ef1e05ea0e668ef0942e0f24774e231102347a25077be9ae94ffda32"} Mar 13 16:29:03 crc kubenswrapper[4907]: I0313 16:29:03.336333 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ba830f46ef1e05ea0e668ef0942e0f24774e231102347a25077be9ae94ffda32" Mar 13 16:29:03 crc kubenswrapper[4907]: I0313 16:29:03.336092 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dhcp-openstack-openstack-cell1-jmbms" Mar 13 16:29:06 crc kubenswrapper[4907]: I0313 16:29:06.783278 4907 scope.go:117] "RemoveContainer" containerID="dfb327c310937cae28a687e8c5c0256a2ab50e59a756669f2586112310e1dc69" Mar 13 16:29:06 crc kubenswrapper[4907]: E0313 16:29:06.784186 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:29:21 crc kubenswrapper[4907]: I0313 16:29:21.795572 4907 scope.go:117] "RemoveContainer" containerID="dfb327c310937cae28a687e8c5c0256a2ab50e59a756669f2586112310e1dc69" Mar 13 16:29:21 crc kubenswrapper[4907]: E0313 16:29:21.796791 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:29:26 crc kubenswrapper[4907]: I0313 16:29:26.506181 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Mar 13 16:29:26 crc kubenswrapper[4907]: I0313 16:29:26.506785 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell0-conductor-0" podUID="647720e3-d21e-4f41-aeb8-d4e8f092cbb5" containerName="nova-cell0-conductor-conductor" containerID="cri-o://c20788b6b1aecba8b32acff82d1977637a4b4a170211b700ca1e5ce58e45cffe" gracePeriod=30 Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.041898 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.042111 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-conductor-0" podUID="eded65eb-7a04-4f90-ad82-be65db9014eb" containerName="nova-cell1-conductor-conductor" containerID="cri-o://e6d65819607c1b5f8189419bb95b4c9b19ca06cb630c59aa03a5474fcce9aa3d" gracePeriod=30 Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.195818 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.196343 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="48adeb28-f850-4e7c-a546-2c0b32d1b16c" containerName="nova-scheduler-scheduler" containerID="cri-o://5de9510392b1c0241095d60bc966293a382f722cf2d981565b650d57c091ce88" gracePeriod=30 Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.216861 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.217145 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="b98ee93f-f32c-4199-8045-62ab95761a6a" containerName="nova-api-log" containerID="cri-o://fa43cc779c11f32664399bd6d5fb18c420f2b54971c3216eb69f8689384ae861" gracePeriod=30 Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.217219 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="b98ee93f-f32c-4199-8045-62ab95761a6a" containerName="nova-api-api" containerID="cri-o://19d030a6fb6fc190b4d45542a5b426271778e2fd40eba0953eec528508e6f63a" gracePeriod=30 Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.231168 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.231502 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="07f475b7-1542-4689-b3c1-f7f4470d4a67" containerName="nova-metadata-metadata" containerID="cri-o://047fde4a00c176dcd1e1415fa32aa7d47b1a371c3c13b528d5f7c550e42353f4" gracePeriod=30 Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.232184 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="07f475b7-1542-4689-b3c1-f7f4470d4a67" containerName="nova-metadata-log" containerID="cri-o://c29e40d545e0c96f28dbfbc899c2ea7a2bf52224e0171c5a63ab9504fa49a30d" gracePeriod=30 Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.309656 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc"] Mar 13 16:29:27 crc kubenswrapper[4907]: E0313 16:29:27.310050 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5764492-ad4a-4e43-8a98-99ea7cfd9248" containerName="neutron-dhcp-openstack-openstack-cell1" Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.310068 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5764492-ad4a-4e43-8a98-99ea7cfd9248" containerName="neutron-dhcp-openstack-openstack-cell1" Mar 13 16:29:27 crc kubenswrapper[4907]: E0313 16:29:27.310091 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ba75cd1-2e4a-4695-b33c-8ed0cbbbb748" containerName="registry-server" Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.310098 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ba75cd1-2e4a-4695-b33c-8ed0cbbbb748" containerName="registry-server" Mar 13 16:29:27 crc kubenswrapper[4907]: E0313 16:29:27.310120 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ba75cd1-2e4a-4695-b33c-8ed0cbbbb748" containerName="extract-utilities" Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.310126 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ba75cd1-2e4a-4695-b33c-8ed0cbbbb748" containerName="extract-utilities" Mar 13 16:29:27 crc kubenswrapper[4907]: E0313 16:29:27.310148 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ba75cd1-2e4a-4695-b33c-8ed0cbbbb748" containerName="extract-content" Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.310154 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ba75cd1-2e4a-4695-b33c-8ed0cbbbb748" containerName="extract-content" Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.310341 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="b5764492-ad4a-4e43-8a98-99ea7cfd9248" containerName="neutron-dhcp-openstack-openstack-cell1" Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.310364 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ba75cd1-2e4a-4695-b33c-8ed0cbbbb748" containerName="registry-server" Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.311074 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc" Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.316001 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.317117 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-cells-global-config" Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.317308 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.317788 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.317867 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.317911 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-44kxw" Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.320113 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.335702 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc"] Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.385395 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-nova-cells-global-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc\" (UID: \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc" Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.385457 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-inventory\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc\" (UID: \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc" Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.385484 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-nova-cell1-compute-config-3\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc\" (UID: \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc" Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.385643 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-nova-cells-global-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc\" (UID: \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc" Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.385690 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc\" (UID: \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc" Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.385726 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-njn9g\" (UniqueName: \"kubernetes.io/projected/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-kube-api-access-njn9g\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc\" (UID: \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc" Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.385861 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-nova-cell1-compute-config-2\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc\" (UID: \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc" Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.385985 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc\" (UID: \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc" Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.386725 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc\" (UID: \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc" Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.386930 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-ssh-key-openstack-cell1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc\" (UID: \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc" Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.386996 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc\" (UID: \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc" Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.387061 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc\" (UID: \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc" Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.387098 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-ceph\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc\" (UID: \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc" Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.489248 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-nova-cell1-compute-config-2\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc\" (UID: \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc" Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.489500 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc\" (UID: \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc" Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.489607 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc\" (UID: \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc" Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.489707 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-ssh-key-openstack-cell1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc\" (UID: \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc" Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.489802 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc\" (UID: \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc" Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.489922 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc\" (UID: \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc" Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.490008 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-ceph\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc\" (UID: \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc" Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.490091 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-nova-cells-global-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc\" (UID: \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc" Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.490198 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-inventory\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc\" (UID: \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc" Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.490288 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-nova-cell1-compute-config-3\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc\" (UID: \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc" Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.490387 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-nova-cells-global-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc\" (UID: \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc" Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.490480 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc\" (UID: \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc" Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.490562 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-njn9g\" (UniqueName: \"kubernetes.io/projected/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-kube-api-access-njn9g\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc\" (UID: \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc" Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.493224 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-nova-cells-global-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc\" (UID: \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc" Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.495758 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-nova-cells-global-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc\" (UID: \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc" Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.502530 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-ssh-key-openstack-cell1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc\" (UID: \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc" Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.505648 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc\" (UID: \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc" Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.509424 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-inventory\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc\" (UID: \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc" Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.509603 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-nova-cell1-compute-config-3\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc\" (UID: \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc" Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.509705 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc\" (UID: \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc" Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.510086 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc\" (UID: \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc" Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.511025 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc\" (UID: \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc" Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.514447 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-ceph\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc\" (UID: \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc" Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.514465 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-nova-cell1-compute-config-2\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc\" (UID: \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc" Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.515674 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc\" (UID: \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc" Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.534664 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-njn9g\" (UniqueName: \"kubernetes.io/projected/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-kube-api-access-njn9g\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc\" (UID: \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc" Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.624521 4907 generic.go:334] "Generic (PLEG): container finished" podID="07f475b7-1542-4689-b3c1-f7f4470d4a67" containerID="c29e40d545e0c96f28dbfbc899c2ea7a2bf52224e0171c5a63ab9504fa49a30d" exitCode=143 Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.624609 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"07f475b7-1542-4689-b3c1-f7f4470d4a67","Type":"ContainerDied","Data":"c29e40d545e0c96f28dbfbc899c2ea7a2bf52224e0171c5a63ab9504fa49a30d"} Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.627020 4907 generic.go:334] "Generic (PLEG): container finished" podID="b98ee93f-f32c-4199-8045-62ab95761a6a" containerID="fa43cc779c11f32664399bd6d5fb18c420f2b54971c3216eb69f8689384ae861" exitCode=143 Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.627099 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b98ee93f-f32c-4199-8045-62ab95761a6a","Type":"ContainerDied","Data":"fa43cc779c11f32664399bd6d5fb18c420f2b54971c3216eb69f8689384ae861"} Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.730153 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc" Mar 13 16:29:28 crc kubenswrapper[4907]: E0313 16:29:28.305328 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5de9510392b1c0241095d60bc966293a382f722cf2d981565b650d57c091ce88" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Mar 13 16:29:28 crc kubenswrapper[4907]: E0313 16:29:28.307162 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5de9510392b1c0241095d60bc966293a382f722cf2d981565b650d57c091ce88" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Mar 13 16:29:28 crc kubenswrapper[4907]: E0313 16:29:28.309043 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5de9510392b1c0241095d60bc966293a382f722cf2d981565b650d57c091ce88" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Mar 13 16:29:28 crc kubenswrapper[4907]: E0313 16:29:28.309095 4907 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="48adeb28-f850-4e7c-a546-2c0b32d1b16c" containerName="nova-scheduler-scheduler" Mar 13 16:29:28 crc kubenswrapper[4907]: I0313 16:29:28.331603 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc"] Mar 13 16:29:28 crc kubenswrapper[4907]: I0313 16:29:28.638387 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc" event={"ID":"de68abba-24e7-49b8-9e1f-adfa05d9d3dd","Type":"ContainerStarted","Data":"315fdd9184d5c2a2780c5ff803490daf8312afa1aa0a6d15f69dbc1f771bf80c"} Mar 13 16:29:29 crc kubenswrapper[4907]: E0313 16:29:29.265093 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e6d65819607c1b5f8189419bb95b4c9b19ca06cb630c59aa03a5474fcce9aa3d" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Mar 13 16:29:29 crc kubenswrapper[4907]: E0313 16:29:29.271375 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e6d65819607c1b5f8189419bb95b4c9b19ca06cb630c59aa03a5474fcce9aa3d" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Mar 13 16:29:29 crc kubenswrapper[4907]: E0313 16:29:29.273580 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e6d65819607c1b5f8189419bb95b4c9b19ca06cb630c59aa03a5474fcce9aa3d" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Mar 13 16:29:29 crc kubenswrapper[4907]: E0313 16:29:29.273634 4907 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell1-conductor-0" podUID="eded65eb-7a04-4f90-ad82-be65db9014eb" containerName="nova-cell1-conductor-conductor" Mar 13 16:29:29 crc kubenswrapper[4907]: I0313 16:29:29.653007 4907 generic.go:334] "Generic (PLEG): container finished" podID="647720e3-d21e-4f41-aeb8-d4e8f092cbb5" containerID="c20788b6b1aecba8b32acff82d1977637a4b4a170211b700ca1e5ce58e45cffe" exitCode=0 Mar 13 16:29:29 crc kubenswrapper[4907]: I0313 16:29:29.653110 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"647720e3-d21e-4f41-aeb8-d4e8f092cbb5","Type":"ContainerDied","Data":"c20788b6b1aecba8b32acff82d1977637a4b4a170211b700ca1e5ce58e45cffe"} Mar 13 16:29:29 crc kubenswrapper[4907]: I0313 16:29:29.655297 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc" event={"ID":"de68abba-24e7-49b8-9e1f-adfa05d9d3dd","Type":"ContainerStarted","Data":"8644ffba95037a0d279e1ad31f88e2f9c3fae28fb5fd51859c1fb3e25cec04cc"} Mar 13 16:29:29 crc kubenswrapper[4907]: I0313 16:29:29.694788 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc" podStartSLOduration=1.983359394 podStartE2EDuration="2.694773164s" podCreationTimestamp="2026-03-13 16:29:27 +0000 UTC" firstStartedPulling="2026-03-13 16:29:28.337268342 +0000 UTC m=+8667.237056031" lastFinishedPulling="2026-03-13 16:29:29.048682082 +0000 UTC m=+8667.948469801" observedRunningTime="2026-03-13 16:29:29.680929935 +0000 UTC m=+8668.580717624" watchObservedRunningTime="2026-03-13 16:29:29.694773164 +0000 UTC m=+8668.594560853" Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.037734 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.156667 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/647720e3-d21e-4f41-aeb8-d4e8f092cbb5-combined-ca-bundle\") pod \"647720e3-d21e-4f41-aeb8-d4e8f092cbb5\" (UID: \"647720e3-d21e-4f41-aeb8-d4e8f092cbb5\") " Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.156955 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/647720e3-d21e-4f41-aeb8-d4e8f092cbb5-config-data\") pod \"647720e3-d21e-4f41-aeb8-d4e8f092cbb5\" (UID: \"647720e3-d21e-4f41-aeb8-d4e8f092cbb5\") " Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.157007 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s7nfv\" (UniqueName: \"kubernetes.io/projected/647720e3-d21e-4f41-aeb8-d4e8f092cbb5-kube-api-access-s7nfv\") pod \"647720e3-d21e-4f41-aeb8-d4e8f092cbb5\" (UID: \"647720e3-d21e-4f41-aeb8-d4e8f092cbb5\") " Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.162951 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/647720e3-d21e-4f41-aeb8-d4e8f092cbb5-kube-api-access-s7nfv" (OuterVolumeSpecName: "kube-api-access-s7nfv") pod "647720e3-d21e-4f41-aeb8-d4e8f092cbb5" (UID: "647720e3-d21e-4f41-aeb8-d4e8f092cbb5"). InnerVolumeSpecName "kube-api-access-s7nfv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.185026 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/647720e3-d21e-4f41-aeb8-d4e8f092cbb5-config-data" (OuterVolumeSpecName: "config-data") pod "647720e3-d21e-4f41-aeb8-d4e8f092cbb5" (UID: "647720e3-d21e-4f41-aeb8-d4e8f092cbb5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.196546 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/647720e3-d21e-4f41-aeb8-d4e8f092cbb5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "647720e3-d21e-4f41-aeb8-d4e8f092cbb5" (UID: "647720e3-d21e-4f41-aeb8-d4e8f092cbb5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.208270 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.259573 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/647720e3-d21e-4f41-aeb8-d4e8f092cbb5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.259604 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/647720e3-d21e-4f41-aeb8-d4e8f092cbb5-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.259615 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s7nfv\" (UniqueName: \"kubernetes.io/projected/647720e3-d21e-4f41-aeb8-d4e8f092cbb5-kube-api-access-s7nfv\") on node \"crc\" DevicePath \"\"" Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.360899 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eded65eb-7a04-4f90-ad82-be65db9014eb-config-data\") pod \"eded65eb-7a04-4f90-ad82-be65db9014eb\" (UID: \"eded65eb-7a04-4f90-ad82-be65db9014eb\") " Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.361671 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hh9bc\" (UniqueName: \"kubernetes.io/projected/eded65eb-7a04-4f90-ad82-be65db9014eb-kube-api-access-hh9bc\") pod \"eded65eb-7a04-4f90-ad82-be65db9014eb\" (UID: \"eded65eb-7a04-4f90-ad82-be65db9014eb\") " Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.361850 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eded65eb-7a04-4f90-ad82-be65db9014eb-combined-ca-bundle\") pod \"eded65eb-7a04-4f90-ad82-be65db9014eb\" (UID: \"eded65eb-7a04-4f90-ad82-be65db9014eb\") " Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.365100 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eded65eb-7a04-4f90-ad82-be65db9014eb-kube-api-access-hh9bc" (OuterVolumeSpecName: "kube-api-access-hh9bc") pod "eded65eb-7a04-4f90-ad82-be65db9014eb" (UID: "eded65eb-7a04-4f90-ad82-be65db9014eb"). InnerVolumeSpecName "kube-api-access-hh9bc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.393972 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eded65eb-7a04-4f90-ad82-be65db9014eb-config-data" (OuterVolumeSpecName: "config-data") pod "eded65eb-7a04-4f90-ad82-be65db9014eb" (UID: "eded65eb-7a04-4f90-ad82-be65db9014eb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.397010 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eded65eb-7a04-4f90-ad82-be65db9014eb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "eded65eb-7a04-4f90-ad82-be65db9014eb" (UID: "eded65eb-7a04-4f90-ad82-be65db9014eb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.464189 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eded65eb-7a04-4f90-ad82-be65db9014eb-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.464224 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hh9bc\" (UniqueName: \"kubernetes.io/projected/eded65eb-7a04-4f90-ad82-be65db9014eb-kube-api-access-hh9bc\") on node \"crc\" DevicePath \"\"" Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.464249 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eded65eb-7a04-4f90-ad82-be65db9014eb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.667155 4907 generic.go:334] "Generic (PLEG): container finished" podID="b98ee93f-f32c-4199-8045-62ab95761a6a" containerID="19d030a6fb6fc190b4d45542a5b426271778e2fd40eba0953eec528508e6f63a" exitCode=0 Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.667476 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b98ee93f-f32c-4199-8045-62ab95761a6a","Type":"ContainerDied","Data":"19d030a6fb6fc190b4d45542a5b426271778e2fd40eba0953eec528508e6f63a"} Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.669369 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.669388 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"647720e3-d21e-4f41-aeb8-d4e8f092cbb5","Type":"ContainerDied","Data":"b4d79e83e6b122d073a9e884515b0623866f06609025b076fcf8513ad8311abc"} Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.669441 4907 scope.go:117] "RemoveContainer" containerID="c20788b6b1aecba8b32acff82d1977637a4b4a170211b700ca1e5ce58e45cffe" Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.672029 4907 generic.go:334] "Generic (PLEG): container finished" podID="eded65eb-7a04-4f90-ad82-be65db9014eb" containerID="e6d65819607c1b5f8189419bb95b4c9b19ca06cb630c59aa03a5474fcce9aa3d" exitCode=0 Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.672082 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"eded65eb-7a04-4f90-ad82-be65db9014eb","Type":"ContainerDied","Data":"e6d65819607c1b5f8189419bb95b4c9b19ca06cb630c59aa03a5474fcce9aa3d"} Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.672102 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"eded65eb-7a04-4f90-ad82-be65db9014eb","Type":"ContainerDied","Data":"4c8aa8b2ae3cb083bba62c66c737dee676b8b626c0c8f53960f20765d91bb7b2"} Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.672156 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.676986 4907 generic.go:334] "Generic (PLEG): container finished" podID="07f475b7-1542-4689-b3c1-f7f4470d4a67" containerID="047fde4a00c176dcd1e1415fa32aa7d47b1a371c3c13b528d5f7c550e42353f4" exitCode=0 Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.677043 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"07f475b7-1542-4689-b3c1-f7f4470d4a67","Type":"ContainerDied","Data":"047fde4a00c176dcd1e1415fa32aa7d47b1a371c3c13b528d5f7c550e42353f4"} Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.741893 4907 scope.go:117] "RemoveContainer" containerID="e6d65819607c1b5f8189419bb95b4c9b19ca06cb630c59aa03a5474fcce9aa3d" Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.772281 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.823864 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.834093 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-0"] Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.878077 4907 scope.go:117] "RemoveContainer" containerID="e6d65819607c1b5f8189419bb95b4c9b19ca06cb630c59aa03a5474fcce9aa3d" Mar 13 16:29:30 crc kubenswrapper[4907]: E0313 16:29:30.878450 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e6d65819607c1b5f8189419bb95b4c9b19ca06cb630c59aa03a5474fcce9aa3d\": container with ID starting with e6d65819607c1b5f8189419bb95b4c9b19ca06cb630c59aa03a5474fcce9aa3d not found: ID does not exist" containerID="e6d65819607c1b5f8189419bb95b4c9b19ca06cb630c59aa03a5474fcce9aa3d" Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.878489 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e6d65819607c1b5f8189419bb95b4c9b19ca06cb630c59aa03a5474fcce9aa3d"} err="failed to get container status \"e6d65819607c1b5f8189419bb95b4c9b19ca06cb630c59aa03a5474fcce9aa3d\": rpc error: code = NotFound desc = could not find container \"e6d65819607c1b5f8189419bb95b4c9b19ca06cb630c59aa03a5474fcce9aa3d\": container with ID starting with e6d65819607c1b5f8189419bb95b4c9b19ca06cb630c59aa03a5474fcce9aa3d not found: ID does not exist" Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.878913 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/07f475b7-1542-4689-b3c1-f7f4470d4a67-logs\") pod \"07f475b7-1542-4689-b3c1-f7f4470d4a67\" (UID: \"07f475b7-1542-4689-b3c1-f7f4470d4a67\") " Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.878965 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07f475b7-1542-4689-b3c1-f7f4470d4a67-combined-ca-bundle\") pod \"07f475b7-1542-4689-b3c1-f7f4470d4a67\" (UID: \"07f475b7-1542-4689-b3c1-f7f4470d4a67\") " Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.878998 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/07f475b7-1542-4689-b3c1-f7f4470d4a67-config-data\") pod \"07f475b7-1542-4689-b3c1-f7f4470d4a67\" (UID: \"07f475b7-1542-4689-b3c1-f7f4470d4a67\") " Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.879093 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7fwbz\" (UniqueName: \"kubernetes.io/projected/07f475b7-1542-4689-b3c1-f7f4470d4a67-kube-api-access-7fwbz\") pod \"07f475b7-1542-4689-b3c1-f7f4470d4a67\" (UID: \"07f475b7-1542-4689-b3c1-f7f4470d4a67\") " Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.882786 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/07f475b7-1542-4689-b3c1-f7f4470d4a67-logs" (OuterVolumeSpecName: "logs") pod "07f475b7-1542-4689-b3c1-f7f4470d4a67" (UID: "07f475b7-1542-4689-b3c1-f7f4470d4a67"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.891713 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/07f475b7-1542-4689-b3c1-f7f4470d4a67-kube-api-access-7fwbz" (OuterVolumeSpecName: "kube-api-access-7fwbz") pod "07f475b7-1542-4689-b3c1-f7f4470d4a67" (UID: "07f475b7-1542-4689-b3c1-f7f4470d4a67"). InnerVolumeSpecName "kube-api-access-7fwbz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.894700 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Mar 13 16:29:30 crc kubenswrapper[4907]: E0313 16:29:30.895159 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07f475b7-1542-4689-b3c1-f7f4470d4a67" containerName="nova-metadata-log" Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.895176 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="07f475b7-1542-4689-b3c1-f7f4470d4a67" containerName="nova-metadata-log" Mar 13 16:29:30 crc kubenswrapper[4907]: E0313 16:29:30.895195 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07f475b7-1542-4689-b3c1-f7f4470d4a67" containerName="nova-metadata-metadata" Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.895202 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="07f475b7-1542-4689-b3c1-f7f4470d4a67" containerName="nova-metadata-metadata" Mar 13 16:29:30 crc kubenswrapper[4907]: E0313 16:29:30.895219 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="647720e3-d21e-4f41-aeb8-d4e8f092cbb5" containerName="nova-cell0-conductor-conductor" Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.895225 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="647720e3-d21e-4f41-aeb8-d4e8f092cbb5" containerName="nova-cell0-conductor-conductor" Mar 13 16:29:30 crc kubenswrapper[4907]: E0313 16:29:30.895235 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eded65eb-7a04-4f90-ad82-be65db9014eb" containerName="nova-cell1-conductor-conductor" Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.895241 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="eded65eb-7a04-4f90-ad82-be65db9014eb" containerName="nova-cell1-conductor-conductor" Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.897382 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="07f475b7-1542-4689-b3c1-f7f4470d4a67" containerName="nova-metadata-metadata" Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.897421 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="07f475b7-1542-4689-b3c1-f7f4470d4a67" containerName="nova-metadata-log" Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.897438 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="eded65eb-7a04-4f90-ad82-be65db9014eb" containerName="nova-cell1-conductor-conductor" Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.897459 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="647720e3-d21e-4f41-aeb8-d4e8f092cbb5" containerName="nova-cell0-conductor-conductor" Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.898285 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.900172 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.918271 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/07f475b7-1542-4689-b3c1-f7f4470d4a67-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "07f475b7-1542-4689-b3c1-f7f4470d4a67" (UID: "07f475b7-1542-4689-b3c1-f7f4470d4a67"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.933473 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/07f475b7-1542-4689-b3c1-f7f4470d4a67-config-data" (OuterVolumeSpecName: "config-data") pod "07f475b7-1542-4689-b3c1-f7f4470d4a67" (UID: "07f475b7-1542-4689-b3c1-f7f4470d4a67"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.940132 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.956969 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.966830 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-0"] Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.975663 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.978676 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.980820 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.981056 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd1f3838-01fe-4e37-9859-8aa28048e81c-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"fd1f3838-01fe-4e37-9859-8aa28048e81c\") " pod="openstack/nova-cell0-conductor-0" Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.981116 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6x2tp\" (UniqueName: \"kubernetes.io/projected/fd1f3838-01fe-4e37-9859-8aa28048e81c-kube-api-access-6x2tp\") pod \"nova-cell0-conductor-0\" (UID: \"fd1f3838-01fe-4e37-9859-8aa28048e81c\") " pod="openstack/nova-cell0-conductor-0" Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.981197 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd1f3838-01fe-4e37-9859-8aa28048e81c-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"fd1f3838-01fe-4e37-9859-8aa28048e81c\") " pod="openstack/nova-cell0-conductor-0" Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.981295 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7fwbz\" (UniqueName: \"kubernetes.io/projected/07f475b7-1542-4689-b3c1-f7f4470d4a67-kube-api-access-7fwbz\") on node \"crc\" DevicePath \"\"" Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.981310 4907 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/07f475b7-1542-4689-b3c1-f7f4470d4a67-logs\") on node \"crc\" DevicePath \"\"" Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.981320 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07f475b7-1542-4689-b3c1-f7f4470d4a67-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.981329 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/07f475b7-1542-4689-b3c1-f7f4470d4a67-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.984105 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.000712 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.082509 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b98ee93f-f32c-4199-8045-62ab95761a6a-combined-ca-bundle\") pod \"b98ee93f-f32c-4199-8045-62ab95761a6a\" (UID: \"b98ee93f-f32c-4199-8045-62ab95761a6a\") " Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.082795 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b98ee93f-f32c-4199-8045-62ab95761a6a-config-data\") pod \"b98ee93f-f32c-4199-8045-62ab95761a6a\" (UID: \"b98ee93f-f32c-4199-8045-62ab95761a6a\") " Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.082869 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t2t6r\" (UniqueName: \"kubernetes.io/projected/b98ee93f-f32c-4199-8045-62ab95761a6a-kube-api-access-t2t6r\") pod \"b98ee93f-f32c-4199-8045-62ab95761a6a\" (UID: \"b98ee93f-f32c-4199-8045-62ab95761a6a\") " Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.082905 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b98ee93f-f32c-4199-8045-62ab95761a6a-logs\") pod \"b98ee93f-f32c-4199-8045-62ab95761a6a\" (UID: \"b98ee93f-f32c-4199-8045-62ab95761a6a\") " Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.083375 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd1f3838-01fe-4e37-9859-8aa28048e81c-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"fd1f3838-01fe-4e37-9859-8aa28048e81c\") " pod="openstack/nova-cell0-conductor-0" Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.083408 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b98ee93f-f32c-4199-8045-62ab95761a6a-logs" (OuterVolumeSpecName: "logs") pod "b98ee93f-f32c-4199-8045-62ab95761a6a" (UID: "b98ee93f-f32c-4199-8045-62ab95761a6a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.083429 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tpn7b\" (UniqueName: \"kubernetes.io/projected/0a6bc64d-334f-46cd-aea9-486ab56d4cf9-kube-api-access-tpn7b\") pod \"nova-cell1-conductor-0\" (UID: \"0a6bc64d-334f-46cd-aea9-486ab56d4cf9\") " pod="openstack/nova-cell1-conductor-0" Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.083500 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a6bc64d-334f-46cd-aea9-486ab56d4cf9-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"0a6bc64d-334f-46cd-aea9-486ab56d4cf9\") " pod="openstack/nova-cell1-conductor-0" Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.083532 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd1f3838-01fe-4e37-9859-8aa28048e81c-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"fd1f3838-01fe-4e37-9859-8aa28048e81c\") " pod="openstack/nova-cell0-conductor-0" Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.083567 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a6bc64d-334f-46cd-aea9-486ab56d4cf9-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"0a6bc64d-334f-46cd-aea9-486ab56d4cf9\") " pod="openstack/nova-cell1-conductor-0" Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.083589 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6x2tp\" (UniqueName: \"kubernetes.io/projected/fd1f3838-01fe-4e37-9859-8aa28048e81c-kube-api-access-6x2tp\") pod \"nova-cell0-conductor-0\" (UID: \"fd1f3838-01fe-4e37-9859-8aa28048e81c\") " pod="openstack/nova-cell0-conductor-0" Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.083682 4907 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b98ee93f-f32c-4199-8045-62ab95761a6a-logs\") on node \"crc\" DevicePath \"\"" Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.085975 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b98ee93f-f32c-4199-8045-62ab95761a6a-kube-api-access-t2t6r" (OuterVolumeSpecName: "kube-api-access-t2t6r") pod "b98ee93f-f32c-4199-8045-62ab95761a6a" (UID: "b98ee93f-f32c-4199-8045-62ab95761a6a"). InnerVolumeSpecName "kube-api-access-t2t6r". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.086664 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd1f3838-01fe-4e37-9859-8aa28048e81c-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"fd1f3838-01fe-4e37-9859-8aa28048e81c\") " pod="openstack/nova-cell0-conductor-0" Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.094518 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd1f3838-01fe-4e37-9859-8aa28048e81c-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"fd1f3838-01fe-4e37-9859-8aa28048e81c\") " pod="openstack/nova-cell0-conductor-0" Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.104360 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6x2tp\" (UniqueName: \"kubernetes.io/projected/fd1f3838-01fe-4e37-9859-8aa28048e81c-kube-api-access-6x2tp\") pod \"nova-cell0-conductor-0\" (UID: \"fd1f3838-01fe-4e37-9859-8aa28048e81c\") " pod="openstack/nova-cell0-conductor-0" Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.114457 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b98ee93f-f32c-4199-8045-62ab95761a6a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b98ee93f-f32c-4199-8045-62ab95761a6a" (UID: "b98ee93f-f32c-4199-8045-62ab95761a6a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.117033 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b98ee93f-f32c-4199-8045-62ab95761a6a-config-data" (OuterVolumeSpecName: "config-data") pod "b98ee93f-f32c-4199-8045-62ab95761a6a" (UID: "b98ee93f-f32c-4199-8045-62ab95761a6a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.185226 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tpn7b\" (UniqueName: \"kubernetes.io/projected/0a6bc64d-334f-46cd-aea9-486ab56d4cf9-kube-api-access-tpn7b\") pod \"nova-cell1-conductor-0\" (UID: \"0a6bc64d-334f-46cd-aea9-486ab56d4cf9\") " pod="openstack/nova-cell1-conductor-0" Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.185313 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a6bc64d-334f-46cd-aea9-486ab56d4cf9-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"0a6bc64d-334f-46cd-aea9-486ab56d4cf9\") " pod="openstack/nova-cell1-conductor-0" Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.185391 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a6bc64d-334f-46cd-aea9-486ab56d4cf9-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"0a6bc64d-334f-46cd-aea9-486ab56d4cf9\") " pod="openstack/nova-cell1-conductor-0" Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.185505 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b98ee93f-f32c-4199-8045-62ab95761a6a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.185521 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b98ee93f-f32c-4199-8045-62ab95761a6a-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.185532 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t2t6r\" (UniqueName: \"kubernetes.io/projected/b98ee93f-f32c-4199-8045-62ab95761a6a-kube-api-access-t2t6r\") on node \"crc\" DevicePath \"\"" Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.189079 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a6bc64d-334f-46cd-aea9-486ab56d4cf9-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"0a6bc64d-334f-46cd-aea9-486ab56d4cf9\") " pod="openstack/nova-cell1-conductor-0" Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.190127 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a6bc64d-334f-46cd-aea9-486ab56d4cf9-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"0a6bc64d-334f-46cd-aea9-486ab56d4cf9\") " pod="openstack/nova-cell1-conductor-0" Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.201580 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tpn7b\" (UniqueName: \"kubernetes.io/projected/0a6bc64d-334f-46cd-aea9-486ab56d4cf9-kube-api-access-tpn7b\") pod \"nova-cell1-conductor-0\" (UID: \"0a6bc64d-334f-46cd-aea9-486ab56d4cf9\") " pod="openstack/nova-cell1-conductor-0" Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.330839 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.346663 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.687757 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"07f475b7-1542-4689-b3c1-f7f4470d4a67","Type":"ContainerDied","Data":"55a13d8a20cd160563d8a2c5f472cc7693d88cf83045ad53675342ed80dab573"} Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.687782 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.687814 4907 scope.go:117] "RemoveContainer" containerID="047fde4a00c176dcd1e1415fa32aa7d47b1a371c3c13b528d5f7c550e42353f4" Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.690592 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b98ee93f-f32c-4199-8045-62ab95761a6a","Type":"ContainerDied","Data":"6c6eae74def50447208be0f6d41599241acef5069efd81dc6d9fc3a28d2dd298"} Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.690719 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.725946 4907 scope.go:117] "RemoveContainer" containerID="c29e40d545e0c96f28dbfbc899c2ea7a2bf52224e0171c5a63ab9504fa49a30d" Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.728906 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.745481 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.778079 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.784806 4907 scope.go:117] "RemoveContainer" containerID="19d030a6fb6fc190b4d45542a5b426271778e2fd40eba0953eec528508e6f63a" Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.805539 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="07f475b7-1542-4689-b3c1-f7f4470d4a67" path="/var/lib/kubelet/pods/07f475b7-1542-4689-b3c1-f7f4470d4a67/volumes" Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.810526 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="647720e3-d21e-4f41-aeb8-d4e8f092cbb5" path="/var/lib/kubelet/pods/647720e3-d21e-4f41-aeb8-d4e8f092cbb5/volumes" Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.812873 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eded65eb-7a04-4f90-ad82-be65db9014eb" path="/var/lib/kubelet/pods/eded65eb-7a04-4f90-ad82-be65db9014eb/volumes" Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.813610 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.813637 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Mar 13 16:29:31 crc kubenswrapper[4907]: E0313 16:29:31.813985 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b98ee93f-f32c-4199-8045-62ab95761a6a" containerName="nova-api-log" Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.813996 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="b98ee93f-f32c-4199-8045-62ab95761a6a" containerName="nova-api-log" Mar 13 16:29:31 crc kubenswrapper[4907]: E0313 16:29:31.814056 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b98ee93f-f32c-4199-8045-62ab95761a6a" containerName="nova-api-api" Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.814064 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="b98ee93f-f32c-4199-8045-62ab95761a6a" containerName="nova-api-api" Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.819009 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="b98ee93f-f32c-4199-8045-62ab95761a6a" containerName="nova-api-api" Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.819050 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="b98ee93f-f32c-4199-8045-62ab95761a6a" containerName="nova-api-log" Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.820522 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.830128 4907 scope.go:117] "RemoveContainer" containerID="fa43cc779c11f32664399bd6d5fb18c420f2b54971c3216eb69f8689384ae861" Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.830386 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.842249 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.855467 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.857426 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.862165 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.867440 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.875932 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.900102 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13364227-6148-4b0d-ba7d-082578dfe1d5-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"13364227-6148-4b0d-ba7d-082578dfe1d5\") " pod="openstack/nova-metadata-0" Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.900187 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/13364227-6148-4b0d-ba7d-082578dfe1d5-config-data\") pod \"nova-metadata-0\" (UID: \"13364227-6148-4b0d-ba7d-082578dfe1d5\") " pod="openstack/nova-metadata-0" Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.900280 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/13364227-6148-4b0d-ba7d-082578dfe1d5-logs\") pod \"nova-metadata-0\" (UID: \"13364227-6148-4b0d-ba7d-082578dfe1d5\") " pod="openstack/nova-metadata-0" Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.900338 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rmdsn\" (UniqueName: \"kubernetes.io/projected/13364227-6148-4b0d-ba7d-082578dfe1d5-kube-api-access-rmdsn\") pod \"nova-metadata-0\" (UID: \"13364227-6148-4b0d-ba7d-082578dfe1d5\") " pod="openstack/nova-metadata-0" Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.934847 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Mar 13 16:29:32 crc kubenswrapper[4907]: I0313 16:29:32.012669 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c11f341d-9a76-4e7e-b1e7-8d63c30bac32-config-data\") pod \"nova-api-0\" (UID: \"c11f341d-9a76-4e7e-b1e7-8d63c30bac32\") " pod="openstack/nova-api-0" Mar 13 16:29:32 crc kubenswrapper[4907]: I0313 16:29:32.013132 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13364227-6148-4b0d-ba7d-082578dfe1d5-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"13364227-6148-4b0d-ba7d-082578dfe1d5\") " pod="openstack/nova-metadata-0" Mar 13 16:29:32 crc kubenswrapper[4907]: I0313 16:29:32.013194 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c11f341d-9a76-4e7e-b1e7-8d63c30bac32-logs\") pod \"nova-api-0\" (UID: \"c11f341d-9a76-4e7e-b1e7-8d63c30bac32\") " pod="openstack/nova-api-0" Mar 13 16:29:32 crc kubenswrapper[4907]: I0313 16:29:32.013233 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/13364227-6148-4b0d-ba7d-082578dfe1d5-config-data\") pod \"nova-metadata-0\" (UID: \"13364227-6148-4b0d-ba7d-082578dfe1d5\") " pod="openstack/nova-metadata-0" Mar 13 16:29:32 crc kubenswrapper[4907]: I0313 16:29:32.013398 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zx7hj\" (UniqueName: \"kubernetes.io/projected/c11f341d-9a76-4e7e-b1e7-8d63c30bac32-kube-api-access-zx7hj\") pod \"nova-api-0\" (UID: \"c11f341d-9a76-4e7e-b1e7-8d63c30bac32\") " pod="openstack/nova-api-0" Mar 13 16:29:32 crc kubenswrapper[4907]: I0313 16:29:32.013446 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/13364227-6148-4b0d-ba7d-082578dfe1d5-logs\") pod \"nova-metadata-0\" (UID: \"13364227-6148-4b0d-ba7d-082578dfe1d5\") " pod="openstack/nova-metadata-0" Mar 13 16:29:32 crc kubenswrapper[4907]: I0313 16:29:32.013515 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c11f341d-9a76-4e7e-b1e7-8d63c30bac32-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"c11f341d-9a76-4e7e-b1e7-8d63c30bac32\") " pod="openstack/nova-api-0" Mar 13 16:29:32 crc kubenswrapper[4907]: I0313 16:29:32.013599 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rmdsn\" (UniqueName: \"kubernetes.io/projected/13364227-6148-4b0d-ba7d-082578dfe1d5-kube-api-access-rmdsn\") pod \"nova-metadata-0\" (UID: \"13364227-6148-4b0d-ba7d-082578dfe1d5\") " pod="openstack/nova-metadata-0" Mar 13 16:29:32 crc kubenswrapper[4907]: I0313 16:29:32.014117 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/13364227-6148-4b0d-ba7d-082578dfe1d5-logs\") pod \"nova-metadata-0\" (UID: \"13364227-6148-4b0d-ba7d-082578dfe1d5\") " pod="openstack/nova-metadata-0" Mar 13 16:29:32 crc kubenswrapper[4907]: I0313 16:29:32.018164 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13364227-6148-4b0d-ba7d-082578dfe1d5-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"13364227-6148-4b0d-ba7d-082578dfe1d5\") " pod="openstack/nova-metadata-0" Mar 13 16:29:32 crc kubenswrapper[4907]: I0313 16:29:32.018471 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/13364227-6148-4b0d-ba7d-082578dfe1d5-config-data\") pod \"nova-metadata-0\" (UID: \"13364227-6148-4b0d-ba7d-082578dfe1d5\") " pod="openstack/nova-metadata-0" Mar 13 16:29:32 crc kubenswrapper[4907]: I0313 16:29:32.029983 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rmdsn\" (UniqueName: \"kubernetes.io/projected/13364227-6148-4b0d-ba7d-082578dfe1d5-kube-api-access-rmdsn\") pod \"nova-metadata-0\" (UID: \"13364227-6148-4b0d-ba7d-082578dfe1d5\") " pod="openstack/nova-metadata-0" Mar 13 16:29:32 crc kubenswrapper[4907]: I0313 16:29:32.115797 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c11f341d-9a76-4e7e-b1e7-8d63c30bac32-logs\") pod \"nova-api-0\" (UID: \"c11f341d-9a76-4e7e-b1e7-8d63c30bac32\") " pod="openstack/nova-api-0" Mar 13 16:29:32 crc kubenswrapper[4907]: I0313 16:29:32.115952 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zx7hj\" (UniqueName: \"kubernetes.io/projected/c11f341d-9a76-4e7e-b1e7-8d63c30bac32-kube-api-access-zx7hj\") pod \"nova-api-0\" (UID: \"c11f341d-9a76-4e7e-b1e7-8d63c30bac32\") " pod="openstack/nova-api-0" Mar 13 16:29:32 crc kubenswrapper[4907]: I0313 16:29:32.115996 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c11f341d-9a76-4e7e-b1e7-8d63c30bac32-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"c11f341d-9a76-4e7e-b1e7-8d63c30bac32\") " pod="openstack/nova-api-0" Mar 13 16:29:32 crc kubenswrapper[4907]: I0313 16:29:32.116070 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c11f341d-9a76-4e7e-b1e7-8d63c30bac32-config-data\") pod \"nova-api-0\" (UID: \"c11f341d-9a76-4e7e-b1e7-8d63c30bac32\") " pod="openstack/nova-api-0" Mar 13 16:29:32 crc kubenswrapper[4907]: I0313 16:29:32.116738 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c11f341d-9a76-4e7e-b1e7-8d63c30bac32-logs\") pod \"nova-api-0\" (UID: \"c11f341d-9a76-4e7e-b1e7-8d63c30bac32\") " pod="openstack/nova-api-0" Mar 13 16:29:32 crc kubenswrapper[4907]: I0313 16:29:32.120148 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c11f341d-9a76-4e7e-b1e7-8d63c30bac32-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"c11f341d-9a76-4e7e-b1e7-8d63c30bac32\") " pod="openstack/nova-api-0" Mar 13 16:29:32 crc kubenswrapper[4907]: I0313 16:29:32.129445 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c11f341d-9a76-4e7e-b1e7-8d63c30bac32-config-data\") pod \"nova-api-0\" (UID: \"c11f341d-9a76-4e7e-b1e7-8d63c30bac32\") " pod="openstack/nova-api-0" Mar 13 16:29:32 crc kubenswrapper[4907]: I0313 16:29:32.131181 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zx7hj\" (UniqueName: \"kubernetes.io/projected/c11f341d-9a76-4e7e-b1e7-8d63c30bac32-kube-api-access-zx7hj\") pod \"nova-api-0\" (UID: \"c11f341d-9a76-4e7e-b1e7-8d63c30bac32\") " pod="openstack/nova-api-0" Mar 13 16:29:32 crc kubenswrapper[4907]: I0313 16:29:32.142815 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 13 16:29:32 crc kubenswrapper[4907]: I0313 16:29:32.221595 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 13 16:29:32 crc kubenswrapper[4907]: I0313 16:29:32.584146 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Mar 13 16:29:32 crc kubenswrapper[4907]: W0313 16:29:32.586856 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod13364227_6148_4b0d_ba7d_082578dfe1d5.slice/crio-1ff2b23316c389fffd85d9082d46b0eb2f2425c20a257e387b3147bcbd87d083 WatchSource:0}: Error finding container 1ff2b23316c389fffd85d9082d46b0eb2f2425c20a257e387b3147bcbd87d083: Status 404 returned error can't find the container with id 1ff2b23316c389fffd85d9082d46b0eb2f2425c20a257e387b3147bcbd87d083 Mar 13 16:29:32 crc kubenswrapper[4907]: I0313 16:29:32.605351 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 13 16:29:32 crc kubenswrapper[4907]: I0313 16:29:32.711268 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"13364227-6148-4b0d-ba7d-082578dfe1d5","Type":"ContainerStarted","Data":"1ff2b23316c389fffd85d9082d46b0eb2f2425c20a257e387b3147bcbd87d083"} Mar 13 16:29:32 crc kubenswrapper[4907]: I0313 16:29:32.713047 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"fd1f3838-01fe-4e37-9859-8aa28048e81c","Type":"ContainerStarted","Data":"2a9b247acff02796096f1440a3319cf378894bfe822c7a2d8316e4926ccfc23a"} Mar 13 16:29:32 crc kubenswrapper[4907]: I0313 16:29:32.713074 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"fd1f3838-01fe-4e37-9859-8aa28048e81c","Type":"ContainerStarted","Data":"fe9e322f499f6530211e73b3ccc8acf23f16f152f1f35fe4ed828c9adb388d70"} Mar 13 16:29:32 crc kubenswrapper[4907]: I0313 16:29:32.713399 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Mar 13 16:29:32 crc kubenswrapper[4907]: I0313 16:29:32.716023 4907 generic.go:334] "Generic (PLEG): container finished" podID="48adeb28-f850-4e7c-a546-2c0b32d1b16c" containerID="5de9510392b1c0241095d60bc966293a382f722cf2d981565b650d57c091ce88" exitCode=0 Mar 13 16:29:32 crc kubenswrapper[4907]: I0313 16:29:32.716135 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 13 16:29:32 crc kubenswrapper[4907]: I0313 16:29:32.716300 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"48adeb28-f850-4e7c-a546-2c0b32d1b16c","Type":"ContainerDied","Data":"5de9510392b1c0241095d60bc966293a382f722cf2d981565b650d57c091ce88"} Mar 13 16:29:32 crc kubenswrapper[4907]: I0313 16:29:32.716344 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"48adeb28-f850-4e7c-a546-2c0b32d1b16c","Type":"ContainerDied","Data":"bba3814d9581f3c03937b4a23ce7f341cbe05c058698e236d16b4dfa33146288"} Mar 13 16:29:32 crc kubenswrapper[4907]: I0313 16:29:32.716362 4907 scope.go:117] "RemoveContainer" containerID="5de9510392b1c0241095d60bc966293a382f722cf2d981565b650d57c091ce88" Mar 13 16:29:32 crc kubenswrapper[4907]: I0313 16:29:32.719715 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"0a6bc64d-334f-46cd-aea9-486ab56d4cf9","Type":"ContainerStarted","Data":"a74f56ca47951bf2450e7cf3e0fb341611dfcfee373d5cc75aa42a2e60567167"} Mar 13 16:29:32 crc kubenswrapper[4907]: I0313 16:29:32.719758 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"0a6bc64d-334f-46cd-aea9-486ab56d4cf9","Type":"ContainerStarted","Data":"32bb38cda79c6777a5df03efdb4d4b0b1b79e95ac6d0ad101b98b5d9b64830ef"} Mar 13 16:29:32 crc kubenswrapper[4907]: I0313 16:29:32.720182 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Mar 13 16:29:32 crc kubenswrapper[4907]: I0313 16:29:32.731915 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/48adeb28-f850-4e7c-a546-2c0b32d1b16c-config-data\") pod \"48adeb28-f850-4e7c-a546-2c0b32d1b16c\" (UID: \"48adeb28-f850-4e7c-a546-2c0b32d1b16c\") " Mar 13 16:29:32 crc kubenswrapper[4907]: I0313 16:29:32.732282 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bm6cq\" (UniqueName: \"kubernetes.io/projected/48adeb28-f850-4e7c-a546-2c0b32d1b16c-kube-api-access-bm6cq\") pod \"48adeb28-f850-4e7c-a546-2c0b32d1b16c\" (UID: \"48adeb28-f850-4e7c-a546-2c0b32d1b16c\") " Mar 13 16:29:32 crc kubenswrapper[4907]: I0313 16:29:32.732547 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/48adeb28-f850-4e7c-a546-2c0b32d1b16c-combined-ca-bundle\") pod \"48adeb28-f850-4e7c-a546-2c0b32d1b16c\" (UID: \"48adeb28-f850-4e7c-a546-2c0b32d1b16c\") " Mar 13 16:29:32 crc kubenswrapper[4907]: I0313 16:29:32.738159 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/48adeb28-f850-4e7c-a546-2c0b32d1b16c-kube-api-access-bm6cq" (OuterVolumeSpecName: "kube-api-access-bm6cq") pod "48adeb28-f850-4e7c-a546-2c0b32d1b16c" (UID: "48adeb28-f850-4e7c-a546-2c0b32d1b16c"). InnerVolumeSpecName "kube-api-access-bm6cq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:29:32 crc kubenswrapper[4907]: I0313 16:29:32.758499 4907 scope.go:117] "RemoveContainer" containerID="5de9510392b1c0241095d60bc966293a382f722cf2d981565b650d57c091ce88" Mar 13 16:29:32 crc kubenswrapper[4907]: E0313 16:29:32.759612 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5de9510392b1c0241095d60bc966293a382f722cf2d981565b650d57c091ce88\": container with ID starting with 5de9510392b1c0241095d60bc966293a382f722cf2d981565b650d57c091ce88 not found: ID does not exist" containerID="5de9510392b1c0241095d60bc966293a382f722cf2d981565b650d57c091ce88" Mar 13 16:29:32 crc kubenswrapper[4907]: I0313 16:29:32.759667 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5de9510392b1c0241095d60bc966293a382f722cf2d981565b650d57c091ce88"} err="failed to get container status \"5de9510392b1c0241095d60bc966293a382f722cf2d981565b650d57c091ce88\": rpc error: code = NotFound desc = could not find container \"5de9510392b1c0241095d60bc966293a382f722cf2d981565b650d57c091ce88\": container with ID starting with 5de9510392b1c0241095d60bc966293a382f722cf2d981565b650d57c091ce88 not found: ID does not exist" Mar 13 16:29:32 crc kubenswrapper[4907]: I0313 16:29:32.766655 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.766636185 podStartE2EDuration="2.766636185s" podCreationTimestamp="2026-03-13 16:29:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 16:29:32.728978856 +0000 UTC m=+8671.628766555" watchObservedRunningTime="2026-03-13 16:29:32.766636185 +0000 UTC m=+8671.666423874" Mar 13 16:29:32 crc kubenswrapper[4907]: W0313 16:29:32.773641 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc11f341d_9a76_4e7e_b1e7_8d63c30bac32.slice/crio-0a9826a17473ee42d705880efe1f102dabc505ebf63c497c5a550f34dd34a194 WatchSource:0}: Error finding container 0a9826a17473ee42d705880efe1f102dabc505ebf63c497c5a550f34dd34a194: Status 404 returned error can't find the container with id 0a9826a17473ee42d705880efe1f102dabc505ebf63c497c5a550f34dd34a194 Mar 13 16:29:32 crc kubenswrapper[4907]: I0313 16:29:32.774379 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.774359343 podStartE2EDuration="2.774359343s" podCreationTimestamp="2026-03-13 16:29:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 16:29:32.742803525 +0000 UTC m=+8671.642591214" watchObservedRunningTime="2026-03-13 16:29:32.774359343 +0000 UTC m=+8671.674147032" Mar 13 16:29:32 crc kubenswrapper[4907]: I0313 16:29:32.775938 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/48adeb28-f850-4e7c-a546-2c0b32d1b16c-config-data" (OuterVolumeSpecName: "config-data") pod "48adeb28-f850-4e7c-a546-2c0b32d1b16c" (UID: "48adeb28-f850-4e7c-a546-2c0b32d1b16c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:29:32 crc kubenswrapper[4907]: I0313 16:29:32.784176 4907 scope.go:117] "RemoveContainer" containerID="dfb327c310937cae28a687e8c5c0256a2ab50e59a756669f2586112310e1dc69" Mar 13 16:29:32 crc kubenswrapper[4907]: E0313 16:29:32.784379 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:29:32 crc kubenswrapper[4907]: I0313 16:29:32.798005 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 13 16:29:32 crc kubenswrapper[4907]: I0313 16:29:32.805354 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/48adeb28-f850-4e7c-a546-2c0b32d1b16c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "48adeb28-f850-4e7c-a546-2c0b32d1b16c" (UID: "48adeb28-f850-4e7c-a546-2c0b32d1b16c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:29:32 crc kubenswrapper[4907]: I0313 16:29:32.835521 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bm6cq\" (UniqueName: \"kubernetes.io/projected/48adeb28-f850-4e7c-a546-2c0b32d1b16c-kube-api-access-bm6cq\") on node \"crc\" DevicePath \"\"" Mar 13 16:29:32 crc kubenswrapper[4907]: I0313 16:29:32.835712 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/48adeb28-f850-4e7c-a546-2c0b32d1b16c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 16:29:32 crc kubenswrapper[4907]: I0313 16:29:32.835771 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/48adeb28-f850-4e7c-a546-2c0b32d1b16c-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 16:29:33 crc kubenswrapper[4907]: I0313 16:29:33.074229 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 16:29:33 crc kubenswrapper[4907]: I0313 16:29:33.084865 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 16:29:33 crc kubenswrapper[4907]: I0313 16:29:33.131921 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 16:29:33 crc kubenswrapper[4907]: E0313 16:29:33.132676 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48adeb28-f850-4e7c-a546-2c0b32d1b16c" containerName="nova-scheduler-scheduler" Mar 13 16:29:33 crc kubenswrapper[4907]: I0313 16:29:33.132695 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="48adeb28-f850-4e7c-a546-2c0b32d1b16c" containerName="nova-scheduler-scheduler" Mar 13 16:29:33 crc kubenswrapper[4907]: I0313 16:29:33.133019 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="48adeb28-f850-4e7c-a546-2c0b32d1b16c" containerName="nova-scheduler-scheduler" Mar 13 16:29:33 crc kubenswrapper[4907]: I0313 16:29:33.134053 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 13 16:29:33 crc kubenswrapper[4907]: I0313 16:29:33.137352 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Mar 13 16:29:33 crc kubenswrapper[4907]: I0313 16:29:33.148269 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 16:29:33 crc kubenswrapper[4907]: I0313 16:29:33.252448 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bv7kh\" (UniqueName: \"kubernetes.io/projected/54123701-97fb-4f4f-baed-4c2183bad203-kube-api-access-bv7kh\") pod \"nova-scheduler-0\" (UID: \"54123701-97fb-4f4f-baed-4c2183bad203\") " pod="openstack/nova-scheduler-0" Mar 13 16:29:33 crc kubenswrapper[4907]: I0313 16:29:33.252526 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54123701-97fb-4f4f-baed-4c2183bad203-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"54123701-97fb-4f4f-baed-4c2183bad203\") " pod="openstack/nova-scheduler-0" Mar 13 16:29:33 crc kubenswrapper[4907]: I0313 16:29:33.252666 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/54123701-97fb-4f4f-baed-4c2183bad203-config-data\") pod \"nova-scheduler-0\" (UID: \"54123701-97fb-4f4f-baed-4c2183bad203\") " pod="openstack/nova-scheduler-0" Mar 13 16:29:33 crc kubenswrapper[4907]: I0313 16:29:33.354722 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/54123701-97fb-4f4f-baed-4c2183bad203-config-data\") pod \"nova-scheduler-0\" (UID: \"54123701-97fb-4f4f-baed-4c2183bad203\") " pod="openstack/nova-scheduler-0" Mar 13 16:29:33 crc kubenswrapper[4907]: I0313 16:29:33.355133 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bv7kh\" (UniqueName: \"kubernetes.io/projected/54123701-97fb-4f4f-baed-4c2183bad203-kube-api-access-bv7kh\") pod \"nova-scheduler-0\" (UID: \"54123701-97fb-4f4f-baed-4c2183bad203\") " pod="openstack/nova-scheduler-0" Mar 13 16:29:33 crc kubenswrapper[4907]: I0313 16:29:33.355213 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54123701-97fb-4f4f-baed-4c2183bad203-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"54123701-97fb-4f4f-baed-4c2183bad203\") " pod="openstack/nova-scheduler-0" Mar 13 16:29:33 crc kubenswrapper[4907]: I0313 16:29:33.362635 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/54123701-97fb-4f4f-baed-4c2183bad203-config-data\") pod \"nova-scheduler-0\" (UID: \"54123701-97fb-4f4f-baed-4c2183bad203\") " pod="openstack/nova-scheduler-0" Mar 13 16:29:33 crc kubenswrapper[4907]: I0313 16:29:33.365471 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54123701-97fb-4f4f-baed-4c2183bad203-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"54123701-97fb-4f4f-baed-4c2183bad203\") " pod="openstack/nova-scheduler-0" Mar 13 16:29:33 crc kubenswrapper[4907]: I0313 16:29:33.377107 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bv7kh\" (UniqueName: \"kubernetes.io/projected/54123701-97fb-4f4f-baed-4c2183bad203-kube-api-access-bv7kh\") pod \"nova-scheduler-0\" (UID: \"54123701-97fb-4f4f-baed-4c2183bad203\") " pod="openstack/nova-scheduler-0" Mar 13 16:29:33 crc kubenswrapper[4907]: I0313 16:29:33.482212 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 13 16:29:33 crc kubenswrapper[4907]: I0313 16:29:33.729950 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c11f341d-9a76-4e7e-b1e7-8d63c30bac32","Type":"ContainerStarted","Data":"961fd42c70e8f82176c365ec5a810b2861304a70794684f34381af8052b330e0"} Mar 13 16:29:33 crc kubenswrapper[4907]: I0313 16:29:33.730326 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c11f341d-9a76-4e7e-b1e7-8d63c30bac32","Type":"ContainerStarted","Data":"f5d734a1b9cc16fa2be7da399d5b434256b5ef06f84511b5c4f404f76cbf3a3d"} Mar 13 16:29:33 crc kubenswrapper[4907]: I0313 16:29:33.730339 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c11f341d-9a76-4e7e-b1e7-8d63c30bac32","Type":"ContainerStarted","Data":"0a9826a17473ee42d705880efe1f102dabc505ebf63c497c5a550f34dd34a194"} Mar 13 16:29:33 crc kubenswrapper[4907]: I0313 16:29:33.736465 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"13364227-6148-4b0d-ba7d-082578dfe1d5","Type":"ContainerStarted","Data":"0ced5e91ca7c08e6c13f6c1fcd52b7cec01a06c9dc08fd56db8b89c560a582f0"} Mar 13 16:29:33 crc kubenswrapper[4907]: I0313 16:29:33.736491 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"13364227-6148-4b0d-ba7d-082578dfe1d5","Type":"ContainerStarted","Data":"8f1587fc1344b2989add0ef318fec5bbb239ecb1b3e8ef09fe25211d417e344b"} Mar 13 16:29:33 crc kubenswrapper[4907]: I0313 16:29:33.749917 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.749898742 podStartE2EDuration="2.749898742s" podCreationTimestamp="2026-03-13 16:29:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 16:29:33.745199179 +0000 UTC m=+8672.644986888" watchObservedRunningTime="2026-03-13 16:29:33.749898742 +0000 UTC m=+8672.649686441" Mar 13 16:29:33 crc kubenswrapper[4907]: I0313 16:29:33.785826 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.785809831 podStartE2EDuration="2.785809831s" podCreationTimestamp="2026-03-13 16:29:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 16:29:33.768611737 +0000 UTC m=+8672.668399426" watchObservedRunningTime="2026-03-13 16:29:33.785809831 +0000 UTC m=+8672.685597520" Mar 13 16:29:33 crc kubenswrapper[4907]: I0313 16:29:33.793712 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="48adeb28-f850-4e7c-a546-2c0b32d1b16c" path="/var/lib/kubelet/pods/48adeb28-f850-4e7c-a546-2c0b32d1b16c/volumes" Mar 13 16:29:33 crc kubenswrapper[4907]: I0313 16:29:33.794452 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b98ee93f-f32c-4199-8045-62ab95761a6a" path="/var/lib/kubelet/pods/b98ee93f-f32c-4199-8045-62ab95761a6a/volumes" Mar 13 16:29:33 crc kubenswrapper[4907]: I0313 16:29:33.949443 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 16:29:33 crc kubenswrapper[4907]: W0313 16:29:33.960449 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod54123701_97fb_4f4f_baed_4c2183bad203.slice/crio-c5374a10aaafdfc348fb5411307fafd2b204901548a49c65dbd3dfb49e078a00 WatchSource:0}: Error finding container c5374a10aaafdfc348fb5411307fafd2b204901548a49c65dbd3dfb49e078a00: Status 404 returned error can't find the container with id c5374a10aaafdfc348fb5411307fafd2b204901548a49c65dbd3dfb49e078a00 Mar 13 16:29:34 crc kubenswrapper[4907]: I0313 16:29:34.750420 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"54123701-97fb-4f4f-baed-4c2183bad203","Type":"ContainerStarted","Data":"a4b11035cfa2df06f9b86ebf95edee9ab566cdaa742b9648bb53b6e9192440e2"} Mar 13 16:29:34 crc kubenswrapper[4907]: I0313 16:29:34.750726 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"54123701-97fb-4f4f-baed-4c2183bad203","Type":"ContainerStarted","Data":"c5374a10aaafdfc348fb5411307fafd2b204901548a49c65dbd3dfb49e078a00"} Mar 13 16:29:34 crc kubenswrapper[4907]: I0313 16:29:34.787759 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=1.782571156 podStartE2EDuration="1.782571156s" podCreationTimestamp="2026-03-13 16:29:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 16:29:34.775043595 +0000 UTC m=+8673.674831354" watchObservedRunningTime="2026-03-13 16:29:34.782571156 +0000 UTC m=+8673.682358885" Mar 13 16:29:38 crc kubenswrapper[4907]: I0313 16:29:38.483627 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Mar 13 16:29:41 crc kubenswrapper[4907]: I0313 16:29:41.395762 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Mar 13 16:29:41 crc kubenswrapper[4907]: I0313 16:29:41.410962 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Mar 13 16:29:42 crc kubenswrapper[4907]: I0313 16:29:42.145032 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Mar 13 16:29:42 crc kubenswrapper[4907]: I0313 16:29:42.145403 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Mar 13 16:29:42 crc kubenswrapper[4907]: I0313 16:29:42.222679 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Mar 13 16:29:42 crc kubenswrapper[4907]: I0313 16:29:42.222756 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Mar 13 16:29:43 crc kubenswrapper[4907]: I0313 16:29:43.228373 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="13364227-6148-4b0d-ba7d-082578dfe1d5" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.0.8:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 16:29:43 crc kubenswrapper[4907]: I0313 16:29:43.228453 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="13364227-6148-4b0d-ba7d-082578dfe1d5" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.0.8:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 16:29:43 crc kubenswrapper[4907]: I0313 16:29:43.310076 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="c11f341d-9a76-4e7e-b1e7-8d63c30bac32" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.11:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 16:29:43 crc kubenswrapper[4907]: I0313 16:29:43.310095 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="c11f341d-9a76-4e7e-b1e7-8d63c30bac32" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.11:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 16:29:43 crc kubenswrapper[4907]: I0313 16:29:43.483817 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Mar 13 16:29:43 crc kubenswrapper[4907]: I0313 16:29:43.524802 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Mar 13 16:29:43 crc kubenswrapper[4907]: I0313 16:29:43.876676 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Mar 13 16:29:45 crc kubenswrapper[4907]: I0313 16:29:45.783669 4907 scope.go:117] "RemoveContainer" containerID="dfb327c310937cae28a687e8c5c0256a2ab50e59a756669f2586112310e1dc69" Mar 13 16:29:45 crc kubenswrapper[4907]: E0313 16:29:45.784267 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:29:50 crc kubenswrapper[4907]: I0313 16:29:50.143114 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Mar 13 16:29:50 crc kubenswrapper[4907]: I0313 16:29:50.144646 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Mar 13 16:29:50 crc kubenswrapper[4907]: I0313 16:29:50.222651 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Mar 13 16:29:50 crc kubenswrapper[4907]: I0313 16:29:50.222715 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Mar 13 16:29:52 crc kubenswrapper[4907]: I0313 16:29:52.146060 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Mar 13 16:29:52 crc kubenswrapper[4907]: I0313 16:29:52.146469 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Mar 13 16:29:52 crc kubenswrapper[4907]: I0313 16:29:52.150015 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Mar 13 16:29:52 crc kubenswrapper[4907]: I0313 16:29:52.152678 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Mar 13 16:29:52 crc kubenswrapper[4907]: I0313 16:29:52.230323 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Mar 13 16:29:52 crc kubenswrapper[4907]: I0313 16:29:52.233902 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Mar 13 16:29:52 crc kubenswrapper[4907]: I0313 16:29:52.243172 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Mar 13 16:29:52 crc kubenswrapper[4907]: I0313 16:29:52.962055 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Mar 13 16:29:59 crc kubenswrapper[4907]: I0313 16:29:59.783162 4907 scope.go:117] "RemoveContainer" containerID="dfb327c310937cae28a687e8c5c0256a2ab50e59a756669f2586112310e1dc69" Mar 13 16:29:59 crc kubenswrapper[4907]: E0313 16:29:59.784263 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:30:00 crc kubenswrapper[4907]: I0313 16:30:00.151049 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556990-mgpgm"] Mar 13 16:30:00 crc kubenswrapper[4907]: I0313 16:30:00.153137 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556990-mgpgm" Mar 13 16:30:00 crc kubenswrapper[4907]: I0313 16:30:00.155109 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 13 16:30:00 crc kubenswrapper[4907]: I0313 16:30:00.156871 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 13 16:30:00 crc kubenswrapper[4907]: I0313 16:30:00.164947 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556990-4jrg4"] Mar 13 16:30:00 crc kubenswrapper[4907]: I0313 16:30:00.167734 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556990-4jrg4" Mar 13 16:30:00 crc kubenswrapper[4907]: I0313 16:30:00.169603 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 16:30:00 crc kubenswrapper[4907]: I0313 16:30:00.169950 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 16:30:00 crc kubenswrapper[4907]: I0313 16:30:00.177980 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 16:30:00 crc kubenswrapper[4907]: I0313 16:30:00.178984 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556990-mgpgm"] Mar 13 16:30:00 crc kubenswrapper[4907]: I0313 16:30:00.192485 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556990-4jrg4"] Mar 13 16:30:00 crc kubenswrapper[4907]: I0313 16:30:00.341544 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-86rp9\" (UniqueName: \"kubernetes.io/projected/5a413954-e701-4cec-8868-4fcdaecfabe8-kube-api-access-86rp9\") pod \"auto-csr-approver-29556990-4jrg4\" (UID: \"5a413954-e701-4cec-8868-4fcdaecfabe8\") " pod="openshift-infra/auto-csr-approver-29556990-4jrg4" Mar 13 16:30:00 crc kubenswrapper[4907]: I0313 16:30:00.341628 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n5k8t\" (UniqueName: \"kubernetes.io/projected/eb5ba94e-890e-4791-b2f6-e30d1d92ed96-kube-api-access-n5k8t\") pod \"collect-profiles-29556990-mgpgm\" (UID: \"eb5ba94e-890e-4791-b2f6-e30d1d92ed96\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556990-mgpgm" Mar 13 16:30:00 crc kubenswrapper[4907]: I0313 16:30:00.341652 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/eb5ba94e-890e-4791-b2f6-e30d1d92ed96-secret-volume\") pod \"collect-profiles-29556990-mgpgm\" (UID: \"eb5ba94e-890e-4791-b2f6-e30d1d92ed96\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556990-mgpgm" Mar 13 16:30:00 crc kubenswrapper[4907]: I0313 16:30:00.341734 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/eb5ba94e-890e-4791-b2f6-e30d1d92ed96-config-volume\") pod \"collect-profiles-29556990-mgpgm\" (UID: \"eb5ba94e-890e-4791-b2f6-e30d1d92ed96\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556990-mgpgm" Mar 13 16:30:00 crc kubenswrapper[4907]: I0313 16:30:00.443768 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-86rp9\" (UniqueName: \"kubernetes.io/projected/5a413954-e701-4cec-8868-4fcdaecfabe8-kube-api-access-86rp9\") pod \"auto-csr-approver-29556990-4jrg4\" (UID: \"5a413954-e701-4cec-8868-4fcdaecfabe8\") " pod="openshift-infra/auto-csr-approver-29556990-4jrg4" Mar 13 16:30:00 crc kubenswrapper[4907]: I0313 16:30:00.443854 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/eb5ba94e-890e-4791-b2f6-e30d1d92ed96-secret-volume\") pod \"collect-profiles-29556990-mgpgm\" (UID: \"eb5ba94e-890e-4791-b2f6-e30d1d92ed96\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556990-mgpgm" Mar 13 16:30:00 crc kubenswrapper[4907]: I0313 16:30:00.443901 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n5k8t\" (UniqueName: \"kubernetes.io/projected/eb5ba94e-890e-4791-b2f6-e30d1d92ed96-kube-api-access-n5k8t\") pod \"collect-profiles-29556990-mgpgm\" (UID: \"eb5ba94e-890e-4791-b2f6-e30d1d92ed96\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556990-mgpgm" Mar 13 16:30:00 crc kubenswrapper[4907]: I0313 16:30:00.443964 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/eb5ba94e-890e-4791-b2f6-e30d1d92ed96-config-volume\") pod \"collect-profiles-29556990-mgpgm\" (UID: \"eb5ba94e-890e-4791-b2f6-e30d1d92ed96\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556990-mgpgm" Mar 13 16:30:00 crc kubenswrapper[4907]: I0313 16:30:00.444915 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/eb5ba94e-890e-4791-b2f6-e30d1d92ed96-config-volume\") pod \"collect-profiles-29556990-mgpgm\" (UID: \"eb5ba94e-890e-4791-b2f6-e30d1d92ed96\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556990-mgpgm" Mar 13 16:30:00 crc kubenswrapper[4907]: I0313 16:30:00.471917 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/eb5ba94e-890e-4791-b2f6-e30d1d92ed96-secret-volume\") pod \"collect-profiles-29556990-mgpgm\" (UID: \"eb5ba94e-890e-4791-b2f6-e30d1d92ed96\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556990-mgpgm" Mar 13 16:30:00 crc kubenswrapper[4907]: I0313 16:30:00.474753 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-86rp9\" (UniqueName: \"kubernetes.io/projected/5a413954-e701-4cec-8868-4fcdaecfabe8-kube-api-access-86rp9\") pod \"auto-csr-approver-29556990-4jrg4\" (UID: \"5a413954-e701-4cec-8868-4fcdaecfabe8\") " pod="openshift-infra/auto-csr-approver-29556990-4jrg4" Mar 13 16:30:00 crc kubenswrapper[4907]: I0313 16:30:00.475651 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n5k8t\" (UniqueName: \"kubernetes.io/projected/eb5ba94e-890e-4791-b2f6-e30d1d92ed96-kube-api-access-n5k8t\") pod \"collect-profiles-29556990-mgpgm\" (UID: \"eb5ba94e-890e-4791-b2f6-e30d1d92ed96\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556990-mgpgm" Mar 13 16:30:00 crc kubenswrapper[4907]: I0313 16:30:00.501167 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556990-mgpgm" Mar 13 16:30:00 crc kubenswrapper[4907]: I0313 16:30:00.511837 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556990-4jrg4" Mar 13 16:30:00 crc kubenswrapper[4907]: I0313 16:30:00.964338 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556990-mgpgm"] Mar 13 16:30:01 crc kubenswrapper[4907]: I0313 16:30:01.056538 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556990-mgpgm" event={"ID":"eb5ba94e-890e-4791-b2f6-e30d1d92ed96","Type":"ContainerStarted","Data":"88f370a6bce1c1e14ee860380282bda5a2ea2cf44f416fcec26c54a386636978"} Mar 13 16:30:01 crc kubenswrapper[4907]: W0313 16:30:01.081283 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5a413954_e701_4cec_8868_4fcdaecfabe8.slice/crio-7efa7e899bdff3abb97dc670393bea14cc20267410e92c626acaeba445e97a71 WatchSource:0}: Error finding container 7efa7e899bdff3abb97dc670393bea14cc20267410e92c626acaeba445e97a71: Status 404 returned error can't find the container with id 7efa7e899bdff3abb97dc670393bea14cc20267410e92c626acaeba445e97a71 Mar 13 16:30:01 crc kubenswrapper[4907]: I0313 16:30:01.086726 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556990-4jrg4"] Mar 13 16:30:02 crc kubenswrapper[4907]: I0313 16:30:02.074316 4907 generic.go:334] "Generic (PLEG): container finished" podID="eb5ba94e-890e-4791-b2f6-e30d1d92ed96" containerID="403679ef0fad99e187e6956860663a7b80495194144d0c59af9ade6f000f4fe6" exitCode=0 Mar 13 16:30:02 crc kubenswrapper[4907]: I0313 16:30:02.074403 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556990-mgpgm" event={"ID":"eb5ba94e-890e-4791-b2f6-e30d1d92ed96","Type":"ContainerDied","Data":"403679ef0fad99e187e6956860663a7b80495194144d0c59af9ade6f000f4fe6"} Mar 13 16:30:02 crc kubenswrapper[4907]: I0313 16:30:02.078136 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556990-4jrg4" event={"ID":"5a413954-e701-4cec-8868-4fcdaecfabe8","Type":"ContainerStarted","Data":"7efa7e899bdff3abb97dc670393bea14cc20267410e92c626acaeba445e97a71"} Mar 13 16:30:03 crc kubenswrapper[4907]: I0313 16:30:03.484830 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556990-mgpgm" Mar 13 16:30:03 crc kubenswrapper[4907]: I0313 16:30:03.617209 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/eb5ba94e-890e-4791-b2f6-e30d1d92ed96-config-volume\") pod \"eb5ba94e-890e-4791-b2f6-e30d1d92ed96\" (UID: \"eb5ba94e-890e-4791-b2f6-e30d1d92ed96\") " Mar 13 16:30:03 crc kubenswrapper[4907]: I0313 16:30:03.617262 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n5k8t\" (UniqueName: \"kubernetes.io/projected/eb5ba94e-890e-4791-b2f6-e30d1d92ed96-kube-api-access-n5k8t\") pod \"eb5ba94e-890e-4791-b2f6-e30d1d92ed96\" (UID: \"eb5ba94e-890e-4791-b2f6-e30d1d92ed96\") " Mar 13 16:30:03 crc kubenswrapper[4907]: I0313 16:30:03.617307 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/eb5ba94e-890e-4791-b2f6-e30d1d92ed96-secret-volume\") pod \"eb5ba94e-890e-4791-b2f6-e30d1d92ed96\" (UID: \"eb5ba94e-890e-4791-b2f6-e30d1d92ed96\") " Mar 13 16:30:03 crc kubenswrapper[4907]: I0313 16:30:03.618531 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eb5ba94e-890e-4791-b2f6-e30d1d92ed96-config-volume" (OuterVolumeSpecName: "config-volume") pod "eb5ba94e-890e-4791-b2f6-e30d1d92ed96" (UID: "eb5ba94e-890e-4791-b2f6-e30d1d92ed96"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 16:30:03 crc kubenswrapper[4907]: I0313 16:30:03.626503 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eb5ba94e-890e-4791-b2f6-e30d1d92ed96-kube-api-access-n5k8t" (OuterVolumeSpecName: "kube-api-access-n5k8t") pod "eb5ba94e-890e-4791-b2f6-e30d1d92ed96" (UID: "eb5ba94e-890e-4791-b2f6-e30d1d92ed96"). InnerVolumeSpecName "kube-api-access-n5k8t". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:30:03 crc kubenswrapper[4907]: I0313 16:30:03.626526 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb5ba94e-890e-4791-b2f6-e30d1d92ed96-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "eb5ba94e-890e-4791-b2f6-e30d1d92ed96" (UID: "eb5ba94e-890e-4791-b2f6-e30d1d92ed96"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:30:03 crc kubenswrapper[4907]: I0313 16:30:03.719996 4907 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/eb5ba94e-890e-4791-b2f6-e30d1d92ed96-config-volume\") on node \"crc\" DevicePath \"\"" Mar 13 16:30:03 crc kubenswrapper[4907]: I0313 16:30:03.720028 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n5k8t\" (UniqueName: \"kubernetes.io/projected/eb5ba94e-890e-4791-b2f6-e30d1d92ed96-kube-api-access-n5k8t\") on node \"crc\" DevicePath \"\"" Mar 13 16:30:03 crc kubenswrapper[4907]: I0313 16:30:03.720040 4907 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/eb5ba94e-890e-4791-b2f6-e30d1d92ed96-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 13 16:30:04 crc kubenswrapper[4907]: I0313 16:30:04.102276 4907 generic.go:334] "Generic (PLEG): container finished" podID="5a413954-e701-4cec-8868-4fcdaecfabe8" containerID="5585587b3788c101f53b09ee74a77a65c4d4a4773da4d3e62432619e4599efa1" exitCode=0 Mar 13 16:30:04 crc kubenswrapper[4907]: I0313 16:30:04.102392 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556990-4jrg4" event={"ID":"5a413954-e701-4cec-8868-4fcdaecfabe8","Type":"ContainerDied","Data":"5585587b3788c101f53b09ee74a77a65c4d4a4773da4d3e62432619e4599efa1"} Mar 13 16:30:04 crc kubenswrapper[4907]: I0313 16:30:04.106565 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556990-mgpgm" event={"ID":"eb5ba94e-890e-4791-b2f6-e30d1d92ed96","Type":"ContainerDied","Data":"88f370a6bce1c1e14ee860380282bda5a2ea2cf44f416fcec26c54a386636978"} Mar 13 16:30:04 crc kubenswrapper[4907]: I0313 16:30:04.106613 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="88f370a6bce1c1e14ee860380282bda5a2ea2cf44f416fcec26c54a386636978" Mar 13 16:30:04 crc kubenswrapper[4907]: I0313 16:30:04.106630 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556990-mgpgm" Mar 13 16:30:04 crc kubenswrapper[4907]: I0313 16:30:04.608007 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556945-vp2lp"] Mar 13 16:30:04 crc kubenswrapper[4907]: I0313 16:30:04.622448 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556945-vp2lp"] Mar 13 16:30:05 crc kubenswrapper[4907]: I0313 16:30:05.575873 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556990-4jrg4" Mar 13 16:30:05 crc kubenswrapper[4907]: I0313 16:30:05.666767 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-86rp9\" (UniqueName: \"kubernetes.io/projected/5a413954-e701-4cec-8868-4fcdaecfabe8-kube-api-access-86rp9\") pod \"5a413954-e701-4cec-8868-4fcdaecfabe8\" (UID: \"5a413954-e701-4cec-8868-4fcdaecfabe8\") " Mar 13 16:30:05 crc kubenswrapper[4907]: I0313 16:30:05.675135 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5a413954-e701-4cec-8868-4fcdaecfabe8-kube-api-access-86rp9" (OuterVolumeSpecName: "kube-api-access-86rp9") pod "5a413954-e701-4cec-8868-4fcdaecfabe8" (UID: "5a413954-e701-4cec-8868-4fcdaecfabe8"). InnerVolumeSpecName "kube-api-access-86rp9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:30:05 crc kubenswrapper[4907]: I0313 16:30:05.769161 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-86rp9\" (UniqueName: \"kubernetes.io/projected/5a413954-e701-4cec-8868-4fcdaecfabe8-kube-api-access-86rp9\") on node \"crc\" DevicePath \"\"" Mar 13 16:30:05 crc kubenswrapper[4907]: I0313 16:30:05.796554 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="329cddab-6958-4dbb-a2d0-baef50885413" path="/var/lib/kubelet/pods/329cddab-6958-4dbb-a2d0-baef50885413/volumes" Mar 13 16:30:06 crc kubenswrapper[4907]: I0313 16:30:06.131087 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556990-4jrg4" event={"ID":"5a413954-e701-4cec-8868-4fcdaecfabe8","Type":"ContainerDied","Data":"7efa7e899bdff3abb97dc670393bea14cc20267410e92c626acaeba445e97a71"} Mar 13 16:30:06 crc kubenswrapper[4907]: I0313 16:30:06.131138 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556990-4jrg4" Mar 13 16:30:06 crc kubenswrapper[4907]: I0313 16:30:06.131152 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7efa7e899bdff3abb97dc670393bea14cc20267410e92c626acaeba445e97a71" Mar 13 16:30:06 crc kubenswrapper[4907]: I0313 16:30:06.625504 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556984-2lfjr"] Mar 13 16:30:06 crc kubenswrapper[4907]: I0313 16:30:06.633944 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556984-2lfjr"] Mar 13 16:30:07 crc kubenswrapper[4907]: I0313 16:30:07.797451 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d0090a6a-15d8-4445-9830-9cab1aec7fdb" path="/var/lib/kubelet/pods/d0090a6a-15d8-4445-9830-9cab1aec7fdb/volumes" Mar 13 16:30:10 crc kubenswrapper[4907]: I0313 16:30:10.783680 4907 scope.go:117] "RemoveContainer" containerID="dfb327c310937cae28a687e8c5c0256a2ab50e59a756669f2586112310e1dc69" Mar 13 16:30:10 crc kubenswrapper[4907]: E0313 16:30:10.784626 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:30:25 crc kubenswrapper[4907]: I0313 16:30:25.783502 4907 scope.go:117] "RemoveContainer" containerID="dfb327c310937cae28a687e8c5c0256a2ab50e59a756669f2586112310e1dc69" Mar 13 16:30:25 crc kubenswrapper[4907]: E0313 16:30:25.784568 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:30:36 crc kubenswrapper[4907]: I0313 16:30:36.782854 4907 scope.go:117] "RemoveContainer" containerID="dfb327c310937cae28a687e8c5c0256a2ab50e59a756669f2586112310e1dc69" Mar 13 16:30:36 crc kubenswrapper[4907]: E0313 16:30:36.784185 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:30:43 crc kubenswrapper[4907]: I0313 16:30:43.496183 4907 scope.go:117] "RemoveContainer" containerID="65a8dfeb436104f422bc6ebeae3f1c135d2d01201103445d064625f2317603ad" Mar 13 16:30:43 crc kubenswrapper[4907]: I0313 16:30:43.567577 4907 scope.go:117] "RemoveContainer" containerID="5fb59125e06d01b85c52baad33f05fd7f12c902ae1f36e7af1e62fdbcb0056b5" Mar 13 16:30:51 crc kubenswrapper[4907]: I0313 16:30:51.802068 4907 scope.go:117] "RemoveContainer" containerID="dfb327c310937cae28a687e8c5c0256a2ab50e59a756669f2586112310e1dc69" Mar 13 16:30:51 crc kubenswrapper[4907]: E0313 16:30:51.803387 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:31:06 crc kubenswrapper[4907]: I0313 16:31:06.783062 4907 scope.go:117] "RemoveContainer" containerID="dfb327c310937cae28a687e8c5c0256a2ab50e59a756669f2586112310e1dc69" Mar 13 16:31:06 crc kubenswrapper[4907]: E0313 16:31:06.784326 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:31:21 crc kubenswrapper[4907]: I0313 16:31:21.796193 4907 scope.go:117] "RemoveContainer" containerID="dfb327c310937cae28a687e8c5c0256a2ab50e59a756669f2586112310e1dc69" Mar 13 16:31:21 crc kubenswrapper[4907]: E0313 16:31:21.797018 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:31:33 crc kubenswrapper[4907]: I0313 16:31:33.782874 4907 scope.go:117] "RemoveContainer" containerID="dfb327c310937cae28a687e8c5c0256a2ab50e59a756669f2586112310e1dc69" Mar 13 16:31:33 crc kubenswrapper[4907]: E0313 16:31:33.783849 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:31:48 crc kubenswrapper[4907]: I0313 16:31:48.782815 4907 scope.go:117] "RemoveContainer" containerID="dfb327c310937cae28a687e8c5c0256a2ab50e59a756669f2586112310e1dc69" Mar 13 16:31:48 crc kubenswrapper[4907]: E0313 16:31:48.783609 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:31:49 crc kubenswrapper[4907]: I0313 16:31:49.105821 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-bfz29"] Mar 13 16:31:49 crc kubenswrapper[4907]: E0313 16:31:49.106575 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a413954-e701-4cec-8868-4fcdaecfabe8" containerName="oc" Mar 13 16:31:49 crc kubenswrapper[4907]: I0313 16:31:49.106601 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a413954-e701-4cec-8868-4fcdaecfabe8" containerName="oc" Mar 13 16:31:49 crc kubenswrapper[4907]: E0313 16:31:49.106632 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb5ba94e-890e-4791-b2f6-e30d1d92ed96" containerName="collect-profiles" Mar 13 16:31:49 crc kubenswrapper[4907]: I0313 16:31:49.106641 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb5ba94e-890e-4791-b2f6-e30d1d92ed96" containerName="collect-profiles" Mar 13 16:31:49 crc kubenswrapper[4907]: I0313 16:31:49.106933 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="eb5ba94e-890e-4791-b2f6-e30d1d92ed96" containerName="collect-profiles" Mar 13 16:31:49 crc kubenswrapper[4907]: I0313 16:31:49.106968 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a413954-e701-4cec-8868-4fcdaecfabe8" containerName="oc" Mar 13 16:31:49 crc kubenswrapper[4907]: I0313 16:31:49.108920 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bfz29" Mar 13 16:31:49 crc kubenswrapper[4907]: I0313 16:31:49.121622 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bfz29"] Mar 13 16:31:49 crc kubenswrapper[4907]: I0313 16:31:49.159552 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xmch8\" (UniqueName: \"kubernetes.io/projected/fed253af-6b1e-49b2-9d21-4467f33ecf94-kube-api-access-xmch8\") pod \"redhat-operators-bfz29\" (UID: \"fed253af-6b1e-49b2-9d21-4467f33ecf94\") " pod="openshift-marketplace/redhat-operators-bfz29" Mar 13 16:31:49 crc kubenswrapper[4907]: I0313 16:31:49.159769 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fed253af-6b1e-49b2-9d21-4467f33ecf94-utilities\") pod \"redhat-operators-bfz29\" (UID: \"fed253af-6b1e-49b2-9d21-4467f33ecf94\") " pod="openshift-marketplace/redhat-operators-bfz29" Mar 13 16:31:49 crc kubenswrapper[4907]: I0313 16:31:49.159956 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fed253af-6b1e-49b2-9d21-4467f33ecf94-catalog-content\") pod \"redhat-operators-bfz29\" (UID: \"fed253af-6b1e-49b2-9d21-4467f33ecf94\") " pod="openshift-marketplace/redhat-operators-bfz29" Mar 13 16:31:49 crc kubenswrapper[4907]: I0313 16:31:49.262430 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xmch8\" (UniqueName: \"kubernetes.io/projected/fed253af-6b1e-49b2-9d21-4467f33ecf94-kube-api-access-xmch8\") pod \"redhat-operators-bfz29\" (UID: \"fed253af-6b1e-49b2-9d21-4467f33ecf94\") " pod="openshift-marketplace/redhat-operators-bfz29" Mar 13 16:31:49 crc kubenswrapper[4907]: I0313 16:31:49.262534 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fed253af-6b1e-49b2-9d21-4467f33ecf94-utilities\") pod \"redhat-operators-bfz29\" (UID: \"fed253af-6b1e-49b2-9d21-4467f33ecf94\") " pod="openshift-marketplace/redhat-operators-bfz29" Mar 13 16:31:49 crc kubenswrapper[4907]: I0313 16:31:49.262608 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fed253af-6b1e-49b2-9d21-4467f33ecf94-catalog-content\") pod \"redhat-operators-bfz29\" (UID: \"fed253af-6b1e-49b2-9d21-4467f33ecf94\") " pod="openshift-marketplace/redhat-operators-bfz29" Mar 13 16:31:49 crc kubenswrapper[4907]: I0313 16:31:49.263113 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fed253af-6b1e-49b2-9d21-4467f33ecf94-catalog-content\") pod \"redhat-operators-bfz29\" (UID: \"fed253af-6b1e-49b2-9d21-4467f33ecf94\") " pod="openshift-marketplace/redhat-operators-bfz29" Mar 13 16:31:49 crc kubenswrapper[4907]: I0313 16:31:49.263538 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fed253af-6b1e-49b2-9d21-4467f33ecf94-utilities\") pod \"redhat-operators-bfz29\" (UID: \"fed253af-6b1e-49b2-9d21-4467f33ecf94\") " pod="openshift-marketplace/redhat-operators-bfz29" Mar 13 16:31:49 crc kubenswrapper[4907]: I0313 16:31:49.295802 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xmch8\" (UniqueName: \"kubernetes.io/projected/fed253af-6b1e-49b2-9d21-4467f33ecf94-kube-api-access-xmch8\") pod \"redhat-operators-bfz29\" (UID: \"fed253af-6b1e-49b2-9d21-4467f33ecf94\") " pod="openshift-marketplace/redhat-operators-bfz29" Mar 13 16:31:49 crc kubenswrapper[4907]: I0313 16:31:49.429108 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bfz29" Mar 13 16:31:49 crc kubenswrapper[4907]: I0313 16:31:49.967756 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bfz29"] Mar 13 16:31:50 crc kubenswrapper[4907]: I0313 16:31:50.272806 4907 generic.go:334] "Generic (PLEG): container finished" podID="fed253af-6b1e-49b2-9d21-4467f33ecf94" containerID="80c7c01d67fbd09368a3d7aec0c7facbc1ce68808d7cf94f486937495c9b6a4c" exitCode=0 Mar 13 16:31:50 crc kubenswrapper[4907]: I0313 16:31:50.272858 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bfz29" event={"ID":"fed253af-6b1e-49b2-9d21-4467f33ecf94","Type":"ContainerDied","Data":"80c7c01d67fbd09368a3d7aec0c7facbc1ce68808d7cf94f486937495c9b6a4c"} Mar 13 16:31:50 crc kubenswrapper[4907]: I0313 16:31:50.273105 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bfz29" event={"ID":"fed253af-6b1e-49b2-9d21-4467f33ecf94","Type":"ContainerStarted","Data":"6843dee463e8aa2139eea25e2bbbae5a7e3cae3bb8c1d0992495315ec9382157"} Mar 13 16:31:52 crc kubenswrapper[4907]: I0313 16:31:52.294818 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bfz29" event={"ID":"fed253af-6b1e-49b2-9d21-4467f33ecf94","Type":"ContainerStarted","Data":"b100747e5d6556c4ee784efb31b617f6c9c90e6075eb0a8d0df0270a93d3d2df"} Mar 13 16:31:55 crc kubenswrapper[4907]: I0313 16:31:55.325789 4907 generic.go:334] "Generic (PLEG): container finished" podID="fed253af-6b1e-49b2-9d21-4467f33ecf94" containerID="b100747e5d6556c4ee784efb31b617f6c9c90e6075eb0a8d0df0270a93d3d2df" exitCode=0 Mar 13 16:31:55 crc kubenswrapper[4907]: I0313 16:31:55.325854 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bfz29" event={"ID":"fed253af-6b1e-49b2-9d21-4467f33ecf94","Type":"ContainerDied","Data":"b100747e5d6556c4ee784efb31b617f6c9c90e6075eb0a8d0df0270a93d3d2df"} Mar 13 16:31:56 crc kubenswrapper[4907]: I0313 16:31:56.341446 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bfz29" event={"ID":"fed253af-6b1e-49b2-9d21-4467f33ecf94","Type":"ContainerStarted","Data":"9064e0377ba2c4d96cb954df5a4d9d8e4208784b7270f9410b0d253f7b5f4410"} Mar 13 16:31:56 crc kubenswrapper[4907]: I0313 16:31:56.379839 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-bfz29" podStartSLOduration=1.901810102 podStartE2EDuration="7.379818269s" podCreationTimestamp="2026-03-13 16:31:49 +0000 UTC" firstStartedPulling="2026-03-13 16:31:50.274424365 +0000 UTC m=+8809.174212054" lastFinishedPulling="2026-03-13 16:31:55.752432532 +0000 UTC m=+8814.652220221" observedRunningTime="2026-03-13 16:31:56.366947947 +0000 UTC m=+8815.266735686" watchObservedRunningTime="2026-03-13 16:31:56.379818269 +0000 UTC m=+8815.279605968" Mar 13 16:31:59 crc kubenswrapper[4907]: I0313 16:31:59.430352 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-bfz29" Mar 13 16:31:59 crc kubenswrapper[4907]: I0313 16:31:59.430797 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-bfz29" Mar 13 16:31:59 crc kubenswrapper[4907]: I0313 16:31:59.783794 4907 scope.go:117] "RemoveContainer" containerID="dfb327c310937cae28a687e8c5c0256a2ab50e59a756669f2586112310e1dc69" Mar 13 16:31:59 crc kubenswrapper[4907]: E0313 16:31:59.784396 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:32:00 crc kubenswrapper[4907]: I0313 16:32:00.145605 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556992-qj2mm"] Mar 13 16:32:00 crc kubenswrapper[4907]: I0313 16:32:00.147008 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556992-qj2mm" Mar 13 16:32:00 crc kubenswrapper[4907]: I0313 16:32:00.154977 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 16:32:00 crc kubenswrapper[4907]: I0313 16:32:00.155071 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 16:32:00 crc kubenswrapper[4907]: I0313 16:32:00.155302 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 16:32:00 crc kubenswrapper[4907]: I0313 16:32:00.159648 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556992-qj2mm"] Mar 13 16:32:00 crc kubenswrapper[4907]: I0313 16:32:00.209644 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d55s6\" (UniqueName: \"kubernetes.io/projected/5d2dfd3f-0bc6-48f9-85f7-100ef8f23476-kube-api-access-d55s6\") pod \"auto-csr-approver-29556992-qj2mm\" (UID: \"5d2dfd3f-0bc6-48f9-85f7-100ef8f23476\") " pod="openshift-infra/auto-csr-approver-29556992-qj2mm" Mar 13 16:32:00 crc kubenswrapper[4907]: I0313 16:32:00.311495 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d55s6\" (UniqueName: \"kubernetes.io/projected/5d2dfd3f-0bc6-48f9-85f7-100ef8f23476-kube-api-access-d55s6\") pod \"auto-csr-approver-29556992-qj2mm\" (UID: \"5d2dfd3f-0bc6-48f9-85f7-100ef8f23476\") " pod="openshift-infra/auto-csr-approver-29556992-qj2mm" Mar 13 16:32:00 crc kubenswrapper[4907]: I0313 16:32:00.338747 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d55s6\" (UniqueName: \"kubernetes.io/projected/5d2dfd3f-0bc6-48f9-85f7-100ef8f23476-kube-api-access-d55s6\") pod \"auto-csr-approver-29556992-qj2mm\" (UID: \"5d2dfd3f-0bc6-48f9-85f7-100ef8f23476\") " pod="openshift-infra/auto-csr-approver-29556992-qj2mm" Mar 13 16:32:00 crc kubenswrapper[4907]: I0313 16:32:00.468261 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556992-qj2mm" Mar 13 16:32:00 crc kubenswrapper[4907]: I0313 16:32:00.490244 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-bfz29" podUID="fed253af-6b1e-49b2-9d21-4467f33ecf94" containerName="registry-server" probeResult="failure" output=< Mar 13 16:32:00 crc kubenswrapper[4907]: timeout: failed to connect service ":50051" within 1s Mar 13 16:32:00 crc kubenswrapper[4907]: > Mar 13 16:32:00 crc kubenswrapper[4907]: I0313 16:32:00.969924 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556992-qj2mm"] Mar 13 16:32:01 crc kubenswrapper[4907]: I0313 16:32:01.416036 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556992-qj2mm" event={"ID":"5d2dfd3f-0bc6-48f9-85f7-100ef8f23476","Type":"ContainerStarted","Data":"8b6a97332a7098fe96c5bc5d67ef2f10bf6399130d709ccb2cc03a0dd3fd7fc1"} Mar 13 16:32:02 crc kubenswrapper[4907]: I0313 16:32:02.426758 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556992-qj2mm" event={"ID":"5d2dfd3f-0bc6-48f9-85f7-100ef8f23476","Type":"ContainerStarted","Data":"a1001aa55a6abdb3dd8474ef6137b7a29548724c2eff0d7fa57396148bbc916a"} Mar 13 16:32:02 crc kubenswrapper[4907]: I0313 16:32:02.443466 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29556992-qj2mm" podStartSLOduration=1.470477152 podStartE2EDuration="2.443449568s" podCreationTimestamp="2026-03-13 16:32:00 +0000 UTC" firstStartedPulling="2026-03-13 16:32:00.955284381 +0000 UTC m=+8819.855072060" lastFinishedPulling="2026-03-13 16:32:01.928256797 +0000 UTC m=+8820.828044476" observedRunningTime="2026-03-13 16:32:02.439464046 +0000 UTC m=+8821.339251745" watchObservedRunningTime="2026-03-13 16:32:02.443449568 +0000 UTC m=+8821.343237257" Mar 13 16:32:03 crc kubenswrapper[4907]: I0313 16:32:03.441606 4907 generic.go:334] "Generic (PLEG): container finished" podID="5d2dfd3f-0bc6-48f9-85f7-100ef8f23476" containerID="a1001aa55a6abdb3dd8474ef6137b7a29548724c2eff0d7fa57396148bbc916a" exitCode=0 Mar 13 16:32:03 crc kubenswrapper[4907]: I0313 16:32:03.441719 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556992-qj2mm" event={"ID":"5d2dfd3f-0bc6-48f9-85f7-100ef8f23476","Type":"ContainerDied","Data":"a1001aa55a6abdb3dd8474ef6137b7a29548724c2eff0d7fa57396148bbc916a"} Mar 13 16:32:04 crc kubenswrapper[4907]: I0313 16:32:04.897512 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556992-qj2mm" Mar 13 16:32:04 crc kubenswrapper[4907]: I0313 16:32:04.924843 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d55s6\" (UniqueName: \"kubernetes.io/projected/5d2dfd3f-0bc6-48f9-85f7-100ef8f23476-kube-api-access-d55s6\") pod \"5d2dfd3f-0bc6-48f9-85f7-100ef8f23476\" (UID: \"5d2dfd3f-0bc6-48f9-85f7-100ef8f23476\") " Mar 13 16:32:04 crc kubenswrapper[4907]: I0313 16:32:04.934124 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5d2dfd3f-0bc6-48f9-85f7-100ef8f23476-kube-api-access-d55s6" (OuterVolumeSpecName: "kube-api-access-d55s6") pod "5d2dfd3f-0bc6-48f9-85f7-100ef8f23476" (UID: "5d2dfd3f-0bc6-48f9-85f7-100ef8f23476"). InnerVolumeSpecName "kube-api-access-d55s6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:32:05 crc kubenswrapper[4907]: I0313 16:32:05.028230 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d55s6\" (UniqueName: \"kubernetes.io/projected/5d2dfd3f-0bc6-48f9-85f7-100ef8f23476-kube-api-access-d55s6\") on node \"crc\" DevicePath \"\"" Mar 13 16:32:05 crc kubenswrapper[4907]: I0313 16:32:05.471334 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556992-qj2mm" event={"ID":"5d2dfd3f-0bc6-48f9-85f7-100ef8f23476","Type":"ContainerDied","Data":"8b6a97332a7098fe96c5bc5d67ef2f10bf6399130d709ccb2cc03a0dd3fd7fc1"} Mar 13 16:32:05 crc kubenswrapper[4907]: I0313 16:32:05.471380 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8b6a97332a7098fe96c5bc5d67ef2f10bf6399130d709ccb2cc03a0dd3fd7fc1" Mar 13 16:32:05 crc kubenswrapper[4907]: I0313 16:32:05.471483 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556992-qj2mm" Mar 13 16:32:05 crc kubenswrapper[4907]: I0313 16:32:05.996536 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556986-nlfck"] Mar 13 16:32:06 crc kubenswrapper[4907]: I0313 16:32:06.028155 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556986-nlfck"] Mar 13 16:32:07 crc kubenswrapper[4907]: I0313 16:32:07.799025 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bfa4f11-b5c4-4767-bb8e-3940faa2a5f4" path="/var/lib/kubelet/pods/1bfa4f11-b5c4-4767-bb8e-3940faa2a5f4/volumes" Mar 13 16:32:09 crc kubenswrapper[4907]: I0313 16:32:09.511270 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-bfz29" Mar 13 16:32:09 crc kubenswrapper[4907]: I0313 16:32:09.577505 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-bfz29" Mar 13 16:32:09 crc kubenswrapper[4907]: I0313 16:32:09.754097 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-bfz29"] Mar 13 16:32:11 crc kubenswrapper[4907]: I0313 16:32:11.545779 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-bfz29" podUID="fed253af-6b1e-49b2-9d21-4467f33ecf94" containerName="registry-server" containerID="cri-o://9064e0377ba2c4d96cb954df5a4d9d8e4208784b7270f9410b0d253f7b5f4410" gracePeriod=2 Mar 13 16:32:12 crc kubenswrapper[4907]: I0313 16:32:12.243094 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bfz29" Mar 13 16:32:12 crc kubenswrapper[4907]: I0313 16:32:12.397375 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fed253af-6b1e-49b2-9d21-4467f33ecf94-utilities\") pod \"fed253af-6b1e-49b2-9d21-4467f33ecf94\" (UID: \"fed253af-6b1e-49b2-9d21-4467f33ecf94\") " Mar 13 16:32:12 crc kubenswrapper[4907]: I0313 16:32:12.397639 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fed253af-6b1e-49b2-9d21-4467f33ecf94-catalog-content\") pod \"fed253af-6b1e-49b2-9d21-4467f33ecf94\" (UID: \"fed253af-6b1e-49b2-9d21-4467f33ecf94\") " Mar 13 16:32:12 crc kubenswrapper[4907]: I0313 16:32:12.397855 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xmch8\" (UniqueName: \"kubernetes.io/projected/fed253af-6b1e-49b2-9d21-4467f33ecf94-kube-api-access-xmch8\") pod \"fed253af-6b1e-49b2-9d21-4467f33ecf94\" (UID: \"fed253af-6b1e-49b2-9d21-4467f33ecf94\") " Mar 13 16:32:12 crc kubenswrapper[4907]: I0313 16:32:12.399243 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fed253af-6b1e-49b2-9d21-4467f33ecf94-utilities" (OuterVolumeSpecName: "utilities") pod "fed253af-6b1e-49b2-9d21-4467f33ecf94" (UID: "fed253af-6b1e-49b2-9d21-4467f33ecf94"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 16:32:12 crc kubenswrapper[4907]: I0313 16:32:12.405808 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fed253af-6b1e-49b2-9d21-4467f33ecf94-kube-api-access-xmch8" (OuterVolumeSpecName: "kube-api-access-xmch8") pod "fed253af-6b1e-49b2-9d21-4467f33ecf94" (UID: "fed253af-6b1e-49b2-9d21-4467f33ecf94"). InnerVolumeSpecName "kube-api-access-xmch8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:32:12 crc kubenswrapper[4907]: I0313 16:32:12.501593 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fed253af-6b1e-49b2-9d21-4467f33ecf94-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 16:32:12 crc kubenswrapper[4907]: I0313 16:32:12.501635 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xmch8\" (UniqueName: \"kubernetes.io/projected/fed253af-6b1e-49b2-9d21-4467f33ecf94-kube-api-access-xmch8\") on node \"crc\" DevicePath \"\"" Mar 13 16:32:12 crc kubenswrapper[4907]: I0313 16:32:12.561330 4907 generic.go:334] "Generic (PLEG): container finished" podID="fed253af-6b1e-49b2-9d21-4467f33ecf94" containerID="9064e0377ba2c4d96cb954df5a4d9d8e4208784b7270f9410b0d253f7b5f4410" exitCode=0 Mar 13 16:32:12 crc kubenswrapper[4907]: I0313 16:32:12.561377 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bfz29" event={"ID":"fed253af-6b1e-49b2-9d21-4467f33ecf94","Type":"ContainerDied","Data":"9064e0377ba2c4d96cb954df5a4d9d8e4208784b7270f9410b0d253f7b5f4410"} Mar 13 16:32:12 crc kubenswrapper[4907]: I0313 16:32:12.561416 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bfz29" event={"ID":"fed253af-6b1e-49b2-9d21-4467f33ecf94","Type":"ContainerDied","Data":"6843dee463e8aa2139eea25e2bbbae5a7e3cae3bb8c1d0992495315ec9382157"} Mar 13 16:32:12 crc kubenswrapper[4907]: I0313 16:32:12.561437 4907 scope.go:117] "RemoveContainer" containerID="9064e0377ba2c4d96cb954df5a4d9d8e4208784b7270f9410b0d253f7b5f4410" Mar 13 16:32:12 crc kubenswrapper[4907]: I0313 16:32:12.561632 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bfz29" Mar 13 16:32:12 crc kubenswrapper[4907]: I0313 16:32:12.571282 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fed253af-6b1e-49b2-9d21-4467f33ecf94-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fed253af-6b1e-49b2-9d21-4467f33ecf94" (UID: "fed253af-6b1e-49b2-9d21-4467f33ecf94"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 16:32:12 crc kubenswrapper[4907]: I0313 16:32:12.597505 4907 scope.go:117] "RemoveContainer" containerID="b100747e5d6556c4ee784efb31b617f6c9c90e6075eb0a8d0df0270a93d3d2df" Mar 13 16:32:12 crc kubenswrapper[4907]: I0313 16:32:12.606273 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fed253af-6b1e-49b2-9d21-4467f33ecf94-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 16:32:12 crc kubenswrapper[4907]: I0313 16:32:12.634012 4907 scope.go:117] "RemoveContainer" containerID="80c7c01d67fbd09368a3d7aec0c7facbc1ce68808d7cf94f486937495c9b6a4c" Mar 13 16:32:12 crc kubenswrapper[4907]: I0313 16:32:12.709541 4907 scope.go:117] "RemoveContainer" containerID="9064e0377ba2c4d96cb954df5a4d9d8e4208784b7270f9410b0d253f7b5f4410" Mar 13 16:32:12 crc kubenswrapper[4907]: E0313 16:32:12.711511 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9064e0377ba2c4d96cb954df5a4d9d8e4208784b7270f9410b0d253f7b5f4410\": container with ID starting with 9064e0377ba2c4d96cb954df5a4d9d8e4208784b7270f9410b0d253f7b5f4410 not found: ID does not exist" containerID="9064e0377ba2c4d96cb954df5a4d9d8e4208784b7270f9410b0d253f7b5f4410" Mar 13 16:32:12 crc kubenswrapper[4907]: I0313 16:32:12.711607 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9064e0377ba2c4d96cb954df5a4d9d8e4208784b7270f9410b0d253f7b5f4410"} err="failed to get container status \"9064e0377ba2c4d96cb954df5a4d9d8e4208784b7270f9410b0d253f7b5f4410\": rpc error: code = NotFound desc = could not find container \"9064e0377ba2c4d96cb954df5a4d9d8e4208784b7270f9410b0d253f7b5f4410\": container with ID starting with 9064e0377ba2c4d96cb954df5a4d9d8e4208784b7270f9410b0d253f7b5f4410 not found: ID does not exist" Mar 13 16:32:12 crc kubenswrapper[4907]: I0313 16:32:12.711644 4907 scope.go:117] "RemoveContainer" containerID="b100747e5d6556c4ee784efb31b617f6c9c90e6075eb0a8d0df0270a93d3d2df" Mar 13 16:32:12 crc kubenswrapper[4907]: E0313 16:32:12.712232 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b100747e5d6556c4ee784efb31b617f6c9c90e6075eb0a8d0df0270a93d3d2df\": container with ID starting with b100747e5d6556c4ee784efb31b617f6c9c90e6075eb0a8d0df0270a93d3d2df not found: ID does not exist" containerID="b100747e5d6556c4ee784efb31b617f6c9c90e6075eb0a8d0df0270a93d3d2df" Mar 13 16:32:12 crc kubenswrapper[4907]: I0313 16:32:12.712266 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b100747e5d6556c4ee784efb31b617f6c9c90e6075eb0a8d0df0270a93d3d2df"} err="failed to get container status \"b100747e5d6556c4ee784efb31b617f6c9c90e6075eb0a8d0df0270a93d3d2df\": rpc error: code = NotFound desc = could not find container \"b100747e5d6556c4ee784efb31b617f6c9c90e6075eb0a8d0df0270a93d3d2df\": container with ID starting with b100747e5d6556c4ee784efb31b617f6c9c90e6075eb0a8d0df0270a93d3d2df not found: ID does not exist" Mar 13 16:32:12 crc kubenswrapper[4907]: I0313 16:32:12.712286 4907 scope.go:117] "RemoveContainer" containerID="80c7c01d67fbd09368a3d7aec0c7facbc1ce68808d7cf94f486937495c9b6a4c" Mar 13 16:32:12 crc kubenswrapper[4907]: E0313 16:32:12.714123 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"80c7c01d67fbd09368a3d7aec0c7facbc1ce68808d7cf94f486937495c9b6a4c\": container with ID starting with 80c7c01d67fbd09368a3d7aec0c7facbc1ce68808d7cf94f486937495c9b6a4c not found: ID does not exist" containerID="80c7c01d67fbd09368a3d7aec0c7facbc1ce68808d7cf94f486937495c9b6a4c" Mar 13 16:32:12 crc kubenswrapper[4907]: I0313 16:32:12.714230 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80c7c01d67fbd09368a3d7aec0c7facbc1ce68808d7cf94f486937495c9b6a4c"} err="failed to get container status \"80c7c01d67fbd09368a3d7aec0c7facbc1ce68808d7cf94f486937495c9b6a4c\": rpc error: code = NotFound desc = could not find container \"80c7c01d67fbd09368a3d7aec0c7facbc1ce68808d7cf94f486937495c9b6a4c\": container with ID starting with 80c7c01d67fbd09368a3d7aec0c7facbc1ce68808d7cf94f486937495c9b6a4c not found: ID does not exist" Mar 13 16:32:12 crc kubenswrapper[4907]: I0313 16:32:12.899500 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-bfz29"] Mar 13 16:32:12 crc kubenswrapper[4907]: I0313 16:32:12.909130 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-bfz29"] Mar 13 16:32:13 crc kubenswrapper[4907]: I0313 16:32:13.796497 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fed253af-6b1e-49b2-9d21-4467f33ecf94" path="/var/lib/kubelet/pods/fed253af-6b1e-49b2-9d21-4467f33ecf94/volumes" Mar 13 16:32:14 crc kubenswrapper[4907]: I0313 16:32:14.782666 4907 scope.go:117] "RemoveContainer" containerID="dfb327c310937cae28a687e8c5c0256a2ab50e59a756669f2586112310e1dc69" Mar 13 16:32:14 crc kubenswrapper[4907]: E0313 16:32:14.783260 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:32:26 crc kubenswrapper[4907]: I0313 16:32:26.783753 4907 scope.go:117] "RemoveContainer" containerID="dfb327c310937cae28a687e8c5c0256a2ab50e59a756669f2586112310e1dc69" Mar 13 16:32:27 crc kubenswrapper[4907]: I0313 16:32:27.749911 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerStarted","Data":"f9156a2a4d423d000cb09f67e16772f675217bb2e483ba02adf57b642ffdffc9"} Mar 13 16:32:43 crc kubenswrapper[4907]: I0313 16:32:43.676293 4907 scope.go:117] "RemoveContainer" containerID="4fc188ddf095682c93e7e1f3f3249b322a5fe91043eda372761cb774c37d43db" Mar 13 16:33:08 crc kubenswrapper[4907]: I0313 16:33:08.223930 4907 generic.go:334] "Generic (PLEG): container finished" podID="de68abba-24e7-49b8-9e1f-adfa05d9d3dd" containerID="8644ffba95037a0d279e1ad31f88e2f9c3fae28fb5fd51859c1fb3e25cec04cc" exitCode=0 Mar 13 16:33:08 crc kubenswrapper[4907]: I0313 16:33:08.224160 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc" event={"ID":"de68abba-24e7-49b8-9e1f-adfa05d9d3dd","Type":"ContainerDied","Data":"8644ffba95037a0d279e1ad31f88e2f9c3fae28fb5fd51859c1fb3e25cec04cc"} Mar 13 16:33:09 crc kubenswrapper[4907]: I0313 16:33:09.706482 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc" Mar 13 16:33:09 crc kubenswrapper[4907]: I0313 16:33:09.841592 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-nova-cell1-compute-config-2\") pod \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\" (UID: \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\") " Mar 13 16:33:09 crc kubenswrapper[4907]: I0313 16:33:09.841653 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-nova-cell1-combined-ca-bundle\") pod \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\" (UID: \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\") " Mar 13 16:33:09 crc kubenswrapper[4907]: I0313 16:33:09.841754 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-njn9g\" (UniqueName: \"kubernetes.io/projected/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-kube-api-access-njn9g\") pod \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\" (UID: \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\") " Mar 13 16:33:09 crc kubenswrapper[4907]: I0313 16:33:09.841805 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-ceph\") pod \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\" (UID: \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\") " Mar 13 16:33:09 crc kubenswrapper[4907]: I0313 16:33:09.841937 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-nova-migration-ssh-key-0\") pod \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\" (UID: \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\") " Mar 13 16:33:09 crc kubenswrapper[4907]: I0313 16:33:09.842009 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-nova-cell1-compute-config-3\") pod \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\" (UID: \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\") " Mar 13 16:33:09 crc kubenswrapper[4907]: I0313 16:33:09.842047 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-nova-cell1-compute-config-0\") pod \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\" (UID: \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\") " Mar 13 16:33:09 crc kubenswrapper[4907]: I0313 16:33:09.842086 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-nova-cells-global-config-0\") pod \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\" (UID: \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\") " Mar 13 16:33:09 crc kubenswrapper[4907]: I0313 16:33:09.842205 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-nova-migration-ssh-key-1\") pod \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\" (UID: \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\") " Mar 13 16:33:09 crc kubenswrapper[4907]: I0313 16:33:09.842259 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-nova-cells-global-config-1\") pod \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\" (UID: \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\") " Mar 13 16:33:09 crc kubenswrapper[4907]: I0313 16:33:09.842300 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-nova-cell1-compute-config-1\") pod \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\" (UID: \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\") " Mar 13 16:33:09 crc kubenswrapper[4907]: I0313 16:33:09.842358 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-inventory\") pod \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\" (UID: \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\") " Mar 13 16:33:09 crc kubenswrapper[4907]: I0313 16:33:09.842447 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-ssh-key-openstack-cell1\") pod \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\" (UID: \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\") " Mar 13 16:33:09 crc kubenswrapper[4907]: I0313 16:33:09.848446 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-ceph" (OuterVolumeSpecName: "ceph") pod "de68abba-24e7-49b8-9e1f-adfa05d9d3dd" (UID: "de68abba-24e7-49b8-9e1f-adfa05d9d3dd"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:33:09 crc kubenswrapper[4907]: I0313 16:33:09.849349 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-kube-api-access-njn9g" (OuterVolumeSpecName: "kube-api-access-njn9g") pod "de68abba-24e7-49b8-9e1f-adfa05d9d3dd" (UID: "de68abba-24e7-49b8-9e1f-adfa05d9d3dd"). InnerVolumeSpecName "kube-api-access-njn9g". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:33:09 crc kubenswrapper[4907]: I0313 16:33:09.852436 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-nova-cell1-combined-ca-bundle" (OuterVolumeSpecName: "nova-cell1-combined-ca-bundle") pod "de68abba-24e7-49b8-9e1f-adfa05d9d3dd" (UID: "de68abba-24e7-49b8-9e1f-adfa05d9d3dd"). InnerVolumeSpecName "nova-cell1-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:33:09 crc kubenswrapper[4907]: I0313 16:33:09.873136 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "de68abba-24e7-49b8-9e1f-adfa05d9d3dd" (UID: "de68abba-24e7-49b8-9e1f-adfa05d9d3dd"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:33:09 crc kubenswrapper[4907]: I0313 16:33:09.875728 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-nova-cells-global-config-0" (OuterVolumeSpecName: "nova-cells-global-config-0") pod "de68abba-24e7-49b8-9e1f-adfa05d9d3dd" (UID: "de68abba-24e7-49b8-9e1f-adfa05d9d3dd"). InnerVolumeSpecName "nova-cells-global-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 16:33:09 crc kubenswrapper[4907]: I0313 16:33:09.894251 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "de68abba-24e7-49b8-9e1f-adfa05d9d3dd" (UID: "de68abba-24e7-49b8-9e1f-adfa05d9d3dd"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:33:09 crc kubenswrapper[4907]: I0313 16:33:09.895769 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "de68abba-24e7-49b8-9e1f-adfa05d9d3dd" (UID: "de68abba-24e7-49b8-9e1f-adfa05d9d3dd"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:33:09 crc kubenswrapper[4907]: I0313 16:33:09.896132 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-inventory" (OuterVolumeSpecName: "inventory") pod "de68abba-24e7-49b8-9e1f-adfa05d9d3dd" (UID: "de68abba-24e7-49b8-9e1f-adfa05d9d3dd"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:33:09 crc kubenswrapper[4907]: I0313 16:33:09.897933 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-ssh-key-openstack-cell1" (OuterVolumeSpecName: "ssh-key-openstack-cell1") pod "de68abba-24e7-49b8-9e1f-adfa05d9d3dd" (UID: "de68abba-24e7-49b8-9e1f-adfa05d9d3dd"). InnerVolumeSpecName "ssh-key-openstack-cell1". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:33:09 crc kubenswrapper[4907]: I0313 16:33:09.904672 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-nova-cells-global-config-1" (OuterVolumeSpecName: "nova-cells-global-config-1") pod "de68abba-24e7-49b8-9e1f-adfa05d9d3dd" (UID: "de68abba-24e7-49b8-9e1f-adfa05d9d3dd"). InnerVolumeSpecName "nova-cells-global-config-1". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 16:33:09 crc kubenswrapper[4907]: I0313 16:33:09.907860 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-nova-cell1-compute-config-3" (OuterVolumeSpecName: "nova-cell1-compute-config-3") pod "de68abba-24e7-49b8-9e1f-adfa05d9d3dd" (UID: "de68abba-24e7-49b8-9e1f-adfa05d9d3dd"). InnerVolumeSpecName "nova-cell1-compute-config-3". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:33:09 crc kubenswrapper[4907]: I0313 16:33:09.909369 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "de68abba-24e7-49b8-9e1f-adfa05d9d3dd" (UID: "de68abba-24e7-49b8-9e1f-adfa05d9d3dd"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:33:09 crc kubenswrapper[4907]: I0313 16:33:09.918096 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-nova-cell1-compute-config-2" (OuterVolumeSpecName: "nova-cell1-compute-config-2") pod "de68abba-24e7-49b8-9e1f-adfa05d9d3dd" (UID: "de68abba-24e7-49b8-9e1f-adfa05d9d3dd"). InnerVolumeSpecName "nova-cell1-compute-config-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:33:09 crc kubenswrapper[4907]: I0313 16:33:09.945662 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-njn9g\" (UniqueName: \"kubernetes.io/projected/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-kube-api-access-njn9g\") on node \"crc\" DevicePath \"\"" Mar 13 16:33:09 crc kubenswrapper[4907]: I0313 16:33:09.945709 4907 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-ceph\") on node \"crc\" DevicePath \"\"" Mar 13 16:33:09 crc kubenswrapper[4907]: I0313 16:33:09.945730 4907 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Mar 13 16:33:09 crc kubenswrapper[4907]: I0313 16:33:09.945748 4907 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-nova-cell1-compute-config-3\") on node \"crc\" DevicePath \"\"" Mar 13 16:33:09 crc kubenswrapper[4907]: I0313 16:33:09.945765 4907 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Mar 13 16:33:09 crc kubenswrapper[4907]: I0313 16:33:09.945783 4907 reconciler_common.go:293] "Volume detached for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-nova-cells-global-config-0\") on node \"crc\" DevicePath \"\"" Mar 13 16:33:09 crc kubenswrapper[4907]: I0313 16:33:09.945798 4907 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Mar 13 16:33:09 crc kubenswrapper[4907]: I0313 16:33:09.945813 4907 reconciler_common.go:293] "Volume detached for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-nova-cells-global-config-1\") on node \"crc\" DevicePath \"\"" Mar 13 16:33:09 crc kubenswrapper[4907]: I0313 16:33:09.945827 4907 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Mar 13 16:33:09 crc kubenswrapper[4907]: I0313 16:33:09.945843 4907 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-inventory\") on node \"crc\" DevicePath \"\"" Mar 13 16:33:09 crc kubenswrapper[4907]: I0313 16:33:09.945858 4907 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-ssh-key-openstack-cell1\") on node \"crc\" DevicePath \"\"" Mar 13 16:33:09 crc kubenswrapper[4907]: I0313 16:33:09.945872 4907 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-nova-cell1-compute-config-2\") on node \"crc\" DevicePath \"\"" Mar 13 16:33:09 crc kubenswrapper[4907]: I0313 16:33:09.945911 4907 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-nova-cell1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 16:33:10 crc kubenswrapper[4907]: I0313 16:33:10.249105 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc" event={"ID":"de68abba-24e7-49b8-9e1f-adfa05d9d3dd","Type":"ContainerDied","Data":"315fdd9184d5c2a2780c5ff803490daf8312afa1aa0a6d15f69dbc1f771bf80c"} Mar 13 16:33:10 crc kubenswrapper[4907]: I0313 16:33:10.249390 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="315fdd9184d5c2a2780c5ff803490daf8312afa1aa0a6d15f69dbc1f771bf80c" Mar 13 16:33:10 crc kubenswrapper[4907]: I0313 16:33:10.249273 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc" Mar 13 16:33:23 crc kubenswrapper[4907]: E0313 16:33:23.493449 4907 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.174:33374->38.102.83.174:33349: write tcp 38.102.83.174:33374->38.102.83.174:33349: write: broken pipe Mar 13 16:34:00 crc kubenswrapper[4907]: I0313 16:34:00.188188 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556994-452h2"] Mar 13 16:34:00 crc kubenswrapper[4907]: E0313 16:34:00.189993 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fed253af-6b1e-49b2-9d21-4467f33ecf94" containerName="extract-utilities" Mar 13 16:34:00 crc kubenswrapper[4907]: I0313 16:34:00.190022 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="fed253af-6b1e-49b2-9d21-4467f33ecf94" containerName="extract-utilities" Mar 13 16:34:00 crc kubenswrapper[4907]: E0313 16:34:00.190053 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d2dfd3f-0bc6-48f9-85f7-100ef8f23476" containerName="oc" Mar 13 16:34:00 crc kubenswrapper[4907]: I0313 16:34:00.190061 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d2dfd3f-0bc6-48f9-85f7-100ef8f23476" containerName="oc" Mar 13 16:34:00 crc kubenswrapper[4907]: E0313 16:34:00.190100 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de68abba-24e7-49b8-9e1f-adfa05d9d3dd" containerName="nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell1" Mar 13 16:34:00 crc kubenswrapper[4907]: I0313 16:34:00.190150 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="de68abba-24e7-49b8-9e1f-adfa05d9d3dd" containerName="nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell1" Mar 13 16:34:00 crc kubenswrapper[4907]: E0313 16:34:00.190198 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fed253af-6b1e-49b2-9d21-4467f33ecf94" containerName="registry-server" Mar 13 16:34:00 crc kubenswrapper[4907]: I0313 16:34:00.190207 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="fed253af-6b1e-49b2-9d21-4467f33ecf94" containerName="registry-server" Mar 13 16:34:00 crc kubenswrapper[4907]: E0313 16:34:00.190217 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fed253af-6b1e-49b2-9d21-4467f33ecf94" containerName="extract-content" Mar 13 16:34:00 crc kubenswrapper[4907]: I0313 16:34:00.190224 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="fed253af-6b1e-49b2-9d21-4467f33ecf94" containerName="extract-content" Mar 13 16:34:00 crc kubenswrapper[4907]: I0313 16:34:00.190829 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="fed253af-6b1e-49b2-9d21-4467f33ecf94" containerName="registry-server" Mar 13 16:34:00 crc kubenswrapper[4907]: I0313 16:34:00.190901 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="5d2dfd3f-0bc6-48f9-85f7-100ef8f23476" containerName="oc" Mar 13 16:34:00 crc kubenswrapper[4907]: I0313 16:34:00.190932 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="de68abba-24e7-49b8-9e1f-adfa05d9d3dd" containerName="nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell1" Mar 13 16:34:00 crc kubenswrapper[4907]: I0313 16:34:00.192629 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556994-452h2" Mar 13 16:34:00 crc kubenswrapper[4907]: I0313 16:34:00.196703 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 16:34:00 crc kubenswrapper[4907]: I0313 16:34:00.197449 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 16:34:00 crc kubenswrapper[4907]: I0313 16:34:00.197645 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 16:34:00 crc kubenswrapper[4907]: I0313 16:34:00.237573 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556994-452h2"] Mar 13 16:34:00 crc kubenswrapper[4907]: I0313 16:34:00.304393 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wqmsz\" (UniqueName: \"kubernetes.io/projected/262afeb8-16de-47e7-a64b-5b4e938aa1f4-kube-api-access-wqmsz\") pod \"auto-csr-approver-29556994-452h2\" (UID: \"262afeb8-16de-47e7-a64b-5b4e938aa1f4\") " pod="openshift-infra/auto-csr-approver-29556994-452h2" Mar 13 16:34:00 crc kubenswrapper[4907]: I0313 16:34:00.407513 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wqmsz\" (UniqueName: \"kubernetes.io/projected/262afeb8-16de-47e7-a64b-5b4e938aa1f4-kube-api-access-wqmsz\") pod \"auto-csr-approver-29556994-452h2\" (UID: \"262afeb8-16de-47e7-a64b-5b4e938aa1f4\") " pod="openshift-infra/auto-csr-approver-29556994-452h2" Mar 13 16:34:00 crc kubenswrapper[4907]: I0313 16:34:00.441703 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wqmsz\" (UniqueName: \"kubernetes.io/projected/262afeb8-16de-47e7-a64b-5b4e938aa1f4-kube-api-access-wqmsz\") pod \"auto-csr-approver-29556994-452h2\" (UID: \"262afeb8-16de-47e7-a64b-5b4e938aa1f4\") " pod="openshift-infra/auto-csr-approver-29556994-452h2" Mar 13 16:34:00 crc kubenswrapper[4907]: I0313 16:34:00.520630 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556994-452h2" Mar 13 16:34:01 crc kubenswrapper[4907]: I0313 16:34:01.076434 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556994-452h2"] Mar 13 16:34:01 crc kubenswrapper[4907]: I0313 16:34:01.091525 4907 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 13 16:34:01 crc kubenswrapper[4907]: I0313 16:34:01.919433 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556994-452h2" event={"ID":"262afeb8-16de-47e7-a64b-5b4e938aa1f4","Type":"ContainerStarted","Data":"a8d9553ed7e2900c7f96d66573bd4bda4a8ba968c5d8cee4a3dbeb326577c98c"} Mar 13 16:34:02 crc kubenswrapper[4907]: I0313 16:34:02.937052 4907 generic.go:334] "Generic (PLEG): container finished" podID="262afeb8-16de-47e7-a64b-5b4e938aa1f4" containerID="237bb05fcc8f490c8ced9263406f8715e053a689a6c929dacb3684e153ea9c27" exitCode=0 Mar 13 16:34:02 crc kubenswrapper[4907]: I0313 16:34:02.937147 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556994-452h2" event={"ID":"262afeb8-16de-47e7-a64b-5b4e938aa1f4","Type":"ContainerDied","Data":"237bb05fcc8f490c8ced9263406f8715e053a689a6c929dacb3684e153ea9c27"} Mar 13 16:34:04 crc kubenswrapper[4907]: I0313 16:34:04.424259 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556994-452h2" Mar 13 16:34:04 crc kubenswrapper[4907]: I0313 16:34:04.608018 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wqmsz\" (UniqueName: \"kubernetes.io/projected/262afeb8-16de-47e7-a64b-5b4e938aa1f4-kube-api-access-wqmsz\") pod \"262afeb8-16de-47e7-a64b-5b4e938aa1f4\" (UID: \"262afeb8-16de-47e7-a64b-5b4e938aa1f4\") " Mar 13 16:34:04 crc kubenswrapper[4907]: I0313 16:34:04.617725 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/262afeb8-16de-47e7-a64b-5b4e938aa1f4-kube-api-access-wqmsz" (OuterVolumeSpecName: "kube-api-access-wqmsz") pod "262afeb8-16de-47e7-a64b-5b4e938aa1f4" (UID: "262afeb8-16de-47e7-a64b-5b4e938aa1f4"). InnerVolumeSpecName "kube-api-access-wqmsz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:34:04 crc kubenswrapper[4907]: I0313 16:34:04.710914 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wqmsz\" (UniqueName: \"kubernetes.io/projected/262afeb8-16de-47e7-a64b-5b4e938aa1f4-kube-api-access-wqmsz\") on node \"crc\" DevicePath \"\"" Mar 13 16:34:04 crc kubenswrapper[4907]: I0313 16:34:04.969828 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556994-452h2" event={"ID":"262afeb8-16de-47e7-a64b-5b4e938aa1f4","Type":"ContainerDied","Data":"a8d9553ed7e2900c7f96d66573bd4bda4a8ba968c5d8cee4a3dbeb326577c98c"} Mar 13 16:34:04 crc kubenswrapper[4907]: I0313 16:34:04.970382 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a8d9553ed7e2900c7f96d66573bd4bda4a8ba968c5d8cee4a3dbeb326577c98c" Mar 13 16:34:04 crc kubenswrapper[4907]: I0313 16:34:04.970469 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556994-452h2" Mar 13 16:34:05 crc kubenswrapper[4907]: I0313 16:34:05.545198 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556988-t5tgt"] Mar 13 16:34:05 crc kubenswrapper[4907]: I0313 16:34:05.558753 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556988-t5tgt"] Mar 13 16:34:05 crc kubenswrapper[4907]: I0313 16:34:05.799321 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6aa9f963-9a2b-4101-88e5-9553783a9963" path="/var/lib/kubelet/pods/6aa9f963-9a2b-4101-88e5-9553783a9963/volumes" Mar 13 16:34:11 crc kubenswrapper[4907]: I0313 16:34:11.862423 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-598sp"] Mar 13 16:34:11 crc kubenswrapper[4907]: E0313 16:34:11.863738 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="262afeb8-16de-47e7-a64b-5b4e938aa1f4" containerName="oc" Mar 13 16:34:11 crc kubenswrapper[4907]: I0313 16:34:11.863754 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="262afeb8-16de-47e7-a64b-5b4e938aa1f4" containerName="oc" Mar 13 16:34:11 crc kubenswrapper[4907]: I0313 16:34:11.864087 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="262afeb8-16de-47e7-a64b-5b4e938aa1f4" containerName="oc" Mar 13 16:34:11 crc kubenswrapper[4907]: I0313 16:34:11.865750 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-598sp" Mar 13 16:34:11 crc kubenswrapper[4907]: I0313 16:34:11.888479 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-598sp"] Mar 13 16:34:12 crc kubenswrapper[4907]: I0313 16:34:12.039499 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j9fgq\" (UniqueName: \"kubernetes.io/projected/92a3a8bb-d3df-4489-8cf5-aacdad7516a9-kube-api-access-j9fgq\") pod \"certified-operators-598sp\" (UID: \"92a3a8bb-d3df-4489-8cf5-aacdad7516a9\") " pod="openshift-marketplace/certified-operators-598sp" Mar 13 16:34:12 crc kubenswrapper[4907]: I0313 16:34:12.039546 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92a3a8bb-d3df-4489-8cf5-aacdad7516a9-catalog-content\") pod \"certified-operators-598sp\" (UID: \"92a3a8bb-d3df-4489-8cf5-aacdad7516a9\") " pod="openshift-marketplace/certified-operators-598sp" Mar 13 16:34:12 crc kubenswrapper[4907]: I0313 16:34:12.039565 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92a3a8bb-d3df-4489-8cf5-aacdad7516a9-utilities\") pod \"certified-operators-598sp\" (UID: \"92a3a8bb-d3df-4489-8cf5-aacdad7516a9\") " pod="openshift-marketplace/certified-operators-598sp" Mar 13 16:34:12 crc kubenswrapper[4907]: I0313 16:34:12.142484 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j9fgq\" (UniqueName: \"kubernetes.io/projected/92a3a8bb-d3df-4489-8cf5-aacdad7516a9-kube-api-access-j9fgq\") pod \"certified-operators-598sp\" (UID: \"92a3a8bb-d3df-4489-8cf5-aacdad7516a9\") " pod="openshift-marketplace/certified-operators-598sp" Mar 13 16:34:12 crc kubenswrapper[4907]: I0313 16:34:12.142540 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92a3a8bb-d3df-4489-8cf5-aacdad7516a9-catalog-content\") pod \"certified-operators-598sp\" (UID: \"92a3a8bb-d3df-4489-8cf5-aacdad7516a9\") " pod="openshift-marketplace/certified-operators-598sp" Mar 13 16:34:12 crc kubenswrapper[4907]: I0313 16:34:12.142564 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92a3a8bb-d3df-4489-8cf5-aacdad7516a9-utilities\") pod \"certified-operators-598sp\" (UID: \"92a3a8bb-d3df-4489-8cf5-aacdad7516a9\") " pod="openshift-marketplace/certified-operators-598sp" Mar 13 16:34:12 crc kubenswrapper[4907]: I0313 16:34:12.143170 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92a3a8bb-d3df-4489-8cf5-aacdad7516a9-utilities\") pod \"certified-operators-598sp\" (UID: \"92a3a8bb-d3df-4489-8cf5-aacdad7516a9\") " pod="openshift-marketplace/certified-operators-598sp" Mar 13 16:34:12 crc kubenswrapper[4907]: I0313 16:34:12.143237 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92a3a8bb-d3df-4489-8cf5-aacdad7516a9-catalog-content\") pod \"certified-operators-598sp\" (UID: \"92a3a8bb-d3df-4489-8cf5-aacdad7516a9\") " pod="openshift-marketplace/certified-operators-598sp" Mar 13 16:34:12 crc kubenswrapper[4907]: I0313 16:34:12.167979 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j9fgq\" (UniqueName: \"kubernetes.io/projected/92a3a8bb-d3df-4489-8cf5-aacdad7516a9-kube-api-access-j9fgq\") pod \"certified-operators-598sp\" (UID: \"92a3a8bb-d3df-4489-8cf5-aacdad7516a9\") " pod="openshift-marketplace/certified-operators-598sp" Mar 13 16:34:12 crc kubenswrapper[4907]: I0313 16:34:12.200299 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-598sp" Mar 13 16:34:12 crc kubenswrapper[4907]: I0313 16:34:12.758702 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-598sp"] Mar 13 16:34:13 crc kubenswrapper[4907]: I0313 16:34:13.068137 4907 generic.go:334] "Generic (PLEG): container finished" podID="92a3a8bb-d3df-4489-8cf5-aacdad7516a9" containerID="49b82ed2d30c3a4ca01343761e05b4b6d372851890749b3a1174afaa9435ef6a" exitCode=0 Mar 13 16:34:13 crc kubenswrapper[4907]: I0313 16:34:13.068235 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-598sp" event={"ID":"92a3a8bb-d3df-4489-8cf5-aacdad7516a9","Type":"ContainerDied","Data":"49b82ed2d30c3a4ca01343761e05b4b6d372851890749b3a1174afaa9435ef6a"} Mar 13 16:34:13 crc kubenswrapper[4907]: I0313 16:34:13.068527 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-598sp" event={"ID":"92a3a8bb-d3df-4489-8cf5-aacdad7516a9","Type":"ContainerStarted","Data":"87a43c517510df3cce3a23c229d4e56247da0404420b2f98fa527a9b52aeadfb"} Mar 13 16:34:15 crc kubenswrapper[4907]: I0313 16:34:15.099511 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-598sp" event={"ID":"92a3a8bb-d3df-4489-8cf5-aacdad7516a9","Type":"ContainerStarted","Data":"8a3c7e325436c7c09192564adf5ec9964f5d304f4b9cf66a467589dd3de42af4"} Mar 13 16:34:16 crc kubenswrapper[4907]: I0313 16:34:16.121600 4907 generic.go:334] "Generic (PLEG): container finished" podID="92a3a8bb-d3df-4489-8cf5-aacdad7516a9" containerID="8a3c7e325436c7c09192564adf5ec9964f5d304f4b9cf66a467589dd3de42af4" exitCode=0 Mar 13 16:34:16 crc kubenswrapper[4907]: I0313 16:34:16.122018 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-598sp" event={"ID":"92a3a8bb-d3df-4489-8cf5-aacdad7516a9","Type":"ContainerDied","Data":"8a3c7e325436c7c09192564adf5ec9964f5d304f4b9cf66a467589dd3de42af4"} Mar 13 16:34:17 crc kubenswrapper[4907]: I0313 16:34:17.134610 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-598sp" event={"ID":"92a3a8bb-d3df-4489-8cf5-aacdad7516a9","Type":"ContainerStarted","Data":"61eedae48e994ea69752606a6aac87ef010a61cdcce61d46631125a9b841b880"} Mar 13 16:34:17 crc kubenswrapper[4907]: I0313 16:34:17.158586 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-598sp" podStartSLOduration=2.685575957 podStartE2EDuration="6.15856231s" podCreationTimestamp="2026-03-13 16:34:11 +0000 UTC" firstStartedPulling="2026-03-13 16:34:13.0706122 +0000 UTC m=+8951.970399889" lastFinishedPulling="2026-03-13 16:34:16.543598553 +0000 UTC m=+8955.443386242" observedRunningTime="2026-03-13 16:34:17.151409468 +0000 UTC m=+8956.051197157" watchObservedRunningTime="2026-03-13 16:34:17.15856231 +0000 UTC m=+8956.058349999" Mar 13 16:34:22 crc kubenswrapper[4907]: I0313 16:34:22.200576 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-598sp" Mar 13 16:34:22 crc kubenswrapper[4907]: I0313 16:34:22.201274 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-598sp" Mar 13 16:34:22 crc kubenswrapper[4907]: I0313 16:34:22.888035 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-598sp" Mar 13 16:34:23 crc kubenswrapper[4907]: I0313 16:34:23.298689 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-598sp" Mar 13 16:34:23 crc kubenswrapper[4907]: I0313 16:34:23.354337 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-598sp"] Mar 13 16:34:25 crc kubenswrapper[4907]: I0313 16:34:25.257841 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-598sp" podUID="92a3a8bb-d3df-4489-8cf5-aacdad7516a9" containerName="registry-server" containerID="cri-o://61eedae48e994ea69752606a6aac87ef010a61cdcce61d46631125a9b841b880" gracePeriod=2 Mar 13 16:34:25 crc kubenswrapper[4907]: I0313 16:34:25.561966 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-kksrr"] Mar 13 16:34:25 crc kubenswrapper[4907]: I0313 16:34:25.564875 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kksrr" Mar 13 16:34:25 crc kubenswrapper[4907]: I0313 16:34:25.578274 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-kksrr"] Mar 13 16:34:25 crc kubenswrapper[4907]: I0313 16:34:25.748459 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-275mr\" (UniqueName: \"kubernetes.io/projected/b70871f8-ca8b-4909-9d67-b7e720821aea-kube-api-access-275mr\") pod \"community-operators-kksrr\" (UID: \"b70871f8-ca8b-4909-9d67-b7e720821aea\") " pod="openshift-marketplace/community-operators-kksrr" Mar 13 16:34:25 crc kubenswrapper[4907]: I0313 16:34:25.748652 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b70871f8-ca8b-4909-9d67-b7e720821aea-catalog-content\") pod \"community-operators-kksrr\" (UID: \"b70871f8-ca8b-4909-9d67-b7e720821aea\") " pod="openshift-marketplace/community-operators-kksrr" Mar 13 16:34:25 crc kubenswrapper[4907]: I0313 16:34:25.749135 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b70871f8-ca8b-4909-9d67-b7e720821aea-utilities\") pod \"community-operators-kksrr\" (UID: \"b70871f8-ca8b-4909-9d67-b7e720821aea\") " pod="openshift-marketplace/community-operators-kksrr" Mar 13 16:34:25 crc kubenswrapper[4907]: I0313 16:34:25.824464 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-598sp" Mar 13 16:34:25 crc kubenswrapper[4907]: I0313 16:34:25.851125 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-275mr\" (UniqueName: \"kubernetes.io/projected/b70871f8-ca8b-4909-9d67-b7e720821aea-kube-api-access-275mr\") pod \"community-operators-kksrr\" (UID: \"b70871f8-ca8b-4909-9d67-b7e720821aea\") " pod="openshift-marketplace/community-operators-kksrr" Mar 13 16:34:25 crc kubenswrapper[4907]: I0313 16:34:25.851491 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b70871f8-ca8b-4909-9d67-b7e720821aea-catalog-content\") pod \"community-operators-kksrr\" (UID: \"b70871f8-ca8b-4909-9d67-b7e720821aea\") " pod="openshift-marketplace/community-operators-kksrr" Mar 13 16:34:25 crc kubenswrapper[4907]: I0313 16:34:25.851597 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b70871f8-ca8b-4909-9d67-b7e720821aea-utilities\") pod \"community-operators-kksrr\" (UID: \"b70871f8-ca8b-4909-9d67-b7e720821aea\") " pod="openshift-marketplace/community-operators-kksrr" Mar 13 16:34:25 crc kubenswrapper[4907]: I0313 16:34:25.852225 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b70871f8-ca8b-4909-9d67-b7e720821aea-catalog-content\") pod \"community-operators-kksrr\" (UID: \"b70871f8-ca8b-4909-9d67-b7e720821aea\") " pod="openshift-marketplace/community-operators-kksrr" Mar 13 16:34:25 crc kubenswrapper[4907]: I0313 16:34:25.853399 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b70871f8-ca8b-4909-9d67-b7e720821aea-utilities\") pod \"community-operators-kksrr\" (UID: \"b70871f8-ca8b-4909-9d67-b7e720821aea\") " pod="openshift-marketplace/community-operators-kksrr" Mar 13 16:34:25 crc kubenswrapper[4907]: I0313 16:34:25.876861 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-275mr\" (UniqueName: \"kubernetes.io/projected/b70871f8-ca8b-4909-9d67-b7e720821aea-kube-api-access-275mr\") pod \"community-operators-kksrr\" (UID: \"b70871f8-ca8b-4909-9d67-b7e720821aea\") " pod="openshift-marketplace/community-operators-kksrr" Mar 13 16:34:25 crc kubenswrapper[4907]: I0313 16:34:25.894494 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kksrr" Mar 13 16:34:25 crc kubenswrapper[4907]: I0313 16:34:25.954166 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92a3a8bb-d3df-4489-8cf5-aacdad7516a9-utilities\") pod \"92a3a8bb-d3df-4489-8cf5-aacdad7516a9\" (UID: \"92a3a8bb-d3df-4489-8cf5-aacdad7516a9\") " Mar 13 16:34:25 crc kubenswrapper[4907]: I0313 16:34:25.954236 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92a3a8bb-d3df-4489-8cf5-aacdad7516a9-catalog-content\") pod \"92a3a8bb-d3df-4489-8cf5-aacdad7516a9\" (UID: \"92a3a8bb-d3df-4489-8cf5-aacdad7516a9\") " Mar 13 16:34:25 crc kubenswrapper[4907]: I0313 16:34:25.954328 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j9fgq\" (UniqueName: \"kubernetes.io/projected/92a3a8bb-d3df-4489-8cf5-aacdad7516a9-kube-api-access-j9fgq\") pod \"92a3a8bb-d3df-4489-8cf5-aacdad7516a9\" (UID: \"92a3a8bb-d3df-4489-8cf5-aacdad7516a9\") " Mar 13 16:34:25 crc kubenswrapper[4907]: I0313 16:34:25.959476 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/92a3a8bb-d3df-4489-8cf5-aacdad7516a9-utilities" (OuterVolumeSpecName: "utilities") pod "92a3a8bb-d3df-4489-8cf5-aacdad7516a9" (UID: "92a3a8bb-d3df-4489-8cf5-aacdad7516a9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 16:34:25 crc kubenswrapper[4907]: I0313 16:34:25.969655 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/92a3a8bb-d3df-4489-8cf5-aacdad7516a9-kube-api-access-j9fgq" (OuterVolumeSpecName: "kube-api-access-j9fgq") pod "92a3a8bb-d3df-4489-8cf5-aacdad7516a9" (UID: "92a3a8bb-d3df-4489-8cf5-aacdad7516a9"). InnerVolumeSpecName "kube-api-access-j9fgq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:34:26 crc kubenswrapper[4907]: I0313 16:34:26.058005 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92a3a8bb-d3df-4489-8cf5-aacdad7516a9-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 16:34:26 crc kubenswrapper[4907]: I0313 16:34:26.058577 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j9fgq\" (UniqueName: \"kubernetes.io/projected/92a3a8bb-d3df-4489-8cf5-aacdad7516a9-kube-api-access-j9fgq\") on node \"crc\" DevicePath \"\"" Mar 13 16:34:26 crc kubenswrapper[4907]: I0313 16:34:26.102377 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/92a3a8bb-d3df-4489-8cf5-aacdad7516a9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "92a3a8bb-d3df-4489-8cf5-aacdad7516a9" (UID: "92a3a8bb-d3df-4489-8cf5-aacdad7516a9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 16:34:26 crc kubenswrapper[4907]: I0313 16:34:26.162304 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92a3a8bb-d3df-4489-8cf5-aacdad7516a9-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 16:34:26 crc kubenswrapper[4907]: I0313 16:34:26.268982 4907 generic.go:334] "Generic (PLEG): container finished" podID="92a3a8bb-d3df-4489-8cf5-aacdad7516a9" containerID="61eedae48e994ea69752606a6aac87ef010a61cdcce61d46631125a9b841b880" exitCode=0 Mar 13 16:34:26 crc kubenswrapper[4907]: I0313 16:34:26.269057 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-598sp" Mar 13 16:34:26 crc kubenswrapper[4907]: I0313 16:34:26.269025 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-598sp" event={"ID":"92a3a8bb-d3df-4489-8cf5-aacdad7516a9","Type":"ContainerDied","Data":"61eedae48e994ea69752606a6aac87ef010a61cdcce61d46631125a9b841b880"} Mar 13 16:34:26 crc kubenswrapper[4907]: I0313 16:34:26.269135 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-598sp" event={"ID":"92a3a8bb-d3df-4489-8cf5-aacdad7516a9","Type":"ContainerDied","Data":"87a43c517510df3cce3a23c229d4e56247da0404420b2f98fa527a9b52aeadfb"} Mar 13 16:34:26 crc kubenswrapper[4907]: I0313 16:34:26.269220 4907 scope.go:117] "RemoveContainer" containerID="61eedae48e994ea69752606a6aac87ef010a61cdcce61d46631125a9b841b880" Mar 13 16:34:26 crc kubenswrapper[4907]: I0313 16:34:26.302494 4907 scope.go:117] "RemoveContainer" containerID="8a3c7e325436c7c09192564adf5ec9964f5d304f4b9cf66a467589dd3de42af4" Mar 13 16:34:26 crc kubenswrapper[4907]: I0313 16:34:26.313745 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-598sp"] Mar 13 16:34:26 crc kubenswrapper[4907]: I0313 16:34:26.333210 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-598sp"] Mar 13 16:34:26 crc kubenswrapper[4907]: I0313 16:34:26.349001 4907 scope.go:117] "RemoveContainer" containerID="49b82ed2d30c3a4ca01343761e05b4b6d372851890749b3a1174afaa9435ef6a" Mar 13 16:34:26 crc kubenswrapper[4907]: I0313 16:34:26.383995 4907 scope.go:117] "RemoveContainer" containerID="61eedae48e994ea69752606a6aac87ef010a61cdcce61d46631125a9b841b880" Mar 13 16:34:26 crc kubenswrapper[4907]: E0313 16:34:26.384879 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"61eedae48e994ea69752606a6aac87ef010a61cdcce61d46631125a9b841b880\": container with ID starting with 61eedae48e994ea69752606a6aac87ef010a61cdcce61d46631125a9b841b880 not found: ID does not exist" containerID="61eedae48e994ea69752606a6aac87ef010a61cdcce61d46631125a9b841b880" Mar 13 16:34:26 crc kubenswrapper[4907]: I0313 16:34:26.384987 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"61eedae48e994ea69752606a6aac87ef010a61cdcce61d46631125a9b841b880"} err="failed to get container status \"61eedae48e994ea69752606a6aac87ef010a61cdcce61d46631125a9b841b880\": rpc error: code = NotFound desc = could not find container \"61eedae48e994ea69752606a6aac87ef010a61cdcce61d46631125a9b841b880\": container with ID starting with 61eedae48e994ea69752606a6aac87ef010a61cdcce61d46631125a9b841b880 not found: ID does not exist" Mar 13 16:34:26 crc kubenswrapper[4907]: I0313 16:34:26.385026 4907 scope.go:117] "RemoveContainer" containerID="8a3c7e325436c7c09192564adf5ec9964f5d304f4b9cf66a467589dd3de42af4" Mar 13 16:34:26 crc kubenswrapper[4907]: E0313 16:34:26.385489 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8a3c7e325436c7c09192564adf5ec9964f5d304f4b9cf66a467589dd3de42af4\": container with ID starting with 8a3c7e325436c7c09192564adf5ec9964f5d304f4b9cf66a467589dd3de42af4 not found: ID does not exist" containerID="8a3c7e325436c7c09192564adf5ec9964f5d304f4b9cf66a467589dd3de42af4" Mar 13 16:34:26 crc kubenswrapper[4907]: I0313 16:34:26.385546 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8a3c7e325436c7c09192564adf5ec9964f5d304f4b9cf66a467589dd3de42af4"} err="failed to get container status \"8a3c7e325436c7c09192564adf5ec9964f5d304f4b9cf66a467589dd3de42af4\": rpc error: code = NotFound desc = could not find container \"8a3c7e325436c7c09192564adf5ec9964f5d304f4b9cf66a467589dd3de42af4\": container with ID starting with 8a3c7e325436c7c09192564adf5ec9964f5d304f4b9cf66a467589dd3de42af4 not found: ID does not exist" Mar 13 16:34:26 crc kubenswrapper[4907]: I0313 16:34:26.385580 4907 scope.go:117] "RemoveContainer" containerID="49b82ed2d30c3a4ca01343761e05b4b6d372851890749b3a1174afaa9435ef6a" Mar 13 16:34:26 crc kubenswrapper[4907]: E0313 16:34:26.387311 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"49b82ed2d30c3a4ca01343761e05b4b6d372851890749b3a1174afaa9435ef6a\": container with ID starting with 49b82ed2d30c3a4ca01343761e05b4b6d372851890749b3a1174afaa9435ef6a not found: ID does not exist" containerID="49b82ed2d30c3a4ca01343761e05b4b6d372851890749b3a1174afaa9435ef6a" Mar 13 16:34:26 crc kubenswrapper[4907]: I0313 16:34:26.387349 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"49b82ed2d30c3a4ca01343761e05b4b6d372851890749b3a1174afaa9435ef6a"} err="failed to get container status \"49b82ed2d30c3a4ca01343761e05b4b6d372851890749b3a1174afaa9435ef6a\": rpc error: code = NotFound desc = could not find container \"49b82ed2d30c3a4ca01343761e05b4b6d372851890749b3a1174afaa9435ef6a\": container with ID starting with 49b82ed2d30c3a4ca01343761e05b4b6d372851890749b3a1174afaa9435ef6a not found: ID does not exist" Mar 13 16:34:26 crc kubenswrapper[4907]: I0313 16:34:26.465352 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-kksrr"] Mar 13 16:34:27 crc kubenswrapper[4907]: I0313 16:34:27.284561 4907 generic.go:334] "Generic (PLEG): container finished" podID="b70871f8-ca8b-4909-9d67-b7e720821aea" containerID="fd08cb1e0e69433e63613bf6f882fd195d81484e09fd17ba0e96e23e939edb66" exitCode=0 Mar 13 16:34:27 crc kubenswrapper[4907]: I0313 16:34:27.284643 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kksrr" event={"ID":"b70871f8-ca8b-4909-9d67-b7e720821aea","Type":"ContainerDied","Data":"fd08cb1e0e69433e63613bf6f882fd195d81484e09fd17ba0e96e23e939edb66"} Mar 13 16:34:27 crc kubenswrapper[4907]: I0313 16:34:27.285247 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kksrr" event={"ID":"b70871f8-ca8b-4909-9d67-b7e720821aea","Type":"ContainerStarted","Data":"bce3053aa4e24f9c102b2e439362111ed2eec33f4e6bd2a0aa4f544909c02204"} Mar 13 16:34:27 crc kubenswrapper[4907]: I0313 16:34:27.795090 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="92a3a8bb-d3df-4489-8cf5-aacdad7516a9" path="/var/lib/kubelet/pods/92a3a8bb-d3df-4489-8cf5-aacdad7516a9/volumes" Mar 13 16:34:28 crc kubenswrapper[4907]: I0313 16:34:28.324389 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kksrr" event={"ID":"b70871f8-ca8b-4909-9d67-b7e720821aea","Type":"ContainerStarted","Data":"1eb85ed4c37b80885e34300324c44dd232a6deab53c7b12db684fc1106d17c93"} Mar 13 16:34:30 crc kubenswrapper[4907]: I0313 16:34:30.350385 4907 generic.go:334] "Generic (PLEG): container finished" podID="b70871f8-ca8b-4909-9d67-b7e720821aea" containerID="1eb85ed4c37b80885e34300324c44dd232a6deab53c7b12db684fc1106d17c93" exitCode=0 Mar 13 16:34:30 crc kubenswrapper[4907]: I0313 16:34:30.350423 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kksrr" event={"ID":"b70871f8-ca8b-4909-9d67-b7e720821aea","Type":"ContainerDied","Data":"1eb85ed4c37b80885e34300324c44dd232a6deab53c7b12db684fc1106d17c93"} Mar 13 16:34:31 crc kubenswrapper[4907]: I0313 16:34:31.365185 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kksrr" event={"ID":"b70871f8-ca8b-4909-9d67-b7e720821aea","Type":"ContainerStarted","Data":"7c4d2e7e2ab554003cdeb6b94cee639bd68a9404b66218366440011f09703ffb"} Mar 13 16:34:31 crc kubenswrapper[4907]: I0313 16:34:31.402279 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-kksrr" podStartSLOduration=2.907058638 podStartE2EDuration="6.402251285s" podCreationTimestamp="2026-03-13 16:34:25 +0000 UTC" firstStartedPulling="2026-03-13 16:34:27.286789392 +0000 UTC m=+8966.186577081" lastFinishedPulling="2026-03-13 16:34:30.781982009 +0000 UTC m=+8969.681769728" observedRunningTime="2026-03-13 16:34:31.387657145 +0000 UTC m=+8970.287444854" watchObservedRunningTime="2026-03-13 16:34:31.402251285 +0000 UTC m=+8970.302039014" Mar 13 16:34:35 crc kubenswrapper[4907]: I0313 16:34:35.895943 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-kksrr" Mar 13 16:34:35 crc kubenswrapper[4907]: I0313 16:34:35.896621 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-kksrr" Mar 13 16:34:36 crc kubenswrapper[4907]: I0313 16:34:36.251535 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-kksrr" Mar 13 16:34:36 crc kubenswrapper[4907]: I0313 16:34:36.491640 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-kksrr" Mar 13 16:34:36 crc kubenswrapper[4907]: I0313 16:34:36.552988 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-kksrr"] Mar 13 16:34:38 crc kubenswrapper[4907]: I0313 16:34:38.481093 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-kksrr" podUID="b70871f8-ca8b-4909-9d67-b7e720821aea" containerName="registry-server" containerID="cri-o://7c4d2e7e2ab554003cdeb6b94cee639bd68a9404b66218366440011f09703ffb" gracePeriod=2 Mar 13 16:34:39 crc kubenswrapper[4907]: I0313 16:34:39.235783 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kksrr" Mar 13 16:34:39 crc kubenswrapper[4907]: I0313 16:34:39.402387 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b70871f8-ca8b-4909-9d67-b7e720821aea-catalog-content\") pod \"b70871f8-ca8b-4909-9d67-b7e720821aea\" (UID: \"b70871f8-ca8b-4909-9d67-b7e720821aea\") " Mar 13 16:34:39 crc kubenswrapper[4907]: I0313 16:34:39.402495 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b70871f8-ca8b-4909-9d67-b7e720821aea-utilities\") pod \"b70871f8-ca8b-4909-9d67-b7e720821aea\" (UID: \"b70871f8-ca8b-4909-9d67-b7e720821aea\") " Mar 13 16:34:39 crc kubenswrapper[4907]: I0313 16:34:39.402595 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-275mr\" (UniqueName: \"kubernetes.io/projected/b70871f8-ca8b-4909-9d67-b7e720821aea-kube-api-access-275mr\") pod \"b70871f8-ca8b-4909-9d67-b7e720821aea\" (UID: \"b70871f8-ca8b-4909-9d67-b7e720821aea\") " Mar 13 16:34:39 crc kubenswrapper[4907]: I0313 16:34:39.403431 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b70871f8-ca8b-4909-9d67-b7e720821aea-utilities" (OuterVolumeSpecName: "utilities") pod "b70871f8-ca8b-4909-9d67-b7e720821aea" (UID: "b70871f8-ca8b-4909-9d67-b7e720821aea"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 16:34:39 crc kubenswrapper[4907]: I0313 16:34:39.411176 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b70871f8-ca8b-4909-9d67-b7e720821aea-kube-api-access-275mr" (OuterVolumeSpecName: "kube-api-access-275mr") pod "b70871f8-ca8b-4909-9d67-b7e720821aea" (UID: "b70871f8-ca8b-4909-9d67-b7e720821aea"). InnerVolumeSpecName "kube-api-access-275mr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:34:39 crc kubenswrapper[4907]: I0313 16:34:39.463848 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b70871f8-ca8b-4909-9d67-b7e720821aea-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b70871f8-ca8b-4909-9d67-b7e720821aea" (UID: "b70871f8-ca8b-4909-9d67-b7e720821aea"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 16:34:39 crc kubenswrapper[4907]: I0313 16:34:39.494346 4907 generic.go:334] "Generic (PLEG): container finished" podID="b70871f8-ca8b-4909-9d67-b7e720821aea" containerID="7c4d2e7e2ab554003cdeb6b94cee639bd68a9404b66218366440011f09703ffb" exitCode=0 Mar 13 16:34:39 crc kubenswrapper[4907]: I0313 16:34:39.494406 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kksrr" event={"ID":"b70871f8-ca8b-4909-9d67-b7e720821aea","Type":"ContainerDied","Data":"7c4d2e7e2ab554003cdeb6b94cee639bd68a9404b66218366440011f09703ffb"} Mar 13 16:34:39 crc kubenswrapper[4907]: I0313 16:34:39.494455 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kksrr" event={"ID":"b70871f8-ca8b-4909-9d67-b7e720821aea","Type":"ContainerDied","Data":"bce3053aa4e24f9c102b2e439362111ed2eec33f4e6bd2a0aa4f544909c02204"} Mar 13 16:34:39 crc kubenswrapper[4907]: I0313 16:34:39.494487 4907 scope.go:117] "RemoveContainer" containerID="7c4d2e7e2ab554003cdeb6b94cee639bd68a9404b66218366440011f09703ffb" Mar 13 16:34:39 crc kubenswrapper[4907]: I0313 16:34:39.496871 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kksrr" Mar 13 16:34:39 crc kubenswrapper[4907]: I0313 16:34:39.505281 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b70871f8-ca8b-4909-9d67-b7e720821aea-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 16:34:39 crc kubenswrapper[4907]: I0313 16:34:39.505323 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b70871f8-ca8b-4909-9d67-b7e720821aea-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 16:34:39 crc kubenswrapper[4907]: I0313 16:34:39.505338 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-275mr\" (UniqueName: \"kubernetes.io/projected/b70871f8-ca8b-4909-9d67-b7e720821aea-kube-api-access-275mr\") on node \"crc\" DevicePath \"\"" Mar 13 16:34:39 crc kubenswrapper[4907]: I0313 16:34:39.543118 4907 scope.go:117] "RemoveContainer" containerID="1eb85ed4c37b80885e34300324c44dd232a6deab53c7b12db684fc1106d17c93" Mar 13 16:34:39 crc kubenswrapper[4907]: I0313 16:34:39.546998 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-kksrr"] Mar 13 16:34:39 crc kubenswrapper[4907]: I0313 16:34:39.559149 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-kksrr"] Mar 13 16:34:39 crc kubenswrapper[4907]: I0313 16:34:39.569493 4907 scope.go:117] "RemoveContainer" containerID="fd08cb1e0e69433e63613bf6f882fd195d81484e09fd17ba0e96e23e939edb66" Mar 13 16:34:39 crc kubenswrapper[4907]: I0313 16:34:39.614703 4907 scope.go:117] "RemoveContainer" containerID="7c4d2e7e2ab554003cdeb6b94cee639bd68a9404b66218366440011f09703ffb" Mar 13 16:34:39 crc kubenswrapper[4907]: E0313 16:34:39.615150 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7c4d2e7e2ab554003cdeb6b94cee639bd68a9404b66218366440011f09703ffb\": container with ID starting with 7c4d2e7e2ab554003cdeb6b94cee639bd68a9404b66218366440011f09703ffb not found: ID does not exist" containerID="7c4d2e7e2ab554003cdeb6b94cee639bd68a9404b66218366440011f09703ffb" Mar 13 16:34:39 crc kubenswrapper[4907]: I0313 16:34:39.615189 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7c4d2e7e2ab554003cdeb6b94cee639bd68a9404b66218366440011f09703ffb"} err="failed to get container status \"7c4d2e7e2ab554003cdeb6b94cee639bd68a9404b66218366440011f09703ffb\": rpc error: code = NotFound desc = could not find container \"7c4d2e7e2ab554003cdeb6b94cee639bd68a9404b66218366440011f09703ffb\": container with ID starting with 7c4d2e7e2ab554003cdeb6b94cee639bd68a9404b66218366440011f09703ffb not found: ID does not exist" Mar 13 16:34:39 crc kubenswrapper[4907]: I0313 16:34:39.615214 4907 scope.go:117] "RemoveContainer" containerID="1eb85ed4c37b80885e34300324c44dd232a6deab53c7b12db684fc1106d17c93" Mar 13 16:34:39 crc kubenswrapper[4907]: E0313 16:34:39.615615 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1eb85ed4c37b80885e34300324c44dd232a6deab53c7b12db684fc1106d17c93\": container with ID starting with 1eb85ed4c37b80885e34300324c44dd232a6deab53c7b12db684fc1106d17c93 not found: ID does not exist" containerID="1eb85ed4c37b80885e34300324c44dd232a6deab53c7b12db684fc1106d17c93" Mar 13 16:34:39 crc kubenswrapper[4907]: I0313 16:34:39.615670 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1eb85ed4c37b80885e34300324c44dd232a6deab53c7b12db684fc1106d17c93"} err="failed to get container status \"1eb85ed4c37b80885e34300324c44dd232a6deab53c7b12db684fc1106d17c93\": rpc error: code = NotFound desc = could not find container \"1eb85ed4c37b80885e34300324c44dd232a6deab53c7b12db684fc1106d17c93\": container with ID starting with 1eb85ed4c37b80885e34300324c44dd232a6deab53c7b12db684fc1106d17c93 not found: ID does not exist" Mar 13 16:34:39 crc kubenswrapper[4907]: I0313 16:34:39.615687 4907 scope.go:117] "RemoveContainer" containerID="fd08cb1e0e69433e63613bf6f882fd195d81484e09fd17ba0e96e23e939edb66" Mar 13 16:34:39 crc kubenswrapper[4907]: E0313 16:34:39.616105 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fd08cb1e0e69433e63613bf6f882fd195d81484e09fd17ba0e96e23e939edb66\": container with ID starting with fd08cb1e0e69433e63613bf6f882fd195d81484e09fd17ba0e96e23e939edb66 not found: ID does not exist" containerID="fd08cb1e0e69433e63613bf6f882fd195d81484e09fd17ba0e96e23e939edb66" Mar 13 16:34:39 crc kubenswrapper[4907]: I0313 16:34:39.616158 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fd08cb1e0e69433e63613bf6f882fd195d81484e09fd17ba0e96e23e939edb66"} err="failed to get container status \"fd08cb1e0e69433e63613bf6f882fd195d81484e09fd17ba0e96e23e939edb66\": rpc error: code = NotFound desc = could not find container \"fd08cb1e0e69433e63613bf6f882fd195d81484e09fd17ba0e96e23e939edb66\": container with ID starting with fd08cb1e0e69433e63613bf6f882fd195d81484e09fd17ba0e96e23e939edb66 not found: ID does not exist" Mar 13 16:34:39 crc kubenswrapper[4907]: I0313 16:34:39.806274 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b70871f8-ca8b-4909-9d67-b7e720821aea" path="/var/lib/kubelet/pods/b70871f8-ca8b-4909-9d67-b7e720821aea/volumes" Mar 13 16:34:43 crc kubenswrapper[4907]: I0313 16:34:43.783213 4907 scope.go:117] "RemoveContainer" containerID="68b2fa8f8bf43cf605cf2bff20ff9f2bc9eb89c3d74246f3836b57e3ae2b5d1c" Mar 13 16:34:48 crc kubenswrapper[4907]: I0313 16:34:48.041410 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 16:34:48 crc kubenswrapper[4907]: I0313 16:34:48.041832 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 16:35:18 crc kubenswrapper[4907]: I0313 16:35:18.041871 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 16:35:18 crc kubenswrapper[4907]: I0313 16:35:18.042312 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 16:35:48 crc kubenswrapper[4907]: I0313 16:35:48.042077 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 16:35:48 crc kubenswrapper[4907]: I0313 16:35:48.042737 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 16:35:48 crc kubenswrapper[4907]: I0313 16:35:48.042799 4907 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" Mar 13 16:35:48 crc kubenswrapper[4907]: I0313 16:35:48.043994 4907 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f9156a2a4d423d000cb09f67e16772f675217bb2e483ba02adf57b642ffdffc9"} pod="openshift-machine-config-operator/machine-config-daemon-hm56j" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 16:35:48 crc kubenswrapper[4907]: I0313 16:35:48.044100 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" containerID="cri-o://f9156a2a4d423d000cb09f67e16772f675217bb2e483ba02adf57b642ffdffc9" gracePeriod=600 Mar 13 16:35:48 crc kubenswrapper[4907]: I0313 16:35:48.405286 4907 generic.go:334] "Generic (PLEG): container finished" podID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerID="f9156a2a4d423d000cb09f67e16772f675217bb2e483ba02adf57b642ffdffc9" exitCode=0 Mar 13 16:35:48 crc kubenswrapper[4907]: I0313 16:35:48.405356 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerDied","Data":"f9156a2a4d423d000cb09f67e16772f675217bb2e483ba02adf57b642ffdffc9"} Mar 13 16:35:48 crc kubenswrapper[4907]: I0313 16:35:48.405707 4907 scope.go:117] "RemoveContainer" containerID="dfb327c310937cae28a687e8c5c0256a2ab50e59a756669f2586112310e1dc69" Mar 13 16:35:49 crc kubenswrapper[4907]: I0313 16:35:49.420714 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerStarted","Data":"7e4d0690f4ac607ea8a268bfa88b5ae6d9a95f8b6013de7747fdffb429f28047"} Mar 13 16:36:00 crc kubenswrapper[4907]: I0313 16:36:00.209963 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556996-hhmgk"] Mar 13 16:36:00 crc kubenswrapper[4907]: E0313 16:36:00.211206 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b70871f8-ca8b-4909-9d67-b7e720821aea" containerName="extract-content" Mar 13 16:36:00 crc kubenswrapper[4907]: I0313 16:36:00.211225 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="b70871f8-ca8b-4909-9d67-b7e720821aea" containerName="extract-content" Mar 13 16:36:00 crc kubenswrapper[4907]: E0313 16:36:00.211241 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b70871f8-ca8b-4909-9d67-b7e720821aea" containerName="registry-server" Mar 13 16:36:00 crc kubenswrapper[4907]: I0313 16:36:00.211251 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="b70871f8-ca8b-4909-9d67-b7e720821aea" containerName="registry-server" Mar 13 16:36:00 crc kubenswrapper[4907]: E0313 16:36:00.211277 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92a3a8bb-d3df-4489-8cf5-aacdad7516a9" containerName="extract-content" Mar 13 16:36:00 crc kubenswrapper[4907]: I0313 16:36:00.211287 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="92a3a8bb-d3df-4489-8cf5-aacdad7516a9" containerName="extract-content" Mar 13 16:36:00 crc kubenswrapper[4907]: E0313 16:36:00.211297 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92a3a8bb-d3df-4489-8cf5-aacdad7516a9" containerName="registry-server" Mar 13 16:36:00 crc kubenswrapper[4907]: I0313 16:36:00.211305 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="92a3a8bb-d3df-4489-8cf5-aacdad7516a9" containerName="registry-server" Mar 13 16:36:00 crc kubenswrapper[4907]: E0313 16:36:00.211342 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92a3a8bb-d3df-4489-8cf5-aacdad7516a9" containerName="extract-utilities" Mar 13 16:36:00 crc kubenswrapper[4907]: I0313 16:36:00.211352 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="92a3a8bb-d3df-4489-8cf5-aacdad7516a9" containerName="extract-utilities" Mar 13 16:36:00 crc kubenswrapper[4907]: E0313 16:36:00.211377 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b70871f8-ca8b-4909-9d67-b7e720821aea" containerName="extract-utilities" Mar 13 16:36:00 crc kubenswrapper[4907]: I0313 16:36:00.211385 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="b70871f8-ca8b-4909-9d67-b7e720821aea" containerName="extract-utilities" Mar 13 16:36:00 crc kubenswrapper[4907]: I0313 16:36:00.211640 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="92a3a8bb-d3df-4489-8cf5-aacdad7516a9" containerName="registry-server" Mar 13 16:36:00 crc kubenswrapper[4907]: I0313 16:36:00.211671 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="b70871f8-ca8b-4909-9d67-b7e720821aea" containerName="registry-server" Mar 13 16:36:00 crc kubenswrapper[4907]: I0313 16:36:00.212563 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556996-hhmgk" Mar 13 16:36:00 crc kubenswrapper[4907]: I0313 16:36:00.227928 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 16:36:00 crc kubenswrapper[4907]: I0313 16:36:00.228294 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 16:36:00 crc kubenswrapper[4907]: I0313 16:36:00.228561 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 16:36:00 crc kubenswrapper[4907]: I0313 16:36:00.244234 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556996-hhmgk"] Mar 13 16:36:00 crc kubenswrapper[4907]: I0313 16:36:00.359028 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vktpm\" (UniqueName: \"kubernetes.io/projected/d0ba1c68-f3a1-4c88-bc6b-db7c438dd2f5-kube-api-access-vktpm\") pod \"auto-csr-approver-29556996-hhmgk\" (UID: \"d0ba1c68-f3a1-4c88-bc6b-db7c438dd2f5\") " pod="openshift-infra/auto-csr-approver-29556996-hhmgk" Mar 13 16:36:00 crc kubenswrapper[4907]: I0313 16:36:00.461690 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vktpm\" (UniqueName: \"kubernetes.io/projected/d0ba1c68-f3a1-4c88-bc6b-db7c438dd2f5-kube-api-access-vktpm\") pod \"auto-csr-approver-29556996-hhmgk\" (UID: \"d0ba1c68-f3a1-4c88-bc6b-db7c438dd2f5\") " pod="openshift-infra/auto-csr-approver-29556996-hhmgk" Mar 13 16:36:00 crc kubenswrapper[4907]: I0313 16:36:00.483394 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vktpm\" (UniqueName: \"kubernetes.io/projected/d0ba1c68-f3a1-4c88-bc6b-db7c438dd2f5-kube-api-access-vktpm\") pod \"auto-csr-approver-29556996-hhmgk\" (UID: \"d0ba1c68-f3a1-4c88-bc6b-db7c438dd2f5\") " pod="openshift-infra/auto-csr-approver-29556996-hhmgk" Mar 13 16:36:00 crc kubenswrapper[4907]: I0313 16:36:00.566162 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556996-hhmgk" Mar 13 16:36:01 crc kubenswrapper[4907]: I0313 16:36:01.107303 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556996-hhmgk"] Mar 13 16:36:01 crc kubenswrapper[4907]: I0313 16:36:01.563569 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556996-hhmgk" event={"ID":"d0ba1c68-f3a1-4c88-bc6b-db7c438dd2f5","Type":"ContainerStarted","Data":"8593f1dbfaa7ee1419c6a64af364f636ad809d767e09866e08ca1068367283d8"} Mar 13 16:36:02 crc kubenswrapper[4907]: I0313 16:36:02.578201 4907 generic.go:334] "Generic (PLEG): container finished" podID="d0ba1c68-f3a1-4c88-bc6b-db7c438dd2f5" containerID="fc2a84d7603e9a5913a999c6beb101ceda891932552390c08d77f4d764a6c116" exitCode=0 Mar 13 16:36:02 crc kubenswrapper[4907]: I0313 16:36:02.578277 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556996-hhmgk" event={"ID":"d0ba1c68-f3a1-4c88-bc6b-db7c438dd2f5","Type":"ContainerDied","Data":"fc2a84d7603e9a5913a999c6beb101ceda891932552390c08d77f4d764a6c116"} Mar 13 16:36:04 crc kubenswrapper[4907]: I0313 16:36:04.092396 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556996-hhmgk" Mar 13 16:36:04 crc kubenswrapper[4907]: I0313 16:36:04.256554 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vktpm\" (UniqueName: \"kubernetes.io/projected/d0ba1c68-f3a1-4c88-bc6b-db7c438dd2f5-kube-api-access-vktpm\") pod \"d0ba1c68-f3a1-4c88-bc6b-db7c438dd2f5\" (UID: \"d0ba1c68-f3a1-4c88-bc6b-db7c438dd2f5\") " Mar 13 16:36:04 crc kubenswrapper[4907]: I0313 16:36:04.268226 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d0ba1c68-f3a1-4c88-bc6b-db7c438dd2f5-kube-api-access-vktpm" (OuterVolumeSpecName: "kube-api-access-vktpm") pod "d0ba1c68-f3a1-4c88-bc6b-db7c438dd2f5" (UID: "d0ba1c68-f3a1-4c88-bc6b-db7c438dd2f5"). InnerVolumeSpecName "kube-api-access-vktpm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:36:04 crc kubenswrapper[4907]: I0313 16:36:04.359515 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vktpm\" (UniqueName: \"kubernetes.io/projected/d0ba1c68-f3a1-4c88-bc6b-db7c438dd2f5-kube-api-access-vktpm\") on node \"crc\" DevicePath \"\"" Mar 13 16:36:04 crc kubenswrapper[4907]: I0313 16:36:04.602434 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556996-hhmgk" event={"ID":"d0ba1c68-f3a1-4c88-bc6b-db7c438dd2f5","Type":"ContainerDied","Data":"8593f1dbfaa7ee1419c6a64af364f636ad809d767e09866e08ca1068367283d8"} Mar 13 16:36:04 crc kubenswrapper[4907]: I0313 16:36:04.602473 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556996-hhmgk" Mar 13 16:36:04 crc kubenswrapper[4907]: I0313 16:36:04.602487 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8593f1dbfaa7ee1419c6a64af364f636ad809d767e09866e08ca1068367283d8" Mar 13 16:36:05 crc kubenswrapper[4907]: I0313 16:36:05.167942 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556990-4jrg4"] Mar 13 16:36:05 crc kubenswrapper[4907]: I0313 16:36:05.183217 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556990-4jrg4"] Mar 13 16:36:05 crc kubenswrapper[4907]: I0313 16:36:05.793961 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5a413954-e701-4cec-8868-4fcdaecfabe8" path="/var/lib/kubelet/pods/5a413954-e701-4cec-8868-4fcdaecfabe8/volumes" Mar 13 16:36:08 crc kubenswrapper[4907]: I0313 16:36:08.068713 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-copy-data"] Mar 13 16:36:08 crc kubenswrapper[4907]: I0313 16:36:08.069398 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/mariadb-copy-data" podUID="6b679563-ed42-418a-80ef-0b24e225d5ef" containerName="adoption" containerID="cri-o://be1062c79f4023088d8eb4fb41e0f370f57e80c1813e6265704fdadc605e804a" gracePeriod=30 Mar 13 16:36:39 crc kubenswrapper[4907]: I0313 16:36:39.013653 4907 generic.go:334] "Generic (PLEG): container finished" podID="6b679563-ed42-418a-80ef-0b24e225d5ef" containerID="be1062c79f4023088d8eb4fb41e0f370f57e80c1813e6265704fdadc605e804a" exitCode=137 Mar 13 16:36:39 crc kubenswrapper[4907]: I0313 16:36:39.014016 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-copy-data" event={"ID":"6b679563-ed42-418a-80ef-0b24e225d5ef","Type":"ContainerDied","Data":"be1062c79f4023088d8eb4fb41e0f370f57e80c1813e6265704fdadc605e804a"} Mar 13 16:36:39 crc kubenswrapper[4907]: I0313 16:36:39.167768 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-copy-data" Mar 13 16:36:39 crc kubenswrapper[4907]: I0313 16:36:39.298667 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mariadb-data\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3b4b3b59-8b2d-44c3-bc88-a79545d23d7c\") pod \"6b679563-ed42-418a-80ef-0b24e225d5ef\" (UID: \"6b679563-ed42-418a-80ef-0b24e225d5ef\") " Mar 13 16:36:39 crc kubenswrapper[4907]: I0313 16:36:39.298785 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rzv64\" (UniqueName: \"kubernetes.io/projected/6b679563-ed42-418a-80ef-0b24e225d5ef-kube-api-access-rzv64\") pod \"6b679563-ed42-418a-80ef-0b24e225d5ef\" (UID: \"6b679563-ed42-418a-80ef-0b24e225d5ef\") " Mar 13 16:36:39 crc kubenswrapper[4907]: I0313 16:36:39.307813 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6b679563-ed42-418a-80ef-0b24e225d5ef-kube-api-access-rzv64" (OuterVolumeSpecName: "kube-api-access-rzv64") pod "6b679563-ed42-418a-80ef-0b24e225d5ef" (UID: "6b679563-ed42-418a-80ef-0b24e225d5ef"). InnerVolumeSpecName "kube-api-access-rzv64". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:36:39 crc kubenswrapper[4907]: I0313 16:36:39.335158 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3b4b3b59-8b2d-44c3-bc88-a79545d23d7c" (OuterVolumeSpecName: "mariadb-data") pod "6b679563-ed42-418a-80ef-0b24e225d5ef" (UID: "6b679563-ed42-418a-80ef-0b24e225d5ef"). InnerVolumeSpecName "pvc-3b4b3b59-8b2d-44c3-bc88-a79545d23d7c". PluginName "kubernetes.io/csi", VolumeGidValue "" Mar 13 16:36:39 crc kubenswrapper[4907]: I0313 16:36:39.401824 4907 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-3b4b3b59-8b2d-44c3-bc88-a79545d23d7c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3b4b3b59-8b2d-44c3-bc88-a79545d23d7c\") on node \"crc\" " Mar 13 16:36:39 crc kubenswrapper[4907]: I0313 16:36:39.402027 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rzv64\" (UniqueName: \"kubernetes.io/projected/6b679563-ed42-418a-80ef-0b24e225d5ef-kube-api-access-rzv64\") on node \"crc\" DevicePath \"\"" Mar 13 16:36:39 crc kubenswrapper[4907]: I0313 16:36:39.435219 4907 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Mar 13 16:36:39 crc kubenswrapper[4907]: I0313 16:36:39.435472 4907 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-3b4b3b59-8b2d-44c3-bc88-a79545d23d7c" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3b4b3b59-8b2d-44c3-bc88-a79545d23d7c") on node "crc" Mar 13 16:36:39 crc kubenswrapper[4907]: I0313 16:36:39.504340 4907 reconciler_common.go:293] "Volume detached for volume \"pvc-3b4b3b59-8b2d-44c3-bc88-a79545d23d7c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3b4b3b59-8b2d-44c3-bc88-a79545d23d7c\") on node \"crc\" DevicePath \"\"" Mar 13 16:36:40 crc kubenswrapper[4907]: I0313 16:36:40.031253 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-copy-data" event={"ID":"6b679563-ed42-418a-80ef-0b24e225d5ef","Type":"ContainerDied","Data":"763a1aa9a599cf6e70f4f78d5d2da41195b4bd8f467a89e3d789f85b36605af1"} Mar 13 16:36:40 crc kubenswrapper[4907]: I0313 16:36:40.031334 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-copy-data" Mar 13 16:36:40 crc kubenswrapper[4907]: I0313 16:36:40.032075 4907 scope.go:117] "RemoveContainer" containerID="be1062c79f4023088d8eb4fb41e0f370f57e80c1813e6265704fdadc605e804a" Mar 13 16:36:40 crc kubenswrapper[4907]: I0313 16:36:40.073424 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-copy-data"] Mar 13 16:36:40 crc kubenswrapper[4907]: I0313 16:36:40.086125 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-copy-data"] Mar 13 16:36:40 crc kubenswrapper[4907]: I0313 16:36:40.823329 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-copy-data"] Mar 13 16:36:40 crc kubenswrapper[4907]: I0313 16:36:40.823662 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-copy-data" podUID="e0c0f09a-8646-4326-86e8-5bd9123d78ca" containerName="adoption" containerID="cri-o://d115558c971717b0e8dd16d1b2ccb9b5004c901031eda65c32db4372508a5b7c" gracePeriod=30 Mar 13 16:36:41 crc kubenswrapper[4907]: I0313 16:36:41.795571 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6b679563-ed42-418a-80ef-0b24e225d5ef" path="/var/lib/kubelet/pods/6b679563-ed42-418a-80ef-0b24e225d5ef/volumes" Mar 13 16:36:43 crc kubenswrapper[4907]: I0313 16:36:43.961446 4907 scope.go:117] "RemoveContainer" containerID="5585587b3788c101f53b09ee74a77a65c4d4a4773da4d3e62432619e4599efa1" Mar 13 16:37:11 crc kubenswrapper[4907]: I0313 16:37:11.351154 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-copy-data" Mar 13 16:37:11 crc kubenswrapper[4907]: I0313 16:37:11.430298 4907 generic.go:334] "Generic (PLEG): container finished" podID="e0c0f09a-8646-4326-86e8-5bd9123d78ca" containerID="d115558c971717b0e8dd16d1b2ccb9b5004c901031eda65c32db4372508a5b7c" exitCode=137 Mar 13 16:37:11 crc kubenswrapper[4907]: I0313 16:37:11.430348 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-copy-data" event={"ID":"e0c0f09a-8646-4326-86e8-5bd9123d78ca","Type":"ContainerDied","Data":"d115558c971717b0e8dd16d1b2ccb9b5004c901031eda65c32db4372508a5b7c"} Mar 13 16:37:11 crc kubenswrapper[4907]: I0313 16:37:11.430378 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-copy-data" event={"ID":"e0c0f09a-8646-4326-86e8-5bd9123d78ca","Type":"ContainerDied","Data":"afb18a83c45725fa475076eb88bc1f59d47ff74d8c364938ed9892b1ba42b1be"} Mar 13 16:37:11 crc kubenswrapper[4907]: I0313 16:37:11.430398 4907 scope.go:117] "RemoveContainer" containerID="d115558c971717b0e8dd16d1b2ccb9b5004c901031eda65c32db4372508a5b7c" Mar 13 16:37:11 crc kubenswrapper[4907]: I0313 16:37:11.430561 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-copy-data" Mar 13 16:37:11 crc kubenswrapper[4907]: I0313 16:37:11.459480 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-data\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1a945566-190e-43e6-ae94-58c483d5c2e4\") pod \"e0c0f09a-8646-4326-86e8-5bd9123d78ca\" (UID: \"e0c0f09a-8646-4326-86e8-5bd9123d78ca\") " Mar 13 16:37:11 crc kubenswrapper[4907]: I0313 16:37:11.459584 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/e0c0f09a-8646-4326-86e8-5bd9123d78ca-ovn-data-cert\") pod \"e0c0f09a-8646-4326-86e8-5bd9123d78ca\" (UID: \"e0c0f09a-8646-4326-86e8-5bd9123d78ca\") " Mar 13 16:37:11 crc kubenswrapper[4907]: I0313 16:37:11.459689 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ctpt2\" (UniqueName: \"kubernetes.io/projected/e0c0f09a-8646-4326-86e8-5bd9123d78ca-kube-api-access-ctpt2\") pod \"e0c0f09a-8646-4326-86e8-5bd9123d78ca\" (UID: \"e0c0f09a-8646-4326-86e8-5bd9123d78ca\") " Mar 13 16:37:11 crc kubenswrapper[4907]: I0313 16:37:11.461351 4907 scope.go:117] "RemoveContainer" containerID="d115558c971717b0e8dd16d1b2ccb9b5004c901031eda65c32db4372508a5b7c" Mar 13 16:37:11 crc kubenswrapper[4907]: E0313 16:37:11.462206 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d115558c971717b0e8dd16d1b2ccb9b5004c901031eda65c32db4372508a5b7c\": container with ID starting with d115558c971717b0e8dd16d1b2ccb9b5004c901031eda65c32db4372508a5b7c not found: ID does not exist" containerID="d115558c971717b0e8dd16d1b2ccb9b5004c901031eda65c32db4372508a5b7c" Mar 13 16:37:11 crc kubenswrapper[4907]: I0313 16:37:11.462248 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d115558c971717b0e8dd16d1b2ccb9b5004c901031eda65c32db4372508a5b7c"} err="failed to get container status \"d115558c971717b0e8dd16d1b2ccb9b5004c901031eda65c32db4372508a5b7c\": rpc error: code = NotFound desc = could not find container \"d115558c971717b0e8dd16d1b2ccb9b5004c901031eda65c32db4372508a5b7c\": container with ID starting with d115558c971717b0e8dd16d1b2ccb9b5004c901031eda65c32db4372508a5b7c not found: ID does not exist" Mar 13 16:37:11 crc kubenswrapper[4907]: I0313 16:37:11.466435 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e0c0f09a-8646-4326-86e8-5bd9123d78ca-kube-api-access-ctpt2" (OuterVolumeSpecName: "kube-api-access-ctpt2") pod "e0c0f09a-8646-4326-86e8-5bd9123d78ca" (UID: "e0c0f09a-8646-4326-86e8-5bd9123d78ca"). InnerVolumeSpecName "kube-api-access-ctpt2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:37:11 crc kubenswrapper[4907]: I0313 16:37:11.468796 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e0c0f09a-8646-4326-86e8-5bd9123d78ca-ovn-data-cert" (OuterVolumeSpecName: "ovn-data-cert") pod "e0c0f09a-8646-4326-86e8-5bd9123d78ca" (UID: "e0c0f09a-8646-4326-86e8-5bd9123d78ca"). InnerVolumeSpecName "ovn-data-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:37:11 crc kubenswrapper[4907]: I0313 16:37:11.488393 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1a945566-190e-43e6-ae94-58c483d5c2e4" (OuterVolumeSpecName: "ovn-data") pod "e0c0f09a-8646-4326-86e8-5bd9123d78ca" (UID: "e0c0f09a-8646-4326-86e8-5bd9123d78ca"). InnerVolumeSpecName "pvc-1a945566-190e-43e6-ae94-58c483d5c2e4". PluginName "kubernetes.io/csi", VolumeGidValue "" Mar 13 16:37:11 crc kubenswrapper[4907]: I0313 16:37:11.562261 4907 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-1a945566-190e-43e6-ae94-58c483d5c2e4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1a945566-190e-43e6-ae94-58c483d5c2e4\") on node \"crc\" " Mar 13 16:37:11 crc kubenswrapper[4907]: I0313 16:37:11.562292 4907 reconciler_common.go:293] "Volume detached for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/e0c0f09a-8646-4326-86e8-5bd9123d78ca-ovn-data-cert\") on node \"crc\" DevicePath \"\"" Mar 13 16:37:11 crc kubenswrapper[4907]: I0313 16:37:11.562323 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ctpt2\" (UniqueName: \"kubernetes.io/projected/e0c0f09a-8646-4326-86e8-5bd9123d78ca-kube-api-access-ctpt2\") on node \"crc\" DevicePath \"\"" Mar 13 16:37:11 crc kubenswrapper[4907]: I0313 16:37:11.588265 4907 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Mar 13 16:37:11 crc kubenswrapper[4907]: I0313 16:37:11.588829 4907 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-1a945566-190e-43e6-ae94-58c483d5c2e4" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1a945566-190e-43e6-ae94-58c483d5c2e4") on node "crc" Mar 13 16:37:11 crc kubenswrapper[4907]: I0313 16:37:11.664250 4907 reconciler_common.go:293] "Volume detached for volume \"pvc-1a945566-190e-43e6-ae94-58c483d5c2e4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1a945566-190e-43e6-ae94-58c483d5c2e4\") on node \"crc\" DevicePath \"\"" Mar 13 16:37:11 crc kubenswrapper[4907]: I0313 16:37:11.777605 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-copy-data"] Mar 13 16:37:11 crc kubenswrapper[4907]: I0313 16:37:11.806389 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-copy-data"] Mar 13 16:37:13 crc kubenswrapper[4907]: I0313 16:37:13.811976 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e0c0f09a-8646-4326-86e8-5bd9123d78ca" path="/var/lib/kubelet/pods/e0c0f09a-8646-4326-86e8-5bd9123d78ca/volumes" Mar 13 16:37:48 crc kubenswrapper[4907]: I0313 16:37:48.041439 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 16:37:48 crc kubenswrapper[4907]: I0313 16:37:48.042195 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 16:38:00 crc kubenswrapper[4907]: I0313 16:38:00.168087 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556998-pzdj2"] Mar 13 16:38:00 crc kubenswrapper[4907]: E0313 16:38:00.169056 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0c0f09a-8646-4326-86e8-5bd9123d78ca" containerName="adoption" Mar 13 16:38:00 crc kubenswrapper[4907]: I0313 16:38:00.169071 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0c0f09a-8646-4326-86e8-5bd9123d78ca" containerName="adoption" Mar 13 16:38:00 crc kubenswrapper[4907]: E0313 16:38:00.169107 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b679563-ed42-418a-80ef-0b24e225d5ef" containerName="adoption" Mar 13 16:38:00 crc kubenswrapper[4907]: I0313 16:38:00.169114 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b679563-ed42-418a-80ef-0b24e225d5ef" containerName="adoption" Mar 13 16:38:00 crc kubenswrapper[4907]: E0313 16:38:00.169126 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0ba1c68-f3a1-4c88-bc6b-db7c438dd2f5" containerName="oc" Mar 13 16:38:00 crc kubenswrapper[4907]: I0313 16:38:00.169132 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0ba1c68-f3a1-4c88-bc6b-db7c438dd2f5" containerName="oc" Mar 13 16:38:00 crc kubenswrapper[4907]: I0313 16:38:00.169310 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="e0c0f09a-8646-4326-86e8-5bd9123d78ca" containerName="adoption" Mar 13 16:38:00 crc kubenswrapper[4907]: I0313 16:38:00.169330 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0ba1c68-f3a1-4c88-bc6b-db7c438dd2f5" containerName="oc" Mar 13 16:38:00 crc kubenswrapper[4907]: I0313 16:38:00.169347 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b679563-ed42-418a-80ef-0b24e225d5ef" containerName="adoption" Mar 13 16:38:00 crc kubenswrapper[4907]: I0313 16:38:00.170101 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556998-pzdj2" Mar 13 16:38:00 crc kubenswrapper[4907]: I0313 16:38:00.174349 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 16:38:00 crc kubenswrapper[4907]: I0313 16:38:00.174558 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 16:38:00 crc kubenswrapper[4907]: I0313 16:38:00.174713 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 16:38:00 crc kubenswrapper[4907]: I0313 16:38:00.189617 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556998-pzdj2"] Mar 13 16:38:00 crc kubenswrapper[4907]: I0313 16:38:00.370140 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zj7j7\" (UniqueName: \"kubernetes.io/projected/2abff695-80c2-4588-9796-44e850ddb65f-kube-api-access-zj7j7\") pod \"auto-csr-approver-29556998-pzdj2\" (UID: \"2abff695-80c2-4588-9796-44e850ddb65f\") " pod="openshift-infra/auto-csr-approver-29556998-pzdj2" Mar 13 16:38:00 crc kubenswrapper[4907]: I0313 16:38:00.473068 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zj7j7\" (UniqueName: \"kubernetes.io/projected/2abff695-80c2-4588-9796-44e850ddb65f-kube-api-access-zj7j7\") pod \"auto-csr-approver-29556998-pzdj2\" (UID: \"2abff695-80c2-4588-9796-44e850ddb65f\") " pod="openshift-infra/auto-csr-approver-29556998-pzdj2" Mar 13 16:38:00 crc kubenswrapper[4907]: I0313 16:38:00.498805 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zj7j7\" (UniqueName: \"kubernetes.io/projected/2abff695-80c2-4588-9796-44e850ddb65f-kube-api-access-zj7j7\") pod \"auto-csr-approver-29556998-pzdj2\" (UID: \"2abff695-80c2-4588-9796-44e850ddb65f\") " pod="openshift-infra/auto-csr-approver-29556998-pzdj2" Mar 13 16:38:00 crc kubenswrapper[4907]: I0313 16:38:00.787738 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556998-pzdj2" Mar 13 16:38:01 crc kubenswrapper[4907]: I0313 16:38:01.110807 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556998-pzdj2"] Mar 13 16:38:02 crc kubenswrapper[4907]: I0313 16:38:02.055943 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556998-pzdj2" event={"ID":"2abff695-80c2-4588-9796-44e850ddb65f","Type":"ContainerStarted","Data":"0fce101df987e00b5ae546936da1d9debe45328102484758ec2c10857209676f"} Mar 13 16:38:03 crc kubenswrapper[4907]: I0313 16:38:03.072030 4907 generic.go:334] "Generic (PLEG): container finished" podID="2abff695-80c2-4588-9796-44e850ddb65f" containerID="406f17bcc9c6733e218df39e7a5238412a849d1df878824a54db3db1358e5bff" exitCode=0 Mar 13 16:38:03 crc kubenswrapper[4907]: I0313 16:38:03.072142 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556998-pzdj2" event={"ID":"2abff695-80c2-4588-9796-44e850ddb65f","Type":"ContainerDied","Data":"406f17bcc9c6733e218df39e7a5238412a849d1df878824a54db3db1358e5bff"} Mar 13 16:38:04 crc kubenswrapper[4907]: I0313 16:38:04.520505 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556998-pzdj2" Mar 13 16:38:04 crc kubenswrapper[4907]: I0313 16:38:04.693040 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zj7j7\" (UniqueName: \"kubernetes.io/projected/2abff695-80c2-4588-9796-44e850ddb65f-kube-api-access-zj7j7\") pod \"2abff695-80c2-4588-9796-44e850ddb65f\" (UID: \"2abff695-80c2-4588-9796-44e850ddb65f\") " Mar 13 16:38:04 crc kubenswrapper[4907]: I0313 16:38:04.698363 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2abff695-80c2-4588-9796-44e850ddb65f-kube-api-access-zj7j7" (OuterVolumeSpecName: "kube-api-access-zj7j7") pod "2abff695-80c2-4588-9796-44e850ddb65f" (UID: "2abff695-80c2-4588-9796-44e850ddb65f"). InnerVolumeSpecName "kube-api-access-zj7j7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:38:04 crc kubenswrapper[4907]: I0313 16:38:04.796150 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zj7j7\" (UniqueName: \"kubernetes.io/projected/2abff695-80c2-4588-9796-44e850ddb65f-kube-api-access-zj7j7\") on node \"crc\" DevicePath \"\"" Mar 13 16:38:05 crc kubenswrapper[4907]: I0313 16:38:05.105693 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556998-pzdj2" event={"ID":"2abff695-80c2-4588-9796-44e850ddb65f","Type":"ContainerDied","Data":"0fce101df987e00b5ae546936da1d9debe45328102484758ec2c10857209676f"} Mar 13 16:38:05 crc kubenswrapper[4907]: I0313 16:38:05.106082 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0fce101df987e00b5ae546936da1d9debe45328102484758ec2c10857209676f" Mar 13 16:38:05 crc kubenswrapper[4907]: I0313 16:38:05.105865 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556998-pzdj2" Mar 13 16:38:05 crc kubenswrapper[4907]: I0313 16:38:05.597374 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556992-qj2mm"] Mar 13 16:38:05 crc kubenswrapper[4907]: I0313 16:38:05.607185 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556992-qj2mm"] Mar 13 16:38:05 crc kubenswrapper[4907]: I0313 16:38:05.793962 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5d2dfd3f-0bc6-48f9-85f7-100ef8f23476" path="/var/lib/kubelet/pods/5d2dfd3f-0bc6-48f9-85f7-100ef8f23476/volumes" Mar 13 16:38:08 crc kubenswrapper[4907]: I0313 16:38:08.638298 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-xndl6/must-gather-8f2lx"] Mar 13 16:38:08 crc kubenswrapper[4907]: E0313 16:38:08.639257 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2abff695-80c2-4588-9796-44e850ddb65f" containerName="oc" Mar 13 16:38:08 crc kubenswrapper[4907]: I0313 16:38:08.639270 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="2abff695-80c2-4588-9796-44e850ddb65f" containerName="oc" Mar 13 16:38:08 crc kubenswrapper[4907]: I0313 16:38:08.639464 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="2abff695-80c2-4588-9796-44e850ddb65f" containerName="oc" Mar 13 16:38:08 crc kubenswrapper[4907]: I0313 16:38:08.640625 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xndl6/must-gather-8f2lx" Mar 13 16:38:08 crc kubenswrapper[4907]: I0313 16:38:08.642450 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-xndl6"/"default-dockercfg-m9xng" Mar 13 16:38:08 crc kubenswrapper[4907]: I0313 16:38:08.642955 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-xndl6"/"kube-root-ca.crt" Mar 13 16:38:08 crc kubenswrapper[4907]: I0313 16:38:08.643136 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-xndl6"/"openshift-service-ca.crt" Mar 13 16:38:08 crc kubenswrapper[4907]: I0313 16:38:08.647225 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-xndl6/must-gather-8f2lx"] Mar 13 16:38:08 crc kubenswrapper[4907]: I0313 16:38:08.786728 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lsrlm\" (UniqueName: \"kubernetes.io/projected/3d121a10-6251-43fa-af3a-0d2520c87c8d-kube-api-access-lsrlm\") pod \"must-gather-8f2lx\" (UID: \"3d121a10-6251-43fa-af3a-0d2520c87c8d\") " pod="openshift-must-gather-xndl6/must-gather-8f2lx" Mar 13 16:38:08 crc kubenswrapper[4907]: I0313 16:38:08.786898 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/3d121a10-6251-43fa-af3a-0d2520c87c8d-must-gather-output\") pod \"must-gather-8f2lx\" (UID: \"3d121a10-6251-43fa-af3a-0d2520c87c8d\") " pod="openshift-must-gather-xndl6/must-gather-8f2lx" Mar 13 16:38:08 crc kubenswrapper[4907]: I0313 16:38:08.889166 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/3d121a10-6251-43fa-af3a-0d2520c87c8d-must-gather-output\") pod \"must-gather-8f2lx\" (UID: \"3d121a10-6251-43fa-af3a-0d2520c87c8d\") " pod="openshift-must-gather-xndl6/must-gather-8f2lx" Mar 13 16:38:08 crc kubenswrapper[4907]: I0313 16:38:08.889937 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lsrlm\" (UniqueName: \"kubernetes.io/projected/3d121a10-6251-43fa-af3a-0d2520c87c8d-kube-api-access-lsrlm\") pod \"must-gather-8f2lx\" (UID: \"3d121a10-6251-43fa-af3a-0d2520c87c8d\") " pod="openshift-must-gather-xndl6/must-gather-8f2lx" Mar 13 16:38:08 crc kubenswrapper[4907]: I0313 16:38:08.891944 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/3d121a10-6251-43fa-af3a-0d2520c87c8d-must-gather-output\") pod \"must-gather-8f2lx\" (UID: \"3d121a10-6251-43fa-af3a-0d2520c87c8d\") " pod="openshift-must-gather-xndl6/must-gather-8f2lx" Mar 13 16:38:08 crc kubenswrapper[4907]: I0313 16:38:08.915050 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lsrlm\" (UniqueName: \"kubernetes.io/projected/3d121a10-6251-43fa-af3a-0d2520c87c8d-kube-api-access-lsrlm\") pod \"must-gather-8f2lx\" (UID: \"3d121a10-6251-43fa-af3a-0d2520c87c8d\") " pod="openshift-must-gather-xndl6/must-gather-8f2lx" Mar 13 16:38:08 crc kubenswrapper[4907]: I0313 16:38:08.962122 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xndl6/must-gather-8f2lx" Mar 13 16:38:09 crc kubenswrapper[4907]: I0313 16:38:09.479498 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-xndl6/must-gather-8f2lx"] Mar 13 16:38:10 crc kubenswrapper[4907]: I0313 16:38:10.159100 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-xndl6/must-gather-8f2lx" event={"ID":"3d121a10-6251-43fa-af3a-0d2520c87c8d","Type":"ContainerStarted","Data":"269636f3f131040884ec72a96a5ea7fa20184677b56a7a261e911c974a24da8b"} Mar 13 16:38:17 crc kubenswrapper[4907]: I0313 16:38:17.272151 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-xndl6/must-gather-8f2lx" event={"ID":"3d121a10-6251-43fa-af3a-0d2520c87c8d","Type":"ContainerStarted","Data":"82747fef3c01173075fddc60efdb72434cb1285ad8c4937002af8a8d281e869f"} Mar 13 16:38:17 crc kubenswrapper[4907]: I0313 16:38:17.272603 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-xndl6/must-gather-8f2lx" event={"ID":"3d121a10-6251-43fa-af3a-0d2520c87c8d","Type":"ContainerStarted","Data":"0fd7545e78a3c44cdc8c41194a9375675f5137f4e55f454ee84c978a332aef19"} Mar 13 16:38:17 crc kubenswrapper[4907]: I0313 16:38:17.289402 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-xndl6/must-gather-8f2lx" podStartSLOduration=2.20949555 podStartE2EDuration="9.289386532s" podCreationTimestamp="2026-03-13 16:38:08 +0000 UTC" firstStartedPulling="2026-03-13 16:38:09.475788773 +0000 UTC m=+9188.375576452" lastFinishedPulling="2026-03-13 16:38:16.555679755 +0000 UTC m=+9195.455467434" observedRunningTime="2026-03-13 16:38:17.288638001 +0000 UTC m=+9196.188425690" watchObservedRunningTime="2026-03-13 16:38:17.289386532 +0000 UTC m=+9196.189174221" Mar 13 16:38:18 crc kubenswrapper[4907]: I0313 16:38:18.041487 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 16:38:18 crc kubenswrapper[4907]: I0313 16:38:18.041737 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 16:38:20 crc kubenswrapper[4907]: I0313 16:38:20.941940 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-xndl6/crc-debug-qt756"] Mar 13 16:38:20 crc kubenswrapper[4907]: I0313 16:38:20.944199 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xndl6/crc-debug-qt756" Mar 13 16:38:21 crc kubenswrapper[4907]: I0313 16:38:21.069500 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5e5575c9-0bdc-472b-9384-fbd1ce5cb480-host\") pod \"crc-debug-qt756\" (UID: \"5e5575c9-0bdc-472b-9384-fbd1ce5cb480\") " pod="openshift-must-gather-xndl6/crc-debug-qt756" Mar 13 16:38:21 crc kubenswrapper[4907]: I0313 16:38:21.069785 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4vjlv\" (UniqueName: \"kubernetes.io/projected/5e5575c9-0bdc-472b-9384-fbd1ce5cb480-kube-api-access-4vjlv\") pod \"crc-debug-qt756\" (UID: \"5e5575c9-0bdc-472b-9384-fbd1ce5cb480\") " pod="openshift-must-gather-xndl6/crc-debug-qt756" Mar 13 16:38:21 crc kubenswrapper[4907]: I0313 16:38:21.171617 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5e5575c9-0bdc-472b-9384-fbd1ce5cb480-host\") pod \"crc-debug-qt756\" (UID: \"5e5575c9-0bdc-472b-9384-fbd1ce5cb480\") " pod="openshift-must-gather-xndl6/crc-debug-qt756" Mar 13 16:38:21 crc kubenswrapper[4907]: I0313 16:38:21.171725 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4vjlv\" (UniqueName: \"kubernetes.io/projected/5e5575c9-0bdc-472b-9384-fbd1ce5cb480-kube-api-access-4vjlv\") pod \"crc-debug-qt756\" (UID: \"5e5575c9-0bdc-472b-9384-fbd1ce5cb480\") " pod="openshift-must-gather-xndl6/crc-debug-qt756" Mar 13 16:38:21 crc kubenswrapper[4907]: I0313 16:38:21.172093 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5e5575c9-0bdc-472b-9384-fbd1ce5cb480-host\") pod \"crc-debug-qt756\" (UID: \"5e5575c9-0bdc-472b-9384-fbd1ce5cb480\") " pod="openshift-must-gather-xndl6/crc-debug-qt756" Mar 13 16:38:21 crc kubenswrapper[4907]: I0313 16:38:21.189444 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4vjlv\" (UniqueName: \"kubernetes.io/projected/5e5575c9-0bdc-472b-9384-fbd1ce5cb480-kube-api-access-4vjlv\") pod \"crc-debug-qt756\" (UID: \"5e5575c9-0bdc-472b-9384-fbd1ce5cb480\") " pod="openshift-must-gather-xndl6/crc-debug-qt756" Mar 13 16:38:21 crc kubenswrapper[4907]: I0313 16:38:21.263065 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xndl6/crc-debug-qt756" Mar 13 16:38:21 crc kubenswrapper[4907]: I0313 16:38:21.310076 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-xndl6/crc-debug-qt756" event={"ID":"5e5575c9-0bdc-472b-9384-fbd1ce5cb480","Type":"ContainerStarted","Data":"f3e01960410a5bad3f4d4d1123a16e5c7cc67a5d538faaf418827fff78354c63"} Mar 13 16:38:34 crc kubenswrapper[4907]: I0313 16:38:34.457779 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-xndl6/crc-debug-qt756" event={"ID":"5e5575c9-0bdc-472b-9384-fbd1ce5cb480","Type":"ContainerStarted","Data":"b1d7df69f877a557d28aa9c5dd1537366e05d49104f113b20536b953f8c2ec90"} Mar 13 16:38:34 crc kubenswrapper[4907]: I0313 16:38:34.476091 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-xndl6/crc-debug-qt756" podStartSLOduration=2.283363906 podStartE2EDuration="14.476070234s" podCreationTimestamp="2026-03-13 16:38:20 +0000 UTC" firstStartedPulling="2026-03-13 16:38:21.299905065 +0000 UTC m=+9200.199692754" lastFinishedPulling="2026-03-13 16:38:33.492611393 +0000 UTC m=+9212.392399082" observedRunningTime="2026-03-13 16:38:34.471182577 +0000 UTC m=+9213.370970266" watchObservedRunningTime="2026-03-13 16:38:34.476070234 +0000 UTC m=+9213.375857923" Mar 13 16:38:39 crc kubenswrapper[4907]: I0313 16:38:39.546546 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-vz6rb"] Mar 13 16:38:39 crc kubenswrapper[4907]: I0313 16:38:39.549416 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vz6rb" Mar 13 16:38:39 crc kubenswrapper[4907]: I0313 16:38:39.561683 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-vz6rb"] Mar 13 16:38:39 crc kubenswrapper[4907]: I0313 16:38:39.602551 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9qbcl\" (UniqueName: \"kubernetes.io/projected/2b31d6ac-276a-4947-a468-e99a784e1ff3-kube-api-access-9qbcl\") pod \"redhat-marketplace-vz6rb\" (UID: \"2b31d6ac-276a-4947-a468-e99a784e1ff3\") " pod="openshift-marketplace/redhat-marketplace-vz6rb" Mar 13 16:38:39 crc kubenswrapper[4907]: I0313 16:38:39.602606 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b31d6ac-276a-4947-a468-e99a784e1ff3-catalog-content\") pod \"redhat-marketplace-vz6rb\" (UID: \"2b31d6ac-276a-4947-a468-e99a784e1ff3\") " pod="openshift-marketplace/redhat-marketplace-vz6rb" Mar 13 16:38:39 crc kubenswrapper[4907]: I0313 16:38:39.602631 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b31d6ac-276a-4947-a468-e99a784e1ff3-utilities\") pod \"redhat-marketplace-vz6rb\" (UID: \"2b31d6ac-276a-4947-a468-e99a784e1ff3\") " pod="openshift-marketplace/redhat-marketplace-vz6rb" Mar 13 16:38:39 crc kubenswrapper[4907]: I0313 16:38:39.704579 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9qbcl\" (UniqueName: \"kubernetes.io/projected/2b31d6ac-276a-4947-a468-e99a784e1ff3-kube-api-access-9qbcl\") pod \"redhat-marketplace-vz6rb\" (UID: \"2b31d6ac-276a-4947-a468-e99a784e1ff3\") " pod="openshift-marketplace/redhat-marketplace-vz6rb" Mar 13 16:38:39 crc kubenswrapper[4907]: I0313 16:38:39.704855 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b31d6ac-276a-4947-a468-e99a784e1ff3-catalog-content\") pod \"redhat-marketplace-vz6rb\" (UID: \"2b31d6ac-276a-4947-a468-e99a784e1ff3\") " pod="openshift-marketplace/redhat-marketplace-vz6rb" Mar 13 16:38:39 crc kubenswrapper[4907]: I0313 16:38:39.705019 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b31d6ac-276a-4947-a468-e99a784e1ff3-utilities\") pod \"redhat-marketplace-vz6rb\" (UID: \"2b31d6ac-276a-4947-a468-e99a784e1ff3\") " pod="openshift-marketplace/redhat-marketplace-vz6rb" Mar 13 16:38:39 crc kubenswrapper[4907]: I0313 16:38:39.705361 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b31d6ac-276a-4947-a468-e99a784e1ff3-utilities\") pod \"redhat-marketplace-vz6rb\" (UID: \"2b31d6ac-276a-4947-a468-e99a784e1ff3\") " pod="openshift-marketplace/redhat-marketplace-vz6rb" Mar 13 16:38:39 crc kubenswrapper[4907]: I0313 16:38:39.705372 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b31d6ac-276a-4947-a468-e99a784e1ff3-catalog-content\") pod \"redhat-marketplace-vz6rb\" (UID: \"2b31d6ac-276a-4947-a468-e99a784e1ff3\") " pod="openshift-marketplace/redhat-marketplace-vz6rb" Mar 13 16:38:39 crc kubenswrapper[4907]: I0313 16:38:39.723484 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9qbcl\" (UniqueName: \"kubernetes.io/projected/2b31d6ac-276a-4947-a468-e99a784e1ff3-kube-api-access-9qbcl\") pod \"redhat-marketplace-vz6rb\" (UID: \"2b31d6ac-276a-4947-a468-e99a784e1ff3\") " pod="openshift-marketplace/redhat-marketplace-vz6rb" Mar 13 16:38:39 crc kubenswrapper[4907]: I0313 16:38:39.867481 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vz6rb" Mar 13 16:38:42 crc kubenswrapper[4907]: I0313 16:38:42.107160 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-vz6rb"] Mar 13 16:38:42 crc kubenswrapper[4907]: W0313 16:38:42.139329 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2b31d6ac_276a_4947_a468_e99a784e1ff3.slice/crio-6aff65ef2fefe4fdcba78cafb69943816320eedcf9d24e99d5c0c1143b301c89 WatchSource:0}: Error finding container 6aff65ef2fefe4fdcba78cafb69943816320eedcf9d24e99d5c0c1143b301c89: Status 404 returned error can't find the container with id 6aff65ef2fefe4fdcba78cafb69943816320eedcf9d24e99d5c0c1143b301c89 Mar 13 16:38:42 crc kubenswrapper[4907]: I0313 16:38:42.537514 4907 generic.go:334] "Generic (PLEG): container finished" podID="2b31d6ac-276a-4947-a468-e99a784e1ff3" containerID="201fc9ca5006a0bcb74f754932e0f1754a8156f25796f4eac96197c357dad124" exitCode=0 Mar 13 16:38:42 crc kubenswrapper[4907]: I0313 16:38:42.537614 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vz6rb" event={"ID":"2b31d6ac-276a-4947-a468-e99a784e1ff3","Type":"ContainerDied","Data":"201fc9ca5006a0bcb74f754932e0f1754a8156f25796f4eac96197c357dad124"} Mar 13 16:38:42 crc kubenswrapper[4907]: I0313 16:38:42.537912 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vz6rb" event={"ID":"2b31d6ac-276a-4947-a468-e99a784e1ff3","Type":"ContainerStarted","Data":"6aff65ef2fefe4fdcba78cafb69943816320eedcf9d24e99d5c0c1143b301c89"} Mar 13 16:38:43 crc kubenswrapper[4907]: I0313 16:38:43.551301 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vz6rb" event={"ID":"2b31d6ac-276a-4947-a468-e99a784e1ff3","Type":"ContainerStarted","Data":"fd46e261e524d075378035d76f6c3b9852c9358bf16927c6ec958aaaf314bb95"} Mar 13 16:38:44 crc kubenswrapper[4907]: I0313 16:38:44.133986 4907 scope.go:117] "RemoveContainer" containerID="a1001aa55a6abdb3dd8474ef6137b7a29548724c2eff0d7fa57396148bbc916a" Mar 13 16:38:44 crc kubenswrapper[4907]: I0313 16:38:44.562473 4907 generic.go:334] "Generic (PLEG): container finished" podID="2b31d6ac-276a-4947-a468-e99a784e1ff3" containerID="fd46e261e524d075378035d76f6c3b9852c9358bf16927c6ec958aaaf314bb95" exitCode=0 Mar 13 16:38:44 crc kubenswrapper[4907]: I0313 16:38:44.562786 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vz6rb" event={"ID":"2b31d6ac-276a-4947-a468-e99a784e1ff3","Type":"ContainerDied","Data":"fd46e261e524d075378035d76f6c3b9852c9358bf16927c6ec958aaaf314bb95"} Mar 13 16:38:45 crc kubenswrapper[4907]: I0313 16:38:45.575601 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vz6rb" event={"ID":"2b31d6ac-276a-4947-a468-e99a784e1ff3","Type":"ContainerStarted","Data":"a54681c8803fa507d1b76e516386bee6ee700712c69fdb45137bd6c2f0e3e8cd"} Mar 13 16:38:45 crc kubenswrapper[4907]: I0313 16:38:45.604123 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-vz6rb" podStartSLOduration=3.988551902 podStartE2EDuration="6.604100368s" podCreationTimestamp="2026-03-13 16:38:39 +0000 UTC" firstStartedPulling="2026-03-13 16:38:42.539860052 +0000 UTC m=+9221.439647741" lastFinishedPulling="2026-03-13 16:38:45.155408518 +0000 UTC m=+9224.055196207" observedRunningTime="2026-03-13 16:38:45.597362618 +0000 UTC m=+9224.497150307" watchObservedRunningTime="2026-03-13 16:38:45.604100368 +0000 UTC m=+9224.503888047" Mar 13 16:38:48 crc kubenswrapper[4907]: I0313 16:38:48.041453 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 16:38:48 crc kubenswrapper[4907]: I0313 16:38:48.043078 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 16:38:48 crc kubenswrapper[4907]: I0313 16:38:48.043223 4907 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" Mar 13 16:38:48 crc kubenswrapper[4907]: I0313 16:38:48.044236 4907 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7e4d0690f4ac607ea8a268bfa88b5ae6d9a95f8b6013de7747fdffb429f28047"} pod="openshift-machine-config-operator/machine-config-daemon-hm56j" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 16:38:48 crc kubenswrapper[4907]: I0313 16:38:48.044413 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" containerID="cri-o://7e4d0690f4ac607ea8a268bfa88b5ae6d9a95f8b6013de7747fdffb429f28047" gracePeriod=600 Mar 13 16:38:48 crc kubenswrapper[4907]: E0313 16:38:48.179659 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:38:48 crc kubenswrapper[4907]: I0313 16:38:48.606463 4907 generic.go:334] "Generic (PLEG): container finished" podID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerID="7e4d0690f4ac607ea8a268bfa88b5ae6d9a95f8b6013de7747fdffb429f28047" exitCode=0 Mar 13 16:38:48 crc kubenswrapper[4907]: I0313 16:38:48.606515 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerDied","Data":"7e4d0690f4ac607ea8a268bfa88b5ae6d9a95f8b6013de7747fdffb429f28047"} Mar 13 16:38:48 crc kubenswrapper[4907]: I0313 16:38:48.606834 4907 scope.go:117] "RemoveContainer" containerID="f9156a2a4d423d000cb09f67e16772f675217bb2e483ba02adf57b642ffdffc9" Mar 13 16:38:48 crc kubenswrapper[4907]: I0313 16:38:48.607749 4907 scope.go:117] "RemoveContainer" containerID="7e4d0690f4ac607ea8a268bfa88b5ae6d9a95f8b6013de7747fdffb429f28047" Mar 13 16:38:48 crc kubenswrapper[4907]: E0313 16:38:48.608108 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:38:49 crc kubenswrapper[4907]: I0313 16:38:49.868031 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-vz6rb" Mar 13 16:38:49 crc kubenswrapper[4907]: I0313 16:38:49.868519 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-vz6rb" Mar 13 16:38:49 crc kubenswrapper[4907]: I0313 16:38:49.934542 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-vz6rb" Mar 13 16:38:50 crc kubenswrapper[4907]: I0313 16:38:50.677865 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-vz6rb" Mar 13 16:38:50 crc kubenswrapper[4907]: I0313 16:38:50.733274 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-vz6rb"] Mar 13 16:38:51 crc kubenswrapper[4907]: I0313 16:38:51.636413 4907 generic.go:334] "Generic (PLEG): container finished" podID="5e5575c9-0bdc-472b-9384-fbd1ce5cb480" containerID="b1d7df69f877a557d28aa9c5dd1537366e05d49104f113b20536b953f8c2ec90" exitCode=0 Mar 13 16:38:51 crc kubenswrapper[4907]: I0313 16:38:51.636492 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-xndl6/crc-debug-qt756" event={"ID":"5e5575c9-0bdc-472b-9384-fbd1ce5cb480","Type":"ContainerDied","Data":"b1d7df69f877a557d28aa9c5dd1537366e05d49104f113b20536b953f8c2ec90"} Mar 13 16:38:52 crc kubenswrapper[4907]: I0313 16:38:52.648181 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-vz6rb" podUID="2b31d6ac-276a-4947-a468-e99a784e1ff3" containerName="registry-server" containerID="cri-o://a54681c8803fa507d1b76e516386bee6ee700712c69fdb45137bd6c2f0e3e8cd" gracePeriod=2 Mar 13 16:38:52 crc kubenswrapper[4907]: I0313 16:38:52.944337 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xndl6/crc-debug-qt756" Mar 13 16:38:52 crc kubenswrapper[4907]: I0313 16:38:52.995152 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-xndl6/crc-debug-qt756"] Mar 13 16:38:53 crc kubenswrapper[4907]: I0313 16:38:53.004550 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-xndl6/crc-debug-qt756"] Mar 13 16:38:53 crc kubenswrapper[4907]: I0313 16:38:53.017296 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4vjlv\" (UniqueName: \"kubernetes.io/projected/5e5575c9-0bdc-472b-9384-fbd1ce5cb480-kube-api-access-4vjlv\") pod \"5e5575c9-0bdc-472b-9384-fbd1ce5cb480\" (UID: \"5e5575c9-0bdc-472b-9384-fbd1ce5cb480\") " Mar 13 16:38:53 crc kubenswrapper[4907]: I0313 16:38:53.017358 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5e5575c9-0bdc-472b-9384-fbd1ce5cb480-host\") pod \"5e5575c9-0bdc-472b-9384-fbd1ce5cb480\" (UID: \"5e5575c9-0bdc-472b-9384-fbd1ce5cb480\") " Mar 13 16:38:53 crc kubenswrapper[4907]: I0313 16:38:53.017633 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5e5575c9-0bdc-472b-9384-fbd1ce5cb480-host" (OuterVolumeSpecName: "host") pod "5e5575c9-0bdc-472b-9384-fbd1ce5cb480" (UID: "5e5575c9-0bdc-472b-9384-fbd1ce5cb480"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 16:38:53 crc kubenswrapper[4907]: I0313 16:38:53.017833 4907 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5e5575c9-0bdc-472b-9384-fbd1ce5cb480-host\") on node \"crc\" DevicePath \"\"" Mar 13 16:38:53 crc kubenswrapper[4907]: I0313 16:38:53.032134 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e5575c9-0bdc-472b-9384-fbd1ce5cb480-kube-api-access-4vjlv" (OuterVolumeSpecName: "kube-api-access-4vjlv") pod "5e5575c9-0bdc-472b-9384-fbd1ce5cb480" (UID: "5e5575c9-0bdc-472b-9384-fbd1ce5cb480"). InnerVolumeSpecName "kube-api-access-4vjlv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:38:53 crc kubenswrapper[4907]: I0313 16:38:53.120568 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4vjlv\" (UniqueName: \"kubernetes.io/projected/5e5575c9-0bdc-472b-9384-fbd1ce5cb480-kube-api-access-4vjlv\") on node \"crc\" DevicePath \"\"" Mar 13 16:38:53 crc kubenswrapper[4907]: I0313 16:38:53.227733 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vz6rb" Mar 13 16:38:53 crc kubenswrapper[4907]: I0313 16:38:53.324249 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b31d6ac-276a-4947-a468-e99a784e1ff3-catalog-content\") pod \"2b31d6ac-276a-4947-a468-e99a784e1ff3\" (UID: \"2b31d6ac-276a-4947-a468-e99a784e1ff3\") " Mar 13 16:38:53 crc kubenswrapper[4907]: I0313 16:38:53.324564 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9qbcl\" (UniqueName: \"kubernetes.io/projected/2b31d6ac-276a-4947-a468-e99a784e1ff3-kube-api-access-9qbcl\") pod \"2b31d6ac-276a-4947-a468-e99a784e1ff3\" (UID: \"2b31d6ac-276a-4947-a468-e99a784e1ff3\") " Mar 13 16:38:53 crc kubenswrapper[4907]: I0313 16:38:53.324599 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b31d6ac-276a-4947-a468-e99a784e1ff3-utilities\") pod \"2b31d6ac-276a-4947-a468-e99a784e1ff3\" (UID: \"2b31d6ac-276a-4947-a468-e99a784e1ff3\") " Mar 13 16:38:53 crc kubenswrapper[4907]: I0313 16:38:53.325393 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2b31d6ac-276a-4947-a468-e99a784e1ff3-utilities" (OuterVolumeSpecName: "utilities") pod "2b31d6ac-276a-4947-a468-e99a784e1ff3" (UID: "2b31d6ac-276a-4947-a468-e99a784e1ff3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 16:38:53 crc kubenswrapper[4907]: I0313 16:38:53.329011 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2b31d6ac-276a-4947-a468-e99a784e1ff3-kube-api-access-9qbcl" (OuterVolumeSpecName: "kube-api-access-9qbcl") pod "2b31d6ac-276a-4947-a468-e99a784e1ff3" (UID: "2b31d6ac-276a-4947-a468-e99a784e1ff3"). InnerVolumeSpecName "kube-api-access-9qbcl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:38:53 crc kubenswrapper[4907]: I0313 16:38:53.362420 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2b31d6ac-276a-4947-a468-e99a784e1ff3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2b31d6ac-276a-4947-a468-e99a784e1ff3" (UID: "2b31d6ac-276a-4947-a468-e99a784e1ff3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 16:38:53 crc kubenswrapper[4907]: I0313 16:38:53.426960 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b31d6ac-276a-4947-a468-e99a784e1ff3-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 16:38:53 crc kubenswrapper[4907]: I0313 16:38:53.427011 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9qbcl\" (UniqueName: \"kubernetes.io/projected/2b31d6ac-276a-4947-a468-e99a784e1ff3-kube-api-access-9qbcl\") on node \"crc\" DevicePath \"\"" Mar 13 16:38:53 crc kubenswrapper[4907]: I0313 16:38:53.427027 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b31d6ac-276a-4947-a468-e99a784e1ff3-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 16:38:53 crc kubenswrapper[4907]: I0313 16:38:53.657513 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f3e01960410a5bad3f4d4d1123a16e5c7cc67a5d538faaf418827fff78354c63" Mar 13 16:38:53 crc kubenswrapper[4907]: I0313 16:38:53.657576 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xndl6/crc-debug-qt756" Mar 13 16:38:53 crc kubenswrapper[4907]: I0313 16:38:53.662345 4907 generic.go:334] "Generic (PLEG): container finished" podID="2b31d6ac-276a-4947-a468-e99a784e1ff3" containerID="a54681c8803fa507d1b76e516386bee6ee700712c69fdb45137bd6c2f0e3e8cd" exitCode=0 Mar 13 16:38:53 crc kubenswrapper[4907]: I0313 16:38:53.662391 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vz6rb" event={"ID":"2b31d6ac-276a-4947-a468-e99a784e1ff3","Type":"ContainerDied","Data":"a54681c8803fa507d1b76e516386bee6ee700712c69fdb45137bd6c2f0e3e8cd"} Mar 13 16:38:53 crc kubenswrapper[4907]: I0313 16:38:53.662421 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vz6rb" event={"ID":"2b31d6ac-276a-4947-a468-e99a784e1ff3","Type":"ContainerDied","Data":"6aff65ef2fefe4fdcba78cafb69943816320eedcf9d24e99d5c0c1143b301c89"} Mar 13 16:38:53 crc kubenswrapper[4907]: I0313 16:38:53.662438 4907 scope.go:117] "RemoveContainer" containerID="a54681c8803fa507d1b76e516386bee6ee700712c69fdb45137bd6c2f0e3e8cd" Mar 13 16:38:53 crc kubenswrapper[4907]: I0313 16:38:53.662572 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vz6rb" Mar 13 16:38:53 crc kubenswrapper[4907]: I0313 16:38:53.720675 4907 scope.go:117] "RemoveContainer" containerID="fd46e261e524d075378035d76f6c3b9852c9358bf16927c6ec958aaaf314bb95" Mar 13 16:38:53 crc kubenswrapper[4907]: I0313 16:38:53.754498 4907 scope.go:117] "RemoveContainer" containerID="201fc9ca5006a0bcb74f754932e0f1754a8156f25796f4eac96197c357dad124" Mar 13 16:38:53 crc kubenswrapper[4907]: I0313 16:38:53.769659 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-vz6rb"] Mar 13 16:38:53 crc kubenswrapper[4907]: I0313 16:38:53.779775 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-vz6rb"] Mar 13 16:38:53 crc kubenswrapper[4907]: I0313 16:38:53.795388 4907 scope.go:117] "RemoveContainer" containerID="a54681c8803fa507d1b76e516386bee6ee700712c69fdb45137bd6c2f0e3e8cd" Mar 13 16:38:53 crc kubenswrapper[4907]: E0313 16:38:53.802230 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a54681c8803fa507d1b76e516386bee6ee700712c69fdb45137bd6c2f0e3e8cd\": container with ID starting with a54681c8803fa507d1b76e516386bee6ee700712c69fdb45137bd6c2f0e3e8cd not found: ID does not exist" containerID="a54681c8803fa507d1b76e516386bee6ee700712c69fdb45137bd6c2f0e3e8cd" Mar 13 16:38:53 crc kubenswrapper[4907]: I0313 16:38:53.802267 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a54681c8803fa507d1b76e516386bee6ee700712c69fdb45137bd6c2f0e3e8cd"} err="failed to get container status \"a54681c8803fa507d1b76e516386bee6ee700712c69fdb45137bd6c2f0e3e8cd\": rpc error: code = NotFound desc = could not find container \"a54681c8803fa507d1b76e516386bee6ee700712c69fdb45137bd6c2f0e3e8cd\": container with ID starting with a54681c8803fa507d1b76e516386bee6ee700712c69fdb45137bd6c2f0e3e8cd not found: ID does not exist" Mar 13 16:38:53 crc kubenswrapper[4907]: I0313 16:38:53.802345 4907 scope.go:117] "RemoveContainer" containerID="fd46e261e524d075378035d76f6c3b9852c9358bf16927c6ec958aaaf314bb95" Mar 13 16:38:53 crc kubenswrapper[4907]: E0313 16:38:53.803000 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fd46e261e524d075378035d76f6c3b9852c9358bf16927c6ec958aaaf314bb95\": container with ID starting with fd46e261e524d075378035d76f6c3b9852c9358bf16927c6ec958aaaf314bb95 not found: ID does not exist" containerID="fd46e261e524d075378035d76f6c3b9852c9358bf16927c6ec958aaaf314bb95" Mar 13 16:38:53 crc kubenswrapper[4907]: I0313 16:38:53.803041 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fd46e261e524d075378035d76f6c3b9852c9358bf16927c6ec958aaaf314bb95"} err="failed to get container status \"fd46e261e524d075378035d76f6c3b9852c9358bf16927c6ec958aaaf314bb95\": rpc error: code = NotFound desc = could not find container \"fd46e261e524d075378035d76f6c3b9852c9358bf16927c6ec958aaaf314bb95\": container with ID starting with fd46e261e524d075378035d76f6c3b9852c9358bf16927c6ec958aaaf314bb95 not found: ID does not exist" Mar 13 16:38:53 crc kubenswrapper[4907]: I0313 16:38:53.803069 4907 scope.go:117] "RemoveContainer" containerID="201fc9ca5006a0bcb74f754932e0f1754a8156f25796f4eac96197c357dad124" Mar 13 16:38:53 crc kubenswrapper[4907]: E0313 16:38:53.803554 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"201fc9ca5006a0bcb74f754932e0f1754a8156f25796f4eac96197c357dad124\": container with ID starting with 201fc9ca5006a0bcb74f754932e0f1754a8156f25796f4eac96197c357dad124 not found: ID does not exist" containerID="201fc9ca5006a0bcb74f754932e0f1754a8156f25796f4eac96197c357dad124" Mar 13 16:38:53 crc kubenswrapper[4907]: I0313 16:38:53.803584 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"201fc9ca5006a0bcb74f754932e0f1754a8156f25796f4eac96197c357dad124"} err="failed to get container status \"201fc9ca5006a0bcb74f754932e0f1754a8156f25796f4eac96197c357dad124\": rpc error: code = NotFound desc = could not find container \"201fc9ca5006a0bcb74f754932e0f1754a8156f25796f4eac96197c357dad124\": container with ID starting with 201fc9ca5006a0bcb74f754932e0f1754a8156f25796f4eac96197c357dad124 not found: ID does not exist" Mar 13 16:38:53 crc kubenswrapper[4907]: I0313 16:38:53.811695 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2b31d6ac-276a-4947-a468-e99a784e1ff3" path="/var/lib/kubelet/pods/2b31d6ac-276a-4947-a468-e99a784e1ff3/volumes" Mar 13 16:38:53 crc kubenswrapper[4907]: I0313 16:38:53.812439 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5e5575c9-0bdc-472b-9384-fbd1ce5cb480" path="/var/lib/kubelet/pods/5e5575c9-0bdc-472b-9384-fbd1ce5cb480/volumes" Mar 13 16:38:54 crc kubenswrapper[4907]: I0313 16:38:54.215950 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-xndl6/crc-debug-gr5ml"] Mar 13 16:38:54 crc kubenswrapper[4907]: E0313 16:38:54.216523 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b31d6ac-276a-4947-a468-e99a784e1ff3" containerName="registry-server" Mar 13 16:38:54 crc kubenswrapper[4907]: I0313 16:38:54.216540 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b31d6ac-276a-4947-a468-e99a784e1ff3" containerName="registry-server" Mar 13 16:38:54 crc kubenswrapper[4907]: E0313 16:38:54.216561 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b31d6ac-276a-4947-a468-e99a784e1ff3" containerName="extract-content" Mar 13 16:38:54 crc kubenswrapper[4907]: I0313 16:38:54.216567 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b31d6ac-276a-4947-a468-e99a784e1ff3" containerName="extract-content" Mar 13 16:38:54 crc kubenswrapper[4907]: E0313 16:38:54.216593 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b31d6ac-276a-4947-a468-e99a784e1ff3" containerName="extract-utilities" Mar 13 16:38:54 crc kubenswrapper[4907]: I0313 16:38:54.216599 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b31d6ac-276a-4947-a468-e99a784e1ff3" containerName="extract-utilities" Mar 13 16:38:54 crc kubenswrapper[4907]: E0313 16:38:54.216618 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e5575c9-0bdc-472b-9384-fbd1ce5cb480" containerName="container-00" Mar 13 16:38:54 crc kubenswrapper[4907]: I0313 16:38:54.216624 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e5575c9-0bdc-472b-9384-fbd1ce5cb480" containerName="container-00" Mar 13 16:38:54 crc kubenswrapper[4907]: I0313 16:38:54.216800 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e5575c9-0bdc-472b-9384-fbd1ce5cb480" containerName="container-00" Mar 13 16:38:54 crc kubenswrapper[4907]: I0313 16:38:54.216820 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="2b31d6ac-276a-4947-a468-e99a784e1ff3" containerName="registry-server" Mar 13 16:38:54 crc kubenswrapper[4907]: I0313 16:38:54.217516 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xndl6/crc-debug-gr5ml" Mar 13 16:38:54 crc kubenswrapper[4907]: I0313 16:38:54.373821 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v5sf2\" (UniqueName: \"kubernetes.io/projected/09517507-85bc-48d4-9fe2-f4a9e862322a-kube-api-access-v5sf2\") pod \"crc-debug-gr5ml\" (UID: \"09517507-85bc-48d4-9fe2-f4a9e862322a\") " pod="openshift-must-gather-xndl6/crc-debug-gr5ml" Mar 13 16:38:54 crc kubenswrapper[4907]: I0313 16:38:54.373925 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/09517507-85bc-48d4-9fe2-f4a9e862322a-host\") pod \"crc-debug-gr5ml\" (UID: \"09517507-85bc-48d4-9fe2-f4a9e862322a\") " pod="openshift-must-gather-xndl6/crc-debug-gr5ml" Mar 13 16:38:54 crc kubenswrapper[4907]: I0313 16:38:54.476402 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v5sf2\" (UniqueName: \"kubernetes.io/projected/09517507-85bc-48d4-9fe2-f4a9e862322a-kube-api-access-v5sf2\") pod \"crc-debug-gr5ml\" (UID: \"09517507-85bc-48d4-9fe2-f4a9e862322a\") " pod="openshift-must-gather-xndl6/crc-debug-gr5ml" Mar 13 16:38:54 crc kubenswrapper[4907]: I0313 16:38:54.476503 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/09517507-85bc-48d4-9fe2-f4a9e862322a-host\") pod \"crc-debug-gr5ml\" (UID: \"09517507-85bc-48d4-9fe2-f4a9e862322a\") " pod="openshift-must-gather-xndl6/crc-debug-gr5ml" Mar 13 16:38:54 crc kubenswrapper[4907]: I0313 16:38:54.476659 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/09517507-85bc-48d4-9fe2-f4a9e862322a-host\") pod \"crc-debug-gr5ml\" (UID: \"09517507-85bc-48d4-9fe2-f4a9e862322a\") " pod="openshift-must-gather-xndl6/crc-debug-gr5ml" Mar 13 16:38:54 crc kubenswrapper[4907]: I0313 16:38:54.502594 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v5sf2\" (UniqueName: \"kubernetes.io/projected/09517507-85bc-48d4-9fe2-f4a9e862322a-kube-api-access-v5sf2\") pod \"crc-debug-gr5ml\" (UID: \"09517507-85bc-48d4-9fe2-f4a9e862322a\") " pod="openshift-must-gather-xndl6/crc-debug-gr5ml" Mar 13 16:38:54 crc kubenswrapper[4907]: I0313 16:38:54.536106 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xndl6/crc-debug-gr5ml" Mar 13 16:38:54 crc kubenswrapper[4907]: I0313 16:38:54.673382 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-xndl6/crc-debug-gr5ml" event={"ID":"09517507-85bc-48d4-9fe2-f4a9e862322a","Type":"ContainerStarted","Data":"d004f3ada10244b8dd65e6d6ddad786d1dfdd84dd59f991b7610b3493b6f4b3e"} Mar 13 16:38:55 crc kubenswrapper[4907]: I0313 16:38:55.687824 4907 generic.go:334] "Generic (PLEG): container finished" podID="09517507-85bc-48d4-9fe2-f4a9e862322a" containerID="97e8dbb17dfb954f08537b16b9fb37c2f747c102c1339b2be558bedc609e5946" exitCode=1 Mar 13 16:38:55 crc kubenswrapper[4907]: I0313 16:38:55.688137 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-xndl6/crc-debug-gr5ml" event={"ID":"09517507-85bc-48d4-9fe2-f4a9e862322a","Type":"ContainerDied","Data":"97e8dbb17dfb954f08537b16b9fb37c2f747c102c1339b2be558bedc609e5946"} Mar 13 16:38:55 crc kubenswrapper[4907]: I0313 16:38:55.732595 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-xndl6/crc-debug-gr5ml"] Mar 13 16:38:55 crc kubenswrapper[4907]: I0313 16:38:55.741743 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-xndl6/crc-debug-gr5ml"] Mar 13 16:38:57 crc kubenswrapper[4907]: I0313 16:38:57.002694 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xndl6/crc-debug-gr5ml" Mar 13 16:38:57 crc kubenswrapper[4907]: I0313 16:38:57.029668 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v5sf2\" (UniqueName: \"kubernetes.io/projected/09517507-85bc-48d4-9fe2-f4a9e862322a-kube-api-access-v5sf2\") pod \"09517507-85bc-48d4-9fe2-f4a9e862322a\" (UID: \"09517507-85bc-48d4-9fe2-f4a9e862322a\") " Mar 13 16:38:57 crc kubenswrapper[4907]: I0313 16:38:57.029756 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/09517507-85bc-48d4-9fe2-f4a9e862322a-host\") pod \"09517507-85bc-48d4-9fe2-f4a9e862322a\" (UID: \"09517507-85bc-48d4-9fe2-f4a9e862322a\") " Mar 13 16:38:57 crc kubenswrapper[4907]: I0313 16:38:57.029917 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/09517507-85bc-48d4-9fe2-f4a9e862322a-host" (OuterVolumeSpecName: "host") pod "09517507-85bc-48d4-9fe2-f4a9e862322a" (UID: "09517507-85bc-48d4-9fe2-f4a9e862322a"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 16:38:57 crc kubenswrapper[4907]: I0313 16:38:57.030279 4907 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/09517507-85bc-48d4-9fe2-f4a9e862322a-host\") on node \"crc\" DevicePath \"\"" Mar 13 16:38:57 crc kubenswrapper[4907]: I0313 16:38:57.035160 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09517507-85bc-48d4-9fe2-f4a9e862322a-kube-api-access-v5sf2" (OuterVolumeSpecName: "kube-api-access-v5sf2") pod "09517507-85bc-48d4-9fe2-f4a9e862322a" (UID: "09517507-85bc-48d4-9fe2-f4a9e862322a"). InnerVolumeSpecName "kube-api-access-v5sf2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:38:57 crc kubenswrapper[4907]: I0313 16:38:57.133343 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v5sf2\" (UniqueName: \"kubernetes.io/projected/09517507-85bc-48d4-9fe2-f4a9e862322a-kube-api-access-v5sf2\") on node \"crc\" DevicePath \"\"" Mar 13 16:38:57 crc kubenswrapper[4907]: I0313 16:38:57.711655 4907 scope.go:117] "RemoveContainer" containerID="97e8dbb17dfb954f08537b16b9fb37c2f747c102c1339b2be558bedc609e5946" Mar 13 16:38:57 crc kubenswrapper[4907]: I0313 16:38:57.711660 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xndl6/crc-debug-gr5ml" Mar 13 16:38:57 crc kubenswrapper[4907]: I0313 16:38:57.793961 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09517507-85bc-48d4-9fe2-f4a9e862322a" path="/var/lib/kubelet/pods/09517507-85bc-48d4-9fe2-f4a9e862322a/volumes" Mar 13 16:38:59 crc kubenswrapper[4907]: I0313 16:38:59.785913 4907 scope.go:117] "RemoveContainer" containerID="7e4d0690f4ac607ea8a268bfa88b5ae6d9a95f8b6013de7747fdffb429f28047" Mar 13 16:38:59 crc kubenswrapper[4907]: E0313 16:38:59.786799 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:39:11 crc kubenswrapper[4907]: I0313 16:39:11.790052 4907 scope.go:117] "RemoveContainer" containerID="7e4d0690f4ac607ea8a268bfa88b5ae6d9a95f8b6013de7747fdffb429f28047" Mar 13 16:39:11 crc kubenswrapper[4907]: E0313 16:39:11.790942 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:39:22 crc kubenswrapper[4907]: I0313 16:39:22.782553 4907 scope.go:117] "RemoveContainer" containerID="7e4d0690f4ac607ea8a268bfa88b5ae6d9a95f8b6013de7747fdffb429f28047" Mar 13 16:39:22 crc kubenswrapper[4907]: E0313 16:39:22.783963 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:39:37 crc kubenswrapper[4907]: I0313 16:39:37.783189 4907 scope.go:117] "RemoveContainer" containerID="7e4d0690f4ac607ea8a268bfa88b5ae6d9a95f8b6013de7747fdffb429f28047" Mar 13 16:39:37 crc kubenswrapper[4907]: E0313 16:39:37.784137 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:39:52 crc kubenswrapper[4907]: I0313 16:39:52.783620 4907 scope.go:117] "RemoveContainer" containerID="7e4d0690f4ac607ea8a268bfa88b5ae6d9a95f8b6013de7747fdffb429f28047" Mar 13 16:39:52 crc kubenswrapper[4907]: E0313 16:39:52.784987 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:40:00 crc kubenswrapper[4907]: I0313 16:40:00.155043 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29557000-b8z24"] Mar 13 16:40:00 crc kubenswrapper[4907]: E0313 16:40:00.161760 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09517507-85bc-48d4-9fe2-f4a9e862322a" containerName="container-00" Mar 13 16:40:00 crc kubenswrapper[4907]: I0313 16:40:00.161798 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="09517507-85bc-48d4-9fe2-f4a9e862322a" containerName="container-00" Mar 13 16:40:00 crc kubenswrapper[4907]: I0313 16:40:00.164644 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="09517507-85bc-48d4-9fe2-f4a9e862322a" containerName="container-00" Mar 13 16:40:00 crc kubenswrapper[4907]: I0313 16:40:00.166059 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29557000-b8z24" Mar 13 16:40:00 crc kubenswrapper[4907]: I0313 16:40:00.170410 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 16:40:00 crc kubenswrapper[4907]: I0313 16:40:00.170770 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 16:40:00 crc kubenswrapper[4907]: I0313 16:40:00.181930 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 16:40:00 crc kubenswrapper[4907]: I0313 16:40:00.235375 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29557000-b8z24"] Mar 13 16:40:00 crc kubenswrapper[4907]: I0313 16:40:00.375275 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9cb9q\" (UniqueName: \"kubernetes.io/projected/ff06fab3-a77c-49ef-8c13-88908b7acf21-kube-api-access-9cb9q\") pod \"auto-csr-approver-29557000-b8z24\" (UID: \"ff06fab3-a77c-49ef-8c13-88908b7acf21\") " pod="openshift-infra/auto-csr-approver-29557000-b8z24" Mar 13 16:40:00 crc kubenswrapper[4907]: I0313 16:40:00.477614 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9cb9q\" (UniqueName: \"kubernetes.io/projected/ff06fab3-a77c-49ef-8c13-88908b7acf21-kube-api-access-9cb9q\") pod \"auto-csr-approver-29557000-b8z24\" (UID: \"ff06fab3-a77c-49ef-8c13-88908b7acf21\") " pod="openshift-infra/auto-csr-approver-29557000-b8z24" Mar 13 16:40:00 crc kubenswrapper[4907]: I0313 16:40:00.499398 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9cb9q\" (UniqueName: \"kubernetes.io/projected/ff06fab3-a77c-49ef-8c13-88908b7acf21-kube-api-access-9cb9q\") pod \"auto-csr-approver-29557000-b8z24\" (UID: \"ff06fab3-a77c-49ef-8c13-88908b7acf21\") " pod="openshift-infra/auto-csr-approver-29557000-b8z24" Mar 13 16:40:00 crc kubenswrapper[4907]: I0313 16:40:00.511189 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29557000-b8z24" Mar 13 16:40:01 crc kubenswrapper[4907]: I0313 16:40:01.029157 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29557000-b8z24"] Mar 13 16:40:01 crc kubenswrapper[4907]: I0313 16:40:01.043482 4907 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 13 16:40:01 crc kubenswrapper[4907]: I0313 16:40:01.625216 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29557000-b8z24" event={"ID":"ff06fab3-a77c-49ef-8c13-88908b7acf21","Type":"ContainerStarted","Data":"3975471bc6fb2efe82da83ec7e799b1b9e617df03d5750fb3e8165d894b63258"} Mar 13 16:40:02 crc kubenswrapper[4907]: I0313 16:40:02.638523 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29557000-b8z24" event={"ID":"ff06fab3-a77c-49ef-8c13-88908b7acf21","Type":"ContainerStarted","Data":"7edb971c3f915cfb8576dc1860f2e0063b3420e5f584dbd096705d5ce436c71e"} Mar 13 16:40:02 crc kubenswrapper[4907]: I0313 16:40:02.665772 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29557000-b8z24" podStartSLOduration=1.5020899349999999 podStartE2EDuration="2.665746056s" podCreationTimestamp="2026-03-13 16:40:00 +0000 UTC" firstStartedPulling="2026-03-13 16:40:01.043089806 +0000 UTC m=+9299.942877525" lastFinishedPulling="2026-03-13 16:40:02.206745957 +0000 UTC m=+9301.106533646" observedRunningTime="2026-03-13 16:40:02.654380796 +0000 UTC m=+9301.554168525" watchObservedRunningTime="2026-03-13 16:40:02.665746056 +0000 UTC m=+9301.565533785" Mar 13 16:40:03 crc kubenswrapper[4907]: I0313 16:40:03.650590 4907 generic.go:334] "Generic (PLEG): container finished" podID="ff06fab3-a77c-49ef-8c13-88908b7acf21" containerID="7edb971c3f915cfb8576dc1860f2e0063b3420e5f584dbd096705d5ce436c71e" exitCode=0 Mar 13 16:40:03 crc kubenswrapper[4907]: I0313 16:40:03.650637 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29557000-b8z24" event={"ID":"ff06fab3-a77c-49ef-8c13-88908b7acf21","Type":"ContainerDied","Data":"7edb971c3f915cfb8576dc1860f2e0063b3420e5f584dbd096705d5ce436c71e"} Mar 13 16:40:03 crc kubenswrapper[4907]: I0313 16:40:03.782964 4907 scope.go:117] "RemoveContainer" containerID="7e4d0690f4ac607ea8a268bfa88b5ae6d9a95f8b6013de7747fdffb429f28047" Mar 13 16:40:03 crc kubenswrapper[4907]: E0313 16:40:03.783388 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:40:05 crc kubenswrapper[4907]: I0313 16:40:05.054271 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29557000-b8z24" Mar 13 16:40:05 crc kubenswrapper[4907]: I0313 16:40:05.201020 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9cb9q\" (UniqueName: \"kubernetes.io/projected/ff06fab3-a77c-49ef-8c13-88908b7acf21-kube-api-access-9cb9q\") pod \"ff06fab3-a77c-49ef-8c13-88908b7acf21\" (UID: \"ff06fab3-a77c-49ef-8c13-88908b7acf21\") " Mar 13 16:40:05 crc kubenswrapper[4907]: I0313 16:40:05.206999 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ff06fab3-a77c-49ef-8c13-88908b7acf21-kube-api-access-9cb9q" (OuterVolumeSpecName: "kube-api-access-9cb9q") pod "ff06fab3-a77c-49ef-8c13-88908b7acf21" (UID: "ff06fab3-a77c-49ef-8c13-88908b7acf21"). InnerVolumeSpecName "kube-api-access-9cb9q". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:40:05 crc kubenswrapper[4907]: I0313 16:40:05.304095 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9cb9q\" (UniqueName: \"kubernetes.io/projected/ff06fab3-a77c-49ef-8c13-88908b7acf21-kube-api-access-9cb9q\") on node \"crc\" DevicePath \"\"" Mar 13 16:40:05 crc kubenswrapper[4907]: I0313 16:40:05.672861 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29557000-b8z24" event={"ID":"ff06fab3-a77c-49ef-8c13-88908b7acf21","Type":"ContainerDied","Data":"3975471bc6fb2efe82da83ec7e799b1b9e617df03d5750fb3e8165d894b63258"} Mar 13 16:40:05 crc kubenswrapper[4907]: I0313 16:40:05.673342 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3975471bc6fb2efe82da83ec7e799b1b9e617df03d5750fb3e8165d894b63258" Mar 13 16:40:05 crc kubenswrapper[4907]: I0313 16:40:05.672951 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29557000-b8z24" Mar 13 16:40:06 crc kubenswrapper[4907]: I0313 16:40:06.129233 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556994-452h2"] Mar 13 16:40:06 crc kubenswrapper[4907]: I0313 16:40:06.138653 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556994-452h2"] Mar 13 16:40:07 crc kubenswrapper[4907]: I0313 16:40:07.807609 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="262afeb8-16de-47e7-a64b-5b4e938aa1f4" path="/var/lib/kubelet/pods/262afeb8-16de-47e7-a64b-5b4e938aa1f4/volumes" Mar 13 16:40:15 crc kubenswrapper[4907]: I0313 16:40:15.782623 4907 scope.go:117] "RemoveContainer" containerID="7e4d0690f4ac607ea8a268bfa88b5ae6d9a95f8b6013de7747fdffb429f28047" Mar 13 16:40:15 crc kubenswrapper[4907]: E0313 16:40:15.783397 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:40:30 crc kubenswrapper[4907]: I0313 16:40:30.783304 4907 scope.go:117] "RemoveContainer" containerID="7e4d0690f4ac607ea8a268bfa88b5ae6d9a95f8b6013de7747fdffb429f28047" Mar 13 16:40:30 crc kubenswrapper[4907]: E0313 16:40:30.784238 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:40:42 crc kubenswrapper[4907]: I0313 16:40:42.782911 4907 scope.go:117] "RemoveContainer" containerID="7e4d0690f4ac607ea8a268bfa88b5ae6d9a95f8b6013de7747fdffb429f28047" Mar 13 16:40:42 crc kubenswrapper[4907]: E0313 16:40:42.783752 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:40:44 crc kubenswrapper[4907]: I0313 16:40:44.289348 4907 scope.go:117] "RemoveContainer" containerID="237bb05fcc8f490c8ced9263406f8715e053a689a6c929dacb3684e153ea9c27" Mar 13 16:40:55 crc kubenswrapper[4907]: I0313 16:40:55.782829 4907 scope.go:117] "RemoveContainer" containerID="7e4d0690f4ac607ea8a268bfa88b5ae6d9a95f8b6013de7747fdffb429f28047" Mar 13 16:40:55 crc kubenswrapper[4907]: E0313 16:40:55.783739 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:41:09 crc kubenswrapper[4907]: I0313 16:41:09.783391 4907 scope.go:117] "RemoveContainer" containerID="7e4d0690f4ac607ea8a268bfa88b5ae6d9a95f8b6013de7747fdffb429f28047" Mar 13 16:41:09 crc kubenswrapper[4907]: E0313 16:41:09.784394 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:41:20 crc kubenswrapper[4907]: I0313 16:41:20.783652 4907 scope.go:117] "RemoveContainer" containerID="7e4d0690f4ac607ea8a268bfa88b5ae6d9a95f8b6013de7747fdffb429f28047" Mar 13 16:41:20 crc kubenswrapper[4907]: E0313 16:41:20.784950 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:41:34 crc kubenswrapper[4907]: I0313 16:41:34.791798 4907 scope.go:117] "RemoveContainer" containerID="7e4d0690f4ac607ea8a268bfa88b5ae6d9a95f8b6013de7747fdffb429f28047" Mar 13 16:41:34 crc kubenswrapper[4907]: E0313 16:41:34.793390 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:41:47 crc kubenswrapper[4907]: I0313 16:41:47.782781 4907 scope.go:117] "RemoveContainer" containerID="7e4d0690f4ac607ea8a268bfa88b5ae6d9a95f8b6013de7747fdffb429f28047" Mar 13 16:41:47 crc kubenswrapper[4907]: E0313 16:41:47.783542 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:42:00 crc kubenswrapper[4907]: I0313 16:42:00.156333 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29557002-tdngs"] Mar 13 16:42:00 crc kubenswrapper[4907]: E0313 16:42:00.157648 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff06fab3-a77c-49ef-8c13-88908b7acf21" containerName="oc" Mar 13 16:42:00 crc kubenswrapper[4907]: I0313 16:42:00.157665 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff06fab3-a77c-49ef-8c13-88908b7acf21" containerName="oc" Mar 13 16:42:00 crc kubenswrapper[4907]: I0313 16:42:00.157979 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff06fab3-a77c-49ef-8c13-88908b7acf21" containerName="oc" Mar 13 16:42:00 crc kubenswrapper[4907]: I0313 16:42:00.158961 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29557002-tdngs" Mar 13 16:42:00 crc kubenswrapper[4907]: I0313 16:42:00.172272 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29557002-tdngs"] Mar 13 16:42:00 crc kubenswrapper[4907]: I0313 16:42:00.192348 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 16:42:00 crc kubenswrapper[4907]: I0313 16:42:00.192483 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 16:42:00 crc kubenswrapper[4907]: I0313 16:42:00.192556 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 16:42:00 crc kubenswrapper[4907]: I0313 16:42:00.225358 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fbw8l\" (UniqueName: \"kubernetes.io/projected/dce83493-bbc5-49ac-a420-229eac93005d-kube-api-access-fbw8l\") pod \"auto-csr-approver-29557002-tdngs\" (UID: \"dce83493-bbc5-49ac-a420-229eac93005d\") " pod="openshift-infra/auto-csr-approver-29557002-tdngs" Mar 13 16:42:00 crc kubenswrapper[4907]: I0313 16:42:00.328204 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fbw8l\" (UniqueName: \"kubernetes.io/projected/dce83493-bbc5-49ac-a420-229eac93005d-kube-api-access-fbw8l\") pod \"auto-csr-approver-29557002-tdngs\" (UID: \"dce83493-bbc5-49ac-a420-229eac93005d\") " pod="openshift-infra/auto-csr-approver-29557002-tdngs" Mar 13 16:42:00 crc kubenswrapper[4907]: I0313 16:42:00.355156 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fbw8l\" (UniqueName: \"kubernetes.io/projected/dce83493-bbc5-49ac-a420-229eac93005d-kube-api-access-fbw8l\") pod \"auto-csr-approver-29557002-tdngs\" (UID: \"dce83493-bbc5-49ac-a420-229eac93005d\") " pod="openshift-infra/auto-csr-approver-29557002-tdngs" Mar 13 16:42:00 crc kubenswrapper[4907]: I0313 16:42:00.510376 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29557002-tdngs" Mar 13 16:42:01 crc kubenswrapper[4907]: I0313 16:42:01.684669 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29557002-tdngs"] Mar 13 16:42:01 crc kubenswrapper[4907]: I0313 16:42:01.790277 4907 scope.go:117] "RemoveContainer" containerID="7e4d0690f4ac607ea8a268bfa88b5ae6d9a95f8b6013de7747fdffb429f28047" Mar 13 16:42:01 crc kubenswrapper[4907]: E0313 16:42:01.790520 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:42:01 crc kubenswrapper[4907]: I0313 16:42:01.852324 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29557002-tdngs" event={"ID":"dce83493-bbc5-49ac-a420-229eac93005d","Type":"ContainerStarted","Data":"cc8a87d050d7d4532f011d62079dba28815de9c2ad3f8a91b6523667a402eab2"} Mar 13 16:42:04 crc kubenswrapper[4907]: I0313 16:42:04.897038 4907 generic.go:334] "Generic (PLEG): container finished" podID="dce83493-bbc5-49ac-a420-229eac93005d" containerID="3ca53b3677f234a39cf1706a0419909431a4370ca20c691d14a98c1cf5bf1d3b" exitCode=0 Mar 13 16:42:04 crc kubenswrapper[4907]: I0313 16:42:04.897257 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29557002-tdngs" event={"ID":"dce83493-bbc5-49ac-a420-229eac93005d","Type":"ContainerDied","Data":"3ca53b3677f234a39cf1706a0419909431a4370ca20c691d14a98c1cf5bf1d3b"} Mar 13 16:42:06 crc kubenswrapper[4907]: I0313 16:42:06.356931 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29557002-tdngs" Mar 13 16:42:06 crc kubenswrapper[4907]: I0313 16:42:06.486836 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fbw8l\" (UniqueName: \"kubernetes.io/projected/dce83493-bbc5-49ac-a420-229eac93005d-kube-api-access-fbw8l\") pod \"dce83493-bbc5-49ac-a420-229eac93005d\" (UID: \"dce83493-bbc5-49ac-a420-229eac93005d\") " Mar 13 16:42:06 crc kubenswrapper[4907]: I0313 16:42:06.495594 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dce83493-bbc5-49ac-a420-229eac93005d-kube-api-access-fbw8l" (OuterVolumeSpecName: "kube-api-access-fbw8l") pod "dce83493-bbc5-49ac-a420-229eac93005d" (UID: "dce83493-bbc5-49ac-a420-229eac93005d"). InnerVolumeSpecName "kube-api-access-fbw8l". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:42:06 crc kubenswrapper[4907]: I0313 16:42:06.589808 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fbw8l\" (UniqueName: \"kubernetes.io/projected/dce83493-bbc5-49ac-a420-229eac93005d-kube-api-access-fbw8l\") on node \"crc\" DevicePath \"\"" Mar 13 16:42:06 crc kubenswrapper[4907]: I0313 16:42:06.923519 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29557002-tdngs" event={"ID":"dce83493-bbc5-49ac-a420-229eac93005d","Type":"ContainerDied","Data":"cc8a87d050d7d4532f011d62079dba28815de9c2ad3f8a91b6523667a402eab2"} Mar 13 16:42:06 crc kubenswrapper[4907]: I0313 16:42:06.923581 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cc8a87d050d7d4532f011d62079dba28815de9c2ad3f8a91b6523667a402eab2" Mar 13 16:42:06 crc kubenswrapper[4907]: I0313 16:42:06.923658 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29557002-tdngs" Mar 13 16:42:07 crc kubenswrapper[4907]: I0313 16:42:07.449207 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556996-hhmgk"] Mar 13 16:42:07 crc kubenswrapper[4907]: I0313 16:42:07.459730 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556996-hhmgk"] Mar 13 16:42:07 crc kubenswrapper[4907]: I0313 16:42:07.798839 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d0ba1c68-f3a1-4c88-bc6b-db7c438dd2f5" path="/var/lib/kubelet/pods/d0ba1c68-f3a1-4c88-bc6b-db7c438dd2f5/volumes" Mar 13 16:42:15 crc kubenswrapper[4907]: I0313 16:42:15.783840 4907 scope.go:117] "RemoveContainer" containerID="7e4d0690f4ac607ea8a268bfa88b5ae6d9a95f8b6013de7747fdffb429f28047" Mar 13 16:42:15 crc kubenswrapper[4907]: E0313 16:42:15.785195 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:42:20 crc kubenswrapper[4907]: I0313 16:42:20.277590 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-s2qr6"] Mar 13 16:42:20 crc kubenswrapper[4907]: E0313 16:42:20.278686 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dce83493-bbc5-49ac-a420-229eac93005d" containerName="oc" Mar 13 16:42:20 crc kubenswrapper[4907]: I0313 16:42:20.278702 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="dce83493-bbc5-49ac-a420-229eac93005d" containerName="oc" Mar 13 16:42:20 crc kubenswrapper[4907]: I0313 16:42:20.279018 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="dce83493-bbc5-49ac-a420-229eac93005d" containerName="oc" Mar 13 16:42:20 crc kubenswrapper[4907]: I0313 16:42:20.280924 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-s2qr6" Mar 13 16:42:20 crc kubenswrapper[4907]: I0313 16:42:20.287997 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-s2qr6"] Mar 13 16:42:20 crc kubenswrapper[4907]: I0313 16:42:20.368301 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/20c838c2-685d-4fa7-bde6-a88be618698f-catalog-content\") pod \"redhat-operators-s2qr6\" (UID: \"20c838c2-685d-4fa7-bde6-a88be618698f\") " pod="openshift-marketplace/redhat-operators-s2qr6" Mar 13 16:42:20 crc kubenswrapper[4907]: I0313 16:42:20.368379 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qhmlq\" (UniqueName: \"kubernetes.io/projected/20c838c2-685d-4fa7-bde6-a88be618698f-kube-api-access-qhmlq\") pod \"redhat-operators-s2qr6\" (UID: \"20c838c2-685d-4fa7-bde6-a88be618698f\") " pod="openshift-marketplace/redhat-operators-s2qr6" Mar 13 16:42:20 crc kubenswrapper[4907]: I0313 16:42:20.368473 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/20c838c2-685d-4fa7-bde6-a88be618698f-utilities\") pod \"redhat-operators-s2qr6\" (UID: \"20c838c2-685d-4fa7-bde6-a88be618698f\") " pod="openshift-marketplace/redhat-operators-s2qr6" Mar 13 16:42:20 crc kubenswrapper[4907]: I0313 16:42:20.470428 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/20c838c2-685d-4fa7-bde6-a88be618698f-catalog-content\") pod \"redhat-operators-s2qr6\" (UID: \"20c838c2-685d-4fa7-bde6-a88be618698f\") " pod="openshift-marketplace/redhat-operators-s2qr6" Mar 13 16:42:20 crc kubenswrapper[4907]: I0313 16:42:20.470777 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qhmlq\" (UniqueName: \"kubernetes.io/projected/20c838c2-685d-4fa7-bde6-a88be618698f-kube-api-access-qhmlq\") pod \"redhat-operators-s2qr6\" (UID: \"20c838c2-685d-4fa7-bde6-a88be618698f\") " pod="openshift-marketplace/redhat-operators-s2qr6" Mar 13 16:42:20 crc kubenswrapper[4907]: I0313 16:42:20.470834 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/20c838c2-685d-4fa7-bde6-a88be618698f-utilities\") pod \"redhat-operators-s2qr6\" (UID: \"20c838c2-685d-4fa7-bde6-a88be618698f\") " pod="openshift-marketplace/redhat-operators-s2qr6" Mar 13 16:42:20 crc kubenswrapper[4907]: I0313 16:42:20.471032 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/20c838c2-685d-4fa7-bde6-a88be618698f-catalog-content\") pod \"redhat-operators-s2qr6\" (UID: \"20c838c2-685d-4fa7-bde6-a88be618698f\") " pod="openshift-marketplace/redhat-operators-s2qr6" Mar 13 16:42:20 crc kubenswrapper[4907]: I0313 16:42:20.471313 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/20c838c2-685d-4fa7-bde6-a88be618698f-utilities\") pod \"redhat-operators-s2qr6\" (UID: \"20c838c2-685d-4fa7-bde6-a88be618698f\") " pod="openshift-marketplace/redhat-operators-s2qr6" Mar 13 16:42:20 crc kubenswrapper[4907]: I0313 16:42:20.491875 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qhmlq\" (UniqueName: \"kubernetes.io/projected/20c838c2-685d-4fa7-bde6-a88be618698f-kube-api-access-qhmlq\") pod \"redhat-operators-s2qr6\" (UID: \"20c838c2-685d-4fa7-bde6-a88be618698f\") " pod="openshift-marketplace/redhat-operators-s2qr6" Mar 13 16:42:20 crc kubenswrapper[4907]: I0313 16:42:20.634187 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-s2qr6" Mar 13 16:42:21 crc kubenswrapper[4907]: I0313 16:42:21.132206 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-s2qr6"] Mar 13 16:42:22 crc kubenswrapper[4907]: I0313 16:42:22.103067 4907 generic.go:334] "Generic (PLEG): container finished" podID="20c838c2-685d-4fa7-bde6-a88be618698f" containerID="0c270278e59534e523470376925881213319928159d9e96183cc72cf4f76ee2e" exitCode=0 Mar 13 16:42:22 crc kubenswrapper[4907]: I0313 16:42:22.103371 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-s2qr6" event={"ID":"20c838c2-685d-4fa7-bde6-a88be618698f","Type":"ContainerDied","Data":"0c270278e59534e523470376925881213319928159d9e96183cc72cf4f76ee2e"} Mar 13 16:42:22 crc kubenswrapper[4907]: I0313 16:42:22.103401 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-s2qr6" event={"ID":"20c838c2-685d-4fa7-bde6-a88be618698f","Type":"ContainerStarted","Data":"2a340d1af9caff0371f5bfd63f2a8323b61e8d31999ba5f5f637876a00501768"} Mar 13 16:42:24 crc kubenswrapper[4907]: I0313 16:42:24.129314 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-s2qr6" event={"ID":"20c838c2-685d-4fa7-bde6-a88be618698f","Type":"ContainerStarted","Data":"167fc2eea8c311e949891f542a0aa9abeb88120496908b41c8378abfe2846f26"} Mar 13 16:42:28 crc kubenswrapper[4907]: I0313 16:42:28.181186 4907 generic.go:334] "Generic (PLEG): container finished" podID="20c838c2-685d-4fa7-bde6-a88be618698f" containerID="167fc2eea8c311e949891f542a0aa9abeb88120496908b41c8378abfe2846f26" exitCode=0 Mar 13 16:42:28 crc kubenswrapper[4907]: I0313 16:42:28.181261 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-s2qr6" event={"ID":"20c838c2-685d-4fa7-bde6-a88be618698f","Type":"ContainerDied","Data":"167fc2eea8c311e949891f542a0aa9abeb88120496908b41c8378abfe2846f26"} Mar 13 16:42:28 crc kubenswrapper[4907]: I0313 16:42:28.783805 4907 scope.go:117] "RemoveContainer" containerID="7e4d0690f4ac607ea8a268bfa88b5ae6d9a95f8b6013de7747fdffb429f28047" Mar 13 16:42:28 crc kubenswrapper[4907]: E0313 16:42:28.784988 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:42:30 crc kubenswrapper[4907]: I0313 16:42:30.209521 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-s2qr6" event={"ID":"20c838c2-685d-4fa7-bde6-a88be618698f","Type":"ContainerStarted","Data":"619dc8758167775cf4458460d2573597c7ee1b199f5844f41d3d53c7ecb99ab5"} Mar 13 16:42:30 crc kubenswrapper[4907]: I0313 16:42:30.249364 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-s2qr6" podStartSLOduration=3.191336437 podStartE2EDuration="10.249343034s" podCreationTimestamp="2026-03-13 16:42:20 +0000 UTC" firstStartedPulling="2026-03-13 16:42:22.107003149 +0000 UTC m=+9441.006790838" lastFinishedPulling="2026-03-13 16:42:29.165009716 +0000 UTC m=+9448.064797435" observedRunningTime="2026-03-13 16:42:30.23210169 +0000 UTC m=+9449.131889389" watchObservedRunningTime="2026-03-13 16:42:30.249343034 +0000 UTC m=+9449.149130733" Mar 13 16:42:30 crc kubenswrapper[4907]: I0313 16:42:30.634577 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-s2qr6" Mar 13 16:42:30 crc kubenswrapper[4907]: I0313 16:42:30.634701 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-s2qr6" Mar 13 16:42:31 crc kubenswrapper[4907]: I0313 16:42:31.695617 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-s2qr6" podUID="20c838c2-685d-4fa7-bde6-a88be618698f" containerName="registry-server" probeResult="failure" output=< Mar 13 16:42:31 crc kubenswrapper[4907]: timeout: failed to connect service ":50051" within 1s Mar 13 16:42:31 crc kubenswrapper[4907]: > Mar 13 16:42:40 crc kubenswrapper[4907]: I0313 16:42:40.698596 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-s2qr6" Mar 13 16:42:40 crc kubenswrapper[4907]: I0313 16:42:40.762854 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-s2qr6" Mar 13 16:42:40 crc kubenswrapper[4907]: I0313 16:42:40.952280 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-s2qr6"] Mar 13 16:42:42 crc kubenswrapper[4907]: I0313 16:42:42.524203 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-s2qr6" podUID="20c838c2-685d-4fa7-bde6-a88be618698f" containerName="registry-server" containerID="cri-o://619dc8758167775cf4458460d2573597c7ee1b199f5844f41d3d53c7ecb99ab5" gracePeriod=2 Mar 13 16:42:43 crc kubenswrapper[4907]: I0313 16:42:43.111916 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-s2qr6" Mar 13 16:42:43 crc kubenswrapper[4907]: I0313 16:42:43.223468 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qhmlq\" (UniqueName: \"kubernetes.io/projected/20c838c2-685d-4fa7-bde6-a88be618698f-kube-api-access-qhmlq\") pod \"20c838c2-685d-4fa7-bde6-a88be618698f\" (UID: \"20c838c2-685d-4fa7-bde6-a88be618698f\") " Mar 13 16:42:43 crc kubenswrapper[4907]: I0313 16:42:43.223568 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/20c838c2-685d-4fa7-bde6-a88be618698f-catalog-content\") pod \"20c838c2-685d-4fa7-bde6-a88be618698f\" (UID: \"20c838c2-685d-4fa7-bde6-a88be618698f\") " Mar 13 16:42:43 crc kubenswrapper[4907]: I0313 16:42:43.224450 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/20c838c2-685d-4fa7-bde6-a88be618698f-utilities\") pod \"20c838c2-685d-4fa7-bde6-a88be618698f\" (UID: \"20c838c2-685d-4fa7-bde6-a88be618698f\") " Mar 13 16:42:43 crc kubenswrapper[4907]: I0313 16:42:43.224992 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/20c838c2-685d-4fa7-bde6-a88be618698f-utilities" (OuterVolumeSpecName: "utilities") pod "20c838c2-685d-4fa7-bde6-a88be618698f" (UID: "20c838c2-685d-4fa7-bde6-a88be618698f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 16:42:43 crc kubenswrapper[4907]: I0313 16:42:43.231133 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20c838c2-685d-4fa7-bde6-a88be618698f-kube-api-access-qhmlq" (OuterVolumeSpecName: "kube-api-access-qhmlq") pod "20c838c2-685d-4fa7-bde6-a88be618698f" (UID: "20c838c2-685d-4fa7-bde6-a88be618698f"). InnerVolumeSpecName "kube-api-access-qhmlq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:42:43 crc kubenswrapper[4907]: I0313 16:42:43.326950 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/20c838c2-685d-4fa7-bde6-a88be618698f-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 16:42:43 crc kubenswrapper[4907]: I0313 16:42:43.326996 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qhmlq\" (UniqueName: \"kubernetes.io/projected/20c838c2-685d-4fa7-bde6-a88be618698f-kube-api-access-qhmlq\") on node \"crc\" DevicePath \"\"" Mar 13 16:42:43 crc kubenswrapper[4907]: I0313 16:42:43.361681 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/20c838c2-685d-4fa7-bde6-a88be618698f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "20c838c2-685d-4fa7-bde6-a88be618698f" (UID: "20c838c2-685d-4fa7-bde6-a88be618698f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 16:42:43 crc kubenswrapper[4907]: I0313 16:42:43.428388 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/20c838c2-685d-4fa7-bde6-a88be618698f-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 16:42:43 crc kubenswrapper[4907]: I0313 16:42:43.537782 4907 generic.go:334] "Generic (PLEG): container finished" podID="20c838c2-685d-4fa7-bde6-a88be618698f" containerID="619dc8758167775cf4458460d2573597c7ee1b199f5844f41d3d53c7ecb99ab5" exitCode=0 Mar 13 16:42:43 crc kubenswrapper[4907]: I0313 16:42:43.537828 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-s2qr6" event={"ID":"20c838c2-685d-4fa7-bde6-a88be618698f","Type":"ContainerDied","Data":"619dc8758167775cf4458460d2573597c7ee1b199f5844f41d3d53c7ecb99ab5"} Mar 13 16:42:43 crc kubenswrapper[4907]: I0313 16:42:43.537855 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-s2qr6" Mar 13 16:42:43 crc kubenswrapper[4907]: I0313 16:42:43.537875 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-s2qr6" event={"ID":"20c838c2-685d-4fa7-bde6-a88be618698f","Type":"ContainerDied","Data":"2a340d1af9caff0371f5bfd63f2a8323b61e8d31999ba5f5f637876a00501768"} Mar 13 16:42:43 crc kubenswrapper[4907]: I0313 16:42:43.537992 4907 scope.go:117] "RemoveContainer" containerID="619dc8758167775cf4458460d2573597c7ee1b199f5844f41d3d53c7ecb99ab5" Mar 13 16:42:43 crc kubenswrapper[4907]: I0313 16:42:43.572602 4907 scope.go:117] "RemoveContainer" containerID="167fc2eea8c311e949891f542a0aa9abeb88120496908b41c8378abfe2846f26" Mar 13 16:42:43 crc kubenswrapper[4907]: I0313 16:42:43.595062 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-s2qr6"] Mar 13 16:42:43 crc kubenswrapper[4907]: I0313 16:42:43.606631 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-s2qr6"] Mar 13 16:42:43 crc kubenswrapper[4907]: I0313 16:42:43.612172 4907 scope.go:117] "RemoveContainer" containerID="0c270278e59534e523470376925881213319928159d9e96183cc72cf4f76ee2e" Mar 13 16:42:43 crc kubenswrapper[4907]: I0313 16:42:43.654920 4907 scope.go:117] "RemoveContainer" containerID="619dc8758167775cf4458460d2573597c7ee1b199f5844f41d3d53c7ecb99ab5" Mar 13 16:42:43 crc kubenswrapper[4907]: E0313 16:42:43.655742 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"619dc8758167775cf4458460d2573597c7ee1b199f5844f41d3d53c7ecb99ab5\": container with ID starting with 619dc8758167775cf4458460d2573597c7ee1b199f5844f41d3d53c7ecb99ab5 not found: ID does not exist" containerID="619dc8758167775cf4458460d2573597c7ee1b199f5844f41d3d53c7ecb99ab5" Mar 13 16:42:43 crc kubenswrapper[4907]: I0313 16:42:43.655825 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"619dc8758167775cf4458460d2573597c7ee1b199f5844f41d3d53c7ecb99ab5"} err="failed to get container status \"619dc8758167775cf4458460d2573597c7ee1b199f5844f41d3d53c7ecb99ab5\": rpc error: code = NotFound desc = could not find container \"619dc8758167775cf4458460d2573597c7ee1b199f5844f41d3d53c7ecb99ab5\": container with ID starting with 619dc8758167775cf4458460d2573597c7ee1b199f5844f41d3d53c7ecb99ab5 not found: ID does not exist" Mar 13 16:42:43 crc kubenswrapper[4907]: I0313 16:42:43.655868 4907 scope.go:117] "RemoveContainer" containerID="167fc2eea8c311e949891f542a0aa9abeb88120496908b41c8378abfe2846f26" Mar 13 16:42:43 crc kubenswrapper[4907]: E0313 16:42:43.656384 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"167fc2eea8c311e949891f542a0aa9abeb88120496908b41c8378abfe2846f26\": container with ID starting with 167fc2eea8c311e949891f542a0aa9abeb88120496908b41c8378abfe2846f26 not found: ID does not exist" containerID="167fc2eea8c311e949891f542a0aa9abeb88120496908b41c8378abfe2846f26" Mar 13 16:42:43 crc kubenswrapper[4907]: I0313 16:42:43.656451 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"167fc2eea8c311e949891f542a0aa9abeb88120496908b41c8378abfe2846f26"} err="failed to get container status \"167fc2eea8c311e949891f542a0aa9abeb88120496908b41c8378abfe2846f26\": rpc error: code = NotFound desc = could not find container \"167fc2eea8c311e949891f542a0aa9abeb88120496908b41c8378abfe2846f26\": container with ID starting with 167fc2eea8c311e949891f542a0aa9abeb88120496908b41c8378abfe2846f26 not found: ID does not exist" Mar 13 16:42:43 crc kubenswrapper[4907]: I0313 16:42:43.656492 4907 scope.go:117] "RemoveContainer" containerID="0c270278e59534e523470376925881213319928159d9e96183cc72cf4f76ee2e" Mar 13 16:42:43 crc kubenswrapper[4907]: E0313 16:42:43.657622 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0c270278e59534e523470376925881213319928159d9e96183cc72cf4f76ee2e\": container with ID starting with 0c270278e59534e523470376925881213319928159d9e96183cc72cf4f76ee2e not found: ID does not exist" containerID="0c270278e59534e523470376925881213319928159d9e96183cc72cf4f76ee2e" Mar 13 16:42:43 crc kubenswrapper[4907]: I0313 16:42:43.657661 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c270278e59534e523470376925881213319928159d9e96183cc72cf4f76ee2e"} err="failed to get container status \"0c270278e59534e523470376925881213319928159d9e96183cc72cf4f76ee2e\": rpc error: code = NotFound desc = could not find container \"0c270278e59534e523470376925881213319928159d9e96183cc72cf4f76ee2e\": container with ID starting with 0c270278e59534e523470376925881213319928159d9e96183cc72cf4f76ee2e not found: ID does not exist" Mar 13 16:42:43 crc kubenswrapper[4907]: I0313 16:42:43.783396 4907 scope.go:117] "RemoveContainer" containerID="7e4d0690f4ac607ea8a268bfa88b5ae6d9a95f8b6013de7747fdffb429f28047" Mar 13 16:42:43 crc kubenswrapper[4907]: E0313 16:42:43.784181 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:42:43 crc kubenswrapper[4907]: I0313 16:42:43.800576 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20c838c2-685d-4fa7-bde6-a88be618698f" path="/var/lib/kubelet/pods/20c838c2-685d-4fa7-bde6-a88be618698f/volumes" Mar 13 16:42:44 crc kubenswrapper[4907]: I0313 16:42:44.427829 4907 scope.go:117] "RemoveContainer" containerID="fc2a84d7603e9a5913a999c6beb101ceda891932552390c08d77f4d764a6c116" Mar 13 16:42:54 crc kubenswrapper[4907]: I0313 16:42:54.783108 4907 scope.go:117] "RemoveContainer" containerID="7e4d0690f4ac607ea8a268bfa88b5ae6d9a95f8b6013de7747fdffb429f28047" Mar 13 16:42:54 crc kubenswrapper[4907]: E0313 16:42:54.783928 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:43:06 crc kubenswrapper[4907]: I0313 16:43:06.783084 4907 scope.go:117] "RemoveContainer" containerID="7e4d0690f4ac607ea8a268bfa88b5ae6d9a95f8b6013de7747fdffb429f28047" Mar 13 16:43:06 crc kubenswrapper[4907]: E0313 16:43:06.784035 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:43:17 crc kubenswrapper[4907]: I0313 16:43:17.783517 4907 scope.go:117] "RemoveContainer" containerID="7e4d0690f4ac607ea8a268bfa88b5ae6d9a95f8b6013de7747fdffb429f28047" Mar 13 16:43:17 crc kubenswrapper[4907]: E0313 16:43:17.784585 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:43:28 crc kubenswrapper[4907]: I0313 16:43:28.782781 4907 scope.go:117] "RemoveContainer" containerID="7e4d0690f4ac607ea8a268bfa88b5ae6d9a95f8b6013de7747fdffb429f28047" Mar 13 16:43:28 crc kubenswrapper[4907]: E0313 16:43:28.783587 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:43:40 crc kubenswrapper[4907]: I0313 16:43:40.783168 4907 scope.go:117] "RemoveContainer" containerID="7e4d0690f4ac607ea8a268bfa88b5ae6d9a95f8b6013de7747fdffb429f28047" Mar 13 16:43:40 crc kubenswrapper[4907]: E0313 16:43:40.784149 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:43:51 crc kubenswrapper[4907]: I0313 16:43:51.788675 4907 scope.go:117] "RemoveContainer" containerID="7e4d0690f4ac607ea8a268bfa88b5ae6d9a95f8b6013de7747fdffb429f28047" Mar 13 16:43:52 crc kubenswrapper[4907]: I0313 16:43:52.231562 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerStarted","Data":"cf78e901eff913d33c2f4e21ad1865cb4b0011402563e6e1fc5c1c3c95de10e9"} Mar 13 16:44:00 crc kubenswrapper[4907]: I0313 16:44:00.148279 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29557004-v4vjf"] Mar 13 16:44:00 crc kubenswrapper[4907]: E0313 16:44:00.149671 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20c838c2-685d-4fa7-bde6-a88be618698f" containerName="extract-utilities" Mar 13 16:44:00 crc kubenswrapper[4907]: I0313 16:44:00.149695 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="20c838c2-685d-4fa7-bde6-a88be618698f" containerName="extract-utilities" Mar 13 16:44:00 crc kubenswrapper[4907]: E0313 16:44:00.149729 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20c838c2-685d-4fa7-bde6-a88be618698f" containerName="extract-content" Mar 13 16:44:00 crc kubenswrapper[4907]: I0313 16:44:00.149741 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="20c838c2-685d-4fa7-bde6-a88be618698f" containerName="extract-content" Mar 13 16:44:00 crc kubenswrapper[4907]: E0313 16:44:00.149795 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20c838c2-685d-4fa7-bde6-a88be618698f" containerName="registry-server" Mar 13 16:44:00 crc kubenswrapper[4907]: I0313 16:44:00.149804 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="20c838c2-685d-4fa7-bde6-a88be618698f" containerName="registry-server" Mar 13 16:44:00 crc kubenswrapper[4907]: I0313 16:44:00.150120 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="20c838c2-685d-4fa7-bde6-a88be618698f" containerName="registry-server" Mar 13 16:44:00 crc kubenswrapper[4907]: I0313 16:44:00.151157 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29557004-v4vjf" Mar 13 16:44:00 crc kubenswrapper[4907]: I0313 16:44:00.154669 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 16:44:00 crc kubenswrapper[4907]: I0313 16:44:00.154840 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 16:44:00 crc kubenswrapper[4907]: I0313 16:44:00.155023 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 16:44:00 crc kubenswrapper[4907]: I0313 16:44:00.164415 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29557004-v4vjf"] Mar 13 16:44:00 crc kubenswrapper[4907]: I0313 16:44:00.279054 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bdlp2\" (UniqueName: \"kubernetes.io/projected/724a5965-1261-4da5-a833-40c8af08fde6-kube-api-access-bdlp2\") pod \"auto-csr-approver-29557004-v4vjf\" (UID: \"724a5965-1261-4da5-a833-40c8af08fde6\") " pod="openshift-infra/auto-csr-approver-29557004-v4vjf" Mar 13 16:44:00 crc kubenswrapper[4907]: I0313 16:44:00.381385 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bdlp2\" (UniqueName: \"kubernetes.io/projected/724a5965-1261-4da5-a833-40c8af08fde6-kube-api-access-bdlp2\") pod \"auto-csr-approver-29557004-v4vjf\" (UID: \"724a5965-1261-4da5-a833-40c8af08fde6\") " pod="openshift-infra/auto-csr-approver-29557004-v4vjf" Mar 13 16:44:00 crc kubenswrapper[4907]: I0313 16:44:00.418511 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bdlp2\" (UniqueName: \"kubernetes.io/projected/724a5965-1261-4da5-a833-40c8af08fde6-kube-api-access-bdlp2\") pod \"auto-csr-approver-29557004-v4vjf\" (UID: \"724a5965-1261-4da5-a833-40c8af08fde6\") " pod="openshift-infra/auto-csr-approver-29557004-v4vjf" Mar 13 16:44:00 crc kubenswrapper[4907]: I0313 16:44:00.476029 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29557004-v4vjf" Mar 13 16:44:01 crc kubenswrapper[4907]: I0313 16:44:01.018024 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29557004-v4vjf"] Mar 13 16:44:01 crc kubenswrapper[4907]: W0313 16:44:01.020806 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod724a5965_1261_4da5_a833_40c8af08fde6.slice/crio-c033b71ea5e962cba3a37e561238afdcae9f063023456a9b5197facd2b95a8db WatchSource:0}: Error finding container c033b71ea5e962cba3a37e561238afdcae9f063023456a9b5197facd2b95a8db: Status 404 returned error can't find the container with id c033b71ea5e962cba3a37e561238afdcae9f063023456a9b5197facd2b95a8db Mar 13 16:44:01 crc kubenswrapper[4907]: I0313 16:44:01.336826 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29557004-v4vjf" event={"ID":"724a5965-1261-4da5-a833-40c8af08fde6","Type":"ContainerStarted","Data":"c033b71ea5e962cba3a37e561238afdcae9f063023456a9b5197facd2b95a8db"} Mar 13 16:44:03 crc kubenswrapper[4907]: I0313 16:44:03.371706 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29557004-v4vjf" event={"ID":"724a5965-1261-4da5-a833-40c8af08fde6","Type":"ContainerStarted","Data":"656ee697d354c492175b1fe789d5e08c0348dae2d403ef138ab40b1d71db2845"} Mar 13 16:44:03 crc kubenswrapper[4907]: I0313 16:44:03.408680 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29557004-v4vjf" podStartSLOduration=2.5342579819999997 podStartE2EDuration="3.408648985s" podCreationTimestamp="2026-03-13 16:44:00 +0000 UTC" firstStartedPulling="2026-03-13 16:44:01.023102398 +0000 UTC m=+9539.922890117" lastFinishedPulling="2026-03-13 16:44:01.897493431 +0000 UTC m=+9540.797281120" observedRunningTime="2026-03-13 16:44:03.392752988 +0000 UTC m=+9542.292540707" watchObservedRunningTime="2026-03-13 16:44:03.408648985 +0000 UTC m=+9542.308436684" Mar 13 16:44:04 crc kubenswrapper[4907]: I0313 16:44:04.392108 4907 generic.go:334] "Generic (PLEG): container finished" podID="724a5965-1261-4da5-a833-40c8af08fde6" containerID="656ee697d354c492175b1fe789d5e08c0348dae2d403ef138ab40b1d71db2845" exitCode=0 Mar 13 16:44:04 crc kubenswrapper[4907]: I0313 16:44:04.392228 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29557004-v4vjf" event={"ID":"724a5965-1261-4da5-a833-40c8af08fde6","Type":"ContainerDied","Data":"656ee697d354c492175b1fe789d5e08c0348dae2d403ef138ab40b1d71db2845"} Mar 13 16:44:05 crc kubenswrapper[4907]: I0313 16:44:05.882018 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29557004-v4vjf" Mar 13 16:44:06 crc kubenswrapper[4907]: I0313 16:44:06.039976 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bdlp2\" (UniqueName: \"kubernetes.io/projected/724a5965-1261-4da5-a833-40c8af08fde6-kube-api-access-bdlp2\") pod \"724a5965-1261-4da5-a833-40c8af08fde6\" (UID: \"724a5965-1261-4da5-a833-40c8af08fde6\") " Mar 13 16:44:06 crc kubenswrapper[4907]: I0313 16:44:06.053325 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/724a5965-1261-4da5-a833-40c8af08fde6-kube-api-access-bdlp2" (OuterVolumeSpecName: "kube-api-access-bdlp2") pod "724a5965-1261-4da5-a833-40c8af08fde6" (UID: "724a5965-1261-4da5-a833-40c8af08fde6"). InnerVolumeSpecName "kube-api-access-bdlp2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:44:06 crc kubenswrapper[4907]: I0313 16:44:06.143157 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bdlp2\" (UniqueName: \"kubernetes.io/projected/724a5965-1261-4da5-a833-40c8af08fde6-kube-api-access-bdlp2\") on node \"crc\" DevicePath \"\"" Mar 13 16:44:06 crc kubenswrapper[4907]: I0313 16:44:06.427332 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29557004-v4vjf" event={"ID":"724a5965-1261-4da5-a833-40c8af08fde6","Type":"ContainerDied","Data":"c033b71ea5e962cba3a37e561238afdcae9f063023456a9b5197facd2b95a8db"} Mar 13 16:44:06 crc kubenswrapper[4907]: I0313 16:44:06.427380 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c033b71ea5e962cba3a37e561238afdcae9f063023456a9b5197facd2b95a8db" Mar 13 16:44:06 crc kubenswrapper[4907]: I0313 16:44:06.427443 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29557004-v4vjf" Mar 13 16:44:06 crc kubenswrapper[4907]: I0313 16:44:06.487250 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556998-pzdj2"] Mar 13 16:44:06 crc kubenswrapper[4907]: I0313 16:44:06.495688 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556998-pzdj2"] Mar 13 16:44:07 crc kubenswrapper[4907]: I0313 16:44:07.808797 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2abff695-80c2-4588-9796-44e850ddb65f" path="/var/lib/kubelet/pods/2abff695-80c2-4588-9796-44e850ddb65f/volumes" Mar 13 16:44:33 crc kubenswrapper[4907]: I0313 16:44:33.283857 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-b9df4"] Mar 13 16:44:33 crc kubenswrapper[4907]: E0313 16:44:33.285757 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="724a5965-1261-4da5-a833-40c8af08fde6" containerName="oc" Mar 13 16:44:33 crc kubenswrapper[4907]: I0313 16:44:33.285776 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="724a5965-1261-4da5-a833-40c8af08fde6" containerName="oc" Mar 13 16:44:33 crc kubenswrapper[4907]: I0313 16:44:33.286145 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="724a5965-1261-4da5-a833-40c8af08fde6" containerName="oc" Mar 13 16:44:33 crc kubenswrapper[4907]: I0313 16:44:33.288093 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-b9df4" Mar 13 16:44:33 crc kubenswrapper[4907]: I0313 16:44:33.302206 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-b9df4"] Mar 13 16:44:33 crc kubenswrapper[4907]: I0313 16:44:33.342630 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1428c7e-574e-4fa0-a364-14a6c85d413a-utilities\") pod \"certified-operators-b9df4\" (UID: \"b1428c7e-574e-4fa0-a364-14a6c85d413a\") " pod="openshift-marketplace/certified-operators-b9df4" Mar 13 16:44:33 crc kubenswrapper[4907]: I0313 16:44:33.343137 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2m8vv\" (UniqueName: \"kubernetes.io/projected/b1428c7e-574e-4fa0-a364-14a6c85d413a-kube-api-access-2m8vv\") pod \"certified-operators-b9df4\" (UID: \"b1428c7e-574e-4fa0-a364-14a6c85d413a\") " pod="openshift-marketplace/certified-operators-b9df4" Mar 13 16:44:33 crc kubenswrapper[4907]: I0313 16:44:33.343229 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1428c7e-574e-4fa0-a364-14a6c85d413a-catalog-content\") pod \"certified-operators-b9df4\" (UID: \"b1428c7e-574e-4fa0-a364-14a6c85d413a\") " pod="openshift-marketplace/certified-operators-b9df4" Mar 13 16:44:33 crc kubenswrapper[4907]: I0313 16:44:33.444821 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1428c7e-574e-4fa0-a364-14a6c85d413a-utilities\") pod \"certified-operators-b9df4\" (UID: \"b1428c7e-574e-4fa0-a364-14a6c85d413a\") " pod="openshift-marketplace/certified-operators-b9df4" Mar 13 16:44:33 crc kubenswrapper[4907]: I0313 16:44:33.445081 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2m8vv\" (UniqueName: \"kubernetes.io/projected/b1428c7e-574e-4fa0-a364-14a6c85d413a-kube-api-access-2m8vv\") pod \"certified-operators-b9df4\" (UID: \"b1428c7e-574e-4fa0-a364-14a6c85d413a\") " pod="openshift-marketplace/certified-operators-b9df4" Mar 13 16:44:33 crc kubenswrapper[4907]: I0313 16:44:33.445191 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1428c7e-574e-4fa0-a364-14a6c85d413a-catalog-content\") pod \"certified-operators-b9df4\" (UID: \"b1428c7e-574e-4fa0-a364-14a6c85d413a\") " pod="openshift-marketplace/certified-operators-b9df4" Mar 13 16:44:33 crc kubenswrapper[4907]: I0313 16:44:33.445307 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1428c7e-574e-4fa0-a364-14a6c85d413a-utilities\") pod \"certified-operators-b9df4\" (UID: \"b1428c7e-574e-4fa0-a364-14a6c85d413a\") " pod="openshift-marketplace/certified-operators-b9df4" Mar 13 16:44:33 crc kubenswrapper[4907]: I0313 16:44:33.445445 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1428c7e-574e-4fa0-a364-14a6c85d413a-catalog-content\") pod \"certified-operators-b9df4\" (UID: \"b1428c7e-574e-4fa0-a364-14a6c85d413a\") " pod="openshift-marketplace/certified-operators-b9df4" Mar 13 16:44:33 crc kubenswrapper[4907]: I0313 16:44:33.470672 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2m8vv\" (UniqueName: \"kubernetes.io/projected/b1428c7e-574e-4fa0-a364-14a6c85d413a-kube-api-access-2m8vv\") pod \"certified-operators-b9df4\" (UID: \"b1428c7e-574e-4fa0-a364-14a6c85d413a\") " pod="openshift-marketplace/certified-operators-b9df4" Mar 13 16:44:33 crc kubenswrapper[4907]: I0313 16:44:33.634229 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-b9df4" Mar 13 16:44:34 crc kubenswrapper[4907]: I0313 16:44:34.136859 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-b9df4"] Mar 13 16:44:34 crc kubenswrapper[4907]: I0313 16:44:34.821873 4907 generic.go:334] "Generic (PLEG): container finished" podID="b1428c7e-574e-4fa0-a364-14a6c85d413a" containerID="406b700ec04569af0a43aa0f00c007b21e68df9b9a2e35767629619ce91ea8f4" exitCode=0 Mar 13 16:44:34 crc kubenswrapper[4907]: I0313 16:44:34.821936 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b9df4" event={"ID":"b1428c7e-574e-4fa0-a364-14a6c85d413a","Type":"ContainerDied","Data":"406b700ec04569af0a43aa0f00c007b21e68df9b9a2e35767629619ce91ea8f4"} Mar 13 16:44:34 crc kubenswrapper[4907]: I0313 16:44:34.822002 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b9df4" event={"ID":"b1428c7e-574e-4fa0-a364-14a6c85d413a","Type":"ContainerStarted","Data":"1e5271fea52afdf222fc307f42b571fe15871546b42ce2ccc995b3c6f4ae18ff"} Mar 13 16:44:35 crc kubenswrapper[4907]: I0313 16:44:35.835395 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b9df4" event={"ID":"b1428c7e-574e-4fa0-a364-14a6c85d413a","Type":"ContainerStarted","Data":"c70795cf0ff18f40beec2e966464d6e8d8e8e805f466ccd0e9155b99894fdc8b"} Mar 13 16:44:37 crc kubenswrapper[4907]: I0313 16:44:37.864810 4907 generic.go:334] "Generic (PLEG): container finished" podID="b1428c7e-574e-4fa0-a364-14a6c85d413a" containerID="c70795cf0ff18f40beec2e966464d6e8d8e8e805f466ccd0e9155b99894fdc8b" exitCode=0 Mar 13 16:44:37 crc kubenswrapper[4907]: I0313 16:44:37.864945 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b9df4" event={"ID":"b1428c7e-574e-4fa0-a364-14a6c85d413a","Type":"ContainerDied","Data":"c70795cf0ff18f40beec2e966464d6e8d8e8e805f466ccd0e9155b99894fdc8b"} Mar 13 16:44:38 crc kubenswrapper[4907]: I0313 16:44:38.877893 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b9df4" event={"ID":"b1428c7e-574e-4fa0-a364-14a6c85d413a","Type":"ContainerStarted","Data":"9083630a661a76583aad7aa2abb1771ff314d367addcdb10ff8aca627760d825"} Mar 13 16:44:38 crc kubenswrapper[4907]: I0313 16:44:38.901470 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-b9df4" podStartSLOduration=2.400423634 podStartE2EDuration="5.901456055s" podCreationTimestamp="2026-03-13 16:44:33 +0000 UTC" firstStartedPulling="2026-03-13 16:44:34.829745512 +0000 UTC m=+9573.729533221" lastFinishedPulling="2026-03-13 16:44:38.330777943 +0000 UTC m=+9577.230565642" observedRunningTime="2026-03-13 16:44:38.898317287 +0000 UTC m=+9577.798104986" watchObservedRunningTime="2026-03-13 16:44:38.901456055 +0000 UTC m=+9577.801243744" Mar 13 16:44:43 crc kubenswrapper[4907]: I0313 16:44:43.748925 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-b9df4" Mar 13 16:44:43 crc kubenswrapper[4907]: I0313 16:44:43.749602 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-b9df4" Mar 13 16:44:44 crc kubenswrapper[4907]: I0313 16:44:44.369664 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-b9df4" Mar 13 16:44:44 crc kubenswrapper[4907]: I0313 16:44:44.426000 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-b9df4" Mar 13 16:44:44 crc kubenswrapper[4907]: I0313 16:44:44.571126 4907 scope.go:117] "RemoveContainer" containerID="b1d7df69f877a557d28aa9c5dd1537366e05d49104f113b20536b953f8c2ec90" Mar 13 16:44:44 crc kubenswrapper[4907]: I0313 16:44:44.616320 4907 scope.go:117] "RemoveContainer" containerID="406f17bcc9c6733e218df39e7a5238412a849d1df878824a54db3db1358e5bff" Mar 13 16:44:44 crc kubenswrapper[4907]: I0313 16:44:44.632363 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-b9df4"] Mar 13 16:44:45 crc kubenswrapper[4907]: I0313 16:44:45.955405 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-b9df4" podUID="b1428c7e-574e-4fa0-a364-14a6c85d413a" containerName="registry-server" containerID="cri-o://9083630a661a76583aad7aa2abb1771ff314d367addcdb10ff8aca627760d825" gracePeriod=2 Mar 13 16:44:46 crc kubenswrapper[4907]: I0313 16:44:46.489285 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-b9df4" Mar 13 16:44:46 crc kubenswrapper[4907]: I0313 16:44:46.652845 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2m8vv\" (UniqueName: \"kubernetes.io/projected/b1428c7e-574e-4fa0-a364-14a6c85d413a-kube-api-access-2m8vv\") pod \"b1428c7e-574e-4fa0-a364-14a6c85d413a\" (UID: \"b1428c7e-574e-4fa0-a364-14a6c85d413a\") " Mar 13 16:44:46 crc kubenswrapper[4907]: I0313 16:44:46.653035 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1428c7e-574e-4fa0-a364-14a6c85d413a-utilities\") pod \"b1428c7e-574e-4fa0-a364-14a6c85d413a\" (UID: \"b1428c7e-574e-4fa0-a364-14a6c85d413a\") " Mar 13 16:44:46 crc kubenswrapper[4907]: I0313 16:44:46.653103 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1428c7e-574e-4fa0-a364-14a6c85d413a-catalog-content\") pod \"b1428c7e-574e-4fa0-a364-14a6c85d413a\" (UID: \"b1428c7e-574e-4fa0-a364-14a6c85d413a\") " Mar 13 16:44:46 crc kubenswrapper[4907]: I0313 16:44:46.653958 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b1428c7e-574e-4fa0-a364-14a6c85d413a-utilities" (OuterVolumeSpecName: "utilities") pod "b1428c7e-574e-4fa0-a364-14a6c85d413a" (UID: "b1428c7e-574e-4fa0-a364-14a6c85d413a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 16:44:46 crc kubenswrapper[4907]: I0313 16:44:46.659895 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b1428c7e-574e-4fa0-a364-14a6c85d413a-kube-api-access-2m8vv" (OuterVolumeSpecName: "kube-api-access-2m8vv") pod "b1428c7e-574e-4fa0-a364-14a6c85d413a" (UID: "b1428c7e-574e-4fa0-a364-14a6c85d413a"). InnerVolumeSpecName "kube-api-access-2m8vv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:44:46 crc kubenswrapper[4907]: I0313 16:44:46.755692 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2m8vv\" (UniqueName: \"kubernetes.io/projected/b1428c7e-574e-4fa0-a364-14a6c85d413a-kube-api-access-2m8vv\") on node \"crc\" DevicePath \"\"" Mar 13 16:44:46 crc kubenswrapper[4907]: I0313 16:44:46.755744 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1428c7e-574e-4fa0-a364-14a6c85d413a-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 16:44:46 crc kubenswrapper[4907]: I0313 16:44:46.971137 4907 generic.go:334] "Generic (PLEG): container finished" podID="b1428c7e-574e-4fa0-a364-14a6c85d413a" containerID="9083630a661a76583aad7aa2abb1771ff314d367addcdb10ff8aca627760d825" exitCode=0 Mar 13 16:44:46 crc kubenswrapper[4907]: I0313 16:44:46.971554 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b9df4" event={"ID":"b1428c7e-574e-4fa0-a364-14a6c85d413a","Type":"ContainerDied","Data":"9083630a661a76583aad7aa2abb1771ff314d367addcdb10ff8aca627760d825"} Mar 13 16:44:46 crc kubenswrapper[4907]: I0313 16:44:46.971593 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b9df4" event={"ID":"b1428c7e-574e-4fa0-a364-14a6c85d413a","Type":"ContainerDied","Data":"1e5271fea52afdf222fc307f42b571fe15871546b42ce2ccc995b3c6f4ae18ff"} Mar 13 16:44:46 crc kubenswrapper[4907]: I0313 16:44:46.971617 4907 scope.go:117] "RemoveContainer" containerID="9083630a661a76583aad7aa2abb1771ff314d367addcdb10ff8aca627760d825" Mar 13 16:44:46 crc kubenswrapper[4907]: I0313 16:44:46.971776 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-b9df4" Mar 13 16:44:47 crc kubenswrapper[4907]: I0313 16:44:47.002209 4907 scope.go:117] "RemoveContainer" containerID="c70795cf0ff18f40beec2e966464d6e8d8e8e805f466ccd0e9155b99894fdc8b" Mar 13 16:44:47 crc kubenswrapper[4907]: I0313 16:44:47.035765 4907 scope.go:117] "RemoveContainer" containerID="406b700ec04569af0a43aa0f00c007b21e68df9b9a2e35767629619ce91ea8f4" Mar 13 16:44:47 crc kubenswrapper[4907]: I0313 16:44:47.070116 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b1428c7e-574e-4fa0-a364-14a6c85d413a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b1428c7e-574e-4fa0-a364-14a6c85d413a" (UID: "b1428c7e-574e-4fa0-a364-14a6c85d413a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 16:44:47 crc kubenswrapper[4907]: I0313 16:44:47.096674 4907 scope.go:117] "RemoveContainer" containerID="9083630a661a76583aad7aa2abb1771ff314d367addcdb10ff8aca627760d825" Mar 13 16:44:47 crc kubenswrapper[4907]: E0313 16:44:47.097202 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9083630a661a76583aad7aa2abb1771ff314d367addcdb10ff8aca627760d825\": container with ID starting with 9083630a661a76583aad7aa2abb1771ff314d367addcdb10ff8aca627760d825 not found: ID does not exist" containerID="9083630a661a76583aad7aa2abb1771ff314d367addcdb10ff8aca627760d825" Mar 13 16:44:47 crc kubenswrapper[4907]: I0313 16:44:47.097266 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9083630a661a76583aad7aa2abb1771ff314d367addcdb10ff8aca627760d825"} err="failed to get container status \"9083630a661a76583aad7aa2abb1771ff314d367addcdb10ff8aca627760d825\": rpc error: code = NotFound desc = could not find container \"9083630a661a76583aad7aa2abb1771ff314d367addcdb10ff8aca627760d825\": container with ID starting with 9083630a661a76583aad7aa2abb1771ff314d367addcdb10ff8aca627760d825 not found: ID does not exist" Mar 13 16:44:47 crc kubenswrapper[4907]: I0313 16:44:47.097299 4907 scope.go:117] "RemoveContainer" containerID="c70795cf0ff18f40beec2e966464d6e8d8e8e805f466ccd0e9155b99894fdc8b" Mar 13 16:44:47 crc kubenswrapper[4907]: E0313 16:44:47.098300 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c70795cf0ff18f40beec2e966464d6e8d8e8e805f466ccd0e9155b99894fdc8b\": container with ID starting with c70795cf0ff18f40beec2e966464d6e8d8e8e805f466ccd0e9155b99894fdc8b not found: ID does not exist" containerID="c70795cf0ff18f40beec2e966464d6e8d8e8e805f466ccd0e9155b99894fdc8b" Mar 13 16:44:47 crc kubenswrapper[4907]: I0313 16:44:47.098346 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c70795cf0ff18f40beec2e966464d6e8d8e8e805f466ccd0e9155b99894fdc8b"} err="failed to get container status \"c70795cf0ff18f40beec2e966464d6e8d8e8e805f466ccd0e9155b99894fdc8b\": rpc error: code = NotFound desc = could not find container \"c70795cf0ff18f40beec2e966464d6e8d8e8e805f466ccd0e9155b99894fdc8b\": container with ID starting with c70795cf0ff18f40beec2e966464d6e8d8e8e805f466ccd0e9155b99894fdc8b not found: ID does not exist" Mar 13 16:44:47 crc kubenswrapper[4907]: I0313 16:44:47.098373 4907 scope.go:117] "RemoveContainer" containerID="406b700ec04569af0a43aa0f00c007b21e68df9b9a2e35767629619ce91ea8f4" Mar 13 16:44:47 crc kubenswrapper[4907]: E0313 16:44:47.098870 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"406b700ec04569af0a43aa0f00c007b21e68df9b9a2e35767629619ce91ea8f4\": container with ID starting with 406b700ec04569af0a43aa0f00c007b21e68df9b9a2e35767629619ce91ea8f4 not found: ID does not exist" containerID="406b700ec04569af0a43aa0f00c007b21e68df9b9a2e35767629619ce91ea8f4" Mar 13 16:44:47 crc kubenswrapper[4907]: I0313 16:44:47.098928 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"406b700ec04569af0a43aa0f00c007b21e68df9b9a2e35767629619ce91ea8f4"} err="failed to get container status \"406b700ec04569af0a43aa0f00c007b21e68df9b9a2e35767629619ce91ea8f4\": rpc error: code = NotFound desc = could not find container \"406b700ec04569af0a43aa0f00c007b21e68df9b9a2e35767629619ce91ea8f4\": container with ID starting with 406b700ec04569af0a43aa0f00c007b21e68df9b9a2e35767629619ce91ea8f4 not found: ID does not exist" Mar 13 16:44:47 crc kubenswrapper[4907]: I0313 16:44:47.179634 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1428c7e-574e-4fa0-a364-14a6c85d413a-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 16:44:47 crc kubenswrapper[4907]: I0313 16:44:47.308256 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-b9df4"] Mar 13 16:44:47 crc kubenswrapper[4907]: I0313 16:44:47.320299 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-b9df4"] Mar 13 16:44:47 crc kubenswrapper[4907]: I0313 16:44:47.815721 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b1428c7e-574e-4fa0-a364-14a6c85d413a" path="/var/lib/kubelet/pods/b1428c7e-574e-4fa0-a364-14a6c85d413a/volumes" Mar 13 16:45:00 crc kubenswrapper[4907]: I0313 16:45:00.158209 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29557005-4k95r"] Mar 13 16:45:00 crc kubenswrapper[4907]: E0313 16:45:00.159146 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1428c7e-574e-4fa0-a364-14a6c85d413a" containerName="registry-server" Mar 13 16:45:00 crc kubenswrapper[4907]: I0313 16:45:00.159161 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1428c7e-574e-4fa0-a364-14a6c85d413a" containerName="registry-server" Mar 13 16:45:00 crc kubenswrapper[4907]: E0313 16:45:00.159176 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1428c7e-574e-4fa0-a364-14a6c85d413a" containerName="extract-content" Mar 13 16:45:00 crc kubenswrapper[4907]: I0313 16:45:00.159185 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1428c7e-574e-4fa0-a364-14a6c85d413a" containerName="extract-content" Mar 13 16:45:00 crc kubenswrapper[4907]: E0313 16:45:00.159212 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1428c7e-574e-4fa0-a364-14a6c85d413a" containerName="extract-utilities" Mar 13 16:45:00 crc kubenswrapper[4907]: I0313 16:45:00.159221 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1428c7e-574e-4fa0-a364-14a6c85d413a" containerName="extract-utilities" Mar 13 16:45:00 crc kubenswrapper[4907]: I0313 16:45:00.159497 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="b1428c7e-574e-4fa0-a364-14a6c85d413a" containerName="registry-server" Mar 13 16:45:00 crc kubenswrapper[4907]: I0313 16:45:00.160471 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29557005-4k95r" Mar 13 16:45:00 crc kubenswrapper[4907]: I0313 16:45:00.163846 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 13 16:45:00 crc kubenswrapper[4907]: I0313 16:45:00.166630 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 13 16:45:00 crc kubenswrapper[4907]: I0313 16:45:00.171977 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29557005-4k95r"] Mar 13 16:45:00 crc kubenswrapper[4907]: I0313 16:45:00.308726 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5a94884c-c438-4536-baa4-fefad79add1a-secret-volume\") pod \"collect-profiles-29557005-4k95r\" (UID: \"5a94884c-c438-4536-baa4-fefad79add1a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29557005-4k95r" Mar 13 16:45:00 crc kubenswrapper[4907]: I0313 16:45:00.309435 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5a94884c-c438-4536-baa4-fefad79add1a-config-volume\") pod \"collect-profiles-29557005-4k95r\" (UID: \"5a94884c-c438-4536-baa4-fefad79add1a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29557005-4k95r" Mar 13 16:45:00 crc kubenswrapper[4907]: I0313 16:45:00.309512 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z5gnq\" (UniqueName: \"kubernetes.io/projected/5a94884c-c438-4536-baa4-fefad79add1a-kube-api-access-z5gnq\") pod \"collect-profiles-29557005-4k95r\" (UID: \"5a94884c-c438-4536-baa4-fefad79add1a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29557005-4k95r" Mar 13 16:45:00 crc kubenswrapper[4907]: I0313 16:45:00.411483 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5a94884c-c438-4536-baa4-fefad79add1a-secret-volume\") pod \"collect-profiles-29557005-4k95r\" (UID: \"5a94884c-c438-4536-baa4-fefad79add1a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29557005-4k95r" Mar 13 16:45:00 crc kubenswrapper[4907]: I0313 16:45:00.412059 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5a94884c-c438-4536-baa4-fefad79add1a-config-volume\") pod \"collect-profiles-29557005-4k95r\" (UID: \"5a94884c-c438-4536-baa4-fefad79add1a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29557005-4k95r" Mar 13 16:45:00 crc kubenswrapper[4907]: I0313 16:45:00.412305 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z5gnq\" (UniqueName: \"kubernetes.io/projected/5a94884c-c438-4536-baa4-fefad79add1a-kube-api-access-z5gnq\") pod \"collect-profiles-29557005-4k95r\" (UID: \"5a94884c-c438-4536-baa4-fefad79add1a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29557005-4k95r" Mar 13 16:45:00 crc kubenswrapper[4907]: I0313 16:45:00.412857 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5a94884c-c438-4536-baa4-fefad79add1a-config-volume\") pod \"collect-profiles-29557005-4k95r\" (UID: \"5a94884c-c438-4536-baa4-fefad79add1a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29557005-4k95r" Mar 13 16:45:00 crc kubenswrapper[4907]: I0313 16:45:00.421006 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5a94884c-c438-4536-baa4-fefad79add1a-secret-volume\") pod \"collect-profiles-29557005-4k95r\" (UID: \"5a94884c-c438-4536-baa4-fefad79add1a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29557005-4k95r" Mar 13 16:45:00 crc kubenswrapper[4907]: I0313 16:45:00.431816 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z5gnq\" (UniqueName: \"kubernetes.io/projected/5a94884c-c438-4536-baa4-fefad79add1a-kube-api-access-z5gnq\") pod \"collect-profiles-29557005-4k95r\" (UID: \"5a94884c-c438-4536-baa4-fefad79add1a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29557005-4k95r" Mar 13 16:45:00 crc kubenswrapper[4907]: I0313 16:45:00.481064 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29557005-4k95r" Mar 13 16:45:01 crc kubenswrapper[4907]: I0313 16:45:01.001127 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29557005-4k95r"] Mar 13 16:45:01 crc kubenswrapper[4907]: I0313 16:45:01.145767 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29557005-4k95r" event={"ID":"5a94884c-c438-4536-baa4-fefad79add1a","Type":"ContainerStarted","Data":"1dea5c9209ab95881c1b9078c066ab6d0957cdb5d7b2ac2325232d2ef8b8cc41"} Mar 13 16:45:02 crc kubenswrapper[4907]: I0313 16:45:02.156537 4907 generic.go:334] "Generic (PLEG): container finished" podID="5a94884c-c438-4536-baa4-fefad79add1a" containerID="5f6d7d4723c79a2c85bc8ce346672905a252464fdb7c28aa0ce2166188a31bd1" exitCode=0 Mar 13 16:45:02 crc kubenswrapper[4907]: I0313 16:45:02.156654 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29557005-4k95r" event={"ID":"5a94884c-c438-4536-baa4-fefad79add1a","Type":"ContainerDied","Data":"5f6d7d4723c79a2c85bc8ce346672905a252464fdb7c28aa0ce2166188a31bd1"} Mar 13 16:45:03 crc kubenswrapper[4907]: I0313 16:45:03.629646 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29557005-4k95r" Mar 13 16:45:03 crc kubenswrapper[4907]: I0313 16:45:03.705090 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z5gnq\" (UniqueName: \"kubernetes.io/projected/5a94884c-c438-4536-baa4-fefad79add1a-kube-api-access-z5gnq\") pod \"5a94884c-c438-4536-baa4-fefad79add1a\" (UID: \"5a94884c-c438-4536-baa4-fefad79add1a\") " Mar 13 16:45:03 crc kubenswrapper[4907]: I0313 16:45:03.705274 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5a94884c-c438-4536-baa4-fefad79add1a-config-volume\") pod \"5a94884c-c438-4536-baa4-fefad79add1a\" (UID: \"5a94884c-c438-4536-baa4-fefad79add1a\") " Mar 13 16:45:03 crc kubenswrapper[4907]: I0313 16:45:03.705967 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5a94884c-c438-4536-baa4-fefad79add1a-config-volume" (OuterVolumeSpecName: "config-volume") pod "5a94884c-c438-4536-baa4-fefad79add1a" (UID: "5a94884c-c438-4536-baa4-fefad79add1a"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 16:45:03 crc kubenswrapper[4907]: I0313 16:45:03.706052 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5a94884c-c438-4536-baa4-fefad79add1a-secret-volume\") pod \"5a94884c-c438-4536-baa4-fefad79add1a\" (UID: \"5a94884c-c438-4536-baa4-fefad79add1a\") " Mar 13 16:45:03 crc kubenswrapper[4907]: I0313 16:45:03.706517 4907 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5a94884c-c438-4536-baa4-fefad79add1a-config-volume\") on node \"crc\" DevicePath \"\"" Mar 13 16:45:03 crc kubenswrapper[4907]: I0313 16:45:03.712856 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5a94884c-c438-4536-baa4-fefad79add1a-kube-api-access-z5gnq" (OuterVolumeSpecName: "kube-api-access-z5gnq") pod "5a94884c-c438-4536-baa4-fefad79add1a" (UID: "5a94884c-c438-4536-baa4-fefad79add1a"). InnerVolumeSpecName "kube-api-access-z5gnq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:45:03 crc kubenswrapper[4907]: I0313 16:45:03.713413 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a94884c-c438-4536-baa4-fefad79add1a-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "5a94884c-c438-4536-baa4-fefad79add1a" (UID: "5a94884c-c438-4536-baa4-fefad79add1a"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:45:03 crc kubenswrapper[4907]: I0313 16:45:03.815506 4907 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5a94884c-c438-4536-baa4-fefad79add1a-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 13 16:45:03 crc kubenswrapper[4907]: I0313 16:45:03.815539 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z5gnq\" (UniqueName: \"kubernetes.io/projected/5a94884c-c438-4536-baa4-fefad79add1a-kube-api-access-z5gnq\") on node \"crc\" DevicePath \"\"" Mar 13 16:45:04 crc kubenswrapper[4907]: I0313 16:45:04.183299 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29557005-4k95r" event={"ID":"5a94884c-c438-4536-baa4-fefad79add1a","Type":"ContainerDied","Data":"1dea5c9209ab95881c1b9078c066ab6d0957cdb5d7b2ac2325232d2ef8b8cc41"} Mar 13 16:45:04 crc kubenswrapper[4907]: I0313 16:45:04.183564 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1dea5c9209ab95881c1b9078c066ab6d0957cdb5d7b2ac2325232d2ef8b8cc41" Mar 13 16:45:04 crc kubenswrapper[4907]: I0313 16:45:04.183380 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29557005-4k95r" Mar 13 16:45:04 crc kubenswrapper[4907]: I0313 16:45:04.741059 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556960-mkbxd"] Mar 13 16:45:04 crc kubenswrapper[4907]: I0313 16:45:04.753856 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556960-mkbxd"] Mar 13 16:45:05 crc kubenswrapper[4907]: I0313 16:45:05.805385 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e1001d4c-6984-4a3c-8d1f-167a6a48ea8a" path="/var/lib/kubelet/pods/e1001d4c-6984-4a3c-8d1f-167a6a48ea8a/volumes" Mar 13 16:45:31 crc kubenswrapper[4907]: I0313 16:45:31.357439 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-wx4sr"] Mar 13 16:45:31 crc kubenswrapper[4907]: E0313 16:45:31.359935 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a94884c-c438-4536-baa4-fefad79add1a" containerName="collect-profiles" Mar 13 16:45:31 crc kubenswrapper[4907]: I0313 16:45:31.359971 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a94884c-c438-4536-baa4-fefad79add1a" containerName="collect-profiles" Mar 13 16:45:31 crc kubenswrapper[4907]: I0313 16:45:31.360470 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a94884c-c438-4536-baa4-fefad79add1a" containerName="collect-profiles" Mar 13 16:45:31 crc kubenswrapper[4907]: I0313 16:45:31.364068 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wx4sr" Mar 13 16:45:31 crc kubenswrapper[4907]: I0313 16:45:31.396721 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-wx4sr"] Mar 13 16:45:31 crc kubenswrapper[4907]: I0313 16:45:31.440188 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/864d645e-16d7-4d3b-82a1-d763ee19600d-catalog-content\") pod \"community-operators-wx4sr\" (UID: \"864d645e-16d7-4d3b-82a1-d763ee19600d\") " pod="openshift-marketplace/community-operators-wx4sr" Mar 13 16:45:31 crc kubenswrapper[4907]: I0313 16:45:31.440269 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/864d645e-16d7-4d3b-82a1-d763ee19600d-utilities\") pod \"community-operators-wx4sr\" (UID: \"864d645e-16d7-4d3b-82a1-d763ee19600d\") " pod="openshift-marketplace/community-operators-wx4sr" Mar 13 16:45:31 crc kubenswrapper[4907]: I0313 16:45:31.440585 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hf9w7\" (UniqueName: \"kubernetes.io/projected/864d645e-16d7-4d3b-82a1-d763ee19600d-kube-api-access-hf9w7\") pod \"community-operators-wx4sr\" (UID: \"864d645e-16d7-4d3b-82a1-d763ee19600d\") " pod="openshift-marketplace/community-operators-wx4sr" Mar 13 16:45:31 crc kubenswrapper[4907]: I0313 16:45:31.543833 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hf9w7\" (UniqueName: \"kubernetes.io/projected/864d645e-16d7-4d3b-82a1-d763ee19600d-kube-api-access-hf9w7\") pod \"community-operators-wx4sr\" (UID: \"864d645e-16d7-4d3b-82a1-d763ee19600d\") " pod="openshift-marketplace/community-operators-wx4sr" Mar 13 16:45:31 crc kubenswrapper[4907]: I0313 16:45:31.543933 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/864d645e-16d7-4d3b-82a1-d763ee19600d-catalog-content\") pod \"community-operators-wx4sr\" (UID: \"864d645e-16d7-4d3b-82a1-d763ee19600d\") " pod="openshift-marketplace/community-operators-wx4sr" Mar 13 16:45:31 crc kubenswrapper[4907]: I0313 16:45:31.543974 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/864d645e-16d7-4d3b-82a1-d763ee19600d-utilities\") pod \"community-operators-wx4sr\" (UID: \"864d645e-16d7-4d3b-82a1-d763ee19600d\") " pod="openshift-marketplace/community-operators-wx4sr" Mar 13 16:45:31 crc kubenswrapper[4907]: I0313 16:45:31.544846 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/864d645e-16d7-4d3b-82a1-d763ee19600d-utilities\") pod \"community-operators-wx4sr\" (UID: \"864d645e-16d7-4d3b-82a1-d763ee19600d\") " pod="openshift-marketplace/community-operators-wx4sr" Mar 13 16:45:31 crc kubenswrapper[4907]: I0313 16:45:31.544843 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/864d645e-16d7-4d3b-82a1-d763ee19600d-catalog-content\") pod \"community-operators-wx4sr\" (UID: \"864d645e-16d7-4d3b-82a1-d763ee19600d\") " pod="openshift-marketplace/community-operators-wx4sr" Mar 13 16:45:31 crc kubenswrapper[4907]: I0313 16:45:31.572681 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hf9w7\" (UniqueName: \"kubernetes.io/projected/864d645e-16d7-4d3b-82a1-d763ee19600d-kube-api-access-hf9w7\") pod \"community-operators-wx4sr\" (UID: \"864d645e-16d7-4d3b-82a1-d763ee19600d\") " pod="openshift-marketplace/community-operators-wx4sr" Mar 13 16:45:31 crc kubenswrapper[4907]: I0313 16:45:31.725072 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wx4sr" Mar 13 16:45:32 crc kubenswrapper[4907]: I0313 16:45:32.319245 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-wx4sr"] Mar 13 16:45:32 crc kubenswrapper[4907]: W0313 16:45:32.324104 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod864d645e_16d7_4d3b_82a1_d763ee19600d.slice/crio-a307370b5b3ddfe67ecb5180a2f6aea9904a507dc0c7d87ae8a5138047252a89 WatchSource:0}: Error finding container a307370b5b3ddfe67ecb5180a2f6aea9904a507dc0c7d87ae8a5138047252a89: Status 404 returned error can't find the container with id a307370b5b3ddfe67ecb5180a2f6aea9904a507dc0c7d87ae8a5138047252a89 Mar 13 16:45:32 crc kubenswrapper[4907]: I0313 16:45:32.577449 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wx4sr" event={"ID":"864d645e-16d7-4d3b-82a1-d763ee19600d","Type":"ContainerStarted","Data":"ecb06530e4c6cf8e5865dec7c587d6387a6ebfa7dea5833ab58b5054e66656b0"} Mar 13 16:45:32 crc kubenswrapper[4907]: I0313 16:45:32.577524 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wx4sr" event={"ID":"864d645e-16d7-4d3b-82a1-d763ee19600d","Type":"ContainerStarted","Data":"a307370b5b3ddfe67ecb5180a2f6aea9904a507dc0c7d87ae8a5138047252a89"} Mar 13 16:45:33 crc kubenswrapper[4907]: I0313 16:45:33.592454 4907 generic.go:334] "Generic (PLEG): container finished" podID="864d645e-16d7-4d3b-82a1-d763ee19600d" containerID="ecb06530e4c6cf8e5865dec7c587d6387a6ebfa7dea5833ab58b5054e66656b0" exitCode=0 Mar 13 16:45:33 crc kubenswrapper[4907]: I0313 16:45:33.592578 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wx4sr" event={"ID":"864d645e-16d7-4d3b-82a1-d763ee19600d","Type":"ContainerDied","Data":"ecb06530e4c6cf8e5865dec7c587d6387a6ebfa7dea5833ab58b5054e66656b0"} Mar 13 16:45:33 crc kubenswrapper[4907]: I0313 16:45:33.596752 4907 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 13 16:45:35 crc kubenswrapper[4907]: I0313 16:45:35.632300 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wx4sr" event={"ID":"864d645e-16d7-4d3b-82a1-d763ee19600d","Type":"ContainerStarted","Data":"38540a9b7efd6bb2a5b35b5d21a8bacf08f2ab0da9a4f992cf84f289f305c6c1"} Mar 13 16:45:37 crc kubenswrapper[4907]: I0313 16:45:37.656512 4907 generic.go:334] "Generic (PLEG): container finished" podID="864d645e-16d7-4d3b-82a1-d763ee19600d" containerID="38540a9b7efd6bb2a5b35b5d21a8bacf08f2ab0da9a4f992cf84f289f305c6c1" exitCode=0 Mar 13 16:45:37 crc kubenswrapper[4907]: I0313 16:45:37.656560 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wx4sr" event={"ID":"864d645e-16d7-4d3b-82a1-d763ee19600d","Type":"ContainerDied","Data":"38540a9b7efd6bb2a5b35b5d21a8bacf08f2ab0da9a4f992cf84f289f305c6c1"} Mar 13 16:45:38 crc kubenswrapper[4907]: I0313 16:45:38.671640 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wx4sr" event={"ID":"864d645e-16d7-4d3b-82a1-d763ee19600d","Type":"ContainerStarted","Data":"4a312a3cf18fab398fbf6c5eaa5295284d64d484a558188758c899a3134be59b"} Mar 13 16:45:38 crc kubenswrapper[4907]: I0313 16:45:38.708116 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-wx4sr" podStartSLOduration=3.244016488 podStartE2EDuration="7.708090076s" podCreationTimestamp="2026-03-13 16:45:31 +0000 UTC" firstStartedPulling="2026-03-13 16:45:33.596477629 +0000 UTC m=+9632.496265328" lastFinishedPulling="2026-03-13 16:45:38.060551207 +0000 UTC m=+9636.960338916" observedRunningTime="2026-03-13 16:45:38.694124943 +0000 UTC m=+9637.593912652" watchObservedRunningTime="2026-03-13 16:45:38.708090076 +0000 UTC m=+9637.607877785" Mar 13 16:45:41 crc kubenswrapper[4907]: I0313 16:45:41.832967 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-wx4sr" Mar 13 16:45:41 crc kubenswrapper[4907]: I0313 16:45:41.833580 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-wx4sr" Mar 13 16:45:42 crc kubenswrapper[4907]: I0313 16:45:42.866268 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-wx4sr" podUID="864d645e-16d7-4d3b-82a1-d763ee19600d" containerName="registry-server" probeResult="failure" output=< Mar 13 16:45:42 crc kubenswrapper[4907]: timeout: failed to connect service ":50051" within 1s Mar 13 16:45:42 crc kubenswrapper[4907]: > Mar 13 16:45:44 crc kubenswrapper[4907]: I0313 16:45:44.772583 4907 scope.go:117] "RemoveContainer" containerID="0c0e8b2fe415ff36180c0f0cbc79e145510c4e8b607c452be50ff3ed2394d856" Mar 13 16:45:51 crc kubenswrapper[4907]: I0313 16:45:51.804479 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-wx4sr" Mar 13 16:45:51 crc kubenswrapper[4907]: I0313 16:45:51.887268 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-wx4sr" Mar 13 16:45:52 crc kubenswrapper[4907]: I0313 16:45:52.397101 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-wx4sr"] Mar 13 16:45:52 crc kubenswrapper[4907]: I0313 16:45:52.836983 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-wx4sr" podUID="864d645e-16d7-4d3b-82a1-d763ee19600d" containerName="registry-server" containerID="cri-o://4a312a3cf18fab398fbf6c5eaa5295284d64d484a558188758c899a3134be59b" gracePeriod=2 Mar 13 16:45:53 crc kubenswrapper[4907]: I0313 16:45:53.445020 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wx4sr" Mar 13 16:45:53 crc kubenswrapper[4907]: I0313 16:45:53.554070 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hf9w7\" (UniqueName: \"kubernetes.io/projected/864d645e-16d7-4d3b-82a1-d763ee19600d-kube-api-access-hf9w7\") pod \"864d645e-16d7-4d3b-82a1-d763ee19600d\" (UID: \"864d645e-16d7-4d3b-82a1-d763ee19600d\") " Mar 13 16:45:53 crc kubenswrapper[4907]: I0313 16:45:53.554405 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/864d645e-16d7-4d3b-82a1-d763ee19600d-catalog-content\") pod \"864d645e-16d7-4d3b-82a1-d763ee19600d\" (UID: \"864d645e-16d7-4d3b-82a1-d763ee19600d\") " Mar 13 16:45:53 crc kubenswrapper[4907]: I0313 16:45:53.554848 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/864d645e-16d7-4d3b-82a1-d763ee19600d-utilities\") pod \"864d645e-16d7-4d3b-82a1-d763ee19600d\" (UID: \"864d645e-16d7-4d3b-82a1-d763ee19600d\") " Mar 13 16:45:53 crc kubenswrapper[4907]: I0313 16:45:53.555414 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/864d645e-16d7-4d3b-82a1-d763ee19600d-utilities" (OuterVolumeSpecName: "utilities") pod "864d645e-16d7-4d3b-82a1-d763ee19600d" (UID: "864d645e-16d7-4d3b-82a1-d763ee19600d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 16:45:53 crc kubenswrapper[4907]: I0313 16:45:53.555800 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/864d645e-16d7-4d3b-82a1-d763ee19600d-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 16:45:53 crc kubenswrapper[4907]: I0313 16:45:53.561188 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/864d645e-16d7-4d3b-82a1-d763ee19600d-kube-api-access-hf9w7" (OuterVolumeSpecName: "kube-api-access-hf9w7") pod "864d645e-16d7-4d3b-82a1-d763ee19600d" (UID: "864d645e-16d7-4d3b-82a1-d763ee19600d"). InnerVolumeSpecName "kube-api-access-hf9w7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:45:53 crc kubenswrapper[4907]: I0313 16:45:53.613196 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/864d645e-16d7-4d3b-82a1-d763ee19600d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "864d645e-16d7-4d3b-82a1-d763ee19600d" (UID: "864d645e-16d7-4d3b-82a1-d763ee19600d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 16:45:53 crc kubenswrapper[4907]: I0313 16:45:53.657598 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hf9w7\" (UniqueName: \"kubernetes.io/projected/864d645e-16d7-4d3b-82a1-d763ee19600d-kube-api-access-hf9w7\") on node \"crc\" DevicePath \"\"" Mar 13 16:45:53 crc kubenswrapper[4907]: I0313 16:45:53.657629 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/864d645e-16d7-4d3b-82a1-d763ee19600d-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 16:45:53 crc kubenswrapper[4907]: I0313 16:45:53.856723 4907 generic.go:334] "Generic (PLEG): container finished" podID="864d645e-16d7-4d3b-82a1-d763ee19600d" containerID="4a312a3cf18fab398fbf6c5eaa5295284d64d484a558188758c899a3134be59b" exitCode=0 Mar 13 16:45:53 crc kubenswrapper[4907]: I0313 16:45:53.856764 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wx4sr" event={"ID":"864d645e-16d7-4d3b-82a1-d763ee19600d","Type":"ContainerDied","Data":"4a312a3cf18fab398fbf6c5eaa5295284d64d484a558188758c899a3134be59b"} Mar 13 16:45:53 crc kubenswrapper[4907]: I0313 16:45:53.856792 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wx4sr" event={"ID":"864d645e-16d7-4d3b-82a1-d763ee19600d","Type":"ContainerDied","Data":"a307370b5b3ddfe67ecb5180a2f6aea9904a507dc0c7d87ae8a5138047252a89"} Mar 13 16:45:53 crc kubenswrapper[4907]: I0313 16:45:53.856812 4907 scope.go:117] "RemoveContainer" containerID="4a312a3cf18fab398fbf6c5eaa5295284d64d484a558188758c899a3134be59b" Mar 13 16:45:53 crc kubenswrapper[4907]: I0313 16:45:53.856918 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wx4sr" Mar 13 16:45:53 crc kubenswrapper[4907]: I0313 16:45:53.895271 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-wx4sr"] Mar 13 16:45:53 crc kubenswrapper[4907]: I0313 16:45:53.899072 4907 scope.go:117] "RemoveContainer" containerID="38540a9b7efd6bb2a5b35b5d21a8bacf08f2ab0da9a4f992cf84f289f305c6c1" Mar 13 16:45:53 crc kubenswrapper[4907]: I0313 16:45:53.904083 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-wx4sr"] Mar 13 16:45:53 crc kubenswrapper[4907]: I0313 16:45:53.937572 4907 scope.go:117] "RemoveContainer" containerID="ecb06530e4c6cf8e5865dec7c587d6387a6ebfa7dea5833ab58b5054e66656b0" Mar 13 16:45:53 crc kubenswrapper[4907]: I0313 16:45:53.978492 4907 scope.go:117] "RemoveContainer" containerID="4a312a3cf18fab398fbf6c5eaa5295284d64d484a558188758c899a3134be59b" Mar 13 16:45:53 crc kubenswrapper[4907]: E0313 16:45:53.979107 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4a312a3cf18fab398fbf6c5eaa5295284d64d484a558188758c899a3134be59b\": container with ID starting with 4a312a3cf18fab398fbf6c5eaa5295284d64d484a558188758c899a3134be59b not found: ID does not exist" containerID="4a312a3cf18fab398fbf6c5eaa5295284d64d484a558188758c899a3134be59b" Mar 13 16:45:53 crc kubenswrapper[4907]: I0313 16:45:53.979158 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a312a3cf18fab398fbf6c5eaa5295284d64d484a558188758c899a3134be59b"} err="failed to get container status \"4a312a3cf18fab398fbf6c5eaa5295284d64d484a558188758c899a3134be59b\": rpc error: code = NotFound desc = could not find container \"4a312a3cf18fab398fbf6c5eaa5295284d64d484a558188758c899a3134be59b\": container with ID starting with 4a312a3cf18fab398fbf6c5eaa5295284d64d484a558188758c899a3134be59b not found: ID does not exist" Mar 13 16:45:53 crc kubenswrapper[4907]: I0313 16:45:53.979194 4907 scope.go:117] "RemoveContainer" containerID="38540a9b7efd6bb2a5b35b5d21a8bacf08f2ab0da9a4f992cf84f289f305c6c1" Mar 13 16:45:53 crc kubenswrapper[4907]: E0313 16:45:53.979587 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"38540a9b7efd6bb2a5b35b5d21a8bacf08f2ab0da9a4f992cf84f289f305c6c1\": container with ID starting with 38540a9b7efd6bb2a5b35b5d21a8bacf08f2ab0da9a4f992cf84f289f305c6c1 not found: ID does not exist" containerID="38540a9b7efd6bb2a5b35b5d21a8bacf08f2ab0da9a4f992cf84f289f305c6c1" Mar 13 16:45:53 crc kubenswrapper[4907]: I0313 16:45:53.979627 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"38540a9b7efd6bb2a5b35b5d21a8bacf08f2ab0da9a4f992cf84f289f305c6c1"} err="failed to get container status \"38540a9b7efd6bb2a5b35b5d21a8bacf08f2ab0da9a4f992cf84f289f305c6c1\": rpc error: code = NotFound desc = could not find container \"38540a9b7efd6bb2a5b35b5d21a8bacf08f2ab0da9a4f992cf84f289f305c6c1\": container with ID starting with 38540a9b7efd6bb2a5b35b5d21a8bacf08f2ab0da9a4f992cf84f289f305c6c1 not found: ID does not exist" Mar 13 16:45:53 crc kubenswrapper[4907]: I0313 16:45:53.979653 4907 scope.go:117] "RemoveContainer" containerID="ecb06530e4c6cf8e5865dec7c587d6387a6ebfa7dea5833ab58b5054e66656b0" Mar 13 16:45:53 crc kubenswrapper[4907]: E0313 16:45:53.980001 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ecb06530e4c6cf8e5865dec7c587d6387a6ebfa7dea5833ab58b5054e66656b0\": container with ID starting with ecb06530e4c6cf8e5865dec7c587d6387a6ebfa7dea5833ab58b5054e66656b0 not found: ID does not exist" containerID="ecb06530e4c6cf8e5865dec7c587d6387a6ebfa7dea5833ab58b5054e66656b0" Mar 13 16:45:53 crc kubenswrapper[4907]: I0313 16:45:53.980037 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ecb06530e4c6cf8e5865dec7c587d6387a6ebfa7dea5833ab58b5054e66656b0"} err="failed to get container status \"ecb06530e4c6cf8e5865dec7c587d6387a6ebfa7dea5833ab58b5054e66656b0\": rpc error: code = NotFound desc = could not find container \"ecb06530e4c6cf8e5865dec7c587d6387a6ebfa7dea5833ab58b5054e66656b0\": container with ID starting with ecb06530e4c6cf8e5865dec7c587d6387a6ebfa7dea5833ab58b5054e66656b0 not found: ID does not exist" Mar 13 16:45:55 crc kubenswrapper[4907]: I0313 16:45:55.799367 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="864d645e-16d7-4d3b-82a1-d763ee19600d" path="/var/lib/kubelet/pods/864d645e-16d7-4d3b-82a1-d763ee19600d/volumes" Mar 13 16:46:00 crc kubenswrapper[4907]: I0313 16:46:00.198596 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29557006-ck2rj"] Mar 13 16:46:00 crc kubenswrapper[4907]: E0313 16:46:00.199692 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="864d645e-16d7-4d3b-82a1-d763ee19600d" containerName="extract-content" Mar 13 16:46:00 crc kubenswrapper[4907]: I0313 16:46:00.199707 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="864d645e-16d7-4d3b-82a1-d763ee19600d" containerName="extract-content" Mar 13 16:46:00 crc kubenswrapper[4907]: E0313 16:46:00.199731 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="864d645e-16d7-4d3b-82a1-d763ee19600d" containerName="extract-utilities" Mar 13 16:46:00 crc kubenswrapper[4907]: I0313 16:46:00.199739 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="864d645e-16d7-4d3b-82a1-d763ee19600d" containerName="extract-utilities" Mar 13 16:46:00 crc kubenswrapper[4907]: E0313 16:46:00.199748 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="864d645e-16d7-4d3b-82a1-d763ee19600d" containerName="registry-server" Mar 13 16:46:00 crc kubenswrapper[4907]: I0313 16:46:00.199758 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="864d645e-16d7-4d3b-82a1-d763ee19600d" containerName="registry-server" Mar 13 16:46:00 crc kubenswrapper[4907]: I0313 16:46:00.200046 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="864d645e-16d7-4d3b-82a1-d763ee19600d" containerName="registry-server" Mar 13 16:46:00 crc kubenswrapper[4907]: I0313 16:46:00.200917 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29557006-ck2rj"] Mar 13 16:46:00 crc kubenswrapper[4907]: I0313 16:46:00.201001 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29557006-ck2rj" Mar 13 16:46:00 crc kubenswrapper[4907]: I0313 16:46:00.206711 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 16:46:00 crc kubenswrapper[4907]: I0313 16:46:00.208394 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 16:46:00 crc kubenswrapper[4907]: I0313 16:46:00.208779 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 16:46:00 crc kubenswrapper[4907]: I0313 16:46:00.339523 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r9swb\" (UniqueName: \"kubernetes.io/projected/0690b5f5-ba57-4fa5-baf4-86a729917817-kube-api-access-r9swb\") pod \"auto-csr-approver-29557006-ck2rj\" (UID: \"0690b5f5-ba57-4fa5-baf4-86a729917817\") " pod="openshift-infra/auto-csr-approver-29557006-ck2rj" Mar 13 16:46:00 crc kubenswrapper[4907]: I0313 16:46:00.442692 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r9swb\" (UniqueName: \"kubernetes.io/projected/0690b5f5-ba57-4fa5-baf4-86a729917817-kube-api-access-r9swb\") pod \"auto-csr-approver-29557006-ck2rj\" (UID: \"0690b5f5-ba57-4fa5-baf4-86a729917817\") " pod="openshift-infra/auto-csr-approver-29557006-ck2rj" Mar 13 16:46:00 crc kubenswrapper[4907]: I0313 16:46:00.473120 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r9swb\" (UniqueName: \"kubernetes.io/projected/0690b5f5-ba57-4fa5-baf4-86a729917817-kube-api-access-r9swb\") pod \"auto-csr-approver-29557006-ck2rj\" (UID: \"0690b5f5-ba57-4fa5-baf4-86a729917817\") " pod="openshift-infra/auto-csr-approver-29557006-ck2rj" Mar 13 16:46:00 crc kubenswrapper[4907]: I0313 16:46:00.528163 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29557006-ck2rj" Mar 13 16:46:01 crc kubenswrapper[4907]: I0313 16:46:01.027159 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29557006-ck2rj"] Mar 13 16:46:01 crc kubenswrapper[4907]: I0313 16:46:01.973296 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29557006-ck2rj" event={"ID":"0690b5f5-ba57-4fa5-baf4-86a729917817","Type":"ContainerStarted","Data":"d7f0cf0bc5d44cbbf4ef528b69bdcaff5527f17372aaf7b2a0cab117b0fe4f68"} Mar 13 16:46:02 crc kubenswrapper[4907]: I0313 16:46:02.989794 4907 generic.go:334] "Generic (PLEG): container finished" podID="0690b5f5-ba57-4fa5-baf4-86a729917817" containerID="29e87ba308a2bea52f09c17c0673a7492d85691ad7ee7034046f4e32cf2bff88" exitCode=0 Mar 13 16:46:02 crc kubenswrapper[4907]: I0313 16:46:02.989927 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29557006-ck2rj" event={"ID":"0690b5f5-ba57-4fa5-baf4-86a729917817","Type":"ContainerDied","Data":"29e87ba308a2bea52f09c17c0673a7492d85691ad7ee7034046f4e32cf2bff88"} Mar 13 16:46:04 crc kubenswrapper[4907]: I0313 16:46:04.519377 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29557006-ck2rj" Mar 13 16:46:04 crc kubenswrapper[4907]: I0313 16:46:04.675628 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r9swb\" (UniqueName: \"kubernetes.io/projected/0690b5f5-ba57-4fa5-baf4-86a729917817-kube-api-access-r9swb\") pod \"0690b5f5-ba57-4fa5-baf4-86a729917817\" (UID: \"0690b5f5-ba57-4fa5-baf4-86a729917817\") " Mar 13 16:46:04 crc kubenswrapper[4907]: I0313 16:46:04.682109 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0690b5f5-ba57-4fa5-baf4-86a729917817-kube-api-access-r9swb" (OuterVolumeSpecName: "kube-api-access-r9swb") pod "0690b5f5-ba57-4fa5-baf4-86a729917817" (UID: "0690b5f5-ba57-4fa5-baf4-86a729917817"). InnerVolumeSpecName "kube-api-access-r9swb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:46:04 crc kubenswrapper[4907]: I0313 16:46:04.778257 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r9swb\" (UniqueName: \"kubernetes.io/projected/0690b5f5-ba57-4fa5-baf4-86a729917817-kube-api-access-r9swb\") on node \"crc\" DevicePath \"\"" Mar 13 16:46:05 crc kubenswrapper[4907]: I0313 16:46:05.021418 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29557006-ck2rj" event={"ID":"0690b5f5-ba57-4fa5-baf4-86a729917817","Type":"ContainerDied","Data":"d7f0cf0bc5d44cbbf4ef528b69bdcaff5527f17372aaf7b2a0cab117b0fe4f68"} Mar 13 16:46:05 crc kubenswrapper[4907]: I0313 16:46:05.021822 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d7f0cf0bc5d44cbbf4ef528b69bdcaff5527f17372aaf7b2a0cab117b0fe4f68" Mar 13 16:46:05 crc kubenswrapper[4907]: I0313 16:46:05.021518 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29557006-ck2rj" Mar 13 16:46:05 crc kubenswrapper[4907]: I0313 16:46:05.612142 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29557000-b8z24"] Mar 13 16:46:05 crc kubenswrapper[4907]: I0313 16:46:05.622967 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29557000-b8z24"] Mar 13 16:46:05 crc kubenswrapper[4907]: I0313 16:46:05.806320 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ff06fab3-a77c-49ef-8c13-88908b7acf21" path="/var/lib/kubelet/pods/ff06fab3-a77c-49ef-8c13-88908b7acf21/volumes" Mar 13 16:46:18 crc kubenswrapper[4907]: I0313 16:46:18.041708 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 16:46:18 crc kubenswrapper[4907]: I0313 16:46:18.042450 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 16:46:44 crc kubenswrapper[4907]: I0313 16:46:44.852857 4907 scope.go:117] "RemoveContainer" containerID="7edb971c3f915cfb8576dc1860f2e0063b3420e5f584dbd096705d5ce436c71e" Mar 13 16:46:48 crc kubenswrapper[4907]: I0313 16:46:48.042559 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 16:46:48 crc kubenswrapper[4907]: I0313 16:46:48.044835 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 16:47:18 crc kubenswrapper[4907]: I0313 16:47:18.041919 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 16:47:18 crc kubenswrapper[4907]: I0313 16:47:18.042542 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 16:47:18 crc kubenswrapper[4907]: I0313 16:47:18.042610 4907 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" Mar 13 16:47:18 crc kubenswrapper[4907]: I0313 16:47:18.043748 4907 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"cf78e901eff913d33c2f4e21ad1865cb4b0011402563e6e1fc5c1c3c95de10e9"} pod="openshift-machine-config-operator/machine-config-daemon-hm56j" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 16:47:18 crc kubenswrapper[4907]: I0313 16:47:18.043851 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" containerID="cri-o://cf78e901eff913d33c2f4e21ad1865cb4b0011402563e6e1fc5c1c3c95de10e9" gracePeriod=600 Mar 13 16:47:18 crc kubenswrapper[4907]: I0313 16:47:18.976914 4907 generic.go:334] "Generic (PLEG): container finished" podID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerID="cf78e901eff913d33c2f4e21ad1865cb4b0011402563e6e1fc5c1c3c95de10e9" exitCode=0 Mar 13 16:47:18 crc kubenswrapper[4907]: I0313 16:47:18.977002 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerDied","Data":"cf78e901eff913d33c2f4e21ad1865cb4b0011402563e6e1fc5c1c3c95de10e9"} Mar 13 16:47:18 crc kubenswrapper[4907]: I0313 16:47:18.977270 4907 scope.go:117] "RemoveContainer" containerID="7e4d0690f4ac607ea8a268bfa88b5ae6d9a95f8b6013de7747fdffb429f28047" Mar 13 16:47:19 crc kubenswrapper[4907]: I0313 16:47:19.987776 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerStarted","Data":"64bd48e5096c07c25a07873a406a3b7bc89c38906dc6f7c7ace8213663361dfc"} Mar 13 16:48:00 crc kubenswrapper[4907]: I0313 16:48:00.140214 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29557008-tqdp7"] Mar 13 16:48:00 crc kubenswrapper[4907]: E0313 16:48:00.141179 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0690b5f5-ba57-4fa5-baf4-86a729917817" containerName="oc" Mar 13 16:48:00 crc kubenswrapper[4907]: I0313 16:48:00.141194 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="0690b5f5-ba57-4fa5-baf4-86a729917817" containerName="oc" Mar 13 16:48:00 crc kubenswrapper[4907]: I0313 16:48:00.141403 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="0690b5f5-ba57-4fa5-baf4-86a729917817" containerName="oc" Mar 13 16:48:00 crc kubenswrapper[4907]: I0313 16:48:00.142230 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29557008-tqdp7" Mar 13 16:48:00 crc kubenswrapper[4907]: I0313 16:48:00.144057 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 16:48:00 crc kubenswrapper[4907]: I0313 16:48:00.144544 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 16:48:00 crc kubenswrapper[4907]: I0313 16:48:00.144698 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 16:48:00 crc kubenswrapper[4907]: I0313 16:48:00.150852 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29557008-tqdp7"] Mar 13 16:48:00 crc kubenswrapper[4907]: I0313 16:48:00.220011 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sjbk4\" (UniqueName: \"kubernetes.io/projected/168ac198-4031-4bf7-bcf6-eb190b28dce6-kube-api-access-sjbk4\") pod \"auto-csr-approver-29557008-tqdp7\" (UID: \"168ac198-4031-4bf7-bcf6-eb190b28dce6\") " pod="openshift-infra/auto-csr-approver-29557008-tqdp7" Mar 13 16:48:00 crc kubenswrapper[4907]: I0313 16:48:00.322409 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sjbk4\" (UniqueName: \"kubernetes.io/projected/168ac198-4031-4bf7-bcf6-eb190b28dce6-kube-api-access-sjbk4\") pod \"auto-csr-approver-29557008-tqdp7\" (UID: \"168ac198-4031-4bf7-bcf6-eb190b28dce6\") " pod="openshift-infra/auto-csr-approver-29557008-tqdp7" Mar 13 16:48:00 crc kubenswrapper[4907]: I0313 16:48:00.343400 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sjbk4\" (UniqueName: \"kubernetes.io/projected/168ac198-4031-4bf7-bcf6-eb190b28dce6-kube-api-access-sjbk4\") pod \"auto-csr-approver-29557008-tqdp7\" (UID: \"168ac198-4031-4bf7-bcf6-eb190b28dce6\") " pod="openshift-infra/auto-csr-approver-29557008-tqdp7" Mar 13 16:48:00 crc kubenswrapper[4907]: I0313 16:48:00.459915 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29557008-tqdp7" Mar 13 16:48:00 crc kubenswrapper[4907]: I0313 16:48:00.923496 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29557008-tqdp7"] Mar 13 16:48:01 crc kubenswrapper[4907]: I0313 16:48:01.608582 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29557008-tqdp7" event={"ID":"168ac198-4031-4bf7-bcf6-eb190b28dce6","Type":"ContainerStarted","Data":"7b6ed12b3c6ca6c521826924c61118611556bd303b2b9c0e550e05e23de1591e"} Mar 13 16:48:03 crc kubenswrapper[4907]: I0313 16:48:03.627493 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29557008-tqdp7" event={"ID":"168ac198-4031-4bf7-bcf6-eb190b28dce6","Type":"ContainerStarted","Data":"feb18e9c5294082e8034bc02e58f3c7d0b3c8a71601de1543217f3616b52ba31"} Mar 13 16:48:03 crc kubenswrapper[4907]: I0313 16:48:03.647486 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29557008-tqdp7" podStartSLOduration=1.512070255 podStartE2EDuration="3.647467128s" podCreationTimestamp="2026-03-13 16:48:00 +0000 UTC" firstStartedPulling="2026-03-13 16:48:00.942505104 +0000 UTC m=+9779.842292833" lastFinishedPulling="2026-03-13 16:48:03.077902017 +0000 UTC m=+9781.977689706" observedRunningTime="2026-03-13 16:48:03.642917443 +0000 UTC m=+9782.542705142" watchObservedRunningTime="2026-03-13 16:48:03.647467128 +0000 UTC m=+9782.547254817" Mar 13 16:48:04 crc kubenswrapper[4907]: I0313 16:48:04.642266 4907 generic.go:334] "Generic (PLEG): container finished" podID="168ac198-4031-4bf7-bcf6-eb190b28dce6" containerID="feb18e9c5294082e8034bc02e58f3c7d0b3c8a71601de1543217f3616b52ba31" exitCode=0 Mar 13 16:48:04 crc kubenswrapper[4907]: I0313 16:48:04.643640 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29557008-tqdp7" event={"ID":"168ac198-4031-4bf7-bcf6-eb190b28dce6","Type":"ContainerDied","Data":"feb18e9c5294082e8034bc02e58f3c7d0b3c8a71601de1543217f3616b52ba31"} Mar 13 16:48:06 crc kubenswrapper[4907]: I0313 16:48:06.132446 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29557008-tqdp7" Mar 13 16:48:06 crc kubenswrapper[4907]: I0313 16:48:06.253944 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sjbk4\" (UniqueName: \"kubernetes.io/projected/168ac198-4031-4bf7-bcf6-eb190b28dce6-kube-api-access-sjbk4\") pod \"168ac198-4031-4bf7-bcf6-eb190b28dce6\" (UID: \"168ac198-4031-4bf7-bcf6-eb190b28dce6\") " Mar 13 16:48:06 crc kubenswrapper[4907]: I0313 16:48:06.264185 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/168ac198-4031-4bf7-bcf6-eb190b28dce6-kube-api-access-sjbk4" (OuterVolumeSpecName: "kube-api-access-sjbk4") pod "168ac198-4031-4bf7-bcf6-eb190b28dce6" (UID: "168ac198-4031-4bf7-bcf6-eb190b28dce6"). InnerVolumeSpecName "kube-api-access-sjbk4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:48:06 crc kubenswrapper[4907]: I0313 16:48:06.360203 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sjbk4\" (UniqueName: \"kubernetes.io/projected/168ac198-4031-4bf7-bcf6-eb190b28dce6-kube-api-access-sjbk4\") on node \"crc\" DevicePath \"\"" Mar 13 16:48:06 crc kubenswrapper[4907]: I0313 16:48:06.662429 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29557008-tqdp7" event={"ID":"168ac198-4031-4bf7-bcf6-eb190b28dce6","Type":"ContainerDied","Data":"7b6ed12b3c6ca6c521826924c61118611556bd303b2b9c0e550e05e23de1591e"} Mar 13 16:48:06 crc kubenswrapper[4907]: I0313 16:48:06.662704 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7b6ed12b3c6ca6c521826924c61118611556bd303b2b9c0e550e05e23de1591e" Mar 13 16:48:06 crc kubenswrapper[4907]: I0313 16:48:06.662727 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29557008-tqdp7" Mar 13 16:48:06 crc kubenswrapper[4907]: I0313 16:48:06.734063 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29557002-tdngs"] Mar 13 16:48:06 crc kubenswrapper[4907]: I0313 16:48:06.745715 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29557002-tdngs"] Mar 13 16:48:07 crc kubenswrapper[4907]: I0313 16:48:07.800927 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dce83493-bbc5-49ac-a420-229eac93005d" path="/var/lib/kubelet/pods/dce83493-bbc5-49ac-a420-229eac93005d/volumes" Mar 13 16:48:19 crc kubenswrapper[4907]: I0313 16:48:19.088227 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_d29f6de0-3229-477c-86a6-550f2d465592/init-config-reloader/0.log" Mar 13 16:48:19 crc kubenswrapper[4907]: I0313 16:48:19.966844 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_d29f6de0-3229-477c-86a6-550f2d465592/init-config-reloader/0.log" Mar 13 16:48:19 crc kubenswrapper[4907]: I0313 16:48:19.989975 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_d29f6de0-3229-477c-86a6-550f2d465592/alertmanager/0.log" Mar 13 16:48:20 crc kubenswrapper[4907]: I0313 16:48:20.031995 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_d29f6de0-3229-477c-86a6-550f2d465592/config-reloader/0.log" Mar 13 16:48:20 crc kubenswrapper[4907]: I0313 16:48:20.178257 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_8ce53a96-2566-4d5c-b4cd-bbfd2516975a/aodh-api/0.log" Mar 13 16:48:20 crc kubenswrapper[4907]: I0313 16:48:20.239183 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_8ce53a96-2566-4d5c-b4cd-bbfd2516975a/aodh-evaluator/0.log" Mar 13 16:48:20 crc kubenswrapper[4907]: I0313 16:48:20.269061 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_8ce53a96-2566-4d5c-b4cd-bbfd2516975a/aodh-listener/0.log" Mar 13 16:48:20 crc kubenswrapper[4907]: I0313 16:48:20.357724 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_8ce53a96-2566-4d5c-b4cd-bbfd2516975a/aodh-notifier/0.log" Mar 13 16:48:20 crc kubenswrapper[4907]: I0313 16:48:20.454220 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-5564857dbd-vn7xm_d83fca2b-ed84-4516-8510-6e0ff4b2cfef/barbican-api/0.log" Mar 13 16:48:20 crc kubenswrapper[4907]: I0313 16:48:20.517841 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-5564857dbd-vn7xm_d83fca2b-ed84-4516-8510-6e0ff4b2cfef/barbican-api-log/0.log" Mar 13 16:48:20 crc kubenswrapper[4907]: I0313 16:48:20.680628 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-68d9445c58-bjlh2_87b4d8b9-57da-4c61-9dd9-d746fc62db84/barbican-keystone-listener/0.log" Mar 13 16:48:20 crc kubenswrapper[4907]: I0313 16:48:20.701511 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-68d9445c58-bjlh2_87b4d8b9-57da-4c61-9dd9-d746fc62db84/barbican-keystone-listener-log/0.log" Mar 13 16:48:20 crc kubenswrapper[4907]: I0313 16:48:20.939094 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-868b698d7f-wbtl4_3599175f-0bbd-4645-ab11-07e26d7fb98d/barbican-worker-log/0.log" Mar 13 16:48:20 crc kubenswrapper[4907]: I0313 16:48:20.962366 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-868b698d7f-wbtl4_3599175f-0bbd-4645-ab11-07e26d7fb98d/barbican-worker/0.log" Mar 13 16:48:21 crc kubenswrapper[4907]: I0313 16:48:21.063182 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-openstack-openstack-cell1-tb9sw_d7346ccb-6154-42a9-9097-bad966681d81/bootstrap-openstack-openstack-cell1/0.log" Mar 13 16:48:21 crc kubenswrapper[4907]: I0313 16:48:21.313269 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_59fb2cd9-5414-4356-9438-d33f90ad084c/ceilometer-notification-agent/0.log" Mar 13 16:48:21 crc kubenswrapper[4907]: I0313 16:48:21.325402 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_59fb2cd9-5414-4356-9438-d33f90ad084c/ceilometer-central-agent/0.log" Mar 13 16:48:21 crc kubenswrapper[4907]: I0313 16:48:21.450923 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_59fb2cd9-5414-4356-9438-d33f90ad084c/proxy-httpd/0.log" Mar 13 16:48:22 crc kubenswrapper[4907]: I0313 16:48:22.225928 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_59fb2cd9-5414-4356-9438-d33f90ad084c/sg-core/0.log" Mar 13 16:48:22 crc kubenswrapper[4907]: I0313 16:48:22.417289 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceph-client-openstack-openstack-cell1-tcvvp_ef037fca-09ee-48ed-b2f7-fc65aea9155a/ceph-client-openstack-openstack-cell1/0.log" Mar 13 16:48:22 crc kubenswrapper[4907]: I0313 16:48:22.922733 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_8c82a596-df2b-4e69-abe9-5f4662a3c6da/cinder-api-log/0.log" Mar 13 16:48:22 crc kubenswrapper[4907]: I0313 16:48:22.970339 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_8c82a596-df2b-4e69-abe9-5f4662a3c6da/cinder-api/0.log" Mar 13 16:48:23 crc kubenswrapper[4907]: I0313 16:48:23.107854 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_d92d5136-2169-44cc-9900-c3335a4b69eb/cinder-backup/0.log" Mar 13 16:48:23 crc kubenswrapper[4907]: I0313 16:48:23.292553 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_d92d5136-2169-44cc-9900-c3335a4b69eb/probe/0.log" Mar 13 16:48:23 crc kubenswrapper[4907]: I0313 16:48:23.356338 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_3a999cb1-5df0-43ae-946f-eae3f2c53848/cinder-scheduler/0.log" Mar 13 16:48:23 crc kubenswrapper[4907]: I0313 16:48:23.364433 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_3a999cb1-5df0-43ae-946f-eae3f2c53848/probe/0.log" Mar 13 16:48:23 crc kubenswrapper[4907]: I0313 16:48:23.628120 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_697762c8-b121-410a-a5db-21916ca1791f/probe/0.log" Mar 13 16:48:23 crc kubenswrapper[4907]: I0313 16:48:23.633705 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_697762c8-b121-410a-a5db-21916ca1791f/cinder-volume/0.log" Mar 13 16:48:23 crc kubenswrapper[4907]: I0313 16:48:23.727061 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-openstack-openstack-cell1-wrvtc_bfe0612b-2d05-4625-b14e-8b9ea02a8c06/configure-network-openstack-openstack-cell1/0.log" Mar 13 16:48:23 crc kubenswrapper[4907]: I0313 16:48:23.919070 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-openstack-openstack-cell1-s9llf_173f7de2-ef0c-4750-a8b3-23e078ab9bde/configure-os-openstack-openstack-cell1/0.log" Mar 13 16:48:23 crc kubenswrapper[4907]: I0313 16:48:23.999520 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-8644fdb9df-gm62m_44fcf282-3688-4a6f-b390-6b321e89b158/init/0.log" Mar 13 16:48:24 crc kubenswrapper[4907]: I0313 16:48:24.208798 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-8644fdb9df-gm62m_44fcf282-3688-4a6f-b390-6b321e89b158/dnsmasq-dns/0.log" Mar 13 16:48:24 crc kubenswrapper[4907]: I0313 16:48:24.240536 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-8644fdb9df-gm62m_44fcf282-3688-4a6f-b390-6b321e89b158/init/0.log" Mar 13 16:48:24 crc kubenswrapper[4907]: I0313 16:48:24.270579 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-openstack-openstack-cell1-84zm8_441cfe9f-03ed-4a4f-b3f6-58ff77c45cc9/download-cache-openstack-openstack-cell1/0.log" Mar 13 16:48:24 crc kubenswrapper[4907]: I0313 16:48:24.488115 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_95435ffc-5322-4bd8-9558-421f953096eb/glance-httpd/0.log" Mar 13 16:48:24 crc kubenswrapper[4907]: I0313 16:48:24.524992 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_70b550c7-5c28-416d-83c3-1ba63ecd5abd/glance-httpd/0.log" Mar 13 16:48:24 crc kubenswrapper[4907]: I0313 16:48:24.548216 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_70b550c7-5c28-416d-83c3-1ba63ecd5abd/glance-log/0.log" Mar 13 16:48:24 crc kubenswrapper[4907]: I0313 16:48:24.581487 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_95435ffc-5322-4bd8-9558-421f953096eb/glance-log/0.log" Mar 13 16:48:24 crc kubenswrapper[4907]: I0313 16:48:24.828681 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-api-7fd69496d-xglm5_a76ba2c0-c077-44ad-a2ed-2ee4f1a0052a/heat-api/0.log" Mar 13 16:48:24 crc kubenswrapper[4907]: I0313 16:48:24.981743 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-cfnapi-7dbbf4d5b8-846ft_1d7cc945-4d68-41c8-9ea0-359af3279cc8/heat-cfnapi/0.log" Mar 13 16:48:25 crc kubenswrapper[4907]: I0313 16:48:25.059347 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-engine-86cbbf5dd8-25c77_3dd72cd1-7342-40aa-8c3c-9d14375b3930/heat-engine/0.log" Mar 13 16:48:25 crc kubenswrapper[4907]: I0313 16:48:25.260856 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-6d79dff969-ssj4t_5315ef54-fd95-400c-a7ef-a356c0f867bc/horizon/0.log" Mar 13 16:48:25 crc kubenswrapper[4907]: I0313 16:48:25.264224 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-6d79dff969-ssj4t_5315ef54-fd95-400c-a7ef-a356c0f867bc/horizon-log/0.log" Mar 13 16:48:25 crc kubenswrapper[4907]: I0313 16:48:25.321443 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-openstack-openstack-cell1-55pht_c32eadac-12ec-49b8-930e-bc6af99631e8/install-certs-openstack-openstack-cell1/0.log" Mar 13 16:48:25 crc kubenswrapper[4907]: I0313 16:48:25.464249 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-openstack-openstack-cell1-vw9h5_cc30c8d7-5f7e-47c8-b89f-6a638b886769/install-os-openstack-openstack-cell1/0.log" Mar 13 16:48:25 crc kubenswrapper[4907]: I0313 16:48:25.714740 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-58b5f55f68-nlrbn_121ed53e-99ad-41a2-b8e6-56e4f56520c6/keystone-api/0.log" Mar 13 16:48:25 crc kubenswrapper[4907]: I0313 16:48:25.726947 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29556961-gshbz_e548a313-fc14-4ce9-b268-bf37c79eed83/keystone-cron/0.log" Mar 13 16:48:25 crc kubenswrapper[4907]: I0313 16:48:25.771611 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_53c58ef1-fd3c-4d26-9b00-5ce0bb0cc592/kube-state-metrics/0.log" Mar 13 16:48:25 crc kubenswrapper[4907]: I0313 16:48:25.979148 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-openstack-openstack-cell1-bdwbh_82c7c5cb-4f6b-4d38-a009-263025faa45b/libvirt-openstack-openstack-cell1/0.log" Mar 13 16:48:26 crc kubenswrapper[4907]: I0313 16:48:26.135336 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_f2518181-dd4d-4d35-9377-061bbd91c811/manila-api/0.log" Mar 13 16:48:26 crc kubenswrapper[4907]: I0313 16:48:26.207101 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_f2518181-dd4d-4d35-9377-061bbd91c811/manila-api-log/0.log" Mar 13 16:48:26 crc kubenswrapper[4907]: I0313 16:48:26.459386 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_2e179ecf-0fc5-4675-89a6-287d3398dc2d/probe/0.log" Mar 13 16:48:26 crc kubenswrapper[4907]: I0313 16:48:26.533541 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_93911045-021f-489b-9817-f0a24066c906/manila-share/0.log" Mar 13 16:48:26 crc kubenswrapper[4907]: I0313 16:48:26.585730 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_2e179ecf-0fc5-4675-89a6-287d3398dc2d/manila-scheduler/0.log" Mar 13 16:48:26 crc kubenswrapper[4907]: I0313 16:48:26.646992 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_93911045-021f-489b-9817-f0a24066c906/probe/0.log" Mar 13 16:48:27 crc kubenswrapper[4907]: I0313 16:48:27.035095 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-564986d4bf-zhh4c_3ce663fe-b525-44ea-a6d4-33f6b2366f46/neutron-httpd/0.log" Mar 13 16:48:27 crc kubenswrapper[4907]: I0313 16:48:27.037271 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-564986d4bf-zhh4c_3ce663fe-b525-44ea-a6d4-33f6b2366f46/neutron-api/0.log" Mar 13 16:48:27 crc kubenswrapper[4907]: I0313 16:48:27.150345 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-dhcp-openstack-openstack-cell1-jmbms_b5764492-ad4a-4e43-8a98-99ea7cfd9248/neutron-dhcp-openstack-openstack-cell1/0.log" Mar 13 16:48:27 crc kubenswrapper[4907]: I0313 16:48:27.397641 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-openstack-openstack-cell1-pwqj6_3dc18cfe-d63a-4960-8c7d-10913cc78896/neutron-metadata-openstack-openstack-cell1/0.log" Mar 13 16:48:27 crc kubenswrapper[4907]: I0313 16:48:27.568057 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-sriov-openstack-openstack-cell1-sqxhp_05d384d7-9635-4d68-8eaf-48fea8bed0b8/neutron-sriov-openstack-openstack-cell1/0.log" Mar 13 16:48:27 crc kubenswrapper[4907]: I0313 16:48:27.805115 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_c11f341d-9a76-4e7e-b1e7-8d63c30bac32/nova-api-api/0.log" Mar 13 16:48:27 crc kubenswrapper[4907]: I0313 16:48:27.956119 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_c11f341d-9a76-4e7e-b1e7-8d63c30bac32/nova-api-log/0.log" Mar 13 16:48:28 crc kubenswrapper[4907]: I0313 16:48:28.096873 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_fd1f3838-01fe-4e37-9859-8aa28048e81c/nova-cell0-conductor-conductor/0.log" Mar 13 16:48:28 crc kubenswrapper[4907]: I0313 16:48:28.257674 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_0a6bc64d-334f-46cd-aea9-486ab56d4cf9/nova-cell1-conductor-conductor/0.log" Mar 13 16:48:28 crc kubenswrapper[4907]: I0313 16:48:28.548121 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_317b6fd4-e642-4858-bc3d-2cce128ea0e7/nova-cell1-novncproxy-novncproxy/0.log" Mar 13 16:48:28 crc kubenswrapper[4907]: I0313 16:48:28.590245 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc_de68abba-24e7-49b8-9e1f-adfa05d9d3dd/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell1/0.log" Mar 13 16:48:28 crc kubenswrapper[4907]: I0313 16:48:28.731672 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-openstack-openstack-cell1-2swxs_563c85ca-38c8-4853-a278-2a46fd0347a3/nova-cell1-openstack-openstack-cell1/0.log" Mar 13 16:48:28 crc kubenswrapper[4907]: I0313 16:48:28.864290 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_13364227-6148-4b0d-ba7d-082578dfe1d5/nova-metadata-metadata/0.log" Mar 13 16:48:28 crc kubenswrapper[4907]: I0313 16:48:28.950618 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_13364227-6148-4b0d-ba7d-082578dfe1d5/nova-metadata-log/0.log" Mar 13 16:48:29 crc kubenswrapper[4907]: I0313 16:48:29.110338 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_54123701-97fb-4f4f-baed-4c2183bad203/nova-scheduler-scheduler/0.log" Mar 13 16:48:29 crc kubenswrapper[4907]: I0313 16:48:29.697929 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-api-6db876fcbd-4wtkj_1c868ec4-5e49-4be1-a1eb-c4748753aecb/init/0.log" Mar 13 16:48:29 crc kubenswrapper[4907]: I0313 16:48:29.936501 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-api-6db876fcbd-4wtkj_1c868ec4-5e49-4be1-a1eb-c4748753aecb/init/0.log" Mar 13 16:48:29 crc kubenswrapper[4907]: I0313 16:48:29.942104 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-api-6db876fcbd-4wtkj_1c868ec4-5e49-4be1-a1eb-c4748753aecb/octavia-api-provider-agent/0.log" Mar 13 16:48:30 crc kubenswrapper[4907]: I0313 16:48:30.213480 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-api-6db876fcbd-4wtkj_1c868ec4-5e49-4be1-a1eb-c4748753aecb/octavia-api/0.log" Mar 13 16:48:30 crc kubenswrapper[4907]: I0313 16:48:30.369252 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-healthmanager-hbbw4_57610eac-b9eb-4628-bd99-b3ebb35b2772/init/0.log" Mar 13 16:48:30 crc kubenswrapper[4907]: I0313 16:48:30.629500 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-healthmanager-hbbw4_57610eac-b9eb-4628-bd99-b3ebb35b2772/init/0.log" Mar 13 16:48:30 crc kubenswrapper[4907]: I0313 16:48:30.641344 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-housekeeping-p69c5_ab5868a8-f083-4417-b7c0-f01dad538927/init/0.log" Mar 13 16:48:30 crc kubenswrapper[4907]: I0313 16:48:30.712224 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-healthmanager-hbbw4_57610eac-b9eb-4628-bd99-b3ebb35b2772/octavia-healthmanager/0.log" Mar 13 16:48:30 crc kubenswrapper[4907]: I0313 16:48:30.884290 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-housekeeping-p69c5_ab5868a8-f083-4417-b7c0-f01dad538927/init/0.log" Mar 13 16:48:30 crc kubenswrapper[4907]: I0313 16:48:30.902130 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-housekeeping-p69c5_ab5868a8-f083-4417-b7c0-f01dad538927/octavia-housekeeping/0.log" Mar 13 16:48:31 crc kubenswrapper[4907]: I0313 16:48:31.506588 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-rsyslog-zbs6z_4df4068e-249f-47ea-a84a-3c178c27bdf3/init/0.log" Mar 13 16:48:31 crc kubenswrapper[4907]: I0313 16:48:31.755534 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-rsyslog-zbs6z_4df4068e-249f-47ea-a84a-3c178c27bdf3/init/0.log" Mar 13 16:48:31 crc kubenswrapper[4907]: I0313 16:48:31.793557 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-rsyslog-zbs6z_4df4068e-249f-47ea-a84a-3c178c27bdf3/octavia-rsyslog/0.log" Mar 13 16:48:31 crc kubenswrapper[4907]: I0313 16:48:31.890836 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-worker-7pld2_f6c0a640-5177-4188-abc3-d934f75624bb/init/0.log" Mar 13 16:48:32 crc kubenswrapper[4907]: I0313 16:48:32.139361 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-worker-7pld2_f6c0a640-5177-4188-abc3-d934f75624bb/init/0.log" Mar 13 16:48:32 crc kubenswrapper[4907]: I0313 16:48:32.150918 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_edb521bb-211e-48c0-b644-8dc44b235feb/mysql-bootstrap/0.log" Mar 13 16:48:32 crc kubenswrapper[4907]: I0313 16:48:32.340913 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-worker-7pld2_f6c0a640-5177-4188-abc3-d934f75624bb/octavia-worker/0.log" Mar 13 16:48:32 crc kubenswrapper[4907]: I0313 16:48:32.351559 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_edb521bb-211e-48c0-b644-8dc44b235feb/mysql-bootstrap/0.log" Mar 13 16:48:32 crc kubenswrapper[4907]: I0313 16:48:32.395414 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_edb521bb-211e-48c0-b644-8dc44b235feb/galera/0.log" Mar 13 16:48:32 crc kubenswrapper[4907]: I0313 16:48:32.569261 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_585e3be7-1e17-443c-b430-0c03727167cb/mysql-bootstrap/0.log" Mar 13 16:48:32 crc kubenswrapper[4907]: I0313 16:48:32.770094 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_2316d199-eef1-4ac1-9f63-cc73cba8d4c7/openstackclient/0.log" Mar 13 16:48:32 crc kubenswrapper[4907]: I0313 16:48:32.804907 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_585e3be7-1e17-443c-b430-0c03727167cb/galera/0.log" Mar 13 16:48:32 crc kubenswrapper[4907]: I0313 16:48:32.867808 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_585e3be7-1e17-443c-b430-0c03727167cb/mysql-bootstrap/0.log" Mar 13 16:48:33 crc kubenswrapper[4907]: I0313 16:48:33.007359 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-xstlm_f9398439-2570-42ec-b6fc-f9770b988c73/openstack-network-exporter/0.log" Mar 13 16:48:33 crc kubenswrapper[4907]: I0313 16:48:33.120790 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-v5b57_34b897ab-ce41-4cd5-bcae-cea3c0dea9e9/ovsdb-server-init/0.log" Mar 13 16:48:33 crc kubenswrapper[4907]: I0313 16:48:33.408975 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-v5b57_34b897ab-ce41-4cd5-bcae-cea3c0dea9e9/ovs-vswitchd/0.log" Mar 13 16:48:33 crc kubenswrapper[4907]: I0313 16:48:33.411443 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-v5b57_34b897ab-ce41-4cd5-bcae-cea3c0dea9e9/ovsdb-server-init/0.log" Mar 13 16:48:33 crc kubenswrapper[4907]: I0313 16:48:33.508476 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-v5b57_34b897ab-ce41-4cd5-bcae-cea3c0dea9e9/ovsdb-server/0.log" Mar 13 16:48:33 crc kubenswrapper[4907]: I0313 16:48:33.673027 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-zb9k2_1cf43c88-f066-4979-8590-b8e6668aab09/ovn-controller/0.log" Mar 13 16:48:33 crc kubenswrapper[4907]: I0313 16:48:33.792215 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_1b95d939-59f1-4ba9-b39a-eb255ab7aa77/openstack-network-exporter/0.log" Mar 13 16:48:33 crc kubenswrapper[4907]: I0313 16:48:33.832438 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_1b95d939-59f1-4ba9-b39a-eb255ab7aa77/ovn-northd/0.log" Mar 13 16:48:34 crc kubenswrapper[4907]: I0313 16:48:34.047188 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-openstack-openstack-cell1-8dwgb_47d984fd-dfa5-4a10-91c3-3c9a13edc91c/ovn-openstack-openstack-cell1/0.log" Mar 13 16:48:34 crc kubenswrapper[4907]: I0313 16:48:34.052906 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_a167eb31-6b82-44f9-bf66-465b64a91226/openstack-network-exporter/0.log" Mar 13 16:48:34 crc kubenswrapper[4907]: I0313 16:48:34.439240 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_a167eb31-6b82-44f9-bf66-465b64a91226/ovsdbserver-nb/0.log" Mar 13 16:48:34 crc kubenswrapper[4907]: I0313 16:48:34.498077 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-1_7e270de1-912f-48f9-bfc7-e415fa5c7b5d/openstack-network-exporter/0.log" Mar 13 16:48:34 crc kubenswrapper[4907]: I0313 16:48:34.612944 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-1_7e270de1-912f-48f9-bfc7-e415fa5c7b5d/ovsdbserver-nb/0.log" Mar 13 16:48:34 crc kubenswrapper[4907]: I0313 16:48:34.739490 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-2_3ca7bf8f-4514-493e-9385-75cfcd0807f9/ovsdbserver-nb/0.log" Mar 13 16:48:34 crc kubenswrapper[4907]: I0313 16:48:34.751214 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-2_3ca7bf8f-4514-493e-9385-75cfcd0807f9/openstack-network-exporter/0.log" Mar 13 16:48:34 crc kubenswrapper[4907]: I0313 16:48:34.966984 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_bffaea76-e210-44d1-8caa-ebf82edb8ef0/openstack-network-exporter/0.log" Mar 13 16:48:34 crc kubenswrapper[4907]: I0313 16:48:34.969673 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_bffaea76-e210-44d1-8caa-ebf82edb8ef0/ovsdbserver-sb/0.log" Mar 13 16:48:35 crc kubenswrapper[4907]: I0313 16:48:35.118900 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-1_e5ec94c1-f60b-4156-b3ba-278db827c9b8/openstack-network-exporter/0.log" Mar 13 16:48:35 crc kubenswrapper[4907]: I0313 16:48:35.192467 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-1_e5ec94c1-f60b-4156-b3ba-278db827c9b8/ovsdbserver-sb/0.log" Mar 13 16:48:35 crc kubenswrapper[4907]: I0313 16:48:35.339631 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-2_b188658f-1e5c-49d4-be19-0a888caddb43/openstack-network-exporter/0.log" Mar 13 16:48:35 crc kubenswrapper[4907]: I0313 16:48:35.391927 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-2_b188658f-1e5c-49d4-be19-0a888caddb43/ovsdbserver-sb/0.log" Mar 13 16:48:35 crc kubenswrapper[4907]: I0313 16:48:35.561197 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-6575d7f87d-t4csd_12e05309-4563-4b78-a598-e4e8ced9cc8d/placement-api/0.log" Mar 13 16:48:35 crc kubenswrapper[4907]: I0313 16:48:35.670477 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-6575d7f87d-t4csd_12e05309-4563-4b78-a598-e4e8ced9cc8d/placement-log/0.log" Mar 13 16:48:35 crc kubenswrapper[4907]: I0313 16:48:35.711276 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_pre-adoption-validation-openstack-pre-adoption-openstack-ckdmlw_044c1d1f-5bd4-4359-9262-e08daf71a4ae/pre-adoption-validation-openstack-pre-adoption-openstack-cell1/0.log" Mar 13 16:48:35 crc kubenswrapper[4907]: I0313 16:48:35.881846 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_3e5a8905-7f71-4979-aee3-f81e772ea46a/init-config-reloader/0.log" Mar 13 16:48:36 crc kubenswrapper[4907]: I0313 16:48:36.053672 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_3e5a8905-7f71-4979-aee3-f81e772ea46a/prometheus/0.log" Mar 13 16:48:36 crc kubenswrapper[4907]: I0313 16:48:36.060676 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_3e5a8905-7f71-4979-aee3-f81e772ea46a/init-config-reloader/0.log" Mar 13 16:48:36 crc kubenswrapper[4907]: I0313 16:48:36.122581 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_3e5a8905-7f71-4979-aee3-f81e772ea46a/thanos-sidecar/0.log" Mar 13 16:48:36 crc kubenswrapper[4907]: I0313 16:48:36.133687 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_8c712e4f-4999-476b-970f-eae3189b4e59/memcached/0.log" Mar 13 16:48:36 crc kubenswrapper[4907]: I0313 16:48:36.134288 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_3e5a8905-7f71-4979-aee3-f81e772ea46a/config-reloader/0.log" Mar 13 16:48:36 crc kubenswrapper[4907]: I0313 16:48:36.329976 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_fc486d5d-ec94-461c-acf2-dcca08b2a1b6/setup-container/0.log" Mar 13 16:48:36 crc kubenswrapper[4907]: I0313 16:48:36.507064 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_fc486d5d-ec94-461c-acf2-dcca08b2a1b6/setup-container/0.log" Mar 13 16:48:36 crc kubenswrapper[4907]: I0313 16:48:36.526120 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_fc486d5d-ec94-461c-acf2-dcca08b2a1b6/rabbitmq/0.log" Mar 13 16:48:36 crc kubenswrapper[4907]: I0313 16:48:36.584430 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_2ae0c4fc-0b3a-47a3-a6db-3e0ae3d64dad/setup-container/0.log" Mar 13 16:48:36 crc kubenswrapper[4907]: I0313 16:48:36.739996 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_2ae0c4fc-0b3a-47a3-a6db-3e0ae3d64dad/setup-container/0.log" Mar 13 16:48:36 crc kubenswrapper[4907]: I0313 16:48:36.793505 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_2ae0c4fc-0b3a-47a3-a6db-3e0ae3d64dad/rabbitmq/0.log" Mar 13 16:48:36 crc kubenswrapper[4907]: I0313 16:48:36.805993 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-openstack-openstack-cell1-x7hvj_cf92c1e7-1021-4405-840e-f2286adea31d/reboot-os-openstack-openstack-cell1/0.log" Mar 13 16:48:36 crc kubenswrapper[4907]: I0313 16:48:36.942064 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-openstack-openstack-cell1-hml5k_34db6153-7522-4aae-9712-eb87cc9aa543/run-os-openstack-openstack-cell1/0.log" Mar 13 16:48:37 crc kubenswrapper[4907]: I0313 16:48:37.051374 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-openstack-phzr9_4aef751c-592b-403e-b581-92c88cf65785/ssh-known-hosts-openstack/0.log" Mar 13 16:48:37 crc kubenswrapper[4907]: I0313 16:48:37.130941 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-openstack-openstack-cell1-gmnh8_b5718b26-a6b1-4b64-812b-a85da3d6f0a4/telemetry-openstack-openstack-cell1/0.log" Mar 13 16:48:37 crc kubenswrapper[4907]: I0313 16:48:37.323676 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-openstack-openstack-cell1-rrpt2_9a4f6803-5919-4737-acd0-5b25e12a7c55/validate-network-openstack-openstack-cell1/0.log" Mar 13 16:48:37 crc kubenswrapper[4907]: I0313 16:48:37.328929 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tripleo-cleanup-tripleo-cleanup-openstack-cell1-fj6qz_02ef8f0f-4ae4-478a-a729-3bde4adade43/tripleo-cleanup-tripleo-cleanup-openstack-cell1/0.log" Mar 13 16:48:44 crc kubenswrapper[4907]: I0313 16:48:44.995670 4907 scope.go:117] "RemoveContainer" containerID="3ca53b3677f234a39cf1706a0419909431a4370ca20c691d14a98c1cf5bf1d3b" Mar 13 16:49:04 crc kubenswrapper[4907]: I0313 16:49:04.159669 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_a25da44700f2da9e511716e9ea5c2d7cc3bdaf6532fb20e8f09ee07298p5ddv_29e80c51-f1f5-4850-8e81-ee52fe060a43/util/0.log" Mar 13 16:49:04 crc kubenswrapper[4907]: I0313 16:49:04.358616 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_a25da44700f2da9e511716e9ea5c2d7cc3bdaf6532fb20e8f09ee07298p5ddv_29e80c51-f1f5-4850-8e81-ee52fe060a43/pull/0.log" Mar 13 16:49:04 crc kubenswrapper[4907]: I0313 16:49:04.368382 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_a25da44700f2da9e511716e9ea5c2d7cc3bdaf6532fb20e8f09ee07298p5ddv_29e80c51-f1f5-4850-8e81-ee52fe060a43/util/0.log" Mar 13 16:49:04 crc kubenswrapper[4907]: I0313 16:49:04.394432 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_a25da44700f2da9e511716e9ea5c2d7cc3bdaf6532fb20e8f09ee07298p5ddv_29e80c51-f1f5-4850-8e81-ee52fe060a43/pull/0.log" Mar 13 16:49:04 crc kubenswrapper[4907]: I0313 16:49:04.610025 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_a25da44700f2da9e511716e9ea5c2d7cc3bdaf6532fb20e8f09ee07298p5ddv_29e80c51-f1f5-4850-8e81-ee52fe060a43/pull/0.log" Mar 13 16:49:04 crc kubenswrapper[4907]: I0313 16:49:04.611764 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_a25da44700f2da9e511716e9ea5c2d7cc3bdaf6532fb20e8f09ee07298p5ddv_29e80c51-f1f5-4850-8e81-ee52fe060a43/extract/0.log" Mar 13 16:49:04 crc kubenswrapper[4907]: I0313 16:49:04.675799 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_a25da44700f2da9e511716e9ea5c2d7cc3bdaf6532fb20e8f09ee07298p5ddv_29e80c51-f1f5-4850-8e81-ee52fe060a43/util/0.log" Mar 13 16:49:05 crc kubenswrapper[4907]: I0313 16:49:05.021151 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-d47688694-tl8t4_863c77e9-4022-4ab9-8ac7-a4006d22813c/manager/0.log" Mar 13 16:49:05 crc kubenswrapper[4907]: I0313 16:49:05.051058 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-66d56f6ff4-7cbb2_8789d213-46a4-4649-add6-51ebe4412c27/manager/0.log" Mar 13 16:49:05 crc kubenswrapper[4907]: I0313 16:49:05.465756 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-5964f64c48-gll76_b7890be5-5773-4562-b1ab-8e647aa16256/manager/0.log" Mar 13 16:49:05 crc kubenswrapper[4907]: I0313 16:49:05.471291 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-77b6666d85-874m2_d3cdd2ca-3e81-480f-876e-2b975d8bc9d8/manager/0.log" Mar 13 16:49:05 crc kubenswrapper[4907]: I0313 16:49:05.719252 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-6d9d6b584d-r7bhq_8ecaf4b8-808e-44b9-8ae7-6ef9519c4a66/manager/0.log" Mar 13 16:49:06 crc kubenswrapper[4907]: I0313 16:49:06.007261 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-5bc894d9b-w5829_19c320fa-6619-45cd-a73e-65834173b786/manager/0.log" Mar 13 16:49:06 crc kubenswrapper[4907]: I0313 16:49:06.367809 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-684f77d66d-7rdzd_7d620a90-0e93-41dc-ab39-fc6ea8c461e7/manager/0.log" Mar 13 16:49:06 crc kubenswrapper[4907]: I0313 16:49:06.499062 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-54dc5b8f8d-lp7kj_64d55221-60c5-4521-ac84-beed4e9b3993/manager/0.log" Mar 13 16:49:06 crc kubenswrapper[4907]: I0313 16:49:06.700655 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-57b484b4df-m8j7m_c55515f1-509e-4a4e-813d-fbc466fca489/manager/0.log" Mar 13 16:49:06 crc kubenswrapper[4907]: I0313 16:49:06.907893 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-5b6b6b4c9f-5m2vw_e8134e91-4b9d-4e3c-9aec-eba18acabdce/manager/0.log" Mar 13 16:49:07 crc kubenswrapper[4907]: I0313 16:49:07.315740 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-776c5696bf-gvgdl_29201cb7-c056-4e4e-b0da-cc775c7ef11f/manager/0.log" Mar 13 16:49:07 crc kubenswrapper[4907]: I0313 16:49:07.864234 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-5f4f55cb5c-sk2l9_5f3d2dd8-6ee2-415a-8cc6-55f6e09cb473/manager/0.log" Mar 13 16:49:07 crc kubenswrapper[4907]: I0313 16:49:07.999244 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-7f84474648-ql4kd_4ab8f569-de56-4fea-9be0-07908063522a/manager/0.log" Mar 13 16:49:08 crc kubenswrapper[4907]: I0313 16:49:08.070727 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-984cd4dcf-w497d_77b1eb71-a455-4ec3-8e17-a847575cf0b0/manager/0.log" Mar 13 16:49:08 crc kubenswrapper[4907]: I0313 16:49:08.134717 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-6f7958d774rr6kf_f915558a-c0ab-4d8e-a427-e56a7588e382/manager/0.log" Mar 13 16:49:08 crc kubenswrapper[4907]: I0313 16:49:08.475728 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-init-6dc56d8cd6-hxpjr_dc51eaa2-313d-4256-a145-3b73b02e2dfe/operator/0.log" Mar 13 16:49:08 crc kubenswrapper[4907]: I0313 16:49:08.667583 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-z98nv_c23245c8-0fba-483e-b126-09349d413030/registry-server/0.log" Mar 13 16:49:08 crc kubenswrapper[4907]: I0313 16:49:08.830367 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-bbc5b68f9-cqmhw_fdddbc99-1a71-4250-ab2a-6f426e744423/manager/0.log" Mar 13 16:49:08 crc kubenswrapper[4907]: I0313 16:49:08.946130 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-574d45c66c-b99f7_0b3db736-b3d4-494f-9755-9103150d267f/manager/0.log" Mar 13 16:49:09 crc kubenswrapper[4907]: I0313 16:49:09.005264 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-xtvzb_44ecb24b-4c5c-4cf6-96e9-4d8b52469d3d/operator/0.log" Mar 13 16:49:09 crc kubenswrapper[4907]: I0313 16:49:09.316796 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-7f9cc5dd44-sd6nr_92718823-f344-4d7b-beff-e75f1574f1e2/manager/0.log" Mar 13 16:49:09 crc kubenswrapper[4907]: I0313 16:49:09.426217 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-6854b8b9d9-dj95f_d139e1bb-8ae8-4760-a1b6-3f7d96f3e9a1/manager/0.log" Mar 13 16:49:09 crc kubenswrapper[4907]: I0313 16:49:09.513686 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5c5cb9c4d7-vzpcd_a2b0825f-f2f1-4b3f-b7b0-5a476f99b4fe/manager/0.log" Mar 13 16:49:09 crc kubenswrapper[4907]: I0313 16:49:09.653010 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-6c4d75f7f9-hjp4p_a89fbf9b-b391-426b-9f9d-91cdba71602b/manager/0.log" Mar 13 16:49:10 crc kubenswrapper[4907]: I0313 16:49:10.584277 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-6484b7b757-qldxp_9ac47edf-03ab-4e93-b238-2aa00765ab06/manager/0.log" Mar 13 16:49:14 crc kubenswrapper[4907]: I0313 16:49:14.444454 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-56j7g"] Mar 13 16:49:14 crc kubenswrapper[4907]: E0313 16:49:14.445440 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="168ac198-4031-4bf7-bcf6-eb190b28dce6" containerName="oc" Mar 13 16:49:14 crc kubenswrapper[4907]: I0313 16:49:14.445453 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="168ac198-4031-4bf7-bcf6-eb190b28dce6" containerName="oc" Mar 13 16:49:14 crc kubenswrapper[4907]: I0313 16:49:14.445669 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="168ac198-4031-4bf7-bcf6-eb190b28dce6" containerName="oc" Mar 13 16:49:14 crc kubenswrapper[4907]: I0313 16:49:14.447266 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-56j7g" Mar 13 16:49:14 crc kubenswrapper[4907]: I0313 16:49:14.460736 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-56j7g"] Mar 13 16:49:14 crc kubenswrapper[4907]: I0313 16:49:14.607471 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kg8v4\" (UniqueName: \"kubernetes.io/projected/8ea81195-fa1d-4422-847d-ae98208f2847-kube-api-access-kg8v4\") pod \"redhat-marketplace-56j7g\" (UID: \"8ea81195-fa1d-4422-847d-ae98208f2847\") " pod="openshift-marketplace/redhat-marketplace-56j7g" Mar 13 16:49:14 crc kubenswrapper[4907]: I0313 16:49:14.607644 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ea81195-fa1d-4422-847d-ae98208f2847-catalog-content\") pod \"redhat-marketplace-56j7g\" (UID: \"8ea81195-fa1d-4422-847d-ae98208f2847\") " pod="openshift-marketplace/redhat-marketplace-56j7g" Mar 13 16:49:14 crc kubenswrapper[4907]: I0313 16:49:14.607950 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ea81195-fa1d-4422-847d-ae98208f2847-utilities\") pod \"redhat-marketplace-56j7g\" (UID: \"8ea81195-fa1d-4422-847d-ae98208f2847\") " pod="openshift-marketplace/redhat-marketplace-56j7g" Mar 13 16:49:14 crc kubenswrapper[4907]: I0313 16:49:14.709796 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ea81195-fa1d-4422-847d-ae98208f2847-utilities\") pod \"redhat-marketplace-56j7g\" (UID: \"8ea81195-fa1d-4422-847d-ae98208f2847\") " pod="openshift-marketplace/redhat-marketplace-56j7g" Mar 13 16:49:14 crc kubenswrapper[4907]: I0313 16:49:14.709936 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kg8v4\" (UniqueName: \"kubernetes.io/projected/8ea81195-fa1d-4422-847d-ae98208f2847-kube-api-access-kg8v4\") pod \"redhat-marketplace-56j7g\" (UID: \"8ea81195-fa1d-4422-847d-ae98208f2847\") " pod="openshift-marketplace/redhat-marketplace-56j7g" Mar 13 16:49:14 crc kubenswrapper[4907]: I0313 16:49:14.710007 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ea81195-fa1d-4422-847d-ae98208f2847-catalog-content\") pod \"redhat-marketplace-56j7g\" (UID: \"8ea81195-fa1d-4422-847d-ae98208f2847\") " pod="openshift-marketplace/redhat-marketplace-56j7g" Mar 13 16:49:14 crc kubenswrapper[4907]: I0313 16:49:14.710378 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ea81195-fa1d-4422-847d-ae98208f2847-utilities\") pod \"redhat-marketplace-56j7g\" (UID: \"8ea81195-fa1d-4422-847d-ae98208f2847\") " pod="openshift-marketplace/redhat-marketplace-56j7g" Mar 13 16:49:14 crc kubenswrapper[4907]: I0313 16:49:14.710406 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ea81195-fa1d-4422-847d-ae98208f2847-catalog-content\") pod \"redhat-marketplace-56j7g\" (UID: \"8ea81195-fa1d-4422-847d-ae98208f2847\") " pod="openshift-marketplace/redhat-marketplace-56j7g" Mar 13 16:49:14 crc kubenswrapper[4907]: I0313 16:49:14.731113 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kg8v4\" (UniqueName: \"kubernetes.io/projected/8ea81195-fa1d-4422-847d-ae98208f2847-kube-api-access-kg8v4\") pod \"redhat-marketplace-56j7g\" (UID: \"8ea81195-fa1d-4422-847d-ae98208f2847\") " pod="openshift-marketplace/redhat-marketplace-56j7g" Mar 13 16:49:14 crc kubenswrapper[4907]: I0313 16:49:14.767525 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-56j7g" Mar 13 16:49:15 crc kubenswrapper[4907]: I0313 16:49:15.274464 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-56j7g"] Mar 13 16:49:15 crc kubenswrapper[4907]: I0313 16:49:15.555466 4907 generic.go:334] "Generic (PLEG): container finished" podID="8ea81195-fa1d-4422-847d-ae98208f2847" containerID="dd4bda3c37e3f835e6ac87c9e4ed30b2ce48b628213cf2a3e9b4d4ba01e7b3e3" exitCode=0 Mar 13 16:49:15 crc kubenswrapper[4907]: I0313 16:49:15.556793 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-56j7g" event={"ID":"8ea81195-fa1d-4422-847d-ae98208f2847","Type":"ContainerDied","Data":"dd4bda3c37e3f835e6ac87c9e4ed30b2ce48b628213cf2a3e9b4d4ba01e7b3e3"} Mar 13 16:49:15 crc kubenswrapper[4907]: I0313 16:49:15.556936 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-56j7g" event={"ID":"8ea81195-fa1d-4422-847d-ae98208f2847","Type":"ContainerStarted","Data":"c2c590297fd6b5f081063b6f2e41e08f60652aec05116d66a170c68ca89264d6"} Mar 13 16:49:16 crc kubenswrapper[4907]: I0313 16:49:16.565604 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-56j7g" event={"ID":"8ea81195-fa1d-4422-847d-ae98208f2847","Type":"ContainerStarted","Data":"a8f2bc7927e473a96dafc57a193fbcc7a34307d796eb4530e9feb1c1071f99e9"} Mar 13 16:49:17 crc kubenswrapper[4907]: I0313 16:49:17.576520 4907 generic.go:334] "Generic (PLEG): container finished" podID="8ea81195-fa1d-4422-847d-ae98208f2847" containerID="a8f2bc7927e473a96dafc57a193fbcc7a34307d796eb4530e9feb1c1071f99e9" exitCode=0 Mar 13 16:49:17 crc kubenswrapper[4907]: I0313 16:49:17.576598 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-56j7g" event={"ID":"8ea81195-fa1d-4422-847d-ae98208f2847","Type":"ContainerDied","Data":"a8f2bc7927e473a96dafc57a193fbcc7a34307d796eb4530e9feb1c1071f99e9"} Mar 13 16:49:18 crc kubenswrapper[4907]: I0313 16:49:18.588457 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-56j7g" event={"ID":"8ea81195-fa1d-4422-847d-ae98208f2847","Type":"ContainerStarted","Data":"a76d9623b30999410f9778c66fe8e242e8278147e679989287e043281592d18d"} Mar 13 16:49:24 crc kubenswrapper[4907]: I0313 16:49:24.768776 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-56j7g" Mar 13 16:49:24 crc kubenswrapper[4907]: I0313 16:49:24.770106 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-56j7g" Mar 13 16:49:24 crc kubenswrapper[4907]: I0313 16:49:24.838287 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-56j7g" Mar 13 16:49:24 crc kubenswrapper[4907]: I0313 16:49:24.858391 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-56j7g" podStartSLOduration=8.460679432 podStartE2EDuration="10.858370709s" podCreationTimestamp="2026-03-13 16:49:14 +0000 UTC" firstStartedPulling="2026-03-13 16:49:15.557545691 +0000 UTC m=+9854.457333380" lastFinishedPulling="2026-03-13 16:49:17.955236968 +0000 UTC m=+9856.855024657" observedRunningTime="2026-03-13 16:49:18.617774739 +0000 UTC m=+9857.517562428" watchObservedRunningTime="2026-03-13 16:49:24.858370709 +0000 UTC m=+9863.758158398" Mar 13 16:49:25 crc kubenswrapper[4907]: I0313 16:49:25.710460 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-56j7g" Mar 13 16:49:25 crc kubenswrapper[4907]: I0313 16:49:25.765241 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-56j7g"] Mar 13 16:49:27 crc kubenswrapper[4907]: I0313 16:49:27.679683 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-56j7g" podUID="8ea81195-fa1d-4422-847d-ae98208f2847" containerName="registry-server" containerID="cri-o://a76d9623b30999410f9778c66fe8e242e8278147e679989287e043281592d18d" gracePeriod=2 Mar 13 16:49:28 crc kubenswrapper[4907]: I0313 16:49:28.258605 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-56j7g" Mar 13 16:49:28 crc kubenswrapper[4907]: I0313 16:49:28.338043 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ea81195-fa1d-4422-847d-ae98208f2847-catalog-content\") pod \"8ea81195-fa1d-4422-847d-ae98208f2847\" (UID: \"8ea81195-fa1d-4422-847d-ae98208f2847\") " Mar 13 16:49:28 crc kubenswrapper[4907]: I0313 16:49:28.338298 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ea81195-fa1d-4422-847d-ae98208f2847-utilities\") pod \"8ea81195-fa1d-4422-847d-ae98208f2847\" (UID: \"8ea81195-fa1d-4422-847d-ae98208f2847\") " Mar 13 16:49:28 crc kubenswrapper[4907]: I0313 16:49:28.338761 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kg8v4\" (UniqueName: \"kubernetes.io/projected/8ea81195-fa1d-4422-847d-ae98208f2847-kube-api-access-kg8v4\") pod \"8ea81195-fa1d-4422-847d-ae98208f2847\" (UID: \"8ea81195-fa1d-4422-847d-ae98208f2847\") " Mar 13 16:49:28 crc kubenswrapper[4907]: I0313 16:49:28.361553 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8ea81195-fa1d-4422-847d-ae98208f2847-utilities" (OuterVolumeSpecName: "utilities") pod "8ea81195-fa1d-4422-847d-ae98208f2847" (UID: "8ea81195-fa1d-4422-847d-ae98208f2847"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 16:49:28 crc kubenswrapper[4907]: I0313 16:49:28.382207 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8ea81195-fa1d-4422-847d-ae98208f2847-kube-api-access-kg8v4" (OuterVolumeSpecName: "kube-api-access-kg8v4") pod "8ea81195-fa1d-4422-847d-ae98208f2847" (UID: "8ea81195-fa1d-4422-847d-ae98208f2847"). InnerVolumeSpecName "kube-api-access-kg8v4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:49:28 crc kubenswrapper[4907]: I0313 16:49:28.387401 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8ea81195-fa1d-4422-847d-ae98208f2847-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8ea81195-fa1d-4422-847d-ae98208f2847" (UID: "8ea81195-fa1d-4422-847d-ae98208f2847"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 16:49:28 crc kubenswrapper[4907]: I0313 16:49:28.442761 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ea81195-fa1d-4422-847d-ae98208f2847-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 16:49:28 crc kubenswrapper[4907]: I0313 16:49:28.442792 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ea81195-fa1d-4422-847d-ae98208f2847-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 16:49:28 crc kubenswrapper[4907]: I0313 16:49:28.442804 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kg8v4\" (UniqueName: \"kubernetes.io/projected/8ea81195-fa1d-4422-847d-ae98208f2847-kube-api-access-kg8v4\") on node \"crc\" DevicePath \"\"" Mar 13 16:49:28 crc kubenswrapper[4907]: I0313 16:49:28.690586 4907 generic.go:334] "Generic (PLEG): container finished" podID="8ea81195-fa1d-4422-847d-ae98208f2847" containerID="a76d9623b30999410f9778c66fe8e242e8278147e679989287e043281592d18d" exitCode=0 Mar 13 16:49:28 crc kubenswrapper[4907]: I0313 16:49:28.690634 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-56j7g" event={"ID":"8ea81195-fa1d-4422-847d-ae98208f2847","Type":"ContainerDied","Data":"a76d9623b30999410f9778c66fe8e242e8278147e679989287e043281592d18d"} Mar 13 16:49:28 crc kubenswrapper[4907]: I0313 16:49:28.690665 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-56j7g" Mar 13 16:49:28 crc kubenswrapper[4907]: I0313 16:49:28.690693 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-56j7g" event={"ID":"8ea81195-fa1d-4422-847d-ae98208f2847","Type":"ContainerDied","Data":"c2c590297fd6b5f081063b6f2e41e08f60652aec05116d66a170c68ca89264d6"} Mar 13 16:49:28 crc kubenswrapper[4907]: I0313 16:49:28.690713 4907 scope.go:117] "RemoveContainer" containerID="a76d9623b30999410f9778c66fe8e242e8278147e679989287e043281592d18d" Mar 13 16:49:28 crc kubenswrapper[4907]: I0313 16:49:28.721453 4907 scope.go:117] "RemoveContainer" containerID="a8f2bc7927e473a96dafc57a193fbcc7a34307d796eb4530e9feb1c1071f99e9" Mar 13 16:49:28 crc kubenswrapper[4907]: I0313 16:49:28.727776 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-56j7g"] Mar 13 16:49:28 crc kubenswrapper[4907]: I0313 16:49:28.741734 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-56j7g"] Mar 13 16:49:28 crc kubenswrapper[4907]: I0313 16:49:28.762425 4907 scope.go:117] "RemoveContainer" containerID="dd4bda3c37e3f835e6ac87c9e4ed30b2ce48b628213cf2a3e9b4d4ba01e7b3e3" Mar 13 16:49:28 crc kubenswrapper[4907]: I0313 16:49:28.795871 4907 scope.go:117] "RemoveContainer" containerID="a76d9623b30999410f9778c66fe8e242e8278147e679989287e043281592d18d" Mar 13 16:49:28 crc kubenswrapper[4907]: E0313 16:49:28.796238 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a76d9623b30999410f9778c66fe8e242e8278147e679989287e043281592d18d\": container with ID starting with a76d9623b30999410f9778c66fe8e242e8278147e679989287e043281592d18d not found: ID does not exist" containerID="a76d9623b30999410f9778c66fe8e242e8278147e679989287e043281592d18d" Mar 13 16:49:28 crc kubenswrapper[4907]: I0313 16:49:28.796282 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a76d9623b30999410f9778c66fe8e242e8278147e679989287e043281592d18d"} err="failed to get container status \"a76d9623b30999410f9778c66fe8e242e8278147e679989287e043281592d18d\": rpc error: code = NotFound desc = could not find container \"a76d9623b30999410f9778c66fe8e242e8278147e679989287e043281592d18d\": container with ID starting with a76d9623b30999410f9778c66fe8e242e8278147e679989287e043281592d18d not found: ID does not exist" Mar 13 16:49:28 crc kubenswrapper[4907]: I0313 16:49:28.796311 4907 scope.go:117] "RemoveContainer" containerID="a8f2bc7927e473a96dafc57a193fbcc7a34307d796eb4530e9feb1c1071f99e9" Mar 13 16:49:28 crc kubenswrapper[4907]: E0313 16:49:28.796585 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a8f2bc7927e473a96dafc57a193fbcc7a34307d796eb4530e9feb1c1071f99e9\": container with ID starting with a8f2bc7927e473a96dafc57a193fbcc7a34307d796eb4530e9feb1c1071f99e9 not found: ID does not exist" containerID="a8f2bc7927e473a96dafc57a193fbcc7a34307d796eb4530e9feb1c1071f99e9" Mar 13 16:49:28 crc kubenswrapper[4907]: I0313 16:49:28.796616 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a8f2bc7927e473a96dafc57a193fbcc7a34307d796eb4530e9feb1c1071f99e9"} err="failed to get container status \"a8f2bc7927e473a96dafc57a193fbcc7a34307d796eb4530e9feb1c1071f99e9\": rpc error: code = NotFound desc = could not find container \"a8f2bc7927e473a96dafc57a193fbcc7a34307d796eb4530e9feb1c1071f99e9\": container with ID starting with a8f2bc7927e473a96dafc57a193fbcc7a34307d796eb4530e9feb1c1071f99e9 not found: ID does not exist" Mar 13 16:49:28 crc kubenswrapper[4907]: I0313 16:49:28.796638 4907 scope.go:117] "RemoveContainer" containerID="dd4bda3c37e3f835e6ac87c9e4ed30b2ce48b628213cf2a3e9b4d4ba01e7b3e3" Mar 13 16:49:28 crc kubenswrapper[4907]: E0313 16:49:28.796818 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dd4bda3c37e3f835e6ac87c9e4ed30b2ce48b628213cf2a3e9b4d4ba01e7b3e3\": container with ID starting with dd4bda3c37e3f835e6ac87c9e4ed30b2ce48b628213cf2a3e9b4d4ba01e7b3e3 not found: ID does not exist" containerID="dd4bda3c37e3f835e6ac87c9e4ed30b2ce48b628213cf2a3e9b4d4ba01e7b3e3" Mar 13 16:49:28 crc kubenswrapper[4907]: I0313 16:49:28.796842 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd4bda3c37e3f835e6ac87c9e4ed30b2ce48b628213cf2a3e9b4d4ba01e7b3e3"} err="failed to get container status \"dd4bda3c37e3f835e6ac87c9e4ed30b2ce48b628213cf2a3e9b4d4ba01e7b3e3\": rpc error: code = NotFound desc = could not find container \"dd4bda3c37e3f835e6ac87c9e4ed30b2ce48b628213cf2a3e9b4d4ba01e7b3e3\": container with ID starting with dd4bda3c37e3f835e6ac87c9e4ed30b2ce48b628213cf2a3e9b4d4ba01e7b3e3 not found: ID does not exist" Mar 13 16:49:29 crc kubenswrapper[4907]: I0313 16:49:29.797401 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8ea81195-fa1d-4422-847d-ae98208f2847" path="/var/lib/kubelet/pods/8ea81195-fa1d-4422-847d-ae98208f2847/volumes" Mar 13 16:49:34 crc kubenswrapper[4907]: I0313 16:49:34.043865 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-djph7_3a661298-271c-4c57-94ea-3d69a27d7ad5/control-plane-machine-set-operator/0.log" Mar 13 16:49:34 crc kubenswrapper[4907]: I0313 16:49:34.246337 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-prlhd_b75dbfe3-8887-4a0b-9541-a4e4000924cb/kube-rbac-proxy/0.log" Mar 13 16:49:34 crc kubenswrapper[4907]: I0313 16:49:34.258464 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-prlhd_b75dbfe3-8887-4a0b-9541-a4e4000924cb/machine-api-operator/0.log" Mar 13 16:49:48 crc kubenswrapper[4907]: I0313 16:49:48.041144 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 16:49:48 crc kubenswrapper[4907]: I0313 16:49:48.041554 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 16:49:48 crc kubenswrapper[4907]: I0313 16:49:48.217008 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-545d4d4674-rn7sp_8a238ffb-ab19-472f-bb14-580e63f0fb7d/cert-manager-controller/0.log" Mar 13 16:49:49 crc kubenswrapper[4907]: I0313 16:49:49.027076 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-6888856db4-tk9wv_3634e065-182f-49b7-af9c-01ea4779b1a6/cert-manager-webhook/0.log" Mar 13 16:49:49 crc kubenswrapper[4907]: I0313 16:49:49.044604 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-5545bd876-v92wj_e1799412-2b44-413c-8c9d-b362432fd536/cert-manager-cainjector/0.log" Mar 13 16:50:00 crc kubenswrapper[4907]: I0313 16:50:00.156558 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29557010-djwrx"] Mar 13 16:50:00 crc kubenswrapper[4907]: E0313 16:50:00.157448 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ea81195-fa1d-4422-847d-ae98208f2847" containerName="extract-utilities" Mar 13 16:50:00 crc kubenswrapper[4907]: I0313 16:50:00.157460 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ea81195-fa1d-4422-847d-ae98208f2847" containerName="extract-utilities" Mar 13 16:50:00 crc kubenswrapper[4907]: E0313 16:50:00.157485 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ea81195-fa1d-4422-847d-ae98208f2847" containerName="registry-server" Mar 13 16:50:00 crc kubenswrapper[4907]: I0313 16:50:00.157491 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ea81195-fa1d-4422-847d-ae98208f2847" containerName="registry-server" Mar 13 16:50:00 crc kubenswrapper[4907]: E0313 16:50:00.157501 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ea81195-fa1d-4422-847d-ae98208f2847" containerName="extract-content" Mar 13 16:50:00 crc kubenswrapper[4907]: I0313 16:50:00.157507 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ea81195-fa1d-4422-847d-ae98208f2847" containerName="extract-content" Mar 13 16:50:00 crc kubenswrapper[4907]: I0313 16:50:00.157711 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ea81195-fa1d-4422-847d-ae98208f2847" containerName="registry-server" Mar 13 16:50:00 crc kubenswrapper[4907]: I0313 16:50:00.158489 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29557010-djwrx" Mar 13 16:50:00 crc kubenswrapper[4907]: I0313 16:50:00.162133 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 16:50:00 crc kubenswrapper[4907]: I0313 16:50:00.162471 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 16:50:00 crc kubenswrapper[4907]: I0313 16:50:00.166232 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 16:50:00 crc kubenswrapper[4907]: I0313 16:50:00.169935 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29557010-djwrx"] Mar 13 16:50:00 crc kubenswrapper[4907]: I0313 16:50:00.368324 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mmh74\" (UniqueName: \"kubernetes.io/projected/6f57b2d7-1aae-4af4-a4fc-1820c229bbed-kube-api-access-mmh74\") pod \"auto-csr-approver-29557010-djwrx\" (UID: \"6f57b2d7-1aae-4af4-a4fc-1820c229bbed\") " pod="openshift-infra/auto-csr-approver-29557010-djwrx" Mar 13 16:50:00 crc kubenswrapper[4907]: I0313 16:50:00.470064 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mmh74\" (UniqueName: \"kubernetes.io/projected/6f57b2d7-1aae-4af4-a4fc-1820c229bbed-kube-api-access-mmh74\") pod \"auto-csr-approver-29557010-djwrx\" (UID: \"6f57b2d7-1aae-4af4-a4fc-1820c229bbed\") " pod="openshift-infra/auto-csr-approver-29557010-djwrx" Mar 13 16:50:00 crc kubenswrapper[4907]: I0313 16:50:00.508111 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mmh74\" (UniqueName: \"kubernetes.io/projected/6f57b2d7-1aae-4af4-a4fc-1820c229bbed-kube-api-access-mmh74\") pod \"auto-csr-approver-29557010-djwrx\" (UID: \"6f57b2d7-1aae-4af4-a4fc-1820c229bbed\") " pod="openshift-infra/auto-csr-approver-29557010-djwrx" Mar 13 16:50:00 crc kubenswrapper[4907]: I0313 16:50:00.519860 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29557010-djwrx" Mar 13 16:50:01 crc kubenswrapper[4907]: I0313 16:50:01.063435 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29557010-djwrx"] Mar 13 16:50:02 crc kubenswrapper[4907]: I0313 16:50:02.074092 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29557010-djwrx" event={"ID":"6f57b2d7-1aae-4af4-a4fc-1820c229bbed","Type":"ContainerStarted","Data":"8a9f4a1e6422f95bb81b0fd7a5c86e2e7db295b65270de46fb22d5b8d2e1cea2"} Mar 13 16:50:03 crc kubenswrapper[4907]: I0313 16:50:03.396084 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-86f58fcf4-wrb6h_46a0badb-4ddd-4f72-a6ae-c415a5485062/nmstate-console-plugin/0.log" Mar 13 16:50:03 crc kubenswrapper[4907]: I0313 16:50:03.575277 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-4drs7_f945fc96-0cb3-469c-ac53-8a37e0b01f51/nmstate-handler/0.log" Mar 13 16:50:03 crc kubenswrapper[4907]: I0313 16:50:03.656665 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-9b8c8685d-fnz7h_5e672ba1-22be-4fb4-b11c-3aa6a2b69c81/kube-rbac-proxy/0.log" Mar 13 16:50:03 crc kubenswrapper[4907]: I0313 16:50:03.715979 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-9b8c8685d-fnz7h_5e672ba1-22be-4fb4-b11c-3aa6a2b69c81/nmstate-metrics/0.log" Mar 13 16:50:03 crc kubenswrapper[4907]: I0313 16:50:03.806012 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-796d4cfff4-jrdzh_443989a2-3471-4c69-a460-c441d90064fe/nmstate-operator/0.log" Mar 13 16:50:03 crc kubenswrapper[4907]: I0313 16:50:03.898847 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f558f5558-cqjnn_57da19c3-e374-4e40-911a-459c1d9da8f7/nmstate-webhook/0.log" Mar 13 16:50:04 crc kubenswrapper[4907]: I0313 16:50:04.093416 4907 generic.go:334] "Generic (PLEG): container finished" podID="6f57b2d7-1aae-4af4-a4fc-1820c229bbed" containerID="f62b5f3a36b4e79cc1d6e83b544fbba0e31137d2e6efeab356e07391a7d3a1aa" exitCode=0 Mar 13 16:50:04 crc kubenswrapper[4907]: I0313 16:50:04.093640 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29557010-djwrx" event={"ID":"6f57b2d7-1aae-4af4-a4fc-1820c229bbed","Type":"ContainerDied","Data":"f62b5f3a36b4e79cc1d6e83b544fbba0e31137d2e6efeab356e07391a7d3a1aa"} Mar 13 16:50:05 crc kubenswrapper[4907]: I0313 16:50:05.501339 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29557010-djwrx" Mar 13 16:50:05 crc kubenswrapper[4907]: I0313 16:50:05.602529 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mmh74\" (UniqueName: \"kubernetes.io/projected/6f57b2d7-1aae-4af4-a4fc-1820c229bbed-kube-api-access-mmh74\") pod \"6f57b2d7-1aae-4af4-a4fc-1820c229bbed\" (UID: \"6f57b2d7-1aae-4af4-a4fc-1820c229bbed\") " Mar 13 16:50:05 crc kubenswrapper[4907]: I0313 16:50:05.610206 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6f57b2d7-1aae-4af4-a4fc-1820c229bbed-kube-api-access-mmh74" (OuterVolumeSpecName: "kube-api-access-mmh74") pod "6f57b2d7-1aae-4af4-a4fc-1820c229bbed" (UID: "6f57b2d7-1aae-4af4-a4fc-1820c229bbed"). InnerVolumeSpecName "kube-api-access-mmh74". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:50:05 crc kubenswrapper[4907]: I0313 16:50:05.705032 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mmh74\" (UniqueName: \"kubernetes.io/projected/6f57b2d7-1aae-4af4-a4fc-1820c229bbed-kube-api-access-mmh74\") on node \"crc\" DevicePath \"\"" Mar 13 16:50:06 crc kubenswrapper[4907]: I0313 16:50:06.115077 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29557010-djwrx" event={"ID":"6f57b2d7-1aae-4af4-a4fc-1820c229bbed","Type":"ContainerDied","Data":"8a9f4a1e6422f95bb81b0fd7a5c86e2e7db295b65270de46fb22d5b8d2e1cea2"} Mar 13 16:50:06 crc kubenswrapper[4907]: I0313 16:50:06.115130 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8a9f4a1e6422f95bb81b0fd7a5c86e2e7db295b65270de46fb22d5b8d2e1cea2" Mar 13 16:50:06 crc kubenswrapper[4907]: I0313 16:50:06.115198 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29557010-djwrx" Mar 13 16:50:06 crc kubenswrapper[4907]: I0313 16:50:06.588024 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29557004-v4vjf"] Mar 13 16:50:06 crc kubenswrapper[4907]: I0313 16:50:06.605003 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29557004-v4vjf"] Mar 13 16:50:07 crc kubenswrapper[4907]: I0313 16:50:07.797785 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="724a5965-1261-4da5-a833-40c8af08fde6" path="/var/lib/kubelet/pods/724a5965-1261-4da5-a833-40c8af08fde6/volumes" Mar 13 16:50:18 crc kubenswrapper[4907]: I0313 16:50:18.042041 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 16:50:18 crc kubenswrapper[4907]: I0313 16:50:18.042661 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 16:50:19 crc kubenswrapper[4907]: I0313 16:50:19.357295 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-68bc856cb9-w2ngb_92caa667-d9ee-4043-95b1-6475bcc7439b/prometheus-operator/0.log" Mar 13 16:50:19 crc kubenswrapper[4907]: I0313 16:50:19.524981 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-7dc9f6d87f-j6dtz_683c61a5-c39a-400b-b673-19bbaf286482/prometheus-operator-admission-webhook/0.log" Mar 13 16:50:19 crc kubenswrapper[4907]: I0313 16:50:19.577536 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-7dc9f6d87f-zlpcq_d188fa9a-103d-49af-9626-90b7a3a5d70c/prometheus-operator-admission-webhook/0.log" Mar 13 16:50:19 crc kubenswrapper[4907]: I0313 16:50:19.694905 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-59bdc8b94-9hfqr_dccfe9f4-59d2-4606-8d8c-a29eb01dbb6d/operator/0.log" Mar 13 16:50:19 crc kubenswrapper[4907]: I0313 16:50:19.755780 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-5bf474d74f-2n9b9_c5a33f26-edc6-4cb0-b9d8-58ee6113fecc/perses-operator/0.log" Mar 13 16:50:35 crc kubenswrapper[4907]: I0313 16:50:35.141785 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-7bb4cc7c98-2g4kt_255dfb62-fe3e-440f-82ed-1ff604426a9b/kube-rbac-proxy/0.log" Mar 13 16:50:35 crc kubenswrapper[4907]: I0313 16:50:35.384938 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6r77v_f5203ff6-d473-484d-ab8f-414e9ca804d4/cp-frr-files/0.log" Mar 13 16:50:35 crc kubenswrapper[4907]: I0313 16:50:35.552605 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-7bb4cc7c98-2g4kt_255dfb62-fe3e-440f-82ed-1ff604426a9b/controller/0.log" Mar 13 16:50:35 crc kubenswrapper[4907]: I0313 16:50:35.579039 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6r77v_f5203ff6-d473-484d-ab8f-414e9ca804d4/cp-metrics/0.log" Mar 13 16:50:35 crc kubenswrapper[4907]: I0313 16:50:35.632086 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6r77v_f5203ff6-d473-484d-ab8f-414e9ca804d4/cp-reloader/0.log" Mar 13 16:50:35 crc kubenswrapper[4907]: I0313 16:50:35.637355 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6r77v_f5203ff6-d473-484d-ab8f-414e9ca804d4/cp-frr-files/0.log" Mar 13 16:50:36 crc kubenswrapper[4907]: I0313 16:50:36.187763 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6r77v_f5203ff6-d473-484d-ab8f-414e9ca804d4/cp-reloader/0.log" Mar 13 16:50:36 crc kubenswrapper[4907]: I0313 16:50:36.412427 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6r77v_f5203ff6-d473-484d-ab8f-414e9ca804d4/cp-reloader/0.log" Mar 13 16:50:36 crc kubenswrapper[4907]: I0313 16:50:36.412544 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6r77v_f5203ff6-d473-484d-ab8f-414e9ca804d4/cp-frr-files/0.log" Mar 13 16:50:36 crc kubenswrapper[4907]: I0313 16:50:36.441237 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6r77v_f5203ff6-d473-484d-ab8f-414e9ca804d4/cp-metrics/0.log" Mar 13 16:50:36 crc kubenswrapper[4907]: I0313 16:50:36.513106 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6r77v_f5203ff6-d473-484d-ab8f-414e9ca804d4/cp-metrics/0.log" Mar 13 16:50:36 crc kubenswrapper[4907]: I0313 16:50:36.675235 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6r77v_f5203ff6-d473-484d-ab8f-414e9ca804d4/cp-frr-files/0.log" Mar 13 16:50:36 crc kubenswrapper[4907]: I0313 16:50:36.681992 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6r77v_f5203ff6-d473-484d-ab8f-414e9ca804d4/cp-reloader/0.log" Mar 13 16:50:36 crc kubenswrapper[4907]: I0313 16:50:36.696723 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6r77v_f5203ff6-d473-484d-ab8f-414e9ca804d4/cp-metrics/0.log" Mar 13 16:50:36 crc kubenswrapper[4907]: I0313 16:50:36.726351 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6r77v_f5203ff6-d473-484d-ab8f-414e9ca804d4/controller/0.log" Mar 13 16:50:36 crc kubenswrapper[4907]: I0313 16:50:36.871651 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6r77v_f5203ff6-d473-484d-ab8f-414e9ca804d4/kube-rbac-proxy/0.log" Mar 13 16:50:36 crc kubenswrapper[4907]: I0313 16:50:36.875329 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6r77v_f5203ff6-d473-484d-ab8f-414e9ca804d4/frr-metrics/0.log" Mar 13 16:50:36 crc kubenswrapper[4907]: I0313 16:50:36.959181 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6r77v_f5203ff6-d473-484d-ab8f-414e9ca804d4/kube-rbac-proxy-frr/0.log" Mar 13 16:50:37 crc kubenswrapper[4907]: I0313 16:50:37.100569 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6r77v_f5203ff6-d473-484d-ab8f-414e9ca804d4/reloader/0.log" Mar 13 16:50:37 crc kubenswrapper[4907]: I0313 16:50:37.243031 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-bcc4b6f68-jlsw9_21498ef9-03ec-44de-ade5-cc56f504499b/frr-k8s-webhook-server/0.log" Mar 13 16:50:37 crc kubenswrapper[4907]: I0313 16:50:37.380576 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-cf9f9b596-sd4ls_1c19fee6-f4bd-4488-acd9-23a4835edd83/manager/0.log" Mar 13 16:50:37 crc kubenswrapper[4907]: I0313 16:50:37.631255 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-bf6b45999-8mfb6_211e8a4b-dd3b-4669-b426-f0ac30bb3ca2/webhook-server/0.log" Mar 13 16:50:37 crc kubenswrapper[4907]: I0313 16:50:37.923169 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-fkplv_5564ca73-f820-49a4-affa-fd718d1ad7af/kube-rbac-proxy/0.log" Mar 13 16:50:38 crc kubenswrapper[4907]: I0313 16:50:38.771403 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-fkplv_5564ca73-f820-49a4-affa-fd718d1ad7af/speaker/0.log" Mar 13 16:50:40 crc kubenswrapper[4907]: I0313 16:50:40.630238 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6r77v_f5203ff6-d473-484d-ab8f-414e9ca804d4/frr/0.log" Mar 13 16:50:45 crc kubenswrapper[4907]: I0313 16:50:45.102815 4907 scope.go:117] "RemoveContainer" containerID="656ee697d354c492175b1fe789d5e08c0348dae2d403ef138ab40b1d71db2845" Mar 13 16:50:48 crc kubenswrapper[4907]: I0313 16:50:48.041942 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 16:50:48 crc kubenswrapper[4907]: I0313 16:50:48.042462 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 16:50:48 crc kubenswrapper[4907]: I0313 16:50:48.042504 4907 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" Mar 13 16:50:48 crc kubenswrapper[4907]: I0313 16:50:48.043253 4907 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"64bd48e5096c07c25a07873a406a3b7bc89c38906dc6f7c7ace8213663361dfc"} pod="openshift-machine-config-operator/machine-config-daemon-hm56j" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 16:50:48 crc kubenswrapper[4907]: I0313 16:50:48.043294 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" containerID="cri-o://64bd48e5096c07c25a07873a406a3b7bc89c38906dc6f7c7ace8213663361dfc" gracePeriod=600 Mar 13 16:50:48 crc kubenswrapper[4907]: E0313 16:50:48.441426 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:50:48 crc kubenswrapper[4907]: I0313 16:50:48.598217 4907 generic.go:334] "Generic (PLEG): container finished" podID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerID="64bd48e5096c07c25a07873a406a3b7bc89c38906dc6f7c7ace8213663361dfc" exitCode=0 Mar 13 16:50:48 crc kubenswrapper[4907]: I0313 16:50:48.598263 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerDied","Data":"64bd48e5096c07c25a07873a406a3b7bc89c38906dc6f7c7ace8213663361dfc"} Mar 13 16:50:48 crc kubenswrapper[4907]: I0313 16:50:48.598297 4907 scope.go:117] "RemoveContainer" containerID="cf78e901eff913d33c2f4e21ad1865cb4b0011402563e6e1fc5c1c3c95de10e9" Mar 13 16:50:48 crc kubenswrapper[4907]: I0313 16:50:48.599407 4907 scope.go:117] "RemoveContainer" containerID="64bd48e5096c07c25a07873a406a3b7bc89c38906dc6f7c7ace8213663361dfc" Mar 13 16:50:48 crc kubenswrapper[4907]: E0313 16:50:48.600056 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:50:53 crc kubenswrapper[4907]: I0313 16:50:53.262229 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874cgs5s_a387235c-ef5d-47cc-8dff-8af5e8b5690f/util/0.log" Mar 13 16:50:53 crc kubenswrapper[4907]: I0313 16:50:53.389469 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874cgs5s_a387235c-ef5d-47cc-8dff-8af5e8b5690f/util/0.log" Mar 13 16:50:53 crc kubenswrapper[4907]: I0313 16:50:53.447607 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874cgs5s_a387235c-ef5d-47cc-8dff-8af5e8b5690f/pull/0.log" Mar 13 16:50:53 crc kubenswrapper[4907]: I0313 16:50:53.510747 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874cgs5s_a387235c-ef5d-47cc-8dff-8af5e8b5690f/pull/0.log" Mar 13 16:50:53 crc kubenswrapper[4907]: I0313 16:50:53.677420 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874cgs5s_a387235c-ef5d-47cc-8dff-8af5e8b5690f/pull/0.log" Mar 13 16:50:53 crc kubenswrapper[4907]: I0313 16:50:53.683943 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874cgs5s_a387235c-ef5d-47cc-8dff-8af5e8b5690f/util/0.log" Mar 13 16:50:53 crc kubenswrapper[4907]: I0313 16:50:53.709955 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874cgs5s_a387235c-ef5d-47cc-8dff-8af5e8b5690f/extract/0.log" Mar 13 16:50:53 crc kubenswrapper[4907]: I0313 16:50:53.831295 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1qkjpl_385226e5-9bb7-4f71-888c-fdd2c3a59b6e/util/0.log" Mar 13 16:50:53 crc kubenswrapper[4907]: I0313 16:50:53.983758 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1qkjpl_385226e5-9bb7-4f71-888c-fdd2c3a59b6e/util/0.log" Mar 13 16:50:54 crc kubenswrapper[4907]: I0313 16:50:54.014988 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1qkjpl_385226e5-9bb7-4f71-888c-fdd2c3a59b6e/pull/0.log" Mar 13 16:50:54 crc kubenswrapper[4907]: I0313 16:50:54.034335 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1qkjpl_385226e5-9bb7-4f71-888c-fdd2c3a59b6e/pull/0.log" Mar 13 16:50:54 crc kubenswrapper[4907]: I0313 16:50:54.198340 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1qkjpl_385226e5-9bb7-4f71-888c-fdd2c3a59b6e/util/0.log" Mar 13 16:50:54 crc kubenswrapper[4907]: I0313 16:50:54.218183 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1qkjpl_385226e5-9bb7-4f71-888c-fdd2c3a59b6e/pull/0.log" Mar 13 16:50:54 crc kubenswrapper[4907]: I0313 16:50:54.220926 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1qkjpl_385226e5-9bb7-4f71-888c-fdd2c3a59b6e/extract/0.log" Mar 13 16:50:54 crc kubenswrapper[4907]: I0313 16:50:54.415812 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5snd4g_1ba163f7-0dce-42a8-ab55-15aba1940065/util/0.log" Mar 13 16:50:54 crc kubenswrapper[4907]: I0313 16:50:54.530262 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5snd4g_1ba163f7-0dce-42a8-ab55-15aba1940065/util/0.log" Mar 13 16:50:54 crc kubenswrapper[4907]: I0313 16:50:54.598258 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5snd4g_1ba163f7-0dce-42a8-ab55-15aba1940065/pull/0.log" Mar 13 16:50:54 crc kubenswrapper[4907]: I0313 16:50:54.644091 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5snd4g_1ba163f7-0dce-42a8-ab55-15aba1940065/pull/0.log" Mar 13 16:50:54 crc kubenswrapper[4907]: I0313 16:50:54.758816 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5snd4g_1ba163f7-0dce-42a8-ab55-15aba1940065/util/0.log" Mar 13 16:50:54 crc kubenswrapper[4907]: I0313 16:50:54.797237 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5snd4g_1ba163f7-0dce-42a8-ab55-15aba1940065/extract/0.log" Mar 13 16:50:54 crc kubenswrapper[4907]: I0313 16:50:54.813257 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5snd4g_1ba163f7-0dce-42a8-ab55-15aba1940065/pull/0.log" Mar 13 16:50:54 crc kubenswrapper[4907]: I0313 16:50:54.927752 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08m9qw2_d91395e4-18e5-4da8-b482-093b71c472a3/util/0.log" Mar 13 16:50:55 crc kubenswrapper[4907]: I0313 16:50:55.123366 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08m9qw2_d91395e4-18e5-4da8-b482-093b71c472a3/util/0.log" Mar 13 16:50:55 crc kubenswrapper[4907]: I0313 16:50:55.123695 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08m9qw2_d91395e4-18e5-4da8-b482-093b71c472a3/pull/0.log" Mar 13 16:50:55 crc kubenswrapper[4907]: I0313 16:50:55.139788 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08m9qw2_d91395e4-18e5-4da8-b482-093b71c472a3/pull/0.log" Mar 13 16:50:55 crc kubenswrapper[4907]: I0313 16:50:55.289820 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08m9qw2_d91395e4-18e5-4da8-b482-093b71c472a3/pull/0.log" Mar 13 16:50:55 crc kubenswrapper[4907]: I0313 16:50:55.296696 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08m9qw2_d91395e4-18e5-4da8-b482-093b71c472a3/util/0.log" Mar 13 16:50:55 crc kubenswrapper[4907]: I0313 16:50:55.307160 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08m9qw2_d91395e4-18e5-4da8-b482-093b71c472a3/extract/0.log" Mar 13 16:50:55 crc kubenswrapper[4907]: I0313 16:50:55.434540 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-slm6d_aa7e1b0d-049f-4775-bd6b-d7430126fbd3/extract-utilities/0.log" Mar 13 16:50:55 crc kubenswrapper[4907]: I0313 16:50:55.636134 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-slm6d_aa7e1b0d-049f-4775-bd6b-d7430126fbd3/extract-content/0.log" Mar 13 16:50:55 crc kubenswrapper[4907]: I0313 16:50:55.657145 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-slm6d_aa7e1b0d-049f-4775-bd6b-d7430126fbd3/extract-content/0.log" Mar 13 16:50:55 crc kubenswrapper[4907]: I0313 16:50:55.659083 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-slm6d_aa7e1b0d-049f-4775-bd6b-d7430126fbd3/extract-utilities/0.log" Mar 13 16:50:55 crc kubenswrapper[4907]: I0313 16:50:55.816552 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-slm6d_aa7e1b0d-049f-4775-bd6b-d7430126fbd3/extract-utilities/0.log" Mar 13 16:50:55 crc kubenswrapper[4907]: I0313 16:50:55.839145 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-slm6d_aa7e1b0d-049f-4775-bd6b-d7430126fbd3/extract-content/0.log" Mar 13 16:50:56 crc kubenswrapper[4907]: I0313 16:50:56.001204 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-bwznz_e3a3b609-3625-4670-b669-d7fd07386be6/extract-utilities/0.log" Mar 13 16:50:56 crc kubenswrapper[4907]: I0313 16:50:56.374245 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-bwznz_e3a3b609-3625-4670-b669-d7fd07386be6/extract-content/0.log" Mar 13 16:50:56 crc kubenswrapper[4907]: I0313 16:50:56.386136 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-bwznz_e3a3b609-3625-4670-b669-d7fd07386be6/extract-utilities/0.log" Mar 13 16:50:56 crc kubenswrapper[4907]: I0313 16:50:56.393233 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-bwznz_e3a3b609-3625-4670-b669-d7fd07386be6/extract-content/0.log" Mar 13 16:50:56 crc kubenswrapper[4907]: I0313 16:50:56.597185 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-bwznz_e3a3b609-3625-4670-b669-d7fd07386be6/extract-utilities/0.log" Mar 13 16:50:56 crc kubenswrapper[4907]: I0313 16:50:56.603551 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-bwznz_e3a3b609-3625-4670-b669-d7fd07386be6/extract-content/0.log" Mar 13 16:50:56 crc kubenswrapper[4907]: I0313 16:50:56.791285 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-h7vlt_05325ebe-cad6-417a-b1fb-af1efdf24b76/marketplace-operator/0.log" Mar 13 16:50:57 crc kubenswrapper[4907]: I0313 16:50:57.375844 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-slm6d_aa7e1b0d-049f-4775-bd6b-d7430126fbd3/registry-server/0.log" Mar 13 16:50:57 crc kubenswrapper[4907]: I0313 16:50:57.631499 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-8dzhn_a4b992e7-af9d-44a5-91ea-4942ad99bce2/extract-utilities/0.log" Mar 13 16:50:57 crc kubenswrapper[4907]: I0313 16:50:57.791688 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-8dzhn_a4b992e7-af9d-44a5-91ea-4942ad99bce2/extract-utilities/0.log" Mar 13 16:50:57 crc kubenswrapper[4907]: I0313 16:50:57.832766 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-8dzhn_a4b992e7-af9d-44a5-91ea-4942ad99bce2/extract-content/0.log" Mar 13 16:50:57 crc kubenswrapper[4907]: I0313 16:50:57.840541 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-8dzhn_a4b992e7-af9d-44a5-91ea-4942ad99bce2/extract-content/0.log" Mar 13 16:50:57 crc kubenswrapper[4907]: I0313 16:50:57.914208 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-bwznz_e3a3b609-3625-4670-b669-d7fd07386be6/registry-server/0.log" Mar 13 16:50:58 crc kubenswrapper[4907]: I0313 16:50:58.016892 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-8dzhn_a4b992e7-af9d-44a5-91ea-4942ad99bce2/extract-utilities/0.log" Mar 13 16:50:58 crc kubenswrapper[4907]: I0313 16:50:58.022687 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-8dzhn_a4b992e7-af9d-44a5-91ea-4942ad99bce2/extract-content/0.log" Mar 13 16:50:58 crc kubenswrapper[4907]: I0313 16:50:58.072340 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-z7sd2_9d354400-9ba4-45b0-a866-09da4921210a/extract-utilities/0.log" Mar 13 16:50:58 crc kubenswrapper[4907]: I0313 16:50:58.248237 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-z7sd2_9d354400-9ba4-45b0-a866-09da4921210a/extract-utilities/0.log" Mar 13 16:50:58 crc kubenswrapper[4907]: I0313 16:50:58.326756 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-z7sd2_9d354400-9ba4-45b0-a866-09da4921210a/extract-content/0.log" Mar 13 16:50:58 crc kubenswrapper[4907]: I0313 16:50:58.340061 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-z7sd2_9d354400-9ba4-45b0-a866-09da4921210a/extract-content/0.log" Mar 13 16:50:58 crc kubenswrapper[4907]: I0313 16:50:58.385485 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-8dzhn_a4b992e7-af9d-44a5-91ea-4942ad99bce2/registry-server/0.log" Mar 13 16:50:58 crc kubenswrapper[4907]: I0313 16:50:58.515779 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-z7sd2_9d354400-9ba4-45b0-a866-09da4921210a/extract-content/0.log" Mar 13 16:50:58 crc kubenswrapper[4907]: I0313 16:50:58.553066 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-z7sd2_9d354400-9ba4-45b0-a866-09da4921210a/extract-utilities/0.log" Mar 13 16:50:59 crc kubenswrapper[4907]: I0313 16:50:59.639115 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-z7sd2_9d354400-9ba4-45b0-a866-09da4921210a/registry-server/0.log" Mar 13 16:51:02 crc kubenswrapper[4907]: I0313 16:51:02.784424 4907 scope.go:117] "RemoveContainer" containerID="64bd48e5096c07c25a07873a406a3b7bc89c38906dc6f7c7ace8213663361dfc" Mar 13 16:51:02 crc kubenswrapper[4907]: E0313 16:51:02.784707 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:51:12 crc kubenswrapper[4907]: I0313 16:51:12.812463 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-68bc856cb9-w2ngb_92caa667-d9ee-4043-95b1-6475bcc7439b/prometheus-operator/0.log" Mar 13 16:51:12 crc kubenswrapper[4907]: I0313 16:51:12.842335 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-7dc9f6d87f-zlpcq_d188fa9a-103d-49af-9626-90b7a3a5d70c/prometheus-operator-admission-webhook/0.log" Mar 13 16:51:12 crc kubenswrapper[4907]: I0313 16:51:12.881366 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-7dc9f6d87f-j6dtz_683c61a5-c39a-400b-b673-19bbaf286482/prometheus-operator-admission-webhook/0.log" Mar 13 16:51:13 crc kubenswrapper[4907]: I0313 16:51:13.012608 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-59bdc8b94-9hfqr_dccfe9f4-59d2-4606-8d8c-a29eb01dbb6d/operator/0.log" Mar 13 16:51:13 crc kubenswrapper[4907]: I0313 16:51:13.021166 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-5bf474d74f-2n9b9_c5a33f26-edc6-4cb0-b9d8-58ee6113fecc/perses-operator/0.log" Mar 13 16:51:14 crc kubenswrapper[4907]: I0313 16:51:14.782573 4907 scope.go:117] "RemoveContainer" containerID="64bd48e5096c07c25a07873a406a3b7bc89c38906dc6f7c7ace8213663361dfc" Mar 13 16:51:14 crc kubenswrapper[4907]: E0313 16:51:14.783248 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:51:27 crc kubenswrapper[4907]: I0313 16:51:27.786347 4907 scope.go:117] "RemoveContainer" containerID="64bd48e5096c07c25a07873a406a3b7bc89c38906dc6f7c7ace8213663361dfc" Mar 13 16:51:27 crc kubenswrapper[4907]: E0313 16:51:27.787353 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:51:39 crc kubenswrapper[4907]: I0313 16:51:39.788112 4907 scope.go:117] "RemoveContainer" containerID="64bd48e5096c07c25a07873a406a3b7bc89c38906dc6f7c7ace8213663361dfc" Mar 13 16:51:39 crc kubenswrapper[4907]: E0313 16:51:39.789164 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:51:54 crc kubenswrapper[4907]: I0313 16:51:54.783050 4907 scope.go:117] "RemoveContainer" containerID="64bd48e5096c07c25a07873a406a3b7bc89c38906dc6f7c7ace8213663361dfc" Mar 13 16:51:54 crc kubenswrapper[4907]: E0313 16:51:54.794593 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:52:00 crc kubenswrapper[4907]: I0313 16:52:00.171522 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29557012-q7b9j"] Mar 13 16:52:00 crc kubenswrapper[4907]: E0313 16:52:00.172325 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f57b2d7-1aae-4af4-a4fc-1820c229bbed" containerName="oc" Mar 13 16:52:00 crc kubenswrapper[4907]: I0313 16:52:00.172342 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f57b2d7-1aae-4af4-a4fc-1820c229bbed" containerName="oc" Mar 13 16:52:00 crc kubenswrapper[4907]: I0313 16:52:00.172648 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="6f57b2d7-1aae-4af4-a4fc-1820c229bbed" containerName="oc" Mar 13 16:52:00 crc kubenswrapper[4907]: I0313 16:52:00.173678 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29557012-q7b9j" Mar 13 16:52:00 crc kubenswrapper[4907]: I0313 16:52:00.176740 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 16:52:00 crc kubenswrapper[4907]: I0313 16:52:00.177032 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 16:52:00 crc kubenswrapper[4907]: I0313 16:52:00.177241 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 16:52:00 crc kubenswrapper[4907]: I0313 16:52:00.183401 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29557012-q7b9j"] Mar 13 16:52:00 crc kubenswrapper[4907]: I0313 16:52:00.297964 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lg8zx\" (UniqueName: \"kubernetes.io/projected/6905e561-08a2-4e2f-a52f-35b597922bc8-kube-api-access-lg8zx\") pod \"auto-csr-approver-29557012-q7b9j\" (UID: \"6905e561-08a2-4e2f-a52f-35b597922bc8\") " pod="openshift-infra/auto-csr-approver-29557012-q7b9j" Mar 13 16:52:00 crc kubenswrapper[4907]: I0313 16:52:00.400840 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lg8zx\" (UniqueName: \"kubernetes.io/projected/6905e561-08a2-4e2f-a52f-35b597922bc8-kube-api-access-lg8zx\") pod \"auto-csr-approver-29557012-q7b9j\" (UID: \"6905e561-08a2-4e2f-a52f-35b597922bc8\") " pod="openshift-infra/auto-csr-approver-29557012-q7b9j" Mar 13 16:52:00 crc kubenswrapper[4907]: I0313 16:52:00.437149 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lg8zx\" (UniqueName: \"kubernetes.io/projected/6905e561-08a2-4e2f-a52f-35b597922bc8-kube-api-access-lg8zx\") pod \"auto-csr-approver-29557012-q7b9j\" (UID: \"6905e561-08a2-4e2f-a52f-35b597922bc8\") " pod="openshift-infra/auto-csr-approver-29557012-q7b9j" Mar 13 16:52:00 crc kubenswrapper[4907]: I0313 16:52:00.504959 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29557012-q7b9j" Mar 13 16:52:01 crc kubenswrapper[4907]: I0313 16:52:01.037543 4907 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 13 16:52:01 crc kubenswrapper[4907]: I0313 16:52:01.038785 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29557012-q7b9j"] Mar 13 16:52:01 crc kubenswrapper[4907]: I0313 16:52:01.448593 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29557012-q7b9j" event={"ID":"6905e561-08a2-4e2f-a52f-35b597922bc8","Type":"ContainerStarted","Data":"a59c71dd36dbcc7069fd86b6a8e8b0604c451db26931d99624f76ff08984d19d"} Mar 13 16:52:02 crc kubenswrapper[4907]: I0313 16:52:02.465311 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29557012-q7b9j" event={"ID":"6905e561-08a2-4e2f-a52f-35b597922bc8","Type":"ContainerStarted","Data":"ad2c3dd13afe4fd2166d55e3e7f0b8ec2cd2be8f0c8a398c03e465676f6e58e6"} Mar 13 16:52:02 crc kubenswrapper[4907]: I0313 16:52:02.503592 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29557012-q7b9j" podStartSLOduration=1.549998928 podStartE2EDuration="2.50356699s" podCreationTimestamp="2026-03-13 16:52:00 +0000 UTC" firstStartedPulling="2026-03-13 16:52:01.037140972 +0000 UTC m=+10019.936928671" lastFinishedPulling="2026-03-13 16:52:01.990709024 +0000 UTC m=+10020.890496733" observedRunningTime="2026-03-13 16:52:02.489158124 +0000 UTC m=+10021.388945853" watchObservedRunningTime="2026-03-13 16:52:02.50356699 +0000 UTC m=+10021.403354719" Mar 13 16:52:03 crc kubenswrapper[4907]: I0313 16:52:03.481022 4907 generic.go:334] "Generic (PLEG): container finished" podID="6905e561-08a2-4e2f-a52f-35b597922bc8" containerID="ad2c3dd13afe4fd2166d55e3e7f0b8ec2cd2be8f0c8a398c03e465676f6e58e6" exitCode=0 Mar 13 16:52:03 crc kubenswrapper[4907]: I0313 16:52:03.481080 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29557012-q7b9j" event={"ID":"6905e561-08a2-4e2f-a52f-35b597922bc8","Type":"ContainerDied","Data":"ad2c3dd13afe4fd2166d55e3e7f0b8ec2cd2be8f0c8a398c03e465676f6e58e6"} Mar 13 16:52:04 crc kubenswrapper[4907]: I0313 16:52:04.987845 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29557012-q7b9j" Mar 13 16:52:05 crc kubenswrapper[4907]: I0313 16:52:05.016937 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lg8zx\" (UniqueName: \"kubernetes.io/projected/6905e561-08a2-4e2f-a52f-35b597922bc8-kube-api-access-lg8zx\") pod \"6905e561-08a2-4e2f-a52f-35b597922bc8\" (UID: \"6905e561-08a2-4e2f-a52f-35b597922bc8\") " Mar 13 16:52:05 crc kubenswrapper[4907]: I0313 16:52:05.023951 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6905e561-08a2-4e2f-a52f-35b597922bc8-kube-api-access-lg8zx" (OuterVolumeSpecName: "kube-api-access-lg8zx") pod "6905e561-08a2-4e2f-a52f-35b597922bc8" (UID: "6905e561-08a2-4e2f-a52f-35b597922bc8"). InnerVolumeSpecName "kube-api-access-lg8zx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:52:05 crc kubenswrapper[4907]: I0313 16:52:05.120242 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lg8zx\" (UniqueName: \"kubernetes.io/projected/6905e561-08a2-4e2f-a52f-35b597922bc8-kube-api-access-lg8zx\") on node \"crc\" DevicePath \"\"" Mar 13 16:52:05 crc kubenswrapper[4907]: I0313 16:52:05.510740 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29557012-q7b9j" event={"ID":"6905e561-08a2-4e2f-a52f-35b597922bc8","Type":"ContainerDied","Data":"a59c71dd36dbcc7069fd86b6a8e8b0604c451db26931d99624f76ff08984d19d"} Mar 13 16:52:05 crc kubenswrapper[4907]: I0313 16:52:05.511029 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a59c71dd36dbcc7069fd86b6a8e8b0604c451db26931d99624f76ff08984d19d" Mar 13 16:52:05 crc kubenswrapper[4907]: I0313 16:52:05.510774 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29557012-q7b9j" Mar 13 16:52:06 crc kubenswrapper[4907]: I0313 16:52:06.085853 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29557006-ck2rj"] Mar 13 16:52:06 crc kubenswrapper[4907]: I0313 16:52:06.100644 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29557006-ck2rj"] Mar 13 16:52:07 crc kubenswrapper[4907]: I0313 16:52:07.782594 4907 scope.go:117] "RemoveContainer" containerID="64bd48e5096c07c25a07873a406a3b7bc89c38906dc6f7c7ace8213663361dfc" Mar 13 16:52:07 crc kubenswrapper[4907]: E0313 16:52:07.783723 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:52:07 crc kubenswrapper[4907]: I0313 16:52:07.795434 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0690b5f5-ba57-4fa5-baf4-86a729917817" path="/var/lib/kubelet/pods/0690b5f5-ba57-4fa5-baf4-86a729917817/volumes" Mar 13 16:52:18 crc kubenswrapper[4907]: I0313 16:52:18.782738 4907 scope.go:117] "RemoveContainer" containerID="64bd48e5096c07c25a07873a406a3b7bc89c38906dc6f7c7ace8213663361dfc" Mar 13 16:52:18 crc kubenswrapper[4907]: E0313 16:52:18.783518 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:52:29 crc kubenswrapper[4907]: I0313 16:52:29.783057 4907 scope.go:117] "RemoveContainer" containerID="64bd48e5096c07c25a07873a406a3b7bc89c38906dc6f7c7ace8213663361dfc" Mar 13 16:52:29 crc kubenswrapper[4907]: E0313 16:52:29.784243 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:52:43 crc kubenswrapper[4907]: I0313 16:52:43.783991 4907 scope.go:117] "RemoveContainer" containerID="64bd48e5096c07c25a07873a406a3b7bc89c38906dc6f7c7ace8213663361dfc" Mar 13 16:52:43 crc kubenswrapper[4907]: E0313 16:52:43.785016 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:52:45 crc kubenswrapper[4907]: I0313 16:52:45.228588 4907 scope.go:117] "RemoveContainer" containerID="29e87ba308a2bea52f09c17c0673a7492d85691ad7ee7034046f4e32cf2bff88" Mar 13 16:52:57 crc kubenswrapper[4907]: I0313 16:52:57.785307 4907 scope.go:117] "RemoveContainer" containerID="64bd48e5096c07c25a07873a406a3b7bc89c38906dc6f7c7ace8213663361dfc" Mar 13 16:52:57 crc kubenswrapper[4907]: E0313 16:52:57.786598 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:53:12 crc kubenswrapper[4907]: I0313 16:53:12.782796 4907 scope.go:117] "RemoveContainer" containerID="64bd48e5096c07c25a07873a406a3b7bc89c38906dc6f7c7ace8213663361dfc" Mar 13 16:53:12 crc kubenswrapper[4907]: E0313 16:53:12.783569 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:53:19 crc kubenswrapper[4907]: I0313 16:53:19.470845 4907 generic.go:334] "Generic (PLEG): container finished" podID="3d121a10-6251-43fa-af3a-0d2520c87c8d" containerID="0fd7545e78a3c44cdc8c41194a9375675f5137f4e55f454ee84c978a332aef19" exitCode=0 Mar 13 16:53:19 crc kubenswrapper[4907]: I0313 16:53:19.470949 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-xndl6/must-gather-8f2lx" event={"ID":"3d121a10-6251-43fa-af3a-0d2520c87c8d","Type":"ContainerDied","Data":"0fd7545e78a3c44cdc8c41194a9375675f5137f4e55f454ee84c978a332aef19"} Mar 13 16:53:19 crc kubenswrapper[4907]: I0313 16:53:19.472046 4907 scope.go:117] "RemoveContainer" containerID="0fd7545e78a3c44cdc8c41194a9375675f5137f4e55f454ee84c978a332aef19" Mar 13 16:53:20 crc kubenswrapper[4907]: I0313 16:53:20.390966 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-xndl6_must-gather-8f2lx_3d121a10-6251-43fa-af3a-0d2520c87c8d/gather/0.log" Mar 13 16:53:26 crc kubenswrapper[4907]: I0313 16:53:26.783147 4907 scope.go:117] "RemoveContainer" containerID="64bd48e5096c07c25a07873a406a3b7bc89c38906dc6f7c7ace8213663361dfc" Mar 13 16:53:26 crc kubenswrapper[4907]: E0313 16:53:26.784113 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:53:28 crc kubenswrapper[4907]: I0313 16:53:28.983749 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-xndl6/must-gather-8f2lx"] Mar 13 16:53:28 crc kubenswrapper[4907]: I0313 16:53:28.985567 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-xndl6/must-gather-8f2lx" podUID="3d121a10-6251-43fa-af3a-0d2520c87c8d" containerName="copy" containerID="cri-o://82747fef3c01173075fddc60efdb72434cb1285ad8c4937002af8a8d281e869f" gracePeriod=2 Mar 13 16:53:28 crc kubenswrapper[4907]: I0313 16:53:28.999031 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-xndl6/must-gather-8f2lx"] Mar 13 16:53:29 crc kubenswrapper[4907]: I0313 16:53:29.511919 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-xndl6_must-gather-8f2lx_3d121a10-6251-43fa-af3a-0d2520c87c8d/copy/0.log" Mar 13 16:53:29 crc kubenswrapper[4907]: I0313 16:53:29.515004 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xndl6/must-gather-8f2lx" Mar 13 16:53:29 crc kubenswrapper[4907]: I0313 16:53:29.548108 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/3d121a10-6251-43fa-af3a-0d2520c87c8d-must-gather-output\") pod \"3d121a10-6251-43fa-af3a-0d2520c87c8d\" (UID: \"3d121a10-6251-43fa-af3a-0d2520c87c8d\") " Mar 13 16:53:29 crc kubenswrapper[4907]: I0313 16:53:29.548304 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lsrlm\" (UniqueName: \"kubernetes.io/projected/3d121a10-6251-43fa-af3a-0d2520c87c8d-kube-api-access-lsrlm\") pod \"3d121a10-6251-43fa-af3a-0d2520c87c8d\" (UID: \"3d121a10-6251-43fa-af3a-0d2520c87c8d\") " Mar 13 16:53:29 crc kubenswrapper[4907]: I0313 16:53:29.560273 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3d121a10-6251-43fa-af3a-0d2520c87c8d-kube-api-access-lsrlm" (OuterVolumeSpecName: "kube-api-access-lsrlm") pod "3d121a10-6251-43fa-af3a-0d2520c87c8d" (UID: "3d121a10-6251-43fa-af3a-0d2520c87c8d"). InnerVolumeSpecName "kube-api-access-lsrlm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:53:29 crc kubenswrapper[4907]: I0313 16:53:29.575565 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-xndl6_must-gather-8f2lx_3d121a10-6251-43fa-af3a-0d2520c87c8d/copy/0.log" Mar 13 16:53:29 crc kubenswrapper[4907]: I0313 16:53:29.575935 4907 generic.go:334] "Generic (PLEG): container finished" podID="3d121a10-6251-43fa-af3a-0d2520c87c8d" containerID="82747fef3c01173075fddc60efdb72434cb1285ad8c4937002af8a8d281e869f" exitCode=143 Mar 13 16:53:29 crc kubenswrapper[4907]: I0313 16:53:29.576000 4907 scope.go:117] "RemoveContainer" containerID="82747fef3c01173075fddc60efdb72434cb1285ad8c4937002af8a8d281e869f" Mar 13 16:53:29 crc kubenswrapper[4907]: I0313 16:53:29.576171 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xndl6/must-gather-8f2lx" Mar 13 16:53:29 crc kubenswrapper[4907]: I0313 16:53:29.646193 4907 scope.go:117] "RemoveContainer" containerID="0fd7545e78a3c44cdc8c41194a9375675f5137f4e55f454ee84c978a332aef19" Mar 13 16:53:29 crc kubenswrapper[4907]: I0313 16:53:29.654002 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lsrlm\" (UniqueName: \"kubernetes.io/projected/3d121a10-6251-43fa-af3a-0d2520c87c8d-kube-api-access-lsrlm\") on node \"crc\" DevicePath \"\"" Mar 13 16:53:29 crc kubenswrapper[4907]: I0313 16:53:29.754146 4907 scope.go:117] "RemoveContainer" containerID="82747fef3c01173075fddc60efdb72434cb1285ad8c4937002af8a8d281e869f" Mar 13 16:53:29 crc kubenswrapper[4907]: E0313 16:53:29.754880 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"82747fef3c01173075fddc60efdb72434cb1285ad8c4937002af8a8d281e869f\": container with ID starting with 82747fef3c01173075fddc60efdb72434cb1285ad8c4937002af8a8d281e869f not found: ID does not exist" containerID="82747fef3c01173075fddc60efdb72434cb1285ad8c4937002af8a8d281e869f" Mar 13 16:53:29 crc kubenswrapper[4907]: I0313 16:53:29.754962 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"82747fef3c01173075fddc60efdb72434cb1285ad8c4937002af8a8d281e869f"} err="failed to get container status \"82747fef3c01173075fddc60efdb72434cb1285ad8c4937002af8a8d281e869f\": rpc error: code = NotFound desc = could not find container \"82747fef3c01173075fddc60efdb72434cb1285ad8c4937002af8a8d281e869f\": container with ID starting with 82747fef3c01173075fddc60efdb72434cb1285ad8c4937002af8a8d281e869f not found: ID does not exist" Mar 13 16:53:29 crc kubenswrapper[4907]: I0313 16:53:29.754984 4907 scope.go:117] "RemoveContainer" containerID="0fd7545e78a3c44cdc8c41194a9375675f5137f4e55f454ee84c978a332aef19" Mar 13 16:53:29 crc kubenswrapper[4907]: E0313 16:53:29.755339 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0fd7545e78a3c44cdc8c41194a9375675f5137f4e55f454ee84c978a332aef19\": container with ID starting with 0fd7545e78a3c44cdc8c41194a9375675f5137f4e55f454ee84c978a332aef19 not found: ID does not exist" containerID="0fd7545e78a3c44cdc8c41194a9375675f5137f4e55f454ee84c978a332aef19" Mar 13 16:53:29 crc kubenswrapper[4907]: I0313 16:53:29.755405 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0fd7545e78a3c44cdc8c41194a9375675f5137f4e55f454ee84c978a332aef19"} err="failed to get container status \"0fd7545e78a3c44cdc8c41194a9375675f5137f4e55f454ee84c978a332aef19\": rpc error: code = NotFound desc = could not find container \"0fd7545e78a3c44cdc8c41194a9375675f5137f4e55f454ee84c978a332aef19\": container with ID starting with 0fd7545e78a3c44cdc8c41194a9375675f5137f4e55f454ee84c978a332aef19 not found: ID does not exist" Mar 13 16:53:29 crc kubenswrapper[4907]: I0313 16:53:29.764212 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3d121a10-6251-43fa-af3a-0d2520c87c8d-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "3d121a10-6251-43fa-af3a-0d2520c87c8d" (UID: "3d121a10-6251-43fa-af3a-0d2520c87c8d"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 16:53:29 crc kubenswrapper[4907]: I0313 16:53:29.796121 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3d121a10-6251-43fa-af3a-0d2520c87c8d" path="/var/lib/kubelet/pods/3d121a10-6251-43fa-af3a-0d2520c87c8d/volumes" Mar 13 16:53:29 crc kubenswrapper[4907]: I0313 16:53:29.857748 4907 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/3d121a10-6251-43fa-af3a-0d2520c87c8d-must-gather-output\") on node \"crc\" DevicePath \"\"" Mar 13 16:53:40 crc kubenswrapper[4907]: I0313 16:53:40.784274 4907 scope.go:117] "RemoveContainer" containerID="64bd48e5096c07c25a07873a406a3b7bc89c38906dc6f7c7ace8213663361dfc" Mar 13 16:53:40 crc kubenswrapper[4907]: E0313 16:53:40.785977 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:53:51 crc kubenswrapper[4907]: I0313 16:53:51.800234 4907 scope.go:117] "RemoveContainer" containerID="64bd48e5096c07c25a07873a406a3b7bc89c38906dc6f7c7ace8213663361dfc" Mar 13 16:53:51 crc kubenswrapper[4907]: E0313 16:53:51.801070 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:54:00 crc kubenswrapper[4907]: I0313 16:54:00.148274 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29557014-24vtx"] Mar 13 16:54:00 crc kubenswrapper[4907]: E0313 16:54:00.149199 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d121a10-6251-43fa-af3a-0d2520c87c8d" containerName="copy" Mar 13 16:54:00 crc kubenswrapper[4907]: I0313 16:54:00.149215 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d121a10-6251-43fa-af3a-0d2520c87c8d" containerName="copy" Mar 13 16:54:00 crc kubenswrapper[4907]: E0313 16:54:00.149231 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6905e561-08a2-4e2f-a52f-35b597922bc8" containerName="oc" Mar 13 16:54:00 crc kubenswrapper[4907]: I0313 16:54:00.149237 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="6905e561-08a2-4e2f-a52f-35b597922bc8" containerName="oc" Mar 13 16:54:00 crc kubenswrapper[4907]: E0313 16:54:00.149263 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d121a10-6251-43fa-af3a-0d2520c87c8d" containerName="gather" Mar 13 16:54:00 crc kubenswrapper[4907]: I0313 16:54:00.149270 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d121a10-6251-43fa-af3a-0d2520c87c8d" containerName="gather" Mar 13 16:54:00 crc kubenswrapper[4907]: I0313 16:54:00.149498 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d121a10-6251-43fa-af3a-0d2520c87c8d" containerName="gather" Mar 13 16:54:00 crc kubenswrapper[4907]: I0313 16:54:00.149523 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d121a10-6251-43fa-af3a-0d2520c87c8d" containerName="copy" Mar 13 16:54:00 crc kubenswrapper[4907]: I0313 16:54:00.149534 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="6905e561-08a2-4e2f-a52f-35b597922bc8" containerName="oc" Mar 13 16:54:00 crc kubenswrapper[4907]: I0313 16:54:00.150238 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29557014-24vtx" Mar 13 16:54:00 crc kubenswrapper[4907]: I0313 16:54:00.152168 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 16:54:00 crc kubenswrapper[4907]: I0313 16:54:00.153286 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 16:54:00 crc kubenswrapper[4907]: I0313 16:54:00.154099 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 16:54:00 crc kubenswrapper[4907]: I0313 16:54:00.173363 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29557014-24vtx"] Mar 13 16:54:00 crc kubenswrapper[4907]: I0313 16:54:00.249195 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zpfhf\" (UniqueName: \"kubernetes.io/projected/7ca3cbcf-c046-4cdd-a742-64e7ab444440-kube-api-access-zpfhf\") pod \"auto-csr-approver-29557014-24vtx\" (UID: \"7ca3cbcf-c046-4cdd-a742-64e7ab444440\") " pod="openshift-infra/auto-csr-approver-29557014-24vtx" Mar 13 16:54:00 crc kubenswrapper[4907]: I0313 16:54:00.352079 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zpfhf\" (UniqueName: \"kubernetes.io/projected/7ca3cbcf-c046-4cdd-a742-64e7ab444440-kube-api-access-zpfhf\") pod \"auto-csr-approver-29557014-24vtx\" (UID: \"7ca3cbcf-c046-4cdd-a742-64e7ab444440\") " pod="openshift-infra/auto-csr-approver-29557014-24vtx" Mar 13 16:54:00 crc kubenswrapper[4907]: I0313 16:54:00.382701 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zpfhf\" (UniqueName: \"kubernetes.io/projected/7ca3cbcf-c046-4cdd-a742-64e7ab444440-kube-api-access-zpfhf\") pod \"auto-csr-approver-29557014-24vtx\" (UID: \"7ca3cbcf-c046-4cdd-a742-64e7ab444440\") " pod="openshift-infra/auto-csr-approver-29557014-24vtx" Mar 13 16:54:00 crc kubenswrapper[4907]: I0313 16:54:00.521693 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29557014-24vtx" Mar 13 16:54:01 crc kubenswrapper[4907]: W0313 16:54:01.074117 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7ca3cbcf_c046_4cdd_a742_64e7ab444440.slice/crio-9355002e7ae95873262837f3c5babc5ec55502ed7b0dec3570a41f0181b4e244 WatchSource:0}: Error finding container 9355002e7ae95873262837f3c5babc5ec55502ed7b0dec3570a41f0181b4e244: Status 404 returned error can't find the container with id 9355002e7ae95873262837f3c5babc5ec55502ed7b0dec3570a41f0181b4e244 Mar 13 16:54:01 crc kubenswrapper[4907]: I0313 16:54:01.102152 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29557014-24vtx"] Mar 13 16:54:01 crc kubenswrapper[4907]: I0313 16:54:01.924018 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29557014-24vtx" event={"ID":"7ca3cbcf-c046-4cdd-a742-64e7ab444440","Type":"ContainerStarted","Data":"9355002e7ae95873262837f3c5babc5ec55502ed7b0dec3570a41f0181b4e244"} Mar 13 16:54:02 crc kubenswrapper[4907]: I0313 16:54:02.939741 4907 generic.go:334] "Generic (PLEG): container finished" podID="7ca3cbcf-c046-4cdd-a742-64e7ab444440" containerID="8f3a8aa06252a3d407b2d28a76807e6a59da9309bbd4f2eb4f731bbd9ebaf501" exitCode=0 Mar 13 16:54:02 crc kubenswrapper[4907]: I0313 16:54:02.939794 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29557014-24vtx" event={"ID":"7ca3cbcf-c046-4cdd-a742-64e7ab444440","Type":"ContainerDied","Data":"8f3a8aa06252a3d407b2d28a76807e6a59da9309bbd4f2eb4f731bbd9ebaf501"} Mar 13 16:54:04 crc kubenswrapper[4907]: I0313 16:54:04.452543 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29557014-24vtx" Mar 13 16:54:04 crc kubenswrapper[4907]: I0313 16:54:04.555772 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zpfhf\" (UniqueName: \"kubernetes.io/projected/7ca3cbcf-c046-4cdd-a742-64e7ab444440-kube-api-access-zpfhf\") pod \"7ca3cbcf-c046-4cdd-a742-64e7ab444440\" (UID: \"7ca3cbcf-c046-4cdd-a742-64e7ab444440\") " Mar 13 16:54:04 crc kubenswrapper[4907]: I0313 16:54:04.582073 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ca3cbcf-c046-4cdd-a742-64e7ab444440-kube-api-access-zpfhf" (OuterVolumeSpecName: "kube-api-access-zpfhf") pod "7ca3cbcf-c046-4cdd-a742-64e7ab444440" (UID: "7ca3cbcf-c046-4cdd-a742-64e7ab444440"). InnerVolumeSpecName "kube-api-access-zpfhf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:54:04 crc kubenswrapper[4907]: I0313 16:54:04.658721 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zpfhf\" (UniqueName: \"kubernetes.io/projected/7ca3cbcf-c046-4cdd-a742-64e7ab444440-kube-api-access-zpfhf\") on node \"crc\" DevicePath \"\"" Mar 13 16:54:05 crc kubenswrapper[4907]: I0313 16:54:05.000166 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29557014-24vtx" event={"ID":"7ca3cbcf-c046-4cdd-a742-64e7ab444440","Type":"ContainerDied","Data":"9355002e7ae95873262837f3c5babc5ec55502ed7b0dec3570a41f0181b4e244"} Mar 13 16:54:05 crc kubenswrapper[4907]: I0313 16:54:05.000431 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9355002e7ae95873262837f3c5babc5ec55502ed7b0dec3570a41f0181b4e244" Mar 13 16:54:05 crc kubenswrapper[4907]: I0313 16:54:05.000492 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29557014-24vtx" Mar 13 16:54:05 crc kubenswrapper[4907]: I0313 16:54:05.530267 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29557008-tqdp7"] Mar 13 16:54:05 crc kubenswrapper[4907]: I0313 16:54:05.544625 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29557008-tqdp7"] Mar 13 16:54:05 crc kubenswrapper[4907]: I0313 16:54:05.796400 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="168ac198-4031-4bf7-bcf6-eb190b28dce6" path="/var/lib/kubelet/pods/168ac198-4031-4bf7-bcf6-eb190b28dce6/volumes" Mar 13 16:54:06 crc kubenswrapper[4907]: I0313 16:54:06.782755 4907 scope.go:117] "RemoveContainer" containerID="64bd48e5096c07c25a07873a406a3b7bc89c38906dc6f7c7ace8213663361dfc" Mar 13 16:54:06 crc kubenswrapper[4907]: E0313 16:54:06.783235 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:54:17 crc kubenswrapper[4907]: I0313 16:54:17.784093 4907 scope.go:117] "RemoveContainer" containerID="64bd48e5096c07c25a07873a406a3b7bc89c38906dc6f7c7ace8213663361dfc" Mar 13 16:54:17 crc kubenswrapper[4907]: E0313 16:54:17.784953 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:54:28 crc kubenswrapper[4907]: I0313 16:54:28.782840 4907 scope.go:117] "RemoveContainer" containerID="64bd48e5096c07c25a07873a406a3b7bc89c38906dc6f7c7ace8213663361dfc" Mar 13 16:54:28 crc kubenswrapper[4907]: E0313 16:54:28.783652 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:54:40 crc kubenswrapper[4907]: I0313 16:54:40.783216 4907 scope.go:117] "RemoveContainer" containerID="64bd48e5096c07c25a07873a406a3b7bc89c38906dc6f7c7ace8213663361dfc" Mar 13 16:54:40 crc kubenswrapper[4907]: E0313 16:54:40.784683 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:54:45 crc kubenswrapper[4907]: I0313 16:54:45.388242 4907 scope.go:117] "RemoveContainer" containerID="feb18e9c5294082e8034bc02e58f3c7d0b3c8a71601de1543217f3616b52ba31" Mar 13 16:54:54 crc kubenswrapper[4907]: I0313 16:54:54.782771 4907 scope.go:117] "RemoveContainer" containerID="64bd48e5096c07c25a07873a406a3b7bc89c38906dc6f7c7ace8213663361dfc" Mar 13 16:54:54 crc kubenswrapper[4907]: E0313 16:54:54.783755 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:55:02 crc kubenswrapper[4907]: I0313 16:55:02.665786 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-xzwk8"] Mar 13 16:55:02 crc kubenswrapper[4907]: E0313 16:55:02.667569 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ca3cbcf-c046-4cdd-a742-64e7ab444440" containerName="oc" Mar 13 16:55:02 crc kubenswrapper[4907]: I0313 16:55:02.667595 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ca3cbcf-c046-4cdd-a742-64e7ab444440" containerName="oc" Mar 13 16:55:02 crc kubenswrapper[4907]: I0313 16:55:02.668035 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ca3cbcf-c046-4cdd-a742-64e7ab444440" containerName="oc" Mar 13 16:55:02 crc kubenswrapper[4907]: I0313 16:55:02.672710 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xzwk8" Mar 13 16:55:02 crc kubenswrapper[4907]: I0313 16:55:02.693446 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xzwk8"] Mar 13 16:55:02 crc kubenswrapper[4907]: I0313 16:55:02.754132 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8474b846-1637-4f0b-ae93-30785f6337dc-catalog-content\") pod \"certified-operators-xzwk8\" (UID: \"8474b846-1637-4f0b-ae93-30785f6337dc\") " pod="openshift-marketplace/certified-operators-xzwk8" Mar 13 16:55:02 crc kubenswrapper[4907]: I0313 16:55:02.754344 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8474b846-1637-4f0b-ae93-30785f6337dc-utilities\") pod \"certified-operators-xzwk8\" (UID: \"8474b846-1637-4f0b-ae93-30785f6337dc\") " pod="openshift-marketplace/certified-operators-xzwk8" Mar 13 16:55:02 crc kubenswrapper[4907]: I0313 16:55:02.754656 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6sbl2\" (UniqueName: \"kubernetes.io/projected/8474b846-1637-4f0b-ae93-30785f6337dc-kube-api-access-6sbl2\") pod \"certified-operators-xzwk8\" (UID: \"8474b846-1637-4f0b-ae93-30785f6337dc\") " pod="openshift-marketplace/certified-operators-xzwk8" Mar 13 16:55:02 crc kubenswrapper[4907]: I0313 16:55:02.857464 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8474b846-1637-4f0b-ae93-30785f6337dc-utilities\") pod \"certified-operators-xzwk8\" (UID: \"8474b846-1637-4f0b-ae93-30785f6337dc\") " pod="openshift-marketplace/certified-operators-xzwk8" Mar 13 16:55:02 crc kubenswrapper[4907]: I0313 16:55:02.857732 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6sbl2\" (UniqueName: \"kubernetes.io/projected/8474b846-1637-4f0b-ae93-30785f6337dc-kube-api-access-6sbl2\") pod \"certified-operators-xzwk8\" (UID: \"8474b846-1637-4f0b-ae93-30785f6337dc\") " pod="openshift-marketplace/certified-operators-xzwk8" Mar 13 16:55:02 crc kubenswrapper[4907]: I0313 16:55:02.857787 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8474b846-1637-4f0b-ae93-30785f6337dc-catalog-content\") pod \"certified-operators-xzwk8\" (UID: \"8474b846-1637-4f0b-ae93-30785f6337dc\") " pod="openshift-marketplace/certified-operators-xzwk8" Mar 13 16:55:02 crc kubenswrapper[4907]: I0313 16:55:02.858257 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8474b846-1637-4f0b-ae93-30785f6337dc-utilities\") pod \"certified-operators-xzwk8\" (UID: \"8474b846-1637-4f0b-ae93-30785f6337dc\") " pod="openshift-marketplace/certified-operators-xzwk8" Mar 13 16:55:02 crc kubenswrapper[4907]: I0313 16:55:02.858433 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8474b846-1637-4f0b-ae93-30785f6337dc-catalog-content\") pod \"certified-operators-xzwk8\" (UID: \"8474b846-1637-4f0b-ae93-30785f6337dc\") " pod="openshift-marketplace/certified-operators-xzwk8" Mar 13 16:55:02 crc kubenswrapper[4907]: I0313 16:55:02.895951 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6sbl2\" (UniqueName: \"kubernetes.io/projected/8474b846-1637-4f0b-ae93-30785f6337dc-kube-api-access-6sbl2\") pod \"certified-operators-xzwk8\" (UID: \"8474b846-1637-4f0b-ae93-30785f6337dc\") " pod="openshift-marketplace/certified-operators-xzwk8" Mar 13 16:55:03 crc kubenswrapper[4907]: I0313 16:55:03.000289 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xzwk8" Mar 13 16:55:03 crc kubenswrapper[4907]: I0313 16:55:03.521662 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xzwk8"] Mar 13 16:55:03 crc kubenswrapper[4907]: I0313 16:55:03.700740 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xzwk8" event={"ID":"8474b846-1637-4f0b-ae93-30785f6337dc","Type":"ContainerStarted","Data":"2392fdf1246963da40e3bc332c1feebc95f83855b723d4fe483e5234daceb34c"} Mar 13 16:55:04 crc kubenswrapper[4907]: I0313 16:55:04.717482 4907 generic.go:334] "Generic (PLEG): container finished" podID="8474b846-1637-4f0b-ae93-30785f6337dc" containerID="74b82b6c06cc647d294be0270216e67f5bc0cee17181904432483ea8365c78ed" exitCode=0 Mar 13 16:55:04 crc kubenswrapper[4907]: I0313 16:55:04.717563 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xzwk8" event={"ID":"8474b846-1637-4f0b-ae93-30785f6337dc","Type":"ContainerDied","Data":"74b82b6c06cc647d294be0270216e67f5bc0cee17181904432483ea8365c78ed"} Mar 13 16:55:07 crc kubenswrapper[4907]: I0313 16:55:07.760472 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xzwk8" event={"ID":"8474b846-1637-4f0b-ae93-30785f6337dc","Type":"ContainerStarted","Data":"a5bc0b77a5a8529773966a6df2adc32a92403a3706d570dbec6f040f99237483"} Mar 13 16:55:07 crc kubenswrapper[4907]: I0313 16:55:07.783652 4907 scope.go:117] "RemoveContainer" containerID="64bd48e5096c07c25a07873a406a3b7bc89c38906dc6f7c7ace8213663361dfc" Mar 13 16:55:07 crc kubenswrapper[4907]: E0313 16:55:07.784162 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:55:09 crc kubenswrapper[4907]: I0313 16:55:09.792697 4907 generic.go:334] "Generic (PLEG): container finished" podID="8474b846-1637-4f0b-ae93-30785f6337dc" containerID="a5bc0b77a5a8529773966a6df2adc32a92403a3706d570dbec6f040f99237483" exitCode=0 Mar 13 16:55:09 crc kubenswrapper[4907]: I0313 16:55:09.801994 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xzwk8" event={"ID":"8474b846-1637-4f0b-ae93-30785f6337dc","Type":"ContainerDied","Data":"a5bc0b77a5a8529773966a6df2adc32a92403a3706d570dbec6f040f99237483"} Mar 13 16:55:10 crc kubenswrapper[4907]: I0313 16:55:10.807974 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xzwk8" event={"ID":"8474b846-1637-4f0b-ae93-30785f6337dc","Type":"ContainerStarted","Data":"76d491557290f49b0a6e5431483402886ab839aaf20944d06fc295bad593294c"} Mar 13 16:55:10 crc kubenswrapper[4907]: I0313 16:55:10.848180 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-xzwk8" podStartSLOduration=3.300704412 podStartE2EDuration="8.848153312s" podCreationTimestamp="2026-03-13 16:55:02 +0000 UTC" firstStartedPulling="2026-03-13 16:55:04.721147277 +0000 UTC m=+10203.620935006" lastFinishedPulling="2026-03-13 16:55:10.268596207 +0000 UTC m=+10209.168383906" observedRunningTime="2026-03-13 16:55:10.828181873 +0000 UTC m=+10209.727969582" watchObservedRunningTime="2026-03-13 16:55:10.848153312 +0000 UTC m=+10209.747941041" Mar 13 16:55:13 crc kubenswrapper[4907]: I0313 16:55:13.001042 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-xzwk8" Mar 13 16:55:13 crc kubenswrapper[4907]: I0313 16:55:13.001668 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-xzwk8" Mar 13 16:55:13 crc kubenswrapper[4907]: I0313 16:55:13.091154 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-xzwk8" Mar 13 16:55:21 crc kubenswrapper[4907]: I0313 16:55:21.789211 4907 scope.go:117] "RemoveContainer" containerID="64bd48e5096c07c25a07873a406a3b7bc89c38906dc6f7c7ace8213663361dfc" Mar 13 16:55:21 crc kubenswrapper[4907]: E0313 16:55:21.790189 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:55:23 crc kubenswrapper[4907]: I0313 16:55:23.691151 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-xzwk8" Mar 13 16:55:23 crc kubenswrapper[4907]: I0313 16:55:23.739585 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xzwk8"] Mar 13 16:55:23 crc kubenswrapper[4907]: I0313 16:55:23.956984 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-xzwk8" podUID="8474b846-1637-4f0b-ae93-30785f6337dc" containerName="registry-server" containerID="cri-o://76d491557290f49b0a6e5431483402886ab839aaf20944d06fc295bad593294c" gracePeriod=2 Mar 13 16:55:24 crc kubenswrapper[4907]: I0313 16:55:24.472530 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xzwk8" Mar 13 16:55:24 crc kubenswrapper[4907]: I0313 16:55:24.575291 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6sbl2\" (UniqueName: \"kubernetes.io/projected/8474b846-1637-4f0b-ae93-30785f6337dc-kube-api-access-6sbl2\") pod \"8474b846-1637-4f0b-ae93-30785f6337dc\" (UID: \"8474b846-1637-4f0b-ae93-30785f6337dc\") " Mar 13 16:55:24 crc kubenswrapper[4907]: I0313 16:55:24.575702 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8474b846-1637-4f0b-ae93-30785f6337dc-catalog-content\") pod \"8474b846-1637-4f0b-ae93-30785f6337dc\" (UID: \"8474b846-1637-4f0b-ae93-30785f6337dc\") " Mar 13 16:55:24 crc kubenswrapper[4907]: I0313 16:55:24.575767 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8474b846-1637-4f0b-ae93-30785f6337dc-utilities\") pod \"8474b846-1637-4f0b-ae93-30785f6337dc\" (UID: \"8474b846-1637-4f0b-ae93-30785f6337dc\") " Mar 13 16:55:24 crc kubenswrapper[4907]: I0313 16:55:24.576980 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8474b846-1637-4f0b-ae93-30785f6337dc-utilities" (OuterVolumeSpecName: "utilities") pod "8474b846-1637-4f0b-ae93-30785f6337dc" (UID: "8474b846-1637-4f0b-ae93-30785f6337dc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 16:55:24 crc kubenswrapper[4907]: I0313 16:55:24.585633 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8474b846-1637-4f0b-ae93-30785f6337dc-kube-api-access-6sbl2" (OuterVolumeSpecName: "kube-api-access-6sbl2") pod "8474b846-1637-4f0b-ae93-30785f6337dc" (UID: "8474b846-1637-4f0b-ae93-30785f6337dc"). InnerVolumeSpecName "kube-api-access-6sbl2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:55:24 crc kubenswrapper[4907]: I0313 16:55:24.634000 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8474b846-1637-4f0b-ae93-30785f6337dc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8474b846-1637-4f0b-ae93-30785f6337dc" (UID: "8474b846-1637-4f0b-ae93-30785f6337dc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 16:55:24 crc kubenswrapper[4907]: I0313 16:55:24.679607 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8474b846-1637-4f0b-ae93-30785f6337dc-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 16:55:24 crc kubenswrapper[4907]: I0313 16:55:24.679642 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8474b846-1637-4f0b-ae93-30785f6337dc-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 16:55:24 crc kubenswrapper[4907]: I0313 16:55:24.679653 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6sbl2\" (UniqueName: \"kubernetes.io/projected/8474b846-1637-4f0b-ae93-30785f6337dc-kube-api-access-6sbl2\") on node \"crc\" DevicePath \"\"" Mar 13 16:55:24 crc kubenswrapper[4907]: I0313 16:55:24.968383 4907 generic.go:334] "Generic (PLEG): container finished" podID="8474b846-1637-4f0b-ae93-30785f6337dc" containerID="76d491557290f49b0a6e5431483402886ab839aaf20944d06fc295bad593294c" exitCode=0 Mar 13 16:55:24 crc kubenswrapper[4907]: I0313 16:55:24.968440 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xzwk8" event={"ID":"8474b846-1637-4f0b-ae93-30785f6337dc","Type":"ContainerDied","Data":"76d491557290f49b0a6e5431483402886ab839aaf20944d06fc295bad593294c"} Mar 13 16:55:24 crc kubenswrapper[4907]: I0313 16:55:24.968465 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xzwk8" event={"ID":"8474b846-1637-4f0b-ae93-30785f6337dc","Type":"ContainerDied","Data":"2392fdf1246963da40e3bc332c1feebc95f83855b723d4fe483e5234daceb34c"} Mar 13 16:55:24 crc kubenswrapper[4907]: I0313 16:55:24.968484 4907 scope.go:117] "RemoveContainer" containerID="76d491557290f49b0a6e5431483402886ab839aaf20944d06fc295bad593294c" Mar 13 16:55:24 crc kubenswrapper[4907]: I0313 16:55:24.968640 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xzwk8" Mar 13 16:55:25 crc kubenswrapper[4907]: I0313 16:55:25.003502 4907 scope.go:117] "RemoveContainer" containerID="a5bc0b77a5a8529773966a6df2adc32a92403a3706d570dbec6f040f99237483" Mar 13 16:55:25 crc kubenswrapper[4907]: I0313 16:55:25.016364 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xzwk8"] Mar 13 16:55:25 crc kubenswrapper[4907]: I0313 16:55:25.032053 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-xzwk8"] Mar 13 16:55:25 crc kubenswrapper[4907]: I0313 16:55:25.035145 4907 scope.go:117] "RemoveContainer" containerID="74b82b6c06cc647d294be0270216e67f5bc0cee17181904432483ea8365c78ed" Mar 13 16:55:25 crc kubenswrapper[4907]: I0313 16:55:25.071419 4907 scope.go:117] "RemoveContainer" containerID="76d491557290f49b0a6e5431483402886ab839aaf20944d06fc295bad593294c" Mar 13 16:55:25 crc kubenswrapper[4907]: E0313 16:55:25.071809 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"76d491557290f49b0a6e5431483402886ab839aaf20944d06fc295bad593294c\": container with ID starting with 76d491557290f49b0a6e5431483402886ab839aaf20944d06fc295bad593294c not found: ID does not exist" containerID="76d491557290f49b0a6e5431483402886ab839aaf20944d06fc295bad593294c" Mar 13 16:55:25 crc kubenswrapper[4907]: I0313 16:55:25.071848 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"76d491557290f49b0a6e5431483402886ab839aaf20944d06fc295bad593294c"} err="failed to get container status \"76d491557290f49b0a6e5431483402886ab839aaf20944d06fc295bad593294c\": rpc error: code = NotFound desc = could not find container \"76d491557290f49b0a6e5431483402886ab839aaf20944d06fc295bad593294c\": container with ID starting with 76d491557290f49b0a6e5431483402886ab839aaf20944d06fc295bad593294c not found: ID does not exist" Mar 13 16:55:25 crc kubenswrapper[4907]: I0313 16:55:25.071894 4907 scope.go:117] "RemoveContainer" containerID="a5bc0b77a5a8529773966a6df2adc32a92403a3706d570dbec6f040f99237483" Mar 13 16:55:25 crc kubenswrapper[4907]: E0313 16:55:25.073057 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a5bc0b77a5a8529773966a6df2adc32a92403a3706d570dbec6f040f99237483\": container with ID starting with a5bc0b77a5a8529773966a6df2adc32a92403a3706d570dbec6f040f99237483 not found: ID does not exist" containerID="a5bc0b77a5a8529773966a6df2adc32a92403a3706d570dbec6f040f99237483" Mar 13 16:55:25 crc kubenswrapper[4907]: I0313 16:55:25.073151 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a5bc0b77a5a8529773966a6df2adc32a92403a3706d570dbec6f040f99237483"} err="failed to get container status \"a5bc0b77a5a8529773966a6df2adc32a92403a3706d570dbec6f040f99237483\": rpc error: code = NotFound desc = could not find container \"a5bc0b77a5a8529773966a6df2adc32a92403a3706d570dbec6f040f99237483\": container with ID starting with a5bc0b77a5a8529773966a6df2adc32a92403a3706d570dbec6f040f99237483 not found: ID does not exist" Mar 13 16:55:25 crc kubenswrapper[4907]: I0313 16:55:25.073173 4907 scope.go:117] "RemoveContainer" containerID="74b82b6c06cc647d294be0270216e67f5bc0cee17181904432483ea8365c78ed" Mar 13 16:55:25 crc kubenswrapper[4907]: E0313 16:55:25.073515 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"74b82b6c06cc647d294be0270216e67f5bc0cee17181904432483ea8365c78ed\": container with ID starting with 74b82b6c06cc647d294be0270216e67f5bc0cee17181904432483ea8365c78ed not found: ID does not exist" containerID="74b82b6c06cc647d294be0270216e67f5bc0cee17181904432483ea8365c78ed" Mar 13 16:55:25 crc kubenswrapper[4907]: I0313 16:55:25.073555 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"74b82b6c06cc647d294be0270216e67f5bc0cee17181904432483ea8365c78ed"} err="failed to get container status \"74b82b6c06cc647d294be0270216e67f5bc0cee17181904432483ea8365c78ed\": rpc error: code = NotFound desc = could not find container \"74b82b6c06cc647d294be0270216e67f5bc0cee17181904432483ea8365c78ed\": container with ID starting with 74b82b6c06cc647d294be0270216e67f5bc0cee17181904432483ea8365c78ed not found: ID does not exist" Mar 13 16:55:25 crc kubenswrapper[4907]: I0313 16:55:25.798718 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8474b846-1637-4f0b-ae93-30785f6337dc" path="/var/lib/kubelet/pods/8474b846-1637-4f0b-ae93-30785f6337dc/volumes" Mar 13 16:55:32 crc kubenswrapper[4907]: I0313 16:55:32.782695 4907 scope.go:117] "RemoveContainer" containerID="64bd48e5096c07c25a07873a406a3b7bc89c38906dc6f7c7ace8213663361dfc" Mar 13 16:55:32 crc kubenswrapper[4907]: E0313 16:55:32.784314 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:55:44 crc kubenswrapper[4907]: I0313 16:55:44.782527 4907 scope.go:117] "RemoveContainer" containerID="64bd48e5096c07c25a07873a406a3b7bc89c38906dc6f7c7ace8213663361dfc" Mar 13 16:55:44 crc kubenswrapper[4907]: E0313 16:55:44.783338 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:55:49 crc kubenswrapper[4907]: I0313 16:55:49.917947 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-mhrkz"] Mar 13 16:55:49 crc kubenswrapper[4907]: E0313 16:55:49.918811 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8474b846-1637-4f0b-ae93-30785f6337dc" containerName="registry-server" Mar 13 16:55:49 crc kubenswrapper[4907]: I0313 16:55:49.918822 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="8474b846-1637-4f0b-ae93-30785f6337dc" containerName="registry-server" Mar 13 16:55:49 crc kubenswrapper[4907]: E0313 16:55:49.918845 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8474b846-1637-4f0b-ae93-30785f6337dc" containerName="extract-content" Mar 13 16:55:49 crc kubenswrapper[4907]: I0313 16:55:49.918851 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="8474b846-1637-4f0b-ae93-30785f6337dc" containerName="extract-content" Mar 13 16:55:49 crc kubenswrapper[4907]: E0313 16:55:49.918915 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8474b846-1637-4f0b-ae93-30785f6337dc" containerName="extract-utilities" Mar 13 16:55:49 crc kubenswrapper[4907]: I0313 16:55:49.918925 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="8474b846-1637-4f0b-ae93-30785f6337dc" containerName="extract-utilities" Mar 13 16:55:49 crc kubenswrapper[4907]: I0313 16:55:49.919121 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="8474b846-1637-4f0b-ae93-30785f6337dc" containerName="registry-server" Mar 13 16:55:49 crc kubenswrapper[4907]: I0313 16:55:49.920604 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mhrkz" Mar 13 16:55:49 crc kubenswrapper[4907]: I0313 16:55:49.936450 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mhrkz"] Mar 13 16:55:49 crc kubenswrapper[4907]: I0313 16:55:49.997025 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4fbc8056-d73e-49f8-b94b-e1ca9827bda8-utilities\") pod \"community-operators-mhrkz\" (UID: \"4fbc8056-d73e-49f8-b94b-e1ca9827bda8\") " pod="openshift-marketplace/community-operators-mhrkz" Mar 13 16:55:49 crc kubenswrapper[4907]: I0313 16:55:49.997075 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jr9pw\" (UniqueName: \"kubernetes.io/projected/4fbc8056-d73e-49f8-b94b-e1ca9827bda8-kube-api-access-jr9pw\") pod \"community-operators-mhrkz\" (UID: \"4fbc8056-d73e-49f8-b94b-e1ca9827bda8\") " pod="openshift-marketplace/community-operators-mhrkz" Mar 13 16:55:49 crc kubenswrapper[4907]: I0313 16:55:49.997147 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4fbc8056-d73e-49f8-b94b-e1ca9827bda8-catalog-content\") pod \"community-operators-mhrkz\" (UID: \"4fbc8056-d73e-49f8-b94b-e1ca9827bda8\") " pod="openshift-marketplace/community-operators-mhrkz" Mar 13 16:55:50 crc kubenswrapper[4907]: I0313 16:55:50.099712 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4fbc8056-d73e-49f8-b94b-e1ca9827bda8-catalog-content\") pod \"community-operators-mhrkz\" (UID: \"4fbc8056-d73e-49f8-b94b-e1ca9827bda8\") " pod="openshift-marketplace/community-operators-mhrkz" Mar 13 16:55:50 crc kubenswrapper[4907]: I0313 16:55:50.099912 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4fbc8056-d73e-49f8-b94b-e1ca9827bda8-utilities\") pod \"community-operators-mhrkz\" (UID: \"4fbc8056-d73e-49f8-b94b-e1ca9827bda8\") " pod="openshift-marketplace/community-operators-mhrkz" Mar 13 16:55:50 crc kubenswrapper[4907]: I0313 16:55:50.099963 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jr9pw\" (UniqueName: \"kubernetes.io/projected/4fbc8056-d73e-49f8-b94b-e1ca9827bda8-kube-api-access-jr9pw\") pod \"community-operators-mhrkz\" (UID: \"4fbc8056-d73e-49f8-b94b-e1ca9827bda8\") " pod="openshift-marketplace/community-operators-mhrkz" Mar 13 16:55:50 crc kubenswrapper[4907]: I0313 16:55:50.100274 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4fbc8056-d73e-49f8-b94b-e1ca9827bda8-catalog-content\") pod \"community-operators-mhrkz\" (UID: \"4fbc8056-d73e-49f8-b94b-e1ca9827bda8\") " pod="openshift-marketplace/community-operators-mhrkz" Mar 13 16:55:50 crc kubenswrapper[4907]: I0313 16:55:50.100337 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4fbc8056-d73e-49f8-b94b-e1ca9827bda8-utilities\") pod \"community-operators-mhrkz\" (UID: \"4fbc8056-d73e-49f8-b94b-e1ca9827bda8\") " pod="openshift-marketplace/community-operators-mhrkz" Mar 13 16:55:50 crc kubenswrapper[4907]: I0313 16:55:50.122785 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jr9pw\" (UniqueName: \"kubernetes.io/projected/4fbc8056-d73e-49f8-b94b-e1ca9827bda8-kube-api-access-jr9pw\") pod \"community-operators-mhrkz\" (UID: \"4fbc8056-d73e-49f8-b94b-e1ca9827bda8\") " pod="openshift-marketplace/community-operators-mhrkz" Mar 13 16:55:50 crc kubenswrapper[4907]: I0313 16:55:50.245654 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mhrkz" Mar 13 16:55:51 crc kubenswrapper[4907]: I0313 16:55:51.473428 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mhrkz"] Mar 13 16:55:52 crc kubenswrapper[4907]: I0313 16:55:52.318113 4907 generic.go:334] "Generic (PLEG): container finished" podID="4fbc8056-d73e-49f8-b94b-e1ca9827bda8" containerID="ae74208e85dfe58e9cd0fd4f63b388596e592e48c797f92d20005d4e46b286f8" exitCode=0 Mar 13 16:55:52 crc kubenswrapper[4907]: I0313 16:55:52.318459 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mhrkz" event={"ID":"4fbc8056-d73e-49f8-b94b-e1ca9827bda8","Type":"ContainerDied","Data":"ae74208e85dfe58e9cd0fd4f63b388596e592e48c797f92d20005d4e46b286f8"} Mar 13 16:55:52 crc kubenswrapper[4907]: I0313 16:55:52.318604 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mhrkz" event={"ID":"4fbc8056-d73e-49f8-b94b-e1ca9827bda8","Type":"ContainerStarted","Data":"a34268e545f27c48c28686dd73a60c951164de53a324312e016b297fe5122100"} Mar 13 16:55:52 crc kubenswrapper[4907]: I0313 16:55:52.648755 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-4zr5z"] Mar 13 16:55:52 crc kubenswrapper[4907]: I0313 16:55:52.650986 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4zr5z" Mar 13 16:55:52 crc kubenswrapper[4907]: I0313 16:55:52.662590 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4zr5z"] Mar 13 16:55:52 crc kubenswrapper[4907]: I0313 16:55:52.762945 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gg57d\" (UniqueName: \"kubernetes.io/projected/40c3d47f-e20b-47a7-be46-a7d30d573327-kube-api-access-gg57d\") pod \"redhat-operators-4zr5z\" (UID: \"40c3d47f-e20b-47a7-be46-a7d30d573327\") " pod="openshift-marketplace/redhat-operators-4zr5z" Mar 13 16:55:52 crc kubenswrapper[4907]: I0313 16:55:52.763134 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/40c3d47f-e20b-47a7-be46-a7d30d573327-catalog-content\") pod \"redhat-operators-4zr5z\" (UID: \"40c3d47f-e20b-47a7-be46-a7d30d573327\") " pod="openshift-marketplace/redhat-operators-4zr5z" Mar 13 16:55:52 crc kubenswrapper[4907]: I0313 16:55:52.763165 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/40c3d47f-e20b-47a7-be46-a7d30d573327-utilities\") pod \"redhat-operators-4zr5z\" (UID: \"40c3d47f-e20b-47a7-be46-a7d30d573327\") " pod="openshift-marketplace/redhat-operators-4zr5z" Mar 13 16:55:52 crc kubenswrapper[4907]: I0313 16:55:52.864961 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gg57d\" (UniqueName: \"kubernetes.io/projected/40c3d47f-e20b-47a7-be46-a7d30d573327-kube-api-access-gg57d\") pod \"redhat-operators-4zr5z\" (UID: \"40c3d47f-e20b-47a7-be46-a7d30d573327\") " pod="openshift-marketplace/redhat-operators-4zr5z" Mar 13 16:55:52 crc kubenswrapper[4907]: I0313 16:55:52.865156 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/40c3d47f-e20b-47a7-be46-a7d30d573327-catalog-content\") pod \"redhat-operators-4zr5z\" (UID: \"40c3d47f-e20b-47a7-be46-a7d30d573327\") " pod="openshift-marketplace/redhat-operators-4zr5z" Mar 13 16:55:52 crc kubenswrapper[4907]: I0313 16:55:52.865183 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/40c3d47f-e20b-47a7-be46-a7d30d573327-utilities\") pod \"redhat-operators-4zr5z\" (UID: \"40c3d47f-e20b-47a7-be46-a7d30d573327\") " pod="openshift-marketplace/redhat-operators-4zr5z" Mar 13 16:55:52 crc kubenswrapper[4907]: I0313 16:55:52.866630 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/40c3d47f-e20b-47a7-be46-a7d30d573327-catalog-content\") pod \"redhat-operators-4zr5z\" (UID: \"40c3d47f-e20b-47a7-be46-a7d30d573327\") " pod="openshift-marketplace/redhat-operators-4zr5z" Mar 13 16:55:52 crc kubenswrapper[4907]: I0313 16:55:52.866954 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/40c3d47f-e20b-47a7-be46-a7d30d573327-utilities\") pod \"redhat-operators-4zr5z\" (UID: \"40c3d47f-e20b-47a7-be46-a7d30d573327\") " pod="openshift-marketplace/redhat-operators-4zr5z" Mar 13 16:55:53 crc kubenswrapper[4907]: I0313 16:55:53.106874 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gg57d\" (UniqueName: \"kubernetes.io/projected/40c3d47f-e20b-47a7-be46-a7d30d573327-kube-api-access-gg57d\") pod \"redhat-operators-4zr5z\" (UID: \"40c3d47f-e20b-47a7-be46-a7d30d573327\") " pod="openshift-marketplace/redhat-operators-4zr5z" Mar 13 16:55:53 crc kubenswrapper[4907]: I0313 16:55:53.322622 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4zr5z" Mar 13 16:55:53 crc kubenswrapper[4907]: I0313 16:55:53.837290 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4zr5z"] Mar 13 16:55:53 crc kubenswrapper[4907]: W0313 16:55:53.845620 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod40c3d47f_e20b_47a7_be46_a7d30d573327.slice/crio-102dad920f7c3e14da000819f2f576fdcf93e506c017ecabf4d72e0cdc9c0160 WatchSource:0}: Error finding container 102dad920f7c3e14da000819f2f576fdcf93e506c017ecabf4d72e0cdc9c0160: Status 404 returned error can't find the container with id 102dad920f7c3e14da000819f2f576fdcf93e506c017ecabf4d72e0cdc9c0160 Mar 13 16:55:54 crc kubenswrapper[4907]: I0313 16:55:54.345046 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mhrkz" event={"ID":"4fbc8056-d73e-49f8-b94b-e1ca9827bda8","Type":"ContainerStarted","Data":"92c11c843af8e1cea03f882c89789b4e329f162f9479f6575e5fd03a8edc069c"} Mar 13 16:55:54 crc kubenswrapper[4907]: I0313 16:55:54.348407 4907 generic.go:334] "Generic (PLEG): container finished" podID="40c3d47f-e20b-47a7-be46-a7d30d573327" containerID="48b68600daee932b5c08e4d390a35167c48346ec15600f7cf99e2ed2d675ff09" exitCode=0 Mar 13 16:55:54 crc kubenswrapper[4907]: I0313 16:55:54.348459 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4zr5z" event={"ID":"40c3d47f-e20b-47a7-be46-a7d30d573327","Type":"ContainerDied","Data":"48b68600daee932b5c08e4d390a35167c48346ec15600f7cf99e2ed2d675ff09"} Mar 13 16:55:54 crc kubenswrapper[4907]: I0313 16:55:54.348494 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4zr5z" event={"ID":"40c3d47f-e20b-47a7-be46-a7d30d573327","Type":"ContainerStarted","Data":"102dad920f7c3e14da000819f2f576fdcf93e506c017ecabf4d72e0cdc9c0160"} Mar 13 16:55:55 crc kubenswrapper[4907]: I0313 16:55:55.364393 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4zr5z" event={"ID":"40c3d47f-e20b-47a7-be46-a7d30d573327","Type":"ContainerStarted","Data":"f8ca9d7c44d6fd79031fa2f3b6f6e28b30d5b22c3aba39efc1c87b1c23e1d048"} Mar 13 16:55:55 crc kubenswrapper[4907]: I0313 16:55:55.370147 4907 generic.go:334] "Generic (PLEG): container finished" podID="4fbc8056-d73e-49f8-b94b-e1ca9827bda8" containerID="92c11c843af8e1cea03f882c89789b4e329f162f9479f6575e5fd03a8edc069c" exitCode=0 Mar 13 16:55:55 crc kubenswrapper[4907]: I0313 16:55:55.370289 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mhrkz" event={"ID":"4fbc8056-d73e-49f8-b94b-e1ca9827bda8","Type":"ContainerDied","Data":"92c11c843af8e1cea03f882c89789b4e329f162f9479f6575e5fd03a8edc069c"} Mar 13 16:55:56 crc kubenswrapper[4907]: I0313 16:55:56.385372 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mhrkz" event={"ID":"4fbc8056-d73e-49f8-b94b-e1ca9827bda8","Type":"ContainerStarted","Data":"185525bfeae44f48e3a3f7a42686da038653faa0563e3ac49793af2b2bbe83e4"} Mar 13 16:55:56 crc kubenswrapper[4907]: I0313 16:55:56.420442 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-mhrkz" podStartSLOduration=3.8465810400000002 podStartE2EDuration="7.420425694s" podCreationTimestamp="2026-03-13 16:55:49 +0000 UTC" firstStartedPulling="2026-03-13 16:55:52.321209401 +0000 UTC m=+10251.220997120" lastFinishedPulling="2026-03-13 16:55:55.895054045 +0000 UTC m=+10254.794841774" observedRunningTime="2026-03-13 16:55:56.408110616 +0000 UTC m=+10255.307898305" watchObservedRunningTime="2026-03-13 16:55:56.420425694 +0000 UTC m=+10255.320213383" Mar 13 16:55:58 crc kubenswrapper[4907]: I0313 16:55:58.783586 4907 scope.go:117] "RemoveContainer" containerID="64bd48e5096c07c25a07873a406a3b7bc89c38906dc6f7c7ace8213663361dfc" Mar 13 16:55:59 crc kubenswrapper[4907]: I0313 16:55:59.460357 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerStarted","Data":"20cccbc7ea5db3596b53a64b518d501aff800021f9e1a58804ae70d6d5732105"} Mar 13 16:56:00 crc kubenswrapper[4907]: I0313 16:56:00.164598 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29557016-8z7s6"] Mar 13 16:56:00 crc kubenswrapper[4907]: I0313 16:56:00.167578 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29557016-8z7s6" Mar 13 16:56:00 crc kubenswrapper[4907]: I0313 16:56:00.170319 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 16:56:00 crc kubenswrapper[4907]: I0313 16:56:00.170488 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 16:56:00 crc kubenswrapper[4907]: I0313 16:56:00.170612 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 16:56:00 crc kubenswrapper[4907]: I0313 16:56:00.189604 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29557016-8z7s6"] Mar 13 16:56:00 crc kubenswrapper[4907]: I0313 16:56:00.246897 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-mhrkz" Mar 13 16:56:00 crc kubenswrapper[4907]: I0313 16:56:00.246940 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-mhrkz" Mar 13 16:56:00 crc kubenswrapper[4907]: I0313 16:56:00.292849 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-mhrkz" Mar 13 16:56:00 crc kubenswrapper[4907]: I0313 16:56:00.340470 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s9khm\" (UniqueName: \"kubernetes.io/projected/ec34d61e-8aef-487f-8637-abcb9318df09-kube-api-access-s9khm\") pod \"auto-csr-approver-29557016-8z7s6\" (UID: \"ec34d61e-8aef-487f-8637-abcb9318df09\") " pod="openshift-infra/auto-csr-approver-29557016-8z7s6" Mar 13 16:56:00 crc kubenswrapper[4907]: I0313 16:56:00.442739 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s9khm\" (UniqueName: \"kubernetes.io/projected/ec34d61e-8aef-487f-8637-abcb9318df09-kube-api-access-s9khm\") pod \"auto-csr-approver-29557016-8z7s6\" (UID: \"ec34d61e-8aef-487f-8637-abcb9318df09\") " pod="openshift-infra/auto-csr-approver-29557016-8z7s6" Mar 13 16:56:00 crc kubenswrapper[4907]: I0313 16:56:00.465515 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s9khm\" (UniqueName: \"kubernetes.io/projected/ec34d61e-8aef-487f-8637-abcb9318df09-kube-api-access-s9khm\") pod \"auto-csr-approver-29557016-8z7s6\" (UID: \"ec34d61e-8aef-487f-8637-abcb9318df09\") " pod="openshift-infra/auto-csr-approver-29557016-8z7s6" Mar 13 16:56:00 crc kubenswrapper[4907]: I0313 16:56:00.472738 4907 generic.go:334] "Generic (PLEG): container finished" podID="40c3d47f-e20b-47a7-be46-a7d30d573327" containerID="f8ca9d7c44d6fd79031fa2f3b6f6e28b30d5b22c3aba39efc1c87b1c23e1d048" exitCode=0 Mar 13 16:56:00 crc kubenswrapper[4907]: I0313 16:56:00.472794 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4zr5z" event={"ID":"40c3d47f-e20b-47a7-be46-a7d30d573327","Type":"ContainerDied","Data":"f8ca9d7c44d6fd79031fa2f3b6f6e28b30d5b22c3aba39efc1c87b1c23e1d048"} Mar 13 16:56:00 crc kubenswrapper[4907]: I0313 16:56:00.491771 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29557016-8z7s6" Mar 13 16:56:01 crc kubenswrapper[4907]: I0313 16:56:01.057244 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29557016-8z7s6"] Mar 13 16:56:01 crc kubenswrapper[4907]: W0313 16:56:01.065549 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podec34d61e_8aef_487f_8637_abcb9318df09.slice/crio-4410f14df484cee29b534bda3f5e72083f5f168595f44341713ad5a9217e4202 WatchSource:0}: Error finding container 4410f14df484cee29b534bda3f5e72083f5f168595f44341713ad5a9217e4202: Status 404 returned error can't find the container with id 4410f14df484cee29b534bda3f5e72083f5f168595f44341713ad5a9217e4202 Mar 13 16:56:01 crc kubenswrapper[4907]: I0313 16:56:01.486992 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29557016-8z7s6" event={"ID":"ec34d61e-8aef-487f-8637-abcb9318df09","Type":"ContainerStarted","Data":"4410f14df484cee29b534bda3f5e72083f5f168595f44341713ad5a9217e4202"} Mar 13 16:56:01 crc kubenswrapper[4907]: I0313 16:56:01.493276 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4zr5z" event={"ID":"40c3d47f-e20b-47a7-be46-a7d30d573327","Type":"ContainerStarted","Data":"5de573655f92b0731cbc8517180a7905cbecd7ac96834604ea0757817741ba48"} Mar 13 16:56:01 crc kubenswrapper[4907]: I0313 16:56:01.525894 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-4zr5z" podStartSLOduration=2.939341145 podStartE2EDuration="9.525855541s" podCreationTimestamp="2026-03-13 16:55:52 +0000 UTC" firstStartedPulling="2026-03-13 16:55:54.351146953 +0000 UTC m=+10253.250934642" lastFinishedPulling="2026-03-13 16:56:00.937661349 +0000 UTC m=+10259.837449038" observedRunningTime="2026-03-13 16:56:01.512102783 +0000 UTC m=+10260.411890472" watchObservedRunningTime="2026-03-13 16:56:01.525855541 +0000 UTC m=+10260.425643230" Mar 13 16:56:02 crc kubenswrapper[4907]: I0313 16:56:02.503505 4907 generic.go:334] "Generic (PLEG): container finished" podID="ec34d61e-8aef-487f-8637-abcb9318df09" containerID="1b26563aa3367027de5284bc99669161edda50757e6ecf3c16064bafe5d673ec" exitCode=0 Mar 13 16:56:02 crc kubenswrapper[4907]: I0313 16:56:02.503666 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29557016-8z7s6" event={"ID":"ec34d61e-8aef-487f-8637-abcb9318df09","Type":"ContainerDied","Data":"1b26563aa3367027de5284bc99669161edda50757e6ecf3c16064bafe5d673ec"} Mar 13 16:56:03 crc kubenswrapper[4907]: I0313 16:56:03.323592 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-4zr5z" Mar 13 16:56:03 crc kubenswrapper[4907]: I0313 16:56:03.325039 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-4zr5z" Mar 13 16:56:03 crc kubenswrapper[4907]: I0313 16:56:03.947417 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29557016-8z7s6" Mar 13 16:56:04 crc kubenswrapper[4907]: I0313 16:56:04.124765 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s9khm\" (UniqueName: \"kubernetes.io/projected/ec34d61e-8aef-487f-8637-abcb9318df09-kube-api-access-s9khm\") pod \"ec34d61e-8aef-487f-8637-abcb9318df09\" (UID: \"ec34d61e-8aef-487f-8637-abcb9318df09\") " Mar 13 16:56:04 crc kubenswrapper[4907]: I0313 16:56:04.135664 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ec34d61e-8aef-487f-8637-abcb9318df09-kube-api-access-s9khm" (OuterVolumeSpecName: "kube-api-access-s9khm") pod "ec34d61e-8aef-487f-8637-abcb9318df09" (UID: "ec34d61e-8aef-487f-8637-abcb9318df09"). InnerVolumeSpecName "kube-api-access-s9khm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:56:04 crc kubenswrapper[4907]: I0313 16:56:04.227381 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s9khm\" (UniqueName: \"kubernetes.io/projected/ec34d61e-8aef-487f-8637-abcb9318df09-kube-api-access-s9khm\") on node \"crc\" DevicePath \"\"" Mar 13 16:56:04 crc kubenswrapper[4907]: I0313 16:56:04.539439 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29557016-8z7s6" event={"ID":"ec34d61e-8aef-487f-8637-abcb9318df09","Type":"ContainerDied","Data":"4410f14df484cee29b534bda3f5e72083f5f168595f44341713ad5a9217e4202"} Mar 13 16:56:04 crc kubenswrapper[4907]: I0313 16:56:04.539486 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4410f14df484cee29b534bda3f5e72083f5f168595f44341713ad5a9217e4202" Mar 13 16:56:04 crc kubenswrapper[4907]: I0313 16:56:04.539553 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29557016-8z7s6" Mar 13 16:56:04 crc kubenswrapper[4907]: I0313 16:56:04.796818 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-4zr5z" podUID="40c3d47f-e20b-47a7-be46-a7d30d573327" containerName="registry-server" probeResult="failure" output=< Mar 13 16:56:04 crc kubenswrapper[4907]: timeout: failed to connect service ":50051" within 1s Mar 13 16:56:04 crc kubenswrapper[4907]: > Mar 13 16:56:05 crc kubenswrapper[4907]: I0313 16:56:05.040303 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29557010-djwrx"] Mar 13 16:56:05 crc kubenswrapper[4907]: I0313 16:56:05.050517 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29557010-djwrx"] Mar 13 16:56:05 crc kubenswrapper[4907]: I0313 16:56:05.808717 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6f57b2d7-1aae-4af4-a4fc-1820c229bbed" path="/var/lib/kubelet/pods/6f57b2d7-1aae-4af4-a4fc-1820c229bbed/volumes" Mar 13 16:56:10 crc kubenswrapper[4907]: I0313 16:56:10.290705 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-mhrkz" Mar 13 16:56:11 crc kubenswrapper[4907]: I0313 16:56:11.858450 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-mhrkz"] Mar 13 16:56:11 crc kubenswrapper[4907]: I0313 16:56:11.860418 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-mhrkz" podUID="4fbc8056-d73e-49f8-b94b-e1ca9827bda8" containerName="registry-server" containerID="cri-o://185525bfeae44f48e3a3f7a42686da038653faa0563e3ac49793af2b2bbe83e4" gracePeriod=2 Mar 13 16:56:12 crc kubenswrapper[4907]: I0313 16:56:12.394939 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mhrkz" Mar 13 16:56:12 crc kubenswrapper[4907]: I0313 16:56:12.537253 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jr9pw\" (UniqueName: \"kubernetes.io/projected/4fbc8056-d73e-49f8-b94b-e1ca9827bda8-kube-api-access-jr9pw\") pod \"4fbc8056-d73e-49f8-b94b-e1ca9827bda8\" (UID: \"4fbc8056-d73e-49f8-b94b-e1ca9827bda8\") " Mar 13 16:56:12 crc kubenswrapper[4907]: I0313 16:56:12.537386 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4fbc8056-d73e-49f8-b94b-e1ca9827bda8-utilities\") pod \"4fbc8056-d73e-49f8-b94b-e1ca9827bda8\" (UID: \"4fbc8056-d73e-49f8-b94b-e1ca9827bda8\") " Mar 13 16:56:12 crc kubenswrapper[4907]: I0313 16:56:12.537501 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4fbc8056-d73e-49f8-b94b-e1ca9827bda8-catalog-content\") pod \"4fbc8056-d73e-49f8-b94b-e1ca9827bda8\" (UID: \"4fbc8056-d73e-49f8-b94b-e1ca9827bda8\") " Mar 13 16:56:12 crc kubenswrapper[4907]: I0313 16:56:12.538521 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4fbc8056-d73e-49f8-b94b-e1ca9827bda8-utilities" (OuterVolumeSpecName: "utilities") pod "4fbc8056-d73e-49f8-b94b-e1ca9827bda8" (UID: "4fbc8056-d73e-49f8-b94b-e1ca9827bda8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 16:56:12 crc kubenswrapper[4907]: I0313 16:56:12.545612 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4fbc8056-d73e-49f8-b94b-e1ca9827bda8-kube-api-access-jr9pw" (OuterVolumeSpecName: "kube-api-access-jr9pw") pod "4fbc8056-d73e-49f8-b94b-e1ca9827bda8" (UID: "4fbc8056-d73e-49f8-b94b-e1ca9827bda8"). InnerVolumeSpecName "kube-api-access-jr9pw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:56:12 crc kubenswrapper[4907]: I0313 16:56:12.600938 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4fbc8056-d73e-49f8-b94b-e1ca9827bda8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4fbc8056-d73e-49f8-b94b-e1ca9827bda8" (UID: "4fbc8056-d73e-49f8-b94b-e1ca9827bda8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 16:56:12 crc kubenswrapper[4907]: I0313 16:56:12.630098 4907 generic.go:334] "Generic (PLEG): container finished" podID="4fbc8056-d73e-49f8-b94b-e1ca9827bda8" containerID="185525bfeae44f48e3a3f7a42686da038653faa0563e3ac49793af2b2bbe83e4" exitCode=0 Mar 13 16:56:12 crc kubenswrapper[4907]: I0313 16:56:12.630163 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mhrkz" Mar 13 16:56:12 crc kubenswrapper[4907]: I0313 16:56:12.630183 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mhrkz" event={"ID":"4fbc8056-d73e-49f8-b94b-e1ca9827bda8","Type":"ContainerDied","Data":"185525bfeae44f48e3a3f7a42686da038653faa0563e3ac49793af2b2bbe83e4"} Mar 13 16:56:12 crc kubenswrapper[4907]: I0313 16:56:12.630402 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mhrkz" event={"ID":"4fbc8056-d73e-49f8-b94b-e1ca9827bda8","Type":"ContainerDied","Data":"a34268e545f27c48c28686dd73a60c951164de53a324312e016b297fe5122100"} Mar 13 16:56:12 crc kubenswrapper[4907]: I0313 16:56:12.630422 4907 scope.go:117] "RemoveContainer" containerID="185525bfeae44f48e3a3f7a42686da038653faa0563e3ac49793af2b2bbe83e4" Mar 13 16:56:12 crc kubenswrapper[4907]: I0313 16:56:12.640659 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jr9pw\" (UniqueName: \"kubernetes.io/projected/4fbc8056-d73e-49f8-b94b-e1ca9827bda8-kube-api-access-jr9pw\") on node \"crc\" DevicePath \"\"" Mar 13 16:56:12 crc kubenswrapper[4907]: I0313 16:56:12.640813 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4fbc8056-d73e-49f8-b94b-e1ca9827bda8-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 16:56:12 crc kubenswrapper[4907]: I0313 16:56:12.640832 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4fbc8056-d73e-49f8-b94b-e1ca9827bda8-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 16:56:12 crc kubenswrapper[4907]: I0313 16:56:12.657053 4907 scope.go:117] "RemoveContainer" containerID="92c11c843af8e1cea03f882c89789b4e329f162f9479f6575e5fd03a8edc069c" Mar 13 16:56:12 crc kubenswrapper[4907]: I0313 16:56:12.667452 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-mhrkz"] Mar 13 16:56:12 crc kubenswrapper[4907]: I0313 16:56:12.676773 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-mhrkz"] Mar 13 16:56:12 crc kubenswrapper[4907]: I0313 16:56:12.699260 4907 scope.go:117] "RemoveContainer" containerID="ae74208e85dfe58e9cd0fd4f63b388596e592e48c797f92d20005d4e46b286f8" Mar 13 16:56:12 crc kubenswrapper[4907]: I0313 16:56:12.726053 4907 scope.go:117] "RemoveContainer" containerID="185525bfeae44f48e3a3f7a42686da038653faa0563e3ac49793af2b2bbe83e4" Mar 13 16:56:12 crc kubenswrapper[4907]: E0313 16:56:12.726598 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"185525bfeae44f48e3a3f7a42686da038653faa0563e3ac49793af2b2bbe83e4\": container with ID starting with 185525bfeae44f48e3a3f7a42686da038653faa0563e3ac49793af2b2bbe83e4 not found: ID does not exist" containerID="185525bfeae44f48e3a3f7a42686da038653faa0563e3ac49793af2b2bbe83e4" Mar 13 16:56:12 crc kubenswrapper[4907]: I0313 16:56:12.726637 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"185525bfeae44f48e3a3f7a42686da038653faa0563e3ac49793af2b2bbe83e4"} err="failed to get container status \"185525bfeae44f48e3a3f7a42686da038653faa0563e3ac49793af2b2bbe83e4\": rpc error: code = NotFound desc = could not find container \"185525bfeae44f48e3a3f7a42686da038653faa0563e3ac49793af2b2bbe83e4\": container with ID starting with 185525bfeae44f48e3a3f7a42686da038653faa0563e3ac49793af2b2bbe83e4 not found: ID does not exist" Mar 13 16:56:12 crc kubenswrapper[4907]: I0313 16:56:12.726662 4907 scope.go:117] "RemoveContainer" containerID="92c11c843af8e1cea03f882c89789b4e329f162f9479f6575e5fd03a8edc069c" Mar 13 16:56:12 crc kubenswrapper[4907]: E0313 16:56:12.727044 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"92c11c843af8e1cea03f882c89789b4e329f162f9479f6575e5fd03a8edc069c\": container with ID starting with 92c11c843af8e1cea03f882c89789b4e329f162f9479f6575e5fd03a8edc069c not found: ID does not exist" containerID="92c11c843af8e1cea03f882c89789b4e329f162f9479f6575e5fd03a8edc069c" Mar 13 16:56:12 crc kubenswrapper[4907]: I0313 16:56:12.727083 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"92c11c843af8e1cea03f882c89789b4e329f162f9479f6575e5fd03a8edc069c"} err="failed to get container status \"92c11c843af8e1cea03f882c89789b4e329f162f9479f6575e5fd03a8edc069c\": rpc error: code = NotFound desc = could not find container \"92c11c843af8e1cea03f882c89789b4e329f162f9479f6575e5fd03a8edc069c\": container with ID starting with 92c11c843af8e1cea03f882c89789b4e329f162f9479f6575e5fd03a8edc069c not found: ID does not exist" Mar 13 16:56:12 crc kubenswrapper[4907]: I0313 16:56:12.727118 4907 scope.go:117] "RemoveContainer" containerID="ae74208e85dfe58e9cd0fd4f63b388596e592e48c797f92d20005d4e46b286f8" Mar 13 16:56:12 crc kubenswrapper[4907]: E0313 16:56:12.727413 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ae74208e85dfe58e9cd0fd4f63b388596e592e48c797f92d20005d4e46b286f8\": container with ID starting with ae74208e85dfe58e9cd0fd4f63b388596e592e48c797f92d20005d4e46b286f8 not found: ID does not exist" containerID="ae74208e85dfe58e9cd0fd4f63b388596e592e48c797f92d20005d4e46b286f8" Mar 13 16:56:12 crc kubenswrapper[4907]: I0313 16:56:12.727447 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ae74208e85dfe58e9cd0fd4f63b388596e592e48c797f92d20005d4e46b286f8"} err="failed to get container status \"ae74208e85dfe58e9cd0fd4f63b388596e592e48c797f92d20005d4e46b286f8\": rpc error: code = NotFound desc = could not find container \"ae74208e85dfe58e9cd0fd4f63b388596e592e48c797f92d20005d4e46b286f8\": container with ID starting with ae74208e85dfe58e9cd0fd4f63b388596e592e48c797f92d20005d4e46b286f8 not found: ID does not exist" Mar 13 16:56:13 crc kubenswrapper[4907]: I0313 16:56:13.799746 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4fbc8056-d73e-49f8-b94b-e1ca9827bda8" path="/var/lib/kubelet/pods/4fbc8056-d73e-49f8-b94b-e1ca9827bda8/volumes" Mar 13 16:56:14 crc kubenswrapper[4907]: I0313 16:56:14.373560 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-4zr5z" podUID="40c3d47f-e20b-47a7-be46-a7d30d573327" containerName="registry-server" probeResult="failure" output=< Mar 13 16:56:14 crc kubenswrapper[4907]: timeout: failed to connect service ":50051" within 1s Mar 13 16:56:14 crc kubenswrapper[4907]: > Mar 13 16:56:23 crc kubenswrapper[4907]: I0313 16:56:23.434561 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-4zr5z" Mar 13 16:56:23 crc kubenswrapper[4907]: I0313 16:56:23.520597 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-4zr5z" Mar 13 16:56:23 crc kubenswrapper[4907]: I0313 16:56:23.862587 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-4zr5z"] Mar 13 16:56:24 crc kubenswrapper[4907]: I0313 16:56:24.791989 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-4zr5z" podUID="40c3d47f-e20b-47a7-be46-a7d30d573327" containerName="registry-server" containerID="cri-o://5de573655f92b0731cbc8517180a7905cbecd7ac96834604ea0757817741ba48" gracePeriod=2 Mar 13 16:56:25 crc kubenswrapper[4907]: I0313 16:56:25.326037 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4zr5z" Mar 13 16:56:25 crc kubenswrapper[4907]: I0313 16:56:25.477988 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/40c3d47f-e20b-47a7-be46-a7d30d573327-utilities\") pod \"40c3d47f-e20b-47a7-be46-a7d30d573327\" (UID: \"40c3d47f-e20b-47a7-be46-a7d30d573327\") " Mar 13 16:56:25 crc kubenswrapper[4907]: I0313 16:56:25.478607 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gg57d\" (UniqueName: \"kubernetes.io/projected/40c3d47f-e20b-47a7-be46-a7d30d573327-kube-api-access-gg57d\") pod \"40c3d47f-e20b-47a7-be46-a7d30d573327\" (UID: \"40c3d47f-e20b-47a7-be46-a7d30d573327\") " Mar 13 16:56:25 crc kubenswrapper[4907]: I0313 16:56:25.478660 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/40c3d47f-e20b-47a7-be46-a7d30d573327-catalog-content\") pod \"40c3d47f-e20b-47a7-be46-a7d30d573327\" (UID: \"40c3d47f-e20b-47a7-be46-a7d30d573327\") " Mar 13 16:56:25 crc kubenswrapper[4907]: I0313 16:56:25.480197 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/40c3d47f-e20b-47a7-be46-a7d30d573327-utilities" (OuterVolumeSpecName: "utilities") pod "40c3d47f-e20b-47a7-be46-a7d30d573327" (UID: "40c3d47f-e20b-47a7-be46-a7d30d573327"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 16:56:25 crc kubenswrapper[4907]: I0313 16:56:25.493970 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/40c3d47f-e20b-47a7-be46-a7d30d573327-kube-api-access-gg57d" (OuterVolumeSpecName: "kube-api-access-gg57d") pod "40c3d47f-e20b-47a7-be46-a7d30d573327" (UID: "40c3d47f-e20b-47a7-be46-a7d30d573327"). InnerVolumeSpecName "kube-api-access-gg57d". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:56:25 crc kubenswrapper[4907]: I0313 16:56:25.580330 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gg57d\" (UniqueName: \"kubernetes.io/projected/40c3d47f-e20b-47a7-be46-a7d30d573327-kube-api-access-gg57d\") on node \"crc\" DevicePath \"\"" Mar 13 16:56:25 crc kubenswrapper[4907]: I0313 16:56:25.580530 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/40c3d47f-e20b-47a7-be46-a7d30d573327-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 16:56:25 crc kubenswrapper[4907]: I0313 16:56:25.609430 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/40c3d47f-e20b-47a7-be46-a7d30d573327-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "40c3d47f-e20b-47a7-be46-a7d30d573327" (UID: "40c3d47f-e20b-47a7-be46-a7d30d573327"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 16:56:25 crc kubenswrapper[4907]: I0313 16:56:25.681803 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/40c3d47f-e20b-47a7-be46-a7d30d573327-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 16:56:25 crc kubenswrapper[4907]: I0313 16:56:25.811046 4907 generic.go:334] "Generic (PLEG): container finished" podID="40c3d47f-e20b-47a7-be46-a7d30d573327" containerID="5de573655f92b0731cbc8517180a7905cbecd7ac96834604ea0757817741ba48" exitCode=0 Mar 13 16:56:25 crc kubenswrapper[4907]: I0313 16:56:25.811101 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4zr5z" event={"ID":"40c3d47f-e20b-47a7-be46-a7d30d573327","Type":"ContainerDied","Data":"5de573655f92b0731cbc8517180a7905cbecd7ac96834604ea0757817741ba48"} Mar 13 16:56:25 crc kubenswrapper[4907]: I0313 16:56:25.811135 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4zr5z" event={"ID":"40c3d47f-e20b-47a7-be46-a7d30d573327","Type":"ContainerDied","Data":"102dad920f7c3e14da000819f2f576fdcf93e506c017ecabf4d72e0cdc9c0160"} Mar 13 16:56:25 crc kubenswrapper[4907]: I0313 16:56:25.811158 4907 scope.go:117] "RemoveContainer" containerID="5de573655f92b0731cbc8517180a7905cbecd7ac96834604ea0757817741ba48" Mar 13 16:56:25 crc kubenswrapper[4907]: I0313 16:56:25.811191 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4zr5z" Mar 13 16:56:25 crc kubenswrapper[4907]: I0313 16:56:25.845134 4907 scope.go:117] "RemoveContainer" containerID="f8ca9d7c44d6fd79031fa2f3b6f6e28b30d5b22c3aba39efc1c87b1c23e1d048" Mar 13 16:56:25 crc kubenswrapper[4907]: I0313 16:56:25.873199 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-4zr5z"] Mar 13 16:56:25 crc kubenswrapper[4907]: I0313 16:56:25.915183 4907 scope.go:117] "RemoveContainer" containerID="48b68600daee932b5c08e4d390a35167c48346ec15600f7cf99e2ed2d675ff09" Mar 13 16:56:25 crc kubenswrapper[4907]: I0313 16:56:25.917856 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-4zr5z"] Mar 13 16:56:25 crc kubenswrapper[4907]: I0313 16:56:25.953929 4907 scope.go:117] "RemoveContainer" containerID="5de573655f92b0731cbc8517180a7905cbecd7ac96834604ea0757817741ba48" Mar 13 16:56:25 crc kubenswrapper[4907]: E0313 16:56:25.954236 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5de573655f92b0731cbc8517180a7905cbecd7ac96834604ea0757817741ba48\": container with ID starting with 5de573655f92b0731cbc8517180a7905cbecd7ac96834604ea0757817741ba48 not found: ID does not exist" containerID="5de573655f92b0731cbc8517180a7905cbecd7ac96834604ea0757817741ba48" Mar 13 16:56:25 crc kubenswrapper[4907]: I0313 16:56:25.954266 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5de573655f92b0731cbc8517180a7905cbecd7ac96834604ea0757817741ba48"} err="failed to get container status \"5de573655f92b0731cbc8517180a7905cbecd7ac96834604ea0757817741ba48\": rpc error: code = NotFound desc = could not find container \"5de573655f92b0731cbc8517180a7905cbecd7ac96834604ea0757817741ba48\": container with ID starting with 5de573655f92b0731cbc8517180a7905cbecd7ac96834604ea0757817741ba48 not found: ID does not exist" Mar 13 16:56:25 crc kubenswrapper[4907]: I0313 16:56:25.954288 4907 scope.go:117] "RemoveContainer" containerID="f8ca9d7c44d6fd79031fa2f3b6f6e28b30d5b22c3aba39efc1c87b1c23e1d048" Mar 13 16:56:25 crc kubenswrapper[4907]: E0313 16:56:25.954586 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f8ca9d7c44d6fd79031fa2f3b6f6e28b30d5b22c3aba39efc1c87b1c23e1d048\": container with ID starting with f8ca9d7c44d6fd79031fa2f3b6f6e28b30d5b22c3aba39efc1c87b1c23e1d048 not found: ID does not exist" containerID="f8ca9d7c44d6fd79031fa2f3b6f6e28b30d5b22c3aba39efc1c87b1c23e1d048" Mar 13 16:56:25 crc kubenswrapper[4907]: I0313 16:56:25.954637 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f8ca9d7c44d6fd79031fa2f3b6f6e28b30d5b22c3aba39efc1c87b1c23e1d048"} err="failed to get container status \"f8ca9d7c44d6fd79031fa2f3b6f6e28b30d5b22c3aba39efc1c87b1c23e1d048\": rpc error: code = NotFound desc = could not find container \"f8ca9d7c44d6fd79031fa2f3b6f6e28b30d5b22c3aba39efc1c87b1c23e1d048\": container with ID starting with f8ca9d7c44d6fd79031fa2f3b6f6e28b30d5b22c3aba39efc1c87b1c23e1d048 not found: ID does not exist" Mar 13 16:56:25 crc kubenswrapper[4907]: I0313 16:56:25.954669 4907 scope.go:117] "RemoveContainer" containerID="48b68600daee932b5c08e4d390a35167c48346ec15600f7cf99e2ed2d675ff09" Mar 13 16:56:25 crc kubenswrapper[4907]: E0313 16:56:25.954941 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"48b68600daee932b5c08e4d390a35167c48346ec15600f7cf99e2ed2d675ff09\": container with ID starting with 48b68600daee932b5c08e4d390a35167c48346ec15600f7cf99e2ed2d675ff09 not found: ID does not exist" containerID="48b68600daee932b5c08e4d390a35167c48346ec15600f7cf99e2ed2d675ff09" Mar 13 16:56:25 crc kubenswrapper[4907]: I0313 16:56:25.954965 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"48b68600daee932b5c08e4d390a35167c48346ec15600f7cf99e2ed2d675ff09"} err="failed to get container status \"48b68600daee932b5c08e4d390a35167c48346ec15600f7cf99e2ed2d675ff09\": rpc error: code = NotFound desc = could not find container \"48b68600daee932b5c08e4d390a35167c48346ec15600f7cf99e2ed2d675ff09\": container with ID starting with 48b68600daee932b5c08e4d390a35167c48346ec15600f7cf99e2ed2d675ff09 not found: ID does not exist" Mar 13 16:56:27 crc kubenswrapper[4907]: I0313 16:56:27.798997 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="40c3d47f-e20b-47a7-be46-a7d30d573327" path="/var/lib/kubelet/pods/40c3d47f-e20b-47a7-be46-a7d30d573327/volumes" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515155040721024445 0ustar coreroot  Om77'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015155040722017363 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015155014264016510 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015155014264015460 5ustar corecore